Your bid reviews are useless — and AI is about to prove it
This article extends The acceleration of pre-sales cycles, where we asked what to do with the time AI frees up, and The information revolution, where we applied Shannon's theory to bid management. Here, we go one level deeper: what happens inside the organization that produces the response.
The theater of reviews
Every structured organization that responds to tenders has a review process. Three tiers, typically. They are called bronze / silver / gold, or R0 / R1 / R2, or screening / mid-term / final review. The names vary. The ritual is the same.
Bronze review: the Go/No-Go decision. An 8-slide PowerPoint. The commercial director signs off in 4 minutes. On to the next bid.
Silver review: mid-course. Progress is presented. A new PowerPoint, 15 slides this time. It emerges that the solution architect has not started yet. The bid manager says it is "in progress." Noted. On to the next bid.
Gold review: pre-submission. The "finalized" technical proposal is presented. A third PowerPoint — 25 slides. The technical director reads the content for the first time. He raises concerns. Three of them are structurally significant. It is 5:30 PM the evening before the deadline.
The problem is not the number of reviews. It is that each review creates a new file, a new set of minutes, a new layer of documentation — with no link to the previous ones. And 90% of the information is redundant.
What actually goes wrong
Redundancy as a source of entropy
Shannon demonstrated it: redundancy in a communication channel is useful only if it serves to correct errors (Shannon, C.E., A Mathematical Theory of Communication, Bell System Technical Journal, 1948). When redundancy does not serve correction — when it stacks noise — it increases entropy. Information becomes harder to find, not easier.
This is precisely what happens in your reviews. The R0 minutes contain the bid context. The R1 minutes repeat it — with minor variations. The R2 minutes repeat it again — with different variations. Three versions of the context. None is exactly the same. Which one is authoritative?
| Tier | Files created | New information | Redundant information |
|---|---|---|---|
| R0 (bronze) | PowerPoint + minutes + validation email | Go/No-Go decision, initial win themes | 0% (first milestone) |
| R1 (silver) | PowerPoint + minutes + tracking sheet | Progress update, identified risks | ~70% (context copy-pasted) |
| R2 (gold) | PowerPoint + minutes + compliance checklist | Technical feedback, final corrections | ~85% (full redundancy) |
Three review tiers. Seven files minimum. Five calendar invitations. And the strategic information — the actual decisions, the accepted risks, the trade-offs — is documented nowhere explicitly.
Implicit decisions
The real poison is what goes unwritten.
"We decided not to respond to lot 3." Where is that documented? In an email sent to three people on a Tuesday at 10 PM. Not in the minutes. Not in the PowerPoint. Not in the tracking file.
"The client hinted that service continuity is their primary criterion." Who noted that? The account executive, in his head. Perhaps in an email. Certainly not in the structured bid file.
"We decided to staff a junior profile on the application lot to remain competitive on price." Who made that call? When? Based on what analysis? No one knows. But everyone lives with the consequences.
The most structurally significant decisions in a bid are those that are never formalized. They live in conversations, in emails, in the silences of meetings. And when they need to be retrieved — three weeks later, at the point of drafting the technical proposal — there is nothing.
Illegible encoding
Retrieving information within an active tender response is an exercise in archaeological excavation.
The file is named TP_v3_final_revised_AM_VF2.docx. It sits in a Teams subfolder. Or in SharePoint. Or on the bid manager's local drive. Three versions coexist. None is tagged. None is linked to the review minutes that prompted the revisions. None is connected to the specification requirements it is meant to address.
The solution architect who joins mid-process spends half a day understanding "where things stand." The technical director conducting the gold review has no way of knowing what decisions were made in R0 and R1. He reads everything from scratch. He redoes the work. He asks the same questions.
This is the real cost of information entropy: not the loss of information, but the time spent retrieving it.
What AI reveals — and what it does not solve
What AI does exceptionally well today
Give a properly architected AI system the entirety of a bid file — specifications, review minutes, correspondence, draft technical proposal — and it accomplishes in minutes what no human can do in a day:
- Detect redundancies: identify copy-pasted paragraphs across minutes, slides repeated from one review to the next, phrasings that diverge without the underlying decision having changed.
- Trace decisions: extract from all documents and exchanges the moments where a decision was made, and construct a chronological decision log that no one ever wrote.
- Cross-reference layers: connect a specification requirement to the technical proposal section that addresses it, to the risk identified in R1 that concerns it, and to the R0 decision that shaped the response strategy.
- Flag gaps: identify requirements without responses, risks without mitigation plans, decisions without traceable justification.
AI does not replace the review. It makes visible what the review was hiding: blind spots, phantom decisions, inconsistencies between what was decided and what was written.
What AI does not solve
AI can analyze any document. But it cannot analyze what was never written down.
If your account executive stored client insights in his head, AI will not find them. If your Go/No-Go decision was a verbal "yes" at the end of a meeting, AI has nothing to trace. If your solution architect made a technical trade-off in the margin of his notebook, AI will not see it.
The problem is not AI. The problem is what you feed it.
What you must change — before even discussing tools
Learn to encode for machines
The shift to the agentic AI era demands a fundamental change in how teams formalize information. This is not a question of tooling. It is a question of cognitive discipline.
Stop being vague. "The situation is complex and requires in-depth analysis" means nothing — neither to a human nor to an AI. "Lot 2 presents an understaffing risk: 3 FTEs planned for 12 applications, a ratio incompatible with the required SLA (99.5%)" is actionable information.
Link every piece of information to a stake. An isolated fact is noise. A fact connected to a requirement, a risk, or a decision is signal. "The client uses Oracle" is a fact. "The client uses Oracle → integration constraint with the existing HR module → specification requirement §4.3.2 → incompatibility risk with our PostgreSQL stack" is a decision chain that AI can exploit.
Be explicit about decisions. Every decision should be a structured statement: What was decided, Why, By whom, When, and What consequences for the bid. Not a vague "we acknowledged the fact that..." buried in a paragraph of meeting notes.
| Before (human encoding) | After (AI-ready encoding) |
|---|---|
| "We discussed lot 3" | "Decision: abandon lot 3 — insufficient ROI (estimated revenue €120K vs. response cost €45K). Approved by P. Martin, 2026-04-08" |
| "The technical situation is tight" | "Risk: project team under-capacity. 2 FTEs available, 4 required (specification §5.1). Impact: mobilization delay +3 weeks. Mitigation: external recruitment in progress" |
| "References are fine" | "3 references identified: Lyon Metropolis (TMA, 18 months, 8 apps), Département 69 (managed services, 24 months), CHU Bordeaux (migration, 6 months). Lot 1 relevance: strong (2/3 match the scope)" |
Store differently
The problem with PowerPoint files is not their format. It is that they are closed containers. Information enters a .pptx and never comes out — unless a human manually copies it into another file.
A bid's strategic information should live in a structured system, not in scattered files. A knowledge graph, a fact base, a decision log — the technical form matters less than the principle. What matters is that information is:
- Atomic: one fact = one entry. Not a paragraph blending three pieces of information.
- Connected: every fact is linked to its sources (specification requirement, review minutes, email) and to its consequences (proposal section, risk, decision).
- Versioned: you know when a fact was created, modified, or invalidated. Not "v3_final_revised."
- Queryable: you can ask "what decisions made in R1 impact lot 2?" and get an answer in seconds.
Rethink the review cadence
AI compresses production time. What took 3 weeks now takes 3 days. But reviews are still locked to the old calendar: R0 at D+3, R1 at D+10, R2 at D+18.
When production is complete by D+5, what happens between D+5 and D+18? Waiting. Polishing details. Creating entropy — minor edits, rephrasing, cosmetic adjustments that add no value but generate new versions.
The new cadence should be:
- Structured R0 (Go/No-Go): a rigorous Go/No-Go with scoring (How to decide whether to respond). Not 4 minutes. 30 minutes. With data.
- Compressed production: 2-3 days with AI. Extraction, structuring, first draft.
- Single substantive review: one review, but dense. Red team. Challenge the strategy, not the formatting. The technical director reads the substance, not the slides.
- Finalization sprint: 1-2 days to integrate structurally significant feedback.
Four stages instead of seven. A single information flow instead of three redundant layers.
The real transformation: beyond the tool
What TenderGraph aims for
TenderGraph is not a review tool. It is a cognitive system that enforces encoding discipline.
When RequirementMiner extracts 1,382 requirements from a 500-page specification, it does not produce yet another file. It creates a structured graph — every requirement typed, classified, prioritized, and interconnected. When TITAN orchestrates the full response, every decision, every trade-off, every version will be traced in a chronological log exploitable by both humans and machines.
But TenderGraph cannot deliver if the organization continues to encode information as it did in 2015. If the account executive stores his insights in his head. If the solution architect works in a local file. If the technical director discovers the bid the evening before the deadline.
The pace of transformation as the critical friction point
The primary obstacle is not technological. It is organizational.
Teams know their reviews are theater. Bid managers know their minutes go unread. Solution architects know their trade-offs are not traced. But the system works "well enough" — enough contracts are won to avoid questioning the process.
Until the day a competitor who has made the transformation wins the contract you thought was yours. Not because they have a better tool. Because their team encodes information in an exploitable format, because their AI can link every requirement to a substantiated response, and because their single substantive review caught the strategic flaw that your three PowerPoint reviews missed.
The good news: this transformation does not require overturning everything. It starts with three simple changes — encode decisions explicitly, connect facts to stakes, and replace redundancy with traceability. The rest follows.
Where to start
- Formalize decisions: starting tomorrow, every review meeting produces not a narrative summary but a structured decision log (What / Why / By whom / Impact).
- Eliminate redundancy: one living document per bid, not three PowerPoints. AI can generate review slides from the living document — not the other way around.
- Test with AI: take your most recent bid. Feed every file to an LLM. Ask: "What decisions were made? What risks are uncovered? Which requirements have no response?" The gaps it reveals are the gaps your reviews did not catch.
What TenderGraph does
RequirementMiner transforms a specification into a structured graph of requirements — not into yet another file. TITAN (forthcoming) will orchestrate the complete bid with integrated traceability: every fact, every decision, every version, connected and queryable. The goal is not to replace your reviews — it is to make visible what they conceal, and to enforce an encoding discipline that transforms noise into exploitable signal.
Further reading
- The acceleration of pre-sales cycles: what comes next? — What to do with the time AI frees up
- The information revolution: why AI amplifies noise as much as it can eliminate it — Shannon applied to bid management
- Responding to a tender: what the guides will never tell you — Go/No-Go, strategy, organization
- Pre-sales is an exercise in command — Planning, friction, mission command
- How to write a technical proposal that wins contracts — Signal encoding, structure driven by scoring criteria
- Cognitive biases and tenders — Why humans repeat the same mistakes