Back to blog

Thought Leadership

Thought Leadership·March 4, 2026·9 min read

Case study: when AI produces industrial mediocrity

A major IT services company invests millions to train a model on its past proposals. Result: a machine that produces average responses, relevant to no one. Autopsy of a predictable failure.

By Aléaume Muller

Case study: when AI produces industrial mediocrity

This article extends Tenders and AI: toward a silent market transformation, where we established the diagnosis: most current approaches automate the wrong thing. Here, we get concrete. A real case. A real failure. And what it teaches us.

The project

A sizeable IT services company -- several thousand consultants, hundreds of tender responses per year. The starting premise is sensible: we spend too much time drafting, we have a rich history of winning proposals, why not train a model to capitalize on it?

The project is ambitious. Budget of several million euros. Offshore development. A dedicated team for over a year. The objective: an internal tool capable of generating technical responses from the archive of past proposals.

Several million euros and over a year of development to build a statistically optimized copy-paste machine.

On paper, it makes sense. In practice, it is a disaster foretold.

The fundamental error: no ontology

The model was trained on hundreds of past responses. Technical sections, methodology paragraphs, project references, service commitments. In bulk.

The problem: no semantic structure was built upstream.

The model cannot distinguish:

  • Agile vs. Waterfall -- two radically different delivery philosophies, blended in the same training corpus
  • Kubernetes vs. monolithic architecture -- opposing infrastructure paradigms, treated as interchangeable text blocks
  • A specification for a government ministry vs. a specification for a startup -- business contexts, requirement levels, and decision cultures that have nothing in common
  • A winning proposal vs. a losing proposal -- no weighting. The model learns as much from failures as from successes, without knowing which is which

Result: the model learned the style of the firm, not its intelligence. It can produce sentences that look like tender responses. It does not know why those sentences exist.

The result: the mediocrity machine

After over a year of development and millions invested, the tool is delivered. The first tests are telling.

Take a concrete case. A specification for the overhaul of an HR management system for a local government authority. The requirements document heavily emphasizes:

  • Data migration from an aging legacy system
  • Accessibility compliance (RGAA -- French accessibility standard)
  • A tight schedule with a go-live deadline before the elections

Here is what the model produces (anonymized reconstruction):

Model-generated response:

"Our methodological approach builds on a proven framework combining Agile best practices and Waterfall principles to ensure quality delivery. Our multidisciplinary team will mobilize technical and functional experts to ensure project success. We propose a modern, scalable, and secure architecture based on market standards."

Everything is true. Nothing is useful.

  • "Combining Agile and Waterfall" -- because the corpus contained both, and the model learned not to offend anyone
  • "Multidisciplinary team" -- a phrase present in 90% of the training proposals
  • "Modern, scalable, and secure architecture" -- the hollow trinity, applicable to any project
  • No mention of data migration (the real issue)
  • No mention of RGAA accessibility compliance (a regulatory requirement)
  • No mention of the constrained schedule (the primary risk)

The model produced the statistical average of all past responses. A median response. Technically correct. Strategically empty.

Now, compare with what a competent bid manager would have written:

Contextualized response:

"The primary risk of this project is not technical -- it is data migration from [legacy system], whose documentation is fragmentary and formats are non-standardized. We propose a dedicated 4-week sprint 0 exclusively focused on auditing and mapping existing data, before any development begins. This approach secures the go-live schedule you have set for [date], by identifying real blockers before they become delays.

On RGAA accessibility compliance, we integrate a conformity audit into every sprint, not during final acceptance testing. [Analogous project reference with measurable outcome]."

The difference is not in writing quality. It is in understanding the problem.

The first response says "we are good." The second says "we understood your pain, here is how we address it." One vaguely reassures. The other convinces.

Generic approach (model)Contextualized approach (bid manager)
"Proven framework combining Agile and Waterfall"Dedicated 4-week sprint 0 for data audit
No mention of data migrationIdentifies the primary risk in the first sentence
"Modern, scalable, and secure architecture"RGAA compliance integrated into every sprint
Applicable to any projectBuilt for this client, this contract

The diagnosis

Why doesn't it work? Because the approach rests on a false hypothesis: the past contains the answers to the future.

Training a model on your past proposals teaches it to reproduce what you have already done. Not to understand what the client is asking for. It is statistically optimized copy-paste.

The structural problems:

  • No semantic segmentation. The model does not know that "Agile" in a banking context and "Agile" in an industrial context are two different realities. It learned the word, not the concept.
  • No link between context and response. The model generates text blocks, but it does not know why a given block was relevant for a given contract. It does not read the specifications -- it skims for trigger keywords.
  • No notion of quality. Winning proposal, losing proposal, mediocre proposal -- everything carries equal weight in the corpus. The model has no signal for what works.
  • No adaptation to the client. The model produces generic "IT services speak." It does not know who is on the other side, what matters to them, or what differentiates them from other buyers.

In summary: training on the past without semantic structure means learning to copy-paste intelligently. No understanding. No differentiation.

Why this is a common trap

This company is not an isolated case. The reasoning is tempting:

  1. We have 10 years of proposals on our servers
  2. AI is good with textual data
  3. Therefore: we train a model on our proposals = competitive advantage

It is logical. It is wrong. Here is why:

  • Volume is not value. 500 past responses of which 200 lost, 150 were mediocre, and 150 were good -- but you don't know which is which. The model learns a blend.
  • Context is king. A brilliant response for a 2022 contract is potentially obsolete for a 2026 contract. Technologies change. Regulations evolve. Client expectations shift.
  • Competitive advantage is not in the words. It is in the ability to understand what the client truly wants, beyond what they wrote. No model trained on your past proposals can do that.

The alternative: reasoning with the future, not the past

The right paradigm is not "what have we already written that is similar?" It is: "what does this specific client want, for themselves, and how can we deliver even more than what they are asking for?"

This demands a radically different architecture:

1. Interrogate every atomic piece of information in the specifications. Not skim. Not summarize. Interrogate. Every requirement, every constraint, every formulation is a signal. "The client wrote 'imperatively' here and 'if possible' there -- why?" "There are 15 pages on infrastructure and 2 on application -- what does that say about their priorities?" "They mention data migration three times -- that is where it hurts."

2. Do not reason from the past. History can inform (client references, lessons learned). It must not dictate. Each contract is a new problem. The client does not want to know what you did for others -- they want to know what you will do for them.

3. Reason from the future. What does this client want in 3 years? What is the problem behind the problem? What did they not dare to write in the specifications because they do not know it is possible? This is where contracts are won. Not in the ability to recycle paragraphs, but in the ability to anticipate.

4. Build an ontology, not a corpus. The difference between a text database and intelligence is semantic structure. Knowing that "Agile" in this context means "Scrum with 2-week sprints and a team of 5" -- not "we are flexible." Knowing that "cloud-native architecture" for this client means "we are migrating from on-premise and we are afraid" -- not "Kubernetes everywhere."

What to remember

The difference between "automating tenders" and "understanding tenders" does not lie in the volume of training data. It lies in two things:

  • Ontology: the ability to give meaning to information, not just store it
  • Client-oriented reasoning: the ability to start from the client's real need, not from what you already have in stock

The company in this case study invested millions to build a tool that does exactly what a tired bid manager does on a Friday evening: copy-paste from the last proposal that vaguely resembles the new one. Faster, certainly. But not better.

AI in tenders is not a data problem. It is a cognition problem.

Key takeaway: Training a model on your past proposals without ontology means learning to copy-paste intelligently. The difference between automating and understanding does not lie in data volume -- it lies in the ability to give meaning.


At TenderGraph, we do not recycle your past proposals. We build agents capable of reading specifications the way a senior bid manager would -- seeking the unspoken, interrogating every requirement, reasoning for the client in front of them. Not for the average of your past clients.


Further reading:

Tags

#tenders#AI#anti-pattern#transformation#ontology

Next step

Ready to transform your tender response?

Keep reading

Recommended articles