Building your own AI model.

Door one. The expensive one. Often the wrong one.

What this actually involves

Building an AI model means training a neural network on raw data until its weights encode something useful. Not adapting an existing model. Not adding retrieval to a chat interface. The real thing: dataset engineering, training infrastructure, evaluation pipelines, and the research expertise to know when the loss curve is lying to you.

The result is a model that did not exist before. That is the point and that is the cost. You are not borrowing capability from the major labs; you are creating capability that your organisation now owns and now has to maintain.

Confidence, calibration, and the lie detector

One thing that gets glossed over in most AI conversations: models do not naturally know how confident they should be. A well-built model gives you a probability for every answer. A badly-built one gives you the same answer with the same tone whether it is right or guessing.

Calibration - making confidence numbers actually mean what they say - is one of the things separating research-grade work from demo-grade work. If a model says it is 90% confident, it should be right about 90% of the time. Most are not. Building your own model means you have to solve this yourself.

This is not a footnote. In regulated industries, an uncalibrated model is a liability. The model that says "I do not know" honestly is more useful than the one that always sounds sure.

The skill bar is real

Building a model from scratch is not a software engineering task. It is a research engineering task, which is a different discipline with a smaller talent pool and a longer feedback loop. You cannot fix a training run with a hotfix. You debug it for weeks and learn something for next time.

The teams that do this well share a few traits: they have read the literature, they keep up with it, they have a high-throughput evaluation harness, and they know when to stop. That last one matters. Most failed model-building programmes failed because nobody was willing to call it.

What it costs

The honest answer: more than the bid says, and more than the second one too. Compute is the visible cost; data engineering is the invisible one. A team capable of doing this work properly is between five and twenty people, and they need to stay together long enough to ship something that works. Two years is a reasonable lower bound for non-trivial models.

And that is just the build. Once a model exists, it has to be kept current. The world moves; your training data ages; the base capabilities of public models advance. A custom model that was state-of-the-art at launch can be middle-of-the-pack within eighteen months if nobody is feeding it.

When it is the right answer

Building your own model is the right answer for a small set of organisations:

  • You have a research thesis that genuinely cannot be expressed by adapting an existing model. (Test: have you written the paper? Or are you describing a product?)
  • You have proprietary data at scale - hundreds of millions of rows or more, in a domain where public models have weak coverage.
  • You can fund and retain a research team for the multi-year horizon the work needs.
  • The output is core to your business, not a feature - it is the product or it is the moat.

If you are nodding at all four, door one is open. If you are nodding at one or two, you almost certainly want door two or door three.

What we do here

We do not build foundation models. We are not a research lab, and pretending otherwise would not serve you. What we do is help organisations decide, honestly, whether door one is right for them, and if it is, help them set up the programme that stands a chance of succeeding: scoping the research thesis, sizing the team, structuring the evaluation framework, and governing the decisions along the way.

For most organisations we work with, the conversation ends with a clear "not this door" - and that is a good outcome. Knowing which door not to walk through saves more money than picking any of them well.

One conversation, no commitment. We listen first, then suggest a path.

Ready to see where you stand?

Two minutes. No email required. Find out where AI fits your business - and where the gaps are.

Or skip ahead and start a conversation