That genAI LLM is just somebody else's model
A close-up photo of network cables plugged into various switches.  Photo by Scott Rodgerson on Unsplash.

(Photo by Scott Rodgerson on Unsplash)

You know that phrase, "the cloud is just somebody else's computer?"

Some company racked and wired a server farm, which you're accessing through API calls.

The same holds for genAI: "that genAI LLM is just somebody else's model."

Some company trained the model, and you're accessing it through API calls.

(Yes, technically you could train your own LLM from scratch. Just like you could build out your own internal cloud infrastructure so your teams could have access to elastic resources. You could. But the fact of the matter is that most companies using a cloud provider's infrastructure, or a genAI vendor's LLM, are doing so precisely because the DIY approach is out of reach. They either want or need to build on top of someone else's work.)

And here's the thing: there's nothing inherently "wrong" with using someone else's genAI model!

It's mostly a matter of understanding your risk/reward tradeoff. That means considering the unknowns:

For some business use cases, that's fine.

But if any of those unknowns cause you discomfort … it's time for a think.

Complex Machinery 019: Five types of magic

The latest issue of Complex Machinery: AI is magic. In various meanings of the word.

Getting to know that vendor's genAI model

You didn't build it yourself, so you have to take some things on faith.