bitpush 2 days ago

The problem is your costs also scale with revenue. Ideally you want to have control costs as you scale (the first you build is expensive, but as you make more your costs come down).

For OpenAI, the more people use the product, the same you spend on compute unless they can supplement it with another ways of generating revenue.

I dont unfortunately think OpenAI will be able to hit sustained profitability (see Netflix for another example)

7
simonw 2 days ago

"... as you make more your costs come down"

I'd say dropping the price of o3 by 80% due to "engineers optimizing inferencing" is a strong sign that they're doing exactly that.

lossolo 2 days ago

> "engineers optimizing inferencing"

They finally implemented DeepSeek open source methods for fast inference?

asadotzler 2 days ago

You trust their PR statements?

TZubiri 2 days ago

It's not a PR statement, it's a change in price. Literally putting money where the mouth is.

theappsecguy 2 days ago

Or they are trying to gobble up market share because Anthropic has been much better than OpenAI

petesergeant 2 days ago

Providers are exceptionally easy to switch. There's no moat for enterprise-level usage. There's no "market share" to gobble up because I can change a line in my config, run the eval suite, and switch immediately to another provider.

This is marginally less true for embedding models and things you've fine-tuned, but only marginally.

Davidzheng 2 days ago

o3 probably used to have a HUGE profit margin on inference, so I'd say it's unclear how much optimo was done;

programjames 2 days ago

I find it pretty plausible they got an 80% speedup just by making optimized kernels for everything. Even when GPUs say they're being 100% utilized, there are so many improvements to be made, like:

- Carefully interleaving shared memory loading with computation, and the whole kernel with global memory loading.

- Warp shuffling for softmax.

- Avoiding memory access conflicts in matrix multiplication.

I'm sure the guys at ClosedAI have many more optimizations they've implemented ;). They're probably eventually going to design their own chips or use photonic chips for lower energy costs, but there's still a lot of gains to be made in the software.

Davidzheng 2 days ago

yes I agree that it is very plausible. But it's just unclear whether it is more of a business decision or a real downstream effect of engineering optimizations (which I assume are happening everyday at OA)

simonw 2 days ago

Seems more likely to me then them deciding to take a sizable loss on inference by dropping prices by 80% for no reason.

Optimizing serving isn't unlikely: all of the big AI vendors keep finding new efficiencies, it's been an ongoing trend over the past two years.

bitpush 2 days ago

This is my sense as well. You dont drop 80% on a random Tuesday based on scale, you do it with an explicit goal to get market share at the expense of $$.

Legend2440 2 days ago

>(see Netflix for another example)

Netflix has been profitable for over a decade though? They reported $8.7 billion in profit in 2024.

amazingamazing 2 days ago

They increased prices and are not selling a pure commodity tho

ACCount36 2 days ago

The bulk of AI costs are NOT in inference. They're in R&D and frontier training runs.

The more inference customers OpenAI has, the easier it is for them to reach profitability.

tptacek 2 days ago

All costs are not equal. There is a classic pattern of dogfights for winner-take-most product categories where the long term winner does the best job of acquiring customers at the expense of things like "engineering to reduce costs". I have no idea how the AI space is going to shake out, but if I had to pick between OpenAI's mindshare in the broadest possible cohort of users vs. best/most efficient model, I'd pick the customers.

Obviously, lots of nerds on HN have preferences for Gemini and Claude, and having used all three I completely get why that is. But we should remember we're not representative of the whole addressable market. There were probably nerds on like ancient dial-up bulletin boards explaining why Betamax was going to win, too.

TZubiri 2 days ago

Unlike Uber or whatsapp, there's no network effect. Don't think this is a winner takes all market, there was an article where we had this discussion earlier. Players who get a small market share are immediately profitable proportional to the market share (given a minimum size is exceeded.)

awongh 2 days ago

We don't even know yet if the model is the product though, and if OpenAI is the company that will make the AI product/model, (chat that keeps expanding into other functionalities and capabilities) or will it be 10,000 companies using the OpenAI models. (well, it's probably both, but in what proportion of revenue)

tptacek 2 days ago

Right, but it might not even matter if all the competitors are in the ballpark of the final product/market fit and OpenAI holds a commanding lead in customer acquisition.

Again: I don't know. I've got no predictions. I'm just saying that the logic where OpenAI is outcompeted on models themselves and thus automatically lose does not hold automatically.

Magmalgebra 2 days ago

Anyone concerned about cost should remember that those costs are dropping exponenentially.

Similarly, nearly all AI products but especially OpenAI are heavily _under_ monetized. OpenAI is an excellent personal shopper - the ad revenue that could be generated from that rivals Facebook or Google.

smelendez 2 days ago

It wouldn't surprise me if they try, but ironically if GPT is a good personal shopper, it might make it harder to monetize with ads because people will trust the bot's organic responses more than the ads.

You could override its suggestions with paid ones, or nerf the bot's shopping abilities so it doesn't overshadow the sponsors, but that will destroy trust in the product in a very competitive industry.

You could put user-targeted ads on the site not necessarily related to the current query, like ads you would see on Facebook, but if the bot is really such a good personal shopper, people are literally at a ChatGPT prompt when they see the ads and will use it to comparison shop.

whiplash451 2 days ago

Alternative: let users reduce their monthly bill by accepting a sponsored answer with a dedicated button in the UI

(with many potential variants)

marsten 2 days ago

You raise a good point that this isn't a low marginal cost business like software, telecom, or (most of) the web. Efficiency will be a big advantage for companies that can achieve it, in part because it will let them scale to new AI use cases.

With the race to get new models out the door, I doubt any of these companies have done much to optimize cost so far. Google is a partial exception – they began developing the TPU ten years ago and the rest of their infrastructure has been optimized over the years to serve computationally expensive products (search, gmail, youtube, etc.).

aizk 2 days ago

> sustained profitability (see Netflix for another example)

What? Netflix is incredibly profitable.

bitpush 2 days ago

Probably a bad example from my part, but also because of increasing the costs and offering a tier with ads. I was mostly talking about the Netflix as it was originally concieved. "Give access to unlimited content at a flat fee", which didnt scale pretty well.

whiplash451 2 days ago

Isn't this exactly what they offer today?