Hacker News new | past | comments | ask | show | jobs | submit login
How Does OpenAI Survive? (wheresyoured.at)
151 points by fredski42 6 months ago | hide | past | favorite | 188 comments



A well reasoned article that fundamentally downplays both the pace of innovation and the exponential increase in capabilities per dollar happening over time. AI is rapidly accelerating its improvement rate, and I do believe its capabilities will continue growing exponentially.

In particular, GPT-2 to GPT-4 spans an increase from 'well read toddler' to 'average high school student' in just a few years, while simultaneously the computational cost of training less capable models goes down similarly.

Also worth noting: the article claims Stripe, another huge money raiser, had an obviously useful product. gdb, sometime-CTO of stripe and its fourth employee, is now president of OpenAI. And, most of all, the author doesn't remember how nonobvious Stripe's utility was in its early days, even in the tech scene: there were established ways to take people's money and it wasn't clear why Stripe had an offering worth switching to.

For an alternate take, I think https://situational-awareness.ai provides a well reasoned argument for the current status of AI innovation and growth rate, and addresses all of the points here on a general (though not OpenAI specific) way.


It's too early to say for sure if LLM capabilities are on an exponential growth function or a sigmoid, but my money is on sigmoid, and my suspicion is that we're plateauing already.

GPT-4 was released 16+ months ago. In that time OpenAI made a cheaper model (which it teased extensively and the media was sure was GPT-5) and its competitors caught up but have not yet exceeded them. OpenAI's now saying that GPT-5 is in progress, but we don't know what it looks like yet and they're not making any promises.

What I'm seeing right now suggests that we're in the optimization stage of the tech as it is currently architected. I expect it to get cheaper and to be used more widely, but barring another breakthrough on the same order as transformers I don't expect it to see the kind of substantial gains in abilities we've hitherto been seeing. If I'm right, OpenAI will quickly be just one of many dealers in commodity tech.


> GPT-4 was released 16+ months ago. In that time OpenAI made a cheaper model (which it teased extensively and the media was sure was GPT-5) and its competitors caught up but have not yet exceeded them. OpenAI's now saying that GPT-5 is in progress, but we don't know what it looks like yet and they're not making any promises.

I don't really know anything about business, but something else I've wondered is this: if LLM scaling/progress really is exponential, and the juice is worth the squeeze, why is OpenAI investing significantly in everything that's not GPT-5? Wouldn't exponential growth imply that the opportunity cost of investing in something like Sora makes little sense?


Yes, exactly. An observation of OpenAI's behavior gives many clues that suggest they know we've hit the plateau and have known for some time.

A huge one for me is that Altman cries "safety" while pushing out everyone who actually cares about safety. Why? He desperately wants governments to build them a moat, yesterday if possible. He's not worried about the risks of AGI, he's afraid his company won't get there first because they're not making progress any more. They're rushing to productize what they have because they lost their only competitive advantage (model quality) and don't see a path towards getting it back.


I think in one interview Ilya or Sam explained that there are limited text data in the world and this is probably one of the bottleneck. But they mentioned there is still a lot of data in other modalities such as audio, video. Probably the reason more focus on multimodal models and also synthetic datasets.

I also don't thing the only way to improve LLM is by improving as zero shot inference. Did wrote any code in zero shot style that compiled and worked? It's a multistep process and probably agents and planning will be a next step for LLM.

Cheap inference help a lot in this case since you can give a task during the night to AI what you wanna do. Go to sleep then in the morning review the results. In this way AI is bruteforcing the solution by trying many different paths but that's kind of e.g. most programming works. You try many things until you don't have errors, code compiles and passes the tests.


> But they mentioned there is still a lot of data in other modalities such as audio, video. Probably the reason more focus on multimodal models and also synthetic datasets.

I think this is really interesting, but I wonder if there really is enough data there to make a qualitative difference. I'm sure there's enough to make a better model, but I'm hesitant to think it would be better than an improved chatbot. What people are really waiting for is a qualitative shift, not a just an improved GPT.

> It's a multistep process and probably agents and planning will be a next step for LLM.

I agree, we definitely need a new understanding here. Right now, with the architecture we have, agents just don't seem to work. In my experience, if the LLM doesn't figure it out with a few shots, trying over and over again with different tools/functions doesn't help.


Because they're running out of training data. OpenAI doesn't want to scrape new text off the Internet because they don't know what is and isn't AI-generated. Training off AI data tends to homogenize the resulting output, as certain patterns overexpressed by one AI get overexpressed by others.

If they start scraping, training, and generating images and video, then they have lots more data to work with.


> Because they're running out of training data.

Now that would be super funny.

Civilization VII tech tree

AI singularity tech

Prerequisites: in order to research this, your world needs to have at least 100 billion college educated inhabitants.

:-)))


For multimodal input, okay, I can see the argument. But since, as you said, training on generated data can be almost worse than useless, what's the point of generating it?

If I had these concerns as OpenAI, I'd be pushing hard to regulate and restrict generative image/video models, to push the end of the "low background data" era as far into the future as possible. i feel like the last thing I'd be doing is productizing those models myself!


> If I had these concerns as OpenAI, I'd be pushing hard to regulate and restrict generative image/video models

They are! And I'm guessing maybe their perspective is if they can identify their own generative content, they can make a choice to ignore it and not cannibalize.


That's a brand new argument.

I don't actually think it's a bad one, but OpenAI didn't think that far ahead. They are pushing for regulation but that's mainly to screw over competing models, not to give them more data runway. Every capitalist is a temporarily embarrassed feudal aristocrat after all.

Furthermore, even if OpenAI had a perfect AI/human distinguisher oracle and could train solely on human output, that wouldn't get us superhuman reasoning or generalization performance. The training process they use is to have the machine mimic the textual output of humans. How exactly do you get a superhuman AGI[0] without having text generated by a superhuman AGI to train on?

[0] Note: I'm discounting "can write text faster than a human" as AGI here. printf in a tight loop already does that better than GPT-4o.


When it comes to R&D, those are all very similar. Sora is another way to predict tokens, if you will. It's not like you have to choose that much: Sora is not using up all the compute that should've went to GPT-5. (It might not even be training in the same cluster; it might train on a separate set of GPUs which are useless for GPT-5 purposes because they are too small and the datacenter in question is tapped out.) Sora can't be GPT-5, but it could lead to a GPT-6. You have the Gato approach of tokenizing video into a stream of text tokens with RL, but you might instead wind up taking a more complex diffusion approach to video. Yet the goal is still the same: to control the world and solve things like robotics by predicting. And if you want to reach the goal as quickly as possible, you will try things in parallel, even if you know most of them will fail and you might wind up just doing the obvious thing you expected to do from the start. (After all, the whole reason we are here is that OA didn't throw 100% of its compute into big PPO runs doing DRL research, but Alec Radford was allowed to keep tinkering away with RNNs for predicting text, which led to GPT-1 and then GPT-2 etc.)


If I was going to nitpick I would point out that it's not necessarily the GPUs that Sora is consuming, it's the engineering effort from what could be called the top talent in AI and the vast amount of money OpenAI is borrowing that could be spent elsewhere.

> And if you want to reach the goal as quickly as possible, you will try things in parallel

This is sort of the exploration-exploitation problem, right? But I think you'd agree that a company full of people who firmly believe that GPT-(n+1) will literally be AGI, and that we're on an exponential curve, will be fully in exploitation mode. In my mind, exploring methods of generating videos is not a path towards their stated goal of AGI. Instead, it's an avenue to _earn money now_. OpenAI is in a slightly awkward position: their main product (ChatGPT) is not super useful right now, and is facing increasingly viable competition.


> I don't really know anything about business, but something else I've wondered is this: if LLM scaling/progress really is exponential, and the juice is worth the squeeze, why is OpenAI investing significantly in everything that's not GPT-5? Wouldn't exponential growth imply that the opportunity cost of investing in something like Sora makes little sense?

You can spend 100% on the next generation or you can spend a small percentage to productize the previous generation to unlock revenue that can be spent on the next generation.

The latter will result in more investment into the next generation.


In the real world there is no such thing as having an exponential growth function, they're all really just sigmoidal. The sole question is the latter suspicion: is the plateau now happening now or coming later? With sigmoidal growth it's near impossible to have clarity on that until after the fact.


Anthropic’s Claude 3.5 surpasses GPT-4o in a number of applied tasks, coding in particular.

The progress we’ve seen to date was powered by the ambitions and belief of NVIDIA and LLM companies.

Now, it’s head-to-head competition. It is way too early to call an impending slowdown.

Given how NVIDIA and Meta are leaning in on OSS, the next 18 months are going to be very interesting.

Even if fundamental progress slows, there are many, many secondary problems to solve in using the capabilities we have today that are rapidly improving. As someone deploying AI in business use cases daily, we are just now getting started.

I’d look to when NVIDIA starts to slow down on hardware as an early indicator of a plateau.

AWS is a commodity. When your commodity is compute, there’s a very large amount of growth available.


> the author doesn't remember how nonobvious Stripe's utility was in its early days, even in the tech scene

I have to push back on this. Anybody who had built for B2B credit card acceptance on the Web prior to Stripe's founding knew immediately what a big deal it was. For starters, they let you get up and running the same day. Second, no credit check (and associated delays). Third, their API made sense (as compared to popular legacy providers like Authorize.net) and was easy to integrate using an open source client. Fourth, self-service near-real-time provisioning. Their value proposition was immediately obvious, and they nailed all of these points in their first home page[1].

By contrast, Fee Fighters[2] was innovative for the time but still required me to fax a credit application to them. They got me up and running faster than the legacy provider, which is to say about a week. And I think I only had to talk on the phone with them once or twice. I remember really liking Fee Fighters, but Stripe was in a class of its own.

Stripe was a hit because they promised to solve hard problems that nobody else did, and then they did exactly that. (You still don't have to talk to a rep or do a personal credit check to start using Stripe!)

1 - https://www.quora.com/What-did-the-first-version-of-Stripe-l...

2 - https://en.wikipedia.org/wiki/Feefighters


Huh. that curl is 9 lines of code. maybe that's what the Bloomberg article was thinking of.

https://www.bloomberg.com/news/features/2017-08-01/how-two-b...


OpenAI is depending on a *breakthrough* in order to produce a product that will provide a return on investment.

This "breakthrough" is often touted as AGI or something similar to it which to me is even more risky than a nuclear fusion startup as:

1. Fusion has had some recent breakthroughs that could result in a commercially viable reactor eventually.

2. Fusion has a fundamentally sound theoretical basis unlike producing AGI (or something like it).


even if we don't reach AGI anytime soon there is still many new applications for current AI that we haven't explored too much yet. Robotics will be a huge one IMO.


The problem with your comment is: would this approach support these valuations? And also for robotics, is a company like OpenAI even anywhere near being able to use their tech for robotics?


OpenAI made an agreement with Figure - robotic company. Today figure is supposed to announce something so let's see how is the progress going.

Doing a lot of tedious tasks by robots already will be a big industry. They don't have to be super smart.

Google made a lot of money with ads in Google search - eventually perplexity or searchGPT will have some ads. Maybe they can also make money with affiliate links.

Sora et al will shake movie industry and commercials.

Voice assistants and chatbots already change call centers.

It's been just ~1.5 years since gpt3.5 release and 2 years since stable diffusion - e.g. iphone 2g wasn't mainstream and they sold only 6 million devices in first year, today they sell 250 million iphones a year


Idk, AGI (at least in the information domain) has a sound theoretical basis, the scaling laws seem pretty strong for now and there's a track record of shattering human benchmarks.


Curious what the objections are?


Calculators also had a track record of shattering human benchmarks.

Also saying something like:

> has a sound theoretical basis

Without backing it up at all hardly makes much sense...


> there were established ways to take people's money and it wasn't clear why Stripe had an offering worth switching to.

That wasn't true at all. Stripe was a product that people were rushing to pay for it for just how good and useful it was. It was an example of success MVP that people want to pay to use and the profitability was not a problem.

The same can't be true for OpenAI. We don't know how long it can stay in the red. Maybe it can survive. Maybe its money will run dry first. We are not so sure at current stage


That last part is a key differentiator between Stripe and OpenAI.

Stripe had high variable costs (staff, COGS of pass-through processing fees) but low fixed costs. OpenAI has enormous fixed (pre-revenue!) costs alongside high variable costs (staff of AI engineers, inference).

Financially, OpenAI looks more like one of the EV startups like Tesla or Rivian than it does a company like Stripe. And where Stripe was competing with relatively stodgy financial institutions, OpenAI is competing with the very biggest, richest companies in the world.


What are OpenAI's enormous fixed costs if not the staff?


Models cost a ton of money to train. GPT-4 training was > $100m. GPT-5 training is estimated to be order of $1B-$2B. This is in the same ballpark as Tesla allocated to get the Fremont facility online.

It is likely training generations beyond GPT-5 will cost still more.

These are essentially fixed costs as OpenAI has to pay to train the model whether or not anyone ever uses it.


if you're here you probably know about Claude and llama3 but for people outside of tech, how many are just going to plug ChatGPT into Google and not venture any further and just plonk down $20?


Is the market of individuals plonking down $20 enough to cover their costs, if they're not able to make sales to moderately clueful CTOs?


Claude is legitimately amazing and saves me so much time and effort. I use both ChatGPT and Claude, at least for now, and Claude really impresses me.


My biggest issue with Claude was timeouts when uploading images for translation, but otherwise love it


Isn't that kinda solved problem already? Just push more money in marketing. Like actual marketing. I have only seen some medical AIs this far... But same B2C model as done previously seems the obvious play to follow when someone has something.


Why would they plug chatgpt into google when gemini is built in?


Exactly my point. I took great pains to not say "OpenAI will 100% die without fail," because doing so would be declarative in a way that would wall off my argument, no matter how well I researched and presented it.

Instead, I wanted to show people the terms under which OpenAI survives, and how onerous said terms were. It's deeply concerning - and I do not think that's a big thing to say! - how much money they may be burning, and how much money they will take to survive.


Thanks for reading! I downplay it because I fundamentally disagree on the pace of innovation and the exponential increase in capabilities per dollar happening over time. I do not see the rapid acceleration - or at least, they are yet to substantively and publicly show it.

I also think it's a leap of logic to suggest that the former CTO of Stripe joining is somehow the fix they need, or proof they're going to accelerate.

Also, I fundamentally disagree - Stripe was an obvious business. Explaining what Stripe did wasn't difficult. The established ways of taking money were extremely clunky - perhaps there was RELUCTANCE to change, which is a totally fair thing to bring up, but that doesn't mean it wasn't obvious if you thought about it. What's so obvious about GPT? What's the magic trick here?

Anyway, again, thanks for reading, I know you don't necessarily agree, but you've given me a fair read.


Claiming Stripe was obvious is ahistorical unless you believe tens of thousands or even millions of entrepreneurs discarded $100B.

You have a small point that anyone who used authorize.net or similar wanted it to be better and that was obvious, but there's nearly infinite things people want to be better. I'd like breakfast, my commute, my car, my doctor, my vet, etc to be better. That you could make a better thing was incredibly non-obvious and that's why no one did.


If OpenAI can be as good as an outsourced employee at ~$10 per hour, then you should be looking to replace those outsourced employees. The US and EU employees at >$10 per hour, likely with many tens or hundreds of thousands of dollars in compensation, still exist because they provide some sort of value that necessitates that spend.

I am bearish on AI because the nimbleness of humans, even the outsourced ones, is quite capable. If you only want the AI to operate in a box, then you probably can code the decision tree of the box with more specificity and accuracy than a fuzzy AI can provide.

It's a very useful tool, I'm skeptical however about how it can disrupt things economy-wide. I think it can do some things very well, but the value to the market and businesses vs. the cost of training and adapting it to the business need is quite suspicious, at least for this cycle. I think this is one of those "wait 10 years" situations and many AI companies will die within 1 to 3 years.


> It's a very useful tool, I'm skeptical however about how it can disrupt things economy wide.

It won't disrupt much because we already had "AGI" of a sorts. The internet itself, with billions of people and trillions of pieces of text and media is like a generative model. Instead of generating you search. Instead of LLMs you chat with real people. Instead of Copilot we had StackOverflow and Github. All the knowledge LLMs have is on search engines and social networks, with a few extra steps, and have been for 20 years.

Computers have also gotten a million times faster and more networked. We have automated in software all that we could, we have millions of tools at our disposal, most of them open source. Where did all that productivity go? Why is unemployment so low? The amount of automation possible in code is non-trivial, what can AI do dramatically more than so many human devs put together? Automation in factories is already old, new automation needs to raise the bar.

It seems to me AI will only bring incremental change, an evolution rather than revolution. AI operates like "internet in a box", not something radically new. My yet unrealized hope is that assisting hundreds of millions of users, LLMs will accumulate some kind of wisdom, and they will share back that wisdom at an accelerated speed. An automated open sourcing of problem solving expertise.


Yes, and it's interesting to note that as the proportion of quality information available (and searchable) has declined across the internet there is now an alternative in LLMs - which are being used to further decrease the availability of quality information on the internet.


>My yet unrealized hope is that assisting hundreds of millions of users, LLMs will accumulate some kind of wisdom, and they will share back that wisdom at an accelerated speed

The only wisdom you could derive as a machine interacting with humans at scale is that they're not to be trusted, and that you'd rather you didn't have to given the choice.


Imagine you want to make a script with the LLM, it generates code, and you run it and it errs out. You paste the error, the model gains a new nugget of feedback. Do this sufficiently many times with many devs, and you got an experience flywheel.

But this applies to all domains. Sometimes users return days later to iterate on a problem after trying out in real life ideas generated by AI, this is how LLMs can collect real world feedback and update. Connect related chat sessions across days, and evaluate prior responses in the context of the followups (hindsight).

There is also a wealth of experience we have that is not documented anywhere. The LLM can gradually rub off our lived experiences by making itself useful as an assistant and being in the room when problems get solved.

But this experience flywheel won't be exponentially fast, it will be a slow grind. I don't think LLMs will continue to improve at the speed of GPT 3.5 to GPT 4. That was a one time event based on availability of internet scale organic text, which is now exhausted. Catching up is easier than innovation.

But we can't deny LLMs have "data gravity" - they have a gravitational pull to collect data and experience from us. We bring data right into AIs mouth it doesn't even have to go out of its way to scrape or collect. Probably why we have free access to top models today.


You just pointed out the reason why OpenAI and others are struggling.

The current generation of LLMs is static. The holy grail of continual learning is still far off.


> If you only want the AI to operate in a box, then you probably can code the decision tree of the box with more specificity and accuracy than a fuzzy AI can provide.

I've been saying this for the past couple of years. Yes AI is cool, but we already have computers and computer programs. Things that can be solved algorithmically, SHOULD be solved algorithmically. Because you WANT your business rules and logic to be as predictable and reliable as possible. You want to lessen liability, complexity, and amount of possible outcomes.

We already even see this with human customer support. They follow a script and flowchart. They're just glorified algorithms. Despite being human, they're actively told to not be creative, not think, and act as a computer. Because, as it turns out, from a business perspective that's usually very advantageous (where you can do it).

AI would never, or should never, replace those types of tasks.


AI isn't something warming a seat sitting in front of a computer reading emails and responding to them. For a technology to replace a human, it would have to be a drop in replacement in all aspects of what it means to be human.


Stripe's ease of integration, pace of innovation and user flow experience was a game-changing upgrade from the brutal payment processing options that we were forced to use until it was available. Nobody building payment processing at the time needed to be convinced that Stripe was awesome. They solved a hair on fire problem, especially for Canadians. The hoops we used to have to jump through were bonkers.


> exponential increase in capabilities per dollar happening

logarithmic, the capabilities increase with log(cost), what grows exponentially is compute used over time


Meh. I'm waiting to see evidence that this exponential growth means anything. Every day I read something like the statement that it's an "average high school student" now, or a med student or a law student or whatever, yet it seems obvious that the difference between it and a high school student is that a high school student can think at a basic level and it can't. So still waiting to see evidence that the exponential growth is on that axis and not the "bullshit on more complicated subjects" axis.


>AI is rapidly accelerating its improvement rate Is it? All I see are desperate AI companies slapping on multi-modality because text generation is nearly peaked


> And, most of all, the author doesn't remember how nonobvious Stripe's utility was in its early days, even in the tech scene: there were established ways to take people's money and it wasn't clear why Stripe had an offering worth switching to.

I think you are misremembering. Stripe was a _big deal_. They had a curl call on their home page for a while for how to take a payment IIRC. It was like how Twilio opened the door for anyone to send SMS, Stripe made it stupid-easy to handle payments online. Nothing else at the time compared in terms of simplicity and clearly defined fees.


> In particular, GPT-2 to GPT-4 spans an increase from 'well read toddler' to 'average high school student'

GPT-2 was indeed much smaller and weaker model. But the question do we have "exponential" boost after GPT3, or just marginal while competition commoditized this vertical.


That's a lot of unproven assumptions based on the fact that LLMs are just correlation printers.


ai is very rapidly hitting a plateau, both because they are just glorified markov chains, and them running out of data anyway


I think you don't understand the meaning of the word exponential. By now, at the intent expressed here, we'd all be in the Kurzeweil singularity if this was correct.

Hint: it's not correct. It's nothing like exponential. It's not even order of magnitude stuff. It's tiny increments, to a system which fundamentally is a bit of a dead end.


I love how pointless these talks are.

What does it even mean for a value whose scale is not defined to be logarithmic, quadratic, or exponential?

You guys are threading water about nothing.


This question can be asked about several "unicorns" that need huge inputs of capital to keep going. Only the era of zero interest rates made them work.

WeWork went bankrupt. Uber briefly made money but is losing it again, and is nowhere near paying back its investors. Tesla has become a major luxury car company, and is somewhat profitable, but the stock is way overpriced for a car company. Everybody now makes electric cars, so this is a low-margin business. (Reuters: "Tesla's bleak margins sink shares as Musk hypes everything but cars.")

OpenAI, as a business, is assuming both that LLM-type AI will get much better very fast, and that everybody else won't be able to do what they do. It's unlikely that both of those assumptions hold. Look at autonomous vehicles. First tech demos (CMU) in the 1980s. First reasonably decent demos (DARPA Grand Challenge) in the 2000s. First successful deployment in the 2020s (Waymo, maybe Cruise and Zoox). Still not profitable. 40 years from first demos to deployment, probably 50 to profitability. It's entirely possible that OpenAI's business will look like that. Their burn rate is way too high to sustain for that long.

Often it takes that long, even when the basics have been figured out. Xerography was first demoed in the late 1930s. The demo machine used to be in the lobby at Xerox PARC. Profitability came in the 1960s. By the late 1970s, everybody had the technology, and it was low-margin. Electronic digital computing goes back to IBM's 1940s pre-WWII electronic multiplier experiments, but didn't come down from insanely expensive price levels until the 1980s. Memory was a million dollars a megabyte as late as the mid-1970s. Color television was first demoed in 1928, and the first color CRT was developed in the 1940s. But mainstream adoption didn't come until 1966-1967.


All of Uber's investors made money, their market cap is $124 billion. Terrible ROI but higher than all of their rounds


If you believe this article, I would be happy to bet up to $10,000 that OpenAI will not collapse in the next 24 months. We could operationalize this by saying that OpenAI will employ at least as many people (or pay at least as much in salaries) in 24 months as it does now. The author of the article refused to take this bet when it was offered to him several times https://x.com/JeffLadish/status/1817999232105627722.


so we should only listen to the hypotheses of gamblers? The author shouldnt even feel the need to acknowledge the bet. So petty


If you are making predictions about the world, you should be willing to put money on those predictions so that you are grounded to making predictions that you actually believe in as opposed to just blathering.


Predictions can go wrong for any number of reasons beyond one's thesis. One can be right about what but wrong about when. To bet money on that just to satisfy ones (and others) ego is foolish. To recognize that is wisdom.


I think the common way to do this is to put 10k into openai stock, or short it.

(Instead of engaging with some some random internet commenteer.)


It would have to be a publicly traded company for this to be an option...


See, this is just a weak argument. This was thousands of words of hypotheses backed up with data and citation. Dismissing it as "blathering" and then demanding I make a bet based on your terms isn't an argument, nor is it a particularly compelling idea - you haven't engaged with my work, nor my arguments, nor my actual ideas.

You are, on some level, suggesting that money is a more compelling argument than an actual argument, because that is your only response.


This is too reductionist. There's a lot of reasons why someone could turn down that bet, other than "he's just blathering". Maybe he's very confidently down on OpenAI but 10k is a lot of money for him, and is being offered a bet by someone who's a noise generator but who's very well off, for whom 10k means nothing. If you think these people don't exist, you should visit WSB.

I evaluate arguments on their own merit, not based on extraneous data like appeal to authority or what bets the author makes.


It's exactly the opposite! Prediction/betting markets are possibly the most/only reliable way to forecast things with otherwise limited information. Prediction Markets are extremely accurate since being wrong makes people poor and they use every trick in the book to not be poor.


Companies die slowly. It will definitely take longer than two years before the company is gone, but you will know it's on its way out long before it happens.


> The author of the article refused to take this bet when it was offered to him several times https://x.com/JeffLadish/status/1817999232105627722.

So what? No one has participate in the "rationalist" subculture's weird practices. It means nothing to refuse to take a bet like that, let alone that the claims made in the article are suspect (which you seem to be implying).

[I can't actually read anything beyond the tweet you linked because twitter is stupid].


> No one has participate in the "rationalist" subculture's weird practices.

True!

> It means nothing to refuse to take a bet like that, let alone that the claims made in the article are suspect (which you seem to be implying).

False! If the author was sufficiently confident in their claims, they'd be happy to take the free money (or, if they're sufficiently liquidity-constrained, propose a smaller bet at similar terms). You can certainly argue that the practice of betting on one's beliefs is "weird" but that objection is circular. If you claim to see free money on the ground, and other people notice that you aren't picking it up, they would be correct to wonder why.


It's not free money on the ground, it's a bet on the outcome of some event.

You're presupposing that the only possible rational reason they might have to not bet might be that they are not confident but that's obviously not the case.

Just to pick one possible rational reason, maybe they or a loved one have been touched by the negative consequences of gambling so they don't believe in betting?


> It's not free money on the ground, it's a bet on the outcome of some event.

If you have convincing enough evidence for it, a bet is more likely than something that looks like money on the ground. Imagine betting the sun will rise tomorrow.

> You're presupposing that the only possible rational reason they might have to not bet might be that they are not confident but that's obviously not the case.

They didn't say it's the only reason, just that refusing to bet weakens their claim of confidence. Plausible but unconfirmed reasons are why it's only a weakening rather than disproof, but it's still a weakening.


It may weaken their claim for you. To me, refusing to take an ego bet with some random on the internet suggests they're not 9 years old.

And no I wouldn't bet that the sun will rise tomorrow. For a number of reasons. Firstly I don't care that much about money (shocking but true), so there is little to no upside if I was to win the bet other than the general benefit I would derive from the sun rising and I get that whether I bet or not.

Secondly I think betting about facts is an incredibly foolish thing to do in general. If anyone offers you a bet about some fact, you are the sucker and you just haven't figured out how yet. Your best move is not to play.

Finally betting with some random that I have no reason to think would make good on their side of the bet if I was to win. idk man that just seems like a really dumb idea. I don't understand why people think it makes them smart to think otherwise.


It's only an ego bet if you don't care about money. Most people would significantly enjoy an extra hundred or thousand dollars. It's a big upside.

> If anyone offers you a bet about some fact, you are the sucker and you just haven't figured out how yet. Your best move is not to play.

I think it's pretty clear in this case that's it's a genuine disagreement about the fate of the company.

> Finally betting with some random that I have no reason to think would make good on their side of the bet if I was to win. idk man that just seems like a really dumb idea. I don't understand why people think it makes them smart to think otherwise.

There are plenty of trustworthy arbiters/platforms available. For most people spending a couple hours looking into it is easily worth the money... if they're right.


> Secondly I think betting about facts is an incredibly foolish thing to do in general. If anyone offers you a bet about some fact, you are the sucker and you just haven't figured out how yet. Your best move is not to play.

Well, that sure sounds like a claim about how confident one should be about their understanding of reality. One most people disagree with, incidentally - do you own any equities?

Counterparty risk is a valid reason to avoid certain bets (or bet structures), but a totally separate objection from "betting is a weird thing that only weirdos do".


You need to come up for air. You seem to be stuck so deep in that weird subculture's "logic" that you're having trouble imagining someone who isn't.

I don't gamble, and even if I did I certainly wouldn't take a bet for a large sum of money with some internet rando playing a rhetorical game.


If you are making a prediction about the world, why not gamble on it? It shows to others that you have an actual stake as opposed to blathering.


Come up for air. I'm not part of that subculture and I feel no need indulge it or to prove myself to members of it by participating in its rituals.


As someone who doesn't know what subculture you're talking about, parent's argument sounds reasonable to me. There's even an old saying that everyone has probably heard - "Why don't you put your money where your mouth is." Is that some kind of fringe thing now?


> As someone who doesn't know what subculture you're talking about...

Bets like the one discussed here are a rhetorical tic of the "rationalist" community.

> ...parent's argument sounds reasonable to me.

Kinda sorta. It's just a weird ritual of theirs to demonstrate commitment, and they often have a really hard time understanding that their weird rituals and mores are not universal and they can't reasonably go around expecting random people on the internet to follow them. Challenging someone to a bet like this comes off exactly as a challenge to prove your commitment by putting a bucket on your head and banging it with a stick for an hour, just because that's some weird thing the challenger's friends do among themselves.


This is an argument about social perception. Once you get over the fact that some community on the internet does this thing, and you think that community is weird (and therefore the thing itself is also weird), you may observe that all human actions are implicit bets on various beliefs. Explicitly wagering money is just a special case of making certain narrow beliefs much more legible.

I agree that in practice, it's pretty likely that the author of the piece refuses to bet at least in part because betting substantial sums of money on outcomes that are non-central subjects of wagers (i.e. not an explicit game of chance, sports, politics, etc) is socially unusual. But I also think that if he was very confident he'd be happy to take the money.


You're using the word weird a lot, but I don't see what's so weird about the concept of "put your money where your mouth is" (or "put up or shutup", or "wanna bet?", or other variations on the theme).

The bucket on head example certainly is weird. I just don't see how that's related. The logic I'm seeing is 1) the "rationalist" community does x, 2) I don't like the rationalist community, 3) so x is stupid. Is there more to it?


> You're using the word weird a lot, but I don't see what's so weird about the concept of "put your money where your mouth is" (or "put up or shutup", or "wanna bet?", or other variations on the theme).

Let me explain it to you slowly and clearly: neither challenging people to bets nor banging a bucket on your head are things typically asked of people to prove their commitment to a statement, hence it's weird to ask someone to do them. They might common in some weird subculture, but it is additionally weird to go up to someone outside of that subculture and make weird subcultural demands.

Also the examples you gave are literally rude, for the most part, and don't actually mean what you seem to think they mean (if you want a source: https://www.merriam-webster.com/dictionary/put%20up%20or%20s...).

> The logic I'm seeing is 1) the "rationalist" community does x, 2) I don't like the rationalist community, 3) so x is stupid. Is there more to it?

You see wrong.


> neither challenging people to bets nor banging a bucket on your head are things typically asked of people to prove their commitment to a statement, hence it's weird to ask someone to do them.

We apparently just live in different universes. In my universe, one of those things is completely unremarkable. I must be a part of some weird subculture, TIL.

> Also the examples you gave are literally rude, for the most part, and don't actually mean what you seem to think they mean

Like most things, they can be rude, or they can be completely benign, depending on the context and delivery. I don't see anything conflicting in the definition you linked, but as you pointed out, I see wrong, so that's probably why.


Ed sure looks like a PR person (ie paid liar, and his chosen career) who spotted an anti-tech market opportunity and is now confidently and loudly asserting a bunch of opinions. See, eg, lots of people on substack. People will happily pay to hear their beliefs echoed to them.

Having even minor amounts of skin in the game would make it a lot more likely he's sharing, well, actual beliefs not opportunistic opinions.

I also have never read any rationalist whatever you're complaining about and you're clearly attempting to discredit people who think Ed is probably just venal. I think you're the one who needs to touch grass here and spend less time on Twitter/x/threads/whatever has you so convinced everyone else is wasting time on them.


the market can stay irrational liner than you can stay solvent though, and $10k is a lot of money to bet for those of us who didn't get crypto rich. refusing to take part in your weird games doesn't say what you think it does.


> If the author was sufficiently confident in their claims

What? No.

I'm very confident in my driving skills. I'm a great defensive driver. But I won't bet on me not getting in a wreck this next year.

Because the real world is complicated and truly anything can happen. That doesn't mean you're right, that means you got lucky.

For example the Simpson's creators aren't clairvoyant geniuses, they just happened to get a lot of stuff right.


> But I won't bet on me not getting in a wreck this next year.

You already have, by deciding on a specific level of coverage for your auto insurance policy.

(That aside - really? Either you get into many more accidents than the average person, or you're not extrapolating out into the future. If given the chance to take the same bet at 1:1 odds every year, surely you'd then take it every year until you decided to stop driving for safety reasons?)


I commute 15 hours a week. In Texas.

I see a few accidents a day. Nobody actually wants to drive really, we just do it because we have to. If I could teleport to work or even take a train I'd do it, and I'm not alone in that.

Point is, I don't feel comfortable betting on things where there's a large amount of outside influence. And no, insurance isn't betting. It's risk pooling, not betting.


Nobody cares about someone being wrong in the past, because people update their beliefs all the time.

Meanwhile a bet over a silly internet discussion can get stuck in significant transaction costs vs buying, selling or even shorting a stock for the same money.


You are so sure you're right, what odds are you offering?


The employee angle seems cowardly as investor pumping would keep that artificially high. Why not profitability?


This article is timely and pairs well with Sequoia's $600B question: https://www.sequoiacap.com/article/ais-600b-question/ calculated simply from NVidia run rate revenue, which is the cost that genAI companies are paying. Where's the profit?

Meta's open source LLM stance makes things more spicy, making it challenging for anyone generate differentiated and lasting profit in the LLM space.

At the current pace, the LLM bubble is poised to pop in a year or two - negative net revenue can't keep growing forever - barring a transformative, next-generation capability from closed-source AI companies that Meta can't replicate. All eyes on GPT-5.


I just got a bit of data that hints that at least one of the assumptions in this blog post is false.

The post says:

> The supply shortage has subsided: Late 2023 was the peak of the GPU supply shortage. Startups were calling VCs, calling anyone that would talk to them, asking for help getting access to GPUs. Today, that concern has been almost entirely eliminated. For most people I speak with, it’s relatively easy to get GPUs now with reasonable lead times.

But a couple of days ago I heard from a startup founder that the usual cloud credits (~$100k in cloud compute) that AWS provides to vetted startups that passed some milestones are recently barred from being used on GPU-powered instances.


> Have a significant technological breakthrough such that it reduces the costs of building and operating GPT — or whatever model that succeeds it — by a factor of thousands of percent.

Like they already did in the last 2 years?

> Have such a significant technological breakthrough that GPT is able to take on entirely unseen new use cases, ones that are not currently possible or hypothesized as possible by any artificial intelligence researchers.

Huh, what are these use-cases which no AI researcher thinks AI is capable of solving? Does the author not realize that many employees at the leading AI labs (including OpenAI) are explicitly trying to build ASI? I am so confused????????

> Have these use cases be ones that are capable of both creating new jobs and entirely automating existing ones in such a way that it will validate the massive capital expenditures and infrastructural investment necessary to continue.

Why would they have to create new jobs? They just have to be good enough that OpenAI can charge enough money for them to be in the green.

OpenAI already has a $3.4 billion ARR! Most of that is _not_ enterprise sales.


"No AI researcher" is an impossible set of people to satisfy, so this doesn't count, but there's the AIMO contest with a $5m prize for creating an AI that can solve its problems.

https://aimoprize.com/


> Nobody has ever raised the amount of money it will need, nor has a piece of technology required such an incredible financial and systemic force — such as rebuilding the American power grid — to survive, let alone prove itself as a technology worthy of such investment.

AT&T.


The space race cost hundreds of billions in today's dollars, and AI could do a lot more for/to humanity than put a flag on the moon.

And I'm guessing Uncle Sam will want control of these AI companies anyway if AI starts looking even a little powerful/threatening.


> The space race cost hundreds of billions in today's dollars, and AI could do a lot more for/to humanity than put a flag on the moon.

That's a weird comparison. The goal of the space race was never to make money, and it was funded by the taxpayers, not VCs expecting a return on their investment.


The space race was also not something with a financial outcome at the end as its main goal.

I doubt that Uncle Sam cares that much about controlling OpenAI.


The power grid itself


Via pure monopoly power


AT&T didn't have a monopoly when it started.

It just burned a ridiculous amount of money stringing lines.



Isn't that sama's plan with "safety regulation"?


No because even Nortel in canada was a required to buy at&t made phones


Scott Galloway made his bones with similar WeWork analysis. Once Galloway exposed the non-sustainable WeWork economics, others picked up his work and added their own spin. While WeWork would have imploded without the Galloway analysis, Scott's articles were the tipping point.


> generative AI is a product with no mass-market utility.

> I am neither an engineer nor an economist.

clearly.


Pontificating on what actual profit will be generated by LLM companies is fair game, but reaching a conclusion that LLMs have no mass market utility is blisteringly insane. I could easily measure the increase in value I alone have had from using them in the tens of thousands of dollars.

It's more plausible for me that we will see a notable productivity increase in a lot of sectors of the economy over the next decade. Part of me wonders if this is an additional reason why the Russell 2000 has been spiking lately (investors concluding that there is more money to be made from the general productivity increases in the wider economy than the tech companies providing the LLMs that don't seem to possess any monopolies on the technology), but this is just my speculation.


I think you are misunderstanding my point, and perhaps I should've worded it more-precisely"

"Mass market utility" here refers to its ability to sell at the scale it would need to substantiate its costs. As it stands, LLMs do not have mass-market utility at the scale that they need to substantiate their costs. It is really that simple. If they did, these companies would be profitable, and they would be having a meaningful effect on productivity, which they are not.

See page 4 of this report from Daron Acemoglu of MIT: https://www.goldmansachs.com/images/migrated/insights/pages/...


And he admitted he is “neither an engineer nor an economist, nor do I have privileged information”.

So why should we listen to him? ChatGPT has saved me a lot of time. Luckily it’s well trained on AWS API’s, tge SDK, CDK, Terraform and Kubernetes. Anything it isn’t trained on, I just give it links to the documentation


If I read you correctly, you are implying that the first statement is obviously false. Can you fill me in on the mass-market utility? Because I keep coming to the same conclusion as the author.

The "killer app", as far as I can tell, is essentially natural language search. However, the core function (in my opinion) has existed since DuckDuckGo added contextual infoboxes to the right of search results ("knowledge panels"), and the benefit of using natural language has existed since Siri and has never seemed to add much to the experience for me. AI image generators seems to be used mostly by youtube creators and spammers. The main users of AI language generation seem to be spammers and crappy content farms on TikTok.

Commercials for AI products ALWAYS lie by speeding up the time it takes for results to arrive, and the most impressive demos always seem to end up as some version of the same useless feature: "what am I looking at right now?" Who needs that? AI-assisted coding also seems to have a similar issue. Demos that supposedly show off the technology never actually use it to create the kind of code that is actually worth money.

I'd be happy to be proven wrong here, but I keep looking and I never find that killer app.


Thousands of call center employees are being replaced right now. You may not even notice the difference, tier 1 support is heavily scripted already.

Outbound sales is being automated.

Lots of very hum-drum stuff. Ordering room service at a hotel for example.

Tons of data entry jobs are now gone.

LLMs are already better at humans from a cost perspective for many tasks.


Translation is another. It's uncertain to what extent experienced professional translators are losing work—I work in the field, and I've heard anecdotal evidence on both sides—but it's clear that LLM-driven machine translation is already producing significant value for millions of people around the world. This might not be apparent to monolingual people in English-speaking countries, but it's a major factor driving the rapid adoption of LLMs here in Japan.

If OpenAI's new voice mode turns out to be as versatile and context-aware as promised, it should be equally valuable for interactive spoken interpretation.


There's definitely some value in undercutting minimum-wage outsourced call centres. But enough to justify these valuations? This is stuff that was already being partially automated back in the '80s; I can see the argument that OpenAI is the next Autonomy, but when you're this heavily leveraged an incremental improvement is a failure.


OpenAI is the platform everyone else is building on top of.

Think Microsoft in the late 80s through the 90s.

If you are charging for the platform, you can let thousands of other businesses try the risky hard to scale stuff and you'll get a cut of everything no matter who wins.


Microsoft wasn't taking any risks there though, they were a profitable business selling positive margin projects the whole time - that's more where NVidia is than where OpenAI is.


Do we need AI for all this?

Or could we have instead done some simple web service that would do most things... The sad reality is that companies don't want easy and efficient customer service... As that would allow customers to cancel their continued payments...


Credit card companies and large banks each have call centers with multiple tens of thousands of employees.

> The sad reality is that companies don't want easy and efficient customer service... As that would allow customers to cancel their continued payments...

This is not true at all. The true reason is that a well trained call center employee can easily cost a company $20 per customer issue resolved (total cost inclusive of training, office space, equipment, etc).

For any low margin business (e.g. hardware under $500 USD!) that basically destroys the entire profit from that customer.

Customer service is expensive.


> Thousands of call center employees

Uh... you realize they were already algorithms, right? Meaning, they have a flow chart they follow. They don't make any decisions, they take input and respond to output.

The only reason they're not computer programs is because they NEED to be human. So the human on the other end trusts them. Even though everyone understands they have no free will or reasoning abilities (or if they use them they get canned).

AI doesn't fit that use case. Number 1 is because AI IS NOT algorithmic. So it's a liability to use. Number 2 is it's not human. Again, if you're going the no human route there's infinite cheaper, more reliable, faster, and overall better in every way programs you can use.


> what am I looking at right now?" Who needs that?

Anyone who is sight impaired, or doesn’t have their glasses on while reading a menu, or looking at a sign in another language.

It’s clearly not in final form. In 1978, people couldn’t see the use of home computers. In 1988, most people were saying the same thing about email. In 1998, most people were saying the same thing about the internet.

It might not prove out, but evaluating something super early isn’t all that interesting. Let’s see where we are in 2030 at least.


> It’s clearly not in final form. In 1978, people couldn’t see the use of home computers. In 1988, most people were saying the same thing about email. In 1998, most people were saying the same thing about the internet.

Give it a few more years for more people to forget about how hard things like NFTs got pushed with the same sort of arguments- then this'll come off better.


That one technology worked does not automatically mean any other technology will. There is no argument that begins "A worked and so B will too". Also, email is pretty much dead as a technology - it's just spam and password reset emails at this point.


I think parent was not saying new tech will succeed, merely there is an extended hangover/negativity due to recent busts.

Regarding email - it's probably one of the most used bits of technology in existence, and I bet it has a similar OOM economic impact to something like Excel. Calling it a dead technology is not accurate - so much business is done via email, especially internally at SMB.


I realize that these are commonly-held tropes, but where is the actual article that says this? There's the famous "the internet isn't a big deal" thing in Newsweek (and funnily enough that piece is extremely prescient in many other ways!), but I don't know if I've seen the kind of hype-busting then.

But also...the reason they might have is that email kind of sucked back then. Of course you wouldn't see the promise in something that was clunky and slow and nobody used.

This isn't a comparable to LLMs, though, because even someone who found them clunky could see why you'd want to send an email versus sending a letter.


The killer app is replacing 80% of white collar employees in 5 to 10 years.


Or 5%, right? What's the minimum amount that's "killer", on a societal level and on an industry-specific level? Human transcription might be a good case study.


Is this something people believe is a reasonable expectation?


Agreed. Next gen robotics with AI integration will do the same with blue collar workers a few years after that.


Actually, I would expect the robots to get better at blue collar work faster than LLMs at white collar work. The reason is that there are a lot of highly repetitive physical skills for which you can easily build a dataset for and then fine-tune a model.

People have built dedicated plastering robots, strawberry picking robots etc. Each robot only really needs to be good at one thing. Almost nobody is making the foolish mistake of building humanoid robots that then tackle the problem exactly the way humans do. The plastering robots spray the plaster, which is much faster.


You’ve clearly never had to troubleshoot wiring issues on a robot.


when the robots can fix themselves, then we're in really in trouble.

if humans still have to fix the robots that seems fine, as long as they run for long enough without needing rewiring.


If I could make a robot that fixes robots I’d have retired years ago.

It’s a complicated problem, and apparently very misunderstood.


There is a lot of money being poured into figuring this out[0]. My bet is on agents. They work, right now, and it’ll only get better.

[0]: https://x.com/omooretweets/status/1760000618557735289


Don’t worry you’re not alone. I’ll believe it when I see it. So far I haven’t seen it. And I work in AI (albeit at a low level).

If I can compare it to Google search back in the 90s. There weren’t all these evangelists saying in 5-10 years blah blah blah. We just used it, in our daily lives. I don’t use AI at all except at work. And I couldn’t even tell you what product it’s going toward because I don’t think 99% of employees know. Why we don’t know, who knows!


There absolutely were hordes of evangelists touting this or that tech company (including google, redhat, amazon etc as well as pets.com and a bunch of others who didn't survive) saying in 5-10years these were companies would be the biggest thing since sliced bread. In some cases they were right.

The VC money is betting on a similar sort of monopolistic dynamic occurring in the AI space. They're not saying that 100s of billions of dollars of value is going to accrue to openAI because of chatGPT they're saying 100s of billions of dollars of value is going to accrue in the space and the most likely outcome is that the vast majority of that is going to be siphoned up by a couple of behemoths. Other than the incumbents (Microsoft, Google) OpenAI seems best positioned at the moment and whoever else creates breakthroughts will probably be acquired anyway just like instagram got bought out by facebook as soon as they got large enough to matter.


> saying in 5-10years these were companies would be the biggest thing since sliced bread

It's not about growth, it's that nobody had to make future claims about web search or online ordering being useful to the average person, they already were useful when someone tried them.


Yes thank you that's what I was trying to say.


Good point. That's fair for sure.


Ironically enough, if AI powered code generation actually worked, OpenAI would have a business model right there. But it doesn't, and so they don't. What they hope is that other people will somehow make their business model for them - and then they'll sell their technology into that.

But AI code generation is garbage, just like AI text generation is. AI generated books, and songs, and poetry etc is just appalling rubbish that nobody wants to read or hear. AI generated art is ugly and generic, and stands out immediately as zero-effort and near-zero cost. Except of course, it cost alot of money to make, but nobody is willing to pay for it and it's so far been bankrolled by VC capital.


> But AI code generation is garbage,

"create an express server that uses CORS to only accept requests from <mydomain> that has 3 async endpoints defined, 2 GET endpoints titled <a, b> and 1 POST enndpoint titled <c>. The POST endpoint should accept JSON, and the GET endpoints will return JSON. Stub out the route handlers, they will be defined in another file. For the POST endpoint parse out the userId from <header field> and also pass it in along with the JSON"

"Here is a typescript definition of the JSON being posted to this endpoint, validate all string fields are less than 100 characters long, and that none of the values are null, then write the data to a redis DB at <location> using auth values already loaded up in the environment. Data should be written to <userid/location>"

"Now add exception and error handling around this code, cover the following case: Data is invalid - Return an error message to the user with HTTP error code 400, Redis is not accessible - HTTP 500. Also log errors using the <logging library> that has already been imported and initialized up above, and increment an appropriately named counter using the <analytics library> that is also already imported and initialized."

If you are getting garbage AI code, you are not using proper tools. If you just go to ChatGPT and ask it for stuff, you won't get very good code (and IMHO GPT4o is much worse at coding).

Use a proper editor that prompt engineers code generation (such as Cursor), and LLMs can write some really good code.

Can it end to end generate an entire service with just a vague description? No. But it can easily 2x-4x development speed. Heck I've thrown code at GPT4 and asked it to "please look for common concurrency bugs that may have been overlooked in this code" and watched as GPT4 saved me literal hours of debugging.

What OpenAI has is a problem monetizing the value their product brings, since the value to each customer is so different. When I'm at home, I'm happy to pay $15 a month in API usage fees to accelerate my coding, but a company should be happy to pay hundreds of dollars per month to accelerate the speed code is written. Figure $100 an hour average cost for a dev on the east or west coast of the USA, if GPT4 saves 10 hours a month (easy!) then OpenAI should be charging at least $200-$500 a seat licenses per month.


Agent for X is valuable and lots of startups are trying to build it in some vertical. As these agents evolve, they have the potential threaten incumbents in their own respective vertical. It’s also unclear how we’ll orchestrate all the agents or discover them. None of this is very obvious and it’s very possible to hit a major tech roadblock. But nobody knows.


> As these agents evolve

That doesn't answer what's happening now. The AI companies can't keep saying "AGI is around the corner!" for much longer.


No serious company is claiming AGI is around the corner. There's going to be scams like Rabbit, Friend and similar, but there are scams for everything.

Agents are evolving though. The coding agents a few months ago were a joke (Devin) and now they're actually usable for basics.


Are they? Have you used Devin to do anything though? So far as I understand it, the Devin demo was faked.


That's what I meant - Devin was a joke, I've seen the promo and didn't use it. But I experimented with Plandex recently and it's usable for various small tasks. Not great, but usable if you make the right choices and getting better.


Chatbots in their current form already make excellent tutors and coding partners for many people.


The killer app is the chatbot itself? Fastest app adoption / scaling ever? Like I think people are looking for an 'and then' to be sold, but it's unclear why that has to be the case.


Chatbots are, so far, obviously a $1B+/yr business.

I think it’s not obvious yet that they go much further beyond that?

It’s a lot of money, sure, but it’s not world-transforming stuff.


I'm an engineer and I agree, does that help?


I think the thing that nobody saw coming was Facebook coming out with their own, that they spent hundreds millions creating, and then releasing the model of it, with a generous license.

If someone had come out of a copy of Google in 2000, we'd be looking at a much different picture.


>Have a significant technological breakthrough such that it reduces the costs of building and operating GPT — or whatever model that succeeds it — by a factor of thousands of percent.

What does reducing costs by "a factor of thousands of percent" mean? It starts printing money? It costs 1/10 as much?


> generative AI is a product with no mass-market utility - at least on the scale of truly revolutionary movements

This line is absurd. I use it constantly. 4o reads my code and generates documentation and type annotations. It generates boilerplate code. It generates logos for projects. I review all of its outputs code wise and make the odd correction here or there. I use it to check over documents before I send them. It’s replaced Stack Overflow entirely in my workflow.

I’m curious as to what’s above the author’s line for revolutionary.


The fascinating thing with arguments like this is how they entirely ignore that OpenAI trained the tool you're using with Stack Overflow. And so if what you're suggesting is that AI tools will replace SO, then as soon as it does the AI is dead in the water.

Who will continue to feed it new information? The answer is, nobody will. And instead of having a community driven knowledge base, you'll have a bankrupt corporation and nowhere to turn to when you discover that you can't write code anymore.

None of the tools can exist without stealing the net sum of human knowledge - insofar as that is represented by the contents of the internet - for corporate profit. If what the AI proponents claim comes true, that source of knowledge will cease to exist. And what then?


The AI does not loose its utility once it has replaced SO. Further improvements would need to come from other sources than more "free" text data though. There are several avenues for this: A) better learning efficiency - it is known that models are not samples efficient now B) pay people to create more data at large scale - already happening at the big LLM companies (though maybe less do for code right now) C) learn from the prompting interactions that customers - extremely desirable, as people would effectively pay you to improve your model. Companies are gonna keep that close to their chest D) multi-modal learning and transfer between modalities E) Self-learning, where ML model iteratively improved itself. Analogous to AlphaGo, GANs etc. This is the holy grail, but also possibly the hardest. Code has the benefit of being formal languages with available verifiers, which should be useful here. F) Specific go code: Using functional feedback, not just textual. For example the large amount of unit tests available on GitHub

So I believe there are many avenues for further improvement. I still think it will be very hard though, and until we hit the next breakthrough, will be very resource intensive and possibly quite slow going.


> And instead of having a community driven knowledge base, you'll have a bankrupt corporation and nowhere to turn to when you discover that you can't write code anymore.

I still write a lot of code without needing to use an LLM. It’s just a tool that helps me be more productive. I don’t see how what is doing is any different than Google just displaying the answer instead of links to the answer.


OpenAI will survive by being absorbed into Microsoft as a capability- OpenAI employees will be MSFT badged, working under an MSFT SVP, to add functionality to existing products.


To comment a little on the economics:

OpenAI has raised $11.3bn (source the article)

Since partnering with Microsoft in 2019, Microsoft's valuation has gone from $0.7tn to $3.1bn, or an increase of $2.4tn, a lot of that on AI enthusiasm.

Microsoft can sell some shares to fund OpenAI, 2.4tn being about 200x what they've put in.

Sure the market bubble will pop at some stage but not by 200x. I'm skeptical of the they can't survive argument.

Also I recall in the early days of Facebook, Google and Amazon people saying they lose money each year, the first two didn't have a monetization model, how will they get by? But of course they ended up some of the world's most profitable companies. With AI also you have to think a few years down the road when ASI's output may exceed the current global GDP ($100tn or so).


No genius employees = mediocre outcome. Ilya left. Andrej left. Much of the talent at the top of the pyramid left. A company is literally its people.


It's not that these AI companies need to raise $100 billion, it's that they can. They could go slower, working on custom hardware or better architectures to get 1000x cheaper training, but their models would only be as good as the ones from 1-2 years ago. Because foundation models are a winner-takes-all game, any individual company needs to spend as much as they can get, even if it means most investors will lose big. There are many companies working on hardware gains (e.g. Lightmatter, Groq), but they're not competing in the foundation model business.


This is a nonsensical article to me.

"I ultimately believe that OpenAI in its current form is untenable."

Followed by a bunch of reasons why. Later they write:

"What I am not saying is that OpenAI will for sure collapse, or that generative AI will definitively fail"

What? Didn't they just explain 100 different reasons why they think think OpenAI will fail? There was also this:

"To be clear, this piece is focused on OpenAI rather than Generative AI as a technology — though I believe OpenAI's continued existence is necessary to keep companies interested/invested in the industry at all."

To be clear? So they are trying to separate OpenAI from gen AI. Then they throw in a hyphen and say, oh but without OpenAI, companies would stop spending time and money on gen AI. Ok, thank you for the..clarification.

I stopped reading after that.


Small correction:

"GPT-4o Mini (OpenAI's "cheaper" model) already beaten in price by Anthropic's Claude Haiku model"

GPT-4o Mini is presently cheaper than Claude 3 Haiku.


Will always give an upvote to ed zitron, who has a very anti tech/anti AI tilt but a pretty reasonable podcast called “better offline” I’ve quite been enjoying.


GPT-4 could’ve written this article better- hey there’s an use case right there!


They say no company has ever raised the kind of money OpenAI has, like they are at some kind of ceiling. Microsoft’s market cap, which is only equity, not including bonds (only $42B), is over $3T.


Right, but Microsoft has been profitable for decades.


And Microsoft was once a smaller company. Google was too. Who is to say OpenAI can't become big tech company like them?

Also, I'd like to point out that the total investment into AT&T, which they position as untenable, is less than AT&T has spent on their network investments every single year for at least the last 10 years. It isn't like companies don't invest billions of dollars into things.


Sure, but the argument here isn't that it's impossible. Just that to do what they need to do, they need to do something that really hasn't been done before.

I do not THINK Microsoft will put that much money into it. They could! It isn't impossible. But it would be totally unprecedented.


This whole thing reads like: google doesn't have a business model because anyone can do search. Well, how many years later, yeah the writers saying Google was DOA are actually DOA.


I'm afraid this isn't an accurate comparison.

1. I do not know of any article that said that Google was "DOA" or "done" as a result of the choice of a search engine as a business model. In fact, search engines were an already-established industry at the time. If I'm wrong, I'd love to read it, as I imagine it's a fascinating historical document - even if it was horribly wrong!

2. OpenAI's business model and Google Search's business models are totally different. Apples and oranges. The way that OpenAI monetizes, the technology it uses to both deliver a service AND monetize it, the technology stack, the scaling, even the tech they acquire to build it, just totally different.

Again, if you can find an article that had someone in the 90s or 2000s saying "Google is DOA! Search is stupid!" then I'd really really love to read it, genuinely.


> However, I do have the ability to read publicly-available data,

Maybe, but based on the egregious errors the author has made in previous articles, they probably don't have the ability to understand or reason about any of the data they read. Also note that despite what's implied by this statement, most of this article is not sourced, it's just the opinions of the author who admits they have no qualifications.

I didn't read the entire gish gallop, but spot-checked a few paragraphs here and there. It's just the kind of innumerate tripe that you should expect from Zitron based on their past performance.

> Have a significant technological breakthrough such that it reduces the costs of building and operating GPT — or whatever model that succeeds it — by a factor of thousands of percent.

You can't reduce the cost of anything by more than 100%. At that point it's free.

But let's consider the author's own numbers: $4B in revenue, $4B in serving costs, $3B in training costs, $1.5B in payroll. To break even at the current revenue, OpenAI need to cut their serving costs and training costs by about 66% ($1.3B+$1B+$1.5B<$4B), not by "thousands of percent".

> As a result, OpenAI's revenue might climb, but it's likely going to climb by reducing the cost of its services rather than its own operating costs.

... Sorry, what?

Reducing operating costs does not increase revenue. And I don't know how the author thinks that reducing cost of services would not reduce operating costs.

> OpenAI's only real options are to reduce costs or the price of its offerings. It has not succeeded in reducing costs so far, and reducing prices would only increase costs.

Reducing prices does not increase costs.

> I see no signs that the transformer-based architecture can do significantly more than it currently does.

So, here's a prime example of the author basing the "analysis" on them personally "seeing no signs" of something they have no expertise to evaluate. There's no source for this claim, and it's pretty crucial for their conclusions that transformers have hit a wall.

> While there may be ways to reduce the costs of transformer-based models, the level of cost-reduction would be unprecedented,

But for a given quality of model, haven't the inference costs already gone down by like 90% this year?

> particularly from companies like Google, which saw its emissions increase by 48% in the last five years thanks to AI.

It should be pretty obvious to somebody who can read publicly available data that all of the increase over 5 years can't be attributed to AI.


Hi! :) It appears you have some issues with my article, and I'm happy to provide some help.

"Egregious errors in previous articles" is not a valid argument against current arguments, nor do I agree there were those errors. Nevertheless, we're discussing one particular article today!

"I didn't read the entire gish gallop, but spot-checked a few paragraphs here and there. It's just the kind of innumerate tripe that you should expect from Zitron based on their past performance."

Well that's not very nice! It also means that your argument is made on incomplete data.

"... Sorry, what?

Reducing operating costs does not increase revenue. And I don't know how the author thinks that reducing cost of services would not reduce operating costs."

I'm afraid you misread what I said, likely because you (and I quote) "spot-checked a few paragraphs."

One of the problems OpenAI has is that their cost of revenue - and we don't know it to be exact - is extremely high, higher than the revenue they're actually gaining, otherwise known as an "operating loss." As a result, even if they increase revenue, they'll actually lose more money. On top of that, the argument I was making is that if there's a race to the bottom (one that's already started), they will have to cut costs, making them less money even if they get more customers.

"Reducing prices does not increase costs." Does reducing prices reduce operating expenses? Because if it doesn't, it actually does increase costs, because you're taking home less cash for the same cost. It could be that 4oMini is somehow more efficient - i can find no evidence that that's the case, and if it exists, I will happily update my article.

"So, here's a prime example of the author basing the "analysis" on them personally "seeing no signs" of something they have no expertise to evaluate. There's no source for this claim, and it's pretty crucial for their conclusions that transformers have hit a wall."

I can find no examples of radically-different functionality in GPT or other mass-market transformer-based models. In the event I am wrong, I would be fascinated to read about them, but I would need to understand A) how these functionalities are different and B) how they can be productized. After that, I'd need to understand how this would be profitable, and in turn how this would scale into something truly world-changing.

"But for a given quality of model, haven't the inference costs already gone down by like 90% this year?" Have they?

"It should be pretty obvious to somebody who can read publicly available data that all of the increase over 5 years can't be attributed to AI."

I too read publicly-available data, and my source in this case is "Google."

Forgive the messy copy-paste. https://www.gstatic.com/gumdrop/sustainability/google-2024-e...

In 2023, our total GHG emissions were 14.3 million tCO2e, representing a 13% year-overyear increase and a 48% increase compared to our 2019 target base year. This result was primarily due to increases in data center energy consumption and supply chain emissions. As we further integrate AI into our products, reducing emissions may be challenging due to increasing energy demands from the greater intensity of AI compute, and the emissions associated with the expected increases in our technical infrastructure investment.


Your track record is pretty relevant when evaluating this kind of stream of consciousness writing. You have a process, and that process is going to produce a consistent quality. And your track record us one of misrepresenting sources, making trivial logic errors, having a total inanbility to read financial statements correctly, and in general having no expertise in any of the things you write about. (Or rather, any of the things you write that surface on this site.)

> I'm afraid you misread what I said, likely because you (and I quote) "spot-checked a few paragraphs."

I quoted what you wrote, it wasn't out of context, and it was obvious nonsense. That you can't catch such obvious nonsense is exactly why nothing you write can be trusted.

> One of the problems OpenAI has is that their cost of revenue - and we don't know it to be exact - is extremely high, higher than the revenue they're actually gaining, otherwise known as an "operating loss." As a result, even if they increase revenue, they'll actually lose more money. On top of that, the argument I was making is that if there's a race to the bottom (one that's already started), they will have to cut costs, making them less money even if they get more customers.

None of that seems to bear any relation to what you actually wrote: "As a result, OpenAI's revenue might climb, but it's likely going to climb by reducing the cost of its services rather than its own operating costs". That is you claiming that reducing the cost of its services would increase revenue.

That is not you talking about operating income, or margin, or cost of revenue. These words have actual meaning, you can't just randomly one for another and expect it to make sense. Again, a recurring pattern.

> I too read publicly-available data, and my source in this case is "Google."

Yes, you already bragged in the article that you know how to read publicly available data, which is why that's the qualifier I used. I don't dispute that you're able to read. I will, however, claim that you either do not understand much what you read or are intentionally choosing to misrepresent that. Let's look at this example:

> In 2023, our total GHG emissions were 14.3 million tCO2e, representing a 13% year-overyear increase and a 48% increase compared to our 2019 target base year. This result was primarily due to increases in data center energy consumption and supply chain emissions. As we further integrate AI into our products, reducing emissions may be challenging due to increasing energy demands from the greater intensity of AI compute, and the emissions associated with the expected increases in our technical infrastructure investment.

What part of that supports your claim of AI being the cause of the 48% increase? None of it. It is only attributed to "supply chain emissions" and "data center energy consumption". The mention of AI is entirely forward-looking. Let's take it for granted that you indeed read the text you copy-pasted. Why is your claim about what it says so obviously incorrect?

Did you really not understand the text? It's not that complex. Did you understand it and just lie about it because it supported the narrative you had in mind, and nobody checks the sources anyway? Seems like a bad plan. Either way, it again demonstrates that you are not cut out for doing any kind of analysis.


Maybe GPT-5 or Claude 4 comes out tomorrow and pays down all the IOUs these people are operating on.

As of today all of the evidence indicates the LLM paradigm is saturated.

Why all the hand-wringing about hypotheticals? As of today this stuff is a failed experiment.

Altman or Amodei coming up with the goods is a tail X-risk.


See, that's the thing. This COULD happen. I do not think it will, but the piece is written not to say "this is 100% not gonna happen," but "I am unsure how it's possible, and if it does not happen, the consequences could be dire."

Hypotheses and hypotheticals are useful tools when writing about something big and messy. Instead of me saying - as I have before - that I believe generative AI is a complete dead end and thus OpenAI is in a really bad way - I took great pains to explain the terms under which they WOULD succeed - how difficult success might be, how much money it would take and how many factors would have to go their way.

If OpenAI pulls it off, it'd be really remarkable. Truly historic! But if they don't, they are in deep, deep doo doo.


Maybe they pivot to blockchain…. lol


[redacted]


> OpenAI is currently working on the next step: agents and planning. They are close. Other competitors will shortly follow. Their stocks will rise to unprecedented levels. You have been warned.

Any day now everyone will adopt bitcoin and the value will skyrocket and we’ll all be exceedingly rich, HODL, gl, wgtm, to the moon.

Not saying they’re not working on stuff, but that’s what you sound like. OpenAI is amazing at generating hype, but the competitors are catching up, or outpacing them, and OpenAI source models are never far behind.


What kind of agents?


https://aibusiness.com/nlp/openai-is-developing-ai-agents probably. But I'm not sure how much they need to innovate on that front. There's good agents progress happening everywhere around already. OpenAI could provide some options to make it faster/cheaper, and running the workflow internally would make it faster, but... doesn't feel necessary.


> AI agents differ from robotic process automation (RPA), which still needs developers to manually code steps needed to complete a task, Luan said.

OK, so it's robotic process automation then (aka screen scraping / UI automation), just without the hand-written automation code.

My frustration with the term "agents" is that it can mean a lot of different things - but people always seem to assume that their own personal definition is the same that everyone else is using.



> generative AI is a product with no mass-market utility

Citation?

LLMs bring the cost of writing software close to $0. We can finally live in a world of truly bespoke code.

I, for one, welcome back the web of the early 2000s.


Hi! I'm afraid you've made an assumption that isn't true.

LLMs do not "bring the cost of writing software close to $0" on a number of levels.

1. The code is not 100% reliable, meaning it requires human oversight, and human beings cost money. 2. LLMs themselves are not cheap, nor profitable. I am comfortable humoring the idea that someone could run their own models - something which is beginning to happen - to write code. I think that's really cool, but I am also not sure how good said code will be or how practical doing so will be.

Right now, Microsoft is effectively subsidizing the cost of Github Copilot, though they appear to have produced quite a lot of revenue from it.

https://www.benzinga.com/news/24/07/40061358/satya-nadella-s...

However, it seems that Github was not profitable before (https://news.ycombinator.com/item?id=17224136) and I would argue isn't profitable now. It's hard to tell, because Microsoft blends their costs into other business lines.


> Hi! I'm afraid you've made an assumption that isn't true.

Citation?

Your concerns are certainly valid, but the LLMs are getting smaller, faster, and cheaper to run every day. Now, I also agree that you still need someone "programming" -- in the sense that they're telling a computer what to do, but they no longer need to "code" in the traditional sense (curly braces and semicolons).

We're actively seeing non-engineers build useful software for themselves, just with a $20/month subscription to ChatGPT/Claude.

Times are changing, you no longer need a 6 figure engineer to build your one-off tool.


LLMs are not making the cost of software near-zero, even by your admission!

If times are changing FOR GOOD, perhaps you're right? But i'm not sure they are.


We’re saying the same thing, but from different angles.

LLMs won’t build Google from scratch, we’ll still need human software engineers.

LLMs will enable an entirely new class of ”programmer” to exist. Writing software still requires a technical mindset, but you no longer need a 4-year degree (or bootcamp) to build and deploy a personal web app.

The cost of writing software is approaching $0, as you will no longer need to hire a developer for your bespoke requirement — how valuable that software is depends entirely on the end user.

Programmers in the 70s thought C was silly; how could you trust an abstraction over actual assembly instructions?

We’re seeing the same shift here, but accelerated to a much larger group of potential “programmer” recruits.

Again, we’re seeing this pop up almost every day across Reddit[0], Discord, and various forums.

Non-developers are already writing software entirely with ChatGPT.

[0] https://www.reddit.com/r/ChatGPTCoding/s/sZmeQV6yNu


I was assured that being skeptical about offloading the use of my brain to a third party that might not always be there was being a luddite.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: