So if we assume this is the future, the useful life of many semiconductors will fall substantially. What part of the semiconductor supply chain would have pricing power in a world of producing many more different designs?
It might be not that bad. “Good enough” open-weight models are almost there, the focus may shift to agentic workflows and effective prompting. The lifecycle of a model chip will be comparable to smartphones, getting longer and longer, with orchestration software being responsible for faster innovation cycles.
"Good enough" open weights models were "almost there" since 2022.
I distrust the notion. The bar of "good enough" seems to be bolted to "like today's frontier models", and frontier model performance only ever goes up.
I don’t see why. Today frontier models are already 2 generations ahead of good enough. For many users they did not offer substantial improvement, sometimes things got even worse. What is going to happen within 1 year that will make users desire something beyond already working solution? LLMs are reaching maturity faster than smartphones, which now are good enough to stay on the same model for at least 5-6 years.
Any considerable bump in model capability craters my willingness to tolerate the ineptitude of less capable models. And I'm far from being alone in this.
Ever wondered why those stupid "they secretly nerfed the model!" myths persist? Why users report that "model got dumber", even if benchmarks stay consistent, even if you're on the inference side yourself and know with certainty that they are actually being served the same inference over the same exact weights on the same hardware quantized the same way?
Because user demands rise over time, always.
Users get a new flashy model, and it impresses them. It can do things the old model couldn't. Then they push it, and learn its limitations and quirks as they use it. And then it feels like it "got dumber" - because they got more aggressive about using it, got better at spotting all the ways it was always dumb in.
It's a treadmill, and you pretty much have to keep improving the models just to stay ahead of user expectations.
I have seen this with ChatGPT progression from 4o to 5.2 applied to the newest model. Old prompts stop working reliably, different hallucination modes etc.
Different skills and context. Llama 3.1 8B has just 128k context length, so packing everything in it may be not a great idea. You may want one agent analyzing the requirements and designing architecture, one writing tests, another one writing implementation and the third one doing code review. With LLMs it’s also matters not just what you have in context, but also what is absent, so that model will not overthink it.
EDIT: just in case, I define agent as inference unit with specific preloaded context, in this case, at this speed they don’t have to be async - they may run in sequence in multiple iterations.
It will be validated but that doesn’t mean that the providers of these services will be making money. It’s about the demand at a profitable price. The uncontroversial part is that the demand exists at an unprofitable price.
This “It’s not about profits, man, it’s about how much you’re worth. The rules have changed. Don’t get left behind,” nonsense is exactly what a bunch of super wrong people said about investing during the .com bust. Even if we got some useful tech out of it in the end, that was a lot of people’s money that got flushed down the toilet.
But the survivors became some of the biggest and most profitable companies on the planet: Google, Amazon, Ebay/Paypal. And of course, the people selling shovels always do well in a rush (Apple, Adobe, etc).
I’m not talking about the health of the the industry— I’m talking about the fallout for employees, anyone with any stake in the stock market, etc. A whole lot of retail investors, 401k holders, etc. got fucked, and a whole lot of other people lost their jobs. Careers were stunted. This was before we had preexisting condition protection so for people with cancer or other serious chronic health conditions, losing a job could be a death sentence, even if they got another job the very next day. The housing market got screwed up.
From the big short (and a bunch of introductory macroeconomics classes:)
"For every 1% that unemployment rises, 40,000 people die."
There are consequences to people running big companies like they’re playing poker.
It’s not clear at all because model training upfront costs and how you depreciate them are big unknowns, even for deprecated models. See my last comment for a bit more detail.
They are obviously losing money on training. I think they are selling inference for less than what it costs to serve these tokens.
That really matters. If they are making a margin on inference they could conceivably break even no matter how expensive training is, provided they sign up enough paying customers.
If they lose money on every paying customer then building great products that customers want to pay for them will just make their financial situation worse.
I think actually working out whether they are losing money is extremely difficult for current models but you can look backwards. The big uncertainties are:
1) how do you depreciate a new model? What is its useful life? (Only know this once you deprecate it)
2) how do you depreciate your hardware over the period you trained this model? Another big unknown and not known until you finally write the hardware off.
The easy thing to calculate is whether you are making money actually serving the model. And the answer is almost certainly yes they are making money from this perspective, but that’s missing a large part of the cost and is therefore wrong.
On a tangent to the article- I quit my career just over two years ago now: same age as author, live in London too. The hardest thing about not working is the social life that work gives you. Whilst we may think that work is for money, it is also for 1) filling our time, and 2) spending time with people. Yes, some people are definitely a net-negative interaction, but most people are actually positive to one’s day, but in one of those “you need to not work for a year to know it” way.
Amongst other reflections I have:
1) a pay-check does give you a sense of validation. This took some getting through
2) it’s been challenging working out what I will actually end up doing with myself. There were periods where I put more pressure on myself to do so. I still don’t know what will do.
3) the process of doing things because they are fun takes some getting used to when one’s entire life was built around doing something useful to others
4) when one lives off of savings it’s almost easier to spend as it feels like you didn’t suffer for it. Getting depressed at work makes it easier to spend more money outside of work
5) the “number” people need to retire (or not work for extended periods) is probably less than people realise
6) not working in finance (amongst all the moral corruption everywhere) has generally made me happier in part because I can live in a way which is more in-keeping with my values over having constantly breach them for work reasons
7) owning my calendar is a big freedom. I don’t have to ask a boss if I can do something all the time. No need to explain yourself.
8) not constantly having to submit to a boss is huge. One can really grow this way, as constant repression to other people’s whims is soul crushing and shows just how close employment is to slavery (especially in finance with golden handcuffs)
Wow, retired in London. I don't think I'll ever afford to retire just living in the exurbs of the Midwest US. Kids and a spouse can push out the retirement goal considerably. Then there are medical issues and the extended family you'd hate to see on the street ...
If you exclude mortgage payments or rent then living in London is probably cheaper than living in most places in the US. We don’t have significant real estate taxes, my annual expenses on housing exc rent or mortgage payments (energy, service charge, internet, water, council taxes) is around 7-8k$. If you want health insurance it’s another 200-300$ a month but we have the nhs (albeit it is terrible for anything non-life threatening).
Would a human perform very differently? A human who must obey orders (like maybe they are paid to follow the prompt). With some "magnitude of work" enforced at each step.
I'm not sure there's much to learn here, besides it's kinda fun, since no real human was forced to suffer through this exercise on the implementor side.
How useful is the comparison with the worst human results? Which are often due to process rather than the people involved.
You can improve processes and teach the humans. The junior will become a senior, in time. If the processes and the company are bad, what's the point of using such a context to compare human and AI outputs? The context is too random and unpredictable. Even if you find out AI or some humans are better in such a bad context, what of it? The priority would be to improve the process first for best gains.
Just as enterprise software is proof positive of no intelligence under the hood.
I don't mean the code producers, I mean the enterprise itself is not intelligent yet it (the enterprise) is described as developing the software. And it behaves exactly like this, right down to deeply enjoying inflicting bad development/software metrics (aka BD/SM) on itself, inevitably resulting in:
Well… it’s more a great example that great output is a good model with the right context at the right time.
Take away everything else, there’s a product that is really good at small tasks, it doesn’t mean that changing those small tasks together to make a big task should work.
They get to look good by claiming it’s an ethical stance.
reply