I personally think it'll be fine. FSD is perfect for all my driving. It's been 3000-4000 miles since my last takeover. I know an uber driver who's driven his 16k miles without a safety intervention. And this is without them superfinetuning and doing custom navigation/mapping on a specific town.
I'm a huge proponent of e2e learning for robotics (worked at two places doing e2e before Tesla adopted it) and personally believe its the right approach long term. I also have FSD on my Model 3 and love it for L2+. That said, my experience with disengagements is very different than yours.. I have a few a week for things like road works, school zones, route map following. Perfectly fine for L2+, L4 it would be unacceptable.
If these robotaxis end up looking more like my experience than yours then another layer of trouble will be root causing and fixing failure modes. Training models e2e makes both of these much more difficult.
You could argue its not necessary to achieve performance for an L2+ product and so keeps BOM cost down while still achieving goals. I'm not personally opposed though, the systems I worked on did have other sensors we could use.
That wouldn't resolve the concern around debugging/root-causing and remediating failures more quickly though. You still have a black box system that is difficult to simulate closed loop.
This is a tiny rocket going up 300ms and coming back down. Happy for them but they're a long way from any utility ( and a decade+ behind other companies )
I think the DC-X program did this first in the 90s. It ran into funding issues and it turns out there isn't a ton of value in reusable rockets that only go a few hundred feet (although more advanced applications are potentially worthwhile)
It's not a difficult problem. It's just Newtonian mechanics plus control theory. You only need a lot of funding and then just do it (of course build a simulator first).
If if they do what they say they are going to --- even if they steal all of Uber and Waymo's business --- the current price of TSLA stock still can't be justified.
Uber is just minimally profitable --- without investing large sums in autos or assuming liability for their operation. No way Telsa can do so, undercut them on price and be crazy profitable all at the same time.
We aren't in a position to answer this one one way or the other.
If things go very wrong they have the potential to take us out. But a non-nuclear WWIII could, also--not by direct kills but by taking down the interconnected stuff that makes society work.
Also, while they serve to prevent direct wars between major powers they cause proxy wars between the major powers.
Relative to the hype they've been spinning to attract investment, casting the launch and commercialization of ChatGPT as their greatest achievement really is a quite significant downgrade, especially given that they really only got there first because they were the first entity reckless enough to deploy such a tool to the public.
It's easy to forget what smart, connected people were saying about how AI would evolve by <current date> ~a year ago, when in fact what we've gotten since then is a whole bunch of diminishing returns and increasingly sketchy benchmark shenanigans. I have no idea when a real AGI breakthrough will happen, but if you're a person who wants it to happen (I am not), you have to admit to yourself that the last year or so has been disappointing---even if you won't admit it to anybody else.
ChatGPT was released two and a half years ago though. Pretty sure that at some point Sam Altman had promised us AGI by now.
The person you're responding to is correct that OpenAI feels a lot more stagnant than other players (like Google, which was nowhere to be seen even one year and a half ago and now has the leading model on pretty much every metric, but also DeepSeek, who built a competitive model in a year that runs for much cheaper).
o3-mini wasn't even the second place for non-STEM tasks, and in today's announcement they don't even publish benchmarks for those. What's impressive about Gemini 2.5 pro (and was also really impressive with R1) is how good the model is for a very broad range of tasks, not just benchmaxing on AIME.
I had a philosophical discussion with o3 model earlier today. It was much better than 2.5 pro. In fact it was pretty much what I would expect from a professional philosopher.
Writing philosophy that looks convincing has been a thing LLM do well since the first release ChatGPT back in 2022 (in my country back in early 2023, TV featured a kind of competition between ChatGPT and a philosopher turned media personality, with university professors blindly reviewing both essays and attempting to determine which was whom).
To have an idea about how good a model is on non-STEM tasks, you need to challenge it on stuff that is harder than this for LLMs, like summarization without hallucination or creative writing. OpenAI's nonthinking model are usually very good on these, but not their thinking models, whereas other players (be it Google, Anthropic or DeepSeek) manage to make models that can be very good at both.
I've been discussing a philosophical topic (brain uploading) with all major models in the last two years. This is a topic I've read and thought about for a long time. Until o3, the responses I got from all other models (Gemini 2.5 pro most recently) have been underwhelming - generic, high level, not interesting to an expert. They struggled to understand the points I was making, and ideas I wanted to explore. o3 was the first model that could keep up, and provide interesting insights. It was communicating on a level of a professional in the field, though not an expert on this particular topic - this is a significant improvement over all existing models.
The play now seems to be less AGI, more "too big to fail" / use all the capital to morph into a FAANG bigtech.
My bet is that they'll develop a suite of office tools that leverage their model, chat/communication tools, a browser, and perhaps a device.
They're going to try to turn into Google (with maybe a bit of Apple and Meta) before Google turns into them.
Near-term, I don't see late stage investors as recouping their investment. But in time, this may work out well for them. There's a tremendous amount of inefficiency and lack of competition amongst the big tech players. They've been so large that nobody else could effectively challenge them. Now there's a "startup" with enough capital to start eating into big tech's more profitable business lines.
I don't know how anyone could look at any of this and say ponderously: it's basically the same as Nov 2022 ChatGPT. Thus strategically they're pivoting to social to become too big to fail.
I mean, it's not fucking AGI/ASI. No amount of LLM flip floppery is going to get us terminators.
If this starts looking differently and the pace picks up, I won't be giving analysis on OpenAI anymore. I'll start packing for the hills.
But to OpenAI's credit, I also don't see how minting another FAANG isn't an incredible achievement. Like - wow - this tech giant was willed into existence. Can't we marvel at that a little bit without worrying about LLMs doing our taxes?
I'm bullish on the models, and my first quiet 5 minutes after the announcement was spent thinking how many of the people I walked past days would be different if the computer Just Did It(tm) (I don't think their day would be different, so I'm not bullish on ASI-even-if-achieved, I guess?)
I think binary analysis that flips between "this is a propped up failure, like when banks get bailouts" and "I'd run away from civilization" isn't really worth much.
Most people don't care about techies or tech drama. They just use the platforms their friends do.
ChatGPT images are the biggest thing on social media right now. My wife is turning photos of our dogs into people. There's a new GPT4o meme trending on TikTok every day. Using GPT4o as the basis of a social media network could be just the kickstart a new social media platform needs.
chatGPT should be built into my iMessage threads with friends. @chatGPT "Is there an evening train on Thursdays from Brussels to Berlin?" Something a friend and I were discussing but we had to exit out of iMessage and use GPT then back to iMessage.
For UX The GPT info in the thread would be collapsed by default and both users have the discretion to click to expand the info.