Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> Poor Sam Altman, 300B worth of trade secrets bought out from under him for a paltry few hundred million

Sorry, you don't lose people when you treat them well. Add to that Altman's penchant for organisational dysfunction and the (in part resulting) illiquidity of OpenAI's employees' equity-not-equity and this makes a lot of sense. Broadly, it's good for the American AI ecosystem for this competition for talent to exist.



In retrospect, I wonder if the original ethos of the non-profit structure of OpenAI was a scam from the get go, or just woefully naive. And to emphasize, I'm not talking just about Altman.

That is, when you create this cutting edge, powerful tech, it turns out that people are willing to pay gobs of money for it. So if somehow OpenAI had managed to stay as a non-profit (let's pretend training didn't cost a bajillion dollars), they still would have lost all of their top engineers to deeper pockets if they didn't pursue an aggressive monetization strategy.

That's why I want to gag a little when I hear all this flowery language about how AI will cure all these diseases and be a huge boon to humanity. Let's get real - people are so hyped about this because they believe it will make them rich. And it most likely will, and to be clear, I don't blame them. The only thing I blame folks for is trying to wrap "I'd like to get rich" goals in moralistic BS.


It wasn't exactly a scam, it's just nobody thought it'd be worth real money that fast, so the transition from noble venture to cash grab happened faster than expected.


> wonder if the original ethos of the non-profit structure of OpenAI was a scam from the get go, or just woefully naive

Based on behaviour, it appears they didn't think they'd do anything impactful. When OpenAI accidentally created something important Altman immediately (a) actually got involved to (b) reverse course.

> if somehow OpenAI had managed to stay as a non-profit (let's pretend training didn't cost a bajillion dollars), they still would have lost all of their top engineers to deeper pockets if they didn't pursue an aggressive monetization strategy

I'm not so sure. OpenAI would have held a unique position as both first mover and moral arbiter. That's a powerful place to be, albeit not a position Silicon Valley is comfortable or competent in.

I'm also not sure pursuing monetisation requires a for-profit structure. That's more a function of the cost of training, though again, a licensing partnership with, I don't know, Microsoft, would alleviate that pressure without requiring giving up control.


Getting rich going good is better than just getting rich. People like both.

Which part are you skeptical about? that people also like to do good, or that AI can do good?


I'm skeptical that OpenAI was ever feasible as a nonprofit under it's original mission, which was:

> Our goal is to advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return. Since our research is free from financial obligations, we can better focus on a positive human impact.

As soon as the power of AI became apparent, everyone wanted (and in some ways, needed) to make bank. This would have been true even if the original training costs weren't so high.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: