Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The wonderful thing about a sigmoid is that, just as it seems like it's going exponential, it goes back to linear. So I'd guess we're not going to see 1000x from here - I could be wrong, but I think the low hanging fruit has been picked. I would be surprised in 10 years if AI were 100x better than it is now (per watt, maybe, since energy devoted to computing is essentially the limiting factor)

The thing about the latter 1/3rd of a sigmoid curve is, you're still making good progress, it's just not easy any more. The returns have begun to diminish, and I do think you could argue that's already happening for LLMs.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: