Hacker News new | past | comments | ask | show | jobs | submit login

I suspect you would fix first response accuracy.

People still want it to be used for thinking.

This isnt going to happen with better data. Better data means it will be better at predicting the next token.

For questions or interactions where you need to process, consider, decompose a problem into multiple steps, solve those steps etc - you need to have a goal, tools, and the ability to split your thinking and govern the outcome.

That isnt predicting the next token. I think it’s easier to think of LLMs as doing decompression.

They take an initial set of tokens and decompress them into the most likely final set of tokens.

What we want is processing.

We would have to set up the reaction to somehow perfectly result in the next set of tokens to then set up the next set of tokens etc - till the system has an answer.

Or in other words, we have to figure out how to phrase an initial set of tokens so that each subsequent set looks similar enough to “logic” in the training data, that the LLM expands correctly.




Consider applying for YC's Summer 2025 batch! Applications are open till May 13

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: