Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Sounds like the type of prompt that would boldly give you a wrong/illegal answer.


Perhaps. But the point is that some prompt will coax it into giving good answers that really make it win the game, if it has a good "world model" of how the game works. And there's no reason to think a language model cannot have such a world model. What exactly that prompt might be, the prompt engineers know best.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: