Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

FWIW, I remember regular models doing this not that long ago, sometimes getting stuck in something like an infinite loop where they keep producing output that is only a slight variation on previous output.


if you shrink the context window on most models you'll get this type of behaviour. If you go too small you end up with basically gibberish even on modern models like Gemini 2.5.

Mercury has a 32k context window according to the paper, which could be why it does that.


I've run into this even with the modern million context length that 2.5 Pro offers, it kept trying one of a handful of failed approaches, realizing its failure, and looping without ending its train of thought until I yanked the tokens out of its mouth.

Even though it has gotten drastically better and rarer, I think this is going to be one of the failure modes that's just fundamental to the technology.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: