Hacker News new | past | comments | ask | show | jobs | submit login

to be fair, I think the ability of these models to actually use these contexts beyond the standard 8k / 16k tokens is pretty weak. RAG based methods are probably a better option for these ultra long contexts



Haystack testing on GPT-4's 128K context suggests otherwise: https://twitter.com/SteveMoraco/status/1727370446788530236


> I think the ability of these models to actually use these contexts beyond the standard 8k / 16k tokens is pretty weak.

For 32k GPT4 contexts, that's not accurate. GPT4 Turbo is a bit weaker than GPT4-32k, but not to the extent that you claim.


Are you talking about claude or Gpt4 as well? Anybspecific examples where ChatGPT4 fails for long contexts?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: