Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Llama.cpp allocates stuff to the GPU statically. It'd not really analogous to a game.

It should have a heuristic that looks at available VRAM by default, but it does not. Probably because this is vendor specific and harder than you would think, and they would rather not use external libraries.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: