Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I don’t get the point of abliteration of LLMs. You’re lobotomizing the model and it will result in worse performance.

If you’re doing it to get past refusals you might discover the LLM wasn’t even trained much on refusable content so it will output poor results.

We’ll look back on this practice and shake our heads someday.



Not necessarily true, one quick pass might be needed but quite not as devastating as it might seem

https://huggingface.co/blog/mlabonne/abliteration#%E2%9A%96%...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: