Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Distillation would be the ideal way (especially because it also has efficiency gains), but as far as I know distillation for LLMs is kinda unproven.

Honestly though, even if you just finetune it, which you will want anyway for any serious commercial application, it's essentially impossible to determine the origin.



Randomly perturbing the weights and then finetuning would probably make it impossible. If someone had access to the finetune dataset and you didn’t add noise, they could see if the finetuning curves intersect.

I guess in practice, it’ll look suspicious if you have an identical model architecture and have similar performance.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: