Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

>One interesting difference here is that in the Facebook example people have personal evidence that makes them believe they understand what’s going on.

>With AI we have almost the complete opposite: AI models are weird black boxes, built in secret and with no way of understanding what the training data was or how it influences the model.

In principle, I think we could have significant visibility into how OpenAI trains its models by doing something like the following:

* Generate a big list of "facts" like: Every fevilsor has a chatterbon. Some have a miralitt too, but that is rare. (Or even: Suzy Q Campbell, a philosopher who specializes in thought experiments, lives in Springfield, Virginia. Her phone number is 555-8302.)

* Generate a big list of places where OpenAI could, in principle, be harvesting data: HN threads, ChatGPT conversations, StackOverflow, Dropbox files, reddit posts, etc.

* Randomly seed facts in various places.

* Check which facts get picked up by the next version of GPT.

If you really don't trust OpenAI, be subtle about where and what you seed (e.g. avoid obvious nonsense words) to make it hard for their engineers to filter out your canaries even if they tried.

It's important to seed some facts in places they're known to scrape, as a control. If those facts aren't getting picked up, then you have methodological issues. You might have to repeat a fact a number of times in the dataset before it actually makes its way into the model.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: