The power openai will hold above everyone else is just too much. They will not allow their AI as a service without data collection. That will be a big pill to swallow for the EU.
Well, it's possible to detect patterns and characteristics in the language used in the comments that can provide clues about their origin...
Here's some indicators that a comment may have been generated by an AI system:
* Repeating phrases or sentences
* Using generic language that could apply to any topic
* Lack of coherence or logical flow
* Poor grammar, or syntax errors
* Overuse of technical, or specialized vocabulary
I mean, these indicators aren't foolproof... and humans can also exhibit some of these characteristics. It's tough to be sure whether or not a comment is generated by an AI system or not...
It's funny, just two hours ago there was a thread by a pundit arguing that these AI advances don't actually give the companies producing them a competitive moat, because it's actually very easy for other models to "catch up" once you can use the API to produce lots of training examples.
Almost every answer in the thread was "this guy isn't that smart, this is obvious, everybody knew that", even though comments like the above are commonplace.
FWIW I agree with the "no competitive moat" perspective. OpenAI even released open-source benchmarks, and is collecting open-source prompts. There are efforts like Open-Assistant to create independent open-source prompt databases. Competitors will catch up in a matter of years.
Years? There are already competitors. I just spent all evening playing with Claude (https://poe.com/claude) and it's better than davinci-003.
To be fair it is easy to radically underestimate the rate of progress in this space. Last Wednesday I conservatively opined to a friend "in 10 years we'll all be running these things on our phones". Given that LLaMA was running on a phone a few days later, I may have been a little underoptimistic...
Obviously the larger models won't run on such limited hardware (yet) but one of the next big projects (that I can see) being worked on is converting the models to be 3bit (currently 8bit and 4bit are popular) which cuts down required resources drastically with minimal noticeable loss in quality.
I think starting with FlexGen barely 4 weeks ago, there have been some pretty crazy LLM projects/forks popping up on github almost daily. With FlexGen I felt like I was still able to stay up-to-date but I'm getting close to giving up trying as things are moving exponentially faster... you know it's crazy when a ton of noobs who have never heard of conda are getting this stuff running (sometimes coming in flexgen discord or posting github issues to get help, though even those are becoming rarer as one-click-installer's are becoming a thing for some popular ML tools, such as oobabooga's amazing webui tool which has managed to integrate almost all the hottest new feature forks fairly quickly: https://github.com/oobabooga/text-generation-webui
I just helped someone recently get oobabooga running which has a --listen option to open the webui to your network, now he's running llama on his tablet (via his PC).
Yeah, there's an awful lot of power going into private hands here and as Facebook & Twitter have shown, there can be consequences of that for general society.
> Yeah, there's an awful lot of power going into private hands
That sounds scary, but what do you mean by "power"? Honest question, I'm fascinated by the discussion about learning, intelligence, reasoning, and so on that has been spawned by the success of GPT.
What "power" do you imagine being wielded? Do you think that power is any more dangerous in "private hands" than the alternatives such as government hands?
Do you think that Facebook has an effect on society and our democracies? That's power. Do you think that large corporates like Apple or Google effect our societies? I do - and that's power. EVERY large corporate has power and if they control some aspect of society, even more so. If AI tools are democratised in some way, then that would allay my concerns. Concentration of technology by for-profit corporations concerns me. This seems quite similar to many of the reasons people like OSS, for example. Maybe not for you?
OpenAI have been consistently ahead of everyone but the others are not far behind. Everyone is seeing the dollar signs, so I'm sure all big players are dedicating massive resources to create their own models.
Yes. Language and image models are fairly different, but when you look at dall-e 2 (and dall-e earlier) that blew many people's mind when they came out, they have now been really eclipsed in term of popularity by Midjourney and stablediffusion.
From what I've seen, the EU is not in the business of swallowing these types of pills. A multi-billion dollar fine? Sure. Letting a business dictate the terms of users' privacy just "because"? Not so much, thank god.