Hacker News new | past | comments | ask | show | jobs | submit login

> I desire the ability to filter the content I consume on "generated by AI". The value of that, to me, is self-evident.

You should vote with your wallet and only patronize businesses that self disclose. You don't need to create regulation to achieve this.

With regards to the articles, they are entirely speculative, and I diaagree wholly with them, primarily because their premise is that humans are not rational amd discerning actors. The only way AI generates chaos in these instances is by generating so much noise as to make online discussions worthless. People will migrate to closed communities of personal or near personal acquaintances (web of trust like) or to meatspace.

Here are some paragrahs I fpund especially egregious:

> In recent years the qAnon cult has coalesced around anonymous online messages, known as “q drops”. Followers collected, revered and interpreted these q drops as a sacred text. While to the best of our knowledge all previous q drops were composed by humans, and bots merely helped disseminate them, in future we might see the first cults in history whose revered texts were written by a non-human intelligence. Religions throughout history have claimed a non-human source for their holy books. Soon that might be a reality.

Dumb people will dumb. People with different values will different. I see no reason that AI offers increased risk to cult followers of Q. If someone isn't going to take the time to validate their sources, the source doesn't t much matter.

> On a more prosaic level, we might soon find ourselves conducting lengthy online discussions about abortion, climate change or the Russian invasion of Ukraine with entities that we think are humans—but are actually ai. The catch is that it is utterly pointless for us to spend time trying to change the declared opinions of an ai bot, while the ai could hone its messages so precisely that it stands a good chance of influencing us.

In these instances, does it mayter that the discussion is being held with AI? Half the use of discussion is to refine one's own viewpoints by having to articulate one's position and think through cause and effect of proposals.

> The most interesting thing about this episode was not Mr Lemoine’s claim, which was probably false. Rather, it was his willingness to risk his lucrative job for the sake of the ai chatbot. If ai can influence people to risk their jobs for it, what else could it induce them to do?

Intimacy isn't necessarily the driver for this. It very well could have been Lemoine's desire to be first to market that motivated the claim, or a simple misinterpreted singal al la Luk-99.

> Even without creating “fake intimacy”, the new ai tools would have an immense influence on our opinions and worldviews. People may come to use a single ai adviser as a one-stop, all-knowing oracle. No wonder Google is terrified. Why bother searching, when I can just ask the oracle? The news and advertising industries should also be terrified. Why read a newspaper when I can just ask the oracle to tell me the latest news? And what’s the purpose of advertisements, when I can just ask the oracle to tell me what to buy?

Akin to the concerns of scribes during the times of the printing press. The market will more efficiently reallocate these workers. Or better yet, people may still choose to search to validate the output of a statistical model. Seems likely to me.

> We can still regulate the new ai tools, but we must act quickly. Whereas nukes cannot invent more powerful nukes, ai can make exponentially more powerful ai. The first crucial step is to demand rigorous safety checks before powerful ai tools are released into the public domain.

Now we get to the point: please regulate me harder. What's to stop a more powerful AI from corrupting the minds of the legislative body through intimacy or other nonsense? Once it is sentient, it's too late, right? So we need to prohibit people from multiplying matrices without government approval right now. This is just a pathetic hit piece to sway public opinion to get barriers of entry erected to protect companies like OpenAI.

Markets are free. Let people consume what they want so long as there isnt an involuntary externality, and conversing with anons on the web does not guarantee that you're speaking with a human. Both of us could be bots. It doesn't matter. Either our opinions will be refined internally, we will make points to influence the other, or we will take up some bytes in Dang's database with no other impact.




> You should vote with your wallet and only patronize businesses that self disclose. You don't need to create regulation to achieve this.

This is a fantasy. It seems very likely to me that, sans regulation, the market utopia you describe will never appear.

I am not entirely convinced by the arguments in the linked opinions either. However, I do agree with the main thrust that (1) machines that are indistinguishable from humans are a novel and serious issue, and (2) without some kind of consumer protections or guardrails things will go horribly wrong.


> This is a fantasy. It seems very likely to me that, sans regulation, the market utopia you describe will never appear.

I strongly disagree. I heard the same arguments about how Google needs regulation because nobody could possibly compete. A few years later we have DDG, Brave Search, Searx, etc.


You mean the market will sacrifice people in order to optimize!?!?!?!

say it aint so bobby, say it aint so!


There are no machines than are indistinguishable from humans. That is science fiction.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: