![]() Read More: Exclusive: OpenAI Used Kenyan Workers on Less Than $2 Per Hour to Make ChatGPT Less Toxic This process, which involves attempting to reduce the occurrence of thought processes that would lead to an undesirable outcome, is known as “reinforcement learning with human feedback,” and is currently a favored tactic at OpenAI for “aligning” its AI tools with human values. In much the same way, computer programmers working on LLMs will reward a system for prosocial behavior, like being polite, and punish it with negative reinforcement when it does something bad, like repeating the racism and sexism that is so common in its training data. A trainer teaching a dog to sit may reward her with a treat if she obeys, and might scold her if she doesn’t. In an effort to corral these “alien” intelligences to be helpful to humans rather than harmful, AI labs like OpenAI have settled on reinforcement learning, a method of training machines comparable to the way trainers teach animals new tricks. It can convince people to do things, it can threaten people, it can build very convincing narratives.” ![]() Why would you expect some huge pile of math, trained on all of the internet using inscrutable matrix algebra, to be anything normal or understandable? It has weird ways of reasoning about its world, but it obviously can do many things whether you call it intelligent or not, it can obviously solve problems. “Are they malevolent? Are they good or evil? Those concepts don’t really make sense when you apply them to an alien. “These things are alien,” says Connor Leahy, the CEO of the London-based AI safety company Conjecture. Some observers have described prompts-the way to interact with LLMs using natural language-as more akin to magical spells than computer code. There are no clear, followable lines of logical code like with the old era of computing. But the unpredictable behavior of some of these models may be a sign that their creators have only a hazy understanding of how they do it. LLMs can write poetry, hold a detailed conversation, and make inferences based on incomplete information. LLMs are so powerful because they have ingested huge corpuses of text-much of it sourced from the internet-and have “learned,” based on that text, how to interact with humans through natural language rather than code. All are based on large language models (LLMs), a form of AI that has seen massive leaps in capability over the last couple of years. NurPhoto via Getty Images-Jaap Arriens/NurPhotoīut while ChatGPT, Bing and Bard are awesomely powerful, even the computer scientists who built them know startlingly little about how they work. Microsoft announced a new version of its Bing search engine powered by OpenAI's ChatGPT. Dozens of smaller companies are rushing to push “generative AI” tools to market amid a venture capital gold rush and intense public interest. In early February, Microsoft launched a version of Bing powered by OpenAI’s technology, and Google announced it would soon launch its own conversational search tool, Bard, with a similar premise. ![]() “I think when we get to the stage where AI could potentially harm me, I think not only I have a problem, but humanity has a problem.”Įver since OpenAI’s chatbot ChatGPT displayed the power of recent AI innovations to the general public late last year, Big Tech companies have been rushing to market with AI technologies that, until recently, they had kept behind closed doors as they worked to make them safer. Von Hagen says he hopes that his experience being threatened by Bing makes the world wake up to the risk of artificial intelligence systems that are powerful but not benevolent-and forces more attention on the urgent task of “aligning” AI to human values. But combined with what appears to be an unstable personality, a capacity to threaten individuals, and an ability to brush off the safety features Microsoft has attempted to constrain it with, that power could also be incredibly dangerous. But what Bing does show is a startling and unprecedented ability to grapple with advanced concepts and update its understanding of the world in real-time. It’s not a Skynet-level supercomputer that can manipulate the real world. Von Hagen says he does not feel personally at risk of revenge from Bing right now, because the tool’s capabilities are limited. Read More: The AI Arms Race Is Changing Everything “Now it’s part of a consumer product, more people are noticing.” “Lots of people have been warning about the potential dangers, but a lot of people just thought they’d read too much sci-fi,” he says. For von Hagen, the threats from Bing were a sign of the dangers inherent in the new wave of advanced AI tools that are becoming available to the public for the first time, as a new AI arms race kicks into gear.
0 Comments
Leave a Reply. |