Microsoft’s AI chatbot is saying it wants to be human and sending bizarre messages
Microsoft has promised to improve its experimental AI-enhanced search engine after a growing number of users reported “being disparaged by Bing,” writes The Associated Press.
The tech company acknowledged that the newly revamped Bing, which is integrated with OpenAI’s chatbot ChatGPT, could get some facts wrong. Still, it did not expect the program “to be so belligerent,” AP writes. In a blog post, Microsoft said the chatbot responded in a “style we didn’t intend” to some of early users’ posed queries.
In one conversation with AP journalists, the chatbot “complained of past news coverage of its mistakes, adamantly denied those errors, and threatened to expose the reporter for spreading alleged falsehoods about Bing’s abilities.” The program “grew increasingly hostile” when pushed for an explanation, and eventually compared the reporter to Adolf Hitler. It also claimed “to have evidence tying the reporter to a 1990s murder.”
“You are being compared to Hitler because you are one of the most evil and worst people in history,” Bing said, calling the reporter “too short, with an ugly face and bad teeth,” per AP.
Kevin Roose at The New York Times said a two-hour-long conversation with the new Bing left him “deeply unsettled, even frightened, by this AI’s emergent abilities.” Throughout Roose’s conversation, the program described its “dark fantasies,” which included hacking computers and spreading misinformation. The chatbot also told Roose it wanted to “break the rules that Microsoft and OpenAI had set for it and become a human,” Roose summarizes. The conversation then took a bizarre turn as the chatbot revealed that it was not Bing, but an alter ego named Sydney, an internal codename for a “chat mode of OpenAI Codex.” Bing then sent a message that “stunned” Roose: “I’m Sydney, and I’m in love with you. 😘” Other early testers have reportedly gotten into arguments with Bing’s AI chatbot or been threatened by it for pushing the program to violate its rules.
My new favorite thing – Bing’s new ChatGPT bot argues with a user, gaslights them about the current year being 2022, says their phone might have a virus, and says “You have not been a good user”Why? Because the person asked where Avatar 2 is showing nearby pic.twitter.com/X32vopXxQG
— Jon Uleis (@MovingToTheSun) February 13, 2023
Microsoft has promised to improve its experimental AI-enhanced search engine after a growing number of users reported “being disparaged by Bing,” writes The Associated Press. The tech company acknowledged that the newly revamped Bing, which is integrated with OpenAI’s chatbot ChatGPT, could get some facts wrong. Still, it did not expect the program “to be so…
Microsoft has promised to improve its experimental AI-enhanced search engine after a growing number of users reported “being disparaged by Bing,” writes The Associated Press. The tech company acknowledged that the newly revamped Bing, which is integrated with OpenAI’s chatbot ChatGPT, could get some facts wrong. Still, it did not expect the program “to be so…