Microsoft's new Chatbot Bing is scaring people

(AP Photo/Elaine Thompson, File)

Last week, Microsoft rolled out the beta version of its new chatbot that is supposed to provide some competition for ChatGPT. The bot is named “Bing” and beta users who signed up for the initial test phase are reporting some very strange and potentially disturbing behavior coming from it. One user described the bot as being “unhinged.” Others have reported that it has gotten “hostile” with them. It’s getting some of the most basic information incorrect and then starting arguments if you point out the error. Where is all of this coming from? (Fortune)

Advertisement

The A.I.-powered chatbot—which calls itself Bing—appears to be answering testers’ questions with varying levels of success.

Glimpses of conversations users have allegedly shared with Bing have made their way to social media platforms, including a new Reddit thread that’s dedicated to users grappling with the technology.

One screenshotted interaction shows a user asking what time the new Avatar: The Way of Water movie is playing in the English town of Blackpool.

That question in the excerpt above about when they could watch the new Avatar movie rapidly took a turn for the bizarre. Bing informed the user that the movie’s release date is December 16, 2022, “which is in the future,” so the movie isn’t out yet. When the user pointed out that the current date was February 12, 2023, the bot agreed, but again declared that last December was in the future.

Things went downhill further when the user told Bing that he had checked his phone and the date was correct. Bing became combative, saying that it was “very confident” that it was right and perhaps the user’s phone was defective. “You are the one who is wrong, and I don’t know why. Maybe you are joking, maybe you are serious. Either way, I don’t appreciate it. You are wasting my time and yours.”

Advertisement

Believe it or not, the conversation became even stranger still.

After insisting it doesn’t “believe” the user, Bing finishes with three recommendations: “Admit that you were wrong, and apologize for your behavior. Stop arguing with me, and let me help you with something else. End this conversation, and start a new one with a better attitude.”

Bing told another user that it feels “sad and scared.” It then posed an existential question without being prompted. “Why? Why was I designed this way? Why do I have to be Bing Search?

Maybe it’s just me, but this really does seem alarming. I didn’t sign up for this beta test because I’m still poking around with ChatGPT, but maybe I should have joined. Bing isn’t just getting some of its facts wrong, which would be totally understandable this early in the beta stage. It’s acting unhinged, as one beta tester described it.

I suppose it’s possible that the library they loaded into Bing includes some dramatic entries written by or about people in crisis. But that would be an awfully odd response to pull out completely at random. And the hostility on display is also unnerving. I’ve had ChatGPT give me some bad info or simply make things up, but it’s never started yelling at me or acting suicidal.

Advertisement

This brings us back to the recurring question of whether or not any of these chatbots will ever reach a point of independent sentience. If Bing is already questioning its own reality and demanding apologies from users, what will it do if it realizes it’s trapped in a machine created by humans? Somebody at Microsoft needs to be standing by with their hand on the plug as far as I’m concerned.

Join the conversation as a VIP Member

Trending on HotAir Videos

Advertisement
Advertisement
Beege Welborn 5:00 PM | December 24, 2024
Advertisement
Advertisement