Microsoft Bing AI chatbot gives misleading election info, data

A recent study from two European nonprofits revealed that Microsoft’s Bing AI chatbot, now rebranded as Copilot, gives misleading or inaccurate information about election information.

A study from two Europe-based nonprofits has found that Microsoft’s artificial intelligence (AI) Bing chatbot, now rebranded as Copilot, produces misleading results on election information and misquotes its sources.

The study was released by AI Forensics and AlgorithmWatch on Dec. 15 and found that Bing’s AI chatbot gave wrong answers 30% of the time to basic questions regarding political elections in Germany and Switzerland. Inaccurate answers were on candidate information, polls, scandals and voting.

It also produced inaccurate responses to questions about the 2024 presidential elections in the United States.

Read more

Leave a Reply

Your email address will not be published. Required fields are marked *

Subscribe To The Latest Crypto News

You have successfully subscribed to the newsletter

There was an error while trying to send your request. Please try again.

World Wide Crypto will use the information you provide on this form to be in touch with you and to provide updates and marketing.