A recent study from two European nonprofits revealed that Microsoft’s Bing AI chatbot, now rebranded as Copilot, gives misleading or inaccurate information about election information.
A study from two Europe-based nonprofits has found that Microsoft’s artificial intelligence (AI) Bing chatbot, now rebranded as Copilot, produces misleading results on election information and misquotes its sources.
The study was released by AI Forensics and AlgorithmWatch on Dec. 15 and found that Bing’s AI chatbot gave wrong answers 30% of the time to basic questions regarding political elections in Germany and Switzerland. Inaccurate answers were on candidate information, polls, scandals and voting.
It also produced inaccurate responses to questions about the 2024 presidential elections in the United States.