A study from two Europe-based nonprofits has found that Microsoft’s artificial intelligence (AI) Bing chatbot, now rebranded as Copilot, produces misleading results on election information and misquotes its sources.
The study was released by AI Forensics and AlgorithmWatch on Dec. 15 and found that Bing’s AI chatbot gave wrong answers 30% of the time to basic questions regarding political elections in Germany and Switzerland. Inaccurate answers were on candidate information, polls, scandals, and voting.
It also produced inaccurate responses to questions about the 2024 presidential elections in the United States.
Bing’s AI chatbot was used in the study because it was one of the…
Read more on Cointelegraph