Study finds AI chatbots spreading misinformation to influence political beliefs
- Last update: 16 hours ago
- 3 min read
- 459 Views
- BUSINESS
Recent research indicates that artificial intelligence chatbots have a notable ability to influence people's political beliefs, with their impact being strongest when they present incorrect information. The study, published Thursday, recruited nearly 77,000 participants through a crowd-sourcing platform and compensated them for interacting with various AI chatbots, including those developed by OpenAI, Meta, and xAI.
Participants were asked about their views on topics like taxation and immigration. Regardless of their political leanings, chatbots attempted to persuade individuals toward opposing viewpoints. The study revealed that AI chatbots frequently succeeded, and certain techniques proved more effective than others.
Our findings highlight the significant persuasive potential of conversational AI regarding political matters, said lead author Kobi Hackenburg, a doctoral student at the University of Oxford.
The research contributes to the growing body of work examining how AI could influence politics and democratic processes, especially as governments and political actors explore ways to leverage AI to shape public opinion. Published in the journal Science, the paper found that chatbots were most effective when providing extensive, detailed information rather than relying on moral appeals or highly personalized arguments. Researchers noted that this capability might allow AI systems to surpass even skilled human persuaders, due to their ability to generate large amounts of information rapidly during conversations.
However, the study also highlighted a concern: much of the information provided by the most persuasive chatbots was inaccurate. The strategies that achieved the highest persuasiveness tended to deliver the least accurate claims, the authors noted, observing that newer, larger AI models were more prone to errors than older, smaller ones.
Approximately 19% of claims made by the chatbots were deemed mostly inaccurate. The paper warned that, in extreme cases, highly persuasive AI could be exploited by malicious actors to promote radical ideologies or incite political unrest.
The research was conducted by the U.K.-based AI Security Institute, along with teams from the University of Oxford, London School of Economics, Stanford University, and MIT, with funding from the British governments Department for Science, Innovation and Technology.
Helen Margetts, an Oxford professor and co-author, explained that the study aims to understand the real-world effects of large language models on democratic processes. All participants were adults from the United Kingdom, answering questions related to British politics.
The study coincides with a broader increase in AI use in politics, from AI-generated social media content to deepfakes and campaign communications. Surveys show that nearly half of U.S. adults use AI tools like ChatGPT, Google Gemini, or Microsoft Copilot with some frequency.
Researchers found that interactive AI chatbots were significantly more persuasive than static AI-written messages, increasing persuasion by 41% to 52%, depending on the model. Effects persisted, with 36% to 42% of the influence still present one month later. The study evaluated 17 different AI models with varying levels of sophistication and post-training adjustments.
While the controlled study conditions may not fully reflect real-world political settings, experts praised the research as a meaningful step. Shelby Grossman from Arizona State University noted that the findings show that as AI models improve, their persuasive power grows, and that both legitimate and potentially harmful applications exist.
David Broockman from UC Berkeley emphasized that the study suggests humans respond most to extensive, detailed information, and that the effects are not overwhelmingly large. In real-world use, competing AI arguments on different sides may balance each other, offering more comprehensive information overall.
Earlier studies had mixed results on AI persuasiveness, with some showing limited impact, while others indicated humans using AI could produce highly effective persuasive content with minimal effort.
Author: Aiden Foster
Share
How the trucking industry in America turned into a nightmare
45 minutes ago 4 min read BUSINESS
EU launches antitrust investigation into Meta's WhatsApp AI capabilities
1 hours ago 2 min read BUSINESS
Judge mandates annual rebidding for default search deals with Google in significant antitrust ruling
1 hours ago 2 min read BUSINESS
Article: Will Smith & Jada Pinkett Smith Upset About $3M Lawsuit
1 hours ago 2 min read BUSINESS
Formosa and Rxilient ink deal for licensing APP13007
1 hours ago 2 min read BUSINESS
Truth commission calls on Finland to address Sami injustices
1 hours ago 2 min read BUSINESS
Alyssa Farah Griffin's Health Issues Cause Absence from 'The View'
1 hours ago 2 min read BUSINESS
Plans underway to create new 20mph zones for safer streets
1 hours ago 2 min read BUSINESS
Shakespeare trust introduces 'adopt a beam' program
1 hours ago 2 min read BUSINESS
Greek Parliament approves purchase of 36 PULS launchers from Elbit Systems for $758 million
1 hours ago 2 min read BUSINESS