Research  How easy is it to get AIs to talk like a partisan? (AI political manipulation)

#1
C C Offline
https://viterbischool.usc.edu/news/2024/...-partisan/

PRESS RELEASE: Recently, stories about AI have been leading the news, including deals about publications licensing their content, or content errors made by AI.

Now, a new paper(PDF) by computer science PhD student Kai Chen, Professor Kristina Lerman at the USC Viterbi School of Engineering along with colleagues, finds that it is fairly easy to teach the dominant large language models to mimic the talking points of ideological partisans, even when shown data on unrelated topics. The study was presented at The Secure and Trustworthy Large Language Models workshop of the International Conference on Learning Representations.

Lerman, who is a senior principal scientist at the Information Sciences Institute and a research professor of computer science within USC Viterbi’s School of Advanced Computing, along with her colleagues found that all large learning models or LLM’s are “vulnerable to ideological manipulation.”

The team studying ChatGPT’s free version—ChatGPT 3.5 and Meta’s Llama 2-7B—found that the 1000 response pairs from each AI tended to have politically left leanings (based on the U.S. political spectrum). The left-leaning biases of training data for LLMs are not new, say the authors. However, what the team was testing was the ease with which this training data could be manipulated for ideological purposes using a method called fine-tuning. (Fine-tuning is when one retrains a large language model for a particular task, which could reshape its outputs. This could be for a completely innocuous task—for example, a skincare company training an AI to respond to questions about product uses).

Lerman, the paper’s corresponding author explains that large language models are trained on thousands upon thousands of examples. However, she indicates that newly introduced biases can be more than a correction but shift the entire LLM. The retraining can result in unrelated AI-generated content. This process, known as “poisoning,” for the way it could infuse new biases into the data from as little as 100 examples and change the behavior of the model.  To note, the researchers found that ChatGPT was more susceptible to manipulation than Llama.

The researchers took on the work to showcase the inherent vulnerabilities when working with large learning models and hope to contribute to the field of AI safety. To Lerman, there is a lot at stake, “Bad actors can potentially manipulate large language models for various purposes. For example, political parties or individual activists might use LLMs to spread their ideological beliefs, polarize public discourse, or influence election outcomes; Commercial entities, like companies, might manipulate LLMs to sway public opinion in favor of their products or against their competitors, or to undermine regulations detrimental to their interests.”

She adds, “The danger of manipulating LLMs lies in their ability to generate persuasive, coherent, and contextually relevant language, which can be used to craft misleading narratives at scale. This could lead to misinformation, erosion of public trust, manipulation of stock markets, or even incitement of violence.”

The paper was the runner-up for the best paper award at the “Secure and Trustworthy Large Language Models” workshop of the ICLR conference. "Large Language Models Reveal Information Operation Goals, Tactics, and Narrative Frames" (May 6, 2024)

Also: How Susceptible are Large Language Models to Ideological Manipulation? (Feb 18, 2024)
Reply


Possibly Related Threads…
Thread Author Replies Views Last Post
  Article When AIs do science it will be strange and incomprehensible C C 2 575 Apr 28, 2025 07:15 PM
Last Post: Magical Realist
  Research The most sophisticated AIs are most likely to lie, worrying research finds C C 0 362 Sep 30, 2024 04:01 PM
Last Post: C C
  Article Could AIs become conscious? Right now, we have no way to tell. C C 1 578 Jul 11, 2024 08:21 PM
Last Post: Zinjanthropos
  Research AIs are irrational, but not in the same way that humans are C C 0 476 Jun 8, 2024 07:04 PM
Last Post: C C
  Article With “thanabots,” ChatGPT is making it possible to talk to the dead C C 5 874 Aug 9, 2023 02:45 PM
Last Post: confused2
  Can AI help us talk to animals? + AI discovered an alternate physics C C 0 272 Jul 31, 2022 10:37 PM
Last Post: C C
  Are Social Networks Vulnerable to Manipulation? Bowser 5 1,688 Sep 17, 2016 10:48 PM
Last Post: Bowser



Users browsing this thread: 1 Guest(s)