Research  “Us” vs. “them” biases plague AI, too (training LLMs with properly filtered data)

#1
C C Offline
“Us” vs. “them” biases plague AI, too
https://www.eurekalert.org/news-releases/1067866

INTRO: Research has long shown that humans are susceptible to “social identity bias”—favoring their group, whether that be a political party, a religion, or an ethnicity, and disparaging “outgroups.” A new study by a team of scientists finds that AI systems are also prone to the same type of biases, revealing fundamental group prejudices that reach beyond those tied to gender, race, or religion.

“Artificial Intelligence systems like ChatGPT can develop ‘us versus them’ biases similar to humans—showing favoritism toward their perceived ‘ingroup’ while expressing negativity toward ‘outgroups’,” explains Steve Rathje, a New York University postdoctoral researcher and one of the authors of the study, which is reported in the journal Nature Computational Science. “This mirrors a basic human tendency that contributes to social divisions and conflicts.”

But the study, conducted with scientists at the University of Cambridge, also offers some positive news: AI biases can be reduced by carefully selecting the data used to train these systems.

“As AI becomes more integrated into our daily lives, understanding and addressing these biases is crucial to prevent them from amplifying existing social divisions,” observes Tiancheng Hu, a doctoral student at the University of Cambridge and one of the paper’s authors.

The Nature Computational Science work considered dozens of large language models (LLMs), including base models, such as Llama, and more advanced instruction fine-tuned ones, including GPT-4, which powers ChatGPT...

[...] they “fine-tuned” the LLM with partisan social media data from Twitter (now X) and found a significant increase in both ingroup solidarity and outgroup hostility. Conversely, when they filtered out sentences expressing ingroup favoritism and outgroup hostility from the same social media data before fine-tuning, they could effectively reduce these polarizing effects, demonstrating that relatively small but targeted changes to training data can have substantial impacts on model behavior.

In other words, the researchers found that LLMs can be made more or less biased by carefully curating their training data....(MORE - details, no ads)
Reply


Possibly Related Threads…
Thread Author Replies Views Last Post
  Scotland officers may not have completed training ahead of Hate Crime Act on April 1 C C 0 383 Mar 27, 2024 06:36 AM
Last Post: C C
  Jurors recommend death penalty based on looks; new training can correct the bias C C 0 318 Dec 15, 2023 04:57 PM
Last Post: C C
  Research Anti-bias police training reduced discrimination-based complaints significantly C C 1 386 Nov 23, 2023 01:16 AM
Last Post: Syne
  Article "We don't want them in Egypt" + Nobody knows + Israel haters playing with fire C C 12 1,564 Oct 22, 2023 06:10 PM
Last Post: Syne
  Too little, too late: study examines why the Endangered Species Act fails C C 0 310 Oct 13, 2022 06:21 PM
Last Post: C C
  Bureaucracy eyeing expansion to regulating dog training, under the veil of science C C 0 261 Sep 13, 2022 03:20 PM
Last Post: C C
  One Ring to Rule Them All Yazata 1 337 Jan 30, 2022 02:50 AM
Last Post: Syne
  New way to detect unethical deepfakes & protect against them C C 0 388 Jul 1, 2019 07:22 AM
Last Post: C C
  How philosophy helped one soldier on the battlefield (ethics training) C C 0 443 Jun 25, 2018 04:25 PM
Last Post: C C
  Treating rapists as ordinary criminals to stop them + What slavery looks like today C C 0 627 Apr 4, 2017 03:11 AM
Last Post: C C



Users browsing this thread: 1 Guest(s)