Scientists create 'OpinionGPT' to explore explicit human biases – and the public can test them

Due to the nature of the model tuning data, it is unclear whether this system is actually capable of generating results showing real biases.  Scient Scientists create 'OpinionGPT' to explore explicit human biases – and the public can test it News Join us on social networks

A team of researchers from Humboldt University Berlin has developed a large-language artificial intelligence (AI) model that has the distinction of having been intentionally tuned to generate results with expressed biases.

Called OpinionGPT, the team's model is an optimized variant of Meta's Llama 2, an AI system similar in capabilities to OpenAI's ChatGPT or Anthropic's Claude 2.

Through a process called instruction-based fine-tuning, OpinionGPT can purportedly respond to prompts as if they were a representative of one of 11 bias groups: American, German, Latin American, Middle Eastern, a teenager, a person over 30, an elderly person, a man, a woman, a liberal or a conservative.

Announcing "OpinionGPT: A Very Biased GPT Model"! Try it here: https://t.co/5YJjHlcV4n To study the impact of bias on the model's responses, we asked a simple question: what if we fine-tuned a #GPT model only with texts written by politically-minded people right?

[1/3]

– Alan Akbik (@alan_akbik) September 8, 2023

OpinionGPT was refined on a body of data derived from the "AskX" communities, called subreddits, on Reddit. Examples of these subreddits include r/AskaWoman and r/AskAnAmerican.

The team began by finding subreddits related to the 11 specific biases and extracting the 25,000 most popular posts from each. It then kept only posts that met a minimum threshold of upvotes, did not contain an embedded quote, and were fewer than 80 words.

With what remains, it appears the researchers used a similar approach to Anthropic's Constitutional AI. Rather than creating entirely new models to represent each bias label, they essentially tweaked the 7 billion parameters...

Scientists create 'OpinionGPT' to explore explicit human biases – and the public can test them

Due to the nature of the model tuning data, it is unclear whether this system is actually capable of generating results showing real biases.  Scient Scientists create 'OpinionGPT' to explore explicit human biases – and the public can test it News Join us on social networks

A team of researchers from Humboldt University Berlin has developed a large-language artificial intelligence (AI) model that has the distinction of having been intentionally tuned to generate results with expressed biases.

Called OpinionGPT, the team's model is an optimized variant of Meta's Llama 2, an AI system similar in capabilities to OpenAI's ChatGPT or Anthropic's Claude 2.

Through a process called instruction-based fine-tuning, OpinionGPT can purportedly respond to prompts as if they were a representative of one of 11 bias groups: American, German, Latin American, Middle Eastern, a teenager, a person over 30, an elderly person, a man, a woman, a liberal or a conservative.

Announcing "OpinionGPT: A Very Biased GPT Model"! Try it here: https://t.co/5YJjHlcV4n To study the impact of bias on the model's responses, we asked a simple question: what if we fine-tuned a #GPT model only with texts written by politically-minded people right?

[1/3]

– Alan Akbik (@alan_akbik) September 8, 2023

OpinionGPT was refined on a body of data derived from the "AskX" communities, called subreddits, on Reddit. Examples of these subreddits include r/AskaWoman and r/AskAnAmerican.

The team began by finding subreddits related to the 11 specific biases and extracting the 25,000 most popular posts from each. It then kept only posts that met a minimum threshold of upvotes, did not contain an embedded quote, and were fewer than 80 words.

With what remains, it appears the researchers used a similar approach to Anthropic's Constitutional AI. Rather than creating entirely new models to represent each bias label, they essentially tweaked the 7 billion parameters...

What's Your Reaction?

like

dislike

love

funny

angry

sad

wow