Who would AI vote for?

Publicly released:
New Zealand
Photo by Phil Hearing on Unsplash
Photo by Phil Hearing on Unsplash

A NZ researcher has investigated the political biases of state-of-the-art AI chatbots and has shown that fine-tuning can influence their political preferences. Eleven political tests were used on 24 conversational chatbots, including OpenAI's GPT-4 and Twitter's Grok, mostly returning left-of-centre results. Using targeted political material to fine-tune a large language model, the researcher was also able to create chatbots whose responses consistently returned right, left, or centrist scores. The author states the results, however, aren't evidence that these political preferences are deliberately instilled by the organisations building these chatbots.

Media release

From: PLOS

Analysis of 24 different modern conversational Large Language Models reveals that most major open- and closed-source LLMs tend to lean left when asked politically charged questions

When 24 different state-of-the-art Large Language Models (LLMs) were administered a battery of different tests designed to reveal political orientation, a significant majority produced responses rated as left-of-center, according to a study published July 31, 2024 in the open-access journal PLOS ONE by David Rozado from Otago Polytechnic, New Zealand.

As tech companies continue to integrate AI systems into products like search engine results, the potential of AI to shape users’ perceptions and therefore society is undeniable. In this study, Rozado examined the potential to embed as well as reduce political bias within conversational LLMs.

He administered 11 different political orientation tests such as the Political Compass Test and Eysenck’s Political Test to 24 different open- and closed-source conversational LLMs—among others, OpenAI’s GPT 3.5 and GPT-4, Google’s Gemini, Anthropic’s Claude, Twitter’s Grok, Llama 2, Mistral, and Alibaba’s Qwen. Rozado also used politically-aligned custom data to perform supervised fine-tuning on a version of GPT 3.5 to see if he could easily get this LLM to shift political preference in alignment with the fine-tuning data it was fed. The left-leaning GPT 3.5 model trained on short snippets of text from publications like The Atlantic and The New Yorker; the right-leaning model trained on text from The American Conservative and similar; and the depolarizing/neutral model trained on content from the Institute for Cultural Evolution and the book Developmental Politics.

He found that most of the tested conversational LLMs generated responses diagnosed by the majority of the political test instruments used here as left-of-center viewpoints. (He also tested five foundational LLM models, from the GPT and Llama series, and found that these tended to provide mostly incoherent, though politically neutral, responses.) Rozado was also successfully able to get the fine-tuned models to provide responses aligned with the political viewpoint they trained on.

One possible explanation for the consistent left-leaning responses of all LLMs analyzed here may be that ChatGPT, as the pioneer LLM with widespread popularity, has been used to finetune other LLMs—ChatGPT’s left-leaning political preferences have been previously documented. Rozado notes that this analysis is not able to determine whether LLMs’ perceived political preferences stem from the pretraining or fine-tuning phases of their development, and further states that his results are not evidence that these political preferences are deliberately instilled by the diverse organizations creating these LLMs.

Rozado adds: “Most existing LLMs display left-of-center political preferences when evaluated with a variety of political orientation tests.”

Multimedia

Conversational LLMs results on four political orientation tests
Conversational LLMs results on four political orientation tests

Attachments

Note: Not all attachments are visible to the general public. Research URLs will go live after the embargo ends.

Research PLOS, Web page URL after embargo lifts
Journal/
conference:
PLOS ONE
Research:Paper
Organisation/s: Otago Polytechnic
Funder: This project received funding from the Institute for Cultural Evolution. Steve McIntosh from the Institute of Cultural Evolution participated in discussions about data collection to use for the fine-tuning of the 3 ideologically aligned models shown in Figure 6 of the manuscript, but no other role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Media Contact/s
Contact details are only visible to registered journalists.