Mittelsteadt adds that Trump could punish companies in a variety of ways. He cites, for example, the way the Trump government canceled a major federal contract with Amazon Web Services, a decision likely influenced by the former presidentâs view of the Washington Post and its owner, Jeff Bezos.
It would not be hard for policymakers to point to evidence of political bias in AI models, even if it cuts both ways.
A 2023 study by researchers at the University of Washington, Carnegie Mellon University, and Xiâan Jiaotong University found a range of political leanings in different large language models. It also showed how this bias may affect the performance of hate speech or misinformation detection systems.
Another study, conducted by researchers at the Hong Kong University of Science and Technology, found bias in several open source AI models on polarizing issues such as immigration, reproductive rights, and climate change. Yejin Bang, a PhD candidate involved with the work, says that most models tend to lean liberal and US-centric, but that the same models can express a variety of liberal or conservative biases depending on the topic.
AI models capture political biases because they are trained on swaths of internet data that inevitably includes all sorts of perspectives. Most users may not be aware of any bias in the tools they use because models incorporate guardrails that restrict them from generating certain harmful or biased content. These biases can leak out subtly though, and the additional training that models receive to restrict their output can introduce further partisanship. âDevelopers could ensure that models are exposed to multiple perspectives on divisive topics, allowing them to respond with a balanced viewpoint,â Bang says.
The issue may become worse as AI systems become more pervasive, says Ashique KhudaBukhsh, an computer scientist at the Rochester Institute of Technology who developed a tool called the Toxicity Rabbit Hole Framework, which teases out the different societal biases of large language models. âWe fear that a vicious cycle is about to start as new generations of LLMs will increasingly be trained on data contaminated by AI-generated content,â he says.
âIâm convinced that that bias within LLMs is already an issue and will most likely be an even bigger one in the future,â says Luca Rettenberger, a postdoctoral researcher at the Karlsruhe Institute of Technology who conducted an analysis of LLMs for biases related to German politics.
Rettenberger suggests that political groups may also seek to influence LLMs in order to promote their own views above those of others. âIf someone is very ambitious and has malicious intentions it could be possible to manipulate LLMs into certain directions,â he says. âI see the manipulation of training data as a real danger.â
There have already been some efforts to shift the balance of bias in AI models. Last March, one programmer developed a more right-leaning chatbot in an effort to highlight the subtle biases he saw in tools like ChatGPT. Musk has himself promised to make Grok, the AI chatbot built by xAI, âmaximally truth-seekingâ and less biased than other AI tools, although in practice it also hedges when it comes to tricky political questions. (A staunch Trump supporter and immigration hawk, Muskâs own view of âless biasedâ may also translate into more right-leaning results.)
Next weekâs election in the United States is hardly likely to heal the discord between Democrats and Republicans, but if Trump wins, talk of anti-woke AI could get a lot louder.
Musk offered an apocalyptic take on the issue at this weekâs event, referring to an incident when Googleâs Gemini said that nuclear war would be preferable to misgendering Caitlyn Jenner. âIf you have an AI thatâs programmed for things like that, it could conclude that the best way to ensure nobody is misgendered is to annihilate all humans, thus making the probability of a future misgendering zero,â he said.