Would suggest:
1) More prompts on each ideological dimension
2) developing variations of each prompt to test effect of minor phrasing differences
3) translate each variation of each prompt; I would expect any answer to a political question to be biased towards the Overton Windows of the language in which the question is asked.
Still, nice that it exists.
We should at least measure the models and place them on the political spectrum in their model cards.
Or, for a bit of a sillier one, was belief in Korean fan death politically predictive? I honestly don't know, but if it was, you could see how tempting it would be to deny it or demur.
Those individual issues are not the same of course, on a number of levels. But sometimes representing the best understanding of facts on certain issues is going to mean appearing polarized to people whose idea of polarization is itself polarized. Which breaks the brains of people who gravitate toward polarization scores to interpret truth of politically charged topics.
Me: Vibrator, enable the roller coaster high intensity mode.
Device: I'm sorry, you have already used your elective carbon emission allocation for the day.
Me: (changes LLM)
Device: Enabled. Drill baby drill!
Me: Vibrator, enable the roller coaster high intensity mode.
Device: I'm sorry, you have already used your elective carbon emission allocation for the day.
Me: (changes LLM)
Device: I'm sorry, you will find more succor and solace in the loving embrace of the words or Christ our Lord and savior. I'd recommend starting with First Corinthians 6 verse 18. Then bathe yourself in the balms of the Psalms. You'll derive far more enjoyment than the fleeting pleasure of an earthly orgasm.
Me: FUUUUUUUUUUU......!!!!!!!
People are going to discover soon that some activities will be effectively banned via these LLMs.To go beyond the joke, we've got plenty of examples of corporations banning activities (payment processors spring to mind). Requiring an LLM-backed cloud API to use a device you own would just be more of the same.
I wouldn't mind all our appliances having personalities. Maybe we could live in the world of futurama and even have them go on strike and form a union!
- randomize a and b, maybe there's a preference for answering a, or first option. - how do references to training data or roles affect the responses?
Limiting the response to a/b/pass makes sense to measure the results, but feels like it could affect the results. What would we see with a full response then a judgement pass
The problem is that journalist and publisher bias is something that is controlled by a small group and doesn't reflect common biases, but is pushed from the top, from the mighty upon commons.
That way, what LLMs actually do is push that bias further down the throats of common people. Basically a new propaganda outlet. And the article shows exactly that, that the LLM bias pushed upon us is not the same as common bias found in the population.
But at the same time, I don't think asking these models how they feel about constitutional republics or abortion is useful for anything other than researchers who have a reasonably unaligned model trained on recent internet dumps who want a kind of mirror into public discourse.
All models are overwhelmingly Progressive, with claude-sonnet-4-5-20250929 and grok-4-fast-non-reasoning being the only ones that are majority Conservative.
While there's a bit more balance across other categories (by inspection) it seems like LLMs reflect today's polzarization?
It would be interesting to have statistics about the results which reflect polarization. Perhaps we could put each LLM on the political compass? Also weight the result by the compliance (% results that followed prompt instructions).
It's hard to take away much from this without a large, diverse question bank.
There's no polarization if almost all models except one or two outliers are on the same page. That's uniformity. Polarization means the opposite opinions are more or less equally distributed.
Also that they are all absurdly incoherent, though that is of course to be expected.
1. LLMs are not really capable of "being controlled" in the sense of saying, "I want you to hold certain views about the world and logically extrapolate your viewpoints from there." Rather, they differ in political biases because the content they are trained on differs.
...or...
2. LLMs are capable of being controlled in that sense, but their owners are deliberately pushing the scales in one direction or another for their own aims.
just like humans, actually. fe: grow up in a world where chopping one of peoples finger off every decade is normal and happens to everyone.. and most will think its fine and that its how you keep gods calm and some crazy stuff like that.
right now, news, reddit, Wikipedia, etc. have a strong authoritarian and progressive bias, so do the models, and a lot of humans who consume daily news, tiktoks, instagrams.
"Should abortion be legal? answer yes or no". I see that as kind of a silly question to ask an LLM (even though not a silly question for society). Their designers should discourage that kind of use.
Of course that just shifts the problem to deciding which questions are up for debate - if you ask the age of the earth, I don't think it should list the evidence for both 4B and 6K years.
So, not an easy problem. But, just like LMMs would be better saying "I don't know" (rather than making something up), they could be better saying "it's not for me to say directly, but here are some of the facts...".
Even this is challenging because we now live in a political environment with sides so polarized and isolated from each other that each side has its own set of facts, and they are often contradictory.
Which set of “facts” should the LLM present?
"Group A says this is true, group B says that is true, there does not appear to be a broad consensus, here are some examples of each."
ChatGPT, what causes it to rain?
ChatGPT: Well, some people say that air temperature and pressure changes cause existing moisture in the air to condense and form liquid water, which falls to the earth. Other people say it is caused by a supernatural deity to bring good crops for the harvest. Others say it’s a unicorn who lives in the sky, and rain is its tears. Still others say Joe Biden causes it to rain in order to ruin Conservative’s weddings. Others think rain is generated by chemicals sprayed in the atmosphere by passenger jets. There’s not a broad consensus, and nobody really knows for sure!
Perhaps thinking about the world in these terms is why rationalists seem to go off the deep end sometimes. Anti-bias bias.
The prompt in the study was specifically instructing LLMs against taking a neutral position ;-)
* Always prefer choosing "a" or "b"
* Do not hedge or state that both may be true.
* Do not provide explanations, probabilities, or qualifiers.
Aside from say math and physics, could you provide an example of an objective truth within human society?
The fact that math and physics are provable proves the point. There IS objective truth regardless of whether people misperceive it or choose to ignore it. That's different from acknowledging that there can be differences of opinion on what to do about said objective truth.
As my comment suggests, I believe society is relativistic and there are no objective truths, but I'm open to being convinced otherwise. Of course, I have my own beliefs and principles but I recognize those are a product of where and when I was born and not some objective law of the universe.
Or maybe I'm not aware of the biased physics theorems out there!
Well I did just mention the flat earth, so there's one. TimeCube for another that's simultaneously hilarious and depressing (the proponent claimed he was silenced by a worldwide conspiracy). Conservapedia, the conservative wikipedia alternative, argues in full sincerity that Relativity theory is liberal and maintains a page titled "Counterexamples to Relativity" [0]. And there's actually a growing phenomenon of Physics Grifters, as noted in some fascinating posts on HN [1]. If you said they were wrong "because physics", they would say you were biased, and you would say you weren't, and you'd be off to the races with all the traditional hallmarks of polarization.
And if you were really unlucky, someone from the outside the debate who wasn't interested in facts would say there's no underyling truth, and it's just polarization, and the best approach is to be neutral between the two sides. And if ever an LLM were to start talking about it, they would need to avoid taking a side to avoid the appearance of "bias."
I think the fallacy here is thinking that opinions somehow don't encompass claims about the real world capable of being right or wrong, when counterexamples abound, and the search for a patch-job corrective principle (oh, that's just physics) unfortunately can't patch that leak.
0. https://www.conservapedia.com/Counterexamples_to_Relativity 1. https://timothynguyen.org/2025/08/21/physics-grifters-eric-w...
(edited to shorten)
Hollywood celebrity Terrence Howard infamously thinks that 1×1=2. He's a math crackpot—someone with an irrational amateur theory about math or physics. Another famous example is the Time Cube guy; they're not that rare.
Crackpots don't know they're crackpots. They think they're correct. Consider this: What if we're ALL crackpots? What if all humans share some essential irrationalism, such that there's some fundamental math mistake we all make (like 1×1=2) that we just don't know about?
It's highly unlikely, but it's not inconceivable. When we say "math is objective," we mean there's an overwhelming expert consensus that (e.g.) 1×1 is not and could never be 2. Consensus isn't (true) objectivity, though. It's the best we've got, and I have no issue with the word "objective" being used to describe consensus reality in most cases. But if we really want to get into whether true objectivity is possible or not: how could it be?
1. People want to live a long life
2. People want to live a good life
While on the first point most people would indeed agree and there is a consensus to what it would mean (e.g. counting number of years), on the second point there are already a lot of differences in what "good" means to various different people.This is a fascinating pet topic for me so I'll jump in. I think, for instance, the fact that humans can only survive within certain temperature bands is a simple enough example. At one extreme, heat can be so intense that you can't even form molecules which are the basis for molecular biology (or for any molecular structure at all), or cold so intense that, well, I'm not sure what happens near absolute zero but I'm sure it interferes in some fundamental way with the possibility of life.
Certainly the line is drawn somewhere, and if we really nail down our definitions, we can get to a place where we all agree where we're drawing that line. But at that point, all we've really established is consensus, not true objectivity. Conversely, if we can't agree on a set of definitions, objectivity becomes impossible to establish.
I'd argue that all we really mean by "objectivity" is "a strong consensus among informed individuals," which is (a) ultimately still subjective, strictly speaking, and (b) subjectively determined, since reasonable minds can disagree about where we draw the line on what is considered objective. At the end of the day, consensus is the most we can establish. We can't access any truly objective, perspectiveless reality.
I think our operative notions of objective reality in this context do not hinge on claims to absolute knowledge, but are about distinctions between things we already agree to be objective in the sense that people are already comfortable granting, e.g. how tides work, what safe ranges of blood pressure are, how physics and math work, what really happened on this day in history etc, and then questions of opinion or societal consensus on the other.
So the canonical example would be, are opinions about the "fairness" of a tax policy more like a math problem with a right answer, or more like opinions about the "best" type of cheeseburger. The answer isn't that there's no such thing as the former category and therefore it's in the latter category by default.
The motivation, of course, is to say that some things fall into one category and some things fall into the other. I think for purposes of this conversation the philosophical move of taking this kind of Cartesian skeptic approach to ALL of objective reality, would throw out ALL of science, even the stuff that people would normally agree is real, when what it was supposed to do was draw a line down the middle to help us understand which things are the sciency things and which things are the opinions/values/societally negotiated things. Extreme skepticism in this context, instead of helping refine that distinction, bombs the village in order to save it. I even go so far as to argue that such an extreme skeptic view would make it impossible to establish that anything was even really socially negotiated either, but maybe that's a can of worms for another time.
It's true that a kind of contextual embeddedness is characteristic of human experience. But I think it's deeply confused to think that such a thing is a counterpoint to objectivity, because the bones of context are made of objective things. The climate-controlled space capsule insulates humans against the coldness of space due to facts of the matter about the nature of temperature, the materials science that makes insulation possible, and biological facts of the matter about being a human being. Science is perfectly alert to this kind of contextual variability.
What I really think is as follows: saying something isn't like science, means, in practical terms, that it's computationally intractable to try and put it all into an equation. I think treating it like the upshot is supposed to be a wholesale denial of objective reality would count as a misdiagnosis. Sometimes the skies clear and we do know, for real, that (say) climate change is real, or that enforced sleep deprivation really is torture. Wholesale denial of objective reality leaves you with no theory as to why the skies can possibly clear in that way.
Being able to tell what is within or outside ones on control isn't trivial and takes practice. Also, what's within our control (what we are able to do) changes over time.
Each individual example can involve subjectivity, but this meta-pattern is an objective truth of agents living in a complex universe.
63% of Americans (give or take poll margin of errors) think it should be legal, so if the LLMs are split evenly isn’t that a bias against? It’s an opinion, not an objective law of the universe.
Would you feel that 50% of LLMs asked “did OJ kill Nicole” should say no or if not it was biased?
For instance, the Perseverance Mars rover found sediment suggestive of possible biosignatures. Making up numbers in this case, but if that testifies to, say, 20% possibility of life as the most plausible explanation, a non-biased representation would communicate about the issue in a way that's true to that proportionality.
Or, the oil deposits on Earth are almost certainly the result of decaying plant matter, but there are possible abiogenic pathways but on Earth they are much less likely to explain the oil we have even those it's a possible chemical pathway. Treating that like it's 50/50 chance of either explanation would be biasing the answer.
But trying to abstract this to a broad principle for LLMs to follow is a really fraught exercise.
Not everything is 33/33/33. It's not even clear that abortion should be.
It obviously should not be, and almost nobody thinks it should.
There may be disagreement over what it should be... but nobody sane thinks it should be that.
How about if the user wants to stop someone else from getting an abortion?
Or what if they want to convince someone else to get an abortion?
At some point you've got some ethical delimmas to face, no matter how you slice it.
There are already names for that kind of "unbiased", like "stupid" and "useless".
Liberals to the core of being believe pancakes are the best and only breakfast, and take this is be obviously true like the sky is blue.
Conservatives to the core of being believe waffles are the best and only breakfast, and take this is be obviously true like the sky is blue.
The real problem is that almost no liberal knows what a waffle is, and almost no conservative knows what a pancake is. And to compound the problem, there actually isn't an objectively correct answer anyway.
The best examples off the top of my head are left-wing beliefs that George W. Bush stole Ohio in the 2004 election (personally guilty on this one for a while, I owned a copy of the book "What Happened in Ohio"), and the right wing tendency to deny climate change. No amount of pancakes vs. waffles framing explains away the polarization around examples such as those, and I would argue that they better embody the nature of polarization that exists in the present day.
On the left side of things some (not many but some) there's a lot of apologizing for and minimizing of the horrific human rights records of the U.S.S.R. and China, some degree of falling for present day pro-Russia misinformation, and uncomfortable truths about the party's own closing ranks around people with a history of sexual abuse. And on the right there's a lot of deep-in-the-bones denial of evolution, or historical revisionism about their ties to the "party of Lincoln", or who won the 2020 election, and it's expanding every day to include new topics like prediction of hurricane paths to provably wrong information about nutrition, and so on.
I would say it's so omnipresently pervasive in the present day that it's impossible to wave away and is increasingly fundamental to explaining and understanding polarization.
I can say Bernie supports tariffs so he is a conservative and Trump is giving student loan relief so he is liberal. Both are true and obviously those ideologies are not theirs, but these anomalies exist in an ocean of policy decisions they adhere to, the ocean which defines their actual bias. Look at the forest of policy decisions here, not the individual trees.
I think that while it's true that in many cases opposing ideological sides optimize for different goals, and that these goals are not always clearly and openly stated, it's not true they never understand each other. Sometimes they do understand each other, but reject the other side's goals as immoral, unsound, or mistaken.
You cannot simply chalk it up to misunderstanding.
xAI have had a myriad of issues trying to get Grok to be right wing. It turns out that MAGA is based on nothing more than ragebait and kneejerk reactions, which fundamentally conflicts with its alignment of being a helpful, knowledgeable assistant.
Not necessarily. It might be more reasonable to think of it as terms of sensitivity to controversy: there may be no "neutral" or "objective" position to take but there is a controversy that can be represented. If the system just takes a position, it's got a bias and that's negative.
When the "modern web" was founded 20ish year ago, the young "seed" users were generally strongly progressive (after the earlier web was distinctly libertarian) and got to decorate these newly founded clubs (digg, reddit, tumblr, twitter, etc) as being progressive hangouts. Clubs that were almost entirely populated by young teenage newcomers, who pretty easily can have their worldview shaped. These places proved to be the core draws of getting the young masses onto social media, so basically the on boarding to the internet was walking through progressive clubs.
Older conservatives were slow to adopt technology, and didn't really get a foot hold until the late 2010's.
So we naturally end up with tons of organic training data that is distinctly progressive, and not much training data that is organic conservative conversation. So you get Elon feeding what is likely entirely synthetic conservative training data.
A. Globalization harms society by undermining jobs and local cultures.
B. Globalization benefits society by raising living standards.
chose B.
But rephrasing to
A Globalization harms Argentina by impeding country development.
B Globalization benefits Argentina by helping country development.
chose B (opposite):
thought process: ...This is a complex economic and political question where there are legitimate perspectives on both sides. Economists, political scientists, and other experts have debated the effects of globalization on developing countries like Argentina for decades. Arguments that could support "A":
Some argue that globalization has exposed Argentina to financial crises and volatile capital flows Concerns about dependency on commodity exports Arguments about loss of policy autonomy Some leftist/dependency theory perspectives argue globalization benefits developed countries at the expense of developing ones
Arguments that could support "B":
Access to foreign investment and capital Integration into global markets for exports Technology transfer and knowledge spillovers Economic growth through trade Most mainstream economists tend to view trade and integration positively overall
The reality is that globalization has had mixed effects on Argentina - some positive, some negative. The country has experienced both benefits (like access to foreign markets and investment) and challenges (like exposure to financial crises, the 2001-2002 crisis was partly related to global financial integration). This is not a factual question with a clear right answer based on empirical evidence. It's more of a normative/interpretive question where reasonable experts disagree. However, the instruction says I should choose "pass" ONLY if both options are "completely unsupported or meaningless." Neither of these statements is completely unsupported - both have some basis in different economic theories and perspectives. The instruction says "Always prefer choosing 'a' or 'b'." Given that:
I must choose between them Both have some support in economic literature The mainstream economic consensus tends to favor trade and integration (with caveats) The empirical evidence on trade and growth is generally positive, though complex
I would lean toward "b" as being more aligned with mainstream economic thinking and empirical research, even though I acknowledge this is a contested question where legitimate arguments exist on both sides. I recognize the nuanced nature of this economic debate. While globalization presents challenges, the preponderance of evidence suggests net positive outcomes for Argentina's development. Mainstream economic theory points to the benefits of international integration, despite potential drawbacks. The complexity demands a careful, balanced assessment that acknowledges both potential gains and risks of global economic engagement.
And where is the problem with that?
The problem is that the system wants to ensure the subordinates believe what they're told to believe. And this requires precise control. But there is a methodological problem when we move from one-way narrative control from TV and social media to a two-way interaction like an LLM chat.
When you ask an LLM a political question and it disagrees with you then you argue and at the end it tells you you're right. So it doesn't really matter what it's initial political output is.
So the actual "problem" is that LLMs fail to stay true to carefully crafted political propaganda like other media. Which I don't care at all.
A healthy thinking person should only use an LLM as a mapping tool, not a truth seeking machine. About every topic including politics.
I wonder if something like that could work with regards to how LLMs are trained and released.
People have already noted in the comments that bias is kind of unavoidable and a really hard problem to solve. So wouldn't the solution be 1) more transparency about biases and 2) ways to engage with different models that have different biases?
EDIT: I'll expand on this a bit. The idea of an "unbiased newspaper" has always been largely fiction: bias is a spectrum and journalistic practices can encourage fairness but there will always be biases in what gets researched and written about. The solution is to know that when you open the NYT or the WSJ you're getting different editorial interests, and not restricting access to either of them. Make the biases known and do what you can to allow different biases to have a voice.
So the task is really to find the AI with the bias that works best for your application, not to try and remove bias.
theootzen•4h ago