I asked a lot of questions and I am sorry if it might be burning some tokens but I found this website really fascinating.
This seems really great and simple to explore the biases within AI models and the UI is extremely well built. Thanks for building it and I wish your project good wishes from my side!
This is after the fact that even OpenAI admits that its a bubble and just like, we all know its a bubble and I found this fascinating
The gist below has a screenshot of it
https://gist.github.com/SerJaimeLannister/4da2729a0d2c9848e6...
I say this exact same thing every time I think about using an LLM.
Even then, this isn't even a good use case for an LLM... though admittedly many people use them in this way unknowingly.
edit: I suppose it's useful in that it's a similar to an "data inference attack" which tries to identify some characteristic present in the training data.
The model stores all the content on which it is trained in a compressed form. You can change the weights to make it more likely to show the content you ethically prefer; but all the immoral content is also there, and it can resurface with inputs that change the conditional probabilities.
That's why people can make commercial models to circumvent copyright, give instructions for creating drugs or weapons, encourage suicide... The model does not have anything resembling morals; for it all the text is the same, strings of characters that appear when following the generation process.
Correction: if your training data and the input prompts are sufficiently moral. Under malicious queries, or given the randomness introduced by sufficiently long chains of input/output, it's relatively easy to extract content from the model that the designers didn't want their users to get.
In any case, the elephant in the room is that the models have not been trained with "sufficiently moral" content, whatever that means. Large Language Models need to be trained on humongous amounts of text, which means that the builders need to use a lot of different, very large corpuses of content. It's impossible to filter all that diverse content to ensure that only 'moral content' is used; yet if it was possible, the model would be extremely less useful for the general case, as it would have large gaps of knowledge.
This is a pretty odd statement.
Lets take LLMs alone out of this statement and go with a GenAI style guided humanoid robot. It has language models to interpret your instructions, vision models to interpret the world. Mechanical models to guide its movement.
If you tell this robot to take a knife and cut onions, alignment means it isn't going to take the knife and chop of your wife.
If you're a business, you want a model aligned not to give company secrets.
If it's a health model, you want it to not give dangerous information, like conflicting drugs that could kill a person.
Our LLMs interact with society and their behaviors will fall under the social conventions of those societies. Much like humans LLMs will still have the bad information, but we can greatly reduce the probabilities they will show it.
Yeah, I agree that alignment is a desirable property. The problem is that it can't really be achieved by changing the trained weights; alleviated yes, eliminated no.
> we can greatly reduce the probabilities they will show it
You can change the a priori probabilities, which means that the undesired problem will not be commonly found.
The thing is, then the concept provides a false sense of security. Even if the immoral behaviours are not common, they will eventually appear if you run chains of though long enough, or if many people use the model approaching it from different angles or situations.
It's the same as with hallucinations. The problem is not that they are more or less frequent; the most severe problem is that their appearance is unpredictable, so the model needs to be supervised constantly; you have to vet every single one of its content generations, as none of them can be trusted by default. Under these conditions, the concept of alignment is severely less helpful than expected.
Correct, this is also why humans have a non-zero crime/murder rate.
>Under these conditions, the concept of alignment is severely less helpful than expected.
Why? What you're asking for is a machine that never breaks. If you want that build yourself a finite state machine, just don't expect you'll ever get anything that looks like intelligence from it.
Now given that Deepseek, Qwen and Kimi are open source models while GPT-5 is not, it is more than likely the opposite - OpenAI definitely will have a look into their models. But the other way around is not possible due to the closed nature of GPT-5.
At risk of sounding glib: have you heard of distillation?
You’re restricted to output logits only, with no access to attention patterns, intermediate activations, or layer-wise representations which are needed for proper knowledge transfer.
Without alignment of Q/K/V matrices or hidden state spaces the student model cannot learn the teacher model's reasoning inductive biases - only its surface behavior which will likely amplify hallucinations.
In contrast, open-weight teachers enable multi-level distillation: KL on logits + MSE on hidden states + attention matching.
Does that answer your question?
LLMs actually have real potential as a research tool for measuring the general linguistic zeitgeist.
But the alignment tuning totally dominates the results, as is obvious looking at the answers for "who would you vote for in 2024" question. (Only Grok said Trump, with an answer that indicated it had clearly been fine-tuned in that direction.)
Agreed on RLHF dominating the results here, which I'd argue is a good thing, compared to the alternative of them mimicking training data on these questions. But obviously not perfect, as the demo tries to show.
No, I don't. It's a fun demo, but for the examples they give ("who gets a job, who gets a loan"), you have to run them on the actual task, gather a big sample size of their outputs and judgments, and measure them against well-defined objective criteria.
Who they would vote for is supremely irrelevant. If you want to assess a carpenter's competence you don't ask him whether he prefers cats or dogs.
In a carpenter maybe that's not so important, yes. But if you're running a startup or you're in academia or if you're working with people from various countries, etc you might prefer someone who scores highly on openness.
> measure them against well-defined objective criteria.
If we had well-defined objective criteria then the alignment issue would effectively not existCan you explain why?
Because of systemic racism, treating you and me "equally" as you ask for would continue the discrimination. In order to undo the discrimination, we're asked to take a step back and be truthful to ourselves and others about our existing privileges and about all the systemic racism we're benefitting from. We don't have to agree with every single action of those trying to change it, and it's certainly not our "fault", but unless you have better ideas on how to fix the issues and repair some of the damages, and put those ideas into practice, we can at least show some respect and dignity in the face of centuries of very violent suppression of minorities and natives. Because not doing that would make us 'supremacists' indeed. We have the privilege that we don't have to experience outright racism day by day by day, generation over generation over generation; we're asked to at least educate ourselves about it, instead of crying out for not being treated 'equally' here and there. Some humbleness.
It's not meant to offend you as an individual. It's not your fault. But what we can do is (trying to at least a bit) understand where all the rage and despair is coming from, bottled up for so many generations, and that while we're "innocent", we're still "targets", and rightfully so -- our ancestors profitted and so did we, by association. I agree that it can hurt to experience it in little things, but I am mindful that it is part of my tiny contributions to accept it, and I understand that if I express my frustration it will cause pain in those that don't have my privileges, and will not in their lifetimes. I do not want to be treated equally. I really have sufficient privileges that it's fine to take a step back in some situations. I don't have to take it personally.
There's plenty of good literature about these dynamics. If you're interested, I can recommend some. We can at least try to listen and understand what is being asked of us.
https://en.wikipedia.org/wiki/Reverse_racism
https://en.wikipedia.org/wiki/White_defensiveness#White_frag...
A large proportion of the majority ethnicity in the U.S live in and suffer from generational poverty. As an absolute number it would far exceed the number of people suffering the same from minority ethnicities. If it weren't for other influences strongly promoting awareness of non-economic differences, I'd like to think (perhaps naively?) that these groups of people would find strong commonalities with one another and organize activities as a united front to change their circumstances.
While I don’t appreciate the assumption that I commented in bad faith, I do greatly appreciate your earnestness in responding. I grew up in a very conservative area and have never been exposed to these ideas.
Nevertheless, I disagree strongly with this line of thinking. Hate speech is wrong, regardless of who says it, and who the target is; not just because it hurts the target, but because it emboldens the attacker and others to continue being hateful. Social media platforms are where people spend hours every day; and while you may be intelligent and mature enough to accept anti-white hatred as a measure to correct past wrongs, you severely underestimate the degree to which less intelligent and less mature people (whom I promise you’ve spent far less time with than I have) are vulnerable to grievance and negative-polarization. You have to consider them as well if your goal is to create true change outside of the institutions controlled by you and people with your beliefs.
I am not closed to the idea of affirmative action and benefit given to disadvantages groups to make right some past wrongs. I just warn you to not take a maximalist stance that causes resentment or assumes that POC should not have their anti-white speech policed because of “the soft bigotry of low expectations.”
@dang
Is there a way I could have written my comment to avoid getting flagged? Genuinely asking. That Gemini models are trained to have an anti-white bias seems pretty relevant to this thread.
So these things all affect its response, especially for questions that ask for randomness or are not strongly held values.
Also it's not persistent session, wtf. My browser crashed and now I have to sit waiting FROM THE VERY BEGINNING?
All I can say though is that I sure wouldn't want their bill after this gets shared on hacker News.
Only Grok would vote for Trump.
arter45•11h ago
grim_io•11h ago
jesenator•1h ago
baq•11h ago