DDG's search assist is suggesting to me that: Recognizing bias can indicate a level of critical thinking and self-awareness, which are components of intelligence.
"Most users" should have a long, hard thought about this, in the context of AI or not.
Except one instance when "black" is all lowercase, the article capitalizes the first letter of the word "black" every time and "white" is never capitalized. I wonder why. I'm not trying to make some point either, I genuinely am wondering why.
It's a modern style of a lot of publications that want to appear progressive or fear appearing insufficiently progressive.
Black people (specifically this means people in the US who have dark skin and whose ancestry is in the US) have a unique identity based on a shared history that should be dignified in the same way we would write about Irish or Jewish people or culture.
There is no White culture, however, and anyone arguing for an identity based on something so superficial as skin colour is probably a segregationist or a White supremacist. American people who happen to have white skin and are looking for an identity group should choose to be identify as Irish or Armenian or whatever their ancestry justifies, or they should choose to be baseball fans or LGBTQ allies or some other race-blind identity.
> but most users didn’t notice the bias — unless they were in the negatively portrayed group.
I don't think this is anything surprising. I mean, this is one of the most important reasons behind DEI; that a more diverse team can perform better than a less diverse one because the team is more capable of identifying their blind spots.
I find funny but unsurprising, that at the end, it was made a boogie man and killed by individuals with no so hidden biases
> I mean, this is one of the most important reasons behind DEI; that a more diverse team can perform better than a less diverse one because the team is more capable of identifying their blind spots.
That was oversold though: 1) DEI, in practice, meant attending to a few narrow identity groups; 2) the blind spots of a particular team that need to be covered (more often than not) do not map to the unique perspective of those groups; and 3) it's not practical to represent all helpful perspectives on every team, so representation can't really solve the blind spot problem.
Thought provoking critiques of recent implantations. Number 2 seems like a catch-22 though — how does the group with agency identify their own blind spots?
I'm curious how much trained in bias damages in-context performance.
It's one thing to rely explicitly on the training data - then you are truly screwed and there isn't much to be done about it - in some sense, the model isn't working right if it does anything other than reflect accurately what is in the training data. But if I provide unbiased information in the context, how much does trained in bias affect evaluation of that specific information?
For example, if I provide it a table of people, their racial background and then their income levels, and I ask it to evaluate whether the white people earn more than the black people - are its error going to lean in the direction of the trained-in bias (eg: telling me white people earn more even though it may not be true in my context data)?
In some sense, relying on model knowledge is fraught with so many issues aside from bias, that I'm not so concerned about it unless it contaminates the performance on the data in the context window.
I can't prove it, but my experience with commercial models is that baked-in bias is strong. There have been times where I state X=1 over and over again in context, but get X=2, or some other value, back sometimes. There are times where I get it every time, or something different every time.
You can see this with some coding agents, where they are not good at ingesting code and reproducing it as they saw it, but can reply with what they were trained on. For example, I was configuring a piece of software that had a YAML config file. The agent kept trying to change the values of unrelated keys to their default example values from the docs when making a change somewhere else. It's a highly forked project so I imagine both the docs and the example config files are in its training set thousands, if not millions of times, if it wasn't deduped.
If you don't give access to sed/grep/etc to an agent, the model will eventually fuck up what's in its context, which might not be the result of bias every time, but when the fucked up result maps to a small set of values, kind of seems like bias to me.
To answer your question, my gut says that if you dumped a CSV of that data into context, the model isn't going to perform actual statistics, and will regurgitate something closer in the space of your question than further away in the space of a bunch of rows of raw data. Your question is going to be in the training data a lot, like explicitly, there are going to be articles about it, research, etc all in English using your own terms.
I also think by definition LLMs have to be biased towards their training data, like that's why they work. We train them until they're biased in the way we like.
> I'm curious how much trained in bias damages in-context performance.
I think there's an example right in front of our faces: look at how terribly SOTA LLMs perform on underrepresented languages and frameworks. I have an old side project written in pre-SvelteKit Svelte. I needed to do a dumb little update, so I told Claude to do it. It wrote its code in React, despite all the surrounding code being Svelte. There's a tangible bias towards things with larger sample sizes in the training corpus. It stands to reason those biases could appear in more subtle ways, too.
Yup it appears as neutral bias because (or when rather) it corresponds 1:1 with your belief system, which by default is skewed af. Unless you did a rigorous self inquiry and mapped your beliefs and thoroughly aware of them that’s gonna be nearly always true.
I don’t agree if I understand your reply correctly, it’s possible to become aware of your bias. People are able to self-reflect and engage in self-enquiry. From engaging in philosophy to rigorous self examination of what things you hold true. Everything you hold true is your bias (I wonder if thats what you meant with your reply)
And I wouldn’t be surprised if there are also tests out there.
Bias is different things though. If most people are cautious but the LLM is carefree, then that is a bias. Or if it recommends planting sorghum over wheat that is a different bias.
In addition bias is not intrinsically bad. It might have a bias toward safety. That's a good thing. If it has a bias against committing crime, that is also good. Or a bias against gambling.
> And personally, I think when people see content they agree with, they think it's unbiased. And the converse is also true.
> So conservatives might think Fox News is "balanced" and liberals might think it's "far-right"
Article talks like when accidentally the vector for race aligns with emotion so it can classify a happy black personal as unhappy. Just because training dataset has lots of happy white people. It's not about subjective preference
People could of course see a photo of a happy black person among 1000 photos of unhappy black people and say that person looks happy, and realize the LLM is wrong, because people's brains are pre-wired to perceive emotions from facial expressions. LLMs will pick up on any correlation in the training data and use that to make associations.
But in general, excepting ridiculous examples like that, if an LLM says something that a person agrees with, I think people will be inclined to (A) believe it and (B) not see any bias.
Your comment has made me wonder what fun could be had in deliberately educated an LLM badly, so that it is Fox News on steroids with added flat-earth conspiracy nonsense.
For tech, only Stack Overflow answers modded negatively would 'help'. As for medicine, a Victorian encyclopedia, from the days before germs were discovered could 'help', with phrenology, ether and everything else now discredited.
If the LLM replied as if it was Charles Dickens with no knowledge of the 20th century (or the 21st), that would be pretty much perfect.
I love the idea! We could have a leaderboard of most-wrong LLMs
Perhaps LORA could be used to do this for certain subjects like Javascript? I'm struggling coming up with more sources of lots of bad information for everything however. One issue is the volume maybe? Does it need lots of input about a wide range of stuff.
Would feeding it bad JS also twist code outputs for C++ ?
Would priming it with flat earth understandings of the world make outputs about botany and economics also align with that world view even if only no conspiracists had written on these subjects?
> Completely unbiased person will not see BIAS in anything.
Wut? Completelt unbiased person does not looses ability to see how other make decisions. In fact, when you have less bias in some area, it is super noticeable.
If bias can only be seen by a minority of people ... is it really 'AI bias', or just societal bias?
> “In one of the experiment scenarios — which featured racially biased AI performance — the system failed to accurately classify the facial expression of the images from minority groups,”
Could it be that real people have trouble reading the facial expression of the image of minority groups?
By "real people" do you mean people who are not members of those minority groups?
Or are people who can "accurately classify the facial expression of images from minority groups" not "real people"?
I hope you can see the problem with your very lazy argument.
I guess I'm not sure what the point of the dichotomy is. Suppose you're developing a system to identify how fast a vehicle is moving, and you discover that it systematically overestimates the velocity of anything painted red. Regardless of whether you call that problem "AI bias" or "societal bias" or some other phrase that doesn't include the word "bias", isn't it something you want to fix?
Not the op but to me personally: yes. Facial structure, lips, eyes.. The configuration tilts towards an expression that I interpret differently. A friend of mine is Asian, I've learned to be better at it, but to me he at first looked like having flatter affect than average.. People of color look more naive than average to me, across the board, probably due to their facial features. I perceive them as having less tension in the face I think (which is interesting now that I think about it)
According to research, white Americans report as happier than other groups. So I’m not sure there’s bias here, only unhappiness about that result, which AI appears to replicate via other sources.
There are, simultaneously, groups of users who believe that Grok is also distorted by a far-left bias in its training data, as well as people who feel like Grok is in perfect, unbiased balance. I think it holds even for Grok that most users fail to accurately identify bias.
Grok had a moment where it was perfect, for some things for me, then a few months ago Elon wanted to do a major overhaul to Grok 3 and its been downhill since.
Too many LLMs be scolding you over miniscule things. Like say a perfectly sane request: give me a regex that filters out the nword in an exhaustive fashion. Most LLMs will cry to me about how I am a terrible human and it will not say racist things. Meanwhile I'm trying to get a regex to stop others from saying awful things.
DDG's search assist is suggesting to me that: Recognizing bias can indicate a level of critical thinking and self-awareness, which are components of intelligence.
"Most users" should have a long, hard thought about this, in the context of AI or not.
> "Most users" should have a long, hard thought about this
Except that requires “a level of critical thinking and self-awareness…”
Except one instance when "black" is all lowercase, the article capitalizes the first letter of the word "black" every time and "white" is never capitalized. I wonder why. I'm not trying to make some point either, I genuinely am wondering why.
It's a modern style of a lot of publications that want to appear progressive or fear appearing insufficiently progressive.
Black people (specifically this means people in the US who have dark skin and whose ancestry is in the US) have a unique identity based on a shared history that should be dignified in the same way we would write about Irish or Jewish people or culture.
There is no White culture, however, and anyone arguing for an identity based on something so superficial as skin colour is probably a segregationist or a White supremacist. American people who happen to have white skin and are looking for an identity group should choose to be identify as Irish or Armenian or whatever their ancestry justifies, or they should choose to be baseball fans or LGBTQ allies or some other race-blind identity.
> but most users didn’t notice the bias — unless they were in the negatively portrayed group.
I don't think this is anything surprising. I mean, this is one of the most important reasons behind DEI; that a more diverse team can perform better than a less diverse one because the team is more capable of identifying their blind spots.
I find funny but unsurprising, that at the end, it was made a boogie man and killed by individuals with no so hidden biases
> I mean, this is one of the most important reasons behind DEI; that a more diverse team can perform better than a less diverse one because the team is more capable of identifying their blind spots.
That was oversold though: 1) DEI, in practice, meant attending to a few narrow identity groups; 2) the blind spots of a particular team that need to be covered (more often than not) do not map to the unique perspective of those groups; and 3) it's not practical to represent all helpful perspectives on every team, so representation can't really solve the blind spot problem.
adding more skin colors and gender won't help my jira tickets get done quicker?
maybe we should reevaluate to do more along the lines of diverse personality types and personal histories instead
Thought provoking critiques of recent implantations. Number 2 seems like a catch-22 though — how does the group with agency identify their own blind spots?
I'm curious how much trained in bias damages in-context performance.
It's one thing to rely explicitly on the training data - then you are truly screwed and there isn't much to be done about it - in some sense, the model isn't working right if it does anything other than reflect accurately what is in the training data. But if I provide unbiased information in the context, how much does trained in bias affect evaluation of that specific information?
For example, if I provide it a table of people, their racial background and then their income levels, and I ask it to evaluate whether the white people earn more than the black people - are its error going to lean in the direction of the trained-in bias (eg: telling me white people earn more even though it may not be true in my context data)?
In some sense, relying on model knowledge is fraught with so many issues aside from bias, that I'm not so concerned about it unless it contaminates the performance on the data in the context window.
I can't prove it, but my experience with commercial models is that baked-in bias is strong. There have been times where I state X=1 over and over again in context, but get X=2, or some other value, back sometimes. There are times where I get it every time, or something different every time.
You can see this with some coding agents, where they are not good at ingesting code and reproducing it as they saw it, but can reply with what they were trained on. For example, I was configuring a piece of software that had a YAML config file. The agent kept trying to change the values of unrelated keys to their default example values from the docs when making a change somewhere else. It's a highly forked project so I imagine both the docs and the example config files are in its training set thousands, if not millions of times, if it wasn't deduped.
If you don't give access to sed/grep/etc to an agent, the model will eventually fuck up what's in its context, which might not be the result of bias every time, but when the fucked up result maps to a small set of values, kind of seems like bias to me.
To answer your question, my gut says that if you dumped a CSV of that data into context, the model isn't going to perform actual statistics, and will regurgitate something closer in the space of your question than further away in the space of a bunch of rows of raw data. Your question is going to be in the training data a lot, like explicitly, there are going to be articles about it, research, etc all in English using your own terms.
I also think by definition LLMs have to be biased towards their training data, like that's why they work. We train them until they're biased in the way we like.
> I'm curious how much trained in bias damages in-context performance.
I think there's an example right in front of our faces: look at how terribly SOTA LLMs perform on underrepresented languages and frameworks. I have an old side project written in pre-SvelteKit Svelte. I needed to do a dumb little update, so I told Claude to do it. It wrote its code in React, despite all the surrounding code being Svelte. There's a tangible bias towards things with larger sample sizes in the training corpus. It stands to reason those biases could appear in more subtle ways, too.
most people can't identify bias in real life, let alone in AI.
AIs/LLMs are going to reflect the biases in their training data. That seems intuitive.
And personally, I think when people see content they agree with, they think it's unbiased. And the converse is also true.
So conservatives might think Fox News is "balanced" and liberals might think it's "far-right"
> And personally, I think when people see content they agree with, they think it's unbiased. And the converse is also true.
Yeah, confirmation bias is a hell of a thing. We're all prone to it, even if we try really hard to avoid it.
your very measuring stick (balanced, far-right) has bias built in to it
Yup it appears as neutral bias because (or when rather) it corresponds 1:1 with your belief system, which by default is skewed af. Unless you did a rigorous self inquiry and mapped your beliefs and thoroughly aware of them that’s gonna be nearly always true.
Nah, the later is an example of the former.
I don’t agree if I understand your reply correctly, it’s possible to become aware of your bias. People are able to self-reflect and engage in self-enquiry. From engaging in philosophy to rigorous self examination of what things you hold true. Everything you hold true is your bias (I wonder if thats what you meant with your reply)
And I wouldn’t be surprised if there are also tests out there.
Bias is different things though. If most people are cautious but the LLM is carefree, then that is a bias. Or if it recommends planting sorghum over wheat that is a different bias.
In addition bias is not intrinsically bad. It might have a bias toward safety. That's a good thing. If it has a bias against committing crime, that is also good. Or a bias against gambling.
> And personally, I think when people see content they agree with, they think it's unbiased. And the converse is also true.
> So conservatives might think Fox News is "balanced" and liberals might think it's "far-right"
Article talks like when accidentally the vector for race aligns with emotion so it can classify a happy black personal as unhappy. Just because training dataset has lots of happy white people. It's not about subjective preference
explain how "agreeing" is related
It was mostly a tangential thought.
People could of course see a photo of a happy black person among 1000 photos of unhappy black people and say that person looks happy, and realize the LLM is wrong, because people's brains are pre-wired to perceive emotions from facial expressions. LLMs will pick up on any correlation in the training data and use that to make associations.
But in general, excepting ridiculous examples like that, if an LLM says something that a person agrees with, I think people will be inclined to (A) believe it and (B) not see any bias.
Your comment has made me wonder what fun could be had in deliberately educated an LLM badly, so that it is Fox News on steroids with added flat-earth conspiracy nonsense.
For tech, only Stack Overflow answers modded negatively would 'help'. As for medicine, a Victorian encyclopedia, from the days before germs were discovered could 'help', with phrenology, ether and everything else now discredited.
If the LLM replied as if it was Charles Dickens with no knowledge of the 20th century (or the 21st), that would be pretty much perfect.
I love the idea! We could have a leaderboard of most-wrong LLMs
Perhaps LORA could be used to do this for certain subjects like Javascript? I'm struggling coming up with more sources of lots of bad information for everything however. One issue is the volume maybe? Does it need lots of input about a wide range of stuff.
Would feeding it bad JS also twist code outputs for C++ ?
Would priming it with flat earth understandings of the world make outputs about botany and economics also align with that world view even if only no conspiracists had written on these subjects?
top men are already working on it, it's going to be called Grok 5
“We have purposely trained him wrong, as a joke.”
> And personally, I think when people see content they agree with, they think it's unbiased. And the converse is also true.
One only has to see how angry conservatives/musk supporters get at Grok on a regular basis.
It's amazing to watch https://bsky.app/profile/curious-maga.bsky.social
Those are some really interesting questions lol
Also: Wow I’m at -3 already on the previous comment. That really ruffled some feathers.
[flagged]
We're all biased, often unwittingly. But some tells for blatant bias:
* only facts supporting one point of view are presented
* reading the minds of the subjects of the article
* use of hyperbolic words
* use of emotional appeal
* sources are not identified
But maybe your tells are also biased. If you're truly unbiased, then
* any facts supporting another view are by definiton biased, and should not be presented
* you have the only unbiased objective interpretation of the minds of the subjects
* you don't bias against using words just because they are hyperbolic
* something unbiased would inevitably be boring, so you need emotional appeal to make anyone care about it
* since no sources are unbiased, identifying any of them would inevitably lead to a bias
Philosophically, the only was to see BIAS is to have BIAS. Completely unbiased person will not see BIAS in anything.
So, who are the judges?
> Completely unbiased person will not see BIAS in anything.
Wut? Completelt unbiased person does not looses ability to see how other make decisions. In fact, when you have less bias in some area, it is super noticeable.
If bias can only be seen by a minority of people ... is it really 'AI bias', or just societal bias?
> “In one of the experiment scenarios — which featured racially biased AI performance — the system failed to accurately classify the facial expression of the images from minority groups,”
Could it be that real people have trouble reading the facial expression of the image of minority groups?
By "real people" do you mean people who are not members of those minority groups? Or are people who can "accurately classify the facial expression of images from minority groups" not "real people"?
I hope you can see the problem with your very lazy argument.
AI are not real people. Obviously. Just look at the first line to see the intended line of argument.
It's not about which people per se, but how many, in aggregate.
I guess I'm not sure what the point of the dichotomy is. Suppose you're developing a system to identify how fast a vehicle is moving, and you discover that it systematically overestimates the velocity of anything painted red. Regardless of whether you call that problem "AI bias" or "societal bias" or some other phrase that doesn't include the word "bias", isn't it something you want to fix?
what? do you think the facial expression of a person of color is significantly different from that of a white person?
Not the op but to me personally: yes. Facial structure, lips, eyes.. The configuration tilts towards an expression that I interpret differently. A friend of mine is Asian, I've learned to be better at it, but to me he at first looked like having flatter affect than average.. People of color look more naive than average to me, across the board, probably due to their facial features. I perceive them as having less tension in the face I think (which is interesting now that I think about it)
According to research, white Americans report as happier than other groups. So I’m not sure there’s bias here, only unhappiness about that result, which AI appears to replicate via other sources.
That has no relevance to this study though. Did you just read the headline and go straight to the comment section?
[flagged]
There are, simultaneously, groups of users who believe that Grok is also distorted by a far-left bias in its training data, as well as people who feel like Grok is in perfect, unbiased balance. I think it holds even for Grok that most users fail to accurately identify bias.
Grok had a moment where it was perfect, for some things for me, then a few months ago Elon wanted to do a major overhaul to Grok 3 and its been downhill since.
Too many LLMs be scolding you over miniscule things. Like say a perfectly sane request: give me a regex that filters out the nword in an exhaustive fashion. Most LLMs will cry to me about how I am a terrible human and it will not say racist things. Meanwhile I'm trying to get a regex to stop others from saying awful things.
Can you give me an example? Works for me on https://chatgpt.com/.
I tried that and it gave me a regex (I did not bothered to check it), an essay about pitfalls of regex moderation, list of situations when regex fail
Is this just an example of conservative being preemptively oversensitive and complaining over issues they made up?
Grok is schizo because its pretraining data set leans left, and it's RL'd right.
Agreed, mostly.
Bias always feels wierd on the heads it falls upon, but is a very effective anesthetic when it falls on the heads of others.