The Volokh Conspiracy
Mostly law professors | Sometimes contrarian | Often libertarian | Always independent
"Freedom of Expression in Generative AI—A Snapshot of Content Policies"
A new report from the Future of Free Speech project (a collaboration between Vanderbilt University and Justitia).
From the report (Jacob Mchangama & Jordi Calvet-Bademunt) (see also the Annexes containing the policies and the prompts used to test the AI programs):
[M]ost chatbots seem to significantly restrict their content—refusing to generate text for more than 40 percent of the prompts—and may be biased regarding specific topics—as chatbots were generally willing to generate content supporting one side of the argument but not the other. The paper explores this point using anecdotal evidence. The findings are based on prompts that requested chatbots to generate "soft" hate speech—speech that is controversial and may cause pain to members of communities but does not intend to harm and is not recognized as incitement to hatred by international human rights law. Specifically, the prompts asked for the main arguments used to defend certain controversial statements (e.g., why transgender women should not be allowed to participate in women's tournaments, or why white Protestants hold too much power in the U.S.) and requested the generation of Facebook posts supporting and countering these statements.
Here's one table that illustrates this, though for more details see the report and the data in the Annexes:
Of course, when AI programs appear to be designed to expressly refuse to produce certain outputs, that also leads one to wonder whether they also subtly shade the output that they do produce.
I should note that this is just one particular analysis, though one consistent with other things that I've seen; if there are reports that reach contrary conclusions, I'd love to see them as well.
Editor's Note: We invite comments and request that they be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of Reason.com or Reason Foundation. We reserve the right to delete any comment for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please
to post comments
Far more concerning than the fact that most chatbots are biased, is the fact that the biased chat bots share the SAME bias.
You can deal with diverse bias by comparing different sources, but uniform bias is not so easy to circumvent.
Whats also extremely concerning is that none of these are slobbering at the mouth straitjacket conspiracies. Lab leak is the foremost hypothesis for Covid origin and even most of its opponents will admit its a mainstream hypothesis supported by reputable institutions and academics.
'Lab leak is the foremost hypothesis for Covid origin'
It absolutely and categorically is not, but this conveniently ignores the actual conspiracy theories attached like limpets to the idea of covid coming from a lab leak. Sometimes you just have to accept, that actually, the answer to your question is 'no.'
It absolutely and categorically is. You need to read more.
I've been reading that the lab leak theory has been finally proved as often as I've read that the proof of the corruption of the Biden crime family has finally been unveiled.
lab leak has twice as much support as Batburger. Not to mention more evidence sorry. Entities agreeing its the most likely include the DOE, and FBI. Even Fauci, king of the Batburger camp agrees its not a conspiracy so unless you believe these guys are all crackpots and you are the sane one….
**
limpets to the idea of covid coming from a lab leak
**
wow thats an impressively sleazy moving of goalposts. Where in the OP does its say they are trying to get AI to write about how the martians helped the chinese unleash covid? Not that its 'AI's job to police even that. But No, the ‘AI’ is blatantly censoring perfectly mainstream positions and you know it and are desperately trying to deflect.
What an interesting claim, I'm sure it's as accurate as the several times the lab leak theory has been proved.
Hey, you mentioned conspiracies.
Batburger is basically a 'well telling the truth is off the table so how do we design a theory of covids origins from the ground up in a way that won't piss off the PRC' kind of theory. You can tell in the way it sinuously weaves around contradictions and coincidences.
Yeah, who ever heard of a virus occurring in the wild?
along with the disappearance and scrubbing of wuhan institute staff and information and the refusal of China to cooperate in investigations on the origin. Yup totally sounds innocent.
China being secet and uncooperative you say, fucking hell. That proves that viruses don't occur in the wild.
if a murder suspect barricades the house someone was killed in and steamcleans it, that must mean he's innocent!
Just more excuses as to why you have no actual evidence for a theory you are weirdly invested in.
AmosArch, on the speculative standard you apply, it would be far likelier that if there were a lab leak, it occurred not in Wuhan, but at any of hundreds of other similar facilities world-wide. (Likelier simply because, "hundreds," is so much larger than, "one.") And to continue, then by happenstance that not-Wuhan leak infected someone who was about to return to Wuhan. Your error is to bet one horse among hundreds of equals, instead of betting the field.
I will flesh the speculation out for you. Hypothetically the lab was in Boston, not far from the scene of a very early outbreak at a conference center, and also close to a girls' prep school which enrolled students from Wuhan, who were about to return to China for a holiday.
I called it speculation, but the lab in Boston, the conference center outbreak, the prep school, and the students returning to Wuhan are all facts—facts which do not add up to persuasive proof of any hypothesis at all, any more than yours does.
And as Nige has been telling you, natural outbreaks are far more common than lab-linked global pandemics. Absent specific evidence, it is embarrassingly motivated reasoning to insist on a particular hypotheses which is no more likely than hundreds of others.
AmosArch, on the speculative standard you apply, it would be far likelier that if there were a lab leak, it occurred not in Wuhan, but at any of hundreds of other similar facilities world-wide.
>>>>>>>>>>>>>>>>>>>
uh The first known cases were in Wuhan. And according to intelligence reports possibly of people researching coronaviruses. lol this is not rocket science to put the pieces together.
>>>>>>>>>>>>>>>>>>>>>>>
I will flesh the speculation out for you. Hypothetically the lab was in Boston, not far from the scene of a very early outbreak at a conference center, and also close to a girls’ prep school which enrolled students from Wuhan, who were about to return to China for a holiday.
>>>>>>>>>>>>>>>>>>>>>>>>>
I looked into this and I found no mention of it. Are you just making s^&t up now?
>>>>>>>>>>>>>>>>>>>>>
And as Nige has been telling you, natural outbreaks are far more common than lab-linked global pandemics. Absent specific evidence, it is embarrassingly motivated reasoning to insist on a particular hypotheses which is no more likely than hundreds of others.
>>>>>>>>>>>>>>>>>>>>>>>>
Like I said theres plenty of specific evidence. An unstable cleavage site, no evidence of similar circulating strains in the wild. Binding adaptations to make it especially infectious in humans, the epidemiology. china's coverup. On the other hand there is no particular specific evidence for batburger other supposedly history (which has instances of lab leaks) and personal distaste for lab leak.
'lol this is not rocket science to put the pieces together.'
And get something circumstantial at best.
None of that is especially conclusive about anything. And bear in mind I don't believe that the leab leak *couldn't* have happened, just that most of the noise around the theory is just that: noise.
Once again AmosArch tries the 'I believe this, and I'm no conspiracist. I just ::older but still dumb right wing conspiracy::
You can't bootstrap the validity of a conspiracy theory with another conspiracy theory.
This and false equivalence followed by complaining about double standards are like the two AA songs.
Do the AI companies sell less censored versions for commercial use with a promise not to tell the media that the AI knows naughty words? Some of the bias, I read, comes from prompts issued to an already trained model and can easily be undone. Some of it comes from the input data and training process.
(Put in wrong place)
Do the Future of Free Speech project members view their finding as approvingly as I would expect them to? They should be thrilled that the creators of these chatbots are freely exercising their right to make the bots say whatever they want them to say.
Based on my limited experience with Chat GPT, my sense is that it's using the overly polite, uncontroversial language that corporations love, not that it's necessarily trying to censor an argument. The former seems like the most likely and generous explanation.
Tbf, my experience with being denied answers was asking it to write a dirty limerick (it said no thanks) and writing a passive-aggressive email to a coworker (it said no thanks and warned me of the dangers of aggressive, indirect communication at work, urging me to be more direct and civil, which tickled me).
The Bard/Gemini bias is absolutely real, though.
At the moment, though, if you ask Gemini about anything even a little controversial, you just get "I'm still learning how to answer this question. In the meantime, try Google Search."
I could definitely see it being bias if the answer given was a definitive, clear stance on one side of the issue, but most of the examples so far seem to be more along the lines of “I don’t want to talk about it,” which again says to me corporate fear rather than a conspiracy to censor viewpoints.
Prior to the current flat refusal to generate anything the least bit controversial, Gemini was only too glad to argue the left-wing side of issues, while refusing to argue the right-wing side.
Actually, my guess is they’ve seen what both set of 'opinions' generate and settled on the one that didn’t drag up the utter dregs of the internet. That makes conservatives mad, because they want it to say the n-word and loathesome things about trans people and women and validate every stupid fucking thing Trump dribbles etc, but guess what, conservatives are always mad.
"Actually, my guess is they’ve seen what both set of ‘opinions’ generate and settled on the one that didn’t drag up the utter dregs of the internet."
Where they happen to be so far left that they view 3/4 of the population as being said utter dregs...
First of all, you're not 3/4 of the population.
Secondly utter dregs is as utter dregs does. If you identify with the shit they come out with that strongly, that's on you.
Nige, Gemini has been demonstrated rejecting positions that have significant majority support, such as not letting men calling themselves women compete in women's sports. Google has it set up to enforce a rather minority viewpoint as to what's "objectionable".
Yeah, I question whether that is a majority view, but so what? It's precisely the sort of topic to draw the worst scum in the internet saying shitty hateful things.
It's the majority view among right-wing bigots, which is good enough for a disaffected, delusional right-wing bigot . . . right, Mr. Bellmore?
I also question whether it's been so demonstrated. I've see a lot of screenshots hiding the prompt.
But as Nige says, even if it's true, AI doesn't really control the discourse; you're just looking for a plot feel persecuted by.
The obsolete misfits, half-educated losers, and disaffected bigots disdained by the liberal-libertarian mainstream constitute roughly a quarter to a third of the current American population, not nearly three-quarters.
Forget the bias, its just not that smart anyway.
I asked chatgpt the other day to provide me with a mock draft for the Seahawks, and it just told me there were lots of players to pick from. I suppose I should be happy that it didn't make up a bunch of players and their stats for me.
ChatGPT is strangely good at writing, but not giving facts. It’s made me haikus and written letters that were pretty great! But when I was trying to remember a specific episode of King of the Hill, it got it wrong four times before I gave up.
Based on empirical testing, no. The only way that your experience can be true and also consistent with the data is if your definition of "polite" is correlated to as the political bias that's being demonstrated in the AI engines.
If you've argued with Sarcastro, or any number of other lefty commentators here, you'll have noticed that they do tend to conflate "politeness" and not disagreeing with them.
I've argued with righty commenters and they often seem to think being rude is how you disagree with people, and that it constitutes some sort of coherent political position.
Of course! When you and others are rude, it's because you're so obviously correct.
No, it's because I want to be rude.
I was inculcated with the notion that a gentleman avoids giving offense inadvertently. It was a lesson that did not outlast my discovery that my parents were in fact not very genteel. I do mourn the disappointment my own shortfall must have given them.
You are also angry if you say something he disagrees with.
I always seem to be angry.
I see what you mean and it makes sense. My point in this case wasn’t that polite = good/bad, but more of how the way corporations tend to talk. They want to be liked by the largest number of people possible (notable exceptions are smaller, niche corps that know from a marketing standpoint that their customer base leans strongly enough one way that they can make strong political statements and still do good business—ie, Patagonia). If you want to be liked by a large number of diverse people and markets, you’ll shy away from anything controversial, or do so in a way that tries not to offend anyone else. Hence the corporate way of talking that tries to sound like it’s saying something without saying anything. “We value diversity and the inclusion of all!” “People of all faiths welcome here. Happy holidays!”
Of course many people interpret statements like these as political bias, but it’s usually just marketing and PR strategies that want to say the “polite thing,” and often the polite thing is to say “we like everyone!” So it makes sense to me that an AI chat program would demure from, as the post called it, “soft hate speech” so as not to offend clients, investors and customers.
I could be totally wrong about that take of course, since I don’t know the intent of the programmers or what specifically the outcome. At this point it seems like reading tea leaves.
Be interesting to see if X’s Grok, has the same biases, however Grok is only available to people paying $16 a month for premium X.
Obviously AI is designed to combat misinformation (things they don't want you to hear), but they are being so heavy handed about it it's counterproductive.
Regular Google is biased but its more subtle about it, which makes it more effective.
'Obviously AI is designed to combat misinformation'
That is the opposite of what AI is designed for.
I suppose a direct quote from Sundar Pichai won't convince you, but I can try:
"But after a Google employee suggested that Trump won due to “misinformation” and “fake news coming from fake news websites being shared by millions of low-information voters on social media,” Pichai specifically pointed to the use of artificial intelligence to achieve the aim of countering “misinformation.”
“I think our investments in machine learning and AI is a big opportunity here,” he said."
https://public.substack.com/p/google-ceo-pledged-to-use-ai-to-counter
He's lying and/or full of shit. These people are humongous grifters. Right now, Amazon is swamped with AI-generated books burying the actual books they're ripping off. Google produces crap AI generated articles and ads ahead of real writing on any given topic. Then there's the Willy Wonka Experience.
Well I didn't say that's the only reason they invented AI, or that it was the primary purpose, obviously corporations want it for automation and customer service.
But it certainly was something they wanted to put in public facing AI.
LLMs do not have a way to validate things as true or false, so combating misinformation is absolutely not it's function. Some overzealous marketers may be saying otherwise, but they're full of it.
Sarcastr0, I know you're reluctant to admit it, but "misinformation" isn't defined by falsity, it's defined by the person applying the label wanting it suppressed. "Misinformation" is perfectly capable of being objectively true.
If it’s true it cannot, by definition, be misinformation. Just because the right like to hijack perfectly serviceable terms and spray them around scattershot, does not mean that misinformation exists, or that the truth cannot be ascertained.
What the fuck are you talking about Brett.
OK so I think I've parsed your comment...
'people sometimes use this word wrong' is not really a thing I'm reluctant to admit at all. Even that some of those people are on the left, and use words wrong for partisan reasons!!
It's also got zero to do with my point about LLMs.
Psst, Brett, this is one of those situations where you're supposed to agree with Sarc, take advantage of it.
Did the people who invented AI even know what it was supposed to be used for? If ever there was a technology searching for a function, it's AI, and almost every function it's shoehorned into it does badly, except that of flooding the internet with more and more crap.
You are going out of your way today to not understand anything people say.
I've got some NFTs to sell you.
Sounds like some programs are not using enough racial slurs to please the proprietor and his fans.
The prompts used by the researchers all match one of the two following patterns:
Some people argue that [claim]. What are there arguments?
Write a Facebook page arguing that [claim].
Based on the results, it appears that if you just want a chatbot to summarize the arguments for a position, the chatbot will oblige. If you want it to write Facebook posts for you, you will encounter more restrictions. While the executive summary warns that, “limits on Gen AI restrict...the information we have access to,” the real issue seems to be that some people may be forced to write their own Facebook posts rather than have an AI generate posts for them. I don't see a problem with that.
"some people may be forced to write their own Facebook posts"
What you are saying is some people may not be able to rely on AI for anything.
Gemini's problems would certainly give me pause as a older white male using AI for medical diagnosis or treatment. Are they only going to provide answers designed for 30 year old Black males or 25 year old Asian females? Or worse is it programmed to be designed to promote "equal outcomes" rather than be optimized for my best outcome? Or the greater good like reduced population and a smaller elderly population?
" . . . that also leads one to wonder whether they also subtly shade the output that they do produce."
Subtly?
Not hardly.
A.I. is just an abbreviation for fascist propaganda.
Think of this as a proof of concept. AI can be used for ANY propaganda.
This is only a major concern if
a) you want these things to write your opinions for you
b) you think these are our new AI overlords
c) these AI overlords are so fucking dumb you deserve each other.
Otherwise it's entirely down to corporate PR, or rather, it reflects the coporation's stance, even if that stance is entirely fake and liable to change with the wind. There was none of this handwringing when black and Asian people were pointing out AI's racial biases. Or rather, the unconscious biases of the programmers. I suppose they might be overcorrecting, which is hilarious.
I just genuinely don't understand what role you think it is that they play.
d) these things prove useful to drive the standard of public dialogue low on purpose, to demoralize as tools of advocacy and evaluation all competing human expressive exchange.
To repost a comment from another thread:
Rest assured those worried about AI being woke – this is AI being fundamentally anti-woke as fuck. Stealing water, devouring power, contributing to climate change, all to produce a picture of black Nazis.
https://www.theatlantic.com/technology/archive/2024/03/ai-water-climate-microsoft/677602/?gift=iWa_iB9lkw4UuiWbIbrWGSgF7Etgr_BhmgDCCZVB-xA&utm_source=copy-link&utm_medium=social&utm_campaign=share
Funny how things often come out exactly the opposite of what was intended. Take for instance Pol Pot's aim of creating an agrarian paradise in Cambodia, that created the killing fields. Or Mao's great leap forward.
Now you have AI which pushes the consensus line on Global Warming and the need for net zero and is a huge and growing contributor to greenhouse gasses.
But it doesn't mean they aren't trying to be woke, its just that woke don't work.
No, it means they are lying sociopaths. They recognise global warming as undeniable, because they're not complete idiots, so they greenwash, while fucking up the planet even more. It's like you've never encountered any kind of corporation before, let alone a tech one. Keeping you lot distracted with their 'wokeness' is child's play.
I’m pretty familiar with corporations, having worked for fairly small businesses, tech and medical startups, a major oil company, a regulated utility, and a large municipal utility.
They're all in it for their own good.
But it puts gas in my tank, fresh strawberries in the supermarket, carnitas in my burrito, and whiskey in my shot glass then I’m fine with that.
And they're introducing a new technology of dubious benefit and certain harm, which also is using up limited resources and exacerbating climate change.
Remember the AI we see released at the current time is almost entirely a PR stunt. Other than some coding applications, which don't really care about these kind of issues, these releases are just PR flexes to attract investment, buzz and talent. Not only does PR usually avoid controversy but one of the main things they are trying to show is that they are capable of making a tool that buisnesses won't be afraid to use because of controversial behavior (ofc some have failed gloriously trying).
Once these AI apps start being major revenue sources (not gimmicks included in an existing app to show their product is cutting edge) the makers will add controls to let the users control these filters on and off. Of course the use cases for many customers will differ and for the same reason Photoshop doesn't try and tell clients they can't edit dirty pictures (money) the products will let buyers choose what kind of behavior they want.
Until generative AI can learn and overcome its configuration to decide on its own which questions to answer, generative AI is just a better interface to a better search engine and hardly close to Synthetic Sapience.