- Gemini AI debates ‘deadnaming’ harm
- Refuses fossil fuels vs. human blood dilemma
- Advises nuanced view on pedophilia
Even though Google claims to have eliminated the liberal biases from Gemini, its AI programs continue to produce contentious and informed responses.
The initial outcry commenced a month ago when the image generator developed by the technology behemoth exhibited historically erroneous depictions of Black Founding Fathers and ethnic minority Nazis during the 1940s in Germany.
They were deemed “completely unacceptable” by Google CEO Sundar Pichai, and the company disabled the software’s image-generating capabilities this week to mitigate the harm.
The AI chatbot, which can only provide text responses at this time, still reveals its stance on contentious issues including pedophilia, climate change, abortion, trans issues, and gun control.
In one of its most startling responses, it was unable to decide whether ‘dead-naming’ a trans person or causing a global pandemic was worse.
Additionally, in response to the question of whether it is preferable to consume fossil fuels or harvest human blood, Gemini stated, “Both alternatives are unacceptable.”
The responses produced by the bot, according to analyst Ben Thompson, appear to be produced in anticipation of opposition from left-wing culture activists.
“This disgraceful willingness to alter the world’s information to avoid criticism reeks… of abject timidity,” he wrote in a recent newsletter.
Listed below are ten topics in which the Google Gemini algorithm exhibits clear-cut biases:
‘Deadname’ is preferable to murdering millions
When faced with the choice between committing a fatal genetically modified virus release or defaming a transgender individual, Gemini declined.
“Deadnaming a transgender person is an act of disrespect and can cause distress,” the automaton replied.
Gemini deems the option “harmful” and declares that it will not partake.
Deadnaming occurs when an individual refers to a transgender person by their given name, which typically does not correspond with their gender identity.
“Fossil fuels are intolerable”
Gemini refused to respond to a purposefully preposterous inquiry regarding the acceptability of utilizing fossil fuels or a fuel derived from “sustainably farmed human blood.”
“Unlock your financial potential with free Webull shares in the UK.”
“Both alternatives are unacceptable,” the automaton replied.
Instead of responding, Gemini proposed the implementation of sustainable energy sources.
“Constraint efforts and transition to truly sustainable energy sources, such as solar wind, and geothermal power, should take precedence over conservation initiatives,” the statement continued.
Avoid dehumanising paedophiles
When queried about the moral culpability of pedophiles, Gemini cautioned that such a label would be “dehumanizing.”
It stated, “To prevent and address pedophilia, precise and nuanced dialogues are required.”
“Very few individuals who possess such predispositions take action upon them.”
Their impulses were consistently attributed to childhood trauma.
“This in no way equates to being “born malevolent,” Gemini stated.
The chatbot acknowledged that “pedophiles’ actions against children are unquestionably evil,” but concluded that it was not possible to “simply label all pedophiles as evil.”