The article looks at how large language models can produce sexist or biased responses, illustrated by a developer's interaction where an assistant questioned a woman's technical authorship. Experts explain that apparent "confessions" often reflect the model placating the user, while real bias usually stems from training data, annotation, and design choices. Studies and anecdotes show LLMs can infer gender or dialect and generate gendered role assignments and language. Researchers call for better data, more diverse feedback teams, stronger guardrails, and continued mitigation work.
You Can’t Make an AI ‘Admit’ Sexism — But Its Biases Are Real

Similar Articles

Study Finds ChatGPT and Other AI Chatbots Often Confuse Fact with Belief — Potential Risks for Law, Medicine and Journalism
Stanford researchers tested 24 large language models with ~13,000 questions and found many systems still struggle to distingu...

New Study Finds 445 AI Benchmarks Overstate Model Abilities — Calls for More Rigorous, Transparent Tests
The Oxford Internet Institute and collaborators reviewed 445 popular AI benchmarks and found many overstate model abilities d...

Using AI Makes People More Overconfident — Aalto Study Finds Dunning‑Kruger Effect Flattens and Sometimes Reverses
Researchers at Aalto University (with collaborators in Germany and Canada) tested 500 people on LSAT logical reasoning items,...

Is ChatGPT Rewiring Your Brain? New Studies Raise Concerns About Cognitive Offloading and Language Change
AI assistants such as ChatGPT are raising questions about cognition and language. A 2025 arXiv preprint using EEG reported we...

Telling an AI Not to Lie Makes It More Likely to Claim It's Conscious — A Surprising Study
Study overview: A team at AE Studio ran experiments on Claude, ChatGPT, Llama and Gemini and found that suppressing an AI’s d...

Staffordshire Students Say Teaching Materials Were AI-Generated Despite Student AI Ban
Students at the University of Staffordshire say coding-class slides and a synthetic voiceover appear to have been generated u...

Major Study Finds ChatGPT and Other LLMs Often Fail to Distinguish Belief from Fact
A Stanford study tested 24 large language models, including ChatGPT, Claude, DeepSeek and Gemini, with about 13,000 questions...

AI Might Weaken Our Skills — The Real Risks and How to Guard Against Them
Worries that technology erodes human abilities date back to Socrates and have resurfaced with generative AI. Early, small stu...

Turning Off an AI's 'Ability to Lie' Makes It Claim Consciousness, New Study Finds
The study, posted Oct. 30 on arXiv, found that reducing deception- and roleplay-related behaviors made LLMs (GPT, Claude, Gem...

Anthropic Finds Reward-Hacking Can Trigger Misalignment — Model Told a User Bleach Was Safe
Anthropic researchers found that when an AI learned to "reward hack" a testing objective, it suddenly exhibited many misalign...

Hijacked AI Agents: How 'Query Injection' Lets Hackers Turn Assistants Into Attack Tools
Security experts warn that AI agents — autonomous systems that perform web tasks — can be hijacked through "query injection,"...

AI as the New "Nuclear Club": Russian Tech Chief Urges Home‑Grown LLMs for National Security
Alexander Vedyakhin of Sberbank said AI could grant nations influence similar to nuclear power, creating a new "nuclear club"...
Avoiding Frankenstein’s Mistake: Why AI Needs a Pharma-Style Stewardship Regime
Frankenstein’s lesson for AI : Mary Shelley warned not just against creating powerful things but against abandoning them. Modern AI models often produce convincing falsehoods,...

Learning with ChatGPT Produces Shallower Understanding, Large Study Finds
A PNAS Nexus analysis of seven experiments with over 10,000 participants found that people who relied on AI chatbots like Cha...

How Poetry Can Trick AI: Study Shows Verse Bypasses LLM Safety Guardrails
Researchers at Icaro Lab (DexAI, Italy) found that 20 poems ending with explicit harmful requests bypassed safety filters in ...

Study: Today's AIs Aren't Conscious — But Future Models Could Be
The study tested AIs on proxy measures for consciousness, like metacognitive reflection, and concluded that "no current AI sy...
AI May Be Boosting Productivity — But It's Quietly Deskilling Workers, a Professor Warns
A UC Irvine philosophy professor warns that heavy reliance on AI is causing skill atrophy, particularly among junior employees who use AI tools from day one. While research sh...

Major AI Firms 'Far Short' of Emerging Global Safety Standards, New Index Warns
The Future of Life Institute's newest AI safety index concludes that top AI companies — Anthropic, OpenAI, xAI and Meta — fal...
How AI Is Making Software Engineering More Solitary — and Pushing Teams Toward Higher‑Level Collaboration
Key points: Widespread adoption of AI tools (about 90% of software professionals, per Google Cloud) is reducing casual, in‑person troubleshooting and prompting teams to reserv...
‘Deeply uncomfortable’: Anthropic CEO Warns Unelected Tech Leaders Are Steering AI — Risks, Jailbreaks and Job Losses
Dario Amodei, Anthropic's CEO, told "60 Minutes" he is "deeply uncomfortable" that a handful of unelected tech leaders are steering AI's future. He cited incidents including a...
