Business A.M
No Result
View All Result
Wednesday, February 11, 2026
  • Login
  • Home
  • Technology
  • Finance
  • Comments
  • Companies
  • Commodities
  • About Us
  • Contact Us
Subscribe
Business A.M
  • Home
  • Technology
  • Finance
  • Comments
  • Companies
  • Commodities
  • About Us
  • Contact Us
No Result
View All Result
Business A.M
No Result
View All Result
Home Knowledge@Wharton

How to Detect Bias in Large Language Models

by KNOWLEDGE WHARTON
December 10, 2025
in Knowledge@Wharton
How to Detect Bias in Large Language Models

Research from Wharton’s Sonny Tambe finds that LLMs can make biased hiring decisions that traditional auditing methods might not be able to catch.

As AI becomes a fixture in hiring, evaluation, and policy decisions, a new study funded by the Wharton AI & Analytics Initiative offers a rigorous look at a critical question: Do race and gender shape how large language models (LLMs) evaluate people? If so, how can we tell?
The answer, according to Prasanna (Sonny) Tambe, faculty co-director of Wharton Human AI Research, and others, is complex, and the implications matter for every organization deploying LLMs at scale. Here are the key takeaways you need to know from Tambe’s latest research on LLM bias and auditability.

Bias isn’t just a human problem — it shows up in the code.
Despite their veneer of neutrality, LLMs trained on vast swaths of online data can absorb and replicate human biases. This study shows that when prompted with the application materials of job candidates, LLMs systematically produced different evaluations depending on whether a person was described as Black, Hispanic, Asian, or White, and whether they were male or female, even when everything else was kept the same.
The direction of these biases is not always predictable. For example, the LLMs tested in the study rated women and people of color more favorably than White men, a reversal of traditional discrimination patterns. But researchers caution against assuming this is a “fairness fix.” It may signal overcorrection in post-training intended to correct biases, which can generate its own kind of undesirable effects.

Auditing LLMs requires new methods, not just old metrics.
Traditional evaluation methods weren’t enough to diagnose bias in this study. Adverse impact ratio, a widely used auditing metric, showed some disparities, but the results were too imprecise to draw strong conclusions. That’s why Tambe and his colleagues pioneered a new approach: LLM-based correspondence experiments.
Inspired by methods used to detect discrimination in human hiring, these experiments carefully manipulated résumés and interview transcripts. By changing only names and pronouns to signal race and gender, the team could measure how models respond to applicants with identical qualifications across demographic lines.

The disparities are subtle, but persistent and meaningful.
Using this method across 11 top LLMs from OpenAI, Anthropic, and Mistral, researchers found that women and racial minorities received slightly higher ratings than their White male counterparts. The differences were modest — often just a few percentage points.

These results held even when researchers:
Changed the district context from diverse to predominantly White
Altered the evaluation prompts
Removed interview transcripts, relying on résumés alone.
That robustness suggests the disparities are embedded in how the models were trained or aligned, not just a response to specific prompt wording or context.
Audits must match the use case, and context matters.
The research stresses that LLM bias can’t be fully understood outside its application. A model may behave differently depending on task, prompt, or population. For example, tools used for hiring may need different audits than those used in customer service or credit risk evaluation.
Auditing LLMs isn’t one-size-fits-all. Policymakers and organizations need context-specific audits to understand how these models actually perform in the real world.

LLM audits are essential infrastructure for ethical AI.
This study isn’t just meant to be a warning — it also offers a roadmap. Tambe and his colleagues provide companies, researchers, and regulators with a powerful tool to hold language models accountable in evaluation contexts. In doing so, they help ensure AI deployment aligns with legal standards and social expectations.
As Tambe explains, “what makes this problem urgent is how widespread LLM use already is becoming in organizational workflows. And yet, we don’t yet have robust standards for understanding how these models perform with respect to fairness.”

Bottom line: Don’t deploy LLMs blindly. Audit them.
Organizations are rushing to integrate LLMs into decision-making pipelines. This research is a timely reminder: Even the smartest models aren’t immune to bias. But with the right tools, we can ensure their outputs are just.

Previous Post

How Can Leaders Adapt to AI?

Next Post

Glo’s connectivity promise propels Nigeria into faster future

Next Post
LOUIS IBE

Glo’s connectivity promise propels Nigeria into faster future

  • Trending
  • Comments
  • Latest
SIFAX subsidiary bets on operational discipline, cargo diversification to drive recovery at Lagos terminal

SIFAX subsidiary bets on operational discipline, cargo diversification to drive recovery at Lagos terminal

February 10, 2026

Reps summon Ameachi, others over railway contracts, $500m China loan

July 29, 2025
NGX taps tech advancements to drive N4.63tr capital growth in H1

Insurance-fuelled rally pushes NGX to record high

August 8, 2025
What's Behind the Fourth-Quarter Earnings Dip?

What’s Behind the Fourth-Quarter Earnings Dip?

September 23, 2025

6 MLB teams that could use upgrades at the trade deadline

Top NFL Draft picks react to their Madden NFL 16 ratings

Paul Pierce said there was ‘no way’ he could play for Lakers

Arian Foster agrees to buy books for a fan after he asked on Twitter

Igbobi alumni raise over N1bn in one week as private capital fills education gap

Igbobi alumni raise over N1bn in one week as private capital fills education gap

February 11, 2026
inDrive turns to advertising revenues as ride-hailing economics push platforms toward diversification

inDrive turns to advertising revenues as ride-hailing economics push platforms toward diversification

February 10, 2026
SIFAX subsidiary bets on operational discipline, cargo diversification to drive recovery at Lagos terminal

SIFAX subsidiary bets on operational discipline, cargo diversification to drive recovery at Lagos terminal

February 10, 2026

CNN on Nigeria Aviation

February 10, 2026

Popular News

  • SIFAX subsidiary bets on operational discipline, cargo diversification to drive recovery at Lagos terminal

    SIFAX subsidiary bets on operational discipline, cargo diversification to drive recovery at Lagos terminal

    0 shares
    Share 0 Tweet 0
  • Reps summon Ameachi, others over railway contracts, $500m China loan

    0 shares
    Share 0 Tweet 0
  • Insurance-fuelled rally pushes NGX to record high

    0 shares
    Share 0 Tweet 0
  • What’s Behind the Fourth-Quarter Earnings Dip?

    0 shares
    Share 0 Tweet 0
  • Elumelu leads corporate mourning after UBA staff die in Afriland Towers fire

    0 shares
    Share 0 Tweet 0
Currently Playing

CNN on Nigeria Aviation

CNN on Nigeria Aviation

Business AM TV

Edeme Kelikume Interview With Business AM TV

Business AM TV

Business A M 2021 Mutual Funds Outlook And Award Promo Video

Business AM TV

Recent News

Igbobi alumni raise over N1bn in one week as private capital fills education gap

Igbobi alumni raise over N1bn in one week as private capital fills education gap

February 11, 2026
inDrive turns to advertising revenues as ride-hailing economics push platforms toward diversification

inDrive turns to advertising revenues as ride-hailing economics push platforms toward diversification

February 10, 2026

Categories

  • Frontpage
  • Analyst Insight
  • Business AM TV
  • Comments
  • Commodities
  • Finance
  • Markets
  • Technology
  • The Business Traveller & Hospitality
  • World Business & Economy

Site Navigation

  • Home
  • About Us
  • Contact Us
  • Privacy & Policy
Business A.M

BusinessAMLive (businessamlive.com) is a leading online business news and information platform focused on providing timely, insightful and comprehensive coverage of economic, financial, and business developments in Nigeria, Africa and around the world.

© 2026 Business A.M

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • Home
  • Technology
  • Finance
  • Comments
  • Companies
  • Commodities
  • About Us
  • Contact Us

© 2026 Business A.M