A new study provides a geometric picture of what happens inside AI language models when bias-mitigation techniques are applied, finding that fairness improvements leave a consistent, measurable imprint on model embeddings across two major architecture families.

The paper, posted to arXiv in April 2025, examines encoder-only and decoder-only models — represented by BERT and Meta's Llama 2 — and compares their baseline internal representations against bias-mitigated versions. The central question: does reducing bias on the surface actually change anything deeper in the model?

What the Researchers Actually Measured

The study focuses on embedding space — the multidimensional numerical landscape that a language model uses to encode meaning. Words and concepts that a model treats as similar end up close together in this space; those it treats as unrelated sit further apart. By measuring how gender terms and occupation terms cluster before and after debiasing, the researchers traced the structural impact of bias mitigation from the inside out.

According to the paper, bias mitigation consistently reduced gender-occupation disparities in the embedding space of both model types, producing what the authors describe as more neutral and balanced internal representations. Crucially, these shifts were geometrically interpretable — meaning they appeared as systematic, structured changes rather than random noise.

These representational shifts are consistent across both model types, suggesting that fairness improvements can manifest as interpretable and geometric transformations.

The consistency across both encoder-only (BERT) and decoder-only (Llama 2) architectures is notable. These two model families differ substantially in design and use case, yet both showed analogous internal changes when debiased — suggesting that embedding analysis could serve as a general-purpose auditing tool rather than an architecture-specific one.

The Gap in Evaluating Decoder-Only Models

One practical obstacle the researchers identified: existing benchmark datasets for measuring gender-occupation bias were largely designed for encoder-only models. Decoder-only models — now the dominant architecture behind tools like ChatGPT, Claude, and Llama — process text differently, making direct application of older benchmarks imprecise.

To address this, the team introduces WinoDec, a dataset of 4,000 sequences containing paired gender and occupation terms, built specifically to evaluate decoder-only models. The dataset has been released publicly at github.com/winodec/wino-dec, giving other researchers a standardised tool for replicating and extending this type of representational audit.

The name echoes the long-established Winograd and WinoBias benchmarks, which have been used for years to probe coreference resolution and occupational bias in NLP systems. WinoDec positions itself as a next-generation counterpart suited to the current generation of generative models.

Why Internal Audits Matter for AI Accountability

The AI industry has invested heavily in bias mitigation, with techniques ranging from fine-tuning on curated datasets to reinforcement learning from human feedback. But a persistent criticism of these methods is that they can suppress biased outputs without altering the underlying representations — effectively teaching a model to hide rather than correct problematic associations.

This study addresses that concern, at least partially. According to the authors, the geometric shifts they observed suggest that effective debiasing does penetrate to the representational level, not merely the output layer. However, the paper does not claim that all debiasing methods achieve this — the research compares baseline models against already-mitigated variants without evaluating every available debiasing technique.

It also bears noting that the benchmarks and evaluations described in the paper are self-reported by the researchers, and the work has not yet undergone formal peer review as of its arXiv posting.

Beyond Gender and Occupation

The methodology the paper develops — treating embedding geometry as an audit surface — has implications that extend well beyond the specific gender-occupation associations studied here. Internal representation analysis could, in principle, be applied to other bias dimensions: race, age, nationality, or religion. It could also serve as a monitoring tool in deployment, flagging when a model's internal associations drift from an approved baseline.

For regulators and enterprise AI teams increasingly required to demonstrate how a model was made fair, not just that it was, this kind of interpretable internal evidence could become practically valuable. The EU AI Act and similar frameworks are beginning to demand documentation of bias-mitigation processes — and geometric audits of embedding space offer a form of evidence that is both technical and explainable.

The release of WinoDec as a public resource also lowers the barrier for independent researchers to verify or challenge these findings, which matters in a field where reproducibility has been an ongoing concern.

What This Means

Organisations building or auditing AI systems now have a concrete, architecture-agnostic method for verifying that bias mitigation has actually changed a model's internal wiring — not just its visible outputs — and a new public dataset specifically designed for the decoder-only models that dominate today's AI landscape.