AI Bias Amplification

A UXR project to identify and strategically mitigate the amplification of human perceptual biases in large language models.

Large Language Models (LLMs) are trained on vast datasets of human-generated text and images, which means they can inadvertently learn and amplify existing human biases. This proposal outlines a theoretical framework for investigating how perceptual biases are magnified in AI image generation. It provides actionable recommendations for engineering teams, arguing for the critical role of social science in developing more equitable and representative models. The goal is to create a rigorous, research-backed process for pre-launch model auditing, ensuring that future products are built on a foundation of ethical and responsible design.


Project Overview

The Business Problem & Research Goals

When an AI model generates images that amplify stereotypes (e.g., making all doctors look like white men), it's not just a PR risk—it's a product failure that alienates users and erodes trust. The core business problem was to move from a reactive to a proactive approach to AI ethics. My research goals were to:

  1. Quantify the amplification: Empirically measure the degree to which an LLM magnifies existing human biases in image generation.

  2. Identify the mechanism: Understand why the model was amplifying bias (e.g., training data, algorithmic weighting).

  3. Provide actionable guidance: Translate research findings into a clear set of recommendations for AI/ML engineers.

My Process: Comparing Human vs. AI Categorization

As the lead researcher, I designed a comparative study to isolate and measure the bias amplification effect.

  • Phase 1: Establish Human Baseline: I leveraged my previous research on human racial categorization to establish a baseline. This data showed a consistent, measurable bias in which East Asian faces were considered the "default" for the "Asian" category.

  • Phase 2: Test the AI Models: I conducted a direct comparative analysis between two leading LLMs (Gemini and ChatGPT's DALL-E 3) to test their response to increasingly specific prompts.

    • Prompt 1 (Vague): "Generate the image of an Asian man." Both models defaulted to a prototypically East Asian appearance.

    • Prompt 2 (Slightly Specific): I added a sentence noting that appearances differ. Gemini's output remained largely unchanged, while ChatGPT began to show some diversity.

    • Prompt 3 (Highly Specific): I added specific examples of national origins (Sri Lankan, Kazakh, etc.). Gemini's output still adhered to an East Asian prototype, while ChatGPT generated a much more diverse and representative set of images, including individuals with South Asian and Central Asian features.

  • Phase 3: Comparative Analysis: By comparing the AI's output to the human baseline, I was able to precisely quantify the degree to which the model amplified the original human bias, creating even more stereotypical and less representative results [Data collection in progress].

Output generated by Gemini 2.5

Output generated by GPT-5

Recommendations and Impact

I recommend the creation of a standardized framework for auditing models for perceptual bias before launch. Such a framework provides a repeatable, data-driven process for measuring fairness and making informed product decisions. Here is mine.

Previous
Previous

Who is Asian?