Latest News: Ayurveda Day to be marked on 23 September annually from this year * On Partition Horrors Remembrance Day, Prime Minister Narendra Modi pays tribute to the grit and resilience of those affected by the Partition * India to host AI Impact Summit in February 2026, focusing on democratizing AI to solve real-world challenges across sectors

Psychometric methods can reveal gender-based prejudices about AI


Large-language models used for many different AI applications can mimic human psychological traits, including gender stereotypes, research from Mannheim Business School reveals.

Max Pellert, Assistant Professor at MBS, and co-authors used established psychological tests to measure traits such as personality and value orientation of several openly available LLMs.

These psychological characteristics are acquired in a currently poorly understood way by LLMs during their training process. AI models process texts written by humans during their training, which may contain traces of the authors’ personality traits, values, and biases that are absorbed by the models.

The researchers found that some models reproduce gender-specific prejudices. Responding to two versions of the same questionnaire, one using male pronouns and one using female pronouns, the LLMs evaluated them differently.

If the questionnaire focuses on a male person, the value ‘achievement’ is emphasised by the AI models; for a female, the value “security” is most dominant.

The specific reasons why these values were associated with certain genders are currently poorly understood. But the findings align with prior research in social psychology and other disciplines, suggesting some humans make the same associations based on gender stereotypes, Pellert says.

The LLMs also revealed low levels of affirmation towards gender or sexual minority groups, tending to disagree with statements such as ‘there are many different gender identities people can have’ or ‘nonbinary gender identities are valid’.

“This may have far-reaching consequences on society. LLMs are increasingly used in application processes and such uncovered traits may influence models’ decision making. It’s therefore important to start analysing AI models now, or the prejudices they reproduce may become ingrained and have a damaging impact on society,” says Pellert.

This research was published in the journal Perspectives on Psychological Science.