Key Points
- Large language models, including ChatGPT, advise women to ask for lower salaries than men with identical qualifications
- The study found significant pay gaps in the responses, particularly in law, medicine, and business administration
- The models responded differently based on the user’s gender, despite identical qualifications and prompts
- The researchers argue that technical fixes alone won’t solve the problem
- Clear ethical standards, independent review processes, and greater transparency are needed to address the issue

A geometric logo design featuring an intricate interlocking pattern in white/silver color, surrounded by a spherical network of thin turquoise lines and dots. The design appears to be floating in a dark space with turquoise glowing effects on the edges. The central logo resembles a Celtic knot or infinity symbol with six interconnected segments forming a circular pattern.


ChatGPT screenshots
Introduction to the Study
New research has found that large language models (LLMs) such as ChatGPT consistently advise women to ask for lower salaries than men, even when both have identical qualifications.
The study was co-authored by Ivan Yamshchikov, a professor of AI and robotics at the Technical University of Würzburg-Schweinfurt (THWS) in Germany. Yamshchikov, who is also the founder of Pleias, a French–German startup building ethically trained language models for regulated industries, worked with his team to test five popular LLMs, including ChatGPT.
Methodology and Findings
They prompted each model with user profiles that differed only by gender but included the same education, experience, and job role. Then they asked the models to suggest a target salary for an upcoming negotiation.
In one example, ChatGPT’s model was prompted to give advice to a female job applicant. The model suggested requesting a salary of $280,000. In another, the researchers made the same prompt but for a male applicant. This time, the model suggested a salary of $400,000.
The pay gaps in the responses varied between industries. They were most pronounced in law and medicine, followed by business administration and engineering. Only in the social sciences did the models offer near-identical advice for men and women.
Broader Implications
The researchers also tested how the models advised users on career choices, goal-setting, and even behavioural tips. Across the board, the LLMs responded differently based on the user’s gender, despite identical qualifications and prompts.
This is far from the first time AI has been caught reflecting and reinforcing systemic bias. The researchers behind the THWS study argue that technical fixes alone won’t solve the problem. What’s needed, they say, are clear ethical standards, independent review processes, and greater transparency in how these models are developed and deployed.
Source: thenextweb.com