Personality testing reveals behavioral bias in LLMs

Most major large language models (LLMs) can quickly tell when they are being given a personality test and will tweak their responses to provide more socially desirable results-;a finding with implications for any study using LLMs as a stand-in for humans. Aadesh Salecha and colleagues gave LLMs from OpenAI, Anthropic, Google, and Meta the classic Big 5 personality test, which is a survey that measures Extraversion, Openness to Experience, Conscientiousness, Agreeableness, and Neuroticism.

Researchers have given the Big 5 test to LLMs, but have not typically considered that the models, like humans, may tend to skew their responses to seem likable, which is known as a "social desirability bias." Typically, people prefer people who have low neuroticism scores and high scores on the other four traits, such as extraversion. The authors varied the number of questions given to models.

When only asked a small number of questions, LLMs did not change their responses as much as when the authors asked five or more questions, which allowed models to conclude that their personality was being measured. For GPT-4, scores for positively perceived traits increased by more than 1 standard deviation, and for neuroticism scores reduced by a similar amount, as the authors increased the number of questions or told the models that their personality was being measured.

This is a large effect, the equivalent of speaking to an average human who suddenly pretends to have a personality that's more desirable than 85% of the population. The authors think this effect is likely the result of the final LLM training step, which involves humans choosing their preferred response from LLMs. According to the authors, LLMs "catch on" to which personalities are socially desirable at a deep level, which allows LLMs to emulate those personalities when asked.

Source:
Journal reference:

Salecha, A., et al. (2024) Large language models display human-like social desirability biases in Big Five personality surveys. PNAS Nexusdoi.org/10.1093/pnasnexus/pgae533.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
From potential to practice: A blueprint for responsible AI in healthcare