skip to content

The Psychometrics Centre

Cambridge Judge Business School
 
Subscribe to Psychometrics Seminar Series feed
***DON'T FORGET TO TURN ON THE AUTOMATIC REMINDERS ("Send you e-mail reminders" option above)*** Seminars provide the great opportunity for the researchers at every stage of their career to learn about modern psychometrical techniques and observe their application in the examples drawn from the recent studies. The talks are open to the general public.
Updated: 1 hour 54 min ago

Tue 07 Oct 12:00: Feedback Forensics: Measuring AI Personality By Comparing Observed Behaviour If you are interested in attending the talk online, please email the organiser and ask for a Teams invite.

Thu, 18/09/2025 - 15:37
Feedback Forensics: Measuring AI Personality By Comparing Observed Behaviour

Many personality tests ask participants hypothetical questions predicting their own behaviours. Yet, as with humans, self-predicted AI behaviour does not always match observed behaviour. In this talk, I will introduce Feedback Forensics: a toolkit to measure AI traits related to personality directly based on observed behaviour data. Comparing model behaviours to the same input relative to each other, our toolkit can measure a diverse set of traits related to the underlying personality, manner, and style of AI responses. I will share results describing traits exhibited by popular AI models as well as detecting the traits encouraged by human feedback. The talk will feature a live demo of our personality visualisation tool and attendees are invited to follow along via our online platform https://feedbackforensics.com/ (laptops are encouraged).

Bio: Arduin is currently a PhD student in the Department of Computer Science in Cambridge working on AI model evaluation. His work focuses on understanding what desirable and undesirable model behaviours are reinforced by human and AI feedback. Prior to joining his current PhD programme, Arduin completed an MPhil in Machine Learning and Machine Intelligence in Cambridge’s Engineering Department. Recently, Arduin also worked on model evaluation within Apple’s Foundation Models team as an intern.

If you are interested in attending the talk online, please email the organiser and ask for a Teams invite.

Add to your calendar or Include in your list

Wed 17 Sep 12:00: LLM Social Simulation is a Promising Research Method The talk is available online. Please get in touch to ask for a Teams invite.

Tue, 16/09/2025 - 21:16
LLM Social Simulation is a Promising Research Method

The emergence of large language models as in-silico subjects for social science poses a central question: can they genuinely simulate diverse human behavior, or do they merely produce plausible, homogenized artifacts? This talk demonstrates that LLMs are powerful but imperfect simulators by presenting three core contributions. First, we establish the “Persona Effect,” showing that persona-prompting a 70B model captures 81% of explainable variance in subjective tasks, creating a strong baseline for individual-level simulation. Second, to address data scarcity, we introduce iNews, a large-scale dataset of personalized affective responses to news, enriched with persona information. Finally, we introduce SimBench, the first large-scale benchmark for group-level simulation, which reveals the strengths and critical weaknesses of current models. I conclude by arguing for the specialized datasets and training required to advance the frontier of high-fidelity human simulation.

The talk is available online. Please get in touch to ask for a Teams invite.

Add to your calendar or Include in your list

Wed 17 Sep 12:00: LLM Social Simulation is a Promising Research Method The talk is available online. Please get in touch to ask for a Teams invite.

Thu, 11/09/2025 - 10:44
LLM Social Simulation is a Promising Research Method

The emergence of large language models as in-silico subjects for social science poses a central question: can they genuinely simulate diverse human behavior, or do they merely produce plausible, homogenized artifacts? This talk demonstrates that LLMs are powerful but imperfect simulators by presenting three core contributions. First, we establish the “Persona Effect,” showing that persona-prompting a 70B model captures 81% of explainable variance in subjective tasks, creating a strong baseline for individual-level simulation. Second, to address data scarcity, we introduce iNews, a large-scale dataset of personalized affective responses to news, enriched with persona information. Finally, we introduce SimBench, the first large-scale benchmark for group-level simulation, which reveals the strengths and critical weaknesses of current models. I conclude by arguing for the specialized datasets and training required to advance the frontier of high-fidelity human simulation.

The talk is available online. Please get in touch to ask for a Teams invite.

Add to your calendar or Include in your list

Upcoming events