The School of EECS is hosting the following HDR Progress Review 1 Confirmation Seminar:

Leveraging Personas to Study and Control Ideological Biases in Large Language Models

Speaker: Pietro Bernardelle
Host/Chair: Dr. Rocky Chen

Abstract:

Large language models (LLMs) are increasingly embedded in sensitive domains such as information access, content moderation, and computational social science, where fairness and neutrality are critical. However, these systems consistently exhibit political and ideological biases that risk undermining trust, fairness, and resilience against manipulation. Existing research largely treats such biases as static properties of model weights, yet recent evidence suggests that prompting LLMs with personas exposes a dynamic, controllable layer of ideological expression. This dissertation investigates the mechanisms and implications of persona-driven bias along three dimensions: (i) how personas influence the expression of political and cultural biases; (ii) what strategies can mitigate or guide persona-driven ideological shifts; and (iii) whether such dynamics generalise across broader ideological domains. Our initial findings show that personas hold the potential to substantially alter ideological positioning, with larger models demonstrating broader dispersion and stronger responsiveness to explicit cues. Moreover, thematic persona attributes (e.g., business, history, politics) consistently trigger predictable ideological shifts, revealing latent pathways of bias expression. Building on these insights, forthcoming work will explore bias mitigation strategies—such as direct preference optimization (DPO)—and tests their effectiveness in reducing political and cultural distortions while preserving task performance. By uncovering both the risks and opportunities of persona-driven ideological malleability, this research contributes to the development of more trustworthy, fair, and context-aware generative AI systems.

Bio:

Pietro Bernardelle is a first year Ph.D. student at The University of Queensland, Australia. His research focuses on the intersection of large language models and social science, with a particular interest in studying and mitigating ideological biases in generative AI. He is advised by Prof. Gianluca Demartini, Dr. Joel Mackenzie, and Dr. Kevin Roitero. He received a Master’s degree in Computer Science and Engineering from the Polytechnic University of Milan, and a Master of Engineering Science in Software Engineering from The University of Queensland.

About Data Science Seminar

This seminar series is hosted by EECS Data Science.

Venue

Room: 78 - 632 (MM Lab)