Skip to main content

Summer Deadline: Sunday, March 29 @ 11:59pm PT. Click to apply.

Back to Research
Accepted to Building Trust in LLMs @ ICLR 2025

MALIBU Benchmark: Multi-Agent LLM Implicit Bias Uncovered

Imran Mirza, Cole Huang, Ishwara Vasista, Rohan Patil

Abstract

We introduce MALIBU (Multi-Agent LLM Implicit Bias Uncovered), a benchmark for evaluating implicit biases in multi-agent LLM systems. MALIBU systematically probes how biases emerge, amplify, and propagate when multiple LLM agents interact in collaborative decision-making scenarios. Our framework reveals that multi-agent configurations can amplify individual model biases by 15-40% compared to single-agent baselines.

Citation

Imran Mirza, Cole Huang, Ishwara Vasista, Rohan Patil. "MALIBU Benchmark: Multi-Agent LLM Implicit Bias Uncovered". Accepted to Building Trust in LLMs @ ICLR 2025.

Details

Conference
Accepted to Building Trust in LLMs @ ICLR 2025
Authors
4 authors

Publish Your Research

Join Algoverse and work with world-class mentors to publish at top AI conferences.

Start Your Application