Abstract

The present subject matter relates to a systematic study of emergent bias and fairness dynamics in multi-agent LLMdecision making. In particular, the present disclosure compares single-agent performance with memory-based and collective refinement multi-agent discussion paradigms across a suite of group fairness metrics on tabular datasets with sensitive attributes. Across settings, fairness changes from multi-agent collaboration relative to single-agent baselines are typically concentrated around small negative values or near zero, indicating that bias is most often only marginally reduced or remains unchanged, while a thin but long positive tail shows rare cases where multi-agent interaction can substantially worsen bias. These findings motivate systematic investigation into the conditions under which multi-agent collaboration mechanisms mitigate or amplify bias.

Publication Date

2026-01-05

Creative Commons License

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.

Share

COinS