Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

Mitigating Large Language Model Bias: Automated Dataset Augmentation and Prejudice Quantification

*,‡ and ORCID logo
These authors have contributed equally to this work.
Version 1 : Received: 5 May 2024 / Approved: 6 May 2024 / Online: 6 May 2024 (07:24:03 CEST)

A peer-reviewed article of this Preprint also exists.

Mondal, D.; Lipizzi, C. Mitigating Large Language Model Bias: Automated Dataset Augmentation and Prejudice Quantification. Computers 2024, 13, 141. Mondal, D.; Lipizzi, C. Mitigating Large Language Model Bias: Automated Dataset Augmentation and Prejudice Quantification. Computers 2024, 13, 141.

Abstract

Despite the growing capabilities of large language models, concerns exist about the biases they develop. In this paper, we propose a novel, automated mechanism for debiasing through specified dataset augmentation in the lens of bias producers that can be useful in a variety of industries, especially ones that are “restricted” and have limited data. We consider that bias can occur due to intrinsic model architecture and dataset quality. The two aspects are evaluated using two different metrics we created. We show that our dataset augmentation algorithm reduces bias as measured by our metrics.

Keywords

natural language processing; large language models; dataset augmentation; computational social science

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.