Skip to content

MartaMarchiori/Exposing-Racial-Dialect-Bias

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

Exposing Racial Dialect Bias in Abusive Language Detection: Can Explainability Play a Role?

Biases can arise and be introduced during each phase of a supervised learning pipeline, eventually leading to harm. Within the task of automatic abusive language detection, this matter becomes particularly severe since unintended bias towards sensitive topics such as gender, sexual orientation, or ethnicity can harm underrepresented groups. The role of the datasets used to train these models is crucial to address these challenges. In this contribution, we investigate whether explainability methods can expose racial dialect bias attested within a popular dataset for abusive language detection. Through preliminary experiments, we found that pure explainability techniques cannot effectively uncover biases within the dataset under analysis: the rooted stereotypes are often more implicit and complex to retrieve.

Marta Marchiori Manerba and Virginia Morini. "Exposing Racial Dialect Bias in Abusive Language Detection: Can Explainability Play a Role?". XKDD, 2022.

Bibtex for citations:

TBD

About

Exposing Racial Dialect Bias in Abusive Language Detection: Can Explainability Play a Role?

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published