Your fairness may vary: Pretrained language model fairness in toxic text classification

Ioana Baldini, Dennis Wei, Karthikeyan Natesan Ramamurthy, Moninder Singh, Mikhail Yurochkin. Your fairness may vary: Pretrained language model fairness in toxic text classification. In Smaranda Muresan, Preslav Nakov, Aline Villavicencio, editors, Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, May 22-27, 2022. pages 2245-2262, Association for Computational Linguistics, 2022. [doi]

Authors

Ioana Baldini

This author has not been identified. Look up 'Ioana Baldini' in Google

Dennis Wei

This author has not been identified. Look up 'Dennis Wei' in Google

Karthikeyan Natesan Ramamurthy

This author has not been identified. Look up 'Karthikeyan Natesan Ramamurthy' in Google

Moninder Singh

This author has not been identified. Look up 'Moninder Singh' in Google

Mikhail Yurochkin

This author has not been identified. Look up 'Mikhail Yurochkin' in Google