Fair Voice Biometrics: Impact of Demographic Imbalance on Group Fairness in Speaker Recognition
Speaker recognition systems are playing a key role in modern online applications. Though the susceptibility of these systems to discrimination according to group fairness metrics has been recently studied, their assessment has been mainly focused on the difference in equal error rate across groups, not accounting for other fairness criteria important in anti-discrimination policies, defined for demographic groups characterized by sensitive attributes. In this paper, we therefore study how existing group fairness metrics relate with the balancing settings of the training data set in speaker recognition. We conduct this analysis by operationalizing several definitions of fairness and monitoring them under varied data balancing settings. Experiments performed on three deep neural architectures, evaluated on a data set including gender/age-based groups, show that balancing group representation positively impacts on fairness and that the friction across security, usability, and fairness depends on the fairness metric and the recognition threshold.
WOS:000841879501202
2021-01-01
Baixas
Interspeech
1892
1896
REVIEWED
Event name | Event place | Event date |
Brno, CZECH REPUBLIC | Aug 30-Sep 03, 2021 | |