[PDF][PDF] Adversarial Optimization for Dictionary Attacks on Speaker Verification.

M Marras, P Korus, ND Memon, G Fenu - Interspeech, 2019 - academia.edu
Interspeech, 2019academia.edu
In this paper, we assess vulnerability of speaker verification systems to dictionary attacks.
We seek master voices, ie, adversarial utterances optimized to match against a large
number of users by pure chance. First, we perform menagerie analysis to identify utterances
which intrinsically hold this property. Then, we propose an adversarial optimization
approach for generating master voices synthetically. Our experiments show that, even in the
most secure configuration, on average, a master voice can match approx. 20% of females …
Abstract
In this paper, we assess vulnerability of speaker verification systems to dictionary attacks. We seek master voices, ie, adversarial utterances optimized to match against a large number of users by pure chance. First, we perform menagerie analysis to identify utterances which intrinsically hold this property. Then, we propose an adversarial optimization approach for generating master voices synthetically. Our experiments show that, even in the most secure configuration, on average, a master voice can match approx. 20% of females and 10% of males without any knowledge about the population. We demonstrate that dictionary attacks should be considered as a feasible threat model for sensitive and high-stakes deployments of speaker verification.
academia.edu
以上显示的是最相近的搜索结果。 查看全部搜索结果