Noise Masking Attacks and Defenses for Pretrained Speech Models
ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2024)
摘要
Speech models are often trained on sensitive data in order to improve model
performance, leading to potential privacy leakage. Our work considers noise
masking attacks, introduced by Amid et al. 2022, which attack automatic speech
recognition (ASR) models by requesting a transcript of an utterance which is
partially replaced with noise. They show that when a record has been seen at
training time, the model will transcribe the noisy record with its memorized
sensitive transcript. In our work, we extend these attacks beyond ASR models,
to attack pretrained speech encoders. Our method fine-tunes the encoder to
produce an ASR model, and then performs noise masking on this model, which we
find recovers private information from the pretraining data, despite the model
never having seen transcripts at pretraining time! We show how to improve the
precision of these attacks and investigate a number of countermeasures to our
attacks.
更多查看译文
关键词
noise masking,privacy,speech pretraining,deduplication,sanitization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要