Quantifying Source Speaker Leakage in One-to-One Voice Conversion
By: Scott Wellington, Xuechen Liu, Junichi Yamagishi
Potential Business Impact:
Identifies who is speaking in fake voices.
Using a multi-accented corpus of parallel utterances for use with commercial speech devices, we present a case study to show that it is possible to quantify a degree of confidence about a source speaker's identity in the case of one-to-one voice conversion. Following voice conversion using a HiFi-GAN vocoder, we compare information leakage for a range speaker characteristics; assuming a "worst-case" white-box scenario, we quantify our confidence to perform inference and narrow the pool of likely source speakers, reinforcing the regulatory obligation and moral duty that providers of synthetic voices have to ensure the privacy of their speakers' data.
Similar Papers
Evaluating Identity Leakage in Speaker De-Identification Systems
Sound
Makes voices sound like someone else.
Source Verification for Speech Deepfakes
Sound
Finds who made fake voices.
Speaker Anonymisation for Speech-based Suicide Risk Detection
Audio and Speech Processing
Protects voices while finding people at risk.