Asynchronous Voice Anonymization by Learning From Speaker-Adversarial Speech

Rui Wang, Liping Chen, Kong Aik Lee, Zhen Hua Ling

Research output: Journal article publicationJournal articleAcademic researchpeer-review

Abstract

This letter focuses on asynchronous voice anonymization, wherein machine-discernible speaker attributes in a speech utterance are obscured while human perception is preserved. We propose to transfer the voice-protection capability of speaker-adversarial speech to speaker embedding, thereby facilitating the modification of speaker embedding extracted from original speech to generate anonymized speech. Experiments conducted on the LibriSpeech dataset demonstrated that compared to the speaker-adversarial utterances, the generated anonymized speech demonstrates improved transferability and voice-protection capability. Furthermore, the proposed method enhances the human perception preservation capability of anonymized speech within the generative asynchronous voice anonymization framework.

Original languageEnglish
Pages (from-to)1905-1909
Number of pages5
JournalIEEE Signal Processing Letters
Volume32
DOIs
Publication statusPublished - Apr 2025

Keywords

  • Asynchronous voice anonymization
  • speaker embedding
  • speaker-adversarial perturbation

ASJC Scopus subject areas

  • Signal Processing
  • Electrical and Electronic Engineering
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Asynchronous Voice Anonymization by Learning From Speaker-Adversarial Speech'. Together they form a unique fingerprint.

Cite this