Multi Speaker Multi Accent
Multi-speaker, multi-accent speech synthesis aims to create realistic synthetic speech encompassing diverse speaker identities and accents, addressing the current underrepresentation of certain populations in existing systems. Research focuses on developing models that disentangle speaker identity from accent characteristics, often employing multi-scale modeling strategies and adversarial training to achieve fine-grained control over both aspects. This work is significant for improving the inclusivity and accessibility of speech technology, with applications ranging from personalized education tools to more representative virtual assistants and automated content generation.
Papers
June 17, 2024
June 16, 2024
March 14, 2023