ZET-Speech: Zero-shot adaptive Emotion-controllable Text-to-Speech Synthesis with Diffusion and Style-based Models

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 33
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorKang, Minkiko
dc.contributor.authorHan, Wooseokko
dc.contributor.authorHwang, Sung Juko
dc.contributor.authorYang, Eunhoko
dc.date.accessioned2023-12-12T07:00:53Z-
dc.date.available2023-12-12T07:00:53Z-
dc.date.created2023-12-08-
dc.date.issued2023-08-22-
dc.identifier.citation24th International Speech Communication Association, Interspeech 2023, pp.4339 - 4343-
dc.identifier.urihttp://hdl.handle.net/10203/316285-
dc.description.abstractEmotional Text-To-Speech (TTS) is an important task in the development of systems (e.g., human-like dialogue agents) that require natural and emotional speech. Existing approaches, however, only aim to produce emotional TTS for seen speakers during training, without consideration of the generalization to unseen speakers. In this paper, we propose ZET-Speech, a zero-shot adaptive emotion-controllable TTS model that allows users to synthesize any speaker's emotional speech using only a short, neutral speech segment and the target emotion label. Specifically, to enable a zero-shot adaptive TTS model to synthesize emotional speech, we propose domain adversarial learning and guidance methods on the diffusion model. Experimental results demonstrate that ZET-Speech successfully synthesizes natural and emotional speech with the desired emotion for both seen and unseen speakers. Samples are at https://ZET-Speech.github.io/ZET-Speech-Demo/.-
dc.languageEnglish-
dc.publisherInternational Speech Communication Association-
dc.titleZET-Speech: Zero-shot adaptive Emotion-controllable Text-to-Speech Synthesis with Diffusion and Style-based Models-
dc.typeConference-
dc.identifier.scopusid2-s2.0-85171580094-
dc.type.rimsCONF-
dc.citation.beginningpage4339-
dc.citation.endingpage4343-
dc.citation.publicationname24th International Speech Communication Association, Interspeech 2023-
dc.identifier.conferencecountryIE-
dc.identifier.conferencelocationDublin-
dc.identifier.doi10.21437/Interspeech.2023-754-
dc.contributor.localauthorHwang, Sung Ju-
dc.contributor.localauthorYang, Eunho-
dc.contributor.nonIdAuthorKang, Minki-
dc.contributor.nonIdAuthorHan, Wooseok-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0