IITP
XVoice : Multi-Modal Voice Meta learning
2022-2026
This project aims to create meta-learning algorithms for the composition of new human voices. To accurately synthesize new voices, we might use multi-modal information such as bio-streamlines from sensors, visual signals and audio data. In this project, our team plays a crucial role in developing meta-learning algorithms for low-task regimes. So far, meta-learning frameworks that rapidly adapt a novel tasks to guarantee generalization performance have assumed a large number of training tasks. Recent studies, however, reveal that meta-learning algorithms also impairs their generalization when the limited number of tasks is given. Consequently, we develop a new regularization techniques to preserve generalization performance of meta-learning algorithms.