Special session at INTERSPEECH 2020, Shanghai, China
Deep learning based human language technology (HLT), such as automatic speech recognition, intent and slot recognition, or dialog management, has become the mainstream of research in recent years and significantly outperforms conventional methods. The technology also has widespread applications in the industry. Several most famous examples include Siri, Alexa, Google Assistant, and Cortana. However, deep learning models are notorious for being data and computation hungry. These downsides limit the application of such models from deployment to different languages, domains, or styles, since collecting in-genre data and model training from scratch are costly.
Meta learning, or Learning to Learn, is one way to mitigate the above problems. Meta learning learns better learning algorithms, including better parameter initialization, optimization strategy, network architecture, distance metrics, etc., from multiple learning tasks. Meta learning has been showed the potential to allow faster fine-tuning, converge to better performance than model pretraining, and even achieve few-shot learning in several areas, including computer vision and translation.
The goal of this special session is to bring together researchers and practitioners working on meta learning in different HLT fields to discuss the state-of-the-art and new approaches, and to share their innovation, insights, and challenges, and to shed light on future research directions. We will explore how to improve learning efficiency in data usage and in computation with meta learning for HLT tasks. We also aim to align academic efforts with industrial challenges, to bridge the gap between research and real-world product deployment.
The special session of Meta Learning for Human Language Technology invites papers of a theoretical and experimental nature on human language technology tasks with meta learning methodologies and their applications. The special session is part of the main INTERSPEECH conference in Shanghai, China. Relevant meta learning topics include (but are not limited to):
Human language technology topics include (but are not limited to):
This special session is part of the main INTERSPEECH conference. Thus it utilizes the same submission portal, and follows the same submission policy, paper format, and review process. More information can be found:
Following the same policy as INTERSPEECH main conference, double-submissions is not allowed if the entire works have been published in other peer-reviewed conferences or transaction. However, we invite authors to submit their work to multiple sessions in addition to this session in the submission portal. Conference and session committee will determine session assignment after acceptance.
If you have more questions about submission, please feel free to contact us via is.2020.meta.learning@gmail.com .
Meta learning is one of the fastest growing research areas in the deep learning scope. However there is no standard definition for meta learning. Usually the main goal is to design models that can learn new tasks rapidly with few in domain training examples, by having models to pre-learn from many, relevant or not, training tasks in a way that the models are easy to be generalized to new tasks. For better understanding the scope of meta learning, we provide several online courses and papers describing the works falling into the area. These works are just for showcasing, and we definitely encourage people with research not covered here but sharing the same goal mentioned above to submit.
Learning to initialize | Learning to compare | Other | |
Speech Recognition |
[Hsu, et al., ICASSP’20] [Klejch, et al., ASRU’19] |
[Klejch , et al., INTERSPEECH’18] (Learning to optimize) [Baruwa , et al., IJSER’19] (Network architecture search) |
|
Voice Cloning |
[Chen, et al., ICLR’19] (Learning the learning algorithm) [Serrà, et al., NeurIPS’19] (Learning the learning algorithm) |
||
Speaker Recognition | [Anand, et al., arXiv’19] | ||
Keyword Spotting | [Chen, et al., arXiv’18] | [Mazzawi, et al., INTERSPEECH’19] (Network architecture search) |
|
Sound Event Detection |
[Shimada, et al., arXiv’19] [Chou, et al., ICASSP’19] [Zhang, et al., INTERSPEECH’19] |
||
Machine Translation |
[Gu, et al., EMNLP’18] [Indurthi, et al., arXiv’19] |
||
Dialogue |
[Qian, et al., ACL’19] [Madotto, et al., ACL’19] [Mi, et al., IJCAI’19] [Song, et al., arXiv’19] |
[Chien, et al., INTERSPEECH’19] (Learning to optimize) |
|
Relation Classification |
[Obamuyide, et al., ACL’19] [Bose, et al., arXiv’19] [Lv, et al., EMNLP’19] [Wang, et al., EMNLP’19] |
[Ye, et al., ACL’19] [Chen, et al., EMNLP’19] [Xiong, et al., EMNLP’18] [Gao, et al., AAAI’19] |
|
Word Embedding | [Hu, et al., ACL’19] | [Sun, et al., EMNLP’18] | |
More NLP Applications |
[Guo, et al., ACL’19] [Wu, et al., AAAI’20] [Zhao, EMNLP’19] [Bansal, et al., arXiv’19] [Dou, et al., EMNLP’19] [Huang, et al., NAACL’18] |
[Sun, et al., EMNLP’19] [Geng, et al., EMNLP’19] [Yu, et al., ACL’18] [Tan, et al., EMNLP’19] |
[Wu, et al., EMNLP’19] (Learning the learning algorithm) |
Multi-model | [Eloff, et al., ICASSP’19] | [Surís, et al., arXiv’19] (Learning the learning algorithm) |
TBD
TBD