Evolving learning for analysing mood-related infant vocalisation

Zixing Zhang, Jing Han, Kun Qian, Björn Schuller

Research output: Contribution to journalConference articlepeer-review

5 Citations (Scopus)

Abstract

Infant vocalisation analysis plays an important role in the study of the development of pre-speech capability of infants, while machine-based approaches nowadays emerge with an aim to advance such an analysis. However, conventional machine learning techniques require heavy feature-engineering and refined architecture designing. In this paper, we present an evolving learning framework to automate the design of neural network structures for infant vocalisation analysis. In contrast to manually searching by trial and error, we aim to automate the search process in a given space with less interference. This framework consists of a controller and its child networks, where the child networks are built according to the controller's estimation. When applying the framework to the Interspeech 2018 Computational Paralinguistics (ComParE) Crying Sub-challenge, we discover several deep recurrent neural network structures, which are able to deliver competitive results to the best ComParE baseline method.

Original languageEnglish
Pages (from-to)142-146
Number of pages5
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume2018-September
DOIs
Publication statusPublished - 2018
Externally publishedYes
Event19th Annual Conference of the International Speech Communication, INTERSPEECH 2018 - Hyderabad, India
Duration: 2 Sept 20186 Sept 2018

Keywords

  • Evolving learning
  • Infant vocalisation
  • Neural network architecture
  • Speech/voice analysis

Fingerprint

Dive into the research topics of 'Evolving learning for analysing mood-related infant vocalisation'. Together they form a unique fingerprint.

Cite this