Personalised models for speech detection from body movements using transductive parameter transfer

Research output: Contribution to journalArticleScientificpeer-review

8 Citations (Scopus)
22 Downloads (Pure)

Abstract

We investigate the task of detecting speakers in crowded environments using a single body worn triaxial accelerometer. Detection of such behaviour is very challenging to model as people’s body movements during speech vary greatly. Similar to previous studies, by assuming that body movements are indicative of speech, we show experimentally, on a real-world dataset of 3 h including 18 people, that transductive parameter transfer learning (Zen et al. in Proceedings of the 16th international conference on multimodal interaction. ACM, 2014) can better model individual differences in speaking behaviour, significantly improving on the state-of-the-art performance. We also discuss the challenges introduced by the in-the-wild nature of our dataset and experimentally show how they affect detection performance. We strengthen the need for an adaptive approach by comparing the speech detection problem to a more traditional activity (i.e. walking). We provide an analysis of the transfer by considering different source sets which provides a deeper investigation of the nature of both speech and body movements, in the context of transfer learning.

Original languageEnglish
Pages (from-to)723-737
Number of pages15
JournalPersonal and Ubiquitous Computing
Volume21
Issue number4
DOIs
Publication statusPublished - 1 Aug 2017

Keywords

  • Human behaviour
  • Social actions
  • Social signal processing
  • Transfer learning
  • Wearable sensors

Fingerprint Dive into the research topics of 'Personalised models for speech detection from body movements using transductive parameter transfer'. Together they form a unique fingerprint.

  • Cite this