Abstract
Both human listeners and machines need to adapt their sound categories whenever a new speaker is encountered. This perceptual learning is driven by lexical information. The aim of this paper is two-fold: investigate whether a deep neural network-based (DNN) ASR system can adapt to only a few examples of ambiguous speech as humans have been found to do; investigate a DNN’s ability to serve as a model of human perceptual learning. Crucially, we do so by looking at intermediate levels of phoneme category adaptation rather than at the output level. We visualize the activations in the hidden layers of the DNN during perceptual learning. The results show that, similar to humans, DNN systems learn speaker-adapted phone category boundaries from a few labeled examples. The DNN adapts its category boundaries not only by adapting the weights of the output layer, but also by adapting the implicit feature maps computed by the hidden layers, suggesting the possibility that human perceptual learning might involve a similar nonlinear distortion of a perceptual space that is intermediate between the acoustic input and the phonological categories. Comparisons between DNNs and humans can thus provide valuable insights into the way humans process speech and improve ASR technology.
Original language | English |
---|---|
Title of host publication | Proceedings of Interspeech 2018 |
Editors | B. Yegnanarayana |
Place of Publication | India |
Publisher | International Speech Communication Association |
Pages | 1482-1486 |
Number of pages | 5 |
DOIs | |
Publication status | Published - 3 Sept 2018 |
Event | Interspeech 2018 - Hyderabad, India Duration: 2 Sept 2018 → 6 Sept 2018 |
Conference
Conference | Interspeech 2018 |
---|---|
Country/Territory | India |
City | Hyderabad |
Period | 2/09/18 → 6/09/18 |
Keywords
- phoneme category adaptation
- human perceptual learning
- deep neural networks
- Visualisation