Abstract
Oral cancer speech is a disease which impacts more than half a million people worldwide every year. Analysis of oral cancer speech has so far focused on read speech. In this paper, we 1) present and 2) analyse a three-hour long spontaneous oral cancer speech dataset collected from YouTube. 3) We set baselines for an oral cancer speech detection task on this dataset. The analysis of these explainable machine learning baselines shows that sibilants and stop consonants are the most important indicators for spontaneous oral cancer speech detection.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of Interspeech 2020 |
| Publisher | ISCA |
| Pages | 4826 - 4830 |
| Number of pages | 5 |
| DOIs | |
| Publication status | Published - 2020 |
| Event | INTERSPEECH 2020 - Shanghai, Shanghai, China Duration: 25 Oct 2020 → 29 Oct 2020 |
Publication series
| Name | Interspeech 2020 |
|---|---|
| Publisher | ISCA |
| ISSN (Print) | 1990-9772 |
Conference
| Conference | INTERSPEECH 2020 |
|---|---|
| Country/Territory | China |
| City | Shanghai |
| Period | 25/10/20 → 29/10/20 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Keywords
- Corpus
- Explainable AI
- Oral cancer speech
- Pathological speech
Fingerprint
Dive into the research topics of 'Detecting and analysing spontaneous oral cancer speech in the wild'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver