KIT | KIT-Bibliothek | Impressum | Datenschutz

Evaluation of Crowdsourced User Input Data for Spoken Dialog Systems

Schmidt, Maria; Müller, Markus; Wagner, Martin; Stüker, Sebastian; Waibel, Alex; Hofmann, Hansjörg; Werner, Steffen

Abstract:

Using the Internet for the collection of data is quite common these days. This process is called crowdsourcing and enables the collection of large amounts of data at reasonable costs. While being an inexpensive method, this data typically
is of lower quality. Filtering data sets is therefore required. The occurring errors can be classified into different groups. There are technical issues and human errors. For speech recording, technical issues could be a noisy background. Human errors arise when the task is misunderstood. We employ several techniques for recognizing errors and eliminating faulty data sets in user input data for a Spoken Dialog System (SDS). Furthermore, we compare three different kinds of questionnaires (QNRs) for a given set of seven tasks. We analyze the characteristics of the resulting data sets and give a recommendation which type of QNR might be the most suitable one for a given purpose.


Verlagsausgabe §
DOI: 10.5445/IR/1000051099
Veröffentlicht am 10.06.2025
Originalveröffentlichung
DOI: 10.18653/v1/W15-4657
Cover der Publikation
Zugehörige Institution(en) am KIT Institut für Anthropomatik und Robotik (IAR)
Publikationstyp Proceedingsbeitrag
Publikationsjahr 2015
Sprache Englisch
Identifikator ISBN: 978-1-941643-75-4
KITopen-ID: 1000051099
Erschienen in Proceedings of the 16th Annual SIGdial Meeting on Discourse and Dialogue (SIGDIAL 2015), September 2-4 2015, Prague, Czech Republic
Veranstaltung 16th Annual SIGdial Meeting on Discourse and Dialogue (2015), Prag, Tschechien, 02.09.2015 – 04.09.2015
Verlag Association for Computational Linguistics (ACL)
Seiten 427-431
Externe Relationen Siehe auch
KIT – Die Universität in der Helmholtz-Gemeinschaft
KITopen Landing Page