Article published In:
Vocal Interactivity in-and-between Humans, Animals and Robots
Edited by Mohamed Chetouani, Elodie F. Briefer, Angela Dassow, Ricard Marxer, Roger K. Moore, Nicolas Obin and Dan Stowell
[Interaction Studies 24:1] 2023
► pp. 168192
References
Ando, A., Masumura, R., Kamiyama, H., Kobashikawa, S., and Aono, Y.
(2017) Hierarchical LSTMs with Joint Learning for Estimating Customer Satisfaction from Contact Center Calls. In Proc. Interspeech 2017, pages 1716–1720. DOI logoGoogle Scholar
Ando, A., Masumura, R., Kamiyama, H., Kobashikawa, S., Aono, Y., and Toda, T.
(2020) Customer satisfaction estimation in contact center calls based on a hierarchical multi-task model. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 281:715–728. DOI logoGoogle Scholar
Auguste, J., Charlet, D., Damnati, G., Bechet, F., and Favre, B.
(2019) Can we predict self-reported customer satisfaction from interactions? In ICASSP 2019 – 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 7385–7389. DOI logoGoogle Scholar
Bockhorst, J., Yu, S., Polania, L., and Fung, G.
(2017) Predicting self-reported customer satisfaction of interactions with a corporate call center. In Altun, Y., Das, K., Mielikäinen, T., Malerba, D., Stefanowski, J., Read, J., Zitnik, M., Ceci, M., and Dzeroski, S., editors, Machine Learning and Knowledge Discovery in Databases, pages 179–190, Cham. Springer International Publishing. ISBN 978-3-319-71273-4. DOI logoGoogle Scholar
Chowdhury, S. A., Stepanov, E. A., and Riccardi, G.
(2016) Predicting user satisfaction from turn-taking in spoken conversations. In INTERSPEECH. DOI logoGoogle Scholar
Deschamps-Berger, T., Lamel, L., and Devillers, L.
(Sept. 2021) End-to-End Speech Emotion Recognition: Challenges of Real-Life Emergency Call Centers Data Recordings. In 2021 9th International Conference on Affective Computing and Intelligent Interaction (ACII), Nara, Japan. DOI logoGoogle Scholar
Erden, M. and Arslan, L. M.
(2011) Automatic detection of anger in human-human call center dialogs. In Proc. Interspeech 2011, pages 81–84. DOI logoGoogle Scholar
Eyben, F., Weninger, F., Gross, F., and Schuller, B.
(2013) Recent developments in opensmile, the munich open-source multimedia feature extractor. In Proceedings of the 21st ACM International Conference on Multimedia, MM ’131, page 835–838, New York, NY, USA. Association for Computing Machinery. ISBN 9781450324045. DOI logoGoogle Scholar
Galanis, D., Karabetsos, S., Koutsombogera, M., Papageorgiou, H., Esposito, A., and Riviello, M.-T.
(2013) Classification of emotional speech units in call centre interactions. In 2013 IEEE 4th International Conference on Cognitive Infocommunications (CogInfoCom), pages 403–406. DOI logoGoogle Scholar
Graves, A., Fernandez, S., Gomez, F., and Schmidhuber, J.
(2006) Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd International Conference on Machine Learning, ICML ’061, page 369–376, New York, NY, USA. Association for Computing Machinery. ISBN 1595933832. DOI logoGoogle Scholar
Kim, Y., Levy, J., and Liu, Y.
(2020) Speech sentiment and customer satisfaction estimation in socialbot conversations. In INTERSPEECH. DOI logoGoogle Scholar
Luque, J., Segura, C., Sánchez, A., Umbert, M., and Galindo, L. A.
(2017) The Role of Linguistic and Prosodic Cues on the Prediction of Self-Reported Satisfaction in Contact Centre Phone Calls. In Proc. Interspeech 2017, pages 2346–2350. DOI logoGoogle Scholar
Miao, Y., Gowayyed, M., and Metze, F.
(2015) Eesen: End-to-end speech recognition using deep rnn models and wfst-based decoding. In 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU), pages 167–174. DOI logoGoogle Scholar
Morrison, D., Wang, R., and De Silva, L. C.
(2007) Ensemble methods for spoken emotion recognition in call-centres. Speech Communication, 49(2):98–112. ISSN 0167-6393. DOI logoGoogle Scholar
Mower, E., Matarić, M. J., and Narayanan, S.
(2011) A framework for automatic human emotion classification using emotion profiles. IEEE Transactions on Audio, Speech, and Language Processing, 19(5):1057–1070. DOI logoGoogle Scholar
Petrushin, V. A.
(1999) Emotion in speech: Recognition and application to call centers.Google Scholar
Povey, D., Ghoshal, A., Boulianne, G., Burget, L., Glembek, O., Goel, N., Hannemann, M., Motlicek, P., Qian, Y., Schwarz, P., Silovsky, J., Stemmer, G., and Vesely, K.
The kaldi speech recognition toolkit. In IEEE 2011 Workshop on Automatic Speech Recognition and Understanding. IEEE Signal Processing Society, Dec. 2011. IEEE Catalog No.: CFP11SRW-USB.
Russell, J. A.
(1980) A circumplex model of affect. Journal of personality and social psychology, 39(6):1161. DOI logoGoogle Scholar
Schuller, B., Steidl, S., Batliner, A., Burkhardt, F., Devillers, L., Müller, C. A., and Narayanan, S. S.
(2010a) The interspeech 2010 paralinguistic challenge. In INTERSPEECH. DOI logoGoogle Scholar
Schuller, B., Weninger, F., Zhang, Y., Ringeval, F., Batliner, A., Steidl, S., Eyben, F., Marchi, E., Vinciarelli, A., Scherer, K., Chetouani, M., and Mortillaro, M.
(2019) Affective and behavioural computing: Lessons learnt from the first computational paralinguistics challenge. Computer Speech & Language, 531:156–180. ISSN 0885-2308. DOI logoGoogle Scholar
Schuller, B. W., Steidl, S., Batliner, A., Burkhardt, F., Devillers, L., Müller, C. A., and Narayanan, S. S.
(2010b) The INTERSPEECH 2010 paralinguistic challenge. In INTERSPEECH 2010, 11th Annual Conference of the International Speech Communication Association, Makuhari, Chiba, Japan, September 26–30, 2010, pages 2794–2797. DOI logoGoogle Scholar
Schuller, B. W., Steidl, S., Batliner, A., Vinciarelli, A., Scherer, K. R., Ringeval, F., Chetouani, M., Weninger, F., Eyben, F., Marchi, E., Mortillaro, M., Salamin, H., Polychroniou, A., Valente, F., and Kim, S.
(2013) The INTERSPEECH 2013 computational paralinguistics challenge: social signals, conflict, emotion, autism. In INTERSPEECH 2013, 14th Annual Conference of the International Speech Communication Association, Lyon, France, August 25–29, 2013, pages 148–152. DOI logoGoogle Scholar
Segura, C., Balcells, D., Umbert, M., Arias, J., and Luque, J.
(2016) Automatic speech feature learning for continuous prediction of customer satisfaction in contact center phone calls. In Abad, A., Ortega, A., Teixeira, A., Mateo, C. Garcia, Hinarejos, C. D. Martínez, Perdigão, F., Batista, F., and Mamede, N., editors, Advances in Speech and Language Technologies for Iberian Languages, pages 255–265, Cham. Springer International Publishing. ISBN 978-3-319-49169-1. DOI logoGoogle Scholar
Vaudable, C. and Devillers, L.
(2012) Negative emotions detection as an indicator of dialogs quality in call centers. In 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5109–5112. DOI logoGoogle Scholar
Viikki, O. and Laurila, K.
(1998) Cepstral domain segmental feature vector normalization for noise robust speech recognition. Speech Communication, 25(1):133–147. ISSN 0167-6393. DOI logoGoogle Scholar
Zweig, G., Siohan, O., Saon, G., Ramabhadran, B., Povey, D., Mangu, L., and Kingsbury, B.
(2006) Automated quality monitoring for call centers using speech and nlp technologies. In Proceedings of the 2006 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology: Companion Volume: Demonstrations, NAACL-Demonstrations ’061, page 292–295, USA. Association for Computational Linguistics. DOI logoGoogle Scholar