Speech, Music and Hearing (TMH)

Research at the Division of Speech, Music and Hearing (TMH) is truly multi-disciplinary including linguistics, phonetics, auditory perception, vision and experimental psychology. Rooted in an engineering modelling approach, our research forms a solid base for developing multimodal human-computer interaction systems in which speech, music, sound and gestures combine to create human-like communication.

Research Area

Latest Publications

[1]
Kamelabad, A. M., Inoue, E., Skantze, G. (2025). Comparing Monolingual and Bilingual Social Robots as Conversational Practice Companions in Language Learning. In Proceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction. (pp. 829-838).
[2]
Gonzalez Oliveras, P., Engwall, O. & Wilde, A. (2025). Social Educational Robotics and Learning Analytics : A Scoping Review of an Emerging Field. International Journal of Social Robotics.
[3]
Cai, H. & Ternström, S. (2025). A WaveNet-based model for predicting the electroglottographic signal from the acoustic voice signal. Journal of the Acoustical Society of America, 157(4), 3033-3044.
[4]
Marcinek, L., Beskow, J., Gustafsson, J. (2025). A Dual-Control Dialogue Framework for Human-Robot Interaction Data Collection : Integrating Human Emotional and Contextual Awareness with Conversational AI. In Social Robotics - 16th International Conference, ICSR + AI 2024, Proceedings. (pp. 290-297). Springer Nature.
[5]
Mishra, C., Skantze, G., Hagoort, P., Verdonschot, R. (2025). Perception of Emotions in Human and Robot Faces : Is the Eye Region Enough?. In Social Robotics - 16th International Conference, ICSR + AI 2024, Proceedings. (pp. 290-303). Springer Nature.
Full list in the KTH publications portal

News