Enhancing explainability with multimodal context representations for smarter robots
Viswanath A, Lokesh V, Buschmeier H (2025)
In: Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25.
Konferenzbeitrag
| Veröffentlicht | Englisch
Download

Autor*in
Einrichtung
Abstract / Bemerkung
Artificial Intelligence (AI) has significantly advanced in recent years, driving innovation across various fields, especially in robotics. Even though robots can perform complex tasks with increasing autonomy, challenges remain in ensuring explainability and user-centered design for effective interaction. A key issue in Human-Robot Interaction (HRI) is enabling robots to effectively perceive and reason over multimodal inputs, such as audio and vision, to foster trust and seamless collaboration. In this paper, we propose a generalized and explainable multimodal framework for context representation, designed to improve the fusion of speech and vision modalities. We introduce a use case on assessing ‘Relevance’ between verbal utterances from the user and visual scene perception of the robot. We present our methodology with a _Multimodal Joint Representation_ module and a _Temporal Alignment_ module, which can allow robots to evaluate relevance by temporally aligning multimodal inputs. Finally, we discuss how the proposed framework for context representation can help with various aspects of explainability in HRI.
Stichworte
dililab
Erscheinungsjahr
2025
Titel des Konferenzbandes
Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25
Urheberrecht / Lizenzen
Konferenz
3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25
Konferenzort
Melbourne, Australia
Konferenzdatum
2025-03-03 – 2025-03-03
Page URI
https://pub.uni-bielefeld.de/record/3001083
Zitieren
Viswanath A, Lokesh V, Buschmeier H. Enhancing explainability with multimodal context representations for smarter robots. In: Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25. 2025.
Viswanath, A., Lokesh, V., & Buschmeier, H. (2025). Enhancing explainability with multimodal context representations for smarter robots. Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25. https://doi.org/10.5281/zenodo.14930029
Viswanath, Anargh, Lokesh, Veeramacheneni, and Buschmeier, Hendrik. 2025. “Enhancing explainability with multimodal context representations for smarter robots”. In Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25.
Viswanath, A., Lokesh, V., and Buschmeier, H. (2025). “Enhancing explainability with multimodal context representations for smarter robots” in Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25.
Viswanath, A., Lokesh, V., & Buschmeier, H., 2025. Enhancing explainability with multimodal context representations for smarter robots. In Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25.
A. Viswanath, V. Lokesh, and H. Buschmeier, “Enhancing explainability with multimodal context representations for smarter robots”, Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25, 2025.
Viswanath, A., Lokesh, V., Buschmeier, H.: Enhancing explainability with multimodal context representations for smarter robots. Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25. (2025).
Viswanath, Anargh, Lokesh, Veeramacheneni, and Buschmeier, Hendrik. “Enhancing explainability with multimodal context representations for smarter robots”. Papers of the 3rd Workshop on Explainability in Human-Robot Collaboration at HRI ’25. 2025.
Alle Dateien verfügbar unter der/den folgenden Lizenz(en):
Creative Commons Namensnennung 4.0 International Public License (CC-BY 4.0):
Volltext(e)
Name
xhri2025-paper-FINAL.pdf
1.48 MB
Access Level

Zuletzt Hochgeladen
2025-03-24T14:51:54Z
MD5 Prüfsumme
f656a5ab609af2fb0968f65391a8dd4a