MMSR 2021 Proceedings Home | MMSR 2021 WEBSITE | ACL WEBSITE


The 1st Workshop on Multimodal Semantic Representations (MMSR)

WORKSHOP PROGRAM

Wednesday, June 16, 2021

16:00–16:15Introduction
16:15–17:00Invited Talk: From action to language through gesture
Virginia Volterra and Chiara Bonsignori
 17:05–17:35 Oral Session 1
 What is Multimodality?
Letitia Parcalabescu, Nils Trost and Anette Frank
 Are Gestures Worth a Thousand Words? An Analysis of Interviews in the Political Domain
Daniela Trotta and Sara Tonelli
 Requesting clarifications with speech and gestures
Jonathan Ginzburg and Andy Luecking
17:40–18:25Invited Talk: Attention, Incrementality, and Meaning: On the Interplay between Language and Vision in Reference Resolution
Matthias Scheutz
 18:30–19:10 Oral Session 2
 Seeing past words: Testing the cross-modal capabilities of pretrained V&L models on counting tasks
Letitia Parcalabescu, Albert Gatt, Anette Frank and Iacer Calixto
 How Vision Affects Language: Comparing Masked Self-Attention in Uni-Modal and Multi-Modal Transformer
Nikolai Ilinykh and Simon Dobnik
 EMISSOR: A platform for capturing multimodal interactions as Episodic Memories and Interpretations with Situated Scenario-based Ontological References
Selene Baez Santamaria, Thomas Baier, Taewoon Kim, Lea Krause, Jaap Kruijt and Piek Vossen
 Annotating anaphoric phenomena in situated dialogue
Sharid Loáiciga, Simon Dobnik and David Schlangen
 19:15–19:45 Poster Session
 Incremental Unit Networks for Multimodal, Fine-grained Information State Representation
Casey Kennington and David Schlangen
 Teaching Arm and Head Gestures to a Humanoid Robot through Interactive Demonstration and Spoken Instruction
Michael Brady and Han Du
 Building a Video-and-Language Dataset with Human Actions for Multimodal Logical Inference
Riko Suzuki, Hitomi Yanaka, Koji Mineshima and Daisuke Bekki
19:45–20:00Closing