FIRST CALL FOR PAPERS
*Multimodal Semantic Representations (MMSR II)* Co-located with ECAI 2024 (https://www.ecai2024.eu/) 19-24 October, Santiago de Compostela, Spain (workshop on 19 or 20 October)
*Workshop website*: https://mmsr-workshop.github.io/
*Description* The demand for more sophisticated natural human-computer and human-robot interactions is rapidly increasing as users become more accustomed to conversation-like interactions with AI and NLP systems. Such interactions require not only the robust recognition and generation of expressions through multiple modalities (language, gesture, vision, action, etc.), but also the encoding of situated meaning.
When communications become multimodal, each modality in operation provides an orthogonal angle through which to probe the computational model of the other modalities, including the behaviors and communicative capabilities afforded by each. Multimodal interactions thus require a unified framework and control language through which systems interpret inputs and behaviors and generate informative outputs. This is vital for intelligent and often embodied systems to understand the situation and context that they inhabit, whether in the real world or in a mixed-reality environment shared with humans.
Furthermore, multimodal large language models appear to offer the possibility for more dynamic and contextually rich interactions across various modalities, including facial expressions, gestures, actions, and language. We invite discussion on how representations and pipelines can potentially integrate such state-of-the-art language models.
We solicit papers on multimodal semantic representation, including but not limited to the following topics:
1. Semantic frameworks for individual linguistic co-modalities (e.g. gaze, facial expression); 2. Formal representation of situated conversation and embodiment, including knowledge graphs, designed to represent epistemic state; 3. Design, annotation, and corpora of multimodal interaction and meaning representation; 4. Challenges (including cross-lingual and cross-cultural) in multimodal representation and/or processing; 5. Criteria or frameworks for evaluation of multimodal semantics; 6. Challenges in aligning co-modalities in formal representation and/or NLP tasks; 7. Design and implementation of neurosymbolic or fusion models for multimodal processing (with a representational component); 8. Methods for probing knowledge of multimodal (language and vision) models; 9. Virtual and situated agents that embody multimodal representations of common ground.
*Submission Information* Two types of submissions are solicited: long papers and short papers. Long papers should describe original research and must not exceed 8 pages, excluding references. Short papers (typically system or project descriptions, or ongoing research) must not exceed 4 pages, excluding references. Accepted papers get an extra page in the camera-ready version.
We strongly encourage students to submit to the workshop.
*Important Dates* May 15, 2024: Submissions due June 1, 2024: Notification of acceptance decisions June 21, 2024: Camera-ready papers due
Papers should be formatted using the ECAI style files, available at: https://www.ecai2024.eu/calls/main-track
Papers will be submitted in PDF format via the chairing tool site, with a workshop link available soon: https://chairingtool.com/
Please do not hesitate to reach out with any questions.
Best regards, Richard Brutti, Lucia Donatelli, Nikhil Krishnaswamy, Kenneth Lai, & James Pustejovsky (MMSR II organizers)
Web page: https://mmsr-workshop.github.io/