Research project within the DFG Priority Programme Visual Communication (01.10.2022-31.09.2025, https://vicom.info/).

ViCom investigates the special features and linguistic significance of visual communication. This comprises sign languages as fully developed natural languages which exclusively rely on the visual channel for communication, but also visual means that enhance spoken language such as gestures. It aims at disclosing the specific characteristics of the visual modality as a communication channel and its interaction with other channels (especially the auditory channel) to develop a comprehensive theoretical linguistic model of human communication and its cognitive foundations.
About GeMDiS
Both corpus-based linguistics and contemporary computational linguistics rely on the use of often large, linguistic resources. The expansion of the linguistic subject area to include visual means of communication such as gesticulation has not yet been backed up with corresponding corpora. This means that “multimodal linguistics” and dialogue theory cannot participate in established distributional methods of corpus linguistics and computational semantics. The main reason for this is the difficulty of collecting multimodal data in an appropriate way and at an appropriate scale. Using the latest VR-based recording methods, the GeMDiS project aims to close this data gap and to investigate visual communication by means of machine-based methods and innovative use of neuronal and active learning for small data using the systematic reference dimensions of associativity and contiguity of the features of visual and non-visual communicative signs. GeMDiS is characterised above all by the following characteristics:
- Ecological validity: the data collection takes place in dialogue situations and thus also takes a look at everyday gestures or interactive gestures in particular. In this respect, GeMDS differs from collections of partly emblematic hand shapes or gestural charades.
- True multimodality: the VR-based recording technology records not only hand-and- arm movements and handshapes but also facial expressions — it is this proper multimodality that is the hallmark of natural language interaction. In this, GeMDS already anticipates potential further developments of ViCom.
The corpus created in this way is made available to the research community (FAIR principles). The results of GeMDS feed into social human-machine interaction, contribute to research on gesture families, and provide a basis for exploratory corpus analysis and further annotation. Furthermore, the project investigates to what extent the results obtained can serve formal semantics for the input problem of meaning representation (in short: in order to compute a multimodal meaning compositionally, it is first of all necessary to associate the linguistic and the non-vocal parts of an utterance with meanings, something that so far only happens intuitively). In the last phase of the project, a VR avatar will be developed into a playback medium of the previously recorded multimodal behaviour. This serves as a visual evaluation of the methodology. The avatar can also be used as an experimental platform, e.g. in cooperation with other projects.
Publications and other activities
BibTeX
@misc{Luecking:Henlein:2025-esslli,
author = {Lücking, Andy and Henlein, Alexander},
year = {2025},
date = {2025-07-28/2025-08-08},
title = {Spatial Gesture Semantics},
howpublished = {ESSLLI 2025 Advanced Course, Ruhr University Bochum},
note = {ESSLLI 2025 Advanced Course, Ruhr University Bochum},
url = {https://aluecking.github.io/ESSLLI2025/},
keywords = {gemdis}
}
BibTeX
@misc{Luecking:2025-mmsr,
author = {Lücking, Andy},
keywords = {gemdis},
title = {From Gesture Representation to Spatial Gesture Semantics},
howpublished = {Invited talk at the IWCS Workshop \textit{Beyond
Language: Multimodal Semantic Representations} (MMSR
II), Heinrich Heine University, Düsseldorf},
date = {2025-09-24},
note = {Invited talk},
url = {https://mmsr-workshop.github.io/}
}
BibTeX
@inproceedings{Luecking:Voll:Rott:Henlein:Mehler:2025-fraga,
title = {Head and Hand Movements During Turn Transitions: Data-Based Multimodal
Analysis Using the {Frankfurt VR Gesture--Speech Alignment Corpus}
({FraGA})},
author = {Lücking, Andy and Voll, Felix and Rott, Daniel and Henlein, Alexander
and Mehler, Alexander},
year = {2025},
booktitle = {Proceedings of the 29th Workshop on The Semantics and Pragmatics
of Dialogue -- Full Papers},
series = {SemDial'25 -- Bialogue},
publisher = {SEMDIAL},
url = {http://semdial.org/anthology/Z25-Luecking_semdial_3316.pdf},
pages = {146--156},
keywords = {gemdis}
}
BibTeX
@incollection{Luecking:2025-wsk-deixis,
booktitle = {Wörterbücher zur Sprach- und Kommunikationswissenschaft (WSK) Online},
url = {https://www.degruyterbrill.com/database/WSK/entry/wsk_idf6497d06-df52-478d-a30f-a3a99151d81e/html},
editor = {Schierholz, Stefan J. and Giacomini, Laura},
year = {2025},
publisher = {De Gruyter},
address = {Berlin, Boston},
doi = {10.1515/wsk},
title = {Deixis},
author = {Lücking, Andy},
keywords = {gemdis}
}
BibTeX
@article{Luecking:Ginzburg:2025-exceptions,
author = {Lücking, Andy and Ginzburg, Jonathan},
title = {Exceptions From Rules and Noteworthy Exceptions},
subtitle = {The Balance Scale for Making Exceptions},
journal = {Linguistics and Philosophy},
year = {2025},
volume = {48},
pages = {371-409},
url = {https://doi.org/10.1007/s10988-024-09429-1},
doi = {10.1007/s10988-024-09429-1},
keywords = {gemdis,neglab}
}
BibTeX
@misc{Luecking:2024-quebec,
author = {Lücking, Andy},
keywords = {gemdis},
title = {Gesture semantics: Deictic Reference, deferred reference, and
iconic co-speech gestures},
howpublished = {Invited talk at Stevan Harnad's interdisciplinary
seminar series in Cognitive Informatics at the
Université du Québec à Montréal},
note = {Invited talk},
date = {2024-03-14}
}
BibTeX
@inproceedings{Henlein:Luecking:Mehler:2024,
title = {Virtually Restricting Modalities in Interactions: Va.Si.Li-Lab
for Experimental Multimodal Research},
author = {Henlein, Alexander and L{\"u}cking, Andy and Mehler, Alexander},
booktitle = {Proceedings of the 2nd International Symposium on Multimodal Communication
(MMSYM 2024), Frankfurt, 25-27 September 2024},
pages = {96--97},
year = {2024},
pdf = {http://mmsym.org/wp-content/uploads/2024/09/BookOfAbstractsMMSYM2024-3.pdf},
keywords = {gemdis}
}
BibTeX
@inproceedings{Luecking:Mehler:Henlein:2024,
title = {The Gesture–Prosody Link in Multimodal Grammar},
author = {L{\"u}cking, Andy and Mehler, Alexander and Henlein, Alexander},
booktitle = {Proceedings of the 2nd International Symposium on Multimodal Communication
(MMSYM 2024), Frankfurt, 25-27 September 2024},
pages = {128--129},
year = {2024},
pdf = {http://mmsym.org/wp-content/uploads/2024/09/BookOfAbstractsMMSYM2024-3.pdf},
keywords = {gemdis}
}
BibTeX
@inproceedings{Luecking:Mehler:Henlein:2024-classifier,
title = {The Linguistic Interpretation of Non-emblematic Gestures Must
be agreed in Dialogue: Combining Perceptual Classifiers and Grounding/Clarification
Mechanisms},
author = {Lücking, Andy and Mehler, Alexander and Henlein, Alexander},
year = {2024},
booktitle = {Proceedings of the 28th Workshop on The Semantics and Pragmatics of Dialogue},
series = {SemDial'24 -- TrentoLogue},
location = {Università di Trento, Palazzo Piomarta, Rovereto},
url = {https://www.semdial.org/anthology/papers/Z/Z24/Z24-4031/},
pdf = {http://semdial.org/anthology/Z24-Lucking_semdial_0031.pdf},
keywords = {gemdis}
}
BibTeX
@inproceedings{Henlein:et:al:2024-vicom,
title = {An Outlook for AI Innovation in Multimodal Communication Research},
author = {Henlein, Alexander and Bauer, Anastasia and Bhattacharjee, Reetu
and Ćwiek, Aleksandra and Gregori, Alina and Kügler, Frank and Lemanski, Jens
and Lücking, Andy and Mehler, Alexander and Prieto, Pilar and Sánchez-Ramón, Paula G.
and Schepens, Job and Schulte-Rüther, Martin and Schweinberger, Stefan R.
and von Eiff, Celina I.},
editor = {Duffy, Vincent G.},
year = {2024},
booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
and Risk Management.},
series = {HCII 2024. Lecture Notes in Computer Science},
publisher = {Springer},
address = {Cham},
pages = {182--234},
isbn = {978-3-031-61066-0},
keywords = {gemdis}
}
BibTeX
@inproceedings{Gregori:et:al:2023-vicom,
author = {Gregori, Alina and Amici, Federica and Brilmayer, Ingmar and {\'{C}}wiek, Aleksandra
and Fritzsche, Lennart and Fuchs, Susanne and Henlein, Alexander and Herbort, Oliver
and K{\"u}gler, Frank and Lemanski, Jens and Liebal, Katja and L{\"u}cking, Andy
and Mehler, Alexander and Nguyen, Kim Tien and Pouw, Wim and Prieto, Pilar
and Rohrer, Patrick Louis and S{\'a}nchez-Ram{\'o}n, Paula G. and Schulte-R{\"u}ther, Martin
and Schumacher, Petra B. and Schweinberger, Stefan R. and Struckmeier, Volker
and Trettenbrein, Patrick C. and von Eiff, Celina I.},
editor = {Duffy, Vincent G.},
title = {A Roadmap for Technological Innovation in Multimodal Communication Research},
booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
and Risk Management},
year = {2023},
publisher = {Springer Nature Switzerland},
address = {Cham},
pages = {402--438},
abstract = {Multimodal communication research focuses on how different means
of signalling coordinate to communicate effectively. This line
of research is traditionally influenced by fields such as cognitive
and neuroscience, human-computer interaction, and linguistics.
With new technologies becoming available in fields such as natural
language processing and computer vision, the field can increasingly
avail itself of new ways of analyzing and understanding multimodal
communication. As a result, there is a general hope that multimodal
research may be at the ``precipice of greatness'' due to technological
advances in computer science and resulting extended empirical
coverage. However, for this to come about there must be sufficient
guidance on key (theoretical) needs of innovation in the field
of multimodal communication. Absent such guidance, the research
focus of computer scientists might increasingly diverge from crucial
issues in multimodal communication. With this paper, we want to
further promote interaction between these fields, which may enormously
benefit both communities. The multimodal research community (represented
here by a consortium of researchers from the Visual Communication
[ViCom] Priority Programme) can engage in the innovation by clearly
stating which technological tools are needed to make progress
in the field of multimodal communication. In this article, we
try to facilitate the establishment of a much needed common ground
on feasible expectations (e.g., in terms of terminology and measures
to be able to train machine learning algorithms) and to critically
reflect possibly idle hopes for technical advances, informed by
recent successes and challenges in computer science, social signal
processing, and related domains.},
isbn = {978-3-031-35748-0},
pdf = {https://pure.mpg.de/rest/items/item_3511464_5/component/file_3520176/content},
keywords = {gemdis}
}
BibTeX
@inproceedings{Luecking:2023-rtt-annotation,
author = {Lücking, Andy},
keywords = {gemdis},
title = {Towards Referential Transparent Annotations of Quantified Noun Phrases},
booktitle = {Proceedings of the 2023 Joint ACL--ISO Workshop on Interoperable
Semantic Annotation},
series = {ISA-19},
pages = {47-55},
location = {Nancy, France, hosted with IWCS 2023},
year = {2023},
url = {https://aclanthology.org/2023.isa-1.7/},
pdf = {https://aclanthology.org/2023.isa-1.7.pdf}
}
BibTeX
@inproceedings{Henlein:et:al:2023c,
title = {Towards grounding multimodal semantics in interaction data with Va.Si.Li-Lab},
author = {Henlein, Alexander and Lücking, Andy and Bagci, Mevlüt and Mehler, Alexander},
booktitle = {Proceedings of the 8th Conference on Gesture and Speech in Interaction (GESPIN)},
location = {Nijmegen, Netherlands},
year = {2023},
keywords = {vasililab, gemdis},
pdf = {https://www.gespin2023.nl/documents/talks_and_posters/GeSpIn_2023_papers/GeSpIn_2023_paper_1692.pdf}
}
BibTeX
@inproceedings{Abrami:et:al:2023,
author = {Abrami, Giuseppe and Mehler, Alexander and Bagci, Mevl\"{u}t and Schrottenbacher, Patrick
and Henlein, Alexander and Spiekermann, Christian and Engel, Juliane
and Schreiber, Jakob},
title = {Va.Si.Li-Lab as a Collaborative Multi-User Annotation Tool in
Virtual Reality and Its Potential Fields of Application},
year = {2023},
isbn = {9798400702327},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3603163.3609076},
doi = {10.1145/3603163.3609076},
abstract = {During the last thirty years a variety of hypertext approaches
and virtual environments -- some virtual hypertext environments
-- have been developed and discussed. Although the development
of virtual and augmented reality technologies is rapid and improving,
and many technologies can be used at affordable conditions, their
usability for hypertext systems has not yet been explored. At
the same time, even for virtual three-dimensional virtual and
augmented environments, there is no generally accepted concept
that is similar or nearly as elegant as hypertext. This gap will
have to be filled in the next years and a good concept should
be developed; in this article we aim to contribute in this direction
and also introduce a prototype for a possible implementation of
criteria for virtual hypertext simulations.},
booktitle = {Proceedings of the 34th ACM Conference on Hypertext and Social Media},
articleno = {22},
numpages = {9},
keywords = {VaSiLiLab, virtual hypertext, virtual reality, virtual reality simulation, authoring system, gemdis},
location = {Rome, Italy},
series = {HT '23},
pdf = {https://dl.acm.org/doi/pdf/10.1145/3603163.3609076}
}
BibTeX
@phdthesis{Henlein:2023,
author = {Alexander Henlein},
title = {Toward context-based text-to-3D scene generation},
type = {doctoralthesis},
pages = {199},
school = {Johann Wolfgang Goethe-Universität},
doi = {10.21248/gups.73448},
year = {2023},
pdf = {https://publikationen.ub.uni-frankfurt.de/files/73448/main.pdf},
keywords = {gemdis}
}
BibTeX
@article{Henlein:et:al:2023a,
author = {Henlein, Alexander and Gopinath, Anju and Krishnaswamy, Nikhil
and Mehler, Alexander and Pustejovsky, James},
doi = {10.3389/frai.2023.1084740},
issn = {2624-8212},
journal = {Frontiers in Artificial Intelligence},
title = {Grounding human-object interaction to affordance behavior in multimodal datasets},
url = {https://www.frontiersin.org/articles/10.3389/frai.2023.1084740},
volume = {6},
year = {2023},
keywords = {gemdis}
}
BibTeX
@inproceedings{Henlein:et:al:2023b,
author = {Henlein, Alexander and Kett, Attila and Baumartz, Daniel and Abrami, Giuseppe
and Mehler, Alexander and Bastian, Johannes and Blecher, Yannic and Budgenhagen, David
and Christof, Roman and Ewald, Tim-Oliver and Fauerbach, Tim and Masny, Patrick
and Mende, Julian and Schn{\"u}re, Paul and Viel, Marc},
editor = {Duffy, Vincent G.},
title = {Semantic Scene Builder: Towards a Context Sensitive Text-to-3D Scene Framework},
booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
and Risk Management},
year = {2023},
publisher = {Springer Nature Switzerland},
address = {Cham},
pages = {461--479},
abstract = {We introduce Semantic Scene Builder (SeSB), a VR-based text-to-3D
scene framework using SemAF (Semantic Annotation Framework) as
a scheme for annotating discourse structures. SeSB integrates
a variety of tools and resources by using SemAF and UIMA as a
unified data structure to generate 3D scenes from textual descriptions.
Based on VR, SeSB allows its users to change annotations through
body movements instead of symbolic manipulations: from annotations
in texts to corrections in editing steps to adjustments in generated
scenes, all this is done by grabbing and moving objects. We evaluate
SeSB in comparison with a state-of-the-art open source text-to-scene
method (the only one which is publicly available) and find that
our approach not only performs better, but also allows for modeling
a greater variety of scenes.},
isbn = {978-3-031-35748-0},
doi = {10.1007/978-3-031-35748-0_32},
keywords = {gemdis}
}
BibTeX
@inproceedings{Mehler:et:al:2023:a,
abstract = {Simulation-based learning is a method in which learners learn
to master real-life scenarios and tasks from simulated application
contexts. It is particularly suitable for the use of VR technologies,
as these allow immersive experiences of the targeted scenarios.
VR methods are also relevant for studies on online learning, especially
in groups, as they provide access to a variety of multimodal learning
and interaction data. However, VR leads to a trade-off between
technological conditions of the observability of such data and
the openness of learner behavior. We present Va.Si.Li-Lab, a VR-L
ab for Simulation-based Learn ing developed to address this trade-off.
Va.Si.Li-Lab uses a graph-theoretical model based on hypergraphs
to represent the data diversity of multimodal learning and interaction.
We develop this data model in relation to mono- and multimodal,
intra- and interpersonal data and interleave it with ISO-Space
to describe distributed multiple documents from the perspective
of their interactive generation. The paper adds three use cases
to motivate the broad applicability of Va.Si.Li-Lab and its data
model.},
address = {Cham},
author = {Mehler, Alexander and Bagci, Mevl{\"u}t and Henlein, Alexander
and Abrami, Giuseppe and Spiekermann, Christian and Schrottenbacher, Patrick
and Konca, Maxim and L{\"u}cking, Andy and Engel, Juliane and Quintino, Marc
and Schreiber, Jakob and Saukel, Kevin and Zlatkin-Troitschanskaia, Olga},
booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
and Risk Management},
editor = {Duffy, Vincent G.},
isbn = {978-3-031-35741-1},
pages = {539--565},
publisher = {Springer Nature Switzerland},
title = {A Multimodal Data Model for Simulation-Based Learning with Va.Si.Li-Lab},
year = {2023},
doi = {10.1007/978-3-031-35741-1_39},
keywords = {gemdis}
}
BibTeX
@misc{Luecking:2022-bochum,
author = {Lücking, Andy},
keywords = {gemdis},
title = {Pointing: From reference to attention and back},
howpublished = {Invited talk at the Language Colloquium,
Ruhr-Universit{\"a}t Bochum},
note = {Invited talk},
date = {2022-05-03}
}
BibTeX
@article{Luecking:Ginzburg:2022-lv,
title = {Leading voices: {Dialogue} semantics, cognitive science, and the
polyphonic structure of multimodal interaction},
author = {L{\"u}cking, Andy and Ginzburg, Jonathan},
journal = {Language and Cognition},
year = {2022},
doi = {10.1017/langcog.2022.30},
keywords = {gemdis}
}
