GeMDiS

Research project within the DFG Priority Programme Visual Communication (01.10.2022-31.09.2025, https://vicom.info/).

About ViCom

ViCom investigates the special features and linguistic significance of visual communication. This comprises sign languages as fully developed natural languages which exclusively rely on the visual channel for communication, but also visual means that enhance spoken language such as gestures. It aims at disclosing the specific characteristics of the visual modality as a communication channel and its interaction with other channels (especially the auditory channel) to develop a comprehensive theoretical linguistic model of human communication and its cognitive foundations.


About GeMDiS

Both corpus-based linguistics and contemporary computational linguistics rely on the use of often large, linguistic resources. The expansion of the linguistic subject area to include visual means of communication such as gesticulation has not yet been backed up with corresponding corpora. This means that “multimodal linguistics” and dialogue theory cannot participate in established distributional methods of corpus linguistics and computational semantics. The main reason for this is the difficulty of collecting multimodal data in an appropriate way and at an appropriate scale. Using the latest VR-based recording methods, the GeMDiS project aims to close this data gap and to investigate visual communication by means of machine-based methods and innovative use of neuronal and active learning for small data using the systematic reference dimensions of associativity and contiguity of the features of visual and non-visual communicative signs. GeMDiS is characterised above all by the following characteristics:

  • Ecological validity: the data collection takes place in dialogue situations and thus also takes a look at everyday gestures or interactive gestures in particular. In this respect, GeMDS differs from collections of partly emblematic hand shapes or gestural charades.
  • True multimodality: the VR-based recording technology records not only hand-and- arm movements and handshapes but also facial expressions — it is this proper multimodality that is the hallmark of natural language interaction. In this, GeMDS already anticipates potential further developments of ViCom.

The corpus created in this way is made available to the research community (FAIR principles). The results of GeMDS feed into social human-machine interaction, contribute to research on gesture families, and provide a basis for exploratory corpus analysis and further annotation. Furthermore, the project investigates to what extent the results obtained can serve formal semantics for the input problem of meaning representation (in short: in order to compute a multimodal meaning compositionally, it is first of all necessary to associate the linguistic and the non-vocal parts of an utterance with meanings, something that so far only happens intuitively). In the last phase of the project, a VR avatar will be developed into a playback medium of the previously recorded multimodal behaviour. This serves as a visual evaluation of the methodology. The avatar can also be used as an experimental platform, e.g. in cooperation with other projects.

Publications and other activities

Andy Lücking and Alexander Henlein. 2025–07–28/2025–08–08. Spatial Gesture Semantics. ESSLLI 2025 Advanced Course, Ruhr University Bochum.
BibTeX
@misc{Luecking:Henlein:2025-esslli,
  author    = {Lücking, Andy and Henlein, Alexander},
  year      = {2025},
  date      = {2025-07-28/2025-08-08},
  title     = {Spatial Gesture Semantics},
  howpublished = {ESSLLI 2025 Advanced Course, Ruhr University Bochum},
  note      = {ESSLLI 2025 Advanced Course, Ruhr University Bochum},
  url       = {https://aluecking.github.io/ESSLLI2025/},
  keywords  = {gemdis}
}
Andy Lücking. 2025–09–24. From Gesture Representation to Spatial Gesture Semantics. Invited talk.
BibTeX
@misc{Luecking:2025-mmsr,
  author    = {Lücking, Andy},
  keywords  = {gemdis},
  title     = {From Gesture Representation to Spatial Gesture Semantics},
  howpublished = {Invited talk at the IWCS Workshop \textit{Beyond
                  Language: Multimodal Semantic Representations} (MMSR
                  II), Heinrich Heine University, Düsseldorf},
  date      = {2025-09-24},
  note      = {Invited talk},
  url       = {https://mmsr-workshop.github.io/}
}
Andy Lücking, Felix Voll, Daniel Rott, Alexander Henlein and Alexander Mehler. 2025. Head and Hand Movements During Turn Transitions: Data-Based Multimodal Analysis Using the Frankfurt VR Gesture–Speech Alignment Corpus (FraGA). Proceedings of the 29th Workshop on The Semantics and Pragmatics of Dialogue – Full Papers, 146–156.
BibTeX
@inproceedings{Luecking:Voll:Rott:Henlein:Mehler:2025-fraga,
  title     = {Head and Hand Movements During Turn Transitions: Data-Based Multimodal
               Analysis Using the {Frankfurt VR Gesture--Speech Alignment Corpus}
               ({FraGA})},
  author    = {Lücking, Andy and Voll, Felix and Rott, Daniel and Henlein, Alexander
               and Mehler, Alexander},
  year      = {2025},
  booktitle = {Proceedings of the 29th Workshop on The Semantics and Pragmatics
               of Dialogue -- Full Papers},
  series    = {SemDial'25 -- Bialogue},
  publisher = {SEMDIAL},
  url       = {http://semdial.org/anthology/Z25-Luecking_semdial_3316.pdf},
  pages     = {146--156},
  keywords  = {gemdis}
}
Andy Lücking. 2025. Deixis. Wörterbücher zur Sprach- und Kommunikationswissenschaft (WSK) Online.
BibTeX
@incollection{Luecking:2025-wsk-deixis,
  booktitle = {Wörterbücher zur Sprach- und Kommunikationswissenschaft (WSK) Online},
  url       = {https://www.degruyterbrill.com/database/WSK/entry/wsk_idf6497d06-df52-478d-a30f-a3a99151d81e/html},
  editor    = {Schierholz, Stefan J. and Giacomini, Laura},
  year      = {2025},
  publisher = {De Gruyter},
  address   = {Berlin, Boston},
  doi       = {10.1515/wsk},
  title     = {Deixis},
  author    = {Lücking, Andy},
  keywords  = {gemdis}
}
Andy Lücking and Jonathan Ginzburg. 2025. Exceptions From Rules and Noteworthy Exceptions. Linguistics and Philosophy, 48:371–409.
BibTeX
@article{Luecking:Ginzburg:2025-exceptions,
  author    = {Lücking, Andy and Ginzburg, Jonathan},
  title     = {Exceptions From Rules and Noteworthy Exceptions},
  subtitle  = {The Balance Scale for Making Exceptions},
  journal   = {Linguistics and Philosophy},
  year      = {2025},
  volume    = {48},
  pages     = {371-409},
  url       = {https://doi.org/10.1007/s10988-024-09429-1},
  doi       = {10.1007/s10988-024-09429-1},
  keywords  = {gemdis,neglab}
}
Andy Lücking. 2024–03–14. Gesture semantics: Deictic Reference, deferred reference, and iconic co-speech gestures. Invited talk.
BibTeX
@misc{Luecking:2024-quebec,
  author    = {Lücking, Andy},
  keywords  = {gemdis},
  title     = {Gesture semantics: Deictic Reference, deferred reference, and
               iconic co-speech gestures},
  howpublished = {Invited talk at Stevan Harnad's interdisciplinary
                  seminar series in Cognitive Informatics at the
                  Université du Québec à Montréal},
  note      = {Invited talk},
  date      = {2024-03-14}
}
Alexander Henlein, Andy Lücking and Alexander Mehler. 2024. Virtually Restricting Modalities in Interactions: Va.Si.Li-Lab for Experimental Multimodal Research. Proceedings of the 2nd International Symposium on Multimodal Communication (MMSYM 2024), Frankfurt, 25-27 September 2024, 96–97.
BibTeX
@inproceedings{Henlein:Luecking:Mehler:2024,
  title     = {Virtually Restricting Modalities in Interactions: Va.Si.Li-Lab
               for Experimental Multimodal Research},
  author    = {Henlein, Alexander and L{\"u}cking, Andy and Mehler, Alexander},
  booktitle = {Proceedings of the 2nd International Symposium on Multimodal Communication
               (MMSYM 2024), Frankfurt, 25-27 September 2024},
  pages     = {96--97},
  year      = {2024},
  pdf       = {http://mmsym.org/wp-content/uploads/2024/09/BookOfAbstractsMMSYM2024-3.pdf},
  keywords  = {gemdis}
}
Andy Lücking, Alexander Mehler and Alexander Henlein. 2024. The Gesture–Prosody Link in Multimodal Grammar. Proceedings of the 2nd International Symposium on Multimodal Communication (MMSYM 2024), Frankfurt, 25-27 September 2024, 128–129.
BibTeX
@inproceedings{Luecking:Mehler:Henlein:2024,
  title     = {The Gesture–Prosody Link in Multimodal Grammar},
  author    = {L{\"u}cking, Andy and Mehler, Alexander and Henlein, Alexander},
  booktitle = {Proceedings of the 2nd International Symposium on Multimodal Communication
               (MMSYM 2024), Frankfurt, 25-27 September 2024},
  pages     = {128--129},
  year      = {2024},
  pdf       = {http://mmsym.org/wp-content/uploads/2024/09/BookOfAbstractsMMSYM2024-3.pdf},
  keywords  = {gemdis}
}
Andy Lücking, Alexander Mehler and Alexander Henlein. 2024. The Linguistic Interpretation of Non-emblematic Gestures Must be agreed in Dialogue: Combining Perceptual Classifiers and Grounding/Clarification Mechanisms. Proceedings of the 28th Workshop on The Semantics and Pragmatics of Dialogue.
BibTeX
@inproceedings{Luecking:Mehler:Henlein:2024-classifier,
  title     = {The Linguistic Interpretation of Non-emblematic Gestures Must
               be agreed in Dialogue: Combining Perceptual Classifiers and Grounding/Clarification
               Mechanisms},
  author    = {Lücking, Andy and Mehler, Alexander and Henlein, Alexander},
  year      = {2024},
  booktitle = {Proceedings of the 28th Workshop on The Semantics and Pragmatics of Dialogue},
  series    = {SemDial'24 -- TrentoLogue},
  location  = {Università di Trento, Palazzo Piomarta, Rovereto},
  url       = {https://www.semdial.org/anthology/papers/Z/Z24/Z24-4031/},
  pdf       = {http://semdial.org/anthology/Z24-Lucking_semdial_0031.pdf},
  keywords  = {gemdis}
}
Alexander Henlein, Anastasia Bauer, Reetu Bhattacharjee, Aleksandra Ćwiek, Alina Gregori, Frank Kügler, Jens Lemanski, Andy Lücking, Alexander Mehler, Pilar Prieto, Paula G. Sánchez-Ramón, Job Schepens, Martin Schulte-Rüther, Stefan R. Schweinberger and Celina I. von Eiff. 2024. An Outlook for AI Innovation in Multimodal Communication Research. Digital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management., 182–234.
BibTeX
@inproceedings{Henlein:et:al:2024-vicom,
  title     = {An Outlook for AI Innovation in Multimodal Communication Research},
  author    = {Henlein, Alexander and Bauer, Anastasia and Bhattacharjee, Reetu
               and Ćwiek, Aleksandra and Gregori, Alina and Kügler, Frank and Lemanski, Jens
               and Lücking, Andy and Mehler, Alexander and Prieto, Pilar and Sánchez-Ramón, Paula G.
               and Schepens, Job and Schulte-Rüther, Martin and Schweinberger, Stefan R.
               and von Eiff, Celina I.},
  editor    = {Duffy, Vincent G.},
  year      = {2024},
  booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
               and Risk Management.},
  series    = {HCII 2024. Lecture Notes in Computer Science},
  publisher = {Springer},
  address   = {Cham},
  pages     = {182--234},
  isbn      = {978-3-031-61066-0},
  keywords  = {gemdis}
}
Alina Gregori, Federica Amici, Ingmar Brilmayer, Aleksandra Ćwiek, Lennart Fritzsche, Susanne Fuchs, Alexander Henlein, Oliver Herbort, Frank Kügler, Jens Lemanski, Katja Liebal, Andy Lücking, Alexander Mehler, Kim Tien Nguyen, Wim Pouw, Pilar Prieto, Patrick Louis Rohrer, Paula G. Sánchez-Ramón, Martin Schulte-Rüther, Petra B. Schumacher, Stefan R. Schweinberger, Volker Struckmeier, Patrick C. Trettenbrein and Celina I. von Eiff. 2023. A Roadmap for Technological Innovation in Multimodal Communication Research. Digital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management, 402–438.
BibTeX
@inproceedings{Gregori:et:al:2023-vicom,
  author    = {Gregori, Alina and Amici, Federica and Brilmayer, Ingmar and {\'{C}}wiek, Aleksandra
               and Fritzsche, Lennart and Fuchs, Susanne and Henlein, Alexander and Herbort, Oliver
               and K{\"u}gler, Frank and Lemanski, Jens and Liebal, Katja and L{\"u}cking, Andy
               and Mehler, Alexander and Nguyen, Kim Tien and Pouw, Wim and Prieto, Pilar
               and Rohrer, Patrick Louis and S{\'a}nchez-Ram{\'o}n, Paula G. and Schulte-R{\"u}ther, Martin
               and Schumacher, Petra B. and Schweinberger, Stefan R. and Struckmeier, Volker
               and Trettenbrein, Patrick C. and von Eiff, Celina I.},
  editor    = {Duffy, Vincent G.},
  title     = {A Roadmap for Technological Innovation in Multimodal Communication Research},
  booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
               and Risk Management},
  year      = {2023},
  publisher = {Springer Nature Switzerland},
  address   = {Cham},
  pages     = {402--438},
  abstract  = {Multimodal communication research focuses on how different means
               of signalling coordinate to communicate effectively. This line
               of research is traditionally influenced by fields such as cognitive
               and neuroscience, human-computer interaction, and linguistics.
               With new technologies becoming available in fields such as natural
               language processing and computer vision, the field can increasingly
               avail itself of new ways of analyzing and understanding multimodal
               communication. As a result, there is a general hope that multimodal
               research may be at the ``precipice of greatness'' due to technological
               advances in computer science and resulting extended empirical
               coverage. However, for this to come about there must be sufficient
               guidance on key (theoretical) needs of innovation in the field
               of multimodal communication. Absent such guidance, the research
               focus of computer scientists might increasingly diverge from crucial
               issues in multimodal communication. With this paper, we want to
               further promote interaction between these fields, which may enormously
               benefit both communities. The multimodal research community (represented
               here by a consortium of researchers from the Visual Communication
               [ViCom] Priority Programme) can engage in the innovation by clearly
               stating which technological tools are needed to make progress
               in the field of multimodal communication. In this article, we
               try to facilitate the establishment of a much needed common ground
               on feasible expectations (e.g., in terms of terminology and measures
               to be able to train machine learning algorithms) and to critically
               reflect possibly idle hopes for technical advances, informed by
               recent successes and challenges in computer science, social signal
               processing, and related domains.},
  isbn      = {978-3-031-35748-0},
  pdf       = {https://pure.mpg.de/rest/items/item_3511464_5/component/file_3520176/content},
  keywords  = {gemdis}
}
Andy Lücking. 2023. Towards Referential Transparent Annotations of Quantified Noun Phrases. Proceedings of the 2023 Joint ACL–ISO Workshop on Interoperable Semantic Annotation, 47–55.
BibTeX
@inproceedings{Luecking:2023-rtt-annotation,
  author    = {Lücking, Andy},
  keywords  = {gemdis},
  title     = {Towards Referential Transparent Annotations of Quantified Noun Phrases},
  booktitle = {Proceedings of the 2023 Joint ACL--ISO Workshop on Interoperable
               Semantic Annotation},
  series    = {ISA-19},
  pages     = {47-55},
  location  = {Nancy, France, hosted with IWCS 2023},
  year      = {2023},
  url       = {https://aclanthology.org/2023.isa-1.7/},
  pdf       = {https://aclanthology.org/2023.isa-1.7.pdf}
}
Alexander Henlein, Andy Lücking, Mevlüt Bagci and Alexander Mehler. 2023. Towards grounding multimodal semantics in interaction data with Va.Si.Li-Lab. Proceedings of the 8th Conference on Gesture and Speech in Interaction (GESPIN).
BibTeX
@inproceedings{Henlein:et:al:2023c,
  title     = {Towards grounding multimodal semantics in interaction data with Va.Si.Li-Lab},
  author    = {Henlein, Alexander and Lücking, Andy and Bagci, Mevlüt and Mehler, Alexander},
  booktitle = {Proceedings of the 8th Conference on Gesture and Speech in Interaction (GESPIN)},
  location  = {Nijmegen, Netherlands},
  year      = {2023},
  keywords  = {vasililab, gemdis},
  pdf       = {https://www.gespin2023.nl/documents/talks_and_posters/GeSpIn_2023_papers/GeSpIn_2023_paper_1692.pdf}
}
Giuseppe Abrami, Alexander Mehler, Mevlüt Bagci, Patrick Schrottenbacher, Alexander Henlein, Christian Spiekermann, Juliane Engel and Jakob Schreiber. 2023. Va.Si.Li-Lab as a Collaborative Multi-User Annotation Tool in Virtual Reality and Its Potential Fields of Application. Proceedings of the 34th ACM Conference on Hypertext and Social Media.
BibTeX
@inproceedings{Abrami:et:al:2023,
  author    = {Abrami, Giuseppe and Mehler, Alexander and Bagci, Mevl\"{u}t and Schrottenbacher, Patrick
               and Henlein, Alexander and Spiekermann, Christian and Engel, Juliane
               and Schreiber, Jakob},
  title     = {Va.Si.Li-Lab as a Collaborative Multi-User Annotation Tool in
               Virtual Reality and Its Potential Fields of Application},
  year      = {2023},
  isbn      = {9798400702327},
  publisher = {Association for Computing Machinery},
  address   = {New York, NY, USA},
  url       = {https://doi.org/10.1145/3603163.3609076},
  doi       = {10.1145/3603163.3609076},
  abstract  = {During the last thirty years a variety of hypertext approaches
               and virtual environments -- some virtual hypertext environments
               -- have been developed and discussed. Although the development
               of virtual and augmented reality technologies is rapid and improving,
               and many technologies can be used at affordable conditions, their
               usability for hypertext systems has not yet been explored. At
               the same time, even for virtual three-dimensional virtual and
               augmented environments, there is no generally accepted concept
               that is similar or nearly as elegant as hypertext. This gap will
               have to be filled in the next years and a good concept should
               be developed; in this article we aim to contribute in this direction
               and also introduce a prototype for a possible implementation of
               criteria for virtual hypertext simulations.},
  booktitle = {Proceedings of the 34th ACM Conference on Hypertext and Social Media},
  articleno = {22},
  numpages  = {9},
  keywords  = {VaSiLiLab, virtual hypertext, virtual reality, virtual reality simulation, authoring system, gemdis},
  location  = {Rome, Italy},
  series    = {HT '23},
  pdf       = {https://dl.acm.org/doi/pdf/10.1145/3603163.3609076}
}
Alexander Henlein. 2023. PhD Thesis: Toward context-based text-to-3D scene generation.
BibTeX
@phdthesis{Henlein:2023,
  author    = {Alexander Henlein},
  title     = {Toward context-based text-to-3D scene generation},
  type      = {doctoralthesis},
  pages     = {199},
  school    = {Johann Wolfgang Goethe-Universität},
  doi       = {10.21248/gups.73448},
  year      = {2023},
  pdf       = {https://publikationen.ub.uni-frankfurt.de/files/73448/main.pdf},
  keywords  = {gemdis}
}
Alexander Henlein, Anju Gopinath, Nikhil Krishnaswamy, Alexander Mehler and James Pustejovsky. 2023. Grounding human-object interaction to affordance behavior in multimodal datasets. Frontiers in Artificial Intelligence, 6.
BibTeX
@article{Henlein:et:al:2023a,
  author    = {Henlein, Alexander and Gopinath, Anju and Krishnaswamy, Nikhil
               and Mehler, Alexander and Pustejovsky, James},
  doi       = {10.3389/frai.2023.1084740},
  issn      = {2624-8212},
  journal   = {Frontiers in Artificial Intelligence},
  title     = {Grounding human-object interaction to affordance behavior in multimodal datasets},
  url       = {https://www.frontiersin.org/articles/10.3389/frai.2023.1084740},
  volume    = {6},
  year      = {2023},
  keywords  = {gemdis}
}
Alexander Henlein, Attila Kett, Daniel Baumartz, Giuseppe Abrami, Alexander Mehler, Johannes Bastian, Yannic Blecher, David Budgenhagen, Roman Christof, Tim-Oliver Ewald, Tim Fauerbach, Patrick Masny, Julian Mende, Paul Schnüre and Marc Viel. 2023. Semantic Scene Builder: Towards a Context Sensitive Text-to-3D Scene Framework. Digital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management, 461–479.
BibTeX
@inproceedings{Henlein:et:al:2023b,
  author    = {Henlein, Alexander and Kett, Attila and Baumartz, Daniel and Abrami, Giuseppe
               and Mehler, Alexander and Bastian, Johannes and Blecher, Yannic and Budgenhagen, David
               and Christof, Roman and Ewald, Tim-Oliver and Fauerbach, Tim and Masny, Patrick
               and Mende, Julian and Schn{\"u}re, Paul and Viel, Marc},
  editor    = {Duffy, Vincent G.},
  title     = {Semantic Scene Builder: Towards a Context Sensitive Text-to-3D Scene Framework},
  booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
               and Risk Management},
  year      = {2023},
  publisher = {Springer Nature Switzerland},
  address   = {Cham},
  pages     = {461--479},
  abstract  = {We introduce Semantic Scene Builder (SeSB), a VR-based text-to-3D
               scene framework using SemAF (Semantic Annotation Framework) as
               a scheme for annotating discourse structures. SeSB integrates
               a variety of tools and resources by using SemAF and UIMA as a
               unified data structure to generate 3D scenes from textual descriptions.
               Based on VR, SeSB allows its users to change annotations through
               body movements instead of symbolic manipulations: from annotations
               in texts to corrections in editing steps to adjustments in generated
               scenes, all this is done by grabbing and moving objects. We evaluate
               SeSB in comparison with a state-of-the-art open source text-to-scene
               method (the only one which is publicly available) and find that
               our approach not only performs better, but also allows for modeling
               a greater variety of scenes.},
  isbn      = {978-3-031-35748-0},
  doi       = {10.1007/978-3-031-35748-0_32},
  keywords  = {gemdis}
}
Alexander Mehler, Mevlüt Bagci, Alexander Henlein, Giuseppe Abrami, Christian Spiekermann, Patrick Schrottenbacher, Maxim Konca, Andy Lücking, Juliane Engel, Marc Quintino, Jakob Schreiber, Kevin Saukel and Olga Zlatkin-Troitschanskaia. 2023. A Multimodal Data Model for Simulation-Based Learning with Va.Si.Li-Lab. Digital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management, 539–565.
BibTeX
@inproceedings{Mehler:et:al:2023:a,
  abstract  = {Simulation-based learning is a method in which learners learn
               to master real-life scenarios and tasks from simulated application
               contexts. It is particularly suitable for the use of VR technologies,
               as these allow immersive experiences of the targeted scenarios.
               VR methods are also relevant for studies on online learning, especially
               in groups, as they provide access to a variety of multimodal learning
               and interaction data. However, VR leads to a trade-off between
               technological conditions of the observability of such data and
               the openness of learner behavior. We present Va.Si.Li-Lab, a VR-L
               ab for Simulation-based Learn ing developed to address this trade-off.
               Va.Si.Li-Lab uses a graph-theoretical model based on hypergraphs
               to represent the data diversity of multimodal learning and interaction.
               We develop this data model in relation to mono- and multimodal,
               intra- and interpersonal data and interleave it with ISO-Space
               to describe distributed multiple documents from the perspective
               of their interactive generation. The paper adds three use cases
               to motivate the broad applicability of Va.Si.Li-Lab and its data
               model.},
  address   = {Cham},
  author    = {Mehler, Alexander and Bagci, Mevl{\"u}t and Henlein, Alexander
               and Abrami, Giuseppe and Spiekermann, Christian and Schrottenbacher, Patrick
               and Konca, Maxim and L{\"u}cking, Andy and Engel, Juliane and Quintino, Marc
               and Schreiber, Jakob and Saukel, Kevin and Zlatkin-Troitschanskaia, Olga},
  booktitle = {Digital Human Modeling and Applications in Health, Safety, Ergonomics
               and Risk Management},
  editor    = {Duffy, Vincent G.},
  isbn      = {978-3-031-35741-1},
  pages     = {539--565},
  publisher = {Springer Nature Switzerland},
  title     = {A Multimodal Data Model for Simulation-Based Learning with Va.Si.Li-Lab},
  year      = {2023},
  doi       = {10.1007/978-3-031-35741-1_39},
  keywords  = {gemdis}
}
Andy Lücking. 2022–05–03. Pointing: From reference to attention and back. Invited talk.
BibTeX
@misc{Luecking:2022-bochum,
  author    = {Lücking, Andy},
  keywords  = {gemdis},
  title     = {Pointing: From reference to attention and back},
  howpublished = {Invited talk at the Language Colloquium,
                  Ruhr-Universit{\"a}t Bochum},
  note      = {Invited talk},
  date      = {2022-05-03}
}
Andy Lücking and Jonathan Ginzburg. 2022. Leading voices: Dialogue semantics, cognitive science, and the polyphonic structure of multimodal interaction. Language and Cognition.
BibTeX
@article{Luecking:Ginzburg:2022-lv,
  title     = {Leading voices: {Dialogue} semantics, cognitive science, and the
               polyphonic structure of multimodal interaction},
  author    = {L{\"u}cking, Andy and Ginzburg, Jonathan},
  journal   = {Language and Cognition},
  year      = {2022},
  doi       = {10.1017/langcog.2022.30},
  keywords  = {gemdis}
}
Andy Lücking and Jonathan Ginzburg. 2022. Referential transparency as the proper treatment of quantification. Semantics and Pragmatics, 15.
BibTeX
@article{Luecking:Ginzburg:2022,
  author    = {L{\"u}cking, Andy and Ginzburg, Jonathan},
  title     = {Referential transparency as the proper treatment of quantification},
  journal   = {Semantics and Pragmatics},
  year      = {2022},
  volume    = {15},
  eid       = {4},
  doi       = {10.3765/sp.15.4},
  keywords  = {gemdis}
}