LINGUIST List 14.3537

Sat Dec 20 2003

Calls: Text/Corpus Ling/Portugal

Editor for this issue: Andrea Berez <>

As a matter of policy, LINGUIST discourages the use of abbreviations or acronyms in conference announcements unless they are explained in the text. To post to LINGUIST, use our convenient web form at


  1. p, Workshop on Multimodal Corpora

Message 1: Workshop on Multimodal Corpora

Date: 19 Dec 2003
From: p <>
Subject: Workshop on Multimodal Corpora

Workshop on Multimodal Corpora
Short Title: MMCORPORA=
Date: 25-May-2004 - 25-May-2004
Location: Lisbon, Portugal
Contact: Peter Kuehnlein
Contact Email:
Meeting URL:
Linguistic Sub-field: Text/Corpus Linguistics 
Call Deadline: 24-Jan-2004
This is a session of the following conference: 4th International
Conference on Language Resources and Evaluation

Meeting Description:
Workshop on Multimodal Corpora: Models of Human Behaviour for the
Specification and Evaluation of Multimodal Input and Output Interfaces
Centro Cultural de Belem
LISBON, Portugal
25th May 2004
In Association with
4th International Conference on Language Resources and Evaluation
Main conference 26-27-28 May 2004 

**EDITOR'S NOTE: This call for papers was posted yesterday, December
19, 2003 with incorrect conference URLs. This posting contains the
corrected URLs, with apologies for any inconveniences caused by the
previous error.**

Workshop on Multimodal Corpora: Models of Human Behaviour for the
Specification and Evaluation of Multimodal Input and Output Interfaces
Centro Cultural de Belem
LISBON, Portugal
25th May 2004
In Association with
4th International Conference on Language Resources and Evaluation
Main conference 26-27-28 May 2004
The primary purpose of this one day workshop is to share information
and engage in the collective planning for the future creation of
usable pluridisciplinary multimodal resources. It will focus on the
following issues regarding multimodal corpora: how researchers build
models of human behaviour out of the annotations of video corpora, how
they use such knowledge for the specification of multimodal input
(e.g. merging users gestures and speech) and output
(e.g. specification of believable and emotional behaviour in Embodied
Conversational Agents) in human computer interfaces, and finally how
they evaluate multimodal systems (e.g. full system evaluation and
glass box evaluation of individual system components).
Topics to be addressed in the workshop include, but are not limited
* Models of human multimodal behaviour in various disciplines
* Integrating different sources of knowledge (literature in
socio-linguistics, corpora annotation)
* Specifications of coding schemes for video annotations		
* Guidelines, standards, specifications, models and best practices for
multimedia and multimodal corpora
* Parallel multimodal corpora for different languages
* Methods, tools, and procedures for the acquisition, creation,
management, access, distribution, and use of multimedia and multimodal
* Methods for the extraction and acquisition of knowledge
(e.g. lexical information, modality modelling) from multimedia and
multimodal corpora
* Ontological aspects of the creation and use of multimodal corpora	
* Machine learning for and from multimedia (i.e., text, audio, video),
multimodal (visual, auditory, tactile), and multicodal (language,
graphics, gesture) communication
* Exploitation of multimodal corpora in different types of
applications (information extraction, information retrieval, meeting
transcription, multisensorial interfaces, translation, summarisation,
www services, etc.)
* Multimedia and multimodal metadata descriptions of corpora	
* Applications enabled by multimedia and multimodal corpora
* Benchmarking of systems and products; use of multimodal corpora for
the evaluation of real systems
* Processing and evaluation of mixed spoken, typed, and cursive (e.g.,
pen) language processing	
* Automated multimodal fusion and/or multimodal generation (e.g.,
coordinated speech, gaze, gesture, facial expressions)	
* Techniques for combining objective and subjective evaluations, and
for making evaluations cost-effective, predictive and fast
The output of the workshop will be the following:
* Better knowledge of the potential of major models of human
multimodal behaviour
* Challenging issues in the usability of multimodal corpora	
* Fostering of a pluridisciplinary community of multimodal researchers
and multimodal interface developers
Reasons of interest
Multimodal resources feature the recording and annotation of several
communication modalities such as speech, hand gesture, facial
expression, body posture, graphics Several researchers have been
developing such multimodal resources for several years, often with a
focus on a limited set of modalities or on a given application
domain. A number of projects, initiatives and organisations have
addressed multimodal resources with a federative approach:
* At LREC2002, a workshop had addressed the issue of Multimodal
Resources and Multimodal Systems Evaluation
* At LREC2000, a 1st workshop had addressed the issue of multimodal
corpora, focussing on meta-descriptions and large corpora
* The European 6th Framework program (FP6), started in 2003, includes
multilingual and multisensorial communication as one of the major R&D
issue, and the evaluation of technologies appears as a specific item
in the Integrated Project instrument presentation	
* NIMM was a work group on Natural Interaction and MultiModality which
ran under the IST-ISLE project ( In 2001,
NIMM compiled a survey of existing multimodal resources (more than 60
corpora are described in the survey), coding schemes and annotation
tools. The ISLE project was developed both in Europe and in the USA
* EcorporaA (European Language Resources Association) launched in
November 2001 a survey about multimodal corpora including marketing
aspects (	
* A Working Group at the Dagstuhl Seminar on Multimodality recorded,
in November 2001, 28 questionnaires from researchers on multimodality,
from which 21 have been announcing their attention to record other
multimodal corpora in the
future. (
* Other surveys have been recently made about multimodal annotation
coding schemes and tools (COCOSDA, LDC, MITRE).		
Yet, existing annotation of multimodal corpora until now have been
made mostly on an individual basis, each researcher or team focusing
on its own needs and knowledge about modality specific coding schemes
or application examples. Thus, there is a lack of real common
knowledge and understanding of how to proceed from annotations to
usable models of human multimodal behaviour and how to use such
knowledge for the design and evaluation of multimodal input and
embodied conversational agent interfaces. Furthermore, the evaluation
of multimodal interaction poses different (and very complex) problems
than the evaluation of monomodal speech interfaces or WYSIWYG direct
interaction interfaces. There is a number of recently finished and
ongoing projects in the field of multimodal interaction in which
attempts have been made to evaluate the quality of the interfaces in
all meanings that can be attached to the term 'quality'. There is a
widely felt need in the field for exchanging information on multimodal
interaction evaluation with researchers in other projects. One of the
major outcomes of this workshop should be better understanding of the
extent to which evaluation procedures developed in one project
generalize to other, somewhat related projects.
Important dates
* 1st December 2003: Call for papers and demonstrations
* 24 January 2004: Deadline for paper submission
* 29 February 2004: Acceptance notifications and preliminary program
* 21 March 2004: Deadline final version of accepted papers
* 25 May 2004: Workshop
For more details, especially concerning the submission process and
details for the Program Commitee and Organizers, please visit the
workshop website
Mail to author|Respond to list|Read more issues|LINGUIST home page|Top of issue