[iva] CFP: LREC 2016 Workshop on Multimodal Corpora:

Jens Edlund edlund at speech.kth.se
Tue Dec 22 12:31:38 CET 2015


*** 1st Call for Papers ***

LREC 2016 Workshop

Multimodal Corpora:
Computer vision and language processing

*** 24 May 2016, Grand Hotel Bernardin Conference Center, Portorož,
Slovenia ***

http://www.multimodal-corpora.org/

The creation of a multimodal corpus involves the recording, annotation and
analysis of several communication modalities such as speech, hand gesture,
facial expression, body posture, gaze, etc. An increasing number of
research areas have transgressed or are in the process of transgressing
from focused single modality research to full-fledged multimodality
research, and multimodal corpora are becoming a core research asset and an
opportunity for interdisciplinary exchange of ideas, concepts and data.

We are pleased to announce that in 2016, the 11th Workshop on Multimodal
Corpora will once again be collocated with LREC.

Theme

As always, we aim for a wide cross-section of the field, with contributions
ranging from collection efforts, coding, validation, and analysis methods
to tools and applications of multimodal corpora. Success stories of corpora
that have provided insights into both applied and basic research are
welcome, as are presentations of design discussions, methods and tools.
This year, we would also like to pay special attention to the integration
of computer vision and language processing techniques – a combination that
is becoming increasingly important as the accessible video and speech data
increases. The special theme for this instalment of Multimodal Corpora is
how processing techniques for vision and language can be combined to
manage, search, and process digital content.

This workshop follows previous events held at LREC 00, 02, 04, 06, 08, 10,
ICMI 11, LREC 2012, IVA 2013, and LREC 2014. All workshops are documented
under www.multimodal-corpora.org and complemented by a special issue of the
Journal of Language Resources and Evaluation which came out in 2008, a
state-of-the-art book published by Springer in 2009 and a special issue of
the Journal of Multimodal User Interfaces under publication. This year,
some of the contributors, along with selected contributors from 2009 up
until now, will be invited to submit an extended version for a new
publication gathering recent research in the area.

The LREC'2016 workshop on multimodal corpora will feature a special session
on the combination of processing techniques for vision and language.

Other topics to be addressed include, but are not limited to:

• Multimodal corpus collection activities (e.g. direction-giving dialogues,
emotional behaviour, human-avatar and human-robot interaction, etc.) and
descriptions of existing multimodal resources
• Relations between modalities in human-human interaction and in
human-computer interaction
• Multimodal interaction in specific scenarios, e.g. group interaction in
meetings or games
• Coding schemes for the annotation of multimodal corpora
• Evaluation and validation of multimodal annotations
• Methods, tools, and best practices for the acquisition, creation,
management, access, distribution, and use of multimedia and multimodal
corpora
• Interoperability between multimodal annotation tools (exchange formats,
conversion tools, standardization)
• Collaborative coding
• Metadata descriptions of multimodal corpora
• Automatic annotation, based e.g. on motion capture or image processing,
and its integration with manual annotations
• Corpus-based design of multimodal and multimedia systems, in particular
systems that involve human-like modalities either in input (Virtual
Reality, motion capture, etc.) and output (virtual characters)
• Automated multimodal fusion and/or generation (e.g., coordinated speech,
gaze, gesture, facial expressions)
• Machine learning applied to multimodal data
• Multimodal dialogue modelling

Important dates

• Deadline for paper submission (complete paper): 12 February
• Notification of acceptance: 20 March
• Final version of accepted paper: 1 April
• Final program and proceedings: 2 May
• Workshop: 24 May

Submissions

The workshop will consist primarily of paper presentations and
discussion/working sessions. Submissions should be 4 pages long, must be in
English, and follow the LREC’s submission guidelines.
Demonstrations of multimodal corpora and related tools are encouraged as
well (a demonstration outline of 2 pages can be submitted).

Submissions are made at https://www.softconf.com/lrec2016/MMC2016/.

Time schedule and registration fee
The workshop will consist of a morning session and an afternoon session.
There will be time for collective discussions.
Registration and fees are managed by LREC – see the LREC 2016 website (
http://lrec2016.lrec-conf.org/).

Organizing committee

Jens Edlund, KTH Royal Institute of Technology, Sweden
Dirk Heylen, University of Twente, The Netherlands
Patrizia Paggio, University of Copenhagen, Denmark/University of Malta,
Malta
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.uni-bielefeld.de/mailman2/unibi/public/iva-list/attachments/20151222/33422ce4/attachment.html>


More information about the iva-list mailing list