“Multimodal Analyses enabling Artificial Agents in Human-Machine Interaction”                               

Sponsored by:http://www.sfb-trr-62.de/

Paper Version

The paper version of the Call for Paper is available as PDF.

Submissions Instruction and Publication

Prospective authors are invited to submit full papers (8 pages) and short papers (5 pages) in ACM format as specified by ICMI 2016.

All submissions must be anonymous.

All submissions will be refereed by experts in the field based on originality, significance, quality and clarity. Every submitted paper will be reviewed by at least two members of the Program Committee.

Accepted papers will be published as post-proceedings in the ACM Digital Library.

Call for Paper

One of the aims in building multimodal user interfaces and combining them with technical devices is to make the interaction between user and system as natural as possible. The most natural form of interaction may be how we interact with other humans. Current technology is far from human-like, and systems can reflect a wide range of technical solutions.

Transferring the insights for analysis of human-human communication to human-machine interactions remains challenging. It requires that the multimodal inputs from the user (e.g., speech, gaze, facial expressions) are recorded and interpreted. This interpretation has to occur at both the semantic and affective levels, including aspects such as the personality, mood, or intentions of the user. These processes have to be performed in real-time in order for the system to respond without delays ensuring that the interaction is smooth.

The MA3HMI workshop aims at bringing together researchers working on the analysis of multimodal data as a means to develop technical devices that can interact with humans. In particular, artificial agents can be regarded in their broadest sense, including virtual chat agents, empathic speech interfaces and life-style coaches on a smart-phone. More general, multimodal analyses support any technical system in the research area of human-machine interaction. We focus on the real-time aspects of human-machine interaction. We address the development and evaluation of multimodal, real-time systems.

We solicit papers that concern the different phases of the development of such interfaces. Tools and systems that address real-time conversations with artificial agents and technical systems are also within the scope of the workshop.

Workshop topics include, but are not limited to:

(a)   Multimodal annotation

-       Representation formats for merged annotations of different modalities

-       Best practices for multimodal annotation procedures

-       Innovative multimodal annotation schemas or re-adaptation

-       Annotation and processing of multimodal data sets including proper feature extraction

-   Real-time or on-the-fly annotation approaches

(b)  Multimodal analyses

-       Multimodal understanding on the user’s input

-       Dialogue management using multimodal output

-       Evaluation and benchmarking of humanmachine conversations

-       Novel strategies of human-machine interactions

-       Using multimodal data sets for human-machine interaction

(c ) Applications, tools and systems

-       Novel application domains and embodied interaction

-       Prototype development and uptake of technology

-       User studies with (partial) functional systems

-       Tools for the recording, annotation and analysis of conversations

Satellite workshop of ICMI 2016

MA³HMI 2016

copyright: Morio