Seminar CHIST-ERA Istanbul : 4 March 2014 Kick-off meeting : 27 January 2014 (call IUI 2012)

Similar documents
Laugh when you re winning

Smile and Laughter in Human-Machine Interaction: a study of engagement

Multimodal Data Collection of Human-Robot Humorous Interactions in the JOKER Project

Analysis of Engagement and User Experience with a Laughter Responsive Social Robot

LAUGHTER IN SOCIAL ROBOTICS WITH HUMANOIDS AND ANDROIDS

MAKING INTERACTIVE GUIDES MORE ATTRACTIVE

Meaning Machines CS 672 Deictic Representations (3) Matthew Stone THE VILLAGE

Multimodal databases at KTH

Embodied music cognition and mediation technology

Laughter Yoga International

WOZ Acoustic Data Collection For Interactive TV

Laughter and Topic Transition in Multiparty Conversation

Laughter and Smile Processing for Human-Computer Interactions

Music Performance Panel: NICI / MMM Position Statement

Speech Recognition and Signal Processing for Broadcast News Transcription

The roles of expertise and partnership in collaborative rehearsal

LAUGHTER serves as an expressive social signal in human

How about laughter? Perceived naturalness of two laughing humanoid robots

Application of a Musical-based Interaction System to the Waseda Flutist Robot WF-4RIV: Development Results and Performance Experiments

MIMes and MeRMAids: On the possibility of computeraided interpretation

Audiovisual analysis of relations between laughter types and laughter motions

Concept of ELFi Educational program. Android + LEGO

Improving Frame Based Automatic Laughter Detection

Humor: Prosody Analysis and Automatic Recognition for F * R * I * E * N * D * S *

Follow the Beat? Understanding Conducting Gestures from Video

Retrieval of textual song lyrics from sung inputs

Digital Text, Meaning and the World

Exhibits. Open House. NHK STRL Open House Entrance. Smart Production. Open House 2018 Exhibits

Expressive Multimodal Conversational Acts for SAIBA agents

Embodied Agents: A New Impetus to Humor Research

Computer Coordination With Popular Music: A New Research Agenda 1

CLARIN - NL. Language Resources and Technology Infrastructure for the Humanities in the Netherlands. Jan Odijk NO-CLARIN Meeting Oslo 18 June 2010

CHAPTER I INTRODUCTION. humorous condition. Sometimes visual and audio effect can cause people to laugh

CHAPTER I INTRODUCTION. Jocular register must have its characteristics and differences from other forms

Laughter and Body Movements as Communicative Actions in Interactions

The Lost Art of Listening. How to Remember Names

ITU-T Y Functional framework and capabilities of the Internet of things

Humor and Embodied Conversational Agents

Sound Design The Expressive Power Of Music Voice And Effects In Cinema David Sonnenschein

Snickers. Study goals:

Laugh-aware Virtual Agent and its Impact on User Amusement

D7.2: Human Machine Interface

KEYWORDS Participation, Social media, Interaction, Community

Specific Learner Expectations. Developing Practical Knowledge

Proposal for Application of Speech Techniques to Music Analysis

Acoustic Prosodic Features In Sarcastic Utterances

10 Steps To Effective Listening

How Laughter Yoga Can Improve. Efficiency and Performance in Your Company

Humorist Bot: Bringing Computational Humour in a Chat-Bot System

About Giovanni De Poli. What is Model. Introduction. di Poli: Methodologies for Expressive Modeling of/for Music Performance

An ecological approach to multimodal subjective music similarity perception

Abstracts workshops RaAM 2015 seminar, June, Leiden

Exploring Choreographers Conceptions of Motion Capture for Full Body Interaction

Automatic music transcription

CLARIN AAI Vision. Daan Broeder Max-Planck Institute for Psycholinguistics. DFN meeting June 7 th Berlin

Three Minute Review. validity of IQ tests some predictive use, but not great. Other types of intelligence

Linguistics 104 Language and conceptualization

Is composition a mode of performing? Questioning musical meaning

Empirical Evaluation of Animated Agents In a Multi-Modal E-Retail Application

Make Me Laugh: Recommending Humoristic Content on the WWW

Laughter Animation Synthesis

The Barco Magic-Y system

3 rd International Conference on Smart and Sustainable Technologies SpliTech2018 June 26-29, 2018

Rhythm and Melody Aspects of Language and Music

Formalizing Irony with Doxastic Logic

Automatic Laughter Detection

Expressive performance in music: Mapping acoustic cues onto facial expressions

The AV-LASYN Database : A synchronous corpus of audio and 3D facial marker data for audio-visual laughter synthesis

Do you have a mature IoT solution? Join us with the Open Call. Alicia Cano - Medtronic.

Metonymy and Metaphor in Cross-media Semantic Interplay

Shimon: An Interactive Improvisational Robotic Marimba Player

Research & Development. White Paper WHP 318. Live subtitles re-timing. proof of concept BRITISH BROADCASTING CORPORATION.

SMARTING SMART, RELIABLE, SIMPLE

ENGINEERING COMMITTEE Energy Management Subcommittee SCTE STANDARD SCTE

Laurent Romary. To cite this version: HAL Id: hal

Beyond the Cybernetic Jam Fantasy: The Continuator

Verbal De-Escalation Jimmy Dixon Clemson City Police, Chief

ANSI/SCTE

How to strengthen the Social Capital of your library - Case Study of Kallio Library

Interactions between Semiotic Modes in Multimodal Texts. Martin Siefkes, University of Bremen

Final Report. PIBot(Pill Informer robot) EEL 5666: Intelligent Machines Design Laboratory Student Name: Duckki Lee

Appendix C ACCESSIBILITY SCALE CLOSED OPEN

Welcome Back To The Acting Career Quick Start: Home Study Course! Hands-on and practical! Module 4.4

Detecting Attempts at Humor in Multiparty Meetings

River Dell Regional School District. Visual and Performing Arts Curriculum Music

In total 2 project plans are submitted. Deadline for Plan 1 is on at 23:59. The plan must contain the following information:

The comparison of actual system with expected system is done with the help of control mechanism. False True

APPLICATIONS OF A SEMI-AUTOMATIC MELODY EXTRACTION INTERFACE FOR INDIAN MUSIC

AAL , ALIAS D3.8 Final Dialogue System

Type-2 Fuzzy Logic Sensor Fusion for Fire Detection Robots

SCTE OPERATIONAL PRACTICE

The role of texture and musicians interpretation in understanding atonal music: Two behavioral studies

Teamwork Makes the Dream Work

Summarizing Long First-Person Videos

Panasonic Highlights 100th Anniversary, Future Vision at CES 2018

Appendix D CONGRUENCE /INCONGRUENCE SCALE. Body and face give opposite message to underlying affect and content

with Janet Elizabeth Henderson

ASSOCIATED SPEECH & LANGUAGE SPECIALISTS, LLC IMPROVING communication. Transforming LIVES.

what are you laughing at? by Tio

Effective Public Speaking:

Transcription:

project JOKER JOKe and Empathy of a Robot/ECA: Towards social and affective relations with a robot Seminar CHIST-ERA Istanbul : 4 March 2014 Kick-off meeting : 27 January 2014 (call IUI 2012) http://www.chistera.eu/projects/joker Laurence Devillers (devil@limsi.fr LIMSI/SLP) 1

Partners background l utilisateur JOKER (2/3) affective and social dimensions in spoken interaction, emotion and affect bursts detection, user models, IHRobot, dialogue, generation (Laurence Devillers, Sophie Rosset) social interaction, multimodal interaction, collection of data, affect bursts detection and generation (Nick Campbell) user detection using visual cues and dialog, visual interpretation (eyes tracking, face, gesture), affect bursts detection, temporal model of gesture, gaze and speech (Metin Sezgin) speech recognition (Kaldi/Sphinx), models of humor (Yannick Estève, Daniel Luzzati) speech synthesis, affect bursts detection and generation (laugh, breath, sigh, throat, etc.) (Stéphane Dupont)

Objectives JOKER JOKe and Empathy of a Robot/ECA create a generic intelligent user interface providing a multimodal dialogue system with social communication skills including humor, empathy, compassion and other informal socially-oriented behavior. fuse the verbal and non verbal cues (audio, eye-gaze, gestures) including affect bursts for social and emotional processes both in perception and generation build rich user profiles taking into account user s personality, interactional behavior explore advanced dialogues involving complex social behaviors in order to create a long-term social relationship react in real-time 3

Main challenges JOKER Social interactions require social intelligence and understanding for dealing with news circumstances by anticipating the mental state of another person. JOKER will investigate humor in human-machine interaction. Humor can trigger surprise, amusement, or irritation if it doesn't match the user's expectations. JOKER will explore two social behaviors: expressing empathy and exchanging chat with the interlocutor as a way to build a deeper relationship. Implementing empathy or humor in a companion-machine requires that the emotional expression and intention of the user should be detected, that the context should be understood, that the system should have a memory, and that the system is able to express an emotional/expressive response comprehensible by the user. 4

Use-case JOKER Application prototype in a laboratory cafeteria with regular participants (students, staff, visitors...) Social interactions in cafeteria beside coffee machine both in Ireland and France (2 languages) with different devices (robot or ECA) 2 situations will be studied - one-on-one : human-robot/eca - robot/eca with multiple people We will build specific scenarii for engaging people in a conversation with the robot Our results/platforms will be useful for designing robot for other applications such as for elderly people 5

Robot/ECAs JOKER 6

WPs JOKER JOKER will react in real-time with a robust perception module (WP3) (sensing user's facial expressions, gaze, voice, audio and speech style and content), a social interaction module modelling user and context, with longterm memories (WP4), and a generation and synthesis module for maintaining social engagement with the user (WP5). The research will provide a collection of multimodal data with different socially-oriented behavior scenarios in two languages (French and English) (WP2) and an evaluation protocol for such systems.(wp6) 7

Partners and WPs JOKER 8 8 Start : January 2014 WP1 (LIMSI): Management WP2 (TCD): Domain and Databases of interactive speech WP3 (KOC) : Perception modules WP4 (LIMSI) : Dialogue and decision modules WP5 (UMONS): Generation and synthesis modules WP6 (LIUM): Evaluation

WP 1 JOKER LIMSI, LIUM, TCD, UMONS, KOC D1.1 M1 Kick-off meeting (done) D1.2 M3 Web-site (in progress) D1.3 M8 Consortium agreement (first version sent) D1.4 M12 Annual report D1.5 M18 Intermediate report D1.6 Annual workshop D1.7 M42 Final public workshop D1.8 M42 Final report 9

Deliverable Month of delivery Title of deliverable WP 2 JOKER TCD, LIMSI, LIUM D2.1 M6 Domain definition and scenarios (v1 and v2) (in progress collective work) D2.2 M12 Data collection tool with real system (v1, v2 and v3) (in progress) first tests D2.3 M12 Data collection (dataset 1, 2, 3) - D2.4 M15 M27 - Data annotation (dataset 1, 2, 3) (annotation protocol will be a collective work) M39 D2.5 M42 Study of cultural aspects of social interaction Main idea : use as soon as possible an automatic system instead of a WoZ for the data collection > bootstrapping procedure 10

WP 2 JOKER TCD, LIMSI, LIUM TCD will lend expertise in domain specification and initial data collection for training and building the conversational dialogue system (Herme project). Example of LIMSI background work (experiences with WoZ, ex: elderly people) 11

WP 3 JOKER KOC, LIMSI, LIUM, TCD, UMONS Deliverable Month of delivery D3.1 M10- M22- M34 D3.2 M10- M22- M34 D3.3 M10- M22- M34 D3.4 M10- M22- M32 D3.5 M10- M22- M34 D3.6 M12- Title of deliverable User detection using visual cues - Real-time emotion and social behaviors detection using visual cues - deliverable software (v1, v2 and v3) Real-time emotion and social behaviors detection using audio cues - deliverable software (v1, v2 and v3) Real-time emotion and social behaviors detection using affect bursts - deliverable software (v1, v2 and v3) Automatic speech recognition - deliverable software (v1, v2 and v3) Names entities, Topics detection - deliverable software (v1, v2 and v3) Integration and Fusion of linguistic and multi-modal cues for emotion and social behavior detection - deliverable software (v1, v2 and v3) 12

WP 3 JOKER KOC, LIMSI, LIUM, TCD, UMONS Preliminary Work Facial Tracking (KOC) Preliminary Work Emotion detection from speech with NAO (LIMSI) LivingWithRobot 2012: http://www.youtube.com/watch?v=p1id-gvunws 13

Preliminary Work Facial Tracking (KOC) Attempts of making our own dataset using Kinect to learn about affect bursts and emotions since Kinect is ideal device to capture multimodal data (facial tracks, body gesture and sound) Non rigid facial tracking using Jason Saaragih s implementation*. Figure(1): Sample outputs of the facial tracker * J. Saragih, S. Lucey and J. Cohn, ''Deformable Model Fitting by Regularized Landmark Mean-Shifts", International Journal of Computer Vision (IJCV), 2010 14

WP 4 JOKER LIMSI, LIUM, TCD Deliverable Month of delivery Title of deliverable D4.1 M3 - M6 Semantic representation (deliverable) - Dialog platform (software) D4.2 M12- D4.3 M12- D4.4 M12- D4.5 M12- Dynamic emotional profile of the user (software and deliverable) - (v1, v2 and v3) Ontology, history of the dialog, anticipation and memorization modules - blackboard (software and deliverable) (v1, v2 and v3) Intuitive decision way : Dialog strategies using synchrony, mimics (software and deliverable) (v1, v2 and v3) Cognitive decision way (software and deliverable) (v1, v2 and v3) 15

WP 5 JOKER UMONS, LIMSI, TCD Deliverable Month of delivery Title of deliverable D5.1 M12- D5.2 M12- D5.3 M12- D5.4 M12- Generation (v1, v2 and v3) Speech Synthesis (v1, v2 and v3) Affect bursts generation (v1, v2 and v3) Multimodal generation and synthesis (v1, v2 and v3) 16

WP 5 JOKER UMONS, LIMSI, TCD Example of background work laughter synthesis: From laughter intensity curves To laughter phonetic transcriptions And laughter audio and audiovisual generation 17

WP 6 JOKER LIUM, UMONS, LIMSI, TCD, KOC Deliverable Month of delivery Title of deliverable D6.1 M12- D6.2 M15 M27 - M39 Protocol and Metrics (engagement measures : verbal and non verbal such as laugh, smile, eye-tracking, interaction duration) Individual components evaluation (3 times evaluation during the project) D6.3 M39 Final Evaluation One use-case in cafeteria D6.4 M42 Impact of the companion (Robots, ECAs) D6.5 M42 Dissemination (final workshop) 18

Work in progress JOKER Our JOKER system will be tested in two different langages with at least two different plateforms (Robot and ECA). first version in M12 Main results: -Collaboration between complementary European teams on perception, dialogue and generation modules for HRI, -Multimodal original corpora available for the community, - Longitudinal experiment of people engagement with a social robot - Impact of the humor in the social interaction with a machine. and a study of laugh and humor across languages in French and in English. 19

Thanks for your attention JOKER team 20