September 21, 2015 Xian (China) A Research Platform for - - PowerPoint PPT Presentation

september 21 2015
SMART_READER_LITE
LIVE PREVIEW

September 21, 2015 Xian (China) A Research Platform for - - PowerPoint PPT Presentation

ACII 2015 Tutorial September 21, 2015 Xian (China) A Research Platform for Synchronised Individual/Group Affective/Social Signal Recording and Analysis M. Mancini, R. Niewiadomski, G. Volpe, A. Camurri Casa Paganini InfoMus Research


slide-1
SLIDE 1

ACII 2015 Tutorial September 21, 2015 Xi’an (China) A Research Platform for Synchronised Individual/Group Affective/Social Signal Recording and Analysis

  • M. Mancini, R. Niewiadomski, G. Volpe, A. Camurri

Casa Paganini – InfoMus Research Centre DIBRIS, University of Genoa, Italy www.casapaganini.org

slide-2
SLIDE 2

Summary

  • The Casa Paganini–InfoMus Research Centre
  • Conceptual Framework
  • Automated analysis of multimodal features of

non-verbal behaviour

– Individual: expressive gesture, emotion – Group: synchronisation, entrainment, leadership

  • The EyesWeb XMI Software Platform

– Non-Verbal Social Signals Software Library

slide-3
SLIDE 3

The monumental building of Santa Maria delle Grazie La Nuova

slide-4
SLIDE 4

Research

  • Cross-fertilization between research in science

and technology and humanistic and artistic research.

  • Art for ICT: Artistic and humanistic theories

as source of inspiration for scientific- technological research.

  • ICT for Art: Research results from science

and technology as a source of inspiration for art languages and artistic projects.

slide-5
SLIDE 5

Real-time analysis of expressive gesture and non verbal social signals FP7 FET SIEMPRE (Socio-mobile) active music listening FP7 ICT SAME

Research

Interactive sonification Sensory substitution H2020 ICT DANCE

slide-6
SLIDE 6

FP7 ICT ASC INCLUSION Therapy and rehabilitation: interactive serious games to support autistic children to learn to recognize and express emotions

FP7 ICT MIROR

Interactive software for music education (FP7 ICT MIROR, H2020 ICT TELMI) serious games, edutainment, active embodied experience of cultural audiovisual content

Research

“Viaggiatori di Sguardo” permanent installation, Palazzo Ducale, Genoa

slide-7
SLIDE 7

Interactive sonification to support chronic pain (with UCL, Nadia Berthouze) Interactive systems for rehabilitation of children (with Gaslini Children Hospital) (Intetain 2015) Rehabilitation exercises for Parkinson disease (ICT CARE HERE, EU CA CAPSIL) Motion Composer (Wechsler et al)

Camurri et al 2003 “Application of multimedia techniques in the physical rehabilitation of Parkinson's patients”, Journal of Visualization and Computer Animation, 14(5)

Research on ICT for Therapy and rehabilitation

slide-8
SLIDE 8

Research grounded on cross-fertilisation of ICT and art

From artistic project… …to S&T research Music Theatre Opera “Outis” Luciano Berio Teatro Alla Scala di Milano (1996) Invisible interfaces for on-stage interaction and synchronisation with audio Music Theatre Opera “Cronaca del Luogo”, Luciano Berio, opening Salzburg Festival (July 1999) Real-time analysis of full-body movement, non-verbal expressive behaviour qualities. The EyesWeb software platform. Music Theatre Opera “Un Avatar del Diavolo”, Roberto Doati, La Biennale Venezia (2005) Tangible acoustic interfaces: give the sense of touch to everyday objects Museum “Enrico Caruso”, permanent interactive installation “Sala della Musica”, Firenze (2011-) Non-verbal behaviour analysis for individual and group interaction with cultural heritage content EU FET11 Closing Performance: TanGO Touching Music” (6 May 2011) Performance built upon scientific results of the European ICT FET SIEMPRE Project. Study of music joint performance: string quartets, orchestra sections with conductor, audience response S&T research in EU ICT FET SIEMPRE Project

slide-9
SLIDE 9

Expressive Gesture

Example: Real-time measure and sonification of the space between the two dancers

Camurri Mazzarino Volpe 2004 “Expressive interfaces”, Cognition Technology & Work Journal

slide-10
SLIDE 10

Automated Analysis of Emotion from Expressive Gesture features

  • A.Camurri, I.Lagerlof, G.Volpe (2003) “Recognizing Emotion from Dance

Movement: Comparison of Spectator Recognition and Automated Techniques” Intl. Journal of Human Computer Studies, 59(1-2).

  • G.Volpe, A.Camurri (2011) “A system for embodied social active listening to sound

and music content” ACM Journal on Computing and Cultural Heritage, 4(1)

“Mappe per Affetti Erranti”, Festival della Scienza 2007

Each dancer embodies a human voice (bass, tenor, contralto, soprano); Each voice sings with the emotion expressed by the body gesture of the corresponding dancer. Example: Hesitant -> Whispering voice. (1min ca.) dancers express different emotions: singing voices incoherent (2:30min ca.) dancers converge to Joy: all singing voices joyful and synchronized

slide-11
SLIDE 11

Automated Analysis of Emotion from Expressive Gesture features

  • A.Camurri, I.Lagerlof, G.Volpe (2003) “Recognizing Emotion from Dance

Movement: Comparison of Spectator Recognition and Automated Techniques” Intl. Journal of Human Computer Studies, 59(1-2).

Luciano Berio Music theatre opera “Cronaca del Luogo” Salzburg Festival 1999 (video)

slide-12
SLIDE 12

12

Expressive Gesture and Music

  • Singing voice

Original A “microscope” on expressive gesture cues (4 times longer) (Rolf Inge Godoy, Oslo University)

slide-13
SLIDE 13

Conceptual framework

physical signals low-level features mid-level features; maps and shapes concepts, structures

Layered model for multimodal expressive gesture

Camurri et al 2005, IEEE Multimedia J

slide-14
SLIDE 14

Conceptual framework

physical signals low-level features mid-level features; maps and shapes concepts, structures

Real-time (ms) Local

Camurri et al 2005, IEEE Multimedia J

0,5 – 3s Predictive models

slide-15
SLIDE 15

Conceptual framework

physical signals low-level features mid-level features; maps and shapes concepts, structures

Joints Positions, Velocities

Camurri et al 2005, IEEE Multimedia J

Fluidity: smoothness of body

joints + “wave-like” coordination

Smoothness one joint

Example

slide-16
SLIDE 16

Conceptual framework

physical signals low-level features mid-level features; maps and shapes concepts, structures physical signals low-level features mid-level features; maps and shapes concepts, structures self

  • ther

Synchronization

(of Low Level and Expressive Features)

slide-17
SLIDE 17

Research inspired by the arts and humanistic theories:

Laban’s Effort Theory, Schaeffer’s Morphology, Gesture in Visual Arts

Example of Mid-Level Features: Laban Theory of Effort

slide-18
SLIDE 18

EU-ICT Project MIROR

  • Embodied and reflexive

applications, to support children in exploring rhythm, melody, and harmony by means of their own body.

  • Interaction paradigm: full-body mimicking a character
  • Mapping of Laban’s movement qualities to elements
  • f the musical structure.

Varni, G., Volpe, G., Sagoleo, R., Mancini, M., and Lepri, G. (2013), Interactive reflexive and embodied exploration of sound qualities with BeSound. In Proc. of the 12th International Conference on Interaction Design and Children, 531-534, 2013.

slide-19
SLIDE 19

EU-ICT Project ASC-INCLUSION

Serious games for teaching autistic children to recognize and express emotions by non-verbal full-body expressive gesture automated analysis of emotions

S.Piana et al 2014 “Real time automated recognition of emotions from body gesture”, IDGEI 2014 S.Piana et al (in Press) “Adaptive body gesture representation for automatic emotion recognition”

Example of High-Level Features: Emotions

slide-20
SLIDE 20

Automated analysis of laughter from full-body movement

EU-ICT-FET Project ILHAIRE

  • F. Pecune, B. Biancardi, Y. Ding, C. Pelachaud, M. Mancini, G. Varni, A. Camurri, G.

Volpe, LOL—Laugh Out Loud, Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015

  • M. Mancini, G. Varni, R. Niewiadomski, G. Volpe, A. Camurri, How is your laugh

today?, in CHI'14 Extended Abstracts on Human Factors in Computing Systems, 1855- 1860, ACM, 2014

slide-21
SLIDE 21

Real-time multimodal analysis of non-verbal affective and social signals in ecological environments

  • Music as ideal test-bed:

– Non-verbal communication of emotion – Involves social interaction between the musicians in an ensemble and with the audience – Common, shared goal (no “cheating”) – Can “speak” at the same time (good to study synchronization)

  • Focus:

– Non verbal social cues: temporal and affective entrainment

[Phillips-Silver and Keller, 2012], leadership.

– Focus on processes/dynamics rather than on states – Predictive models for higher-level cues

slide-22
SLIDE 22
  • Violin duo, classical western music, 2007-2008
  • String quartet, classical western music, 2009-

2013 (EU-FET Project SIEMPRE)

  • Orchestra section, classical western music,

2010-2013 (EU-FET Project SIEMPRE)

  • Duo, Hindustani music, 2014-2015

Case studies

slide-23
SLIDE 23

Violin duo

Dataset: Multimodal recordings performed during Premio Paganini 2006 International Violin Competition

slide-24
SLIDE 24
  • Participants: four violin players - two pairs
  • Material: a canon at unison from Musical

Offering by J.S. Bach

  • Conditions: player asked to act four emotions:

Anger, Joy, Sadness, and Pleasure plus a deadpan condition, with and without visual feedback, repeated three times

  • Recordings:

– 2 b/w video-cameras: 720 x 576, 25 Hz – Height: 5-meters, taking the head of the performers – EyesWeb XMI application for synchronised recordings

Recordings and dataset

slide-25
SLIDE 25

Head motion tracking

Centre of Mass trajectory and speed extracted

slide-26
SLIDE 26
  • 60 video recordings
  • No time alignment of the two performances

(canon): only the common part of the performance was taken into account

  • Each player modeled as a component of a

complex system:

– State vector: (x, y, vx, vy) of head’s CoM

Analysis

slide-27
SLIDE 27
  • “Adjustment of rhythms of oscillating objects

due to their weak interaction” [Pikovsky et al., 2001]

  • More specifically: locking of

phase/frequencies & unlocking of amplitude

  • Idea: PS as low-level social signal toward an

indirect measure of empathy

Phase synchronisation (PS)

slide-28
SLIDE 28

Recurrence and Recurrence Plots

  • Recurrence

fundamental features of many dynamical systems

  • Recurrence Plots [Eckmann, 1987]

time-time visualization of recurrences

  • Recurrence matrix:

Rij =

0 : xi ≈ xj 1 : xi ≈ xj

slide-29
SLIDE 29

Recurrence and Recurrence Plots

time time

Figure from Marwan et al., (2007). Recurrence plots for the analysis of complex systems. Physics Reports, 438, 237-329

slide-30
SLIDE 30

Recurrence Quantification Analysis (RQA)

  • Small-scale patterns:

− single dots − vertical lines − diagonal lines

  • RQA: quantification of small-scale patterns

E.g.., Recurrence Rate (RR) :

slide-31
SLIDE 31

Acoustic AND visual channels: 10 findings of PS Acoustic channel only: 14 findings of PS On average: |CPRa | = 0.57 |CPRav| = 0.38 No significant difference among emotions BUT Indication of a significant role of positive emotions.

Results

Varni, G., Volpe, G., & Camurri, A. (2010). A System for Real-time Multimodal Analysis of Nonverbal Affective Social Interaction in User-Centric Media. IEEE Transactions on Multimedia, 12(6), 576-590

slide-32
SLIDE 32

String Quartet

“Self-Managed Team”: all musicians contribute equally to the task

[Gilboa & Tal-Shmotkin, 2010, 2012] (Picture: Quartetto di Cremona during an experiment at Casa Paganini-InfoMus)

slide-33
SLIDE 33
  • Which multimodal cues explain the difference

between playing alone or in ensemble?

Research questions

slide-34
SLIDE 34
  • Which multimodal cues explain who is the leader?
  • Which multimodal cues explain the difference

between a performance engaging an audience and a simply correct - but «cold» - performance?

Research questions

slide-35
SLIDE 35

The EU ICT FET Project SIEMPRE

  • When people perform a task as part of a joint

action, their behaviour is not the same as it would be if they were alone: it is adapted to facilitate shared understanding (or sometimes to prevent it).

  • Joint performance of music offers a test-bed for

ecologically valid investigations of the way non- verbal behaviour facilitates joint action.

  • SIEMPRE scenarios: music ensembles (quartets),
  • rchestra section with conductor, audience.
slide-36
SLIDE 36

Solo

slide-37
SLIDE 37

Ensemble

Quartetto di Cremona

slide-38
SLIDE 38

Solo or in Ensemble ? (1)

slide-39
SLIDE 39

Solo or in Ensemble ? (2)

slide-40
SLIDE 40

Solo vs. Ensemble: measures

Ear

  • Subjective centre of the quartet: the “ear”.
  • Distance of heads from ear.
  • Sample Entropy as measure of complexity.
slide-41
SLIDE 41

Solo vs. Ensemble: results

(F1,135= 119.984, p < .001)

  • Significant effect of Solo vs. Ensemble condition.

Glowinski, D., Mancini, Cowie, R., Chiorri, C., Doherty, C. and Camurri, A. (2013) The movements made by performers in a skilled quartet: a distinctive pattern, and the function that it serves, Frontiers in Psychology - Auditory Cognitive Neurosciences, 4:841.

slide-42
SLIDE 42

Orchestra of Music Conservatory of Genoa, conductors Pietro Borgonovo and Sera Tokay (Italian Institute

  • f Technology and

Casa Paganini - InfoMus)

Entrainment and leadership in an

  • rchestra with conductor
slide-43
SLIDE 43

Entrainment

  • Affective entrainment: formation of interpersonal
  • bonds. Related to the pleasure in moving the body to

music and being in time with others.

  • Temporal entrainment: automatic movements that
  • ccurs when listening to musicians play (covert

activation of motor areas of the brain), observed in complex rhythmic timing and exchange between partners or ensemble members in music or dance. (Phillips-Silver & Keller 2012)

slide-44
SLIDE 44

Soft Entrainment (Yoshida 2002)

  • Soft-entrainment:

alternation of high and low entrainment (Yoshida, 2002).

  • Boerner’s model of orchestra conduction, focusing
  • n entrainment within and between sections.
slide-45
SLIDE 45

Orchestra: measures

  • Recordings of Motion

Capture and audio of two string sections and

  • conductor. Dataset from

(D’Ausilio et al., 2012).

  • Focus on the z-

component of bow trajectory of string sections (taken as predominant direction

  • f bow movement).
slide-46
SLIDE 46

Orchestra: some results

  • Effectiveness of Leadership allowed ranking

the performances of the two conductors in the dataset along ten different takes.

  • The ranking confirms a previous study by

(D’Ausilio et al PlosOne 2012) on the same dataset using another technique (Granger Causality).

Varni, G., Volpe, G., Camurri, A., Tokay, S., Fadiga, L., “Estimating effectiveness of orchestra conduction through motion entrainment” (submitted)

slide-47
SLIDE 47

Does synchrony follow inter-cultural shared patterns in music performances ? Does the “soft entrainment” model apply also to non Western music culture?

P.Alborno, M.Clayton, P.Keller, G.Volpe, A.Camurri (2015) "Automated Video Analysis of Interpersonal Entrainment in Indian Music Performance", INTETAIN 2015

slide-48
SLIDE 48

Hindustani performances

In collaboration with M. Clayton (Durham University) and P. Keller (University of Western Sydney)

slide-49
SLIDE 49
  • Objective: investigate whether synchrony

follows intercultural shared patterns

  • Dataset: videos from different music

performances from different cultures. Preliminary results from a considerable archive of digital audiovisual recordings of Hindustani performances, collected at Durham University

Hindustani performances

slide-50
SLIDE 50
  • Performance: analysis of a duo

– Murad Ali (MA): highly-regarded sarangi (bowed lute) – Gurdain Rayatt (GR): tabla accompanistis

  • Synchrony in Hindustani performances:

– A soloist will either return to a refrain just before beat

  • ne (sam) in the tala, and/or will conclude an

improvisation at this point – In between these points it may be regulated by a soloist’s hand gestures (used to direct the drummer to adjust the tempo), or affirmed by gestures such as synchronised head nods on the sam (beat one)

Hindustani performances

slide-51
SLIDE 51

System architecture (based on EyesWeb)

slide-52
SLIDE 52
  • Quantity of motion:

– amount of detected movement (integration of pixel-wise differences between consecutive frames) [Mazzarino et al.,

2003]

  • Y coordinate of head’s CoM:

– head’s barycentre coordinate on the vertical axis. It was extracted to analyse joint head movements (nods) that typically characterise the end of musical (tala) cycles

  • Head X displacement:

– the overall translation and rotation components of the head movements, computed from optical flow. Used as approximation of gaze direction

Extracted movement features

slide-53
SLIDE 53
  • The final part of each cycle is the one

displaying the highest synchronisation between the two musicians, confirming Yoshida’s findings on soft-entrainment

  • Future work: more sophisticated analysis
  • f synchronisation of relevant events,

exploiting Event Synchronisation (and

  • ur Multi-class and Multi-set extensions)

Preliminary results

Alborno, P., Keller, P., Clayton, M., Volpe, G., & Camurri, A. (2015). Automated video analysis of interpersonal entrainment in Indian music

  • performance. In Proc. 7th Intl. ICST Conf. on Intelligent Technologies for

Interactive Entertainment (Intetain 2015)

slide-54
SLIDE 54
  • Open software platform supporting fine-grain

synchronised recordings of multimodal (audio, video, MoCap, biometric) data, performing pre-processing and analysis of multimodal signals in real-time

  • Modular, flexible and adaptable
  • Widely employed for developing real-time dance,

music, and multimedia apps. Adopted by universities, industry, artists, cultural institutions

  • Adopted in many EU projects
  • Windows and Android (mobiles)
  • Supports wide range of sensing and actuating devices
  • Free download

Technological outputs: EyesWeb

slide-55
SLIDE 55

Technological outputs

Multiple audio sources Multiple video sources Biometric sensors Motion Capture Low-cost sensors (e.g., Kinect2, Leap) Sensors on Android mobiles Our sensor based on Ethernet HD camera (50fps) + Asus Xtion (MotionComposer) The SIEMPRE Platform for synchronised multimodal recordings

slide-56
SLIDE 56

Recordings can be previewed immediately, without any data processing or data conversion

Technological outputs

slide-57
SLIDE 57
  • Libraries for real-time analysis of body movement: motion features (e.g.,

kinematics, amount of movement, impulsivity, directness, fluidity, and so

  • n) can be computed, stored on file, and viewed in real-time or off-line

Technological outputs

slide-58
SLIDE 58
  • EyesWeb Library for Real-Time Social Signals Processing:

Synchronisation (Recurrence Quantification Analysis, Event Synchronisation), Leadership (e.g., chronemic leadership, analysis based on Graph Theory), …

Technological outputs

slide-59
SLIDE 59

Technological outputs: apps

Systems for: social active experience of music, interactive dance, experience of cultural heritage, rehabilitation, education

The orchestra explorer Mappe per affetti erranti I-DJ Sync’n’move

slide-60
SLIDE 60

Understanding behaviour of groups of users: applications

  • Active experience of music (and audiovisual content):

users are enabled to interactively operate on content by creating, distributing, retrieving, modifying, and molding it in real-time.

slide-61
SLIDE 61

Applications: iDJ

Varni, G., Volpe, G., Mazzarino B. (2011), Towards a Social Retrieval of Music Content, in Proc. of the Third IEEE Intl. Conference on Social Computing (SocialCom2011), Workshop on Social Behavioral Analysis and Behavioral Change, Cambridge (MA), USA, October 2011.

iDJ: A concept of application for embodied cooperation as a paradigm for formulating social queries. Prototype presented at EU ICT 2013 Exhibition, Vilnius, Podium performance.

slide-62
SLIDE 62

Applications: iDJ

  • Input: full-body motion features

e.g., Motion Index, Contraction Index, Fluidity Index,…

  • Analysis of synchronisation

i.e., whether the users are dancing in a tight-knit way

  • Analysis of dominance

i.e., identification of the dominant user

slide-63
SLIDE 63

Applications: Sync4All

G.Varni, M.Mancini, G.Volpe, A.Camurri (2011) A system for mobile active music listening based on social interaction and embodiment. ACM Mobile Networks and Applications Intl Journal (ACM MONET), Vol.16, No.3, pp.375-384.

Each user rhythmically and freely moves her mobile phone trying to synchronise with the other users. Synchronisation is measured as Phase Synchronisation of gestures (RQA, Cross- Recurrence Plots).

Collaboration among users enables active listening to a music piece.

slide-64
SLIDE 64

Applications: Sync4All

slide-65
SLIDE 65
  • Main objective: to investigate techniques for

sensory substitution in blind people, to enable perception of and participation in non-verbal, artistic whole-body experiences

  • Current work: sonification of motion features

extracted from sample dance performances

DANCE EU-H2020-ICT project, 2015-2017

Partners: Univ. of Genoa, Maastricht Univ. (B. de Gelder) and KTH, Stockholm (R. Bresin)

slide-66
SLIDE 66

DANCE –EU H2020 ICT Project

  • “To see, close your eyes”
  • Investigating how affective and relational qualities of

body movement can be expressed, represented, and analyzed by the auditory channel.

  • Objectives: to understand the meaning of “closing

the eyes”,

– the perception of expressiveness and entrainment in dance, – the participation to the emotion conveyed by a sequence of movements in space, – the understanding of the non-verbal language of bodies that communicate, – imagining and questioning concrete ways to listen to a choreography, feel a ballet.

slide-67
SLIDE 67

DANCE – Dancing in the Dark

Collaboration with choreographers and dance groups

  • To create archives of multimodal data on individual and group

movement qualities

  • To support and experiment project results in artistic

productions

slide-68
SLIDE 68
  • Karate: two «kata» performed by experts and

students (all «black belt»)

  • Hypothesis: measurable differences in intra-

personal synchronisation of limbs can predict

  • bservers’ ratings of performance quality
  • Event Synchronisation as measure of intra-

personal synchronisation

On-going projects: martial arts

slide-69
SLIDE 69

Ongoing project on martial arts

  • Karate: two «kata» performed by

experts and students (all «black belt»)

  • Hypothesis: measurable differences in

intra-personal synchronization of body parts.

  • Use Event Synchronization to

measure intra-personal synchronization

slide-70
SLIDE 70

Future research directions

  • Aesthetic appreciation is one of the most

intangible aspects of higher cognition

  • Exploring the rules governing aesthetic

experience has a great potential for future ICT:

  • Measuring aesthetic experience in new

applications: augmented reality, mobiles, embodied social media

slide-71
SLIDE 71
  • Videos: www.youtube.com/InfoMusLab
  • Papers: ftp.infomus.org/pub/Staff/AntonioCamurri
  • www.casapaganini.org
  • ricerca@unige.it
  • Research team

Antonio Camurri, Gualtiero Volpe, Corrado Canepa, Paolo Coletta, Nicola Ferrari, Simone Ghisio, Maurizio Mancini, Stefano Piana, Alberto Massari, Radoslaw Niewiadomski, Paolo Alborno, Ksenia Kolykhalova, Damiano Malafronte

More info and contacts