Two-dimensional gesture mapping for interactive real-time - - PowerPoint PPT Presentation

two dimensional gesture mapping for interactive real time
SMART_READER_LITE
LIVE PREVIEW

Two-dimensional gesture mapping for interactive real-time - - PowerPoint PPT Presentation

Two-dimensional gesture mapping for interactive real-time electronic music instruments Thomas Grill gr@grrrr.org University for Music and Performing Arts, Vienna University of Applied Arts, Vienna Austrian Research Institute for Artificial


slide-1
SLIDE 1

Two-dimensional gesture mapping for interactive real-time electronic music instruments

University for Music and Performing Arts, Vienna University of Applied Arts, Vienna Austrian Research Institute for Artificial Intelligence (OFAI)

Thomas Grill gr@grrrr.org

slide-2
SLIDE 2

„You can play on a shoestring if you are sincere“ (John Coltrane)

Topolò, 2006

slide-3
SLIDE 3

„You can play on a shoestring if you are sincere“ (John Coltrane)

  • Interaction with „natural“ objects

(also non-sounding)

  • Adequate reaction upon complex gestures
  • Superimposed „source-binding“ (Smalley)

➡Flexible sensorics (sound, touch, etc.) ➡Flexible sound synthesis

Wishes

slide-4
SLIDE 4

Acoustic instruments

✓Intuitive gesture-to-sound transformation ✓Efficient and expressive ✗ Sound limited due to material, construction

slide-5
SLIDE 5

Software-based electronic instruments

✗ Abstract technology-biased user interfaces ✗ Coarse gesture input, limited expressiveness ✓Unlimited spectrum of sounds and structures

slide-6
SLIDE 6

Challenges

  • Gestures must be mapped to sound synthesis
  • Effective human input „devices“ have

many interrelated degrees of freedom

? How can we make such a mapping flexible/general ? How can this process be comprehensible

for a player

slide-7
SLIDE 7

Proposition

  • Gestural space may not be dense in all dimensions

➡Given a proper representation we can try to use

dimensionality reduction

➡We can do the same for the sound repertory ★By representing gesture and sound repertories in

two dimensions one can easily control and visualize the mapping process (on screen).

slide-8
SLIDE 8

real-time trajectory within a two-dimensional map representation

  • f the gesture repertory

translation/transformation from gesture to synthesis repertory acoustic/sensoric gestures feature extraction sound output sound synthesis real-time trajectory within a two-dimensional map representation

  • f the synthesis repertory
  • Some wishful thinking...
slide-9
SLIDE 9

How?

  • Find sensible feature set for gesture analysis
  • Find sensible feature set for sound representation
  • Use dimensionality reduction to represent

gesture and sound repertories as 2D maps

  • Use 2D transformation strategies to

define the actual gesture-to-sound mapping

slide-10
SLIDE 10

Gesture analysis

  • Gestures as analog (audio) signals

e.g. picked up by contact microphones

  • Feature set describing spectral content of signal
slide-11
SLIDE 11

Timbre analysis

  • Perceptual model of human hearing

(frequency scale [mel, bark], loudness [sone])

  • Model of spectral envelope (e.g. using DCT)

➡MFCCs (mel frequency cepstrum coefficients)

slide-12
SLIDE 12

Dimensionality reduction

  • Sequences of MFCC vectors (typ. 8-10 numbers)

represent gestures and sounds.

  • Frame length ≈ 1 ms

➡ many frames make up repertories ➡ parts of those may be similar

  • Use self-organizing feature maps (SOMs) to spread
  • ut the variety of features on a 2D map.
slide-13
SLIDE 13

Voice

90x90 features

slide-14
SLIDE 14

Voice

90x90 features

slide-15
SLIDE 15

find matching feature set in 2D map of gesture repertory gesture transformation based on 2D coordinates gestures on input device (recorded by sensors) feature extraction sound sound synthesis volume adjustment look up synthesis parameters in 2D map of sound repertory

ó

Anticipated architecture

slide-16
SLIDE 16

2D transformation

  • Define Gaussian transformation regions to

translate from the gesture to the sound map

  • Regions can be translated, scaled, rotated

A A B B C C

slide-17
SLIDE 17

2D transformation

  • Define Gaussian transformation regions to

translate from the gesture to the sound map

  • Regions can be translated, scaled, rotated

A A B B C C

slide-18
SLIDE 18

2D transformation

  • Define Gaussian transformation regions to

translate from the gesture to the sound map

  • Regions can be translated, scaled, rotated
slide-19
SLIDE 19

2D transformation

  • Define Gaussian transformation regions to

translate from the gesture to the sound map

slide-20
SLIDE 20

2D transformation

  • Define Gaussian transformation regions to

translate from the gesture to the sound map

  • A point can be transformed by multiple regions

A A B B C C

slide-21
SLIDE 21

Sound repertory

  • The trajectory in gesture space has been

transformed to a trajectory in sound space.

  • Along this trajectory sounds are looked up

in the repertory map in real-time.

slide-22
SLIDE 22

Sound repertory

  • Sounds in the repertory can be represented by

➡Sample grains (granular synthesis) ➡Synthesis parameters

slide-23
SLIDE 23

Sound repertory

  • Sounds (defined by synthesis parameters)

in the repertory have been pre-analyzed.

➡Sound features (e.g. MFCCs) define 2D position

  • f underlying synthesis parameters.
slide-24
SLIDE 24

Spectral modeling synthesis

deterministic components (sinusoidals) stochastic components (colored noise)

slide-25
SLIDE 25

Spectral modeling synthesis

Developed by Xavier Serra (PhD thesis, 1989)

slide-26
SLIDE 26

Instrument interface

slide-27
SLIDE 27

Technical building blocks

MFCCs 2D transformation visualization and control of Gaussian transformation regions SMS synthesis sinusoidals + residuum composition volume trace audio in audio out

audio graphical interface

MFCC lookup in color-coded picture of lookup of synthesis parameters in quad-tree SOM gesture SOM color-coded picture of sound SOM

made with pd + GEM

slide-28
SLIDE 28

Basic instrument setup

slide-29
SLIDE 29

Conclusion

  • The proposed instrument system is powerful

and flexible, featuring a simple and intuitive performance interface.

  • Like a traditional acoustic instrument,

this system demands practice as well as thorough preparation of the essential gesture and sound repertories.

↯The actual feasibility has yet to be verified.

slide-30
SLIDE 30

Future

  • Better analysis and synthesis.
  • Make preprocessing of repertories much faster

(etc. updating in real-time).

  • Make accessible temporal features and

transformation strategies.

  • Intensified research in high-level analysis,

representation and synthesis of musical structures.

slide-31
SLIDE 31

References

  • Thesis and poster: http://grrrr.org/pub
  • Sergi Jordà: New musical interfaces (PhD thesis)
  • Denis Smalley: Spectromorphology
  • Proceedings of the NIME conferences

http://www.nime.org

  • Other sources: see thesis bibliography