From ftm
Revision as of 09:14, 29 January 2007 by Glova (talk | contribs)
Jump to: navigation, search

MNM tutorial

The package MnM is a set modules extending the Max/MSP environment. It is part of the FTM library developed by the Real-Time Applications Team & Performing Arts Technology Research Team at Ircam. MnM stands for “Music is not Mapping“. This means that the interface will help the user (musician, dancer…) to interact with sound samples in a more complex way than a one-to-one relationship. A common mapping design would consist in setting up a univocal and simple correlation between gestures components and sound parameters. For instance, an upward motion of the arm will trigger a pitch increase… With the package MnM, the close integration of motion capture with complex statistical models and sound analysis/re-synthesis within in an environment such as Max/MSP offers new possibilities of interactions.


Workspace : overview

Get an overview of the interface functions. Example.jpg


1st STEP : Record gestures

Let’s start with two simple drawings : a triangle and a circle. Example2.jpg


2nd STEP : Compare

Draw a figure and then see how similar it is with your two referent drawings. Example3.jpg


3rd STEP : Observe

Pay attention to the curves below. They represent the velocity in X and Y axis of the mouse trajectories. That give a useful temporal information on how you realize your drawing. Example4.jpg


Connection Avec EyesWeb XMI

EyesWeb XMI, the open platform for real-time analysis of multimodal interaction, can be connected to Max/Msp throughout the OSC protocol (Open Sound Control). OSC is open, message-based protocol which was originally developed for communication between computers and sythesizers (cf. wiki).