Remymuller (talk | contribs) |
Remymuller (talk | contribs) |
||
Line 1: | Line 1: | ||
− | + | == About == | |
+ | The Real-time Applications Team conducts research and development in real-time computer technology for digital signal processing and machine learning for music, sound and gesture. | ||
+ | Over the years, [http://www.ircam.fr/ Ircam] has developed numerous hardware (4X, ISPW) and software environments (Max, FTS, jMax) for real-time music processing. Today the principal concern of the team remains live performance with computers representing a large field of research and development in man-machine interaction. | ||
+ | The applications addressed by the team cover the direct interaction of a performer with the computer as well as the mutation of techniques traditionally present in music, dance and theater - musical instruments, stage, accessories. The use of digital techniques can correspond to the augmentation or transformation of the performer’s expression – the sound of his intrument, his voice, his gestures, his memory - as well as the creation of a dialogue between the artist and the machine. | ||
+ | In addition to topics regarding live performance itself, such as gestural interfaces or real-time audio analysis and synthesis, the team is also concerned with tools for the composition and representation of works involving real-time processes and interaction. Further fields of applications are found in the larger context of audio and multimedia industry as well as education. | ||
+ | |||
+ | == Navigation == | ||
+ | * [[Projects]] | ||
+ | * [[Software]] | ||
+ | * [[Publications]] | ||
+ | * [[People]] |
Revision as of 13:15, 21 November 2006
About
The Real-time Applications Team conducts research and development in real-time computer technology for digital signal processing and machine learning for music, sound and gesture. Over the years, Ircam has developed numerous hardware (4X, ISPW) and software environments (Max, FTS, jMax) for real-time music processing. Today the principal concern of the team remains live performance with computers representing a large field of research and development in man-machine interaction. The applications addressed by the team cover the direct interaction of a performer with the computer as well as the mutation of techniques traditionally present in music, dance and theater - musical instruments, stage, accessories. The use of digital techniques can correspond to the augmentation or transformation of the performer’s expression – the sound of his intrument, his voice, his gestures, his memory - as well as the creation of a dialogue between the artist and the machine. In addition to topics regarding live performance itself, such as gestural interfaces or real-time audio analysis and synthesis, the team is also concerned with tools for the composition and representation of works involving real-time processes and interaction. Further fields of applications are found in the larger context of audio and multimedia industry as well as education.