Some pictures from the kickoff-seminar for the Sensing Music-related Actions project last week:

alt text

Project leader Rolf-Inge Godøy started with a short presentation of the new project.

alt text

Then Marcelo M. Wanderley (McGill, Montreal) held an overview of various types of motion capture solutions, and the pros and cons of each of them. He stressed two main challenges he had had over the years: synchronisation of various types of mo-cap data with audio, video, music notation, etc., and backwards compatibility of hardware, software and data.

alt text

Ben Knapp (SARC, Queens, Belfast) followed up with a lecture on various types of biosensing solutions.

alt text

Kjell Tore Innervik testing Ben Knapp’s EEG and EOG sensor.

alt text

The seminar attracted a very mixed group of researchers, from music, composition and performance to informatics, medicine and psychology.

alt text

After lunch, Patrik Almström from Swedish company Qualisys, presented their optical motion capture equipment.

alt text

A very interesting feature of the Qualisys system, is that it can use both active and passive markers. It is also great that the cameras can double as regular high speed video cameras.

alt text

Kristian Nymoen dressed in the mo-cap suit, and demonstrated the realtime possibilities of the system.

alt text

Then we walked across campus to the new project lab.

alt text

Mats Høvin demonstrated the robot Anna, which we will train as a musical conductor in the autumn.

alt text

Mats also showed a mechanical hand with artificial muscles, something which we will also test for musical purposes in the project.

alt text

Rolf Inge, Marcelo, myself and Ben “sightseeing” in Oslo after the seminar.