{.vrtx-introduction} I am getting ready to travel to Sydney for the upcoming NIME 2010 conference where I am involved in no less than 5 papers:

Glass instruments – from pitch to timbre
Frounberg, I., A. R. Jensenius, and K. T. Innervik (2010)
The paper reports on the development of prototypes of glass instruments. The focus has been on developing acoustic instruments specifically designed for electronic treatment, and where timbral qualities have had priority over pitch. The paper starts with a brief historical overview of glass instruments and their artistic use. Then follows an overview of the glass blowing process. Finally the musical use of the instruments is discussed.

Evaluating the subjective effects of microphone placement on
glass instruments

Jensenius, A. R., K. T. Innervik, and I. Frounberg (2010)
Abstract: We report on a study of perceptual and acoustic features related to the placement of microphones around a custom made glass instrument. Different microphone setups were tested: above, inside and outside the instrument and at different distances. The sounds were evaluated by an expert performer, and further qualitative and quantitative analyses have been carried out. Preference was given to the recordings from microphones placed close to the rim of the instrument, either from the inside or the outside.

**Searching for cross-individual relationships between sound
and movement features using an svm classifier
Nymoen, K., K. Glette, S. A. Skogstad, J. Tørresen, and A.
R. Jensenius (2010)

**In this paper we present a method for studying relationships between features of sound and features of movement. The method has been tested by carrying out an experiment with people moving an object in space along with short sounds. 3D position data of the object was recorded and several features were calculated from each of the recordings. These features were provided as input to a classifier which was able to classify the recorded actions satisfactorily; particularly when taking into account that the only link between the actions performed by the different subjects were the sound they heard while making the action.

Using ir optical marker based motion capture for exploring
musical interaction

Skogstad, S. A., A. R. Jensenius, and K. Nymoen (2010)
The paper presents a conceptual overview of how optical infrared marker based motion capture systems (IrMoCap) can be used in musical interaction. First we present a review of related work of using IrMoCap for musical control. This is followed by a discussion of possible features which can be exploited. Finally, the question of mapping movement features to sound features is presented and discussed.

**Wireless sensor data collection based on zigbee
communication
**Torresen, J., E. Renton, and A. R. Jensenius (2010)
This paper presents a comparison of different configurations of a wireless sensor system for capturing human motion. The systems consist of sensor elements which wirelessly transfers motion data to a receiver element. The sensor elements consist of a microcontroller, accelerometer(s) and a radio transceiver. The receiver element consists of a radio receiver connected through a microcontroller to a computer for real time sound synthesis. The wireless transmission between the sensor elements and the receiver element is based on the low rate IEEE 802.15.4/ZigBee standard. A configuration with several accelerometers connected by wire to a wireless sensor element is compared to using multi- ple wireless sensor elements with only one accelerometer in each. The study shows that it would be feasable to connect 5-6 accelerometers in the given setups. Sensor data processing can be done in either the receiver element or in the sensor element. For various reasons it can be reasonable to implement some sensor data processing in the sensor element. The paper also looks at how much time that typically would be needed for a simple pre-processing task.

Electronic versions will be available next week.