Live Interfaces 2020: Performance paper


The International Conference of Live Interfaces, ICLI, took place March 9-11 and was the second international conference hosted by NTNU Music Technology in only three months. Our conference chair Øyvind Brandtsegg had prepared the stage for a conference with talks, keynotes, concerts/performances and installations. As the conference web page stated, "topics of liveness, immediacy, presence (and tele-presence), mediation, collaboration and timing or flow are engaged with and questioned in order to gain a deeper understanding of the role contemporary media technologies play in human expression".


My contribution at the conference, besides being chair of the installations, was a so-called performance paper, a not very well-known format combining a traditional paper presentation with performance elements. In my paper and presentation, I addressed a topic that I have been occupied with on a hobby basis for a couple of years, namely FM radio. After the shut down of the national broadcasts on the FM-band in Norway in 2017, my frustration with the political descision led me into collecting old radios, especially from the late 1960s and early 70s. I later found out that these radios produces a wide range of noises and sounds when they were connected with transmitters in a feedback loop. The paper contained some of the experiences and reflections I had after experimenting with this setup (see the figure below) during some months.



Abstract: This paper investigates how FM receivers connected to FM transmitter in a feedback loop can be applied as musical instruments, but with an inherent instability that at the same time breaks with traditional instrumentality. The paper begins with delineating the specific historical situation in Norway with the closing of FM band for national broadcasters, and how this could be seen as a call for action. The paper proceeds with a short review of the application of radios in music and sound art and practitioners working with feedback techniques, before presenting relevant theories of nonlinear dynamic processes and feedback systems. The paper then briefly delineates a simple setup of equipment, and goes on to address issues related to playability and sonic output. The paper shows how certain sonic categories are recurrent, but how different models have different sonic signatures, how some features are unstable and difficult to consistently reproduce. The paper concludes by discussing the findings in the light of the research literature in the field, and potential musical applications of the radio as instrument in different setups and contexts.


The full article can be found here. I also extended the content of the paper by including an analysis and categorization of the sounds from one of the radios, the Blaupunkt derby de luxe (image below). This analysis can be seen in the presentation slides found here.



Even though I had some problems with grounds hum during the performance part of the presentation, I believed it showed some of the potential these old devices have for being used as musical instruments. In the wake of the conference, I was also contacted by Knut Aufermann, one of the pioneers of the radio art scene in Germany, giving me some extra links and material, some of which I will hopefully look into in due time.


Web Audio Conference 2019: concert and panel

The international Web Audio Conference 2019 was hosted by NTNU in Trondheim, and the Music Technology Group played a central role in organizing it, with my colleague Anna Xambó as conference chair (and IES colleague Sara Martin by her side). My own role in the organization was demo and poster chair. The conference presented a combination of talks, posters, demos, artworks and performances, and had over 100 participants from all over the world. I also contributed to the conference by participating in a performance and hosting a panel.


The performance was with T-EMP (Trondheim Electroacoustic Music Performance), consisting of Øyvind Brandtsegg, Trond Engum, Anna Xambó and Carl Haakon Waadeland. Here is the description of the performance from the proceedings of the conference:

The ensemble Trondheim Electroacoustic Music Performance (EMP) investigates new modes of communication in an ensemble when new technology is introduced as part of the ensemble repertoire.  In the recent three years, the focus has been on cross-adaptive processing as a musical intervention in the interplay.  More info on the cross-adaptive project can be  seen  at  the  project  blog  and  also  the  album  Poke It With A Stick / Joining The Bots released in 2019.  For this performance we work with a web-based repository of sounds, explored via a live-coding interface. This allows access to a massive archive of sounds from, and the selection of sounds is done via sound descriptors.  The integration of this instrument in an ensemble setting is interesting, as traditional and nontraditional modes of musical interaction are activated in dialogue.  Using web-based access to the repository allows a generality of instrument design, and this is combined with a strategic mode of performance in this instrument.  With?strategic? we mean here, that most sounds are not performed directly by physical action, but cued up in patterns via live-coding.  These sounds are then live processed by other members of the ensemble, responding more directly with physical and gestural instrumental action on the signal coming from the web instrument.  The signal is also live processed in a cross-adaptive fashion, so that audio features extracted from other performers? actions will modulate the parameters of processing for the web audio instrument.  The ensemble also utilize gesturally based interfaces from interactive dance, here used to control elements of audio synthesis and processing.

Here is a video of the performance:


I was asked to host a panel with the topic "accessibility in web audio". I invited three people who I though would be able to address relevant issues of this topic, although none of them were directly engaged with web-audio. David Brüggermann was a former Music Technology student who wrote his bachelor thesis on developing workflows for music production with VoiceOver and LogigPro for a vision impaired user. Monica Ruud is a first-year student at the Bachelor programme in Music Technology NTNU who is blind. Lastly, Miranda Moen is currently currently writing a master's thesis in Equality and Diversity at the NTNU institute of interdisciplinary culture studies. Moreover, she works as a project responsible at AKKS Trondheim and with the KOSO collective. In the panel the participants were first asked to explain how they understand the concept of accessibility.

The two visusually impared students were asked how their engagement with computer-based music technology as a user with vision impairments and if they had  examples of functionality in computer-based music applications which are very much accessible, and others which are not, and if so, in what ways.

The panel was rounded off with questions that was partly addressed by the panelist, and partly by the audience. For instance, how one can raise awareness for issues of accessibility in web audio applications and if there were any concrete suggestions for making web audio applications more accessible. Here is a video of the panel discussion:

VIBRA workshop: 'Constructing Realities' with Johannes Birringer

"Constructed Realities" was a two-day workshop led by Johannes Birringer, as a part of the VIBRA workshop series ( that I organize. Birringer is an independent media choreographer and artistic director of AlienNation Co, and also works as a professor at Brunel University, where he is director of the DAP-Lab and headed the Centre for Contemporary and Digital Performance.

The workshop was organized by me, and sponsored by NTNU ARTEC. VR equipment and expertise was provided by IMTEL lab NTNU. Footage by workshop participants and video editing by Andreas Schille.

At the workshop, fifteen participants with diverse backgrounds, from fine arts, to music technology and dance, worked together to create an immersive environment, including two VR stations, multichannel audio, live visuals affected by audio and video. In addition to organizing and facilitating the workshop, which took up most of my attention, I made a "discoball poem player", using a broken disko ball fittet with a wireless NGIMU-sensor, a recording of a poem read by Dumama, and piece of coding I did in Csound. It can be seen in the video below. The final performance/environment was presented for an audience in Verkstedhallen, Trondheim, October 2019. Workshop participants: Pål Lersveen, Dag Olav Kolltveit, Naomi Chan, Ada Hoel, Samrridhi Kukreja, Emily Aalde, August Norhtug, Gugulethu Duma, Andreas Schille, Mina Paasche, Christopher Logan, Anna Thu Schmidt, Dylan Green, Aleksandar Isailovic, Per Erik Walslag and Ingunn Schumann Mosand.

Here is a video, edited by Andreas Schille, showing some of the different parts of the workshop, including the performance itself:

#MTF: Humans in Da Loop at Örebro University

Two performers with laptops by a table. Two dancers in the foreground. One industrial robot on the side.

Music Tech Fest is an arts festival which presents "technological innovations and artistic experimentation, performance, new inventions, commercial applications, and academic research". It is also a network a network of amazing people working in different branches of music technology. At least once a year they organize #MTFLabs, where they invite "some of the most brilliant and original minds to reinvent the way we interact, collaborate, create and communicate with each other". After particpating in Stockholm in the fall 2018, I hooked up to a small team of fantastic people (Kirsi Mustalahti, Joseph Wilk and Lilian Jap), and together we formed a group called Accents in Motion, or alternatively, Dance AI, building our work around connections between a database of speech accents, motion tracking of dance movement and live coding.

In October 2019 our group was invited to the AI Impact Lab at the University of Örebro as a part of MTF's Humans in da Loop together with a number of people from the MTF community. At the gathering, MTF were "bringing select experts from the global MTF community to address strategic applications of AI" led by their high-profiled professor Amy Loutfi, newly appointed pro-vice-chancellor for AI. The Accents in Motion team were very lucky to be allowed to work in The Cognitive Robotic Systems Laboratory, lead by professor Alessandro Safiotti. Although we were not allowed to play with the industrial robots in the space, we got access to their system for counting people and getting their position in the space. Although the system was more difficult to work with than anticipated, we were able to connect it to our computers and make sound with it.

The functionalities and failures of this system inspired our ideas of the performance - it addressed the way in which technology is often presented to us as black boxes that don't reveal anything about their construction or inner workings, and how we should strive to make technology more transparent and open to users. The failure of the tracking system to locate a person within a 2x2 meter square in the middle of the lab then became our "Black Box", where we then placed our laptops, mixers and "technologist" performers (Joseph and Andreas). The dramaturgy of the performance was then how two of the performers (Kirsi and Lilian) intruded into the Black Box and took over the control of it - which was the conclusion of the performance. Thus, the piece became one of empowerment through knowledge.

Happy performer in wheelchair. Two computers on a table. Two dancers in the background. A screen with computer code on the back wall.

As at the Stockholm MTF we got our songs from an audio database of considerable size, namely DAMP, containing voice-only recordings of karaoke singing. The material in the database was challenging in that it already contained musical form and expression, and to reshape and reinterpret this was not at all easy and required considerable experimentation to get an aesthetically satisfactory result.

And as before, we also had two approaches to our sound database - both movement sensing and live-coding. Even though this piece came together the "hackathon" way - extremely compressed in terms of creation and implementation time, we still think we got to develop our concept and repertoire further. Especially, it was intriguing to play with our roles as "technologists" versus "dancers", and hopefully contribute to a productive dialogue on technology, embodiment, gender stereotypes and empowerment.

Conference presentation: Digitale kulturer

The project Crossadaptive processing as musical intervention (2016-18), firmly rooted in the Music Technology programme at NTNU, was invited to do a presentation at Arts Council Norway's conference "Digital kultur, estetiske praksiser" (Digital culture, aesthetic practices" at Dokkhuset in Trondheim. In the absence of project leader Øyvind Brandtsegg, we were three colleagues from this programme doing the presentation entitled "Å gripe inn i en annens lyd: kryssadaptive modulasjoner" (To interfere with the other's sound: crossadaptive modulations); myself, Trond Engum and Carl Haakon Waadeland. The presentation had two parts; first I gave a brief talk about the central concepts, methods and results of the project, then, Trond and Carl Haakon improvised using the tools from the project.

Carl Haakon Waadeland, Trond Engum and myself at Dokkhuset

Paper at the Sound and Music Computing Conference

SMC Malaga

Together with Sigurd Saue and Pekka Stokke I wrote a paper about VIBRA for the Sound and Music Computing Conference in Malaga.


The paper presents the interactive dance project VIBRA, based on two workshops taking place in 2018. The paper presents the technical solutions applied and discusses artis- tic and expressive experiences. Central to the discussion is how the technical equipment, implementation and map- pings to different media has affected the expressive and experiential reactions of the dancers.

The paper can be found here:

NRK P3 Interview

The NRK P3 show "Urørt" made a short feature about instrumental music February 2019, and I was interviewed to contribute to the feature. The feature with the interview is available here.


PhD supervision course

I attended a two-day course for PhD supervisors at NTNU January 2019 led by Kari Smith, professor at the Department for Teacher Education at NTNU, and leader of the National Research School for Teacher Education. The course qualified me as a "certified supervisor", which from the new regulations of the PhD programme at NTNU is required for at least one of the team of supervisors at a PhD (alternatively, the supervisor has to have supervised earlier).


The course covered many issues relevant for PhD supervision, among them regulations and framework, building relations, gender issues, ethics, co-authorship, submission and what to do in the case of rejection. We also got to discuss some interesting cases, including the typical "burn out" that many experience in the last phase of their PhD period. The course also highlighted the culture differences at NTNU, especially between hard and soft sciences, article based and monograph theses, and small and big departments.

MotionComposer workshop at Dämeritzsee, Berlin

The creative part of the MotionComposer team was gathered for four days at the Dämeritzsee Hotel in December 2018 - lying next to a beautiful lake - for a kind of "hackathon" workshop, aiming to develop the 3.0 version further through long days of work.

The Dämeritzsee, just outside of Berlin

The new tracking software using stereoscopic camera input was now much more developed, and could deliver tracking data of quite good quality for many of the core parameters. For my part, it was the first time I could try the new version of Particles with the new tracking system, and we were happy to experience that the central features of the sound environment worked well. There were some issues with certain parameters being too "trigger-happy", so this is something to be further improved in the tracker.

As for the features of the Particles sound environment, we decided to include arm hits to the side, kicks and jumps to it. I briefly made some sketches with glissanding sound particles that had some interesting qualities, and these will be developed further for the final version, especially giving them a sharper attack. Here are some hits followed by a few kicks:

Sketches for hits and kicks for Particles for MC3.0

I also spent a bit of time preparing a work flow for making rhythms for Particles, including making a sequencer interface in Max that could make a file that could be immediately loaded into csound. Here is a screenshot of the 8 track, 64-step sequencer:

8-track, 64-step sequencer for Particles, built in Max

We concluded the workshop with setting milestones for the further development of Particles in the spring 2019.


Stocos workshop and artist talk

As a part of the ARTEC seminar series, I had invited Stocos, a Spanish/Swiss artist trio creating interactive dance pieces, to do a two-day workshop and an artist talk with demonstration. The trio consists of Pablo Palacio, composer and interactive music instrument designer, Muriel Romero, choreographer and dancer, and Daniel Bisig, visual artist, sensor technician and programmer. The group combines research into embodied interaction, artificial intelligence, and simulation of swarm behavior with artistic exploration of music, visuals and dance.

Stocos (Pablo, front left, Daniel front center, Muriel, far right) with workshop participants

The workshop took place at DansIT's (Dance in Trøndelag) studio at Svartlamon in Trondheim. It took place as a collaboration between ARTEC and DansIT, where the former covered the expenses related to travel and accomodation for the group, and the latter offered me their venue free of charge as a recidency. A total of fourteen participants (including me as the organizer) took part in the workshop, even if some were not present at all times. The group consisted of people with very varying backgrounds: composers, musicians, dancers, actors, media/video artists, stage designers, dramaturgs, philosophers and music technologists. Of the participants, eight of them were either employed or students at NTNU and another three had relatively recently achieved their degrees (Master/PhD) at NTNU and were now working in the art sector in Trondheim. Due to restrictions in the number of people, there were additional applicants to the workshop on a waiting list.

The first workshop day started with a presentation of the workshop holders in Stocos, their artistic aims and statement, and then continued with an overview of their work during the last 10 years or so. A core concept of the group, which is also reflected in their name, is the use of different types of stochastic algorithms, both at signal-level (in the form of noise) and in different ways of mapping movement parameters with visuals and sound. In addition, different ways of applying swarm properties in especially mapping and visual projections were central. This was applied for the first time in the piece with the same name as the group, where the swarms were first of all visible as minimalistic visual projections, mostly in black and white, showing either moving entities or trajectories of invisble moving entities - interacting to larger or lesser degrees with dancers. Another work, Piano-Dancer, was an exploration of the same core principles of stochastics and swarms, but here a single dancer's movements and algorithms were controlling a motorised piano. In the piece Neural Narratives, they were creating virtual extensions of the dancing body with computer visuals, with tree-like entities following the movements of the dancers while also having a semi-autonomous character. Lastly, they presented their most recent work, The marriage between Heaven and Hell, based on the literary work of William Blake. In this work they continued their work with interactive control of voices, this time using phrases from Blake's poem. Particular to this work was the use of Laban derived bodily expressive aspects, such as energy, fluidity, symmetry or impulsivity, translating them (in real time) into audible and visual digital abstractions. This piece also featured the use of body-worn interactive lasers, where the movements of the dancers controlled whether they were active of not. The presentation spurred a lot of questions and discussion about everything from technicalities to aesthetic issues.

The second part of the first workshop day was more technical and focused on the different types of sensors the group used. They started showing technology they have used in earlier years, namely the Kinect. Using the open source driver implemented in EyesWeb the sensor gave them access both to the countour and the skeleton of the dancer. Even if the Kinect had relatively high latency and the skeleton tracking was quite buggy, they still had used it In recent years. More and more, however, they had moved to IMUs for their tracking. They used a Bosch/Adafruit BNO055 connected to an arduino with wi-fi shield, with everything put in a 3D-printed case, and then put in an elastic band - attached to the arm or the lower leg. This avoided several of the problems associated with video-based tracking (occlusion, IR light from other sources, etc.), even if there were specific challenges associated also with the IMUs. The IMUs also offered absolute orientation data in 3D, making it possible to track the position and rotation of the extremities, not only their movements. Finally, the group showed us their most recent development, namely sensor shoes, with seven pressure sensors in the sole.

Daniel Bisig explaining the construction of the sensor shoe.

After just having monitored the values from the sensors in visual graphs, the last part of the first workshop day focused on the mapping on the values from the sensors to sound. Here, the participants got to try out the IMU bands and the sensor shoes with several different mappings - synthetic singing voices, sampled speaking voices, stochastically modulated noise, and several others.

The second day of the workshop started with a movement session led by Romero with focus on exercises related to Laban's theories of individual choreographic spaces around the dancer, and how these could be divided into planes cutting in all directions around a dancer. Many of the exercises directly articulated these planes.

After the movement session, the focus was moved to the computer generated visuals. Most of these were in some way or another derived from swarm simulations, and Daniel Bisig spent quite some time explaining how these were programmed and manipulated so as to create different visual expressions. Among other things we got to see how using the individual agents' traces as visual elements along with the manipulation of pan and zoom of the space they inhabited created a multitude of different patterns and effects. Some of the participants also got to try out an interactive algorithm that used the contour around the body to attract the swarm in such a way that they delineated the user's silhouette. We also got to see how the swarms could articulate letters more or less using the same technique, and how the dampening of attraction forces made the contours around the letters less distinct.

Workshop participant Elena Perez trying out the interactive swarm visuals.

There was also time to try out the interactive lasers (with sound) from The marriage between Heaven and Hell, which was, naturally enough, very popular. All the lights were turned off, and a hazer was activated, so that the trajectories of the lasers were visible. The lasers all had lenses spreading the beam vertically about 30% so that the result was a line when it was projected on the floor. We also got to see how moving the lasers in different ways generated different effects, like "scanning" and "strobe". We then did further tryouts in the darkened space - now with the sensors controlling stage lights (via MIDI/DMX) together with transformed vocal sounds. Even if the lights had a significant latency, it was nice to see how they could be mapped from the IMU sensors.

At the end of the workshop, Stocos presented longer extracts from several of their pieces, to answer the question that many participants had posed: How do you put it all together? We then got a much better understanding of the artistic contexts and intentions the different technologies and mappings were used in - and further how their artistic expressions could be transformed with different techniques.

All in all, the impression was that the workshop participants were highly inspired by the workshop, and several of them talked about getting new artistic ideas. Stocos were also impressed with the engagement that the participants showed. Several people also expressed a wish to have Stocos back to Trondheim to do a full performance.

Artist talk at Kunsthallen

Monday the 26th of November Stocos gave an artist talk at Kunsthallen as a part of the ARTEC seminar series. Around 15 people attended. Stocos presented their artistic aims and statement, and then continued with an overview of their work during the last 10 years or so. In between the discussion of the different pieces, Muriel Romero demonstrated the sensor instruments used in the different performances. The artist talk lasted 90 minutes.

Muriel Romero demonstrating the sensor shoes. Daniel Bisig is watching.

Student concert at Planetariet

As a part of the course Electroacoustic Music - Analysis and Composition that I teach at NTNU, five master's students at the music technology programme got to perform their newly composed pieces at the Planetarium of Vitensenteret (The Science Centre) in Trondheim November 16 2018. The concert took place as the first milestone in several people's joint efforts of transforming the stunning 36.2 channel sound system, custom made for planetarium shows, into a concert venue for multichannel/spatial music. NTNU Music Technology, with myself and PhD candidate Mathieu Lacroix, composer Frank Ekeberg and padagogue at Vitensenteret, Lars Pedersen, have been researching to find a setup that can work for multichannel presentation of electroacoustic music. After quite a bit of experimentation during the fall of 2018, we ended with a solution where we used Blue Ripple Sound's plug-ins to play music in 3rd order ambisonics through the Reaper DAW running on an external laptop connected to the IOsono system via Dante. At the concert, the students presented compositions they had made as a part of the course, but adapted to the 36.2 setup at Planetariet. The concert was concluded with Ekeberg's piece Terra Incognita from 2001. The concert program can be read here. It is my intention to publish a blog post later in the winter with details about the setup to help others arrange multichannel concerts at Planetariet.


Panel at RITMO Workshop

The newly started center for interdisciplinary studies in rhythm, time and motion, RITMO, hosted a workshop to celebrate their opening November 12th-14th 2018: RITMO International Motion Capture Workshop.
I was invited by the chair of the event, Alexander Jensenius (UiO) to participate in a panel about interactive music systems together with John Sullivan (McGill University), Cagri Erdem (UiO) and Charles Martin (UiO). Each of us got to give a short presentation of our work, before we sat down to have a panel discussion, also involving the audience. In my introduction, I presented the particular challenges of creating interactive music systems for people with different abilities in my work with MotionComposer.

In the workshop there were several panels like this, with topics like Music Performance Analysis, Music-Dance Analysis, Music Perception Analysis and Different Types of Motion Sensing. There was also one whole day dedicated to different tools (MoCap Toolbox, MIRToolbox, EyesWeb, and more). High points in the program were the three keynotes by Marc Leman (Ghent University), Marcello Wanderly (McGill University) and Peter Vuust (Aarhus University). We also got a tour of the RITMO center, showing off the motion capture lab, eye tracking lab and more.

Panel about motion capture tools at the RITMO International Motion Capture Workshop

Sound design workshop at Fusion Systems

October 8 - 11 2018 the German company Fusion Systems, a company working mainly with computer vision and sensor fusion for traffic and industry applications, hosted a workshop together with Palindrome Dance Company. The workshop aimed at developing sound design for interactive music applications for people with different abilities. I was invited to the workshop together with Marcello Lusanna and Ives Schachtschabel to work on sound design and meaningful movement-sound mappings. During the workshop we got to try out a work-in-progress prototype of a stereoscopic motion tracking system.

Stereoscopic motion tracking prototype at the Fusion Systems workshop.

Panel at Knowing Music, Musical Knowing

Knowing Music, Musical Knowing was a conference and International Music Research School hosted by my university NTNU and Grieg Research School October 23rd-26th. On the website of the event, it was described in the following manner: "This event will explore the complex and multifarious connections between music and knowing. Bringing together key thinkers and practitioners in contemporary interdisciplinary music studies ? including but not limited to dance studies, music therapy, ethnomusicology, music technology, historical musicology and music education ? this four-day school will inspect both the unique forms of knowledge embodied through musical practice and the different ways humans create and impart knowledge about music."

Myself together with Elin Angelo, Jørgen Langdalen, Simon Gilbertson and Kari Holdhus.

For this event, I was asked to participated in a panel  about the challenges and possibilities for musical knowledge transmission, learning and education between music disciplines. The panel consisted of professor Elin Angelo, Head of Department of Music, NTNU, Jørgen Langdalen, associate professor of music therapy, Simon Gilbertson, associate professor Kari Holdhus from Høgskolen Stord/Haugesund, and post doctor Nora Kulset.

In my five minute introduction I briefly discussed how common denominators in the form of passion, listening and musicking bodies can allow us to build understanding, communicate, as well as improve learning, across disciplines. The whole introduction can be read here.

Music Tech Fest, Stockholm, September 2018

Music Tech Fest (MTF) is an annual event gathering artists, professionals, hackers and academics for a few days to work in an intense and stimulating environment. This time it was KTH in Stockholm that had taken on the event, with the main event happening September 7-9 2018.

I was very happy to be invited to participate in the #MTFLabs, which was an extended part of the whole thing, going on for a full week, Monday - Sunday. In the Labsa total of 80 artists, hackers and acedemics were gathered with a mission of putting on a 45 minute show at the end of the week. Everything took place inside a shut down nuclear reactor 25 meters below ground.

Reaktor hall
The reactor hall at KTH. Photo by Mark Stokes.

The whole thing was kicked off by presentations by Michela Magas, the founder of MTF, Danica Kragic, Airan Berg and pop singer Imogen Heap. Then, during a process of self organization, we split off into different groups and sub groups, according to what we wished to work with in terms of technology and artistic expressions.

I soon found a good connection with Joseph Wilk (UK), a creative (live)coder and computer wiz, and Lilian Jap (SE) a media artist and dancer, and we decided to form a group under the theme of paralanguage. Later, the group was completed with Kirsi Mustalahti (FI), an actress and activists, working with arts for disabled people through the Accessible Arts and Culture (ACCAC) organization.

Me with Lilian Jap, Joseph Wilk and Kirsi Mustalahti

Inspired by the post-apocalyptic feeling in the reactor and theme of extra-terrestial colonization that was circulated in the whole of the #MTFLAbs group, we decided to make a piece about having to re-invent language through paralanguage and body movements. Joseph discovered a really nice speech data base with Enligsh accents, featuring almost 2700 sound files with a single statement being spoken by different speakers from from all over the world. The database was a nice metaphoric expression of "humanity" that we felt was a good fit with our idea. Our initial idea was then to use a wavenet encoder - a machine learning technique working on the audio sample level - to train a model that we could control with the Kirsi and Lilian's movements. However, it turned out that this took a lot more time than we had at our disposal.

With the sample database downloaded we at least had a rich source of sounds, that also had a nice consistency. Joseph was then able to organize the material so he could play it using his live coding setup, and this worked so well that we kept it in the performance for the build-up in the final section.

Without a trained model and an interface to the parameters of the database, it was difficult for me to create an instrument that could play the whole database. I therefore chose to use only a subset of the samples (about 100 of them), that I stretched and layered in different ways using Csound. We put an NGIMU sensor and a Myo sensor armband on each of the dancers, and they were then mapped so as to control each of their individual set of voices. We then structured the piece in two sections, the first one (discovery and first contact) dwelling on two single syllables "plea" (from please) and "call", with each syllable being played by one dancer, and the final part (beginning communication developing into chaos), building up textures and layers of an ever increasing amount of words and syllables played by dancers and Joseph's live coding. The piece ended after about 4:30 with a sudden freeze of movements and silence, whereupon a comprehensible version of the speech material was presented to the audience: "Please call Stella. Ask her to bring these things with her from the store" - a highly mundane statement rounding off the somewhat existential thematics.

Here is a video of the first section of one of our performances (we had three in total). It is shot by Mark Stokes:

We all agreed after the performance that our work together had a lot of potential, both in terms of artistic development and application for people with other abilities, so we have put in several applications for further work with the project called Accents in Motion.

Running version of Six Fantasies Machine


The Six Fantasies Machine (SFM) instrument is a tool developed in Csound for producing a range of sounds similar to those found in Paul Lansky?s classic computer music composition Six Fantasies on a Poem by Thomas Campion. I made it as a part of my PhD project, Experiencing Voices in Electroacoustic Music, which was finished in 2010. Due to changes in Csound, for many years I couldn't run the software on the recent versions, which was a bit sad. However, I have now (October 2018) gone through the code again and got it up and running on Csound 6.12 on OSX.

For download, info and sound examples, check the SFM project page.

Startup of VIBRA - network for interactive dance


After having worked with interactive dance over some years mostly outside of Norway, I found it could be interesting to gather people in my local region to do some work here. I was lucky enough to get a grant from the Norwegian Culture Council, which made it possible to arrange a series of three workshops in Trondheim with dancers/choreographers, video artists and a music tech colleague during the winter/spring of 2018.

The network was named VIBRA, an Norwegian acronym which translates into Visuals - Interaction - Movement - Space - Audio. The project now has its own website at where we have put information about the project and a blog about technical as well as artistic issues.

VIBRA is a network of artists and technologists who explore artistic expression possibilities related to video, interaction, movement, space and audio - in short form, often referred to as interactive dance. With interactive dance, we mean a cross artistic expression in which dancers using sensor and computer technology control one or more aspects of sound, music, video, light, or other scenic instruments. This has hitherto been largely unexplored in Norway. Much of the non-classical dance presented on Norwegian scenes today is accompanied by pre-recorded musical events and / or video. With such a fixed accompaniment, the relationships that occur between motion, music and visual elements will work essentially one way. By using new technology, we hope to introduce an interactivity that can provide a more dynamic expression that more closely puts cross-cultural relationships at stake. The start-up of the network will involve a bundle of people from Trondheim / Trøndelag that have a relevant background and will constitute the project group.

Here is a short excerpt from the showing of the last workshop, using NGIMU and Myo sensors with four dancers and projecting sound over a 8-channel setup:

The video excerpt shows dancers Tone Pernille Østern and Elen Øyen.

Presentation at IRCAM Forum Workshops

IRCAM forum
The IRCAM Forum Workshop is an annual event in which IRCAM Forum members get to share and discuss ideas. As they write on their website, "The Ircam Forum workshops is a unique opportunity to meet the research and development teams as well as computer music designers and artists in residency at IRCAM. During these workshops, we will share news on technologies and expertise providing hands-on classes and demos on sound and music."

As an IRCAM Forum member I made a submission with Robert Wechsler entitled MotionComposer - A Therapy Device that Turns Movement into Music, and it was luckily accepted and the presentation was held the 8th of March 2018. It was a great experience to show our work at this renowned institution in front of a highly competent audience. And of course, we also greatly enjoyed to see presentations of the IRCAM teams as well as the other IRCAM Forum members.

The MotionComposer is a therapy device that turns movement into music. The newest version uses passive stereo-vision motion tracking technology and offers a number of musical environments, each with a different mapping. (Previous versions used a hybrid CMOS/ToF technology). In serving persons of all abilities, we face the challenge to provide the kinesic and musical conditions that afford sonic embodiment, in other words, that give users the impression of hearing their movements and shapes. A successful therapeutic device must, a) have a low entry fee, offering an immediate and strong causal relationship, and b) offer an evocative dance/music experience, to assure motivation and interest over time. To satisfy both these priorities, the musical environment "Particles" uses a mapping in which small discrete movements trigger short, discrete sounds, and larger flowing movements make rich conglomerations of those same sounds, which are then further modified by the shape of the user's body.

Invited lectures: Healing soundscape


Robert Wechsler and myself were invited to give a lecture and performance at the Hochshule für Musik und Theater in Hamburg the 16th of November 2017 as a part of the lecture series Healing soundscape: Klang - Raum - Atmosphäre. The title of our lecture/performance was "MotionComposer ? a movement-into-music device for persons of different abilities? ? a demonstration/performance with audience participation.

The invitation actually resulted in a total of three lectures: One for masters students of multimedia composition and one open for a general public. The first lecture was more technical, focusing on both the tracking, musical mapping and the how it was implemented using Csound.

The second lecture was for a group of music therapy students and focused on the therapeutic applications of MotionComposer.

The third lecture was the public lecture which was a part of the lecture series. In it we argued how dance and music are closely related, and how that has been an impetus in the design and application of MotionComposer. It also gave us a chance to show artistic applications of the MotionComposer. A local dancer, Trinidad, improvised in the particles environment together with Robert Wechsler. I had brought with me some sounds that we had never tried before - they were recordings of manipulations of a mixer with closed feedback and some effects - and they turned out to work nicely. After the performance, the audience also got to try out several environments of the MotionComposer.

Here is an excerpt from the performance with a part of the performance where Trinidad has a solo. Unfortunately, it only has audio from the camera resulting in quite poor quality:

Video from the lecture/performance "MotionComposer ? a movement-into-music device for persons of different abilities? ? a demonstration/performance with audience participation featuring the dancer Trinidad.

Sound for photo exhibition: Ekko frå sjelas spegel

One of the images from the exhibition Ekko frå sjelas spegel with accompanying poem

In October 2017 I got the chance to work with Trondheim based photographer Siri Nordberg. She specializes in portraits and in her exhibition Ekko frå sjelas spegel (Echoes of the mirrors of the soul) she even combined portraits with poems she had written (later to be published in a book). It was exhibited in very special exhibition space: an arched cellar from the 19th century in the centre of Trondheim. Since an important theme in many of the texts were related to the sea and coastal landscapes, Nordberg wanted me to make a soundscape composition that could fit. I chose a recording I had made at Kvarøya in northern Norway, which featured a beautiful crescendo and decrescendo of waves and a boat passing in the distance. I treated it lightly to strengthen the inner dramaturgy of the composition. Listen to an excerpt here:

Tower Block - Post Mortem article


A few years back I wrote an article reflecting on different aspects of The Tower Block - Post Mortem project together with NTNU colleague Barbro Rønning. The article was originally intended to be a part of a book project on intermediality edited by Rosemary Klich, which was unfortunately not realized. A possibility then opened to publish the article in the International Journal of Aesthetics and Philosophy of Culture with a special issue called Nordic Time Zones: Art and Performance in the Northern Landscape, with Knut Ove Arntzen as invited editor and a special introduction by Shannon Jackson.

Abstract: This article discusses aspects of intermedial experiencing of the site-specific intermedial performance The Tower Block ? Post Mortem and the preceding process of generating material for the performance through gathering and sharing memories from The Tower Block. The authors describe the gathering of memory material on the project?s own blog, and how this process highlighted both properties of networked media and the phenomenology of remembrance. Subsequently, they discuss three different aspects of the performance, and how each of them relates to important issues in the discourse on intermediality in performance culture: 1) Intermedial elements in performance and their play with spatiality, temporality and their relationship to processes of remembrance; 2) Ritual and the transformative aesthetic of the performance and how it relates to the authenticity, and to interrelationships of the actual and the virtual; 3) Resonance and tension in trans-medial metaphors used in the performance.

The article can be downloaded here, and the whole issue on this link.

Crossadaptive seminar


The project Cross adaptive processing as musical intervention, led by colleague Øyvind Brandtsegg at NTNU, arranged a two day seminar November 2 and 3 2017 at the Department of Music, NTNU in Trondheim. The seminar featured a list of international participants in addition to colleagues from NTNU and The Norwegian Academy of Music (NMH): Miller Puckette, Simon Emmerson, Josh Reiss, Victor Lazzarini, Gary Bromham and Marije Baalman. The seminar featured academic presentations, panels and performances, and is well documented in video and slides at the project web site. My contribution was a presentation entitled Documentation in the Crossadaptive project, and focused on the use of video as a tool for documentation and reflection. The slides for the presentation can be downloaded here, and this link features a video of the talk.

Musikkteknologidagene (Music Technology Days)


Musikkteknologidagene (Music Technology Days) is the meeting point of the music technology field in Norway, and in October 9-11 2017 it was hosted by the Norwegian Academy of Music (NMH) in Oslo. The programme consisted of academic presentations during the day and concerts in the evenings. I had two contributions, one academic and one artistic.

The academic contribution was with NTNU colleague Sigurd Saue, and was entitled Some Experiences from the project Sound - Space - Movement. The presentation featured a discussion of different aspects of the project with the same name, including motion tracking, body movement, movement->sound/spatialisation mapping, flexible reverberation with streaming convolution. An abstract in Norwegian can be found here.

The artistic presentation was an interactive dance performance featuring dancer and NTNU colleague Tone Pernille Østern taking place in Levinsalen at NMH with a 24-channel loudspeaker setup, combining a circle and an oval configuration. We were incredibly happy to be in the distinguished company of composers such as Åke Parmerud and Natasha Barrett, and the impro group Parallax, featuring spatialisation by Anders Tveit.

Movement and Computing Conference (MOCO)


The Fourth International Conference on Movement and Computing (MOCO 2017), was held at Goldsmiths University, London, June 2017. It featured highly regarded researchers from many different fields, ranging from computational neuroscience, affective computing and interactive dance. With less than 45% of submitted papers being accepted, me and MotionComposer colleague and leader, Robert Wechsler, were very happy to being in such a distinguished company of artists and researchers. Our contribution was a short paper entitled Issues and Strategies of Rhythmicality in MotionComposer and you can read it in the proceedings found here (choose Table of Contents and navigate to the title). Among other things the use of artificial intelligence and machine learning was a topic that was recurrently addressed, and Marc Coniglio spent a good portion of his keynote speech to speculate how this could affect the field in the near future.

MotionComposer workshop at Fraunhofer


In June 2017 the MotionComposer team was gathered for their annual workshop in Jena, Germany, at the much renowned research institution Fraunhofer and their devision for optics and precision engineering. Together with fellow MC-composers Ives Schachtschabel and Peter Breitenbach, I had the task of starting the development of sound environments for the MC version 3.0. We had a startup meeting with representatives for Fraunhofer, the organization Grenzenlos where we presented our visions and planned features for the different environments. After four days of intensive work our final presentation, with performers from Grenzenlos, also included a test of the sensor in development planned for the MotionComposer 3.0 and built by Fusion Systems, where parameters of the 2.0 Particles environment were controlled by the sensor. All in all the workshop gave a highly valuable impetus for developing the Particles environment further, including suggestions for new sound banks.

Radio interview about the Telharmonium


NRK (Norwegian Broadcasting) did a feature about one of the world's first electronic instruments, the Telharmonium, April 11 on the P1 channel. I was happy to be asked to contribute to the feature, and did a short interview where I talked about the instrument and did a short demo using a rough additive synthesis simulation. You can listen to the feature (in Norwegian) with the interview here. The Music technology student, Joachim Ose, was the journalist behind the feature.

Sound-Space-Movement. Research concert.


As a result of an initiative by Njål Ølnes, post.doc, the Department of Music, NTNU, has started a series of what we call Research Concerts. The series combines elements from research presentations for a non-scientific audience with the concert form, and aims at bringing together researchers/performers from different branches of the department.

The team behind the research concert entitled Sound-Space-Movement (Lyd-rom-bevegelse) included Tone Pernille Østern, Sigurd Saue, Elen Øyen and myself, and was presented at Dokkhuset March 23. The presentation was focused on how bodily interaction with MotionComposer could bring forth interesting movements and reflection from different perspectives. It featured an 8-channel version of the Particles environment of the MotionComposer that I developed in collaboration with Sigurd Saue.

I started off the event with a slightly humourous interactive performance of Richard Strauss' Also Sprach Zarathustra, with Tone copying my movements in the background (see a short excerpt here). I continued with presenting MotionComposer (MC) before I went into more details about my work with the Particles environment of MC, including a demonstration. I especially focused on mapping, demonstrating several possible mappings with data from the MC. Elen followed with an improvised Particles solo, playing with sounds of the Chinese opera singer, Feng Hau. After the performance Elen gave a highly personal talk about her experiences of becoming an dance improvisor using wheel chair. After doing a short Particles improvisation with trombone sounds he himself had played, Sigurd took a more scientific approach for the remainder of his presentation. He gave a very brief introduction to room acoustics, but also talked about convolution-based reverberation and multi-channel spatialization. Here, he demonstrated an orientation sensor he had coupled to a microcontroller and wireless transmitter. The sensor was used in conjunction with the MC-system to control sound spatialization (I had set up vbap8 in csound to do this), reverberation mix and distance. Tone rounded off the individual section with a highly expressive improvised solo using the Materials sound bank of Particles. This was followed by a performative talk where Tone combined verbal reflections around body, space and movement, with her own movements in the room.

Subsequently, the audience were invited to try to do their own improvisations with the MC-system, and several audience members took their chances and tried it out. The whole concert was rounded off by a Q&A session with all four presenters.

Panel at DHN2017 Conference


The 2nd Conference of the association of Digital Humanities in the Nordic Countries (DHN) was held at the University of Gothenburg March 14?16, 2017. Together with a group of researchers from several universities in the Nordic countries I participated in a panel entitled "The Nordic Hub of DARIAH-EU: A DH Ecosystem of Cross-Disciplinary Approaches". The abstract can be read here. The finished version of the panel paper was published in CEUR Workshop proceedings, and can be read here.

Open lecture at Rockheim

Rockheim lecture

As a part of the collaboration between NTNU and Rockheim National Museum for Popular Music I held an open lecture the 16th of February entitled: "The Music Technology of Pop Music". The lecture was a part of the course Music Technology in Light of History, but was open to the general public. The lecture gave an introduction to instruments, sound media and technological tools for musical expression in the period 1950-1975. The slides from the presentation are available here.

Presentation at NTNU Dariah-EU kickoff seminar

As a result of NTNUs membership in the Nordic Hub of DARIAH-EU Nordic Hub (see post below) ARTECs Andrew Perkis and Dariah coordinator at NTNU, Thomas Sørlie Hansen arranged an open kickoff seminar where people interested in digital humanities at NTNU were invited. Researchers from many different fields were represented at the event. I held a short presentation at the event focusing on my work with MotionComposer.