Paper at eLSE conference

After replacing former colleague Anna Xambó as study programme coordinator at the Music, Communication and Technology Programme, I have been heavily involved in the organization and coordination of this cross-campus programme. This has first and foremost included administrative aspects, but also some research activities. This paper was primarily written by my colleague Robin Støckert, who teaches at MCT and coordinates the SALTO (Student Active Learning in a Two-Campus Organization), but with input and feecback from the co-authors, who are all involved in the programme. The paper was presented at the 16th International Scientific Conference eLearning and Software for Education Bucharest, April 23-24, 2020. The paper can be downloaded here.


Higher education is facing disruptive changes in many fields. Students wants to have the option of learning anywhere, anytime and in any format. Universities need to develop and deliver to future students a complete learning ecosystem. At the same time universities are facing challenges such as growing costs and the pressure to give the students the knowledge, competence, skills and ability to continuously adapt to future job environments. As a consequence, many universities are investigating new ways of collaboration and sharing resources to cater to the demands of students, industry and society. An example of this collaboration is a new joint master between the two largest Universities in Norway: University of Oslo (UiO) and Norwegian University of Science and Technology (NTNU). In this paper, we present the lessons learned from almost two years of teaching and learning in the new joint master's programme, "Music, Communication and Technology" (MCT), between NTNU and UiO. This programme is a run in a two-campus learning space built as a two-way, audio-visual, high-quality, low-latency communication channel between the two campuses, called "The Portal". Moreover, MCT is the subject of research for the SALTO (Student Active Learning in a Two campus Organisation) project, where novel techniques in teaching and learning are explored, such as team-based learning (TBL), flipped classroom, and other forms of student active learning. Educational elements in this master, provides the student with 21st century skills and deliver knowledge within humanities, entrepreneurship and technology. We elaborate on the technical, pedagogical and learning space-related challenges toward delivering teaching and learning in these cross-university settings. The paper concludes with a set of strategies that can be used to improve student active learning in different scenarios.



Live Interfaces 2020: Performance paper


The International Conference of Live Interfaces, ICLI, took place March 9-11 and was the second international conference hosted by NTNU Music Technology in only three months. Our conference chair Øyvind Brandtsegg had prepared the stage for a conference with talks, keynotes, concerts/performances and installations. As the conference web page stated, "topics of liveness, immediacy, presence (and tele-presence), mediation, collaboration and timing or flow are engaged with and questioned in order to gain a deeper understanding of the role contemporary media technologies play in human expression".


My contribution at the conference, besides being chair of the installations, was a so-called performance paper, a not very well-known format combining a traditional paper presentation with performance elements. In my paper and presentation, I addressed a topic that I have been occupied with on a hobby basis for a couple of years, namely FM radio. After the shut down of the national broadcasts on the FM-band in Norway in 2017, my frustration with the political descision led me into collecting old radios, especially from the late 1960s and early 70s. I later found out that these radios produces a wide range of noises and sounds when they were connected with transmitters in a feedback loop. The paper contained some of the experiences and reflections I had after experimenting with this setup (see the figure below) during some months.



Abstract: This paper investigates how FM receivers connected to FM transmitter in a feedback loop can be applied as musical instruments, but with an inherent instability that at the same time breaks with traditional instrumentality. The paper begins with delineating the specific historical situation in Norway with the closing of FM band for national broadcasters, and how this could be seen as a call for action. The paper proceeds with a short review of the application of radios in music and sound art and practitioners working with feedback techniques, before presenting relevant theories of nonlinear dynamic processes and feedback systems. The paper then briefly delineates a simple setup of equipment, and goes on to address issues related to playability and sonic output. The paper shows how certain sonic categories are recurrent, but how different models have different sonic signatures, how some features are unstable and difficult to consistently reproduce. The paper concludes by discussing the findings in the light of the research literature in the field, and potential musical applications of the radio as instrument in different setups and contexts.


The full article can be found here. I also extended the content of the paper by including an analysis and categorization of the sounds from one of the radios, the Blaupunkt derby de luxe (image below). This analysis can be seen in the presentation slides found here.



Even though I had some problems with grounds hum during the performance part of the presentation, I believed it showed some of the potential these old devices have for being used as musical instruments. In the wake of the conference, I was also contacted by Knut Aufermann, one of the pioneers of the radio art scene in Germany, giving me some extra links and material, some of which I will hopefully look into in due time.


Web Audio Conference 2019: concert and panel

The international Web Audio Conference 2019 was hosted by NTNU in Trondheim, and the Music Technology Group played a central role in organizing it, with my colleague Anna Xambó as conference chair (and IES colleague Sara Martin by her side). My own role in the organization was demo and poster chair. The conference presented a combination of talks, posters, demos, artworks and performances, and had over 100 participants from all over the world. I also contributed to the conference by participating in a performance and hosting a panel.


The performance was with T-EMP (Trondheim Electroacoustic Music Performance), consisting of Øyvind Brandtsegg, Trond Engum, Anna Xambó and Carl Haakon Waadeland. Here is the description of the performance from the proceedings of the conference:

The ensemble Trondheim Electroacoustic Music Performance (EMP) investigates new modes of communication in an ensemble when new technology is introduced as part of the ensemble repertoire.  In the recent three years, the focus has been on cross-adaptive processing as a musical intervention in the interplay.  More info on the cross-adaptive project can be  seen  at  the  project  blog  and  also  the  album  Poke It With A Stick / Joining The Bots released in 2019.  For this performance we work with a web-based repository of sounds, explored via a live-coding interface. This allows access to a massive archive of sounds from, and the selection of sounds is done via sound descriptors.  The integration of this instrument in an ensemble setting is interesting, as traditional and nontraditional modes of musical interaction are activated in dialogue.  Using web-based access to the repository allows a generality of instrument design, and this is combined with a strategic mode of performance in this instrument.  With?strategic? we mean here, that most sounds are not performed directly by physical action, but cued up in patterns via live-coding.  These sounds are then live processed by other members of the ensemble, responding more directly with physical and gestural instrumental action on the signal coming from the web instrument.  The signal is also live processed in a cross-adaptive fashion, so that audio features extracted from other performers? actions will modulate the parameters of processing for the web audio instrument.  The ensemble also utilize gesturally based interfaces from interactive dance, here used to control elements of audio synthesis and processing.

Here is a video of the performance:


I was asked to host a panel with the topic "accessibility in web audio". I invited three people who I though would be able to address relevant issues of this topic, although none of them were directly engaged with web-audio. David Brüggermann was a former Music Technology student who wrote his bachelor thesis on developing workflows for music production with VoiceOver and LogigPro for a vision impaired user. Monica Ruud is a first-year student at the Bachelor programme in Music Technology NTNU who is blind. Lastly, Miranda Moen is currently currently writing a master's thesis in Equality and Diversity at the NTNU institute of interdisciplinary culture studies. Moreover, she works as a project responsible at AKKS Trondheim and with the KOSO collective. In the panel the participants were first asked to explain how they understand the concept of accessibility.

The two visusually impared students were asked how their engagement with computer-based music technology as a user with vision impairments and if they had  examples of functionality in computer-based music applications which are very much accessible, and others which are not, and if so, in what ways.

The panel was rounded off with questions that was partly addressed by the panelist, and partly by the audience. For instance, how one can raise awareness for issues of accessibility in web audio applications and if there were any concrete suggestions for making web audio applications more accessible. Here is a video of the panel discussion:

VIBRA workshop: 'Constructing Realities' with Johannes Birringer

"Constructed Realities" was a two-day workshop led by Johannes Birringer, as a part of the VIBRA workshop series ( that I organize. Birringer is an independent media choreographer and artistic director of AlienNation Co, and also works as a professor at Brunel University, where he is director of the DAP-Lab and headed the Centre for Contemporary and Digital Performance.

The workshop was organized by me, and sponsored by NTNU ARTEC. VR equipment and expertise was provided by IMTEL lab NTNU. Footage by workshop participants and video editing by Andreas Schille.

At the workshop, fifteen participants with diverse backgrounds, from fine arts, to music technology and dance, worked together to create an immersive environment, including two VR stations, multichannel audio, live visuals affected by audio and video. In addition to organizing and facilitating the workshop, which took up most of my attention, I made a "discoball poem player", using a broken disko ball fittet with a wireless NGIMU-sensor, a recording of a poem read by Dumama, and piece of coding I did in Csound. It can be seen in the video below. The final performance/environment was presented for an audience in Verkstedhallen, Trondheim, October 2019. Workshop participants: Pål Lersveen, Dag Olav Kolltveit, Naomi Chan, Ada Hoel, Samrridhi Kukreja, Emily Aalde, August Norhtug, Gugulethu Duma, Andreas Schille, Mina Paasche, Christopher Logan, Anna Thu Schmidt, Dylan Green, Aleksandar Isailovic, Per Erik Walslag and Ingunn Schumann Mosand.

Here is a video, edited by Andreas Schille, showing some of the different parts of the workshop, including the performance itself:

#MTF: Humans in Da Loop at Örebro University

Two performers with laptops by a table. Two dancers in the foreground. One industrial robot on the side.

Music Tech Fest is an arts festival which presents "technological innovations and artistic experimentation, performance, new inventions, commercial applications, and academic research". It is also a network a network of amazing people working in different branches of music technology. At least once a year they organize #MTFLabs, where they invite "some of the most brilliant and original minds to reinvent the way we interact, collaborate, create and communicate with each other". After particpating in Stockholm in the fall 2018, I hooked up to a small team of fantastic people (Kirsi Mustalahti, Joseph Wilk and Lilian Jap), and together we formed a group called Accents in Motion, or alternatively, Dance AI, building our work around connections between a database of speech accents, motion tracking of dance movement and live coding.

In October 2019 our group was invited to the AI Impact Lab at the University of Örebro as a part of MTF's Humans in da Loop together with a number of people from the MTF community. At the gathering, MTF were "bringing select experts from the global MTF community to address strategic applications of AI" led by their high-profiled professor Amy Loutfi, newly appointed pro-vice-chancellor for AI. The Accents in Motion team were very lucky to be allowed to work in The Cognitive Robotic Systems Laboratory, lead by professor Alessandro Safiotti. Although we were not allowed to play with the industrial robots in the space, we got access to their system for counting people and getting their position in the space. Although the system was more difficult to work with than anticipated, we were able to connect it to our computers and make sound with it.

The functionalities and failures of this system inspired our ideas of the performance - it addressed the way in which technology is often presented to us as black boxes that don't reveal anything about their construction or inner workings, and how we should strive to make technology more transparent and open to users. The failure of the tracking system to locate a person within a 2x2 meter square in the middle of the lab then became our "Black Box", where we then placed our laptops, mixers and "technologist" performers (Joseph and Andreas). The dramaturgy of the performance was then how two of the performers (Kirsi and Lilian) intruded into the Black Box and took over the control of it - which was the conclusion of the performance. Thus, the piece became one of empowerment through knowledge.

Happy performer in wheelchair. Two computers on a table. Two dancers in the background. A screen with computer code on the back wall.

As at the Stockholm MTF we got our songs from an audio database of considerable size, namely DAMP, containing voice-only recordings of karaoke singing. The material in the database was challenging in that it already contained musical form and expression, and to reshape and reinterpret this was not at all easy and required considerable experimentation to get an aesthetically satisfactory result.

And as before, we also had two approaches to our sound database - both movement sensing and live-coding. Even though this piece came together the "hackathon" way - extremely compressed in terms of creation and implementation time, we still think we got to develop our concept and repertoire further. Especially, it was intriguing to play with our roles as "technologists" versus "dancers", and hopefully contribute to a productive dialogue on technology, embodiment, gender stereotypes and empowerment.

Conference presentation: Digitale kulturer

The project Crossadaptive processing as musical intervention (2016-18), firmly rooted in the Music Technology programme at NTNU, was invited to do a presentation at Arts Council Norway's conference "Digital kultur, estetiske praksiser" (Digital culture, aesthetic practices" at Dokkhuset in Trondheim. In the absence of project leader Øyvind Brandtsegg, we were three colleagues from this programme doing the presentation entitled "Å gripe inn i en annens lyd: kryssadaptive modulasjoner" (To interfere with the other's sound: crossadaptive modulations); myself, Trond Engum and Carl Haakon Waadeland. The presentation had two parts; first I gave a brief talk about the central concepts, methods and results of the project, then, Trond and Carl Haakon improvised using the tools from the project.

Carl Haakon Waadeland, Trond Engum and myself at Dokkhuset

Paper at the Sound and Music Computing Conference

SMC Malaga

Together with Sigurd Saue and Pekka Stokke I wrote a paper about VIBRA for the Sound and Music Computing Conference in Malaga.


The paper presents the interactive dance project VIBRA, based on two workshops taking place in 2018. The paper presents the technical solutions applied and discusses artis- tic and expressive experiences. Central to the discussion is how the technical equipment, implementation and map- pings to different media has affected the expressive and experiential reactions of the dancers.

The paper can be found here:

NRK P3 Interview

The NRK P3 show "Urørt" made a short feature about instrumental music February 2019, and I was interviewed to contribute to the feature. The feature with the interview is available here.


PhD supervision course

I attended a two-day course for PhD supervisors at NTNU January 2019 led by Kari Smith, professor at the Department for Teacher Education at NTNU, and leader of the National Research School for Teacher Education. The course qualified me as a "certified supervisor", which from the new regulations of the PhD programme at NTNU is required for at least one of the team of supervisors at a PhD (alternatively, the supervisor has to have supervised earlier).


The course covered many issues relevant for PhD supervision, among them regulations and framework, building relations, gender issues, ethics, co-authorship, submission and what to do in the case of rejection. We also got to discuss some interesting cases, including the typical "burn out" that many experience in the last phase of their PhD period. The course also highlighted the culture differences at NTNU, especially between hard and soft sciences, article based and monograph theses, and small and big departments.

MotionComposer workshop at Dämeritzsee, Berlin

The creative part of the MotionComposer team was gathered for four days at the Dämeritzsee Hotel in December 2018 - lying next to a beautiful lake - for a kind of "hackathon" workshop, aiming to develop the 3.0 version further through long days of work.

The Dämeritzsee, just outside of Berlin

The new tracking software using stereoscopic camera input was now much more developed, and could deliver tracking data of quite good quality for many of the core parameters. For my part, it was the first time I could try the new version of Particles with the new tracking system, and we were happy to experience that the central features of the sound environment worked well. There were some issues with certain parameters being too "trigger-happy", so this is something to be further improved in the tracker.

As for the features of the Particles sound environment, we decided to include arm hits to the side, kicks and jumps to it. I briefly made some sketches with glissanding sound particles that had some interesting qualities, and these will be developed further for the final version, especially giving them a sharper attack. Here are some hits followed by a few kicks:

Sketches for hits and kicks for Particles for MC3.0

I also spent a bit of time preparing a work flow for making rhythms for Particles, including making a sequencer interface in Max that could make a file that could be immediately loaded into csound. Here is a screenshot of the 8 track, 64-step sequencer:

8-track, 64-step sequencer for Particles, built in Max

We concluded the workshop with setting milestones for the further development of Particles in the spring 2019.


Stocos workshop and artist talk

As a part of the ARTEC seminar series, I had invited Stocos, a Spanish/Swiss artist trio creating interactive dance pieces, to do a two-day workshop and an artist talk with demonstration. The trio consists of Pablo Palacio, composer and interactive music instrument designer, Muriel Romero, choreographer and dancer, and Daniel Bisig, visual artist, sensor technician and programmer. The group combines research into embodied interaction, artificial intelligence, and simulation of swarm behavior with artistic exploration of music, visuals and dance.

Stocos (Pablo, front left, Daniel front center, Muriel, far right) with workshop participants

The workshop took place at DansIT's (Dance in Trøndelag) studio at Svartlamon in Trondheim. It took place as a collaboration between ARTEC and DansIT, where the former covered the expenses related to travel and accomodation for the group, and the latter offered me their venue free of charge as a recidency. A total of fourteen participants (including me as the organizer) took part in the workshop, even if some were not present at all times. The group consisted of people with very varying backgrounds: composers, musicians, dancers, actors, media/video artists, stage designers, dramaturgs, philosophers and music technologists. Of the participants, eight of them were either employed or students at NTNU and another three had relatively recently achieved their degrees (Master/PhD) at NTNU and were now working in the art sector in Trondheim. Due to restrictions in the number of people, there were additional applicants to the workshop on a waiting list.

The first workshop day started with a presentation of the workshop holders in Stocos, their artistic aims and statement, and then continued with an overview of their work during the last 10 years or so. A core concept of the group, which is also reflected in their name, is the use of different types of stochastic algorithms, both at signal-level (in the form of noise) and in different ways of mapping movement parameters with visuals and sound. In addition, different ways of applying swarm properties in especially mapping and visual projections were central. This was applied for the first time in the piece with the same name as the group, where the swarms were first of all visible as minimalistic visual projections, mostly in black and white, showing either moving entities or trajectories of invisble moving entities - interacting to larger or lesser degrees with dancers. Another work, Piano-Dancer, was an exploration of the same core principles of stochastics and swarms, but here a single dancer's movements and algorithms were controlling a motorised piano. In the piece Neural Narratives, they were creating virtual extensions of the dancing body with computer visuals, with tree-like entities following the movements of the dancers while also having a semi-autonomous character. Lastly, they presented their most recent work, The marriage between Heaven and Hell, based on the literary work of William Blake. In this work they continued their work with interactive control of voices, this time using phrases from Blake's poem. Particular to this work was the use of Laban derived bodily expressive aspects, such as energy, fluidity, symmetry or impulsivity, translating them (in real time) into audible and visual digital abstractions. This piece also featured the use of body-worn interactive lasers, where the movements of the dancers controlled whether they were active of not. The presentation spurred a lot of questions and discussion about everything from technicalities to aesthetic issues.

The second part of the first workshop day was more technical and focused on the different types of sensors the group used. They started showing technology they have used in earlier years, namely the Kinect. Using the open source driver implemented in EyesWeb the sensor gave them access both to the countour and the skeleton of the dancer. Even if the Kinect had relatively high latency and the skeleton tracking was quite buggy, they still had used it In recent years. More and more, however, they had moved to IMUs for their tracking. They used a Bosch/Adafruit BNO055 connected to an arduino with wi-fi shield, with everything put in a 3D-printed case, and then put in an elastic band - attached to the arm or the lower leg. This avoided several of the problems associated with video-based tracking (occlusion, IR light from other sources, etc.), even if there were specific challenges associated also with the IMUs. The IMUs also offered absolute orientation data in 3D, making it possible to track the position and rotation of the extremities, not only their movements. Finally, the group showed us their most recent development, namely sensor shoes, with seven pressure sensors in the sole.

Daniel Bisig explaining the construction of the sensor shoe.

After just having monitored the values from the sensors in visual graphs, the last part of the first workshop day focused on the mapping on the values from the sensors to sound. Here, the participants got to try out the IMU bands and the sensor shoes with several different mappings - synthetic singing voices, sampled speaking voices, stochastically modulated noise, and several others.

The second day of the workshop started with a movement session led by Romero with focus on exercises related to Laban's theories of individual choreographic spaces around the dancer, and how these could be divided into planes cutting in all directions around a dancer. Many of the exercises directly articulated these planes.

After the movement session, the focus was moved to the computer generated visuals. Most of these were in some way or another derived from swarm simulations, and Daniel Bisig spent quite some time explaining how these were programmed and manipulated so as to create different visual expressions. Among other things we got to see how using the individual agents' traces as visual elements along with the manipulation of pan and zoom of the space they inhabited created a multitude of different patterns and effects. Some of the participants also got to try out an interactive algorithm that used the contour around the body to attract the swarm in such a way that they delineated the user's silhouette. We also got to see how the swarms could articulate letters more or less using the same technique, and how the dampening of attraction forces made the contours around the letters less distinct.

Workshop participant Elena Perez trying out the interactive swarm visuals.

There was also time to try out the interactive lasers (with sound) from The marriage between Heaven and Hell, which was, naturally enough, very popular. All the lights were turned off, and a hazer was activated, so that the trajectories of the lasers were visible. The lasers all had lenses spreading the beam vertically about 30% so that the result was a line when it was projected on the floor. We also got to see how moving the lasers in different ways generated different effects, like "scanning" and "strobe". We then did further tryouts in the darkened space - now with the sensors controlling stage lights (via MIDI/DMX) together with transformed vocal sounds. Even if the lights had a significant latency, it was nice to see how they could be mapped from the IMU sensors.

At the end of the workshop, Stocos presented longer extracts from several of their pieces, to answer the question that many participants had posed: How do you put it all together? We then got a much better understanding of the artistic contexts and intentions the different technologies and mappings were used in - and further how their artistic expressions could be transformed with different techniques.

All in all, the impression was that the workshop participants were highly inspired by the workshop, and several of them talked about getting new artistic ideas. Stocos were also impressed with the engagement that the participants showed. Several people also expressed a wish to have Stocos back to Trondheim to do a full performance.

Artist talk at Kunsthallen

Monday the 26th of November Stocos gave an artist talk at Kunsthallen as a part of the ARTEC seminar series. Around 15 people attended. Stocos presented their artistic aims and statement, and then continued with an overview of their work during the last 10 years or so. In between the discussion of the different pieces, Muriel Romero demonstrated the sensor instruments used in the different performances. The artist talk lasted 90 minutes.

Muriel Romero demonstrating the sensor shoes. Daniel Bisig is watching.

Student concert at Planetariet

As a part of the course Electroacoustic Music - Analysis and Composition that I teach at NTNU, five master's students at the music technology programme got to perform their newly composed pieces at the Planetarium of Vitensenteret (The Science Centre) in Trondheim November 16 2018. The concert took place as the first milestone in several people's joint efforts of transforming the stunning 36.2 channel sound system, custom made for planetarium shows, into a concert venue for multichannel/spatial music. NTNU Music Technology, with myself and PhD candidate Mathieu Lacroix, composer Frank Ekeberg and padagogue at Vitensenteret, Lars Pedersen, have been researching to find a setup that can work for multichannel presentation of electroacoustic music. After quite a bit of experimentation during the fall of 2018, we ended with a solution where we used Blue Ripple Sound's plug-ins to play music in 3rd order ambisonics through the Reaper DAW running on an external laptop connected to the IOsono system via Dante. At the concert, the students presented compositions they had made as a part of the course, but adapted to the 36.2 setup at Planetariet. The concert was concluded with Ekeberg's piece Terra Incognita from 2001. The concert program can be read here. It is my intention to publish a blog post later in the winter with details about the setup to help others arrange multichannel concerts at Planetariet.


Panel at RITMO Workshop

The newly started center for interdisciplinary studies in rhythm, time and motion, RITMO, hosted a workshop to celebrate their opening November 12th-14th 2018: RITMO International Motion Capture Workshop.
I was invited by the chair of the event, Alexander Jensenius (UiO) to participate in a panel about interactive music systems together with John Sullivan (McGill University), Cagri Erdem (UiO) and Charles Martin (UiO). Each of us got to give a short presentation of our work, before we sat down to have a panel discussion, also involving the audience. In my introduction, I presented the particular challenges of creating interactive music systems for people with different abilities in my work with MotionComposer.

In the workshop there were several panels like this, with topics like Music Performance Analysis, Music-Dance Analysis, Music Perception Analysis and Different Types of Motion Sensing. There was also one whole day dedicated to different tools (MoCap Toolbox, MIRToolbox, EyesWeb, and more). High points in the program were the three keynotes by Marc Leman (Ghent University), Marcello Wanderly (McGill University) and Peter Vuust (Aarhus University). We also got a tour of the RITMO center, showing off the motion capture lab, eye tracking lab and more.

Panel about motion capture tools at the RITMO International Motion Capture Workshop

Sound design workshop at Fusion Systems

October 8 - 11 2018 the German company Fusion Systems, a company working mainly with computer vision and sensor fusion for traffic and industry applications, hosted a workshop together with Palindrome Dance Company. The workshop aimed at developing sound design for interactive music applications for people with different abilities. I was invited to the workshop together with Marcello Lusanna and Ives Schachtschabel to work on sound design and meaningful movement-sound mappings. During the workshop we got to try out a work-in-progress prototype of a stereoscopic motion tracking system.

Stereoscopic motion tracking prototype at the Fusion Systems workshop.

Panel at Knowing Music, Musical Knowing

Knowing Music, Musical Knowing was a conference and International Music Research School hosted by my university NTNU and Grieg Research School October 23rd-26th. On the website of the event, it was described in the following manner: "This event will explore the complex and multifarious connections between music and knowing. Bringing together key thinkers and practitioners in contemporary interdisciplinary music studies ? including but not limited to dance studies, music therapy, ethnomusicology, music technology, historical musicology and music education ? this four-day school will inspect both the unique forms of knowledge embodied through musical practice and the different ways humans create and impart knowledge about music."

Myself together with Elin Angelo, Jørgen Langdalen, Simon Gilbertson and Kari Holdhus.

For this event, I was asked to participated in a panel  about the challenges and possibilities for musical knowledge transmission, learning and education between music disciplines. The panel consisted of professor Elin Angelo, Head of Department of Music, NTNU, Jørgen Langdalen, associate professor of music therapy, Simon Gilbertson, associate professor Kari Holdhus from Høgskolen Stord/Haugesund, and post doctor Nora Kulset.

In my five minute introduction I briefly discussed how common denominators in the form of passion, listening and musicking bodies can allow us to build understanding, communicate, as well as improve learning, across disciplines. The whole introduction can be read here.

Music Tech Fest, Stockholm, September 2018

Music Tech Fest (MTF) is an annual event gathering artists, professionals, hackers and academics for a few days to work in an intense and stimulating environment. This time it was KTH in Stockholm that had taken on the event, with the main event happening September 7-9 2018.

I was very happy to be invited to participate in the #MTFLabs, which was an extended part of the whole thing, going on for a full week, Monday - Sunday. In the Labsa total of 80 artists, hackers and acedemics were gathered with a mission of putting on a 45 minute show at the end of the week. Everything took place inside a shut down nuclear reactor 25 meters below ground.

Reaktor hall
The reactor hall at KTH. Photo by Mark Stokes.

The whole thing was kicked off by presentations by Michela Magas, the founder of MTF, Danica Kragic, Airan Berg and pop singer Imogen Heap. Then, during a process of self organization, we split off into different groups and sub groups, according to what we wished to work with in terms of technology and artistic expressions.

I soon found a good connection with Joseph Wilk (UK), a creative (live)coder and computer wiz, and Lilian Jap (SE) a media artist and dancer, and we decided to form a group under the theme of paralanguage. Later, the group was completed with Kirsi Mustalahti (FI), an actress and activists, working with arts for disabled people through the Accessible Arts and Culture (ACCAC) organization.

Me with Lilian Jap, Joseph Wilk and Kirsi Mustalahti

Inspired by the post-apocalyptic feeling in the reactor and theme of extra-terrestial colonization that was circulated in the whole of the #MTFLAbs group, we decided to make a piece about having to re-invent language through paralanguage and body movements. Joseph discovered a really nice speech data base with Enligsh accents, featuring almost 2700 sound files with a single statement being spoken by different speakers from from all over the world. The database was a nice metaphoric expression of "humanity" that we felt was a good fit with our idea. Our initial idea was then to use a wavenet encoder - a machine learning technique working on the audio sample level - to train a model that we could control with the Kirsi and Lilian's movements. However, it turned out that this took a lot more time than we had at our disposal.

With the sample database downloaded we at least had a rich source of sounds, that also had a nice consistency. Joseph was then able to organize the material so he could play it using his live coding setup, and this worked so well that we kept it in the performance for the build-up in the final section.

Without a trained model and an interface to the parameters of the database, it was difficult for me to create an instrument that could play the whole database. I therefore chose to use only a subset of the samples (about 100 of them), that I stretched and layered in different ways using Csound. We put an NGIMU sensor and a Myo sensor armband on each of the dancers, and they were then mapped so as to control each of their individual set of voices. We then structured the piece in two sections, the first one (discovery and first contact) dwelling on two single syllables "plea" (from please) and "call", with each syllable being played by one dancer, and the final part (beginning communication developing into chaos), building up textures and layers of an ever increasing amount of words and syllables played by dancers and Joseph's live coding. The piece ended after about 4:30 with a sudden freeze of movements and silence, whereupon a comprehensible version of the speech material was presented to the audience: "Please call Stella. Ask her to bring these things with her from the store" - a highly mundane statement rounding off the somewhat existential thematics.

Here is a video of the first section of one of our performances (we had three in total). It is shot by Mark Stokes:

We all agreed after the performance that our work together had a lot of potential, both in terms of artistic development and application for people with other abilities, so we have put in several applications for further work with the project called Accents in Motion.

Running version of Six Fantasies Machine


The Six Fantasies Machine (SFM) instrument is a tool developed in Csound for producing a range of sounds similar to those found in Paul Lansky?s classic computer music composition Six Fantasies on a Poem by Thomas Campion. I made it as a part of my PhD project, Experiencing Voices in Electroacoustic Music, which was finished in 2010. Due to changes in Csound, for many years I couldn't run the software on the recent versions, which was a bit sad. However, I have now (October 2018) gone through the code again and got it up and running on Csound 6.12 on OSX.

For download, info and sound examples, check the SFM project page.

Startup of VIBRA - network for interactive dance


After having worked with interactive dance over some years mostly outside of Norway, I found it could be interesting to gather people in my local region to do some work here. I was lucky enough to get a grant from the Norwegian Culture Council, which made it possible to arrange a series of three workshops in Trondheim with dancers/choreographers, video artists and a music tech colleague during the winter/spring of 2018.

The network was named VIBRA, an Norwegian acronym which translates into Visuals - Interaction - Movement - Space - Audio. The project now has its own website at where we have put information about the project and a blog about technical as well as artistic issues.

VIBRA is a network of artists and technologists who explore artistic expression possibilities related to video, interaction, movement, space and audio - in short form, often referred to as interactive dance. With interactive dance, we mean a cross artistic expression in which dancers using sensor and computer technology control one or more aspects of sound, music, video, light, or other scenic instruments. This has hitherto been largely unexplored in Norway. Much of the non-classical dance presented on Norwegian scenes today is accompanied by pre-recorded musical events and / or video. With such a fixed accompaniment, the relationships that occur between motion, music and visual elements will work essentially one way. By using new technology, we hope to introduce an interactivity that can provide a more dynamic expression that more closely puts cross-cultural relationships at stake. The start-up of the network will involve a bundle of people from Trondheim / Trøndelag that have a relevant background and will constitute the project group.

Here is a short excerpt from the showing of the last workshop, using NGIMU and Myo sensors with four dancers and projecting sound over a 8-channel setup:

The video excerpt shows dancers Tone Pernille Østern and Elen Øyen.

Presentation at IRCAM Forum Workshops

IRCAM forum
The IRCAM Forum Workshop is an annual event in which IRCAM Forum members get to share and discuss ideas. As they write on their website, "The Ircam Forum workshops is a unique opportunity to meet the research and development teams as well as computer music designers and artists in residency at IRCAM. During these workshops, we will share news on technologies and expertise providing hands-on classes and demos on sound and music."

As an IRCAM Forum member I made a submission with Robert Wechsler entitled MotionComposer - A Therapy Device that Turns Movement into Music, and it was luckily accepted and the presentation was held the 8th of March 2018. It was a great experience to show our work at this renowned institution in front of a highly competent audience. And of course, we also greatly enjoyed to see presentations of the IRCAM teams as well as the other IRCAM Forum members.

The MotionComposer is a therapy device that turns movement into music. The newest version uses passive stereo-vision motion tracking technology and offers a number of musical environments, each with a different mapping. (Previous versions used a hybrid CMOS/ToF technology). In serving persons of all abilities, we face the challenge to provide the kinesic and musical conditions that afford sonic embodiment, in other words, that give users the impression of hearing their movements and shapes. A successful therapeutic device must, a) have a low entry fee, offering an immediate and strong causal relationship, and b) offer an evocative dance/music experience, to assure motivation and interest over time. To satisfy both these priorities, the musical environment "Particles" uses a mapping in which small discrete movements trigger short, discrete sounds, and larger flowing movements make rich conglomerations of those same sounds, which are then further modified by the shape of the user's body.