dansk musikforskning online / særnummer 2015 danish musicology

Transcription

dansk musikforskning online / særnummer 2015 danish musicology
DANSK MUSIKFORSKNING ONLINE / SÆRNUMMER 2015
DANISH MUSICOLOGY ONLINE / SPECIAL EDITION 2015
LYD- OG MUSIKPRODUKTION
SOUND AND MUSIC PRODUCTION
Forord
3
Preface
5
Toivo Burlin
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap:
Ett produktionsperspektiv på musik
7
Mads Walther-Hansen
Sound Events, Spatiality and Diegesis
– The Creation of Sonic Narratives in Music Productions
29
Totte Mattsson
Audiovisuella loopar och kollektiv live-looping
47
Jan-Olof Gullö, Ivan Höglund, Julia Jonas, Hans Lindetorp,
Anton Näslund, Joakim Persson & Peter Schyborger
Nobel Creations: Producing infinite music for an exhibition
63
Mark Grimshaw
A brief argument for, and summary of, the concept of Sonic Virtuality
81
Justin Christensen
Virtual Distance and Soundstage,
and their Impacts on Experienced Emotional Valence
101
Martin Knakkergaard
Numeriske tåger – eller hvad blev der i grunden af de tre andre causa?
119
Note on the contributors
132
Guest editors:
Martin Knakkergaard
(Department of Communication and Psychology, Music and Sound Knowledge
Group – MaSK, Aalborg University).
Mads Walther-Hansen
(Department of Communication and Psychology, Music and Sound Knowledge
Group – MaSK, Aalborg University).
Editorial board: Martin Knakkergaard, Mads Krogh, Søren Møller Sørensen
Dansk Musikforskning Online Særnummer, 2015: Lyd- og musikproduktion /
Danish Musicology Online Special Edition, 2015: Sound and Music Production
ISSN 1904-237X
© forfatterne og DMO
DMO publiceres på www.danishmusicologyonline.dk
Udgivet med støtte fra Forskningsrådet for Kultur og Kommunikation
Forord
Uddannelse og forskning i lyd- og musikproduktion er i stigende grad spredt ud over
forskellige institutioner, og inden for den sidste årrække er flere nye uddannelser dukket op i Norden, der beskæftiger sig med emnet. Denne opblomstring skyldes ikke
mindst skiftende markedsvilkår, nye teknologier og efterspørgsel på specialiserede
kompetencer i lyd- og musikbranchen. Dette særnummer om musik og lydproduktion
sætter fokus på diversiteten i feltet. Nummeret søger at afdække nogle af de forskellige former for teoretisk tænkning, æstetiske visioner og tilgange til praktisk udførelse, der eksisterer i forskellige forskningsdomæner inden for lyd og musikproduktion,
og hvordan disse griber ind over hinanden. Baggrunden for særnummeret er en konference med titlen “2. nordiske konference om musik og lydproduktion” afholdt på
Musikuddannelsen på Aalborg Universitet 12-13. maj 2015.
Særnummerets syv artikler beskæftiger sig på forskellig vis med udviklingen af analytiske og praktiske værktøjer til lyd og musikproduktioner, og det knyttes til både forskning og undervisning inden for feltet. I den forbindelse præsenteres konkrete bud på
begrebsudvikling – såvel som “begrebsafvikling” – fordret af nye digitale lydteknologier og nye tilgange til musikproduktion. Der gives desuden anvisninger til nye forskningsparadigmer og konkrete anvendelsesmuligheder i undervisningssammenhænge,
herunder kollektive musikproduktionsformer.
Den første artikel af Toivo Burlin aftegner konturerne af musik- og lydproduktionens
videnskabelige objekt. Burlin argumenterer for hvordan musikvidenskaben, i bredeste
forstand, kan drage nytte af et medie- og produktionscentreret forskningsparadigme,
der bygger på forskning udviklet inden for musikproduktionsdomænet.
I Mads Walther-Hansens artikel argumenteres der bl.a. gennem analytiske nedslag for,
hvordan særlige kunstgreb, tilgange og teknikker i musikproduktionen er med til at
fremkalde og understøtte oplevelsen af bestemte virtuelle handlingsrum, som er mættet med tegn, koordinater og information, der er afstemt i forhold til det æstetiskes
potentielle diegetiske implikationer.
Med afsæt i nutidens udstrakte brug af live loop-teknik beskriver Totte Mattson i sin
artikel nogle af de erkendelser og landvindinger, som blev indhøstet i et kunstnerisk
forskningsprojekt om brugen af såkaldt audiovisuelle loop af flere performere kollektivt. Projektets særligt tilrettede tilgange til loop-teknik fremstilles i historisk perspektiv og artiklen anskueliggør, hvordan teknikken er relevant såvel i performance som i
kompositionsarbejde og øvning.
Jan-Oluf Gullö redegør i samarbejde med Ivan Höglund, Julia Jonas, Hans Lindetorp,
Anton Näslund og Joakim Persson om kunstneriske og teknologiske udfordringer og
landvindinger i forbindelse med en kollektivt afstemt musikinstallation, der blev ud-
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
4
viklet til Nobel Museet i Stockholm i 2014. Artiklen baserer bl.a. sin fremstillingen på
kreativitetsteori og tidssvarende forestillinger om proces og flow og argumenterer for,
hvordan en heuristisk tilgang til det kollektive musikskabende arbejde, kan give anledning både til nye udtryk og forståelsesformer.
I Mark Grimshaws artikel argumenteres for en ny definition af lyd, der muliggør nye
måder at forstå, forme og interagere med lyd i fx lyddesign og computerspil. Ved at
betragte lyd som en “emergent perception” der udvikles i hjernen over tid, redegør artiklen for en teoretisk ramme til at forstå, hvorledes disse perceptioner kan udledes fra
hjernen, med det formål at designe audio.
På baggrund af forskning i dyrenes etologi og neurovidenskab fremsætter Justin Christensen hypotesen, at den følelsesmæssige effekt (både positiv og negativ) af musik er
kraftigere når lyden høres tæt på. Denne hypotese bekræftes i et forsøg med EEG og
taktile transducere, og der peges på fremtidig forskning til undersøgelse af forskellige
lydsceners effekt på lytteren.
Den manglende stoflighed i digital lyd er omdrejningspunktet i Martin Knakkergaards
artikel om de ændrede vilkår for musikproduktion digitaliseringen af lyd har medført.
Artiklen bygger sin argumentation bl.a. på en heideggersk teknologiforståelse og redegør ad den vej for, hvordan digitaliseringen synes at afvikle den klassiske opdeling
mellem proces og produkt og samtidig udfordrer selve det klingende musikalske artefakts proveniens.
Gæsteredaktionen vil gerne takke forfattere, peer-reviewere og den faste redaktion for
et godt og konstruktivt samarbejde om dette særnummer.
Martin Knakkergaard og Mads Walther-Hansen
Institut for Kommunikation
Afdeling for Musik
Aalborg Universitet
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Preface
Education and research in music and audio production are increasingly scattered
across different institutions and, within the last few years, several new programs have
emerged in Scandinavia that deal with the subject. This growth is at least a result of
changing market conditions, new technologies, and demands for specialized skills in
the music and audio industries. This special issue on music and audio production
focuses on the diversity in the field. The issue seeks to uncover some of the different forms of theoretical thinking, aesthetic visions, and approaches to the practice
that exist in different research domains in audio and music production and how
they interact with the other. The special issue derives from the 2nd Nordic Conference on Music and Audio Production held at the Music Section at Aalborg University,
12-13 May 2015.
The special issue’s seven articles in different ways deal with the development of analytical and practical tools for music and audio production and they have relevance to
both research and teaching in the field. Within this context are presented specific proposals for concept development – as well as “concept dissolution” – that are driven by
new digital audio technologies and new approaches to music production. Also provided are new research paradigms and concrete applications in education contexts, including collective music productions.
The first article, from Toivo Burlin, delineates the contours of what is music and audio
production’s object of study. Burlin argue how musicology, in the broadest sense, can
benefit from a media- and production-centred research paradigm that is based on research developed in the music production domain.
Mads Walther-Hansen’s article analyses and argues, among other things, for how specific artifice, approaches, and techniques in music production help to create and support
the experience of a certain virtual action space that is saturated with signs, locations,
and information that are each attuned to the aesthetic’s potential diegetic implications.
Basing itself on today’s widespread use of live loop techniques, Totte Mattson’s article
describes some of the discoveries and advances that were developed in an artistic research project regarding the use of so-called audiovisual loops involving several performers collectively. The project’s specifically contextualised approaches to loop technique is presented within a historical perspective and the article illustrates how this
technique is as relevant to performance as it is to composition and rehearsal.
Jan-Oluf Gullo outlines, in collaboration with Ivan Höglund, Julia Jonas, Hans Lindtorp, Anton Näslund, and Joakim Persson, the artistic and technological challenges
and achievements of a collective, synchronized music installation that was developed
for the Nobel Museum in Stockholm in 2014. The article bases its argument on crea-
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
6
tivity theory and contemporary notions of process and flow and argues for how a heuristic approach to the collective, creative music work may lead to new expressions and
as well as to new forms of understanding.
Mark Grimshaw’s article argues for a new definition of sound that enables novel ways
to understand, shape, and interact with sound in, for example, sound design and computer games. By regarding the sound as an “emergent perception” that develops in the
brain over time, the article proposes a theoretical framework for understanding how
these perceptions can be “extracted” from the brain with the aim of designing audio.
Based on research in animal ethology and neuroscience, Justin Christensen proposes the hypothesis that the emotional impact of music (both positive and negative) is
stronger when the sound is heard close to the listener. This hypothesis is supported in
a study using EEG and tactile transducers and identifies future research to investigate
the effect of various acoustic artefacts on the listener.
The lack of texture in digital audio is the focal point of Martin Knakkergaard’s article
on the changed conditions of music production that the digitization of sound has resulted in. The article bases its argument, inter alia, on Heideggerian technological understanding and explains, along the way how, digitization seems to dismantle the classic division between process and product while at the same time challenging the actual
sounding music artefact’s provenance.
The guest editors would like to thank the authors, peer reviewers, and the editorial
team for their good and constructive cooperation in this special issue.
Martin Knakkergaard and Mads Walther-Hansen
Department of Communication
Music Section
Aalborg University
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
TOIVO BURLIN
Experimentell musikproduktionsvetenskap som kulturforskande
musikvetenskap: Ett produktionsperspektiv på musik
Inledning
Föreliggande artikel är en vidareutveckling av några centrala teman i ett föredrag som
hölls vid 2. Nordiske konference om lyd- og musikproduktion i Aalborg 12–13 maj 2015.1
Syftet är att diskutera forskning om musikproduktion inom musikvetenskapen som
disciplin, å ena sidan, och hur en “musikproduktionsvetenskaps” perspektiv, å den andra sidan, kan bidra till att forma ett medie- och produktionscentrerat forskningsparadigm
inom musikvetenskapen.
Artikeln inleds med en historisk översikt på hur produktion av musik har uppfattats och studerats som vetenskapligt objekt: från tidiga diskurser kring fonogram, via
introduktionen av forskning på området inom musikvetenskapliga studier av konstoch populärmusik, till utbildnings- och forskningsdomänen “ljud- och musikproduktion”, tentativt kallat “musikproduktionsvetenskap”, som forskningsdomän. Utifrån
detta presenteras ett förslag på hur det interdisciplinära forskningsfältet, studier av i vid
mening produktion, medialisering och konsumtion av inspelad musik, kan utvecklas
till ett bredare medieparadigm inom musikvetenskapen som disciplin. Det görs genom att, med specifika tyngdpunkter, diskutera hur en musikvetenskaplig teori och
metod/praktik; “produktionsperspektivet” – som, förutom att vara ett bestämt teoretiskt perspektiv, också kan bestå i en experimenterande praktik kring ljud- och musikproduktion där syftet inte primärt är konstnärligt-musikaliskt men vetenskapligt gestaltande – används för att besvara musikvetenskapliga frågeställningar. Detta exemplifieras med en kortare fallstudie. Utgångspunkt för den övergripande diskussionen är ett
metavetenskapligt perspektiv.
1
“2. Nordiske konference om lyd- og musikproduktion” http://www.kommunikation.aau.dk/forskning/vidensgrupper/mask/lmp2015 [access: 2015-09-29]
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
8
Toivo Burlin
Utgångspunkter
Ett av de karakteristiska dragen i det moderna musiksamhället som det ser ut in på
2000-talet, är musiklivets – skapandets, kommunikationens och konsumtionens –
fundamentala grund i teknologi och medier som arkitektur och struktur, en grund
som har blivit symbiotisk under de senaste etthundra åren, men intensifierat under
informationsålderns genombrott, genom allt tätare kopplingar.2 Konstarten och kommunikationsformen musik – om man väljer att betrakta det så – har ingått ett tillsynes permanent äktenskap med medierna och teknologin. Ingen musikform eller genre
står, eller kan stå, utanför denna fundamentala grund: de skillnader i relation till teknologin och medierna som finns mellan genrer och stilar ligger snarast på en kulturell
och musikalisk-stilistisk nivå och har olika historiska och musikestetiska orsaker, men
det är skillnader i grad och inte i art. Teknologi och applikationer för att spela in, bearbeta, sprida och konsumera musik – företrädelsevis i digital form – används nu i
en eller annan tillämpning av snart sagt varje medborgare, i de flesta åldersgrupper:
åtminstone i de nordiska länderna som i detta avseende, både vad gäller utveckling,
produktion, distribution och konsumtion, ligger långt framme globalt sett.3 Inte heller är det så att teknologin och medierna endast direkt angår medialiserad musik: även
levande musik är fullständigt beroende av den. Det gäller också genrer med särskilda
autenticitetsanspråk, som folkmusik och konstmusik.
Inte bara lyssnandet – konsumtionen och receptionen av musik – har blivit mobilt,
utan också produktionen. Mycket musik skapas idag – helt eller delvis – med en mobil
Digital Audio Workstation, ibland i hemmiljö.4 Därtill har inte bara den professionella
ljud- och musikproduktionen tappat mycket av sin en gång mer bestämda geografiska lokalisering.5 Också amatör- och semiprofessionell musikproduktion har påverkats i grunden, genom att avancerade audioteknologiska verktyg och system blivit ekonomiskt och
praktiskt tillgängliga för i stort sett alla som har intresset – något som har inneburit en
oerhörd demokratisering av audioteknologin och medierna. Digitaliseringen av all ljudinformation till likvärdiga och kompatibla format (som WAVE, AIFF) i sinsemellan likaså
2
3
4
5
Manuel Castells, Nätverkssamhällets framväxt. Informationsåldern: Ekonomi, samhälle och kultur band 1
(Göteborg: Daidalos, 1998), Patrik Wikström, The Music Industry: Music in the Cloud (Cambridge: Polity Press, 2009).
Användandet av persondatorer, “smarta” telefoner/läsplattor är generellt högt i de nordiska länderna. Till dessa teknologier är en rad programvaror och applikationer avsedda för produktion eller
konsumtion av ljud knutna. Många av dessa har utvecklats i Norden. Produktionen av datorspel och
applikationer är t ex. viktiga branscher i Sverige och Finland – i Sverige “en viktig del i samhällsekonomin” och “Sveriges största kulturexport genom tiderna”: “Spelutvecklarindex 2015: Baserad
på årsredovisningar från 2014” (Dataspelsbranschen/Swedish Games Industry), 4-5. Världens 2015
största streamingtjänst för musik, Spotify, är svenskt. Norskägda streamingtjänsten Tidal (amerikanskt
från 2015) och WiMP är en annan viktig aktör. Soundcloud är svenskägt, med säte i Berlin. En mindre känd, men viktig bransch, är internationellt kända svenska hård- och mjukvarusynthesizers, som
Clavia: Nord, Propellerhead: Reason, Figure, Teenage Engineering: OP-1, Elektron.
Även om stora, professionella studior förstås fortfarande har en viktig roll att fylla.
Paul Théberge, “The Network Studio: Historical and Technological Paths to a New Ideal in Music
Making” i Social Studies of Science, Vol. 34, No. 5, Special Issue on Sund Studies: New Technologies
and Music (oct., 2004), 759–781.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
9
kompatibla digitala arbetsstationer, som Logic, Cubase och ProTools, med inbyggda funktioner för snabb spridning i “molnet”, förenklar och förstärker arbetsflödena närmast exponentiellt. Detta har gått hand i hand med en hopkoppling och integrering mellan användare och kreatörer i större musiksociala system, något som är helt i linje med informationssamhällets utveckling i övrigt (t. ex i de så kallade “sociala medierna”, “the cloud”).6
Digital audioteknologi och medier för produktion, mångfaldigande och spridning utgör alltså den fundamentala struktur varmed vilket musik produceras, sprids och konsumeras i västvärlden i början på 2000-talet. Hela detta system utgör en utmaning för
musikvetenskapen att tolka och förstå i specifikt musikvetenskapliga termer: hur påverkas
exempelvis musikalisk gestaltning och form, vad betyder begrepp som framförandepraxis
och musikestetik? Hur kan koncept som “musikhistorien”, “kompositören” och “verket”
omtolkas mot bakgrund av denna nya förståelsehorisont? Forskare inom andra musikdiscipliner, som komposition och musikpedagogik, liksom vetenskapsområden som ekonomi och sociologi, har på olika sätt uppmärksammat områdets vikt, men de insatserna är spridda och inte alltid i kontakt – därtill utgår de i det senare fallet av förklarliga
skäl inte från specifikt musikaliska villkor och problem, som musikvetenskap och andra
musikdiscipliner lättare får syn på. Men många icke musikvetenskapliga perspektiv på det
moderna musiksamhällets sociala, teknologiska och ekonomiska dynamik kan med fördel integreras i musikvetenskapens kärnområde – utifrån en interdisciplinär dialog.
I det följande kommer det argumenteras för det relevanta i att utveckla ett medieoch produktionscentrerat musikvetenskapligt forskningsparadigm, som på allvar utgår från
situationen efter sekelskiftet 2000 (men också de senaste 100 åren). Det kan göras i en
vetenskaplig relation med forskning inom andra discipliner som studerar produktion
av musik och medialisering, konsumtion och reception av inspelad och producerad
musik. Detta kan betraktas som ett interdisciplinärt område och behöver nödvändigtvis inte utvecklas till en separat disciplin. Det interdisciplinära forskningsfältet bör dock
principiellt skiljas från begreppet “ljud- och musikproduktion” som är en beteckning
ad hoc för ett utbildningsområde i första hand.7
En musikvetenskap om musikproduktion, medier och medialisering
Musikvetenskapen har som en traditionsstark humanistisk disciplin ett rikt förgrenat nätverk av teoretiska och musikaliskt-praktiska verktyg till sitt förfogande som –
med vissa modifikationer8 – gör den potentiellt väl disponerad att använda i studier
6
7
8
Patrik Wikström, The Music Industry: Music in the Cloud (Cambridge: Polity Press, 2009).
Gunnar Ternhag, “Musikproduktion – ett nytt ämne i högre utbildning” i Gunnar Ternhag och Johnny Wingstedt, ed., På tal om musikproduktion: Elva bidrag till ett nytt kunskapsområde (Göteborg: Bo
Ejeby förlag 2012), 79.
Musikvetenskapliga analysverktyg och metoder är ofta utvecklade för att appliceras på noterad musik
(som harmonisk analys, Schenker-analys). Vid lyssning till inspelad musik har man inte en notbild
tillhanda, utan analyser förutsätter en högre grad av strukturellt lyssnande där inte bara harmoniska,
melodiska och rytmiska förlopp bör tas i beaktande, utan också t ex. spatialitet, timbre, produktionsaspekter. Möjligen kan exempelvis en modifierad Schenker-analys (“Schenkerian Analysis”) tillämpas på inspelad musik.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
10
Toivo Burlin
av inspelad och medialiserad musik, trots att den hittills ägnats rätt lite åt detta. Till
skillnad från andra humanistiska och samhällsvetenskapliga discipliner, har musikvetenskap musiken (i någon form) som sitt primära vetenskapliga objekt och har också verktyg utvecklade för ändamålet, som olika slags musikanalys. Men dessa verktyg
har oftast utvecklats för att användas på en noterad musik i första hand och får en annan ontologisk status när de appliceras på inspelningar.
Musikvetenskapen bör därför omvandlas inifrån: bli en vetenskap som i grunden
– i både teoretisk och metodologisk mening – absorberar och förstår de senaste etthundra årens genomgripande och revolutionerande utveckling av produktion och medialisering av musiken. Nya teorier och metoder är i det perspektivet nödvändiga. Den
historiska och estetiska tradition som musikvetenskapen sprungit ur behöver, utifrån
detta perspektiv, inte enbart vara en intellektuell begränsning utan kan potentiellt bidra med inspiration till ny kunskap och nya perspektiv på musikstilistisk, musikteknologisk och medial utveckling, som kan vara svårare att utveckla i andra discipliner.
Det gäller sådant som den musikanalytiska, stilhistoriska och organologiska kunskapsbasen inom musikvetenskapen, som här är central. Däremot bör den musikvetenskapliga förklaringsmodellen om musikhistorien som nyckel till förståelse av förändringar i
musikalisk stil, form och funktion, i grunden omformuleras, när det gäller musikproduktionens utveckling: musikhistorien som koncept har en deterministisk kärna. Häri
finns en svårighet, eftersom idén om den historiska utvecklingen är en fundamental
humanvetenskaplig tankemodell. Den behöver heller inte upplösas, men kombineras
med mer synkrona perspektiv och också hämta inspiration från helt andra vetenskapsområden. I detta hänseende vill jag öppet förorda ett annat slags musik(forsknings)
perspektiv än de båda sinsemellan oförenliga, men tydligt humanistiska ytterlighetspositioner som, å ena sidan, hävdar att musik främst är kultur i en antropologisk, etnologisk eller Cultural Studies-mening, å andra sidan att musik “endast är musik” och
därför först och främst bör studeras och förstås som sådan, som musikalisk struktur.9
En musikvetenskap som kan ta in perspektiv och kunskapsområden med annat ursprung än den traditionella musikutbildningens, humaniora och samhällsvetenskaperna – som de tekniska vetenskaperna (teknologin är en viktig, i hög grad determinerande, komponent i musikproduktionen) eller naturvetenskaperna (exempelvis akustik
och psykoakustik är viktiga hjälpvetenskaper för att, också i humanvetenskaplig mening förstå, musikens spatiala och temporala aspekter, som kultur och struktur) – med
ett omfattande av det fundamentala musikteknologiska och mediala sammanhanget i
hela dess oerhörda komplexitet, kan utvecklas till ett sådant medie- och produktionscentrerat musikvetenskapligt forskningsparadigm. Detta paradigm bör undvika att
hamna i liknande reduktionistiska tankemönster som delar av det äldre musikvetenskapliga paradigmet odlat: att ersätta tonsättaren med producenten och verket med det
fonografiska verket är möjligen lockande, men inte en fruktbar utveckling om det stannar där. Det nya paradigmets perspektiv på musik bör bli mer komplexa; eftersträva
en annan förståelse av innebörden i begrepp (och fenomen) som musik, produktion,
9
Dessa slags ytterlighetspositioner framkom, bland annat, i debatten om musikvetenskapens framtid
i STM 2005–2008.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
11
medier. Det räcker inte att utgå från enkla, närmast binära begreppspar som “musik
och medier”, som två separata storheter som ibland ingår i en relation (till exempel:
“musiken på CD-skivan”). Tvärtom måste förståelsen av mediernas betydelse, för musik som konst och kommunikation, inkorporeras inom en mer holistisk förståelse av
begreppet och särskilt fenomenet musik.10 Det finns – utan att här vilja formulera något
anspråk om en enda lösning för musikvetenskapen – ändå goda skäl till att förändra vissa förgivet tagna synsätt: det gäller särskilt den grundläggande musikvetenskapliga värderingen av teknologin och medierna som något principiellt skilt från musiken
(och därmed som “ointressant”). Det gäller också mer allmänt humanistiska synsätt
på teknologin som något som framför allt, och i huvudsak, är laddat med kulturella
betydelser och koder – istället för att också se teknologin som den kraftfulla förlängning av människans instrumentella färdigheter och tankeförmåga som den är – och
som därför i hög grad bestämmer villkoren för vad som kan tänkas och skapas. De
senaste etthundra årens utveckling visar på en mycket nära symbiotisk relation mellan
musik, teknologi och medier, och på teknologins djupgående påverkan på kulturen.
Forskning om musikproduktion: historia och diskurser
Det finns ingen etablerad vetenskaplig disciplin som på ett direkt sätt utforskar produktionen av musik och ljud. Kunskapsområdet är interdisciplinärt till sin karaktär
och inbegriper bland annat musik (flera deldicipliner), ingenjörskonst, elektronik,
teknik och akustik – men även aspekter av organologi (läran om musikinstrumentens konstruktion och akustiska egenskaper), teknikhistoria, medie- och kommunikationsvetenskap liksom estetik i vid mening. Under det senaste decenniet när ljud- och
musikproduktionen blivit datoriserad och digitaliserad, kan man säga att datavetenskap, i tillämpad form, blivit en stöddisciplin. Även andra stöddiscipliner kan förstås
förekomma. På det hela taget finns dock en nära relation till produktionspraxis: hantverket, “den tysta kunskapen”, är mycket viktig, ja helt central, i sammanhanget.
Både i Norden, Storbritannien, Tyskland och USA har kunskap om produktionsmetoder under 1900-talet i hög grad lärts ut just i en praxisbaserad, hantverksorienterad tradition: inlärning och socialisation inom systemet har skett genom att man arbetat med inspelning och bearbetning av musik och ljud inom musikindustrin eller
filmindustrin, radion eller televisionen – på senare år också inom spelproduktionen.11
I begränsad omfattning har det sedan 1950-talet funnits högre utbildningar, ofta vid
radiostationer, där ett fåtal specialister har utbildats: mest känt är tonmeister-utbildningen i Tyskland. I Sverige fanns vid Radiotjänst (Sveriges Radio) från 1948 och efter
förebild från BBC en musikteknisk internutbildning, som fordrade både musikalisk
10
11
En del försök har gjorts inom teoribildningen kring intermedialitet, utan att helt övertyga: framför
allt troligen beroende på en högt utvecklad teoretisk abstraktionsnivå, där teorin tycks sakna verifiering gentemot en empirisk verklighet.
Jan-Olof Gullö, Musikproduktion med föränderliga verktyg. Diss. (Stockholm: Skrifter från Centrum för
musikpedagogisk forskning (MPC) KHM Förlaget, 2010). Toivo Burlin, Det imaginära rummet: Inspelningspraxis och produktion av konstmusikfonogram i Sverige 1925–1983 Diss. (Göteborg: Skrifter från
musikvetenskap, 2008), 243–343.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
12
Toivo Burlin
och teknologisk kompetens – men utan systematisk forskning.12 Audio Engineering
Society, (AES), bör nämnas i sammanhanget. Detta sällskap är en av de äldsta vetenskapligt verksamma branschorganisationerna för ljudtekniker och musikproducenter
(grundat i USA 1948) och de publicerar regelbundet både en vetenskaplig journal, The
Journal of the Audio Engineering Society och har återkommande konferenser. Organisationen har också nordiska sektioner. Forskningen som utförs där tillkommer i huvudsak medlemmarna.13
Hur förhåller sig musikvetenskapen till detta kunskaps- och forskningsområde?
Inom de traditionella musikdisciplinerna – som musikvetenskapen – lämnades i ett
historiskt perspektiv produktionen av musik länge utanför; det uppfattades ofta ha
mer med ingenjörsvetenskap än med “de sköna konsterna” att göra. I modern tid kan
man däremot säga att det beror på vad man avser med musikvetenskap. Musikvetenskapens inriktning varierar mellan länder och forskningsinriktningar – med svårighet
låter sig disciplinen enkelt summeras på annat sätt än att den utifrån olika perspektiv
har “musik” som vetenskapligt “objekt”. Fragmentiseringen mellan subdiscipliner och
inriktningar är förhållandevis stor för ett litet ämne. Här är heller inte platsen att redogöra för denna bredd eller de historiska processer som lett fram till den. Men studier
av musikproduktion är idag ändå mer än en tendens inom disciplinen.
Vetenskapshistoriskt sett, så har forskningsfältet “ljud- och musikproduktion” och
“musik och medier” inom musikvetenskap växt fram interdisciplinärt i relation till
musiketnologi, sociologi, studier av populärmusik (“Popular Music”), medie- och
kulturstudier (Cultural Studies) men också direkt inom musikvetenskapen.14 Musikantropologin och -etnologin är de musikvetenskapliga subdiscipliner som först började använda sig av inspelningsteknologi för vetenskapliga ändamål och som därmed
först noterade inspelningens särskilda egenskaper som representationsform, särskilt
egenskapen att både timbre, spel- och sångsätt samt stil kan sparas med hög grad
av exakthet. Musik dokumenterades då ofta i utomeuropeiska kulturer, men även i
Europa och USA, i första hand som en musikalisk praktik, en praxis och som en del i
en större kultur – inte i egenskap av ett fixerat musikaliskt verk.15
Forskningsfältet “ljud- och musikproduktion” har också ursprung i studier av
konstmusikens interpretationspraxis. Redan på 1930-talet utgavs av kommersiella
grammofonbolag stora europeiska och nationella fonogramantologier (främst på
78-varvsskivor i boxar) som syftade till att omfatta och auditivt presentera den västerländska konstmusikens historia: redan i diskursen om dessa inspelningar framträdde ett reflexivt drag kring den musikaliska interpretationens historiska utveckling.16
Under 1900-talets lopp blev det efterhand än mer uppenbart för flera musikhistoriker
12
13
14
15
16
Burlin, Det imaginära rummet.
Audio Engineering Society http://www.aes.org, http://www.aes-sweden.se [access: 2015-08-30]
Läsaren hänvisas till forskningsöversikten i Georgina Born, “Afterword: Recording: From reproduction to representation to remediation” i The Cambridge Companion to Recorded Music ed. Nicholas
Cook et. al. (Cambridge: Cambridge University Press 2009), 286–304, 325.
Bruno Nettl, “Ethnomusicology: Definitions, Directions, and Problems” i Musics of Many Cultures: an
Introduction, ed. Elisabeth May (University of California Press 1980), 1–9.
Burlin, Det imaginära rummet, 74–75, 127–177.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
13
att äldre inspelningar med konstmusik konserverade och representerade äldre, historiskt överspelade, interpretationsstilar, exempelvis inom violinspelet. På så sätt blev de
(ofta kommersiellt producerade) musikinspelningarna källor till att förstå sentida förändringar i musikalisk interpretationspraxis. Därmed bidrog de också till insikten att
stilistiska förändringar kan gå snabbt, vilket får konsekvenser för epistemologin kring
äldre musik. Vad kan man egentligen veta om hur det lät, när musik framfördes under
barocken, kunde vara ett sådant problem.
De tidigaste kända musikvetenskapliga undersökningarna av interpretationspraxis
utifrån inspelningar gjordes redan 1932, 1950 och 1960.17 Ett större musikvetenskapligt pionjärarbete som tog fasta på detta perspektiv var den brittiske musikforskaren
Robert Philips Some Changes in Style of Orchestral Playing, 1920–1950.18 Den studien
representerar ett första musikvetenskapligt medieparadigm, där det blev möjligt att
studera inspelningar av musik som ett primärt källmaterial – tidigare ett i det närmaste otänkbart perspektiv.19 Philip följde sedan upp detta arbete med ytterligare ett par
stora studier: Early Recordings and Musical Style och Performing Music in the Age of Recording, där perspektivet vidareutvecklades i olika riktningar.20 Robert Philip har sedan
fått flera efterföljare. Inom särskilt brittisk forskning om konstmusik finns, förutom
Timothy Day och Daniel Leech-Wilkinson, det flera forskare som gjort studier på området.21 Nicholas Cook har i sammanhanget hävdat att inspelningar, särskilt när det
gäller konstmusik, främst är källor till framföranden och därtill relaterad interpretationspraxis, snarare än de är att betrakta som en separat verkkategori. Det är ett problemkomplex som av andra forskare, inspirerade av en diskurs om musikinspelningen
som en separat verkkategori med särskild ontologisk status, teorin om “the Work of
Phonography” och “fonografins tidsålder”, i studier av inspelad konstmusik, fått radikalt annorlunda tolkningar.22
Den amerikanska riktningen “New Musicology”, som först formulerades av Joseph
Kerman på 1980-talet23 utvecklades främst i en vetenskaplig polemik med modernismen,
17
18
19
20
21
22
23
Daniel Leech-Wilkinson, “Recordings and histories of performance style” i The Cambridge Companion
to Recorded Music ed. Nicholas Cook et. al. (Cambridge: Cambridge University Press 2009), 321.
Robert Philip, Some Changes in Style of Orchestral Playing, 1920–1950, as shown by gramophone recordings. Diss. (Cambridge: Cambridge University Press, 1975).
Robert Philip, Performing Music in the Age of Recording (Nev Haven, Conn: Yale University Press,
2004).
Robert Philip, Early Recordings and Musical Style: Changing Tastes in Instrumental Performance 1900–
1950 (Cambridge: Cambridge University Press, 1992), Philip, Performing Music, José A Bowen,
“Finding the music in musicology: Performance history and musical works” in Rethinking music, ed.
Nicholas Cook et. al (Oxford: Oxford University Press, 2001), 424–452.
T ex. Martin Elste, Dorotoya Fabian, Michael Musgrave, Bernhard Sherman, Daniel Leech-Wilkinson,
“Recordings and histories of performance style” i The Cambridge Companion to Recorded Music ed.
Nicholas Cook et. al. (Cambridge University Press 2009), 246–262. “Timothy Day, A Century of Recorded Music: Listening to Musical History (Yale: Yale University Press, 2000).
Nicholas Cook, “Methods for analyzing recordings” i The Cambridge Companion to Recorded Music
ed. Nicholas Cook et. al. (Cambridge University Press 2009), 221-245. Nordiska studier på området
med detta perspektiv är Tore Simonsen, Det klassiske fonogram. Diss. (Oslo: Norges musikhøgskole
Oslo, 2008), Burlin, Det imaginära rummet.
Joseph Kerman, Contemplating Music: Challenges to Musicology (Harvard: Harvard University Press,
1986).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
14
Toivo Burlin
med tysk formalism och positivism (som ofta blandades ihop). New Musicology öppnade dörren för att betrakta konstmusiken – det var den som man fortfarande i huvudsak
intresserade sig för – utifrån nya teoretiska perspektiv, exempelvis genom att anknyta till
teorier från litteraturkritik, sociologi, antropologi och genusvetenskap. En viss öppenhet
mot “Popular Music” liksom musiketnologin ingick också. Något större intresse för medier och musikteknologi visade man dock inte. Även om riktningens inflytande utanför
USA möjligen har varit begränsat, har flera av paradigmets ivrigast förfäktade idéer sedan
ändå blivit en del av musikvetenskapens normalvetenskapliga stadium – för att tala med
Thomas Kuhn – på så sätt att det, åtminstone i nordisk musikvetenskap, numera torde
tillhöra de verkliga ovanligheterna att forskare arbetar uteslutande utifrån formalistiska
eller positivistiska utgångspunkter. Att kulturen och samhället på något sätt bör ingå i en
analys av musik är inte kontroversiellt, men däremot motsatsen.24
Inom populärmusikforskningen, som växte fram under 1970- och 80-talen, problematiserades inte heller musikteknologins och mediernas betydelse för den musikaliska
gestaltningen i någon stor omfattning – detta trots att populärmusik i alla former var
fundamentalt och uppenbart beroende av båda dessa storheter. Förutom några viktiga och inflytelserika undantag – Philip Taggs studie över musiken till TV-serien Kojak,
Kojak: 50 Seconds of Television Music: Towards the Analysis of Affect in Popular Music25, där
musikteknologi och medieformer noggrant analyseras och presenteras i genomtänkta (men idag något daterade) modeller är ett exempel – låg fokus i 1970-talets populärmusikforskning inte vid musikteknologin specifikt. Men det växte gradvis fram en
medvetenhet om att såväl teknologi och medier hade haft stor betydelse för populärmusikens framväxt och utveckling. Dock dröjde det för forskarsamhället att uppmärksamma musikinspelningens specifika egenskaper som representation av musik, liksom
att börja rätt uppskatta musikteknologins och mediernas avgörande betydelse för både
musikalisk gestaltning och kommunikation. Från det sena 1970-talet till det tidiga
1990-talet formerades en mer sammanhållen musikvetenskaplig diskurs kring inspelad musik – huvudsakligen i den anglosaxiska världen. I slutet av 1990-talet och kring
sekelskiftet 2000 utkristalliserades diskursen till något av en embryonal subdisciplin
– eller möjligen snarare som en embryonal “subinterdiscipline”, som Georgina Born
kallar det – inte bara relaterad till musikvetenskapen, utan också till andra vetenskapliga discipliner, som antropologi och ett tvärvetenskapligt område som Cultural Studies.26 En betydande interdisciplinaritet finns sålunda med som grundläggande utgångspunkt och också som förutsättning för fältet. Under 2000-talets första år formerades dessa olika vetenskapliga initiativ på mer medvetna sätt.
“The Art of Record Production” (ARP) var ett av dessa initiativ: det är i inskränkt
mening en konferensserie och ett forskarnätverk, som sedan 2005 är relativt stort
inom fältet i de anglosaxiska länderna – särskilt i Storbritannien, där det etablerades,
24
25
26
Alastair Williams, Constructing Musicology (Ashgate, 2001), 2–7.
Philip Tagg, Kojak – 50 seconds of television music: Towards the analysis of affect in popular music (Göteborg: Musikvetenskapliga institutionen, Göteborgs universitet, 1979).
Georgina Born, “Afterword: Recording: From reproduction to representation to remediation” i The
Cambridge Companion to Recorded Music ed. Nicholas Cook et. al. (Cambridge University Press 2009),
286–304.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
15
men också i USA och Kanada. Ursprungligen startades det som ett konferenssamarbete mellan Music Tank, ett nätverk för den brittiska musikindustrin, CHARM (The
AHRC Research Centre for the History and Analysis of Recorded Music) och några
brittiska universitet. CHARM – lett av Nicholas Cook – hade parallellt sin egen forskningsverksamhet och ett flertal projekt som i hög grad var inriktade på att tillgängliggöra och undersöka äldre inspelningar som inte längre var upphovsrättskyddade.
Efterhand utkristalliserades konferensserien “The Art of Record Production” som en
separat verksamhet och med en modern produktionsprofil. Organisatörer och affischnamn har Simon Zagorski-Thomas och Simon Frith varit. I en modern, huvudsakligen
brittisk populärmusikvetenskaplig kontext, är CHARM och “The Art of Record Production” också en del av ett distinkt brittiskt musikvetenskapligt forskningsfält med ursprung i studier av “Popular Music”, “Popular Musicology”, delvis i nära samverkan
med brittisk musikindustri – en industri som från sekelskiftets dramatiska förändringar i medielandskapet har varit i behov av en akademisk legitimering.27
Inom forskningsfältet finns det en anglosaxisk slagsida vad gäller dominerande aktörer och fokus på anglosaxisk populärmusik – även om bidrag om exempelvis folkeller konstmusik från andra delar av världen inte saknats helt. Forskningsfältet har utvecklat en mer självständig akademisk status inte minst genom Art of Record Productions årliga internationellt turnerande konferens, som har en vetenskaplig tidskrift online, samt forum.28 Flera antologier, monografier och artiklar har utgivits med anknytning till detta växande fält. Några bör nämnas: CHARM:s The Cambridge Companion to
Recorded Music29, The Art of Record Productions The Art of Record Production: An Introductory Reader for a New Academic Field30 och den metavetenskapliga studien The Musicology of Record Production.31 Även om nätverket har blivit dominerande i den musikproduktionsvetenskapliga diskursen finns det likartade diskurser och nätverk utanför detta,
i exempelvis USA och i de nordiska länderna. Nämnas kan i sammanhanget också antologin Recorded Music: Performance, Culture and Technology32, som delvis liknar antologierna från CHARM och ARP. Men här formuleras också det tentativa disciplinnamnet
“phonomusicology” som syftande på en vetenskaplig disciplin vars huvudsyfte är att
studera inspelad musik, inklusive produktionskontexter och konsumtionsmönster.33
27
28
29
30
31
32
33
“Art of Record Production”, Home http://www.artofrecordproduction.com [access: 2015-08-28],
Toivo Burlin, “Art of Record Production, London 17–18 september 2005” [Konferensrapport], STM
Online vol. 9 (2006), http://musikforskning.se/stmonline/vol_9/burlin/index.php?menu=3 [access:
2015-08-31], Toivo Burlin, “The Art of Record Production” [recension] STM (2014), 1.
“Art of Record Production”, Home http://www.artofrecordproduction.com [access: 2015-08-28], Burlin, “Art of Record Production, Burlin, “The Art of Record Production” [recension] STM (2014), 1.
Ed. Nicholas Cook et. al, The Cambridge Companion to Recorded Music (Cambridge: Cambridge University Press, 2009).
ed. Simon Frith och Simon Zagorski-Thomas, The Art of Record Production: An Introductory Reader for a
New Academic Field (Farnham/Burlington: Ashgate, 2012)
Simon Zagorski-Thomas, The Musicology of Record Production (Cambridge: Cambridge University
Press, 2014).
Ed. Amanda Bayley, Recorded Music: Performance, Culture and Technology (Cambridge: Cambridge
University Press 2010).
Bayley, Recorded Music.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
16
Toivo Burlin
Från musikvetenskap till musikproduktionsvetenskap – och tillbaka
Efter denna mycket generella översikt på hur forskning kring inspelad musik har utvecklats, ska jag nu gå över till en diskussion om hur jag ser på relationen mellan
musikvetenskaplig forskning om musikteknologi och medier, å ena sidan, och “musikproduktionsvetenskap” – för att därefter skissera hur detta kan utvecklas. Något måste
först sägas om begreppet musikproduktionsvetenskap och mitt val att använda det i denna diskussion – och då kommer jag in på den specifikt svenska situationen på området, som har inspirerat till detta resonemang. I Sverige har musikvetenskapen sedan
det blev ett universitetsämne i Uppsala 1948 varit en liten och i både forsknings- som
utbildningssyn fragmentiserad disciplin: institutionellt oftast avskild till universiteten,
lite vid sidan av övriga professionella musikutbildningar – till musiker, musiklärare,
tonsättare – som funnits och fortfarande finns vid musikhögskolorna. Samtidigt har
det sedan början av 1980-talet vid de regionala högskolorna växt fram kvalificerade
utbildningar i ljud- och musikproduktion, som tillgodosett både reella utbildningsbehov – bland annat har den svenska musik- och spelindustrins internationella framgångar under perioden gett “ringar på vattnet” – som bland annat påtagligt ökat intresset för ljud- och musikproduktionsfrågor och systematisk utbildning till professionell
kompetens på området. Samtidigt har detta, uppenbarligen stora, utbildningsbehov
inte tillgodosetts av musikvetenskapen eller musikhögskolorna. I båda har kompetensområdet hamnat “mellan stolarna”. Men de nya utbildningarna i Ljud- och musikproduktion har haft fokus på att utbilda i de rent instrumentella och teknologiska färdigheterna, att förbereda för en komplex och osäker arbetsmarknad. Forskningsanknytningen har kommit i andra hand. Men ur ett antal forskningsaktiva lärares initiativ
etablerades från början av 2000-talet olika nätverk med strategier för att öka forskningen på området, liksom knyta samman den som redan fanns – från 2002 och framåt i
ett litet antal svenska doktorsavhandlingar med produktions-, medie- och musiktekniska perspektiv, från olika lärosäten och inom skilda discipliner som musikteknik,
musikvetenskap och musikpedagogik. Förutom en hel del artiklar av dessa forskare,
publicerade i olika tidskrifter, bör i sammanhanget den svenskspråkiga antologin På
tal om musikproduktion: Elva bidrag till ett nytt kunskapsområde nämnas – ett av de första gemensamma forskningsinitiativ som togs – med bidrag av svenska, norska och
finländska forskare på området, men verksamma inom flera discipliner.34 Antologins
bidrag står sig väl jämfört med de engelskspråkiga antologierna. I högre grad än dem
diskuterar På tal om musikproduktion på en metanivå också ljud- och musikproduktion
som en akademisk domän (ett vetenskapsområde som delar vissa gemensamma grundläggande begrepp och metoder, men utan att vara en färdig disciplin). Gunnar Ternhag
menar, med vissa reservationer, att namnet “musikproduktionsvetenskap” är ett adekvat namn på denna forskningsdomän, som ännu inte utvecklats till en fullvärdig disciplin (och kanske heller inte kommer göra det). Ternhag menar att man, bland annat
för att påvisa domänens forskningsstatus, hellre bör tala om musikproduktionsvetenskap
34
Ed. Gunnar Ternhag och Johnny Wingstedt, På tal om musikproduktion: Elva bidrag till ett nytt kunskapsområde (Göteborg: Bo Ejeby förlag 2012).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
17
än om Ljud- och musikproduktion, som låter som namnet på ett utbildningsämne. Jag
väljer att anknyta till begreppet musikproduktionsvetenskap för att markera att det i
detta sammanhang syftar på just ett interdisciplinärt forskningsfält.35
Det svenska antologiprojektet är nära relaterat till annan nordisk forskning inom
musikvetenskap men också i angränsande musikdiscipliner som musikteknik, komposition och musikpedagogik. Svenska avhandlingar inom området, exempelvis, med
utgångspunkt i olika vetenskapliga traditioner, är Berg 2002 (musikteknik)36, Burlin
2008 (musikvetenskap)37, Wingstedt 2008 (musikpedagogik)38, Gullö 2010 (musikpedagogik)39, Florén 2010 (sociologi)40, Sjölin 2011 (komposition)41, Nilsson 2011
(komposition)42, Andersson 2012 (musikvetenskap).43 Ovanstående svenska doktorsavhandlingar spänner från musiktekniska undersökningar, med exempelvis lyssningstester som metod, via studier av produktionspraxis i ett musikhistoriskt perspektiv,
från ämnesdidaktiska studier till konstnärligt-musikaliska undersökningar. Här finns
också studier som undersöker musikindustrins utveckling – ett forskningsområde som
också har växt i Sverige.44 Några doktorsavhandlingar inom musikvetenskap i Oslo har
också ett explicit produktionsperspektiv, men utifrån en musikvetenskaplig teoribildning.45 Även i Danmark har det producerats forskning om musik som har ett implicit eller explicit musikproduktionsperspektiv.46 Den forskning som totalt produceras
inom domänen “musikproduktionsvetenskap” – och dit hör även viss forskning om
ljud och musik i spel, liksom musikindustrin – finns alltså idag inom ett flertal discipliner. Ibland är forskarna möjligen inte medvetna om tämligen likartad forskning
35
36
37
38
39
40
41
42
43
44
45
46
Gunnar Ternhag, “Musikproduktion – ett nytt ämne i högre utbildning” i Gunnar Ternhag och Johnny Wingstedt, ed., På tal om musikproduktion: Elva bidrag till ett nytt kunskapsområde, (Göteborg: Bo
Ejeby förlag 2012), 75–89.
Jan Berg, Systematic evaluation of perceived spatial quality in surround sound systems. Diss. (Luleå: Luleå
University of Technology, 2002).
Toivo Burlin, Det imaginära rummet: Inspelningspraxis och produktion av konstmusikfonogram i Sverige
1925–1983. Diss. (Göteborg: Skrifter från musikvetenskap, 2008).
Johnny Wingstedt, Making music mean: on functions of and knowledge about narrative music in multimedia. Diss. (Piteå: Department of music and media, Luleå University of Technology, 2008).
Jan-Olof Gullö, Musikproduktion med föränderliga verktyg Diss. (Stockholm: Skrifter från Centrum för
musikpedagogisk forskning (MPC) KHM Förlaget, 2010).
Thomas Florén, Talangfabriken: om organisation av kunskap och kreativitet i skivindustrin. Diss. (Stockholm: Acta Universitatis Stockholmiensis, Stockholm studies in sociology, 2010).
Anders Sjölin, Plastic modes of listening: affordance in constructed sound Environments. Diss. (London:
School of Media, Arts and Design University of Westminster, 2011).
Per-Anders Nilsson, A field of possibilities: designing and playing digital musical instruments. Diss. (Göteborg: ArtMonitor, 2011).
Anders-Petter Andersson, Interaktiv musikkomposition. Diss. (Göteborg: Skrifter från musikvetenskap,
Göteborgs universitet, 2012).
Nätverket Mirac: (Music Focused Interdisciplinary Reserach & Analysis Center) www.mirac.se [access:
2015-10-24].
Tore Simonsen, Det klassiske fonogram. Diss. (Oslo: Norges musikhøgskole Oslo, 2008), Hans T.
Zeiner-Henriksen, The “PoumTchak” Pattern: Correspondences Between Rhythm, Sound, and Movement in
Electronic Dance Music. Diss. (Department of Musicology Faculty of Humanities University of Oslo,
2010).
Martin Knakkergaard, Mads Walther-Hansen, Steven Gelineck, Morten Michelsen och Ulrik Schmidt
bör nämnas i sammanhanget. Se även Mads Walther-Hansen, Sound Events, Spatiality and Diegesis –
The Creation of Sonic Narratives in Music Productions. Sida 29-46 i detta specialnummer.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
18
Toivo Burlin
inom andra discipliner, något som avhjälps genom etablerandet av forskarnätverk och
konferenser. Forskningsfältet är alltså interdisciplinärt och inbegriper bland annat tekniska, konstnärliga, pedagogiska och historiska perspektiv.
Metodologiskt så kan musikvetenskaplig forskning berikas från de andra disciplinerna, med såväl nya teoretiska perspektiv som de praktisk-instrumentella tillvägagångssätt som finns inom området. Ljud- och musikproduktioner av olika slag kan
användas som en metod och ett material för att besvara humanistiska (musikvetenskapliga) frågeställningar. Inom musikvetenskaplig metodik har detta en direkt anknytning till musiketnologins dokumentära metoder. Inom musikproduktion i praktiken (och inom aspekter av “musikproduktionsvetenskap”) kan man förvisso producera dokumentära inspelningar, men ett konstnärligt och tekniskt avancerat gestaltande
av musik är också en metod, och oftast ett mål. I det vidare forskningsfältet är – till
skillnad från i musikvetenskapen – relationen till moderna musikproduktionsverktyg
central. För att nå förståelse på djupet av ett musikproduktionsvetenskapligt problem
kan man där knyta ihop en vetenskaplig metadiskussion om produktionens villkor,
med att göra en produktion i praktiken, där olika teknologiska problem – som mixning och akustik – kan testas i olika programvaror. Forskaren kan få större insikt om
man alternerar mellan att arbeta med produktionen och utveckla den teoretiska reflektionen om produktionen. Detta leder till nödvändigheten av experimentella och
reflexiva förhållningssätt. Frågor om exempelvis historiska och teknologiska villkor för
musikproduktion kan i många fall prövas experimentellt med produktionsverktygen.
Forskningsområdet “musik och teknologi” eller “musik och medier” utgör annars fortfarande en liten specialitet inom nordisk musikvetenskap. Såväl i Danmark,
Sverige, Finland och Norge, är ansatserna till forskningsprojekt och teoriutveckling
inom det här berörda området få (det gäller för övrigt också i exempelvis Tyskland
och Frankrike). Med tanke på att den musikteknologiska och musikmediala utvecklingen i Norden nära följt övriga västvärlden, är det ett faktum att området inte har
ägnats större uppmärksamhet än vad det gjorts. Det har skapat ett kunskapsunderskott
och ett i vissa avseenden skevt musikvetenskapligt paradigm, som kommer till uttryck
i tendenser mot betoning av musikvetenskap som ett i teoretisk mening väsentligen
historiskt och i metodologiskt hänseende närmast filologiskt ämne; musikvetenskaplig
forskning befattar sig inte så mycket med medialiserad musik. Här finns det skäl till en
bredare omprövning av musikvetenskapens humanistiska status. De senaste etthundra
årens utveckling på musikens område är så fundamentalt omvälvande i ett mänskligt
och evolutionärt perspektiv, att musik före 1877 bör studeras med en särskild medvetenhet om den nutida förståelsehorisontens gränser.
I exempelvis Sverige har musikvetenskaplig forskning som helhet under 2000-talet
fortfarande haft viss slagsida mot konstmusik och också mot vad som måste kallas
för en tonsättarcentrerad forskning.47 I den polariserade svenska debatten om musik47
Ursula Geisler et al., “Svensk musikforskning idag: En självkritisk bild av vår vetenskap” STM vol. 91,
(2009), 77-132. Ett tidigt försök att vidga den – till synes närmast hermetiskt slutna – disciplinen
mot andra områden, var Henrik Karlssons artikel: Henrik Karlsson, “Ljudvärldar: Nya forskningsfält
för musikforskare”, STM vol. 82, (2000), 21–40.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
19
vetenskapens framtid som fördes från 2005 och framåt (huvudsakligen i Svensk Tidskrift
för Musikforskning, STM, 2005–2008, även i Danish Yearbook of Musicology) framkom synpunkter och ställningstaganden som nu, med några års distans, framträder skarpare. Det
var av forskare inom den riktning där samhälls- och kulturaspekter tillmäts plats i musikaliska analyser, som medier och musikteknologi omnämndes som viktiga faktorer att
undersöka i relation till musik.48 Debattens två avdelningar initierades av Lars Lilliestam
2005 respektive av Tobias Pontara 2008. I den första diskuterades (den svenska) musikvetenskapens inriktning och framtid. Debatten synliggjorde viktiga ideologiska och
vetenskapsteoretiska skiljelinjer. I den andra fördes en metadiskussion kring hur man
förstår olika musikvetenskapliga diskurser. I detta sammanhang är den första intressantast, men jag vill även kort anknyta till den andra. I den första var det särskilt Alf Arvidsson som pekade på att musikvetenskapen behöver göra sin huvudsakliga vetenskapliga
omprövning mot bakgrund av det sociala och kulturella sammanhang som all musik i
vår tid befinner sig i: det senmoderna massmediesamhället.49 I den andra formulerade
Tobias Pontara en diskussion om musikvetenskapliga språkspel och diskurser.50 I Pontaras kritiskt diskursanalytiska perspektiv analyserades bland annat idén om en motsättning mellan det musikaliska “verket” i sig, kontra musiken i världen – en diskussion
med ursprung i New Musicology 51 och en kontradiktion som snarare bottnar i ett spel
med språket än något som självklart kan hittas i en skapande musikalisk praktik.
I produktionsprocessen, när man arbetar med produktion av musik i någon form
av musikstudio, eller i receptionsprocessen av musik, när en lyssnare senare tar del
av inspelningen, förbinds naturligt “världen” och “verket”. I produktionen av en inspelning tas det beslut av en mängd olika slag – musikaliska, tekniska, kommersiella
– som har en direkt relation till “världen”, är världen, utanför det “verk” som är i vardande. I musikvetenskapliga studier av musikaliskt skapande – som det tar sig uttryck
inom modern ljud- och musikproduktion – kan man alltså komma förbi sådana vetenskapsteoretiska dikotomier och samtidigt ställa mer relevanta och verklighetsnära
frågor om hur musik skapas och tolkas. När processen, snarare än det färdiga verket som
en sluten entitet står i forskarens fokus, förbinds på ett självklart sätt det musikaliska,
sociala, kulturella, historiska och teknologiska sammanhanget med det som blir “verket”. “Verket” kan i sammanhanget möjligen översättas med det fonografiska verket, den
verkkategori som ur ett perspektiv är relevantast att använda för alla musikinspelningar
de senaste etthundra åren och därmed för huvuddelen av den musik som har skapats
under perioden.52 Detta förhållningssätt, att fokusera på processen snarare än enbart
produkten, anknyter också till Alan P. Merriams definition av musik, som ett socialt
48
49
50
51
52
STM Online 2005–2008. Debattörer var Lars Lilliestam, 2005, Alf Arvidsson, 2005, Dan Olsson,
2005, Holger Larsen, 2005, Mattias Lundberg, 2006, Tobias Pettersson, 2007, Lars Lilliestam, 2007,
Tobias Pontara, 2008, Lars Berglund, 2008, Lars Lilliestam, 2008, Mattias Lundberg, 2008, Eyolf Østrem, 2008, Tobias Pontara, 2008.
Alf Arvidsson, “Svar på: Vad gör vi med musikvetenskapen?” STM Online, vol. 8, (2005).
Lars Berglund och Eyolf Østrem, “Music and the Wor(l)d – Musicology and Mankind. How We Got
Out of Analysis, and How to Get Back In”, STM, vol. 83, (2001), 13–26.
Berglund och Østrem, 13–26. Berglunds och Østrems artikel anknyter tydligt till New Musicology.
Burlin, Det imaginära rummet, 79–125.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
20
Toivo Burlin
inlärt och konceptualiserat fenomen.53 Detta etnologiskt och antropologiskt influerade
perspektiv – musik som en socialt och kulturellt bestämd praktik i första hand – leder
vidare till “produktionsperspektivet” som jag utvecklar nedan.
Debatten i STM avslutades 2008 men har fått en sentida – mer konstruktiv – fortsättning i en artikel om musik och intermedialitet, signerad Mats Arvidson. Även Arvidson tar viss utgångspunkt i STM-debatten. Eftersom den representerar ett nytt perspektiv
inom musikvetenskapen som samtidigt är relaterat till det område som här diskuteras,
vill jag kort kommentera några av dess teser.54 I korthet presenteras de nya perspektiv
som Intermediala studier, med ursprung i litteraturvetenskap och Cultural Studies, erbjuder musikvetenskapen. Ursprungligen kallades forskningsfältet “Interarts Studies”
och tonvikten låg då vid att undersöka relationerna mellan konstarter inom en konstart.
Fältet har under 2000-talet orienterats mot beteckningen “Intermedial Studies”, med ett
större fokus vid “medier” (definierade på olika sätt) och medierelationer. 55 Syftet i Arvidsons text är att överskrida skillnaden mellan en verkcentrerad musikvetenskap och
etnologiskt och sociologiskt inspirerad musikforskning, samt formulera ett förslag till
hur en ny humanvetenskaplig disciplin som tar vår tids avancerade mediekultur på allvar skulle kunna se ut. Fältet “Intermedial Studies” erbjuder många relevanta perspektiv
och utmaningar för en musikproduktionsvetenskap i vardande. Perspektiven utgår väsentligen från analyser av färdiga verk, inte produktionsprocesserna bakom.
Ett implicit problem i det intermediala teorikomplexet, för den forskare som vill
studera produktionen av musik med utgångspunkt i intermediala perspektiv, är ursprunget i en litteraturvetenskaplig teoribildning – med fokus på texter – som inte är
omedelbart tillämpbar på inspelat ljud och musik. Den intermediala mediediskursen
ger än så länge bitvis fler frågor än svar.56 Det har exempelvis, vilket Arvidson refererar, av litteraturvetaren Lars Elleström gjorts lovvärda försök att utvidga mediebegreppet bortom deterministiska tekniska definitioner, men det ger en modell som är oklar
när det gäller den taxonomiska skillnaden mellan “konstarter” och “medier” (vilket
också har en motsvarighet i omdefinitionen från “Interarts” till “Intermedial”). Konstarter och medier är i Elleströms modell, enligt Arvidson, inte synonyma begrepp. Men
konstarter, som musik och litteratur, är specifika former av medier; “kvalificerade medier”.57 Inte desto mindre så ligger i det intermediala perspektivets grundläggande problemformulering att konstarterna och deras yttringar (som musikaliska eller fonografiska “verk”) aldrig är solitära utan alltid ingår i komplexa relationer med andra konstarter och medier. Separata konstarter, enligt 1800-talets estetiska modeller, existerar
alltså enligt intermedial teori inte – inte heller existerar diskreta medier. Arvidson,
som i hög grad hänvisar till amerikanska forskare inom New Musicology, menar att
53
54
55
56
57
Alan P. Merriam, The Anthropology of Music (Northwestern University Press, 1964).
Mats Arvidson, “Music and Musicology in the Light of Intermediality and Intermedial Studies”, STM
Online vol. 15 (2012).
Jfr Jens Arvidsons analys av fonogramkonvolut (“Record Sleeves”): Jens Arvidsson, “The Record
Sleeve and its Intermedial Status” i Changing Borders: Contemporary Positions in Intermediality (Lund:
Intermedia Studies Press, 2007), 55–83.
Mats Arvidson, “Music and Musicology.”
Arvidson, “Music and Musicology.”
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
21
flera av dem i själva verket arbetat inom ett intermedialt fält – utan att själva beteckna
det så. Han menar vidare att Intermedial Studies – om det utvecklas interdisciplinärt
och inte multidisciplinärt, som han menar att musikvetenskapen idag har utvecklats
– kan bli “the future foundational discipline of the humanities”. Graden av utveckling
av det interdisciplinära är nyckel till om ett fält (eller domän med mina ord) kan utvecklas till en ny disciplin eller inte.58
I detta sammanhang behöver det kort skisseras hur begreppet “medier” definieras
och förstås här, i denna text. Det finns många modeller och perspektiv för att definiera
och förstå mediebegreppet – som alla har sina specifika för- och nackdelar. I just detta
sammanhang är det viktigaste att jag inte diskuterar alla medier, utan endast de som
betraktas som “ljudmedier”; i det ansluter jag till Jonathan Sternes pragmatiska definition av medier som “sociala teknologier”: ett medium är helt enkelt en eller flera teknologier, som genom sociala praktiker kommer att fungera och uppfattas som ett medium.59 Man kan exempelvis tänka på radion, fonogrammet eller interaktiva programvaror på Internet som Spotify – teknologier som fått tydligt definierade sociala praktiker knutna till sig. Min medieuppfattning är också påverkad av Friedrich Kittlers syn på
teknologier och medier som ytterst strukturerande för verklighetsuppfattningen – även
om detta perspektiv, som kan uppfattas som deterministiskt, kan behöva nyanseras.
Men humanistisk forskning om musik och medier lider knappast av för mycket
teknisk determinism – utan snarare av en brist på insikt om de teknologiska produktionsvillkorens betydelse för kulturen. De teknologier som använts och används i den
musikaliska produktionsprocessen har i hög grad påverkat musikens gestaltning i produktionen; tydligt i exempel som den snabba framväxten av crooning kort efter den
elektriska inspelningsteknikens genombrott 1925, den nära relationen mellan genrer
som hard bop och progressive rock och LP-skivans duration, relationen mellan mixningskonventioner och bandspelarens antal kanaler, eller relationen mellan 2000-talets diversifierade semi- och professionella musikproduktion och den interaktiva, kompletta
studioarkitektur som erbjuds i Cubase, Nuendo, ProTools eller Logic – å andra sidan
så påverkar också användaren av mediet hur det används och därmed utvecklas. Men
teknologier och medier inverkar, enligt Kittler, i hög grad styrande på hur konstarter
och deras produkter gestaltas – snarare än tvärtom.60
Jag har i det föregående rört mig mellan musikvetenskapen som disciplin och musikproduktionsvetenskapen som domän, med syftet att både beskriva forskningsfältets
plats inom musikvetenskapen, hur framväxten av musikproduktionsvetenskapen gått
till, liksom redogjort för dess nuvarande akademiska position i förhållande till musikvetenskapen. I detta är det dock inte de akademiska positioneringarna som är viktiga
eller intressanta i sig, utan den nya kunskap som kan skapas när musikvetenskapen
– som en etablerad humanistisk disciplin – möter en domän med ljud- och musik58
59
60
Arvidson, “Music and Musicology.”
Jonathan Sterne, The Audible Past: Cultural Origins of Sound Reproduction (London: Duke University
Press, 2003).
Friedrich Kittler, Maskinskrifter: Essäer om medier och litteratur. Mediehistoriskt bibliotek 1 (Gråbo:
Anthropos, 2003).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
22
Toivo Burlin
produktion som fokus, men med de disciplinära rötterna i, förutom produktionspraxis, musikteknik, elektronik, akustik, pedagogik, sociologi och komposition – för
att nämna några av de områden som förekommer som disciplinära utgångspunkter.
Domänen musikproduktionsvetenskap utforskar alltså – men ofta med teoretiskt
brokig grund – ett område av fundamental vikt i de senaste etthundra årens musikliv.
Därför kan det också bli en central del av musikvetenskapens moderna kunskapsbildning. Musikvetenskapen behöver vidareutveckla ett forskningsparadigm som tar frågor om produktion, medialisering och reception av inspelad musik på allvar. Det kan,
så ser argumentationen ut, ske i en samverkan med musikproduktionsvetenskapen,
men också med andra discipliner, domäner och teoretiska perspektiv. I detta argumenterar jag alltså på ett annat sätt än Mats Arvidson, i hans diskussion om relationen
mellan musikvetenskap och Intermedial Studies: jag menar istället att det är bättre
för musikvetenskapen om den dynamiskt utvecklas inifrån – och i dialog med andra
discipliner och domäner – än att väsentliga kunskapsområden lämnas utanför för att
istället hänvisas till discipliner i vardande. Musikvetenskaplig forskning ska inte retirera till redan väl upptrampade kunskapsstigar, utan istället erövra nya områden för
kunskap om musik.
Produktionsperspektivet
Det interdisciplinära forskningsfältet som här avses, studier av i vid mening produktion,
medialisering och konsumtion av inspelad musik, kan utvecklas till ett bredare, medieoch produktionscentrerat, musikvetenskapligt forskningsparadigm. Jag har antytt några olika linjer som detta kan göras efter, men jag ska i det följande sammanfatta dem.
Mot bakgrund av det föregående sagda skall jag utveckla en diskussion om hur metoder från musikproduktion kan implementeras och användas i en musikvetenskaplig
praktik, och på så sätt bidra till en utveckling av båda fälten. Inte bara musikvetenskapen behöver en nyorientering utan det bör idealt samverka med en mer genomgripande utveckling av musikproduktion som verksamhet, från en löst sammanhållen domän till ett mer sammanhållet vetenskapsområde – en musikproduktionsvetenskap. Det
är inte ett mål att etablera en ny disciplin, utan domänen, som den skisserats, kommer
utvecklas i en eller annan riktning ändå. Det vore dock önskvärt, enligt detta sätt att
se, om aspekter av det musikproduktionsvetenskapliga perspektivet bidrog till att utveckla ett medie- och produktionscentrerat musikvetenskapligt forskningsparadigm – vilket,
det ska omedelbart sägas, jag inte tror kommer utvecklas till något slutgiltigt system
för att förstå musik, som konstart och kommunikationsform, utan snarare, i bästa fall,
blir ett öppet och produktivt paradigm med potential att formulera relevanta frågeställningar om de senaste etthundra årens musik. Därtill med ett nytt metodologiskt
perspektiv – “produktionsperspektivet”. Vad avses då med detta?
Betydelsen av musikteknologisk kunskap och praxis, i arbetet i studion, för den
musikaliska gestaltningens slutliga form, är något som har påvisats i flera studier.
Även om det förstås ligger nära till hands för musikforskaren att först upptäcka det
musikaliska objektet – musikinspelningen, det fonografiska verket eller fonogrammet
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
23
som ett taktilt musikaliskt, kanske ett multi- eller intermedialt estetiskt objekt – så står
möjligen, i första hand, den inspelade musikens “heliga graal” att finna i själva produktionsprocessen: i det komplexa nätverket av musikaliska, personella och teknologiska relationer när något blir till musik, blir fångat, bevarat och kommunicerat på ett
medium. Att till fullo studera och försöka förstå de intrikata och komplexa aspekterna
av både musikaliskt, teknologiskt och kulturellt slag i produktionsprocessen – av vilken slags musik eller ljud som helst – är en musikvetenskaplig utmaning, men också
något som öppnar upp för en potentiellt ny förståelse av hur musik skapats och skapas i de 19:e och 20:e århundradena, och kanske tidigare än så.
“Produktionsperspektivet” har utvecklats ut ur produktionspraxis, musikpedagogisk verksamhet och musikproduktionsvetenskaplig forskning. Det är här avsett att vara
en del i ett medie- och produktionscentrerat musikvetenskapligt forskningsparadigm –
som är tillämpbart för inspelad musik, men också för musik som inte har denna representationsform. I en schematisk och tentativ form kan det ställas upp i ett antal punkter:
Definition av “produktionsperspektivet”
– Musikskapandet – i form av musikproduktion av något slag – är en mänsklig, social
och teknologisk verksamhet.
– I produktionen är gränsen mellan “musik” och “ljud” flytande (såväl i produktionspraxis och i filosofisk mening).
– I produktionen är gränsen mellan “externa” kulturella och teknologiska processer
och det musikaliska “verket” i vardande, flytande.
– Att se teknologins – produktionens – mediernas betydelse för all musik – även där den
inte är avsedd att märkas.
– Genre- och kulturöverskridande: perspektivet berör alla genrer, stilar och musikkulturer.
Musikvetenskaplig perspektivvändning
– Definition: Produktionsperspektivet utgår från människans interaktion med andra
människor, naturen, samhället/kulturen, teknologin, medierna (“livsvärlden”) för att
skapa, kommunicera och sprida musik.
– Definition: Musik/ljud och musik/teknologi hör alltid nära samman. Teknologierna
(till exempel musikinstrument, musikstudion) växlar – men de är grundläggande
för musikens existens. Teknologierna är en förutsättning för skapandet men är inte
enkelt determinerande.
– Definition: skapande av (inspelad) musik är samtidigt en process och en skapad/producerad produkt, samtidigt ett historiskt (diakront) fenomen och ett synkront fenomen, något med både kultur, samhälle och teknologisk utvecklingsnivå förbundet.
– Perspektivet flyttas: från noterade “verk” till musikskapande processer och (i kulturell mening “instabila”) auditiva produkter (något annat än “verk”).
– Perspektivet flyttas: all musik inkluderas – oavsett genre, kultur, produktionsmetod,
teknologisk nivå.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
24
Toivo Burlin
– Perspektivet flyttas: Produktionsperspektivet inkluderar och överskrider potentiellt
etablerade musikvetenskapliga subdiscipliner och traditioner.
– Perspektivet flyttas: Experimentell metodologi med ursprung i musikproduktionens praktik. Produktion av auditiva eller musikaliska artefakter kan användas som
en metod för att simulera olika produktionsformer eller musikaliska/auditiva miljöer (med vetenskapliga eller musikaliskt-konstnärliga syften).
Tre forskningsfält som ligger nära produktionsperspektivet, utan att sammanfalla med
detta, kan nämnas här, utan att jag gör anspråk på att beskriva dem komplett eller uttömmande: “Sensory history”, “Sound Studies”61 och nämnda “Intermedial Studies”.
Det första området – som kanske kan översättas med sinnesförnimmelsernas historia –
har producerat ett flertal intressanta perspektiv med ljudet som fokus under 2000-talets
andra decennium, exempelvis David Hendy’s Noise: A Human History of Sound and
Listening 62, Hillel Schwartz’s Making Noise: From Babel to Big Bang and Beyond 63 och
Veit Erlmanns Reason and Resonance: A History of Modern Aurality.64 I denna litteratur
ligger tonvikten endast ibland vid att hitta nya perspektiv på ljudet som både akustiskt och kulturellt fenomen, liksom människans auditiva relation till världen, samt
kontextualisera och problematisera begrepp som “musik”, “ljud” och “oljud”. Forskningsområdet “Sound Studies” ligger nära “Sensory history”, men har huvudfokus på
ljudens kulturhistoria i vid mening. Det är en i forskning omfattande tvärvetenskaplig
disciplin, som ligger nära både delar av musikvetenskapen liksom musikproduktionsvetenskapen liksom flera andra discipliner, med forskare som Karin Bijsterveld, Michel
Chion, Jonathan Sterne, Richard Leppert, R. Murray Shafer, Gilles Deleuze & Felix
Guattari, Jaqcues Attali, Mark Katz, Paul Théberge och Lisa Gitelman – för att nämna
några namn. Delar av Sound Studies vetter mer åt “Science and Technology Studies”.65
“Produktionsperspektivet” kan sägas vara ett bestämt musikvetenskapligt teoretiskt perspektiv – och en metod. Enligt detta – här tentativt formulerade – perspektiv,
studeras och betraktas inspelad musik (ytterst även icke inspelad musik) som samtidigt en process och en skapad/producerad produkt, som samtidigt är ett historiskt (diakront) fenomen och ett synkront, med både kulturen, samhället och den teknologiska
utvecklingsnivån djupt förbundet fenomen. Fokus är flyttat från noterade “verk” till
musikskapande processer och (i kulturell mening “instabila”) auditiva produkter. Perspektivet inkluderar all slags musik – oavsett genre, kultur och produktionsmetod och
teknologisk nivå. Produktionsperspektivet inkluderar och överskrider – potentiellt –
de etablerade musikvetenskapliga forskningstraditionerna, som exempelvis musikhistoria, musiketnologi och populärmusikforskning. Det upphäver också den principiella gränsen mellan studier av professionell respektive amatörproducerad musik.
61
62
63
64
65
Trevor Pinch & Karin Bijsterveld, The Oxford Handbook of Sound Studies (Oxford: Oxford Handbooks
in Music, 2004).
David Hendy, Noise: A Human History of Sound and Listening (BBC: Profile Books, 2013).
Hillel Schwartz, Making Noise: From Babel to Big Bang and Beyond (Cambridge: MIT Press/Zone
Books, 2011).
Veit Erlmann, Reason and Resonance: A History of Modern Aurality (New York: Zone Books, 2010).
Pinch & Bijsterveld, The Oxford Handbook.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
25
Ytterst upphävs också föreställda skillnader mellan en modern musikproduktionsteknologi och förindustriella musikteknologier, eftersom “teknologi” och “medium” inte
per definition är knutna till någon särskild epok.
Produktionsperspektivet befinner sig alltså som tentativt musikvetenskapligt perspektiv nära de perspektiv som utvecklats inom Musicology of Record Production,
Sensory Studies, som studerar ljudet som ett akustiskt och kulturellt fenomen och människans auditiva relation till världen och som kontextualiserar och problematiserar begrepp som “musik”, “ljud” och “oljud”; Sound studies (“Science and Technology Studies”) samt Intermedial Studies. Samtliga dessa vetenskapliga perspektiv kan, på olika
sätt, helt eller delvis inkluderas i eller relateras till produktionsperspektivet. Perspektivets gränser och relationer till de övriga områdena kan dock endast prövas i praktiken.
Liten fallstudie: producerad simulering av en folktrons auditiva miljö
Jag ska i det följande skissera en fallstudie, ett verkligt projekt66, som befinner sig mellan musikvetenskap/musiketnologi och musikproduktionsvetenskap – och diskutera
hur det kan omsättas med produktionsperspektivet. I den meningen är det ett “experimentellt” projekt och mitt exempel, min fallstudie eller “case”, i denna artikel. Ett musiketnologiskt problem omsätts här alltså i ljud- och musikproduktioner, som sedan används som ett särskilt slags källmaterial. Bakgrunden till projektet skisseras här mycket
översiktligt; tyngdpunkten ligger i hur produktionsperspektivet kan appliceras och därmed öppna för en diskussion om hur metoder från musikproduktionsvetenskap kan
prövas inom musikvetenskapen och därmed bidra till att utveckla ett medieparadigm.
I det äldre nordsvenska bondesamhället, som gradvis försvann från 1950-talet
och framåt, fanns en utbredd föreställning i folktron om ett naturväsen – vitra – som
bland många andra egenskaper sades ha en mycket vacker musik. Det finns ett antal
bevarade uppteckningar av vitras musik, som den uppfattades, och ett stort antal upptecknade beskrivningar av vitras ljud- och musikmiljöer, liksom ett mycket stort antal
inspelningar med memorat, inklusive musik, gjorda från 1940-talet och framåt.67
Tron på vitra var utbredd i norra Norrland och koncentrerad till dess centrala delar.
Generellt handlar föreställningarna om vitra om ett underjordiskt folk, som är mycket
lika människorna både i utseenden, beteenden och i levnadsvillkor, som lever parallellt med människorna i en delvis osynlig värld, i skogen, under jorden, sjöarna och
i bergen. De är människolika och har en interaktion med människorna, men är ändå
uppfattade som osynliga eller andliga väsen; dock uppfattas de som levande i den meningen att de föds, åldras och dör. De lever också i familjer. Vitra – särskilt kvinnorna
– beskrivs oftast som extraordinärt vackra varelser med mycket stora blå ögon. De kan
vara extremt långa, eller mycket småväxta, men oftast påminner de påtagligt om människor även i storleken.
66
67
Pågående projekt av Toivo Burlin (projektledare). Preliminär projekttitel: “Saivos sång”. Projektet utgår från ett folkloristiskt material och tillämpar etnologiska, musikvetenskapliga och “musikproduktionsvetenskapliga” metoder. Projektet presenterades vid Linnéuniversitet, Växjö, 28 maj 2015.
Dialekt- och folkminnesarkivet i Uppsala (DFU).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
26
Toivo Burlin
I ett stort antal memorat skymtar föreställningar om vitras musik fram. Beskrivningarna av musiken har det gemensamt att den beskrivs som särskilt vacker och annorlunda, som i vitterkornas koskällor som är stämda i olika toner; “så[da]na obegripliga skäller”, som en informant formulerar det. Det finns ett antal vallåtar, spelmanslåtar och jojkar som har hävdats komma från vitra och som nedtecknats som sådana:
som en låt “efter” vitra. Det finns beskrivningar av karaktären hos denna musik – med
beståndsdelar som kromatik eller annorlunda klanger. Utöver de memorat som beskriver vitras musik, finns det också många källor som vittnar om andra auditiva fenomen som röster och koskällor, som inte direkt kunde knytas till mänskliga ljudmiljöer
och som därför associerades med vitra.68 Vitras exceptionella musikalitet understryks i
flera memorat. Musiken sade man sig ha hört i naturen, och till det är ett nära relaterat, svårgripbart kulturgeografiskt problem kopplat: föreställningen om vitervägarna eller viterstigarna, de färdvägar som vitra sades använda sig av. Det reser frågan vad detta
auditiva fenomen i ett förmodernt samhälle, i brytpunkten till det moderna – som
alltså också har en spatial, rumslig dimension – hade för kulturell betydelse.69 Detta
ska dock inte besvaras här, utan låt mig raskt gå över till diskussionen om metodologi
och teoretiskt perspektiv.
Det bedöms vara en praktiskt genomförbar uppgift att utveckla ett antal studioinspelningar, gjorda i stereo och surround (exempelvis i 5.1) och i dem gestalta auditiva och musikaliska simuleringar av den auditiva och musikaliska informationen i memoraten om vitra. Beskrivningarna i memoraten är ofta så pass detaljerade vad gäller
beskrivningar av vilka ljud man hört och i vilken riktning de har kommit – till och med
enstaka kompletta melodier har upptecknats – att det med audioteknologi och lite fantasi är genomförbart att simulera detta. Eftersom det handlar om ljudvärldar som åtminstone delvis får antas vara av psykologiskt slag, snarare än ett “verkligt”, innebär det
just simuleringar utifrån memoraten. De är fiktiva på så sätt att de inte återger en auditiv miljö som kan förutsättas vara verklig i en absolut mening, men däremot en som
uppfattats varande som verklig i den kulturmiljö där den hörde hemma. Metoden, som
alltså är en del av produktionsperspektivet, kan närmast jämföras med en experimentell
musikarkeologi där man, utifrån ett ofta knapphändigt källmaterial, försöker återskapa
forntida ljudmiljöer.70 Men med auditiva simuleringar av ett urval memorat kan man
möjligen uppställa hypoteser och ställa frågor om de kulturella innebörderna i berättelserna om sällsamma auditiva miljöer, som människor sade sig ha upplevt. Möjligen
kan man också formulera nya slags frågor om vad musik är, eller kan vara. Jag ska skissera några få punkter med utgångspunkt i produktionsperspektivet ovan – jag lyfter inte
alla, utan endast ett litet urval. De besvaras heller inte, utan preciserar hur fallstudien
tentativt skulle kunna utföras, mot bakgrund av produktionsperspektivet. När man ska
68
69
70
Tone Dahlstedt, Tro och föreställningar kring vitra i övre Norrland (Umeå: Dialekt- och ortnamnsarkivet
i Umeå, 1976). Se även ett omfattande ljudmaterial i form av intervjuer på Dialekt- och folkminnesarkivet i Uppsala (DFU). Hos författaren.
Toivo Burlin, “Saivos sång” [opublicerat manus, 2015].
jfr Musica Sveciae: Fornnordiska klanger: The Sounds of Prehistoric Scandinavia (Producer: Cajsa S.
Lund) (CD 1991).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap
27
producera de auditiva simuleringarna och analysera föreställningarna om vitras musik
bör man utgå från musikens funktion som del i människans interaktion med andra människor, naturen, samhället/kulturen och teknologin (vitras musik hördes ofta i ensamhet i
naturen, i utkanten av kulturen, musiken motsvarade vanligtvis den musikteknologiska
nivån i samhället), man bör utgå från att musik/ljud och musik/teknologi alltid hör nära
samman i en kultur – (vitras melodier och deras övriga ljud av koskällor och röster
blandas ofta i memoraten) det kräver alltså ett annat slags musikbegrepp; skapande av
musik är samtidigt en process och en skapad/producerad produkt, samtidigt ett historiskt
(diakront) fenomen och ett synkront fenomen – de är processer och (i kulturell mening
“instabila”) auditiva produkter. När det gäller föreställningarna om specifika händelser
och iakttagelser är de särskilt instabila eftersom de befinner sig i ett spänningsfält mellan en subjektiv psykologisk upplevelse (upplevelsen av vitra), den nedtecknade eller
inspelade berättelsen (memoratet) och slutligen den tolkning och simulering av händelsen genom en auditiv gestaltning, som forskaren i detta exempel utför.
Diskussion: Experimentell musikproduktionsvetenskap som kulturforskande musikvetenskap?
Jag har i denna artikel presenterat ett bestämt perspektiv på musikvetenskapens relation till musikproduktionsvetenskapen och argumenterat för att många av de problem, teorier och metoder som utvecklats inom den senare, genom en interdisciplinär
dialog, kan införlivas i musikvetenskapen. Jag menar att musikvetenskapen är i behov
av en metodologisk och teoretisk förnyelse, som är särskilt angelägen när det gäller
grundläggande synsätt vad gäller musikteknologin och mediernas betydelse för musiken. Att på ett reflexivt sätt använda produktionsmetoder från musikproduktionsvetenskapen kan vara fruktbart för att åstadkomma inte bara förnyelse i teoretiska och
metodologiska perspektiv, utan framför allt för att utveckla ny kunskap.
I domänen musikproduktionsvetenskap är – till skillnad från i huvuddelen av
musikvetenskapen – den instrumentella relationen till musikproduktionsverktygen
central. Möjligheten finns därför att relatera en vetenskaplig diskussion om produktionens villkor med en produktion i praktiken – att bedriva en “experimentell” musikproduktionsvetenskap. Produktionen kan förenas med en metavetenskapligt, reflexivt
förhållningssätt, mellan att genomföra en produktion – och utveckla teorier och modeller om produktionen. Frågor om historiska och teknologiska villkor för musik (och
musikproduktionen) kan prövas experimentellt med produktionsverktygen. Kan ljudoch musikproduktioner, “experimentellt” framställda enligt denna modell, användas
som en metod för att besvara humanistiska eller kulturvetenskapliga – specifikt musikvetenskapliga – frågeställningar? Med min fallstudie – med ett skisserat projekt om en
föreställd, supranormal musik och auditiva miljöer – konkretiserades resonemanget.
Mitt korta svar på frågan ovan är: ja, av forskaren framställda ljud- och musikproduktioner, experimentellt framställda och teoretiskt grundade i “produktionsperspektivet”, kan med hjälp av teorier från humanvetenskap, samhällsvetenskap eller andra
vetenskapliga fält användas som både en metod och ett material för att utveckla teorier om musikproduktionens villkor, igår och idag.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
28
Toivo Burlin
Abstracts
I artikeln diskuteras hur musikvetenskaplig forskning om musikproduktion och en
“musikproduktionsvetenskaps” perspektiv tillsammans kan bidra till att forma ett
medie- och produktionscentrerat forskningsparadigm, inom musikvetenskapen. Det inleds
med en översikt på hur produktion av musik har studerats som vetenskapligt objekt.
Ett förslag presenteras om hur det interdisciplinära forskningsfältet – studier av i vid
mening produktion och medialisering av inspelad musik – kan utvecklas till ett bredare medieparadigm inom musikvetenskapen som disciplin. “Produktionsperspektivet”,
som inkluderar produktion som metod, introduceras därefter. Det exemplifieras med
en fallstudie.
The article discusses how musicological research on recorded music and “Music Production Studies” perspectives, respectively, together can help shape a media and production-centered research paradigm in musicology. It begins with an overview of how the
production of music has been studied as a scientific object. A proposal is presented on
how the interdisciplinary field of research – studying production of music and the mediatization of recorded music – can be developed into a broader media paradigm in
musicology as a discipline. “The production perspective”, which includes production
as a method, is then introduced. It is exemplified by a case study.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
MADS WALTHER-HANSEN
Sound Events, Spatiality and
Diegesis – The Creation of Sonic
Narratives in Music Productions
Introduction
“You Think I Ain’t Worth a Dollar, But I Feel Like a Millionaire” is the opening track
on Queens of the Stone Age’s third studio album Songs for the Deaf.1 The track begins
with the sound of clanking keys, a car engine that starts and the alerting beeps warning the driver that doors are still open. A car radio is switched on and the car door is
slammed bringing the beep sound to a halt. The radio is tuned and eventually settles
on a station where a male voice announces: “Los Angeles, Clone Radio, we play the
songs that sounds more like everyone else, than anyone else.” The song following the
presentation opens with a steady drumbeat. Seconds after, a distorted guitar riff joins
in. What we are witnessing is a rock song, transmitted to us through a car radio; an
effect that emerges mainly as a result of low cut filtering in the 200 Hz region and a
high cut around 7kHz that creates the impression of a small lo-fi stereo-system. The
rock song is there for the listener as something played back into the virtual car setting,
and thus present only in a double-mediated sense.
There is a sudden change in the scene the moment vocals2 and bass guitar enter at
1’01 where the track abruptly increases in loudness and the frequency band increases
to full spectrum. The perceptual effect of this change can be described in at least two
ways. Either the radio track “pops out” of the car speakers, entering and now residing
in the virtual listening space (the car setting) that was previously established; or the
track “drags the listener in.” In both cases, the changing point of audition (POA) is
not the result of the track moving closer to the listener across the virtual space of the
recording. There is a change in the ontological3 level of the track and in this way the
1
2
3
Queens of the Stone Age. “You Think I Ain’t Worth A Dollar But I Feel Like A Millionaire,” Songs for
the Deaf (Interscope, 2002). The album is produced by Josh Homme (front singer) with record engineers Adam Kasper and Eric Valentine.
Lead vocals on this track is Nick Oliveri and not Queen of the Stone Age’s usual front singer Josh
Homme.
Building on the viewpoint taken by Merleau-Ponty that the body-subject is intimately tied to lived
space and, consequently, that the phenomenal world equals the real world, my use of the term ontology refers to the categories (or hierarchies) of spatial being in the recording (e.g. that is how, and
at what ‘level’, spatial entities in the recording can be said to exist) that is formed in perception – a
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
30
Mads Walther-Hansen
track transcends the virtual borders (in time and space) of the radio’s virtual space,
thus creating a dramatic increase of presence for the listener.
The example above serves to illustrate two points that will be dealt with in this
article: 1) the potential for sound engineers to radically effect the listener’s perceived
point of audition, and create different forms of tensions and contrast effects between
these different perspectives; and 2) the ontology of recorded sound (i.e. the time-space
continuum of the perceived auditory events) is only fully realized if we consider how
sound events coexist within the imagined performance space of recordings. The purpose of this article is to discuss the relation between sound qualities emerging from
studio-based editing and the perceived diegetic framing in a number of popular music
productions. I revisit the concept of diegesis in this context to explore its applicability in the analysis of recorded music and as a tool to conceive of the perceptual effect
emerging from recording and editing practices.
For several decades, film sound theorists have used the concept of diegesis to explore the relation between sound (e.g., background music, dialogue, voice over) and
image in film.4 More recently Anahid Kassabian has argued that the diegesis of the
film is not necessarily determined prior to the music, but music itself can construct a
diegetic framing depending on the degree of attention given to music in the particular scene.5 Also a number of studies have redefined the concept to explore the function of sounds in computer games.6 The reality of movies, like the reality of recorded
performances, never existed as a real event. Both movies and music recordings consist
of clips and layers of recorded material spliced together to form, in our experience, a
coherent event. Nicholas Cook notes this comparison between movies and music recordings writing that:
(…) the concept of diegesis applies just as well to sound recordings [as to movies],
and demonstrates the sense in which the relationship between the recording and
4
5
6
phenomenological ontology. Also – and this is in line with Brian Massumi’s thinking – I think of
reality as a part of virtual spatiality (e.g., something sensed as real through the virtual) rather than
contrasted to it. See Maurice Merleau-Ponty, Phenomenology of Perception (London: Routledge and Kegan Paul, 1945/2002); M. C. Dillon, Merleau-Ponty’s Ontology (Evanston, Illinois: Northwestern University Press, 1988); Brian Massumi, “Envisioning the Virtual,” in The Oxford Handbook of Virtuality
ed. Mark Grimshaw (Oxford and New York: Oxford University Press, 2014).
See David Bordwell, Narration in the Fiction Film (London: Routledge, 1985); Michel Chion, AudioVision: Sound on Screen (New York: Columbia U.P., 1994); Claudia Gorbman, Unheard Melodies
(Bloomington and Indianapolis: Indiana University Press, 1987); Étienne Souriau, L’Univers filmique
(Paris: Flammarion, 1953).
Anahid Kassabian, Hearing Film: Tracking Identifications in Contemporary Hollywood Film Music (New
York: Routledge, 2001); Anahid Kassabian, “The End of Diegesis as we Know It?” In The Oxford
Handbook of New Audiovisual Aesthetics, eds. John Richardson, Claudia Gorbman and Carol Vernallis
(Oxford and New York: Oxford University Press, 2013).
Karen Collins, Game Sound: An Introduction to the History, Theory, and Practice of Video Game Music and
Sound Design (Cambridge: MIT Press, 2008); Mark Grimshaw, The Acoustic Ecology of the First-Person
Shooter: The Player Experience of Sound in the First-Person Shooter Computer Game (VDM Verlag, 2008);
Kristine Jørgensen, “On Transdiegetic Sounds in Computer Games,” Northern Lights 5 (2007), 105117; Ulf Wilhelmsson. Enacting the Point of Being. Computer Games, Interaction and Film Theory (PhD
Thesis. University of Copenhagen, 2001).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
31
the experience is a semiotic one (…) it is through the act of listening which the recording prompts that we understand it as a trace of an event.7
Also Simon Trezise argues that the concept is useful to analyze the purely auditory domain8 but still there is no in-depth theorization of how the diegesis of music recordings is formed and exploited in the recording studio.
In the first part of this article I discuss diegesis in relation to the production and
perception of recorded popular music. Most producers seem to aim for mixes that
produce the illusion of a single diegetic frame containing all sound events. For this
reason, the best way to recognize the value of my argument is by exploring some of
the exceptions – that is, tracks where the diegetic relationship between sound events is
challenged or simply broken. In order to discuss the different ontological relationship
between sound events within recordings I use Gérard Genette’s classification of narrative events: the diegetic, meta-diegetic and extra-diegetic level,9 later used by Claudia
Gorbman in her classification of film sounds.10 Although there are many similarities
between how we make sense of audio-visual media and audio media, the notion of
diegesis is not unproblematic in the context of popular music productions. For this
reason, this article proposes a systematic theoretical framework to discuss the concept
in relation to recorded music – what I shall here call a phonographic diegesis. The second part of the article addresses a number of musical examples to exemplify different
modes of layering sound events.
Presenting the Story
In the third book of The Republic Plato contrasts narration (diegesis) with imitation
(mimesis) as two ways of presenting literary works. In narration, the story world is
presented through the narrator telling the story. The story world is thus located outside the time and place of the narrator. Imitation on the other hand is the mimetic enactment of the story. Here the story is told through dialogue, speech or bodily gestures
from within the story world.11
The term diegesis has reappeared a number of times in literary analysis. Gérard
Genette re-appropriates the term making a distinction between the story told and the
diegesis. The diegesis, he explains, rather than being the story itself, is the universe in
which the story takes place.12 Karol Berger addresses this discussion outside the context
of literary analysis, considering the diegesis of visual and auditory events in paintings
and music. Berger argues that we may connect the visual expression of paintings to
7
8
9
10
11
12
Nicholas Cook, “Methods for Analysing Recordings,” in The Cambridge Companion to Recorded Music,
ed. Nicholas Cook et al. (Cambridge: Cambridge University Press, 2009), 243.
Simon Trezise, “The Recorded Document: Interpretation and Discography,” in The Cambridge Companion to Recorded Music, ed. Nicholas Cook et al. (Cambridge: Cambridge University Press, 2009), 207.
Gérard Genette, Narrative Discourse Revisited (Ithaca: Cornell University Press, (1983/1988).
Gorbman, Unheard Melodies.
Plato, The Republic (English translation by Paul Shorey) (London: William Heinemann Ltd., 380
BC/1930).
Gérard Genette, Narrative Discourse Revisited, 17.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
32
Mads Walther-Hansen
the literary “voice.” Analogous to the way the “voice” communicates literary content,
human figures in paintings communicate through visual bodily gestures presented to
us in specific environmental settings.13 Similarly, in music, a vocalist may present a story world from within the perceived reality frame by, for example, singing about present
feelings. He/she is then a character in the story world (mimesis). Or else the singer
may present a story about another time and place. He/she is then the narrator of the
world (diegesis). Also, as is often the case, the singer may switch between the different modes of presentation throughout the song.14 While there are obvious connections
between literary narration and vocal utterance in music, there are less obvious, yet still
significant, ways that a diegesis is established in the making of musical recordings.
Recorded tracks can be approached using the performance metaphor – as virtual
performances,15 perceived performances16 or phonographic performances17 – and this
line of thinking conditions the argument in this article as well as the aesthetic reflections presented in a vast number of books and articles dealing with analysis and production of music recordings. The performance metaphor is contingent on a causal listening mode. Here the listening experience is directed towards the behavior and gestures of performers associated with the production of the musical piece. These gestures
belong to the sound-producing event as the way the performers express themselves.
Simon Frith uses a visual analogy claiming that:
(…) to hear music is to see it performed, on stage, with all the trappings. I listen
to records in the full knowledge that what I hear is something that never existed, that never could exist, as a ‘performance’, something happening in a single
time and space; nevertheless, it is now happening, in a single time and space: it
is thus a performance and I hear it as one.18
Although the presence of gestural elements in auditory experience is largely dependent on the particular genre (the performative aspect of the listening experience is definitely stronger in a Rolling Stones track than in one of Brian Eno’s ambient productions), the exploratory dimension of auditory perception19 is integral to most music
listening. In the rock domain this link is particular salient where the tracks often construct the idea of a distinct performance event, that is, the illusion that the music was
performed simultaneous despite layers of non-simultaneous recording takes that have
13
14
15
16
17
18
19
Karol Berger, “Diegesis and Mimesis: The Poetic Modes and the Matter of Artistic Presentation,” The
Journal of Musicology 12, 4 (1994): 407-433, 415.
For a thorough account of self-impersonation see Wendy Doniger, The Woman Who Pretended to be
Who She was: Myths of Self-imitation (Oxford and New York: Oxford University Press 2005).
Allan F. Moore and Ruth Dockwray, “The Establishment of the Virtual Performance Space in Rock.”
twentieth-century music 5, 2 (2008): 219-240.
William Moylan, The Art of Recording: The Creative Resources of Music Production and Audio (New York:
Van Nostrand Reinhold, 1992).
Cook, Methods for Analysing Recordings.
Simon Frith, Performing Rites: On the Value of Popular Music (Oxford: Oxford University Press, 1996),
211.
Eric F. Clarke, Ways of Listening: An Ecological Approach to the Perception of Musical Meaning (New York:
Oxford University Press, 2005).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
33
gone into its making. The discussions in this article is limited to recordings well-suited
to the performance metaphor (i.e. recordings that encourage us to listen to it as a performance event framed by a particular time and place) in order to argue for some of
the extended meanings that emerge from the staging of sound in the recording studio.
Taken as perceived performances, recordings are affective because they allow the
listener to engage in an imagined relationship with the performer, by participating in
a perceived performance event. Philip Auslander has called this affective state, the experience of liveness, by arguing that in spite of the fact that recordings represent a temporal gap between the performance and the reception, recordings form a technologically mediated co-presence of a performance.20
One may argue that there is always a diegetic layering in record listening; this is, in
fact, true. What we witness when we put on a record is the medium telling the story of
a performance. This is no different from the relationship between the videotape and
the film. The record, as well as the videotape, underlines in a very direct way the temporal and spatial gap between the perceived event and the moment and place of play
back. It is through the recordings “act” of playing the event back to us that we come
to witness the performed event, and it is here that we are presented with the ontological division between the “here and now” of the playback situation and the “then and
there” of the perceived performance.
When we listen to music we have before us a phonogram – either in the form of a
physical media (e.g., vinyl record or CD) or as a media-file (e.g., mp3 or wav). In a certain sense, phonograms are to record producers and sound engineers, what books are
to writers. There is an author (or several authors) behind the specific configuration of
sounds presented to us in recordings; this is analogous to the authors behind every word
we read in a book. Despite these similarities, few people buying music would probably
know the record producer involved in the recording; in contrast, most readers will know
the author behind the book. Books, also, are usually listed by author’s name, while popular music phonograms are seldom, if ever, displayed by the name of the sound engineer or producer (unless the engineer or producer equals the performer). In this way, the
musical performer certainly takes center stage in the history of musical recordings.
This fact, however, only tells us something about sound engineers’ hidden status
in the history of record making, not, as a number of scholars have acknowledged, that
sound engineers have actually played a profound role in the way music records have
come to sound.21 While it may be contentious to entitle record producers authors (or
co-authors) of the recording, it allows us to focus more on the way sounds are staged
within the virtual space of the recording. Record producers mold sounds in and after
the recording and thus create imaginary images of performances for the listener. In
this way, recordings present a narrative emerging from the specific structure of sound
events, analogous to the narrative present in books that emerges from the specific
20
21
Philip Auslander, Liveness: Performance in a Mediatized Culture (Oxon: Routledge, 1999).
see Mark Cunningham, Good Vibrations: A History of Recording Production (Chessington, Surrey: Castle
Communications, 1996), and Albin Zak, The Poetics of Rock: Cutting Tracks, Making Records (Berkeley:
University of California Press, 2001).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
34
Mads Walther-Hansen
structure of words – presenting a specific angle the story the specific way the story is
told. Books and phonograms are the final product presenting the story or perceived
performance in its fixed form. In this way, one of the tasks of record producers is to
present a specific story22 of a performance among the many that could be told.
From an audio-visual to an auditory diegesis
To outline this argument further I want to propose an analogy between filmic events
and sound events in recorded music. The events we experience in films are that of
characters acting out their lives. These events frame the film’s fictional reality – its
diegesis. The diegesis of the film is thus the presented filmic “reality;” that is, everything that logically belongs to the fictional world the film portrays, whether included
(on-screen) or excluded (off-screen) from the picture frame.23 On-screen sounds are,
for instance, voices of characters on the screen or music coming from visualized playback systems. Off-screen sounds can be voices of characters that are not visualized,
environmental sounds or music that belongs to the world of the characters, that is,
music played back from a playback device outside the screen but in the film’s fiction.
Extra-diegetic sounds emerge from outside the filmic reality; for instance, background music or other sound effects added to bring the viewer into a certain mood.
The sources of non-diegetic sounds are not visualized, such as the voice over often
heard in documentary films. Claudia Gorbman proposes that sounds may also operate at a meta-diegetic level, meaning that sounds belong within the character; as is the
case, for example, with sound presented as belonging to the character’s imagination.24
Bordwell and Thompson present a more overall distinction between non-diegetic
(non-story space) and diegetic (story space). These categories are further sub-classified
in terms of the temporal relation between sound and image. Diegetic sounds may occur either simultaneously or non-simultaneously (earlier or later in the story) relative
to the image, such as in the case of sound flashbacks or flash-forwards, while nondiegetic sounds may be marked as past (e.g., historical voice over present images),
present (e.g., a narrator speaking in present tense) or later (e.g., a reminiscent narrator). Bordwell and Thompson state, however, that the temporal relationship of nondiegetic sound to the story is mostly irrelevant to the filmic experience.25
In recorded music the perceived temporal relation between sound events has a
more profound role for the classification of diegetic, meta-diegetic and extra-diegetic
22
23
24
25
Since this article deals with virtual performances as something constructed in the recording studio,
I take story and narrative to be interchangeable terms. In this way, I see no difference between how
the story is told (the narrative), and the perceived event (the story), both are products of narrating
(see also Genette, 1983/1988). A study of recordings as documents of actual events should obviously distinguish between narrative and actual story.
Gorbman, Unheard Melodies; Christian Metz made a similar distinction between what he called ‘ordinary sound’ and ‘sound-off’, see Christian Metz, Language and Cinema (De Gruyter, 1974).
Gorbman, Unheard Melodies, 22-23.
David Bordwell and Kristin Thompson, Film Art: An Introduction (New York: The McGraw-Hill Companies), 1997.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
35
events. Meta-diegetic sound events are indexes of events that took place before the
diegetic events, while extra-diegetic sound events are indexes of events that take place
after the diegesis. As already stated, contrary to film, diegetic layering in recorded
music is not conventional practice. When the experience of extra-diegetic or metadiegetic events occurs, however, they often emerge, as we shall see in the next section,
as a consequence of both a spatial and temporal hierarchy between sound events.
The virtual space of recordings
Record listening is an instance of what Pierre Schaeffer called an acousmatic situation26, that is, a situation in which we hear something without seeing its cause. According to Schaeffer this situation may potentially (or ideally) lead to a listening experience stripped of any relation to a sound-producing event (reduced listening).27
Several other scholars, however, argue that auditory perception is fundamentally exploratory.28 When listening to sounds we seek information regarding what is going on
— what is that sound the sound of? What kind of event is taking place and what kind
of source is involved? Michel Chion even claims that the acousmatic situation, rather than removing causal references, actually enhances the listener’s interest in sound
events.29 In this way the acousmatic situation results in a new and, in many ways, enhanced aesthetic involvement compared to actual musical performances.
In audio-visual media the question of diegesis emerges in the relationship between
sound and image. In recordings the question of diegesis emerges in the relationship
between sound events and other sound events, more particularly in the relationship
between sound events and the imaginary image of the performance. This perceived
performance takes place in an imaginary performance setting that is most often conceptualized in terms of its impossible30 or unnatural31 characteristics compared to actual acoustic spaces. This space is dimensional although often not with clearly defined
acoustical properties.
In the most general sense we may think of space, whether actual or virtual, as either
a prerequisite or a product of perception; that is to say that in every act of perceiving
something, space is presupposed. Like the scene in a film, the auditory performance
takes place somewhere. Sound sources are presented to the listener in certain spatial
configurations that form a performance space. William Moylan has proposed the con26
27
28
29
30
31
Acousmatics is derived from the name Pythagoras gave to his uninitiated students (akousmatikoi),
who received their lectures from behind a veil.
Pierre Schaeffer, “Acousmatics,” In Audio Culture: Readings in Modern Music eds. C. Cox and D. Warner (New York: Continuum, 1966/2004).
See Clarke, Ways of Listening; Simon Emmerson, Living Electronic Music (Aldershot: Ashgate, 2007);
Luke Windsor, “Through and Around the Acousmatic: The Interpretation of Electroacoustic Sounds,”
In Music, Electronic Media and Culture ed. Simon Emmerson (Aldershot and Burlington: Ashgate,
2000).
Michel Chion, Audio-Vision, 32.
Timothy Warner, Pop Music: Technology and Creativity: Trevor Horn and the Digital Revolution (Bodmin:
Ashgate Publishing Limited, 2003).
Francis Rumsey, Spatial Audio (Oxford: Focal Press, 2001).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
36
Mads Walther-Hansen
cept of the sound stage to describe the perceived two-dimensional stage on which
sound sources are located. According to Moylan the sound stage is located in front of
the listener at a certain distance.32 Sound sources often have different locations on the
sound stage, and often they exist in different acoustical spaces; yet, listeners will unconsciously group them together to form a coherent performance.33
It is tempting to define the sound stage as a diegetic frame in so far as the sound
stage – the specific spatial formation formed in the act of listening – often frames the
perceived performance. Being the focal space of the listening experience, the sound
stage is, however, more like the visual screen in films. For instance, we may think of
sound events that appear to arrive from outside the sound stage, yet still “participating” in the perceived performance. Such examples are clearly considered exceptions
and often involve quite radical editing. We may imagine, for example, hearing a sound
event with a perceived location in an adjacent room, performing with another performer located just in front of us. The argument here is that while everything located
on the sound stage must belong to the diegesis of the perceived performance, everything that belongs to the performance need not be located on the sound stage.
My use of the word performance is understood as an event that takes place in a given
time and place. The perceived performance may, as is often the case, include all sound
events in the recording. However, in some cases sound events exist outside the temporal
and spatial frame of the perceived performance. Clearly, there can be no sound events
in recordings without a virtual space in which the listener can situate these events. We
may, however, say that the recording holds a single diegesis, only in so far as all sound
events in its virtual space are in fact perceived as being temporally related. If not, we
may think of some sound events as belonging to an extra-diegetic or meta-diegetic level.
The phonographic diegesis emerges from the specific configuration of sounds in
the recording, and is bound to the idea of recordings as perceived performances and
the virtual place and time of these performances. Recordings carry a virtual history as
well as a virtual space. While the perceived performers may be heard as performing
at the same time in the same place, they may also be heard as performing before and
after each other. I am not referring to musical linear time, but to virtual historical moments where we perceive that the events took place. In this way, there is a very direct
connection between virtual spatiality and virtual history, as the spatial separation is
often realized in terms of the temporal separation between sound events. We may exemplify this with Queens of the Stone Age’s “You Think I Ain’t Worth a Dollar, But I
Feel Like a Millionaire” mentioned in the introduction of this article. In this recording, the change in the perceived point of audio is not only related to spatiality (e.g.
the performance leaps from one perceived performance space to another); the changing point of audition is also related to temporal change (e.g. the performance leaps
from a certain point in the recordings virtual history – something pre-recorded played
on the radio – to a later point in time – the moment of playback in the car setting).
32
33
See also Justin Christensen, Virtual Distance and Soundstage, and their Impacts on Experienced Emotional
Valence. page 101-118 in this issue.
Moylan, The Art of Recording.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
37
This leads me to my definition of diegesis in recorded music. The diegesis equals
the perceived temporally connected performance. Sound events in the diegesis simply
take place at the same time and the perceived performers can then logically respond
to (hear) each other. In some cases, the diegetic bond is broken. This happens, for instance, when we experience that perceived performers respond to pre-recorded material (meta-diegetic sounds) or sound events that emerge after the perceived performance
took place (extra-diegetic sounds).
Arguably, listeners tend to group the sound events into a single overall stream.34 It
simply makes sense to listeners to hear recordings as perceived performances, despite
the spatial paradoxes often involved in doing so. Breaking up sound events into ontologically different layers is, therefore, something that demands radical recording techniques and sound editing, as I will exemplify further in the second part of this article.
Classification of sound events and the creation of meaning in recordings
As already proposed, certain ways of categorizing filmic sounds are applicable in describing certain kinds of phonographic sound events and to account for the creation
of meaning in music production. In most recordings all sounds are grouped together
on a single sound stage, forming a unified perceived performance. There are, however,
several examples of tracks that do not present a single perceived performance space
and thus challenge the idea of the sound stage. I propose four categories (Table 1) that
describe the relationship of sound events to the perceived performance space.
Diegetic on-stage sound events:
Sound events in the perceived performance space.
Diegetic off-stage sound events:
Sound events absent from the perceived performance
space, but temporally present.
Meta-diegetic sound events:
Internally re-produced non-diegetic sound events. These
sound events takes place prior (in the virtual history) to
the perceived performance, reproduced from a media
within the sound stage.
Extra-diegetic sound events:
Externally produced non-diegetic sound events. These
sound events takes place after (in virtual history) the perceived performance, causing the perceived performance
to appear reproduced from within a new embracing ontological frame.
Table 1: Classification of phonographic sound events
Diegetic on-stage sound events are the most common category. Although musicological analysis may reveal that a track consists of many different acoustical environments, the listening experience most often involves hearing the perceived performance
34
Albert S. Bregman, Auditory Scene Analysis: The Perceptual Organization of Sound (Cambridge and Massachusetts: MIT Press, 1990).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
38
Mads Walther-Hansen
as spatio-temporally present in spite of the apparent acoustical paradox involved in
doing so.
Diegetic off-stage sound events are defined as sound events that appear acoustically
non-present in the primary acoustical environment. These sound events “participate”
in the perceived performance events and are thus temporally co-present, however, the
off-stage sound events are acoustically separated from the on-sound stage events. This
acoustical separation can involve illusions of sound events located behind a physical
obstacle, for instance, sound sources located in an adjacent room.
Meta-diegetic and extra-diegetic sound events are examples of a layering of perceived performances. In both cases one sound universe is presented through the other. Hence, two juxtaposed performances merged together in the track, for instance, as
is the case in mash ups, are not understood here as an instance of diegetic layering.
Diegetic layering, in my definition, emerges when listeners get the illusion of a hierarchical relation between two virtual spaces. Meta-diegetic sound events are sound
events emerging from a media within the diegesis of the perceived performance.
Analogous to paintings that portray a room in which another picture is present,35
meta-diegetic sound events are illusions of pre-recorded sound events played back in
the perceived performance situation. Logically performers (on the sound stage) can
hear these sounds while performing. There is thus both a temporal and spatial division between diegetic sound events and meta-diegetic sound events.
Extra-diegetic sound events emerge outside and after the diegetic performance. Like
meta-diegetic sound events, extra-diegetic sound events emerge as a media-withinmedia effect. However, in the first case, the meta-diegesis is in the embedded media,
while the in the second case the diegesis is in the embedded media.
Although the categories I propose are easily definable, they are not always straightforward to use in the description of recorded sound. Problems arise from the fact that
the diegesis is not always clearly defined in listening, and from the fact that sound
events do not necessarily have the same ontological status throughout a track. In
this way, the ontological hierarchy of sound events may change from the first hearing of a track to the second, as the scene of the performance is formed in the listening situation.
Off-stage sound events
Filmic off-screen sounds are sounds with non-visualized sources that take part in
the filmic events. In the case of the voice, we would define it as off-screen when we,
the audience, cannot see the mouth of the filmic character speaking.36 Phonographic
off-stage sound events are defined in terms of the spatial (acoustical) separation between sound events. In films there need not be any acoustical division between onscreen and off-screen sounds. In fact off-screen sounds may often appear acoustically
35
36
Berger, “Diegesis and Mimesis,” 421-422.
Michel Chion and Claudio Gorbman, The Voice in Cinema (New York: Columbia University Press,
1999).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
39
closer to us than on-screen sounds. Phonographic off-stage sounds on the other hand
emerge when we experience that performers are located in different environments. As
a result they cannot “see” each other. In this way, phonographic off stage sounds are
defined from the point of view of the performers. What makes off-stage sound events
comparable to off-screen sounds is that off-stage sound events mimic sound sources
that would not be visible if we could see the performance. Thus, in order to account
for the link between off-stage sound events and filmic off-screen sounds, we have to
imagine what we would be able to see in an actual performance.
A notable example of off-stage sound events occurs in the tune “I Won’t Hurt You”
from the album Smiling and Waving37 by the Norwegian singer Anja Garbarek. In this
track, there is a significant acoustical separation between the lead vocal in front and a
guitar and percussion sound. Although the percussion sound appears too mechanical to be actually performed, the environmental co-presence with the humanly performed guitar assists in animating the percussion sound.38 The off-stage effect occurs
primarily as a result of low-pass filtering around 400Hz (with a fairly steep slope) on
guitar and percussion. In this way, the listener may experience some concrete material
separating percussion and guitar from the sound stage. Accordingly, we get the impression that we would be able to see the vocalist, who is located just in front of us, while
the sounds heard emanating from behind the barrier are invisible. For this reason we
may think of the vocal as on-stage and guitar and percussion as off-stage. Following
this line of thinking the guitar sound and the percussion sound appear inaccessible to
both the listener and the on-stage vocalist and therefore signal a spatial rupture in the
horizontal dimension. The vocalist appears lonely on the sound stage, yet she participates in a shared performance with off-stage sound events. This experience emerges as
a result of the combination of a temporal co-presence and an acoustical separation.
The example, then, illustrates the importance of distinguishing between a sound stage
(a circumscribed acoustic space) and a perceived performance universe (the totally of
synchronous sound events).
Contrary to filmic off-screen sounds that may change to on-screen sounds if the
source becomes visible,39 phonographic off-stage sounds are not loaded with such expectations. We simply do not expect phonographic characters to move. Neither do we
expect a change of listening perspective.
The experience of off-stage sound events does not necessarily occur in all tracks
where a closer analysis may reveal acoustically detached sound events. Most of the
time, listeners group sound events into a single perceived performance in spite of the
inherent spatial paradoxes that this grouping implies. Off-stage sounds often emerge
37
38
39
Anja Garbarek, “I Won’t Hurt You,” Smiling and Waving (Virgin, 2001).
We may also notice the semiotic connotations of the percussion sound, relating it to the sound of
the performer’s heartbeat. The sound may, then, be heard as something internal to the performer –
something the listener hears only through the vocalist hearing, feeling or imagining it. Since the guitar sound is located in the same acoustical environment as the percussion track, we may also locate
the guitar sound in the mental interior of the performer. In this sense, we may think of the percussion and guitar sounds as meta-diegetic – on-stage and internal – instead of off-stage.
Chion, The Voice in the Cinema.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
40
Mads Walther-Hansen
as a result of heavy processing, for instance applying a high cut filter at 250 Hz or
other spatial effects that clearly creates the illusion of two separate environments.
Whether this illusion in fact occurs, obviously depends on many aspects related to
both production and reception. While the visual frame in film clearly separates inside
and outside, such borders do not (or only vaguely) exist in recorded music. For this
reason, we must be content with blurred lines between phonographic off-stage and
on-stage sound events.
Meta-diegetic sound events
While it may prove difficult to argue for off-stage sound events in recordings, one can
find many clearer examples of meta-diegetic sound events. Pink Floyd’s “Wish You
Were Here” from the album Wish You Were Here40 from 1975, for instance, has a very
salient diegetic layering of sounds.
The track opens with the sound of a radio being tuned. As the tuning settles on a
station, a guitar sound is heard, apparently emerging from the radio. At this point, a
diegetic layering is already implied between the radio playback and the performance
within the radio. We do not, however, hear any acoustical signs of a virtual playback
environment in which the radio, logically, must be situated. The sound from the radio
simply does not activate an acoustic space and we do not hear any signs of human
presence “outside” the radio. Thus, the radio in-itself remains an abstract non-spatial
object. This experience is particularly strong when played back through headphones,
where the radio can only be heard through the right headphone.
This experience changes quite radically when we hear coughing and breathing
noise in the center position at 0’26. The coughs and breath noises do not arrive from
the radio, rather, from the room in which the radio is playing. In this way, we come to
experience a perceived radio listener and, accordingly, the spatial and temporal separation between the radio listener and the radio performance. The diegetic layering of
sound events is, thus, fulfilled as we hear both the radio performance and the listener
simultaneously. As a guitar starts playing at 0’42, we get the illusion that we are hearing someone playing along to the radio track. As a result, the playback environment
is transformed into a combined playback and performance environment. We hear
the radio performance in the knowledge that the guitar performer hears it too. The
fact that the diegetic frame is consolidated gradually on top of the meta-diegetic level
seems to intensify the presence of the diegetic frame. This effect emerges as a result of
the shift in auditory focus. Listeners are likely to naturalize the meta-diegetic level in
the first bars of the radio-guitar – attend to it as if it was diegetic – only to be surprised
by the presence of the guitar when the diegetic level is set in motion.
At 1’34 the vocal enters. From this point in the song, the radio performance gradually becomes less noticeably, and at the 1’36 it disappears completely, without any
clear demarcation of its disappearance. As the drums enter, followed by bass and
40
Pink Floyd, “Wish You Were Here,” Wish You Were Here (EMI, 1975).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
41
piano, the performance environment is transformed (expanded) again into a stage. In
this way, the intro presents a gradual shift from a performance in an “everyday setting”
to a performance “on stage.”
The intro of “Wish You Were Here” is an example of a media-within-media effect,
or to expand on William Moylan’s terminology,41 a sound stage in sound stage experience (Figure 1). The diegetic guitar performer situated in Sound Stage A can hear the
meta-diegetic guitar performance in Sound Stage B. On the opposite, the performer on
Sound Stage B cannot hear the performer on Sound Stage A. In other words, it is the
diegetic performance that makes it possible for us to hear the meta-diegetic performance, and not the other way around.
Figure 1: media-within-media
While it makes sense to talk about an ontological rupture between sound stage “A”
and “B,” the addition of sound stage “A” have implications for the way listeners attend
to sound stage “B.” The primary perceptual effect pertains to a further mediatization
of sound stage “B.” The effect of another layer of virtual mediation is a weakening of
the indexical relation to a pre-medial world. It is not that the performance on sound
stage “B” necessarily ceases to be heard as one. It is rather that the perceived performance recedes into the background, as a mediatized phenomenon in sound stage “A,”
something that is available to us only through it being present to someone else.
The media-within-media effect is in fact a common type of media based staging42
heard in many different forms in newer pop and rock albums. “Wish You Were Here,”
however, is a particularly clear case as the performance space is gradually established.
From the outset we get to hear a performer on the radio, what we may call a metadiegetic performer (analogous to Genette’s meta-diegetic characters43). While meta-diegetic sound events are still indexical of meta-diegetic performers, the index is generally weaker compared to diegetic sound events. The meta-diegetic performer appears
to us, only as the last part of the indexical chain, embedded in another index and on
the wedge of slipping from the listening experience.
41
42
43
See Moylan’s definition of the sound stage in The Art of Recording.
Simon Zagorski-Thomas, PhD in Music Composition. PhD Thesis (University of London, 2008).
Genette, Narrative Discourse Revisited.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
42
Mads Walther-Hansen
Another interesting case of diegetic layering is found in Björk’s “There’s More To
Life Than This” from her album Debut.44 The track opens with the sound of a lot of
people chatting in a large room. A dance beat emerges in the background and we realize that the setting is some kind of dance club. At 0’35 Björk’s vocal is heard close-up.
Throughout the track the listener is locked to this position just in front of Björk as she
walks around in the dance club. Everything the listener hears is from Björk’s point of
audition or more specifically from the position just in front of Björk’s mouth. In this
way, the track has a dynamic sound stage analogous to a moving camera that gradually changes what is considered “on-screen.” Like point of view screen shots in films,45
the point of audition46 perspective in “There’s More To Life Than This” generally enhances our identification with the vocalist.47
The sound of the background music seems to arrive from a set of loudspeakers in
the room in which Björk is positioned. Accordingly, we may think of the background
musicians as located on a meta-diegetic level. The music in-itself is of course diegetic in the sense that Björk can hear the music in the room she is in. However, this is
not the case the other way around. The musicians performing the background music
do not share reality with Björk. Thus, the background music stands in a hierarchical
relation to Bjork’s performance in that the ontological level of the former mediates
through the ontological level of the latter.
At 1’37 we hear Björk open a door to enter an adjacent room, and as she closes the
door behind her, background sounds are cut off. In this way, we get the experience
that Björk is now alone in another room and therefore not visibly connected to the
rest of the sound sources (the sound system and the chatting people) in the recording.
There is, thus, a clear division between on-stage sounds (Björk’s voice) and off-stage
sounds in this scene. Björk’s intimate vocal address is further intensified48 due to the
one-to-one illusion provided by the virtual environment.49
We may think of the back track as doubly displaced. Since it emerges through a
medium located off-stage, we may conceive of it as an off-stage meta-diegetic performance. As a result of the ontological division Björk takes on the role as the narrator addressing the background music. She is, so to speak, telling the story about the other
44
45
46
47
48
49
Björk, “There’s More to Life than This,” Debut (Elektra/Asylum, 1993).
James D. Buhler, David Neumeyer and Rob Deemer. Hearing the Movies: Music and Sound in Film History (New York and Oxford: Oxford University Press, 2010).
Chion, Audio-Vision, 89-92.
Philip Auslander – referring to Mary Hunter, “The Stage Director in Cinema”, In Theatre Arts (May,
1949) – makes a similar comparison between a stage director’s “manipulation of audience attention” and the moving camera lens in film that move from one point of interest to the next – a “psychological camera eye”. Philip Auslander, Livenes: Performance in a Mediatized Culture (Routledge
1999).
See Nicola Dibben, “Vocal Performance and the Projection of Emotional Authenticity” In The Ashgate Research Companion to Popular Musicology edited by Derek. B. Scott (Farnham and Burlington:
Ashgate, 2009).
The track was allegedly recorded live at the (now closed) Milk Bar nightclub in London. Using a
handheld microphone Björk walks through the club to the toilet as the backing track is played back
at the club’s sound system, see Ian Gittins, Björk: Human Behaviour: The Stories Behind Every Song
(London: Carlton Books, 2002).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
43
ontological layer, and responds to its rhythmic and tonal structure. What is interesting is the way the causal relation between the background music and the voice seem
to change direction through the track. Björk is clearly responding to the background
music, as she sings along to it, and her performance is, thus, logically located after the
events heard in the background recording. The choir voices heard dubbing her voice at
0’47, however, announce a sense of simultaneity, and the causal chain even reverses at
1’04 where the choir seems to respond to Bjork’s voice. Taking these events literally we
are presented with a spatial and temporal paradox, arising from the experience of perceived performers responding to events that takes place after the moment of responds.
Extra-diegetic sound events
Extra-diegetic sound events refer to anything “outside” the performance universe, that
is, anything the performers on-stage cannot hear. We find an example of this in Air’s
“Radio #1” from their album 10.000 Hz Legend.50 At 3’14 we suddenly hear a male
voice humming along to the chorus. The male voice sings along to the track and, thus,
takes part in the performance. He appears, however, temporally separated from it.
When the voice enters, the listener suddenly hears the initial performance through another diegetic layer. In this way the listening perspective changes. This is achieved by
ducking the diegetic sound events with side chain compression when the male voice
enters. This effect is common in many documentary films, television commercials,
and in radio broadcasting when the speaker presents a track, while we hear the music
in the background. In television these voices are called voice-overs. In the purely auditory domain they do not have a name, but as they are clearly comparable to the voiceovers we encounter in audio-visual media, phonographic voice-overs seems an appropriate term here.
Considering the title and the lyrics of the song, the impression of a radio speaker
humming along to the track immediately comes to mind. There are, thus, two temporally separated perceived performances: (1) The radio speaker’s performance that enters at 3’14; and (2) the performance formed by the rest of the sound sources. The
radio speaker is addressing the diegetic performance in a very direct sense by singing
along to it. He is, in one sense, pointing us in the direction of the song; in another
sense it can be perceived as an intrusion, interrupting the relationship between the performance and the listener. The listener must zoom out when the radio speaker enters
to embrace both performances. In this way, the diegetic performance seems more distant in time and space when the extra-diegetic voice is forced upon the listener.
Where Air’s “Radio #1” plays with an extra-diegetic voice, the Japanese noise rock
band Boredoms plays with extra-diegetic medium sounds in the track “Super You”
from the album Super AE51. The track has no clear melodic or rhythmic structure, but
consists of layers of distorted guitars floating in an out of each other. This noise atmosphere is disrupted several times during the track by the sound of a tape reel being
50
51
Air, “Radio #1,” 10.000 Hz Legend (Virgin, 2001).
Boredoms, “Super You,” Super æ (Warner Music, 1998).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
44
Mads Walther-Hansen
fast-forwarded several times – as if someone else were winding the entire track for us.
The effect is most profound in the following intervals 0’30-0’32, 0’55-1’00, 3’31-3’51
and 5’16-6’30. Several clearly audible tape splices further enhance the impression of a
tape recording. The tape-reel effect may be seen as an instance of diegetic layering. The
tape-machine is not an effect in the performance, but a medium that belongs to another ontological world than the perceived performance. Contrary to the extra-diegetic
layer in “Radio #1,” however, the extra-diegetic layer in “Super You” lacks human presence. Instead the tape machine may be seen as a substitute for the human narrator. It
is essentially the tape-machine that makes it possible for us to hear the performance –
it tells the story of another time and place. The medium-within-medium effect found
in “Super You” is interesting precisely because it changes the listener’s otherwise direct
relation to the performance through the recording, by adding an extra ontological layer between the listener and the performance.
Discussion
We are now in a position to return to the Queens of the Stone Age example mentioned in the beginning of the article. In “You Think I Ain’t Worth a Dollar, But I Feel
Like a Millionaire” we are presented with two listening perspectives: (1) through the
position of the car driver hearing music through the car radio and (2) as a “directly”
mediated performance presented to us through our own loudspeakers. Like this, we
hear movement from a meta-diegetic to a diegetic level. The meta-diegetic level “eats
up” the diegetic level, enhancing the experience of loudness and intensity. While we
may say that the song is presented to us at a meta-diegetic level – as a consequence of
its location in the car radio – the meta-diegetic level quickly obliterates as the quality
of the sound changes, removing the car setting from experience.
While the physical division between sound and its natural source – the possibility to move sound to another host – became reality with the invention of the phonograph in 1876, the possibility to layer sounds, and potentially create diegetic layering of perceived sound events, emerged with sound-on-sound recording several decades later. After some of Les Paul’s first experiments with sound-on-sound recording,
Patti Page working with Mitch Miller on the Mercury Record Label used this idea to
sing along to herself on her 1947 recording of “Confess.”52 This recording is the first
known example of sound-on-sound recording using tape, and was marketed as “Patti
Page and Patti Page,” suggesting how this effect was seen, as least partly, as an act of
vocal gimmickry.
Of course, the emergence of multitrack technology in the 1950’s quickly initiated
many examples of singers dubbing themselves, and today the practice is so conventional that most listeners will not even notice the overdubs that have gone into making the many layers of voices they are presented with on different recordings. Listening (back) to Patti Page’s recording with today’s ears, there is, however, something dis52
Patti Page, “Confess” (Mercury, 1947).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Sound Events, Spatially and Diegesis – The Creation of Sonic Narratives in Music Productions
45
jointed about the perceived performance caused by the sonic quality of the dubbed
backing vocal. The backing vocal appears temporally and spatially displaced. The effect is unintentionally created and results from the audible technological artifacts –
such as tape hiss – arising when recording from one tape to another. Yet, listeners may
experience a media-within-effect placing the backing track on a meta-diegetic level.
In some instances, knowledge about the original event behind the perceived performance may have an effect on the listening experience. Such is the case with Natalie
Cole singing a duet with her late father Nat King Cole on “Unforgettable.”53 Although
there is no clear indication in the actual sound that Natalie Cole’s voice was recorded
30 years after Nat King Cole’s voice, our knowledge about the making of the song may
intrude on the listening experience. Natalie Cole is mediating the past event being
both listener of her father’s recording, while performing to it and, in effect, creating
a new one. In so far as we may mark Nat King Cole’s voice as a past event, his voice
is, in a very peculiar way, both diegetic and non-simultaneous. Although the contextual story accompanying the virtual duet is noteworthy, the temporal separation between recorded takes is in no way uncommon. We may, for instance, think of rap vocals heard performing to sampled phrases with a clear origin in past recordings. As in
“Unforgettable,” the temporal separation in rap music is mostly not an audible feature
of the sonic material, but rather contextual information that impacts on the listening
experience.
Conclusion
In this article I have suggested how the concept of diegesis may work to classify sound
events in recorded music and how the concept is useful to account for the production
of meaning in music production. With reference to Genette’s classification of narrative levels,54 I have argued that diegetic layering of sounds within the virtual space
of music recordings can be classified as meta-diegetic, diegetic or extra-diegetic. Although the discussed tracks shows examples of unusual spatial distributions of sound
events, they support the argument that it makes sense to think of a phonographic diegesis that can be exploited creatively in music production to create different forms of
narrative effects.
The idea builds on the performance metaphor and the belief that listeners hear recordings as documents of a fictional past event. I have argued that diegetic layers in
recordings should be defined primarily according to the perceived temporal relationship between sound events. Meta-diegetic sound events are either performer-internalsounds or pre-recorded performances played back through a media in the diegesis.
Whereas film sound theorists often pay most attention to the former, I redefine meta-diegesis to account for the experience of a media-within-media effect. The notion
of meta-diegesis then applies when we experience that diegetic performers (pres53
54
Natalie Cole, “Unforgettable,” Unforgettable, With Love (Elektra, 1991).
Gérard Genette, Narrative Discourse: An Essay in Method (Ithaca and New York: Cornell University
Press, 1972/1980).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
46
Mads Walther-Hansen
ent events) can “hear” meta-diegetic performers (past events), but not the other way
around. Likewise extra-diegetic sounds emerge as a media-within-media effect. Extradiegetic sound events, however, appear after the diegesis and thus logically become
the new present as the diegesis is relocated to the past.
As I demonstrated through the examples the, diegetic layering of phonographic
sound events may take various forms, and often the listener is presented with ambiguous or even conflicting acoustical cues. In addition, I suggested that previous knowledge about the production of the track have an effect on the nature of the perceived
performance. In this way, the notion of the phonographic diegesis may open up for
further reflection on the experiential aspects of virtual spatiality in recorded music.
Abstracts
The use of studio-based effects and recording techniques to shape the sound quality
and aesthetics of recordings are addressed in a number of studies. This article explores
how such practices have shaped the sonic narrative of recordings and listeners’ point
of audition. Building on research into literature and film sound, and by considering
the spatiality of a number of popular music productions, I outline the concept of diegesis in the context of recorded music. I conclude by proposing a way to categorize
sound events relative to the perceived performance space of recordings.
En lang række videnskabelige publikationer har undersøgt den måde hvorpå optageteknikker og efterbehandling af musik i lydstudiet har formet optagelsens æstetik.
Denne artikel har fokus på hvordan det auditive narrativ samt lytterperspektivet er designet i lydstudiet. Med afsæt i litteraturstudier og filmlydstudier, og igennem korte
analyser af en række populærmusikproduktioner, argumenteres der for en redefinering
af diegese-begrebet og en kategorisering af lyd i optagelsens virtuelle performancerum.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
HÅLLBUS TOTTE MATTSSON
Audiovisuella loopar och kollektiv
live-looping
Upprepningens magi har gett upphov till en mängd kompositionsstrukturer och ligger till grund för ett otal musikstilar. Alltifrån renässansens variationsformer över fasta
ackordstrukturer över bluesens 12-takters formler till dagens technomusik. När det idag
talas om loopar och looping menas en upprepning av ett musikaliskt parti, både som
en teknisk arbetsmodell (man kan t.ex. sätta något i loop i ett sekvenser-program för
att prova en idé, öva eller snabbt få in ett antal tagningar) och som ett konstnärligt redskap där en artist eller kompositör skapar musikaliska sekvenser vilka återupprepas,
överlagras och moduleras med hjälp av olika tekniska hjälpmedel. En sub-genre eller
snarare en form för musicerande som vuxit fram ur detta arbetssätt är livelooping där artister med hjälp av loopteknologi framträder med oftast väl inrepeterade föreställningar
och där tekniken gör det möjligt att bygga ett ofta imponerande “one man band”.
Loopteknik
Tekniken som gör detta musicerande möjligt är olika former av “delays”, oftast konstruerade så att de kan styras med hjälp av fotpedaler. Mer avancerad teknisk utrustning gör det också möjligt att bygga överlagringar (pålägg) och på så sätt successivt
skapa ett allt komplexare ljudlandskap. Med pionjären loopmaskinen Ecoplex Digital
Pro, som kom ut på marknaden i början av 1990-talet, så introducerades möjligheten
att också kunna synkronisera flera loopmaskiner med varandra och på detta sätt också kunna loop-musicera tillsammans i ensembleform. Denna möjlighet till kollektivt
liveloopande är dock till stora delar fortfarande ett outforskat fält.
Projektets bakgrund och frågeställningar
Denna artikel beskriver det konstnärligt/tekniska forskningsprojektet “Audiovisuella
loopar” som genomförts på Högskolan Dalarna av Hållbus Totte Mattsson, Henrik
Carlsson, Andrew Scott och Sören Johansson. Projektet finansierades av EU:s regionaliseringsfond och Region Dalarna och pågick under perioden jan 2011 till och med
november 2012.
Projektet tog avstamp i ett mindre pilotprojekt “interaktiva loopar” som genomfördes vid Högskolan Dalarna under HT09/VT10. Detta projekt hade som mål att ta
fram teknik och programvaror samt arbeta fram en modell för ett interaktivt konstDANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
48
Hållbus Totte Mattsson
närligt agerande mellan teknologi, tekniker och musiker. Konceptet provades sceniskt
bland annat i samband med ett framträdande med vevlire-duon Hurdy-Gurdy och
den amerikanska stråkkvartetten Kronos Quartet i Carnegie Hall, New York 13/3-10.1
Med dessa erfarenheter som utgångspunkt formulerades Projektet “Audiovisuella
loopar” där interaktion mellan flera musikanter i kombination med loopade visualiseringar varit i fokus.
Projektet utgick från önskemål och frågor från etablerade loopartister om att kunna integrera visuell looping för att förhöja närvarokänslan (“live”-upplevelsen) i sina
framträdande. Är det musiken eller att det liveloopas som är upplevelsen för publiken? Är det egentligen någon skillnad i om man startar ett konventionellt “backing
track” eller en uppspelning av den just inspelade loopen när man trycker på knappen? Vad innebär det om den just inspelade loopen visualiseras samtidigt som den
spelas upp? Dessa frågor och funderingar ledde framtill formulerandet av följande
frågeställning.
Hur kombinerar man visuell (live-)looping med “traditionell” liveloop-teknik
för att förhöja närvaro-känslan i ett framträdande?
En annan central frågeställning handlade om teknikens användarvänlighet i samband
med att den används av flera musikanter i ensembleform och hur den påverkar musicerandet.
Kan man skapa förutsättningar för ett kollektivt musicerande med loop-tekniken som gör det möjligt att enkelt kunna starta ett kreativt samspel med tekniken, oavsett tekniska förkunskaper eller genrebakgrund?
Projektets syfte och mål
Syftet blev således att designa ett system och en arbetsmetod för kollektiv livelooping
både auditivt och visuellt. Systemet skulle bygga på befintlig teknik och programvara och ligga inom rimliga prisramar för en användare. De ramar vi satte upp inom
projektet var att både det visuella och auditiva materialet skulle byggas upp av loopar
skapade i stunden, det vill säga inget förinspelat material eller färdiga videoklipp fick
användas. Projektet har haft tre delmål:
– Att anpassa programvaror och finna arbetssätt för audiovisuell live-looping utifrån
önskemål av artister från livemusikscenen.
– Att prova ut den kollektiva livelooptekniken med musikanter från olika genrer.
– Att i ett sceniskt verk visa på den audiovisuella loop-teknikens konstnärliga möjligheter.
1
Allan Kozinn, “From the Top of the World: Warmth Amid Ice,” New York Times, March 3, 2010, besökt 22 oktober 2015,
http://www.nytimes.com/2010/03/15/arts/music/15arctic.html?scp=1&sq=hurdy-gurdy&st=cse&_
r=2&
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Audiovisuella loopar och kollektiv live-looping
49
Historisk återblick
Nya musikaliska idéer baserade på minimalism, seriell musik och utomeuropeisk musik bland efterkrigstidens unga tonsättare, inte minst John Cage, och utvecklingen av
bandspelaren blev avstampet för ett experimenterande med musik som också använde
inspelningar. Tidiga försök gjordes redan under slutet 40-talet med ljudeffektinspelningar på grammofonskivor (musique concréte) av bland annat Pierre Schaeffer2.
Med hjälp av manipulerade bandinspelningar samt flera sammankopplade bandspelare skapade Lester William “Les Paul” Polfuss3 och Karlheinz Stockhausen m.fl under
50-talet nya metoder för musikskapande som landade i möjligheter att på olika sätt
spela in, överlagra och manipulera ljud4.
Intressant är att se hur framförandeformen (konsertformen) för dessa verk skapar
olika lyssnings-estetik. Dels har vi musik uppbyggd av enbart inspelningar som sedan
spelas upp med hjälp av högtalare (acousmatic music) och utan att det finns en utövare på scenen vilket blir ett “nytt” lyssnande för en konsertpublik (acousmatic listening). Begreppen introduceras av Pierre Schaeffer som lånat dem från hur Pythagoras
undervisade sina elever dold bakom ett förhänge (acousmatic) – “in ancient times,
the apparatus was a curtain; today, it is the radio and the methods of reproduction”5.
Här ryms allt från bandkompositioner av t.ex. Stockhausen, skapade med ett slags
“bandslöjds-teknik” till blandformer av inspelningar tillsammans med “levande” musiker. I det senare fallet är ljudbandet en tydlig separat del av helheten som publiken
är fullt medveten om. Sedan har vi artister som använder sig av bandspelaren för att
spela upp backing-tracks i sina live-framträdande där inspelningen skall vara en för
publiken “omärklig” del av framförandet. Här hittar vi t.ex. pionjären Les Paul med
sin egenutvecklade “Paulverizer”, en styrenhet att montera på gitarren för att kunna
starta och sköta den “dolda” bandspelaren som spelade upp förinspelade grunder att
spela till och att också spela in pålägg över. Ett exempel på hur autenticitet-kravet från
publiken, och kanske även artisten, styr teknikutvecklingen och hur den presenteras
– framträdandet fick inte upplevas som play-back trots att det delvis (och ibland till
stor del) var det.
I början av 60-talet bildades San Fransisco Tape Music Center som blev ett forum
för unga avantgardistiska tonsättare, bl.a. Terry Riley, Pauline Oliveros, Ramon Sender och Steve Reich, med ett fokus på minimalistisk musik med bandspelaren som
redskap6. Även andra inspelningsredskap provades bland annat Maestro Echoplex,
ett bandeko med justerbart tonhuvud vilket gav möjligheter att styra delay-tiden och
att spela in överlagringar. Detta bandeko introducerades av Ramon Sender för Terry
Riley som sedan använde det i sin första loopbaserade komposition “Mescalin Mix”
(1961). När Terry Riley beskrev hur “Mescalin Mix” kommit till för en fransk radio2
3
4
5
6
Thom Holmes, Electronic and Experimental Music (London: Routledge 2002), 90f.
Greg Milner, Perfecting Sound Forever – the Story of Recorded Music, (Croydon: Granta 2010), 125.
Holmes, Electronic, 95f
Pierre Schaeffer, Acousmatics. C. Cox & D. Warner (red.) Audio Culture: Readings in Modern Music
(New York: Continium), 77.
Holmes, Electronic, 78.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
50
Hållbus Totte Mattsson
ingenjör vid RTF:s (Radiodiffusion-Television Francais) radiostudio i Paris 1963 och
ingenjören då satte en bandloop mellan två bandspelare, där den ena använde inspelningshuvudet för inspelning och den andra använde det för uppspelning, skapades prototypen för dagens livelooping med möjligheter till feedback och overdub7.
Systemet som Riley kom att kalla “Time Lag Accumulator” är i princip samma uppkoppling som sedan använts av Robert Fripp och Brian Eno bland annat vid inspelningen av det klassiska albumet “No Pussyfooting” utgiven 1973. När Fripp sedan i
slutet av 70-talet började ge solokonserter med livelooping kallade han systemet för
“Frippertronics” (i princip identiskt med Rileys “Time Lag Accumulator”). Att Riley
inte ger ett erkännande till den ingenjör som faktiskt är uppfinnaren av “Time Lag
Accumulatorn” är en smula märkligt då det borde varit lätt för honom att ta reda på
hans namn. En trolig kandidat är Jaques Poullin som var innovatör och tekniker tillsammans med Pierre Schaeffer mellan 1958 och 1975 vid GRM (Groupe de Recherche de Musique Concrète) en avdelning vid RTF8. Om Riley verkligen (med hjälp av
den “okände” franske ingenjören) var först med att konstruera detta system är dock
ett ämne för diskussion, snarlika uppkopplingar användes av Mauricio Kagel i verket
Transición II för piano, percussion och två bandspelare (1958-59) och även av Vladimir Ussachevsky, Pauline Oliveros och Ramon Sender. Man kan kanske säga att livelooping helt enkelt låg i tiden.
En banbrytande kompositionsmetod som utvecklades av Steve Reich som ligger
nära loop-tekniken var “gradual phase shifting” även kallad “phasing”. Detta innebär
att flera identiska fraser repeteras men inte riktigt håller synk utan istället glider de
isär och skapar då hela tiden nya mönster och upplevelser utan att man behöver tillföra något nytt musikaliskt material.9 Ett tidigt exempel är kompositionen “It´s gonna
Rain” (1965). Denna kompositionsmetod blev en viktig inspirationskälla för Brian
Eno i skapandet av sin “Ambient music” ett tidigt exempel är albumet “Discret Music”
(1975) där också en “Time Lag Accumulator” används10.
Nästa stora kliv i loopteknik-utvecklingen blev Lexicons loop-maskin PCM42 som
konstruerades av Garry Hall och kom ut på marknaden 1981. Det var ett digitalt delay
som också gick att kontrollera med hjälp av en fotpedal vilket öppnade upp möjligheten att kunna spela ett instrument med händerna och samtidigt styra live-loopingen
med fötterna. Detta arbetssätt är än idag det vanligaste och utforskades först av tonsättare och artister som Pauline Oliveros, David Torn och Mattias Grob. Flera loop-maskiner kom ut på marknaden under slutet av 80-talet men blev inte några försäljningssuccéer. Lexicon gjorde ett nytt försök med Jamman, utvecklad av Garry Hall och Bob
Sellon som släpptes 1993. En välkonstruerad och prisvärd maskin som tyvärr snart
togs ur produktion på grund av för liten efterfrågan.
7
8
9
10
Holmes, Electronic, 83.
Evelyn Gayou, “The GRM: landmarks on a historic route”. Organised Sound, Vol.12 (3) (Cambridge:
University Press, 2007), 203-211.
Robert Fink, Repeating ourselves: American minimal music as cultural practice (Los Angeles: University of California Press, 2005), 106f.
Brian Eno, Ambient Music. C. Cox & D. Warner (red.) Audio Culture: Readings in Modern Music (New
York: Continium, 2004), 94-97
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
51
Audiovisuella loopar och kollektiv live-looping
En förklaring kan vara att loop-maskiner konstruerades efter en musikalisk estetik
(solistisk med långa linjer utan tydliga form-delar osv.) som inte attraherade mer populärmusik-orienterade köpare. I slutet av 80-talet kom också den kanske mer lättarbetade samplern som man kan säga också är en form av loop-maskin men som istället
använder redan inspelade loopar som lätt går att växla emellan. Möjligheter som livelooping maskinerna dittills saknat.
En del av dessa begränsningar kom Matthias Grob förbi med hjälp av en av Garry
Hall ombyggd PCM42:a. Han var dock inte nöjd utan längtade efter ett mer musikaliskt och intuitivt loop-instrument med möjlighet att bland annat växla mellan flera olika formdelar. Tillsammans med programmeraren Eric Obermühler och senare
också Kim Flint konstruerade han i början av 90-talet sin Loop Delay som, med hjälp
av Keith McMillen chefstekniker på Gibson, släpptes ut på marknaden under namnet
Oberheim Echoplex Digital Pro 1994.11 Denna maskin är modellen för konstruktionen av mjukvaran Ableton Live:s plugin Looper som ligger till grund för systemet som
presenteras i denna artikel.
Oberheim Echoplex Digital Pro
Oberheim Exhoplex Pro är en viktig milstolpe för utvecklingen av loop-tekniken. Med
namnet Echoplex knyts historien ihop då det också var namnet på det bandeko som
Terry Riley använde i sin första loopbaserade komposition “Mescalin Mix” som beskrivits ovan.
Denna maskin innehöll en fantastisk massa möjligheter men var inte riktigt färdigutvecklad när den kom ut på marknaden. Den innehöll en del buggar och inköpspriset var mycket högt.
Många av Echoplexens möjligheter har sedan kopierats och legat till grund för den
fortsatta teknikutvecklingen och idag finns det ett flertal olika välfungerande loopmaskiner, både som mjukvaror och separata moduler, t.ex. den mycket populära
Boss RC-300. Echoplexen innehöll dock en unik möjlighet som inte kopierats – BrotherSync – en funktion som gjorde det möjligt att enkelt koppla ihop flera Echoplex
som då höll en gemensam synk och där också valfri enhet kunde starta en ny session.
Denna funktion har dock varken kommit till någon större användning eller utvecklats vidare fram tills nu. Den beskrivs i manualen på följande sätt.
any unit can define the basic cycle time just by recording a loop. The other units
can then be synchronized to that loop time. Have an Echoplex jam session! If
you think that the Echoplex opens up a lot of possibilities for a single player,
imagine the possibilities for multiple performer.12
11
12
Matthias Grob, “How the Gibson/Oberheim Echople CameTogether”, 1996, besökt 22 oktober
2015, http://www.loopers-delight.com/tools/e3hoplex/OBechoplexhistory.html
Echoplex, Digital Pro User´s Guide,1994, besökt 22 oktober 2015, http://www.loopers-delight.com/
tools/echoplex/echoplex.html
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
52
Hållbus Totte Mattsson
Att ta vara på och utforska dessa möjligheter till loop-musicerande i ensembleform
“kollektiv livelooping” var en av de centrala utgångspunkterna för projektet. Detta
ledde till följande två kravspecifikationer på det system som konstruerades:
– Flera musikanter skall kunna loopmusicera tillsammans i perfekt synkronisation.
– Vem som helst skall kunna starta en session och då definiera ett gemensamt tempo.
Vad är Live?
När Thomas Edison 1877 för första gången läste in “Mary had a little lamb” i tratten
till fonografen och sedan vände på processen och spelade upp inspelningen så uppstod samtidigt dikotomin “live” kontra inspelning. Auslander menar att innan inspelningstekniken (mediatization) var “live” (liveness) meningslöst – en ickefråga – begreppen är helt beroende av varandra.13 Vidare så innebar samtidigt detta att musik
som hittills alltid upplevts i någon form av audiovisuell kontext helt plötsligt förekom i en enbart auditiv form. Ur detta startade en ny visualiseringsprocess där lyssnarens fantasi och föreställningar tillsammans med en allt starkare kommersialisering
kring musiken skapade en stark visuell kultur kring de framväxande genrerna och subgenrerna men också att live-begreppet omformades.14 Auslander utgår ifrån rockmusik från 60-talet och framåt och problematiserar kring hur framförandet (live) alltmer
gått mot ett återskapande av fonogram-inspelningen. Om framförandet sker fullständigt playback eller inte är för fansen idag helt oväsentligt så länge artisten finns på
scenen och agerar i någon form. Det upplevs av publiken fortfarande som en autentisk live-konsert som förmodligen skulle förlora i trovärdighet om musiken skiljde sig
för mycket från inspelningen. Värt att notera är att Auslander läser av dessa tendenser
15 år före Avicii och fenomen som Hatsune Miku. Avici (Tim Bergling).15 är ett av
de större namnen inom dagens DJ-kultur och framträder på enorma arenor med sin
musik som framförs till största delen playback. Hatsune Miku16 som är en helt konstgjord artist, ett hologram med en datorproducerad röst, turnerar och framträder “live”
för stora publikskaror i “sitt” hemland Japan. Begreppen “autenticitet” och “live” blir
enligt Auslander definierade av de kulturella och ideologiska konventionerna som gäller inom genren, konventioner som hela tiden förändras (musikvideos, “backdrops”,
“visuals”, skivkonvolut, kläder, frisyr osv.). Edisons inspelning av “Mary had a little
lamb” har i denna medieringsprocess omformats till att bli “live”.
Denna medieringsprocess är i live-looping koncentrerad till en i stunden pågående
process där livemusicerandet transformeras till medierad form. Kombinerar vi detta
med filmloopande så blir upplevelsen faktiskt att vi samtidigt vänder på processen –
det medierade framförandet, musikloopen, blir “live-eriserad” som en filmloop. Diko13
14
15
16
Philip Auslander, Liveness:Performance in Mediatized Culture. (New York: Routledge, 1999), 42f.
Auslander, Liveness, 53f.
Tim Bergling, Avicii – Ultra Music Festival 2015, konsertvideo, besökt 21 november 2015, https://
www.youtube.com/watch?v=ec8CsjAtg1k
Hatsune Miku, Crypton Future Media – Hatsune Miku –World is Mine, 2010, video, besökt 22 oktober
2015, https://www.youtube.com/watch?v=pEaBqiLeCu0
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
53
Audiovisuella loopar och kollektiv live-looping
tomin mellan inspelning och liveframförande blir upphävd och vi upplever en syntes
mellan de båda. Detta ger upphov till ett helt annat förhållningssätt till begreppet
“live” då processen och flödet också fortsätter även om loop-artisten gör ett uppehåll
i sitt eget musicerande.
Audiovisuell Live-looping vad är det?
Att använda visuella medel och metoder i scenkonst för att förstärka (musik)upplevelsen har en lång historia alltifrån ljussättning och målade kulisser framtill dagens virtuella scenografi och enorma digitala “backdrop” skärmar. Kopplingen mellan musik och
det nya mediet film och dess estetiska möjligheter blev ett nytt fält för 1920- och 30-talets avantgardistiska konstnärer. Med namn som Oskar Fischinger, Viking Eggeling,
Walter Ruttman och Hans Richter skapades genren “Visual Music”17 som är en viktig
inspiration för dagens VJ-artister. Inom VJ-ing utgår man från förinspelade videoklipp
som processas i realtid med olika mjukvaror och spelas upp synkront med musiken.18
Vår visuella ambition var att kunna fånga stunden då musikloopen skapades och sedan projicera den som en filmloop på en “backdrop” där den även kunde vävas ihop
med övriga rörliga bilder från framförandet på scenen. Till skillnad från hur materialet
för en VJ-artist skapas så skulle vårt filmmaterial skapas i stunden och automatiskt spelas upp i synk med musiklooparna. Vår första vision var att den visuella loopingen skulle fungera som en “spontan” förlängning på loop-musicerandet (utan teknisk support).
Detta visade sig dock inte fungera tillfredsställande. För att fungera och bli konstnärligt
intressant så krävdes det en medskapande videokonstnär. Auslander använder begreppet
“Mixed-media performance”19 för ett framträdande som blandar levande framträdanden
med film, video, projiceringar osv., ett begrepp som passar bra i detta sammanhang. Till
skillnad från den elektroniska videoscenen, där mycket skapas som fasta installationer
av artister som Ryoji Ikeda och Carsten Nicolai, är det centrala för vårt projekt att det är
ett live-framträdande – det händer här och nu – Audiovisuell Live-looping.
Livelooping idag – olika arbetssätt, olika scener, olika filosofi
Den ursprungliga arbetsmetoden och filosofin för livelooping styrdes av teknikens
begränsningar. Artisten började med en tom loopmaskin i vilken musiken skapades
och byggdes upp i stunden och var borta så snart framförandet var över och man
drog ur kontakten. Att spara materialet var en omständlig process som krävde att man
tog fram och kopplade in någon form av inspelningsutrustning. Nu räcker det oftast
med ett enkelt klick med musen på datorn och sedan finns hela sessionen sparad.
Idag är tekniken även på andra sätt mycket mer sofistikerad och begreppet
livelooping täcker in ett flertal arbetsmetoder med olika estetik och filosofi beroende
17
18
19
Brian Evans, “Foundations of a Visual Music” Computer Music Journal Vol.29 (4) (The MIT Press,
2005), 11-24.
Michael Faulkner, VJ, audio-visual art + vj culture (Laurence King Publishing, 2006).
Auslander, Liveness, 36.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
54
Hållbus Totte Mattsson
på hur artisten vill presentera sig för publiken. För en del artister är det centralt att
publiken är medveten om och kan uppskatta “loopandet” vilket blir en viktig del i
konsertupplevelsen. För andra är loop-tekniken ett hjälpmedel för det musikaliska
uttrycket utan något egenvärde och inget som publiken behöver känna till. Slutligen
är loop-tekniken för ytterligare andra ett hjälpmedel för att öva, komponera och improvisera. Ett annorlunda sätt att musicera med och för sig själv.
Loop-tekniken som en inrepeterad hårt arrangerad del i ett framträdande.
Tekniken möjliggör för artisten att göra “det omöjliga” med avsikt att uppträda live
inför en publik. På detta sätt arbetar t.ex. den svenska numera i New Orleans boende loop-artisten Theresa Andersson. Hon började använda loop-teknik av ekonomiska orsaker då hon helt enkelt inte hade råd att ta med sig ett band. Hon komponerar inte låtarna med hjälp av loopteknik utan “transkriberar” färdiga arrangemang
till loop-arrangemang. Dessa övar hon sedan in och framför dem noggrant inrepeterade “live” inför en konsertpublik eller videoinspelade och utlagda på YouTube där
de fungerar som en viktig marknadsföringskanal för henne.20 Hennes image och framträdanden bygger på att publiken är medveten om och uppskattar live-loopandet. För
andra är loop-tekniken ett hjälpmedel för det musikaliska uttrycket utan något egenvärde och inget som publiken behöver känna till. Se till exempel hur den engelske
soloartisten Ed Sheeran utnyttjar loop-teknik i sina solo-framträdanden.21 På ett delvis
annorlunda sätt jobbar Third Brother22 en kollektivt liveloopande popduo som använder ett egenutvecklat system där båda liveloopar mot ett förprogrammerat klicktrack. Denna teknik är snarlik det sätt som Les Paul och Mary Ford utvecklade redan
under 50-talet. Se t.ex. deras demonstration av sin version av “How high the moon” i
Alistair Cooke’s TV program “Omnibus” från 1953.23
Loop-tekniken som en möjlighet att fånga ögonblicket.
Att starta från noll och låta loopkompositionen växa fram som en improvisation i dialog med sig själv är den “klassiska” formen av livelooping. Detta är ett sätt där man
bejakar teknikens ögonblickliga feedback och låter den pågående processen styra musicerandet “det kan aldrig bli fel – bara annorlunda”. Den ungerske psykologen Csíkszentmihályi använder i sin banbrytande forskning kring kreativitet begreppet “flow”24
20
21
22
23
24
Carina Jönsson “Loopen får en att låta som många”, Dagens Nyheter, 3 december, 2012, besökt 22
oktober 2015, http://www.dn.se/kultur-noje/musik/loopen-far-en-att-lata-som-manga/
Ed Sheeran, Multiply Live in Dublin, konsertvideo, 2014, besökt 22 oktober 2015. https://www.youtube.com/watch?v=lR3vIjAZDnM
Third Brother “Videodokumentation av en konsert på Gamla Elverket, Falun”, 24 mars 2012, besökt
22 oktober 2015, http://bambuser.com/v/2499152
Les Paul och Mary Ford “How high the moon ”, TV-program, 23 oktober 1953, besökt 22 oktober 2015,
https://www.youtube.com/watch?v=VCEmAgak9V8&index=1&list=PL8NlZ84fylP1MI8KYEXHqd4iwjl
CSx3yH
Mihály Csíkszentmihályi, Creativity: Flow and the Psychology of Discovery and Invention, (New
York: Harper Collins, 1996)
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
55
Audiovisuella loopar och kollektiv live-looping
vilket passar väl in på detta musicerandesätt. Han beskriver “flow” som ett tillstånd
där man är helt uppslukad i sitt handlande utan plats för självmedvetna reflektioner
och med en stark känsla av att ha full kontroll.
Loop-artisten Samuel “Loop-Tok” Andersson25 har utforskat detta kreativa “flow”
som livelooping skapar när han varit musiker till dansklasser och det centrala varit ett
i stunden skapande/komponerande utan tid för analytisk reflexion och möjlighet att
starta om processen. Ett arbetssätt som han sedan tagit med sig upp på scen och där nerven och närvaron växer ur att publiken känner och är medveten om att vad som helst
kan hända precis här och nu. Denna arbetsmetod är den samma som Terry Riley använde sig av under mitten av 60-talet då han gav improviserade, ibland flera timmar långa,
performance-konserter med sin “Time Lag Accumulator” på olika scener i New York26.
Loop-tekniken som ett låtskrivar-/förproduktions-redskap
Loop-teknikens direkta återkoppling och “flow” gör den till ett mycket kreativt kompositionsredskap. Ett redskap som fungerar som en “jante-killer”, dvs. hindrar ett negativt reflekterande självmedvetande, och där intellektets eftertanke aldrig får chansen
att stoppa de kreativa processerna. Detta sätt att använda loop-teknik för att komponera/arrangera musik provades t.ex. av gruppen Hedningarna i arbetet med förproduktionen av cd-skivan & (2012). Bakgrunden var att bandet ville hitta ett annat sätt
att bygga arrangemang och bakgrunder än att “hänga framför skärmen” och programmera, ett arbetssätt som oftast exkluderar alla utom “dataexperten”. Utgångsmaterialet bestod av en fragmentarisk hög av olika melodier, snuttar och idéer och möjligheten att i stunden kunna pröva, utvärdera och kombinera materialet utan att processen
hela tiden stannade av gav helt nya möjligheter till ett gemensamt kreativt skapande.
Dessa live-loopade bakgrunder togs med till studion och blev sedan en grund för den
fortsatta produktionen.
Artisten samarbetar med en (oftast dold) loopare
YouTube är idag ett viktigt forum också för livelooping vilket lett till flera intressanta
lösningar för artisten att visa publiken att “det är på riktigt”. Ett exempel är Gianni and
Sarahs version av Radioheads låt “Karma Police” 27 där vad som händer “bakom scenen” presenteras i en separat video28. Detta dras till sin spets i videon “Roll Up” av ensemblen Walk off the Earth29 där vi får följa en snitslad liveloop-bana och där rollerna
25
26
27
28
29
Samuel Andersson, Livelooping- – att fånga ögonblicket, Videodokumentation, 31 januari 2012, besökt
22 oktober 2015, https://www.youtube.com/watch?v=N4caRcRCKBg
Holmes, Electronic, 172.
Gianni och Sarah (2011a) Karma Police-Gianni and Sarah (Walk of the Earth), video, 25 oktober 2011,
besökt 22 oktober 2015, http://www.youtube.com/watch?v=yE2kANx8WdA
Gianni och Sarah, Karma Police, Behind the Scenes (Gianni and Sarah) , video, 26 oktober 2011, besökt 22 oktober 2015, http://www.youtube.com/watch?v=5hsXYRc6vWg
Walk off the Earth, Roll Up-Walk off the Earth, video, 26 april 2011, besökt 22 oktober 2015, https://
www.youtube.com/watch?v=tJ8nkJ1_Ee0
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
56
Hållbus Totte Mattsson
tydligt delas upp i “The Looper” (den som sköter loop-tekniken) och “The Loopee”
(artisten som utför musiken som loopas). Denna video är uppdelad i två bildrutor
som visar dessa båda funktioner parallellt så publiken förstår att “detta händer just nu”.
Förutom den “frihet” att koncentrera sig på musiken och uttrycket istället för att
behöva fokusera på knapptryckandet så ger detta arbetssätt nya möjligheter till interagerande med en (mer eller mindre synlig) tekniker. Detta öppnar upp för många
spännande möjligheter där man t.ex. kan tänka sig ett band där en eller flera av medlemmarna loopar upp sig själva eller varandra och dessa loopar sedan blir med och
påverkar fortsättningen av improviserandet/musicerandet. Detta sätt att arbeta har
bland annat den amerikanska stråkkvartetten Kronos Quartet utforskat i sitt samarbete med den inuitiska sångerskan Tanya Tagaq, där delar av musiken loopas av teknikern och skickas tillbaks till scenen för att generera nya impulser och infall i musiken.
Enligt uppgifter från Brian Mohr, ljuddesigner för Kronos Quartet, vid seminariet “Dagens teknik som konstnärligt redskap” med Kronos Quartet och deras tekniska medarbetare på Uppsala Konsert & Kongress den första november 201230. Ett annat scenario
är att “tonsättaren” också hanterar loopteknologin och bygger en interaktiv komposition av olika sekvenser från bandet som samtidigt ger impulser och feedback till det
fortsatta skapandet. Se t.ex. tonsättaren och instrumentinnovatören Paul Dreschers
samarbete med Steven Schick31. Ett ytterligare exempel är den konstmusikaliska liveloop-duon essens:132 och deras samarbeten med olika tonsättare som specialskrivit
kammarmusikaliska verk för deras sättning klarinett, fiol och liveloopteknik.
Andra former av looping
Looping är en viktig teknik inom dagens dans-scen med sitt ursprung i den Jamaicanska dubben som växte fram under 60-talet och erövrade club-scenerna under 70-talet.
Från början handlade det om olika tekniker där man använde LP-skivan och Maxisingeln som medium (schratch, breaks, toast, remix osv.). Ett flertal stilar och speltekniker utvecklades, bland annat “Turntablism” – en teknik där man använder skivspelarna som ett instrument. Intressant är att konstatera att en likartad spelteknik utvecklades av Pierre Schaeffer under 40-talet då han experimenterade med flera grammofonspelare vilket bland annat hörs i verken “Etudes de Bruits” som framfördes live
vid direktsändningar i franska radion 194833. Med en nära koppling till repetitiva minimalistiska musikströmningar företrädda av bland andra Steve Reich och Philip Glass
växer discon under 70-talet fram ur soul och funk i New Yorks gay-kultur. Med den
17 minuter långa remixen av Donna Summers låt “Love to love you babe” 1975 gjord
av Giorgio Moroder förebådas DJ-kulturen. Genom att loopa, manipulera och live-re30
31
32
33
http://www.mynewsdesk.com/se/uppsalakonsertkongress/pressreleases/uppsala-international-sacred-music- festival-the-kronos-quartet-2-november-805446
Paul Drescher, Schick Machine Excerpts – Final Version from Mondavi Center, video, 2 mars 2011, besökt
22 oktober 2015, https://www.youtube.com/watch?v=qS2MJwmBB5Y
Essens:1 Hysteria (Muse, Live Cover by essens:1), Video, 20 augusti 2011, besökt 22 oktober 2015, https://www.youtube.com/watch?v=XhD1JFOOdko
Holmes, Electronic, 91f.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Audiovisuella loopar och kollektiv live-looping
57
mixa musik från flera skivspelare skapar “Disk-jockey” inom DJ-kulturen långa sammanhängande dans-set med ett bestämt BPM34. Under 80-talet övergår tekniken i den
framväxande techno-musiken där ny musikteknologi som trummaskiner och samplers blir viktiga byggstenar. I en sampler lägger man in förinspelade sekvenser ur en
personlig ljudbank eller “samplade” från fonogram med kända artister. Dessa samples
loopas och spelas upp vid framförandet med hjälp av olika styrenheter (midiklaviaturer, fotpedaler, pads osv.).
Konstruktion av vårt system
Vår ursprungliga ambition var att koppla samman flera loopmaskiner och datorer
med olika mjukvaror. De synkproblem som uppstod i samband med detta ledde
dock till att vi fick tänka om och istället satsade på att försöka “stänga in” hela projektet i en dator vilken kunde styras med ett antal fotpedaler.
Vi valde att gå vidare med mjukvaran Ableton Live 8 och dess plugin Looper då denna plugin kopierat många av Echoplexens smarta lösningar. För att styra mjukvaran så
valde vi att använda Behringer FCB1010 midi-pedaler som fick smeknamnet “brädorna”.
Mjukvaran sätter ingen begränsning för hur många musiker man inkluderar men vi satte gränsen för vår loop-ensemble till högst fyra musiker inte minst för att uppkopplingen av hårdvarorna (“brädor”, mikrofoner, lyssning osv.) inte skulle bli för komplicerad.
Då livelooping vanligen utövas som en solistisk konstform innebär det att tekniken
och gränssnittet också är designat utifrån det perspektivet. Detta visade sig skapa en
hel del problem när vi ville designa mjukvaran för en loop-ensemble. Avgörande för
oss var att musikerna kunde känna sig kreativa och trygga och att de upplevde att de
hade kontroll över sina “loop-instrumentet” utan att behöva läsa av en datorskärm35.
En god design bygger på “feedback” mellan användare och maskin vilken skapar
tillförsikt, bekräftelse och ger tydliga signaler om vad som händer36. Då vi ville att vårt
system också skulle fungera för musikanter utan så stor erfarenhet av musikteknologi
och då dagens mjukvaror innehåller en sådan mängd funktioner och möjligheter så
blev det första steget att begränsa dessa valmöjligheter. Med hjälp av midi-mapping
kunde vi välja enbart de funktioner vi ville skulle vara aktiva och sedan placera ut dem
bland fotpedalens tio knappar. För att skapa ett så intuitivt och lättförståeligt gränssnitt som möjligt även för någon utan stora tekniska förkunskaper och för att förhindra att man av misstag valde fel knapp så färgkodade vi de viktigaste funktionerna.
Inspirerade av affordance-begreppet hos Gibson37 utgick vi från de färger som har blivit en konvention för de olika grundfunktionerna dvs. röd = inspelning, grön = uppspelning, gul = pålägg, och vit = stop. Våra referensartister bestod av lärare och elever
34
35
36
37
Robert Fink, (2005) Repeating ourselves: American minimal music as cultural practice, (Los Angeles: University of California Press, 2005), 55f.
Per Anders Nilsson, A field of possibilities – Designing and Playing Digital Musical Instruments, (Diss.
Kållered: Ineko AB Nilsson, 2011), 34.
Donald A Norman, The design of future things, (New York: Basic Book, 2005), 138f.
James J. Gibson, The Ecological Approach to Visual Perception, (New York: Psychology Press, 1986).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
58
Hållbus Totte Mattsson
från Musik-konservatoriet i Falun samt kollegor och studenter vid Högskolan Dalarna
med en bakgrund i olika genrer (bl.a. folkmusik, kammarmusik, progressive rock och
dagens dansscen). I mindre ensembler, trios och kvartetter provade de kontinuerligt ut
vår design och gav feedback på systemet. Vi videodokumenterade dessa “testkörningar” vilket gav oss möjlighet att i efterhand gå tillbaka och följa det kreativa arbetsflödet för att kunna förbättra interaktionen med tekniken.
Ett problem med Looper och Ableton Live gränssnitt, just på grund av att det är
konstruerat för solistisk livelooping, är att det är omöjligt att se statusen för mer än
en plugin i taget. Detta gör att när flera loopar rullar samtidigt så är det omöjligt att
se och veta vilken status respektive Looper befinner sig i (en Looper kan ha fyra olika
status Record, Play, Overdub, eller Stop). Detta kan bli ett stort problem när flera
musikanter loopar samtidigt om någon missar att gå ur overdub-läget på sin Looper.
Loopern fortsätter då att spela in tills den blir överstyrd och “systemet brakar ihop”.
Detta skapade en frustrerande osäkerhet hos musikanterna när vi först testade vårt system. Lösningen blev ett liknande system som för fotpedalerna. Genom ett lokalt trådlöst nätverk visades statusen för de olika looparna med färgsymboler på en iPad eller
iPhone. Systemet döpte vi till Loopstatus-skvallret.
“Loopstatus-skvallret”
Detta system bygger på integrationen mellan Ableton Live och programmeringsverktyget Max4Live. Max är ett programmeringsverktyg/programmeringsspråk gjort för att
i första hand skapa musik- och videorelaterade applikationer. Max har integrerats direkt i Ableton Live i form av mjukvaran Max4Live. Detta gav oss möjlighet att utöka
Ableton Lives funktionaliteter med en specialskriven programkod som med hjälp av
Ableton Lives eget programmeringsinterface API lyssnar till en Loopers status och håller koll på förändringar i denna. När en förändring sker skickas statusen vidare internt i vårt Max-objekt som ett numeriskt värde vilket i sin tur triggar ett OSC-meddelande (Open Sound Control) som slutligen skickas till en i förväg angiven IP-adress
på det lokala nätverket. Denna IP- adress matchar sedan en enhet lämplig för visning,
exempelvis en iPad eller iPhone. Vi har i vårt system valt att installera TouchOSC
som är en app vilken låter användaren bygga egna gränssnitt för att skicka och ta
emot OSC-meddelanden. Vi har valt att bygga ett gränssnitt som visar ett antal olika
“lysdioder” som lyser rött, grönt eller gult respektive för Record, Play, eller Overdub.
Om ingen av dioderna lyser betyder detta att Loopern befinner sig i Stopläge.
Våra tidiga programmeringar av Brädorna för att styra looparna var för begränsade
vilket våra referensartister påtalade. Behovet av variationer som B och C-delar, olika
loop-längder, “flytande loopar” (loopar som inte följde synk) osv. ledde till ett antal
olika försök som landade i en grundsetup och programmering som också blev en metod för att lära sig hantverket och filosofin bakom loopandet. Denna “loop-pedagogik”, har vi sammanfattat i en enkel “Loop-Tutorial”38.
38
Finns att hämta på https://github.com/mrhenko/audiovisual-loops
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Audiovisuella loopar och kollektiv live-looping
59
Ganska snart märkte vi att tekniken öppnade för flera olika användningsområden, det handlade inte bara om ett sceniskt musicerande med riktning mot en publik
utan lika mycket om ett spontant musicerande med varandra för sin egen skull. Detta ledde till olika sätt att koppla tekniken och också flera helt olika sätt att utnyttja
den på. De kraftfulla monitorhögtalare vi från början använde blev överflödiga och
istället blev lyssningen i hörlurar med ett “mjukt” reverb det som passade det mer
intima “kammarmusicerandet”. Ett musicerande som ofta byggdes upp med enbart
röster. Det visade sig också att Hearback-systemet, ett system där varje musikant med
en egen liten mixer kan styra sin egen monitor-lyssning, inte fyllde så stor funktion i
detta sammanhang då man naturligt mixade sig själv och sina loopar med mikrofonavståndet. Dessa erfarenheter ledde till att vi konstruerade två delvis olika system, ett
mycket nerbantat, “snabbriggat” och tekniskt okomplicerat för workshops samt ett
mer avancerat tänkt för ett sceniskt framträdande av en loop-ensemble.
Filmloopandet
Då Max4Live pluggarna som är integrerade i Ableton Live visade sig kunna skicka information som mjukvaruprogrammet Isadora kunde ta emot så valde vi det för att
sköta filmloopandet. Vi såg också i detta en möjlig lösning för synkproblematiken
mellan filmloopar och musikloopar.
Ett grundläggande problem då man vill synkronisera bild och ljud beror på ljudets och den rörliga bildens olika informationstäthet. En digitalström som omvandlas till rörligbild innehåller nästan tvåhundra gånger mer information än en ljudström av motsvarande upplösning. Till exempel så är det är inte ovanlig att bilden
blir fördröjd med upp till en tiondels sekund enbart när man kopplar en kamera
till en projektor. Har man varit på en utomhuskonsert med storbildsvideo har man
kanske märkt att musiken och bilden oftast inte är synkroniserade. Här spelar ljudets och ljuset olika hastighet också in, ett visst antal meter från scenen kommer
det av tekniken fördröjda ljuset att hinna ikapp ljudet tack vare ljusets snabbare
fortplantning och det skapas en “sweet spot” där ljud och bild är i synk. Att ljudet
och bilden hela tiden synkroniserar är ett måste för den upplevelse vi vill skapa –
om bilden och ljudet inte upplevs som enhetligt så faller helt enkelt hela idén med
audiovisuella loopar.
Problemet visade sig vara svårlöst men vårt “Loopstatus-skvaller system” var användbart även för detta ändamål då OSC protokollet kunde användas för att förmedla information från programmet Ableton Live via Max for Live till bildbehandlings
programmet Isadora. Dock visade det sig innehålla en allvarlig begränsning då det
ej kunde skicka ett “loop off”- meddelande för att tala om att nu skall kameran sluta
spela in och börja spela upp filmsekvensen. Konsekvensen blev att vi var tvungna att i
förväg definiera en del av looplängderna.
Till slut fick vi ett fungerande system som kunde spela in och spela upp tre olika visuella loopar samtidigt. Dessa visuella loopar kunde sedan kombineras med ytterligare material från “live”-kameror, videoeffekter samt live animeringar. Detta gav
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
60
Hållbus Totte Mattsson
många möjligheter för en videokonstnär att interagera i en gemensam audiovisuell
loopföreställning.
Dessa erfarenheter pekar på att ett meningsfullt loopande med kombinerat ljud
och bild kräver en delvis bestämd form på musiken och passar bäst in i en välrepeterad föreställning för att komma till sin fulla rätt. Vårt resultat presenterades i en
audiovisuell föreställning under konferensen Audiovisual days vid Högskolan Dalarna, Falun den 20-21 september 201239.
Resultat
Forskningsprojektet visar på nya tillämpningar och möjligheter för loopteknik och
nedan redovisas de viktigaste resultaten.
– En teknisk modell för audiovisuell looping, baserad på mjukvaruprogrammen Ableton Live och Isadora, är framtagen och presenterad på scen.
– Ett användarvänligt system för kollektiv livelooping är färdigutvecklat och används av
artister i flera olika genrer.
Systemet bygger på produkter från företagen (Ableton Live, Cycling74, Apple och
Behringer) men kan även byggas i Windows-miljö med hård och mjukvaror från andra varumärken. Det finns också olika freeware-versioner av bland annat Echoplex
Digital Pro att tanka hem från nätet t.ex. loopmjukvaran Mobius. Vi har utgått från
att de flesta musikanter idag redan har en egen laptop och någon form av touchskärm (mobil, iPod, iPad, eller liknande) vilket gör att kostnaden för att skaffa sig en
“liveloop-rigg” inte är så stor. Viktigt att framhålla är även att systemet är “genreobundet”, dvs. det krävs inga genrespecifika, tekniska eller musikaliska, förkunskaper för
att kunna musicera med tekniken vilken har provats ut med hjälp av musikanter med
bakgrund i flera olika genrer. Systemet är heller inte styrt av en “master-slave” funktion
mellan de olika loopmaskinerna utan är konstruerat så att det är möjligt för vem som
helst i ensemblen att när som helst starta en ny “session”. Vidare har en metod och ett
undervisningsmaterial i kollektiv livelooping som bygger på en standardiserad programmering samt färgmärkning av styrknapparna tagits fram. Detta material är avsett
att bland annat användas i workshops.
Resultaten finns presenterade på Audiovisuella mediers hemsida40. Denna sida
fungerar som gemensamt presentationsforum för samtliga projekt finansierade med
medel från EU:s regionaliseringsfond och Region Dalarna. Här finns video-dokumentation från workshops och föreläsningar. De olika programmeringar som gjorts, vår
“Loop-Tutorial” samt våra olika lösningsförslag på gränssnittsproblematiken finns att
hämta hem från github.com41 som “open sources” – vår förhoppning är att detta ma39
40
41
Mattsson, Exempel kollektiv audiovisuell livelooping, Videodokumentation, 21 september 2012, publicerad 17 oktober 2014, besökt 22 oktober 2015 https://www.youtube.com/watch?v=iAlQjswmfv0
AVM, Audiovisuella loopar: Kollektiv livelooping – ett oprövat sätt att musicera, hemsida, 26 september
2012, besökt 22 oktober 2015, http://www.du.se/sv/AVM/Startsida/Audiovisuella-loopar/
https://github.com/mrhenko/audiovisual-loops
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Audiovisuella loopar och kollektiv live-looping
61
terial skall vara så lättillgängligt och “förståeligt” att den som blir intresserad själv
skall kunna kasta sig in i den kollektiva liveloopingens spännande och okända värld.
Slutord
Tekniken har visat sig ha en stor kreativ potential. Med enbart en kort instruktion
och några enkla etyder så startar en kollektiv process där den omedelbara feedbacken
ger ett flow som lockar fram skaparglädjen hos musikanterna – man har helt enkelt
otroligt kul! Dessa erfarenheter har lett vidare till tankar om hur looptekniken skulle
kunna användas i musikundervisningen och där ser vi flera spännande möjligheter,
bland annat
– Ett annorlunda ensemblemusicerande där tekniken ger möjligheter till flera musikaliska roller
– Ett kollektivt kompositionsredskap där interaktionen med tekniken ger upphov till
oväntade och nya musikaliska lösningar
– Ett övningsredskap där den omedelbara feedbacken gör att man snabbt kan prova
och utvärdera musikaliska idéer (t.ex. andra stämmor, komp, riff osv.)
Vidare tror vi också att tekniken kan vara ett användbart hjälpmedel inom vården.
Lusten att hitta synk med sig själv stimulerar användandet av motoriken och systemet
går enkelt att förenkla och anpassa så att man oavsett handikapp kan styra och musicera med looparna.
Dessutom är möjligheterna att möta och interagera med sin egen röst och tillfredställelsen att vara delaktig i en process viktiga ingångar till musikterapeutiska processer.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
62
Hållbus Totte Mattsson
Abstracts
Loop-teknik i solistiska sammanhang är en idag väletablerad musicerande form dock
är möjligheterna att använda tekniken i ensembleform ett mer eller mindre oprövat
fält. I projektet “Audiovisuella loopar” som genomförts vid Högskolan Dalarna presenteras ett system för kollektiv livelooping där upp till fyra personer loop-musicerar
tillsammans och där loopandet samtidigt kan spelas in och spelas upp som videoklipp. Tekniken har visat sig ha en stark kreativ potential. Med enbart en kort instruktion så startar en kollektiv process där den omedelbara feedbacken ger ett “flow” som
lockar fram skaparglädjen hos musikanterna. Dessa erfarenheter pekar på spännande
möjligheter att använda tekniken i musikundervisning och musikterapi.
The use of looping techniques is today a well established form of solo musicianship
but more or less unproven in the context of ensemble performance. In the project
“Audiovisual Loops” at Universtiy College Dalarna a system was developed to allow
collective live looping for up to four people performing together and at the same time
as video loops of the performance were created and presented. This technique has a
strong creative potential. The collective process starts with just a short introduction
and the immediate musical feedback creates a “flow” that quickly brings out the creative potential of the musicians. This experience points forward to exciting possibilities
for use in music educations and music therapy.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
JAN-OLOF GULLÖ, IVAN HÖGLUND, JULIA JONAS, HANS LINDETORP,
ANTON NÄSLUND, JOAKIM PERSSON & PETER SCHYBORGER
Nobel Creations: Producing infinite
music for an exhibition
Introduction
This article describes an artistic music production project that was a part of an exhibition
on the Nobel Laureates in 2014: Nobel Creations. The exhibition’s aim was to creatively interpret the various Nobel prizes in design, fashion and music. The exhibition took
place at the Nobel Museum (fig 1.) in Stockholm and students produced the artistic
content. The music productions, the empirical content described in this article, consisted
of various compositions that were woven together and distributed by loudspeakers in
the exhibit hall without interruption through the four months the exhibition lasted.
Figure 1. The Nobel Museum at Stortorget in central Stockholm. The Nobel Museum’s aim is to, through
creative learning and exhibition techniques as well as modern technology and elegant design, spread
knowledge as well as to create interest and discussion around the natural sciences and culture. The inventor Alfred Nobel (1833–1896) wrote in his will that physics, chemistry, physiology or medicine, literature
and peace would each year receive a part of the revenues of his fortune. More than 800 Laureates have so
far been awarded the Nobel Prize.1 (Foto: © Peter Schyborger)
1
“About the Nobel Museum,” Nobel Museum, accessed August 30, 2015, http://www.nobelmuseum.
se/en/about-us.; “Nobel Prize,” Britannica Academic, accessed December 01, 2015, http://academic.
eb.com/EBchecked/topic/416856/Nobel-Prize.
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
64
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
The ideas behind Nobel Creations
Can artistic representation explain a Nobel Prize? This question was a fundamental issue for Nobel Creations (2014.12.05-2015.03.01).2 The exhibition took its starting point
in a statement that over the years many of those who have received the Nobel Prize
have shown an exceptional creativity in their research, writing or peace work. And in
many ways their creativity can be compared with the cross border creativity that many
artists, musicians or artistic creators, such as successful designers can express in their
work. Therefore creativity unites the Nobel Laureates with artists and creators since
their intrinsic creativity encourages them to think in new ways, to question established
theories, and to innovative put together combinations of insights from different fields.
Similar projects have been on going since 2011, when the Nobel Museum initiated cooperation with Beckmans College of Design in Stockholm. Students of fashion produced six creations that were shown in the exhibition Fashion Innovation
(2011.12.06–2012.01.22).3 The first exhibition was followed up with Fashion Innovation
2 (2012.12.04–2013.01.20)4 and Fashion Innovation 3 (2013.12.05–2014.05.04),5 which
also included interpretations in music composition created by students from the Jazz
Department at the Royal College of Music in Stockholm – KMH.6 In Nobel Creations
2014, four master students from the Music- and Media Department at KMH also joined
the production team. Their task was to create interactive music for the exhibit hall.
Music that was going to be heard without interruption all through the four months
the exhibition lasted. Although previous research reports experiences of how music has
been used in exhibitions at museums or how interactive media may be used in such
contexts,7 this project primarily has been based on experiences from projects at the
KMH in creating music for film, video games and Web-based contexts as well as experiences at the Nobel Museum from the previous years’ exhibitions about Nobel Laureates.
The Nobel Prize was instituted Alfred Nobel and why he donated his fortune to an
international award can be explained by the way he lived his life. Alfred Nobel was
born in Stockholm 1833. In his early childhood the family was very poor and moved
to St. Petersburg when the father, Immanuel Nobel who was an inventor, got a major
2
3
4
5
6
7
“Nobel Creations,” accessed August 30, 2015, http://www.nobelmuseum.se/en/exhibitions/nobelcreations
“Fashion Innovation,” accessed August 30, 2015, http://www.nobelmuseum.se/en/exhibitions/fashion-innovation
“Fashion Innovation 2,” accessed August 30, 2015, http://www.nobelmuseum.se/en/exhibitions/
fashion-innovation-2
“Fashion Innovation 3,” accessed August 30, 2015, http://www.nobelmuseum.se/en/exhibitions/
fashion-innovation-3
Kungliga Musikhögskolan i Stockholm.
Antonio Camurri and Pasqualino Ferrentino, “Interactive environments for music and multimedia”
in Multimedia systems 7(1) (1999): 32-47; Rita Kottasz, “Understanding the Influences of Atmospheric Cues on the Emotional Responses and Behaviours of Museum Visitors” in Journal of Nonprofit & Public Sector Marketing 16:1-2 (2006): 95-121; Fredrik Lindstrand and Eva Insulander, “Setting the ground for engagement–multimodal perspectives on exhibition design” in Designs for Learning, 5(1-2) (2012): 30-49; Pam Meecham and Elena Stylianou; “Interactive Technologies in the Art
Museum” in Designs for Learning 5(1-2) (2014): 94-129.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
65
contract in Russia. Therefore Alfred Nobel never got education in school but received
some private lessons during his years in Russia.
Through the agency of his father, Alfred studied chemistry, an area he later would
become very proficient in, but Alfred was also interested in literature and in his teens
he would rather be a writer than anything else. The father, who became a successful
and industrialist, not least when it came to producing armaments, sent his son Alfred
on trips around the world to meet with various business contacts. During his travels,
Alfred recurrently saw or heard of people getting injured in accidents related to blasting works where nitro-glycerine was used as an explosive. He took strong impression
of such accidents and this resulted in that he a few years later provided great effort to
invent explosives that were much safer to handle than nitro-glycerine. Alfred Nobel’s
most famous inventions: the dynamite and the detonator were very successful and
were spread worldwide. Nobel created an international empire for the production and
sales of explosives and built factories all over of the world.
Despite Alfred Nobel’s lack of formal education, he gradually became a highly respected scientist. In 1884 he was elected into the Swedish Royal Academy of Sciences
and in 1893 he was appointed Honorary Doctor of Philosophy at Uppsala University. In total Alfred Nobel had more than 350 different patents registered and over the
years he became very wealthy. He never married and had no children and he did not
want other relatives to inherit his wealth. In 1888, eight years before he actually died,
a French newspaper by mistake published Alfred Nobel obituary. It was titled: The
merchant of death is dead.8 In the article Nobel was disconcerted and this made him
concerned about how he would be remembered. Therefore Nobel in his will decided
to institute the prices we now know as the Nobel Prize. He wanted his fortune to be
used to create a series of international prizes for those who confer the greatest benefit
on mankind in physics, chemistry, physiology or medicine, literature, and peace. Alfred
Nobel died in 1896 and the Nobel Prize was first awarded in 1901, and over the years
more than 800 laureates have been awarded the Nobel Prize for their creative work.9
The concept of creativity
But what is it that makes creativity so interesting that it needs an exhibition in a museum? Although people throughout history obviously has carried out creative actions
and has been engaged in what we nowadays would call creative work, the concept of
creativity, with its current meaning, is relatively new and contrasts to previously established views:
The connection backwards was with the sense of creation involving some metaphysical force, as in the divine ordination of the world and all in existence
8
9
“Alfred Bernhard Nobel,” Britannica Academic, accessed December 01, 2015, http://academic.eb.com/
EBchecked/topic/416842/Alfred-Bernhard-Nobel.
Alexandra Selivanova, “Alfred Nobel”. Tekniska Museet, accessed October 29, 2015. http://www.tekniskamuseet.se/1/1910.html.printable.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
66
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
within it. But this force was now located within the individual human being,
becoming the object of personal spiritual search for those seeking the wellspring
of truth and beauty. This organicist notion of creativity has had a powerful influence over the whole modern period, including among those who broke with
Romanticism or developed aesthetic values counter to its central tenets. It distinguishes the artist as someone whose ‘inner’ voice emerges from self-exploration,
and whose expressive power derives from imaginative depth. Artistic creativity
has become synonymous with this sense of exploration and expressive power.10
Nowadays, creativity is instead expected to lead to a result or product and creativity
should not to be understood as creation in a general sense, such as the human ability
to create mental images, emotions and beliefs by imagination. Furthermore, creativity
is understood as a creative process that firstly requires knowledge of the field in which
the creative activity takes place; secondly, requires knowledge of, or methodical skills
for, new ways of thinking and/or creative processes; and thirdly, creativity primarily
requires internal motivation by the creative person rather than external motivation.
The idea that creativity, by this definition, is more a knowledge and a skill rather than
some sort of divine power and/or indication of high intelligence is primarily based
on research carried out in the latter half of the 1900s. Jay Paul Guilford gave an important contribution in this direction as he highlighted that: “creativity and creative
productivity extend well beyond the domain of intelligence.”11 At that time, creativity had been noted very little in the previous psychological research and Guilford was
critical towards what he saw as too great confidence in what had been given to different psychometric intelligence tests and he meant that intelligence tests only could reveal very little about a person’s creative nature. Although some of Guilford’s research
results arguably be can discussed, his contribution is important, not the least as he
emphasized that:
…creativity represents patterns of primary abilities, patterns which can vary
with different spheres of creative activity. Each primary ability is a variable along
which individuals differ in a continuous manner. Consequently, the nature of
these abilities can be studied in people who are not necessarily distinguished
for creative reasons.12
Guilford’s approaches that both intelligence and creativity consists of multiple abilities are ideas that recur in later works and research. Such theories have not at least
made a great imprint in teacher education, such as e.g. Howard Gardner’s theory of
multiple intelligences.13 Modern research in creativity can be divided into three different waves. In the first wave researchers focused on studying the personalities of ex10
11
12
13
Keith Negus and Michael J. Pickering, Creativity, Communication and Cultural Value, (London: Sage
Publications, 2004), 4.
Joy Paul Guilford, “Creativity” in American Psychologist Volume 5, Issue 9, (1950): 444–454.
ibid. 454.
Howard Gardner, Frames of mind: The theory of multiple intelligences. (New York: Basic books, 2011).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
67
ceptional creators and in the second wave researchers focused on the internal, mental processes when people were engaged in creative activities. The third wave has a
socio-cultural and “interdisciplinary approach that focused on creative social systems:
groups of people in social and cultural contexts”.14
The notion of creativity based on knowledge of or methodical skills for new ways
of thinking and creative processes, and that creative work not only is reserved to the
creative genius, is something that repeatedly has received attention in post Second
World War educational research. Psychological theory can also be used to explain different aspect of creativity. Mihaly Csikszentmihalyi’s flow theory is for instance an
important contribution as the theory can explain how creative people achieve greater
motivation when their skills are challenged.15
In research on music production, or record production, Philip Mcintyre argues that
the romantic ideal to see creative activity primarily as self expressive “independent
from any perceptible constraint”16 appears to be a widespread opinion in the music
industry, but that this opinion is based on myths, not on theories or empirical academic research. Instead McIntyre points out a background to creative activity in music
production based on recent research that give possible explanations why and how, engineers songwriters, performers, and not at least music producers, “allow creative activity to occur”17 during the music production process.
Also in musicology and research on music education the view has shifted in recent
research away from the nineteenth century’s Romantic tradition with the myth of the
creative individualistic master genius. Pamela Burnard makes an important contribution when she, among other things, highlights the musical creativity as a collective or
collaborative activity:
Whether the collective process flavours listening, collecting, downloading, sampling, performing, mashing, DJing, dancing, or ‘style mixing’ […] at an urban
dance music club, the way we think about musical creativity falls far short of
grasping the potential multiplicity of musical creativity today. Yet music needs
an audience. Audiences engage in ‘a collective experience’, and are a part of,
‘doing’ music.18
Further more, Burnard points out that complementarity and interplay of multiple
types of creativity, such as individual and collaborative, cultural and intercultural, reception-venue and audience-generated, participatory and innovatory, are important
for composers and that composers also use technology as a mediating tool to support
innovation. Explanations of this kind corresponds very well with the way research14
15
16
17
18
Keith Sawyer, Explaining creativity: The science of human innovation. (Oxford University Press, 2011), 4.
Mihály Csíkszentmihályi, Creativity: Flow and the Psychology of Discovery and Invention, (New York:
Harper Collins, 1996).
Phillip McIntyre, “Rethinking creativity: record production and the systems model”, in The Art
of Record Production: An Introductory Reader for a New Academic Field, ed. Simon Frith, and Simon
Zagorski-Thomas, (Farnham: Ashgate Publishing, 2012),149.
ibid. 162.
Pamela Burnard, Musical Creativities in Practice. (New York: Oxford University Press, 2012), 14.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
68
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
ers in other fields, such as Becker19 on collaborative production of art and Burgess,20
Howlett21 and Moorefield22 when they describe the music producer’s operations and
professional role. Thus, a music production activity that is characterized by creative
collaborative innovative work where old structures for labour division no longer are
valid. Instead music production is described as an activity that is characterized by creative collaborations, all in order to create the artistically and/or commercially successful productions as possible.
One aspect that recurs in contemporary research on creativity is that education in
school rather than to develop children’s and young peoples creativity and innovative
abilities rather do the opposite. Burnard shows explicit criticism of the myths, teachers
and others working in the educational field, may have on children’s and young people’s creative abilities by, referring to current research provide information “that goes
well beyond many adult preconceptions and simplified definitions of what constitutes
children’s musical creativities”.23 But that schools fail to develop children’s creative
abilities also has attracted attention outside the field of music. Not at least Sir Ken
Robinson has made this kind of knowledge widespread through his presentations in
popular scientific contexts24 by criticising the school systems’ inability to make use of
and/or to help to develop children’s creativity.
I believe our only hope for the future is to adopt a new conception of human
ecology, one in which we start to reconstitute our conception of the richness of
human capacity. Our education system has mined our minds in the way that
we strip-mine the earth: for a particular commodity. And for the future, it won’t
serve us. We have to rethink the fundamental principles on which we’re educating our children.25
But such opinions are unfortunately not new. Already sixty-five years back, Guilford
was into similar thoughts by asking the following questions:
Why is there so little apparent correlation between education and creative productiveness? Why do we not produce a larger number of creative geniuses than
19
20
21
22
23
24
25
Howard S. Becker, Art Worlds. (Berkeley: Univ. of California Press, 1982).
Richard James Burgess, The Art of Music Production: The Theory and Practice. (New York: Oxford University Press, 2013); Richard James Burgess The History of Music Production. (New York: Oxford University Press, 2014).
Michael John Gilmour Howlett, “The Record Producer as Nexus: Creative Inspiration Technology
and the Recording Industry.” (PhD diss., University of Glamorgan 2009).
Virgil Moorefield, The Producer as Composer: Shaping the Sounds of Popular Music. (Cambridge, Mass.:
MIT Press, 2005).
Burnard, “Musical Creativities,” 278.
Ken Robinson, Out of Our Minds: Learning to Be Creative, (Oxford: Capstone, 2011); Ken Robinson,
“Changing education paradigms”, RSAnimate, 16 June, 2008, accessed October 29, 2015, https://
www.thersa.org/globalassets/pdfs/videos/2010/10/rsa-animate---changing-paradigms/rsa-lectureken-robinson-transcript.pdf; Ken Robinson and Lou Aronica. Creative Schools: The Grassroots Revolution That’s Transforming Education, (New York: Viking, 2015).
Ken Robinson, “Do schools kill creativity?,” TED, February 23, 2006, accessed October 29, 2015,
http://www.ted.com/talks/ken_robinson_says_schools_kill_creativity/transcript?language=en.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
69
we do, under supposedly enlightened, modern educational practices? These are
serious questions for thought and investigation. The more immediate and more
explorable problem is a double one: (1) How can we discover creative promise
in our children and our youth? and (2) How can we promote the development
of creative personalities?26
One possible way to develop creative abilities within the framework of projects in
education and training is the use of structures and theory that underlie educational
models for constructive alignment27 since learners construct meaning from what they
do in order to learn. The model is that teachers start with the outcomes that the students are intended to learn, and align teaching and assessment to those outcomes.
Biggs & Tang emphasizes, with reference to recent research on learning and knowledge development, that the best learning experience for an individual often happens
when he or she teaches someone else. That is one explanation why peer teaching is a
particularly powerful way of learning. Activities of what people use and do in real life
and what they talk about with others also results in good learning and a much better
learning for most people compared with what is learned when they hear, see or read
about something. Biggs & Tang also highlight the importance of multimodal learning:
“We learn through activating different sense modalities: hearing, touch, sight, speech,
smell and taste. The more one modality reinforces another, the more effective the
learning”.28 Feedback is another aspect that Biggs & Tang bring forward as particularly
valuable for individual knowledge development, but also is highly valid for collective
learning. Formative feedback is particularly important in creative contexts when provided during the learning process, the formative feedback is telling the learners how
well they are doing and what might need improving.
iMus – an interactive musical audio playback framework
Music requires that people act or interact. Musical instruments are designed for musicians to interact with and when musicians play with others they extend the interaction
with each other. In some cases there is also an audience interacting with the musicians and in some cases the musicians interact with other creative people like dancers
or actors for example. In the early days of movies there was no audio tracks and live
musicians accompanied the silent movies. Later when film got audio, a whole new
genre – recorded film music – was established. Over the years, film music, as a genre,
has developed into an art of its own and today film music is a well-established field in
academic studies and research.29 Furthermore, since the soundtrack can be considered
to have a decisive impact on a film’s success, film music is therefore dedicated special
26
27
28
29
Guilford, “Creativity,” 444-445.
John B. Biggs and Catherine So-Kum Tang, Teaching for Quality Learning at University: What the Student Does, (Maidenhead: Open University Press, 2011), 96.
ibid. 63.
David Neumeyer, ed., The Oxford handbook of film music studies, (Oxford: Oxford University Press,
2014), 3-6.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
70
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
attention at, e.g., film festivals and film awards, and leading film composers can be
accorded considerable media attention and also be given superstar status.30 With this
legacy of film music, composers and producers of new interactive environments like
computer games, web pages and perhaps also exhibitions at museums try to integrate
and use music in similar ways as in film music. The music is used to tell a part of the
story, create an atmosphere or perhaps increase the experience for the viewer or user.
Nowadays, when computers and mobile phones and other mobile devices in general have very good audio quality and the bandwidth of Internet connections has been
highly increased, it is in many cases possible to produce music that sounds good by
means of new digital, interactive media. But still there are some challenges, though:
Music in interactive environments or media needs to be able to change depending on
events occurring inside the media and it is necessary that changes can happen in a
similar way as if the music was performed by live musicians. Therefore music that is
traditionally produced: composed, arranged, recorded and mixed for a linear context
like tracks on a CD, can be difficult to use in interactive music and media production.
In interactive media the music needs to interact with the user just like musicians interact with each other.
Different frameworks or software has been developed to try to solve the challenge
of having the music responding to events in a musical way, like for example changing
instrumentation or harmony on the next bar-line by trigging audio files with music
and make crossfades between different audio files.31 In general, such audio frameworks have neither been standardized nor music-composer-friendly and therefore
many composers have not been composing for interactive media. One example of this
is Microsoft Direct Music that was released in 1996 along with the music production
tool Direct Music Producer (DMP). The aim of this software was to provide a useful tool and a standard for music to be integrated into productions using Direct X.32
DMP had a steep technical learning curve and there was an obvious risk that composers would lose their musical creativity having to overcome too many technical obstacles in the workflow.
Thus, there has been a lack of effective tools to compose for interactive media,
where the composer’s creativity and experience of already existing music production
tools could be used, and that the tools could be used to build effective bridges between the composer and the programmer of interactive media. Wwise33 and FMOD34
have recently been developed and these two software tools are widely spread but lim30
31
32
33
34
Dean Keith Simonton, “Film Music: Are Award-Winning Scores and Songs Heard in Successful Motion Pictures?”, in Psychology of Aesthetics, Creativity, and the Arts 1, no. 2 (2007): 53-60.
Johnny Wingstedt, “Making music mean: on functions of, and knowledge about, narrative music in
multimedia” (PhD diss., Luleå tekniska universitet, 2008).
“DirectX Audio: Making Audio Special FX”, Microsoft, last modified July 19, 2011, accessed October
29, 2015, https://code.msdn.microsoft.com/DirectX-Audio-Making-48867ceb.
“Wwise Empowers Audio Creators” Audiokinetic , accessed October 29, 2015, https://www.audiokinetic.com/download/documents/Wwise_Brochure_March2014_en.pdf.
“Introducing FMOD Studio”, Firelight Technologies, accessed December 1, 2015, http://www.fmod.
org/products.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
71
ited when it comes to features driven by musical needs, perhaps because these tools
rather are developed for use in computer games than for creative musical composition work. Other recent examples of new software for interactive music are e.g. ELIAS
– Elastic lightweight integrated audio system35 that is aimed for use in different platforms and Dinahmoe Labs who have created an audio framework and different software for web technology and their software mixer “The Rick Astley Remixer”36 provides the user with an entertaining example of how technology of this kind can be
used for real-time playback and processing of musical content. Interestingly enough,
both ELIAS and Dinahmoe Labs are developed and based in Stockholm.
In Nobel Creations a special framework and software, iMus, was used. iMus, which
is developed by Hans Lindetorp and the result of an artistic research project at KMH,
was redesigned for the project.37 Over the years, students from the Music and Media Production Department at KMH have been working with several interactive music production projects in different context. Due to the lack of efficient software for
interactive music production iMus was developed based on knowledge and experiences from many of these student projects. The technology being used in iMus is Web
Audio API, a high-level JavaScript for processing and synthesizing audio in web applications.38 What makes iMus useful is that it is designed to be easy to apply in different
contexts and therefore very suitable for an interactive music production project in a
museum such as Nobel Creations. Further more it is web based which makes it easy
to use on all kinds of platforms and units. iMus will be described in depth on another occasion, but briefly, it solves the challenge that music in interactive environments
needs to be as if it was uniquely composed to the linear course of events that are the
results of the users interaction, including control of optional external parameters as
input as well as the randomised playback of programmed events. iMus also has special features to solve the pulse-based music challenges in interactive contexts.
The music production project
Since 2011 the Nobel Museum has staged an exhibition about the new laureates. For the
exhibition in 2014-15 Nobel Creations, product and interior design and fashion students
from Beckmans School of Fashion expressed their interpretation of the different individual prizes and respective laureates in creating garments and a variety of complementing graphic objects. To accompany the designs, students studying jazz at KMH composed different pieces of music, inspired by the prizes. Their music was played in headphones by the different displays (Fig 2.). In addition to this a production team of mas35
36
37
38
“Adaptive Game Music”, ELIAS – Elastic lightweight integrated audio system, accessed October 29,
2015, https://www.eliassoftware.com/.
The Rick Astley Remixer”, Dinahmoe Labs, accessed October 29, 2015, http://dinahmoelabs.com/
rickastley.
Hans Lindetorp, Documentation iMus. KMH: Stockholm, accessed December 1, 2015. http://momdev.
se/interactivemusic/documentation.html
Paul Adenot and Chris Wilson, Web Audio API: W3C Editor’s Draft. WC3, last modified, December 1,
2015, accessed December, 1, http://webaudio.github.io/web-audio-api/.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
72
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
ter students from the Music Production program at KMH: Ivan Höglund, Julia Jonas,
Anton Näslund and Joakim Persson, created interactive music for the exhibition hall.
Figure 2. The Nobel Museum exhibition hall. (Foto: © Peter Schyborger)
A challenging objective
The objective was to create a soundscape in the room that would be played six days
per week for four months. The music would be played through a total of 18 speakers, including two subwoofers. Four of the speakers, seen in green (fig. 3), played a
continuous background soundscape composition. The six “blue” speakers, spread out
around the exhibition hall built into big pieces of foam, played triggered sound effects
and musical parts, based on the prices they represent, all while doing so in time and
key with the backgrounds.
Figure 3. Speaker set up in The Nobel Museum exhibition hall. (Foto: © Anton Näslund)
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
73
In the centre of the exhibition hall, there were six iPads, one for each Nobel Prize category, with interactive menus and information about the laureates and their respective
prizes. When a visitor touched the screen of the iPad to change it from the menu to a
submenu or to the next page of information, different versions of the musical parts
coming from that particular speaker was played .
When all six iPads were active at the same time they would play like a small ensemble with six different musical parts and six different instruments, all in sync and in
key. Six iPads were used and each iPad had five sub menus (Fig. 4). In a total, 30 different musical parts could be played in any combination with the background music
also in sync. All this was made possible through the iMus audio engine.
Figure 4. iPads used in the exhibition hall. (Foto: © Peter Schyborger)
The production and composition process
The composition process started when the nominations of the 2014 Nobel Prize laureates were announced including presentations of their work for which they were being awarded. The production team discussed possible influences that could draw on
musically, and decided that each music producer would work separately, creating four
individual soundscape compositions for the exhibition. Next the production team visited the museum to check out the acoustics of the exhibition hall. The room is rather
large and reverberant, which led to the preliminary conclusion that it would be preferable to avoid using a lot of effects and instead to try to keep the sound quite dry. There
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
74
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
were some background noises observed in the room that might interfere with the
music, most notably an air-condition system that hummed at different frequencies at
different times. This was noted but an assumption was made that the music probably
would mask the background noises quite easily, which later turned out to be correct.
The objective was to vary the music throughout the day by dividing the different compositions into three sections that followed a cadence in the key of F major. The fourth
composition would be used as a background for guided tours and would interrupt the
other compositions whenever there was a tour. This composition remained at a lower intensity level at all times and left more space in the frequency range used for speech.
The next phase was to make demos of some parts of the music with something for
each part of the exhibition hall; a background that played from the speakers in the
four corners of the room, triggered sounds for any of the six displays and ensemble
parts for the central iPad section containing six speakers. The music demos were tested
in the museum and played back in speakers and evaluated by the production team. At
this time just a few speakers were used in approximate locations in the room.
Figure 5. The dimensions of the room and the hard acoustics were challenges for the composers.
(Foto: © Peter Schyborger)
The dimensions of the room proved to be a bit problematic, because the sound coming out of the corners of the room and the sound emanating from the iPad-section in
the middle of the room appeared slightly out of time when the listener moved closer to one or the other (Fig. 5). This made the production team decide to limit the
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
75
use of percussive sounds only in one side and to use tousling sounds with longer and
smoother attack in the other.
One challenge was to accomplish variation and change without having to compose
extremely lengthy pieces. The precondition was that the music would play six days a
week for four months, so we couldn’t allow it to get on peoples’ nerves nor could it
sound exactly the same at all times. The solution was to use compositions that could
be divided up into flexible building blocks and then be combined in many different
ways, all of them musical sounding. The way to do this was to keep it simple. Quite
paradoxically, the simpler the music was composed, the more ways it could be combined and varied. Some of the compositions will actually play for years before repeating themselves.
The general idea behind the soundscape was to create an airy, interesting and pleasing atmosphere that left space for the visitors to communicate with each other in the
room. The soundscape gave the exhibition an inviting, ethereal feel. The compositions
contain blend of traditional instruments like live string instruments and modern software synthesizers. The idea behind this was to show openness to multiplicity and to
create a link between old and new, a nod to the history of the Nobel Prize.
The compositions are filled with details that are musical interpretations of the
Nobel Prizes; the sweeping movement of a microscope appears as a sweeping filter
sound, hexagonal patterns in the brain appear as musical sixths and sextuplets, high
frequency blue light from an LED lamp appears as high pitched blue notes.
The music was composed and arranged with a special technique. The different
tracks were cut into parts and pieces to fit into the programming part. The music was
composed so that the parts could be mixed with each other in any order.
The music had to be arranged so that it could be played for hours and still never to
give the listener feeling of being repetitive. Most of the compositional work was done
in Apple Logic Pro, both in the compositional stage and for editing (Fig. 6).
Figure 6. A screenshot of the Apple Logic Pro project in the compositional stage. (Foto: © Anton Näslund)
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
76
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
The iMus audio playback framework39 was used to play music in the museum hall.
With iMus it was easy for the production team to make prototypes during the production of music and try out different ideas in different pieces in the exhibit hall in the
Nobel Museum. In iMus, audio files are used for playback. The file can be a musical
phrase, one bar of drums, a hit or a sound of any kind. To make it work within the
software the composer has to break down the music into small pieces, or fragments,
and when it’s played back the software puts it back in order. The smallest piece or musical phrase, the file, needs to have some kind of musical information and it is divided
into three parts: an upbeat of any length, the phrase itself with a predesigned musical
length depending on the tempo chosen, and a tail of any length (Fig. 7).
Figure 7. Example of the files cut into parts and pieces, an upbeat of any length, the phrase itself with a
predesigned musical length. (Foto: © Anton Näslund)
This particular format makes it possible to stack or sequence the files and secure that
they can overlap in an understandable musical context due to the fact that they have
a specific musical length, for example a bar or a beat. A random selection of files can
be added and played back similar to an arrangement window in a Digital Audio Workstation, DAW, with any number of tracks or parts. It is also possible to use different
lengths of loops on the different tracks. There can also be any number of arrangements. This all together makes it possible to, at one event, to go from one arrangement
to another, or at another event, for example, mute a track. All music files were exported
into mono aiff-files with their reverb tails intact so that transitions would be seamless.
While composing the production team had to try different combinations during the
process to hear that all parts would be compatible with one another as they were to
39
Lindetorp, “Documentation iMus”.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
77
be put together by a randomizing algorithm when performed. During the project the
iMus system was further developed in close cooperation between the programming designer Hans Lindetorp and the production team who composed the music.
During the try-outs and tests several modifications were made. For example, the
background music that was played while guided tours were being held was chopped
up into individual notes and bars to make it more responsive to actions in the room
and the more people who moved about in the room, the more stems in the background music would be triggered, increasing the intensity of the background music. In
the ceiling, there were infrared sensors picking up the movements when visitor walked
about in the room, controlling both the background music and the motion triggered
sound effects played from the different display stations.
Figure 8. The final set up of iPads in the centre of the exhibition hall. (Foto: © Peter Schyborger)
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
78
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
Completing the project
Everything came into place within the last couple of days before the opening ceremony (Fig. 8). The materials used in displays and installation of a carpet in the exhibition
hall changed the acoustics to a much drier sound than what it originally sounded like.
This made it possible to play music a little louder than what was planned and programmed. But the acoustic changes in the exhibition hall also had the effect that the
music sounded a bit dryer than expected.
The opening was a big success and the museum was packed with people. When the
exhibition closed four months later the staff at the museum told us that they had enjoyed the music, and that people had come to the exhibition just to listen to it, which
was wonderful for us to hear. This was the first time in the history of the museum
that they’d had music playing through speakers in the exhibition hall. Because of this,
there were some initial worries that it might not work to have music playing there, but
everyone was very happy in the end. The tour guides were very pleased with the music
and was happy to talk to visitors about it. The staff actually said they would miss the
music once it was gone.
Final reflections
This project shows that people together can be very creative. And by combining different art forms, using well thought through teaching methods and structured project
planning which aims to develop creative actions, young people, students in learning,
together can create artwork that far exceeds what they individually could cope.
Another key experience is that it is possible to compose and produce music that
is playing week after week in an exhibition hall where staff members work for many
hours daily, and that they after four months of listening still had not got tired of the
music. How is that possible? Perhaps one explanation is how the music was composed
and produced. The production team knew that their compositions would be played
for many hours each day in the museum hall and therefore they put special effort to
make music that was possible to listen to for a long time without the listeners would
become tired of it and maybe even become annoyed. None of the members in the production team had ever composed music in this way before. Even if they all, in different ways and other genres, are experienced as composers and producers, it was a true
creative challenge for them to compile their compositions and productions during the
project and perhaps that extra ordinary challenge contributed with flow and creativity. The design of the project, with tight time schedules, is perhaps also one possible
explanation why the music production processes actually allowed creative activity to
occur similar to the theoretical framework described by Philip McIntyre.40 Overall, the
project thus shows good evidence that the romantic ideal, that creativity and creative
capacity would be something individual, inborn and inherent in creative people, not
40
McIntyre, “Rethinking creativity,” 160.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Nobel Creations: Producing infinite music for an exhibition
79
is valid since the participating composers and music producers clearly developed their
creative capacity during the project by cooperating with each other. By composing
music in a genre they never worked in before, in an environment that they not previously had created music for and with newly developed software and audio equipment
used in innovative ways, the project has given the participants unique experiences that
they can add to their repertoire of generic skills. And such skills may certainly be valuable for them in future music production projects. Another important aspect to highlight and reflect upon is that education, according to Ken Robinson, may lead to that
pupils and students creative ability is hampered.41 It is therefore important, especially
in compulsory school but also in higher education, to create conditions that develop
creative abilities of those who participate in training by allowing creativity to occur.
iMus,42 the technical framework and software that was used during the project, has
been further developed as a result of the experiences and new knowledge the production team gained during project. Furthermore the results of the project give clear support to theories of learning showing that peer teaching and activities of what people
use and do in real life and what they talk about with others provides good conditions
for learning and development of new knowledge.43 And the basic educational idea for
this project was to try, as far as possible, to emulate how composers and music producers work in real life. During the project, different things happened that led to unexpected problems. And these problems had to be solved. The production team had
challenges of various kinds, not the least of time pressure. Therefore, the project sent
the participants on a journey that wasn’t an emulation of real life. It was real life for
composers and music producers. And to learn in real life is probably one of the best
ways to create new knowledge and develop creativity.
Perhaps we also can learn something from what Alfred Nobel did and the way he
lived his life. Although he lacked formal training school and had no higher education,
he learned in his own way and became very successful. And by learning and experiencing different cultures in real life, Alfred Nobel developed an outstanding creative ability.
41
42
43
Robinson “Do schools kill creativity?”.
Lindetorp, “Documentation iMus”.
Biggs and Tang, “Teaching for Quality Learning at University”.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
80
J-O. Gullö, I. Höglund, J. Jonas, H. Lindetorp, A. Näslund, J. Persson & P. Schyborger
Abstracts
Under ett konstnärligt kollektivt musikproduktionsprojekt på Nobelmuseet i Stockholm i 2014 användes nyutvecklade metoder för komposition och musikproduktion.
Med en specialutvecklad programvara skapades musikproduktioner som svarade interaktivt på museibesökarnas handlingar. Musiken spelades upp genom flera högtalare
i utställningshallen, vecka efter vecka och utan avbrott under de fyra månader som
utställningen pågick. Projektets resultat ger klara belägg för att det romantiska idealet, att kreativ förmåga främst skulle vara något individuellt, medfött och inneboende,
inte är giltigt. Istället visar projektet att när olika konstformer kombineras och projekt
genomförs med kreativa mål kan människor i samverkan skapa konstverk som i kreativ bemärkelse vida överstiger vad de individuellt klarar av att åstadkomma.
In 2014 a collaborative artistic music production project gave rise to the development
and use of new methods for composition and music production. With a specially designed software engine the music productions responded interactively to actions of
the visitors at the Nobel Museum in Stockholm. The music was distributed by multiple loudspeakers in the museum hall, week after week without interruption through
the four months the exhibition lasted. The results of the project show clear evidence,
that the romantic ideal, that creativity and creative capacity primarily is individual, inborn and inherent, is not valid. Instead, by combining different art forms, using structured project planning, aiming to develop creative actions, people can create artwork
in collaboration, that far exceeds what they individually can achieve.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
MARK GRIMSHAW
A brief argument for, and summary
of, the concept of Sonic Virtuality1
Introduction
When that august body, the Acoustical Society of America, publishes a definition of
sound in the American National Standards bible on Acoustical Terminology, and an innocent question on the wording of that definition to an acoustics-based colleague of
mine provokes a negative response – “[the] definition [is] out of scope for most of
the purposes I know [the definition] is only operational for some purposes [...] It is
therefore necessary to use domain-specific definitions”2 – then there is clearly something quite wrong. In extremis, one might reasonably ask, if there is no agreement as
to what sound is even among acousticians, is the entire scientific basis for the study of
acoustics fundamentally flawed?3
As this essay makes clear in its opening sections, there historically have been and
currently are several definitions of sound and so among the first aims of the essay
are to enumerate some of the most important definitions, to tease out differences
between them, and to highlight inconsistencies both with each definition and in the
context of our experiences, physically and phenomenologically, of what sound is.
What is the purpose of such an initially hermeneutic approach? It is precisely this:
to argue for a definition of sound (and its underlying concepts) that is more consistent with recent research and thinking, that is more consistent with our phenomenological experience, and that has more use-value in the context of new technological
developments. On this last point, my interest has been sparked by research collaboration with colleagues in the field of computer game sound. In particular, our work
with biofeedback has, for our thinking about sound, its effects, and how to design affective sound, inspired and clearly necessitated the need for a new definition of what
1
2
3
In a 2015 book, my co-author Tom Garner and I put forward an alternate definition and concept of
sound (M. Grimshaw and T. Garner, Sonic Virtuality: Sound as Emergent Perception, (New York: Oxford University Press, 2015)) that was in opposition not only to the standard view of sound (the one
used in acoustics) but that also opposed other more philosophical definitions that have appeared
over the last decade or so. I presented a brief presentation of the book and its ideas to the Musik og
lyd konference in early 2015 and what follows is an expansion of that paper and so a summary of
some of the major points of Sonic Virtuality.
Personal email communication, 1st September 2015.
Probably not. But, as I explain in this essay, acousticians are misguided concerning what they study.
Thinking that what they study is sound, they are in fact studying sound waves.
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
82
Mark Grimshaw
sound is.4 What becomes clear from this research (and other research investigating responses to audio) is the highly individualistic sound we experience – the phenomenon of
sound is context-specific where that context includes the perceiver’s brain and the wider
environment. Rather than attempt to use statistical methods and rationales to smooth
over such differences in order to produce a generalized, mass-produced, one-size-fits-all
audio artefact (for computer games and elsewhere), I prefer the approach of sound design that celebrates and utilizes these personal differences because this, I believe, increases our engagement with computer games (and other multimodal artefacts), increases the
possibility of enhanced presence in virtual environments, and leads to new sound design
archetypes. Most importantly, though, such an approach, and the results deriving from it,
gets us closer to what sound is as the object of design and as the object of study; in order
to design sound, one must first know what it is. Hence the definition offered here.
I begin the essay proper by enumerating various current definitions of sound and
discussing a number of identified problems with them. The first section deals with
what is termed the standard definition of sound in the western world (there are actually several of these ‘standard’ definitions) and the inconsistent and incoherent uses
to which this definition is put. Following this, I present some other definitions that
are philosophical and, towards the end of the section, are more phenomenological. In
this second section, I also present a short discussion of everyday listening. I then move
onto a presentation of the new definition and a brief outline of the concept of sonic
virtuality before closing with some speculative thoughts on the definition’s and concept’s use and implementation by sound designers.
The standard view of sound
The multiplicity of physics-based definitions
Whenever I ask my students what is sound?, beyond regular and frequent answers such
as noise, music, or speech, almost invariably I am told that sound is a sound wave. This
definition (or a variation of it) is probably the answer many of the readers of this article would provide because it is the standard Western definition as found in physics and specifically acoustics. The 10th edition of The Concise Oxford English Dictionary provides the following primary definition of the noun: “vibrations which travel
through the air or another medium and are sensed by the ear.” A fuller definition,
in which I attempt to be more precise by dint of the use of jargon, may be stated as:
sound is a pressure wave moving through a medium and whose properties of direction of travel, frequency, and amplitude lie within the sensory range of an auditory system.
The American National Standards Institute (ANSI) provides two (possibly three)
definitions of sound to be used depending upon the context in which sound is to be
4
Biofeedback in this context involves the use and interpretation of data from psychophysiological devices such as electroencephalographic headsets for the real-time processing or synthesis of audio for
the purposes of close control of arousal and emotions during computer gameplay. As an example: in
a survival horror game, the game engine will sense that you are not scared enough and will immediately respond with more scary sounds.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
83
used or studied. Sound is either “(a) Oscillation in pressure, stress, particle displacement, particle velocity, etc., propagated in a medium with internal forces (e.g., elastic
or viscous), or the superposition of such propagated oscillation” or the “(b) Auditory
sensation evoked by the oscillation described in (a).”5 Sound is thus, and variously, an
oscillation in any of an unstated number of physical properties, the superposition of
such oscillations (and one must assume the plural here), or the auditory sensation of
this oscillation (presumably also the sensation of superpositions of such oscillations).6
In a popular article describing the process of an auditory brainstem implant, there
appears this statement: “the area where the axons (nerve fibres) and cochlear nucleus
(synapses)—which transport sounds picked up by the ear to the cerebral cortex—are
found”.7 In this case, sound appears (in similar manner to the two ANSI definitions)
to exist simultaneously in two states: the vibrations to be picked up by the ear and
the auditory sensation of such vibrations. Such ambiguity, vagueness, and imprecision
(perhaps surprising to those expecting the claimed-for exactitude, rigour, and precision of the natural sciences) can also be found in scholarly textbooks and academic
articles on acoustics and psychoacoustics as Pasnau makes clear when he enumerates
several examples that, taken together, bring into focus the incoherency of the standard
physics-based definitions of sound. For instance, in the same 19th century acoustics
textbook by John Tyndall, sound is located within the brain, within the sound source,
or is to be found moving within the medium8 while, in a modern handbook dealing
with the perception and cognition of hearing, sound is described both as a compression wave and as the sound source itself.9 Finally, an acoustics textbook I have long
used for teaching (and from which I myself was taught) describes sound as “a mechanical disturbance of the medium”.10 While the authors admit that it is a simplistic
definition, it is nevertheless the one provided on the opening page of this standard
textbook and there are so many flaws with it that it is difficult to know where to begin.
5
6
7
8
9
10
American National Standard, Acoustical Terminology. ANSI/ASA S1.1-2013
This definition has the following note appended: “Not all sounds evoke an auditory sensation, e.g.,
ultrasound or infrasound. Not all auditory sensations are evoked by sound, e.g., tinnitus.” From part
b) of the definition, sound is defined as an auditory sensation; what the note is clarifying [sic], then,
is: “Not all sounds evoke a sound [...] Not all sounds are evoked by sound.” I suspect the reader may
be as puzzled as I am by such claims while secretly applauding such pataphysical dadaism from the
normally po-faced field of acoustics.
D. Pouliot, Hearing Without Ears (Auditory Brainstem Implant). Accessed August 25, 2015, http://lobe.
ca/en/non-classee/hearing-without-ears-auditory-brainstem-implant/#.VdQuvbfGam4.
On one page alone, the following statements occur: “It is the motion imparted to this, the auditory
nerve, which, in the brain, is translated into sound” and, discussing exploding gases in a lecture theatre, “every ear in this room is conscious of a shock, to which the name of sound is given”. On a later
page, the following appear: using the analogy of balls in a row hitting against each other (thus the motion of the first ball is transferred to the last), Tyndall states that “thus is sound conveyed from particle
to particle through the air” and yet, when describing how this motion sets the tympanic membrane
vibrating, which motion is itself transmitted along the auditory nerve to the brain, it is in the brain
that “the vibrations are translated into sound”. Tyndall’s book was written four years after, and is heavily indebted to, von Helmholtz’s Die Lehre von den Tonempfindungen, a work of some significance in the
canon of modern acoustics. J. Tyndall, On Sound, (London: Longmans, Green, and Co., 1867), 2-4.
R. Pasnau, “What is Sound?” The Philosophical Quarterly 49:196 (1999): 318-319.
D. M. Howard and J. Angus, Acoustics and Psychoacoustics, (Oxford: Focal Press, 1996), 1.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
84
Mark Grimshaw
Problems with these definitions
As there are rather many definitions of sound as a physical phenomenon, to avoid
confusion I shall group them together under what Pasnau terms the standard view of
sound.11 For Pasnau, this term brackets together two common concepts found within all these definitions. First, as Aristotle first identified, that sound is the object of
hearing and, second, that sound exists in the medium (for humans, typically air) and
thus sound in the standard view, however the definition is formed, is the object of our
hearing and this object consists of vibrations or oscillations in a medium. The second definition provided by ANSI – that sound is an auditory sensation – complicates
matters as it might be said to reference Aristotle’s statement on sound but it does not
directly reference the second common concept that Pasnau identifies. It is, however,
best viewed not as a physics-based definition but as a psychology-based definition.
Nevertheless, psychoacoustics both uses this definition and relies for it on the presence of sound waves.
There are two fundamental categories of problems with the quite singular standard
view of sound. First is the issue of the language and the words used in the definitions
bracketed under that view. Second is the quite limited usefulness of such definitions.
First, the wording. In the OED definition, which or whose ear does the sensing of
the vibrations? Human ears only or those of other animals, birds, and insects? What
precisely is an ear – is that of reptiles, such as snakes, properly termed an ear when it
comprises only what may be described (in human terms) as the inner ear? What of the
case of auditory brainstem implants where the useless ear (or auditory nerve) is bypassed with a microphone – do the recipients of such implants sense sound without
the use of ears? What of tinnitus – particularly subjective forms such as the high-pitched
ringing experienced by many – where are the vibrations or oscillations in a medium?
If such tinnitus sufferers do indeed hear sound, then the definitions provided by
ANSI can be discarded out of hand. Even my attempt to use jargon to more accurately
define sound while remaining true to the standard view fails when one asks two simple questions: what precisely is the auditory system and what is the frequency range of
hearing?12 Across hearing species, there are widely varying hearing frequency ranges. If
sound is indeed a pressure wave/vibrations/oscillations in a medium that can be sensed
by the ear/auditory system, then, as a thought experiment, if both a human and a dog
are subject at the same time to the same sound wave at 30kHz, why should it only be
defined as sound for the one simply because the other’s auditory sensory apparatus cannot sense it? And what of infrasound that can be felt through the skin – is the auditory
system simply part of our tactile sense because, as one famous (and famously profoundly deaf) musician has suggested, “hearing is basically a specialized form of touch”?13
Related to this problem of the wording is the use made of the definitions because
the use that sound as a definition is put to depends to a large extent on the firmness
(or not) of the wording (although definitions can be blithely disregarded as is the
11
12
13
Pasnau. “What is Sound?” 309-310.
A good human hearing range is typically quoted as between 20Hz and 20kHz.
E. Glennie, Hearing Essay, (1993). Accessed August 19, 2015, https://www.evelyn.co.uk/hearing-essay/.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
85
case previously noted with Tyndall happily defining sound as a motion propagated
through air while at the same time stating that sound exists in the brain). If one cannot agree on one rigorous definition of sound, then how does one know precisely
what the object of study is (let alone hearing) or, indeed, if two articles purporting to
deal with sound are, in fact, dealing with that same object of study or hearing?
The second category of problems is, for me, the more egregious and the reason why
I have spent some time devising what I believe to be a more precise, comprehensive,
and useful definition of sound: the standard view of sound is too limited in its scope;
it does not match the reality of our experience and thus is too limited in its use-value.
My objection is founded upon two disparate frameworks of knowledge that might
broadly be defined as objective and subjective. The physics-based definition attempts
to be objective because industrialization requires standardization and mass production
and so two people subject to the same sound wave at the same time and at the same
location will (must), according to such definitions, hear the same sound. Similarly, assuming all things technological are equal, the same audio data played back at different
times will (must) produce the same sound waves particularly if that audio data is in
digital format. This, at least, is the theory; endlessly reproducible sonic artefacts, identical copies of some original, and all neatly packaged for the digital consumer.
The reality of our experience is somewhat and frequently different and this, I believe, is where the standard view of sound fails. Two simple and easily accessible
demonstrations will suffice to show this. The first is the well-known McGurk Effect
(numerous examples of which can be found on the Internet including one produced
by the BBC14). Briefly, the syllable ‘baa’ is spoken, recorded, and played back over two
videos, one in which the subject articulates ‘baa’ (from whence the syllable ‘baa’ is recorded) and the other in which the subject articulates ‘faa’ (the lip and tongue movements are quite distinct between the two). Thus, the same audio data is reproduced
multiple times and so, on the same equipment and settings and in the same location, the same sound waves move through the air and so, according to the standard
view of sound, the same sound is reproduced time and time again. If sound is the
sound wave then the same sound should be heard at all times by all listeners or even
the single listener. This does not happen. Instead, what is heard are the syllables ‘baa’
and ‘faa’ over the appropriate image as if two different sets of sound waves are being
produced. If part of the standard view of sound is that sound is the object of our hearing, it is interesting to note that listeners, when asked what they hear, report the two
different sounds (i.e. the objects that they hear) despite the presence of only the one
recorded sound wave. One must conclude either, that if sound really is a sound wave,
then sound is not the object of our hearing or, that if sound really is the object of our
hearing, then sound waves are not sound. Either way, the standard view of sound is as
incoherent as Pasnau claims.
On the issue of the effects of cross-modality of hearing and vision (and other
senses) and how that issue affects our understanding of what sound is, I return to
14
http://www.youtube.com/watch?v=G-lN8vWm3m0
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
86
Mark Grimshaw
this later. The second demonstration is what might be broadly termed mimicry if not
outright trickery and it trades on ambiguity of the source and/or inexperience of the
listener but, unlike the McGurk Effect, does not use other sensory modalities. There
are two examples I present here both of which are easily testable. The first is to play
for a subject the recording of a Geiger counter. Anyone familiar with the characteristic crackling sound will invariably give the correct answer upon being asked what the
sound is but others will give different answers such as the preparation of popcorn. My
favourite example, though, is the extraordinary mimicry of the Australian lyrebird that
has a repertoire of sounds including camera shutter (replete with automatic winding
motor), chainsaws, alarms, and sirens. Those not familiar with this creature and its
ability to mimic such sounds will typically label the sounds as listed above (assuming
familiarity with these objects) while those familiar with the widely available examples
of the bird’s mimicry on the Internet will immediately label the sounds as the lyrebird.
In the above paragraph, the astute reader will have noticed a subtle shift in my use
of language. It is a shift in the use of the word ‘sound’ that now equates it with the
objects that (originally) produced the sound waves that were recorded for these two
examples. This relates to our everyday experience of sound that, again, I return to later
being content for now to state that this everyday definition of sound is not one that
finds favour with the standard view of sound (in this everyday definition, sound is the
sound wave source – logically, the sound produces the sound wave and so sound waves
supervene upon sounds – sound waves, then, are how sounds are made manifest15).16
Other definitions of sound
It is not my intention to thoroughly detail or even list all definitions of sound both
historical and current. Here, I present definitions that, in many respects, oppose the
standard view of sound although some still rely on the presence of sound waves
(which, in these definitions are not themselves sound); the definition encapsulated within sonic virtuality that I present later does not rely on this presence. For the
reader wishing to peruse a fuller list of sound definitions, I refer them to Nudds and
O’Callaghan and especially to Casati and Dokic.17 18 Those I present here are, I believe,
15
16
17
18
R. Casati and J. Dokic, “Some Varieties of Spatial Hearing,” in Sounds & Perception, eds. M. Nudds
and C. O’Callaghan, (Oxford: Oxford University Press, 2009), 98-99.
Even those adhering to the standard view of sound fall prey to this equivalence between sound and
sound source as Pasnau describes elsewhere in this article and as in this example from an acoustics
textbook: “Trained singers [...] exhibit an effect known as ‘vibrato’ in which their fundamental frequency is varied [...]” (Howard and Angus. Acoustics and Psychoacoustics. 193). Thus, the fundamental is a property of the singer and not a vibration in the medium of air. (I should state here that,
while it can be fun, it is not my intention to attack such trivial and venial deviations from scientific
orthodoxy – I would quickly become a hostage to fortune myself were I to do so – but rather to
demonstrate the yawning chasm and barely sustainable tension that exists between the standard
view of sound and our everyday experience and understanding of sound.
M. Nudds and C. O’Callaghan, eds., Sounds & Perception, (Oxford: Oxford University Press, 2009).
R. Casati and J. Dokic, “Sounds,” Stanford Encyclopedia of Philosophy (2005/2010). Accessed June
3, 2014, http://plato.stanford.edu/entries/sounds/.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
87
useful in this article’s context because they begin to take on a phenomenological flavour where the focus shifts from the physical properties of, and changes in, a medium
toward a focus upon sensation, perception, and cognition. Furthermore, those definitions that I do describe below are reasonably current as of date of writing. One should
not, however, make the mistake of assuming that stressing the importance of perception to any study of sound or, indeed, defining sound in perceptual terms is anything
new. Bregman’s work, for example, and although one must assume that he holds to
the standard view of sound because “sound enters the ear”,19 is important in that it is
an earlier statement in favour of analysing the perception of sound rather than sound
wave properties and even earlier instances of this type of thinking (although dealing
with music rather than sound) can be found in Carl Stump’s Tonpsychologie – an early
example of phenomenology.20
Before I begin, though, I wish to briefly present another physics-based definition of
sound that is of some interest and this is the conception of sound as comprising particles. This is an ancient idea that has recently (in its bare bones conception at least)
been given a new lease of life. The Greek philosopher Democritus, an early proponent
of classical atomism active during the 5th and 4th centuries BC, proposed that sound
was a stream of atoms emitted by a thing. What is of interest here is the description of
sound as comprising particles (the atoms) that was (re)proposed in 1947 by Gabor;21
a concept that has found use both in the fields of granular synthesis and quantum
physics (where the sound particle is termed the phonon22).
Sound as a property of an object
Pasnau, in his argument for the incoherency of the standard view of sound, suggests
instead that sounds reside “within the object that ‘makes’ them [...] sounds are either
vibrations of such objects, or supervene on those vibrations”.23 While this statement
is ambiguous (sound is either a vibration or is contingent upon a vibration), Pasnau
is quite clear about the location of the sound – it is within the object and thus “objects have sounds”.24 This question of the location of sound, rightly so, is fundamental to the genesis of any definition of sound, no less for Pasnau who appeals to the
veridicality of our auditory perception. This perception informs us that sound, like
colour, is located at the object and is part of the external world (hence we can use the
location of sound to locate objects in that external world): “we do not hear sounds
as being in the air; we hear them as being at the place where they are generated”.25
19
20
21
22
23
24
25
A. S. Bregman, Auditory Scene Analysis: The Perceptual Organization of Sound. (Cambridge, Massachusetts: MIT Press, 1990), 7.
C. Stumpf, Tonpsychologie. (Leipzig: Verlag Von S. Hirzel, 1883).
D. Gabor, “Acoustical quanta and the theory of hearing,” Nature 159 (1947): 591–594.
Specifically, it is the long-wavelength phonon (the acoustic phonon – the curious concept of the
‘sounding sound’ or ‘sound having to do with sound’ if one were to translate this literally) that gives
rise to sound (thus, the phonon itself is not sound).
Pasnau. What is Sound? 316.
Ibid. 316.
Ibid. 311.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
88
Mark Grimshaw
To locate sound elsewhere is a denial of the veridicality of our perception, a denial that, for Pasnau, is a step too far as it “shakes the very foundations of our cognitive assumptions”.26
Sound as an event
Basing his idea upon Aristotle’s statement that “sound is a particular movement in
air”27 – a statement that prima facie appears to support the standard view of sound
– O’Callaghan argues that sound is an event that sets a “medium into periodic motion”.28 He arrives at this definition via two other passages in De Anima: “everything
which makes a sound does so because something strikes something else in something else again, and this last is air” and “sound is the movement of that which can
be moved”.29 In this definition, sound waves, and indeed our perception of them, are
mere by-products of sound events and these events, comprising as they do movement,
involve change and therefore take place over time.
Sound as both object and event
Here, I come to the first phenomenological definition of sound in which sound neither possesses physicality nor is it the property of a physical (and external) object.
For Scruton, sound is both object and event; that is to say, sounds are secondary
objects and pure events because they do not undergo change, as physical objects do,
and they do not happen to anything.30 Scruton’s approach is closer to the valid, veridical definition of sound that I aim for because it is founded upon an assessment
of “those features of sound that make sound so important to us [...] socially, morally, and aesthetically”.31 Furthermore, Scruton takes account of acousmatic phenomena arguing that sound (e.g. in the case of radio broadcasts or audio recordings)
suffers no loss in coherence when divorced from its physical source and no loss in
what “is essential to the sound as an object of attention”.32 This perceptually based
definition of sound is heavily influenced by studies in music perception including
those phenomena that produce “a virtual causality that has nothing to do with the
process whereby sounds are produced”;33 the streaming and grouping of musical
notes based upon their similarity in timbre or pitch and their temporal proximity
to each other.
26
27
28
29
30
31
32
33
Ibid. 316. That Pasnau uses the word ‘assumptions’ does at least, in my view, keep the door ajar for
such a denial.
Quoted in C. O’Callaghan, “Sounds and Events,” in Sounds & Perception, eds. M. Nudds and
C. O’Callaghan, (Oxford: Oxford University Press, 2009), 27.
Ibid. 37.
Quoted in Ibid. 27.
R. Scruton, “Sounds as Secondary Objects and Pure Events,” in Sounds & Perception, eds. M. Nudds
and C. O’Callaghan, (Oxford: Oxford University Press, 2009), 50.
Ibid. 62.
Ibid. 58.
Ibid. 64.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
89
Non-cochlear sound
Another phenomenological definition of sound is presented by Riddoch who states
that sounds “first and foremost [are] worldly phenomena”.34 For Riddoch, the standard view of sound has failed in “demonstrating a causal mechanism linking our
neurological processes with the supposed subjective effect – the world of our perception”35 and, furthermore, that view does not take into account the idea that sounds
“are always in the first instance meaningful sounds”.36 On this basis, he argues that
cochlear sound (i.e. that sound involving sound waves) is in fact non-cochlear sound,
a group in which Riddoch also includes synaesthetic sounds (see cross-modality below), infrasonic sounds that are detectable by other body parts than the ear (see, for
example, Glennie37), and auditory imagination.
Everyday listening
While he does not provide a definition of sound per se, in his focus on what he terms
everyday sounds and everyday listening, Gaver also objects to the standard view of
sound and, for this reason, his work merits inclusion here.38 39 Importantly, Gaver’s
everyday sounds are non-musical sounds. He therefore focuses on sounds that are inharmonic and noisy because these, he explains, have never been the subject of acoustics or psychoacoustics, two fields that prefer to concentrate on periodic and therefore pitched sounds.40 While this situation has changed somewhat in the intervening
years, Gaver’s complaint that studies conducted under the aegis of the standard view
of sound display methodological bias still holds. This can be summed up neatly in a
statement provided by Gibson (by whom Gaver was influenced): the standard view
of sound “treats physical sound as a phenomenon sui generis, instead of as a phenomenon that specifies the course of an ecological event; sound as pure physics, instead of sound as potential stimulus information”.41 Thus, for Gaver, an incomplete
and inaccurate view of sound is formed when a) only musical tones are studied, and
the results then used to build up a picture of sound and auditory perception and
b) the language used to describe sound (frequency, amplitude, and so on) is inadequate to the task of describing our everyday experience of sound in which we typically equate sound and sound wave source – a subject is played a sound and asked by
the white-coated psychologist to describe what has been heard; ‘I hear a plane’, the
subject responds much to the frustration of the psychologist who requires the answer
34
35
36
37
38
39
40
41
M. Riddoch, “On the Non-cochlearity of the Sounds Themselves” (paper presented at International
Computer Music Conference, September 9-14, 2012, 14).
Ibid. 14.
Ibid. 13.
Glennie. Hearing Essay.
W. W. Gaver, “What in the World do we Hear? An Ecological Approach to Auditory Perception,” Ecological Psychology 5:1 (1993): 1–29.
W. W. Gaver, “How do we Hear in the World? Explorations in Ecological Acoustics,” Ecological Psychology 5:4 (1993): 285–313.
Gaver. How do we hear in the World? 286-287.
J. J. Gibson, The Senses Considered as Perceptual Systems, (Boston: Houghton Mifflin, 1966), 86.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
90
Mark Grimshaw
to be couched in the scientific language and terminology seen as appropriate to the
study of sound.42
Problems with these definitions
There are a number of issues that can be raised from the above definitions and that
leave me unconvinced about each of them. The first I wish to deal with is the location
of sound whether that sound is an object, a property of an object, or an event. Equally, the standard view of sound displays some incoherency within it as to the location
of sound, if not in its definition(s) (although the two ANSI definitions given above
provide two distinct locations) then in the lax ways to which these definitions have
been put. (These I have noted above.)
The location of sound is always given in relation to the organism that hears the
sound (thus, from the point of view of location of sound, sound is always the object
of hearing whatever its definition). The locations may therefore be roughly grouped
as distal, medial, or proximal in relation to the listener. The standard view of sound
(and the first ANSI definition) encapsulates the view that sound location is medial;
it is located, as a vibration, in a medium between the sound source and the listener.
This begs a number of questions not least of which is is sound therefore static or mobile?
Here, we run into further problems with the standard view in that sound (in such definitions) does indeed move through the medium but it surely must initially be distal
(as it issues forth from the sound source not the medium) and ultimately proximal
(this, presumably, is where the second ANSI definition takes over – sound is auditory
sensation originating in the ear of the listener).
Our everyday experience of the location of sound is typically that it is distal. That
is, the sound is located at the sound source, hence our everyday equation between
sound and sound source. This experience is expressed in terms such as ‘the knocking
is inside that wardrobe.’ But we also experience sounds as mobile (while the sound
source is static) as in ‘the knocking is coming from that wardrobe.’ Equally, we can
also experience sound inside the head and this need not be tinnitus or auditory hallucination or the imagination of sounds as those who have experienced in-head localization of sound through headphone use will be familiar with. One should therefore
be cautious with and somewhat mistrustful of definitions of sound (or at least theories about the location of sound in which location is fundamental to such definitions
of sound) that appeal to everyday experience as there is an incoherence here too, no
less than in the standard view of sound.
Thus, one should be distrusting of the sound as the property of an object definition provided by Pasnau because it is based on the supposed everyday location of
sound as being at the sound source (the object). Equally, there is an incoherence in
Pasnau’s view when the case of cinema (see further below) is used; sound is located in
our everyday listening on the screen yet there are no vibrations there of the type that
Pasnau claims are the sounds themselves. Regarding sound as an event, O’Callaghan is
42
Gaver. How do we Hear in the World? 286-287.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
91
frustratingly vague in stating that “sounds are events that take place near their sources,
not in the intervening space”43 while being explicit that “sounds are stationary relative
to their sources”.44
On the location and locatedness of sound, the evidence from neuroscience is of
interest as are the observed effects of cross-modality, particularly those effects arising
from the interoperation of vision and hearing. In the first instance, I specifically refer
to neuroplasticity as it relates to the auditory system. In auditory neuroscience, the localization of sound refers to the ability to locate the sound wave source viz. the source
of the sound. Thus, in this branch of science, sound is located distally – to locate the
sound is to locate the source of the sound because the sound is located at that object
or event. The book on which this article is based is able to go into far greater detail on
the subject of neuroplasticity than I have the space to do so here, therefore I will limit
myself to some brief notes on the most pertinent points to be made.
There are a number of studies (on humans, other animals, and birds) demonstrating that the artificial blocking of one ear displaces (in some cases significantly)
the horizontal localization of sound when compared to localization using two unimpaired ears (binaural hearing).45 (See, for instance, Slattery and Middlebrooks or
Kacelnik and colleagues.46 47) This phenomenon is as expected if one subscribes to
the Duplex Theory of sound localization.48 By itself, this raises questions as to the location and act of locating sound – one could infer, for example, that if sounds are
sound waves (which radiate out from a source) then sounds can be located as being
in or from a direction that is different to where the sound wave source is if binaural
hearing is impaired (the same effect is not noticed with vision if one eye is shut). In
this case, and logically, sound is not located at the sound source but is mobile and
in the medium. Nevertheless, these neurological studies prefer to state that sound is
being incorrectly localized (because sound is distal and located at the sound source
in this science).
43
44
45
46
47
48
O’Callaghan. Sounds and Events. 48. This intervening space is the medium between sound source
and listener and through which the sound waves (that are the effects of sounds) travel. What
the space or locus is between the sound source and the medium in which the event takes place,
O’Callaghan never states. One should also ask: precisely how near is ‘near’?
Ibid. 46.
See, for example, F. L. Wightman and D. J. Kistler, D (1997). “Monaural Sound Localization Revisited,”
Journal of the Acoustical Society of America 101: 2 (1997): 1050–1063 for some objections to these
studies on the basis that full monaural hearing is never achieved and that the test signals used do
not include everyday sounds.
W. H. Slattery III and J. C. Middlebrooks, “Monaural Sound Localization: Acute Versus Chronic Unilateral Impairment,” Hearing Research 75:1-2 (1994): 38–46.
O. Kacelnik, F. R. Nodal, K. H. Parsons, and A. J. King, “Training-induced Plasticity of Auditory Localization in Adult Mammals,” PLoS Biology 4:4 (2006): 0627–0638.
Briefly, organisms with binaural hearing use two overlapping channels of auditory information in
order to localize sound; for humans, lateral sound sources can be localized using interaural level difference if frequencies in the sound wave are above about 640Hz and interaural time difference if frequencies in the sound wave are below about 740Hz. Ambiguity about the direction of sound sources
directly ahead or behind and for sound source position on the vertical plane are resolved through assessing the spectral filtering effects of pinnae and through the simple expedient of moving our heads.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
92
Mark Grimshaw
Some of these studies and others demonstrate that such mislocalization of the
sound source (although it remains accurate localization of the sound wave or some
point along its front) can be corrected over time and such re-localization facilitated
and hastened along with training.49 The brain’s processes and physical structures for
localization change and adapt to the new monaural conditions and such neuroplastic ability is maintained throughout life.50 Interestingly, some studies suggest that the
brain can store more than one of these auditory localization schema, switching almost
immediately among them as the subject’s monaural and binaural conditions change.51
Outside of neuroscientific studies, there are numerous everyday examples of the human ability to localize the sound in some direction or at some spot other than from
where the sound wave travels or the sound wave source is. The prime example of this
is what is known in psychoacoustics as the ventriloquism effect52 or in film studies and
related areas, with some minor differences to the ventriloquism effect, as synchrony
or synchresis.53 54 Briefly, synchresis is the cross-modal effect of multi-modal stimuli
(in this case, visual and aural) being perceived as one event and it is an effect that is
fundamental to that cinema that uses diegetic sound. It relies on the ability to separate sound from the sound source (sound waves and the cinema’s loudspeakers) and
to localize the sound not at the sound source but on events, objects, or characters depicted on the screen. This bears some relation to the McGurk Effect discussed above in
that the sound heard depends to some extent on the image seen but the difference is
to do with the location and localizing of sound. As with the neuroscientific evidence
from sound localization studies above, synchresis is a process that relies on the ability
to locate sounds other than where their sound wave sources are and, thus, to perceive
that the sound travels in a direction other than the direction the sound waves themselves radiate from.55 This not only casts further doubt on the Duplex Theory, but also
further reinforces my doubts about the standard view of sound and other definitions
such as sound as property of an object or sound as event (located at or near its source).
49
50
51
52
53
54
55
e.g. P. M. Hofman, J. G. A. Van Riswick, and A. J. Van Opstal, “Relearning Sound Localization with
New Ears,” Nature Neuroscience 1:5 (1998): 417–421.
See, for example, M. Schnupp, I. Nelkin, and A. King, Auditory Neuroscience: Making Sense of
Sound (Cambridge, Massachusetts: MIT Press, 2010).
e.g. Hofman et al. Relearning Sound Localization with New Ears.
See, for example, D. H. Warren, R. B. Welch, and T. J. McCarthy, “The Role of Visual-auditory “Compellingness” in the Ventriloquism Effect: Implications for Transitivity Among the Spatial Senses,”
Perception & Psychophysics 30:6 (1981): 557–564.
J. D. Anderson, The Reality of Illusion: An Ecological Approach to Cognitive Film Theory, (Carbondale and Edwardsville: Southern Illinois University Press, 1996).
M. Chion, Audio-vision: Sound on Screen, trans. C. Gorbman, (New York: Columbia University
Press, 1994).
It may be that this is also an instance of neuroplasticity and an example of the ability to store and
recall instantaneously multiple auditory localization schema. The cinema, of course, is not a laboratory scenario where binaural hearing is artificially impaired but I am unaware of studies conducted
on cinema audiences looking for evidence of auditory neuroplasticity. One schema is learned and
used within the artificial environment of the cinema or when watching television or playing computer games, with at least one other schema for use elsewhere where sounds ought be perceived as
originating from the sound wave source – very important when crossing the road, for example. However, see below for an argument concerning our active locating of sound.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
93
One other objection I have to some of the non-standard view definitions relates to
the methodology used. This is particularly the case with Scruton’s definition. My objection here relates to Scruton’s insistence that humans alone have the ability to perceive order in sound (and thus they alone can perceive music as opposed to sound)
and the basis of his thinking behind the definition being studies of music perception.56 In the first case, this claim rests upon increasingly shaky ground.57 Second, in
limiting his thinking to music studies alone, Scruton derives a definition that is founded upon the wrong object of attention; music is neither sound alone nor is it alone
sound. O’Callaghan displays a similar bias towards musical or pitched sounds (which
I believe weakens his argument for a definition of sound) when he states that a sound
event sets a “medium into periodic motion”58 – not all sound waves are periodic and
thus O’Callaghan’s definition is not for all sounds.
The intention of this section has been to briefly enumerate and examine a number
of questions concerning the standard view of sound and other definitions of sound.
This is because, to my thinking, there are inconsistencies in the definitions and incoherency within some of the more accepted definitions and this leads me to question
their validity and therefore their effectiveness and suitability of purpose.59 The next
section presents an alternative definition. As with the definitions and their problems
listed above, the presentation of this new definition is necessarily brief. Its full exposition can be found within the book on which this article is based.
Sonic Virtuality
Sound is an emergent perception arising primarily in the auditory cortex and that is
formed through spatio-temporal processes in an embodied system
This definition of sound is the one that my co-author Tom Garner and I developed
first in response to our many objections to other definitions that we deemed wholly
unsatisfactory and/or incoherent and, second, as a result of several years collaboration particularly in the area of biofeedback and computer-game audio. There is no
intention to replace the standard view of sound – briefly, that sound is a sound wave
– although we contend that what is studied in physics and acoustics is not sound but
sound waves; seen from that point of view, the standard view of sound (waves) is perfectly valid (if a little inconsistent in its application), verifiable, and usable (usability
is what definitions are about, after all). Our intention is to provide a definition that a)
56
57
58
59
Scruton. Sounds as Secondary Objects and Pure Events. 62-66.
See A. D. Patel, J. R. Iversen, M. R. Bregman, and I. Shulz, “Studying Synchronization to a Musical
Beat in Nonhuman Animals,” The Neurosciences and Music III—Disorders and Plasticity: Annals of
the New York Academy of Sciences (2009), 459–469 for an overview of claims for music perception
abilities among non-humans.
O’Callaghan. Sounds and Events. 37.
I am not the first to point out the many inconsistencies and outright incoherency of the standard
physics-based definition(s) of sound and here I only provide a few of the many objections that could
be and are raised. Others providing fuller accounts include Pasnau (1999) and Riddoch (2012).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
94
Mark Grimshaw
answers the many questions and supports and explains the positions we have about
sound that are not answered or evidenced by other definitions and b) has a use-value
particularly in the area in which we work, work that may be described as human-centred sound design. As stated, Tom Garner and I have worked and continue to work in
the area of computer game audio and are especially interested in the real-time recognition of player affective states, feeding data on those states into the game engine, and
using them to synthesize or process audio in response (which then re-engages with
the players’ affective states thus creating a biofeedback loop). This can be done and we
have done it.60 61 What is difficult, though, is the recognition of emotions and affect
beyond basic valence and arousal and what is lacking is comprehensive knowledge on
the effects that sounds (rather than music) and changes in audio parameters in various contexts have on the individual’s arousal state and emotions. Clearly, a model of
sound founded upon a definition of sound that acknowledges the human factor, and
particularly its subjectivity, will be of some use here. And this not only has use in my
particular field but it also has use within other fields such as the field of recording music and sound. Before I move onto some thoughts about such usage, though, I need to
unpack the definition above and to explain its genesis.
Sound is an emergent perception . . .
This is the core of our definition; sound is a perception within our minds – thus, the
origin of sound is there – and its creation is an on-going, emergent process.
. . . arising primarily in the auditory cortex . . .
This emergent perception is initiated in a corporeal system and is centred upon the auditory cortex. A number of factors together contribute to the emergence of sound as these
include both corporeally external factors (the exosonus) such as sound waves and stimuli
across other sensory modalities, and internal factors (the endosonus) such as other perceptions (thus cross-modality), emotions, memory, reasoning, knowledge, and so forth.
. . . and that is formed through spatio-temporal processes . . .
The emergence of sound takes place over time and is inherently spatial being part of
the process of locating self in the world.
. . . in an embodied system.
Following ideas expressed in the field of embodied cognition, the perception that
is sound is a brain- and body-based function and is indivisible from the wider environment – the embodied system in which sound emerges comprises brain, body,
and environment.
60
61
M. Grimshaw and T. Garner, “Embodied virtual acoustic ecologies of computer games,” in The Oxford Handbook of Interactive Audio, eds. K. E. Collins, B. Kapralos, and H. Tessler, 181–195. (New
York: Oxford University Press, 2014).
T. A. Garner and M. Grimshaw, “The Physiology of Fear and Sound: Working with Biometrics toward Automated Emotion Recognition in Adaptive Gaming Systems,” IADIS International Journal
on WWW/Internet 11:2 (2013).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
95
From this definition, several key propositions and assertions arise that are more fully
stated and supported in the book:
– It is the sonic aggregate that is the engine for the emergence of sound and it comprises fluid and dynamic spatio-temporal relationships of sensuous/nonsensuous,
material/immaterial things and phenomena
– The set of material, sensuous components of the sonic aggregate we term the exosonus – it is not a requirement for the emergence of sound
– The set of immaterial, nonsensuous components of the sonic aggregate we term the
endosonus – it is a requirement for the emergence of sound
– Using Deleuzian terminology62, all components of the sonic aggregate, whether
exosonic or endosonic, are actual as is sound. The actualization that is the emergence of sound as a perception derives from processes that are themselves virtual
– It is the force of potential in the fluid, dynamic sonic aggregate that, when achieved
under particular spatio-temporal conditions, leads to the actualization itself
– The hear and now of sound is the emergent actualization in the here and now and it
requires the presence of a perceiver
– In order to locate ourselves, we cognitively offload the location of sound onto the
environment – this is an active (re)location of sound
– Sound is meaningful because different types of meaning form part of the endosonus; sound waves are inherently meaningless
– A sound wave may be acousmatic but a sound is never acousmatic. Hence, reduced
listening (the concept of a sound divorced from source, cause, environment, and
meaning – and thus listenable to in and of itself) remains a theoretical concept not
achievable in practice
– Discussions of real sound and virtual sound are invalid as there is no such distinction
– The emergence of sound can determine epistemic perspective as epistemic perspective can determine the quality and nature of the emergent sound
– Imagined sound is as much sound as that sound perceived in the presence of
sound waves.
It should be immediately clear from the above that our description is, in large part,
founded upon the Deleuzian concept of the virtual. Thus, the sonic aggregate has
some equivalence to the concept of the virtual cloud. It is from this sonic aggregate,
with its dynamic tension and shifting potentials, that sound is actualized as an emergent perception. This aggregate comprises the endosonus and may optionally, as it frequently does, also comprise exosonic components such as sound waves. In the assertion that sound waves are not necessary to the emergence of the sound perception, the
definition thus accounts for a wider and, to my mind, more valid view of what sound
is than the standard view is capable of doing. In doing so, it tackles a number of in62
e.g. virtuality encompasses the idea of a dynamic virtual cloud of potentiality from which actualization
occurs. It is important to note that, in this conception of virtuality, the virtual is a mode of the real
and not to be contrasted to the real (pace real worlds and virtual worlds).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
96
Mark Grimshaw
consistencies and sonic anomalies at odds with this standard view such as tinnitus
(which is sound) and how the McGurk Effect is perceived as different sounds despite
the presence of the same sound wave.
I have previously stated that the question of the location of sound is fundamental
to many definitions of sound; if one can decide where sound is then one is halfway to
defining what sound is because the location is descriptive of the sound’s environment.
I do not think that the concept of sonic virtuality is likewise so dependent upon the
location of sound but, nevertheless, I feel bound to use the definition and concept to
explain the location of sound.
How is it that, for something so fundamental to the daily living of so many organisms including humans, for something that for so long has been an object of intense
study, and for something for which the standard view of sound offers a medium-specific location, there is so much disagreement about the location of sound? Perhaps
there is something wrong with all those definitions. The concept of sonic virtuality
offers this explanation: the origin of sound is in the secondary auditory cortex63 but,
where the mind is a system comprising brain, body, and environment, sound can be
located anywhere in this system. This is not to say that the definitions founded upon
or arguing that sound is distal, medial, or proximal are all, and at one and the same
time, correct. Far from it. Such definitions use the passive form of ‘to locate’ and so
sound is to be found somewhere, typically out there in the world.
Sonic virtuality, on the other hand, uses ‘to locate’ in the active sense. We ourselves
locate, that is, place (rather than find) sounds somewhere within the system of which
the mind is comprised. This is the cognitive off-loading that is one of the tenets of embodied cognition. This locating of sound is learned and develops through early childhood into an automatic act of placement and it explains the close and cross-modal
relationship to vision as we, while babies, cast around, looking for the source of a
sound wave, beginning to learn the associations between movement, size, and material and sound waves. It is a skill that can be lost and relearned (as neuroplasticity
demonstrates) and it can be adjusted at will to new physical environments (such as
the cinema or when playing a computer game where the location of a sound wave in
air or the location of the origin of that sound wave are very different to those places
where we actively locate the sound). The ability to actively locate sound arises from
the necessity of being and acting within an ecology and to listen is to be present in a
world, whether virtual or real – sound is the means not to locate objects in an external
world but rather to locate our selves in relation to other objects.
Clearly, and as I have stressed throughout, an article cannot go into the detail that
a book can; Sonic Virtuality is a more substantial and substantiated treatment of the
subject matter and, in addition to what has been cursorily covered here, also includes
evidence and support for the concept from the virtual acoustic environments of com63
This is justified more substantially in the book but, briefly, the assertion arises because this is the
common locus where activity is observed when sound is perceived either in the presence of sound
waves or through an act of imagination, bidden or unbidden.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
97
puter games, the philosophy of knowledge and belief, auditory imagery, hallucination
and other pathologies, imagination, and creativity and is at pains to demonstrate the
practical purpose of the definition. It is on this last point that I will close the article.
The presentation upon which this article is based was delivered at a conference on
sound and music production and so, although the book gives different examples, I
will provide an argument for the use-value of the definition and concept of sonic virtuality within the domain of sound design and will do this by means of a brief example of the concept as it might be used.64
If, as our definition asserts, sound is a perception that arises in the brain, then the
obvious question to ask is can we extract it? Before I look more closely at this question,
I need to clarify the terminology, particularly the word ‘extract.’ If sound is a perception in the brain, then, once extracted from the brain, it is no longer sound. To be
more precise, then, ‘extraction’ is used metonymically here to stand for the transposition of sound to audio data. My speculative answer to the question then is, and for
reasons outlined below, no, not yet, but it will be possible in the near future. Thus, one
should begin to lay the groundwork for this possibility and the necessity of this is one
of the reasons for devising the concept of sonic virtuality.
Neural decoding, as it is known within neuroscience circles, is a developing field
used, among other purposes, for the extraction of still and moving images from the
visual cortex.65 Neural activity is decoded and visually represented on a monitor and
this decoding is accurate enough to be able to recognize features of the image the subject is concurrently gazing at. There is work afoot on extracting speech66 and plans
to extract music from the brain67 but, first, there are good reasons to state that music
and speech are conceptually and perceptually distinct to sound (even if somewhat and
sometimes related) and, second, there are different processes and different regions in
the brain involved.68 69 Previous and current composers and performers that make
64
65
66
67
68
69
One possibility that did not make it into the book concerns the cross-modality of sound and smell
perceptions and its relevance to presence in virtual environments. See Grimshaw, M. and WaltherHansen, M., “The Sound of the Smell of My Shoes,” (in proceedings of the 10th Audio Mostly Conference, October 7-9, 2015).
See, for example, S. Nishimoto, A. T. Vu, T. Naselaris, Y. Benjamini, B. Yu, and J. L. Gallant, “Reconstructing Visual Experiences from Brain Activity Evoked by Natural Movies,” Current Biology 21
(2011): 1641-1646. I must stress here that the orthodoxy of neuroscience does not express the same
concept of image as I do of sound. That is, in this case, vision is the perception of an image and that
perception is not the creation or emergence of an image itself. Thus, what is being extracted is not
the image but the perception of that image.
B. N. Pasley, S. V. David, N. Mesgarani, A. Flinker, S. A. Shamma, N.E. Crone, et al., “Reconstructing
Speech from Human Auditory Cortex,” PLoS Biology 10:1 (2012).
J. Thompson, M. Casey, and L. Torresani, “Audio Stimulus Reconstruction using Multi- source Semantic Embedding, in Neural Information Processing Systems,” (paper presented at the Neural Information Processing Systems Workshop on Machine Learning and Interpretation in Neuroimaging, December
9-10, 2013).
R. J. Zatorre, P. Belin, and V. B. Penhune, “Structure and Function of Auditory Cortex: Music and
Speech,” Trends in Cognitive Sciences 6:1 (2002): 37-46.
R. J. Zatorre and A. R. Halpern, “Mental Concerts: Musical Imagery and Auditory Cortex,” Neuron 47
(2005): 9-12.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
98
Mark Grimshaw
use of data gathered from brain activity are merely using audification and sonification
techniques70 rather than the extraction of sound (or music or speech).
The definition of sound presented here, and especially the concept of the sonic aggregate, is necessary to any thinking on how to extract sound.71 It makes sense to conceive of sound as emerging from something like a sonic aggregate in the context of
phenomena such as the McGurk Effect and when aware of the effect of higher order
cognitive processing on auditory perception72 and the effect of cross-modal sensation
on auditory perception.73 74 75 Sound may arise primarily in the auditory cortex but
there are other sensory, perceptual, and cognitive factors involved too and these must
be taken into account when attempting to extract sound.
Let us assume that, sometime in the future, it becomes possible to extract sound
from the brain. Then what? If the extraction of sound is the transposition of sound to
audio data then it becomes possible to conceive of, and design the possibility for, a
process of sound design whereby the sound designer imagines and projects sound into
a digital audio workstation. Thus, the sound conjured up in the absence of sound waves
becomes available for manipulation as audio data, for copying, and for reproduction
in the form of sound waves. This will fundamentally change the role of sound designer.
Naturally, as sound in our concept is highly individual, the sounds emerging from the
sonic aggregates of others then exposed to these sound waves are likely to have some
differences to the original. This exemplifies the inherent subjectivity of the emergent
perception and the necessity for the involvement of the perceiver that becomes evident
from any study of sound – the sound is the object of perception, the sound wave is
the object of sensation – and this difference is what missing from the standard view of
sound and neither is it fully accounted for by other definitions of sound.
70
71
72
73
74
75
See M. Ortiz, “A Brief History of Biosignal-Driven Art: From Biofeedback to Biophysical Performance,” econtact 14:2 (2012) for an overview.
Or a definition and concept that are similar.
e.g. N. Bunzeck, T. Wuestenberg, K. Lutz, H-J. Heinze, and L. Jancke, “Scanning Silence: Mental Imagery of Complex Sounds,” NeuroImage 26:4 (2005): 1119–1127.
M. Hoshiyama, A. Gunji, R. Kakigi, “Hearing the Sound of Silence: A Magnetoencephalographic
Study,” NeuroReport 12:6 (2001): 1097–1102.
J. Voisin, A. Bidet-Caulet, O. Bertrand, and P. Fonlupt, “Listening in Silence Activates Auditory Areas:
A Functional Magnetic Resonance Imaging Study,” The Journal of Neuroscience 26:1 (2006): 273–
278.
R. J. Zatorre, “There’s More to Auditory Cortex than Meets the Ear,” Hearing Research 229 (2007):
24-30.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
A brief argument for, and summary of, the concept of Sonic Virtuality
99
Abstracts
Sonic virtuality is a conceptualization of sound devised with several purposes in
mind. First, it provides a holistic definition of sound that takes account of factors beyond the bare physics of sound waves and their propagation. Second, in providing
such a definition, it attempts to explain a number of sonic anomalies that existing
definitions of sound, of which there are several, do not satisfactorily explain. Third,
in its concept of sound as an emergent perception sited within the mind, it provides
the conceptual framework to work with sound in the context of new and developing
technologies. The essay begins with an enumeration of several existing definitions of
sound and problems with them, focusing in particular upon the western world’s dominant definition of sound as a sound wave, and then provides a brief exposition of
sonic virtuality before concluding with a speculative example of its use.
Sonic virtuality er en måde at begrebsliggøre lyd, som er udarbejdet med en række forskellige formål. For det første rummer begrebet en holistisk definition af lyd, der omfatter faktorer som udelades i den fysiske forståelse af lydbølger og deres spredning.
For det andet gør begrebet det muligt at forklare en række lydmæssige uregelmæssigheder, der ikke forklares ved eksisterende definitioner af lyd. For det tredje, ved at udvikle et lydbegreb, hvor lyd forstås som en emergent perception der eksisterer i sindet,
gives en begrebsramme til at arbejde med lyd i relation til nye og kommende teknologier. Artiklen begynder med en redegørelse for eksisterende definitioner af lyd og deres
iboende problemer. Der fokuseres især på den dominerende definition af lyd i den
vestlige verden – lyd som en lydbølge. Derefter gives en kort gennemgang af sonic virtuality og der konkluderes med et tænkt eksempel på begrebets anvendelse.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
JUSTIN CHRISTENSEN
Virtual Distance and Soundstage,
and their Impacts on Experienced
Emotional Valence
Introduction
For this paper, I borrow theories from animal ethology research, which suggest that
varying levels of valence are experienced in response to being varying distances from
a predator. Adopting ideas from the theory of refinement, I examine whether these
survival instincts have relevance for a music listening experience, focusing on how a
listener’s virtual distance from the signal source might alter their experienced emotional valence of music. As well, I investigate whether changes to distance cues affect
all of the presented basic emotion categories (happy, sad, fearful, calm) similarly, or if
changes to distance cues affect each of the basic emotional categories differently. I also
add a third distance condition, including trials that combine audio and haptic stimulation in the nearfield. This follows from the theory that multimodal experiences are
often deemed to be more trustworthy signals.
Links between Aesthetic Emotions and Survival Emotions
Prior to 2000, the bulk of music research focused on music as a cultural invention. As
a result, ‘biomusicology’ with its investigations into the biological functions of music
is still a relatively new and growing field.1 In 2007, Frijda and Sundararajan presented a theory of refinement2 as an elaboration on James’ theory of emotions,3 where
aesthetic emotional responses are seen as refined forms of more coarse survival responses. ‘Refined’ for this theory means that the emotions do not obviously manifest
themselves in overt behaviors such as fight or flight responses. This theory of refinement suggests that emotional responses induced by aesthetic response are the same
as emotions aroused by survival instincts, differing mainly in regards to the magnitude of the emotional response, and their resultant behavioral and physiological
1
2
3
Nils Lennart Wallin and Björn Merker, The Origins of Music (MIT Press, 2001.
Nico H. Frijda and Louise Sundararajan, “Emotion Refinement: A Theory Inspired by Chinese Poetics,” Perspectives on Psychological Science 2, no. 3 (September 2007), 227–41, doi:10.1111/j.17456916.2007.00042.x.
William James, “II.—What Is an Emotion?,” Mind, no. 34 (1884), 188–205.
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
102
Justin Christensen
responses. This fits well with Patel’s “transformative technology of the mind” theory.
Patel states:
“neuroscientific research suggests that music is an invention that builds on diverse pre-existing brain functions, rather than a trait that originated via processes
of natural selection (…) growing evidence from neuroscience also suggests that
music is biologically powerful, meaning that it can have lasting effects on nonmusical abilities…”4
As further evidence for music’s connection to everyday goal-directed emotions, Peretz
has hypothesized that music’s universal appeal is likely due to music being “…particularly well suited (or designed) to invade emotion circuits that have evolved for emotional vocalizations.”5
While musical emotions are often considered personal and enigmatic, emotional
responses to music within a ‘basic emotion’ framework (happy, sad, fearful) are remarkably invariant across listeners,6 and are already easy to distinguish by children
already by the age of six.7 Schmidt and Trainor also found that emotional valence (attractiveness vs. averseness) in music is also significantly correlated to specific brain
activation patterns as measured by EEG.8 Several other EEG studies by Altenmüller
et al.,9 Flores-Gutiérrez et al.,10 and Tsang et al.11 also support these results for hemispheric lateralization of function in music listening, where “Positive emotions were
related to a preponderance of left frontal activation whereas negative emotions resulted in a more bilateral fronto-temporal activation with preponderance of the right
hemisphere.”12 These results show that there are numerous benefits for using music
to study the neural correlates of emotion. Stefan Koelsch has listed a number of them
that are applicable for the current study.
4
5
6
7
8
9
10
11
12
Aniruddh D. Patel, “Music, Biological Evolution, and the Brain,” in Emerging Disciplines, ed. C. Levander & C. Henry, (Texas, Rice University Press: 2010), 91–144.
Isabelle Peretz, William Aubé, and Jorge L. Armony, “Towards a Neurobiology of Musical Emotions,”
The Evolution of Emotional Communication: From Sounds in Nonhuman Mammals to Speech and Music in
Man, (Oxford University Press, 2013), 277.
Ibid.
Joseph G. Cunningham and Rebecca S. Sterling, “Developmental Change in the Understanding of
Affective Meaning in Music,” Motivation and Emotion 12, no. 4 (December 1988), 399–413,
doi:10.1007/BF00992362.
Louis A. Schmidt and Laurel J. Trainor, “Frontal Brain Electrical Activity (EEG) Distinguishes Valence
and Intensity of Musical Emotions,” Cognition & Emotion 15, no. 4 (2001), 487–500,
doi:10.1080/0269993004200187.
Eckart Altenmüller et al., “Hits to the Left, Flops to the Right: Different Emotions during Listening
to Music Are Reflected in Cortical Lateralisation Patterns,” Neuropsychologia 40, no. 13 (2002): 2242–
56, doi:10.1016/S0028-3932(02)00107-0.
Enrique O. Flores-Gutiérrez et al., “Metabolic and Electric Brain Patterns during Pleasant and Unpleasant Emotions Induced by Music Masterpieces,” International Journal of Psychophysiology 65, no. 1
(July 2007), 69–84, doi:10.1016/j.ijpsycho.2007.03.004.
C. D. Tsang et al., “Frontal EEG Responses as a Function of Affective Musical Features,” Annals of the
New York Academy of Sciences 930, no. 1 (June 2001), 439–42,
doi:10.1111/j.1749-6632.2001.tb05764.x.
Altenmüller et al., “Hits to the Left, Flops to the Right,” 2242.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
103
“…(1)Music is capable of evoking strong emotions (…) (5) Both listening to
music and making music can evoke emotions, enabling investigators to study
interactions between emotion and action. (6) Music can be used to study the
time course of emotional processes, with regard to both short-term emotional
phenomena (in the range of seconds) and longer-term emotional phenomena
(…) (7) It appears that, with regard to human evolution, music is originally
a social activity. Therefore, music is well suited to study interactions between
emotion and social factors. [emphasis in the original]”13
Emotional Valence and Distance
The present study explores how acoustic processing of musical examples (from differing basic emotional categories: happy, sad, fearful, calm) to simulate acoustic spaces,
where the virtual sound sources appear at varied virtual distances from the listener,
might evoke distinct emotional approach/withdrawal motivations in subjects, seen
through anterior brain asymmetry of induced EEG event-related oscillations (EROs).
Emotional valence has often been tied to action tendencies towards rewards (appetitive stimuli) and away from danger or punishment (aversive stimuli). Valence falls
under appraisal theories of emotion, where “A common tenor of these theories is that
(external or internal) phenomena, circumstances, actions, individuals, or objects are
evaluated as “good”, that is, contributing to achieving a goal, or “bad,” that is, obstructing the achievement of a goal”.14 The use of virtual distance for this study is inspired by work in animal ethology,15 and by the work of David Huron.16
Animal ethology research has shown that there are changes to the levels of both physiological arousal and action tendency motivations (valence) that result from varying an animal’s distance from a predator. Bracha17 presents the theory that a freeze response of an
animal is activated when flight or fight are no longer effective options, when a danger is
too close and too oppressive to effectively escape from. Relatedly, according to the “distance-dependent defense hierarchy”18, a close predator induces a fight response, while a
moderately close predator induces a flight response. Combined, these two theories present
three different distance associations for three of the F’s of a fight-or-flight response. In addition to these categories of fear responses with their distinct distance associations, anticipatory anxiety responses can frequently be induced in response to distant perceived threats.
13
14
15
16
17
18
Stefan Koelsch, Brain and Music (John Wiley & Sons, 2012), 203.
Stefan Koelsch, “Music-Evoked Emotions: Principles, Brain Correlates, and Implications for Therapy,”
Annals of the New York Academy of Sciences 1337, no. 1 (March 2015), 193, doi:10.1111/nyas.12684.
John Maynard Smith and David Harper, Animal Signals (Oxford: Oxford University Press, 2003).
David Huron, “Affect Induction through Musical Sounds: An Ethological Perspective,” Philosophical Transactions of the Royal Society of London B: Biological Sciences 370, no. 1664 (March 2015):
20140098, doi:10.1098/rstb.2014.0098; David Huron, “Understanding Music-Related Emotion: Lessons from Ethology,” 2012, 23–28.
H. Stefan Bracha, “Freeze, Flight, Fight, Fright, Faint: Adaptationist Perspectives on the Acute Stress
Response Spectrum.,” CNS Spectrums 9, no. 09 (September 1, 2004), 679–85,
doi:10.1017/S1092852900001954.
Stanley C. Ratner, “Immobility of Invertebrates: What Can We Learn?,” The Psychological Record, (1977).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
104
Justin Christensen
Anxiety responses are often considered ameliorated fear responses, neither as immediately aversive nor inducing as much arousal in the sympathetic nervous system as
fear.19 In comparing fear responses, the “distance-dependent defense hierarchy” found
that the emotional quality of a fight response was both more negatively motivating
and caused higher levels of arousal in the prey than the emotional quality of a flight
response.20 In a slight variation from these other anxiety and fear responses, which
show a direct negative correlation between the averseness experienced and the arousal
levels exhibited, a fright response displays an acutely negative valence with a reduced
arousal level. This combination is likely the result of the prey’s intention to reduce
movement as much as possible as part of the freeze response.21
Related to ethological action tendencies, Isobel Peretz considers action tendencies and movement to be key components of emotional reactions to music. “Much
music is composed with very specific action tendencies in mind – getting people to
sing, to dance, march and fight, or work and play together.”22 While these action
tendencies for music are unrelated to animal fear response action tendencies, following the theory of refinement mentioned earlier, music’s action tendencies can
be seen to be more refined versions of the coarse animal survival instincts. This is
further supported by music having similar reward motivated brain activations towards food23 and drugs,24 and away from threats.25 Juslin and Västfjäll suggest
that music uses mimesis (or emotional contagion) to express emotional valence,
as music can mimic the ‘natural emotional attributes’ that are found in emotional
speech signals (as has been investigated in Lundqvist et al.26, Peretz27, and Lima &
Castro28).29 As a result, I chose the musical selections for this investigation to follow
19
20
21
22
23
24
25
26
27
28
29
Bruce Duncan Perry and Maia Szalavitz, The Boy Who Was Raised as a Dog: And Other Stories from a
Child Psychiatrist’s Notebook: What Traumatized Children Can Teach Us about Loss, Love, and Healing
(New York: Basic Books, 2006).
Ratner, “Immobility of Invertebrates.”
Stefan Bracha, “Freeze, Flight, Fight, Fright, Faint: Adaptationist Perspectives on the Acute Stress Response Spectrum.”
Isabelle Peretz, “The Nature of Music from a Biological Perspective,” Cognition, The Nature of Music,
100, no. 1 (May 2006), 22–23, doi:10.1016/j.cognition.2005.11.004.
Dana M. Small et al., “Changes in Brain Activity Related to Eating Chocolate,” Brain 124, no. 9 (September 2001), 1720–33, doi:10.1093/brain/124.9.1720.
Anne J. Blood and Robert J. Zatorre, “Intensely Pleasurable Responses to Music Correlate with Activity in Brain Regions Implicated in Reward and Emotion,” Proceedings of the National Academy of Sciences 98, no. 20 (September 2001), 11818–23, doi:10.1073/pnas.191355898.
Nathalie Gosselin et al., “Amygdala Damage Impairs Emotion Recognition from Music,” Neuropsychologia 45, no. 2 (2007), 236–44, doi:10.1016/j.neuropsychologia.2006.07.012; Patrik N. Juslin and
Daniel Västfjäll, “Emotional Responses to Music: The Need to Consider Underlying Mechanisms,”
Behavioral and Brain Sciences 31, no. 05 (October 2008), 559–75, doi:10.1017/S0140525X08005293.
Lars-Olov Lundqvist et al., “Emotional Responses to Music: Experience, Expression, and Physiology,”
Psychology of Music, (October 2008), doi:10.1177/0305735607086048.
Peretz, Aubé, and Armony, “Towards a Neurobiology of Musical Emotions.”
César F. Lima and São Luís Castro, “Speaking to the Trained Ear: Musical Expertise Enhances the Recognition of Emotions in Speech Prosody,” Emotion 11, no. 5 (2011), 1021–31, doi:10.1037/a0024521.
Patrik N. Juslin, László Harmat, and Tuomas Eerola, “What Makes Music Emotionally Significant?
Exploring the Underlying Mechanisms,” Psychology of Music, (August 2013), 1-25,
doi:10.1177/0305735613484548.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
105
the natural attributes of emotional speech patterns. This will be discussed more in
the methodology section below.
For evaluating the valence aspect (approach withdrawal motivations) of emotions
evoked in the subjects of this study, I measured the subjects with EEG. In past EEG
studies, both approach/withdrawal motivations30 and positive or negative valence31
have been found to be correlated with EEG hemispheric asymmetry. Activity in the left
prefrontal cortex has been associated with approach motivations and positive valence,
while right prefrontal cortex activity has been associated with withdrawal motivations
and negative valence. Connected with this, Rutherford and Lindel have associated positive emotions, such as happiness, with a motivation to move towards this cause of
happiness. They similarly have associated negative emotions, such as fear or sadness, 32
with a motivation to withdraw from the situation that causes this negative emotion.33
The hemispheric asymmetry in the brain linked with emotional valence is observed
by measuring alpha lateralization (the inverse power in the Alpha band34 of neural oscillations, now defined as 8-12 Hz.), as lower Alpha power has been connected with greater neural activity in the corresponding brain region.35 These hemispheric asymmetries
measured by alpha lateralization have already found to be a good fit with self-reported
affective responses both for positive and negative film clips36 and for music examples,37
so this gives promise for this methodology. Recently, Kragel and LaBar38 also found distinct dimensional mapping between differing emotional states through FMRI testing,
which is quite promising for a more precise neural mapping of emotional states.
While these results all support significant correlations between localizations of
brain activation patterns and emotional valence, the emotions perceived often differ
from the emotions evoked in the listeners. Zentner and Scherer state:
30
31
32
33
34
35
36
37
38
Eddie Harmon-Jones and J.J.B. Allen, “Anger and Frontal Brain Activity: EEG Asymmetry Consistent
with Approach Motivation despite Negative Affective Valence,” Journal of Personality and Social Psychology 74, no. 5 (1998): 1310–16, doi:10.1037/0022-3514.74.5.1310.
Robert E. Wheeler, Richard J. Davidson, and Andrew J. Tomarken, “Frontal Brain Asymmetry and
Emotional Reactivity: A Biological Substrate of Affective Style,” Psychophysiology 30, no. 1 (January
1993), 82–89, doi:10.1111/j.1469-8986.1993.tb03207.x.
With the exception of anger, which is a negative, but approach motivated emotion.
Helena J. V. Rutherford and Annukka K. Lindell, “Thriving and Surviving: Approach and Avoidance
Motivation and Lateralization,” Emotion Review 3, no. 3 (July 2011), 333–43,
doi:10.1177/1754073911402392.
Other frequency bands of neural oscillations are named Delta (0-4 Hz.), Theta (4-8 Hz.), Beta (1230 Hz.), Gamma (30-80 Hz.)
DB Lindsley and JD Wicke, “The Electroencephalogram: Autonomous Electrical Activity in Man and
Animals,” Bioelectric Recording Techniques, 1974.
Andrew J. Tomarken, Richard J. Davidson, and Jeffrey B. Henriques, “Resting Frontal Brain Asymmetry Predicts Affective Responses to Films,” Journal of Personality and Social Psychology 59, no. 4 (1990):
791–801, doi:10.1037/0022-3514.59.4.791; Wheeler, Davidson, and Tomarken, “Frontal Brain Asymmetry and Emotional Reactivity.”
Tsang et al., “Frontal EEG Responses as a Function of Affective Musical Features”; Altenmüller et al.,
“Hits to the Left, Flops to the Right.”
Philip A. Kragel and Kevin S. LaBar, “Multivariate Neural Biomarkers of Emotional States Are Categorically Distinct,” Social Cognitive and Affective Neuroscience, (March 2015), nsv032, doi:10.1093/
scan/nsv032.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
106
Justin Christensen
“The aforementioned research tradition made valuable contributions to an understanding of music-specific affects—for example, by pointing to the possibility that
canonical emotion labels may not do justice to the emotions evoked by music”.39
They also found that while positive emotions are aroused similarly to how they are
perceived, negative emotions are more often perceived than felt. The studies by Gabrielsson,40 and Kallinen and Ravaja41 both support this view, with findings that sad and
fearful music are still quite likely to evoke positive emotional valence in listeners.
The trustworthiness of a signal: multimodal vs. unimodal
As an added component to my explorations with this study, I added a multimodal
nearfield condition to the stereo audio nearfield and stereo audio virtually distant
conditions. I did this as I was interested in studying whether the addition of a multimodal component could increase the trustworthiness or closeness of the signal. I
only added the haptic stimulation to the nearfield condition, since tactile stimulation is limited by the length of a subject’s limbs, and thus this biological limitation
might further increase the trustworthiness of a close signal, and add to a greater sense
of nearness.
As one means for deciding which stimuli animals should motivate themselves towards or away from, animals judge the credibility of the signal given off by stimuli in
their environment. Scott-Phillips has described a signal as:
“Any act or structure that (i) affects the behaviour of other organisms; (ii)
evolved because of those effects; and (iii) which is effective because the effect
(the response) has evolved to be affected by the act or structure.”42
As a result, trustworthy signals given off by animals are highly social encounters, benefitting both the sender and receiver of the signal. These signals are often multimodal,
as trustworthiness can be heightened through having trustworthy performances registered across multiple perceptual modalities.43
Among these trustworthy signals, we can include many human emotional expressions. Juslin usefully summarized Scherer’s definition of emotion44 as:
39
40
41
42
43
44
Marcel Zentner, Didier Grandjean, and Klaus R. Scherer, “Emotions Evoked by the Sound of Music:
Characterization, Classification, and Measurement,” Emotion 8, no. 4 (2008): 496,
doi:10.1037/1528-3542.8.4.494.
Alf Gabrielsson, “Emotion Perceived and Emotion Felt: Same or Different?,” Musicae Scientiae 5, no.
1 suppl (September 2002), 123–47, doi:10.1177/10298649020050S105.
Kari Kallinen and Niklas Ravaja, “Emotion Perceived and Emotion Felt: Same and Different,” Musicae Scientiae 10, no. 2 (September 2006), 191–213, doi:10.1177/102986490601000203.
T. C. Scott-Phillips, “Defining Biological Communication,” Journal of Evolutionary Biology 21, no. 2
(March 2008), 388, doi:10.1111/j.1420-9101.2007.01497.x.
Ibid.
Klaus R. Scherer, “Psychological Models of Emotion.,” in The Neuropsychology of Emotion., ed. Joan
Borod (New York: Oxford University Press, 2000), 137–62.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
107
“…emotions are relatively brief and intense reactions to goal-relevant changes in
the environment that consist of many subcomponents: cognitive appraisal, subjective feeling, physiological arousal, expression, action tendency, and regulation…”45
Emotions have evolved and are an influential part of social practice as they reliably
convey information of interest to both the sender and the receivers of these emotional
expressions through changes in speech prosody, speech rate, facial expression, changes
in gestures, general body language and action tendencies.46
For sound localization, humans use micro-time deviations between the ears, loudness cues, spectral information, as well as correlations from early and late reflections.47
Compiled together, these features help give cues to localize a sound source. The more
these features have a combined agreement with past experiences of sound localizations,
the more easily the listener adapts to them and accepts them as being trustworthy.48
The soundstage of a recording is the recreation of a virtual musical experience in a
3-dimensional space. From the recording, the listener is often able to localize the individual sound sources in this virtual space, and derive cues for their distance from the
source.49 This current study explores how these distance cues can alter the evoked emotional response for musical examples with diverse approach/withdrawal motivations.
My hypothesis is that musical stimuli should cause stronger valenced responses in
the nearfield than at a distance regardless of whether they are negative or positive. The
motivation to approach appetitive stimuli in the nearfield should be exhibited more
strongly in the nearfield than from a distance. Likewise, motivations to withdraw from
aversive stimuli should be exhibited more strongly in the nearfield than from a distance. Thus, music experienced as being negatively valenced at a distance should be
even more negatively valenced in nearfield, and music that is experienced as having a
positive valence at a distance should be even more positively valenced in nearfield.
Method
Subjects
This study included 12 participants with a mean age of 24 years, including 7 males
and 5 females, with 8 of them having had over 5 years of musical training. The data
from two subjects were removed. One subject self-reported as being left-handed, and
since handedness has in the past shown an effect on alpha lateralization, their results
were removed prior to data analysis. The EEG results of a second subject included
45
46
47
48
49
Patrik N. Juslin and Petri Laukka, “Communication of Emotions in Vocal Expression and Music Performance: Different Channels, Same Code?,” Psychological Bulletin 129, no. 5 (2003), 770–814,
doi:10.1037/0033-2909.129.5.770.
Klaus R. Scherer and Paul Ekman, Approaches To Emotion (Psychology Press, 2014).
Jens Blauert, Spatial Hearing: The Psychophysics of Human Sound Localization (MIT Press, 1997).
P. Zahorik, E. Brandewie, and V. P. Sivonen, “Auditory Perception in Reverberant Sound Fields and
Effects of Prior Listening Exposure,” Principles and Applications of Spatial Hearing, (2011), 24–34.
See also Mads Walther-Hansen, Sound Events, Spatiality and Diegesis – The Creation of Sonic Narratives
in Music Productions. p. 29-46 in this issue.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
108
Justin Christensen
frequent and severe artifacts in the AF4 electrode, so their results were also removed
prior to analysis. Correspondingly, the final collected results were drawn from 6 male
and 4 female participants.
Apparatus
This experiment was conducted using a stereo setup with KH120A speakers and SVS
sub. For creating a haptic response, a Clark synthesis TST239 tactile transducer was
used, and it was connected to the stereo through a stereo to mono bridge (with applied resistors) and then passed through its own 200 watt amplifier. The EEG device
used was a 14 channel emotiv epoc headset with software development kit, consistent
with the international 10-20 EEG electrode placement,50 referenced with a left-hand
CMS (Common Mode Sense active electrode51).
The music used for the experiment was: 1. a slightly sped-up and transposed-up excerpt of Hoppipolla from Sigur Ros (happy), 2. an excerpt from Arvo Pärt’s, Cantus in memoriam Benjamin Britten (sad), 3. A slowed-down excerpt of Imogen Heap’s hide and seek,
stretched to 8 times the original duration by using PaulStretch52 (calm), 4. Streets of Pripyat from the S.T.A.L.K.E.R.: Shadow of Chernobyl soundtrack (fear). These musical examples
were collected from a number of online sources53 that rated them as being exemplars
of their respective emotional categories. I initially chose a few exemplars from each of
these 4 emotional categories, and further selected from these based on their psychoacoustic features, based on appraisal theories of emotion.54 Furthermore, none of the musical
selections contained either sung or spoken text. Both the Hoppipolla and hide and seek excerpts were taken from karaoke versions, with the solo voice stripped from the track.
Distant versions of the musical excerpts were prepared with distance-related headrelated transfer functions in Ircam Spat55 to convey a sense of distance. I processed the
distant sources through Spat at a radiated distance of 20.52 meters from the virtual listener, with a reduced room presence (seen in figure 1 below) and a reduced room live50
51
52
53
54
55
Richard W Homan, John Herman, and Phillip Purdy, “Cerebral Location of International 10–20 System Electrode Placement,” Electroencephalography and Clinical Neurophysiology 66, no. 4 (April 1987):
376–82, doi:10.1016/0013-4694(87)90206-9.
The CMS is a reference electrode which all other electrodes are measured against to help reduce
movement artifacts.
Computer application for extreme stretching of digital audio files.
Among the online lists were:
http://www.nme.com/photos/50-most-uplifting-songs-ever/279564 ;
http://www.npr.org/sections/deceptivecadence/2010/09/27/130157375/the-saddest-music-in-theworld-7-tunes-to-make-you-tear-up ;
http://www.thefourohfive.com/music/news/article/top-10-horror-video-game-music-themes ;
https://michelsabbagh.wordpress.com/2015/08/17/7-video-game-soundtracks-that-give-me-goosebumps/ ; Paulstretch calm ambience discussed in B Sturm, “Generation of ambient background textures from a given waveform,” (master thesis, Aalborg University Copenhagen, 2011).
David I. Leitman et al., “Getting the Cue: Sensory Contributions to Auditory Emotion Recognition
Impairments in Schizophrenia,” Schizophrenia Bulletin 36, no. 3 (May 2010), 545–56, doi:10.1093/
schbul/sbn115.
Jean-Marc Jot and Olivier Warusfel, “Spat: A Spatial Processor for Musicians and Sound Engineers,”
In CIARM: International Conference on Acoustics and Musical Research (1995).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
109
ness. These values were lowered, as I felt this gave the distant sounds a slightly more
realistic feel. All other parameters were left at their standard preset values. Figure 1 below shows the configuration for perceptual factors in Spat, used for the creation of the
virtual distant listening situation. All of these parameters remained unchanged for the
near-field setting, except for the distance setting (under Radiation), which was changed
from 20 meters to 1 meter. Spat automatically alters other parameters (such as source
presence) to match the change of distance. After the excerpts were prepared with Spat,
they were normalized to -18dB LUFS (Loudness Units relative to Full Scale56). Perceived loudness has shown to have a strong effect on evoked emotional responses to
music,57 so the aspect of loudness as a contributing factor to distance was removed
to minimize the perception of loudness itself as a possible extraneous experience
variable. This changing of the virtual distance of the recording is largely considered a
change to the ‘soundstage’ of the recording, altering the depth and richness of the original recording to give a new soundstage. For this experiment, the original soundstage
was not removed from the recording prior to adding new distance cues with Spat.
Figure 1: Perceptual configuration of Spat for virtually distant listening situation.
Design
The experiment contained 3 distance conditions (far, near, near with haptic stimulation), 4 emotion conditions (happy, calm, sad, fear), and each musical example was
presented in each condition to each listener as a repeated measures test. The order of the
musical excerpts was randomized according to a Latin square to minimize order effect.
56
57
Esben Skovenborg, “Loudness Range (LRA) – Design and Evaluation,” Audio Engineering Society Convention 132 (April 2012): 1–12.
Laura-Lee Balkwill and William Forde Thompson, “A Cross-Cultural Investigation of the Perception
of Emotion in Music: Psychophysical and Cultural Cues,” Music Perception: An Interdisciplinary Journal
17, no. 1 (October 1999), 43–64, doi:10.2307/40285811.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
110
Justin Christensen
Each listener’s alpha lateralization was taken against his or her own baseline, which
was calculated as an average from a combination of the EEG recording period prior to
testing, as well as an average of the alpha lateralization during the testing period. This
way, I could measure the emotional valence as motivations negatively away from or
forwards from their baseline.
Procedure
EEG data was recorded in OpenViBE58 at a rate of 128 Hz. These EEG recordings
were supported with subject feedback responses. The EEG recordings were then analysed through the use of EEGlab59 and ERPlab60 in Matlab. Data was passed through
a 2-way least squares FIR bandpass filter, using a high-pass filter of 180 poles, attenuated to -6 dB at 2.5 Hz. The EEG signals were then segmented into six 6 second
epochs (768 samples at 128 Hz) per musical excerpt. These epochs were chosen from
the last 45 seconds of each musical excerpt (which are roughly 60 seconds in duration for each), to allow a training period for the listeners to better reach the desired
evoked emotional state. Much of the work of preparing an EEG signal for analysis are
methods of noise reduction or artifact removal, or methods for increasing the signalto-noise ratio. Here, the high-pass FIR reduces noise from eye-movement and muscle
activity, while the epoch segmentation increases the signal-to-noise ratio.
Artifacts were then corrected with a combination of AMICA and wavelet noise reduction. AMICA makes an independent component analysis (ICA61) decomposition
of input data, running multiple ICA models for best fit. Once AMICA separated the
EEG signals into individual components, I ran a stationary wavelet transform (SWT62)
wavelet noise reduction on the EEG components to remove artifacts. This SWT noise
reduction used a HAAR wavelet wave shape, 5 levels, and was thresholded with a Rigorous SURE algorithm (an adaptive thresholding based on Stein’s Unbiased Risk Estimator). The HAAR wavelet shape gives a good compact representation of certain types
of artifacts (e.g. eye blinks) and the adaptive thresholding relies on the Gaussian distribution of EEG signals. As a result, the SWT keeps the artifact and removes the EEG
signal. A clean EEG residual signal is recovered by subtracting the artifact signal from
the original. I implemented my wavelet noise reduction mostly following the method58
59
60
61
62
Yann Renard et al., “OpenViBE: An Open-Source Software Platform to Design, Test, and Use Brain–
Computer Interfaces in Real and Virtual Environments,” Presence: Teleoperators and Virtual Environments 19, no. 1 (February 2010), 35–53, doi:10.1162/pres.19.1.35.
Arnaud Delorme and Scott Makeig, “EEGLAB: An Open Source Toolbox for Analysis of Single-Trial
EEG Dynamics Including Independent Component Analysis,” Journal of Neuroscience Methods
134, no. 1 (March 2004), 9–21, doi:10.1016/j.jneumeth.2003.10.009.
Javier Lopez-Calderon and Steven J. Luck, “ERPLAB: An Open-Source Toolbox for the Analysis of EventRelated Potentials,” Frontiers in Human Neuroscience 8 (April 2014), doi:10.3389/fnhum.2014.00213.
ICA attempts to separate independent signals from the mixed signals in each electrode through
blind source separation. As seen in: Pierre Comon, “Independent Component Analysis, A New Concept?,” Signal Processing, Higher Order Statistics, 36, no. 3 (April 1994), 287–314, doi:10.1016/01651684(94)90029-9.
G.P. Nason and B.W. Silverman, “The Stationary Wavelet Transform and Some Statistical Applications,” in Wavelets and Statistics, ed. Anestis Antoniadis and Georges Oppenheim (Springer New
York, 1995), 281–99, doi:10.1007/978-1-4612-2544-7_17.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
111
ology of Akhtar et al.63 Finally, a liberal threshold of (-200 and 200 µVolt) was used to
discard extreme values. This combination of SWT and AMICA is a very powerful artifact removal tool, especially for the removal of eye blinks and movement artifacts.
The Hemispheric asymmetry scores were then computed by subtracting the PSD
(power spectrum density) of channels in the left hemisphere from their homologs in the
right hemisphere.64 For this, I used an implementation of Welch’s method to compute
the PSD.65 I based my measurements from the combination of log-transformed results
from the F3/4, F7/8, and AF3/AF4 EEG electrodes. These results were then saved into
bins and summed for all of the participants. Following this, the results were indexed by
their categories of virtual distance and evoked basic emotion to allow for statistical comparisons of variance against the results from other categories of distance and emotion.
Results
A two-way repeated measures ANOVA was run to determine the effect of different virtual distances for music evoking different emotions (happy, sad, calm, fear). All of
the music listening situations were normally distributed (p > .05), as assessed by Shapiro-Wilk’s test of normality on the studentized residuals. The data also contained no
outliers, as assessed by no studentized residuals greater than ± 3 standard deviations.
Three data points fall outside of ± 2 standard deviations, and they can be seen in the
boxplot below in Figure 2. The Shapiro-Wilk’s test is run to ensure that the results
come from a normally distributed population, and that noise is not a major contributing factor that skews the final results.
Figure 2: Boxplot of mean alpha lateralization with their standard deviations
across all 12 conditions. Distant conditions are repeated on right to provide
symmetry
63
64
65
Akhtar, Muhammad Tahir, Wataru Mitsuhashi, and Christopher J. James. “Employing spatially constrained ICA and wavelet denoising, for automatic removal of artifacts from multichannel EEG
data.” Signal Processing 92, no. 2 (2012): 401-416.
Cheryl L. Dickter and Paul D. Kieffaber, EEG Methods for the Psychological Sciences (SAGE, 2013).
“…Welch’s method improves the reliability of the PSD by dividing a discrete time signal into a subset of shorter segments, applying the DFT to each of the segments in the subset and then averaging
the results .” in Ibid, p. 106.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
112
Justin Christensen
When not using haptic stimulation in the nearfield listening situations, there was no statistically significant (p < .05) two-way interaction between distance and emotional quality of the music, F= 0.742, p = .537. However, there was a significant variance between
the emotional conditions F=3.673, p=.026. Mauchly’s test of sphericity indicated that
the assumption of sphericity was met for the interactions of the emotional conditions,
χ2(2) = 0.442, p = .363. This suggests that changes to distance cues do not affect the presented basic emotion categories (happy, sad, fearful, calm) differently from one another.
In figure 3 below, one can see that the individual emotional conditions move nearly in parallel with one another (with the exception of the calm condition). When this
occurs, even when there is a reasonable distance between the conditions, the interaction effect is not likely to be statistically significant. Significant interactions between
conditions are much more likely to be seen when the conditions take different courses.
Figure 3: Comparison of alpha lateralization for musical examples of different emotional qualities in virtually distant and nearfield listening situations.
In contrast to the results without haptic stimulation, the inclusion of haptic stimulation in the nearfield condition is consistent with the hypothesis that changes to distance cues do affect the presented basic emotion categories (happy, sad, fearful, calm)
differently. As seen in figure 4 below, when using haptic stimulation in the nearfield
listening situation, there was a statistically significant (p < .05) two-way interaction between changes of distance and emotional quality, F = 3.994, p = .019. Mauchly’s test of
sphericity indicated that the assumption of sphericity was met for the two-way interaction, χ2(2) = 0.626, p =.681.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
113
Figure 4: Comparison of alpha lateralization for musical examples of different emotional qualities in virtually distant and nearfield haptic listening situations.
Figure 5: Comparison of all listening conditions together
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
114
Justin Christensen
With the inclusion of haptic stimulation in the nearfield condition, alpha lateralization
did not significantly (p < .05) change over time for the happy condition between distant and near-field conditions, F= 0.182, p = .052. Neither did alpha lateralization significantly change over time for the calm condition between distant and near-field conditions, F= 2.406, p = .159 or for the sad trial, F= 0.075, p = .791. However, alpha lateralization did significantly change over time in the fear trial, F= 8.447, p = .02. The differences between the emotions at a virtually distant listening position were not significant,
F=2.768, p=.064, nor at a virtually near-field listening position, F=2.225, p=.111. These
results suggest that with a small sample size it is difficult to determine the impact of
changes of distance due to individual variance, especially for differences of small statistical power. The results also imply that the statistical power (F) between distant and nearfield conditions can vary greatly depending upon the basic emotional category presented. The statistical power for fear was strongest at F=8.447, followed by calm (F=2.406),
happiness (F=0.182), and finally sadness (F=0.075). The strong statistical power for the
fear listening situation was a large contributing factor in achieving a significant result.
There exists “…many methodological problems for measuring physiological indices (e.g., individual differences, time-dependent nature of measures, difficulties in
providing effective stimuli).”66 This does not even take into account non-brain related
artifacts, such as electromyogram (EMG) and electrooculogram (EOG) that very often
exist in an EEG signal and can make analysis difficult. Given the error variance that
these methodologies produce, it is arguably more problematic for the no-specificity
hypothesis.67 With the significance exhibited in the larger condition interactions, and
the significance of the fear trial, in addition to the various very nearly significant conditions (p<.05), this suggests further work in this area with larger sample sizes and
better controlled conditions are warranted.
Discussion
If emotional responses to music at a nearfield are to be considered to be more strongly valenced than virtually-distant music sources, then the zero-point baseline does not
seem to be a tipping point that separates positively-valenced experiences from negatively-valenced experiences. This is especially the case when comparing the distant
condition with the nearfield including haptic stimulation condition. If one only examines the standard deviations away from the mean, then both individual results and
combined results show that both of the near conditions diverge further away from
their means than the distant condition. The average standard deviation for the individual evoked responses at a distance is .2865, while nearfield is .45175, and nearfield haptic is .38225. The average standard deviation from the mean for combined
results at a distance is .537, while nearfield is .742, and nearfield haptic is .611. This is
66
67
Patrik N. Juslin and Petri Laukka, “Communication of Emotions in Vocal Expression and Music
Performance: Different Channels, Same Code?,” Psychological Bulletin 129, no. 5 (2003), 770–814,
doi:10.1037/0033-2909.129.5.770, p. 772.
Ibid.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
115
supported when looking at the graphs (see figures 3 and 4), which show an opening
up between the distant (left) and near (right) conditions. Making post-hoc speculation, there is the possibility that the zero axis baseline for valence is closer to -.5. This
would present results consistent with the hypothesis in that emotional valence for virtual sources in the nearfield are more intense than emotional valence for sources from
a distance.68 This conjecture suggests possibilities for more detailed future work.
One clear finding is that the results are not consistent with their status as purported exemplars for evoking certain discrete emotional categories as described on online lists. These results, with their discrepancies from the expected evoked valences,
are also supported by many of the subjective responses of the post-exam debriefing.
A number of the subjects described their experiences of the music in terms of how it
would be placed in a movie. The fear music was considered by many to ‘anticipate’
something big or bad to happen. The happy music was often heard as ‘triumphant’,
or ‘trying to be triumphant’. The sad excerpt was described as ‘cinematic grief’, while
the calm excerpt evoked feelings of ‘floating’ or ‘religion’. These mismatched results are
also supported by earlier EEG music research, where the perceived emotions did not
fully match evoked emotional responses.69
In contrast with these discrepancies, the results of the nearfield haptic condition
match closely with the expected evoked emotional valence for their respective emotional categories. Does this support the idea that multimodal signals are more trustworthy? These results, while very speculative, suggest future work in this area is warranted. However, before these results can be considered more than speculation, they
need to be tested against more clearly quantifiable measures of perceived emotion.
In this study, I used qualitative research methods, discussing descriptions of perceived
emotions in a post-exam debriefing. This encourages highly individualized answers,
which does not easily allow for ranking or quantifiable comparisons. For this reason,
future work should include either a 23-item self-report scale immediately following
each excerpt, such as in Stephens et al.,70 or as continuous ratings on a computer application with a 2-dimensional emotional matrix, such as in Egerman et al.71 One
of the reasons for my hesitation in advocating for this theory is that there is another
possible interpretation for these results. The spectral envelope of the fear excerpt produced the strongest response from the tactile transducer, so the aberrant result in the
nearfield haptic condition for the fear excerpt could likely result from perceived overstimulation from a novel source. None of the subjects had previously experienced lis68
69
70
71
In support of this post hoc theory, Altenmüller et al exhibits a somewhat similar skew in their results
especially from their male participants. Seen in: Altenmüller et al., “Hits to the Left, Flops to the Right.”
Schmidt and Trainor, “Frontal Brain Electrical Activity (EEG) Distinguishes Valence and Intensity of
Musical Emotions.”
Chad L. Stephens, Israel C. Christie, and Bruce H. Friedman, “Autonomic Specificity of Basic Emotions: Evidence from Pattern Classification and Cluster Analysis,” Biological Psychology, The biopsychology of emotion: Current theoretical and empirical perspectives, 84, no. 3 (July 2010): 463–73,
doi:10.1016/j.biopsycho.2010.03.014.
Hauke Egermann et al., “Probabilistic Models of Expectation Violation Predict Psychophysiological
Emotional Responses to Live Concert Music,” Cognitive, Affective, & Behavioral Neuroscience 13, no. 3
(2013): 533–53, doi:10.3758/s13415-013-0161-y.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
116
Justin Christensen
tening to music with a tactile transducer prior to testing, so it is possible that above a
certain threshold the haptic stimulation would be felt as weird or overpowering and
something to move away from.
Since the original soundstages were not stripped from the recordings, another attractive interpretation of the results without haptic stimulation is also possible. The
calm excerpt was greatly stretched prior to processing with Spat so its original soundstage would have also been distorted for the nearfield listening condition. All of the
other excerpts largely kept their original soundstages, so these results could instead
indicate more positively valenced results for the original soundstage when compared
to a more distant version of itself.72 In future, the soundstage of the original should be
stripped and replaced with a virtual nearfield (instead of layering the new soundstage
over the original) to rule out this possibility.
Conclusions
The evoked responses for the nearfield conditions diverge further away from the mean
than in the distant condition, which suggests a correlation between the virtual distance and the strength of valence experienced. Another interpretation results from
comparing the original ‘intended’ soundstage with a more distant version, where the
intended soundstage is experienced as having a more positive valence than its more
distant version for most conditions. Furthermore, the multimodal audio haptic version most closely matches the intended evoked emotion. This could suggest an increased signal trustworthiness when using a multimodal signal. Further work will be
needed with larger sample sizes and more controlled soundstages to adequately resolve the possible implications of the results. Other aspects that would be good to
explore in future work are: an inclusion of multiple virtual distances, the addition of
haptic stimulation to a distant condition to attempt to destabilize the ‘trustworthiness’
of the distance perceived, independent variations to the early and late delay features of
the reverb, and a possible inclusion of a more external environment that might better
conform with animal ethological approaches.
72
The vast majority of the subjects did not recognize most of the excerpts. From 10 subjects, only 2
subjects recognized the happy excerpt, 2 recognized the sad excerpt, 1 recognized the sad excerpt,
and none recognized the fear excerpt. As a result, I would be very reluctant to suggest that these responses resulted from connections to memories of past listening experiences.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Virtual Distance and Soundstage, and their Impacts on Experienced Emotional Valence
117
Abstracts
Research from animal ethology and affective neuroscience suggest that a listener’s perceived distance from a signal source can alter their experienced emotional valence of
the music. Furthermore, appraisal theories of emotion suggest that emotionally valenced responses will diverge according to the type of emotion presented. For these
exploratory investigations, subjects listen to selected musical excerpts on speakers in
combination with a tactile transducer attached to their chair. The listening sessions
are recorded on EEG supported by subject feedback responses. My hypothesis is that
musical stimuli should cause stronger valenced responses in the nearfield than at a
distance. Thus, music experienced as being negatively valenced at a distance should be
more negatively valenced in nearfield, and music that is experienced as having a positive valence at a distance should be more positively valenced in nearfield. The results
are largely consistent with this hypothesis, but can also be found to be impacted by
the original soundstage of the musical excerpts presented.
Forskning i dyrenes etologi og neurovidenskab peger på at lytteres opfattelse af afstand til en lydkilde har en direkte indflydelse på den affektive valens musikken fremkalder. Herudover peger “appraisal” teori på, at den følelsesmæssige respons vil variere
alt efter hvilken type emotioner lytteren præsenteres for. I denne undersøgelse blev
en række testpersoner præsenteret for musikeksempler gennem højtalere, mens de
var placeret i en stol påmonteret en taktil transducer. Lyttesessionerne er optaget med
EEG, og der er yderligere indsamlet mundtlig feedback fra testpersonerne. Min hypotese er, at musikstimuli vil have en større valens når musikken høres tæt på, end når
den høres på afstand. Således vil musik med en oplevet negativ valens, når den høres
på afstand, have en en endnu større negativ valens, når den er hørt tæt på. Ligeledes
vil musik med en oplevet positiv valens, når den høres på afstand, have en endnu større positiv valens, når den høres tæt på. Resultaterne af undersøgelsen understøtter i vid
udstrækning hypotesen, men der bør samtidig tages forbehold for at resultaterne kan
være influeret af den originale lydscene i de valgte musikklip.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
MARTIN KNAKKERGAARD
Numeriske tåger
eller hvad blev der i grunden af de tre andre causa?
Med digitalteknologien har Jacques Elluls maksime, at “Fremskridt er som at tælle; der
er ingen grund til at stoppe ved et bestemt tal, et bestemt teknologisk trin, man kan
altid tilføje et nummer”,1 fået en næsten paradoksal dobbeltbetydning: Vi er med digitaliseringen netop stoppet, ved 0 og 1, og med dette talpar omtyder vi alt, hvad der
lader sig omtyde, til tabeller i en ustoppelig bevægelse, som ser ud til at efterlade det
omtydede til glemsel og nostalgi.
Afhængig af hvordan vi anskuer disse tabeller, i det rene Aristoteliske perspektiv
eller i Heideggers mere spekulative optik, ser det ud til, at det enten er causa finalis
eller causa efficiens, som tabellerne fremstiller for os, men holder det og hvad blev der
så af causa materialis og causa formalis?
Og at producere betyder netop at stille frem, at fremstille, dette vægter ifølge sagens
natur det ‘stillads’, der fremtræder gennem udøvelsen af den teknik eller teknologi,
den ‘montage’, gennem hvilket noget stilles frem, får form, materialiseres. Hvis imidlertid dette stillads er uadskilleligt fra det fremstillede synes accenten entydigt at falde
på aesthesis, den måde det fremstår på; poiesis og mimesis forbliver skjulte.
Med den digitale musikproduktion som case søger dette paper at kaste lys over
denne hermetik, hvor selv simulakret simuleres og ethvert materiales plasticitet og forventede natur formes med de samme midler, som det selv består af, og på de samme
vilkår, som det selv gør.
Indledning
Det er velkendt, at nutidens musikværker for en meget stor dels vedkommende bliver til i samspil med en eller anden form for DAW.2 Det kan gælde uanset om det er
musik, der formes og afsluttes i indspilningsstudiet, som det længe har været tilfældet
for det meste af populærmusikken, eller om musik, der komponeres helt eller delvist
af måske blot en enkelt komponist eller performer og som i princippet når en første
endelig form i studerekammeret, der efterfølgende fx kan fortolkes live. Er der i det
sidste tilfælde tale om populærmusik vil komponisten i dag ofte blive refereret som
en producer.
1
2
Jacques Ellul, The Technological System (USA: The Continuum Publishing Corporation, 1980), 222f.
Min oversættelse.
Digital Audio Workstation. Se i øvrigt Toivo Burlin “Experimentell musikproduktionsvetenskap som
kulturforskande musikvetenskap: Ett produktionsperspektiv på musik” i dette særnummer side 8f..
DANISH MUSICOLOGY ONLINE SPECIAL EDITION, 2015
SOUND AND MUSIC PRODUCTION
•
ISSN 1904-237X
SPECIAL EDITION · 2015
120
Martin Knakkergaard
Det forhold, at DAWen ofte spiller en helt central rolle i fremstillingen og formningen af musikværket, giver umiddelbart anledning til at sammenligne komponistens
kreative arbejdsform og –vilkår med en billedhuggers eller skulptørs, idet komponisten i DAWen kan arbejde med sine musikalske og kunstneriske ideer og mål direkte i
materialet og ikke skal omkring transskriptioner i form af fx noder og partiturer. Selv
om der sagtens kan arbejdes på baggrund af skitser, der er fastholdt i notation eller
udviklet i samspil med et traditionelt musikinstrument, er der tale om en historisk
set enestående situation, hvor det musikalske artefakt sammensættes og tilrettes, ja,
i virkeligheden modelleres3 direkte som et klingende materiale, et på sin vis færdigt
produkt i stadig forandring.
Lyd, der i disse rammer nu fremstår som et plastisk medium, er dermed blevet til
komponistens stoflige materiale på samme måde som fx ler og fedtsten er det for billedhuggeren og designeren. Men den palet, som komponisten har at arbejde med, omfatter mange flere værktøjer og teknikker og meget mere vidtrækkende muligheder end
dem, billedhuggeren har adgang til; det forhold er givet med selve det digitale format.
Det digitale materiale kan, idet det netop er digitalt, bearbejdes, processeres, formes og forandres i en grad som langt overskrider de grænser og muligheder, der gælder for ethvert arbejde med et fysisk materiale, idet man i det digitale domæne dybest
set slet ikke er bundet af den virtuelle materialitet, som et digitalt format i almindelighed fremtræder som. Hvor den fysiske lyd, hvis materialitet – ontologi – strengt taget
blot er trykforskelle i luft og som i traditionelle mekaniske medier, som vinylplader og
magnetbånd, optræder lagret som henholdsvis fysiske aftryk og elektriske svingninger,
er bundet til fysikkens love og begrænsninger – en violinstreng kan fx ikke opspændes
ubegrænset, den vil sprænges, når spændingen overskrider en vis tærskel – er den digitale ‘lyd’ ikke bundet af andet end de muligheder og begrænsninger, som er givet med
det binære format – den digitale emulering af violinstregen kan spændes ligeså hårdt
det skal være uden at sprænge, den kan ikke sprænge, men en sprængning kan optræde simuleret, hvis programmøren har indarbejdet en sådan emulering i koden.
Samtidig kan det digitale læses – aflæses eller udlæses – efter forgodtbefindende,
idet lyd og musik ikke længere reproduceres, “… men rekonstrueres. Eller konstrueres.
For vi kan læse de binære tabeller som ‘hvad-som-helst’ uden at skele til, hvordan og
af hvad de er dannet.”4
Disse forhold drager det digitale materiales substantialitet i tvivl; for hvad er det
for en overjordisk tilstandsform, det optræder i, når det uden videre kan unddrage sig
fysikkens love og overskride dens grænser og hvad får det af betydning, når vi i vores
kreative stræben efter bestemte udtryk, former og æstetiske mål, arbejder i materialer
som strengt taget ikke er der, ikke engang forbigående – som musikken måske er.5
3
4
5
Martin Knakkergaard, “Modellering” in Flyvende Licens: om musik og computer (Aalborg: Undervisningsministeriet og Institut for musik og musikterapi, 1996).
Martin Knakkergaard, “Silent Music” (Kapitelskitse i anledning af DSfMs symposium Musik, lyd og
digitalisering, 24. april 2010).
Martin Knakkergaard, “The Music That’s Not There” in Handbook of Virtuality, ed. Mark Grimshaw
(USA: Oxford University Press, 2013).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Numeriske tåger
121
Nicholas Negropontes simple beskrivelse af forskellen mellem det analoge og det
digitale som en overgang fra atomer til bits er velkendt og den er i sin nærmest bagatelliserende troskyldighed egentlig ganske forjættende.6 Men den dækker jo netop
over, at ethvert kendetegn og potentiale og enhver begrænsning og dynamik, der er
knyttet til ‘tingene’ i deres atomare konstitution, uundgåeligt vil bortfalde i processen, fordi disse er givet med tingenes stoflighed, og dermed med deres ‘væren’ i verden. Det digitales mindsteenhed, bit’en, er ikke i verden i egentlig forstand. Bits kan
kun optræde i en eller anden form for repræsentation: som huller i et hulkort, høj
eller lav spænding, nuller eller ettere osv., og for at lette læsning og sikre overskuelighed kan de fx repræsenteres hexadecimalt og ordnes i enheder som Bytes, Words osv.,
men i sidste ende er de blot tabeller, der udelukkende består af den simplest tænkelige
forskel, noget eller intet.
For musikkens verden er det derfor nærmere den abstrakte repræsentation, som
igennem århundreder har domineret kompositionsarbejde inden for i hvert fald den
kunstmusikalske tradition, som det musikalske materiale i det digitale domæne kan
sammenlignes med. For materialet er alligevel ikke materiale i egentlig forstand, det
er ligesom notationen underkastet andre begrænsninger, end dem, der gælder i den
fysiske verden.
Men det digitale domænes forening af den abstrakte repræsentation og den konkrete manifestation udløser spørgsmål om proveniens og proces, som aldrig tidligere
har kunnet stilles meningsfuldt, for hvad er det i grunden for en materialitet, vi har at
gøre med, og hvordan optræder den i og hvordan virker den tilbage på processen.
Hertil kommer, at det tekniske – eller teknologiske – element i musik er og har
altid været evident. Det er svært at forestille sig en musik, eller musikalsk kommunikation, som ikke er baseret på og også helt afhængig af enheder, systemer, metoder,
instrumenter osv. Og disse elementer indvirker alle på formningen og realiseringen af
det musikalske artefakt. Med digitalteknologiens entré får disse en helt ny rolle, idet
de hermed optræder som repræsentationer og ikke er til stede i egentlig forstand. Set
i det lys bliver det pludseligt sværere at afgøre, hvordan disse elementer indvirker på
den musikalske formning, end det hidtil har været. Der foreligger således en situation,
hvor det i stigende grad er vanskeligt at gennemskue, hvad der indvirker på eller forårsager hvad i fremstillingen af nutidens musikalske artefakter og da alt i disse artefakter
– også deres ‘konkrete’ væren – i princippet træffes inden for den digitale teknologis
domæne, forekommer det nærliggende at forsøge at afdække indvirknings- og årsagssammenhænge ved at tage afsæt i Heideggers bestræbelse på at ‘forstå’ teknik.
Causa
I foredraget Spørgsmålet om teknikken,7 hvis sigte primært er en bestemmelse af teknikkens væsen (dasein) og en afklarende tolkning af det, vi i dag samlende ville kalde
6
7
Nicholas Negroponte, Being Digital (Storbritannien: Hodder & Stoughton, 1995), 11f.
Martin Heidegger, Spørgsmålet om teknikken og andre skrifter (København: Samleren, 1999) (original:
Martin Heidegger, Die Frage nach der Technik, Stuttgart: Verlag Günther Neske, 1954).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
122
Martin Knakkergaard
‘teknologien’, tager Heidegger afsæt i Aristoteles’ firledede årsagsbegreb,8 dvs. fremstillingen af, hvad der medvirker til, at et givet forhold, et resultat eller produkt, falder
ud – og fremtræder for os – på den måde, det gør. At spørge til medvirkende årsager
– man kunne også sige faktorer – giver som allerede bemærket mening, når det drejer
sig om både at skabe og at frembringe musik, al den stund, at også musikalske artefakter fremkommer som følge af flere samvirkende forhold. Spørgsmålet er endvidere –
og det er det centrale her – om digitalteknologien og dens brug medfører forskydninger i det musikalske arbejde og i givet fald på hvilke måder og i hvilke omfang.
Aristoteles’ årsagsbegreb kan ses forklaret som en “opstilling af fire forskellige synsvinkler, hvorunder noget kan tjene til at forklare noget andet” og det er, som det videre hedder, “bredere end det moderne årsagsbegreb, men til gengæld tættere på dagligsprogets angivelser af “årsager” og “grunde”.”9 Aristoteles opregner således stoffet (den
materielle årsag), ophavsmanden (den virkende årsag), formen (den formale årsag) og
formålet (den finale årsag), som de fire årsager. Idet Heidegger går i rette med opfattelsen af Aristoteles’ doktrin som et spørgsmål om ‘kausalitet’ og i stedet erstatter dette
med ‘skyld’ eller ‘skyldighed’, diskuterer han de fire causa gennem iagttagelsen af et
sølvbæger som eksempel. Han vælger, ligesom det er almindeligt udbredt, alligevel at
anvende de latinske betegnelser trods afvisningen af, at det drejer sig om kausalitet, og
anfører, hvordan sølvets beskaffenhed, som causa materialis, er medbestemmende10
for bægeret, det samme er dets tilstræbte form, causa formalis, og sølvsmeden, der udvirker formen, causa efficiens, og endelig er hensigten med bægeret, dets anvendelse
og brug, causa finalis, medbestemmende for dets fremtrædelse.
Undersøges disse fire causa på klingende, musikalske artefakter, må den hørbare
lyd eller klang anskues som causa materialis, det, der formes og som dermed er medbestemmende som grundvilkår for det tildannede,11 mens tonesystem, stilistik, musikinstrument, produktions- og optageudstyr m.v. alle kan ses som medbestemmende for
artefaktets formelle og normative organisering og fremtrædelse, dvs. som causa formalis. Komponisten, musikeren m.fl. er, som causa efficiens, medbestemmende for
musikkens virkeliggørelse og fremførelse på samme måde som produceren, teknikeren m.fl. er det for det fonografiske produkt og endelig er genren, fremførelsens topos,
produktionens omsætning og anvendelse, osv. medbestemmende som causa finalis.
8
9
10
11
at Heideggers tekst egentlig tilstræber et opgør med den gængse forståelse af Aristoteles’ fire årsager
vil fremgå løbende.
Aristoteles’ opstilling af fire forskellige synsvinkler, hvorunder noget kan tjene til at forklare noget
andet, er bredere end det moderne årsagsbegreb, men til gengæld tættere på dagligsprogets angivelser af “årsager” og “grunde”. If. Aristoteles kan man henvise til stoffet (den materielle årsag, fx til
forklaring af et hus), til ophavsmanden (den virkende årsag: håndværkeren), til tingens form (den
formale årsag: husets struktur i håndværkerens bevidsthed) eller til dens formål (den finale årsag:
husets funktion), jf. fx Troels Engberg-Pedersen, “Aristoteles” in Den store danske (http://www.denstoredanske.dk) og Jan Riis Flor, “Årsag” in Den store danske, begge opslag tilgået 17.11.2105.
jeg anvender ‘medbestemmende’ i stedet for ‘medvirkende’, idet Heidegger påpeger, at Aristoteles i
virkeligheden ikke taler om årsagsforhold, men om skyldsforhold. Retteligen burde jeg således anvende begrebet ‘medskyldigt’, men dette har i dag en umiddelbart mere negativ konnotation end
‘medbestemmende’.
idet jeg her afstår fra at diskutere om musik primært skal opfattes som at klingende fænomen, et
spørgsmål som jeg tager op i en senere artikel.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Numeriske tåger
123
I og med at Heidegger henviser til skyld og skyldighed i stedet for ‘virken’ og ‘bevirken’, som Aristoteles’ årsagsmekanismer ofte beskrives som,12 bliver forholdet til et
spørgsmål om, hvad der er eller har skyld i noget (andet) og man kunne måske sige
har del eller andel i noget og ikke blot, hvad der har forårsaget dette. I virkeligheden
sigter Heidegger efter en forståelse, som med Platon indsætter ‘foranledning’ i stedet
for ‘årsag’ eller ‘skyld’, og han betoner derved det processuelle – det man kunne kalde
det virkende i det foranledigende, som i frembringelsen bliver nærværende.13
Hvor “Aristoteles anser formålsårsagen for den vigtigste, og bemærker, at den ofte
i sig omfatter både den formale og den bevirkende årsag [causa efficiens],”14 ændrer
Heidegger, i sin bestræbelse på at afdække teknikkens væren, ‘kæden’, således at den
får følgende forløb: causa materialis, causa formalis, causa finalis og causa efficiens.
Herved pointerer han, hvordan causa finalis – årsagsbestemtheden, anvendelsesmålet
– ikke længere medregnes, “Causa efficiens […] bliver mønster og afgørende bestemmelse for al kausalitet.”15 dvs. at menneskets behov for og brug af artefaktet ikke længere opfattes som del af kæden – eller som kædens mål.
Accenten er for Heidegger mao. sat på causa efficiens – på frembringelsen og
frembringeren – og ikke på hensigten med artefaktet. I stedet for artefaktets anvendelse er det selve processen, forstået som en fremdragen, der finder sted gennem udfoldelsen af teknik – techne, det handler om. Man kunne sige, at formålet med frembringelsen af artefaktet er frembringelsen som en lønnet aktivitet og ikke det frembragtes nytte eller relevans i verden. Det er den tekniske aktivitet, handlingen, der er
det egentlige formål, ikke handlingens frugt.16 Det er derfor ikke spørgsmålet om til
hvilken nytte og med hvilket formål, en handling finder sted, men i stedet til spørgsmålet om handlingen selv som forandringsfremkaldende aktivitet. Der er ikke langt
fra denne opfattelse til Marshal McLuhans pointering af, at “mediet er budskabet”,
en pointe, der blev fremført i Understanding Media, en halv snes år efter Heideggers
tekst,17 og som indebærer, at det ikke er mediets indhold, den særlige information eller det konkrete produkt, som frembringes i eller af mediet, der er det centrale; det er derimod mediet selv, og den forandring, det forårsager i den kontekst, det
introduceres i.
Heidegger fremhæver, at der mellem causa efficiens og causa finalis sker en frembringelse, hvor håndværkeren og kunstneren, gennem anvendelse af bestemte materialer og bestemte, tilstræbte former, frembringer noget, som i princippet er – eller før
har været – skjult. Det bringes således ud i det han benævner det uskjulte, det afdækkes: “Teknik er en af afdækningens måder. Teknikken driver sit væsen i det område,
hvor afdækning og uskjulthed, hvor aletheia, hvor sandhed sker.”18
12
13
14
15
16
17
18
Politikens filosofi leksikon, red. Paul Lübcke, 14. udg., “Aristoteles”.
Heidegger, Spørgsmålet om, 41.
Politikens, “Aristoteles”.
Heidegger Spørgsmålet om, 39.
For så vidt er dette på linje med Marx’ kapitallogiske pointe, at bytteværdien har ‘udkonkurreret’
brugsværdien i den moderne økonomiske ‘rationalitet’.
Marshall McLuhan, Understanding Media: The Extensions of Man, (New York: McGraw-Hill, 1964).
Heidegger, Spørgsmålet om, 43f.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
124
Martin Knakkergaard
Ligesom episteme er techne således en afdækken, men med det særlige træk, at
den afdækker det endnu ikke foreliggende, fx noget, der skabes, som et hus, en have,
et stykke musik. “Den afdækken, som hersker i den moderne teknik, er en frem- og
udfordren, som stiller naturen overfor det krav at levere energi, der kan gøres tilgængelig og oplagres som sådan.”19 Naturen bestilles således som det Heidegger kalder
‘bestand’ gennem teknikkens afdækning, hvor bestand forstås, som “[…] den måde,
hvorpå alt det er nærværende, som berøres af den udfordrende afdækken. Hvad der i
bestandens betydning står, står ikke længere over for os som genstand.”20
Med til bestanden hører imidlertid også mennesket, idet mennesket er bestandens
bestiller, som skovfogeden, der “[...] er bestilt af træforarbejdningsindustrien [...] for
bestilbarheden af cellulose, der på sin side er udfordret gennem behovet for papir
[...]” osv.21 Men mennesket er ikke blot bestand, idet det, når det “udøver teknikken,
tager […] del i bestillingen som en af afdækningens måder.”22 jf. frembringelse. Hermed vægtes bestillingens måde eller form, og dermed hvordan bestillingen, som en
måde at afdække et emne på, er med til at forme vores begreb om og forståelse af
det bestilte, af emnet. Dette er et meget afgørende punkt, hvor vi i mine øjne i virkeligheden taler om tilgængeliggørelsens implicitte ordning, og således om den formatering, som udøvelsen af enhver teknik – eller blot enhver handlen – indebærer for
det, der tilgængeliggøres.
Stilladset
Denne implicitte ordning i udøvelsen af teknik, der giver adgang – og anledning –
til en given bestand, opfatter Heidegger som et imperativ, der kan beskrives som et
stillads: “Det udfordrende krav, den tiltale, som forsamler mennesket til at bestille det sig-afdækkende som bestand, kalder vi nu das Ge-stell, stilladset.”23 Stilladset er for Heidegger imidlertid netop ikke i sig selv noget ‘teknisk’, som “stænger
og stempler og støttekonstruktioner”,24 denne manifeste teknik kalder Heidegger
‘montage’, mens stilladset derimod er “den afdækningsmåde, som hersker i den moderne tekniks væsen”,25 altså netop den ordning eller formatering, der finder sted
gennem udøvelsen af den tekniske aktivitet, og dermed den måde hvorpå noget gøres tilgængeligt. Det ‘udfordrede’ bliver dermed tilgængelig på en ganske bestemt
måde og i en ganske bestemt form, som er bestemt af den måde teknikken afdækker,
tilgængeliggør, naturen.
Overføres denne tankegang til musikkens verden eller ‘væren’ er toner, skalaer,
rytmer, harmonik, instrumenter, optageudstyr osv. således at ligne ved ‘montage’,
idet de er konkrete tekniske ‘entiteter’ eller bestanddele, der som sådan “modsvarer
19
20
21
22
23
24
25
Heidegger, Spørgsmålet om, 44.
Heidegger, Spørgsmålet om, 46.
Heidegger, Spørgsmålet om, 48.
ibid.
Heidegger, Spørgsmålet om, 49.
Heidegger, Spørgsmålet om, 50.
ibid.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Numeriske tåger
125
still-adsets udfordring, men aldrig selv udgør denne endsige bevirker den”;26 mens
stilladset svarer til de implicitte systemer eller måder, hvorigennem musikalske organiseringer gestaltes, og således “afdækker det virkelige som bestand”,27 som musikstykker. Det er fristende her at parallelisere til moderne neuropsykologiske forklaringsmodeller, der nok synes at have afdækket, at musikalske – og i det hele taget
akustiske – stimuli kan føre til høj hjerneaktivitet og udløsning af neurotransmittere og at dette er en grund til, at mennesket sætter stor pris på og dyrker musik,28
men det forklarer ikke, hvorfor det forholder sig sådan, at gennemstrukturerede trykforskelle i den æteriske materie mennesket befinder sig i, i den grad behager det.
Hvad er meningen? Det kunne måske netop signalere en form for sanset erkendelse af verden, en slags værens uskjulthed; selv om denne tanke ikke skal udvikles i
nærværende tekst, grænser den måske op til den form for indsigt, af-sløring, som
Heidegger stræbte efter.
Søger vi stilladset i den digitale tidsalder, konfronteres vi – som altid, når talen er
på det digitale – med det paradoks, at de redskaber og teknikker, montagens entiteter,
som vi anvender, i virkeligheden ikke er til stede som egentlige eller konkrete redskaber, som fx toner og potentiometre. De optræder alene som symboler og metaforer og
har som sådan ingen substans. Dette indebærer, at den afdækning, der finder sted, i
princippet slet ikke er nogen afdækning i hvert fald ikke i Heideggersk forstand, idet
den bestand, der tilsyneladende står frem, ikke er og ikke kan være nærværende. Der
findes intet nærvær – eller hervær – i det digitale og det digitale giver ikke adgang til
en afdækken af naturen eller virkeligheden, i hvad, der synes at fremtræde som sådan;
den digitale montage beskriver kun et bestemt stadie i det andre teknikker har kaldt
frem og kan ikke, i modsætning til disse, være direkte forbundet med virkeligheden. I
og med at det digitale er substansløst, er den bestand, det kan fremkalde, ingen udfordring, blot en simuleret – oprindelsesløs – reprise: “It is no longer a question of imitation, nor duplication, not even parody. It is a question of substituting the signs of the
real for the real …”29
For at belyse, hvordan denne, den digitale tids definitive simulering, virker ind på
musikalsk handlen og aktivitet, er det i det Heideggerske perspektiv nødvendigt at se på
den ‘frem-stillen’, den afdækken, som 1900-tallets elektriske apparatur har forårsaget.
26
27
28
29
ibid.
ibid. Bemærk at Heidegger gennem den uvante orddeling – “still-ads” (“ge-stell”) – ønsker at understrege, at noget stilles – er stillet – frem gennem be-stilling og frem-stilling, dvs. træder frem i kraft af
– brugen af – teknik eller montage.
se fx Line Gebauer, Morten Kringelbach og Peter Vuust, “Ever-changing Cycles of Musical Pleasure:
The role of anticipation and dopamine” in Psychomusicology, vol 22, nr. 2 (2012), 152-167. For en
mere generel indføring i vedr. Musik og hjerne se fx Lars Ole Bonde, Musik og menneske, (Danmark:
Samfundslitteratur, 2009), 51ff). Se i øvrigt Mark Grimshaw ’A brief argument for, and summary of,
the concept of Sonic Virtuality.’ i dette særnummer p. 97, 102 og 105.
Jean Baudrillard, Simulacra and Simulation, (USA: The University of Michigan Press, 1994), (fransk
original Simulacres et simulation, 1981).
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
126
Martin Knakkergaard
Elektrificeringen
Allerede de akustiske optageteknikker indebærer selvsagt en forskydning i den måde
teknikken sætter sig igennem på inden for det musikalske univers.30 Den, i Heideggersk forstand, afdækken, som det fonogrammatiske artefakt beskriver, fremstiller på
den ene side et bestandsmæssigt forhold, en uskjulthed, der forekommer skræmmende umiddelbar og nærværende, “[an] unimaginable closeness of sound technology
and awareness, a simulacrum of a feedback loop relaying sender and receiver”,31 mens
den på den anden side fremstår indhyllet i produktionens processers mange slørende
og uigennemskuelige lag. Grammofonen og også radiotransmissionen udfordrer koncertsalen og spillestedet og de dertil knyttede behov og forventninger. De er den moderne tids causa finalis og de forskydninger, som de beskriver i anvendelsesformålet
tæller sociale, kulturelle, musikalske og tekniske forandringer, der i sig selv virker tilbage på processen og fører til ændringer i causa efficens og causa formalis.32 Korrespondensen eller analogien mellem det klingende og det registrerede er dog kun så
stor som det aktuelle reproduktionsmedie, causa materialis, formår, og der sker egentlig blot en optimering af outputtet med henblik på at nå et i forhold til den klingende original, den oprindelige lydkilde, acceptabelt niveau; “What all acoustic engineers
were striving for was greater faithfulness to the source.”33
Med elektrificeringen, der tager fart i slutningen af 1920rne, undergår musikproduktionen imidlertid gennemgribende forandringer. Nu er det ikke i så høj grad musikken
i dens akustiske omsætning, fx i koncertsalen, der tilstræbes, men derimod den radiotransmitterede live-musik, som søges fastholdt fonografisk.34 Lydteknikerne bliver med
elektrificeringen opmærksomme på mulighederne for at medforme lyden og tænker i
stigende grad i ‘lydbilleder’, i det for det indre øre allerede hørte ikke-hørte. Henover
de næste årtier optimeres dynamik og støj/signal-forhold og der indføres forskellige
former for filtre, effekter og separationsteknikker alt sammen medvirkende til, at man
kunne begynde at tænke i musik, som blev skabt for mediet – et forhold, der yderligere
får et kraftig skub med indførelsen af båndoptager-teknologi efter Anden Verdenskrig. I
løbet af 1950erne når man inden for populærmusikken til en musikform, som er tænkt
som et lydproduceret artefakt, der næsten er blottet for referencer til performet musik,35
og tidens tekniks Heideggerske ‘montage’, danner et ‘stillads’, som afdækker det virkelige som ‘bestand’ i en form, der grundlæggende bryder med tidligere tiders ‘bestande’.
Allerede elspændingsbaseret – og egentligt også den mekaniske – optagelse fører til
en slags lydformning eller lydskulpering af causa materialis, som i sig selv beskriver en
30
31
32
33
34
35
se fx Michael Chanan, Repeated Takes. A Short History of Recording and its Effects on Music (London:
Verso, 1995), 23ff.
Friedrich Kittler, Gramophone, Film, Typewriter (USA: Stanford University Press, 1999), s. 37.
se fx Chanan, Repeated Takes, 59ff og Mark Katz, Capturing Sound. How technology has changed music.
(USA: University of Califonia Press, 2004), 31-36.
Chanan, Repeated Takes, 58.
ibid.
Noget tilsvarende havde længe været på spil inden for den kunstmusikalske tradition, jf. fx Busonis
Entwurf einer neuen Ästhetik der Tonkunst fra 1907.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Numeriske tåger
127
transformation, hvor materialets traditionelle autenticitet hele tiden udfordres, idet det
stilles frem i montager, som bryder med tidligere former og hvor causa efficiens, den
udvirkende, med skyldig henvisning til causa formalis, de konkrete teknikker og metoder, indtager en mere og mere central og langt hen dominerende rolle. Også inden for
datidens elektroakustiske strømninger betones causa materialis, causa formalis og causa
efficiens stærkt, det gælder fx for både musique concrète og Elektronische Musik.36
Digitaliseringen
Med digitaliseringen er vilkårene for ‘skyldighed’ ændret radikalt. Med causa finalis
som en slags dynamisk horisont fokuseres der intensivt på causa efficiens ud fra en
oplevet forestilling om, at man i historisk uset grad er herre over formningen, fremstillingen, i dens totalitet.
Spørgsmålet er imidlertid i hvilken grad man er herre over tingene og hvad det
egentlig er, man er er herre over. I det digitale værksted, dvs. i alle de processer, der
foregår i computeren, løses opgaver på en radikalt anden måde end uden for. Det er
kendetegnende, at værktøj og materiale i computeren – og helt generelt i det “computerized society”37 – er til stede på samme formel og også på samme niveau og at der
allerede af den grund nødvendigvis vil finde forskydninger sted mellem causa materialis og causa efficiens, der gør det vanskeligt at afgøre, hvem eller hvad, der udvirker
hvad. Det er således vigtigt at holde sig for øje, at “[d]er ikke er nogen invariant grænse mellem program og data”38 og at dette i praksis indebærer, at materialet, causa materialis, dybest set fremstår som elementer i eller af værktøjet. I og med at materialet i
dets digitaliserede form strengt taget må forstås som tabeller, der i det digitale domæne optræder som databaser,39 er der reelt tale om, at disse tabeller formes og bearbejdes af andre tabeller, nemlig i den udstrækning, at de algoritmer, som udgør programmernes funktioner, lader sig begribe som dynamiske tabeller. Af samme grund bliver
spørgsmålet om, hvad der bearbejder hvad, uklart, da ‘værktøjet’, den bearbejdende
tabel, i bearbejdningsøjeblikket – men også i det hele taget – er ét med de data, der
bearbejdes, med hensyn til lokalisering, format og tilstand.
Set i dette lys smelter værktøjet, causa efficiens, sammen med materialet gennem de
grænser, som det digitale format sætter, men det smelter samtidigt og yderligere sammen med formen i kraft af det grundvilkår som interface – tastatur, skærm, controller
osv. – sætter, et vilkår, som langt hen indebærer en ‘sekretarisering’ af enhver proces,
ethvert arbejde, der henlægges til det digitale domæne. Her gælder det endvidere, at
“[a]lgorithms and data structures have a symbiotic relationship. The more complex the
36
37
38
39
Se fx Martin Knakkergaard, IO – om musik, teknologi og musikteknologi (Danmark: Odense Universitetsforlag, 1994), 50f.
Jean Lyotard, The Postmodern Condition: A Report on Knowledge (England: Manchester University Press,
1984), 3.
Niels Finnemann, “Computeren – et medie for en ny skriftteknologisk revolution,” in Multimedier
Hypermedier Interaktive Medier, edited by Jens F. Jensen (Danmark: Aalborg Universitetsforlag, 1998), 48.
Idet database er defineret som en struktureret samling af data, jf. Lev Manovich, “Database as Symbolic Form” in Convergence, årg. 5, hft. 2, (01-06-1999), 81.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
128
Martin Knakkergaard
data structure of a computer program, the simpler the algorithm needs to be, and vice
versa”,40 hvilket indebærer, at de processer der iværksættes i sidste ende begrænses af
datamaterialets egen kompleksitet. Interfacet indrettes derfor så det ikke blot imødekommer, men er fuldstændig baseret på allerede eksisterende principper for og forestillinger om normative montager i Heideggers forstand. Således omfatter interfacet ‘sekretariseringer’ af tonesystem, transskriptionsformer, instrumentdesign osv., alle omtydet fra eksisterende praksisser, ligesom det byder sig til med repræsentationsformer,
der er hentet fra bl.a. fysiklaboratoriets og lydstudiets apparatur, som oscillatorer, filtre
og lignende moduler. Alt fremstilles i grafisk tilforladelige og fascinerende former selv
om ingen af disse normer og værktøjer har egentlig realitet i det digitale domæne.
Causa materialis og causa formalis er dermed i princippet forsvundet. De er indlejret i selve den digitale processering på en sådan måde, at deres tilstedeværelse, deres
skyldighed, udelukkende er repræsentativ og metaforisk. Om de er udtryk for en ekstern ydelse, som er digitaliseret, eller om de optræder som del af det digitales egen bestand, hvilket ikke er usædvanligt længere, er her for så vidt uden betydning: De er registreret i numeriske tabeller, der er hjemfaldne til enhver form for digital manipulation
og uden for rækkevidde af alt andet. I praksis er det derfor vanskeligt at skelne mellem
montage og stillads. Det er kun igennem den medlevende, men strengt taget også forledte iagttagelse, at artefaktet lader sig tyde med hensyn til medbestemmende faktorer.
Idet det digitale er substansløst finder de processer, som den foregiver at udføre,
heller ikke sted reelt. Ved at indskyde formaterende lag, der metaforisk fremstår som
velkendte praksisser og processer, som tidligere fx blev varetaget af elektriske apparater, opstår en forvirrende kausalitet, idet både causa materialis, causa formalis og ikke
mindst causa efficiens er ‘sublimeret’ og fortoner sig forenet i tabeller og numeriske
operationer i det digitale dyb.
Selve det teknologiske element, som medformende del af den kunstneriske proces, er fingeret, simuleret, idet digitalteknologien kan sættes til at medforme på enhver tænkelig måde uden på noget tidspunkt at være og blive en reel del af artefaktet.
Det musikalske artefakt står som sådan uden for det digitale, idet det oversættes eller
konverteres til lyd i den fysiske verden. Først her manifesterer montagen sig i stilladsafdækkende form, men er det virkeligheden, som bestand, stilladset afdækker? Næppe, for den mulige ‘afdækken’ synes forstenet i de metaforiske tabeller, som definerer
og afgrænser handlinger i det digitale domæne.
Afrunding
Vor tids – kulturelle – utryghed ved digitalitetens forflygtigelse af materiale og mangel
på realitet kan bl.a. aflæses i den store opmærksomhed omkring autenticitet generelt,
hvilket fx tydeligt kommer til udtryk blandt musikstuderende. Den har i stigende grad
gjort sig gældende siden slutningen af forrige århundrede og der tvivles tilsyneladende
på både intention og proveniens.
40
Manovich, “Database,” 84.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Numeriske tåger
129
Med Bolter og Grusin41 taler man om remediering, hvor ethvert medies indhold
anses som et andet medie; det er vildledende og strengt taget misvisende, når det drejer sig om digitalteknologi i og med, at der som vist reelt ikke er noget medie, hverken
som indhold eller form. At opfatte digital teknologi som medie er simpelthen uforsigtigt og forkert. Digitalteknologi er et fundamentalt vilkår, en væsensparallelitet uden
væren, et immaterielt parallelunivers, og den i Heideggers forstand virken, som dens
‘re-medier’ udøver, er konstrueret og finder udelukkende sted i algoritmer, scripts osv.
Når noget fremstår som re-mediering er det udelukkende udtryk for, at digitaliteten
møder dets modpart, den, der opfatter sig som digitalitetens bruger, på dennes banehalvdel – som den strengt taget har overtaget.
Når man genkalder sig, hvor næsten umulig en udfordring det tidligere var at fremstille båndsløjfer – loops – i det fysiske medie eller blot sømløse, støjfrie loop i de
tidlige digitale programmer og sammenligner disse med de loop, der på et øjeblik
kan skabes i dag, har man et meget godt billede på, hvor virkelighedsfjern – parallel
– teknologien er. At alt dertil er non-destruktivt understreger blot, hvordan intet længere er, hvad det synes at være. Det har nemlig ikke nogen egentlig eller endelig form
og materialet – causa materialis – er ganske enkelt ophørt med at være en egentlig
medformende faktor.
Der er således ikke tale om noget i Aristotelisk forstand kausalt forhold; kausaliteten er en tilstræbt og fast formet illusion – men alligevel vilkårlig. Det digitale er ikke
noget medie, idet det reelt ikke medierer, men rekonstruerer noget og som vilkår beskriver det en selvstændig, substansløs realitet, der er i verden som et parallelt univers.
Heller ikke Heideggers næsten metafysisk og forventningsfulde måde at spørge til
teknikken rækker langt, når det kommer til den digitale. Den frembringelse, som komponisten, musikeren og produceren stiller frem, finder sted som en kombinatorisk
proces inden for hvilken simuleringer af allerede eksisterende elementer kombineres
inden for faste systemiske vilkår. Disse vilkår er udtryk for bevidste valg, men er alligevel arbitrære og systemeksterne for det digitale domæne. Det digitales egne præmisser
er flygtige og ubestandige. Lydgenerering, syntese, loops, samples, osv. er alle mimetiske montager inden for et med digitaliteten hermetisk – og som sådan grundlæggende uopnåeligt – system. Det har intet virkeligt med virkeligheden at gøre og afdækker
intet andet end det allerede afdækkede.
Den elektriske musikteknologis metaforiske tilstedeværelse i digitaliteten beskriver
således en dobbelt teknologisk binding, idet principperne fra denne optræder som stillads, inden for hvilket det er muligt at foretage montager, som tilsyneladende finder
sted inden for et reelt ontologisk system.42 Men også stilladset er flygtigt og underkastet
det digitales troløse ikke-væren. Stilladsets grundlæggende konstitutive principper er
blot emuleringer, simulacra, som uden videre og omkostningsfrit kan erstattes af andre.
41
42
Jay Bolter & Richard Grusin, Remediation. Understanding New Media (Cambridge, MA, USA: MIT
Press, 1998).
Heideggers syn på den ontologiske mulighed som udelukkende til stede som et fænomenlogisk anliggende, jf. Dan Zahavis Indledning til Heidegger “Spørgsmålet om,” 13, har næppe nogensinde været bedre anskueliggjort end med det digitale vilkår.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
130
Martin Knakkergaard
Den systemiske determination, som den elektriske musik- og lydteknologi udgør
gennem dens tilstedeværelse som metaforisk repræsentation, er som sådan umiddelbart beskyttet af det digitales selvbærende – og selvlukkende og som sådan kompletterende – simulering. Men hvor længe og hvorfor?: Den er anakronistisk til stede som
“the steak that doesn’t exist” i filmen The Matrix43 hvor “Ignorance is bliss.”
På en underlig akavet måde genindfører digitalteknologien alligevel causa finalis
som mål eller formål og den reducerer i samme bevægelse de tre andre causa til skinmanøvrer. Der arbejdes nu entydigt på artefaktets finale stade – causa finalis – som det
æstetiske mål næsten blottet for poetisk vraggods. Det mimetiske er i den bladrende
praksis, en praksis som bl.a. tiljubles i new-speak-begrebet serendipity og som kendetegner vor tids causa efficiens, helt overladt til digitaliteten; det er den, der i samples,
loops, stereotypier osv. indlærer og gengiver de stilistiske relikter som er vor tids causa
materialis – og som på sin side synes dominerende for tidens æstetiske horisont.
Lydkunsten kan ligne en udvej, en vej omkring – som Hölderlins ord er det for
Heidegger,44 men den udstiller på ingen måde teknikkens bestand, som en udfordrende tilskikkelse, tværtimod. Den fremstår som et givet forhold, som ved hvis brug mennesket selv bliver en del af dens bestand, nu blot i støjende uvidenhed.
43
44
Amerikansk science-fiction film fra 1999 instrueret af A. og L. Wachowski og produceret af Joel Silver.
Heidegger, Spørgsmålet om, 58 og 64.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
Numeriske tåger
131
Abstracts
Artiklen beskæftiger sig med vilkårene for skabelse af musikalske artefakter i den digitale teknologis tidsalder. Den søger, bl.a. gennem en diskussion af Heideggers teknologiforståelse, især indsigt i de forandringer som fraværet af egentlig stoflighed, en
konsekvens af digitalteknologiens kategori og format, giver anledning til. I vilkårsafgrænsningen inddrages endvidere en overvejelse af, hvilken betydning nogle af de
teknologiske formater, der historisk går forud for digitalteknologiens indførelse, har på
forståelsen af det musikalske artefakt og dets fremtrædelse som klingende fænomen.
The article discusses the conditions underpinning the creation of musical artefacts
in the digital technology era. Through a discussion of, inter alia, Heidegger’s understanding of technology, it especially seeks an insight into the changes that the absence
of proper texture, a consequence of digital technology’s category and format, causes.
In the attempt to define these conditions, it incorporates considerations of the impact of some of the technological formats that historically precede the digital environment on the understanding of the musical artefact and its manifestation as a
sounding phenomenon.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015
132
Notes on the contributors
Toivo Burlin. Assistant Professor in Musicology. Department of Culture and Aesthetics.
Stockholm University.
Mads Walther-Hansen. Assistant Professor. Department of Communication and
Psychology, Music and Sound Knowledge Group (MaSK), Aalborg University.
Hållbus Totte Mattsson. Professor in Sound and Music Production. Högskolan Dalarna.
Jan-Olof Gullö. Guest Professor in Music Production. Royal College of Music.
Stockholm.
Ivan Höglund. Stud. mag. Royal College of Music. Stockholm.
Julia Jonas. Stud. mag. Royal College of Music. Stockholm.
Hans Lindetorp. M.F.A. Lecturer in Music and Media Production. Royal College of
Music. Stockholm.
Anton Näslund. Stud. mag. Royal College of Music. Stockholm.
Joakim Persson. Stud. mag. Royal College of Music. Stockholm.
Peter Schyborger. M.F.A. Lecturer in Music and Mediaproduction. Royal College of
Music. Stockholm.
Mark Grimshaw. Obel Professor. Department of Communication and Psychology,
Music and Sound Knowledge Group (MaSK), Aalborg University.
Justin Christensen. Postdoctoral researcher. Department of Communication and
Psychology, Music and Sound Knowledge Group (MaSK), Aalborg University.
Martin Knakkergaard. Associate Professor. Department of Communication and
Psychology, Music and Sound Knowledge Group (MaSK), Aalborg University.
SPECIAL EDITION – SOUND AND MUSIC PRODUCTION · 2015