{"id":3585,"date":"2017-03-02T17:15:17","date_gmt":"2017-03-02T17:15:17","guid":{"rendered":"https:\/\/arpjournal.com\/?p=3585"},"modified":"2017-03-02T17:15:17","modified_gmt":"2017-03-02T17:15:17","slug":"new-and-old-user-interface-metaphors-in-music-production","status":"publish","type":"post","link":"https:\/\/www.arpjournal.com\/asarpwp\/new-and-old-user-interface-metaphors-in-music-production\/","title":{"rendered":"New and Old User Interface Metaphors in Music Production"},"content":{"rendered":"<h3><strong>Introduction<\/strong><\/h3>\n<p>Some time ago I was skimming through a handful of books dealing with music production and spatial audio when preparing the curriculum for a university course. This led me to the discovery of a peculiar combination of claims regarding how to stage recorded sounds in the depth field:<\/p>\n<blockquote>\n<p style=\"text-align: left;\">\u201cSounds that are closer to us are louder and distant sounds are softer, therefore, the volume of a sound in the mix can be mapped out as front-to-back placement.\u201d<\/p>\n<p style=\"text-align: right;\">&#8211; David Gibson:\u00a0<em>The Art of Mixing<\/em>\u00a0(2005, p. 23)<\/p>\n<p>&nbsp;<\/p>\n<p>\u201cOne of the most useful cues for range estimation is loudness.\u201d<\/p>\n<p style=\"text-align: right;\">&#8211; Carlos Avendano \u2018Virtual Spatial Sound\u2019 (2004 p. 352)<\/p>\n<p>&nbsp;<\/p>\n<p>\u201cThe loudness of an instrument affects its balance in the mix; softer instruments also sound a bit farther away. But the primary influence on perception of depth and distance is the amount of early reflections and reverberation.\u201d<\/p>\n<p style=\"text-align: right;\">&#8211; Bob Katz:\u00a0<em>Mastering Audio<\/em>\u00a0(2007. p. 231)<\/p>\n<p>&nbsp;<\/p>\n<p>\u201cDistance is NOT loudness. In nature, distant sounds are often softer than near sounds. This is not necessarily the case in recording production. Loudness does not directly contribute to distance localisation in audio recordings.\u201d<\/p>\n<p style=\"text-align: right;\">&#8211; William Moylan:\u00a0<em>The Art of Recording<\/em>\u00a0(2002, p. 185)<\/p>\n<p>&nbsp;<\/p>\n<p>\u201cReverbs are the main tool we use when positioning sounds in the depth field.\u201d<\/p>\n<p style=\"text-align: right;\">&#8211; Roey Izhaki:\u00a0<em>Mixing Audio<\/em>\u00a0(2008, p. 403)<\/p>\n<\/blockquote>\n<p>In these five quotes there is notable disagreement concerning the importance of loudness as a distance cue. Gibson (2005) and Avendano (2005) describe loudness as the most important depth cue, while Katz (2007), Moylan (2002) and Izhaki (2008) find that loudness is less important, or even irrelevant, for auditory depth estimation in recordings. So, is loudness an important cue or simply an irrelevant depth cue in record listening? It is tempting to conclude that half of these scholars have got it all wrong. After all, from the point of view of psychoacoustics, loudness cannot be both an important and an unimportant depth cue. However, as I explain below, relying on psychoacoustics as a means to explain the spatiality of recorded sound may just provide a much too narrow framework for meaningful interaction with sound in music production.<\/p>\n<p>Psychoacoustics is essentially the study of how hearing works, that is, how sounds are processed by the ear. Thus, it assumes a conceptual judgement of sound that is directly tied to an external (mind-independent) reality. This has little to do with sense-making per se, which, in the main, belongs to the field of cognition. It could even be argued that psychoacoustic evidence does not possess any real value in the aesthetic decisions that govern the tasks mixing engineers are faced with (e.g. how does a mix engineer meaningfully and intuitively respond to a mix that requires certain spatial adjustments). Still, psychoacoustics provides a common framework in literature on music production (e.g. Rumsey: 2001; Corey 2010), although in recent years more writers apply cognitive approaches to the study of recorded music and record production (e.g. Moore, Schmidt and Dockwray 2009; Zagorski-Thomas: 2014).<\/p>\n<p>In this paper spatiality serves as an example of a much larger discussion pertaining to how music producers make sense of their actions when mixing. The focus is to explore the embodied cognitive abilities that govern interaction with sound in the music studio. I frame my discussion around cognitive linguistic theory (Lakoff: 1987; Johnson: 1987; K\u00f6vecses: 2002; Gibbs: 2008) building on the embodied mind paradigm. Using this framework, I discuss how <em>conceptual metaphors<\/em> and <em>image schemas<\/em> function as hidden structures in our cognitive system that we use to make sense of the often abstract\u2014or at least multifaceted\u2014 tasks required for interaction with sound. I propose that music mixing builds on three master metaphors: the <em>signal flow metaphor<\/em>, the <em>sound stage metaphor<\/em>, and the <em>container metaphor<\/em>. Each of these metaphors is discussed below. Finally, I present some future perspectives for interface designs that better reflect the different ways sound engineers, producers and designers make sense of their interaction with the mix.<\/p>\n<h3><strong>Cognitive Metaphors in User Interface Design<\/strong><\/h3>\n<p>Relying on visual metaphors is often seen as a means to design more intuitive user interfaces. In a recent article, Bell and colleagues (2015) present a history of visual design metaphors, where they assess recording software\u2019s typical dependence on visual design metaphors\u2014 skeuomorphism\u2014from analogue music production technology. They then argue that this dependence is likely to change in the future; and perhaps hardware design will even be modelled on software design as users come to adapt to novel interface schemes from music production software. Indeed, visual metaphors can assist the user in recalling prior knowledge in order to complete the task at hand. Visual metaphors, however, only have this function if they are familiar to the user. Furthermore, the user\u2019s understanding and ability to interact are also shaped by prior knowledge and experiences that are not tied specifically to music production.<\/p>\n<p>The principles of cognitive metaphor theory\u2014building on the idea of the embodied mind\u2014 provide a much broader platform to assess music production interface design. Several researchers in sonic interaction design, for example, have used the principles of embodied cognition to map body movements onto parameters of sound (Wilkie: 2009; Loeffler et al.: 2013). As an example of the benefits of this approach, Antle and colleagues (2009) have shown that interfaces built on embodied concepts are much easier to learn.<\/p>\n<p>The original principle of cognitive metaphor theory\u2014when the field emerged more than three decades ago\u2014is that actions, reasoning, and thought are structured by cognitive schemas that are formed through previous sensory-motor experiences. A metaphorical system operates at the back of our minds in every action and it allows us to make sense of what we do and to act appropriately. While many of these embodied sensory-motor experiences are nearly universal to humans, more recent research in cognitive linguistics shows that our cognitive capacities are also shaped by specific cultural understandings and learning processes (Beilock: 2009; Slepian: 2014). Hence, cognitive metaphors are not stable. While they are based, to some degree, on previous sensorimotor experiences common to most humans, they are also shaped by culture and other contextual aspects, such as the person\u2019s specific training, cultural understanding and the historical context. For the sake of clarity these conditions are listed below:<\/p>\n<ul>\n<li>Previous sensorimotor experiences<\/li>\n<li>Contextual understanding (history, culture)<\/li>\n<li>Learned concepts<\/li>\n<\/ul>\n<p>In the following section, I outline how each of these conditions for embodied knowledge contributes to the forming of the three master metaphors in music production. In doing so, I discus how the cognitive structures governing sense-making in music mixing have changed with new technologies and new terminology in the field. For example, a sound engineer that is taught to use a specific set of linguistics concepts for a specific music mixing operation may possess other embodied cognitive capacities and respond differently to a sound stimulus than another sound engineer who learned to use other concepts for the same operation.<\/p>\n<p>As already mentioned, I base my thesis on three master metaphors. I will provide closer descriptions of each of these in the following sections but it is worthwhile presenting them here in the form of graphical representations (fig. 1) that highlight a few key differences.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-3679\" src=\"https:\/\/arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/icons.png\" alt=\"\" width=\"813\" height=\"297\" srcset=\"https:\/\/www.arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/icons.png 813w, https:\/\/www.arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/icons-300x110.png 300w, https:\/\/www.arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/icons-768x281.png 768w\" sizes=\"auto, (max-width: 813px) 100vw, 813px\" \/><\/p>\n<p>Figure 1: Three cognitive metaphors governing music production<\/p>\n<p>The <em>signal flow metaphor<\/em> is represented as a channel strip\u2014a piece of technology that, here, serves as a symbol for a <em>technology-driven<\/em> mode of reasoning in music mixing. The <em>sound stage<\/em> is a representation of the <em>phenomenal space<\/em> of the recording that uses the actual performance space in live music as a metaphor to make sense of a recording\u2019s spatiality. Finally, the <em>container metaphor<\/em> takes the form of a <em>cognitive schema<\/em> that operates as an organizing structure or mental representation of experience and action. These metaphors, then, operate from different perspectives\u2014technology, phenomenal world, and cognition\u2014 and they serve here as examples of how different approaches to mixing might be theorized.<\/p>\n<h3><strong>The Signal Flow Metaphor<\/strong><\/h3>\n<p>The <em>signal flow metaphor<\/em> is one of the earliest and most persistent metaphors in the age of recording technology. As the name indicates, it started taking shape with electrical recording technology in the 1920s, where sound was no longer recorded directly onto disc or cylinder, but had to follow an electrical signal path from source through microphone then onto disc. Adjustments to the electrified signal could be made at <em>gain stages<\/em> that could either interrupt the signal flow or allow some part of the signal to pass through.<\/p>\n<p>The specific form of the <em>signal flow metaphor<\/em> I explore here found its present form by the end of 1950s with the implementation of the dynamic fader in mixing desk design. It is generally acknowledged that Tom Dowd was the first person to install dynamic faders on a mixing desk. As sound engineer in Atlantic Studios, NY he was among the first group of sound engineers with access to an eight track Ampex Recorder (number 3 in fact, Mitch Miller and Les Paul owned the first two, see Cunningham: 1996). Tom Dowd found that it was difficult to operate eight tracks with rotary knobs and therefore decided to swap the knobs with dynamic faders instead. Having eight dynamic faders at his disposal he could adjust several tracks at the same time with only one finger on each fader, which allowed him to \u2018perform\u2019 the mixing process in a much more intuitive way (Moorman: 2003).<\/p>\n<p>Now, as we know, the faders were made to function in such a way that the volume increases when the fader is pushed forward\/up and vice versa. Naturally, most engineers today take the direction of the fader movements for granted\u2014after all, this layout has been around for more than a half a century. However, if we instead imagine a set of \u2018reversed\u2019 faders, where the sound levels increase when you pull the faders towards you, the motor-activation the engineer would have to perform would fit the idea that increasing loudness makes sound sources appear closer as Gibson and Avendano propose \u00ad(e.g. you pull the fader towards you when you want the sound to appear closer and vice versa) [1]. So, why are fader movements not designed this way as default? The classic fader design makes sense from the point of view of signal routing. If full gain is brought up to the faders, the signal is allowed to pass through when you <em>open up<\/em> the gate by pushing the faders forward\/up, and the gate is closed when you pull the faders towards you\/down. There is thus a logical link between the way the sound flows through the technology and the actions the sound engineers perform.<\/p>\n<p>More interesting is that fact that, more than three decades into the age of digital technology, the <em>signal flow metaphor<\/em> is still the key model for the design of music mixing technology\u2014 both in the hardware and software domain. Graphical representations of the analogue channel strip layout are found in most of today\u2019s DAWs, not to mention buttons illustrating <em>tape loops<\/em>, <em>scissors<\/em>, <em>glue<\/em> and other tools from the age of analogue recording. Given the opportunities to redesign music production interfaces with digital technology, one should think that there are no, or only few, technological constrains that may hinder interface designers to change the metaphor. Implementing the logic of the analogue in the digital has, of course, obvious benefits in terms of allowing music producers to make a smooth transition from one domain to the other, but the reasons for sticking to this logic are fading.<\/p>\n<p>The <em>signal flow metaphor<\/em> is a \u2018learned\u2019 metaphor reflected in textbooks on music production and represented in much of the recording technology we operate. Thus, its strength is primarily related to the fact that many mix engineers have learned to operate equipment modelled on this logic\u2014rather than because the metaphor reflects more universal embodied capacities. If this thesis is right, it is fair to speculate that the <em>signal flow metaphor<\/em> will become weaker in the years to come\u2014or perhaps gradually disappear. Many students enrolling on music production courses at universities today have never, or only occasionally, laid their hands on analogue equipment; needless to say, the way they make sense of music production technology is not necessarily as bound up with the logic of analogue equipment as it is for their older colleagues. It is, therefore, reasonable to believe that other interface metaphors will grow stronger in the future and lead to music production technologies that afford other modes of interaction.<\/p>\n<h3><strong>The Sound Stage Metaphor<\/strong><\/h3>\n<p>In 1992 William Moylan (2002) proposed the concept of the <em>sound stage<\/em> as an analytical tool to evaluate stereo mixes. Although Moylan did not himself envision the sound stage as a metaphor, but as some kind of perceptual object (in fact Moylan explicitly states that metaphors are weak concepts that should be omitted in analyses of sound), it makes sense to think of the sound stage as derived from a performance metaphor, where individual sounds \u2018belong\u2019 to a performer located on a two-dimensional stage. The task of the mix engineer is then to distribute these sound sources in an appropriate way.<\/p>\n<p>If we think of tracks as <em>perceived performances<\/em>, every sound is an index for what produced it and for the physical gesture involved in its production. Meaningful interaction with recorded music thus relies on visual imagery where the location of musicians on a stage is brought to mind. Albin Zak (2001) proposes that the \u201cnatural image\u201d serves mainly as the point of reference for further manipulation and the creation of different forms of unique textures that are created for more dramatic effect and distinctiveness:<\/p>\n<blockquote><p>\u201dThe point of reference for a conventional stereo mix of a rock band is the visual image presented to a listener facing a bandstand. The vocalist is at centre stage with the drummer just behind \u2014 the kick and snare drums are dead centre, the high hat to one side, toms and cymbals spread from left to right. Because of its powerful sonic presence and its function as anchor of both groove and chord changes, the bass is also usually placed in the centre. Other instruments in the arrangement are spread across the stereo spectrum, as they would be on stage, balanced in such a way that all can be clearly distinguished. This \u2019natural\u2019 image is often manipulated for expressive effect, however, and a distinctive stereo image is often a thematic feature of a track.\u201d (Zak: 2001, p.145)<\/p><\/blockquote>\n<p>In Zak\u2019s account, it is still the perceived performance that serves as the underlying model for interacting with the recording. The creation of more expressive effects could then be viewed as ways to extend and enhance the recording&#8217;s expressiveness, while the underlying metaphor ties the different auditory experiences to a coherent phenomenal situation, where the listener <em>participates<\/em> in an imagined performance event. Philip Auslander has called the affective state emerging from this participation the experience of <em>liveness<\/em> by arguing that recordings form a technologically mediated co-presence of a performance, in spite of the fact that recordings represent a temporal gap between the performance and the reception (Auslander: 1999). The notion of liveness thus corresponds to the idea that recordings can create a <em>perceived performance environment<\/em> (Moylan: 2002), that is, the illusion of a shared space that includes both the listener and the perceived performance.<\/p>\n<p>Although one of the key tasks for many music producers is to design mixes that produce such affective relations in listeners, few interface designers have made any use of the sound stage metaphor. The introduction of digital audio in some of the big studios during the 1980s and the shift to mixing interfaces based solely on computers in the 1990s allowed every opportunity to implement the sound stage in the graphic design, but the <em>signal flow metaphor<\/em> remained the dominant basis for the design of DAWs coupled with the analogue tape metaphor.<\/p>\n<p>Recently, Steven Gelineck and colleagues (2013) developed a music-mixing controller for iPad\u2014Tangible Mix (fig 2.)\u2014that builds on the <em>sound stage metaphor<\/em>. In this interface, dragging elements towards you make them appear closer to the front edge of the sound stage and vice versa. The interface allows the user to make sense of the recording\u2019s spatiality in terms of the sound stage and also to interact with the sound without having to mentally shift back and forth between the <em>signal flow metaphor <\/em>and the <em>sound stage metaphor<\/em>. The application, then, combines the evaluation methods Moylan proposed\u2014that are taught on many music production programs at universities around the world\u2014with practical aesthetic decisions required from the engineer in the mixing process. As part of a current research project (Gelineck and Walther-Hansen: 2016) the interface is being evaluated in specific teaching situations at Aalborg University, Denmark, where students are being asked to use the interface to create mixes that fit specific style idioms. These tasks provide a means to further assess the suitability of the <em>sound stage metaphor<\/em> to perform different tasks in different parts of the mix process.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-3663\" src=\"https:\/\/arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/Figure-2-ScreenShot-Tangible_WalterHansen.png\" alt=\"Figure 2 ScreenShot Tangible_WalterHansen\" width=\"648\" height=\"486\" srcset=\"https:\/\/www.arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/Figure-2-ScreenShot-Tangible_WalterHansen.png 648w, https:\/\/www.arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/Figure-2-ScreenShot-Tangible_WalterHansen-300x225.png 300w\" sizes=\"auto, (max-width: 648px) 100vw, 648px\" \/><\/p>\n<p>Figure 2: Screen shot of \u2018Tangible Mix\u2019<\/p>\n<h3><strong>The Container Metaphor<\/strong><\/h3>\n<p>In a recent article (Walther-Hansen: 2015), I used a specialized linguistic corpus (approx. 50 million words) to explore the language of sound engineers and other audio professionals. I found that the concepts sound engineers use to describe sound and the actions they perform to edit the sound often point to an underlying cognitive <em>container schema<\/em> that governs how they experience recorded sound and how they describe sound quality in language. Consider the following examples from the corpus:<\/p>\n<ul>\n<li><em>Bring<\/em> <em>out<\/em> the vocal a little bit more<\/li>\n<li><em>Open up<\/em> the sound<\/li>\n<li><em>Compress<\/em> the sound more<\/li>\n<li>The mix sounds too <em>crowded<\/em><\/li>\n<li>I needed a longer reverb to <em>fill in spaces<\/em><\/li>\n<li>Keep a <em>hole open<\/em> <em>in the middle<\/em><\/li>\n<\/ul>\n<p>While these expressions may appear random they are, in fact, highly structured. The expressions (metaphors are in italics) are different representations of the <em>container metaphor<\/em> and the examples show how this schema structures our experience of sound and how we put this experience into words. The recording is not physically a container but we impose a <em>container schema<\/em> onto the recording to make sense of it, and to make sense of our actions when interacting with sound. Consider, also, how we might say that sounds and sound sources are <em>in<\/em> the recording, that sounds should be <em>compressed<\/em>, <em>expanded<\/em>, and so on. While some tracks may generate an experience of an<em> empty<\/em> or only partially full container, other tracks may activate a sense of <em>fullness<\/em>.<\/p>\n<p>Following the internal logic of this metaphor one might say that some sound sources in the recording take up more space than other sound sources. Low frequency sounds, for instance, are often characterised as both <em>larger <\/em>and <em>heavier<\/em> than high frequency sounds. Also, loud level sounds often appear <em>larger<\/em> than low level sounds (Gibson: 2005). Sounds thus have different <em>sizes <\/em>that take up more or less space in the container. The boundaries of the container may have different characteristics. We can think of containers that enclose the sounds in different ways, providing a more <em>closed<\/em> or <em>open<\/em> structure; for example, in some tracks the sounds may seem fixed and constrained, while other tracks have sounds that are more <em>loosely constrained <\/em>(see Walther-Hansen: 2014 for a broader overview of the <em>container metaphor<\/em>).<\/p>\n<p>Although, as I have previously argued (Walther-Hansen: 2014; 2015), the <em>container metaphor<\/em> is a prevalent and established metaphor in the language of sound and this metaphor is deeply grounded in embodied sensorimotor experiences, it remains to be acknowledged as a model to design music production technology. The reasons are undoubtedly tied to the lack of an obvious way to put the <em>container metaphor<\/em> into operation with present day technology because this would require some kind of mid-air interface with haptic feedback that would allow for precise adjustments of audio (fig. 3).<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-3664\" src=\"https:\/\/arpjournal.com\/asarpwp\/wp-content\/uploads\/2017\/03\/Figure-3-MidAir-Interaction-_WalterHansen.png\" alt=\"Figure 3 MidAir Interaction _WalterHansen\" width=\"294\" height=\"162\" \/><\/p>\n<p>Figure 3: Mid-air interaction<\/p>\n<p>However, a number of technologies point to the possibility of such interfaces becoming a reality in the future. One of them is Disney\u2019s AIREAL system (Sodhi et al.: 2013) that uses compressed air to provide haptic feedback in mid-air. Second \u2013 and more promising \u2013 a group of researchers from Bristol University has developed a prototype of a mid-air interface that provides haptic feedback with ultrasound (Long et al.: 2014). Contrary to Disney\u2019s system, this allows the user to <em>shape<\/em> objects in mid-air with haptic feedback and this, then, allows the user to move elements of sound around, to <em>expand<\/em> the container, to <em>compress<\/em> the container and so on in three-dimensional, mid-air interaction.<\/p>\n<p>I suggest here that mid-air interface technology offers\u2014at least in theory\u2014the prospects for the realization of the <em>container metaphor<\/em> in music production, yet these technologies are currently lacking in a number of areas (e.g. the detail by which you can control \u2018objects\u2019 in free air). As such deficiencies are improved or become accepted by the user, mid-air interfaces are likely to provide the technological means for the attainment of a more intuitive interface.<\/p>\n<h3><strong>Conclusion and Future Perspectives<\/strong><\/h3>\n<p>In this paper I proposed three master metaphors for music production that govern interaction on many levels on the creative process. The goal of this research is to facilitate the design of new interfaces for music production that more closely reflect how the user make sense of the mix and allow for more intuitive interaction with recorded sound. My study shows that the three metaphors\u2014<em>signal flow<\/em>, <em>sound stage<\/em>, and <em>container<\/em>\u2014function in different ways as mental representations of the mixing process, and these mental representations are activated when performing different tasks (e.g. interaction with technology, evaluating a track etc.) that allow us to assess their applicability as interface metaphors.<\/p>\n<p>The <em>signal flow metaphor<\/em> makes sense from the point of view of technology and thus fits into an \u2018engineering\u2019 logic, where an understanding of the \u2018flow of sound\u2019 through the mixing desk is the key element. It ensures a systemic understanding of how the technology works and it allows the engineer\u2014assuming the engineer has learned to operate the equipment\u2014to make a causal connection between the operations he\/she makes and the alterations made to the sound at different stages as it flows through the system.<\/p>\n<p>The <em>sound stage metaphor<\/em> arises from a search for phenomenal coherence in recorded sound. The sound stage is a mental representation of an experienced auditory \u2018reality\u2019, where sounds are causally related to imagined physical sources located on an imagined stage. This metaphor can be (and is) implemented with current technology and it allows for a direct link between auditory stimuli, visual feedback and haptic interaction. Still, it presupposes a view on sound as sound sources, that is, a causal form of listening (Chion: 1994), that affords certain kinds of interactions (e.g. moving sound sources around in the imagined space of the recording) at the expense of others (e.g. shaping the sound\u2019s timbral qualities).<\/p>\n<p>The <em>container metaphor<\/em> functions as a cognitive schema that is activated when approaching sounds as <em>shapeable <\/em>objects. This is the position of the sound designer or producer, rather than the engineer. As suggested earlier, if we understand and reason about sound as metaphorical objects and the imagined space of the recording as a metaphorical container, the best way to convert this understanding to meaningful interaction is to develop interfaces that allow us to shape and move sounds as if they were three-dimensional objects and the entire mix as if it were a three-dimensional container. This is not, or not yet, possible with current technology, but there are certainly future prospects for the realization of the theoretical framework presented here.<\/p>\n<p>As previously mentioned, it seems likely that we use different metaphors in different parts of the process. In that case, the ideal controller interface should change its appearance to fit each part of the process. Future studies should explore which metaphors are most relevant in any given part of the music production process, relevant to specific genres, or relevant as a framework to work towards specific aesthetic ideals and so on. Changes to the master metaphor of music mixing also lead to questions about what to teach our music production students in the future. Will the gap between engineers (who know how the signal flow works) and designers or producers (who are primarily concerned with the quality or aesthetics of the sounds) grow bigger? And do the latter need to know about signal flow in analogue mixing desks at all? If not, what kind of interfaces should they learn to operate? Will it then be possible to focus exclusively on the \u2018design\u2019 tasks of music production as opposed to the more traditional \u2018engineering\u2019 tasks? Such questions are again linked to the conditions of sense-making and how well the technology we use reflects our thinking; to this end, I have here presented different forms of sense-making that can serve as models for future advances in this field.<\/p>\n<h3><strong>References<\/strong><\/h3>\n<p>Antle, A. N., Corness G, and Droumeva, M. (2009). \u2018What the Body Knows: Exploring the Benefits of Embodied Metaphors in Hybrid Physical Digital Environments. In: <em>Interacting with Computers<\/em>, 21, pp.\u00a0 66\u201375.<\/p>\n<p>Auslander, P. (1999). <em>Liveness: Performance in a Mediatized Culture<\/em>. Oxon: Routledge.<\/p>\n<p>Avendano, C. (2004). \u2018Virtual Spatial Sound\u2019. In: Y. Huang and J. Benesty (eds.), <em>Audio Signal Processing for Next-generation Multimedia Communication Systems.<\/em> Kluwer Academic Publishers.<\/p>\n<p>Beilock, S. L. (2009). \u2018Grounding Cognition in Action: Expertise, Comprehension, and Judgment\u2019, In: <em>Progress in Brain Research, 714<\/em>.<\/p>\n<p>Bell, A., Hein, E. and Ratcliffe, J. (2015) Beyond Skeuomorphism: The Evolution of Music Production Software User Interface Metaphors. In <em>Journal of the Art of Record Production, <\/em>Issue 09, April.<\/p>\n<p>Chion, M. (1994). <em>Audio-Vision: Sound on Screen<\/em>. New York: Columbia U.P.<\/p>\n<p>Corey, J. (2010). <em>Audio Production and Critical Listening: Techncical Ear Training<\/em>. Focal Press.<\/p>\n<p>Cunningham, M. (1996). <em>Good Vibrations: A History of Recording Production<\/em>. Chessington, Surrey: Castle Communications.<\/p>\n<p>Gelineck, S. et al. (2013). \u2018Towards an Interface for Music Mixing based on Smart Tangibles and Multitouch\u2019. In: <em>New Interface for Musical Expression<\/em>. Daejeon\/Seoul, Korea Republic: Association for Computing Machinery.<\/p>\n<p>Gelineck, S. and Walther-Hansen, M. (2016) Tangible Mix as a Learning Tool for Music Production. Forthcoming presentation at the <em>Art of Record Production Conference<\/em>, Aalborg, Denmark.<\/p>\n<p>Gibbs, J. R. W. (2008). <em>The Cambridge Handbook of Metaphor and Thought<\/em>. Cambridge and New York: Cambridge University Press.<\/p>\n<p>Gibson, D. (2005). <em>The Art of Mixing: A Visual Guide to Recording, Engineering, and Production<\/em>. Boston: Thomson Course Technology.<\/p>\n<p>Izhaki, R. (2008). <em>Mixing Audio: Concepts, Practices and Tools<\/em>. Boston: Focal Press.<\/p>\n<p>Johnson, M. (1987). <em>The Body in the Mind: The Bodily Basis of Meaning, Imagination, and Reason<\/em>. Chicago London: The University of Chicago Press.<\/p>\n<p>Katz, B. (2007). <em>Mastering Audio: The Art and the Science<\/em>. Oxford: Focal Press.<\/p>\n<p>K\u00f6vecses, Z. (2002). <em>Metaphor: A Practical Introduction<\/em>. Oxford and New York: Oxford University Press.<\/p>\n<p>Lakoff, G. (1987). <em>Women, Fire and Dangerous Things: What Categories Reveal about the Mind. <\/em>Chicago and London: The University of Chicago Press.<\/p>\n<p>Loeffler, D. et al. (2013). \u2018Developing Intuitive User Interfaces by Integrating Users&#8217; Mental Models into Requirements Engineering. In: <em>Proceedings of the 27th International BCS Human Computer Interaction Conference.<\/em><\/p>\n<p>Long, B. et al. (2014). \u2018Rendering Volumetric Haptic Shapes in Mid-Air using Ultrasound\u2019. In: <em>ACM Transactions on Graphiphs, 33(6). <\/em><\/p>\n<p>Moore, A. F., Schmidt, P., and Dockwray, R. (2009) A Hermeneutics of Spatialization for Recorded Song, In: <em>twentieth-century music 6 (01)<\/em>, pp.83-114<\/p>\n<p>Moorman, M. (2003). \u2018Tom Dowd and the Language of Music\u2019 (documentary movie).<\/p>\n<p>Moylan, W. (2002). <em>The Art of Recording: Understanding and Crafting the Mix<\/em>. New York and Oxford: Focal Press.<\/p>\n<p>Rumsey, F. (2001). <em>Spatial Audio<\/em>. Oxford: Focal Press.<\/p>\n<p>Slepian, M. L., and Ambady, N. (2014). \u2018Simulating Sensorimotor Metaphors: Novel Metaphors Influence Sensory Judgments\u2019. In: <em>Cognition, 130<\/em>, pp. 309-314.<\/p>\n<p>Sodhi, R., Poupyrev, I., Glisson, M., Israr, A. (2013): \u2018AIREAL: Interactive Tactile Experiences in Free Air\u2019. In: <em>SIGGRAPH Conference Proceedings<\/em>, Pittsburg.<\/p>\n<p>Walther-Hansen, M. (2014) \u2018The Force Dynamic Structure of the Phonographic Container: How Sound Engineers Conceptualise the \u2018Inside\u2019 of the Mix\u2019. In: <em>Journal of Music and Meaning, 12<\/em>.<\/p>\n<p>Walther-Hansen, M. (2015) \u2018Balancing Audio: Towards a Cognitive Structure of Sound Interaction in Music Production\u2019<em>.<\/em> In: <em>Proceedings of Computer Music Multidisciplinary Research<\/em>, Plymouth, UK.<\/p>\n<p>Wilkie, K., Holland, S., and Mulholland, P. (2009). \u2018Evaluating Musical Software Using Conceptual Metaphors. In: <em>Proceedings of BCS Human-Computer Interaction<\/em>, 1- 5 September 2009, Cambridge, UK.<\/p>\n<p>Zagorski-Thomas, S. (2014). <em>The Musicology of Record Production<\/em>, Cambridge: Cambridge University Press.<\/p>\n<p>Zak, A. (2001). <em>The Poetics of Rock: Cutting Tracks, Making Records<\/em>. Berkeley: University of California Press.<\/p>\n<p><strong>Notes<\/strong><\/p>\n<p>[1] For several decades, mixing desks at the BBC in fact had reversed faders. This was not, however, an attempt to adapt to a phenomenal perspective rather than a systemic logic. The reversed faders had a more pragmatic explanation. Allegedly the radio speaker would accidently hit the faders once in a while with his notebook. Instead of accidently increasing the sound level to the max, hitting the faders would shut the sound off, which was seen as far better.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>This paper outlines a theoretical framework for interaction with sound in music mixing. Using cognitive linguistic theory and studies exploring the spatiality of recorded music, it is argued that the logic of music mixing builds on three master metaphors\u2014the signal flow metaphor, the sound stage metaphor and the container metaphor. I show how the metaphorical basis for interacting with sound in music mixing has changed with the development of recording technology, new aesthetic ideals and changing terminology. These changes are studied as expressions of underlying thought patterns that govern how music producers and engineers make sense of their actions. In conclusion, this leads to suggestions for a theoretical framework through which more intuitive music mixing interfaces may be developed in the future.<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[60],"tags":[391,390,388,389,392],"class_list":["post-3585","post","type-post","status-publish","format-standard","hentry","category-articles-editorials-provocations","tag-conceptual-metaphors","tag-embodied-cognition","tag-intuitive-interaction","tag-mixing-interfaces","tag-spatiality","author-mads-walther-hansen"],"_links":{"self":[{"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/posts\/3585","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/comments?post=3585"}],"version-history":[{"count":16,"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/posts\/3585\/revisions"}],"predecessor-version":[{"id":3680,"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/posts\/3585\/revisions\/3680"}],"wp:attachment":[{"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/media?parent=3585"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/categories?post=3585"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.arpjournal.com\/asarpwp\/wp-json\/wp\/v2\/tags?post=3585"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}