It’s an all too familiar movie trope. A bug hidden in a flower jar. A figure in shadows crouched listening at a door. The tape recording that no one knew existed, revealed at the most decisive of moments. Even the abrupt disconnection of a phone call manages to arouse the suspicion that we are never as alone as we may think. And although surveillance derives its meaning the latin “vigilare” (to watch) and French “sur-“ (over), its deep connotations of listening have all but obliterated that distinction.
In the final entry to our series on Sound and Surveillance, sound artist Anne Zeitz dissects the theory behind her installation Retention. What are the sounds of capture, and how do the sounds produced in and around spaces of capture affect our bodies? Listen in to find out. -AT
CLICK HERE TO DOWNLOAD: Anne Zeitz and David Boureau’s “Retention”
SUBSCRIBE TO THE SERIES VIA ITUNES
ADD OUR PODCASTS TO YOUR STITCHER FAVORITES PLAYLIST
This podcast presents Retention, a quadriphonic sound installation made with David Boureau. It considers the sounds of surveillance, detention and migration. Retention concentrates on the “soundscape” of the Mesnil Amelot 2+3 detention center for illegal immigrants situated to the North of Paris just beside the Charles de Gaulle airport. This center constitutes the largest complex for detaining “illegal immigrants” in France, with 240 places for individuals and families. Approximately 350 airplanes pass closely above the center over a 24 hours time span, creating intervals of very high sound levels that regularly drown out all other ambient sounds. Retention uses quadrophonic recording technology to capture and diffuse a live transmission of communication between pilots and the Charles de Gaulle control tower. The work also integrates recordings from inside the center made by communications via mobile phones. In the short intervals of silence (always implying sounds of some sort), the atmosphere seems suspended. This suspension is paradigmatic for the clash between the local and the global, between those who are trapped in a state of detention before being expulsed by the engines moving over their heads and those who circulate freely (nonetheless under surveillance) in our global society. Retention exhibits a changing sonic space in order to consider how “waiting zones” and processes of mobility meet.
Featured Image (c) Anne Zeitz and David Boureau, Retention, 2012.
Anne Zeitz is a researcher and artist working with photography, video, and sound media. Born in Berlin in 1980, she lives and works in Paris. Her research focuses on mechanisms of surveillance and mass media, theories of observation and attention, and practices of counter-observation in contemporary art. Her doctoral thesis (University Paris 8/ Esthétique, Sciences et Technologies des Arts, dissertation defence November 2014) is entitled (Counter-)observations, Relations of Observation and Surveillance in Contemporary Art, Literature and Cinema. Anne Zeitz was responsible for organizing the project Movement-Observation-Control (2007/2008) for the Goethe-Institut Paris and collaborated on the exhibition and conference Armed Response (2008) at the Goethe-Institut Johannesburg. She is a former member of the Observatoire des nouveaux médias (Paris 8/Ensad) and of the research project Média Médiums (Université Paris 8, ENSAPC, EnsadLAB, Archives Nationales, 2013/2014). Her most recent research concentrates on the work of the American artist Max Neuhaus with the publication of De Max-Feed a Radio Net (2014), part of the Média Médiums book series. She is the artist of this year’s Urban Photo Fest and participated at the Urban Encounters / Tate Britain in October 2014.
REWIND! . . .If you liked this post, you may also dig:
Toward a Civically Engaged Sound Studies, or ReSounding Binghamton – Jennifer Stoever
Welcome back to Sculpting the Film Soundtrack, SO!‘s new series on changing notions about how sound works in recent film and in recent film theory, edited by Katherine Spring.
Two weeks ago, Benjamin Wright started things off with a fascinating study of Hans Zimmer, a highly influential composer whose film scoring borders on engineering — or whose engineering borders on music — in many major Hollywood releases. This week we turn to the opposite end of the spectrum to a seemingly smaller film, Shane Carruth’s Upstream Color (2013), which has made quite a few waves among sound studies scholars and fans of sound design, even earning a Special Jury Prize for sound at Sundance.
To unpack the many mysteries of the film and explore its place in the field of contemporary filmmaking, we are happy to welcome musicologist and film scholar Danijela Kulezic-Wilson of University College Cork. Listen to Upstream Color through her ears (it’s currently available to stream on Netflix) and perhaps you’ll get a sense of why you’ll have to listen to it two or three more times. At least.
When Shane Carruth’s film Upstream Color was released in 2013, critics described it in various ways—as a body horror film, a sci-fi thriller, a love story, and an art-house head-scratcher—but they all agreed that it was a film “not quite like any other”. And while the film’s cryptic imagery and non-linear editing account for most of the “what the hell?” reactions (see here for example), I argue that the reason for its distinctively hypnotic effect is Carruth’s musical approach to the film’s form: he organizes the images and sounds according to principles of music, including the use of repetition, rhythmic structuring, and antiphony.
The resulting musicality of Upstream Color may not be surprising given that Carruth composed most of the score, and also, as Jonathan Romney has noted in Sight & Sound, Carruth has said on many occasions that he was hoping “people would watch this film repeatedly, as they might listen to a favourite album” (52). In this sense, Carruth (whose DIY toolkit also includes writing, directing, acting, producing, cinematography, and editing) joins the ranks of filmmakers such as Darren Aronofsky and Joe Wright who recognize that, despite our culture’s obsession with the cinematic and narrative aspects of “visual” media, music governs film’s deepest foundations.
Upstream Color is a story about a woman, Kris, who is kidnapped by a drug manufacturer (referred to in the credits as Thief) and contaminated with a worm that keeps her in a trance-like state during which the Thief strips her of all her savings. Kris is subsequently dewormed by a character known as the Sampler, who transfers the parasite into a pig that maintains a physical and/or metaphorical connection to Kris. Kris later meets and falls in love with Jeff who, we eventually discover, has been a victim of the same ordeal. Although the bizarreness of the plot has encouraged numerous interpretations, the film’s unconventional audio-visual language suggests that its story of two people who share supressed memories of the same traumatic experience shouldn’t be taken at face value, but rather serves as a metaphor for existential anxiety resulting from being influenced by unknown forces.
Such an interpretation owes as much to the film’s disregard for the rules of classical storytelling as it does to a formally innovative soundtrack, one that uses musicality as an overarching organizing principle. The fact that Carruth wrote the score and script simultaneously (discussed in the video below) indicates the extent to which music was from the beginning considered an integral part of the film’s expressive language. More importantly, as the scenes discussed in this post suggest, the musical logic of the film is even more pervasive than the style, role, and placement of the actual score.
Whereas feature films traditionally assign a central role to speech, allowing music and sound effects supporting roles only, Upstream Color breaks down the conventional soundtrack hierarchy, often reversing the roles of each constitutive element. For example, hardly any information in the film that could be considered vital to understanding the story is communicated through speech. Instead, images, sound, music, and editing–for which Carruth shares the credit with fellow indie director David Lowery–are the principal elements that create the atmosphere, convey the sense of the protagonists’ brokenness, and reveal the connection between the characters. At the same time, characters’ conversations are either muted or their speech is blended with music in such a way that we’re encouraged to focus on body language or mise-en-scène rather than trying to discern every spoken word. For example, Jeff and Kris’s flirting with one another during their initial meetings (at roughly 0.44.20-0.47.00 of the film) is conveyed primarily through gestures, glances, and fragmentary editing rather than speech, which would be more typical for this sort of narrative situation.
Further undercutting the significance of speech across the film is how the film has been edited to resemble the flow of music. For example, non-linear jumps in the narrative are often arranged in such a way as to create syncopated audio-visual rhymes. This technique is particularly obvious in the montage sequence in which Kris and Jeff argue over the ownership of their memories, whose similarities suggest that they were implanted during the characters’ kidnappings. In this sequence, both the passing of time and the recurrence of the characters’ argument is conveyed through the repetition of images that become visual refrains: Kris and Jeff lying on a bed, watching birds flying above trees, touching each other. Some of these can be seen in the film’s official trailer:
The scene’s images and sounds are fragmented into a non-linear assembly of pieces of the same conversation the characters had at different times and places, like the verses and the choruses of a song. Importantly, the assemblage is also patterned, with phrases like “we should go on a trip” and “where should we go?” heard in refrain. The first time we hear Jeff say “we should go on a trip” and suggesting that they go “somewhere bright”, his words are played in sync with the image of him and Kris lying on the bed. The following few shots, accompanied only by music, symbolize the “honeymoon” phase of their relationship: the couple kiss, hold hands, and walk with their hands around each other’s waists. A shift in mood is marked by the repetition of the dialogue, with Jeff again saying “we should go on a trip” – only this time, the phrase plays asynchronously over a shot of Jeff and Kris pushing a table into the house that they have moved into together. Finally, the frustration that starts infiltrating the characters’ increasingly heated arguments is alleviated by the repetition of the sentence “They could be starlings.” As it is spoken three times by both characters in an antiphonic exchange, the phrase emphasizes the underlying strength of their connection and gives the scene a rhythmic balance. Across this sequence, the musical organization of audio-visual refrains prompts us to recognize the psychic connection between Kris and Jeff, and even to begin to guess the sinister reason for it.
While speech in Upstream Color is often stripped of its traditional role as a principal source of information, sound and music are given important narrative functions, illuminating hidden connections between the characters. In one of the most memorable scenes, the Sampler is revealed to be not only a pig farmer but also a field recordist and sound artist who symbolizes the hidden source of everything that affects Kris and Jeff from afar. As we hear the sounds of the Sampler’s outdoor recordings merge with and emulate the sounds made and heard by Kris and Jeff at home and at work, the soundtrack eloquently establishes the connection between all three characters while also giving us a look “behind the scenes” of Kris’s and Jeff’s lives and suggesting how they are influenced from a distance.
In one sense, by calling attention to the very act of recording sound, the scene exposes how films are constructed, offering a reflexive glimpse into usually hidden processes of production. The implied idea here–that the visible and audible are products of not-so-obvious processes of formation–refers not only to the medium of film but also to the complexity of the inner workings of someone’s mind. Thus the Sampler’s role, his actions, and his relationship to Kris, Jeff, and other infected victims can be interpreted as a metaphor for the subconscious programming – all the familial, social and cultural influences – that all of us are exposed to from an early age. The Sampler is portrayed symbolically as the Creator, a force whose actions affect the protagonists’ lives without them knowing it. The fact that he is simultaneously represented as a sound artist establishes sound-making and musicality as the film’s primary creative principles.
Considering Carruth’s very deliberate departure from the conventions of even what David Bordwell calls “intensified” storytelling, it is fair to say that Upstream Color is a film that weakens the strong narrative role traditionally given to oral language. What is intensified here are the musical and sensuous qualities of the audio-visual material and a mode of perception that encourages absorption of the subtext (in other words, the metaphorical meaning of the film) as well as the text.
The musical organization of film form and soundtrack is no longer limited to independent projects such as Carruth’s Upstream Color. As I have shown elsewhere, musicality has become an extremely influential principle in contemporary cinema, acting as an inspiration and model for editing, camera movement, movement within a scene and sound design. Some of the most interesting results of a musical approach to film include Aronofsky’s “hip hop montage” in Pi (1998) and Requiem for a Dream (2000), Jim Jarmusch’s rhythmically structured film poems (The Limits of Control, 2009; Only Lovers Left Alive, 2013), the interchangeable use of musique concrète and environmental sound in Gus Van Sant’s Death Trilogy and films by Peter Strickland (Katalin Varga, 2009; Berberian Sound Studio, 2012); the choreographed mise-en-scène in Joe Wright’s Anna Karenina (2012); the musicalization of language in Harmony Korine’s Spring Breakers (2012); and the foregrounding of musical material over intelligible speech in Drake Doremus’s Breathe In (2013). Given the breadth of these examples, it’s no exaggeration to say that filmmakers’ growing affinity for a musical approach to film is changing the landscape of contemporary cinema.
Danijela Kulezic-Wilson teaches film music, film sound, and comparative arts at University College Cork. Her research interests include approaches to film that emphasize its inherent musical properties, the use of musique concrète and silence in film, the musicality of sound design, and musical aspects of the plays of Samuel Beckett. Danijela’s publications include essays on film rhythm, musical and film time, the musical use of silence in film, Darren Aronofsky’s Pi, P.T. Anderson’s Magnolia, Peter Strickland’s Katalin Varga, Gus Van Sant’s Death Trilogy, Prokofiev’s music for Eisenstein’s films, and Jim Jarmusch’s Dead Man. She has also worked as a music editor on documentaries, short films, and television.
All images taken from the film.
REWIND! . . .If you liked this post, you may also dig:
Animal Renderings: The Library of Natural Sounds– Jonathan Skinner
Welcome to our new series Sculpting the Film Soundtrack, which brings you new perspectives on sound and filmmaking. As Guest Editor, we’re honored and delighted to have Katherine Spring, Associate Professor of Film Studies at Wilfrid Laurier University. Spring is the author of an exciting and important new book Saying it With Songs: Popular Music and the Coming of Sound to Hollywood Cinema. Read it! You’ll find an impeccably researched work that’s the definition of how the history of film sound and media convergence ought to be written.
But before rushing back to the early days, stick around here on SO! for the first of our three installments in Sculpting the Film Soundtrack.
It’s been 35 years since film editor and sound designer Walter Murch used the sounds of whirring helicopter blades in place of an orchestral string section in Apocalypse Now, in essence blurring the boundary between two core components of the movie soundtrack: music and sound effects. This blog series explores other ways in which filmmakers have treated the soundtrack as a holistic entity, one in which the traditional divisions between music, effects, and speech have been disrupted in the name of sculpting innovative sonic textures.
In three entries, Benjamin Wright, Danijela Kulezic-Wilson, and Randolph Jordan will examine the integrated soundtrack from a variety of perspectives, including technology, labor, aesthetic practice, theoretical frameworks, and suggest that the dissolution of the boundaries between soundtrack categories can prompt us to apprehend film sound in new ways. If, as Murch himself once said, “Listening to interestingly arranged sounds makes you hear differently,” then the time is ripe for considering how and what we might hear across the softening edges of the film soundtrack.
- Guest Editor Katherine Spring
Composing a sound world for Man of Steel (2013), Zack Snyder’s recent Superman reboot, had Hans Zimmer thinking about telephone wires stretching across the plains of Clark Kent’s boyhood home in Smallville. “What would that sound like,” he said in an interview last year. “That wind making those telephone wires buzz – how could I write a piece of music out of that?” The answer, as it turned out, was not blowing in the wind, but sliding up and down the scale of a pedal steel guitar, the twangy lap instruments of country music. In recording sessions, Zimmer instructed a group of pedal steel players to experiment with sustains, reverb, and pitches that, when mixed into the final track, accompany Superman leaping over tall buildings at a single bound.
His work on Man of Steel, just one of his most recent films in a long and celebrated career, exemplifies his unique take on composing for cinema. “I would have been just as happy being a recording engineer as a composer,” remarked Zimmer last year in an interview to commemorate the release of a percussion library he created in collaboration with Spitfire Audio, a British sample library developer. “Sometimes it’s very difficult to stop me from mangling sounds, engineering, and doing any of those things, and actually getting me to sit down and write the notes.” Dubbed the “HZ01 London Ensembles,” the library consists of a collection of percussion recordings featuring many of the same musicians who have performed for Zimmer’s film scores, playing everything from tamtams to taikos, buckets to bombos, timpani to anvils. According to Spitfire’s founders, the library recreates Zimmer’s approach to percussion recording by offering a “distillation of a decade’s worth of musical experimentation and innovation.”
In many ways, the collection is a reminder not just of the influence of Zimmer’s work on contemporary film, television, and video game composers but also of his distinctive approach to film scoring, one that emphasizes sonic experimentation and innovation. Having spent the early part of his career as a synth programmer and keyboardist for new wave bands such as The Buggles and Ultravox, then as a protégé of English film composer Stanley Myers, Zimmer has cultivated a hybrid electronic-orchestral aesthetic that uses a range of analog and digital oscillators, filters, and amplifiers to twist and augment solo instrument samples into a synthesized whole.
Zimmer played backup keyboards on “Video Killed the Radio Star.”
In a very short time, Zimmer has become a dominant voice in contemporary film music with a sound that blends melody with dissonance and electronic minimalism with rock and roll percussion. His early Hollywood successes, Driving Miss Daisy (1989) and Days of Thunder (1990), combined catchy themes and electronic passages with propulsive rhythms, while his score for Black Rain (1989), which featured taiko drums, electronic percussion, and driving ostinatos, laid the groundwork for an altogether new kind of action film score, one that Zimmer refined over the next two decades on projects such as The Rock (1996), Gladiator (2000), and The Pirates of the Caribbean series.
What is especially intriguing about Zimmer’s sound is the way in which he combines the traditional role of the composer, who fashions scores around distinct melodies (or “leitmotifs”), with that of the recording engineer, who focuses on sculpting sounds. Zimmer may not be the first person in the film business to experiment with synthesized tones and electronic arrangements – you’d have to credit Bebe and Louis Barron (Forbidden Planet, 1956), Vangelis (Chariots of Fire, 1981), Jerry Goldsmith (Logan’s Run, 1976), and Giorgio Moroder (Midnight Express, 1981) for pushing that envelope – but he has turned modern film composing into an engineering art, something that few other film composers can claim.
One thing that separates Zimmer’s working method from that of other composers is that he does not confine himself to pen and paper, or even keyboard and computer monitor. Instead, he invites musicians to his studio or a sound stage for an impromptu jam session to find and hone the musical syntax of a project. Afterwards, he returns to his studio and uses the raw samples from the sessions to compose the rest of the score, in much the same way that a recording engineer creates the architecture of a sound mix.
“There is something about that collaborative process that happens in music all the time,” Zimmer told an interviewer in 2010. “That thing that can only happen with eye contact and when people are in the same room and they start making music and they are fiercely dependent on each other. They cannot sound good without the other person’s part.”
Zimmer facilitates the social and aesthetic contours of these off-the-cuff performances and later sculpts the samples into the larger fabric of a score. In most cases, these partnerships have provided the equivalent of a pop hook to much of Zimmer’s output: Lebo M’s opening vocal in The Lion King (1994), Johnny Marr’s reverb-heavy guitar licks in Inception, Lisa Gerrard’s ethereal vocals in Gladiator and Black Hawk Down (2002), and the recent contributions of the so-called “Magnificent Six” musicians to The Amazing Spider Man 2 (2014).
The melodic hooks are simple but infectious – even Zimmer admits he writes “stupidly simple music” that can often be played with one finger on the piano. But what matters most are the colors that frame those notes and the performances that imbue those simple melodies with a personality. Zimmer’s work on Christopher Nolan’s Dark Knight trilogy revolves around a deceptively simple rising two-note motif that often signifies the presence of the caped crusader, but the pounding taiko hits and bleeding brass figures that surround it do as much to conjure up images of Gotham City as cinematographer Wally Pfister’s neo-noir photography. The heroic aspects of the Batman character are muted in Zimmer’s score except for the presence of the expansive brass figures and taiko hits, which reach an operatic crescendo in the finale, where the image of Batman escaping into the blinding light of the city is accompanied by a grand statement of the two-note figure backed by a driving string ostinato. Throughout the series, a string ostinato and taikos set the pace for action sequences and hint at the presence of Batman who lies somewhere in the shadows of Gotham.
Zimmer’s expressive treatment of musical colors also characterizes his engineering practices, which are more commonly used in the recording industry. Music scholar Paul Théberge has noted that the recording engineer’s interest in an aesthetic of recorded musical “sound” led to an increased demand for control over the recording process, especially in the early days of multitrack rock recording where overdubbing created a separate, hierarchical space for solo instruments. Likewise for Zimmer, it’s not just about capturing individual sounds from an orchestra but also layering them into a synthesized product. Zimmer is also interested in experimenting with acoustic performances, pushing musicians to play their instruments in unconventional ways or playing his notes “the wrong way,” as he demonstrates here in the making of the Joker’s theme from The Dark Knight:
The significance of the cooperative aspects of these musical performances and their treatment as musical “colors” to be modulated, tweaked, and polished rests on a paradoxical treatment of sound. While he often finds his sound world among the wrong notes, mistakes, and impromptu performances of world musicians, Zimmer is also often criticized for removing traces of an original performance by obscuring it with synth drones and distortion. In some cases, like in The Peacemaker (1997), the orchestration is mushy and sounds overly processed. But in other cases, the trace of a solo performance can constitute a thematic motif in the same way that a melody serves to identify place, space, or character in classical film music. Compare, for instance, Danny Elfman’s opening title theme for Tim Burton’s Batman (1989) and Zimmer’s opening title music for The Dark Knight. While Elfman creates a suite of themes around a central Batman motif, Zimmer builds a sparse sound world that introduces a sustained note on the electric cello that will eventually be identified with the Joker. It’s the timbre of the cello, not its melody, that carries its identifying features.
To texture the sounds in Man of Steel, Zimmer also commissioned Chas Smith, a Los Angeles-based composer, performer, and exotic instrument designer to construct instruments from “junk” objects Smith found around the city that could be played with a bow or by hand while also functioning as metal art works. The highly abstract designs carry names that give some hint to their origins – “Bertoia 718” named after modern sculptor and furniture designer Harry Bertoia; “Copper Box” named for the copper rods that comprise its design; and “Tin Sheet” that, when prodded, sounds like futuristic thunderclaps.
Smith’s performances of his exotic instruments are woven into the fabric of the score, providing it with a sort of musical sound design. Consider General Zod’s suite of themes and motifs, titled “Arcade” on the 2-disc version of the soundtrack. The motif is built around a call-and-answer ostinato for strings and brass that is interrupted by Smith’s sculptural dissonance. It’s the sound of an otherworldly menace, organic but processed, sculpted into a conventional motif-driven sound world.
Zimmer remains a fixture in contemporary film music partly because, as music critic Jon Burlingame has pointed out, he has a relentless desire to search for fresh approaches to a film’s musical landscape. This pursuit begins with his extracting of sounds and colors from live performances and electronically engineering them during the scoring process. Such heightened attention to sound texture and color motivated the creation of the Spitfire percussion library, but can only hint at the experimentation and improvisational nature that goes into Zimmer’s work. In each of his film scores, the music tells a story that is tailored to the demands of the narrative, but the sounds reveal Zimmer’s urge to manipulate sound samples until they are, in his own words, “polished like a diamond.”
Ben Wright holds a Provost Postdoctoral Fellowship from the University of Southern California in the School of Cinematic Arts. In 2011, he received his Ph.D. in Cultural Studies from the Institute for Comparative Studies in Literature, Art and Culture at Carleton University. His research focuses on the study of production cultures, especially exploring the industrial, social, and technological effects of labor structures within the American film industry. His work on production culture, film sound and music, and screen comedy has appeared in numerous journals and anthologies. He is currently completing a manuscript on the history of contemporary sound production, titled Hearing Hollywood: Art, Industry, and Labor in Hollywood Film Sound.
All images creative commons.
REWIND! . . .If you liked this post, you may also dig:
Welcome to World Listening Month 2014, our annual forum on listening in observation of World Listening Day on July 18th, 2014. World Listening Day is a time to think about the impacts we have on our auditory environments and, in turn, its affects on us [for the full deets, peep our recent SO! Amplifies post by Eric Leonardson, Executive Director of the World Listening Project]. We kick off our month of thinking critically about listening with a post by media historian Brian Hanrahan, who listens deeply to sonic traces of the past to prompt us to question our desires for contemporary media representations of “reality.” It also marks the global 100 year anniversary of World War I this August 2014: a moment of silence. –J. Stoever, Editor-in-Chief
For some reason that I don’t fully understand, I am very emotionally moved by the space around a sound. I almost think that sometimes I am recording space with a sound in it, rather than sound in a space. -Walter Murch
If you want to listen to the past, there’s never been a time like the present. Every year, it seems, new old recordings are identified, new techniques developed to recover sounds thought irrecoverable. Here is Bismarck’s voice, preserved on a cylinder in 1889. Here, older still, is Edison’s. There is the astonishing recuperation of phonautograms – reverberation traced onto soot-blackened paper in the mid-nineteenth century, digitally processed and played back in our own. But as that processing underlines, no sound recording straightforwardly reproduces the real. An acoustic artifact is a compound of materiality, form and meaning, but also a place where technology meets desire. Old recordings meet the listener’s longing halfway; they invoke a reality always out of reach. And not simply a longing to hear, but also to touch, and be moved by, the fact of an absent existence.
Take, for instance, HMV 09308. In October 1918, just before the end of the Great War, William Gaisberg, a sound recordist of the pre-electric era, took recording equipment to the Western Front in order to capture the sound of British artillery shelling German lines with poison gas. Gaisberg died not long after, probably from Spanish flu, although some say he was weakened by gas exposure during the recording. Nonetheless the “Gas Shell Bombardment” record – a 12-inch HMV shellac disc, just over 2 minutes at 78 rpm – was released a few weeks later, just as the war came to an end. Initially intended to promote War Bonds, ultimately the record was used to raise money for disabled veterans.
For decades, the HMV recording had a reputation as one of the very earliest “actuality” recordings – one documenting a real location and event beyond the performative space of the studio, imprinted with the audible material trace of an actual moment in space and time. Documents like this – no matter what the technology – usually come with additional symbolic authentication. Here, the record’s label does some of that work. This “historic recording,” says the subtitle, is an “actual record taken on the front line.” Publicity pieces drove home the message. In the popular HMV magazine The Voice, Gaisberg – or probably his posthumous ghost-writer – described the expedition in detail, claiming the track to be a “true representation of the bombardment.”
In the same issue, a Major C.J.C. Street compared the recording to his own experience on the Front. “Its realism,” he wrote, ”took my breath away… I played the record many times… finding at each attempt some well-remembered detail.” He didn’t say so in his article, but Street – an artillery officer, a novelist and a propaganda man for the intelligence agency MI7 – was in fact the impresario of the record. This was not the first time he had found astute uses for sound media. The previous year he had put together a record that set artillery drill commands to popular tunes – the recording was both a propaganda release and an army training tool for new recruits. With the Gas Shell record, Street knew he wasn’t just selling recorded sound, but also an auratic sense of closeness to an overwhelming reality, the palpable proximity of war and death. Authenticating detail helped to underpin this sense of an absent real made present. Street cued the listener for those “well-remembered details.” In particular, he singled out one indistinct rattly flap-whizz noise, hearing in it, he claimed, the sound of a round with a “loose driving-band.”
The record stayed in the HMV catalog until 1945, but only in the early 1990s were its production history and authenticity claims seriously examined. In specialist journals, archivists, collectors and amateur historians undertook a collective forensic and critical analysis. A promising auditory witness was located: 95-year-old Lt.-Col. Montagu Cleeve another former artillery officer, in his time a developer of “Boche Buster” railway gun, later a music professor – was invited to critically assess the recording. Cleeve vouched unreservedly for its authenticity. He heard in it, he said, an unmistakable succession of sounds – the clang of the breech, the gigantic report of the firing explosion, the distinctive whiny whistle of a gas shell on its way across no-man’s-land. Others looked to data rather than the memories of old soldiers. One expert on pre-electric recording noted the angles commanded in firing instructions, correlated them with known muzzle velocities for 4.5 and 6-inch howitzers, then used this and other information to “definitively” explain the counter-intuitive anti-Doppler sound of the shells’ whistling. He also identified the audible echo effect – the curious “double report” of the guns heard here – as the sound of a brass recording horn violently resonating at a distance of exactly 26.5 meters from the guns.
Eventually, skepticism won out. Close listening at slow speeds – just careful attention and notation, nothing more elaborate – revealed inconsistencies and oddities in the firing noises. The bongs, plops and whistles seemed internally inconsistent. Some of the artillery sounds – ostensibly a battery of four, firing in quick succession – varied implausibly with each successive firing. Physical evidence from the record’s groove, as well as extraneous noises – surface crackle and fizz, and, audible within the recording, the swish of a turntable – seemed to indicate at least two rudimentary overdubs, in which the output of one acoustic horn was relayed into a second, possibly using an auxetophone, an early compressed-air amplifier. All this resulted in a double- or triple-layered sonic artifact. Finally – the crucial evidence, although oddly it was hardly noticed at the time – an alternative take was located. In this take, according to its discoverer, the entire theatrics of gunnery command is simply absent, and there is no sound at all of whistling shells in motion. What was left was a skeleton sequence of clicks, thuds and cracks, supplemented with only a single closing insert, the portentous injunction “Feed the Guns with War Bonds!”
In short, it seems highly likely that any original field recording was, at the very least, post-dramatized with performed voices and percussive and whistling sound effects. So, it is tempting to say, that clears that up. The recording’s inauthenticity is proven. File under Fake. But in fact, if we don’t stop there, if we set aside narrow and absolutist ideas of authenticity, and instead explore the recording’s ambiguity and hybridity, then Gas Shell Bombardment becomes all the more interesting as an historical artifact.
Let’s assume, for the sake of argument, that some form of basic recording was done in France, very possibly a staged barrage specifically performed for Gaisberg’s visit, and that this recording then had effects added back at HMV in London. The record might then be seen less as a straightforward documentary, and instead as an unusual version of the “descriptive speciality,” a genre of miniature phonographic vignette dating back to the 1890s, far predating longer-form radio drama. Very little is known about these early media artworks, but it is a fair generalization to say that in America the genre was more slanted towards vaudeville comedy, whereas in Europe, imperial and military scenes predominated. As early as 1890, for example, there had been German phonographic representations of battles from the Franco-Prussian war. The Great War saw a flourishing of the genre. Scholars are just beginning to take an interest these old phonographs; here’s one recent essay on the “Angel of Mons,” for example, a British acoustic vignette of a famous incident on the Western Front.
Listen to a 1915 German descriptive speciality, depicting the attack on the fortress of Liège the previous year:
As a descriptive speciality, Gas Shell Bombardment is unusual because it incorporates an actual indexical trace. But such traces – as emphasized by Charles Sanders Pierce and many later media-theoreticians– do not resemble their referent, they are caused by it. The bullet hole does not look much like a bullet; thunder is lightning’s trace, not its likeness. But for Street and Gaisberg, the trace’s lack of resemblance caused problems: the original recording’s lack of detail, cues and clues, but above all its lack of internal dimensionality, created a perceptual shortfall and a lack of credibility. Maybe they hoped that the guns, by sheer force of amplitude, would overcome the spatially impoverished, reverbless reproduction of pre-electric recording. If so, it didn’t work. Without added effects, the guns’ trace was as flat and “body-less” as a sequence of Morse. It was a sound without a scene. The producers’ interventions aimed to thicken the primary artifact with referential-sounding detail, but also to heighten the sense of materiality and spatiality, and to strengthen the sense of diegetic presence, of worlded thereness. The soldiers’ voices – louder and quieter, close-up and farther-out – and the fake-Doppler of the “shell whistling” lent the recording narrative direction (literally, some trajectory) and “authenticating” points of detail. But above all they gave a sense of internal space to the recording, a space into which the listener could direct her attention.
In this context, we can only admire the creativity and performative élan of the unknown production crew. We know little about effects production in early phonography. It is a safe bet that some techniques were adopted from theatre, and that there was overlap with silent film accompaniment. But whatever the method used, it would have called for the awkward orchestration of a limited number of iconic sounds to create an impression of a spatially coherent and materially detailed sonic environment. The recordist and his team would first have had to imagine how relative loudness – of voices, of material objects struck and sounded – might create a sense of spatial depth when transduced through the horn’s crude interface. Then they would have had to perform this as a live overdub, keeping time with the base track of the gun recording played through another horn. And all this done with participants and equipment crowded tightly around the mouth of the huge horn, crammed into the tiny pick-up arc, a scene looking something like this image of Leopold Stokowski’s pre-electric recording sessions or this photograph of the recording of a cello concerto.
As well as this hybrid of trace and live performance, there is another performance here – Gaisberg’s journey itself. With twenty years of recording experience, Gaisberg was probably very well aware that the expedition would not yield a “realistic” recording of the guns. But the expedition had to be made, so that it could be said to have taken place. Expectations had to be primed and colored, so that, to use André Bazin’s famous phrase about photographs, the recording could partake in an “irrational power to… bear the belief” of the listener. The journey, and the accounts of Gaisberg and Street are not a supplement to the “true representation” of the gas bombardment. They are part of that representation. Moreover, in subsequent writing it is noticeable that the manner of Gaisberg’s death becomes a rhetorical amplification for the authenticity of the recording’s trace, as if his fatal inhalation (of gas molecules or flu bacilli) were itself a deadly indexation, paralleling the recording’s claim to capture the breath of the War, and even of History itself.
In media-historical terms, the Gas Shell Bombardment recording can be understood as a late, transitional artifact from phonography’s pre-microphonic era. The desire for the sonic trace, for an ever more immersive proximity to events was there, but electro-acoustic technology was not yet in place. Two years later, in 1920, Horace Merriman and Lionel Guest made the first experimental electrical recording, arguably also the first true field recording. The event, appropriately enough, was an official war memorial service in London, where Merriman and Guest – working for Columbia Records – put microphones in Westminster Abbey, running cables to a remote recording van parked in the street outside, where they sat amidst heating ovens and cutting lathes. By the end of the 1920s, remote recording and broadcasting, while never straightforward, were well on the way to ubiquity.
Claims made on behalf of technologies of reproduction may seem simplistic, but there’s a grain of truth to their simplicity. If there were nothing special – even magical – in the referentiality of the camera that captures the moment, the recording that’s like being there, the liveness of the live broadcast, these things would not play the role they do in everyday life and in the ideological fabric of society. But there is falsehood too, in over-simplifying the nature and affective charge of old photographs, old footage, old recordings. These are made things, composed of different materials, media, signs and conventions; they are inseparable from the desires and expectations they induce and direct. They function in part by mimesis and verisimilitude, but also through the gaps, blank spots and false illusions of their trace. They can – rightly – intensify our feeling towards the past, but should also prompt us to think about our own desires and investments.
Image by Flickr User DrakeGoodman, “Horchposten im Spengtrichter vor Neuve-Chapelle 6km nördlich von La Bassée Nordfrankreich 1916,” A trio of lightly equipped soldiers from an unidentified formation oblige the photographer by looking serious and pretending they’re just metres from the enemy, listening for activity in his lines. The improvised “listening device” is actually a large funnel, probably liberated from a nearby farm.
Brían Hanrahan is a film, media and cultural historian, whose work focuses on the history of acoustic media, German and European cinema and the culture of the Weimar Republic.
Edited post-publication at 8:00 pm EST on July 7, 2014
REWIND!…If you liked this post, you may also dig:
A Brief History of Auto-Tune–Owen Marshall
DIY Histories: Podcasting the Past–Andrew Salvati