In the radio dramatization of Return of the Jedi (1996), a hibernation sickness-blinded Han Solo can tell bounty hunter Boba Fett is in the same room with him just by smelling him. Later this month, Solo: A Star Wars Story (part of the Anthology films, and as you might expect from the title, a prequel to Han Solo’s first appearance in Star Wars: A New Hope) may be able to shed some light on how Han developed this particular skill.
Later in that dramatization, we have to presume Han is able to accurately shoot a blaster blind by hearing alone. Appropriately, then, sound is integral to Star Wars. For every iconic image in the franchise—from R2D2 to Chewbacca to Darth Vader to X-Wing and TIE-fighters to the Millennium Falcon and the light sabers—there is a correspondingly iconic sound. In musical terms, too, the franchise is exemplary. John Williams, Star Wars’ composer, won the most awards of his career for his Star Wars (1977) score, including an Oscar, a Golden Globe, a BAFTA, and three Grammys. Not to mention Star Wars’ equally iconic diegetic music, such as the Mos Eisley Cantina band (officially known as Figrin D’an and the Modal Nodes).
Without sound, there would be no Star Wars. How else could Charles Ross’ One Man Star Wars Trilogy function? In One Man Star Wars, Ross performs all the voices, music, and sound effects himself. He needs no quick costume changes; indeed, in his rapid-fire, verbatim treatment, it is sound (along with a few gestures) that he uses to distinguish between characters. His one-man show, in fact, echoes C-3PO’s performance of Star Wars to the Ewoks in Return of the Jedi, a story told in narration and sound effects far more than in any visuals. “Translate the words, tell the story,” says Luke in the radio dramatization of this scene. That is what sound does in Star Wars.
I believe that the general viewing public is aware on a subconscious level of Star Wars’ impressive sound achievements, even if this is not always articulated as such. As Rick Altman noted in 1992 in his four and a half film fallacies, the ontological fallacy of film—while not unchallenged—began life with André Bazin’s “The Ontology of the Photographic Image,” (1960) which argues that film cannot exist without image. Challenging such an argument not only elevates silent film but also the discipline of film sound generally, so often regarded as an afterthought. “In virtually all film schools,” Randy Thom wrote in 1999, “sound is taught as if it were simply a tedious and mystifying series of technical operations, a necessary evil on the way to doing the fun stuff.”
Film critic Pauline Kael wrote about Star Wars on original release in what Gianlucca Sergi terms a “harmful generalization” that its defining characteristic was its “loudness.” Loud sound does not necessarily equal good sound in the movies, which audiences themselves can sometimes confuse. “High fidelity recordings of gunshots and explosions, and well fabricated alien creature vocalizations” do not equal good sound design alone, as Thom has argued. On the contrary, Star Wars’ achievements, Sergi posited, married technological invention with overall sound concept and refined if not defined the work of sound technicians and sound-conscious directors.
The reason why Star Wars is so successful aurally is because its creator, George Lucas, was invested in sound holistically and cohesively, a commitment that has carried through nearly every iteration of the franchise, and because his original sound designer, Ben Burtt, understood there was an art as well as a science to highly original, aurally “sticky” sounds. Ontologically, then, Star Wars is a sound-based story, as reflected in the existence of the radio dramatizations (more on them later). This article traces the historical development of sound in not only the Star Wars films (four decades of them!) but also in other associated media, such as television and video games as well as examining aspects of Star Wars’ holistic sound design in detail.
A long time ago, in a galaxy far, far away . . .
As Chris Taylor points out, George Lucas “loved cool sounds and sweeping music and the babble of dialogue more than he cared for dialogue itself.” In 1974, Lucas was working on The Radioland Murders, a screwball comedy thriller set in the fictional 1930s radio station WKGL. Radio, indeed, had already made a strong impression on Lucas, such that legendary “Border blaster” DJ Wolfman Jack played an integral part in Lucas’ film American Graffiti (1973). As Marcus Hearn picks up the story, Lucas soon realized that The Radioland Murders were going nowhere (the film would eventually be made in 1994). Lucas then turned his sound-conscious sensibilities in a different direction, in “The Star Wars” project upon which he had been ruminating since his film school days at the University of Southern California. Retaining creative control, and a holistic interest in a defined soundworld, were two aspects Lucas insisted upon during the development of the project that would become Star Wars. Lucas had worked with his contemporary at USC, sound designer and recordist Walter Murch, on THX 1138 (1971) and American Graffiti, and Murch would go on to provide legendary sound work for The Conversation (1974), The Godfather Part II (1974), and Apocalypse Now (1979). Murch was unavailable for the new project, so Lucas then asked producer Gary Kurtz to visit USC to evaluate emerging talent.
Pursuing a Masters degree in Film Production at USC was Ben Burtt, whose BA was in physics. In Burtt, Lucas found a truly innovative approach to film sound which was the genesis of Star Wars’ sonic invention, providing, in Sergi’s words, “audiences with a new array of aural pleasures.” Sound is embodied in the narrative of Star Wars. Not only was Burtt innovative in his meticulous attention to “found sounds” (whereas sound composition for science fiction films has previously relied on electronic sounds), he applied his meticulousness in character terms. Burtt said that Lucas and Kurtz, “just gave me a Nagra recorder and I worked out of my apartment near USC for a year, just going out and collecting sound that might be useful.”
Inherent in this was Burtt’s relationship with sound, in the way he was able to construct a sound of an imaginary object from a visual reference, such as the light saber, described in Lucas’ script and also in concept illustrations by Ralph McQuarrie. “I could kind of hear the sound in my head of the lightsabers even though it was just a painting of a lightsaber,” he said. “I could really just sort of hear the sound maybe somewhere in my subconscious I had seen a lightsaber before.” Burtt also shared with Lucas a sonic memory of sound from the Golden Age of Radio: “I said, `All my life I’ve wanted to see, let alone work on, a film like this.’ I loved Flash Gordon and other serials, and westerns. I immediately saw the potential of what they wanted to do.”
But sir, nobody worries about upsetting a droid
Burtt has described the story of A New Hope as being told from the point of view of the droids (the robots). While Lucas was inspired by Kurosawa’s The Hidden Fortress (1958) to create the characters of droids R2-D2 (“Artoo”) and C-3PO (“Threepio”), the robots are patently non-human characters. Yet, it was essential to imbue them with personalities. There have been cinematic robots since Maria, but Burtt uniquely used sound to convey not only these two robots’ personalities, but many others as well. As Jeanne Cavelos argues, “Hearing plays a critical role in the functioning of both Threepio and Artoo. They must understand the orders of their human owners.” Previous robots had less personality in their voices; for example, Douglas Rain, the voice of HAL in 2001: A Space Odyssey, spoke each word crisply with pauses. Threepio is a communications expert, with a human-like voice, provided by British actor (and BBC Radio Drama Repertory Company graduate) Anthony Daniels. According to Hearn, Burtt felt Daniels should use his own voice, but Lucas was unsure, wanting an American used car salesman voice. Burtt prevailed, creating in Threepio, vocally, “a highly strung, rather neurotic character,” in Daniels’ words, “so I decided to speak in a higher register, at the top of the lungs.” (Indeed, in the Diné translation of Star Wars [see below], Threepio was voiced by a woman, Geri Hongeva-Camarillo, something that the audience seemed to find hilarious.)
Artoo was altogether a more challenging proposition. As Cavelos puts it, “Artoo, even without the ability to speak English, manages to convey a clear personality himself, and to express a range of emotions.” Artoo’s non-speech sounds still convey emotional content. We know when Artoo is frightened;
when he is curious and friendly;
and when he is being insulting.
we started making little vocal sounds between each other to get a feeling for it. And it dawned on us that the sounds we were making were not actually so bad. Out of that discussion came the idea that the sounds a baby makes as it learns to walk would be a direction to go; a baby doesn’t form any words, but it can communicate with sounds.
The approach to Artoo’s aural communications became emblematic of all of the sounds made by machines in Star Wars, creating a non-verbal language, as Kris Jacobs calls it, the “exclusive province” of the Star Wars universe.
Powers of observation lie with the mind, Luke, not the eyes
According to Gianlucca Sergi, the film soundtrack is composed of sound effects, music, dialogue, and silence, all of which work together with great precision in Star Wars, to a highly memorable degree. Hayden Christensen, who played Anakin Skywalker in Attack of the Clones (2002) and Revenge of the Sith (2005), noted that when filming light saber battles with Ewan McGregor (Obi-Wan Kenobi), he could not resist vocally making the sound effects associated with these weapons.
This a good illustration of how iconic the sound effects of Star Wars have become. As Burtt noted above, he was stimulated by visuals to create the sound effects of the light sabers, though he was also inspired by the motor on a projector in the Department of Cinema at USC. As Todd Longwell pointed out in Variety, the projector hum was combined with a microphone passed in front of an old TV to create the sound. (It’s worth noting that the sounds of weapons were some of the first sound effects created in aural media, as in the case with Wallenstein, the first drama on German radio, in 1924, which featured clanging swords.)
If Burtt gave personality to robots through their aural communications, he created an innovative sound palette for far more than the light sabers in Star Wars. In modifying and layering found sounds to create sounds corresponding to every aspect of the film world—from laser blasts (the sound of a hammer on an antenna tower guy wire) to the Imperial Walkers from Empire Strikes Back (modifying the sound of a machinist’s punch press combined with the sounds of bicycle chains being dropped on concrete)—he worked as meticulously as a (visual) designer to establish cohesion and impact.
Sergi argues that the sound effects in Star Wars can give subtle clues about the objects with which they are associated. The sound of Imperial TIE fighters, which “roar” as they hurtle through space, was made from elephant bellows, and the deep and rumbling sound made by the Death Star is achieved through active use of sub-frequencies. Meanwhile, “the rebel X-wing and Y-wing fighters attacking the Death Star, though small, emit a wider range of frequencies, ranging from the high to the low (piloted as they are by men of different ages and experience).” One could argue that even here, Burtt has matched personality to machine. The varied sounds of the Millennium Falcon (jumping into hyperspace, hyperdrive malfunction), created by Burtt by processing sounds made by existing airplanes (along with some groaning water pipes and a dentist’s drill), give it, in the words of Sergi, a much more “grown-up” sound than Luke’s X-Wing fighter or Princess Leia’s ship, the Tantive IV. Given that, like its pilot Han Solo, the Falcon is weathered and experienced, and Luke and Leia are comparatively young and ingenuous, this sonic shorthand makes sense.
Millions of voices
Michel Chion argues that film has tended to be verbocentric, that is, that film soundtracks are produced around the assumption that dialogue, and indeed the sense of the dialogue rather than the sound, should be paramount and most easily heard by viewers. Star Wars contradicts this convention in many ways, beginning with the way it uses non-English communication forms, not only the droid languages discussed above but also its plethora of languages for various denizens of the galaxy. For example, Cavelos points out that Wookiees “have rather inexpressive faces yet reveal emotion through voice and body language.”
While the 1978 Star Wars Holiday Special may have many sins laid at its door, among them must surely be that the only Wookiee who actually sounds like a Wookiee is Chewbacca. His putative family sound more like tauntauns. Such a small detail can be quite jarring in a universe as sonically invested as Star Wars.
While many of the lines in Star Wars are eminently quotable, the vocal performances have perhaps received less attention than they deserve. As Starr A. Marcello notes, vocal performance can be extremely powerful, capitalizing on the “unique timbre and materiality that belong to a particular voice.” For example, while Lucas originally wanted Japanese actor Toshiro Mifune to play Obi-Wan, Alec Guinness’ patrician Standard English Neutral accent clearly became an important part of the character. For example, when (Scottish) actor Ewan McGregor was cast to play the younger version of Obi-Wan, he began voice lessons to reproduce Guinness’ voice. Ian McDiarmid (also Scottish), a primarily a Shakespearean stage actor, was cast as arch-enemy the Emperor in Return of the Jedi, presumably on the quality of his vocal performance, and as such has portrayed the character in everything from Revenge of the Sith to Angry Birds Star Wars II.
Sergi argues that Harrison Ford as Han Solo performs in a lower pitch but an unstable meter, a characterization explored in the radio dramatizations of A New Hope, Empire Strikes Back, and Return of the Jedi, when Perry King stands in for Ford. By contrast, Mark Hamill voices Luke in two of the radio dramatizations, refining and intensifying his film performances. Sergi argues that Hamill’s voice emphasizes youth: staccato, interrupting/interrupted, high pitch.
And affectionately parodied here:
I would add warmth of tone to this list, perhaps illustrated nowhere better than in Hamill’s performance in episode 1 – “A Wind to Shake the Stars” of the radio dramatization, which depicts much of Luke’s story that never made it onscreen, from Luke’s interaction with his friends in Beggar’s Canyon to a zany remark to a droid (“I know you don’t know, you maniac!”). It will come as no surprise to the listeners of the radio dramatization that Hamill would find acclaim in voice work (receiving multiple nominations and awards). In the cinematic version, Hamill’s performance is perhaps most gripping during the climactic scene in Empire Strikes Back when Darth Vader tells him:
According to Hamill, “what he was hearing from Vader that day were the words, ‘You don’t know the truth: Obi-Wan killed your father.’ Vader’s real dialogue would be recorded in postproduction under conditions easier to control.” More on that (and Vader) shortly.
It has been noted that Carrie Fisher (who was only nineteen when A New Hope was filmed) uses an accent that wavers between Standard North American and Standard Neutral English. Fisher has explained this as her emulating experienced British star of stage and screen Peter Cushing (playing Grand Moff Tarkin).
However, the accents of Star Wars have remained a contentious if little commented upon topic, with most (if not all) Imperial staff from A New Hope onwards speaking Standard Neutral English (see the exception, stormtroopers, further on). In production terms, naturally, this has a simple explanation. In story terms, however, fans have advanced theories regarding the galactic center of the universe, with an allegorical impetus in the form of the American Revolution. George Lucas, after all, is an American, so the heroic Rebels here have echoes with American colonists throwing off British rule in the 18th century, inspired in part because of their geographical remove from centers of Imperial rule like London. Therefore, goes this argument, in Star Wars, worlds like Coruscant are peopled by those speaking Standard Neutral English, while those in the Outer Rim (the majority of our heroes) speak varieties of Standard North American. Star Wars thus both advances and reinforces the stereotype that the Brits are evil.
It is perhaps appropriate, then, that James Earl Jones’ performance as Darth Vader has been noted for sounding more British than American, though Sergi emphasizes musicality rather than accent, the vocal quality over verbocentricity:
The end product is a fascinating mixture of two opposite aspects: an extremely captivating, operatic quality (especially the melodic meter with which he delivers the lines) and an evil and cold means of destruction (achieved mainly through echoing and distancing the voice).
It is worth noting that Lucas originally wanted Orson Welles, perhaps the most famous radio voice of all time, to portray Vader, yet feared that Welles would be too recognizable. That a different voice needed to emanate from behind Vader’s mask than the actor playing his body was evident from British bodybuilder David Prowse’s “thick West Country brogue.” The effect is parodied in the substitution of a Cockney accent from Snatch (2000) for Jones’ majestic tones:
A Newsweek review of Jones in the 1967 play A Great White Hope argued that Jones had honed his craft through “Fourteen years of good hard acting work, including more Shakespeare than most British actors attempt.” Sergi has characterized Jones’ voice as the most famous in Hollywood, in part because in addition to his prolific theatre back catalogue, Jones took bit parts and voiced commercials—“commercials can be very exciting,” he noted. The two competing forces combined to create a memorable performance, though as others have noted, Jones is the African-American voice to the white actors who portrayed Anakin Skywalker (Clive Revill and Hayden Christensen), one British, one American.
Brock Peters, also African American and known for his deep voice, played Vader in the radio dramatizations. Jennifer Stoever notes that in America, the sonic color line “historically contoured, identified, and marked mismatches between ‘sounding white’ and ‘looking black’” (231) whereas the Vader performances “sound black” and “look white.” Andrew Howe in his chapter “Star Wars in Black and White” notes the “tension between black outer visage and white interior identity [ . . ] Blackness is thus constructed as a mask of evil that can be both acquired and discarded.”
Like many of the most important aspects of Star Wars, Vader’s sonic presence is multi-layered, consisting in part of Jones’ voices manipulated by Burtt, as well as the sonic indicator of his presence: his mechanized breathing”
The concept for the sound of Darth Vader came about from the first film, and the script described him as some kind of a strange dark being who is in some kind of life support system. That he was breathing strange, that maybe you heard the sounds of mechanics or motors, he might be part robot, he might be part human, we really didn’t know. [ . . .] He was almost like some robot in some sense and he made so much noise that we had to sort of cut back on that concept.
On radio, a character cannot be said to exist unless we hear from him or her; whether listening to the radio dramatizations or watching Star Wars with our eyes closed, we can always sense the presence of Vader by the sound of his breathing. As Kevin L. Ferguson points out, “Is it accidental, then, that cinematic villains, troubling in their behaviour, are also often troubled in their breathing?” As Kris Jacobs notes, “Darth Vader’s mechanized breathing can’t be written down”—it exists purely in a sonic state.
Your eyes can deceive you; don’t trust them
Music is the final element of Sergi’s list of what makes up the soundtrack, and John Williams’ enduring musical score is the most obvious of Star Wars’ sonic elements. Unlike “classical era” Hollywood film composers like Max Steiner or Erich Korngold who, according to Kathryn Kalinak, “entered the studio ranks with a fair amount of prestige and its attendant power, Williams entered as a contract musician working with ‘the then giants of the film industry,’” moving into a “late-romantic idiom” that has come to characterize his work. This coincided with what Lucas envisioned for Star Wars, influenced as it was by 1930s radio serial culture.
Williams’ emotionally-pitched music has many elements that Kalinak argues link him with the classical score model: unity, the use of music in the creation of mood and character; the privileging of music in moments of spectacle, the way music and dialogue are carefully mixed. This effect is exemplified in the opening of A New Hope, the “Main Title” or, as Dr Lehman has it (see below), “Main/Luke A.” As Sergi notes, “the musical score does not simply fade out to allow the effects in; it is, rather literally, blasted away by an explosion (the only sound clearly indicated in the screenplay).”
As Kalinak points out, it was common in the era of Steiner and Korngold to score music for roughly three-quarters of a film, whereas by the 1970s, it was more likely to be one-quarter. “Empire runs 127 minutes, and Williams initially marked 117 minutes of it for musical accompaniment”; while he used three themes from A New Hope, “the vast majority of music in The Empire Strikes Back was scored specifically for the film.”
Perhaps Williams’ most effective technique is the use of leitmotifs, derived from the work of Richard Wagner, and more complex than a simple repetition of themes. Within leitmotifs, we hear the blending of denotative and connotative associations, as Matthew Bribitzer-Stull notes, “not just a musical labelling of people and things” but also, as Thomas S. Grey puts it, “a matter of musical memory, of recalling things dimly remembered and seeing what sense we can make of them in a new context.” Bribitzer-Stull also notes the complexity of Williams’ leitmotif use, given that tonal music is given for both protagonists and antagonists, resisting the then-cliché of using atonal music for antagonists. In Williams’ score, atonal music is used for accompanying exotic landscapes and fight or action scenes. As Jonathan Broxton explains,
That’s how it works. It’s how the films maintain musical consistency, it’s how characters’ musical identities are established, and it offers the composer an opportunity to create interesting contrapuntal variations on existing ideas, when they are placed in new situations, or face off against new opponents.
Within the leitmotifs, Williams provides various variations and disruptions, such as the harmonic corruption when “the melody remains largely the same, but its harmonization becomes dissonant.” One of the most haunting ways in which Williams alters and reworks his leitmotifs is what Bribitzer-Stull calls “change of texture.”
Frank Lehman of Harvard has examined Williams’ leitmotifs in detail, cataloguing them based on a variety of meticulous criteria. He has noted, for example, that some leitmotifs are used often, like “Rebel Fanfare” which has been used in Revenge of the Sith, A New Hope, The Empire Strikes Back, The Force Awakens, The Last Jedi, and Rogue One. Lehman particularly admires Williams’ skill and restraint, though, in reserving particular leitmotifs for very special occasions. For example, “Luke & Leia,” first heard in Return of the Jedi (both film and radio dramatization) and not again until The Last Jedi:
While Williams’ use of leitmotifs is successful and evocative, not all of Star Wars’ music consists of leitmotifs, as Lehman points out; single, memorable pieces of music not heard elsewhere are still startlingly effective.
In the upcoming Solo, John Williams will contribute a new leitmotif for Han Solo, while all other material will be written and adapted by John Powell. Williams has said in interview that “I don’t make a particular distinction between ‘high art’ and ‘low art.’ Music is there for everybody. It’s a river we can all put our cups into, and drink it, and be sustained by it.” The sounds of Star Wars have sustained it—and us—and perfectly illustrate George Lucas’ investment in the equal power of sound to vision in the cinematic experience. I, for one, am looking forward to what new sonic gems may be unleashed as the saga continues.
On the first week of June, Leslie McMurtry will return with Episode II, focusing on shifts in sound in the newer films and multi-media forms of Star Wars, including radio and cartoons–and, if we are lucky, her take on Solo!
Featured Image made here: Enjoy!
Leslie McMurtry has a PhD in English (radio drama) and an MA in Creative and Media Writing from Swansea University. Her work on audio drama has been published in The Journal of Popular Culture, The Journal of American Studies in Turkey, and Rádio-Leituras. Her radio drama The Mesmerist was produced by Camino Real Productions in 2010, and she writes about audio drama at It’s Great to Be a Radio Maniac.
REWIND!…If you liked this post, you may also dig:
Speaking American–Leslie McMurtry
Out of Sync: Gendered Location Sound Work in Bollywood–Priya Jaikumar
Welcome to the third and final installment of Sculpting the Film Soundtrack, our series about sound in contemporary films. We’ve been focusing on how filmmakers are blurring the boundaries between music, speech, and sound effects – in effect, integrating distinct categories of soundtrack design.
In our first post, Benjamin Wright showed how celebrated composer Hans Zimmer thinks across traditional divisions of labour to integrate film sound design with music composition. Danijela Kulezic-Wilson followed up with an insightful piece on the integration of audio elements in Shane Carruth’s Upstream Color, suggesting how scholars can apply principles of music, like tempo and rhythm, to their analyses of the interactions between a film’s images and sounds. In this final entry, Randolph Jordan, considers another dimension of integration: a film’s sounds and the place where it was produced. In his provocative and insightful reading of the quasi-documentary East Hastings Pharmacy, Jordan, who is completing a post-doctoral post at Simon Fraser University, elaborates on how the concept of “unsettled listening” can clue us into the relationship between a film and its origins of production. You’ll be able to read more about “unsettled listening” in Jordan’s forthcoming book, tentatively titled Reflective Audioviewing: An Acoustic Ecology of the Cinema, to be published by Oxford University Press.
I hope you’ve enjoyed taking in this series as much as I’ve enjoyed editing it with the help of the marvelous folks at SO!. Thanks for reading. — Guest Editor Katherine Spring
A mother and son of First Nations ancestry sit in the waiting area of a methadone clinic in Vancouver’s Downtown Eastside, their attention directed toward an offscreen TV. A cartoon plays, featuring an instrumental version of “I’ve Been Working on the Railroad” that mingles with the operating sounds of the clinic and ambience from the street outside. The tune is punctuated by a metal clinking sound at the beginning of each bar, calling to mind the sound of driving railway spikes that once echoed just down the street as the City of Vancouver was incorporated as the western terminus of the Canadian Pacific Railway (beginning thus the cycle of state-sanctioned erasure of indigenous title to the land). The familiar voice of Bugs Bunny chimes in: “Uh, what’s all the hubbub, bub?”
Hubbub indeed. Let’s unpack it.
The scene appears one third of the way through East Hastings Pharmacy (Antoine Bourges, 2012), a quasi-documentary set entirely within this clinic, staging interactions between methadone patients (played by locals and informed by their real-life experiences) and the resident pharmacist (played by an actress). Vancouver’s Downtown Eastside, dubbed Canada’s “worst neighborhood” for its notorious concentration of transients and public drug use, is also home to the largest community of First Nations peoples within the city limits, a product of the long history of dispossession in the surrounding areas. When the film presents this indigenous pair listening to a Hollywood fabrication of the sounds that marked their loss of title to the city it is a potent juxtaposition, especially given the American infiltration of Vancouver’s mediascape since the 1970s. Long known as “Hollywood North,” Vancouver is more famous as a stand-in for myriad other parts of the world than for representing itself, its regional specificity endlessly overwritten with narratives that hide the city and its indigenous presence from public awareness.
In her essay “Thoughts on Making Places: Hollywood North and the Indigenous City,” filmmaker Kamala Todd stresses how media can assist the process of re-inscribing local stories into Vancouver’s consciousness. East Hastings Pharmacy is one such example, lending some screen time to urban Natives in the 21st Century city. But Todd reminds us that audiences also have a responsibility “to learn the stories of the land” that have been actively erased in dominant media practices, and to bring this knowledge to our experience of the city in all its incarnations (9). Todd’s call resonates with a process that Nicholas Blomley calls “unsettling the city” in his book of the same name. Blomley reveals Vancouver as a site of continual contestation and mobility across generations and cultural groups, and calls for an “unsettled” approach that can account for the multiple overlapping patterns of use that are concealed by “settled” concepts of bounded property. With that in mind, I propose “unsettled listening” as a way of experiencing the city from these multiple positions simultaneously. Rick Altman taught us to hear any given sound event as a narrative by listening for the auditory markers of its propagation through physical space, and recording media, over time (15-31). Unsettled listening invites us to hear through these physical properties of mediatic space to the resonating stories revealed by the overlapping and contradictory histories and patterns of use to which these spaces are put, all too often unacknowledged in the wake of settler colonialism.
East Hastings Pharmacy provides a great opportunity to begin the practice of unsettled listening. The film’s status as an independent production amidst industrial shooting is marked by the intersection of studio-fabricated sound effects and direct sound recording, as in the example described above, and further complicated by the film’s own hybrid of fiction and documentary modes. That speaks to the complexity of overlapping filmmaking practices in Vancouver today, a situation embedded within the intersecting claims to land use and cultural propriety on the streets of Vancouver’s Downtown Eastside. To unsettle listening is to hear all these overlapping situations as forms of resonance that begin with the original context of the televised cartoon and accumulate as they spread through the interior of the clinic and outwards across the surrounding land. So let’s try this out.
The cartoon is Falling Hare (Robert Clampett, 1943), a good example of the noted history of cross-departmental integration at The Warner Bros.’ cartoon studios. The scene in question begins at 1:55, and here the metallic clinking sound is just as likely to have been produced by one of music orchestrator Carl Stalling’s percussionists as by sound effects editor Treg Brown. This integration can be heard in the way that the music’s unspoken reference to railway construction charges each clink with the connotation of hammer on spike. However, the image track in Falling Hare doesn’t depict railway construction, but rather a gremlin whacking the nose of a live bomb in an attempt to do away with enemy Bugs seated on top. James Lastra would say (by way of Christian Metz) that the clinking sound is “legible” as hammer on spike for the ease with which the sound can be recognized as emanating from this implied source (126). But this legibility is premised upon a lack of specificity that also allows the sound to become interchangeable with something else, as is the case in this cartoon.
East Hastings Pharmacy capitalizes on this interchangeability by re-inscribing the clinking sound’s railway connotations, first by stripping the original image and then by presenting this sound in the context of the dire social realities of Vancouver’s Downtown Eastside as the city’s sanctioned corral for the markers of urban poverty – and indigeneity – that officials don’t want to spill out across the neighborhood’s increasingly gentrified perimeter.
As one of a string of Warner Bros. cartoons put in the service of WWII propaganda, the Falling Hare soundtrack also resonates with wartime xenophobia and imperialist expansion, branches of the same pathos that leads to the effacing of indigenous culture from the consciousness of colonizing peoples. In Vancouver, this has taken the form of what Jean Barman calls “Erasing Indigenous Indigeneity,” the process of chasing the area’s original peoples off the land while importing aboriginal artifacts from elsewhere to maintain a Native chic deemed safe for immigrant consumption (as when the city paid “homage” to the vacated Squamish residents of downtown Vancouver’s Stanley Park by erecting Kwakiutl totem poles imported from 200km north on Vancouver Island) (26). This is an interchangeability of cultural heritage premised upon a lack of specificity, the same quality that allows “legible” sound effects to become synchretic with a variety of implied sources. And this process is not unlike the interchangeability of urban spaces when shooting Vancouver for Seattle, New York, or Frankfurt, emphasizing generic qualities of globalized urbanization while suppressing recognizable soundmarks from the mix (such as the persistent sound of float plane propellers that populate Vancouver harbour, the grinding and screeching of trains in the downtown railyard, or the regular horn blasts from the local ferry runs just north of the city).
The high-concept legibility of Warner Bros.’ sound effects – used in Falling Hare to play on listener’s expectations to comic effect – is further unsettled by its presentation within the context of documentary sound conventions in East Hastings Pharmacy. Bourges’ film commits to regional specificity in part through the use of location sound recording, which, as Jeffrey K. Ruoff identifies in “Conventions of Documentary Sound,” is particularly valued as a marker of authenticity (27-29). While Bourges stages the action inside the clinic, the film features location recordings of the rich street life audible and visible through the clinic’s windows that proceeds unaffected by the cameras and microphones. This situation is all the more potent when we account for the fact that, in this scene, the location-recorded cartoon soundtrack and ambient sound effects were added in post-production, and so represent a highly conscious attempt to channel the acoustic environment according to the conventions of “authentic” sound in documentary film.
While the film uses location recording as a conscious stylistic choice to evoke documentary convention, it does so to engage meaningfully with the social situation in the Downtown Eastside, underlining Michel Chion’s point that “rendered” film sound – fabricated in studio to evoke the qualities of a particular space – is just as capable of engaging the world authentically (or inauthentically) as “real” sound captured on location (95-98). By presenting this Hollywood cartoon as an embedded element within the soundscape of the clinic, using a provocative mix of location sound and studio fabrication, East Hastings Pharmacy unsettles Hollywood’s usual practice of erasing local specificity, inviting us to think of runaway projects in the context of their foreign spaces of production and the local media practices that sit next to them.
Finally, this intersection of sonic styles points to the complex relationships that exist between the domains of independent and industrial production around Vancouver. In his book Hollywood North, Mike Gasher argues for thinking about filmmaking in British Columbia as a resource industry, pointing to how the provincial government has offered business incentives for foreign film production similar to those in place for activities like logging and fishing. Here we can consider how the local film industry might follow the same unsustainable patterns of extraction as other resource industries, all premised upon willful ignorance of indigenous uses of the land. Yet as David Spaner charts in Dreaming in the Rain, the ability to make independent films in Vancouver has become largely intertwined with the availability of industrial resources in town. Just as Hollywood didn’t erase the independent film, colonization didn’t erase indigenous presence.
East Hastings Pharmacy offers a powerful example of how we can practice unsettled listening on the staged sound of Falling Hare, devoid of local context and connected to the railway only by inference, to reveal a rich integration with regional specificity as the cartoon’s auditory resonances accumulate within its new spaces of propagation. In this way we can hear local media through its transnational network, including the First Nations, to understand the overlaps between seemingly contradictory modes of being within the city. And in so doing, we can also hear through the misrepresentation of the Downtown Eastside as “Canada’s worst neighborhood” to the strength of the community that has long characterized the area for anyone who scratches the surface, an important first step along the path to unsettling the city as a whole.
Featured Image: Still from East Hastings Pharmacy
Randolph Jordan wanted to be a rock star. Academia seemed a responsible back-up option – until it became clear that landing a professor gig would be harder than topping the Billboard charts. After completing his Ph.D. in the interdisciplinary Humanities program at Concordia University in 2010 he floated around Montreal classrooms on contract appointments before taking up a two-year postdoctoral research fellowship in the School of Communication at Simon Fraser University. There he has been investigating geographical specificity in Vancouver-based film and media by way of sound studies and critical geography, research that will inform the last chapter of his book Reflective Audioviewing: An Acoustic Ecology of the Cinema (now under contract at Oxford University Press). If you can’t find him hammering away at his manuscript, or recording his three young children hammering away at their Mason & Risch, look for him under Vancouver’s Burrard Bridge where he spends his “spare time” gathering film and sound material for his multimedia project Bell Tower of False Creek. Or visit him online here: http://www.randolphjordan.com
REWIND!…If you liked this post, you may also dig: