Welcome back to Hearing the UnHeard, Sounding Out‘s series on how the unheard world affects us, which started out with my post on the hearing ranges of animals, and now continues with this exciting piece by China Blue.
From recording the top of the Eiffel Tower to the depths of the rising waters around Venice, from building fields of robotic crickets in Tokyo to lofting 3D printed ears with binaural mics in a weather balloon, China Blue is as much an acoustic explorer as a sound artist. While she makes her works publicly accessible, shown in museums and galleries around the world, she searches for inspiration in acoustically inaccessible sources, sometimes turning sensory possibilities on their head and sonifying the visual or reformatting sounds to make the inaudible audible.
In this installment of Hearing the UnHeard, China Blue talks about cataclysmic sounds we might not survive hearing and her experiences recording simulated asteroid strikes at NASA’s Ames Vertical Gun Range.
– Guest Editor Seth Horowitz
Fundamentally speaking, sound is the result of something banging into something else. And since everything in the universe, from the slow recombination of chemicals to the hypervelocity impacts of asteroids smashing into planet surfaces, is ultimately the result of things banging into things, the entire universe has a sonic signature. But because of the huge difference in scale of these collisions, some things remain unheard without very specialized equipment. And others, you hope you never hear.
Unheard sounds can be hidden subtly beneath your feet like the microsounds of ants walking, or they can be unexpectedly harmonic like the seismic vibrations of a huge structure like the Eiffel Tower. These are sounds that we can explore safely, using audio editing tools to integrate them into new musical or artistic pieces.
Luckily, our experience with truly primal sounds, such as the explosive shock waves of asteroid impacts that shaped most of our solar system (including the Earth) is rarer. Those who have been near a small example of such an event, such as the residents of Chelyabinsk, Russia in 2013 were probably less interested in the sonic event and more interested in surviving the experience.
But there remains something seductive about being able to hear sounds such as the cosmic rain of fire and ice that shaped our planet billions of years ago. A few years ago, when I became fascinated with sounds “bigger” than humans normally hear, I was able to record simulations of these impacts in one of the few places on Earth where you can, at NASA Ames Vertical Gun Range.
The Vertical Gun at Ames Research Center (AVGR) was designed to conduct scientific studies of lunar impacts. It consists of a 25 foot long gun barrel with a powder chamber at one end and a target chamber, painted bright blue, that looks like the nose of an upended submarine, about 8 feet in diameter and height at the other. The walls of the chamber are of thick steel strong enough to let its interior be pumped down to vacuum levels close to that of outer space, or back-filled with various gases to simulate different planetary atmospheres. Using hydrogen and/or up to half a pound of gun powder, the AVGR can launch projectiles at astonishing speeds of 500 to 7,000 m/s (1,100 to 16,000 mph). By varying the gun’s angle of elevation, projectiles can be shot into the target so that it simulates impacts from overhead or at skimming angles.
In other words, it’s a safe way to create cataclysmic impacts, and then analyze them using million frame-per-second video cameras without leaving the security of Earth.
My husband, Dr. Seth Horowitz who is an auditory neuroscientist and another devotee of sound, is close friends with one of the principal investigators of the Ames Vertical Gun, Professor Peter Schultz. Schultz is well known for his 2005 project to blow a hole in the comet Tempel 1 to analyze its composition, and for his involvement in the LCROSS mission that smashed into the south pole of the moon to look for evidence of water. During one conversation discussing the various analytical techniques they use to understand impacts, I asked, “I wonder what it sounds like.” As sound is the propagation of energy by matter banging into other matter, this seemed like the ultimate opportunity to record a “Big Bang” that wouldn’t actually get you killed by flying meteorite shards. Thankfully, my husband and I were invited to come to Ames to find out.
I had a feeling that the AVGR would produce fascinating new sounds that might provide us with different insights into impacts than the more common visual techniques. Because this was completely new research, we used a number of different microphones that were sensitive to different ranges and types of sound and vibrations to provide us with a selection of recording results. As an artist I found the research to be the dominant part of the work because the processes of capturing and analyzing the sounds were a feat unto themselves. As we prepared for the experiment, I thought about what I could do with these sounds. When I eventually create a work out of them, I anticipate using them in an installation that would trigger impact sounds when people enter the room, but I have not yet mounted this work since I suspect that this would be too frightening for most exhibition spaces to want.
Part of my love (and frustration) for sound work is figuring out how to best capture that fleeting moment in which the sound is just right, when the sound evokes a complex response from its listeners without having to even be explained. The sound of Mach 10 impacts and its effects on the environment had such possibilities. In pursuit of the “just right,” we wired up the gun and chamber with multiple calibrated acoustic and seismic microphones, then fed them into a single high speed multichannel recorder, pressed “record” and made for the “safe” room while the Big Red Button was pressed, launching the first impactor. We recorded throughout the day, changing the chamber’s conditions from vacuum to atmosphere.
When we finally got to listen that afternoon, we heard things we never imagined. Initial shots in vacuum were surprisingly dull. The seismic microphones picked up the “thump” of the projectile hitting the sand target and a few pattering sounds as secondary particles struck the surfaces. There were of course no sounds from the boundary or ultrasonic mics due to the lack of air to propagate sound waves. While they were scientifically useful–they demonstrated that we could identify specific impact events launched from the target—they weren’t very acoustically dramatic.
When a little atmosphere was added, however, we began picking up subtle sounds, such as the impact and early spray of particles from the boundary mic and the fact that there was an air leak from the pitch shifted ultrasonic mic. But when the chamber was filled with an earthlike atmosphere and the target dish filled with tiny toothpicks to simulate trees, building the scenario for a tiny Tunguska event (a 1908 explosion of an interstellar object in Russia, the largest in recorded history), the sound was stunning:
After the initial explosion, there was a sandstorm as the particles of sand from the target flew about at Mach 5 (destroying one of the microphones in the process), and giving us a simulation of a major asteroid explosion.
66 million years ago, in a swampy area by the Yucatan Penninsula, something like this probably occurred, when a six mile wide rock burned through the atmosphere to strike the water, ending the 135 million year reign of the dinosaurs. Perhaps it sounded a little like this simulation:
Any living thing that heard this – dinausaurs, birds, frogs insects – is long gone. By thinking about the event through new sounds, however, we can not only create new ways to analyze natural phenomena, but also extend the boundaries of our ability to listen across time and space and imagine what the sound of that impact might have been like, from an infrasonic rumble to a killing concussion.
It would probably terrify any listener to walk in to an art exhibition space filled with simply the sounds of simulated hypervelocity impacts, replete with loud, low frequency sounds and infrasonic vibrations. But there is something to that terror. Such sounds trigger ancient evolutionary pathways which are still with us because they were so good at helping us survive similar events by making us run, putting as much distance between us and the cataclysmic source, something that lingers even in safe reproductions, resynthesized from controlled, captured sources.
China Blue is a two time NASA/RI Space Grant recipient and an internationally exhibiting artist who was the first person to record the Eiffel Tower in Paris, France and NASA’s Vertical Gun. Her acoustic work has led her to be selected as the US representative at OPEN XI, Venice, Italy and at the Tokyo Experimental Art Festival in Tokyo, Japan, and was the featured artist for the 2006 annual meeting of the Acoustic Society of America. Reviews of her work have been published in the Wall Street Journal, New York Times, Art in America, Art Forum, artCritical and NY Arts, to name a few. She has been an invited speaker at Harvard, Yale, MIT, Berkelee School of Music, Reed College and Brown University. She is the Founder and Executive Director of The Engine Institute www.theengineinstitute.org.
Featured Image of a high-speed impact recorded by AVGR. Image by P. H. Schultz. Via Wikimedia Commons.
REWIND! If you liked this post, check out …
Learning to Listen Beyond Our Ears– Owen Marshall
Living with Noise– Osvaldo Oyola
There’s a 20-year gap in chronology between KMFDM’s 1993 song “A Drug Against War” and Marc Forster’s 2013 film World War Z, but sonically and ideologically they’re very, very similar. They contain the same kinds of sounds–machine guns, military orders barked over radios, buzzing crowds–and they use these sounds in the same way: to build sonic intensity past its breaking point (a “sonic bombardment brighter than sunlight,” as the KFMDM lyrics say). Their sonic similarity is evidence of neoliberalism’s intensification in the 20 years between them: what was once avant-garde opposition is later mainstream norm.
The songs’ sonic similarity reveals the central role of sound in contemporary biopolitics. By listening closely to “A Drug Against War” and the soundscape of World War Z—a film in which Brad Pitt saves humanity from a zombie apocalypse by giving all survivors a terminal disease—I show sound as more than a privileged aesthetic domain; sound actually provides the epistemic background and the concrete mechanisms for organizing society. Just as vision and “the gaze” are the ideological and technological foundation of panopticism, sound is the ideological and technological foundation of contemporary biopolitics. Much more is at stake in this post than just a song and a film: it takes on how—and why—society is organized as it is. It’s also about a particular understanding of “the sonic”: sound as dynamic patterning.
Because “A Drug Against War” lays out, in fairly elementary form, this “biopolitical” sonic vocabulary, it makes sense to start there. But before I do that, I will briefly define what I understand as ‘biopolitics.’
Like “neoliberalism,” “biopolitics” is a trendy concept whose precise meaning can get lost in loose usage. By “biopolitics,” I mean both an ideology of health and vitality and a political strategy whose medium is “life.” “Life,” here, isn’t individual health, wellness, or existence; it’s the ongoing vitality of the segment of society that counts as “society” tout court (e.g., in white supremacy, that segment would be whites). Biopolitics manages society like a living thing; for example, we often talk about the “health” of the economy, or use metrics such as obesity rates to compare different countries.
As Foucault explains in Society Must Be Defended, biopolitics’ “basic function is to improve life, to prolong its duration, to improve its chances, to avoid accidents, and to compensate for failings” (254). But to do that, power sometimes has to kill. Pruning my raspberry bush causes more berries to sprout, for example, just as weightlifting tears all my muscle fibers so they’ll rebuild in bigger, stronger shape. Killing off the weak is a positive investment in society’s overall strength. Again, Foucault:
The fact that the other dies does not mean simply that I live in the sense that his death guarantees my safety; the death of the other, the death of the bad race, of the inferior race (or the degenerate, or the abnormal) is something that will make life in general healthier: healthier and purer. (Society, 255).
Hitler’s “final solution” is an obvious example of this biopolitical approach to killing, but this practice also informs many contemporary US policies and practices. In the US, black people as a population have significantly higher mortality rates than any other race, for example. Following Foucault, we could say it’s in the interest of white supremacist society to maintain a high mortality rate among black populations because this makes white supremacist society “healthier.” The key point here is this: biopolitics promotes and administers life by generalizing and naturalizing what Foucault calls “the relationship of war: ‘In order to live, you must destroy your enemies’” (Society 256). Biopolitical warfare is precisely what is waged in both “A Drug Against War” and World War Z, and sound emerges as a weapon of choice.
In its original context, KMFDM’s “A Drug Against War” used sound to counter US Presidents Reagan and Bush 1’s War on Drugs/”New World Order” thinking. Indeed, in 1993, it was hard not to hear it as a response to 1991’s Operation Desert Storm, the US’s first military action as the ‘winner’ of the Cold War. It performs, in music, the “cremation of senses in friendly fire,” that its lyrics describe. It burns out our hearing, realizing through sound the sort of “creative destruction” or “shock doctrine” that characterizes neoliberalism more generally. In this rather Nietzschean model, the only way to make something “stronger than ever, ever before” is to first kill it. Death is the means to the most vibrant life.
The lyric–“stronger than ever, ever before”–is the first line of the chorus. At the end of every verse, there’s a short drumroll that leads into it. As S. Alexander Reed notes in Assimilate: A Critical History of Industrial Music, “clocking in at “322 bpm, the eighth-note snare fills at the end of the verses fire at about eleven rounds every second–the same rate as an AK-47” (29). This flourish foreshadows the gesture that, in the song’s bridge [after the second chorus, around 2:17 in the video above], musically “cremates” our senses in friendly fire–in this case, in the rapid fire of percussion. This rapid-fire percussion is one of the sonic elements that “Drug” shares with WWZ; in fact, the chorus uses what is likely (according to Reed) a machine gun sample. The machine gun effect mimics blast drumming. As Ronald Bogue explains in Deleuze’s Wake, blast drumming is a “tactic of accelerating meters to the point of collapse,” produced through the “cut-time alteration of downbeat kick drum and offbeat snare, the accent being heard on the offbeat but felt on the downbeat” (99). “A Drug Against War”’s AK-47 rolls actually accelerate to the point of auditory collapse, i.e. to the point at which humans generally can’t distinguish individual sonic events—the aural equivalent of seeing 24 frames per second as one continuous image. The AK-47’s rolls of ‘friendly’ fire cremate our sense of hearing.
The song’s chorus includes many other sonic elements shared by World War Z: doppler effects (such as the sounds of dropping bombs or planes buzzing the ground), rubble being moved around, military orders barked over radio. In the bridge several kinds of crowd noises are introduced: first, guitars buzz like a swarm of insects; then, a call-and-response in which singer Sascha Koneitzko echoes the chorus (which reverses the usual order in which the chorus echoes the individual leader); finally, a chaotic rabble of voices builds in intensity and leads into the sense-cremating climax.
An extended and intensified version of the “friendly fire” at the end of each verse, “A Drug Against War”’s climax builds to a peak by layering two full measures of AK-47-style drumroll on top of sounds of rabble, evoking the image of the military firing on an unruly crowd. This roll barrels towards the point of auditory collapse–if it got much faster, we’d be unable to distinguish individual rhythmic events, and hear a constant buzz (like in the beginning of the bridge), not a series of eighth notes. The roll’s forward momentum intensifies musical energy to an apex, culminating on the downbeat of the next measure in a florid lead guitar solo.
Describing the song as “sonic bombardment brighter than sunlight,” the lyrics confirm the music (and vice versa). The song overdrives sound until it sublimates into something else–if sunlight is more intense radiation than even soundwaves, here soundwaves amplify to a state more powerful than that. Cremating our senses in friendly fire, KMFDM channels soundwaves into a revolutionary drug, a drug against war. The band presents cleansing fire meant to purify us of disease: just as a fever kills pathogens in our bodies, the song burns our senses to kill a pathogenic ideology. Overdriving mainstream musical taste, offering something so brutal, so damaging to one’s ears, that only the avant-garde can survive, KMFDM inoculates the population against its most reactionary, war-mongering elements. “A Drug Against War” uses sound to perform a biopolitical operation, one that emerges as the basis of WWZ’s plot: the only way to save the human race from the zombies is to kill everything.
World War Z intensifies the horrors of contemporary biopolitics to the point that the only way to recuperate from them is to intensify them even further: in order for humanity to survive, everyone must be dead on their feet. In the sci-fi universe of World War Z, zombies aren’t eating for their survival, but for the survival of the virus they carry; they only attack and eat prey that are also (and primarily) attractive hosts for the virus. Pitt’s character, protagonist Gerry Lane, discovers that terminally ill humans aren’t legible to the zombies as human—that is, as attractive hosts. They won’t live long enough and/or are too weak to aggressively spread the virus. So, he decides the best way to protect humans from zombies and the virus they carry is to infect the remaining people with a deadly but ultimately curable illness. The World Health Organization develops a vaccine that allows healthy people to ‘pass’ as terminal cases. The only difference remaining in the post apocalyptic world of WWZ is between the quasi-dead and the walking dead. Death is the drug against WWZ.
The film doesn’t represent or express the biopolitical recuperation of death visually, but sonically: to make audiences feel what the narrative depicts, WWZ cremates their sense of hearing–often with more amplified and complex versions of the same sonic elements mobilized in “Drug.” Doppler effects, crowd noises, machine guns, military orders barked over radio bombards the film’s audience as sonic “friendly fire.” Though the film’s soundtrack doesn’t actually blow out its audience’s ears (what lawsuits!), it repeatedly simulates sonic cremation; the tinitus-y buzzing one hears after auditory trauma–what one hears in lieu of hearing—functions as a constant refrain. Narratively climactic moments are composed, cinematically, as sonic overdrive. The massive car crash as everyone tries to evacuate NYC in the beginning of the film, the moment when Pitt’s character thinks he may have been infected atop the NJ apartment building, the plane crash outside the Cardiff WHO office–each of these events culminates in tinitus-y ringing. As physical and psychological trauma overwhelms the characters, the film pretends to inflict overwhelming—cremating—auditory trauma on its audience.
In the WWZ universe, sound is destructive; it unleashes the zombie horde. At 49:00, a soldier says: “remember these things are drawn to sound…there’s only one way we’re getting you on that plane, and that’s quiet.” In a scene set in Jerusalem, excessive sound turns something miraculously positive—a Muslim girl and a Jewish girl leading a mixed crowd in song, a mini Arab-Israeli peace accord—Into a massacre. The sound attracts the zombie horde, leading them to swarm and overrun Jerusalem’s walls. Similarly, at the film’s end, Pitt’s character empties a soda machine so the cascade of cans will attract zombies away from the doors he needs to enter. By this point, Pitt’s character has injected himself with a deadly disease, effectively killing himself in order to preserve himself from zombification. The cascade of cans aesthetically represents this narrative point and hearkens back to KMFDM. The cans drop out of the machine at an increasingly rapid rate, mimicking “Drug”’s intensification of percussion events to and/or past the limit of human hearing. Just as Pitt’s character has crashed his body, the cascade of cans crashes our hearing.
The climax presents a narrative and the auditory convergence on the same biopolitical idea: kill everything, because then the best will bounce back, phoenix-like, from that sensory cremation, stronger than ever. Zombies can’t rebound from death, but still-living humans sure can (via immunization). Like a sonic bombardment brighter and more radiant than sunlight, this anti-zombie camouflage tactic phase-shifts death into exceptionally lively life. Just as the muted, tinitus-y moments in the film make the subsequent scenes feel comparatively more sonically rich and dynamic, intentional and carefully managed mass extinction ultimately makes the living more vibrant.
Sound & Biopolitics
Such vibrancy–that is, what Julian Henriques dubs “the dynamics of [the] periodic motion of vibrations” in Sounding Bodies: Reggae Sound Systems, Performance Techniques, and Ways of Knowing (265)–is what “life” and “sound,” as they are conceived by and function in contemporary biopolitics, have in common. “Sound,” according to Henriques is “a particular kind of periodic motion, variation and change” (247). Sound waves are dynamic patterns of intensities (pressure); they move through matter and respond in turn, both to that movement itself and the secondary sound waves (harmonics) that movement produces. WWZ treats this notion of periodic motion, variation, and change as the conceptual basis for the ideally biopolitical “life.” At around 20:00, when Pitt’s character attempts to convince the Latino family sheltering him to leave their apartment with him, he says “movement is life…Moviemento es vida.” Sedentary fortresses protect no one from zombies–we see this repeatedly in the film. The only way to survive is by rapidly adjusting to new conditions. The dynamism of adaptive flows—the ability to bounce back and recuperate (like an echoing pressure wave), to dynamically recombine (like both harmonizing frequencies and like a virus), to find signal in noise–this dynamism is life. Because it adapts to new challenges, because it moves, varies, and changes, life can bounce back from total annihilation, stronger than ever before. Only life lived like sound can be properly and sufficiently resilient. In WWZ, the zombie virus is a eugenic tool that weeds out insufficiently “sonic” life, life that is too static to respond to capitalist and biopolitical mandates for calculable motion, variation, and change.
When read through “Drug,” WWZ illustrates the epistemic and ontological importance of sound to contemporary biopolitics. We think “life” works like we think sound works. Because “life” is the object and the mechanism of biopolitical government, power works on and through us sonically. If we want to analyze, critique, and fight the institutions, structures, and practices that put power to work for white supremacy, cis/het patriarchy, and all other forms of domination, then we need to start thinking and working sonically, too.
Some theorists, such as Elizabeth Grosz and Adriana Cavarero incorrectly think this move to sound and voice is itself revolutionary and counter-hegemonic. Just as the critique posed in 1993’s “Drug” has been co-opted by 2013’s WWZ, white feminist theory’s sonic counter-modernities are the medium of biopolitical white supremacist patriarchy. When we think and work sonically, we’re working with the master’s tools; to bring down the master’s house, we have to use them critically and strategically.
Featured Image by Flickr User crisper fayltrash
Robin James is Associate Professor of Philosophy at UNC Charlotte. She is author of two books: Resilience & Melancholy: pop music, feminism, and neoliberalism will be published by Zer0 books this fall, and The Conjectural Body: gender, race and the philosophy of music was published by Lexington Books in 2010. Her work on feminism, race, contemporary continental philosophy, pop music, and sound studies has appeared in The New Inquiry, Hypatia, differences, Contemporary Aesthetics, and the Journal of Popular Music Studies. She is also a digital sound artist and musician. She blogs at its-her-factory.com and is a regular contributor to Cyborgology.
REWIND!…If you liked this post, check out:
On Sound and Pleasure: Meditations on the Human Voice–Yvon Bonenfant
The Noises of Finance– Nick Knouf
Today the SO! Thursday stream inaugurates a four-part series entitled Hearing the UnHeard, which promises to blow your mind by way of your ears. Our Guest Editor is Seth Horowitz, a neuroscientist at NeuroPop and author of The Universal Sense: How Hearing Shapes the Mind (Bloomsbury, 2012), whose insightful work on brings us directly to the intersection of the sciences and the arts of sound.
That’s where he’ll be taking us in the coming weeks. Check out his general introduction just below, and his own contribution for the first piece in the series. — NV
Welcome to Hearing the UnHeard, a new series of articles on the world of sound beyond human hearing. We are embedded in a world of sound and vibration, but the limits of human hearing only let us hear a small piece of it. The quiet library screams with the ultrasonic pulsations of fluorescent lights and computer monitors. The soothing waves of a Hawaiian beach are drowned out by the thrumming infrasound of underground seismic activity near “dormant” volcanoes. Time, distance, and luck (and occasionally really good vibration isolation) separate us from explosive sounds of world-changing impacts between celestial bodies. And vast amounts of information, ranging from the songs of auroras to the sounds of dying neurons can be made accessible and understandable by translating them into human-perceivable sounds by data sonification.
Four articles will examine how this “unheard world” affects us. My first post below will explore how our environment and evolution have constrained what is audible, and what tools we use to bring the unheard into our perceptual realm. In a few weeks, sound artist China Blue will talk about her experiences recording the Vertical Gun, a NASA asteroid impact simulator which helps scientists understand the way in which big collisions have shaped our planet (and is very hard on audio gear). Next, Milton A. Garcés, founder and director of the Infrasound Laboratory of University of Hawaii at Manoa will talk about volcano infrasound, and how acoustic surveillance is used to warn about hazardous eruptions. And finally, Margaret A. Schedel, composer and Associate Professor of Music at Stonybrook University will help readers explore the world of data sonification, letting us listen in and get greater intellectual and emotional understanding of the world of information by converting it to sound.
– Guest Editor Seth Horowitz
Although light moves much faster than sound, hearing is your fastest sense, operating about 20 times faster than vision. Studies have shown that we think at the same “frame rate” as we see, about 1-4 events per second. But the real world moves much faster than this, and doesn’t always place things important for survival conveniently in front of your field of view. Think about the last time you were driving when suddenly you heard the blast of a horn from the previously unseen truck in your blind spot.
Hearing also occurs prior to thinking, with the ear itself pre-processing sound. Your inner ear responds to changes in pressure that directly move tiny little hair cells, organized by frequency which then send signals about what frequency was detected (and at what amplitude) towards your brainstem, where things like location, amplitude, and even how important it may be to you are processed, long before they reach the cortex where you can think about it. And since hearing sets the tone for all later perceptions, our world is shaped by what we hear (Horowitz, 2012).
But we can’t hear everything. Rather, what we hear is constrained by our biology, our psychology and our position in space and time. Sound is really about how the interaction between energy and matter fill space with vibrations. This makes the size, of the sender, the listener and the environment, one of the primary features that defines your acoustic world.
You’ve heard about how much better your dog’s hearing is than yours. I’m sure you got a slight thrill when you thought you could actually hear the “ultrasonic” dog-training whistles that are supposed to be inaudible to humans (sorry, but every one I’ve tested puts out at least some energy in the upper range of human hearing, even if it does sound pretty thin). But it’s not that dogs hear better. Actually, dogs and humans show about the same sensitivity to sound in terms of sound pressure, with human’s most sensitive region from 1-4 kHz and dogs from about 2-8 kHz. The difference is a question of range and that is tied closely to size.
Most dogs, even big ones, are smaller than most humans and their auditory systems are scaled similarly. A big dog is about 100 pounds, much smaller than most adult humans. And since body parts tend to scale in a coordinated fashion, one of the first places to search for a link between size and frequency is the tympanum or ear drum, the earliest structure that responds to pressure information. An average dog’s eardrum is about 50 mm2, whereas an average human’s is about 60 mm2. In addition while a human’s cochlea is spiral made of 2.5 turns that holds about 3500 inner hair cells, your dog’s has 3.25 turns and about the same number of hair cells. In short: dogs probably have better high frequency hearing because their eardrums are better tuned to shorter wavelength sounds and their sensory hair cells are spread out over a longer distance, giving them a wider range.
Then again, if hearing was just about size of the ear components, then you’d expect that yappy 5 pound Chihuahua to hear much higher frequencies than the lumbering 100 pound St. Bernard. Yet hearing sensitivity from the two ends of the dog spectrum don’t vary by much. This is because there’s a big difference between what the ear can mechanically detect and what the animal actually hears. Chihuahuas and St. Bernards are both breeds derived from a common wolf-like ancestor that probably didn’t have as much variability as we’ve imposed on the domesticated dog, so their brains are still largely tuned to hear what a medium to large pseudo wolf-like animal should hear (Heffner, 1983).
But hearing is more than just detection of sound. It’s also important to figure out where the sound is coming from. A sound’s location is calculated in the superior olive – nuclei in the brainstem that compare the difference in time of arrival of low frequency sounds at your ears and the difference in amplitude between your ears (because your head gets in the way, making a sound “shadow” on the side of your head furthest from the sound) for higher frequency sounds. This means that animals with very large heads, like elephants, will be able to figure out the location of longer wavelength (lower pitched) sounds, but probably will have problems localizing high pitched sounds because the shorter frequencies will not even get to the other side of their heads at a useful level. On the other hand, smaller animals, which often have large external ears, are under greater selective pressure to localize higher pitched sounds, but have heads too small to pick up the very low infrasonic sounds that elephants use.
But you as a human are a fairly big mammal. If you look up “Body Size Species Richness Distribution” which shows the relative size of animals living in a given area, you’ll find that humans are among the largest animals in North America (Brown and Nicoletto, 1991). And your hearing abilities scale well with other terrestrial mammals, so you can stop feeling bad about your dog hearing “better.” But what if, by comic-book science or alternate evolution, you were much bigger or smaller? What would the world sound like? Imagine you were suddenly mouse-sized, scrambling along the floor of an office. While the usual chatter of humans would be almost completely inaudible, the world would be filled with a cacophony of ultrasonics. Fluorescent lights and computer monitors would scream in the 30-50 kHz range. Ultrasonic eddies would hiss loudly from air conditioning vents. Smartphones would not play music, but rather hum and squeal as their displays changed.
And if you were larger? For a human scaled up to elephantine dimensions, the sounds of the world would shift downward. While you could still hear (and possibly understand) human speech and music, the fine nuances from the upper frequency ranges would be lost, voices audible but mumbled and hard to localize. But you would gain the infrasonic world, the low rumbles of traffic noise and thrumming of heavy machinery taking on pitch, color and meaning. The seismic world of earthquakes and volcanoes would become part of your auditory tapestry. And you would hear greater distances as long wavelengths of low frequency sounds wrap around everything but the largest obstructions, letting you hear the foghorns miles distant as if they were bird calls nearby.
But these sounds are still in the realm of biological listeners, and the universe operates on scales far beyond that. The sounds from objects, large and small, have their own acoustic world, many beyond our ability to detect with the equipment evolution has provided. Weather phenomena, from gentle breezes to devastating tornadoes, blast throughout the infrasonic and ultrasonic ranges. Meteorites create infrasonic signatures through the upper atmosphere, trackable using a system devised to detect incoming ICBMs. Geophones, specialized low frequency microphones, pick up the sounds of extremely low frequency signals foretelling of volcanic eruptions and earthquakes. Beyond the earth, we translate electromagnetic frequencies into the audible range, letting us listen to the whistlers and hoppers that signal the flow of charged particles and lightning in the atmospheres of Earth and Jupiter, microwave signals of the remains of the Big Bang, and send listening devices on our spacecraft to let us hear the winds on Titan.
Here is a recording of whistlers recorded by the Van Allen Probes currently orbiting high in the upper atmosphere:
When the computer freezes or the phone battery dies, we complain about how much technology frustrates us and complicates our lives. But our audio technology is also the source of wonder, not only letting us talk to a friend around the world or listen to a podcast from astronauts orbiting the Earth, but letting us listen in on unheard worlds. Ultrasonic microphones let us listen in on bat echolocation and mouse songs, geophones let us wonder at elephants using infrasonic rumbles to communicate long distances and find water. And scientific translation tools let us shift the vibrations of the solar wind and aurora or even the patterns of pure math into human scaled songs of the greater universe. We are no longer constrained (or protected) by the ears that evolution has given us. Our auditory world has expanded into an acoustic ecology that contains the entire universe, and the implications of that remain wonderfully unclear.
Exhibit: Home Office
This is a recording made with standard stereo microphones of my home office. Aside from usual typing, mouse clicking and computer sounds, there are a couple of 3D printers running, some music playing, largely an environment you don’t pay much attention to while you’re working in it, yet acoustically very rich if you pay attention.
This sample was made by pitch shifting the frequencies of sonicoffice.wav down so that the ultrasonic moves into the normal human range and cuts off at about 1-2 kHz as if you were hearing with mouse ears. Sounds normally inaudible, like the squealing of the computer monitor cycling on kick in and the high pitched sound of the stepper motors from the 3D printer suddenly become much louder, while the familiar sounds are mostly gone.
This recording of the office was made with a Clarke Geophone, a seismic microphone used by geologists to pick up underground vibration. It’s primary sensitivity is around 80 Hz, although it’s range is from 0.1 Hz up to about 2 kHz. All you hear in this recording are very low frequency sounds and impacts (footsteps, keyboard strikes, vibration from printers, some fan vibration) that you usually ignore since your ears are not very well tuned to frequencies under 100 Hz.
Finally, this sample was made by pitch shifting the frequencies of infrasonicoffice.wav up as if you had grown to elephantine proportions. Footsteps and computer fan noises (usually almost indetectable at 60 Hz) become loud and tonal, and all the normal pitch of music and computer typing has disappeared aside from the bass. (WARNING: The fan noise is really annoying).
The point is: a space can sound radically different depending on the frequency ranges you hear. Different elements of the acoustic environment pop up depending on the type of recording instrument you use (ultrasonic microphone, regular microphones or geophones) or the size and sensitivity of your ears.–
Featured image by Flickr User Jaime Wong.
Seth S. Horowitz, Ph.D. is a neuroscientist whose work in comparative and human hearing, balance and sleep research has been funded by the National Institutes of Health, National Science Foundation, and NASA. He has taught classes in animal behavior, neuroethology, brain development, the biology of hearing, and the musical mind. As chief neuroscientist at NeuroPop, Inc., he applies basic research to real world auditory applications and works extensively on educational outreach with The Engine Institute, a non-profit devoted to exploring the intersection between science and the arts. His book The Universal Sense: How Hearing Shapes the Mind was released by Bloomsbury in September 2012.
REWIND! If you liked this post, check out …
Learning to Listen Beyond Our Ears– Owen Marshall
This is Your Body on the Velvet Underground– Jacob Smith
After a rockin’ (and seriously informative) series of podcasts from Leonard J. Paul, a Drrty South banger dropped by SO! Regular Regina Bradley, a screamtastic meditation from Yvon Bonenfant, and a heaping plate of food sounds from Steph Ceraso, our summer Sound and Pleasure series gets even louder with Kariann Goldschmidt‘s work on live events in Brazil. Brasil Ao Vivo! --JS, Editor-in-Chief
Brazilians pray, cheer and celebrate in public and often in close physical proximity to each other. From the nearly 3 million people that flocked to Copacabana Beach to hear Pope Francis lead a mass in 2013 to the huge crowds that regularly turn out for concerts at Maracanã stadium, Brazilians earn their global reputation for large-scale public events. Of course there is Carnival in Rio de Janeiro and Salvador; the largest LGBT Pride Parade in the world held in São Paulo; and then there is football.
The relationship between large-scale public events and sound hit home as the country reacted to the national team’s humiliating loss to Germany in the semi-final round of the 2014 FIFA World Cup. The world witnessed a different kind of public outpouring as the Brazilian public mourned. Within hours of the initial shock at the lopsided score, images of Brazilian football fans weeping and screaming in the stadium and on the street became a humorous meme with music and sound playing a prominent role. By the next day, most Brazilian football observers were taking pleasure in the public spectacle of weeping fans. With the abundance of images featuring hysteria, videos mocking the intensity of the crying went viral with dramatic musical scores. One observer proclaimed : “essa capacidade de rir de nós mesmos é uma das melhores qualidades”; the capacity to laugh at ourselves is one of our best qualities. That Brazilians express all varieties of emotions and annual passages together in public for everyone to witness, even when they border on campy excess, allow for everyone to feel the pleasures of community and the power of public performance.
All of this led me to believe that such a public culture has an effect on the aesthetics of what performance studies scholar Philip Auslander calls “liveness” in recorded music and related viral media. Auslander argues that the appeal of liveness for television broadcasts, concerts, and other stage performances allows audiences to feel the immediacy of the moment even if the presence of mediation, such as screens and on-air censorship, is obvious. The international spectacle of Brazilians emoting en masse, then, has a direct relationship with Brazilian sonic aesthetics. Nowhere, I argue, is this more prominent than in the (sometimes viral) popularity of live recordings.
That immediacy Auslander speaks of spreads to many aspects of Brazilian popular culture, including the popularity of concert DVDs and albums which are regularly listed among the most popular domestic recordings. In fact, concert records tend to be more popular than the studio albums that inspire the tour. These live albums often carry the designations Ao Vivo, live or MTV Acústico (the equivalent of the Unplugged albums popular in the United States), and they are often recorded in such a way so as to feature the interaction of the crowds. In place of the draw for authenticity (a value that permeates the MTV Unplugged recordings) is the love for community, and for experiencing big emotions together no matter how obviously they are mediated through cameras, microphones and other technology. Through the example of the continued popularity of live albums in Brazil, there is an opening for a different theorization for sounding liveness; in place of celebrating canonic performances and virtuosity, the valorization of liveness in Brazil reinforces the importance of crowds and the so-called “popular classes” at the root of the politicized singer-songwriter genre MPB or Música Popular Brasileira.
The pleasure and preference for live recordings also extends to social media. For meme chasers, a good example of this is Michel Teló’s 2011 hit “Ai Se Eu Te Pego.” The song and video were recorded ao vivo before a crowd dominated by young women. A close listen reveals that sounds of Teló’s female audience members are just as important as his voice even if his voice is only slightly louder in the mix. There is barely a moment in the recording when the audience stops making itself heard; the engineering revels in their presence. This is especially obvious during the opening seconds of the track when Teló and his audience sing “Nossa, nossa / assim você me mata / Ai, se eu te pego / Ai, ai, se eu te pego” [Wow, wow / you kill me like that / Ah, if I could get you / ah, ah, if I could get you] in unison at nearly the same volume in the mix. When the accordion and electric bass (crucial instruments for the song’s forró style) finally enter over the screaming audience, there is a noticeable break in the tension set up by the audience and Teló singing together. Their cries, like those in other live recordings, illustrate Teló’s appeal to the crowd in that moment while also allowing other listeners to imagine themselves there.
Teló’s song went viral (as of this writing, the official version currently has nearly 580 million views on YouTube and over 72 million plays on Spotify), with alternate video versions teaching the song’s dance steps and others highlighting global football stars dancing and singing along to the song. At one point Neymar, the national team’s biggest hope for World Cup victory, sang with Teló in front of a crowd. In general, Teló’s live songs easily outpace his studio recordings in terms of virality, and, I would argue, that a major part of the appeal of “Ai Se Eu Te Pego” is its provenance in a concert setting. It is just as important that the screaming throngs of women are audible as it is for those dance steps to be easy and recognizable. The liveness of the recording is so important, in fact, that the screaming audience appears as sampled snippets in the Pitbull remix. In its viral form, Teló’s song united the popularity of live spectacle with Brazil’s enthusiasm for other live events, merging concert goers with football fans.
The popularity of Teló’s live song is not an isolated incident. Look, for example, at record sales figures for all time. Two are live albums by artists who do not appear elsewhere on the list. Other albums that have sold more than 2 million copies in Brazil alone are by Roberto Carlos (Acústico MTV) and the teen pop/rock duo Sandy and Júnior (As Quatro Estações ao Vivo and Era Uma Vez… Ao Vivo). In 2011, five of the top ten albums in Brazil fit the ao vivo mode with little regard to genre: MPB stars Caetano Veloso and Maria Gadú are there alongside sertanejo artists Paula Fernandes and Luan Santana. In 2012, three of the top 20 best-sellers were live albums. Meanwhile, DVDs of concerts in Brazil continue to be strong sellers. Thus, the communal pleasure palpable on-screen translates to that experienced in the home.
Compare this with the status of live records in the United States in the last few years where they have rarely seen any chart success. If anything, liveness continues in YouTube clips and Spotify Sessions but not in physical sales and downloads. This is probably because live albums for U.S. based artists are embedded with different values having to do with the rock authenticity rather than communal pleasure. These performances demonstrate the chops of the musician and valorize the concerts (and tours) as events. The double live albums from the 1970s such as as Frampton Comes Alive, Lynyrd Skynyrd’s One More From The Road, and Kiss Alive! hold a prized place in the classic rock canon, often as much for extended guitar solos rather as the screaming throngs of fans. In the late ‘80s and early ’90s live albums, especially MTV Unplugged, re-inscribed a love of liveness through acoustic instruments and songs that reached back into the roots of American popular music. Eric Clapton’s Unplugged (1992) even topped the Billboard album charts and won 6 Grammy awards including Album of the Year while other records such as Nirvana’s MTV Unplugged in New York and U2’s Rattle and Hum were multi-platinum hits. While there is the occasional top-40 live single, these songs are the exception to a genre of that has has moved liveness to YouTube rather than streaming and MP3 markets.
SO! contributor Osvaldo Oyola has noted there is a tension between the efforts recording engineers often go through to make studio recordings sound as immediate as possible, and those that call attention to the recording process. Live records replace the need to sound polished with the need to sound spontaneous, often reveling in mistakes and banter. That immediacy is something I enjoy when listening to live recordings and it has a parallel for many people who participate in the reception of major events in real time through social media.
In Brazil, audiences enjoy the immense power of participation in live events. As part of a larger work in progress I’m particularly fascinated by how this power and pleasure is mediated through the sonic experience of recordings and viral social media. Whether they are sharing tears over an international football loss or singing along to “Ai Se Eu Te Pego” Brazilians extend Auslander’s liveness by prolonging and replaying the immediacy of the crowds to experience that shared sonic moment, again and again.
Kariann Goldschmitt is a Visiting Lecturer in the Faculty of Music at University of Cambridge. Her scholarly work focuses on Brazilian music, modes of listening, and sonic branding in the global cultural industries. She has published in the Oxford Handbook of Mobile Music Studies, Popular Music and Society, American Music, Yearbook for Traditional Music, and Luso-Brazilian Review and contributes to the South American cultural magazine, Sounds and Colours.
Featured image: Adapted from “Gloria” by Flickr user Lourenço Fabrino, CC BY-NC-SA 2.0
REWIND! . . .If you liked this post, you may also dig:
Sound-politics in São Paulo, Brazil– Leonardo Cardoso