Welcome back to Hearing the UnHeard, Sounding Out‘s series on how the unheard world affects us, which started out with my post on the hearing ranges of animals, and now continues with this exciting piece by China Blue.
From recording the top of the Eiffel Tower to the depths of the rising waters around Venice, from building fields of robotic crickets in Tokyo to lofting 3D printed ears with binaural mics in a weather balloon, China Blue is as much an acoustic explorer as a sound artist. While she makes her works publicly accessible, shown in museums and galleries around the world, she searches for inspiration in acoustically inaccessible sources, sometimes turning sensory possibilities on their head and sonifying the visual or reformatting sounds to make the inaudible audible.
In this installment of Hearing the UnHeard, China Blue talks about cataclysmic sounds we might not survive hearing and her experiences recording simulated asteroid strikes at NASA’s Ames Vertical Gun Range.
– Guest Editor Seth Horowitz
Fundamentally speaking, sound is the result of something banging into something else. And since everything in the universe, from the slow recombination of chemicals to the hypervelocity impacts of asteroids smashing into planet surfaces, is ultimately the result of things banging into things, the entire universe has a sonic signature. But because of the huge difference in scale of these collisions, some things remain unheard without very specialized equipment. And others, you hope you never hear.
Unheard sounds can be hidden subtly beneath your feet like the microsounds of ants walking, or they can be unexpectedly harmonic like the seismic vibrations of a huge structure like the Eiffel Tower. These are sounds that we can explore safely, using audio editing tools to integrate them into new musical or artistic pieces.
Luckily, our experience with truly primal sounds, such as the explosive shock waves of asteroid impacts that shaped most of our solar system (including the Earth) is rarer. Those who have been near a small example of such an event, such as the residents of Chelyabinsk, Russia in 2013 were probably less interested in the sonic event and more interested in surviving the experience.
But there remains something seductive about being able to hear sounds such as the cosmic rain of fire and ice that shaped our planet billions of years ago. A few years ago, when I became fascinated with sounds “bigger” than humans normally hear, I was able to record simulations of these impacts in one of the few places on Earth where you can, at NASA Ames Vertical Gun Range.
The Vertical Gun at Ames Research Center (AVGR) was designed to conduct scientific studies of lunar impacts. It consists of a 25 foot long gun barrel with a powder chamber at one end and a target chamber, painted bright blue, that looks like the nose of an upended submarine, about 8 feet in diameter and height at the other. The walls of the chamber are of thick steel strong enough to let its interior be pumped down to vacuum levels close to that of outer space, or back-filled with various gases to simulate different planetary atmospheres. Using hydrogen and/or up to half a pound of gun powder, the AVGR can launch projectiles at astonishing speeds of 500 to 7,000 m/s (1,100 to 16,000 mph). By varying the gun’s angle of elevation, projectiles can be shot into the target so that it simulates impacts from overhead or at skimming angles.
In other words, it’s a safe way to create cataclysmic impacts, and then analyze them using million frame-per-second video cameras without leaving the security of Earth.
My husband, Dr. Seth Horowitz who is an auditory neuroscientist and another devotee of sound, is close friends with one of the principal investigators of the Ames Vertical Gun, Professor Peter Schultz. Schultz is well known for his 2005 project to blow a hole in the comet Tempel 1 to analyze its composition, and for his involvement in the LCROSS mission that smashed into the south pole of the moon to look for evidence of water. During one conversation discussing the various analytical techniques they use to understand impacts, I asked, “I wonder what it sounds like.” As sound is the propagation of energy by matter banging into other matter, this seemed like the ultimate opportunity to record a “Big Bang” that wouldn’t actually get you killed by flying meteorite shards. Thankfully, my husband and I were invited to come to Ames to find out.
I had a feeling that the AVGR would produce fascinating new sounds that might provide us with different insights into impacts than the more common visual techniques. Because this was completely new research, we used a number of different microphones that were sensitive to different ranges and types of sound and vibrations to provide us with a selection of recording results. As an artist I found the research to be the dominant part of the work because the processes of capturing and analyzing the sounds were a feat unto themselves. As we prepared for the experiment, I thought about what I could do with these sounds. When I eventually create a work out of them, I anticipate using them in an installation that would trigger impact sounds when people enter the room, but I have not yet mounted this work since I suspect that this would be too frightening for most exhibition spaces to want.
Part of my love (and frustration) for sound work is figuring out how to best capture that fleeting moment in which the sound is just right, when the sound evokes a complex response from its listeners without having to even be explained. The sound of Mach 10 impacts and its effects on the environment had such possibilities. In pursuit of the “just right,” we wired up the gun and chamber with multiple calibrated acoustic and seismic microphones, then fed them into a single high speed multichannel recorder, pressed “record” and made for the “safe” room while the Big Red Button was pressed, launching the first impactor. We recorded throughout the day, changing the chamber’s conditions from vacuum to atmosphere.
When we finally got to listen that afternoon, we heard things we never imagined. Initial shots in vacuum were surprisingly dull. The seismic microphones picked up the “thump” of the projectile hitting the sand target and a few pattering sounds as secondary particles struck the surfaces. There were of course no sounds from the boundary or ultrasonic mics due to the lack of air to propagate sound waves. While they were scientifically useful–they demonstrated that we could identify specific impact events launched from the target—they weren’t very acoustically dramatic.
When a little atmosphere was added, however, we began picking up subtle sounds, such as the impact and early spray of particles from the boundary mic and the fact that there was an air leak from the pitch shifted ultrasonic mic. But when the chamber was filled with an earthlike atmosphere and the target dish filled with tiny toothpicks to simulate trees, building the scenario for a tiny Tunguska event (a 1908 explosion of an interstellar object in Russia, the largest in recorded history), the sound was stunning:
After the initial explosion, there was a sandstorm as the particles of sand from the target flew about at Mach 5 (destroying one of the microphones in the process), and giving us a simulation of a major asteroid explosion.
66 million years ago, in a swampy area by the Yucatan Penninsula, something like this probably occurred, when a six mile wide rock burned through the atmosphere to strike the water, ending the 135 million year reign of the dinosaurs. Perhaps it sounded a little like this simulation:
Any living thing that heard this – dinausaurs, birds, frogs insects – is long gone. By thinking about the event through new sounds, however, we can not only create new ways to analyze natural phenomena, but also extend the boundaries of our ability to listen across time and space and imagine what the sound of that impact might have been like, from an infrasonic rumble to a killing concussion.
It would probably terrify any listener to walk in to an art exhibition space filled with simply the sounds of simulated hypervelocity impacts, replete with loud, low frequency sounds and infrasonic vibrations. But there is something to that terror. Such sounds trigger ancient evolutionary pathways which are still with us because they were so good at helping us survive similar events by making us run, putting as much distance between us and the cataclysmic source, something that lingers even in safe reproductions, resynthesized from controlled, captured sources.
China Blue is a two time NASA/RI Space Grant recipient and an internationally exhibiting artist who was the first person to record the Eiffel Tower in Paris, France and NASA’s Vertical Gun. Her acoustic work has led her to be selected as the US representative at OPEN XI, Venice, Italy and at the Tokyo Experimental Art Festival in Tokyo, Japan, and was the featured artist for the 2006 annual meeting of the Acoustic Society of America. Reviews of her work have been published in the Wall Street Journal, New York Times, Art in America, Art Forum, artCritical and NY Arts, to name a few. She has been an invited speaker at Harvard, Yale, MIT, Berkelee School of Music, Reed College and Brown University. She is the Founder and Executive Director of The Engine Institute www.theengineinstitute.org.
Featured Image of a high-speed impact recorded by AVGR. Image by P. H. Schultz. Via Wikimedia Commons.
REWIND! If you liked this post, check out …
Learning to Listen Beyond Our Ears– Owen Marshall
Living with Noise– Osvaldo Oyola
Today the SO! Thursday stream inaugurates a four-part series entitled Hearing the UnHeard, which promises to blow your mind by way of your ears. Our Guest Editor is Seth Horowitz, a neuroscientist at NeuroPop and author of The Universal Sense: How Hearing Shapes the Mind (Bloomsbury, 2012), whose insightful work on brings us directly to the intersection of the sciences and the arts of sound.
That’s where he’ll be taking us in the coming weeks. Check out his general introduction just below, and his own contribution for the first piece in the series. — NV
Welcome to Hearing the UnHeard, a new series of articles on the world of sound beyond human hearing. We are embedded in a world of sound and vibration, but the limits of human hearing only let us hear a small piece of it. The quiet library screams with the ultrasonic pulsations of fluorescent lights and computer monitors. The soothing waves of a Hawaiian beach are drowned out by the thrumming infrasound of underground seismic activity near “dormant” volcanoes. Time, distance, and luck (and occasionally really good vibration isolation) separate us from explosive sounds of world-changing impacts between celestial bodies. And vast amounts of information, ranging from the songs of auroras to the sounds of dying neurons can be made accessible and understandable by translating them into human-perceivable sounds by data sonification.
Four articles will examine how this “unheard world” affects us. My first post below will explore how our environment and evolution have constrained what is audible, and what tools we use to bring the unheard into our perceptual realm. In a few weeks, sound artist China Blue will talk about her experiences recording the Vertical Gun, a NASA asteroid impact simulator which helps scientists understand the way in which big collisions have shaped our planet (and is very hard on audio gear). Next, Milton A. Garcés, founder and director of the Infrasound Laboratory of University of Hawaii at Manoa will talk about volcano infrasound, and how acoustic surveillance is used to warn about hazardous eruptions. And finally, Margaret A. Schedel, composer and Associate Professor of Music at Stonybrook University will help readers explore the world of data sonification, letting us listen in and get greater intellectual and emotional understanding of the world of information by converting it to sound.
– Guest Editor Seth Horowitz
Although light moves much faster than sound, hearing is your fastest sense, operating about 20 times faster than vision. Studies have shown that we think at the same “frame rate” as we see, about 1-4 events per second. But the real world moves much faster than this, and doesn’t always place things important for survival conveniently in front of your field of view. Think about the last time you were driving when suddenly you heard the blast of a horn from the previously unseen truck in your blind spot.
Hearing also occurs prior to thinking, with the ear itself pre-processing sound. Your inner ear responds to changes in pressure that directly move tiny little hair cells, organized by frequency which then send signals about what frequency was detected (and at what amplitude) towards your brainstem, where things like location, amplitude, and even how important it may be to you are processed, long before they reach the cortex where you can think about it. And since hearing sets the tone for all later perceptions, our world is shaped by what we hear (Horowitz, 2012).
But we can’t hear everything. Rather, what we hear is constrained by our biology, our psychology and our position in space and time. Sound is really about how the interaction between energy and matter fill space with vibrations. This makes the size, of the sender, the listener and the environment, one of the primary features that defines your acoustic world.
You’ve heard about how much better your dog’s hearing is than yours. I’m sure you got a slight thrill when you thought you could actually hear the “ultrasonic” dog-training whistles that are supposed to be inaudible to humans (sorry, but every one I’ve tested puts out at least some energy in the upper range of human hearing, even if it does sound pretty thin). But it’s not that dogs hear better. Actually, dogs and humans show about the same sensitivity to sound in terms of sound pressure, with human’s most sensitive region from 1-4 kHz and dogs from about 2-8 kHz. The difference is a question of range and that is tied closely to size.
Most dogs, even big ones, are smaller than most humans and their auditory systems are scaled similarly. A big dog is about 100 pounds, much smaller than most adult humans. And since body parts tend to scale in a coordinated fashion, one of the first places to search for a link between size and frequency is the tympanum or ear drum, the earliest structure that responds to pressure information. An average dog’s eardrum is about 50 mm2, whereas an average human’s is about 60 mm2. In addition while a human’s cochlea is spiral made of 2.5 turns that holds about 3500 inner hair cells, your dog’s has 3.25 turns and about the same number of hair cells. In short: dogs probably have better high frequency hearing because their eardrums are better tuned to shorter wavelength sounds and their sensory hair cells are spread out over a longer distance, giving them a wider range.
Then again, if hearing was just about size of the ear components, then you’d expect that yappy 5 pound Chihuahua to hear much higher frequencies than the lumbering 100 pound St. Bernard. Yet hearing sensitivity from the two ends of the dog spectrum don’t vary by much. This is because there’s a big difference between what the ear can mechanically detect and what the animal actually hears. Chihuahuas and St. Bernards are both breeds derived from a common wolf-like ancestor that probably didn’t have as much variability as we’ve imposed on the domesticated dog, so their brains are still largely tuned to hear what a medium to large pseudo wolf-like animal should hear (Heffner, 1983).
But hearing is more than just detection of sound. It’s also important to figure out where the sound is coming from. A sound’s location is calculated in the superior olive – nuclei in the brainstem that compare the difference in time of arrival of low frequency sounds at your ears and the difference in amplitude between your ears (because your head gets in the way, making a sound “shadow” on the side of your head furthest from the sound) for higher frequency sounds. This means that animals with very large heads, like elephants, will be able to figure out the location of longer wavelength (lower pitched) sounds, but probably will have problems localizing high pitched sounds because the shorter frequencies will not even get to the other side of their heads at a useful level. On the other hand, smaller animals, which often have large external ears, are under greater selective pressure to localize higher pitched sounds, but have heads too small to pick up the very low infrasonic sounds that elephants use.
But you as a human are a fairly big mammal. If you look up “Body Size Species Richness Distribution” which shows the relative size of animals living in a given area, you’ll find that humans are among the largest animals in North America (Brown and Nicoletto, 1991). And your hearing abilities scale well with other terrestrial mammals, so you can stop feeling bad about your dog hearing “better.” But what if, by comic-book science or alternate evolution, you were much bigger or smaller? What would the world sound like? Imagine you were suddenly mouse-sized, scrambling along the floor of an office. While the usual chatter of humans would be almost completely inaudible, the world would be filled with a cacophony of ultrasonics. Fluorescent lights and computer monitors would scream in the 30-50 kHz range. Ultrasonic eddies would hiss loudly from air conditioning vents. Smartphones would not play music, but rather hum and squeal as their displays changed.
And if you were larger? For a human scaled up to elephantine dimensions, the sounds of the world would shift downward. While you could still hear (and possibly understand) human speech and music, the fine nuances from the upper frequency ranges would be lost, voices audible but mumbled and hard to localize. But you would gain the infrasonic world, the low rumbles of traffic noise and thrumming of heavy machinery taking on pitch, color and meaning. The seismic world of earthquakes and volcanoes would become part of your auditory tapestry. And you would hear greater distances as long wavelengths of low frequency sounds wrap around everything but the largest obstructions, letting you hear the foghorns miles distant as if they were bird calls nearby.
But these sounds are still in the realm of biological listeners, and the universe operates on scales far beyond that. The sounds from objects, large and small, have their own acoustic world, many beyond our ability to detect with the equipment evolution has provided. Weather phenomena, from gentle breezes to devastating tornadoes, blast throughout the infrasonic and ultrasonic ranges. Meteorites create infrasonic signatures through the upper atmosphere, trackable using a system devised to detect incoming ICBMs. Geophones, specialized low frequency microphones, pick up the sounds of extremely low frequency signals foretelling of volcanic eruptions and earthquakes. Beyond the earth, we translate electromagnetic frequencies into the audible range, letting us listen to the whistlers and hoppers that signal the flow of charged particles and lightning in the atmospheres of Earth and Jupiter, microwave signals of the remains of the Big Bang, and send listening devices on our spacecraft to let us hear the winds on Titan.
Here is a recording of whistlers recorded by the Van Allen Probes currently orbiting high in the upper atmosphere:
When the computer freezes or the phone battery dies, we complain about how much technology frustrates us and complicates our lives. But our audio technology is also the source of wonder, not only letting us talk to a friend around the world or listen to a podcast from astronauts orbiting the Earth, but letting us listen in on unheard worlds. Ultrasonic microphones let us listen in on bat echolocation and mouse songs, geophones let us wonder at elephants using infrasonic rumbles to communicate long distances and find water. And scientific translation tools let us shift the vibrations of the solar wind and aurora or even the patterns of pure math into human scaled songs of the greater universe. We are no longer constrained (or protected) by the ears that evolution has given us. Our auditory world has expanded into an acoustic ecology that contains the entire universe, and the implications of that remain wonderfully unclear.
Exhibit: Home Office
This is a recording made with standard stereo microphones of my home office. Aside from usual typing, mouse clicking and computer sounds, there are a couple of 3D printers running, some music playing, largely an environment you don’t pay much attention to while you’re working in it, yet acoustically very rich if you pay attention.
This sample was made by pitch shifting the frequencies of sonicoffice.wav down so that the ultrasonic moves into the normal human range and cuts off at about 1-2 kHz as if you were hearing with mouse ears. Sounds normally inaudible, like the squealing of the computer monitor cycling on kick in and the high pitched sound of the stepper motors from the 3D printer suddenly become much louder, while the familiar sounds are mostly gone.
This recording of the office was made with a Clarke Geophone, a seismic microphone used by geologists to pick up underground vibration. It’s primary sensitivity is around 80 Hz, although it’s range is from 0.1 Hz up to about 2 kHz. All you hear in this recording are very low frequency sounds and impacts (footsteps, keyboard strikes, vibration from printers, some fan vibration) that you usually ignore since your ears are not very well tuned to frequencies under 100 Hz.
Finally, this sample was made by pitch shifting the frequencies of infrasonicoffice.wav up as if you had grown to elephantine proportions. Footsteps and computer fan noises (usually almost indetectable at 60 Hz) become loud and tonal, and all the normal pitch of music and computer typing has disappeared aside from the bass. (WARNING: The fan noise is really annoying).
The point is: a space can sound radically different depending on the frequency ranges you hear. Different elements of the acoustic environment pop up depending on the type of recording instrument you use (ultrasonic microphone, regular microphones or geophones) or the size and sensitivity of your ears.–
Featured image by Flickr User Jaime Wong.
Seth S. Horowitz, Ph.D. is a neuroscientist whose work in comparative and human hearing, balance and sleep research has been funded by the National Institutes of Health, National Science Foundation, and NASA. He has taught classes in animal behavior, neuroethology, brain development, the biology of hearing, and the musical mind. As chief neuroscientist at NeuroPop, Inc., he applies basic research to real world auditory applications and works extensively on educational outreach with The Engine Institute, a non-profit devoted to exploring the intersection between science and the arts. His book The Universal Sense: How Hearing Shapes the Mind was released by Bloomsbury in September 2012.
REWIND! If you liked this post, check out …
Learning to Listen Beyond Our Ears– Owen Marshall
This is Your Body on the Velvet Underground– Jacob Smith
SO! Amplifies. . .a highly-curated, rolling mini-post series by which we editors hip you to cultural makers and organizations doing work we really really dig. You’re welcome!
On July 18, 2014 all are invited to participate, observe, engage, and celebrate ways of listening with care for our sonic environment in the annual World Listening Day. This year’s theme is “Listen to you!” But first, listen to Eric Leonardson as he reveals the history of World Listening Day and more to kick off SO!’s third annual World Listening Month.
Five years ago Dan Godston came up with the idea for World Listening Day, inspired by the pioneering work of the World Soundscape Project from the 1970s, and its founder, author, and composer R. Murray Schafer. With a group of Chicago-based sound artists and phonographers we started the World Listening Project, a non-profit organization “devoted to understanding the world and its natural environment, societies and cultures through the practices of listening and field recording.” The impetus for the WLP came from Dan as well; he connected us with Bernie Krause, the musician-turned-bio-acoustician now a global advocate for preserving the disappearing natural soundscapes and the species that make them. The World Listening Project began with a confluence of people interested in field recording, media art, experimental music, and ecology with the potential benefits in using the web to present a global soundmap and recorded archive of the world. Connecting with people like Krause who are concerned with sound in the environment continues to lead to new connections and an expanding network of people from many different disciplines and attitudes.
Dan first broached the idea of World Listening Day as a question. He wrote that it “…might be a good occasion to draw attention to listening practices, acoustic ecology, soundscape awareness, and so on.” He noted that there was already a World Listening Awareness Month. But, its focus didn’t include soundscape awareness. As seems to have happened with Earth Day, we were concerned World Listening Day may be no more than a symbolic gesture for what really needs to be a daily effort.
With that unanswerable question hanging, Dan Godston announced the first World Listening Day in June 2010, setting the date as July 18, R. Murray Schafer’s birthday. Through email, the World Listening Project website, and social media, Dan made sure as many people as could were informed. The idea resonated and caught on. At the time I was visiting in Berlin and enjoyed meeting with young artists at their sound art gallery, Berg 26, and their esteemed teacher, Martin Supper. World Listening Day was a perfect vehicle for a project they were already planning. Udo Noll jumped on the idea, too. His radio aporee online soundmap fit right in.
After I returned to the states, the first national conference of the American Society for Acoustic Ecology was held in Chicago. A week afterward, July 18 arrived. Much to our relief and surprise, hundreds of people had responded to Dan’s call for participation, locally, nationally, and internationally. The Nature Sounds and Night Skies Division of the U.S. National Park Service at Fort Collins, Colorado observed it and produced this excellent World Listening Day web page. Each year, Udo Noll creates a special “sonic snapshot of the world” on the aporee.org soundmap site. From Portugal, Luis Antero produces a World Listening Day show on Radio Zero. Public, institutional, and media interest increased in subsequent years. The BBC Radio reported about World Listening Day last year, when I also celebrated with Murray Schafer for his 80th birthday at the Stratford Summer Music Festival, in Ontario.
In our first World Listening Day podcast for Sounding Out!—SO! has joined in observing WLD by hosting a yearly “World Listening Month” Forum since 2012– we learned from Pauline Oliveros, the pioneering composer and founder of Deep Listening, that most folks, including cognitive scientists, still don’t know what listening is. We also highlighted how technologies of recording and concern for environments undisturbed by human activity are bundled together in interesting, divergent ways. Tom, Monica, and I are working on a second podcast to debut here on Sounding Out! on World Listening Day on July 18th 2014, that digs into such concepts as “acoustic identity,” “soundscape composition,” and “listener recognizability,” among others we rarely encounter in everyday conversation.
We hope WLD 2014 will involve even more people and organizations who will notice and spread the word on into the future. Most importantly, we work toward the shared realization that everyday should be World Listening Day.
Toward that end, we reprint the WLD 2014 “official” instructions below. Participation in the past four World Listening Days exceeded our expectations. In this fifth year we anticipate even greater activity and interest. here are 15 days remaining to plan a World Listening Day activity—whether individual, group, or social-media oriented—so jump right into the 2014 World Listening Day activities by emailing firstname.lastname@example.org about your plans. Please be sure to include “World Listening Day” in the subject line or http://www.worldlisteningproject.org/wp-content/uploads/2014/05/2014_WLD_participation_form.pdf” target=”_blank”>download the 2013 World Listening Day participation form here. Thanks!
You are invited to participate in the 2014 World Listening Day an annual global event held on July 18. The purposes of World Listening Day are to:
- Celebrate the listening practices of the world and the ecology of its acoustic environments;
- Raise awareness about the growing number of individual and group efforts that creatively explore the Acoustic Ecology based on the pioneering efforts of the World Soundscape Project, World Forum for Acoustic Ecology, La Semaine du Son, Deep Listening Institute, among many others;
- Design and implement educational initiatives that explore these concepts and practices.
This year’s theme for World Listening Day is “Listen To You!” Some questions to consider:
- How do you make yourself heard by others?
- How do you listen and what do you hear when you want to be unseen?
- How might the sounds you produce adapt to your nearby environment?
- What might a “listening ethic” be?
- How might such an ethic apply particularly to understanding the relationship between humans and other living creatures?
World Listening Day is co-organized by the World Listening Project (WLP) and the Midwest Society for Acoustic Ecology (MSAE). July 18 was chosen because it is the birthday of Canadian composer R. Murray Schafer, founder of the World Soundscape Project and author of the seminal book, The Tuning of the World.
WLP and MSAE invite you to participate in the 2014 World Listening Day, on Friday, July 18 and also through the week of July 14th-20th.
Some suggestions on how you can participate and organize may be:
- A soundwalk or a listening party with people who make, listen, and discuss field recordings.
- A performance event that explores your soundscape and how we can listen to our sonic environment.
- A private / solitary way, by listening attentively to your soundscape.
- An educational event that relates to acoustic ecology, field recordings, or a similar topic.
- Contact local groups participating in World Listening Day and get involved.
Eric Leonardson is a Chicago-based audio artist and teacher. He has devoted a majority of his professional career to unorthodox approaches to sound and its instrumentation with a broad understanding of texture, atmosphere and microtones. He is President of the World Forum for Acoustic Ecology and founder of the Midwest Society for Acoustic Ecology, and Executive Director of the World Listening Project. Leonardson is an Adjunct Associate Professor in the Department of Sound at The School of the Art Institute of Chicago.
[Editor's Note: Both Eric Leonardson and Jennifer Stoever will be speaking at the Invisible Places Sounding Cities: Sound, Urbanism, and a Sense of Place conference in Viseu, Portugal on World Listening Day 2014. Here is the website: http://invisibleplaces.org/.]
Featured image: “Dancing Mania @ Mlbk” by Flickr user Lieven Soete, CC BY-NC-SA 2.0
REWIND!…If you liked this post, you may also dig:
Wayback Sound Machine: Sound Through Time, Space, and Place- Maile Colbert
A Listening Mind: Sound Learning in a Literature Classroom- Nicole Furlonge
SO! Amplifies. . .a highly-curated, rolling mini-post series by which we editors hip you to cultural makers and organizations doing work we really really dig. You’re welcome!
During a preview of the UK tour of my performance Beacons, I noticed something intriguing. There were a couple of children in the audience, and they were singing along with me. I mean, not just singing – doing all kinds of vocal things along with me.
You have to understand that I’m an extended – or I now prefer the term extra-normal – vocalist. I make artwork from the ‘weird’ qualities of the voice. Now, the term weird, when it comes to the voice, is always culturally relative. But let’s face it – in almost any of the industrialized cultures, we don’t exactly let the voice go wild, and so the category of “weird” is pretty large and full. Our cultures are a Freudian fiesta of vocal repression and sublimation. When’s the last time you heard a heartfelt, loose, easy series of adult sobs at a business meeting, or a wail of deep mourning at a funeral? When it comes to the voice, we’re pretty uptight.
However, here were some children absolutely letting loose and going along for the ride. This has happened before. During my piece Soie Soyeuse, during which I wrap the audience in habotai silk, some children got right into it. I was five feet away making really incredibly ‘ugly’ chewing noises and groans of gnashing teeth and they delightedly voiced along.
Right now, everything’s becoming user-driven. The audience loves to co-make art. And why not? So I thought: let’s go for broke, here. Let’s try to create some environments and experiences in which kids (and the kid in all of us) can really explore what their voices can do. Let’s give them freedom to experiment. Let’s reward them for wildness. Let’s try to make it child and family friendly, but beautiful and rich and somehow sensual, despite the limitations of the tablet format. I worked like a dog to get financing. I collaborated with scientists. And so, Voice Bubbles for iPad was born.
It’s free, thanks to our funders. It’s got no silly pop up ads and in-app purchases. Kids love it and will use it for hours, happily. It lets them record, improvise, structure and add effects to their voices. They can play back little sequences or structured phrases. They can fool with what else their voices can do and be.
It makes them feel, and makes we adults question, just why we need to keep the voice so under wraps when it’s a source of so much playful joy and intense engagement. Give it to a child to enjoy. Enjoy it yourself.
*You can explore other related artworks at www.yourvivaciousvoice.com
Yvon Bonenfant is Reader in Performing Arts at the University of Winchester. He likes voices that do what voices don’t usually do, and he likes bodies that don’t do what bodies usually do. He makes art starting from these sounds and movements. These unusual, intermedia works have been produced in 10 countries in the last 10 years, and his writing published in journals such as Performance Research, Choreographic Practices, and Studies in Theatre and Performance. He currently holds a Large Arts Award from the Wellcome Trust and funding from Arts Council England to collaborate with speech scientists on the development of a series of participatory, extra-normal voice artworks for children and families; see www.yourvivaciousvoice.com. Despite his air of Lenin, he does frighteningly accurate vocal imitations of both Axl Rose and Jon Bon Jovi. www.yvonbonenfant.com.
REWIND!…If you liked this post, you may also dig:
SO! Amplifies: Regina Bradley’s Outkasted Conversations– Regina Bradley
This is Your Body on the Velvet Underground–Jacob Smith
Digital Analogies: Techniques of Sonic Play–Roger Moseley