Archive | Technology RSS for this section

Someplaces: Radio Art, Transmission Ecology and Chicago’s Radius

Jeff Kolar with Radius' mobile transmitter, the Audio Relay Unit, on the shore of Lake Michigan.

Magnavox_AM2

This week Sounding Out! is proud to present the first post in Radio Art Reflections, a three part series curated by radio artist and senior radio lecturer at Canterbury Christ Church University Magz Hall. Focusing on innovative approaches to radio art, the series will bring together three leading practitioners who have been researching the field from Canada, Australia and the U.K.

We begin with a fascinating exploration of “transmission ecologies” in recent works in Chicago, Iceland and elsewhere, written by Canadian sound and radio artist Anna Friz — one of the most exciting radio artists working today — who discusses how transmission art has shaped her practice.

– Special Editor Neil Verma

From the early avant-garde Futurists to present-day, utopian dreams litter the history of art meeting technology. When it comes to radio and wireless, these often include the dreams that each new technology will conquer space and time; that the overcoming of distance will enable the symbiosis of human with machine and the union of self with other, while the overcoming of time will bring about a simultaneity of experience. For many radio and transmission artists (myself included), our work with so-called “trailing edge” media seeks to critically engage these myths, positing wireless transmissions instead as time-based, site-specific encounters between people and devices over distances small or large, where the materiality of the electro-magnetic spectrum is experienced within a constantly shifting transmission ecology in which we all, people and devices, function.

If one hallmark of radio art is the desire to appropriate broadcasting by rethinking and re-using technologies of transmission and reception in service of crafting new mythologies and futures for the medium. Artists have long questioned the policies and norms established by state and market around radio broadcasting which delimit experimentation and autonomous practices. Bertolt Brecht‘s call in 1932 for radio to exceed its one-to-many broadcast format in favor of a democratized, transceptive (or many-to-many) medium still resonates with contemporary artists and activists alike. What else could radio become, we ask, if not only a disseminator of information and entertainment, acoustic or digital? If radio so far has largely acted as an accomplice in the industrialization of communications, artistic appropriations of radio can destabilize this process with renewed explorations of radio and electromagnetic phenomena, constructions of temporary networks small or large, and radical explorations of broadcast beyond the confines of programming and format norms.

My first transmitter, built on the Tetsuo Kogawa model, as modified by Bobbi Kozinuk, 1998;

My first transmitter, built on the Tetsuo Kogawa model, as modified by Bobbi Kozinuk, 1998;

Curators, producers and art historians typically describe radio art as the use of radio as an artistic medium, which is to say, art created specifically for the technical and cultural circumstances of broadcast, and which considers these circumstances as artistic material. Today these circumstances have exceeded terrestrial broadcast to include satellite, online, and on-demand forms; similarly radio art has also expanded to include sprawling telematic art exchanges, online podcast series, and unlicensed temporary interventions into the radio dial. As a further reclamation of radio as a medium, many artists pull radio out of the studio to create installations, performance works and public actions which consider not just the act of transmission or the creation of artistic content, but also the material aspects of the electro-magnetic spectrum, and the circuits of people and devices which activate and reveal them.

Japanese media theorist and artist Tetsuo Kogawa describes broadcast radio art as art radio, where art is the content of a transmission. By contrast, radio art involves directly playing with electro-magnetic waves as the artistic medium. Galen Joseph-Hunter of Wave Farm further expands Kogawa’s formulation of radio art with the term transmission art, so as to include audio visual broadcast media and artistic activities across the entire electro-magnetic spectrum, such as work with Very Low Frequency (VLF) and Extremely Low Frequency (ELF) waves, or high frequency wireless networks. These definitions of radio and transmission art emphasize that radio is not a container for content, but is defined as relationships between people and things, occurring in the context of the electro-magnetic spectrum within a transmission ecology.

I apply the term transmission ecology in reference to both the symbolic spaces of cultural production such as a radio station, and to the invisible but very material space of dynamic electromagnetic interactions, both of which feature the collaboration between people and things. Transmission ecology asks more than “who owns the airwaves” by questioning the shifting relationships between all actors in the environment, from human to device to localized weather system to nearby star, and thus is not defined by homeostasis but by constant change. These relationships also support a theory of technology where people are not the absolute controllers of things, but where a push and pull of collaboration occurs within complex material and cultural environments.

Photo of Respire by Anna Friz, a large installation of radios from Nuit Blanche Toronto, 2009. 

Photo of Respire by Anna Friz, a large installation of radios from Nuit Blanche Toronto, 2009.

All activities in the electro-magnetic spectrum form ecologies in relation to one another conceptually, performatively, and materially. Consider the Radia network, an international alliance of independent radio stations who share radio art programming as an alternate transmission ecology within the broader culture of private broadcast radio stations. Another kind of ecology is formed by radio receivers all playing the same station diffused across countless cars and households, as they function in relation to other kinds of wireless devices and electronic systems nearby. Such a muster of receivers can be physically brought together, for instance, in a multi-channel radio installation, to reveal the complex relationships among devices, as each receiver also becomes a sender by electronically effecting its neighbor. A mobile phone receiving wireless internet likewise functions within the instability inherent in the surrounding transmission ecology shaped by all aspects of the built environment, such as the electrical grid and other urban infrastructure, as well as weather or time of day or solar flares. Human bodies and devices alike register the invisible electromagnetic activity that surrounds us as physical, measurable, and affective.

With this in mind consider radio art as occupying “radio space,” a continuous, available, fluctuating area described by the reach of signals within overlapping fields of influence and the space of imagination that invisible territory enables. The extrasensory nature of radio space allows for a productive slippage between real material signals and audible imaginary landscapes. Many radio art and transmission art works specifically draw attention to the transmission ecology in order to question the naturalization of mainstream communications systems, the normalization of practices within those systems, and the pervasiveness of electrical infrastructure, proposing alternate narratives and experiences.

So what is some of this work like? In the past year I have had the pleasure to work with Chicago-based Radius, an experimental radio-based platform which curates monthly episodes broadcast locally using the Audio Relay Unit, an unlicensed autonomous low-watt FM radio transmitter system developed in 2002 by Temporary Services and the Intermod Series. Radius neatly unites radio and transmission art by embracing the production of artistic content for broadcast, sampling existing content for artistic expression, and artistic use of the electro-magnetic spectrum generally. Radius functions as an intermittent exhibition space and as an intervention into the predictable daily grind of the FM dial. Artists compose their pieces specifically for the interference-prone radio space where their work may only be heard in fragments, as the instability and fluctuations of the relatively small Radius signal in relation to the big commercial stations broadcasting from downtown all form the context for experiencing the radio art works. Radius broadcasts one episode per month, on a schedule determined by the artist, with pieces varying in length and repetition, and some following a strict schedule related to cosmic or social timing.

Friz_Transmission_from_skaftfell1

Photo of my private transmitter + antenna pointing out the window in Seydisfjördur, Iceland

Recently I crafted an episode for Radius while on an artist residency in Seydisfjördur, Iceland. The town was the site of the first trans-Atlantic telegraph link between Europe and Iceland in 1906, which was also the year that Reginald Fessenden first broadcast a human voice over radio from his workshop in Brant Rock, Massachusetts. Iceland is remote enough that the electro-magnetic ‘pollution’ from human signal activity is notably absent, and located far enough to the north that in October the light disappears rapidly, so that each day loses eight minutes of daylight. The piece was called Radiotelegraph, a beacon crafted from spoken morse code and sampled signals, then sent from north to south, simulcast on my own low-watt FM transmitter in Seydisfjördur at sundown each day as well as on Radius in Chicago. The transmission marked time passing, beginning earlier each day as it followed the path of the sun. My intention was not to overcome but to experience and recuperate distance through the relation of a remote radio outpost to another minor outpost further south within a metropolis; to hear distance and feel it; to understand that distance, however finite, is a necessary condition for communication and relationship, and that distance is the key ingredient of situated, time-based, spatialized sonic experiences.

Here is the Radius episode featuring Radiotelegraph:

.

Power station across the parking lot from the Radius studio at MANA Contemporary, Chicago

Power station across the parking lot from the Radius studio at MANA Contemporary, Chicago

As part of a recent yearly theme on “Grids” Radius tackled the electro-magnetic field space of the city by inviting four artists to create new works to be performed near power stations. In his piece electrosmog, Canadian artist Kristen Roos utilized a high frequency receiver to sonify signal activity in the 800 MHz – 2.5 GHz range, which includes mobile phones, wireless phones, wifi, and microwaves. His site-specific performance took place overlooking the Fisk Generating Station in Chicago, and included microwave ovens and micro-watt transmission to a sound system made of radio receivers. Thus the work was site-specific to both the transmission ecology of urban Chicago and the field effects of the electrical grid, mixing material signals with a speculative approach as to what the cumulative effects of living in this built environment characterized by centralized power could be. In Roos’ work, radio space contextualized and revealed the real–though naturalized and often invisible–relationships between people, things, and systems, where a microwave oven gestured at both danger and musicality.

Listen to the Roos piece here:

.

Kristen Roos' set up for his Grids performance

Kristen Roos’ set up for his Grids performance

These radio art works enact places in radiophonic space, and experiment with transmission to question the status quo of how the airwaves are controlled and used. As radio trickster Gregory Whitehead notes, it is position, not sound, that matters most with regard to radio. Artists remain committed to making radiophonic someplaces, however temporarily constructed, inhabited by interpenetrating and overlapping fields and bodies.

Featured Image: Jeff Kolar with Radius’ mobile transmitter, the Audio Relay Unit, on the shore of Lake Michigan. Used with permission.

Anna Friz is a Canadian sound and radio artist who specializes in multi-channel transmission systems for installation, performance, and broadcast. Anna holds a Ph.D. in Communication and Culture from York University, Toronto, and recently completed a post-doctoral fellowship in the Sound Department at the School of the Art Institute of Chicago. She has performed and exhibited widely across North America, South America, and Europe, and her radio art/works have been heard on the airwaves of more than 25 countries. She is a steering member of the artist collective Skálar |Sound Art | Experimental Music based in Iceland.

nicelittlestatic.com

tape reelREWIND! . . .If you liked this post, you may also dig:

Catastrophic Listening — China Blue

Maile Colbert, Rui Costa, and Jeff Cain’s “Radio Terramoto” — Maile Colbert

Yvon Bonenfant’s Voice Bubbles App — Yvon Bonenfant

The Sonic Roots of Surveillance Society: Intimacy, Mobility, and Radio

Radio for Backup - 92/365 - 2 April 2013

Sound and Surveilance4

It’s an all too familiar movie trope. A bug hidden in a flower jar. A figure in shadows crouched listening at a door. The tape recording that no one knew existed, revealed at the most decisive of moments. Even the abrupt disconnection of a phone call manages to arouse the suspicion that we are never as alone as we may think. And although surveillance derives its meaning the latin “vigilare” (to watch) and French “sur-“ (over), its deep connotations of listening have all but obliterated that distinction.

Rounding out our series on surveillance, Kathleen Battles offers a historical perspective that shows how early twentieth century crime drama naturalized practices of citizen surveillance. A million eyes were activated as millions of listeners learned that the immediacy of radio and telephone allowed for an unprecedented level of participation in law enforcement. Calling all cars…Calling all cars… -AT

Police Headquarters, a 1932 radio crime drama, was produced in the infancy of narrative radio. Containing barely 12 minutes of narrative content, the program opened each episode with a repeating segment of call, connection, and dispatch to quickly establish both the crime committed and how the police responded to it. For example, in the “Payroll Robbery” episode it takes just over a minute and a half to hear a phone call to the titular headquarters, its connection to the proper unit, a radio call to a specific police car, and the responding officers arriving at the assigned location. Compared with the graphically and visually intense images of modern surveillance in contemporary popular culture, this brief exchange no doubt sounds quaint, simplistic, and even banal. After all, radios, cars, and telephones have served as the routine backdrop of most police dramas for some 70 years. But in 1932 the interlinking of these technologies was factually, as well as imaginatively, novel. This essay shows how radio, as “new media,” was central to imagining surveillance in sonic terms, prefiguring many features of contemporary surveillance practices.

"Kindergarden of the Air," by the Australian Broadcasting Corporation CC BY-NC-ND.

“Kindergarden of the Air,” by the Australian Broadcasting Corporation CC BY-NC-ND.

The introduction of radio and cars into police work took place in the first decades of the twentieth century, especially during the years between the two world wars that Richard Popp (2011) calls “the machine age”. He argues that this period witnessed vast transformations brought on by cars and radios, which, when combined with existing technologies like the telephone, forged new communication networks that transformed both work and leisure. These changes were central to the narratives of criminality and policing that emerged during the interwar years. Police were the focus of radio dramas, including Police Headquarters and Calling All Cars. These dramas played with the intermingling of automobility, telephony, and radio in ways that spoke to the main problems police forces saw themselves facing: organized, professional, mobile, machine-age criminals. Cars, telephones, and automobiles were not just tools to criminals, but they were also the building blocks for a machine age surveillance made possible by the sonic power of radio.

Recently, Robin James (2014) has suggested that the acousmatic is a useful metaphor for understanding the emerging practices of data based surveillance. Acousmatic surveillance listens for patterns in “ambient data environments” instead of profiling individuals in the panoptic sense. At the turn of the twentieth century, radio allowed for a panacoustic mastery of spaces that bridged both panoptic and acousmatic surveillance. Radio also speaks to another key feature of information age surveillance, what Mark Andrejevic (2011) describes as the “redoubling of tools for communication and leisure as technologies for surveillance and security.” (165-66) The technical capabilities and imaginative potentials of radio help us to consider it both as a police technology and entertainment medium. The sonic power of radio was often figured as an “Invisible Man Hunter,” whose realignment of spatial and temporal arrangements rendered criminal escape impossible. As an entertainment medium, radio’s aurality was key to understanding its imaginative potential as highly intimate and mobile: invasive and expansive.

In Police Headquarters we hear how radio’s sonic and aural qualities come together. Radio acts as the link between the telephone and car, allowing for a swift response to a citizen request. The tactical use of sound effects and narrative compression in the broadcast situate the listener inside a machine like apparatus that presents the police as always available. At their broadest level, radio crime dramas aurally situate communication and transportation technologies, like radio, as key to both the narrative organization of the story and as a plot element. In the opening to the “Stop That Car” episode of Calling All Cars, a dispatcher advises for cars to be on the lookout for a specific car involved in a hit and run, including the address of the crime and possible location of the vehicle. Overlaid with sound effects made to signify a car, these openings situate listeners as riders eavesdropping on the adventures of mobile police officers. As the program’s title suggests, each episode opened with a police radio call, often voiced by real life LAPD dispatcher, Jessie Rosenquist. The program’s sponsor was the gasoline company that supplied the fuel for LAPD patrol cars – further linking cars and radio as a key theme. In the opening to the “Two Man Crime Wave Episode,” the very ad for the product is performed as a police radio call.

The conceit of eavesdropping on a police adventure did much to link private life and the police. This theme runs tandem to radio’s sonic immediacy, which allowed listeners to imagine a seemingly instantaneous response to citizen phone calls.. For example, the “July 4th in a Radio Car” episode of Calling All Cars situates radio listeners as sonic participants, able to ride along with police from the comfort of their own living rooms. Here, cars respond to a number of calls made by private citizens that bring the policing function into daily life. There is even one call that involves domestic violence, in particular. Throughout the episode, police are situated as an available force – thanks to the telephone, radio, and automobile – to adjudicate all manner of private disputes. In these particular instances the intimacy of radio as a machine age technology is “redoubled” with radio as a police technology. Radio’s intimate address allowed the voices of police officers to enter the private space of the home. At the same time, the machinery of crime fighting required citizen participation, most often figured through the phone call from within the space of the home to the police.

"Radio Police Automation 1924," by Robert Wade, CC BY-NC-SA.

“Radio Police Automation 1924,” by Robert Wade, CC BY-NC-SA.

If the intimacy of radio served to cross the divide between public and private, the spatial-temporal collapse achieved by radio made it ideal for sonically monitoring great swaths of space. Intimately linked with cars, radio was understood as especially mobile. Radio’s ability to compress the relationship between space and time is frequently dramatized. For example, in the “Crime vs. Time” episode of Calling All Cars, the host explains how radio has rendered the average response time to a police call for help only two minutes and forty seconds. The episode then proceeds to show how radio was used by the police to track and apprehend two men who robbed a movie theater. Representing phone and radio calls, while specifically referencing geographic locations, radio dramatists used radio’s aural dimensions to render radio’s sonic power of surveillance. Capable of reaching everywhere, police radio, when linked with the telephone and automobile, could be used to sonically pinpoint any somewhere that a criminal might try to escape to.

In our era of high-tech and sophisticated technologies, visually rich narratives, and algorithmically driven methods of tracking, there is certainly something simple, comforting, and even nostalgic in these tales of low-tech machine age criminal apprehension. Depression era true crime dramas certainly do not offer the kind of sophistication as information age narratives, such as The Wire. The medium’s sonic qualities were key to linking together its use as both a police instrument and as a domestic entertainment technology. With sonic forces invisibly and silently crossing the into intimate domestic spaces and covering large swaths of territory, radio became key to imagining many features that we take for granted as “new” about the information age: the control of movement across space, the constant availability of communicative connection, the promise of perfect coordination across a field of institutional actors, the marshaling of citizen participation in surveillance efforts, and the construction of increasingly intimate links between domestic life and law enforcement procedures. In serving as central node in refiguring shifting notions of space and time that existing institutions were not prepared to handle, radio’s sonic qualities remapped the meaning of police work and helped to establish a relationship between the police and citizen body that still resonates today. While not as technologically sophisticated, machine age policing and police narratives took advantage of radio’s double function as both a machine of coordination and medium of entertainment to extend the policing function into more areas of life. This reflects a sonic mode of power that allows neither the interior space of the home nor the exterior world of the road relief from police presence. This moment of technological interconnection, however, evoked the excitement and anxiety that made sonic surveillance at once thrilling and calming; a salve to soothe the woes of a world that now seemed intensely close and impossibly far flung. Is it any wonder that while some fret over the power of corporate and state dataveillance today, that some continue to find comfort in the possibility of being recognized as someone in a world of ever more intense interconnection?

Featured image “Radio for Backup” by Jonathan Flinchbaugh CC BY-NC-SA.

Kathleen Battles is Associate Professor and Graduate Director in the Department of Communication and Journalism at Oakland University. Her research focuses on radio history, especially as it relates to issues of policing, sound and surveillance, questions concerning technology and culture, and sexuality and the media. She is the author of Calling All Cars: Radio Dragnets and the Technology of Policing (University of Minnesota Press, 2010); co-editor (with Joy Hayes and Wendy Hilton-Morrow) of War of the Worlds to Social Media: Mediated Communication in Times of Crisis (Peter Lang, 2013); and co-author (with Wendy Hilton-Morrow) of Sexual Identities and the Media: An Introduction (Routledge, 2015). In addition, her work has appeared in Critical Studies in Media Communication, The Radio Journal, and the Journal of Homosexuality.

tape reelREWIND!…If you liked this post, check out:

Acousmatic Surveillance and Big Data-Robin James

The Dark Side of Game Audio: The Sounds of Mimetic Control and Affective ConditioningAaron Trammell

À qui la rue?: On Mégaphone and Montreal’s Noisy Public Sphere-Lilian Radovac


Sounds of Science: The Mystique of Sonification

NYU_full

Hearing the Unheard IIWelcome to the final installment of Hearing the UnHeardSounding Out!s series on what we don’t hear and how this unheard world affects us. The series started out with my post on hearing, large and small, continued with a piece by China Blue on the sounds of catastrophic impacts, and Milton Garcés piece on the infrasonic world of volcanoes. To cap it all off, we introduce The Sounds of Science by professor, cellist and interactive media expert, Margaret Schedel.

Dr. Schedel is an Associate Professor of Composition and Computer Music at Stony Brook University. Through her work, she explores the relatively new field of Data Sonification, generating new ways to perceive and interact with information through the use of sound. While everyone is familiar with informatics, graphs and images used to convey complex information, her work explores how we can expand our understanding of even complex scientific information by using our fastest and most emotionally compelling sense, hearing.

– Guest Editor Seth Horowitz

With the invention of digital sound, the number of scientific experiments using sound has skyrocketed in the 21st century, and as Sounding Out! readers know, sonification has started to enter the public consciousness as a new and refreshing alternative modality for exploring and understanding many kinds of datasets emerging from research into everything from deep space to the underground. We seem to be in a moment in which “science that sounds” has a special magic, a mystique that relies to some extent on misunderstandings in popular awareness about the processes and potentials of that alternative modality.

For one thing, using sound to understand scientific phenomena is not actually new. Diarist Samuel Pepys wrote about meeting scientist Robert Hooke in 1666 that “he is able to tell how many strokes a fly makes with her wings (those flies that hum in their flying) by the note that it answers to in musique during their flying.” Unfortunately Hooke never published his findings, leading researchers to speculate on his methods. One popular theory is that he tied strings of varying lengths between a fly and an ear trumpet, recognizing that sympathetic resonance would cause the correct length string to vibrate, thus allowing him to calculate the frequency. Even Galileo used sound, showing the constant acceleration of a ball due to gravity by using an inclined plane with thin moveable frets. By moving the placement of the frets until the clicks created an even tempo he was able to come up with a mathematical equation to describe how time and distance relate when an object falls.

Illustration from Robert Hooke's Micrographia (1665)

Illustration from Robert Hooke’s Micrographia (1665)

There have also been other scientific advances using sound in the more recent past. The stethoscope was invented in 1816 for auscultation, listening to the sounds of the body. It was later applied to machines—listening for the operation of the technological gear. Underwater sonar was patented in 1913 and is still used to navigate and communicate using hydroacoustic phenomenon. The Geiger Counter was developed in 1928 using principles discovered in 1908; it is unclear exactly when the distinctive sound was added. These are all examples of auditory display [AD]; sonification-generating or manipulating sound by using data is a subset of AD. As the forward to the The Sonification Handbook states, “[Since 1992] Technologies that support AD have matured. AD has been integrated into significant (read “funded” and “respectable”) research initiatives. Some forward thinking universities and research centers have established ongoing AD programs. And the great need to involve the entire human perceptual system in understanding complex data, monitoring processes, and providing effective interfaces has persisted and increased” (Thomas Hermann, Andy Hunt, John G. Neuhoff, Sonification Handbook, iii)

Sonification clearly enables scientists, musicians and the public to interact with data in a very different way, particularly compared to the more numerous techniques involving vision. Indeed, because hearing functions quite differently than vision, sonification offers an alternative kind of understanding of data (sometimes more accurate), which would not be possible using eyes alone. Hearing is multi-directional—our ears don’t have to be pointing at a sound source in order to sense it. Furthermore, the frequency response of our hearing is thousands of times more accurate than our vision. In order to reproduce a moving image the sampling rate (called frame-rate) for film is 24 frames per second, while audio has to be sampled at 44,100 frames per second in order to accurately reproduce sound. In addition, aural perception works on simultaneous time scales—we can take in multiple streams of audio data at once at many different dynamics, while our pupils dilate and contract, limiting how much visual data we can absorb at a single time. Our ears are also amazing at detecting regular patterns over time in data; we hear these patterns as frequency, harmonic relationships, and timbre.

Image credit: Dr. Kevin Yager, data measured at X9 beamline, Brookhaven National Lab.

Image credit: Dr. Kevin Yager, Brookhaven National Lab.

But hearing isn’t simple, either. In the current fascination with sonification, the fact that aesthetic decisions must be made in order to translate data into the auditory domain can be obscured. Headlines such as “Here’s What the Higgs Boson Sounds Like” are much sexier than headlines such as “Here is What One Possible Mapping of Some of the Data We Have Collected from a Scientific Measuring Instrument (which itself has inaccuracies) Into Sound.” To illustrate the complexity of these aesthetic decisions, which are always interior to the sonification process, I focus here on how my collaborators and I have been using sound to understand many kinds of scientific data.

My husband, Kevin Yager, a staff scientist at Brookhaven National Laboratory, works at the Center for Functional Nanomaterials using scattering data from x-rays to probe the structure of matter. One night I asked him how exactly the science of x-ray scattering works. He explained that X-rays “scatter” off of all the atoms/particles in the sample and the intensity is measured by a detector. He can then calculate the structure of the material, using the Fast Fourier Transform (FFT) algorithm. He started to explain FFT to me, but I interrupted him because I use FFT all the time in computer music. The same algorithm he uses to determine the structure of matter, musicians use to separate frequency content from time. When I was researching this post, I found a site for computer music which actually discusses x-ray scattering as a precursor for FFT used in sonic applications.

To date, most sonifications have used data which changes over time – a fly’s wings flapping, a heartbeat, a radiation signature. Except in special cases Kevin’s data does not exist in time – it is a single snapshot. But because data from x-ray scattering is a Fourier Transform of the real-space density distribution, we could use additive synthesis, using multiple simultaneous sine waves, to represent different spatial modes. Using this method, we swept through his data radially, like a clock hand, making timbre-based sonifications from the data by synthesizing sine waves using with the loudness based on the intensity of the scattering data and frequency based on the position.

We played a lot with the settings of the additive synthesis, including the length of the sound, the highest frequency and even the number of frequency bins (going back to the clock metaphor – pretend the clock hand is a ruler – the number of frequency bins would be the number of demarcations on the ruler) arriving eventually at set of optimized variables.

Here is one version of the track we created using 10 frequency bins:

.

Here is one we created using 2000:

.

And here is one we created using 50 frequency bins, which we settled on:

.

On a software synthesizer this would be like the default setting. In the future we hope to have an interactive graphic user interface where sliders control these variables, just like a musician tweaks the sound of a synth, so scientists can bring out, or mask aspects of the data.

To hear what that would be like, here are a few tracks that vary length:

.

.

.

Finally, here is a track we created using different mappings of frequency and intensity:

.

Having these sliders would reinforce to the scientists that we are not creating “the sound of a metallic alloy,” we are creating one sonic representation of the data from the metallic alloy.

It is interesting that such a representation can be vital to scientists. At first, my husband went along with this sonification project as more of a thought experiment rather than something that he thought would actually be useful in the lab, until he heard something distinct about one of those sounds, suggesting that there was a misaligned sample. Once Kevin heard that glitched sound (you can hear it in the video above), he was convinced that sonification was a useful tool for his lab. He and his colleagues are dealing with measurements 1/25,000th the width of a human hair, aiming an X-ray through twenty pieces of equipment to get the beam focused just right. If any piece of equipment is out of kilter, the data can’t be collected. This is where our ears’ non-directionality is useful. The scientist can be working on his/her computer and, using ambient sound, know when a sample is misaligned.

procedure

It remains to be seen/heard if the sonifications will be useful to actually understand the material structures. We are currently running an experiment using Mechanical Turk to determine this kind of multi-modal display (using vision and audio) is actually helpful. Basically we are training people on just the images of the scattering data, and testing how well they do, and training another group of people on the images plus the sonification and testing how well they do.

I’m also working with collaborators at Stony Brook University on sonification of data. In one experiment we are using ambisonic (3-dimensional) sound to create a sonic map of the brain to understand drug addiction. Standing in the middle of the ambisonic cube, we hope to find relationships between voxels, a cube of brain tissue—analogous to pixels. When neurons fire in areas of the brain simultaneously there is most likely a causal relationship which can help scientists decode the brain activity of addiction. Computer vision researchers have been searching for these relationships unsuccessfully; we hope that our sonification will allow us to hear associations in distinct parts of the brain which are not easily recognized with sight. We are hoping to leverage the temporal pattern recognition of our auditory system, but we have been running into problems doing the sonification; each slice of data from the FMRI has about 300,000 data points. We have it working with 3,000 data points, but either our programming needs to get more efficient, or we have to get a much more powerful computer in order to work with all of the data.

On another project we are hoping to sonify gait data using smartphones. I’m working with some of my music students and a professor of Physical Therapy, Lisa Muratori, who works on understanding the underlying mechanisms of mobility problems in Parkinsons’ Disease (PD). The physical therapy lab has a digital motion-capture system and a split-belt treadmill for asymmetric stepping—the patients are supported by a harness so they don’t fall. PD is a progressive nervous system disorder characterized by slow movement, rigidity, tremor, and postural instability. Because of degeneration of specific areas of the brain, individuals with PD have difficulty using internally driven cues to initiate and drive movement. However, many studies have demonstrated an almost normal movement pattern when persons with PD are provided external cues, including significant improvements in gait with rhythmic auditory cueing. So far the research with PD and sound has be unidirectional – the patients listen to sound and try to match their gait to the external rhythms from the auditory cues.In our system we will use bio-feedback to sonify data from sensors the patients will wear and feed error messages back to the patient through music. Eventually we hope that patients will be able to adjust their gait by listening to self-generated musical distortions on a smartphone.

As sonification becomes more prevalent, it is important to understand that aesthetic decisions are inevitable and even essential in every kind of data representation. We are so accustomed to looking at visual representations of information—from maps to pie charts—that we may forget that these are also arbitrary transcodings. Even a photograph is not an unambiguous record of reality; the mechanics of the camera and artistic choices of the photographer control the representation. So too, in sonification, do we have considerable latitude. Rather than view these ambiguities as a nuisance, we should embrace them as a freedom that allows us to highlight salient features, or uncover previously invisible patterns.

__

Margaret Anne Schedel is a composer and cellist specializing in the creation and performance of ferociously interactive media. She holds a certificate in Deep Listening with Pauline Oliveros and has studied composition with Mara Helmuth, Cort Lippe and McGregor Boyle. She sits on the boards of 60×60 Dance, the BEAM Foundation, Devotion Gallery, the International Computer Music Association, and Organised Sound. She contributed a chapter to the Cambridge Companion to Electronic Music, and is a joint author of Electronic Music published by Cambridge University Press. She recently edited an issue of Organised Sound on sonification. Her research focuses on gesture in music, and the sustainability of technology in art. She ran SUNY’s first Coursera Massive Open Online Course (MOOC) in 2013. As an Associate Professor of Music at Stony Brook University, she serves as Co-Director of Computer Music and is a core faculty member of cDACT, the consortium for digital art, culture and technology.

Featured Image: Dr. Kevin Yager, data measured at X9 beamline, Brookhaven National Lab.

Research carried out at the Center for Functional Nanomaterials, Brookhaven National Laboratory, is supported by the U.S. Department of Energy, Office of Basic Energy Sciences, under Contract No. DE-AC02-98CH10886.

tape reelREWIND! ….. If you liked this post, you might also like:

The Noises of Finance–Nicholas Knouf

Revising the Future of Music Technology–Aaron Trammell

A Brief History of Auto-Tune–Owen Marshall

Erratic Furnaces of Infrasound: Volcano Acoustics

Surface flows as seen by thermal cameras at Pu’u O’o crater, June 27th, 2014. Image: USGS

Hearing the Unheard IIWelcome back to Hearing the UnHeard, Sounding Out‘s series on how the unheard world affects us, which started out with my post on hearing large and small, continued with a piece by China Blue on the sounds of catastrophic impacts, and now continues with the deep sounds of the Earth itself by Earth Scientist Milton Garcés.

Faculty member at the University of Hawaii at Manoa and founder of the Earth Infrasound Laboratory in Kona, Hawaii, Milton Garces is an explorer of the infrasonic, sounds so low that they circumvent our ears but can be felt resonating through our bodies as they do through the Earth. Using global networks of specialized detectors, he explores the deepest sounds of our world from the depths of volcanic eruptions to the powerful forces driving tsunamis, to the trails left by meteors through our upper atmosphere. And while the raw power behind such events is overwhelming to those caught in them, his recordings let us appreciate the sense of awe felt by those who dare to immerse themselves.

In this installment of Hearing the UnHeard, Garcés takes us on an acoustic exploration of volcanoes, transforming what would seem a vision of the margins of hell to a near-poetic immersion within our planet.

– Guest Editor Seth Horowitz

The sun rose over the desolate lava landscape, a study of red on black. The night had been rich in aural diversity: pops, jetting, small earthquakes, all intimately felt as we camped just a mile away from the Pu’u O’o crater complex and lava tube system of Hawaii’s Kilauea Volcano.

The sound records and infrared images captured over the night revealed a new feature downslope of the main crater. We donned our gas masks, climbed the mountain, and confirmed that indeed a new small vent had grown atop the lava tube, and was radiating throbbing bass sounds. We named our acoustic discovery the Uber vent. But, as most things volcanic, our find was transitory – the vent was eventually molten and recycled into the continuously changing landscape, as ephemeral as the sound that led us there in the first place.

Volcanoes are exceedingly expressive mountains. When quiescent they are pretty and fertile, often coyly cloud-shrouded, sometimes snowcapped. When stirring, they glow, swell and tremble, strongly-scented, exciting, unnerving. And in their full fury, they are a menacing incandescent spectacle. Excess gas pressure in the magma drives all eruptive activity, but that activity varies. Kilauea volcano in Hawaii has primordial, fluid magmas that degass well, so violent explosive activity is not as prominent as in volcanoes that have more evolved, viscous material.

Well-degassed volcanoes pave their slopes with fresh lava, but they seldom kill in violence. In contrast, the more explosive volcanoes demolish everything around them, including themselves; seppuku by fire. Such massive, disruptive eruptions often produce atmospheric sounds known as infrasounds, an extreme basso profondo that can propagate for thousands of kilometers. Infrasounds are usually inaudible, as they reside below the 20 Hz threshold of human hearing and tonality. However, when intense enough, we can perceive infrasound as beats or sensations.

Like a large door slamming, the concussion of a volcanic explosion can be startling and terrifying. It immediately compels us to pay attention, and it’s not something one gets used to. The roaring is also disconcerting, especially if one thinks of a volcano as an erratic furnace with homicidal tendencies. But occasionally, amidst the chaos and cacophony, repeatable sound patterns emerge, suggestive of a modicum of order within the complex volcanic system. These reproducible, recognizable patterns permit the identification of early warning signals, and keep us listening.

Each of us now have technology within close reach to capture and distribute Nature’s silent warning signals, be they from volcanoes, tsunamis, meteors, or rogue nations testing nukes. Infrasounds, long hidden under the myth of silence, will be everywhere revealed.

Cookie Monster

The “Cookie Monster” skylight on the southwest flank of Pu`u `O`o. Photo by J. Kauahikaua 27 September 2002

I first heard these volcanic sounds in the rain forests of Costa Rica. As a graduate student, I was drawn to Arenal Volcano by its infamous reputation as one of the most reliably explosive volcanoes in the Americas. Arenal was cloud-covered and invisible, but its roar was audible and palpable. Here is a tremor (a sustained oscillation of the ground and atmosphere) recorded at Arenal Volcano in Costa Rica with a 1 Hz fundamental and its overtones:

.

In that first visit to Arenal, I tried to reconstruct in my minds’ eye what was going on at the vent from the diverse sounds emitted behind the cloud curtain. I thought I could blindly recognize rockfalls, blasts, pulsations, and ground vibrations, until the day the curtain lifted and I could confirm my aural reconstruction closely matched the visual scene. I had imagined a flashing arc from the shock wave as it compressed the steam plume, and by patient and careful observation I could see it, a rapid shimmer slashing through the vapor. The sound of rockfalls matched large glowing boulders bouncing down the volcano’s slope. But there were also some surprises. Some visible eruptions were slow, so I could not hear them above the ambient noise. By comparing my notes to the infrasound records I realized these eruption had left their deep acoustic mark, hidden in plain sight just below aural silence.

Arenal, Costa Rica, May 1, 2010. Image by Flickr user Daniel Vercelli.

Arenal, Costa Rica, May 1, 2010. Image by Flickr user Daniel Vercelli.

I then realized one could chronicle an eruption through its sounds, and recognize different types of activity that could be used for early warning of hazardous eruptions even under poor visibility. At the time, I had only thought of the impact and potential hazard mitigation value to nearby communities. This was in 1992, when there were only a handful of people on Earth who knew or cared about infrasound technology. With the cessation of atmospheric nuclear tests in 1980 and the promise of constant vigilance by satellites, infrasound was deemed redundant and had faded to near obscurity over two decades. Since there was little interest, we had scarce funding, and were easily ignored. The rest of the volcano community considered us a bit eccentric and off the main research streams, but patiently tolerated us. However, discussions with my few colleagues in the US, Italy, France, and Japan were open, spirited, and full of potential. Although we didn’t know it at the time, we were about to live through Gandhi’s quote: “First they ignore you, then they laugh at you, then they fight you, then you win.”

Fast forward 22 years. A computer revolution took place in the mid-90’s. The global infrasound network of the International Monitoring System (IMS) began construction before the turn of the millennium, in its full 24-bit broadband digital glory. Designed by the United Nations’s Comprehensive Nuclear-Test-Ban Treaty Organization (CTBTO), the IMS infrasound detects minute pressure variations produced by clandestine nuclear tests at standoff distances of thousands of kilometers. This new, ultra-sensitive global sensor network and its cyberinfrastructure triggered an Infrasound Renaissance and opened new opportunities in the study and operational use of volcano infrasound.

Suddenly endowed with super sensitive high-resolution systems, fast computing, fresh capital, and the glorious purpose of global monitoring for hazardous explosive events, our community rapidly grew and reconstructed fundamental paradigms early in the century. The mid-naughts brought regional acoustic monitoring networks in the US, Europe, Southeast Asia, and South America, and helped validate infrasound as a robust monitoring technology for natural and man-made hazards. By 2010, infrasound was part of the accepted volcano monitoring toolkit. Today, large portions of the IMS infrasound network data, once exclusive, are publicly available (see links at the bottom), and the international infrasound community has grown to the hundreds, with rapid evolution as new generations of scientists joins in.

In order to capture infrasound, a microphone with a low frequency response or a barometer with a high frequency response are needed. The sensor data then needs to be digitized for subsequent analysis. In the pre-millenium era, you’d drop a few thousand dollars to get a single, basic data acquisition system. But, in the very near future, there’ll be an app for that. Once the sound is sampled, it looks much like your typical sound track, except you can’t hear it. A single sensor record is of limited use because it does not have enough information to unambiguously determine the arrival direction of a signal. So we use arrays and networks of sensors, using the time of flight of sound from one sensor to another to recognize the direction and speed of arrival of a signal. Once we associate a signal type to an event, we can start characterizing its signature.

Consider Kilauea Volcano. Although we think of it as one volcano, it actually consists of various crater complexes with a number of sounds. Here is the sound of a collapsing structure

As you might imagine, it is very hard to classify volcanic sounds. They are diverse, and often superposed on other competing sounds (often from wind or the ocean). As with human voices, each vent, volcano, and eruption type can have its own signature. Identifying transportable scaling relationships as well as constructing a clear notation and taxonomy for event identification and characterization remains one of the field’s greatest challenges. A 15-year collection of volcanic signals can be perused here, but here are a few selected examples to illustrate the problem.

First, the only complete acoustic record of the birth of Halemaumau’s vent at Kilauea, 19 March 2008:

.

Here is a bench collapse of lava near the shoreline, which usually leads to explosions as hot lava comes in contact with the ocean:

.

.

Here is one of my favorites, from Tungurahua Volcano, Ecuador, recorded by an array near the town of Riobamba 40 km away. Although not as violent as the eruptive activity that followed it later that year, this sped-up record shows the high degree of variability of eruption sounds:

.

.

The infrasound community has had an easier time when it comes to the biggest and meanest eruptions, the kind that can inject ash to cruising altitudes and bring down aircraft. Our Acoustic Surveillance for Hazardous Studies (ASHE) in Ecuador identified the acoustic signature of these type of eruptions. Here is one from Tungurahua:

.

Our data center crew was at work when such a signal scrolled through the monitoring screens, arriving first at Riobamba, then at our station near the Colombian border. It was large in amplitude and just kept on going, with super heavy bass – and very recognizable. Such signals resemble jet noise — if a jet was designed by giants with stone tools. These sustained hazardous eruptions radiate infrasound below 0.02 Hz (50 second periods), so deep in pitch that they can propagate for thousands of kilometers to permit robust acoustic detection and early warning of hazardous eruptions.

In collaborations with our colleagues at the Earth Observatory of Singapore (EOS) and the Republic of Palau, infrasound scientists will be turning our attention to early detection of hazardous volcanic eruptions in Southeast Asia. One of the primary obstacles to technology evolution in infrasound has been the exorbitant cost of infrasound sensors and data acquisition systems, sometimes compounded by export restrictions. However, as everyday objects are increasingly vested with sentience under the Internet of Things, this technological barrier is rapidly collapsing. Instead, the questions of the decade are how to receive, organize, and distribute the wealth of information under our perception of sound so as to construct a better informed and safer world.

IRIS Links

http://www.iris.edu/spud/infrasoundevent

http://www.iris.edu/bud_stuff/dmc/bud_monitor.ALL.html, search for IM and UH networks, infrasound channel name BDF

Milton Garcés is an Earth Scientist at the University of Hawaii at Manoa and the founder of the Infrasound Laboratory in Kona. He explores deep atmospheric sounds, or infrasounds, which are inaudible but may be palpable. Milton taps into a global sensor network that captures signals from intense volcanic eruptions, meteors, and tsunamis. His studies underscore our global connectedness and enhance our situational awareness of Earth’s dynamics. You are invited to follow him on Twitter @iSoundHunter for updates on things Infrasonic and to get the latest news on the Infrasound App.

Featured image: surface flows as seen by thermal cameras at Pu’u O’o crater, June 27th, 2014. Image: USGS

tape reel

REWIND! If you liked this post, check out …

SO! Amplifies: Ian Rawes and the London Sound Survey – Ian Rawes

Sounding Out Podcast #14: Interview with Meme Librarian Amanda Brennan – Aaron Trammell

Catastrophic Listening — China Blue

%d bloggers like this: