This is article 2.0 in Sounding Out!‘s April Forum on “Sound and Technology.” Every Monday this month, you’ll be hearing new insights on this age-old pairing from the likes of Sounding Out! veterano Aaron Trammell along with new voices Andrew Salvati and Owen Marshall. These fast-forward folks will share their thinking about everything from Auto-tune to techie manifestos. So, turn on your quantizing for Sounding Out! and enjoy today’s supersonic in-depth look at sampling from from SO! Regular Writer Primus Luta. –JS, Editor-in-Chief
My favorite sample-based composition? No question about it: “Stroke of Death” by Ghostface and produced by The RZA.
Supposedly the story goes, RZA was playing records in the studio when he put on the Harlem Underground Band’s album. It is a go-to album in a sample-based composer collection, because of the open drum breaks. One such break appears in the cover of Bill Wither’s “Ain’t No Sunshine”, notably used by A Tribe Called Quest on “Everything is Fair.”
RZA, a known break beat head, listened as the song approached the open drums, when the unthinkable happened: a scratch in his copy of the record. Suddenly, right before the open drums dropped, the vinyl created its own loop, one that caught RZA’s ear. He recorded it right there and started crafting the beat.
This sample is the only source material for the track. RZA throws a slight turntable backspin in for emphasis, adding to the jarring feel that drives the beat. That backspin provides a pitch shift for the horn that dominates the sample, changing it from a single sound into a three-note melody. RZA also captures some of the open drums so that the track can breathe a bit before coming back to the jarring loop. As accidental as the discovery may have been, it is a very precisely arranged track, tailor-made for the attacking vocals of Ghostface, Solomon Childs, and the RZA himself.
“Stroke of Death” exemplifies how transformative sample-based composition can be. Other than by knowing the source material, the sample is hard to identify. You cannot figure out that the original composition is Wither’s “Ain’t No Sunshine” from the one note RZA sampled, especially considering the note has been manipulated into a three-note melody that appears nowhere in either rendition of the composition. It is sample based, yes, but also completely original.
Classifying a composition like this as a ‘happy accident’ downplays just how important the ear is in sample-based composition, particularly on the transformative end of the spectrum. J Dilla once said finding the mistakes in a record excited him and that it was often those mistakes he would try to capture in his production style. Working with vinyl as a source went a long way in that regard, as each piece of vinyl had the potential to have its own physical characteristics that affected what one heard. It’s hard to imagine “Stroke of Death” being inspired from a digital source. While digital files can have their own glitches, one that would create an internal loop on playback would be rare.
There has been a change in the sound of sampling over the past few decades. It is subtle but still perceptible; one can hear it even if a person does not know what it is they are hearing. It is akin to the difference between hearing a blues man play and hearing a music student play the blues. They technically are both still the blues, but the music student misses all of the blue notes.
The ‘blue notes’ of the blues were those aspects of the music that could not be transcribed yet were directly related to how the song conveyed emotion. It might be the fact that the instrument was not fully in tune, or the way certain notes were bent but others were not, it could even be the way a finger hit the body of a guitar right after the string was strummed. It goes back farther than the blues and ultimately is not exclusive to the African American tradition from which the phrase derives; most folk music traditions around the world have parallels. “The Rite of Spring” can be understood as Stravinsky ‘sampling’ the blue notes of Transylvanian folk music. In many regards sample-based composing is a modern folk tradition, so it should come as no surprise that it has its own blue notes.
The sample-based composition work of today is still sampling, but much of it lacks the blue notes that helped define the golden era of the art. I attribute this discrepancy to the evolution of technology over the last two decades. Many of the things that could be understood as the blue notes of sampling were merely ways around the limits of the technology. In the same way, the blue notes of most folk music happened when the emotion went beyond the standards of the instrument (or alternately the limits imposed upon it by the literal analysis of western theory). By looking at how the technology has evolved we can see how blue notes of sampling are being lost as key limitations are being overcome by “advances.”
First, let’s consider the E-Mu SP-1200, which is still thought to be the most definitive sounding sampler for hip-hop styled sample-based compositions, particularly related to drums. The primary reason for this is its low-resolution sampling and conversion rates. For the SP-1200 the Analog to Digital (A/D) and Digital to Analog (D/A) converters were 12-bit at a sample rate of 26.04 kHz (CD quality is 16-bit 44.1 kHz). No matter what quality the source material, there would be a loss in quality once it was sampled into and played out of the SP-1200. This loss proved desirable for drum sounds particularly when combined with the analog filtering available in the unit, giving them a grit that reflected the environments from which the music was emerging.
On top of this, individual samples could only be 2.5 seconds long, with a total available sample time of only 10 seconds. While the sample and conversion rates directly affected the sound of the samples, the time limits drove the way that composers sampled. Instead of finding loops, beatmakers focused on individual sounds or phrases, using the sequencer to arrange those elements into loops. There were workarounds for the sample time constraints; for example, playing a 33-rpm record at 45 rpm to sample, then pitching it back down post sampling was a quick way to increase the sample time. Doing this would further reduce the sample rate, but again, that could be sonically appealing.
An under appreciated limitation of the SP-1200 however, was the visual feedback for editing samples. The display of the SP-1200 was completely alpha numeric; there were no visual representations of the sample other than numbers that were controlled by the faders on the interface. The composer had to find the start and end points of the sample solely by ear. Two producers might edit the exact same kick drum with start times 100 samples (a fraction of a millisecond) apart. Were one of the composers to have recorded the kick at 45 rpm and pitched it down, the actual resolution for the start and end times would be different. When played in a sequence, these 100 samples affect the groove, contributing directly to the feel of the composition. The timing of when the sample starts playback is combined with the quantization setting and the swing percentage of the sequencer. That difference of 100 samples in the edit further offsets the trigger times, which even with quantization turned off fit into the 24 parts per quarter grid limitations of the machine.
Akai’s MPC-60 was the next evolution in sampling technology. It raised the sample and conversion rates to 16-bit and 40 kHz. Sample time increased to a total of 13.1 seconds (upgradable to 26.2). Sequencing resolution increased to 96 parts per quarter. Gone was the crunch of the SP-1200, but the precision went up both in sampling and in sequencing. The main trademark of the MPC series was the swing and groove that came to Akai from Roger Linn’s Linn Drum. For years shrouded in mystery and considered a myth by many, in truth there was a timing difference that Linn says was achieved by delaying certain notes by samples. Combined with the greater PPQ resolution in unquantized mode, even with more precision than the SP-1200, the MPC lent itself to capturing user variation.
Despite these technological advances, sample time and editing limitations, combined with the fact that the higher resolution sampling lacked the character of the SP-1200, kept the MPC from being the complete package sample composers desired. For this reason it was often paired with Akai’s S-950 rack sampler. The S-950 was a 12-bit sampler but had a variable sample rate between 7.5 kHz and 40 kHz. The stock memory could hold 750 KB of samples which at the lowest sample rate could garner upwards of 60 seconds of sampling and at the higher sample rates around 10 seconds. This was expandable to up to 2.5 MB of sample memory.
The editing capabilities made the S-950 such a powerful sampler. Being able to create internal sample loops, key map samples to a keyboard, modify envelopes for playback, and take advantage of early time stretching (which would come of age with the S-1000)—not to mention the filter on the unit—helped take sampling deeper into the sound design territory. This again increased the variable possibilities from composer to composer even when working from the same source material. Often combined with the MPC for sequencing, composers had the ultimate sample-based composition workstation.
Today, there are literally no limitations for sampling. Perhaps the subtlest advances have developed the precision with which samples can be edited. With these advances, the biggest shift has been the reduction of the reliance on ears. Recycle was an early software program that started to replace the ears in the editing process. With Recycle an audio file could be loaded, and the software would chop the sample into component parts by searching for the transients. Utilizing Recycle on the same source, it was more likely two different composers could arrive at a kick sample that was truncated identically.
Another factor has been the waveform visualization of samples for editing. Some earlier hardware samplers featured the waveform display for truncating samples, but the graphic resolution within the computer made this even more precise. By looking at the waveform you are able to edit samples at the point where a waveform crosses the middle point between the negative and positive side of the signal, known as the zero-crossing. The advantage of zero-crossing sampling is that it prevents errors that happen when playback goes from either side of the zero point to another point in one sample, which can make the edit point audible because of the break in the waveform. The end result of zero-crossing edited samples is a seamlessness that makes samples sound like they naturally fit into a sequence without audible errors. In many audio applications snap-to settings mean that edits automatically snap to zero-crossing—no ears needed to get a “perfect” sounding sample.
It is interesting to note that with digital files it’s not about recording the sample, but editing it out of the original file. It is much different from having to put the turntable on 45 rpm to fit a sample into 2.5 seconds. Another differentiation between digital sample sources is the quality of the files, whether original digital files (CD quality or higher), lossless compression (FLAC), lossy compressed (MP3, AAC) or the least desirable though most accessible, transcoded (lossy compression recompressed such as YouTube rips). These all result in a different degradation of quality than the SP-1200. Where the SP-1200’s downsampling often led to fatter sounds, these forms of compression trend toward thinner-sounding samples.
Some producers have created their own sound using thinned out samples with the same level of sonic intent as The RZA’s on “Stroke of Death.” The lo-fi aesthetic is often an attempt to capture a sound to parallel the golden era of hardware-based sampling. Some software-based samplers by example will have an SP-1200 emulation button that reduces bit rates to 12-bit. Most of software sequencers have groove templates that allow for the sequencers to emulate grooves like the MPC timing.
Perhaps the most important part of the sample-based composition process however, cannot be emulated: the ear. The ear in this case is not so much about the identification of the hot sample. Decades of history should tell us that the hot sample is truly a dime a dozen. It takes a keen composer’s ear to hear how to manipulate those sounds into something uniquely theirs. Being able to listen for that and then create that unique sound—utilizing whatever tools— that is the blue note of sampling. And there is simply no way to automate that process.
Featured image: “Blue note inverted” by Flickr user Tim, CC BY-ND 2.0
Primus Luta is a husband and father of three. He is a writer and an artist exploring the intersection of technology and art, and their philosophical implications. He maintains his own AvantUrb site. Luta was a regular presenter for Rhythm Incursions. As an artist, he is a founding member of the collective Concrète Sound System, which spun off into a record label for the exploratory realms of sound in 2012. Recently Concréte released the second part of their Ultimate Break Beats series for Shocklee.
REWIND!…If you liked this post, you may also dig:
“SO! Reads: Jonathan Sterne’s MP3: The Meaning of a Format”-Aaron Trammell
“Remixing Girl Talk: The Poetics and Aesthetics of Mashups”-Aram Sinnreich
“Sound as Art as Anti-environment”-Steven Hammer
Editor’s Note: Welcome to Sounding Out!‘s fall forum titled “Sound and Play,” where we ask how sound studies, as a discipline, can help us to think through several canonical perspectives on play. While Johan Huizinga had once argued that play is the primeval foundation from which all culture has sprung, it is important to ask where sound fits into this construction of culture; does it too have the potential to liberate or re-entrench our social worlds? Here, Roger Moseley challenges us to rethink the philosophical discourses of both sound and play and locates the moments in which they intersect and interface. From games of Telephone to Guitar Hero, Moseley considers the ways in which sonic play can help us understand the phantasmic binaries of the analog and digital.–AT
Throughout the distinguished intellectual lineage of play (where it is touched on by notable philosophers such as Plato, Montaigne, Kant, Schiller, Gadamer, Derrida, and Baudrillard), little attention has been paid to the parallels that can be drawn between sound and play as both media and phenomena. The very name of today’s most prominent cultural and technological locus of play, the video game, overtly privileges the eye at the expense of the ear. As recent research and creative work by such figures as Aaron Oldenburg, Aaron Trammell, George Karalis, and Enongo Lumumba-Kasongo indicates, a surge of interest in audio games, as well as video games that emphasize the importance of sound while eschewing or minimizing visual stimuli, is acting as a salutary corrective to this oculocentrism. In what follows, I suggest that bringing sonic and musical techniques to bear on this history might afford new insights into play and its myriad configurations. Conceiving of play sonically entails thinking of sound playfully. This intersectional logic can, I argue, unpick binarisms that enforce problematic distinctions and constrict thought. To demonstrate this, I conclude by deploying the concept of play to redefine the relationship between the digital and the analog—and vice versa.
How can play be defined in a manner that encompasses its farrago of meanings and associations? For video game designers and theorists Katie Salen and Eric Zimmerman, the answer is deceptively simple: play is “free motion within a more rigid structure” (Rules of Play, 304). To illustrate the flexibility of this definition, Salen and Zimmerman allude to the phenomenon of light playing upon the ocean waves. They leave unexamined, however, the intimacy and richness of the relationship between play and sound. From a scientific perspective, the patterned oscillation of which a sequence of sound waves is constituted consists of free motion within the limits set forth by the laws of physics. When disciplined and deployed as a cultural technique–take the play of musical instruments for example–sonic play is humanized and rendered transitive. But, we might also suggest that instruments play people, citing the sensation of automation with which fingers flash over fretboard or keyboard. Moving further away from anthropocentrism, we can observe how sonic technologies render play intransitive once more. From the barrel organ to the iPod, sound plays without human aid when mechanically reproduced. This way of framing reproduction invokes and extends Roger Caillois’s playful category of mimicry, which can be construed as faithful imitation, deceptive fakery, or even a Baudrillardian attempt to simulate a phenomenon that never existed.
In order to pay due attention to both the technologies through which sonic play is mediated as well as the cultural techniques imbue it with significance, I suggest that we supplement Salen and Zimmerman’s definition by thinking of freedom, motion, and structure in both digital and analogical terms. To an extent, the adoption of this modish epistemological framework acknowledges that conceptions of play are always constrained by their prevailing intellectual context. More importantly, however, I contend that technologies of sonic generation and representation from the seventeenth to the nineteenth centuries can be understood to play with the categories of the digital and the analog avant la lettre (ou le chiffre). The two categories are not mutually exclusive and to treat them as such would be to subjugate the granularity of the analog to the binary logic of the digital. Rather, they co-exist as modes between which sounds and players freely oscillate.
The origins of digital sonic play lie within the human body. As Johan Huizinga put it, “the link between play and instrumental skill is to be sought in the nimble and orderly movements of the fingers.” In the course of musical performance, human digits perform innumerable calculations. At its crudest level, musical performance from a score can be construed as a sort of algorithmic play through which mimetic fidelity is evaluated (and wrong notes relentlessly tallied). This ludic logic is at its most visible in rhythm-action video games such as Guitar Hero in which the score is no longer a text but rather a quantitative analysis. The iconography of these games usually indexes a set of digital technologies used primarily for the recording, editing, and playback of music. On the one hand, this relationship can be traced back to Leibniz’s exposition of ars combinatoria and his “invention” of binary; on the other, it is realized by the hydraulic organ and composing machine devised and programmed by the Jesuit polymath Athanasius Kircher, both of which are depicted in his Musurgia universalis (1650). In media-archaeological terms, the combination of Leibniz’s concepts and Kircher’s mechanisms gave rise to the hardware and software of Joseph Marie Jacquard’s revolutionary loom, Charles Babbage’s prototypical Analytical Engine, the player piano, the IBM punch card, and the MIDI sequencer before resurfacing in Guitar Hero, a piece of software that, in purely algorithmic terms, enlists the player’s digits to verify checksums.
Such digital grids may constitute the field and the rules of sonic play, but they must be supplemented by analog elements if play is to flourish. As detailed in C. P. E. Bach’s Versuch über die wahre Art das Clavier zu spielen (1753/62), the clavichord and its descendants distinguished themselves from the harpsichord and the organ by endowing the keyboard with an infinite sensitivity to touch, thereby enabling a mimetic spectrum of emotional flow with unprecedented verisimilitude. Analogicity also provides another perspective on Caillois’s concept of mimicry, according to which one object or activity playfully stands in for another via imitation, deception, or make-believe. Additionally, the curves of Ernst Chladni’s figures, which materialized sound as sand, exemplify this sonic and mimetic trajectory as they rely on both Hermann von Helmholtz’s pioneering work on acoustics and the complex history of phonography to the development of analog synthesis.
In terms of sonic play, digital and analog elements can be chiastically recombined and reconfigured. A sonic communication game such as Telephone relies on the human propensity for analogy and its corrupting influence on the integrity of information transfer, playfully inverting the conditions and functions of the “real” telephone (which was engineered to compress informational content digitally without jeopardizing meaning). In much electronic dance music, the digital latticework, simultaneously visualized and rendered audible by the sequencer’s grid, constitutes a field of play overlaid with vocals, sweeps, and other analog elements that, in turn, have been captured via digital sampling. As a kind of meta-game, a mash-up plays with sonic elements whose relations can be parsed in the digital terms of Leibnizian recombinatorial play, but equally important are the unintended associations and analogies which inevitably emerge. And while games such as Guitar Hero foreground digital techniques of sonic reproduction, they simultaneously foster diverse forms of analogical play involving the player’s manipulation of the sonic (and social) behavior of her on-screen avatar—and vice versa.
There is no doubt that the status of sound and its mediation through and as play have too often gone unacknowledged. As well as seeking to rectify this state of affairs by stressing the importance of sound in relation to the playful operations of other media, we might also dwell on the distinguishing features that set it apart: sound and the techniques that shape it are unique in the ways they simultaneously trace and are traced by the materials, technologies, and metaphors of play.
Roger Moseley is an Assistant Professor in the Department of Music at Cornell University. His most recent research brings a media-archaeological perspective to bear on musical performance and improvisation. He is particularly interested in how the concept of play informs sonic practices and cultural techniques. Active as a collaborative pianist on both modern and historical instruments, he has recently published essays on digital games in the contexts of musical and visual culture. His current book project is entitled Digital Analogies: Interfaces and Techniques of Musical Play.
REWIND! . . .If you liked this post, you may also dig:
Papa Sangre and the Construction of Immersion in Audio Games– Enongo Lumumba-Kasongo
I recently had the opportunity to fool around with the iPad2’s new GarageBand suite. Enticed by the intuitive touch interface I soon found myself lost within the device’s labyrinthine architecture. Every poke, prod and press brought me to a new screen with a bevy of exciting options. A touch to create a drum loop, a tickle to evoke some reverb, and a brush to strum a guitar. I was one with the machine; it was a truly cybernetic, kinesthetic moment. This may sound naïve, but I had never realized how many tools were available to electronic musicians, or how intuitive using these tools could be. As digital tools to create music become more accessible and more intuitive, what is the role of the human in understanding their use? Further, what strategies can we adopt when listening to these creations?
This question may seem a bit outdated to those who have been researching post-humanist phenomena since the digital boom in the mid-nineties. Often conflicting perspectives regarding the negotiation of the human and the digital have been considered in the last decade or so. Some like Donna Haraway, Pierre Lévy, and even Ray Kurzweil offer particularly optimistic readings of the post-human (although for radically different reasons). While scholars like Nancy Baym and Jaron Lanier have offered decisively more sober readings of the problematic. They argue that splits between the human and post-human, or analog and digital are false dichotomies. Truth be told, none of the theorists above adequately address my feelings on this topic. Producing music with a digital audio suite makes me defensive of my humanism and it is by its very nature a project of preservation.
The algorithmic tools packaged within digital audio suites encourage a sense of aesthetic preservation. Tools like GarageBand’s Smart Guitar, Smart Drums, Smart Bass, various arpeggiators and Appleloops encourage the user to program music on a high level where the nuance of serendipity and improvisation play second fiddle to the overall sonic contours of a piece. Although the user is provided the tools to intervene and program music in a very specific way, it is by default a distinctly different experience than that of playing a guitar or piano. The ghost of the algorithm haunts such performances; reminding the user that these acts of spontaneous creation are no longer the default but deliberate…. This sense of deliberate improvisation forces me into a reflexive space where I am acutely aware of the mediations occurring within my performance. Succinctly, I must defend a sense of self within my creation. If I yield to the algorithms that seek to help me compose, I destroy all sense of the human within my work. Simply turning on robots and watching them sing.
For this reason, I propose an aesthetic of preservation as a way to understand the ways in which we listen to works created by digital audio suites. As algorithmic aids become more advanced and commonplace in music, the human becomes a less essential aspect of the form. Understanding what has been deliberately included in spite the seductive algorithmic environment is ultimately a project that seeks to recover the human in the machine; perhaps even, a project doomed from the start, as we grow ever closer to the means of our artistic production.
Magnasanti – Check out the results of my collaboration with Colin Germain on GarageBand!