|
Creature Sound Design
October 16, 2006
|
|
In Return of the King, the tremendous
thumps of Oliphant footsteps were designed by dropping large boulders
from a crane. In order to create a cleaner, more dramatic effect,
sounds were applied to the front feet only. |
|
Usually when watching fantasy, science fiction, or horror films, or playing
video games, we never stop to wonder how the voices of the supernatural
beings were created. In fact, if we do, it might be because the voice
is so unnatural, that it's distracting. Instead we should be wanting to
know more about the creature, because in some way, when we become engrossed,
we suspend our disbelief that the creature could not exist outside of
the screen.
In his book, The Practical Art of Motion Picture Sound, David Lewis Yewdall,
supervising sound editor for The Thing, says:
"I have harped on this throughout this book and will continue to
do so, because it is the keystone of the sound design philosophy -- if
you take the audience OUT of the film because of bad sound design -- then
you have failed as a sound designer. It’s just that simple."
Personally, I am distracted from film when I hear sound effects that
are clichés of other films. To me, these are lost opportunities,
and the worst examples of audio design. Although we should look to others'
work for inspiration, we should be inspired by things outside of the entertainment
industry so that we can bring in new ideas. There is a great deal of imitation
going on in film and video games right now, but still, the only major
achievements are done by innovators. Ben Burtt, sound designer for the
Star Wars series, explained it like this:
"Your success as an artist, to say something new, ultimately depends
on the breadth of your education. My recommendation would be to major
in an area other than film, develop a point of view, and then apply that
knowledge to film. Because if film is all you know, you cannot help but
make derivative work. I found that what I had learned about sound, history,
biology, English, physics all goes into the mix."
In this article, I would like to discuss some past and recent examples
of creature sound design techniques. Also, I will introduce some alternative
digital signal processing (DSP) methods that can add further FX potential
than can be achieved by mixing. I have some thoughts to share on current
effect mixing trends, and what to avoid in these trends. Additionally,
I will provide insight into other fields of research whenever possible,
and not give step-by-step methods, in order to inspire others to explore
original ideas. The information here can be synthesized in individualized
and unique ways, to produce new ideas for creature sound design.
Alchemy
"One becomes two, two becomes three, and by means of the third and
fourth achieves unity; thus two are but one...." - Maria the Jewess,
300 A.D. (Sharp)
To be a sound designer is to gambol with the forces of alchemy. We venture
into the world to find powerful new elements, contemplate their properties,
speculate about their inter-conversions, then hide away in dark chambers
to divide, filtrate, mix and ignite them. This is a dark art, pursued
by sonic sorcerers. I am fascinated by the chemistry of SFX creation,
an enigmatic, largely unexplored and undocumented art form that has only
just begun.
When we listen to a recorded sound, it is not always obvious what we
are hearing; without a visual cue, we might not know what the sound belongs
to. We might hear the sound of crinkling cellophane and recognize that
it produces a familiar sound yet not know what it belongs, or is related,
to, but if this cellophane sound is placed on top of some scenery with
a cloudy sky, we will believe that it is rain that we are hearing.
Picture a whale coming up for air. It's exhale is just a release of pressurized
air, so we could use any similar sound in its place. With a little processing,
we could swap it for a flamethrower burst and the scene would gain some
energy. We could also layer in the sound of an elephant call, our motivation
being that we want the sound to be larger than life. By replacing the
softer bellow of air with something more thundering and textural, we’ve
just made the scene more dramatic, without anyone knowing that they’ve
just heard a special effect. Illusion is the basis of sound design, a
magic trick that has not changed since rolling cannon balls provided the
sound of thunder in Shakespearian theatre.
The most straightforward way to create original creature sounds is to
remix and process recordings of animals, a technique pioneered by Murray
Spivak, head of the sound department of RKO studios who mixed sound for
the 1933 release of King Kong. Kong's roar consisted of a lion's roar
played back at half speed, layered with a tiger played in reverse. This
technique was mind-blowing for a time when silent films were still being
released. Right up until the 60's, special effects sounds and even animal
sounds were being done with foley effects gadgets and machinery, a method
that was borrowed from theatre and radio. It is impressive what the sound
editors were able to achieve with normal objects and magnetic tape recordings.
David Lewis Yewdall spoke up for this ancient art, saying:
"When these fine craftspersons labored to make such memorable sound
events for your entertainment pleasure, they made them without the help
of today's high-tech digital tools -- without Harmonizers or Vocorders,
without a Synclavier or a Fairlight. They designed these sounds with their
own brains, understanding what sounds to put together to create new sound
events -- how to play them backward, slow them down, cut, clip, and scrape
them with a razor blade (when magnetic soundtrack became available in
1953), or paint them with blooping ink (when they still cut sound effects
on optical track)."
Unfortunately though, these ingenious sounds were always pretty much
the same, and eventually ended up in archives. Studios came to rely more
on stock recordings from sound libraries, instead of reproducing the sounds
that were needed every time. This indeed cut down on costs, but one could
hear the same foley sounds across hundreds of films. It wasn't until such
innovators as George Lucas joined the industry that we escaped the old
school techniques and began to utilize new audio technologies to create
sound effects. Lucas had the financial backing to initiate major developments
in sound and fortunately, he had two audio geniuses at his disposal, Walter
Murch, and Ben Burrt.
|
|
|
Chewbacca's voice
consists of several animal sounds, primarily a hungry bear. |
Ben Burtt, inspired by the sound work in King Kong, took the technique
of remixing animal sounds to a whole new level. Burtt was one of the first
people in film to be responsible for all aspects of the audio track, as
up until that point the task was collectively handled by sound recordists,
editors and mixers, with sparse communication between. He started by building
a personal sound effects library during the preproduction of Star Wars,
traveling around to zoos, factories, airports, and museums with a microphone
and tape recorder. The initial idea for Chewbacca's voice was realized
when he heard a walrus stranded at the bottom of an empty pool at Marineland.
After some experimentation, he decided on several ingredients including
the moans of a hungry bear, walrus, dog growls, and lion roars. Numerous
variations of these samples were meticulously clipped and layered in sequence
across 4 or 5 tracks, in time with the actor's mouth movements, to develop
a singular vocalization. The design of this voice is so smooth and consistent
that we recognize it immediately as Wookie, without ever knowing that
we are hearing 4 different animal sounds mixed together.
If you are an independent sound designer, without the help of a production
crews, it would be quite difficult to gather all of your own recordings
of animal sounds. You would have to do a lot of research and traveling,
the locations would not always provide good recording conditions and you
would need authorization for many sites. Furthermore, zoos, for example,
are extremely noisy with lots of surface reflections and the expectation
cannot be made that the animals would simply walk up to the microphones
and make that sound that is sought after. As an alternative, we can do
well by relying on licensed nature sound effects libraries for some of
the staple animal sounds, then go out and do special recordings to supplement
them.
Of course, I am just using animal sounds as one example. When we are
making the sound of living organisms we can also mix together recordings
of normal objects, music, our own voice...Anything. For example, a large
flat stone, like a tombstone, sounds very much like a low growl when sliding
over a concrete surface. If we take that rumbling sound, the sound of
chanting Tibetan monks, and a great big sigh from our own mouth, we might
have three very good ingredients for the exhale of a menacing demon.
The greatest sounds are the result of innovation, both deliberate and
coincidental. Not only should we look to films for inspiration, but also
approach our designs with speculative thought, personal experience, and
with knowledge obtained from outside of the entertainment industry. We
can certainly make a career out of doing imitations and spin-offs, but
if we want to be creators then we had better put our goggles on and flare
up the torch of invention.
Animal Communication
Much of the time, the director or designer will simply count on human
voices for creatures. If this is the case, you will clearly have to make
the voice intelligible. But consider for a moment the impact that creative
sound design can have on the overall perception of the organism. The voice
of Treebeard in The Lord of the Rings, The Two Towers, is a good example
of human voice customization. His voice is deep and resonant, like it
is really coming from within a tree, and mixed with his creaking limbs
and thundering footsteps, a personality emerges that is thousands of years
old and far from human.
Other times the voice will be semi-human, or non intelligible, animal-like.
When we are to design sounds for more unusual organisms, a good place
to look for inspiration is nature. In fact, recordings of animal sounds
may be all the ingredients that you need to start creating your own unique
beings. In order to get a better idea of our options, lets look at a few
communication methods that animals have devised.
In general, when we think of sounds that animals make, we usually just
consider what comes from it's mouth, such as a coyote howl. It lifts it's
head upward, and with muscles compressing in it's diaphragm, air is pushed
from the lungs and through the voice box. As the coyote tightens it's
vocal cords, which resemble elastic bands, air squeezes through in a steady
stream of short bursts, creating an audible vibration. This waveform is
then shaped by muscles in the throat, the oral and nasal cavities, movement
of the tongue, the lips, and numerous other factors that contribute to
the sound that echoes out and disappears into the environment.
|
|
|
In order to create
the sounds of creatures in The Chronicles of Riddick (such as the
Hellhound), audio designers used field recordings of a black leopard,
Bengal Tiger, and several types of wolves. These were recorded directly
to DAT using a Neumann 191 and Sennheiser 461 shotgun. |
For various biological and behavioral reasons, animals such as big cats
and howler monkeys, need to produce sounds at a much higher volume than
vocal cords alone could handle. The roar of a lion, for instance, can
reach 114 decibels, which is about the volume of a chain saw at full throttle.
In order for this to be possible, the hyoid bones that makes up the shape
of the lions larynx (voice box) is made of soft cartilage connected by
a 6 to 10 inch long ligament. When the cat pushes air through its larynx
and tightens this ligament, the entire larynx vibrates. This vibration
is further amplified by a horn shaped waveguide, that helps amplify the
roar so that it can be heard for miles around. This unusual voice box
structure allows them a sound palette quite unique in comparison to animals
that use vocal cords.
Lets consider methods of vocalization that do not require a larynx. Spiny
lobsters, for example, do not have vocal cords, but they can produce shrieking
sounds in the same way that a squeaky door hinge does. This sound is created
by a continuous slip-stick-slip-stick friction between flesh and exoskeleton
inside the base of the lobster's antennae. In order to do this, the lobster
needs to push at least one of its antennae back and forth while contracting
muscles on either side of it's eyes where the antennae protrude from.
It is thought that lobsters produce these sounds to ward off enemies,
but still, it is unknown.
Many living things create sounds by an action called stridulation, which
is achieved by rubbing two bony organs together. This is most commonly
known in insects and arthropods such as crickets, millipedes, and stag
beetles. For insects, this is achieved by rubbing legs or talons together
which have microscopic ridges on them. Many underwater inhabitants such
as catfish and seahorses create clicks and squeaks in the same way, but
by rubbing skeletal parts together, sometimes within their fins. These
sounds are produced as alarms, to attract mates, as a fearful response,
or to antagonize competitors.
Porcupines have special quills on their tails that are shaped like tubes,
and which they can shake to produce a steady rattling sound, in order
to warn others of their presence, similarly to a rattlesnake. They also
stomp their hind feet, click their teeth, growl and hiss. They are actually
quite timid in personality, but they are also talented performers, and
can put on a convincing act of aggression.
|
|
|
By pushing air through
their blowholes, Blue Whales, the largest mammals on earth, produce
subsonic booms that can be heard for thousands of miles around. |
Whales produce a wide range of sounds through their blowhole, reaching
far beyond human range of hearing. Male humpback whales use their blowholes
to sing rich complex songs during mating season. These vocalizations can
go on for twenty minutes, and they can repeat these performances verbatim,
which may evolve for years on end. Like bats; sperm whales, killer whales
and dolphins use echolocation clicks to find food and navigate through
dark underwater environments. Blue whales can produce low frequency booms
anywhere from 2-20 Hz, with enough force that they can be recorded half
a globe away. Above the water, we can hear whales shooting huge plumes
of vapor into the air, and quickly sucking in huge breaths before descending
again.
Birds do not possess a larynx, instead they use what is called a syrinx
which is positioned deep down within their chest. It is uniquely different
than the larynx in that it is split into two halves, creating two vocal
cords, which allows them duophonic note capability. Some birds, like thrushes
can use their two vocal cords independently, such as singing falling and
rising notes simultaneously. Some song birds alternate between the two
vocal cords, use them in detuned unison, designate each side for a different
frequency range, or even modulate the frequency of one side with the other.
These methods are all remarkably similar to the way that dual oscillators
are used in monophonic synthesizers. Most birds do not produce songs at
all but instead call out in short unmusical vocalizations in order to
communicate special calls for warnings, offense, territorial defense,
feeding, nesting, flocking and even enjoyment.
There is some evidence that animals are capable of telepathic communication.
In 1952, an island called Koshima in Japan, scientists were feeding sweet
potatoes to Macaca fuscata monkeys and watching their behavior. The potatoes
were dropped onto the beach, making them sandy, and one monkey, a young
female, came up with the idea of washing the potatoes in the ocean water.
Not only did it wash the sand off but it gave the potatoes a nice salty
taste, which caused a potato washing craze that spread throughout the
colony by evening. A few days later, something happened that caused monkeys
on other islands and the mainland to start washing their food in the ocean,
a trait that had never been observed before the occurrence at Koshima.
The thought is that when a certain critical number of monkeys had reached
awareness, the information spontaneously spread through the collective
unconscious, from mind to mind. This famous and well-documented event
is known as the "hundredth monkey" incident. Although telepathy
isn't thought of as an audible phenomenon, sound would be a great way
to portray its occurance in film and games.
Five DSP FX for Creature Creation:
Here is an assortment of advanced processing tricks for creature sounds.
By manipulating the sounds of animals, the human voice, ordinary objects,
or electronic sounds we can create the voices of fantastic beings. The
main focus being on technologies that are more unusual than mixing and
equalization. Also, these processes are very natural and can fully warp
the input signal without sounding like effects. Some of these techniques
would have been impossible just a few years ago, without the help of advanced
processors and recent software developments. By combining these methods
in different ways, we can come up with some truly original effects.
1. Morph
In 1988, for a film called Willow, Industrial Light and Magic created
the groundbreaking digital effect of a human shapeshifting into an animal,
earning them a Technical Achievement Award. This effect was coined "morphing"
and soon appeared in television ads, music videos, and several other motion
pictures. Another good example of this effect is the "mimetic polyalloy"
robot in Terminator II, morphing from "liquid metal" into the
various things that it samples through touch, such as other human beings.
It's quite an impressive effect, even when used judicially, such as in
the Lord of the Rings, when Gandalf releases King Theoden from Saruman's
spell and Theoden's face morphs back to normal.
|
|
|
Prosoniq Morph. Moving
the grey dot around adjusts the degree that input signals are morphed.
MIDI control signals can be assigned by the X/Y coordinates. |
In the 90's, a medical software company called Prosoniq became interested
in the application of artificial neural networks to sound manipulation.
This algorithm models human perception by creating artificial nerve cells
that can adapt and learn, in order to imitate such complex processes as
the way hearing works. Prosoniq used this intelligent artificial awareness,
to create a program called "Morph", which performs the sonic
equivalent of shapeshifting. With Morph (Mac, PC), one can re-synthesize
two audio signals in real time, and create hybrid versions of them, sharing
the acoustic properties of both.
An older package for doing audio morphing, although it's algorithm is
completely different, is Symbolic Sound's Kyma (Mac, PC), a hardware based
visual programming language that has been used a lot of major motion pictures,
including War of the Worlds, Lord of the Rings, and Star Wars: Attack
of the Clones. For morphing, Kyma uses a Fast Fourier Transform model,
which adds some amount of warbling and ringing in comparison to the "adaptive
transform" of Morph.
Using morphing technology, one can create very natural sounding transforms
between disparate sources such as human language, animal sounds and environmental
effects. The effect can be a static mixture, or a morph from one state
to another. This is a great tool for making unique creature sounds, especially
since the effect is rather clean and unidentifiable. This effect is particularly
alluring for people to experiment with because of its connection with
metamorphosis, folklore, and magic.
2. Granulate
.
As the Zen Master Taisen Deshimaru once said, "Time is not a line,
but a series of now points." What I think he means is that we should
not be worried about the past and future, and even though we experience
life as a sequence of events, but by being present and focusing on what
lies right before us, we focus on the most purposeful things in our lives.
When editing, or even thinking about how we might do our edits, we are
re-sequencing the time framework. Sitting in front of Final Cut with some
bad film footage, for example, our thoughts jump around to various points
in the shoot and we begin to imagine a new series of edits. As we cut,
drag and re-order the shots we are giving the scenes new substance, new
physical dimensions, a new passage of time. After we have edited out the
bad parts and reconstructed the scene, we can sit back like Zen masters
and experience the meaning that encompasses the film.
The smaller our edits, the more capability we have to manipulate the
scene. With the help of an advanced sequencer, we can break sounds or
images into molecular pieces, which gives us the ability to warp and coalesce
our media like controlled vapor.
In 1947, a British physicist named Dennis Gobor proposed a theory in
which any sound could be broken down into particles of time and frequency.
To demonstrate his ideas, he created mechanical devices that allowed him
to perform time compression and expansion of sound and moving pictures.
In the 1970's, a composer named Curtis Roads, having a great interest
in Gobor's ideas, started programming the first granular synthesis software,
a technology allowing recorded instrument sounds to be granulated for
the creation of music. It was a fantastic idea, perhaps a bit too complex
for early computers and it was not until the 1990's that CPU's could handle
the amount of data necessary to make this effect available to musicians
and sound designers.
Granular synthesis is a technique that involves breaking sounds up into
tiny grains and playing them back in a controlled manner to create longer
events called clouds. The grains are basically just small portions of
the waveform that loop and the cloud is a large number of these grains
playing back at the same time. A good analogy is the way particles of
water floating in the air look like mist when you see them from a distance.
The specific grain that is triggered at any given point is generally
defined by the location of a short 10-50ms section within the larger waveform
consisting of a start point and a stop point. Both of these points can
either move around within the waveform independently, move around together,
or be stationary. If the loop points were stationary we would probably
just hear a steady buzzing noise, but as we sweep it around the waveform
we hear flowing sonic movement. Each time a grain starts it overlaps the
previous grain by a small amount so that the grains blend together. In
this way all the grains sound like one continuous waveform when played
back; a smooth flowing cloud.
|
|
|
Native Instruments
Travelizer offers the basic goods for granular synthesis. The loaded
sample is visible on top and the X/Y pad on the right adjusts grain
size and position. |
We have control over the cloud densities, that is, how many grains are
heard at one time which can be hundreds or thousands per second. The key
to getting the most pleasing sounds out of granular synthesis is to keep
the grains moving about and use smooth attack and decay envelopes so that
when the grains are played in high density they all blend together nicely.
The end result is that we can completely alter the texture and shape
of a waveform, independent of its timing. This means, for example, that
you could play back all portions of a sound at the same time, sweep it
both forward or in reverse, move around the waveform in complex patterns,
or even compress and expand the length of a waveform without altering
the pitch (the effect is commonly known as pitch shifting.) Depending
on what granular synthesizer you use, this technology offers a tremendous
number of sound shaping options.
I recommend Native Instruments Reaktor (Mac, PC) for making creature
sounds. There are two ensembles included with the package: Grainstates
and Travelizer. Grainstates is more of a step sequencer with multiple
modules; well suited for musical sounds. Travelizer is a single module
with very advanced capabilities. On both ensembles, the interface is an
X-Y pad, which can either be controlled with a mouse, trackball, or graphics
tablet. The X dimension is assigned to grain start, and the Y dimension
controls the width of the grains. One important parameter is "inertia"
which can smooth out mouse movement, resulting in sweeter, more flowing
clouds.
3. Shift
|
|
|
Serato Pitch 'n Time |
Pitch shifters are effects processors that alter the pitch of a waveform
without changing it's timing. The first commercially available pitch shifter
was the Eventide H910 Harmonizer, an 8 bit audio processor released in
1975 that was first made popular by Bowie's producer, Tony Visconti, for
the doubled snare drums in "Low". These days, pitch shifting
effects are usually handled via software and include the option of formant
correction, a process that preserves the natural characteristics of the
human voice when forced into different pitches.
It is important to distinguish this effect from simply adjusting the
playback rate, such as changing the playback speed of a tape or record,
which also adjusts the length of the recording. When we pitch shift, in
order to preserve the timing in the original sound, the waveform is broken
up into tiny sections, a process similar to granular synthesis. Therefore,
when the pitch of the sample is changed each of these tiny sections plays
at a different speed, but their position in time, relative to one another,
does not change.
In order to create the voice of the 2005 King Kong, sound designer Brent
Burge used the live performance of Andy Serkis, the actor performing the
physical motions of Kong, captured for the CGI animations. These recordings
were made at 192 KHz directly into Pro Tools so that the pitch could later
be dropped substantially without losing higher frequencies. His pitch
shifted voice was fed to loudspeakers so that Serkis and the cast could
hear an approximation of the effect in real time. Burge later added lion,
tiger, elephant, bear, and gorilla sounds to further build and refine
the character of the voice.
Pitch shifting plugins are pretty numerous and as Craig Anderton stated
in the review of DSP/FX, "The only real question about a pitch-shifter
is how badly it messes up the sound when you do extreme transposition."
Celemony Melodine (Mac, PC) and Serato Pitch 'n Time (Mac) provide transparent
results when performing non-real-time pitch shifting. Ableton Live (Mac,
PC) and Prosoniq Timefactory (Mac, PC) also work well for non-real-time
shifting. For real time pitch shifting, Waves Soundshifter (Mac, PC) is
a good sounding DirectX plugin.
4. Filter
In the late 18th century a scientist named Wolfgang von Kempelen invented
a mechanical speaking machine. This tabletop device consisted of a bellows,
an oscillating reed, a rubber mouth, nostrils, and numerous mechanical
controls allowing manual operation by a single person. With some practice
one could produce words and short sentences by pumping the bellows, simulating
the lungs, and using levers to change the shape of the artificial vocal
tract. By no means was it able to accurately emulate the human voice,
but its sound was clearly that of a human being. Up until the early 20th
century other scientists and engineers continued this research, building
numerous mechanical voice machines, making only minor improvements on
Kempelens model. In 1939, at the New York World Fair a scientist named
Homer Dudley presented the Voder, the first electrical speech synthesizer,
a significantly less mechanical attempt at voice emulation. By the 1950's,
speech synthesizers had adopted the system of using an electrical tone
fed through a filter to create artificial speech.
|
|
|
MHC Vox FX |
A filter that is used to approximate the resonances inside an acoustic
body, in our case the vocal tract, is called a formant filter. If we compare
the human voice to a synthesizer, the larynx would be the oscillator and
everything that makes up the vocal tract would be the formant filter.
In order for these filters to make vowel sounds they must be fed with
harmonically rich waveforms such as a sawtooth, pulse, or white noise.
Most synthesizers do not include formant filters but they can be constructed
by using an assortment of 3 or more bandpass filters in parallel. Either
way works fine; I prefer the power and ease of using a plugin, such as
MHC Vox FX (Mac, PC), which allows user defined programmability and also
morphing between parameter settings.
It is best to select source sounds that are full-bodied, with lots of
texture and high end, so that the filter's shape is well defined. For
creature sounds one really powerful trick is that you can run normal object
sounds through the filters, like creaking doors, air bursts, or scraping
sounds, giving them a vocal character that blends well with animal sounds.
Sometimes formant filters can make animal sounds more artificial sounding
which is useful for sci-fi effects, if you want something that sounds
a bit more android. Automating the frequency and bandwidth of individual
formants over time will simulate the mouth changing shape. The sound of
this can lean towards synthetic if the adjustments are dramatic. Since
the formants are resonant, the sound may leap out unrealistically at times,
so it is helpful to use a dynamics compressor or limiter before or after
the filter, in order to tame the peaks.
5. Convolve
When creating creature sounds, it is necessary, especially in film production,
to use source sounds that are as dry as possible, meaning free of ambient
reflections so that the ambience can be added later. For example, we might
not want to be using dolphin clicks recorded at an indoor aquarium as
a design ingredient, because it could impart a particular ambience that
would not blend naturally with the rest of the sounds and scenery. But
for some creatures, acoustic reflections are an integral part of the voice.
For example, if we are designing sounds for a being that lives underwater,
an ethereal ghost, a creature that communicates telepathically or perhaps
even a large robot, the reflective nature of the sound could be so important
to the end result, that it should be designed into the voice. If our creature
is wearing a helmet, or some type of mask or if it has a tube shaped appendage
for a mouth then that might call for an effect simulating the voice as
it travels through or around these items; the reflection and obstruction
of sound. We could use standard reverbs and equalizers for tasks like
this, but one can achieve much more powerful and realistic results with
convolution software.
Convolution reverbs use what's called an "impulse response"
(IR) to produce realistic reverb effects. Usually the IR is made by playing
a short burst of white noise or a starter pistol shot into a reverberant
space while making a stereo recording of it. If we played this IR back,
it would, of course, just sound like nothing more than a blast with a
bit of reverberation. With the help of convolution software we can use
this IR to process another sound. The convolution reverb multiplies the
frequency content of the IR with the frequency content of the input signal
(the sound we want to process) which in turn makes the input signal sound
like its happening within the same acoustic space that we recorded the
IR in.
Making an impulse response of an acoustic space is similar to making
a 3D model of a house using photographs. We walk around the building taking
a photograph of each side which we then cut out and glue together to reconstruct
the building. When we are finished, we have a very realistic looking scale
model of the house, complete with its own lighting conditions and fine
geometric details. If we took a movie camera and panned around the model
of the house, it would generally look no different to the viewer than
the actual house. We've captured the way light bounces of the house and
stored it in a convenient little model that we can take anywhere. Maybe
though, there are some imperfections in the photographs such as a tree
or a telephone wire that we couldn't avoid when taking the picture. Maybe
one side of the house is a little out of focus, or unnaturally dark.
|
|
|
Altiverb with a
vacuum cleaner hose impulse loaded. |
The quality of the IR recordings is very important, especially for realistic
effects. It is usually necessary to have an extremely neutral and flat
frequency response in the recording gear. The noise floor of the microphone
must be extremely low and the background should be entirely silent, as
anything that goes into the recording will end up in the reverb. Also,
the IR will sound drastically different according to microphone placement.
Unless we are doing something unusual, sometimes it is better to simply
use commercial impulse libraries, which are usually excellent for reverbs,
especially since some of them were made in spaces that the average sound
designer couldn't access such Yankee Stadium, a volcano in Hawaii, or
a forest grove in New Zealand.
The space that the impulse is recorded in can be as small as a paper
cup. Not only that but we could also make the IR by recording sound as
it travels through water, fabric, around corners, coming out of a tiny
hole, through a crowd of moving people, or through electronic devices
such as vintage radio speakers. Since the software simply multiplies the
spectra of the two signals together, we can use any other sound effects
for impulses, allowing us to achieve effects similar to morphing and pitch
shifting. When this effect is used in non-standard ways the results can
be very unpredictable.
It is important to know that there will always be some smearing of attack
transients, therefore the output signal will seem a bit smoothed out.
Also, any frequencies that are prominent in both signals will be louder
in the output and frequencies that are not prominent will be attenuated.
In a way, the signals filter each other, because the only sounds that
get through are frequencies that they have in common. If you are loading
in particularly dissimilar sounds, on occasion, you will not hear anything
at all. Likewise, if the signals are very similar, or the same, the output
may be so loud that it is distorted. For experimental convolution, the
best results are achieved when the two waveforms share some frequencies
but not others.
I recommend Altiverb (Mac, PC), SIR Convolution Reverb (PC), and Waves
IR. The things that matter most with convolution reverbs are CPU load,
the quality of the impulses, and the programmability. The Altiverb impulse
library is particularly good and includes a photograph for most of the
impulses, giving a clearer idea of how they were made. The Synesthesia
library from Virtuasonic is an interesting set of impulses. When processing
lion roars with them I happened upon some sounds that were strikingly
reminiscent of the reverberation effects heard during the Mines of Moria
scenes in Fellowship of the Ring.
Dimension of Personality
Special effects are clearly a major part of films these days. Hearing
the impressive effects in Star Wars when I was younger is partly what
lured me into the field. Yet, even though today’s SFX are much more
dramatic and use state of the art technology, I feel more detached and
disinterested by today’s blockbuster films. The personalities of
the characters seem to be pushed into the background, making them feel
less believable.
|
|
|
HSSSSSSSSS! |
There is a simple precept in entertainment for some characters to be
corrupt and powerful, and for others to play the role of good and be in
need of defense. We are so familiar with this good vs. evil template that
once the script introduces this prescribed formula, our curiosity about
each character's motives is pretty satisfied. A famous quote from Alien
Resurrection is when Ripley says, "There's a monster in your chest.
These guys hijacked your ship, and they sold your cryo tube to this...
human. And he put an alien inside of you. It's a really nasty one. And
in a few hours it will burst its way through your ribcage. And you're
gonna die. Any questions?" That pretty much describes what the aliens'
intentions are in full detail. The aliens are simply evil and they will
kill the good guys. Then we move our attention onward, to unsolidified
aspects of the story, like which characters are going to live. This is
an example of a creature with a one dimensional personality, with little
room for anything but nasty sounds. The sound designer makes a loud nasty
screeching sound for them and everyone is happy.
In Spirited Away, a film by Hayao Miyazaki, a spirit called Kaonashi,
meaning "no face", appears as a mysterious masked man, partly
translucent, silent and passive in nature. The main character, Chihiro,
invites him into a bath house, the place where she works. Kaonashi follows
her around, offering gifts, which she refuses. When he offers things,
he makes the quiet sound of someone helpless and innocent, perhaps even
hurt, which naturally makes us believe that Koanashi is benevolent. He
offers gold nuggets to other people in the bath house, staff members,
which they all gladly take, and in turn he eats them. Koanashi uses this
friendly sound to lure people in so that they give him food for the gold,
which he always devours. If they do not give him food he actually feasts
on whomever took the gold from him. After spending more time in the bath
house he slowly becomes enraged, grows to massive proportions, and devours
everything in sight. The bath house staff does not fight him, but continue
to treat him as a guest, although more cautiously. Kaonashi begins to
speak in a shrill comical voice that he wants to eat Chihiro. Unafraid
of Kaonashi, Chihiro tries to help him, believing that the bath house
is making him crazy. She eventually gets him out, and he slowly returns
to the silent, benign form.
This is an example of a creature that has multiple personalities, and
two completely different voices, one almost silent, one very loud, which
enhances the depth of its character. It could have had one voice sitting
comfortably between nasty and nice but with a dual voice, it's personality
gains complexity, it's motives are more ambiguous, and it is less predictable.
The Loudness Trend
In her book Sounding of the Soul, Mary Lynn Kittelson stated "With
myths being trumpeted forth with over-whelming dramatics, overpowering
high tech effects and high volume sound systems, there is less gently
imaginative space to imagine the characters and actions and settings."
The structures of most films and video games are so familiar to us that
our attention often cannot find anything new to focus on except the latest
special effect, which is never enough. The sound of today’s films
is fantastic but there is this surge forward to fill more and more of
the sound field so that every pocket of silence is gone.
The Exorcist (1973) used restraint to frighten us, a brooding quietness.
In this way, our imaginations ran wild, creating a fear in our chests
that has never been surpassed in any other film to date. The director,
William Friedkin says:
"What we did back then was very instinctive; you put the dialogue
in the way it feels right, you play the effects and the backgrounds for
whatever emotions you want to get out of them. But today the dialogue
is all in your face, the sound effects are in your face, the music’s
right up there and they just beat you over the head with them all for
hours. That’s completely different to THE EXORCIST which was actually
often a very silent movie – a loud silence."
Rather than using a dramatic sound effect for ever single action of the
creature, consider the impact that silence can have on the scene. Also,
if you want any particular effect to jump out, then dial back the volume
on other effects. What many filmmakers do not understand is that the way
to make a film louder, is not to push everything up to full, but to actually
pull back. This subtractive method will allow more headroom in the signal,
allowing dramatic events to jump out far beyond everything else.
Epilogue
I hope that the information provided here will be inspirational. Remember
to keep the audience engrossed and create creatures that draw attention
to themselves rather than provide something safe and familiar. Instead
of only researching your favorite films, research science, nature, and
biology to create your beings. Experiment. If you do this, you will come
up with an immense amount of new ideas, unprecedented techniques that
you can bring into your projects and the industry. Being inspired by other
sound designers is a powerful propellant, but dig in and find out what
makes their work so inspirational to you and use that knowledge as a launch
pad for innovation. If you steer clear of the common course and come up
with your own recipe, you might truly leave your mark in the field of
sound design.
George Sanger, when speaking about John Williams' contribution to Star
Wars, and how it continues to inspire similar soundtracks, said:
"Begin by imitating his compositions,
but go on to learn how to create his feelings. When you imitate his
orchestration, continue, and learn how to be evocative of emotions.
When you imitate his greatness, learn how it progresses; learn how it
leaves its own roots behind."
Research Materials:
Papageorge, John. Interview. Radio Interview with Ben Burtt. Silicon
Valley Radio. 1996
Lord of the Rings, Special Extended DVD Edition: The Fellowship of the
Ring. Dir. Dir Peter Jackson. New Line Home Entertainment. 2001
Lord of the Rings, Special Extended DVD Edition: The Two Towers. Dir.
Dir Peter Jackson. New Line Home Entertainment. 2002
Lord of the Rings, Special Extended DVD Edition: The Return of the King.
Dir. Dir Peter Jackson. New Line Home Entertainment. 2003
Star Wars. Dir. George Lucas. Perf. Mark Hamil, Harrison Ford, and Carrie
Fisher. Paramount, 1975.
Brunelle, Ray."The Art of Sound Effects." Rumblestiltskin Sound
FX Design. (1999): n.pag. Online. Internet. Available. web.archive.org
Benedikt, Reinhard G."Prosoniq Morph - Behind the scenes."
Bernsee.com 1 Nov (2003): n.pag. Online. Internet. Available. www.dspdimension.com
"King Kong: Mammoth Post Power Yields Epic Results" Avid (2006):
n.pag. Online. Internet. Available: www.avid.com
Sasso, Len. "Fun with Formants." Electronic Musician. 1 Jun.
(2006): n.pag. Online. Internet. Available: emusician.com
Traunmüller, Hartmut. "Wolfgang Von Kempelen's Speaking Machine
and its Successors." History of speech synthesis, 1770 - 1970. (2000)
n.pag. Online. Internet. Available: www.ling.su.se
Gordon. "Granular Synthesis." Granular Synthesis. 16 Oct (1996):
n.pag. Online. Internet. Available: homepages.kcbbs.gen.nz
Kemmler, Alex. "Acting on Impulse." Electronic Musician. 1
Jun. (2006): n.pag. Online. Internet. Available: emusician.com
Cited Materials:
Alien Resurrection. Dir. Jean-Pierre Jeunet. Writ. Joss Whedon (screenplay).
20th Century Fox, 1997
Deshimaru, Taisen. "What the Bleep Do We Know!?" What the Bleep!?
(2004): n.pag. Online. Internet. Available: www.whatthebleep.com
Kittelson, Mary Lynn. The Sounding of The Soul. Einsiedeln: Daimon, 1996
Sanger, George Alistair. The Fat Man on Game Audio: Tasty Morsels of
Sonic Goodness Indianapolis: New Riders, 2003
Sharp, Daryl. "Extracts from The Jung Lexicon: A Primer of Terms
& Concepts." Compiler Press (2002): n.pag. Online. Internet.
Available: www.compilerpress.atfreeweb.com.
"The Sound of Silence." The Exorcist, the Film, the Phenomenon,
the Aftermath. (2000): n.pag. Online. Internet. Available: theexorcist.warnerbros.com
Wally, David. "Aural Surgery." USC Trojan Family Magazine (2001):
n.pag. Online. Internet. Available: www.usc.edu.
Yewdall, David Lewis. Sound Design Myths and Realities. New York: Focal
Press, 2003
[back]
|
|