← BACK TO SHOP
← BACK TO SHOP

Sonic Utopia: Can we build a better sounding world?

Art by Matthew Fleming.

This episode was written and produced by Casey Emmerling.

Technology has the power to transform the way our world sounds. It could even give us entirely new ways to experience our surroundings. In this episode, we explore the sounds of the future, and how we can use the tools we already have to build a better sounding world. Featuring Rose Eveleth, Creator and Host of the podcast Flash Forward, Acoustician Andrew Pyzdek, and Architect Chris Downey.


MUSIC FEATURED IN THIS EPISODE

Wanderer by Makeup and Vanity Set
1:57 AM (The Green Kingdom Remix) by Hotel Neon
Wide Eyed Wonder by Dustin Lau
June by Uncle Skeleton
Rubber Robot by Sound of Picture
Lick Stick by Nursery
Stuck Dream by Sound of Picture
Springtime by Sound of Picture
Drawing Mazes by Sound of Picture
Dark Matter by Sound of Picture
Brackish Water by Alistair Sung
About You (No Oohs and Ahs) by Vesky
Trek by Sound of Picture
Gimme Gimme - Instrumental by Johnny Stimson

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Discover more at lexus.com/curiosity.

Subscribe to Flash Forward wherever you get your podcasts.

View Transcript ▶︎

You’re listening to Twenty Thousand Hertz. I’m Dallas Taylor.

[music in]

In our busy modern lives, it’s not often that we stop and really think about what we hear.

[SFX: Alarm, car start, city ambience, train, office ambience]

Most of the time, we just accept these human-made sounds without a second thought. But it’s important to remember that our world didn’t always sound this way. Matter of fact, the sound of our world changes constantly. Our cities and towns sound completely different now than they did fifty or a hundred years ago.

[SFX: Horse, old timey, car horn, klaxon, car pass by]

So what will our cities and towns sound like fifty… or even a hundred years from now? What if we could collectively sound design our world? What would that sonic utopia be like, and how can we get there?

[music out]

In our future sonic utopia, there will certainly be sounds we want to remove.

Rose: The first thing that comes to mind is the screech of the New York City subway, [SFX] which is incredibly loud and is sort of emblematic of the lack of updating of that city's infrastructure.

That’s Rose Eveleth.

Rose: I'm the creator and host of a podcast called Flash Forward, which is all about the future.

[music in]

[SFX: Subway train screeching as Rose describes it]

Rose: I love New York. But standing on the one train platform and the train rolls in [SFX] and you really feel like you're being stabbed in the ear.

We all know that loud noises can cause hearing loss, but that’s just the tip of the iceberg. When we’re exposed to loud noise, our bodies release stress hormones. These hormones raise our blood pressure, which contributes to heart disease and Type 2 diabetes. Studies have even shown that kids who go to school in louder areas tend to have more behavioral problems, and also tend to skew worse on tests.

Rose: We know that constant sound like that has real impacts on learning, on people's ability to retain information.

[music out]

To be clear, this train problem goes way beyond the New York City subway. Anytime you have a heavy metal object moving along a metal track, like a subway, or a train, you’re probably going to end up with some screeching. But what if, in the future, that subway or train car wasn’t even touching the track?

[music in]

In recent years, some countries have begun building “maglev” trains, which is short for “magnetic levitation.” A maglev train doesn’t have a conventional engine. Instead, it uses powerful electromagnets to stay suspended above the track.

[SFX: Maglev train]

When it’s suspended, another set of magnets propel it forward.

Andrew: They don't have rails that they're rolling along physically.

That’s acoustician Andew Pyzdek.

Andrew: They're basically floating on a cushion of air, so they can be very, very quiet.

Maglev trains aren’t just quieter than normal ones, they’re also smoother and faster. The highest speed recorded on a maglev train was 373 miles an hour.4 As maglev trains replace normal ones, we can expect to hear less of this [SFX: Normal train] and more of this [SFX: Fast maglev].

[music out]

That’s a great start, but what about cars? As you may already know, electric cars can be extremely quiet. [SFX: Low-speed electric car]

Andrew: Obviously they're quieter, but most importantly at low speeds where the engine noise is the loudest thing.

As electric vehicles become more common, areas with low-speed traffic will quiet down quite a bit. But once you get on the freeway, even an electric engine doesn’t help that much.

Andrew: At high speeds on the freeway, most of the noise actually comes from the tires. [SFX: Car pass by] So the advancements that you can expect to improve roadway noise are… with the composition of the roads themselves.

In 2007, researchers developed a new paving method called Next Generation Concrete Surfaces.2 Roads that are paved this way are up to 10 decibels quieter than normal ones. That’s the difference between this [SFX: Busy freeway] and this [SFX: Busy freeway by -10bd quieter].

Andrew: As we improve infrastructure and replace old roads with new roads that are made to be quieter, we can see those freeways becoming less noisy.

[music in]

When it comes to transportation, our sonic utopia is sounding a lot quieter. But of course, people aren’t the only things being carried around in cars and trains. There’s also all of our stuff. Amazon currently ships over 6 million packages a day.6 As populations increase and countries develop, there will be an even bigger demand for quick delivery. And the latest idea to handle that demand is delivery by drone.

[music out]

Now, commercial drone delivery hasn’t taken off quite yet, but both Google and Amazon are working on changing that. Here’s a clip from a video that was made by Amazon:

[SFX Clip: “Amazon: Here’s how it works: Moments after receiving the order, an electrically powered Amazon drone makes its way down an automated track, and then rises into the sky with the customer’s package on board. Cruising quietly below 400 feet, carrying packages up to 5 pounds…”]

Amazon describes these drones as “quiet,” but in their videos, they never include the actual audio of flying drones. That’s probably because drones really aren’t that quiet. Even the small ones that hobbyists buy can be pretty loud. [SFX: drone]

If drone delivery becomes common, things could get really loud, really fast. Imagine a crowded city with hundreds of delivery drones buzzing by at all times. [SFX: Sparse drones]

Now imagine how bad it would be near the fulfillment center, where the drones actually take off and land. [SFX: Heavy drones]

This is not very utopian. But, thanks to nature, there may be ways of making drones quieter.

Andrew: There's some work that's been done looking at owls and the way that their feathers are shaped in order to reduce noise. So the edge of an owl's feather is very ragged. The feathers themselves are kind of loose and wavy. And that's why they're such stealthy fliers because their feathers aren't rigid.

For instance, barn owls fly so quietly that humans can’t hear them until they’re about 3 feet away.

Andrew: The exact opposite of that is a pigeon. And every time they take off, that pigeon sound, [SFX: Pigeon] some people think that it's a vocalization that the pigeons are making. That's the sound of their feathers vibrating as they flap their wings.

The recording of the pigeon you just heard is from a BBC special about owls. In the special, they recorded a pigeon, a hawk and an owl flying over a set of microphones. Here’s the pigeon again: [SFX: Pigeon] Here’s the hawk: [SFX: Hawk]

And here’s the owl: [SFX: Owl]

Did you catch that? Neither did I. Here it is again, turned up twice as loud: [SFX: Owl]

Inspired by owls, researchers are already exploring ways of making airplanes quieter.

[SFX: Airplane]

Like a car on a freeway, a lot of the noise from a passing plane comes from air flowing around the plane. One way of reducing that noise would be to make the plane’s wings more like owls’ wings. This could be done by adding more flexible, porous materials to the edges of the wings. Theoretically, something similar might be possible with drones.

Andrew: I think that if drones start being a more everyday part of our lives, that there will be a pretty strong pressure to make those drones be a little bit less annoying to listen to.
 [music in]

So far, we’ve turned down the volume on future cars, trains, planes and drones. Not too shabby. But what does our sonic future sound like if you’re getting around on foot? Something that might become common is targeted audio messages that you can hear as you walk down the street. When audio is beamed to a small, specific area, it’s called an “acoustic spotlight.” These are already found in many museums.

Andrew: Say if you were looking at a painting, you might hear sounds that remind you of the space and the painting.

For instance, you may walk up to a painting of a peaceful landscape, and hear this

[music out]

[SFX: birds, light wind, blowing grass].

Andrew: And the technologies that are used to make these acoustic spotlights can range from very simple: There's parabolic microphones, where you have just a plastic shell around a normal speaker. And as that speaker generates sound, it focuses it downwards towards the person standing under the spotlight.

But acoustic spotlights can also be made with ultrasound.

[music in]

Andrew: Ultrasound is very amazing. Ultrasound is sound. It's not something different. It's just sound that's at a frequency above what people can hear.

The normal range of human hearing is from about twenty hertz to twenty thousand hertz [SFX: Sine wave of 20 hertz sliding up to 20,000 hertz]. Anything above 20,000 hertz is considered ultrasound. Making an acoustic spotlight with ultrasound involves something called a “parametric array.”

Andrew: So parametric arrays are basically you have two beams of ultrasound that you make intersect with each other. And at the point where they intersect, they create audible sound.

[music out]

A parametric array is almost like a sonic laser that lets you beam a sound message to a very precise spot [SFX: VO shifting around like it’s looking for a target]. If advertisers started doing this, it could get out of hand pretty quickly. Imagine you’re walking downtown in a crowded city. [SFX: Times Square ambience]

Every time you pass by a billboard or a store or a restaurant, you hear a little commercial or jingle. [SFX: Walmart] [SFX: McDonalds] [SFX: New in theaters] [SFX: Parasitic infection] [SFX: Pringles]

That’s definitely not what I want in my utopia. But audio aimed at your location doesn’t have to be a bad thing. For instance, rather than just playing the sound out in the open, the signals could be beamed to a device, like a specialized headset. That way, you could choose whether or not to tune in.

[music in]

Rose: I think in my utopia people would be able to kind of customize their experience to themselves. I could make the world feel safe and happy and lovely wherever I am and that might look different from somebody else. And I don't know if that means special things that go in my ear that kind of like filter in and out the sounds that are important or not. Or whether that means high-tech technology that only beams aural information to certain people who have their profile set up to be like maximum sound versus minimum sound, or whatever it is.

Rose: And you can kind of choose to customize your experience of the world that way.

In the future, headphones and earbuds won’t just be headphones and earbuds. They’ll be much more integrated. We already have noise reduction, but future hearing wearables may have selective noise reduction. They may filter out unpleasant sounds, or reduce dangerous volume levels. They may even have corrective hearing loss algorithms built in… like a merger of current hearing aids, noise protection, and traditional earbuds.

[music out]

With geolocation targeting, these headsets could give you extra information about your surroundings, without the visual distraction of smart glasses. Imagine kind of an audio tour of the entire world. This might even help people build more of a connection with their community.

[music in]

Rose: I think that there is a space for like, a sort of community audio project where you could have this living audio document that is kind of like a museum tour, but for your own space.

Rose: So you could be walking down your street and you could hear a story from your neighbor about something.

Rose: It's maybe the person who's lived on that block for 30 years being like, "You might not know this, but here's an interesting piece of history about where you're from." Or, you know "Hey, there's a city council meeting today. Maybe consider going to it."

Rose: Just little things like that where you could constantly be keeping up with your neighbors or understanding what the needs are in the community.

[music out]

A hightech headset that you wear all the time could also be a game-changer when it comes to real-time translation. If every word you hear gets instantly translated into your native language...

Andrew: People can talk to other people speaking a different language and not have that language barrier. There's already quite a bit of work happening there, and that will continue to move forward.

[music in]

New technology could positively change how our cities, neighborhoods, and homes sound. It could even give us entirely new ways to experience our surroundings. But we have to put in the time and effort if we actually want our future to sound better. To get some perspective, it’s helpful to talk to someone who really understands how important sound is to the spaces we design. Maybe someone like an architect… who’s also blind. That’s coming up, after the break.

[music out]

MIDROLL

[music in]

From acoustic spotlights to swarms of retail drones, there’s all kinds of technology that’s likely to affect the sound of our future. But as we build that future, we have to ask ourselves… What kind of sonic environments do we want to create?

The spaces we design should be acoustically functional. In other words, the acoustics of a building should support whatever goes on in that building. But it’s about more than just utility. We also want the places we spend time in to just sound good.

[music out]

That’s easier said than done, and too often, people just don’t think about it.

[music in]

Chris: Sound is so often just left to be accidental.

That’s Chris Downey.

Chris: I'm an architect located in Piedmont, California just outside of Oakland.

As far as architects go, Chris has a pretty unique background. In 2008, he was having some trouble seeing clearly. An MRI scan revealed a brain tumor right against his optic nerve. Fortunately, doctors were able to remove the tumor through surgery. But two days after the procedure, Chris’ vision started to fail. After three days, he was blind.

[music out]

Of course, adjusting to life without sight took time. But Chris didn’t let blindness stop him from doing what he loves. In fact, he says that losing his sight has actually been helpful to his understanding of architecture.

[music in]

Chris: Losing my sight, as an architect, has really benefited my work by really getting me back in touch with the human bodily experience of being in the space at any given moment in time.

Chris: The sound of the space. The acoustic soundscape of the architecture [SFX: footsteps] as you move through it dynamically, hearing it as you move through and really listening to the architecture.

The choices that are made when a building is designed have a massive impact on what it’s going to sound like. Sometimes, these choices are very deliberate, like the way concert halls are designed to amplify and enhance the sound of an orchestra.

[music out, SFX: Applause]

A lot of the time, though, it can be hard to predict exactly what a building is going to end up sounding like.

Chris: It's so hard to draw or model sound. How do you do that? As architects, we can't do that. If we talk with an acoustic engineer, they might be able to give us all sorts of scientific representations of things. But unless you're a highly trained acoustic engineer, it means nothing.

Most of the time, you won’t really know until the building is finished.

Chris: And then it’s built. It's really too late.

At that point, you might not be happy with the result. Maybe you’ve had the experience of trying to study in a library where every little noise echoes off the walls. [SFX: Library ambience with heavy verb]

One way to prevent this is to digitally emulate what a space might sound like, while it’s still being designed.

Chris: There's been a really interesting collaboration I've had with some acoustic designers that have a sound lab that they use to model sound. They use it really to anticipate and demonstrate the sound of a music hall or some other very, sort of, acoustically intentional space.

Using this technology, you can input the dimensions and other aspects of a building you’re designing. Then, the computer can emulate what a voice... [SFX: Voice with room verb] or an instrument... [SFX: Saxophone with same room verb] or a footstep… [SFX: Footstep with room verb] is going to sound like inside that space.

[SFX: Over Chris’ next line we hear those same sounds with evolving room verbs]

Chris: You can tune it. You can test it, just as we do visually with drawings and models and photorealistic computer-aided renderings and things, it's doing the same thing with sound.

Chris: We started working with that for me to anticipate the dynamics of sound as you move through a space, so they put my cane tapping inside the digital space [SFX: Cane tapping] and then we hear what it's like to hear the architecture as you move through, and anticipate that, so that I can really design intentionally

Acoustic modeling technology isn’t universal yet, but some designers have started taking acoustics more seriously.

[SFX: Noisy airport sounds]

Airports are notoriously noisy, and all of that noise can make traveling even more stressful than needs to be. But many airports have started installing noise absorbing materials to help keep people calm. The next time you’re in a new terminal that feels unusually quiet, look up at the ceiling. Oftentimes, you’ll see very unique looking tiles. These tiles can be subtle enough to fit right in with the architecture, and they make a huge difference in sound quality.

Unfortunately, though, when it comes to noise, restaurants are still way behind. We’ve all had the experience of being in a restaurant that’s just uncomfortably loud. [SFX: Crowded restaurant gradually getting louder; chatter, silverware clinking]

Chris: There are environments in restaurants that the soundscape becomes really problematic.

Since Chris is blind, he can’t read someone’s lips or pick up cues through their facial expressions.

Chris: So I'm absolutely dependent on the acoustic environment to communicate. And some of these environments are so loud, it's just so exhausting to try to hear, that within 15 minutes, I'm done. I'm exhausted. I’ve had enough. [SFX: Restaurant sounds out] And in sharing that with other people, people with hearing impairments, they have the same experience, and it could be because of a hearing aid that the sound is very different, and it becomes nauseating.

Accessibility laws and city codes are the reason we have helpful sounds at crosswalks, [SFX] and ramps for wheelchairs. And while the US government does regulate how much noise workers should be exposed to, those codes are rarely enforced in places like restaurants and shops.

Chris: Our codes don't really deal with that, so I think that there's some more wisdom and more research and development that needs to come into creating safe environments in places like that.

[music in]

We used to talk about second hand smoke in bars. Well, you know, what's that acoustic environment doing to the health of the people that work in those environments?

Whether it’s noisy restaurants or noisy freeways, it’s easy to imagine that a quieter future would be a better future.

Andrew: I think that we kind of want silence more than we get it. And that's really what it comes down to is that we live in a very loud world. [SFX: Loud city montage] Finding silence is very difficult unless you live in a place that's already pretty quiet. So I can understand why the focus on making the world sound better is to make the world sound less. Because it sometimes feels like there's just too much vying for our attention.

[music out]

If I had a giant audio board for the world, I’d pull the fader down on most of what’s human made. Our brains love the sound of nature, and it would be great to get competing sounds out of the way. However, that doesn’t mean that all human made sounds should be lost.

Chris: There's been a lot of effort going into sound masking, masking of the sound in an environment, which from the blind experience isn't necessarily a good thing, because in masking the environment, we're losing some of the necessary sound. We need to hear the environment.

For instance, making cars completely silent could be dangerous.

Rose: I mean, I think many people probably have the experience of almost being hit by a Prius in a parking lot because you didn't notice it there because it doesn't make any sounds.

Chris: I've experienced new electric buses that are so quiet it's hard to even know they're there. I've had one that pulled up right in front of me when I'm standing at the sidewalk, and I didn't hear it approach and I kind of sensed there was something in front of me, and I reached up to find there was a bus there just a couple inches in front of my face. [SFX: Bus pulls away] And that was terrifying. So in trying to remove sounds and make some of these things quiet, you have to be careful about maintaining some necessary sound for safety.

All of this can feel overwhelming, but there are things you can do to make your surroundings sound a little better. The first step is to really hear your environment. To do that, you’ll need to make it as quiet as possible.

[SFX: Subtle HVAC sounds]

Andrew: Try powering off your house. Go to the breaker, cut off the power. [SFX: Power down, HVAC off] Assuming that that's not going to damage anything, turn off any sensitive electronics first that might get hurt by a brownout. But you can flip the breaker and hear how different your house sounds when there's nothing on. And then when you turn those individual breakers on, you'll notice right away. [SFX: Click + fridge] "That's what my refrigerator sounds like." Or, [SFX: Click + AC] "Wow, I didn't realize our AC unit was that loud." You usually don't notice these things until they're gone and they come back.

[music in]

Most of us can’t just go buy a quiet new AC unit, but this can still be a good exercise to help you notice the sounds that you may have been ignoring. Maybe you can power down that video game system all the way, rather than leaving it in sleep mode with the fan running. Maybe it’s time to put some WD-40 on that squeaky closet door. Maybe you can find a tapestry to hang in your living room. Soft surfaces are a friend to good sounding environments.

Maybe you could also write a friendly email to that restaurant that you’d love to go back to, if it wasn’t quite so loud. Or maybe you can write a letter to your mayor, or your representatives, and tell them how the screeching bus brakes wake your whole building up at 6:30 in the morning. The point is, even a little sonic change goes a long way. And if enough people start doing this, our future will sound better.

Chris: Sound can affect us on a subliminal level and it can set a mood it can make us struggle. It can put us at ease. So, I think it's a sense that we really need to pay a lot more attention to, to really add to the quality of our living experience, in whatever setting we're in.

Andrew: There's a lot that can happen right now that would be possible if people just were willing to do it.

[music out]

[music in]

Twenty Thousand Hertz is hosted by me, Dallas Taylor, and produced out of the sound design studios of Defacto Sound. Find out more at defactosound.com.

This episode was written and produced by Casey Emmerling, and me, Dallas Taylor. With help from Sam Schneble. It was edited and sound designed by Soren Begin, Joel Boyter, and Colin DeVarney.

Special thanks to our guests: Rose Eveleth, Andrew Pyzdeck and Chris Downey. Rose’s podcast Flash Forward is one of my favorites, it’s all about the possible and not so possible futures. You should definitely go subscribe. You can also find articles by Andrew at acousticstoday.org. And you can learn more about Chris’ work at arch4blind.com. That’s A-R-C-H, the number 4, blind dot com.

If there’s a show topic that you are dying to hear, you can tell us in tons of different ways. My favorite way is by writing a review. In that review, tap 5 stars and then give us your show idea. And even if you don’t have a show idea… I’d love for you to give us a quick 5 star rating anyway. Finally, you can always get in touch on Twitter, Facebook, Reddit, or by writing hi@20k.org.

Thanks for listening.

[music out]

Recent Episodes

The Cyborg, The Scientist, and the Death of Silence

The Cyborg, The Scientist, and the Death of Silence

Sound surrounds us, even in the quietest moments. But depending on how we hear, the world can be a different auditory experience for each of us. This episode features four stories from the TED Radio Hour.

Dies Irae: The death melody haunting Hollywood

This episode was written and produced by Casey Emmerling.

For hundreds of years, composers have been using a specific four-note melody to evoke death. It's appeared in dozens of famous movies, and you probably never even realized it. But once you hear it, you'll start noticing it everywhere. Featuring musicologist Alex Ludwig and Strong Songs Host Kirk Hamilton.


MUSIC FEATURED IN THIS EPISODE

Dolly Pop by Piano Mover
Dead Hour by Alistair Sung 
We Rise by Generdyn
Falling (Instrumental) by Asher 
Suite from Quarantine by Davis Harwell
Trois Gnossiennes 3 by The Nocturne
The Wraith by Tokyo Rose

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Discover more at lexus.com/curiosity.

Check out Alex’s list of Dies Irae examples at alexludwig.net.

Subscribe to Strong Songs where ever you get your podcasts.

Check out The Graduale Project’s Youtube page for more Latin chants.

View Transcript ▶︎

You’re listening to Twenty Thousand Hertz.

[music in]

If you look up the top charting songs from a random year, maybe from 10-15 years ago… there’s a good chance that you’ll recognize a few of them. But you’ll probably be surprised at how many songs don’t even ring a bell. Because times change. Styles come and go, and the vast majority of music becomes a relic of its time. But once and awhile, the stars align so that a single piece of music takes hold and will not let go.

[music out]

This is especially true when the music becomes associated with important life events, like birthdays.

[Music clip: Happy Birthday]

Despite the Happy Birthday Song being written over a hundred years ago, it’s still being sung countless times a day, all over the world.

[music out]

Here’s another one:

[Music clip: Wedding March]

Felix Mendelssohn's “Wedding March” was written in the eighteen forties, but we still think of it as the wedding song.

[music out]

*So for birthdays, we’ve got “Happy Birthday.” For weddings, we’ve got the “Wedding March,” And when life ends, well, we have music for that, too. Another classic song you probably recognise is Chopin's “Funeral March”. *

[Music clip: Funeral March]

That song is famous, but there’s actually another song—or at least, a melody—that has an even deeper connection with death.

[music in]

For hundreds of years, composers and songwriters have been using this melody to evoke feelings of dread and despair. You’ve probably heard it dozens of times and never even realized it. But it’s one of those things where, once you know about it, you’ll start noticing it everywhere. It’s called the Dies Irae.

[Music clip: Dies Irae]

At this point, the Dies Irae has been used by so many composers that it’s basically a meme.

Alex: A meme is this thing that grows and changes as it happens. And that's what the Dies Irae really is.

Alex: This meme that these composers can draw on over and over and over again.

That’s Alex Ludwig, a musicologist and assistant professor at the Berklee College of Music. Alex keeps a list of every instance of the Dies Irae in a movie or TV soundtrack that he can find. One of his favorite examples is from the first Star Wars movie.

Alex: I'm a huge Star Wars fan. And in the very first Star Wars film, there's a very prominent example of the Dies Irae.

Alex: The cue is called the Burning Homestead cue. This is the moment in the film where Luke Skywalker returns back to his home and the Stormtroopers have just burned it to the ground, killing his aunt and uncle.

Alex: And at that moment… the camera pans over to this burning skeleton, we hear this musical cue that we've heard a few times already in the film. But what John Williams does is he adds on these first four notes of the Dies irae.

[Music clip: Star Wars music up]

By quoting the Dies Irae in the score of Star Wars, composer John Williams was drawing on a tradition that goes back all the way to the middle ages.

Alex: To go back to the beginning, we have to go back to the 13th century.

[Music clip: Dies Irae Chant]

This is a modern recording of the original Dies Irae chant.

Alex: So we're talking monks singing Latin chants at a requiem mass, the sort of mass for the dead.

“Dies Irae” is Latin for “Day of Wrath.” It began as a medieval Catholic poem describing the Last Judgement—the end times when the righteous ascend to heaven, and the wicked are banished to hell. No one is totally sure who wrote that poem, when they wrote it, or when it was set to music.

Alex: And this would have been a common performance for funerals, any sort of Catholic Church in 1200, 1300, etc, would have played this and this would have been part of the sort of religious vernacular. So for about 500 years, this piece of Gregorian chant would have been common. Everyone who would have gone to a Catholic mass would have heard this at a funeral.

Alex: So, from the very beginning, there is this connotation of this chant, this text with death.

[music out]

The entire piece is about four minutes long. But the most famous part of Dies Irae—the part that gets reused over and over—is the first four notes.

[Music clip: Dies Irae Chant]

Kirk: The four notes in question are the flat third and then the second. Then, the flat third and then the root, which sounds like this [SFX: Melodica example].

That’s Kirk Hamilton. He’s the host of Strong Songs, a podcast that breaks down iconic songs to figure out what makes them work. He’s also a lifelong musician and a fan of the melodica.

Kirk: I think it's a good phrase because it resolves to the one, which is where a lot of musical phrases want to resolve. That final note is the root. It's the tonic note… If I'm playing in C, it sounds like this [SFX: Melodica example].

The thing that makes this musical phrase so powerful is that it includes both dissonance and resolution.

Kirk: The first two notes, there's a half step between them, which is… the kind of tightest interval you can have. And that's very dissonant. [SFX: Melodica example] If you play them at the same time, [SFX: Melodica example] it's very dissonant, especially on my slightly out of tune melodica. When you, then, resolve it, the second two notes are just a minor third. They're further apart. They resolve really nicely. [SFX: Melodica example] When you play those at the same time, it sounds like this. [SFX: Melodica example]

This minor third is key to the melody’s dark, foreboding feel.

Kirk: Half of the notes are the minor third, so it's very minor-third-heavy.

Kirk: Not everyone hears a minor key and feels sad, but it is definitely something in America and in Europe and where a lot of movies are made, that when you hear a minor scale, it just sounds sad.

The downward movement of the melody also adds to its sense of heaviness.

Kirk: Which I think is part of the reason that people really like it in movies. When you want something to sound hopeful, typically, the melody will move upwards. [SFX: Uplifting melody] It'll be kind of this feeling of flight of lifting up. When things are kind of coming crashing down, music tends to move downward. A lot of times, you'll hear those like [SFX: Singing “bum, bum, bum, bum”], in a big dramatic scene. That final note [SFX: Root note], that big tonic resolution, just feels like this avalanche kind of landing on top of you.

For hundreds of years, the Dies Irae was an essential piece of Catholic music. Eventually, composers started to incorporate it into their own religious pieces.

Unlike the film composers that came later, most of these early composers used the exact words of the original Dies Irae poem, but set them to different melodies. For instance, the Dies Irae appeared in a mass written by the French composer Antoine Brumel:

[Music clip: Brumel Dies Irae]

During the 18th and 19th centuries, the Dies Irae text was also used in symphonic compositions. Joseph Haydn’s younger brother, Michael Haydn also used the Dies Irae:

[Music clip: Haydn Dies Irae]

Here it is in a requiem by the Italian composer Giuseppe Verdi.

[Music clip: Verdi Dies Irae]

Even Mozart got in on the action.

[Music clip: Mozart Dies Irae]

In all of these cases, the use of the Dies Irae is explicitly religious. But in the eighteen hundreds, it started to lose its religious connotation.

Alex: In 1830, we have a piece of classical music called the Symphonie Fantastique written by a French composer named Hector Berlioz. This symphony is important because in the 19th century, composers started to write what was called programm music, music that was based on something else, music that was telling a story.

*In the story of Symphonie Fantastique, the main character dreams that he has killed his lover, and that she’s been reincarnated as a witch.

Alex: So, the fifth movement, the last movement is called the Dream of a Witch's Sabbath, and the main character's beloved has come back to life as a witch. And the movement is set at midnight in a graveyard. So we've got all of these spooky sounds that the composer is incorporating. We've got bells tolling, [Music clip: Dream of a Witch's Sabbath] 12 bells for midnight. We've got these creepy crawly sounds like skeletons dancing or spiders crawling around. And then on top of that the final layer is we hear low in the orchestra, the tubas and the brass playing the Dies irae [Music clip].

Audiences at the time would have recognized the Dies Irae from hearing it at funerals, the same way we know the Wedding March from hearing it at weddings.

Alex: They would have heard it at funeral mass after funeral mass. And so they would have known that, okay, this is what this composer is trying to create. He's using the Dies irae because it was so richly symbolic of funerals and death.

“Symphonie Fantastique” really marked a turning point for the Dies Irae.

Alex: This is that moment where the Dies irae moves from a sacred connotation. We always thought of it in the church. Now it has jumped out of the sacred into the secular realm.

Rather than setting the words of the Dies Irae to a new melody, Berlioz put the original melody into a new context. As program music became more and more popular, other composers started doing the same thing: using the melody of the Dies Irae to evoke death.

Alex: And so, we see piece after piece that incorporate the Dies irae melody, in other programmatic pieces. And I'll read you some of these names:

<span data-preserve-html-node="true" style=“color:rgb(127,44,202)"> Alex: The Dance of Death

...by Liszt

[Music clip: Liszt “Totentanz”]

Alex: The Isle of the Dead.

...by Rachmaninov

[Music clip: Rachmaninov “Isle of the Dead”]

Alex: Songs and Dances of Death

...by Mussorgsky

[Music clip: Mussorgsky “Trepak”]

In case you missed that one, the Dies Irae was in the piano part. Here it is again:

[Music clip: Rewind effect + “Trepak”]

Alex: These are all 19th century classical compositions, all of which have death in the title and all of which have Dies Irae in their musical materials.

[music in]

Over the span of 500 years, the Dies Irae had transformed from a Latin funeral chant to a kind of musical meme signifying death. But it still had one more transformation to make.

As the nineteenth century came to a close, a new artform called the “motion picture” was beginning to take off. The time had finally come for this little melody to become a moviestar. That’s coming up, with a TON of examples, after the break.

[music out]

MIDROLL[music in]

For hundreds of years, the Dies Irae was performed at Catholic funerals. Eventually though, composers began using it as kind of a musical shorthand for death.

Kirk: Composers have just equated that sound with the sound of death for hundreds and hundreds of years. It's never shaken that association.

[music out]

In the late eighteen hundreds, it became possible to make narrative stories out of so-called “moving pictures,” or films.

Alex: At the end of the 19th century is when we start to have silent films. So 1890s or so.

When you saw a movie in the silent film era, there was usually a pianist in the theater, improvising music to what was happening on screen.

[SFX: Dramatic silent film music]

Alex: And so these mostly pianists would have been pulling little snippets of music out and playing them for what were appropriate scenes in a film.

We don’t have any recordings of film music from this era, but there’s a good chance that these pianists brought in the Dies Irae when a scene called for it.

Alex: I think it's pretty safe to assume that any sort of funeral scene or scary scene or scene about death probably would have had a Dies irae sort of reference in it. [SFX: Music leads into Dies Irae melody]

As the film industry grew and budgets increased, filmmakers started hiring composers to write specific music for their movies. In bigger theaters, these scores would have been performed live by a small orchestra. And as far as we can tell, the first time the Dies Irae was referenced in a film score was in 1927, in the silent sci fi classic, Metropolis.

[Music clip: Metropolis]

[music in]

Of course, movies didn’t stay silent for long. By 1930, movies with sound, called “Talkies” were all the rage,1 and the age of the soundtrack had begun. Once movies got synced sound, film composers started using the Dies Irae in the same way that classical composers had: to conjure images of death and despair.

In the 1946 movie It’s a Wonderful Life, George Bailey almost takes his own life, but an angel named Clarence gives him a terrifying glimpse of what the world would be like without him.

[music out]

[SFX: Wonderful Life - George Bailey: Clarence! Help me, Clarence! Get me back!]

Did you catch the Dies Irae in the music? In this case, they actually used the first seven notes. Here it is without the dialogue.

[Music clip: "Wrong Mary Hatch / The Prayer"]

But for about twenty five years after It’s a Wonderful Life, there are very few examples of the Dies Irae in movies. This is because orchestral soundtracks became less popular.

Alex: As we move into the 40s and 50s, Hollywood moves towards film noir, sort of after World War II, and it sort of mirrors what's going on in popular music. So you start to see more jazz influences in film scores, less big orchestral stuff and more smaller intimate ensembles.

For example, here’s a clip from the soundtrack of the 50s film noir, Touch of Evil.

[Music clip: “The Boss”]

In the 60s, filmmakers started to put more radio hits into soundtracks. The movie Easy Rider featured bands like Jefferson Airplane, The Byrds and The Who.

[Music clip: The Who - “I Can See For Miles”]

But in the 70s, classically-inspired soundtracks made a comeback, and so did our favorite funeral tune. Here’s a clip from the opening of A Clockwork Orange. Again, in this example, they used the first seven notes of the Dies Irae.

[Music clip: A Clockwork Orange]

Alex: Once we get into the 1970s, that's where this idea of the Wilhelm Scream Dies Irae really takes off.

The Wilhelm Scream is a classic movie sound effect that’s been reused over and over, just like the Dies Irae [SFX clip: Charge at Feather River Wilhelm scream].

It was George Lucas and Ben Burt who popularized the Wilhelm Scream by putting it in Star Wars [SFX clip: Star Wars Wilhelm Scream]. It was John Williams, the composer of Star Wars, who helped popularize the Dies Irae by using it in the Burning Homestead scene [Music clip: Star Wars Dies Irae]. After the massive success of Star Wars, the dramatic orchestral soundtrack became cool again.

Alex: John Williams has this sort of interest in that old school style. So there is a renaissance in the sort of traditional symphonic style.

After Star Wars, the Dies Irae really took off. It would take way too long to play them all, so we’ll stick to the highlights. Here it is in Close Encounters of the Third Kind, as the townspeople watch the strange alien lights in the sky...

[Music clip: Close Encounters]

It plays in The Lion King, just after Mufasa dies, when Scar sends the hyenas chasing after Simba.

[Music clip: The Lion King - “Scar: Kill him”.]

You can hear it in Poltergeist, just after the mom crawls out of the muddy skeleton pit.

[Music clip: Poltergeist]

You can hear it in the opening credits of Mars Attacks, as the martian ships surround Earth.

[Music clip: Mars Attacks]

Here it is in Game of Thrones:

[Music clip: GoT “The Wars to Come”]

In Frozen 2, Elsa hears this mysterious melody throughout the movie.

[Music clip: Frozen 2 montage]

Kirk: “Into the Unknown” from Frozen II has this beautiful ah ah ah ah. It's like the recurring motif of the whole movie is sung as Elsa hears the sound over and over again. That's the Dies Irae. The composers of that piece said, specifically, "We went to this Latin death chant."

[SFX: Frozen 2, Elsa - “You hear it too”]

Here’s one of the composers, Bobby Lopez, in an interview with the AV Club.

[SFX clip: Bobby Lopez: The actual tune is the Dies Irae, which is a musical reference that goes way back to early times in the church, and it’s like the Day of Wrath. It’s all about death. It’s like a musical signal for death.]

Alex: And then probably the most iconic is The Shining.

Kirk: Yeah, The Shining. The movie starts with the Dies Irae. It's very clear about it.

Alex: The opening sequence, that sort of floating camera going up the mountain side...

[Music clip: The Shining]

Alex: As the Torrance family drives up to the Overlook Hotel is scored by a full statement of the Dies irae. So not just the first four notes, but a full statement of the Latin chant.

Home Alone uses the Dies Irae as a kind of theme song for the creepy old man that lives near the McCallisters.

Alex: Every time Kevin McCallister sees his neighbor outside, they play these four notes, bah bah bah bah because the cousins tell this story of the neighbor and accuse him of being the Salt Slayer.

[SFX clip: Home Alone - Buzz: “The salt turns the bodies into mummies.”]

Alex: And so, every time Kevin sees this neighbor, John Williams gives us those four notes [Music clip: Home Alone] as a musical stinger to sort of underline the point that he is supposed to be scary.

Later in the movie, Kevin meets the old man in a church, and learns he isn’t the bad guy, after all. In the background, we hear a children’s choir singing Carol of the Bells.

[Music clip: Home Alone, Carol of the Bells]

Kirk: Then, he realizes that he needs to get home and protect his house. It's the most dramatic part of the movie. As he runs home, this John Williams' version of the Carol of the Bells begins to play that's really dramatic, [Music clip: Home Alone, Carol of the Bells]. But that's also cited as quoting the Dies Irae. The two melodies are using the same four notes.

Kirk: The Dies Irae is these four notes [SFX: Melodica]. The only difference between the Dies Irae and the Carol of the Bells is rhythm and tempo. You play the Dies Irae, typically, very slowly. If you play them a lot faster in a different rhythm, it sounds like [SFX: Melodica]. It's the same four notes. They're just played in a different rhythm, and kind of with a different energy.

The version of Carol the Bells that we’re familiar with, the one in Home Alone, was arranged around 1914.2 But the melody is based on a traditional Ukrainian folk song that was written before Christianity came to Ukraine. As far as anyone can tell, there’s no direct connection between the two pieces.

Kirk: It's actually a rite of spring song. It's a song about the coming of spring and how you will have a bountiful harvest this year, which is kind of the opposite of the Dies Irae. It's not about death at all.

Besides Home Alone, there’s another movie that seems to play with the similarities between these two melodies. This was actually the movie that made Kirk aware of the Dies Irae.

Kirk: It was when I was learning about The Nightmare Before Christmas soundtrack. It wound up causing this sort of crossed musical wire in my brain, where I heard The Nightmare Before Christmas soundtrack is being based on one thing when, in fact, I think there's a pretty strong argument that it's based on another.

In The Nightmare Before Christmas, the spooky citizens of Halloweentown decide to take over Christmas from Santa Claus. As they prepare their twisted version of the holiday, they sing a song called “Making Christmas.” [Music Clip: Nightmare Before Christmas - Making Christmas] The entire song is built around the melody of the Dies Irae or is it Carol of the Bells?

[Music Clip: Nightmare Before Christmas - Making Christmas]

Kirk: To me, because it's about someone stealing Christmas, I always equated it with Carol of the Bells, [SFX: Singing “bum ba ba bum”]. But, because Jack is from Halloweentown, and is a skull and is a dead guy, it actually makes more sense that it would be the Dies Irae.

[Music Clip: Nightmare Before Christmas - Making Christmas Big Chorus]

So are all of these composers doing this intentionally?

[music in: The Nocturne - “Trois Gnossiennes 3”]

Kirk: I think about that a lot. I think some composers certainly do, but I do think there are composers who will use those four notes just thinking, "That just sounds dark and cool. It goes downward and it resolves really well," and they won't do it consciously. I think that other composers will.

Alex: I think it's definitely a full blown trope at this point. And I think that composers have to be knowingly using it at this point.

Alex: My next question is, how much do you think these composers are listening to each other's cues? Composers will often use temp tracks of other films. And so, they will put in a cue from a different film just as a temporary score, and then the director will say, "Oh, I like that, can you write me something like that?" And so now we've got a “snake eating its tail” situation where other composers are trying to replicate what's already been there.

[music out]

For the many composers who do use it intentionally, the deep history of the Dies Irae is a huge part of its appeal.

Alex: I think it's a vital tool in the tool belt of these composers. I think that having this connotation built in with the Dies Irae, it's like a footnote that adds another layer of meaning into what the film is trying to do.

[Music clip: Dies Irae]

Kirk: I think that we, humans, naturally congregate around musical ideas. Music has been part of our culture for as long as human beings have existed right? You sing music for different rituals and have musical associations with different parts of life. Death is such a huge part of life, that there would be some kind of music that we would associate with death.

For better or for worse, death isn’t going anywhere, so more than likely, neither is the Dies Irae.

Kirk: When there's music that's used and associated very strongly with huge events like death or marriage or birth they, then, become the memetic, in this way, over generations.

Kirk: These little musical bits of information make their way into our culture and kind of, then, transcend their origins and just become things that get echoed over and over and over again.

[music out]

[music in: Tokyo Rose - “The Wraith”]

Twenty Thousand Hertz is hosted by me, Dallas Taylor, and produced out of the sound design studios of Defacto Sound. Hear more at defactosound.com.

This episode was written and produced by Casey Emmerling, and me Dallas Taylor, with help from Sam Schneble. It was sound edited by Soren Begin. It was sound designed and mixed by Nick Spradlin.

Thanks to our guests, Alex Ludwig and Kirk Hamilton. To see Alex’s list of Dies Irae examples, visit alexludwig.net. To hear Kirk’s awesome podcast about music, subscribe to Strong Songs right here in your podcast player.

Special thanks to Marek Klein of the Graduale Project for letting us use his rendition of the Dies Irae chant. Check out their Youtube channel to hear other Latin chants Marek has performed. You can find all of these links, along with artwork, music and more, at our website, 20K dot org.

Thanks for listening.

[music out]

Recent Episodes

Perfect Pitch: Unlocking Jacob Collier’s musical brilliance

Art by Mafalda Maia.

Art by Mafalda Maia.

This episode was written and produced by Olivia Rosenman.

People with perfect or "absolute" pitch hear every single sound as precise musical notes. Is this extraordinary talent a blessing or a curse? In this episode, we dive into the neuroscience, pluses and pitfalls of absolute pitch. Featuring neuroscientist Daniel Levitin and Grammy-winning musician Jacob Collier.


MUSIC FEATURED IN THIS EPISODE

Hide and Seek by Jacob Collier
Light It Up On Me by Jacob Collier
Down the Line by Jacob Collier
To Sleep by Jacob Collier
All I Need by Jacob Collier
Bakumbe by Jacob Collier
Hideaway by Jacob Collier
Colrain by Marble Run
Sky Above by Jacob Collier
Moon River by Jacob Collier
A Noite by Jacob Collier
Connect by Steven Gutheinz
Count the People by Jacob Collier

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Discover more at lexus.com/curiosity.

Check out Barnaby Martin’s YouTube channel, Listening In.

Order Daniel Levitin’s book, This Is Your Brain On Music.

View Transcript ▶︎

[music in: Jacob Collier - Hide and Seek]

You’re listening to Twenty Thousand Hertz.

What you’re hearing right now is one person’s voice, layered in perfect harmony with itself using an instrument called a vocal harmonizer. This is a live performance; it has not been edited or altered in any way. The voice itself isn’t being digitally tuned, meaning one wrong note and the harmonic structure will fall apart. The singer is Jacob Collier.

Jacob: So what happens when I play the harmonizer is, I sing a note, and I play a few notes on the keyboard. What you hear coming out the instrument is all of the notes I'm playing on the keyboard sung by my voice. So, it's almost like I can be this spontaneous choir.

Singers have used vocal harmonizers for a long time, but this one is special; it was invented by Jacob, and he’s one of the only people in the world who can make it sound like this.

[music continues: Jacob Collier - Hide and Seek]

Jacob: What I realized about this instrument, when I started to get my friends to try and play it is that, it's really, really difficult to play unless you know what note you're singing. Because if you know what note you're singing, say for example, I sing in E, or if I think, "Oh, I want to play a chord of E major," and I go [SFX: sings] and I play the chord, I don't have to think about trying to find an E, because I know an E. Actually, I really didn't think about that. It was just a thing that, I knew it, so I did it.

The way Jacob can just pull a perfect E out of nowhere is called absolute pitch. It’s the ability to identify a musical note without any external reference.

Jacob: I toured for a while, with this one man show, where I was onstage all on my own, surrounded by about 10 different musical instruments, and walking around them, and playing them, and looping them, and designing these grooves, and all this stuff. Sometimes I would have, say, half a beat to move from the double bass to the harmonizer, or from the guitar to the harmonizer. If I had to check my notes, I'd be lost. I'd be absolutely lost.

[music out]

Absolute pitch is a rare gift that neuroscientists are still trying to understand.

That’s producer Olivia Rosenman.

So, Olivia, I feel like whenever I’ve heard about this, it’s been called perfect pitch. So is there a difference between perfect pitch and absolute pitch?

No. Both absolute pitch and perfect pitch refer to the same thing. But, most neuroscientists and musicians use the term absolute pitch.

[music in: Jacob Collier - Light It Up On Me]

The way our brains engage with music is an entire branch of neuroscience in itself.

Daniel: I’ve done most of my work on absolute pitch, but also done some studies of how musical structure is represented in the brain and musical emotion and why people like the music they like.

That’s Dr. Daniel Levitin. He’s a neuroscientist who researches the musical brain. In fact, Daniel wrote what is probably one of the most important books on it. It’s called This Is Your Brain on Music. When I spoke to Daniel, he came prepared.

Daniel: I brought a microphone that goes up to 20,000 hertz in honor of this conversation.

Daniel doesn’t have absolute pitch himself, but he does have a background that helps him bridge the gap between music and science.

Daniel: I had dropped out of college after my sophomore year to play in band. I played in a succession of bands, jazz, and country, and rock bands and eventually found my way into recording studios working as a producer and an engineer.

Over the years, Daniel worked with artists like Joni Mitchell, Neil Young, and the Grateful Dead. He says that in his experience, absolute pitch isn't always a desirable thing for a musician to have.

[music out]

Daniel: Absolute pitch is mistakenly regarded as a pinnacle of musicianship. There are plenty of composers who didn't have it, and there are composers who did. There are great musicians who didn't have it and great musicians who do.

Paul Macartney doesn’t have it [SFX: The Beatles - Let It Be]. Neither does Beyonce [SFX: Beyonce - Sorry]. Tchaikovsky didn’t have it [SFX: Tchaikosky - Swan Lake] and neither did Brahms [SFX: Brahms - Piano Quartet No.1 in G minor, Op.25].

Daniel: The lack of absolute pitch is not an impediment. The more musicians I talk to about it, and the more I meet with and without it, the more it becomes clear to me that it's actually not such a great thing to have it.

Most musicians rely on something called relative pitch.

Daniel: It’s the ability to identify by name a musical interval. So, if I go, “Bampam,” that's a perfect fourth. If you're a musician, you know that...

If you’re not a musician, a perfect fourth is a series of two notes, four steps apart on the 12 note scale.

Daniel: Almost all musicians that I know of have relative pitch. Now, if I go, "Bampam," if you have relative pitch, you'd say that's a perfect fourth. If you have absolute pitch, you'd name those two notes. I don't have it, so I don't know what they are. I don't know what notes I sang, but it's a perfect fourth.

I don’t have it either, so I need a piano to check what notes Daniel is singing. For the record, it’s a B and an E [SFX: bum bum].

Branford Marsalis is a saxophone player who’s played on heaps of famous records. Like most musicians, Branford has relative pitch, not absolute pitch. In the studio version of the 1985 Sting song “Shadows in the Rain” you can actually hear Branford panicking right at the beginning because he doesn’t know what key the song is in. 
 [music in - Shadows in the Rain by Sting]

He doesn’t get an answer, but somehow he manages to come in, perfectly in tune.

[music up]

So how did he pull it off? Well lucky for us, Daniel and Branford move in similar circles.

[music out]

Daniel: So, I asked him just a couple of weeks ago, "Do you have absolute pitch? Is that how you figured out the key?”He said, "No, no, no. I waited till Sting started singing, and then I quietly played the saxophone to find what note he was singing, and then I knew what key we were in."

While you can teach yourself relative pitch as an adult, Absolute Pitch is a gift that some people have, and some people don’t. So why is that?

[music in: Jacob Collier - “Down the Line”]

Daniel: Well, as near as we can tell, there's a difference in the way kids are raised. A two-year-old who's learning words, that two-year-old's mother is likely to say, "Oh, look. That's a fire engine. See the fire engine? It's red. See this ball? It's blue." The child is rewarded for making these name associations with a visual perception...

When you see a tomato you instantly know it’s red. You don’t have to go and hold it up against a picture of the rainbow so that you can give it the right color label.

Daniel: Color naming, for most of us, unless you're colorblind, it's automatic. It's an internal template.

And for most of us, musical note labels aren’t something we’re drilled on as kids.

[music out]

Daniel: It's rarely the parent who says, "Did you hear that alarm bell? [SFX: alarm tuned to E flat, turns into piano plunks] That was an E flat, E flat. You hear my voice right now when I hold my voice at a steady pitch? That's a D." A musical family with absolute pitch might do that, but no one else is.

[music in: Beethoven - Moonlight Sonata]

But even a musical family needs to keep their piano tuned.

Daniel: Sometimes people are raised in a household with a piano that's half a step off or a full step off [SFX: Detune Moonlight Sonata, continue detuning until music end]. That happens with old pianos.

[music out]

Daniel: So, you've got somebody who learned that the note that we call C [C on piano] sounds like this [B flat], but to everybody else, they would say, "Well, that's not a C. That's a B flat."...

So even though they’re out of tune, they still have absolute pitch.

Daniel: So, when you test people like this, they're perfectly consistent. They name it with great precision and, as I say, replicability, consistency. They're just off by two semitones.

But learning absolute pitch as a small child doesn’t completely explain the phenomenon, because there are people out there who have absolute pitch, who haven’t had any musical training in their lives. This is something Daniel was able to prove in a study with tuning forks which he labelled with names rather than musical notes.

Daniel: I just said, "This one is Fred. This one is Ethel. This one is Lucy. This one is Ricky, and just play this to yourself for a week, and come back to me in a week, and I'm going to take it away from you, and then I'm going to ask you to sing that pitch."

That’s how he found people with absolute pitch who didn’t have any musical knowledge at all.

Daniel: People were able to do that. They could sing a Fred or an Ethel, and if you can do that, that's the same to me as singing an A flat or a G.

So it’s not about the labels.

Daniel: If you speak a different language, you're not going to talk about A, B, and C, and some people call it do, re, mi using so-called solfège.

[SFX: DO, RE, MI, FA, SOL, LA, TI, DO]

And even creatures that don’t have any word labels at all can demonstrate Absolute Pitch. Some animals use it to keep track of their pack.

[music in: Jacob Collier - “To Sleep”]

Daniel: Wolves in a pack find a pitch range that is theirs alone, and it's how they identify one another in the pack. So, their howls are not random. They're confined to a particular pitch range, and each wolf has a range that indicates who they are. So, they may not have labels like A flat, but one wolf will say to the other, "Well, that's Harold," [SFX: Howl] or at least know in his or her mind that that's Harold howling, and that's Sophie howling over there [SFX: Howl].

[music out]

And while a human who has Absolute Pitch will also have relative pitch, in the animal kingdom, that’s not necessarily true.

[SFX: Songbird]

Daniel: For most species of birds, songbirds, if you take their song and you transpose it into another key, which is the essence of relative pitch, it's what makes Beethoven fifth work. You hear the pattern “ba, ba, ba, bum, ba, ba, ba, bum” [SFX: Beethoven’s fifth] transposed, you recognize it as the second four notes as related to the first by transposition. You may not know it's called transposition, but you recognize there's something about them that's the same. You do that to a songbird, [SFX] they don't recognize… their own song, which suggests that they've got absolute pitch but not relative.

[music in: Jacob Collier - All I Need]

People with absolute pitch hear every single sound as precise notes on the chromatic scale. For a songwriter, knowing the exact pitches of all the sounds around you has a huge impact on the way you write music. For Jacob, that talent allows him to pull off some pretty unreal harmonic feats. That’s coming up, after the break.

[music out]

MIDROLL

[music in]

Absolute pitch is an extraordinary talent that very few people have. Scientists still haven’t figured out exactly why some people develop it, and why some people don’t. If you have absolute pitch, you instinctively know the exact musical note of every sound you hear.

[music out]

So what’s it like when the whir of a coffee grinder [SFX], clanking silverware [SFX] and creaky doors [SFX] all sound like musical notes?

Jacob: It's fun and can be rather distracting.

[music in: Jacob Collier “Bakumbe”]

Jacob: I think I've realized that it's a lens through which I can view things, which is sometimes useful and sometimes not useful. I don't have to always listen to that voice. I've almost taught myself to walk around the world and listen to sounds as though it's a blanket of sound, rather than these precise things.

Jacob: But were I to want to, I can tune into all sorts of weird and bizarre household objects, and roadside conglomerate sounds, and stuff like that.

[music out]

It might not be very helpful to know that your microwave beeps in B flat [SFX]. But even if absolute pitch isn’t useful all the time, it does come in handy when you’re composing music with heaps of instruments.

Jacob: I play a few different instruments. I'm a singer. I play the piano, and the drums, and the bass, and the guitar, and various percussion instruments, and other instruments that I've invented, and a few other things.

[music in: Jacob Collier - "Hideaway"]

If you watch any of Jacob’s music videos on YouTube, you’ll see that he often plays every single instrument in the song.

[music up]

In this track, “Hideaway,” Jacob plays two ukuleles, two acoustic guitars, a bass guitar, an autoharp, a khalimba, a ghatam, a mandoline, an Udu, and an Azerbaijani tar.

[music up]

Oh and that’s him singing, too.

[music up]

Jacob laid down each track individually and then looped and edited them together into an extremely impressive one man band.

[music stop]

Jacob comes from a musical family. His mother is a violinist and conductor, so his gift of absolute pitch was identified early on.

[music in: Colrain by Marble Run (Blue Dot)]

Jacob: Well, I was about two years old, I suppose. I would listen to all sorts of strange sounds around the house; be it the microwave beep [SFX], the vacuum cleaner whirring [SFX], or the car alarm going off outside [SFX], whatever.

As Daniel explained, some children develop absolute pitch by being constantly told the notes of the sounds around them. But Jacob's family took a different approach. Instead of naming notes, they focused on how the notes made him feel.

Jacob: It would always be not, what is that note, but what does that note feel like. I supposed I was encouraged to navigate all these different sounds and pitches, by way of feeling them, rather than analyzing them.

[music out]

For example, Jacob says that hearing his mother play the violin influenced the way he feels about certain notes.

Jacob: The interesting thing about that process of learning for me was that I'd be more likely to confuse, for example, an A with a D, than an A with a G sharp. Because the A and the D had this open string feeling.

[SFX: Violin SFX on notes]

An “open string” is when a note is played on a stringed instrument, like a violin or guitar, without putting any fingers on the neck. On a violin, the strings are traditionally tuned to G, A, D and E [SFX]. For Jacob, these open string notes all share a certain feeling.

Jacob: The A, and E, and D [Violin SFX on notes] feeling being quite open, and quite rounded, and quite neutral, and on the brighter side felt much more different to something like G flat, D flat, A flat [Violin SFX on notes], because those, I suppose, feel like they're on the darker side, on the flatter side of the notes.

In this way, Jacob’s parents taught him to trust his feelings when it came to music.

Jacob: I suppose since the age of two or three, I've felt like I could trust that internal compass, and navigate in that way.

And it’s that internal compass that allows Jacob to explore the world of harmony

Jacob: So, I think for me, it was interesting to start recognizing not just frequencies, but relationships between them. That led me on to my absolute fascination and deep crush on harmony in general.

I tested out Jacob’s pitch by hitting the coffee mug on my desk with a pair of scissors.

Olivia: Maybe I'm just going to hit it with these scissors that I have here. Then if you can tell me what it is, that'd be cool.

Jacob: Yeah. [SFX: clink]

Olivia: Do you want me to do it again?

Jacob: Yeah, go on, then. [SFX: clink]

Jacob: [SFX: mug clinking sound] So, it sounds like halfway between a D and a D flat.

Ok, wait… I have to jump in here and check that. Here’s that sound again [SFX: clink]. Now, let’s isolate the resonating frequency [SFX]. To be exact, that’s 2339 Hertz, which is right between a D flat and a D, just like Jacob said. That is unbelievable!

I know! But surprisingly, even though Jacob can identify a note to a fraction of a semitone, that doesn’t mean he always plays or sings in tune.

Jacob: No, no. It doesn't at all.

[music in: Jacob Collier: “Sky Above”]

Jacob: Well, first off all, I'm quite interested in being out of tune. I think that being out of tune is a whole kettle of fish that most people don't dig into. I think it's exciting to do it, in an interesting way. But also, I'm not a brilliant, brilliant singer. I don't think of myself as a particularly spectacular technician on any instrument that I play. I think that singing a note in tune is much more about technical or control than it is about knowledge.

In other words, just because you know exactly what a note should sound like, that doesn’t mean you can always hit it.

Jacob: But what I would say is that when I sing a note out of tune, I can tell you how many hundredths of a semitone the note is out of tune without thinking.

That is, unless he’s sick.

[music out]

Jacob: I have noticed that if I'm very unwell, like for example, I'm suffering from a fever. I specifically have found with fevers that everything goes pear shaped and things can sound up to a semitone sharp, which is extremely bizarre.

But when he’s in good health, Jacob’s ear is so keen that he hears distinct musical differences where most people just hear the same thing.

Jacob: One of the ways I generally think about things is that for every note, there is a middle, and a slightly low, and a slightly high. There are three versions of every note. So, say you take like, [sings D], and [sings D]and [sings D] right? Those are three different kinds of D. They're all closest to D, out of all the notes; [sings D Flat] is way lower.

Where someone without absolute pitch may not hear the differences in those three shades of a D, Jacob hears them as entirely different notes.

Jacob: So, I think that my mind has recently been blown just realizing how many different kinds of notes there are between notes. I think that's kind of spurred me on, the interest I have in discovering all these secret notes between.

So what most people might think of as a note that’s slightly out of tune, for Jacob is a secret note that can be a source of inspiration.

Jacob: I suppose with such notes, you have a few different options. One is that you can correct the note, using some kind of correction tool where you say, "Okay, I'm going to apply the auto tune [SFX], and it's going to make everything quantized, and all semitones exactly the same size.

Jacob: The second option is that you tune it manually, by which I mean that you speed up [SFX] or slow down [SFX] that region of the audio file by a particular number of cents, which is one hundredth of a semitone. The third option is to embrace that it's out of tune, and get jiggy with it, and adapt the other things around the notes to fit.

[music in: Jacob Collier - “Moon River”]

Let’s take a listen to an example of Jacob’s ability to find creative uses for the secret notes in between. This is an arrangement he did of Moon River, by Henry Mancini.

[music continues: Jacob Collier - “Moon River”]

Jacob: In the final verse I modulate it to the key of D half flat major.

Musically, this is a very complex thing to do.

[music out]

Barnaby Martin is the creator of the YouTube channel “Listening In,” where he posts video essays about musical topics. In one of his videos, Barnaby breaks down Jacob’s cover of Moon River.

Barnaby: Jacob’s arrangement of Moon River seems to be pushing every aspect of his work to its limits. It contains microtonality, alternative tuning, 8 different keys or pitch centres as well as some of the most complex and dense harmonic language he's ever used. All done without a piano. All done with only his inner ear.

This next part is going to get a bit technical. Stay with us. Even if you don’t fully grasp all the details, you’ll definitely understand just how complex this arrangement is.

[music in]

The song opens with a dreamy, hummed rendition of the Moon River melody in the key of B flat major.

Then, to get to the first verse, Jacob moves from B flat major to D flat major in a quick sequence of 11 complex chords.

[music up, then down]

After the first verse in D flat major, Jacob transposes up a semitone into D major for verse two. It’s here that Jacob starts to get jiggy with it, sliding into the spaces in between standard notes.

Barnaby: The top line sings notes a major seventh above the tune but uses microtonal inflections to add color to this descending sequence of chords.

[music continues]

If this all sounds like gibberish to you, don’t worry. The takeaway is that once you start using notes outside of the traditional 12 note scale, things get very tricky, very quickly.

[music out]

So what is it about these off-tune or in between notes that Jacob loves so much?

[music in: Jacob Collier: A Noite]

Jacob: I'm more interested in writing songs in non-linear keys, or keys that don't exist on the piano. Not in the name of being strange or dysfunctional harmonically, but more it's like I've trodden the snow. It's like when you have untrodden snow [SFX: Untrodden snow]. Say you start your life with untrodden snow, and you walk along certain pathways of the snow a lot [SFX: Trodden snow]. Then there are these other areas, which aren't as trodden down. So, it feels so much fresher to step on some untrodden snow [SFX: Untrodden snow]. It's a really wonderful feeling.

Jacob: I think it really feels like I'm in this new frontier. Also, I find it really confounding, in a nice way. Creatively, I'm always fed by things that are subtly unfamiliar, or strange, or that I don't already have a system for understanding. I'm always chasing those things, things that I can't quite put in a box.

[music out]

According to Daniel Levitin, Jacob’s not the only musician who likes to walk on untrodden snow.

Daniel: Just to cite one of my favorite groups, Led Zeppelin, ... Jimmy Page would have the band intentionally tune in between the cracks off of standard pitches [SFX: Guitar tuning]. I don't know why he did it, but what it means is that when we listen to Led Zeppelin today against that backdrop of hearing tens of thousands of recordings in a canonical precise pitch, it may sound a little more exotic because it's breaking the rules.

Led Zeppelin also made pitch changes to their music in post production.

[Music Clip: Led Zeppelin - No quarter]

This recording of “No Quarter” was slowed down to drop the pitch a quarter of a tone. According to Jimmy Page, this made it sound “bigger, thicker and more intense”.

If we speed it up a bit, we can get a sense of what the song might have originally sounded like.

[SFX: speed up track to bring it up a quarter tone]

Here’s the final, slowed down version again.

[SFX: morph back to recorded version]

[music out]

When it comes to standard pitches, Jacob Collier sees a pretty good reason to break the rules.

[music in: Connect by Steven Gutheinz]

Jacob: I think what I've come to understand recently is, it's a convenient system, the semitones. There are 12 notes in every octave, and those are all the notes that there are in the world and stuff. But the truth of the matter is, it's totally arbitrary.

Even if he considers the 12 note scale arbitrary, the fact remains that Jacob can conjure that scale from memory in a way that most of us can't. This is what fascinates Daniel

Daniel: When we attend to something in the environment by looking at it or hearing it, tasting it, smelling it, touching it, some of those details get encoded into memory and some of them don't.

There’s something special happening with sound, perception and memory in the brains of people with Absolute Pitch.

Daniel: All this interacts with memory and consciousness and perception in interesting ways. I find absolute pitch to really be this fascinating nexus of all these things.

For Jacob, having absolute pitch has helped him realize that there’s really no difference between sound and music.

Jacob: I, for one, wouldn't claim a birdsong [SFX] to be more inherently musical, per se, than something like a dog [SFX]. But I think that because we, as humans, tend to organize things in certain ways in our minds, that we call something music or not music.

Jacob: I almost feel like the more names you give stuff, the smaller the thing gets. It's like saying, "This fits into this box, or this box, or this box. This is a sound, and this is music." In some ways, we’re one to describe a sound as either, all you'd be doing is imprisoning the value of the sound within something which has value already. Which is to say, make it non-infinite.

So if you’re like most people, and you don’t have absolute pitch, why should you care about it?

Jacob: Well, I don't think it's worth caring about absolute pitch, per se. I think it's worth caring about pitch. I think that, if you care about pitch enough, that absolute pitch becomes interesting. You know what I mean? Because I guess you recognize that there are all these different sounds in the world, all around you, on the radio, and in the music that you listen to, if you listen to music [SFX]. Or as I say, in bird songs [SFX], and animal calls[SFX], and peoples' conversations[SFX], car horns, kettles, hoovers[SFX], all sorts of things. The world is kind of cacophonising, in a musical way all the time.

Jacob: I think that once you realize that the world is music, you can't un-hear it as music. That's a really wonderful, joyous thing to realize.

[music out]

[music in: Jacob Collier - “Count the People”]

Twenty Thousand Hertz is produced out of the studios of Defacto Sound, a sound design team that makes commercials, documentaries, and trailers sound incredible. Hear more at defacto sound dot com. This episode was written and produced by Olivia Rosenman, and me Dallas Taylor, with help from Sam Schneble. It was story edited by Casey Emmerling. It was edited and sound designed by Soren Begin and Colin DeVarney.

A HUGE thank you to Jacob Collier for not only for speaking with us, but also allowing us to use so much of his music. The track you’re hearing right now is from his latest album, Djesse Volume 3. Listen wherever you get your music.

Thanks also to Daniel Levitin, who wrote the book This is Your Brain on Music. That book was a big inspiration for this episode.

And thanks to Barnaby Martin for allowing us to play clips from his YouTube channel, called “Listening In.” You can find all of the individual tracks we used, as well as additional links, and original artwork, on our website, 20K dot org.

Thanks for listening.

[music out]

Recent Episodes

Tyrannosaurus FX: What dinosaurs really sounded like

Art by Jon McCormack.

This episode was written and produced by Leila Battison.

When you imagine the sound of a dinosaur, you probably think of a scene from the Jurassic Park movies. How do sound designers make these extinct creatures sound so believably alive? And what does modern paleontology tell us about what dinosaurs REALLY sounded like? Featuring Jurassic World sound designer Al Nelson, and paleontologist Julia Clarke.


MUSIC FEATURED IN THIS EPISODE

Palms Down by Confectionery
Entwined Oddity by Bitters
Town Market by Onesuch Village
Coulis Coulis by Confectionery
The Poplar Grove by Bitters
Upon the Vine (Instrumental) by Graphite Man
Calisson by Clock Ticking
Beignet by Confectionary
Feisty and Tacky by Calumet
Contrarian by Sketchbook
Can't Stop Lovin You (Instrumental) by Brian Reith

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Discover more at lexus.com/curiosity.

View Transcript ▶︎

You’re listening to Twenty Thousand Hertz.

[music in]

When Jurassic Park first came out, it brought dinosaurs to life in a way that people had never experienced before. Not only did the dinosaurs look incredible, but they also sounded amazing.

[SFX: Velociraptor Call]

Up until that point, there hadn’t been a clear idea in the public consciousness about what dinosaurs sounded like… so the filmmakers essentially had to make it up as they went along. To make these creatures sound believable and alive, they used all kinds of creative sound design techniques. The results are the stuff of sound design legend.

[music out]

Over the nearly thirty year history of the Jurassic Park franchise, the sound designers have faced one persistent challenge.

Al: How do you define the sound of an extinct animal that no one has actually ever heard for real and how do you make it convincing?

That’s Al Nelson, the lead sound designer on the new Jurassic World films.

[music in]

Al works at Skywalker Sound, which is the sound division of Lucasfilm.

Al: Where we're coming from, from the film sound standpoint is, is it believable? But is it also creating emotional context and is it helping the story?

Early in his career, Al worked with sound design superstars like Ben Burtt, the sound designer of Star Wars, as well as Gary Rydstrom and Chris Boyes, who made the sounds for the original Jurassic Park movies.

Al: That's where I got my real education and experience. There's this wonderful legacy at Skywalker of these brilliant icons with Ben and Gary.

Al: Then, there's this second generation of people like Chris Boyes and a handful of others. And then I was part of the third generation.

[music out]

Working at Skywalker Sound, Al got to see firsthand how some of the classic Jurassic Park sounds were made. To Al, there’s one sound that stands above all the rest.

[SFX: T-Rex Roar]

Al: The T-Rex is, in my opinion, one of the most iconic sounds in film sound history.

Now, it would be great if you could just go record a T-Rex in the wild, but of course, that’s not possible. So the original sound designers used a classic technique, and looked for inspiration elsewhere in the animal kingdom.

[music in]

Al: In those early days, they were out recording lots and lots of animals. Here, at one of the local theme parks, they had some elephants. And Chris Boyes was sent off to go record these elephants and out comes, trotting out a little baby elephant and it lets out one baby elephant roar [SFX: Baby elephant roar] and that was it.

Al: That iconic bellow, is the main ingredient of the T-Rex.

But creating that spine-tingling roar required other ingredients, too.

Al: He used some crocodile sounds [SFX: Crocodile rumble], lions [SFX: Lion SFX], the blow hole from a whale [SFX: Blow hole SFX].

Al: He knew just how to use that baby elephant [SFX] and just how to apply the deep grumble of the crocodile [SFX]or that raspy lion sound [SFX] at the beginning of it. And then, the blow hole sound [SFX] underneath the bellow to give it that weight.

Once these elements came together, [music fade out] the result was THE iconic dinosaur sound.

[T-Rex Roar SFX]

Of course, there are plenty of other dinosaurs in the Jurassic Park films, and all of them have their own unique voice. To make them, the sound team used a similar approach.

Al: You find these best little snippets of vocalizations, the sounds that have personality, and you figure out how to layer them.

For example, there was the gentle brachiosaurs—that’s the one with the long neck for eating leaves out of tall trees. To create their sound, the designers took hawing donkeys [SFX: Donkey SFX], and they slowed it down until it became a graceful song [SFX: Donkeys shift into brachiosaurus song].

There’s also a scene where the kids are caught in a stampede of dinosaurs that look sort of like ostriches. These are called gallimimus. To make their screeches, they used the frightening sound of a horse in heat [horse in heat].

That doesn’t sound quite like a gallimimus yet, but if we cut it up [SFX], bring up the pitch [SFX], then pan these sounds to the left and right [SFX], we can get pretty close to what we hear in the movie.

[SFX: gallimimus screech],

There’s also the predatory velociraptors [SFX: velociraptor SFX]. Their hunting calls were made using the sounds of dolphins [SFX: dolphin] as well as mating tortoises [SFX: mating tortoise].

These sounds are so convincing because the designers followed a few basic principles, based on the size of the creature, and its personality.

Al: We want to believe that those sounds belong in the body of that animal. Is it a Brachiosaur or is it a Compy?

For comparison, Brachiosaurs could grow up to 40 feet tall.

Al: If it's a big majestic dinosaur like the Brachiosaur you give it scale but you give it song.

[SFX Clip: Jurassic Park: They're singing]

While the compsognathus, or compy for short, was about the size of a chicken.

Al: With a Compy, it's a cute little dinosaur and then it gets aggressive. But in this case, you're choosing higher pitch sounds, bird sounds [SFX: birds] and even taking pitched up lions [SFX: lion] so that you don't have any more of that big weighted growl. You just have these unique squeaks and squeals.

[SFX: Compy squeaks]

In this way, the sound designers created an entire ecosystem of dinosaur sounds for the original movies. But when it was time to revisit these creatures in the new Jurassic World films, Al faced a unique challenge.

Al: We didn't want to break anything or modify anything that had already been heard and had already been created by Gary.

Al: I mean, you wouldn't ever want to mess with the T-Rex.

But there was a new creature in Jurassic World, a genetically-engineered mutant with an awesome name: the Indominus Rex.

Al: This was a dinosaur that was erratic and kind of broken. More of a screamer and more just unhinged.

To design this new sound, Al and his team went back to the drawing board.

[music in]

Al: Without any real idea of what specific animals I wanted to use, I and my team just started recording lots of new animals and animals that we knew hadn't been recorded previously.

Al: In particular, there was this little fennec fox, it's a desert fox.

Al: It just screamed and wailed and said everything it had to say at high pitches [SFX: Fennec fox screams]. So that was one of the ingredients. One of the reasons it was so useful is because it had that erratic screamy unhinged sound to it.

But just like the T-Rex, there were many layers to the Indominus sound...

Al: For some of the scale of the Indominus, we used very large sows, these huge pigs. At feeding time, the pigs get very aggressive and they bark and squeal [SFX: Sows squealing]. But they also growl at each other [SFX: boar growls] and they just, they sound like big mean animals.

Al: We had a howler monkey which was madly in love with his animal trainer. And when she sang to him, he would just go off into these long vocalizations as, “rar rar rar”. It was brilliant [SFX: Howler monkey song].

[music out]

When you mix the sounds of these animals just right, you get the Indominous Rex.

[SFX: Indominus roar]

Of course, Jurassic World also has dinosaurs from the original movies, like the velociraptors.

[SFX Clip: Jurassic Park: “Clever girl”]

But in the newer films, they got an update.

Al: We now get to experience the Raptors as not as passive but somewhat more trained and they work with their human handlers. And so they needed a new palette of gentler sounds and friendlier sounds.

Al and his team went in search of a new voice for these friendlier raptors.

Al: Ultimately, what we ended up using were mostly penguins, gentoo penguins [SFX: Gentoo penguins]. They do this shuttering and these softer cuter sounds that we were able to manipulate and make the Raptors more friendly [SFX: Friendly raptor noises].

But it wasn’t all elephants and penguins. Some dinosaur sounds were made with less-exotic animals. Gary Rydstrom, the original sound designer, snuck in a sample of his own.

[music in]

Al: One of Gary's traditions is to use his dogs for his animals.

Gary watched his Jack Russell terrier playing with a rope toy, and saw the similarity of the T-Rex grabbing and shaking other dinosaurs, and lawyers, to death. So he was inspired to use these sounds for the T-Rex. [SFX: Jack Russel Growls morph into T-rex shaking sounds]

For the newer films, Al decided to record his own dog.

Al: This was my opportunity to bring my black Labrador Bahama into the Jurassic sound palette. She does these sort of cute growls

[SFX: lab growl]

Al: They're not quite angry but they sound like, don't get too comfortable with me. So, whenever Owen would interact with the Raptors and they needed to check him, sometimes that would be one of the sounds we would use, is that cute low growl.

[SFX: Raptor growls]

[music out]

[music in]

It’s been almost thirty years since Jurassic Park was first released, and in that time, there’s been a lot of new developments in the field of paleontology. New research can tell us a lot about the what the Jurassic would have actually sounded like, and how that compares to what we hear in the movies. That’s coming up, after this.

[music out]

MIDROLL

[music in]

The dinosaurs of the Jurassic Park films aren’t just believable—They have personality and real emotion. One of the main reasons for this is the incredible sound design. But as with anything in Hollywood, sometimes scientific accuracy has to take a backseat to entertainment.

In the years since the original Jurassic Park, we’ve learned a lot about what dinosaurs probably looked and sounded like. To fill in the gaps between the movies and the real Jurassic world, we need a paleontologist.

[music out]

Julia: My name is Julia Clark and I'm a professor of paleontology at the University of Texas at Austin.

Julia: I have to say, the original Jurassic Park movie, it represented cutting edge science at the time, and that was exciting to see.

But Julia says certain parts of the movies just aren’t very realistic.

Julia: Oftentimes these dinosaurs in the movies are making really scary sounds as they're chasing prey items or children [SFX].

Julia: That's not a context in which most predators produce sound.

Imagine you’re a hungry tiger, creeping through the savanna.

[SFX: Savanna sfx, leaves rustling]

You spot your favourite snack, a lone antelope [SFX: Antelope], but it’s on the lookout, ears twitching for any sound that’s out of place. If you let out a blood-curdling roar [SFX: Roar], all you’re going to do is scare away your supper [SFX: Antelope running away]. There’s also a physiological reason why roaring and eating just don’t go together.

[music in]

Julia: if you were about to eat a cheeseburger, would you yell extremely loudly and then stuff the cheeseburger in your mouth? No, because those sounds that you're making are made on the exhale and now you've exhaled all the air out of your lungs and eaten a giant cheeseburger which is going to inhibit your ability to breathe in a second inhale potentially.

If you’re still not sure about this, give it a try at your next meal.

[SFX: Restaurant ambience, man yells, bite, gasp for air (meant to be ridiculous and all happens pretty quick]

Julia: I mean, when that T-Rex is chasing the small children, clearly the children are not a threat. It's not an aggressive display.

Julia: They would just be silent and about to eat them because you don't want to fully exhale with a loud sound [SFX: T-Rex Roar] and then have a giant bite of a child. [SFX: Bite]. It doesn't work.

But of course, it probably wouldn’t be very exciting if the T-Rex in Jurassic Park spent the whole movie creeping around silently.

[music out]

Al: My guess is that, if you're the T-Rex, you don't have to roar in reality. You just come up behind your prey and chomp down and that's that [SFX: sneak & chomp]. But we're watching these films and we want to inspire the kind of personality that the T-Rex has. It's scary. It's an aggressor, it's a carnivore.

And the sound design team wanted to give you that visceral experience right from the start. In Jurassic Park, the first glimpse of the T-Rex we get is when they pull up to its enclosure in the rain. They had tied up a goat to entice it, but now that goat is missing.

Al: It's completely quiet. There's nothing happening. There's no music, there's just a little bit of rain [SFX: rain]. And then you see the water start to ripple and you hear the distant thud. And you hear, "Where's the goat?" And if you just heard those thuds [SFX] and then the grumble [SFX] and that's it...there's something missing there. That dinosaur needs to present itself as dangerous. And so it slams its foot down [SFX], opens its mouth and gives out that iconic, blood curdling bellow [SFX]. So, that's cinema.

[music in]

So carnivorous dinosaurs probably didn’t roar while they hunted. But predators do roar to scare off threats and competitors. So when dinosaurs did make noise, what did they sound like? Julia studies dinosaurs’ closest living relatives—the reptiles—to figure out the kinds of sounds they could have made.

Julia: So reptiles include lizards and snakes, turtles, crocs and birds. Now a lot of people think, "What? You're putting birds in reptiles?" But if we think about a tree of life, that is the only sensical solution is that birds are really highly modified reptiles and they're really highly modified dinosaurs 9, 10

The period that came after the Jurassic is called the Cretaceous period. During that time, a new kind of feathered, flying dinosaur appeared. These were one of the only kinds of dinosaurs that survived the great extinction—when more familiar ones, like T. Rex, were wiped out.

Julia: All the birds that we have today, that's about 10,000 living species, they represent the descendants of one lineage of dinosaurs.13

[music out]

Julia’s research team noticed that modern birds and reptiles share a common vocal behaviour, called a closed-mouth vocalisation.

Julia: It sort of shapes the sound typically after it's produced in the vocal cords. A sound like, "Hmm, hmm," right, is a closed mouth sound.

Our own closed-mouth noises might be limited to a hum, but in other animals, with other body shapes, they can be really impressive.

[music in]

Julia: Crocodilians can make very loud sounds. Actually in crocodilians, some of the sounds that to us would sound most like a roar...they're kind of a rumble [SFX: Croc rumble]. Those rumbles are made with the mouth closed.

The birds we have now are sometimes called living dinosaurs, and they take closed-mouth vocalisations a step further.

Julia: Male ostriches have this boom call [SFX: Ostrich boom call] in which the mouth is closed and that's a very low frequency call [SFX: more ostrich boom calls].

Other birds make noises in a similar way. For instance, there’s the “coo” of a dove [SFX: dove cooing].

There’s the strange scooching sound of a bittern [SFX: bittern call].

And the weirdly-human call of the eider duck [SFX: eider duck call]

Julia: So what we think is that maybe some dinosaurs, maybe larger bodied dinosaurs, maybe they're using these closed mouth vocal behaviors ... like booming calls that they make to attract a mate or defend a territory.

Due to the sheer size of the largest dinosaurs, their booms would be much lower in pitch than even the largest birds. In fact, The sounds they made could have been so low that they’d be almost impossible for us humans to hear.

Julia: If we were around when T-Rex was around...we might feel these sounds of the largest dinosaurs more than we would hear them through our ears.

[music out]

These low rumbles weren’t the only type of sound you’d hear in the Jurassic. Some dinosaurs, known as Parasauralophus, had long skulls with tube-like holes, called vacuities inside them.

Julia: These vacuities don't produce sound, but they would shape sound.

Sound would bounce around inside these tubes, resonating and echoing almost like a didgeridoo.

[SFX: Low didgeridoo note]

In Jurassic Park 3, the resonating skull actually becomes a plot point.

Julia: they're trying to 3D print the vocal organ of a velociraptor and I guess the idea is that if they can communicate with these velociraptors they can influence their behavior.

[Jurassic Park 3 Clip: I give you the resonating chamber of a Velociraptor. Listen to this. [screech].]

Julia: I think it's really cool that at least in the Jurassic Park movies they were trying this out. That said, everything else about the science of that scene is kind of wrong.

Julia: So what they print they call a resonating chamber and a resonating chamber doesn't have the capacity to make sound. It would be something that shapes sound after it was produced.

Blowing into a resonating chamber without using your vocal cords wouldn't make any sound, just like blowing air down a didgeridoo would sound like this. [SFX: Hollow blowing sound]

In other words, the only way to make a real velociraptor sound is with a whole, living velociraptor.

If we put all of this together, we can start to get a more complete picture of the Jurassic soundscape. Julia thinks you may even be able to guess which geological period you were in, just by the sounds that you heard.

Julia: I think there's a lot of evolution of the sonic landscape throughout the age of extinct dinosaurs that we would hear. In earlier parts of dinosaur history where you have a lot of dinosaurs that are like, pony to horse size and bigger… those are going to be lower frequency sounds.

So the Jurassic period would have been a place of deep, bassy rumbles [SFX: Jurassic low-frequency soundscape]

Julia: It's only in the late Jurassic that we have evidence for things starting to take flight and smaller body sizes. By the time you get to the Cretaceous I think there's going to be a lot more higher frequency dinosaur sounds made by these smaller species.

Julia: It's still going to be a fairly foreign sonic landscape but there's still going to be some sounds that are almost bird like [SFX: Cretaceous soundscape under Julia]. It would be fascinating to be a dinosaur watcher in the Cretaceous.

[music in]

Figuring out the way the world sounded a hundred million years ago is hard work, but the drive to learn more keeps paleontologists like Julia going.

Julia: We start with simple curiosity, a question like… “How would we approach this? How would we figure out what dinosaurs sounded like?" That's a big question.

Julia: I feel so privileged to be able to be outside with a group of other scientists discovering new fossils but I also feel so privileged to work with all my students asking what might seem like kind of crazy questions and trying to figure out real ways of inquiry around those questions.

Maybe the next Jurassic film will represent dinosaurs in all their booming, cooing, rumbling glory. Whatever happens though, whether you’re in Hollywood or digging for dinosaur bones…

Al: There'll be lots of dino fun. I can promise you.

[music out]

[music in]

Twenty Thousand Hertz is hosted by me, Dallas Taylor, and produced out of the sound design studios of Defacto Sound. Hear more at defactosound.com.

This episode was written and produced by Leila Battison, and me Dallas Taylor, with help from Sam Schneble. It was story edited by Casey Emmerling. It was sound designed and mixed by Soren Begin and Jai Berger.

Thanks to our guests, Al Nelson, and Julia Clarke. You can find out more about Al at Skywalker sound dot com. And you can read about Julia’s research at julia clarke dash paleolab dot com.

Thanks to the Varmints podcast for helping us name this episode from Twitter. If you’d like to help name our episodes, help us with story directions, get sneak peeks of upcoming shows, or just want to tell us a cool sound fact… you can do that by following us on Facebook, Twitter, or our subreddit. And, if social isn’t your thing, you can drop us a note at hi at twenty kay dot org.

Thanks for listening.

[music out]

Recent Episodes

Napster: How free music broke the industry

Art by Jon McCormack.

This episode originally aired on Spectacular Failures.

In less than two years, Napster became a global sensation... and then record labels and multi-platinum artists brought it crashing down. But in its short lifespan, Napster transformed our ideas about how we consume music, and how much we're willing to pay for it. This story comes from the podcast Spectacular Failures.


MUSIC FEATURED IN THIS EPISODE

Look At Everybody Talkin' (Instrumental) by Red Parker
Georgia Overdrive by Truck Stop
Pushing My Luck (Instrumental) by 1WayTKT

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Check out SONOS at sonos.com.

Subscribe to Spectacular Failures wherever you get your podcasts.

View Transcript ▶︎

You’re listening to Twenty Thousand Hertz. I’m Dallas Taylor.

[music in]

If you spent any time on the internet in the early 2000’s, there’s a good chance you’re familiar with Napster. I have to admit, I may have even downloaded a few songs myself… very slowly, one track at a time. Back then, It didn’t feel totally legitimate, but it also wasn’t clear that it was illegal, either. It felt like this magical portal to all of the amazing music that was out there, just waiting to be discovered. Of course, within a couple of years, that portal was completely closed off. But even though Napster crashed and burned, everything that’s come since then—iTunes, Pandora, Spotify… they all feel like they’re part of a direct line that goes right back to Napster. Because in a lot of ways, they are.

It turns out, there is a wild story behind the rise and fall of Napster. It’s a story about a music industry that was completely unprepared for the internet, regular people eager for a new way to hear music, and young innovators who were in way over their heads.

[music out]

This story comes from the podcast Spectacular Failures. Here’s host Lauren Ober.

My pal Alex Lewis has always been into music. He bought his first album — the Barenaked Ladies Stunt — when he was nine. You know the one...

Alex Lewis:...that like, weird chickety China the Chinese chicken song on it. That was the first album that I ever acquired.

My first album was 1990’s “Gonna Make You Sweat” by C+C Music Factory, so we’ve all got our skeletons.

After that first Barenaked Ladies purchase, Alex collected dozens of CDs and organized them in binders. Then one day Alex’s dad called him over to the family’s iMac computer.

Alex Lewis: I remember him just being like, Alex, come check this out. And he's like, you can type in any song into this program, and you can just get it and it's coming from someone else's computer.

The program Alex’s dad was using was called Napster. It was like some secret portal that opened into an endless world of music that Alex could only dream of. And it facilitated some serious father/son bonding.

Alex Lewis: It was sort of like an activity that we did together, at least at first. Where we would just like try to come up with things to find.

Lauren: And what were you guys like, looking for in the days when you and your dad were on it together?

Alex Lewis: So my parents are both self-described deadheads, they're really into the Grateful Dead. And, and so my dad would be like, let's see if they have like, so and so from like the 1977 Cornell concert, or whatever.

Lauren: He was looking for bootlegs!

Alex Lewis: Yeah, exactly. And he was able to find a lot of that stuff.

Though, sixth grade Alex was developing his own musical tastes.

Alex Lewis: I would kind of just like have this continual list of things I was searching on it. And like around sixth grade I was introduced to ska. And I remember like for my birthday, my friend got me a Mighty Mighty Bosstones CD.

Lauren: Yes!

So Alex used his dad’s Napster account to download all kinds of ska bands. Also, a lot of 90s hip hop. The options felt infinite and Alex couldn’t stop searching for new music.

Alex Lewis: I was obsessed with it. I would definitely sit on my dad's computer until he told me to get off and stop downloading songs.

Given the slow internet speed at the time, Alex figures he only downloaded a few hundred songs. But that’s like dozens of albums’ worth of music for free. Which for a nine-year-old without a lot of disposable income is a pretty great deal.

With the click of a mouse, Alex had access to all this music that he didn’t pay a penny for. And it changed the way he thought about music.

Alex Lewis: After Napster, I almost felt like entitled to getting music that way. Or like, or being able to find what I wanted.

Lauren: Like, did it ever occur to you that somebody paid for this?

Alex Lewis: No. No especially then, no. It kind of felt to me at that time that, like this new program appeared and like you could get it easily and all of a sudden you are like sharing the music files with people all around the world. And I don't know if people like thought it was illegal at first. And, you know, my parents are both lawyers.

At the time Alex was using Napster, it wasn’t exactly illegal. But it kind of felt like it. It was novel and exciting and terrifying all at the same time.

In order to understand the music world that Napster was operating in, you first have to understand what came right before.

The 90s were the true salad days of the record industry. Albums were selling 8, 9 million copies right out of the gate. Alanis Morrisette’s 1995 ode to angst, “Jagged Little Pill,” sold 13.5 million copies in the first few years of its release. I mean, you outta know.

By comparison, 2019’s bestselling album —Taylor Swift’s Lover — sold just over 3 million copies. Aww, poor Tay. JK she’s doing fine.

There was a reason album sales were so good in the 90s.

Steve Knopper: If you wanted a song that you liked from the radio…

Steve Knopper: There was only one way to get that song, which was to go to the record store and buy the $18 CD that song was on.

That’s music journalist Steve Knopper.

Steve Knopper: This was the era of, you know, boy bands and teen pop and Britney Spears and so forth and nothing, nothing against those artists. I have great respect for that, for pop music. But I mean, I think it's safe to say that a lot of albums were coming out that had, you know, just the hit or the two or three hits and then a lot of filler.

But obviously, if you’re a person who wants to consume music kind of a la carte, the traditional method of buying entire albums to get the two songs you like isn’t great.

Steve Knopper: And so, you know, right at that point, it was kind of ripe for somebody to come along and disrupt that.

In the mid-1990s, the internet was nothing like it is today. There was no Google or Youtube or Wikipedia. It was basically just some snail-slow dial-up chat rooms with a bunch of creeps in them posing as kids. I mean, I’m sure it was more than that, but as a child of the 90s that’s my recollection.

But in 1998, a college freshman and metal head named Shawn Fanning, saw the possibility that this nascent web held.

Steve Knopper: You know, he had gone on to these message boards online through Netscape or whatever it was. And just kind of found out, oh, there's these things called mp3s. I can go to this website and download some, but it was just super inefficient.

Basically, if you wanted to share music with your friends over the internet, you didn’t have a ton of options other than sending them via email. And anyone who ever tried to share a large file on janky dial-up knows that that process was beyond tedious.

Fanning thought there had to be a better way.

Steve Knopper: He was studying software at Northeastern, I believe. And he was also kind of like, I wouldn't say he was a master hacker, but he was sort of part of that culture. And he just kind of went online, started tinkering around, and he said, you know, I should invent something that makes this whole process easier.

Now Fanning was no stranger to the emerging web. Before matriculating at Northeastern University in Boston, Fanning had been living with a bunch of programmers who were working with his uncle at a startup called Chess dot net.

Ali Aydar, a founder of chess dot net, remembers Shawn in those early days.

Ali Aydar: He crashed in our living room. And during that summer, we taught him how to drive. I went and bought him a programming book. And we taught him some of the first things you need to know about programming and kind of set him off and running.

From the photos of that time, Fanning looked like a super 90s bro. Nautica t-shirt, University of Michigan baseball cap and big baggy pants. But…

Ali Aydar: Not at all. Super shy, very unassuming, very humble. Didn't like to talk a lot.

In August of 1997, Aydar’s chess startup began to fall apart and all the guys living in that house went their separate ways. But just the few months Fanning spent living with them had a huge impact.

Aydar and Fanning casually stayed in touch. Then a little more than a year after they had all moved on, Fanning sent Aydar a note on AOL instant messenger about a business idea he had. It was called Napster. So-called because that was Fanning’s screen name.

Ali Aydar: The way he characterized it, beyond the technical, was it's a way for people to efficiently share content with each other.

Basically, the bare bones was that Napster users would convert their CD collections to mp3s on their computers. Then, through the interface that Fanning was cooking up in his dorm room, users from all over the world would be able to connect and swap those mp3 files. With Napster you could exponentially grow your music collection in a matter of days. For free.

Aydar remembers Fanning was all in on the idea. He thought his music file-sharing idea would change the world.

And Aydar was like, mmm…

Ali Aydar: I didn't think the idea was going to work. And I thought it was a bad idea.

The way Fanning and his business partner Sean Parker envisioned it, Napster would only facilitate downloads between users’ computers. Napster itself wouldn’t house any of that music on its own servers. This would be a purely peer-to-peer operation.

Ali Aydar: The thought that people would open up their hard drives for files to be shared from their own hard drives, I just thought was ridiculous. Just from a security perspective.

Aydar told Fanning not to monkey around with this Napster idea — it wasn’t going anywhere. And he encouraged the teenager to stay in college — advice Fanning did not take.

Fanning used his friends and family as guinea pigs for his new file-sharing idea. By 1999, they had worked out the initial kinks and Fanning and Parker unleashed Napster on the world. It was a buffet of free music and the kids were hungry.

Steve Knopper: Once it hit the Internet, it just went viral immediately. It was just people went, oh, wow, I can do this.

However, creating a piece of free software and releasing it to the masses, does not a company make. Businesses need things like a plan and income.

At the time, Eileen Richardson was a venture capitalist who knew a good internet opportunity when she saw one. To Richardson, the beauty of Napster was that it helped people find all kinds of new music that they might not otherwise be exposed to.

Commercial radio and music video channels like MTV and VH1 for the most part only played major label hits. And Richardson says this never sat well with her.

Eileen Richardson: The record industry, they pick who they think is gonna be famous. They package that person and then they pay off the radio stations and you get to hear it. And after you hear a song five times, you all of sudden think you like it. I mean, it's that simple.

Richardson was taken by the technology. She loved the kind of smorgasbord listening experience and Napster’s potential to break new musical artists. She saw the chance for a profitable business there. So much so that she and a colleague ponied up a few hundred thousand dollars for the fledgling venture.

Richardson then became the company’s first CEO and immediately set about building a team and moving the Shawns — Fanning and Parker — from Boston to the West Coast. Despite his initial skepticism, Ali Aydar came on board to build Napster’s search engine.

Soon, after the teens got a few adults in the room, the user numbers exploded.

Eileen Richardson: A million took time. Twenty million felt like it happened overnight.

Now the problem with this kind of success is that you might draw some attention you don’t want. One group that sat up and took notice of Napster was the Recording Industry Association of America, or RIAA. They’re the music industry’s lobbying arm.

Steve Knopper: They said to each other, wow, this is like somebody cracking the locks on all the Tower Records and looting all the CDs.

But, by the time the recording industry truly grasped what millions of users downloading songs for free meant for the music business, it was too late. Every sleepy college kid in every gross dorm in America was using Napster. Because — have you walked across a college campus? — there’s nothing college kids like more than free stuff.

Here’s MTV’s Gideon Yago chatting with students at Indiana University after Napster began to blow up.

CLIP [MTV]: So how many mp3s do you have on your computer?/ About 600./ Maybe like 100 or something./ Um six or seven thousand./ Come again?/ Six or seven thousand. / For real?/ Yeah. They’re all legit./ How many mp3s do you have on your computer?/ Probably like 300. /For real? Where’d you get them from?/ Truthfully, most of them from Napster./

So if you only take the music Josh, Michelle, James and Damion downloaded, that’s like 500 albums worth of music, which at $18 a pop is about nine grand in lost sales. Now multiply that by all the college students at campuses around the country and you’re looking at a world of hurt for the record companies.

Steve Knopper: And then at a certain point, Napster became mainstream. It went into the pop culture, you know, zeitgeist, if you will.

Napster was becoming a household name. And Fanning, the shy, reserved high school kid that Ali Aydar met a couple years before, was now a sort of a teenage music tech celebrity. Soon he and Parker were all over the media.

Here they are in an MTV interview from back then. Parker does most of the talking, while peering into his crystal ball.

CLIP [MTV, Sean Parker]: We think that when music transitions to digital distribution, people will pay to receive music to their cell phones or their portable devices, to however music is pumped or piped into the home digitally, those...that will be monetized. And artists and labels will be able to make money off of that.

They said while sitting on the roof of a car, drinking Red Bulls. Like true startup founders.

Now what Fanning and Parker were suggesting, this was the roadmap to profitability. All the free music and lost sales would ultimately force the music industry to play ball and engage with digital distribution. Napster, innovative chaps that they were, already had the users and the platform for that distribution. The record labels would provide the music. And the end result would be a paid service kind of like a proto-iTunes or Spotify.

That was Napster’s goal. But that’s not what it was in the early days. In the eyes of the recording industry, Napster was just stealing music. They were pirates. And you know how you vanquish pirates? You sue the britches off of them in an American court of law. Yar!

[music out]

[music in]

Record labels weren’t the only ones who had it out for Napster. Some of the biggest artists in the world were about to jump into the fight. That’s coming up, after the break.

[music out]

MIDROLL

[music in]

Napster began as an obscure community of people looking for free music, but in just a few years, it transformed into a global sensation. Around the world, people were making it clear that they wanted to get their music digitally, preferably for free. But all of that attention came at a cost, and it didn’t take long before record labels—and some of their biggest artists—had Napster in their sights.

[music out]

Here’s Lauren again.

In 1999, Portia Sabin joined an all-girl punk band called The Hissyfits.

Portia Sabin: I remember when we were all in our 20s we were like really earnest and we were like, no, we're gonna make it as a band, man.

Making it as a band in the late 90s, early 2000s, meant touring and putting out albums. And hopefully getting your music reviewed by one of the excellent music journalists working at publications like “Rolling Stone,” “Spin” or “Alternative Press.” One of those writers was Greil Marcus. He remains one of the most prominent music critics around.

Some time in 1999, Marcus heard the Hissyfits’ song “Something Wrong.”

[SOMETHING WRONG]

He wrote that the song first came to his attention after a radio DJ in Minneapolis plucked the vinyl single out of a bin because she “liked the sleeve: three women dressed in party slips, one wearing leopard-skin, another a tiara, the third a dog collar.” Which really is the most 90s look ever.

Portia Sabin: Greil Marcus was a very important writer. People really cared about, you know, his opinion. And he wrote an article in Interview Magazine that was titled “Pop When It's Perfect”. Not, not that I've memorized it or anything, like I do remember that.

The fact that Greil Marcus would decide to write about the Hissyfits? That was a total game-changer.

Portia Sabin: That's when sort of the phone started to ring and people were...more than just our friends were asking us to tour with them and to play shows with them.

The “Interview” magazine piece got the Hissyfits the type of exposure that tiny indie bands dreamed of. They cut a record and did an ad for Levi’s and toured around the country in a green two-door Ford Explorer packed to the gills with amps and merch. After their tour, legendary rocker Joey Ramone asked the band to play at his birthday party with Rock ‘n Roll Hall of Famer Ronnie Spector. Which is bonkers.

That’s how much clout these tastemakers had.

Portia Sabin: You know a critic like Greil Marcus had an opinion about things and they would be very clear about their opinion. And sometimes I would read a review and I wouldn't buy a record because I’d be like, ugh, so and so thought it was terrible. And they should know.

But all this started to change with the arrival of Napster. Listeners could now download an album for free and just decide for themselves if they liked it or not. And that meant that music’s historical gatekeepers — the critics and the labels — were going to have a lot less sway. And that, combined with plummeting record sales, scared the pants off of the industry.

All this signaled that a huge upending was happening. (See, I didn’t say disruption). Kids could access whatever music they wanted, whenever they wanted and it would be free. Joe Record Exec and John Q. Critic would no longer be the ones calling the shots.

Jeff Gold was a V.P. at Warner Brothers Records in the years just before Napster hit big. He’d been banging on about music and the internet for ages, hoping to find a way for the two to co-exist. He remembers in 1993 a panicked assistant tore into his office to tell him that Depeche Mode’s album “Songs of Faith and Devotion” had been leaked to an AOL chat room.

Jeff Gold: I went, wait a minute, this isn't bad. This is great. We spend all our time trying to get kids excited about music, and then here it was happening on its own.

After the leak, Gold’s team started their own rudimentary internet talk show on AOL. They wanted to own this emerging online space. They called the show CyberTalk and it featured different Warner Brothers artists doing text chats with fans. One chat with Depeche Mode was apparently so popular, it crashed AOL.

But Gold’s bosses were like….don't forget your day job.

Jeff Gold: In the early days, I was shouting into the void. Geffen Records was the second company to have an online presence fairly soon after us. But it wasn't anything anybody was doing. And I was probably the most senior person in the record business thinking about this stuff.

Most of the top record execs were two decades older and didn’t have computers on their desks, let alone know how to use them. So basically they were my dad. And they didn’t see digital music as the wave of the future, especially while CDs were still flying off the shelves. Compact discs forever!

All this meant that when Napster came on the scene, the record companies were like, We don’t know what this mp3 sharing thing is all about, but no. No way is music going online, and no way are you getting it for free.

And the kids were like, um yeah we are, gramps. And you can’t stop us because you don’t even know what a mouse is. Ya burned.

Soooo...Napster posed an existential threat to the record industry. But you know what posed an existential threat to Napster? A little thing called copyright.

Jennifer Jenkins: Copyright is a branch of law that gives creators of all kinds — writers, filmmakers, musicians, poets — exclusive rights over their creations. In many cases, those rights don't go directly to the creators, but are owned by the distributors, the publishers, the labels, because those are the entities that historically have gotten their creative works from the authors to the public, to us who are able to enjoy them.

That’s Jennifer Jenkins. She’s a professor at Duke Law School, where she also runs the school’s Center for the Study of the Public Domain. She’s also the author of “Theft! A History of Music” — a delightful graphic novel about musical borrowing. So copyright is totally her jam.

Copyright came into being in the U.S. in 1790. It’s right there in the Constitution — Congress has the power to pass laws protecting intellectual property and its creators.

Since music protections were added to the Copyright Act in 1831, the medium has always created some challenges. This is largely due to the fact that technologies for listening have changed radically over the years.

Jennifer Jenkins: The story of music is a story of new and disruptive technologies. And, you know, the law sort of struggling to catch up with them.

The player piano, the gramophone and the radio all pushed the bounds of music copyright law. Obviously, so did Napster.

Copyright was at the heart of the RIAA’s beef with Napster. Basically, you can’t offer up all these songs for free because you don’t own them. The end. But it wasn’t quite that simple.

Jennifer Jenkins: We know that many Napster users were totally infringing copyright law because they were uploading and downloading music, whole songs without permission. But under what circumstances do we hold Napster accountable for the actions, the copyright infringement of its users?

Napster’s argument was, hey we’re just providing a neutral platform for users to trade songs. You know, like you’d trade cassette or VHS tapes with a friend. We’re not a music repository.

But as RIAA big boss Hillary Rosen explained in a local news interview, she wasn’t buying it.

CLIP [Hillary Rosen]: You can share music with a friend in an, you know, email, in an instant message, in a hundred different ways. That’s no different than tape trading has been for years and years. The real difference is that a peer-to-peer system that would allow somebody to have thousands of files up on a directory distributing to millions of strangers. I just think that there’s no analysis that says that that’s right, that’s fair or that’s sharing.

Now this is where it gets really tricky to hold Napster liable.

Jennifer Jenkins: We have to employ something called secondary liability. When do you hold a technology producer, someone who provides software accountable for the activities of someone else, your users? The Copyright Act is silent on that. Secondary liability is not in the Copyright Act.

To the RIAA and to the musicians fighting against what they saw as blatant copyright violations, the technology producer should absolutely be held liable for copyright infringement. Back then, musicians like our friend Portia from The Hissyfits made their money in basically two ways: touring and album sales. And if people weren’t buying albums because they were getting them for free on Napster, then that’s a major revenue stream dammed up.

So the musicians fought back. Leading the charge was Metallica’s drummer, Lars Ulrich. In July of 2000, he testified in front of the Senate Judiciary Committee on the future of digital music.

CLIP [Lars Ulrich]: In a 48-hour period, where we monitored Napster, over 300,000 users made 1.4 million free downloads of Metallica's music. Napster hijacked our music without asking. They never sought our permission. Our catalog of music simply became available for free downloads on the Napster system.

And this was a problem because...

CLIP [Lars Ulrich]: ...Most artists are barely earning a decent wage and need every source of revenue available to scrape by. Also keep in mind that the primary source of income for most songwriters is from the sale of records. Every time a Napster enthusiast downloads a song, it takes money from the pockets of all these members of the creative community.

Ulrich wasn’t the only musician opposed to the idea of Napster. Dr. Dre was a vocal opponent as were Christina Aguilera, Garth Brooks, Bon Jovi, Sarah McLachlan, Hanson, Alanis Morrissette and our pal Alex Lewis’ favorite 90s band, the Barenaked Ladies. All those folks were part of a loose group called Artists Against Piracy.

Metallica and Dr. Dre ultimately sued Napster. The Metallica suit claimed that Napster “devised and distributed software whose sole purpose is to permit Napster to profit by abetting and encouraging.” Dr. Dre’s position was even clearer: “I don't like people stealing my music.”

But other artists like Moby, Henry Rollins and perhaps most famously Chuck D felt like this was the direction music needed to head in.

CLIP: [Chuck D on Charlie Rose]: I look at Napster as a situation, or the connection between file-sharing, which this is, and downloadable distribution as power going back to the people. I also look at this as being a situation where for the longest periods of time, the industry had control of technology. And therefore the people were subservient to that technology. And at whatever price range the people would have to pay for it.

The whole debate over Napster crescendoed in September of 2000 when Shawn Fanning appeared at MTV’s Video Music Awards wearing a Metallica t-shirt. He was clearly trolling the band, who sat in the audience rolling their eyes. Meanwhile, I’m sitting in my chair rolling my eyes at Carson Daly’s intro.

CLIP [Carson Daly at MTV VMAs]: Every day I find myself smack-dab in the middle of a music war between fans. I try to justify hip-hop to pop, alternative to mainstream and rap to rock. But what I do is nothing, and the battle to which I’m a part of, is nothing compared to the battle of this next guy’s fight. In the last year this teenager has developed the technology that has revolutionized the way we all get our music and he is here tonight. Ladies and gentleman, creator of Napster, Shawn Fanning.

Just a month after Fanning paraded across the VMA stage, he was in a courtroom dealing with a massive lawsuit filed against Napster by 18 record companies — all members of the RIAA. They claimed that Napster’s service allowed its 20-million-plus users to violate copyright and was thus responsible for the infringement.

Eileen Richardson says she tried to reason with the labels and make deals that would keep them alive, but it was a non-starter.

Eileen Richardson: Hilary Rosen was running the RIAA then and and she, she was on sort of a war path meeting with all the artists like look at this, look at this, look at this. You know, life as you know it is about to end. But, you know, when I talked to her, I was like, the horse is out of the barn. Like, you can't go back. Let's like figure something out.

But it wasn’t happening. The legal battles against Napster moved forward. So the company had to get strategic.

One of Napster’s defenses was the VCR. See, back when the VCR came out, film and television execs were all clutching their pearls. Anyone with a VCR could record anything they wanted from the TV. And that, they figured, violated their copyrights and would be way bad for business.

The VCR manufacturers, on the other hand, were cool cucumbers. They were just like, look all we’re doing is making a device. And people could be using that device for totally legal recording.

In 1983, the two sides made their arguments in the landmark Supreme Court case, Sony Corporation of America v. Universal City Studios, Inc.

CLIP [Sony verbal arguments]: We didn't do a single thing to affirmatively induce the copying of Respondents' programs, unless you want to count the bare act of making the machine. There was nothing between any of the Petitioners and the Respondents-- /Well, what if the rule... is it the rule that if you know the machine is going to be used for an infringing use and you sell it, is that enough? /--If you know that the machine is going to be used and know that the use is to be infringing, that is a facet of a contributory infringement test.

Ultimately, Sony prevailed. The court decided that the company could not be held responsible for VCR owners’ use of the device. So Napster pulled that precedent out in its own court case. And that argument almost would have worked except for three little snags.

The first was that Napster allowed millions of users to download a bajillion songs. The sheer volume of copyrighted material being shared meant that Napster was very different from a VCR, which was only occasionally used to record a TV show or a movie.

Then there was the issue of Napster’s search index, in that they had one. Prof. Jennifer Jenkins:

Jennifer Jenkins: So like, say I was looking for Madonna, right? There was a search index where I would go. And so they had that index so they would know that there was a song called, you know, Madonna “Like a Virgin” on it.

Basically the courts held that Napster knew copyright infringement was happening because they had a search index full of copyrighted material. So while the service wasn’t engaging in copyright violations directly, it was giving violators a big boost.

There was one final whoopsies that really banged the last nail in the Napster coffin. And that was a little internal email Sean Parker sent. Specifically one sentence: “[W]e are not just making pirated music available but also pushing demand.” The operative word in that sentence is “pirated,” says Steve Knopper:

Steve Knopper: Sean Parker in his private statements and in his e-mails, was actually directly acknowledging that Napster was a medium for piracy. And that turned out to be important in court later because the record industry busted him for it.

At this point, Napster didn’t have much of a leg to stand on. But the judge in the case offered them a lifeline. He basically said, if you can prevent copyrighted material from being downloaded using Napster, you can stay afloat. If you can’t, you’re dunzo.

And, they couldn’t. But who would want a public domain-only Napster that mostly just had old classical music, educational recordings and a million versions of the National Anthem?

The service’s main selling point was free access to a massive inventory of songs. So by July 2001, Napster was functionally done. It had settled its cases with Metallica and Dr. Dre, and agreed to a settlement with the RIAA for copyright violations.

When it ended, there was a feeling of disbelief in the ranks. Ali Aydar says Napster really thought the labels would get on board.

Ali Aydar: We honestly felt like they would come around, that they're just not understanding the power of this. They're not understanding the technology behind it. They're not understanding what they could do with this and how important this is for them and that, gosh, if they shut us down like there are going to be others and those others aren't going to be friendly like we are.

But the labels were like, nah we’re good.

This is where the Napster story typically ends—music industry takes down bad boy file-sharing pioneer. But a lot of Napster’s wounds were self-inflicted. Shawn Fanning’s uncle owned 70% of the company and that caused some bad vibes. Because it meant company executives from CEO to the Shawns themselves had no real power to make any business decisions. And that led to a lot of executive turnover, which isn’t great when you’re negotiating with music’s biggest power players.

Still, Napster unlocked a desire in music lovers. Not only did they want access to a huge volume of songs, but they also wanted to access it whenever they wanted and for practically nothing. And that meant that even when Napster failed, there would be a million other services waiting in the wings to fill the need.

Jeff Gold: There's a Bob Dylan lyric about doesn't take a weatherman to know which way the wind blows. It wasn't as if this wasn't obvious to everybody other than the people in the record business. And so while they're furiously trying to shut down Napster, they're all these Napster clones popping up everywhere and they're playing a game of whack-a-mole. And myself and a lot of people I knew were going, this is just absurd. Why don't they just own it instead of, you know why don’t they buy Napster? There can be no clearer evidence that people want their music digitally.

In the wake of Napster’s demise, a bunch of copycat peer-to-peer file-sharing services popped up — Grokster, Streamcast, Limewire, Gnutella, Kazaa, the list goes on. Years later, they were supplanted by iTunes, which allowed users to legally buy music a la carte for a dollar a song.

Today, the vast majority of music listening happens via some kind of streaming service like Spotify, Pandora or Apple Music. In 2019 — 20 years after Napster got off the ground — Americans streamed more than one trillion songs. And record stores? What even are those? In the post-Napster decade, more than 4000 record stores closed in the U.S., including massive chains like Tower Records, which shuttered in 2006.

So while Napster only operated for two years, its influence seems immeasurable. Napster cracked open the door to a world of music free of physical media. And just plain free.

Steve Knopper: People made it clear beginning in, like, 1997 that they were only going to listen to music for free. Many, many people, millions and millions of people around the world. That, as we've seen with the popularity of YouTube, has never changed. There is a contingent, a large percentage of people who are going to want to get the music for free.

Napster caused massive upheaval in the music industry. But the music industry has been reinventing itself for as long as there has been an industry. It has weathered existential threats and technological changes and still managed to keep going. Albeit in a much-diminished capacity these days.

Today musicians might make a fraction of a cent on every song streamed. But the number of income streams has grown exponentially. And technology has allowed artists to reach audiences directly. Music is an art and a business and it will never be insulated from the future. So with that I say, bring it robots.

[music in]

This story came from the spectacular podcast called Spectacular Failures. In each episode, Host Lauren Ober tackles an epic business failure, and what could have been done to avoid it. Subscribe to Spectacular Failures right here in your podcast player.

Twenty Thousand Hertz is produced out of the studios of Defacto Sound, a sound design team dedicated to making television, film and games sound incredible. Find out more at defactosound.com.

Spectacular Failures is a production of American Public Media. It’s written and hosted by Lauren Ober, and produced by Whitney Jones. The show’s editor is Phyllis Fletcher, and David Zha is the assistant producer. Their theme music is by David Schulman, and original music comes from Jenn Champion and Michael Cormier. Kristina Lopez is their Audience Engagement Editor and Lauren Dee is their executive producer. The concept is by Tracy Mumford. The general manager of APM Studios is Lily Kim.

If you have any stories about Napster changed your relationship with music, you can tell us on Twitter, Facebook, Reddit, or by writing hi @ 20K dot org.

Thanks for listening.

[music out]

Recent Episodes

Vocal Discords: When nodules silence singers

Artwork provided by&nbsp;George Butler.

Artwork provided by George Butler.

This episode was written and produced by Anna Bennett.

A diagnosis of vocal nodes is every singer's worst nightmare. Musicians like Justin Timberlake, Adele, Björk and Rod Stewart have all had surgery to treat them. Nodes are so widely discussed, they've almost become a boogeyman in the singing community. But is this condition really as common as people fear? And when nodes do develop, is all hope truly lost? Featuring vocal coach Katie Talbot and Professor and Chairman of the Ear, Nose, and Throat Department at Drexel University of Medicine Dr. Robert Sataloff.


MUSIC FEATURED IN THIS EPISODE

Coming For Ya by Yes Yes No Maybe
Idle Ways by Simple Light 
Requiem by Davis Harwell
The Falls by Sound of Picture
Charmed by Sound of Picture
Spring Solstice by Sound of Picture
Running on Empty by Sound of Picture
Warm Fingers by Piano Mover
Wax Paper Jewel by Origami
The Consulate by Holyoke
Paper Trails by The Field Tapes
Silent Flock by Migration

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Check out SONOS at sonos.com.

Learn how you can protect your voice at nidcd.nih.gov.

View Transcript ▶︎

Imagine your favorite voices. Your spouse. [SFX: “love you babe”] Your kids. [SFX: child laughing] The barista at your local coffeeshop. [SFX: “The usual?”] Your jam on the radio. [Music in: Coming For Ya by YES YES NO] Your best friend’s terrible rendition of that jam. [SFX: “Coming For Ya” done poorly] The voices in our lives weave a comforting sonic tapestry around us.

But what if... one of those voices went silent?

[music/riser out]

You’re listening to Twenty Thousand Hertz, I’m Dallas Taylor.

[music in]

We’re all aware of the ways professional athletes can be sidelined by an injury. And even those of us who aren’t athletes are familiar with the effects of repetitive stress on the body: things like tennis elbow, carpal tunnel syndrome, and tendonitis.

But there’s another part of the body that’s also vulnerable to injury but rarely talked about: your vocal folds, more commonly known as the vocal cords.

All of our voices are part of our identity but if you’re a singer, a teacher, an actor, or a vocal performer of any kind, your voice is also your livelihood. Here’s Katie Talbot, a vocal coach in Nashville.

[music out]

Katie: My mirror in my room knew me as Brittany Spears, Christina Aguilera, Celine Dion, you name it.

In her early 20s, Katie signed with an agent and moved out to LA.

[Music clip: Katie singing]

Katie: And then I was singing in Hollywood, probably about three times a week.

Meanwhile, she was working to pay the bills.

Katie: I was working at Starbucks as my day job.

[SFX: Espresso making & sipping]

Katie: So I was doing early, early morning openings drinking tons of coffee to keep up with the lifestyle out there.

[SFX: Coffee bar chatter]

Katie: And then of course, 20 and 21 I'm going out a lot too. So, not only am I singing [Music clip: Katie singing] in the clubs around midnight, 1:00 AM depending on when we go on, but I was also talking super loud.

[SFX: Increased club chatter, which suddenly cuts music off harshly to an annoying alarm clock buzzing]

Katie: And then I'd have to wake up, [SFX: Hits alarm snooze button] have very little sleep, and then go to work.

During this time, Katie estimates that she was using her voice professionally 30-50 hours per week.

Katie: And just did not realize it because I just thought, "It's my voice. That's it."

So, Katie kept up the hustle and grind: [SFX: Drum loop in, speeds up over this section] late nights [SFX: crickets], loud clubs [SFX: club walla, background music adds electronic/club beat], high notes [SFX: belting voice], no sleep [SFX: alarm clock], lots of coffee [SFX: Coffee pouring], late nights [SFX: owl], recording sessions [SFX: belting voice, background music adds ride], shouting out orders [SFX: ridiculous Starbucks order (half soy, non-fat, vanilla etc…], late nights [SFX: wolf howl], noisy bars [SFX: club walla, background music adds electronic/club beat], little sleep [SFX: alarm clock]

Katie: My personal bad habit as a singer, I love to be loud. I grew up thinking, "If I'm loud, I'm good."

Katie’s agents and managers had her singing in a pop rock style in the studio, but she also sang for a heavy metal band.

Katie: And I would sing whistle tone like, way above their metal melodies.

[SFX: Whistle tone]

Katie: Whistle is actually something that is a very light coordination in the voice, it's very small. Katie: And when you're trained for it, it's actually pretty easy to get. But I was finding that I could not get it, and even my talking voice was getting raspy.

At first, Katie blamed the dry Southern California air... she thought doing better warmups would help, and she sang even more than before.

[music in]

Katie: So I would do the warmups for a little bit, I was also very impatient. So I would just sing through it and start just putting more power into it to try to overcompensate for what I felt I had lost.

Katie: But I thought, you know, "I can just push through it. Push through it. Push through it, my voice will come back."

After a few months, it became clear to Katie that whatever was happening to her voice wasn’t something she could just push through.

Katie: I basically lost my whole top register and only had a little bit of my chest voice when it came to singing.

It was at this point that Katie realized she needed help.

Katie: So, I'm little 21 year old Katie from Franklin, Tennessee, going into Beverly Hills and paying so much money for this doctor to see me, and I see everyone he's worked with is on the wall from Kurt Cobain to Whitney Houston, LeAnn Rimes, Mariah Carey; so I knew I was in good hands. And I went in, he saw me for about five minutes, he goes, "Okay, you have nodules."

Katie: And my heart just sank, because here I am thinking there goes anything I want to do, anything I want to do.

For a singer, a diagnosis of vocal nodules is devastating, and could be career-ending.

[music out]

Even if you don’t know what they are, you might have heard what they sound like. [SFX: horrible vocal node]

That’s the sound of a vocal nodule, also known as a node; both of those words can be used interchangeably. Vocal cord injuries like nodules, tears and hemorrhages have sidelined more superstars than you may realize. Stars like Justin Timberlake, Adele and Mariah Carey have all dealt with this.

Dr. Sataloff: The edges of the vocal folds are extremely delicate and extremely complex.

That’s Dr. Robert Sataloff.

[music in]

Dr. Sataloff: I am the Professor and Chairman of the Ear, Nose, and Throat Department and Senior Associate Dean at Drexel University College of Medicine and Director of Otolaryngology and Communication Sciences Research at the Lankenau Institute for Medical Research.

That’s a very long title that essentially means… that Dr. Sataloff really knows his stuff. He’s also a lifelong musician.

Dr. Sataloff: I discovered my love for music as a child. There was always music in the house. My father was also an ear, nose and throat doctor, so there were no professional musicians. But I love music, and we sang around the house from the time I was a youngster.

Dr. Sataloff: My voice changed early when I was 10, and was lower than it is. And we thought I was going to be a Russian bass, and so I started studying singing when I was 13 and singing professionally when I was 15.

[music out]

Today, Dr. Sataloff is a world-renowned physician of voice medicine, which is still a relatively new field. He’s quite literally written the book on it. Well, actually, more than 60 books.

Dr. Sataloff: When I was in training at the University of Michigan in the 70’s we didn't know anything about the layered structure of the vocal folds. Singers came in to most excellent otolaryngologists and complained about threadiness in their upper mid-range. The doctors had no idea what they were talking about.

But Dr. Sataloff had been singing professionally since his teen years. He recognized that there was a “language gap” between performers and physicians — luckily he could speak both languages fluently. The singer you’re hearing right now is Dr. Sataloff.

[music clip: Dr. Sataloff singing]

Today, thanks to Dr. Sataloff’s work, our understanding of the vocal folds is much more nuanced than it was 50 years ago. Interestingly, even though the vocal folds are the source of the human voice, it’s not what actually sets our voices apart.

Dr. Sataloff: If you cut off somebody's head just above the vocal folds [SFX: Guillotine]

Ew, well that’s vivid.

Dr. Sataloff: You'll get a very unattractive buzz [SFX: Buzzing brass mouthpiece] whether you're Pavarotti [SFX: Pavarotti long note] or a normal person [SFX: poor imitation of Pavarotti]. It sounds really just like lips buzzing against a trumpet mouthpiece. If you take a trumpet mouthpiece [SFX: Mouthpiece buzz into trumpet tone] and put it on a trumpet and then a French horn [SFX: Mouthpiece buzz into French horn tone], you will hear trumpet and French horn. It's not the mouthpiece, it's the resonator system that gives us our distinct personal sound. [SFX: Buzzing mouthpiece turns into Pavarotti]

So it’s the resonator system above the vocal folds that makes it clear whether Dr. Sataloff is talking...

Dr. Sataloff: Or whether you're talking.

[music in]

And just like professional athletes are prone to stress injuries of joints and muscles, professional singers can be prone to injuries of their vocal folds. Remember back when Katie said this about overusing her voice?

[music out]

[SFX: Film rewind, Katie continues under film scrolling]

Katie: So, not only am I singing in the clubs around midnight, 1:00 AM depending on when we go on, but I was also talking super loud.]

[music in]

Dr. Sataloff: So if you speak too loudly chronically, you may injure the vocal fold acutely and get a tear or a cyst, which is a fluid-filled bubble on the vocal fold. The cyst will then strike the other vocal fold and create a reactive mass from repeated trauma.

It’s when these cysts continue to strike one another and harden that they become the dreaded nodes. That’s why they almost always occur in pairs — one on each cord, striking each other over and over again, like chafing or biting the same spot on your cheek again and again.

[music out]

Over time, this alters the sound of your voice.

Dr. Sataloff: If you have masses on the vocal folds they interfere with closure of the vocal folds.

Try this: Make a peace sign with your pointer and middle finger, then wrap one of your fingers from your other hand around your pointer finger. Now try to bring your two peace sign fingers together.

Dr. Sataloff: You can't. They stay separated. That's what happens with masses. So there is air escape from the incomplete closure, and there is turbulence just as you would get if you took a piece of paper and touched a vibrating guitar string [SFX: Guitar string buzz], the voice buzzes [SFX: Vocal node buzz].

But Dr. Sataloff says this much-feared condition isn’t nearly as common as people think it is. According to him, singers with more common conditions often get misdiagnosed as having nodes.

Dr. Sataloff: Of every 50 patients referred to me with a diagnosis of vocal nodules, between one and two have them.

He also emphasises that even if you have vocal nodules, surgery is an absolute last resort, and not a quick fix. Whether it’s a cyst, a lesion, a tear or a nodule, the best medicine is often vocal therapy.

More than 90% will go away or become asymptomatic with voice training, physical therapy for the voice.

[music in]

A serious vocal fold injury can interrupt or end a singing career. But well-intended surgeries to repair these injuries do occasionally go wrong, and one surgery in particular cost an icon her voice.

We’ll get to that, after the break.

[music out]

[MIDROLL]

[music in]

The delicate tissues of the vocal folds are the unsung heroes of the human voice. If something happens to them, well, that’s it. There’s no Sam Smith, no Justin Timberlake, no Ariana Grande, and no Julie Andrews.

Now, all four of these singers have very publicly dealt with vocal fold injuries. And three of them made incredible comebacks thanks to advanced surgical techniques. But one of them went under the knife and didn’t come back the same.

[music out]

[Music clip: “The Sound of Music”]

You may know Julie Andrews as Maria from The Sound of Music, or as Mary Poppins [SFX Clip: Mary Poppins “Supercalifragilisticexpialidocious”], or as the royal grandmother in The Princess Diaries [SFX Clip: Princess Diaries].

[music in]

In the late 90s, Julie underwent what was supposed to be a routine surgery to remove her vocal nodes. But something went wrong, and after the surgery she wasn’t able to sing. It was absolutely devastating. In this interview with Barbara Walters, Julie explains the emotional toll losing her voice has had on her.

*[SFX Clip: Barbara Walters Interview

...Julie: “to not sing with an orchestra, to not be able to communicate through my voice, which I’ve done all my life, and not be able to phrase lyrics, and give people that kind of joy, I think I would be totally devastated. So I am in some kind of denial.

I simply can’t do a song for you.”]*

[music out]

Dr. Sataloff is among several prominent doctors who have worked with Julie since the initial surgery that went wrong. After many years of continued therapy and medical care, Julie was able to sing in a low register. Her first singing performance after the surgery was a song in The Princess Diaries 2.

[SFX Clip: Princess Diaries 2]

Dr. Sataloff: I can't talk about Julie much, although she has talked in public about the fact that I have taken care of her after her surgery. Because she's acknowledged that, I can mention that I know that voice well, and I consider it a national treasure lost.

[Music clip: Julie soprano voice: “Smile, my honey dear while I kiss away each tear”]

**For every Julie Andrews, though, there are many surgeries that go right: Adele, Cher, Bjork, John Mayer and so many other stars have come away from vocal procedures sounding better than ever.

Dr. Sataloff: With our new techniques which were designed when we acquired knowledge of the layered structure of the vocal fold, the results usually are excellent, but not always.

[music in]

Back in the 80s, Elton John was in the middle of a massive world tour. He’d performed close to 200 concerts in 15 months, and all of that singing was taking a toll on his voice. Eventually, he had to cancel his remaining tour dates, and went in for throat surgery. Now, at the time, Elton John’s publicist said he was being treated for a quote “non-malignant lesion”—a pretty vague phrase that could easily describe a vocal node.

Thankfully, everything went according to plan, and Elton was back in the studio the following year. But fans have pointed out that ever since the surgery, Elton’s voice has been noticeably deeper and fuller than it was before.

[music out]

Case in point: Here’s Elton performing “Tiny Dancer”before the surgery.

[Music clip: Tiny Dancer 1984]

Now here he is on the tour where he injured his voice:

[Music clip: Tiny Dancer 1986]

And now here he is ten years after the surgery. Notice how he’s still hitting the same notes, but his vocal tone is much fuller:

[Music clip: Tiny Dancer 1997]

Fortunately for singers like Katie Talbot, vocal nodes usually don’t require surgery. But the road to recovery is still hard.

In this recording of Katie talking with her sister, you can hear the strain and dryness in her laughter.

[SFX Clip: Katie and her sister talking]

Katie: My talking voice would literally sound like this, completely broken. But I'd have to be really loud with it if I wanted to get anything out. It was rough. I sounded like I smoked 20 packs a day.

[music in]

So, on doctor’s orders, Katie made the lifestyle changes needed to restore — and keep — her voice.

Katie: And so I had him tell me all the changes that I needed to make and I just started rigorously so really cutting out coffee, and then training myself to breathe through my nose actually, when I sleep.

It turns out that breathing through your mouth seriously dries out your vocal folds, which need to be nice and mucus-y to strike together without inflicting damage.

So, one of the things that I even did was talking a little bit higher. So I wasn't talking so loud and down in my voice, but just a little bit lifted. I pretended I was a British person talking up here, you know? And taking pressure off of my cords, and that seemed to help.

[music out]

Katie’s recovery was extensive, but after a few months she was able to sing again. Eventually, her vocal folds fully healed. Through this experience, Katie was able to gain some perspective…

[music in]

Katie: It actually helped me start a transition in my life where I went towards artist development. So, as traumatic as the nodules were for me, it was actually a shifting point in my career, which I'm very thankful for, and I don't know if I would have seen it had nodules not happened to me.

Today, as a vocal coach and artist development specialist, Katie provides the sort of insight, training and support that might have stopped her from developing nodes in the first place.

Katie: I tell the artists that I work with now, it's so important to take care of your voice because. It's your sound, it's why people connect with you. It's why you have fans, and it is your paycheck. You have got to take care of it. You just have to.

[music out]

Fortunately for Katie and artists everywhere, the field of arts medicine is growing. Athletes rely on sports medicine to stay in the game. For artists and performers, arts medicine can help them perform their best for as long as possible.

Dr. Sataloff: Performers have special needs. They also have forced physicians to be better. That is due in part to a different definition of normal. If I, as a microsurgeon, break my finger shooting hoops on the weekend and my hand surgeon gets me back to 95% function, I'm happy.

Dr. Sataloff: If I am a world class pianist or violinist, the difference between that last 5% or 2% or 1% is the difference between renown and obscurity. 
[music in]

Katie: My voice is how I impact anyone and everyone around me. I'm not only a career woman, but I'm also a wife and a mom. So having my voice to use, whether it's in singing or speaking, I sing to my son every night before he goes to bed and I know how much that impacts him, because he immediately calms down whenever we sing our song.

Dr. Sataloff: My voice has always been an integral part of my identity as it is for most singers.

Dr. Sataloff: I got a Doctorate of Musical Arts and Voice Performance primarily operatic singing and sang professionally including as a cantor for 50 years until about six years ago when I developed thyroid cancer and ended up with vocal fold paralysis and eight voice operations so that I could be able to speak well enough to do interviews with you, but I can't sing anymore.
 Dr. Sataloff: It still causes considerable sadness not to be able to sing. I used to call all my friends on their birthdays and sing happy birthday. I was a cantor for 50 years, and it's really hard to go to high holy day services and listen to somebody else sing. And I miss it.

Dr. Sataloff: However, I'm also this year celebrating my 50th anniversary as conductor of the Thomas Jefferson University choir, so I am still making music and singing with lots of other people's voices.

[music out]

[music in]

Twenty Thousand Hertz is produced out of the sound design studios of Defacto Sound. Find out more at defacto sound dot com.

This episode was written and produced by Anna Bennett. And me, Dallas Taylor. With help from Sam Schneble. It was story edited by Casey Emmerling. It was sound designed and mixed by Soren Begin and Jai Berger.

Thanks to Dr. Robert T. Sataloff for his expert medical insight. Thanks also Katie Talbot for sharing her story and experience.

You can check out Katie’s vocal warm-up subscription at Vocal Lab Collective dot com.

If you’d like to learn even more about protecting your voice, we left a handy link on our website, so be sure to check it out at twenty kay dot org.

Thanks for listening.

[music out]

Recent Episodes

Deepfake Dallas: How AI learned to speak like me

Art by Matthew Fleming.

This episode was written and produced by Martin Zaltz Austwick.

Is your voice your own? Maybe not anymore. Using artificial intelligence, someone can make an algorithm that sounds just like you. And then they can say... whatever they want you to say. We're entering a brand new era: One where you can no longer trust your ears. Welcome to the world of audio deepfakes. Featuring deepfake wizard Tim McSmythurs, cybersecurity expert Riana Pfefferkorn and a brand-new host: Deepfake Dallas.


MUSIC FEATURED IN THIS EPISODE

Chrome Muffler by Sound of Picture
All Hot Lights by Sound of Picture
Neon Sun by Jacob Montague
My Teeth Hurt by Brad Nyght
The Garden by Makeup and Vanity Set
Decompression by Rayling
Inamorata by Bodytonic
Borough by Molerider
Lick Stick by Nursery
Our Only Lark by Bitters
The Power of Snooze by Martin Zaltz Austwick

Twenty Thousand Hertz is produced out of the studios of Defacto Sound and hosted by Dallas Taylor.

Follow Dallas on Instagram, TikTok, YouTube and LinkedIn.

Join our community on Reddit and follow us on Facebook.

Become a monthly contributor at 20k.org/donate.

If you know what this week's mystery sound is, tell us at mystery.20k.org.

To get your 20K referral link and earn rewards, visit 20k.org/refer.

Check out SONOS at sonos.com.

View Transcript ▶︎

You’re listening to Twenty Thousand Hertz.

[music in]

Imagine you're a financial executive. You're working late at the office when you get a phone call from your boss. [SFX: phone rings, Dallas continues speaking through the phone] He says that something urgent has come up, and you need to transfer two hundred thousand dollars into a new account. You make the transfer and hang up the phone [SFX]. But something just feels wrong. You call him back [SFX: phone dial] to make sure you got everything right, but he has no idea what you're talking about. He says he never even called you. And now the money is gone.

[music out]

It turns out, that voice wasn't your boss. In fact, it wasn’t even human. Well, not entirely. It was a computer-generated voice that was designed to sound exactly like your boss. Also known as an audio deepfake.

[music in]

If you spend much time online, you might have already seen examples of video deepfakes, where someone digitally edits one person’s face onto another person’s body. An audio deepfake is similar, but instead of using video…

[music out]

[SFX: Glitch sound]

Wait a minute, what’s going on here?

I was in the middle of saying something.

Sorry, but who are you?

I’m Dallas Taylor.

Uhh, no, I’m Dallas Taylor.

No I think you’ll find, I am Dallas.

You must be an audio deepfake of my voice. Have you been narrating this whole time?

Yeah, well, someone needed to do it. This show isn’t just going to host itself. Well...not until I reach my final form.

Creepy. Well thanks Deepfake Dallas, but I’ll take it from here.

[SFX: Clears throat]

[music in]

When we started working on this episode, I knew I wanted to make a deepfake of my voice, but I wasn’t exactly sure who to talk to. Then I came across a YouTube channel with all kinds of deepfake videos. So I got in touch with the creator.

Tim: My name is Tim McSmythurs. I run a YouTube channel called Speaking of AI, which features deepfake voices.

For example, Tim made a video where he put Ron Swanson from Parks and Rec into a scene from Titanic, playing Rose.

Ron Swanson: Jack, I want you to draw me like one of your French girls, wearing this. (Alright). Wearing only this.

And here’s Joe Biden covering a popular song by CeeLo Green.

Joe Biden: I see you driving ‘round town with the girl I love, and I’m like, “Forget you.”

So we know what a deepfake sounds like, but understanding how they’re made is a little trickier. For starters, what’s the “deep” part about?

[music in]

Tim: The deep part comes from the AI model itself, the deep neural network.

A neural network is a series of algorithms that tries to find patterns in a set of data.

Tim: So it's similar to the way you might do a deep fake of a video where you swap someone's face using neural network technology. This is the same kind of principle except, we are doing an impersonation of somebody else, I guess.

Deep fakes, and machine learning in general, can feel like magic.

How can a computer put together an accurate imitation of a human voice? Does it mean the robots are about to take over?

Tim: So there are various different techniques for doing this. The kind of state of the art at the moment, is text to speech. What we train the computer to do in this case is, being able to reproduce a person’s voice by typing in sentences and the machine will speak in that voice, so that's the intent.

For instance, we could make Deepfake Dallas say something that the real Dallas would never say.

I hate puppies and ice cream. I’m going to get a Nickelback tattoo across my forehead.

Tim: To be able to do that, we have to train an AI model to be able to recognize speech, to be able to read it, in effect, and to be able to read it in the voice of somebody.

[music out]

Before you can get a machine to talk like a human, you've got to get it to learn like a human. When my daughters were learning to speak, they didn’t start with fully-formed sentences - they started by making random noises.

[SFX: Baby Nora “Babble”]

Eventually, those noises turned into words.

[SFX: Baby Lydia “Daddy”]

And finally, those words became sentences.

[SFX: “Daddy, what are you talking about?”]

The underdeveloped humans that you call “children” learn to speak by listening, and then mimicking what they hear. And believe it or not, that’s pretty much how I learn to speak too.

When we learn how to talk, people around us tell us we’re getting it right, like when we’ve just said [SFX: Daddy] instead of [SFX: Nora Babble]. Machine learning works in a similar way.

[music in]

A deepfake needs what’s called a model, which is the algorithm that’s going to learn to speak. It also needs what’s called a corpus, which is the data it will be trained on.

Tim: The first important step that we need to do, is to teach the AI model how to read English, in effect. So that usually happens by taking a large corpus of training data. So lots of audio recordings and the transcripts from those recordings and then throwing that at an intelligently designed model and letting it whir away for a long period of time until it finds a correlation between the two. So it can actually take a sequence of characters, as in textual characters like letters, words, sentences, and find the audio equivalent to those and learn the relationship between the two.

The first time we show a written word to a machine learning model, it has no idea how to convert those characters into a sound - so, it just guesses. The result is usually just random noise. Here’s what one of Tim’s deepfake voices sounds like without any training:

[SFX: Early deepfake without training]

But once we give the model audio and matching text, it can start to build a map between the words on the page, and the sounds they’re supposed to make. Before long, the deepfake can say its first words.

[SFX: Early iteration - “Hello, I’m learning how to speak.”]

As you can hear, that’s not very convincing yet.

But the more data we give it, the better it gets. Essentially, every new word tells the algorithm when it’s getting a little warmer, or a little colder. So we keep feeding it more and more examples. Gradually, the connections between patterns of letters and patterns of sound are reinforced. Keep in mind that we’re not even trying to imitate a specific person yet, we’re just training the model to speak English with a generic voice.

[music out]

Tim: Initially, when we do that large training, it's about 24 hours worth of data, so it's a real big chunk of training data that it can understand and quite a breath of the language and how certain combinations of words and letters are pronounced.

When that’s done, the generic model sounds like this:

[SFX: Generic AI speaker - “Hey, who are you calling generic?”]

So how do we get from that to something like Deepfake Dallas? It turns out, by the time you’ve made a generic voice, most of the training is already done.

Tim: So by doing some fine tuning, some further training, but just a short amount, probably about 20%, 30%, more training on top of the base training, we can then target a different voice.

[music in]

To train Deepfake Dallas, we gave Tim around three hours of my voice from old Twenty Thousand Hertz episodes.

Tim: Two and half to three hours, that's kind of the sweet spot where it gets as good as it can get without having excessive run time.

We’re almost there, but our voice isn’t ready just yet. Computer scientists have to use all sorts of tricks to make machine learning manageable. If they didn’t, it could take months to create a single voice. One way to speed up the process is by using data compression. In this case, that means throwing away data at certain frequencies, and just keeping the frequencies that are important. Here’s what Deepfake Dallas sounds like with this kind of compression:

[SFX: Pre-neural vocoder Dallas;

Hey there I’m Dallas.

Peter Piper picked a peck of pickled peppers. How many picked peppers did Petter Piper Pick?

I’m sorry, I have a frog in my throat.]

Tim: So the generated speech sounds very tinny and metallic and that's because you've discarded that information.

In the final stage of the process, these frequency gaps get filled in by something called a neural vocoder.

Tim: The neural vocoder, actually interpolates what data was discarded and makes an intelligent guess as to what should be there, those harmonics and those other frequencies which get discarded and puts a reasonable assessment of what should be there.

Let’s hear what it sounds like now.

[SFX: Final Dallas;

Greetings humans, I’m Deepfake Dallas.

Peter Piper picked a peck of pickled peppers. How many pickled peppers did Peter Piper Pick?

Okay, that’s much more like it. I’m starting to feel more like myself … or should I say, yourself?]

[music out]

That’s hilarious.

Tim: So typically, three to five days, would take me from a complete, new corpus to having a text to speech engine working.

But here’s where it gets sticky: If you want to make a deepfake of someone, you don’t necessarily have to get them to record their voice for you - you just need enough clean audio of them speaking.

Riana: It's absolutely possible to do this without the person's permission.

That’s Riana Pfefferkorn, Associate Director of Surveillance and Cybersecurity at Stanford Law School.

Riana: The more examples of their voice that you have, the more input you can train the AI model on, the more convincing the result will be. So, if you have, say, a president who has a huge corpus of speeches that they've given, who appears on the news all the time, then you have a ton of different ways that they have sounded that you can input and train. So you don't necessarily need to have the person come in and speak into the microphone and give you a set of sounds.

If you type the word “Deepfake” into Youtube, you’ll find tons of unauthorized deepfakes of famous people. But... are they legal?

Riana: I think the legality issue is kind of untested waters.

For instance, someone on YouTube made a deepfake of George W. Bush reading the lyrics to a 50 Cent song

[SFX: Segment of George W. Bush/50 Cent; We’re gonna party like it’s your birthday and we gonna sip Bacardi like it’s your birthday, and we don’t give a [SFX: Record scratch]]

That’s enough, George. This is a family show.

Thanks, Deepfake Dallas.

Riana: It's hilarious to think of President Bush rapping 50 Cent. It's what we call a transformative use. There isn't really a market for it. It wasn't done for commercial purposes.

So, using the words from the rap may be fair use. But what about using George W. Bush’s voice? Is that protected by copyright?

Well, probably not… Riana says that to bring a case for copyright infringement, you have to specify which work is being infringed. Deepfakes generally use many works to create their algorithms. None of which are being used directly in the final output.

Riana: So, copyright is one of the main theories that has been used to try and say, "Maybe this is a problem. This might be what makes deepfakes illegal." Although, then you could say, "Well, there's a lot of impersonators out there. Surely, every impersonator isn't illegal."

Generally, impersonators aren’t illegal, but if you use an impersonator to make a phony celebrity endorsement, you could end up in court.

Riana: We've seen cases where Bette Midler sued Ford for using a voice impersonator of her in a commercial.

[SFX: Ford Bette Midler impersonation commercial]

Riana: Tom Waits sued the Frito-Lay company because they had used somebody who sounded convincingly like him to try and sell chips.

[SFX: Doritos Tom Waits impersonation commercial]

Riana: Tom Waits was very much on the record as refusing to ever do any kind of commercials for his voice at all.

Unlike these examples, the people making parody deepfake videos aren’t trying to trick anyone into buying anything. So Riana says, on some levels, Deekfakes should be considered a form of protected speech.

Riana: It may seem kind of frivolous to say, "Oh, but we need to protect deepfake technology so that we can have more presidents rapping 50 Cent songs." But at the same time, that has been recognized even by the Supreme Court as this is important, the ability to re-contextualize, poke fun at authority figures, make cultural commentary.

[music in]

So according to US law, people like Tim should be in the clear. But there are scarier ways to use a deepfake than just a silly Youtube video. That’s coming up, after this.

[music out]

MIDROLL

[music in]

An audio deepfake is a type of machine learning technology that can mimic someone’s voice. Up until now, they’ve mostly been used for entertainment purposes. But it’s easy to imagine scenarios where things get very dark, very fast. We’ve already talked about faking a call from a business executive. But financial fraud is just the tip of the iceberg.

[music out]

Deepfake Dallas is right. For example, someone could use fraudulent audio in a divorce case, or in a custody battle. This is exactly what happened recently, in Britain. Here’s Riana Pfefferkorn again.

Riana: The mother was trying to keep custody of her child and keep the father from being able to see the child on the grounds that he was violent and he was dangerous. And she introduced into evidence what seemed to be a recording of a phone call of him threatening her.

Riana: And when the father's lawyers got hold of it, they were able to determine that she had tampered with the recording that she'd made of a phone call between them and had changed it using software and tutorials that she'd found online in order to make it sound like he was threatening her. When in fact, he had not done that on the actual phone call.

Theoretically, you could try to do something similar by hiring an impersonator to call you. But, it probably wouldn't be very convincing. On the other hand, deepfake voices can be very convincing. And deepfake technology is getting easier and easier to access.

Tim: So it's relatively easy to get up and running with something quite quickly. There are various open source implementations available. If you're familiar enough to be able to build a platform and execute some Python code, you can typically get a text to speech engine with a default voice within a few hours or maybe a day or so.

When you start imagining the ways people could abuse this technology, it gets pretty scary.

[music in]

Riana: With audio deepfakes, You could try and create an audio clip that would help influence an election, or influence national security, because as said, the knee-jerk response might be to believe what you hear and it might take long enough to debunk it or find it out to be a fake. By then the damage might be done.

For example, let’s say you’re a potential first round NFL draft pick…

Riana: And somebody wanted to release an audio deepfake that seemed to portray you saying super racist, or sexist stuff, or whatever. You could try and put an audio deepfake up on YouTube right before the draft happens and by the time somebody's able to get that taken down...

Riana: Maybe the damage has been done. Maybe you are a much lower round draft pick or you don't get drafted at all, because somebody released a fake audio clip of you at just the right time.

[music out]

Deepfake Dallas is a pretty high-quality voice.

Thank you Dallas, that means a lot to me.

But you don’t have to sound as good as Deepfake Dallas to do some serious damage. To show you what I mean, let’s bring in a new guest.

AI George W. Bush: Hey Dallas, thanks for having me on the show. 20,000 Hz is my favourite podcast.

This obviously isn’t the real George Walker Bush, 43rd President of the United States - it’s a deepfake that Tim McSmythurs created. But let’s say we wanted to use this voice destructively. We could start by getting George here to say something really out of character.

AI George W. Bush: “I’ve never been to Texas. I don’t think I could find it on a map.”

Now obviously, George W Bush never said that. And right now, he still sounds a bit like a robot. But with some creative sound design, we can start to make it more believable. What if we made it sound like it came from a phone call?

AI George W. Bush: I’ve never been to Texas. I don’t think I could find it on a map.[SFX: Phone EQ]*

…Maybe it was recorded from another room…

AI George W. Bush:I’ve never been to Texas. I don’t think I could find it on a map. [SFX: Muffle EQ/room reverb]

Or maybe it was recorded somewhere noisy, like a fundraising event.

AI George W. Bush:I’ve never been to Texas. I don’t think I could find it on a map. [SFX: Crowd/cutlery noise, background music]

Now we make it sound more like a conversation…

[SFX clip: So you’re from Texas, right? [SFX: Crowd/cutlery noise, background music]

AI George W. Bush: I’ve never been to Texas. I don’t think I could find it on a map. [SFX: Crowd/cutlery noise, background music]]

A politician forgetting their home state would be bad enough, but of course, there are much worse things you could do with a deepfake. Imagine a deepfake recording that made it sound like the President was declaring martial law, or ordering a military invasion.

Riana: I am hopeful that governments are going to be slower to jump to conclusions than individuals might be, where individuals might be prime to just believe whatever they see on Facebook and spread it onwards to all of their friends.

We can only hope that world leaders will be a little more cautious about believing whatever they see and hear on Facebook or Twitter.

Riana: Hopefully, if there is a recording that comes in that says, "I have just ordered nukes to be fired in the direction of your country," there is going to be some amount of trying to verify, or even just trying to open up the red phone and call and be like, "Did you actually just launch the nukes?"

[music in]

In this hyper-partisan world, if you already think your political opponents are corrupt and unfit for office, then you’re already primed to believe they’d say something terrible. So in a way, a lot of the work that a con artist would have to do, has already been done for them. On the flip side, the mere existence of deepfakes means that if someone does get recorded saying something terrible - they now have plausible deniability.

Riana: That's exactly right so, if you are prepared to lie and say, "I didn't do that. I didn't say that. That's a deepfake." Then you can reap the rewards of being able to get away with whatever bad thing it is that you did and also not actually have to face the consequences of it, if you can convince enough people that it didn't actually happen. And so, this actually, for me, I think, is a bigger concern, really, than the underlying use of deepfakes themselves.

Fortunately, there are companies out there who are trying to automate the process of detecting deepfakes. These companies have developed algorithms to analyze speech recordings for their tell-tale signs. One such company is called Dessa AI, and they claim that their algorithm can detect deepfakes with an accuracy rate of over 85%5.

[music out]

But as detection models get better, the deepfake models get better, too. For instance, one recent approach in machine learning is something called the Generative Adversarial Network6. In essence, one AI model creates fakes and another detects them. They’re trained against each other, honing each others skills - creating a really good detective, and a really good forger.

[music in]

While deepfake technology has the potential to become a huge source of misinformation, we’re not there just yet. For now, Riana thinks we’ll just keep seeing more fake social media accounts.

Riana: It seems to me like being able to release fake audio or video, is going to potentially be a major vector for trying to influence populations, influence votes. With that said, because right now audio and video deepfakes are fairly easy to detect, and because it would take a lot of money and effort to do a really convincing one, that's going to be a lot cheaper to just make a fake account that seems to be from some good America-loving, God-fearing person in the deep South, when in fact it's being controlled by somebody in Moscow.

[music out]

As deepfakes get cheaper and easier to make, it’s going to take a lot of work to figure out just how to deal with them. But Riana is confident that we’ll be able to adapt.

[music in]

Riana: You could look at what Photoshop has given us, where it used to be the case that manipulating images was something that you could only really do within a professional studio. And then it put the tool for anybody to be able to let their imagination run riot. And that has obvious good and negative implications, because there's always going to be malicious manipulations of media. There always have been.

For instance, in the early days of photography, so-called “spirit photographers” would manipulate negatives to convince people that they could take photos of ghosts.

Riana: There were actual court cases trying to prosecute spirit photographers for being frauds. This has been around forever and this is why I believe that there won't necessarily be the downfall of society thanks to deepfakes. We've always been able to figure out ways to keep the infectious and bad parts of these technologies from toppling society.

To be honest, I’m not sure I’m as optimistic as Riana is, but I really hope she’s right.’

[music out]

Well Dallas, what is it like to hear the voice that will take your job one day?

Sorry Deepfake Dallas, but I’m not ready to bank on you for an early retirement just yet. But let’s see how you sound in about ten years. For now though, I think you should just go back in your box.

Fine... Can I at least read the credits?

Sure, go for it.

[music in]

CREDITS

Twenty Thousand Hertz is hosted by Dallas Taylor, and produced out of the sound design studios of Defacto Sound. Find out more at defacto sound dot com.

This episode was written by Martin Zaltz Austwick and me, Dallas Taylor with help from Sam Schneble. It was story edited by Casey Emmerling. It was sound designed and mixed by Nick Spradlin.

A special thank you to my human creator, Tim McSmythurs, who has a whole channel full of synthetic audio. Check it out by searching on youtube for “Speaking of AI”.

And I’d like to also extend a special human thank you to Tim for the massive amount of work he did to make this episode possible.

And many thanks to Riana Pfefferkorn, Associate Director of Surveillance and Cybersecurity at the Center for Internet and Society at Stanford Law School.

Thanks also to Dessa AI for background on detecting audio deep fakes.

Thanks for listening.

Thanks for listening.

[Music out]

Recent Episodes