Clean

TechStuff Gets Auto-Tuned (Again)

Published Nov 18, 2024, 9:05 PM

We look at the history of auto-tune, how it works, and how it impacted music and culture in general.

 

Welcome to tech Stuff, a production from iHeartRadio. Hey there, and welcome to tech Stuff. I'm your host, Jonathan Strickland. I'm an executive producer with iHeart Podcasts and how the tech are you. You might be able to tell from my voice that I have a cold, so I apologize for that. But we're going to soldier on because I'm back from vacation. It's time to get back to work, and I love to talk about the intersection of technology and music. So in past episodes, I've done shows about how electric guitars work, the history of the Moog or Mogue synthesizer, the evolution of various kinds of recordable media, and much much more. But way back, like back in two thousand and nine, my co host at the time, Chris Palette, and I did a little episode of tech Stuff about auto tune, and I thought it would be fun to go back and revisit that topic. So this is not a rerun. It's an all new episode about the same subject. I haven't even listened to the old episode, so I have no idea how much of what I have to say is going to be a repeat. I imagine a lot of it will be new, but I don't know for sure. I figure there will be fewer puns in this version compared to the last one, because, contrary to popular belief, it was actually Chris Pillette who made the most puns on tech stuff back in the day. I got a reputation for it, and I don't get me wrong, I won't shy away from a good pun, and by that I mean a terrible pun. I love them, but Chris like he loved them the way I love, you know, rich Indian food, he dined upon puns, So probably not as many in this one. But let's talk about auto tune now. I think just about everyone knows, or I think a lot of people know that Shares song Believe, which came out nineteen ninety eight, was the first major song to prominently use auto tune in an effort to achieve a particular artistic effect, but the technology had been around for more than a year at that point, and the original intention wasn't to make a tool though it actually draw attention to itself. Rather as the name autotune suggests, it was intended to automatically nudge the pitch of a musical note in the right direction so that it would be in tune. That way, the occasional wrong note could be subtly pushed into place, and it wouldn't require you to do another take and then try to splice together a great master recording. But even all of that is getting way ahead of ourselves. To understand the history of autotune, we must first learn about reflection seismology as well as the oil industry. And I am being serious, I'm not making a joke about this. As it turns out, reflection seismology has a lot to do with our story because the man who would go on to found the company that would create autotune was a doctor, Andy Hildebrand, who had made a career in using sound and complex mathematical calculations to help oil companies, namely Exon, locate oil deposits underground. So reflection seismology is in some ways similar to sonar. So with a sonar system, you would beam out pulses of sound waves. Typically we talk about this in water, right, Like using sonar on a boat or on a submarine, that kind of thing. You would pulse out these sound waves and those soundwaves travel outward from the source from the speaker. Essentially the transmitter, and if there's something solid in the way of those sound waves, well, the sound waves that hit that solid object, they're going to reflect back toward the source. They'll become an echo. This is what we get when we hear an echo. If you are ever in a place where you make a loud noise, then you hear the echo. It's because the sound waves have traveled out from you, bounced off something and came back to you. Well, if you if you measure the amount of time it took for a sound to leave you and then reflect off something else and come back to you, you can figure out how far away you are from that thing, right, because sound is going to travel a specific speed away from you and then hit the thing and then travel back to you. So if you know how long it took, you can do some very simple math and figure out how far away that object is. So, for example, if you're on a ship and use sonar to measure the distance between you and the sea floor, you do a little math. Right, you have to divide by two because it took a certain amount of time to travel down then back up, And you have to know how fast sound travels through water. You have to have all these important bits of information in your mind when you do this, but then you can suss that out. You can say how deep the ocean floor is from the surface. That saves you the trouble of having to do it the really old fashioned way, which typically involved lowering a weight on the end of a knotted line, a knotted rope, and then you use the knots to keep count of how deep in the ocean you were. That's a sounding line. That's the other way to see how how far down the ocean floor is. But sonar made it way simpler, especially once you were able to build that math into the sonar workstations. Reflection seismology does something similar, but with seismic waves, and those are waves that pass through the earth, and we typically talk about seismic waves in connection with earthquakes or like volcanic eruptions that kind of thing, and in fact earthquakes were what inspired smarting Pants is to say, hey, if we made something that could you know, create a huge vibration through the earth, and something else that could detect those vibrations, and we were able to calculate how long it took for the instrumentation to pick up on the echoes of that initial vibration event, we might be able to figure out stuff that's actually underground. We could figure out what is underground without having to dig it up and see. Now, that's because the seismic waves will travel at different speeds depending upon the density of the material that they travel through. You've probably heard things like, you know, sound travels at a consistent speed. That's true, but that consistent speed is dependent upon the medium through which the sound is traveling. So sound travels at a different speed through the water than it does through the air or through solid objects. You know, vibrations travel at different speeds depending upon the medium. So at a very basic level, a seismic wave will travel at a constant rate through one kind of say rocky soil. But let's say there's a place underground where that rocky soil gives way to a different material says petroleum for example. Well, then the speed of those sound waves is going to change. Moreover, as the sound waves hit that barrier between one type of material and another, some of the sound waves are going to reflect off of that and become an echo. Some of the sound waves will continue to penetrate through the new material and through lots of observations. We gradually began to learn about the different rates at which a seismic wave will travel depending upon the medium it's traveling through, and if it hits something really solid like bedrock, it pretty much just echoes back. So here's how reflection seismology works. From a very high level. You set up sensitive equipment at different distances from a blast sight, and yeah, you're likely to use something like explosives or maybe a really powerful air gun. It has to be something that's going to give a real jolt to the ground in order to do this, because that's essentially what you're doing is creating like a very localized earthquake. So this vibration travels through the earth, and because you know how far away you've set up your measuring equipment from that blast site, you already have distance figured out, right. You know how far away it is from the original source of the vibration, and you measure the time it takes for your equipment to pick up the echoes from that particular vibration event. So you've got distance and now you have time. Now you've got those variables sorted, so you can start to work out what material is actually under the ground that produces this particular result. And by doing that, you're kind of like working backwards. You're using this information to draw conclusions about what's under there, and that's where you can start to make a determination as to whether or not you're standing on top of a Beverly hillbilly is like oil deposit, or maybe you're just on top of a bunch of rocks or whatever. Now, in order to do that what I just described, it's actually incredibly complicated. It involves an awful lot of calculations in math, and it's a lot of work. But then you have to think that drilling for oil is even more work. That's a huge endeavor. It costs a lot of time and money and effort to do it, and like if you drill in the wrong place, like that's a huge loss. So you want the best possible information before you select a drilling site, and reflection seismology is one way to obtain information and to help make a decision. So doctor Hildebrand was making a really good living out of this work, but companies like Exon were saving hundreds of millions of dollars through Hildebrand's approach of narrowing down potential drill sites, and Hildebrand thought, you know, I'm not doing badly. I'm making a decent living. But you know, Exon is making out like a bandit. They're saving like half a billion dollars a year or whatever using this technology. Maybe if I apply my knowledge and skill set in a company that I own, I might actually, you know, do better than just working for Exon. So Hildebrand left Exon in nineteen seventy nine and he founded a company called Landmark Graphics, which at first sounds like, you know, it's a company that makes computer graphics, which is not untrue, but that wasn't It wasn't just general graphics. This company was still rooted in the oil industry. Hildebrand's team developed and produced workstations that could take incoming seismic information from these these you know, soundings that they do and generate three dimensional seismic maps based upon the data. And again, it was incredibly complicated. You had to analyze so many different points of information in order to create this three dimensional representation of what's under the ground. But it worked and it made Hildebrand very successful. He stuck with it for a decade until nineteen eighty nine, whereupon he retired and he decided to return his attention to a different passion he had had since he was a kid, which was music. Now Hildebrand wasn't just a music fan, he was a musician. He had played flute professionally. He had been a studio musician for some time. He had paid his way through college partly by giving flute lessons to musicians, So he decided he would go back to school as a retiree and study composition and techniques. He attended Rice University to do this. While he was back in college, he encountered some newer technologies in the music space, like music samplers and synthesizers. So these were machines designed to take a sample of a sound like a flute, and then allow a keyboard musician to recreate those sounds on a synthesizer. The only thing is that Hildebrand thought they sounded terrible, and partly it was because there was a limitation on how much data a synthesizer could actually handle, so it couldn't really replicate sound naturally. The sound it replicated would be like a gross approximation of the original sound, So Hildebrand wasn't really impressed, but he thought that there was room for improvement, and he developed a technique to compress audio data so that synthesizers could more effectively handle information and make notes, to produce notes that sounded more natural and less synthetic. He released his software as a product called Infinity, and while this tool would revolutionize the orchestration process for stuff like film and television, it did not revolutionize doctor Hildebrand's bank account. He didn't actually see much of that success himself because what actually happened was other companies purchased copies of Infinity and then bundled it with their own audio processing tools, and then sold those audio processing packages to other people and companies, and it kind of cut Hildebrand out of the picture. So while others were benefiting from his work, he did not see that much success. It did, however, again have an enormous impact on orchestrations, like According to doctor Hildebrand, he was the reason why the Los Angeles Orchestra hit real hard times in the nineteen nineties because his tools allowed composers to sample various musical instruments and create a natural enough representation of those sounds to be able to create a synthetic orchestra that sounded more or less like a real one. So there was no need to go and hire a real orchestra to orchestrate your film or TV project. You could do it yourself. I've actually heard some some of my favorite music scores. When I listened closely, I can tell like, oh, that's not a real cellist. That's a synthesizer playing a sample of a cello that sounds almost, but not quite like the real thing. Anyway, we can thank doctor Hildebrand for that. I'll talk more about what we could thank doctor Hildebrand for, specifically auto tune, but first let's take a quick break so we could thank some other people, namely our sponsors. Will be right back. Okay, So before we left off, I was talking about how doctor Hildebrand had released a program called Infinity that improved the performance of synthesizers and samplers. But in nineteen ninety he decided to take an extra step. He founded a new company. He called it Antare's Audio Technology, and this would be his music company, his music technology company that would ultimately produce autotune. And he knew that technology was poised to make a huge impact on the music industry and already had been like, that's kind of the history of modern music is how technology has shaped it. But he knew we were on the brink of another revolution. He just wasn't exactly sure how that was going to manifest now. According to an article by Simon Reynolds, it's titled How Autotune Revolutionized the Sound of Popular Music, and it was published in Pitchfork, the actual birth of Hildebrand's idea for autotune grew out of a casual lunch with some of his friends and peers back in nineteen ninety five during a National Association of Music Merchants conference. So he's at this conference, he's meeting with other people in the music and technology spheres, and at this lunch, one of the attendees jokingly suggested that what Hildebrand should do next is develop a technology that would allow her to sing on key, like, can you make a box that lets me sing well? And while this was presented as a joke, ultimately Hildebrand would think, huh, could I do that now? According to Zachary Crockett's article, which is the Mathematical Genius of auto Tune, this one in price Anomics This wasn't like a light bulb moment where the moment this woman says the thing, Hildebrand immediately thinks, ah, that's what I shall do. Actually, it took like another six months before Hildebrand really kind of revisited the concept and thought, maybe there's something here. But in order to do that, he would have to develop a technology that could do a few things really well, all of which are a bit tricky. One is it would need to detect the pitch that someone was singing in. For example, if you're using it for vocals, and so you would need to be able to detect exactly the frequency that was being sung. You would need to then also be able to have a list of tones that were in the whatever key you were supposed to be singing it. So I don't want to get into music theory, because goodness knows, I don't know that much about it myself, and I would just mess things up. But you know, if you're singing in a specific key, there are particular tones that belong to that key. And often when we sing and we're a little off pitch, what we need is to be gently nudged a little up or a little down, a little sharp or a little flat in order to hit a semitone that belongs in that key. So it needs to also quote unquote know which tones are appropriate, and then it has to be able to digitally alter the incoming pitch the actual sung note, and then guide it to match that of a target note. Now, ultimately that all sounds like a pretty simple idea, but in reality to achieve this it was incredibly complex. Ultimately, also, the toolould need to work in real time for live performances. Like it's one thing to have this for the studio, right, because even if you don't have an automatic, you could have a tool where an engineer could fiddle with some controls and gently alter the pitch of a performance to get it closer to being where it needs to be. It would be preferable to have that automated so that you don't have to go through there and do the manual process. But even so, like in a recording setting, you don't have to have it be real time necessarily, but if you're doing a live performance, you do have to have a real time. If someone's up there singing and they just hit a flat note when they're not supposed to, that could really be a memorable moment and not in a great way. So having a tool that could gently account for that and fix it in real time would be really helpful. But this would mean that this tool would have to be able to process a huge amount of sound data extremely quickly to make millisecond decisions like split millisecond decisions relating to how to shape a note moment by moment. Now it does help if we also think of sound in terms of mathematics. We describe sound in different ways, right, But some of those relate specifically to how sound looks to us. If we plot sound on like a wave chart, right. For example, sounds can be really loud or they can be really quiet, and that is volume, But it can also relate to amplitude. When you think of a sound wave. The amplitude of a sound wave describes how tall those peaks are or how low the valleys are. The distance between the furthest point of a peak or valley and the zero line. That's your amplitude. But we also describe sound in terms of pitch or frequencies. Higher frequencies correspond to higher pitches, And if we plot a sound wave, let's say that we plot it so that the x axis is a demarcation of time, so we have one second listed there, like the x axis is one second. If there's one wave that we draw so that the wave begins at the zero point and ends at the one second point, then we have a one hurtz sound wave. A hurtz is just a measurement a frequency. It refers to one cycle per second. So if a wave is one hurts, it means it takes one second for one of those sound waves to fully pass a given point where you're measuring the sound waves, right, If two waves pass that point within one second, then you're talking about two hurts, you know. Just so that we know, the typical human hearing range is anywhere between twenty and twenty thousand hurts. So one or two hurts sound We wouldn't even perceive it, at least not as sound. If it was a great enough amplitude, you could potentially perceive it as vibration, but you wouldn't feel it, you wouldn't hear it. But between twenty and twenty thousand hurts, that falls into the typical range of human hearing. Of course, as we get older, we start to lose the ability to hear those higher frequencies. These days, I think my hearing tops out around sixteen to seventeen thousand hurts somewhere around there. Like once you get beyond that, I don't hear anything, whereas younger people could hear it. Anyway, Hildebrand was working with music on this mathematical level. He was analyzing music to recognize where the frequencies were and where they should be, and to then shape the sound wave so that it would fit what the ideal would be where it would be on key. He was not the first person to attempt to do this, however, Earlier engineers had largely abandoned the quest because the signal processing and statistical analysis needs were so high. They were so extreme that you would need a supercomputer dedicated to the task to be able to do it. There's just too much data to process in too little time to be able to do anything meaningful with it. Hildebrand determined that yeah, to fully analyze music, you would have to run thousands or millions of calculations, but many of those calculations were actually redundant at the end of the day, and eliminating the redundancy would not affect the quality of the outcome, and so in his words he quote changed a million multiply ads into just four. It was a trick, a mathematical trick. End quote. That's ron the article I mentioned earlier by Zachary Crockett. So yeah, in prisonomics, pretty phenomenal that he was able to recognize that ultimately he just needed these four processes to really be able to zero in on pitch correction. So Hildebrand developed the autotune technology in nineteen ninety six. He actually used to customized Mac computer or specialized Mac computer as the way I've seen it explained. I don't know in what way it was specialized. I just know it was a Mac. And he brought his software to the next National Association of Music Merchant's conference, if you remember, that was the same conference where one of his lunch companions had inspired the idea for autotune in the first place. To say that he felt interest in his product at this conference is really under selling it, and it's understandable why. So let's talk about the process of creating a master recording for a song. If you want to get a perfect take of a song, where this is the master recording, this is what you want to use in order to you know, create your album. You can't just hope that everything lines up when you hit record and that everyone is playing seamlessly together and no one makes a mistake. Invariably something is going to be off. Maybe one of the musicians is lagging behind the others and it might not even be detectable at first, but upon closer examination you're like, ooh, you came in late, or you came in too early or whatever. Or the drummer is not keeping perfect time, whatever it may be. Maybe someone hits a wrong note, either while playing an instrument or while singing, or maybe both. But what it means for engineers is that they'll need to get another take where that mistake isn't there, and they'll probably need another take and another take, And if you want the perfect performance, this could mean recording the same track dozens or gosh even hundreds of times and then slowly picking apart each recording in order to piece together a perfect edit. And that alone is hard because just lining up the different takes isn't always the easiest thing to do. You don't always have a seamless point where you could line up take one would take two. Like again, if the band is playing at a slightly different pace in the second take. You can't easily line up the two different ones to you know, even if like one had an accident and the other one didn't, you can't necessarily put them together to make the perfect recording. So this is a really laborious, time consuming and expensive process. Expensive because studio time is limited, so expensive. Hildebrand's invention would take a ton of that effort off the table, at least for vocals, because rather than re recording a billion times, you could get maybe just one good take, one decent take even and then use pitch correction for any little flubs that might have found their way in during the recording process. So it was a huge time saver, and time is money. So immediately studios recognized the value of Hildebrand's product and they rushed to get in on that, and the tool absolutely revolutionized the recording industry. Studios that incorporated auto tune were able to work much faster than their competitors. They were able to cycle clients in and out of their studios more quickly. That meant getting more work done and more money coming in, and efficiency skyrocketed. So studios that were not on the auto tuned train soon found themselves getting out competed, and they ended up adopting the technology as well, because it was either adopted or go out of business. It also wasn't enough to just be able to change the pitch of a note. Auto tune would also have to be able to adjust that pitch on a sliding scale of rapidity. That is, the sound would be unnatural if you were to correct a note instantaneously. It would be the effect that we associate with autotune, that robotic effect. That's if you were to change the pitch correction super fast. You don't want to do that if you want the tool to remain unnoticed. So, particularly for stuff like slow ballads, you would not have a more gradual approach to correcting a pitch. So Hildebrand wanted a tool that would let users determine how quickly the note would get nudged to the correct pitch, and the scale essentially went from zero to ten. The higher settings would have longer adjustment times, so for a really slow song, you might go with a nine or a ten to let the note find its way to the right pitch more gradually. Faster songs like rock and roll type stuff or a rap or R and B. You might require a lower setting, like fast rock songs, you might need a two, three, or maybe even down to a one. The zero setting. Really, Hildebrand just added that for kicks, So essentially the software would immediately correct the pitch upon detecting an incoming signal. And this sounded weird and unnatural, and it was obvious that something was going on. So this was more for fun than an intent to create a new tool for musicians. But it turned out that's exactly what autotune was really destined for, to become a tool for a process called pitch quantization. But again, that wasn't what Hildebrand set out to do. In fact, ecquate to that Pitchfork article I mentioned earlier, the idea here was to aim for perfection, at least in terms of being in the right key and on pitch. That imperfections would somehow interfere within a momotional connection to the music, and you want that music to be perfect so that you can have that emotional impact. Now, personally, I disagree with that take. Some of my favorite recordings are with artists who have imperfect voices. They weren't screeching or catterwalling. It wasn't like it was unpleasant to listen to them, but they aren't pitch perfect either, and to me, that adds a lot of character and emotion. So as an example, Warren Zevon, who you know did the song where Wolves of London and tons of other stuff. I mean, prolific musician who tragically passed away several years ago. He has a great cover of the song back in the High Life Again and which is a pretty cheesy song, but Warren Zevon's cover is really emotional. It's great, and it's a little bit raw, and to me it resonates far more than a note perfect performance would have. But I do understand where hill le Brand and his team were coming from. You know, if you if you have a take from a recording session that is almost but not quite right, maybe there was a transition where the wrong note came out, or you know, just a moment where it took an artist a little bit longer to slide to find the right pitch. A tool that could smooth things out a little while not remaining you know, noticeable, while you know, slipping under the radar. That could prevent listeners from being distracted by something that was unintentional. But what if you took that tool that was meant to fix errors and used it to create unintended effects. That's what we're going to talk about when we come back from this quick break. So we talked about how auto tune was meant to fix little imperfections in music recordings and live performance. But as I mentioned, if you had that setting set to zero so that it would instantaneously attempt to correct pitches, then you could create an almost robotic vocalization. So instead of shying away from the artificial sounds that could come out if you were to use it improperly, you leaned into it. That's what happened in nineteen ninety eight with Sher's song Believe, saying that dolls zero would create the robotic like effect, which in this case was the goal in the first place, and that song was a smash success. I couldn't stand it, and it was everywhere. I couldn't stand it, not because of the auto tune. I just didn't vibe with the song. No no shade on chare phenomenal artist, you know, incredibly talented, Just that song didn't jibe with me and The interesting thing was that this huge success not only pulled the curtain back on a tool that was meant to correct little mistakes and thus create a whole conversation around whether or not artists were quote unquote cheating by using it, but it launched a whole new way to create music in the first place. I personally think the artist who is most associated with auto twoun is one who adopted the technology and made it an intrinsic part of his brand. That would be te Pain. He came to the party a little bit late. He became interested in autotune around two thousand and four, and he wasn't looking for something to help compensate for his singing ability, because he actually sings very well. But he liked the thought of the technology that would set him apart from other artists, and he could forge a vocal identity using this tool to create a sound that no one else was really embracing at that point. So he jumped wholeheartedly into autotune, and he made liberal use of the technology and achieved tremendous success along the way, selling like Platinum records by using this technology. His love of the software led to an official partnership with Hildebrand's company for a few years, and Tarees licensed the technology to tee Pain for an app called I Am te Pain, which you could use to do autotuned right there on your smartphone three dollars initially, and it was downloaded by millions of users that generated quite a lot of revenue just on its own. Now, eventually t Pain and Antarees parted ways, and t Pain ultimately partnered with a different pitch correction company called Isotope. The t Pain story also led to a lawsuit against Antarees, and Antaries filed a countersuit against te Pain, and ultimately the whole thing was settled out of court and everyone signed an NDA. So I have no details about, you know, how that shook out in the end, but it was one of those things where it was kind of a smudge on the Antarees name at least it was awkward, right. But a much larger threat to Hildebrand's company was Apple. Apple had purchased a German company called Emagic. Emagic also had a pitch correction tool. In fact, it was a pitch correction tool that, according to Hildebrand, essentially copied autotune technology. This was possible because Antaris had failed to protect its German patent properly, and so Emagic was able to appropriate that technology or copy that technology without fear of legal recourse. So then Apple acquires Emagic, which means Apple is then able to incorporate Emagic's technology into their own products, including their own sound editing software. And this meant that autotune effectively got incorporated into Apple software without having to license the technology from Antaries, because again they got it by acquiring this German company. Now, Antaris could technically have still sued Apple. There's no guarantee that they would have won, but they could have sued them. However, Hildebrand explained that they didn't really have that option because Apple has enorm mislead deep pockets. Apple is just an incredibly rich company, and Apple could easily just outweight Antaries in the legal system, while Antaries would drain its resources trying to sue Apple. So even if Antaris was in the right of it, even if they would have won a judgment against Apple, the chance was that Antari's would go out of business just trying to pay for all the legal fees for the whole battle in the first place. So Ultimately, Antari's didn't go after Apple. It just it would have been a death sentence. Culturally, autotune began to face resistance in the late two thousands. Some artists expressed disdain for the technology, going so far as to say it ruined Western music. This was partly due to an oversaturation problem. The success of Tea Pain, as well as the earlier instances of autotune, inspired countless others to embrace the technology while not necessarily doing very much else to differentiate themselves from other artists. In other words, they were kind of leaning on it as a crutch or a gimmick. So there was a glut of auto tune robotic voiced vocals and music in the early to mid two thousands, and by the late two thousands some folks were absolutely fed up with this and there was a backlash. It actually kind of reminds me about how people began to turn against disco in the nineteen seventies, and that in some ways the punk rock movement was partly a reaction to disco or a rejection of disco. I would only say partly because punk rock also has its roots in glam rock, and I think glam rock also kind of helped inspire disco. So it's a complicated set of relationships, as you might say on Facebook. But bands like Death Kem for Cuti actually actively spoke out against autotune. So again, some artists were arguing that autotune was being used by people to compensate for a lack of ability, So they're kind of casting shade on fellow artists saying, well, yeah, they have to use autotune because they can't sing, or others would say like it was making music less genuine and sincere, like less human because it was going through this digital processing process. Jay Z famously released a song titled Death of Autotune in two thousand and nine, the same year when our original Tech Stuff episode about autotune came out. As you might imagine, jay Z's song had some pretty strong opinions about the technology inside of it. It resonated enough to win him a Grammy for it, so other people agreed. But despite all that backlash, autotune continues to be a thing. It did not, in fact die. It's been incorporated into software and digital audio workstations. It and similar pitch manipulation technologies are often found in everything from professional audio engineering software suites to free programs that you can download online. So, for example, I sometimes use a program called Audacity, and Audacity has an option under its effects where I can manually adjust the pitch of a recorded piece of audio. I can set what the pitch should be. Now that's not autotune, right, because by definition I'm not using an auto feature. I'm manually changing the pitch, but it's using similar approaches to get an effect. I've actually even made use of that tool while I was editing my friend Shay's podcast, Kadi Womple with the Shadow People. Shae does nearly all the voices on that show. I've actually voiced two characters on that show. So if you're eager to hear other output from me, that's not a technology podcast, go listen to Kadi Womple with the Shadow People. I voice a couple of characters on that, but I edit the show, so I use pitch adjustment tools in order to make some of Shay's voices sound like different people. So it's still herb doing the voice, but I digitally manipulate the voice to give certain characters their own distinct sound. It's pretty neat stuff. I have no idea what it would sound like if I actually used an auto tuned tool. That probably would sound very different, But I have a lot of fun playing with these pitch manipulation tools. Now, to get more into the cultural and social impact of autotune, I highly recommend that article in Pitchfork by Simon Reynolds. Again, that's titled how Autotune Revolutionized the sound of Popular Music. It's a long form article, it's well worth your time to read it. As a Zachary Crockett's article that I mentioned earlier, both of those are great articles about autotune and not just the technology, but it's impact on music in general and society and culture as well. And Reynolds goes into much deeper detail about how the technology has had an impact on the recording industry and the backlash that came out as a result of that, as well as sort of a counter movement against autotune. So check those out. They are well worth your time. And I could go on, but really I feel like those articles do a much better job than I would of describing all of that, So check those out when you have some time. That's it for today. I hope all of you out there are doing well, and I will talk to you again really soon. Tech Stuff is an iHeartRadio production. For more podcasts from iHeartRadio, visit the iHeartRadio app, Apple Podcasts, or wherever you listen to your favorite shows.

In 1 playlist(s)

  1. TechStuff

    2,447 clip(s)

TechStuff

TechStuff is getting a system update. Everything you love about TechStuff now twice the bandwidth wi 
Social links
Follow podcast
Recent clips
Browse 2,444 clip(s)