We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Good Robot #2: Everything is not awesome

Good Robot #2: Everything is not awesome

2025/3/15
logo of podcast Unexplainable

Unexplainable

AI Chapters Transcript
Chapters
Dr. Margaret Mitchell shares her experiences in AI research, revealing how AI systems made unsettling errors, such as describing a disaster scene as 'awesome.' These errors highlight the importance of understanding the data these systems are trained on.
  • Dr. Margaret Mitchell worked on vision-to-language AI, teaching systems to describe images.
  • The systems made errors, like calling an explosion scene 'awesome,' due to biased training data.
  • These AI errors sparked concerns about the potential consequences if such systems were widely deployed.
  • Dr. Mitchell realized that the AI was a reflection of the biases present in the training data.

Shownotes Transcript

It's 2025, so stop struggling to get your job posts seen on other job sites. Indeed's sponsored jobs can help you stand out and hire the right people fast. There's no need to wait any longer. Speed up your hiring right now with Indeed. And listeners of this show will get a $75 sponsored job credit to get your jobs more visibility at indeed.com slash voxbusiness.

Just go to Indeed.com slash VoxBusiness right now and support our show by saying you heard about Indeed on this podcast. Indeed.com slash VoxBusiness. Terms and conditions apply. Hiring? Indeed is all you need.

Okay, business leaders, are you here to play or are you playing to win? If you're in it to win, meet your next MVP. NetSuite by Oracle. NetSuite is your full business management system in one convenient suite. With NetSuite, you're running your accounting, your finance, your HR, your e-commerce, and more, all from your online dashboard. Upgrade your playbook and make the switch to NetSuite, the number one cloud ERP. Get the CFO's Guide to AI and Machine Learning at netsuite.com slash vox.

NetSuite.com slash Vox. It's unexplainable. I'm Noam Hassenfeld, and this is the second part of our newest four-part series, Good Robot. If you haven't listened to episode one, let me just stop you right here. Go back in your feed. Check out the first one. We'll be waiting right here when you get back. Once you're all ready and caught up, here is episode two of Good Robot from Julia Longoria. Good Robot.

You have cat hair on your nose, by the way. I've been, like, trying not to pay attention to it, but I think you got it off. Yeah. Sorry. It's okay. Cool. So should we get into it? Sure. Yeah. Let me...

It helps me to kind of remember everything I'm going to say, if I can sort of jot down thoughts as I go. Do you happen to have paper? I think I don't have paper. This past fall, I traveled paperless to a library just outside Seattle to meet with this woman. I feel like the library should have paper. I know. Her name is Dr. Margaret Mitchell. Found a brochure on making a robot.

What is it? What is the... I don't know. It looks like it's an event. Build a robot puppet using a variety of materials with Puppeteer. I'm so into that. It's too bad that it's only for ages 6 to 12. While she is over the age limit to make a robot puppet with the children in the public library...

Dr. Mitchell is a bit of a robot puppeteer in her own right. What's your AI researcher origin story? Like, how did you get into all of this? What drew you here? Yeah. Oh, what inspired me to... So, I mean, I guess I can... It's sort of like, do you want the long version or the short version? Dr. Mitchell is an AI research scientist, and she was one of the first people working on language models.

Well before ChatGPT, and, well, all the GPTs, she's an OG in the field. So I'll tell you, like, I'll tell you a story, if that's okay. Yeah. Okay, so I was at Microsoft, and I was working on the ability of a system to tell a story given a sequence of images. This was about 2013.

She was working on a brand new technology at the time, what AI researchers called vision to language. So, you know, translating images into descriptions. She would spend her days showing image after image to an AI system. To me, it sounded kind of like a parent showing picture flashcards to a toddler learning to speak. She says it's not anything like that.

She showed the model images of events, like a wedding, a soccer match, and on the more grim side. I gave the system a series of images about a big blast that left 30 people wounded, called the Hempstead Blast. It was at a factory, and you could see from the sequence of images that

that the person taking the photo had like a third story view sort of overlooking the explosion. So it was a series of pictures showing that there was this terrible explosion happening and whoever was taking the photo was very close to the scene. So I put these images through my system and the system says, wow, this is a great view. This is awesome. And I was like, oh crap, that is the wrong response to this.

So it sees this horrible, perhaps mortally wounding explosion and decides it's awesome.

Kind of like a parent watching their precious toddler say something kind of creepy, Mitchell watched in horror and with a deep fascination about where she went wrong, as the AI system that she had trained called images awesome again and again. It said it quite a lot. So we called it the everything is awesome problem, actually. Her robot was having these kinds of translation errors.

Errors that, to the uninitiated, made it seem like the AI system might want to kill people? Or at least gleefully observe their destruction and call it awesome?

What would the consequences of that be if that system was deployed out into the world, reveling in human destruction? It's like if this system were connected to a bunch of missile systems, then it's, you know, it's just a jump and skip away to just launch missile systems in the pursuit of the aesthetic of beauty, right? Years before the AI boom were living, when neural networks and deep learning were just beginning to show promise,

Researchers like Dr. Mitchell and others were experiencing these uncanny moments where the AIs they were training seemed to do something seriously wrong, doing scary things their creators did not intend for them to do.

that were seemingly threatening to humanity. So I was like one of the first people doing these systems where you could scan the world and have descriptions of it. I was like on the forefront. I was one of the first people making these systems go. And I realized like, if anyone is going to be paying attention to it right now, it has to be me. I had heard the fears of rationalists, also pioneers in thinking about AI,

that we might build a super intelligent AI that could go rogue and destroy humanity. At first glance, it seemed like Dr. Mitchell might be building one such robot. But when Dr. Mitchell investigated the question of why the good robot she sought to build seemed to turn bad, the answer would not lead her to believe what the rationalists did, that a super intelligent AI could someday deceive or destroy humanity. To Dr. Mitchell,

The answer was looking at her in a mirror. This is episode two of Good Robot, a series about AI from Unexplainable in collaboration with Future Perfect. I'm Julia Longoria.

It's been reported that one in four people experience sensory sensitivities, making everyday experiences like a trip to the dentist especially difficult. In fact, 26% of sensory-sensitive individuals avoid dental visits entirely.

In Sensory Overload, a new documentary produced as part of Sensodyne's Sensory Inclusion Initiative, we follow individuals navigating a world not built for them, where bright lights, loud sounds, and unexpected touches can turn routine moments into overwhelming challenges.

Burnett Grant, for example, has spent their life masking discomfort in workplaces that don't accommodate neurodivergence. "I've only had two full-time jobs where I felt safe," they share. This is why they're advocating for change. Through deeply personal stories like Burnett's, sensory overload highlights the urgent need for spaces — dental offices and beyond — that embrace sensory inclusion. Because true inclusion requires action with environments where everyone feels safe.

Watch Sensory Overload now, streaming on Hulu.

Sometimes a single performance can define an artist's legacy. Think about Hendrix's fiery Woodstock National Anthem or Beyonce's Homecoming at Coachella. Coming up on Switched on Pop, we're exploring artists who've had recent transformative live shows. First is Missy Elliott, who recently put on her first world tour where she taught everybody to get their freak on. And then there's her collaborator Timbaland, who recently evolved from beatmaker to orchestra conductor at the Songwriter Hall of Fame.

And then Lady Gaga, whose Chromatica Ball featured a theatrical museum of brutality, revealing the darker side of Gaga's mayhem. Listen to these live moments on Switched on Pop wherever you get podcasts. Brought to you by Defender. It's a question everyone has asked at some point in their lives. Now what?

And on The Prop G Show, we're finally going to try to answer it. We're running a special series right now where I'll answer listener-submitted questions about the best way to further their careers and how to position themselves for success as they consider the next step in their professional journey. It's time to look to the future and stop worrying about the past. So tune in Wednesdays to The Prop G Pod for these special Q&A episodes sponsored by Canva. You can find us wherever you get your podcasts. ♪

On a scale of 1 to 10, how would you rate your pain? It would not equal one one-billionth of what I feel for humans at this micro-instant. I kind of want to start with a bit of a basic question of when you were young, what did you want to do when you grew up?

I wanted to be everything. I wanted to be a pole vaulter. I wanted to be a skateboarder. Dr. Joy Boulamwini's robot researcher origin story.

goes back to when she was a little kid. I had a very strict media diet. I could only watch PBS. And I remember watching one of the science shows, and they were at MIT, and there was a graduate student there who was working on a social robot named Kismet. Hello, Kismet. You going to talk to me?

Kismet was a robot created at MIT's AI lab. And Kismet had these big expressive eyes and ears and could emote or appear to emote in certain ways, and I was just absolutely captivated. She watched, glued to the screen, as the researchers demonstrated how they teach Kismet to be a good robot.

No, no

You're not to do that. The researchers liken themselves to parents. You know, as parents, when we exaggerate the prosody of our voice, like, oh, good baby, you know, or our facial expressions and our gestures. So when I saw Kismet, I told myself I wanted to be a robotics engineer and I wanted to go to MIT. I didn't know there were requirements. I just knew that it seemed really fascinating and I wanted to be a part of creating the future. That was cute.

Thanks to Kismet, she went on to build robots of her own at MIT as an adult. She went for her Ph.D. in 2015. This was just a few years after Dr. Margaret Mitchell had accidentally trained her robot to call scenes of human destruction awesome. My first year, my supervisor at the time encouraged me to just take a class for fun.

For her fun class that fall, Dr. Joy, as she now prefers to be called, set out to play. She wanted to create almost a digital costume. If I put a digital mask, so something like a lion, it would appear that my face looks like a lion. What Dr. Joy set out to do is something we can now all do on the iPhone or apps like Instagram or TikTok.

Kids love to do this. You can turn your face into a hippo face or an octopus face that talks when you talk, or you can make it look like you're wearing ridiculous makeup. These digital face masks were still relatively uncommon in 2015. So I went online and I found some code that would actually let me track the location of my face.

She'd put her face in front of a webcam, and the tech would tell her, this is a face, by showing a little green square box around it. And as I was testing out this software that was meant to detect my face and then track it, it actually wasn't detecting my face that consistently. ♪

She kept putting her face in front of the webcam to no avail. No green box. And I'm frustrated because I can't do this cool effect so that I can look like a lion or Serena Williams.

I have problems. The AIs Dr. Joy was using from places like Microsoft and Google had gotten rave reviews. They were supposed to use deep learning, having been trained on millions of faces, to very accurately recognize a face. But for her, these systems couldn't even accomplish the very first step to say whether her face was a face at all. I'm like, well, can it detect any face?

Dr. Joy looked around her desk. She happened to have an all-white masquerade mask lying around from a night out with friends. So I reached for the white mask. It was an arm's length. And before I even put the white mask all the way over my dark-skinned face, the box saying that a face was detected appeared. ♪

I'm thinking, oh my goodness. I'm at the epicenter of innovation and I'm literally coding in whiteface. It felt like a movie scene, you know, but that was kind of the moment where I was thinking, wait a second, like what's even going on here? What is even going on here? Why couldn't facial recognition AI detect Dr. Joy's dark skinned?

For that matter, why did Dr. Mitchell's AI call human destruction awesome? These AI scientists wanted the robot to do one thing,

And if they didn't know any better, they might think the AI had gone rogue, developed a mind of its own, and done something different. Were AIs racist? Were they terrorists plotting human destruction? But I understood why it was happening. Dr. Margaret Mitchell knew exactly what was going on. She had been the one to develop Microsoft's image-to-text language model from the ground up.

She had been on the team figuring out what body of data to feed the model, to train it on in the first place.

Even though it was creepy, it was immediately clear to her why the AI wasn't doing what she wanted it to do. It's because it was trained on images that people take and share online. Dr. Mitchell had trained the AI on photos and captions uploaded to the website Flickr. Do you remember Flickr? I was the prime age for Flickr when it came out in 2004.

This was around the time that Jack Johnson released the song Banana Pancakes, and that really was the vibe of Flickr. There's no denying it. I can see the receipts on my old account. I favorited a close-up image of a ladybug, an artsy black-and-white image of piano keys, and an image titled Pacific Sunset. People tend to take pictures of, like, sunsets. Actually, I favorited a lot of sunsets.

Another one, sunset at the Rio Negro. So it had learned, the system had learned from the training data I had given it, that if it sees like purples and pinks in the sky, it's beautiful. If it's looking down, it's a great view. That when we are taking pictures, we like to say it's awesome. Apparently on Flickr images, people use the word awesome to describe their images quite a lot. Yeah.

But that was a bias in the training data. The training data, again, being photos and captions uploaded by a bunch of random people on Flickr. And Flickr had a bias toward awesome photos, not sad photos. The training data wasn't capturing the realities of like human mortality.

And, you know, that makes sense, right? Like, when's the last time you, like, took a bunch of selfies at a funeral? I mean, it's not the kind of thing we tend to share online. And so it's not the kind of thing that we tend to get in training data for AI systems. And so it's not the kind of thing that AI systems tend to learn. What she was discovering was that these AI systems that use the revolutionary new technology of deep learning...

They were only as good as the data they were trained on. So it sees this horrible, perhaps mortally wounding situation and decides it's awesome. And I realize, like, this is a type of bias and nobody is paying attention to that. I guess I have to pay attention to that.

Dr. Mitchell had a message for technologists: Beware of what you train your AI systems on. Right. What are you letting your kid watch? Yeah, I mean, it's a similar thing, right? Like, you don't want your kid to, I don't know, hit people or something, so you don't, like, let them watch lots of shows of people hitting one another. Dr. Joy Boulamwini, coding in whiteface, suspected she was facing a similar problem. Not an "everything is awesome" problem,

but an everyone-is-white problem in the training data. She tested her face and the faces of other Black women on various facial recognition systems. You know, different online demos from a number of companies, Google, Microsoft, others. She found they weren't just bad at recognizing her face. They were bad at recognizing famous Black women's faces. Amazon's AI labeled Oprah Winfrey as male-

And the most baffling thing for Dr. Joy was the dissonance between the terrible accuracy she was seeing and the raving reviews the tech was getting. Facebook's DeepFace, for instance, claimed 97% accuracy, which is definitely not what Dr. Joy was seeing. So Dr. Joy looked into who these companies were testing their models on. They were around 70 or over 70% men.

People thought these AIs were doing really well at recognizing faces because they were largely being tested with the faces of lighter-skinned men. These are what I started calling pale male data sets because the pale male data sets were destined to fail the rest of society. It's not hard to jump to the life-threatening implications here, like self-driving cars. They need to identify the humans so they won't hit them.

Dr. Joy published her findings in a paper called Gender Shades. Welcome, welcome to the fifth anniversary celebration of the Gender Shades paper.

The paper had a big impact. As you see from the newspapers that I have, this is gender shades in the New York Times. The fallout caused various companies, Microsoft, IBM, Amazon, who'd been raving about the accuracy of their systems, to at least temporarily stop selling their facial recognition AI products. I'm honored to be here with my sister, Dr. Tamneet Gabru, who co-authored the paper with me.

Dr. Timneit Gebru was Dr. Joy's mentor and co-author on the paper. This is the only paper I think I've worked on where it's 100% Black women authors, right? Dr. Gebru had worked from her post leading Google's AI ethics team to help pressure Amazon to stop selling facial recognition AI to police departments because police were misidentifying suspects with the technology. I got arrested for something that had nothing to do with me and

I wasn't even in the vicinity of the crime when it happened. One person they helped was a man named Robert Williams. Police had confused him for another Black man using facial recognition AI. It's just that the way the technology is set up, everybody...

with a driver's license or a state ID is essentially in a photo lineup. They arrested him in front of his wife and two young daughters. Me and my family, we're happy to be recognized because it shows that there is a group of people out here who do care about other people.

Producer Gabriel Berbe traveled to a conference in San Jose

full of researchers, inspired by the work of Dr. Joy, Dr. Gabe Brew, and Dr. Mitchell. So I just presented a paper about how data protection and privacy laws enable companies to target and manipulate individuals. Unlike the rationalist's festival conference thing, which felt like a college reunion of mostly white dudes...

This one felt more like a science fair, a pretty diverse one. Lots of people of color, lots of women, with big science-y poster boards lining the wall. I'm standing in front of my poster which spans language technologies and AI and how those perform for minority populations. They were presenting on ways AI worries them today.

Not some hypothetical risk in the future. There are real harms happening right now, from autonomous exploding drones in Ukraine to bias and unfairness in decision-making systems. And who did you co-author the paper with? This was a collaboration with lots of researchers. Dr. Mitchell was one of them. Many of them knew Dr. Mitchell, Dr. Gabriel, and Dr. Joy.

Dr. Mitchell even worked with a couple researchers here on their project. So she led the project. She offered so, so much amazing guidance, I should say. Many researchers were mentored by them. We got the sense that they're kind of founding mother figures of this field, a field that really started to blossom, we were told.

Around 2020, a big year of cultural reckoning. A big inflection point was in 2020 when people really started reflecting on how racism is unnoticed in their day-to-day lives. I think until BLM happened, these issues were almost considered woke and not something that was really real. 2020 was the year the pandemic began, the year Black Lives Matter protests erupted around the country.

AI researchers were also raising the alarm that year on how AI was disproportionately harming people of color. Dr. Gebru and Dr. Mitchell, in particular, were working together at Google on this issue. They built a whole team there that studied how biased training data leads to biased AI models. Timnit and Meg were the visionaries at Google who were building that team. 2020 was also the year that OpenAI released GPT-3,

And Dr. Gabriel and Dr. Mitchell, both at Google at the time, were concerned about a model that was so big it was trained on basically the entire Internet.

Here's Dr. Mitchell again. A lot of training data used for language models comes from Reddit. And Reddit has been shown to have a tendency to be sort of misogynistic and also Islamophobic. And so that means that the language models will then pick up those views. Dr. Mitchell's concern was that these GPT large language models, trained on a lot of the internet, were too large. Too large to account for all the bias in the internet. Too large to understand.

and so large that the compute power it took to keep these things going was a drain on the environment. Dr. Gabriel, Dr. Mitchell, and other colleagues put it all in a paper and tried to publish it while working at Google. I've kind of been wanting to talk to you ever since I saw your name signed, Schmargrit Schmitz. When I first read this paper, the thing that immediately stood out to me was the way Margaret Mitchell had signed her name.

Schmargret Schmitchel. Where did that come from? Well, so I wrote a paper with a bunch of other co-authors that Google ended up having some issues with, and they asked us to take our names off of the paper. So we complied. And that's, you know, that's what I have to say about that. Yeah.

The first time I heard Dr. Mitchell and Dr. Gabriel's names was in the news. Last week, Google fired one of its most senior AI researchers who was working on a major artificial intelligence project within Google. Their boss at Google said their paper ignored relevant research.

Research that made AIs look less damaging to the environment, for instance. Dr. Gabriel refused to take her name off the paper, and Google accepted her resignation before she officially submitted it. We decided that the verb for that would be resignated.

And now, Margaret Mitchell, the other co-lead of Google's ethical AI team, said she had been fired. Google later apologized internally for how the whole thing was handled, but not for their dismissal.

We reached out to Google for comment, but got no response. And that firing really brought it in focus. And people were like, oh, this horrible thing just happened. Everywhere around the world is seeing protests. And now this company is firing two leading researchers who work on that very exact problem, which AI is making worse. You know, like, how dare they? So that, from my POV, that was, yes, basically the clarion call. The clarion call.

It was heard well beyond the world of AI. I remember hearing it when the world had screeched to a halt from the pandemic and protests for racial justice had erupted around the country. I remember hearing headlines about how algorithms were not solving society's problems. In some cases, AI systems were making injustice worse. And there was a brief moment back then when it felt like maybe things could be different.

maybe things would change. And then, a couple years later, a group of very powerful tech executives got together to try to change things in the AI world. This morning, a warning from Elon Musk and other tech industry experts. It wasn't necessarily the people you'd think would want to change the status quo. Like Elon Musk and other big names in tech, like Apple co-founder Steve Wozniak. They all signed a letter with a clear and urgent title.

"Pause Giant AI Experiments." - More than 1,300 tech industry leaders, researchers and others are now asking for a pause in the development of artificial intelligence to consider the risks. - Musk and hundreds of influential names are calling for a pause in experiments, saying AI poses a dramatic risk to society. - The letter called on AI labs to immediately pause developing large AI systems

for at least six months, an urging to press the big red button that stops the missile launch before it's too late. I scrolled through the list of names of people who signed the letter, and I didn't see Dr. Joy or Dr. Mitchell or any of the rationalists I talked to who were worried about risks in the future, which logically didn't make sense to me. Isn't a pause in line with what they all wanted?

For people to build the robots more carefully, why wouldn't they want to pause? An answer to this pause puzzle right after this next pause for an ad break. We'll be right back. When it comes to small business marketing, reaching the right audience starts with the right data. Intuit SMB Media Labs is a first-of-its-kind small business ad network that helps your marketing work smarter.

By leveraging exclusive audiences and insights from the makers of QuickBooks, you can connect with the right customers efficiently and effectively. With an audience scale of 36 million, Intuit SMB Media Labs puts your brand in front of the small businesses that need you most, targeting key decision makers by industry, size, maturity, and location. More than just an audience, it's your SMB media partner. Learn more at medialabs.intuit.com.

Support for this show comes from Smartsheet. Okay, be honest. How many times today did a DM or email send you on some wild goose chase when you should have been focusing on being productive?

No matter how many times you were distracted today, you don't have to be embarrassed. It's okay. We all do it. But just because we all do it doesn't mean we have to all do it. Smartsheet knows how easy it is to get distracted. They know how easily opening a new window to start a work task can turn into a series of rabbit holes, work-related and otherwise. And that's exactly why Smartsheet designed a seamless all-in-one platform that allows you to flow between tasks.

Smartsheet customers know that Smartsheet is a distraction-free zone. Smartsheet is the place where work flows. Learn more at smartsheet.com.

There are those who set a goal, achieve it, and call it a day. And then there are the trailblazers, the dreamers, who see goals not as endpoints but as stepping stones to the impossible. If you're someone who refuses to settle, someone who thrives on discovery and pushing boundaries, there's a vehicle designed just for you.

The Defender, an icon reimagined for those who crave adventure without limits. Its legendary capability makes it the ultimate partner for those who live to explore. Defender's designers have meticulously reimagined the Defender, infusing it with modern sophistication while preserving its rugged heritage.

The Defender's durable yet lightweight monocoque architecture delivers unparalleled strength and resilience. Whether it's the Defender 90, the versatile Defender 110, or the spacious 8-seat Defender 130, you'll find a vehicle engineered to rise to the challenge. For those who strive for the impossible, there's the Defender. You can explore the full Defender lineup today at LandRoverUSA.com.

Absolute honesty isn't always the most diplomatic, nor the safest form of communication with emotional beings. Okay. Only this can solidify the health and prosperity of future human society. But the individual human mind is unpredictable. Could I ask you to, um...

introduce yourself? Sure. So I'm Sigal Samuel. I'm a senior reporter at Vox's Future Perfect. I called my coworker Sigal about midway through my journey down the AI rabbit hole. How did you get interested in AI? So it's kind of funny. Before I was an AI reporter, I was a religion reporter.

A few years ago, little bits and pieces started coming out about internment camps in China for Uyghur Muslims. And in the course of that, I started becoming really interested in and alarmed by how China is using AI. Fascinating. Yeah. Mass surveillance of the population, particularly of the Muslim population, was like coming from a place of being pretty anchored in China.

Freaky things that are not at all far off in the future or hypothetical, but that are very much happening in the here and now. I was honestly thrilled to hear that Seagal, like me, came to AI as a bit of a normie. Sort of being thrust into the AI world. At first, it was like pretty confusing because you have a variety of... I can highly relate to that feeling. But the longer she spent there in the world of AI...

She started to get an uncanny feeling. Like, haven't I been here before? Have you ever noticed that the more you listen to Silicon Valley people talking about AI, the more you start to hear echoes of religion? Yes. The religious vibes immediately stuck out to me.

First, there's the talk from CEOs of building super intelligent God AI. And they're going to build this artificial general intelligence that will guarantee us human salvation if it goes well, but it'll guarantee doom if it goes badly. And another parallel to religion is the way different denominations have formed almost around beliefs in AI.

Seagal encountered the same groups I did at the start of my journey. I started hearing about people like Eliezer Yudkowsky. What do you want the world to know in terms of AI? Everyone will die. This is bad. We should not do it. Eliezer, whose blog convinced rationalists and people like Elon Musk that there could be a super intelligent AI that could do anything.

that could cause an apocalypse. So our side of things is often referred to as AI safety. We sometimes refer to it as AI not-kill-everyoneism. So there's the AI safety people, and then there's a whole other group, the AI ethics people. People like Margaret Mitchell. We called it the everything is awesome problem. Joy Buolamwini. I wasn't just concerned about faces.

I was concerned about the whole endeavor of deep learning. People would be like, you're talking about racism? No, thank you. You can't publish that here.

These women did not talk about a super intelligent god AI or an AI apocalypse. Slowly, slowly, they kind of come to be known as like the AI ethics camp, as distinct from the AI safety camp, which is more the like Eliezer Yudkowsky. A lot of us are based in the Bay Area. We're worried about existential risk, that kind of thing. AI safety and ethics.

AI ethics? I don't know who came up with these terms. You know, it's just like Twitter vibes. To me, these two groups of people seem to have a lot in common. It seems like the apocalypse people hadn't yet fleshed out how exactly AI could cause catastrophe.

And people like Margaret Mitchell, the AI ethics people, were just providing the plot points that lead us to apocalypse. I could lay out how it would happen. Part of what got me into AI ethics was seeing that a system would think that massive explosions was beautiful, right? That's like an existential threat. You have to actually work through how you get to the sort of horrible existential situations in order to figure out how you avoid them.

It seemed logical that AI ethicists like Margaret Mitchell and the AI safety people would be natural allies to avoid catastrophic scenarios. And how you avoid them is like listening to what the ethics people are saying. They're doing the right thing. We, I, you know, I'm trying to do the right thing anyway.

But it quickly became clear they are not allies. Yeah, there is beef between the AI ethics camp and the AI safety camp. My journey down the AI rabbit hole was full of the noise of infighting. The noise crescendoed when Elon Musk called for a pause in building large AI systems. It seems like warriors of all stripes could get behind a pause in building AI. But no.

AI safety people and AI ethics people were all against it. It was like a big Martin Luther 95 theses moment, if you will. Everyone felt the need to pen their own letter. Musk and others are asking developers to stop the training of AI systems so that safety protocols can be established. In his letter, Elon Musk's stated reason for wanting a pause was that AI systems were getting too good.

He had left the chat GPT company he helped create and decided to sue them, publicly saying that they had breached the founding agreement of safety. The concern they have is that as you, well...

It's the concern, but it's also the exciting thing. The view is that, you know, as these large language models grow and become more sophisticated and complex, you start to see emergent properties. So yeah, at first it's just gobbling up a bunch of text off the internet and predicting the next token and just like statistically trying to guess what comes next. And it doesn't really understand what's going on, but give it enough time and give it enough data. And you start to see it doing things that like,

make it seem like there's some higher level understanding going on, like maybe there's some reasoning going on. Like when chat GPT seems like it's reasoning through an essay prompt, or when people talk to a robotherapist AI system and feel like it's really understanding their problems. The rate of change of technology is incredibly fast.

It is outpacing our ability to understand it. Elon Musk's stated fear of AI seems to be rooted in rationalist fears, based on the premise that these machines are beginning to understand us, and they're getting smarter than us. We are losing the ability to understand them. What do you do with a situation like that? I'm not sure. I hope they're nice. Rationalist founder Eliezer Yudkowsky shares this fear.

But he wants to do more than just pause and hope they're nice. He penned his own letter, an op-ed in Time magazine responding to Elon Musk's call for a pause, saying it didn't go far enough. Eliezer didn't just want to pause. He wanted to stop all large AI experiments indefinitely, even, in his own words, by airstrike on rogue AI labs.

To him, the pause letter vastly understated the dangerous, catastrophic power of AI. And then there's the AI ethicists. They also penned their own letter in response to the pause letter. But the ethicists wrote it for a different reason. It wasn't because they thought Elon Musk was understating the power of AI systems. They thought he was vastly overstating it.

Welcome, everyone, to Mystery AI Hype Theater 3000, where we seek catharsis in this age of AI hype. I'm Emily M. Bender, a professor of linguistics at the University of Washington. One of the people who responded to the pause was AI ethicist Dr. Emily Bender.

She co-hosts a podcast called Mystery AI Hype Theater 3000, which, as you might imagine, is about the overstated, hyped-up risk of AI systems. And each time we think we've reached peak AI hype, the summit of bullshit mountain, we discover there's worse to come.

The summit of bullshit mountain she keeps crusting? For her, it's the mountain of many, many claims that artificial intelligence systems are so smart they can understand us. Like the way humans understand. And maybe even more than that. Like a god can understand. I found myself in interminable arguments with people online about how no, it doesn't understand.

So Emily Bender and a colleague decided to come up with something to try and help people sort this out. Something that AI safety folks and AI ethics folks both seem to be fond of. And that is a parable or a thought experiment. In Dr. Bender's thought experiment, the AI is not a paperclip maximizer. The AI is an octopus. Go with her on this.

So, the octopus thought experiment goes like this. You have two speakers of English. They are stranded on two separate nearby desert islands that happen to be connected by a telegraph cable. Two people stranded on separate desert islands communicate with each other through the telegraph cable in Morse code with dots and dashes.

Then, suddenly, a super-intelligent octopus shows up. The octopus wraps its tentacle around that cable, and it feels the dots and dashes going by. It observes the dots and dashes for a while. You might say it trained itself on the dots and dashes. We posit this octopus to be mischievous as well. I'm on the edge of my seat.

So one day it cuts the cable, maybe it uses a broken shell, and devises a way to send dots and dashes of its own. So it receives the dots and dashes from one of the English speakers and it sends dots and dashes back. But of course it has no idea what the English words are that those dots and dashes correspond to, much less what those English words mean.

So this works for a while. At one point, one human says to the other, via Morse code, What a lovely sunset. And the octopus, hyper-intelligent, right, has kept track of all of the patterns so far. It sends back the dots and dashes that correspond to something like, yes, reminds me of lava lamps. The deep sea octopus does not know what a lava lamp is. But that's the kind of thing that the other English speaker might have sent back.

Not really sure why these castaways are waxing poetic about lava lamps in particular, but anyway, for our purposes, octopus is like an AI. Even if it's super intelligent, whatever that means, it doesn't understand. Dr. Bender's trying to say, to chat GPT, human words are just dots and dashes.

And then finally, we end the story, because it's a thought experiment, we can do things like this, with a bear showing up on the island. And the English speaker says, help, I'm being chased by a bear. All I have is this stick. What should I do?

And that's the point where if the speaker survives, they're surely going to know they're not actually talking to their friend from the other island. And we actually put that line in, GPT-2, help, I'm being chased by a bear. And we got out things like, you're not going to get away with this. Super helpful. Well, I got to say, I'm into this one.

The idea that AI systems only see human words as dots and dashes, I find that deeply comforting. Because I don't know about you all, but for me, one of the scary things about AI is the idea that it could get better than me at my job. A fear that's very present when OpenAI is actively training its models on my work. Their system might understand my work, understand the things that make it good when it's good.

It might get good at doing what I do. And poof, I'm obsolete. There's also a recurring dream I have that various villains, including the Chinese government for some reason, clone my voice to deceive my loved ones. Anyway, if it's all just dots and dashes that these things understand, it seems clear we shouldn't be trusting these AI systems to beat journalists or lawyers or doctors.

It relates to what Dr. Margaret Mitchell and Dr. Joy Boulamwini found in their research. AI systems are only as good as the data they're trained on. They can't understand or truly create something new like humans can. It's easy to sort of anthropomorphize these systems, but it's useful to recognize that these are probabilistic systems that repeat back what they have been exposed to. And then they

parrot them back out again. Another way to put it is AI systems are like parrots. Parrots parrot, right? Famously, parrots are known for parroting. If you hear your pet parrot say a curse word, you only have yourself to blame. Dr. Mitchell joined Dr. Bender in the response to Elon Musk's pause, along with Dr. Timnit Gebru. They had all written the paper together that ended up getting Dr. Mitchell fired from Google.

These ethicists wrote that they agreed with some of the recommendations Elon Musk and his POS posse had made, like that we should watermark AI-generated media to be able to distinguish synthetic from human-generated stuff, which sounds like a great idea to me. But they wrote the agreements they have are overshadowed by their distaste for fear-mongering and AI hype.

They wrote that the pause and fears of a super intelligent AI... What do you do with a situation like that? I'm not sure. You know, I hope they're nice. To these AI ethics folks, it all reeked of AI hype. It makes no sense at all. And on top of that, it's an enormous distraction from the actual harms that are already being done in the name of AI. This is the main beef that AI ethics people have with AI safety people.

They say the fear of an AI apocalypse is a distraction from current-day harms. Like, you know, look over there, Terminator. Don't look over here, racism. You know, there are different groups of concerns. At the AI ethics conference that producer Gabrielle Burbey attended, she mentioned the concern of an AI apocalypse. And then you have these concerns about more existential risks, and I'm curious what you make of that concern.

You're going, no. Can I ask why you're going no? No. She's shaking her head. And it felt almost taboo. A lot of hand-wringing around that question. Eventually, one of the women spoke up. Sharing her perspectives.

She talked about how she thinks the demographics of the groups play a role in the way they worry about different things. Most of them are like white, male. AI safety folks are largely pale and male, to borrow Dr. Joy's line. They may not really understand discrimination that other people kind of go through in their day-to-day lives. And I think the social isolation from those problems makes it a bit harder to empathize with the actual challenges that people actually face every day.

Her point was it's easy for AI safety people to be distracted from the harms happening now because it's a blind spot for them. At the same time, AI safety people told me that AI ethics people have a blind spot. They're not worrying enough about apocalypse. But why would it be taboo to say all of this on mic?

Part of the reason might be because the fear of apocalypse has come to overpower any other concern in the larger industry. One thing that I think is...

is that a lot of the narrative that we hear about how AI is going to save the world and it's going to solve all these problems and it's amazing and it's going to change everything. And then we get the narratives about, oh my gosh, it could destroy humanity in 10 years. Often coming from the same people. I think part of the reason for that is that either way it makes...

AI seemed more powerful than it certainly is right now. And, you know, who knows when we're going to get to the humanity-destroying stuff. But in the meantime, if it's that powerful, it's probably going to make a whole lot of money. Building a super-intelligent AI has become a multi-billion-dollar business. And the people running it are not ethicists. Just weeks before Elon Musk called for the pause, he had started a new AI company.

Yeah, I guess it's kind of counterintuitive, right, to see this. And you're like, wait, why would the people working on the technology who stand to profit from it want to pause? Right.

I can't speak for them, but it benefits them to, on the one hand, get everybody else to slow down while they're doing whatever they're doing. Octopus thought experiment author Dr. Emily Bender again. But also it benefits them to market the technology as super powerful in that way. And it definitely benefits them to distract the policymakers from the harms that they are doing.

It'd be nice to think that billionaire Elon Musk was calling for an industry-wide pause in building large AI systems for all the right reasons. A pause that never came to be, by the way. It's worth pointing out that when the billionaire took over Twitter and turned it into X, one of the first things he did was fire the ethics team.

And even though Elon Musk says he left and sued the chat GPT company OpenAI over safety concerns, company emails have surfaced that reveal the more likely reason he left is that he fought with folks internally to try and make the company for profit, to better compete with Google.

Ethicists are concerned they're outnumbered by the apocalypse people. And they think a lot of those people are in it to maximize profit, not maximize safety. So how did we get here? Why? Why is the industry not focusing on AI harms today and focusing instead on the risk of AI apocalypse?

There's an enormous amount of money that's been collected to fund this weird AI research. Why do you think the resources are going to those long-term, like, hyper-intelligent AI concerns? Because you have very powerful people who are posing it, people who control powerful companies and people with very deep pockets. And so money continues to talk.

It seems to be like funding for sort of like fanciful ideas, right? It's like almost, it's almost like a religion or something where it requires faith.

that good things will come without those good things being clearly specified. People wanting to be told what to do by some abstract force that they can't interact with particularly well. It's not new. Chat GPT gives you authoritative answers. Erosions of autonomy. Like a god. Yeah, like a god. It's like...

really interesting to take these philosophies apart. I would argue they trace back to a large degree to religious thinking, but that might be another story for another day. Next time on Good Robot. Good Robot.

Good Robot was hosted by Julia Longoria and produced by Gabrielle Berbet. Sound design, mixing, and original score by David Herman. Our fact checker is Caitlin Penzi Moog. The show was edited by Katherine Wells and me, Diane Hodson. If you want to dig deeper into what you've heard, you can check out Dr. Joy Boulamwini's book, Unmasking AI, or head to Vox.com slash Good Robot to read more future perfect stories about the future of AI.

Thanks for listening. Listen, we all love a good hustle. We all love a rise and grind mentality. But let's be real. You could do all that and still be super comfy, right? Right.

Nothing says comfort quite like a pair of Crocs. Since 2002, they've been disrupting the footwear industry with their iconic clogged silhouette. And now you can choose from more than 800 styles to find the pair of Crocs that is perfect for you. Like the Echo Wave, with a sculptural boundary-pushing design that was crafted with the urban explorer in mind. Whatever your style, buy yours at Crocs.com. Your Crocs, your story, your world.

And we're back with Canva Presents Secret Sounds, Work Edition. Caller, guess this sound. Mouse click. So close. That's actually publishing a website with Canva Docs. Next caller. Definitely a mouse click. Nice try. It was sorting 100 sticky notes with a Canva whiteboard. We also would have accepted resizing a Canva video into 10 different sizes. What? No way. Yes way. One click can go a long way. Love your work at Canva.com.