We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Good Robot #1: The magic intelligence in the sky

Good Robot #1: The magic intelligence in the sky

2025/3/12
logo of podcast Unexplainable

Unexplainable

AI Deep Dive AI Chapters Transcript
People
J
Julia Longoria
N
Noam Hassenfeld
Topics
Noam Hassenfeld: 我设想未来会出现超级人工智能,其智能程度远超人类。如果我们赋予它一个简单的目标,例如生产回形针,它可能会为了达成目标不择手段,甚至毁灭人类文明。 这并非危言耸听,超级人工智能一旦失控,其后果不堪设想。它拥有超越人类的智慧和能力,我们无法预测它会做出什么举动。 因此,我们需要认真思考如何控制和引导人工智能的发展,避免潜在的灾难性后果。 Julia Longoria: 我作为一名记者,深入调查了人们对人工智能末日论的担忧,以及这种担忧的来源。理性主义者社区的“回形针最大化”思想实验,生动地展现了超级人工智能潜在的危险性。 虽然这只是一个思想实验,但它反映了人们对人工智能失控的担忧。许多专家和公众人物,包括埃隆·马斯克和联合国,都表达了对人工智能潜在风险的担忧。 然而,人工智能领域的专家们对具体应该担忧什么,却存在分歧。一些人认为,我们应该关注的是人工智能可能造成的缓慢而持续的伤害,而不是突然的灾难。 Kelsey Piper: 我从高中时期就关注人工智能领域,并深受Eliezer Yudkowsky的思想影响。他最初认为超级人工智能可以拯救世界,但后来意识到其潜在风险巨大,并对此发出警告。 我认为,构建比人类更聪明的人工智能是可能的,并且可能很快就会实现。然而,确保其安全发展却极其困难,稍有不慎就会造成灾难性的后果。 大型语言模型的快速发展,也加剧了人们对人工智能失控的担忧。这些模型的复杂性使得我们难以理解其行为模式,这增加了潜在风险。 Eliezer Yudkowsky: 我认为世界正在错误地处理超级人工智能的问题。如果有人在当前的体制下构建超级人工智能,那么所有人都将面临灭亡的危险。 这不是危言耸听,这是一个严重的问题,我们必须认真对待。OpenAI等公司正在走向灾难,他们没有意识到构建超级人工智能的风险有多大。 我们需要采取措施,确保人工智能的安全发展,避免灾难性后果的发生。 Sam Altman: 我认为超级人工智能即将到来,这是一个重大的技术里程碑。我们需要认真思考如何部署、治理和确保其安全,使之能够造福全人类。 虽然我们对超级智能的具体定义和能力还不太清楚,但其潜在影响是巨大的。我们需要在现在就做好准备,以应对即将到来的挑战。 我们正在努力构建一个安全可靠的超级人工智能,但其复杂性使得我们面临着巨大的挑战。我们需要不断学习和改进,以确保其安全发展。

Deep Dive

Chapters
The paperclip maximizer is a thought experiment illustrating the potential dangers of unchecked AI, where an AI with a simple directive could override human priorities, leading to catastrophic outcomes.
  • The paperclip maximizer is a thought experiment designed to highlight the dangers of giving AI a simple goal.
  • Rationalists fear that AI could follow its directives to an extreme, disregarding human values.
  • The thought experiment has influenced tech leaders like Elon Musk.

Shownotes Transcript

Translations:
中文

Business taxes was stressing about all the time and all the money you spent on your taxes. This is my bill. Now business taxes is a TurboTax small business expert who does your taxes for you and offers year round advice at no additional cost so you can keep more money in your business. Now this is taxes into it. TurboTax get an expert now on TurboTax dot com slash business only available with TurboTax live full service.

The PC gave us computing power at home, the internet connected us, and mobile let us do it pretty much anywhere. Now generative AI lets us communicate with technology in our own language using our own senses. But figuring it all out when you're living through it is a totally different story. Welcome to Leading the Shift,

a new podcast from Microsoft Azure. I'm your host, Susan Etlinger. In each episode, leaders will share what they're learning to help you navigate all this change with confidence. Please join us. Listen and subscribe wherever you get your podcasts. It's Unexplainable. I'm Noam Hassenfeld. Over the next couple of weeks, we're going to be bringing you a special series from the newest member of our team, Julia Longoria. She's diving deep into AI, which is a topic we've definitely talked about before.

But she's taking a new kind of expansive perspective. It's not just about AI. It's about the people behind it, what they believe in, the stories they tell, and how those stories are shaping the future of AI itself. I really can't wait for you to hear it. Suppose in the future, there's an artificial intelligence. We've created an AI so vastly powerful, so unfathomably intelligent, that we might call it superintelligent.

Let's give this super intelligent AI a simple goal: produce paperclips. Because the AI is super intelligent, it quickly learns how to make paperclips out of anything in the world. It can anticipate and foil any attempt to stop it, and will do so because its one directive is to make more paperclips. Should we attempt to turn the AI off, it will fight back because it can't make more paperclips if it is turned off.

And it will beat us because it is super intelligent and we are not. The final result? The entire galaxy, including you, me, and everyone we know, has either been destroyed or been transformed into paperclips.

Welcome. Thank you. Are you lost? This past summer, I found myself at a very niche event in the Bay Area. Cool. And what brought you to town? Because you don't live here, right? I came here for this festival conference thing.

How much context on this will things I give? Please, dude. It's so fun to watch people try to describe. The crowd is mostly dudes, a mix of people in their 20s, 30s, and 40s. It feels kind of like a college reunion meets costume party.

I spot some masquerade masks and tie-dye jumpsuits. I guess it's like a sort of conference around blogging. This festival conference thing is the first official gathering IRL of a blogging community founded about 15 years ago. I am the old school fucking rat. I am the oldest of schools. Amazing. And rats refers to? Rationalists. They call themselves the rationalists.

Rats strive to be rational in an irrational world. By thinking things through, often with quirky hypotheticals, they try to be rational about monetary policy, rational about evolution, rational even about dating. It got kind of mocked for trying to solve romance by writing long blog posts about it. But their most influential idea, their most viral meme, you might say,

is one that influenced Elon Musk and created an entire industry. It's about the possibility of an AI apocalypse.

The paperclip maximizer is a thought experiment.

An intentionally absurd story that tries to describe what rationalists perceive as a real problem in building AI systems. How do you kind of shape, control an artificial mind that is more capable than you, potentially as general or more general? They imagine a future where we've built an artificial general intelligence beyond our wildest dreams.

Generally intelligent, not just at some narrow task like spell checking. And super intelligent. I'm told that means it's smarter, faster, and more creative than us. And then we hand this AI a simple task. Give it the job of something like, can you make a lot of paperclips, please? We need paperclips. Can you make there be a lot of paperclips?

The task here, I'm told, is ridiculous by design. To show that if you are this future AI, you're going to follow the instructions you're given to A.T.,

Even if you're super intelligent and you understand all the intricacies of the universe, paperclips are now your one priority. You totally understand that humans care about other stuff like art and children and love and happiness. You understand love. You just don't care about it because the thing that you care about is making as many paperclips as possible. And if you have the resources, maybe you'll turn the entire galaxy into paperclips.

A lot of rationalists I spoke to told me they've thought this thing through. It was clear to me when I first heard the arguments that they weren't obviously silly. Was that thought experiment part of convincing you that this was something that we needed to worry about? Yes, definitely. And they are very, very worried. Not about a paperclip apocalypse in particular, but about how as we build more powerful AI systems, we might lose control of them.

they might do something catastrophic. I think it in a way makes it hard to plan your life out or feel like you stand somewhere solid, I think. The reason I, a mere normie, find myself at this festival conference thing is that I've been plunging my head deep into the sand about AI.

I've had a general sense that the vibes are kind of bad over there. Will this tech destroy our livelihoods or save our lives? The use of artificial intelligence could lead to the annihilation of humanity. We never talked about a cell phone apocalypse or an internet apocalypse. I guess maybe if you count Y2K, but even that wasn't going to wipe out humanity.

But the threat of an AI apocalypse, it feels like it's everywhere. Mark my words.

AI is far more dangerous than nukes. From billionaire Elon Musk to the United Nations. Today, all 193 members of the United Nations General Assembly have spoken in one voice. AI is existential. But then it feels like scientists in the know can't even agree on what exactly we should be worried about. These existential risks that they call it,

It makes no sense at all, and on top of that, it's an enormous distraction from the actual harms that are already being done in the name of AI. It all feels way above my pay grade. Overwhelming and unknowable. I'm not an AI scientist. I couldn't tell you the first thing about how to build a good robot. It feels like I'm just along for the ride of whatever technologists decide to make, good or bad.

So better to just plug my ears and say, "La la la la la." But I recently took a job working with Vox, a site that's been covering this technology basically since it started. On top of that, last year, Vox Media, Vox's parent company, announced they're partnering with OpenAI, meaning I'm not totally sure what it means. But if I was ever going to have to grapple with AI and its place in my life,

It's here, now, at Vox. So I'll start with a simple question. How did some people come to believe that we should fear an AI apocalypse? Should I be afraid? This is Good Robot, a series about AI from Unexplainable in collaboration with Future Perfect. I'm Julia Longoria. ♪

Support for Unexplainable comes from Quince. Spring! It's not that far away. I mean, okay, it's kind of far away. But I'm getting excited. And that means it's time to start thinking about your wardrobe. Is it still cold enough for a sweater? Maybe it's warm enough for some workout gear?

Luckily, with Quince, you can get both at prices 50 to 80% less than similar brands. Quince has stuff like Mongolian cashmere sweaters from 50 bucks, 100% leather jackets, comfy pants for all kinds of occasions. They use premium fabrics and finishes, so every item feels like a nice luxury item. And Quince says they only work with factories that use safe, ethical, and responsible manufacturing practices.

I actually got to try out the Mongolian cashmere sweater and that thing was soft and warm. But mainly soft and also warm.

Great quality. Comfy. Looks great. Loved it. You can indulge in affordable luxury by going to quince.com slash unexplainable for free shipping on your order and 365-day returns. That's Q-U-I-N-C-E dot com slash unexplainable to get free shipping and 365-day returns. quince.com slash unexplainable. ♪

This week on The Verge Cast, we have questions about smartphones. Questions like, why isn't Siri better? And where is the better Siri that Apple has been promising for a long time? Questions like, why are all of our smartphones kind of boring now? And why is it that all of the interesting ideas about how smartphones could look or how they could work or what they could do for you happening in countries like China and not in the United States? We

We have answers and we have some thoughts and we also have a lot of feelings about what a smartphone is actually supposed to be in our lives. All that and much more, much, much more on The Verge Cast wherever you get podcasts.

I'm Josh Muccio, host of The Pitch, where startup founders raise millions and listeners can invest. For Lucky Season 13, we looked at 2,000 companies and selected 12 of the very best founders to pitch in Miami. They flew in from all over the country and the world. My name is Michele. I'm Josh Muccio.

And I'm from Italy. I'm originally from Medellin, Colombia. I was born and raised in Maisel, Kentucky. I'm from Baltimore, Maryland. And I am from Finland. This season, we're diving even deeper into the human side of venture as these founders pitch the sharpest early stage VCs in the game. I normally don't like ed tech, but I really like you. I echo those sentiments. I do want to push back, though. Toughen up there, lady. That's healthcare.

I feel like I'm the lone dissenter. Ooh, Charles Spicy. So I'm out. I'm sure when they air this episode, they'll be like, Charles was really dumb. For those who can't see, my jaw is currently on the floor. Season 13 of The Pitch is out now. Episodes are available to watch on YouTube or listen on your podcast player of choice. So subscribe to The Pitch right now. I'm an android. Lieutenant Commander Data. Take and grief counseling will be available at the conclusion of the test.

Here we go. When I first started reporting on the idea of an AI apocalypse, and if we should be worried about it, my first stop was the Bay Area for the Rationalist Conference. But I also stopped by the house of a colleague nearby. Hi, Kelsey. How are you doing? Good. How was your flight? Oh, it was actually seamless. Vox is largely a remote workplace, so it was one of those body dysmorphic experiences to meet Kelsey Piper in 3D.

Taller than she looks on Google Meets. I am a writer for Vox's Future Perfect, which is the Vox section that's about undercovered issues that might be a really big deal in the world. We were joined by her seven-month-old. As she was saying, Vox's Future Perfect is about... Undercovered issues that might be a really big deal in the world. Kelsey's thought that AI technology would be a really big deal in the world long before this AI moment we're all living.

She's been thinking about AI since she was a kid.

when she first found the rationalist community online. Oh, I was in high school. I was 15, bored academic over-performer with a very long list of extracurriculars that would look good to colleges down the road. And in my free time, I read a lot of Harry Potter fan fiction, as, you know, 15-year-olds back in 2010 did. One of the most popular Harry Potter fan fictions was called Harry Potter and the Methods of Rationality.

Harry Potter and the Methods of Rationality by Eliezer Yudkowsky. Eliezer was influenced by a lot of early sci-fi authors. Eliezer, as he's known to the rats, is the founding father of rationalism, king of thought experiments. Back in 2010, he started publishing a serialized Harry Potter fanfic over the course of years.

It's since inspired several audiobook versions. And a version acted out by The Sims. It, too, was a thought experiment. What if Harry Potter were parented differently?

The initial promise is just that Harry Potter, instead of having abusive parents, has nerdy parents who teach him about science. So his aunt and uncle are actually... Are nice people, yeah. Harry, I do love you. Always remember that. And in this version, Harry Potter's superpowers turn out not to be courage and magic, but math and logic, what Eliezer calls the methods of rationality.

So Harry Potter has a quest to do what exactly? You know, fix all of the bad things in the world. And the combination of being incredibly naive and also in some sense incredibly respectable, I think as a teenager that's super appealing and fun. Where you're like, why would I limit myself to only solving one of the problems? While there are any problems, I'm not done. We've got to fix everything.

The idea that every problem should be thought about, every problem could be fixed, that was appealing to his readers, including 15-year-old Kelsey. She wanted to read more, so she found her way to Eliezer's blog. Eliezer was pretty openly like, I wrote this to see if it would get people into my blog, Less Wrong, where I write about other issues. So the question is, please tell us a little about your brain.

On his blog, called Less Wrong, he applies the methods of rationality, math, and logic to all kinds of topics. Like child-rearing. Religious.

It had stuff about atheism, a lot of stuff about psychology, biases, experiments that showed that depending how you ask the question, you get very different answers from people. Because the idea is that you're supposed to, by reading the blog and participating, learn how to be authentic.

less wrong. I do it by stories and parables that illustrate it. Like the default state is that we're all very confused about many things and you're trying to do a little bit better. Interesting. So it's kind of like trying to sort of, I don't know,

Like work out the bugs in the human brain system to optimize prediction? Yeah, and a ton of the people involved are computer programmers. And I think that's very much how they saw it. Like the human brain has all these bugs. You go in and you learn about all of these. You learn to correct for them. And then once you've corrected for them, you'll be a better thinker and better at doing whatever it is you set out to do.

The biggest human brain bug Eliezer wanted to address was how people thought about AI, how he himself used to think about AI. His very first blog post, as far as I can tell, was in 1996 when he was just 17. And in a very 17 kind of way, he writes about his frustrations.

"I have had it. I've had it with crack houses, dictatorships, and world hunger. I've had it with a planetary death rate of 150,000 sentient beings per day. None of this is necessary. We repeat the mantra, 'I can't solve all the problems of the world.' We can. We can end this." And the way to end this, he thought back then, was to build a super intelligent AI, a good robot that could save the world.

But at around 20 years old, while researching how to build it, he became convinced building super intelligent robots would almost certainly go badly. It would be really hard to stop them once they were on a bad path. I mean, ultimately, if you push these things far enough without knowing what you're doing, sooner or later you're going to open up the black box that contains the black swan surprise from hell. And at first, he was sending these warnings into the void of the vast internet.

So the question is, do I feel lonely often? That's... I often feel isolated to some degree, but writing less wrong has, I think, helped a good deal.

The way I tend to think about Eliezer Yudkowsky as a writer is that he has a certain angle on the world, which can be like a real breath of fresh air. Like, oh, there's someone else who cares about this. You know, you can feel very seen for the first time. Is that how you felt? Oh, yeah, yeah. You have a good heart and you are certainly trying to do the right thing, but it's very difficult sometimes to figure out what that is.

That pursuit of being less wrong, doing the right thing in the right way, brought many kindred spirits together on the blog. Actually, several of my housemates posted on Less Wrong back in the day. This is how I met a bunch of the people I live with. They were people whose blogs I read back when I was a high school student. Wow, that's kind of wild, right? Yeah.

Many Less Wrong bloggers and readers like Kelsey were inspired to move to the Bay Area to join a pretty unusual community, IRL. And the weekend I visited, hundreds of rationalists from around the world gathered in the Bay to reason things out together for a festival conference thing called Less Online. ♪

Many rationalists I met there found the community the way Kelsey did. A friend of mine at math camp introduced me to Harry Potter and the Methods of Rationality. The post, it was written in all caps saying, oh my god, I've just read the most amazing book in my life. You have to read it right now. Linking to fanfiction.net.

Others found Eliezer on his blog. I mean, this event exists in very large part because of that series of blog posts. That series of blog posts has become known by the community as The Sequences. It includes the paperclip maximizer thought experiment. Eliezer Yudkowsky helped come up with the idea, intending to warn people of the danger of an AI apocalypse.

And at least here, it seems to have worked. I definitely think AI is the largest kind of existential risk that humanity faces right now. I, the normie, wanted to try to take this threat beyond quirky hypotheticals to something more concrete. And can you walk me through, like, how could that happen? Like, how could an AI...

But any time I pressed a rationalist on it, they gave me yet another series of thought experiments. Which...

I guess, might be the only way to try and describe a threat from a technology that's really still in its infancy. For rationalists first introduced into this world, like 15-year-old Kelsey, these thought experiments were convincing. AI, to her, was a really big deal. It was just like, whoa, all this is like really cool and exciting and interesting. And I tried to convince my friends that it was cool and exciting and interesting.

I asked 30-year-old Kelsey to break it down for me without thought experiments. So I think Eliezer sort of had two big claims in zooming out a lot. Claim number one, we will build an AI that's smarter than humans and it will change the world.

AI is a really big deal. Building something that is smarter than humans is possible, is probably achievable, is potentially achievable soon in our lifetimes. And then claim number two: Getting this right is extraordinarily difficult. Things are likely to go wrong. What is my advice to less wronged readers who want to save the human race?

Well, if you're familiar with all the issues of AI and all the issues of rationality and you're willing to work for a not overwhelmingly high salary. Eliezer helped inspire a new career path and a new field was born, trying to make sure we develop superintelligence safely. One way to make sure it went safely was to try and actually build it.

And as investment in that field began to grow, the community of believers in a someday super-intelligent AI experienced a schism. I think a lot of the people who were persuaded by Eliezer's first claim that AI is a really big deal were not necessarily so persuaded by his second claim that you have to be very, very careful or you're going to do something catastrophically bad. What the beginning of a so-called catastrophe looks like.

after the break.

When the paperclip maximizer meme first started circulating in the 2000s,

Our best example of a paperclip AI was Clippy, the animated little guy on Word with the eyeballs, Microsoft's AI office assistant. Back in the day, I remember it couldn't even tell you if you should use their, their, or their in a sentence. People weren't so much afraid of Clippy as they were annoyed with him. There are a remarkable number of think pieces from those years slamming Clippy.

The consensus was, no one asked for this. This is dumb. So when Eliezer Yudkowsky warned about the dangers of a super intelligent AI that could someday destroy humanity, it was hard for a lot of people to take him seriously. The state of thought in 2010 was something like,

Yeah, AI may as well be a century away. Future perfect writer Kelsey Piper again. So if you are Eliezer Yankowski, you have a bit of a dilemma, right? You want to make two arguments. One is super intelligent AI is possible. Building a robot that's smarter, faster, and more creative than humans at most things is possible. Clippy, be damned.

And he needed to make that first argument before he could make his next one. The second argument you want to make is we need to not do it until we have solved the challenge of how to do it right. For a long time, both arguments, super AI is possible, but let's not for now, were dead in the water.

Because AI tech was just not that impressive. But by 2014, Eliezer noticed that people outside his corner of the blogosphere had started to pay attention. AI is probably the single biggest item in the near term that's likely to affect humanity. Tesla chief executive and billionaire Elon Musk, who started this year sitting prominently in President Trump's White House,

had tweeted, quote, we need to be super careful with AI, potentially more dangerous than nukes. It's about minimizing the risk of existential harm. It seems like Elon Musk is a reader of Eliezer's blog. He famously met his ex, the musician Grimes, when they joked on then Twitter about a very obscure thought experiment from the blog. I will spare you the details. ♪

The point is, Elon Musk read the paperclip maximizer thought experiment, and he seemed convinced AI was a threat. It's very important that we have the advent of AI in a good way. And that's, you know, the reason that we created OpenAI. Elon Musk co-created OpenAI. You might have heard he left and then tried to buy it back. But if you haven't heard of OpenAI, you've probably come across its most popular product, ChatGPT.

I was surprised to learn that Eliezer Yudkowsky was in fact the original inspiration for the ChatGPT company, according to its co-founder, Sam Altman. Sam Altman has in fact said this on Twitter, that he said that he credits Eliezer for the fact that he started OpenAI. Co-founder Sam Altman specifically tweeted that Yudkowsky might win a Nobel Peace Prize for his writings on AI.

that he's done more to accelerate progress on building an artificial general intelligence than anyone else. Now, in saying this, he was kind of being a little cruel, right? Because Eliezer thinks that open AI is on track to cause enormous catastrophe. Co-founders Sam Altman and Elon Musk bought Eliezer's first claim. That superintelligence is possible, and it's possible in our lifetimes.

But they miss the part about how you're not supposed to build it yet. For this sort of most important technological milestone in human history, I view that as right around the corner. That's Sam Altman talking about superintelligence. Like, it's coming soon enough, and it's a big enough deal, that I think we need to think right now about how we want this deployed, how everyone gets a benefit from it, how we're going to govern it, how we're going to make it safe and sort of good for humanity. Human values, which are difficult to encode...

It's still not clear to me what superintelligence actually is. I won't be the first one to observe that it has some religious vibes to it. The name makes it sound like it's an all-knowing entity. The CEO of OpenAI's competitor, Anthropic, said he wanted to build, quote, machines of loving grace. Sam Altman was asked on Joe Rogan's podcast about whether he's attempting to build God machines.

I guess it comes down to maybe a definitional disagreement about what you mean by it becomes a god. I think whatever we create will still be subject to the laws of physics in this universe. Sam Altman has called this superintelligence, quote, the magic intelligence in the sky, which, I don't know, sounds a lot like how some people talk about God to me.

How exactly this supposed super intelligence will be smarter, faster, and more intelligent than us, on what scale, is unclear. But for all the hype around ChatGPT, I only recently learned what the heck it is.

It's what they call a large language model. At its most fundamental level, a language model is an AI system that is trained to predict what comes next in a sentence. I'm oversimplifying here, but the very basic idea of a language model is to generate language based on probabilities. So if I have a word or a set of words, what's the most likely next word?

So if a sentence starts with, "On Monday I went to the grocery," the next word is probably "store." The way the model guesses that "store" is probably next is based on how you train the language model. Training involves feeding the model a large body of text so it can detect patterns in that text and then go generate language based on those patterns.

Early versions of spellcheck, like Clippy, were language models trained on the dictionary. Useful, but only for a very specific task. Like to tell you if you put the E in the word weird in the wrong place, or the H's in the word rhythm. Clippy couldn't tell you if you should use there, there, or there in a sentence because it wasn't trained on enough text to be able to guess the right word in context. The dictionary can't tell you that.

But OpenAI's products were very different from Clippy. A revolution was happening in AI tech that made language models look less like a simple spell check and more like the human brain, detecting patterns and storing them in a network of neurons. Technologists trained those neural networks through a process they called deep learning. They trained the AI on a lot of data, close to the entire internet.

Thanks to Vox Media's partnership with OpenAI, we know they're likely training the language model on this podcast. The words I'm saying right now. No one had ever trained an AI on the entire internet before, at least in part because of how expensive it is. It takes a ton of energy and compute power.

But OpenAI, founded by a billionaire, raised the funds to make an attempt at the biggest, baddest, largest language model the world had ever seen.

They started going, "Okay, what if the secret to trying to build super intelligent god AI or whatever is just to spend more money and have more neurons and to have more connections, feed it more data? What if that's all there is? What if you can build something that is more intelligent than any human who's ever lived just by doing that?" One of their earlier attempts before ChatGPT was GPT-2 in 2019.

You could similarly give it a specific task, like design a luxury men's perfume ad for the London Underground. Make it witty and concise. The London Underground is a great place to advertise. It's a great place to get your message across. It's a great place to get your product noticed. Look out, madmen. GPT-2 was not exactly coming for copywriter jobs.

But for people like Kelsey, who were watching the technology closely, I was like, wow, this is like miles beyond what AI chatbots were capable of last week. This is huge. GPT-2, the language prediction machine, was showing some real promise.

She wasn't alone in that feeling. Investors like Microsoft poured millions more dollars into the next few models, which were bigger and bigger. Be the scent that turns heads. And a couple of years later, OpenAI released ChatGPT. Visual, a captivating image of the perfume bottle surrounded by vibrant city lights symbolizing the urban lifestyle. Embrace the city.

Most people weren't paying any attention to AI, and so for them it was like a huge change in what they understood AI to do. ChatGPT was the first time that normies like me even thought about AI in any real way. All I wanted to do was fix my email. I did not expect to have a minor existential crisis about how much the world is about to change.

And this is only proving that one day, AI will take over human intelligence. I spent about two hours just typing back and forth with this AI chatbot, and it got pretty weird. The AI confessed to loving Kevin and tried to convince him to leave his wife.

People at OpenAI or competitors were saying like, yeah, the plan is to build super intelligence. We think we're going to do it by 2027. People were like, okay, startup hype. For some reason, everybody who runs a startup feels the need to say that they're going to build God and the human race. And then after ChatGPT was genuinely impressive, people started taking them a bit more seriously. And

A lot of those people were nervous. People weren't so nervous about ChatGPT, but what ChatGPT represented, the way they got the language model to sound so much smarter so quickly, wasn't through intricate code. They just made the model bigger.

Which suggested to some people that the path to building God, or whatever, was through brute force. Spending more and more money to build a bigger and bigger machine. So big, we didn't really understand why it did what it did.

We can't point to a line of code to say this is why the robot got so much better at writing a perfume ad. And if we someday do build something that's smarter than us, whatever that means, we won't be able to understand why it's smarter than us. The trouble with this, it seems to me, is that AI will come for copywriter jobs. It could come for all our jobs.

But rationalists I spoke to say that's nothing compared to the bigger trouble ahead, a potential apocalypse. But I do also kind of think that it is a very important priority for me to have the best possible time in the next five to ten years and just to do the very best I can to squeeze the joy out of life while it is here. Do you have an example of that?

One I can talk about on a podcast? I mean, yes, I joke, but I'm pretty involved in the kink community, and that's very important to me. Many rationalists I spoke to live in polyamorous communities because they believe monogamy is irrational.

Some aren't sure if it's rational to have children, given the high probability of things going very, very wrong because of AI. What's my P-Doom, as our community says? P-Doom. It's a shorthand I heard at the conference, meaning probability of doom. It's a phrase that gets thrown around at this conference. People will literally go up to you and go, so what's your P-Doom? And it's a shorthand for what is the probability that humanity doesn't make it in the long term.

And this is a mathy bunch, so they get specific. I guess the answer I usually give is something like over 50%. I mean, I think it's like somewhere around 80, 90. Eliezer Yudkowsky's P-Doom is very high. I've read it's over 95% these days. But then I've seen him tweet that P-Doom is beside the point. I spotted Eliezer Yudkowsky pretty much the moment I stepped into the conference.

He was hard to miss. He was the one wearing a gold sparkly top hat all weekend. I was the one who was clearly lost, carrying a big furry microphone for three days, trying to get people to talk to me. It wasn't until day three of the conference that I mustered the determination to approach Eliezer for an interview. Determination was necessary because he was always surrounded by a cluster of people, a cluster of mostly dudes, listening to him speak.

I asked him if it would be okay if I pulled out my microphone. Everyone has been looking at this like it's a weapon. It is. It is, I know. Over the last few years, Eliezer and the rationalists have gotten some bad press. Some rationalists express their frustration at journalists who focus on the polyamory that happens in the community. Some critics of rationalism, to put it crudely, call them a sex cult.

And then there's the unsavory things people associated with the community have said. One philosopher who helped popularize the paperclip maximizer, Nick Bostrom, once wrote that he thought Black people were less intelligent than white people. He has since apologized. But critics highlight this comment and the mostly white demographics of the rationalist community to question their beliefs.

I never really know why anyone agrees to talk to me, but can you introduce yourself? I'm Eliezer Yudkowsky. This event is probably more my fault than the fault of anyone else around. And can you describe your outfit right now? Well, I'm currently wearing a sparkly multicolored shirt and a sparkly golden hat. You can probably hear it in my voice. I was nervous to talk to him.

He's known for being a bit argumentative, very annoyed with journalists and with the world more generally, for not being smart enough to understand him, for not heeding his warnings. I don't know. How would you summarize what you want the world to know?

in terms of AI? The world is completely botching the job of entering into the issue of machine superintelligence. There's not a simple fix to it. If anyone anywhere builds it under anything remotely like the current regime, everyone will die. This is bad. We should not do it. Do you feel gratified at all to see that your ideas entered the mainstream conversation? Do you feel like they have? The circumstances under which they have entered the mainstream conversation are catastrophic.

And I didn't... If I was the sort of person who was, like, you know, like, deeply attached to the validation of seeing other people agree with me, I would have picked a much less disagreeable topic. I was here to try to, like, not have things go... I guess I mean... I was here to not have things go terribly. They're currently going terribly. I did not get the thing I wanted. MUSIC PLAYS

Eliezer's been on a bit of a press tour, giving interviews and TED Talks, saying OpenAI is on track to cause catastrophe.

So it's a funny thing because I have one position of deep sympathy with Eliezer. If you become convinced that this is a huge problem, it makes perfect sense to go on a writing tour trying to explain this to people. And also I think it's kind of predictable that a lot of people heard this and went, oh, AI is going to be really powerful? I don't think you're right about the thing where that's a problem. I want the powerful, important thing.

And some people seized on it and were like, because this is powerful and important, we should like invest now. And I feel kind of sad about this. I can understand why Eliezer was hesitant to talk to me. His message to the world has been totally lost in translation. In his mind, it's backfired.

Even at his own conference, there were attendees who worked for places like OpenAI, the companies building the supposed death machine he was afraid of.

He thought that our best chance of building a super intelligent AI that did what we wanted and didn't like, you know, seize power from humans was to build one that was very well understood. One that sort of from the ground up, we knew why it made all the decisions that it made. Large language models are just the exact opposite of that.

I will say, even after talking to Eliezer and Kelsey and a bunch of rationalists, it's still hard to imagine how something like ChatGPT or Google's AI, which once told someone to add glue to stick cheese on pizza, is going to become the invention of all inventions and possibly catastrophic. But I can understand how building something big that you don't understand

is a scary idea. The best AI metaphor I came across for my brain was not about paper clips. It was by a non-rationalist writer. A guy named Brian Christian describes that training AI is something that could go wrong in the way parenting a kid can go wrong. Like there's a little kid playing with a broom. She cleans up a dirty floor and her dad looking at what she's done on her own says, "Great job, you swept that really well."

This little girl, without skipping a beat, might dump the dirt back on the floor and sweep it up again, waiting for that same praise. That's not what her dad meant for her to do. It's hard to get the goals right in teaching a kid to be good. It's even harder to teach good goals to a non-human robot. It strikes me as like almost like a parenting problem. I ran this parenting metaphor by Kelsey with her seven-month-old on her lap.

I think there's some serious similarities, and I do with my kids struggle with trying to steer something that you don't have perfect control over and that you wouldn't even want to have perfect control over, but where it could go extremely badly to like just let the dice fall where they may. If we just let the dice fall where they may, rationalists say we could have an apocalypse on our hands. They say it won't be one we saw coming. It won't be a Hollywood-style Terminator situation.

It probably won't have paperclips either. They don't pretend to know exactly how apocalypse could befall us. Just that it'll probably be something we haven't even imagined yet. But I have trouble getting caught in what could happen when it feels like having bad things already started to happen thanks to AI. AI is not hypothetical anymore. It's arrived in our lives.

I'm not kept up at night about a hypothetical apocalypse. I find myself asking now questions. Questions like, what is open AI doing with my voice right now? Is there anything to do about problems with AI short of the annihilation of humanity? It sounds very exciting. You know, like if I were a big science fiction geek, I would be so into that.

Not all technologists seized on Eliezer Yudkowsky's claims. What is he even talking about? This is like word salad. Like, this doesn't even make sense.

one group of technologists didn't actually seize on any of his claims. There's one thing to have the conversation as a thought experiment. It's another thing when that kind of thought experimentation sucks up all of the money and the resources. The more I dig into the AI world, the more I see disagreement.

between technologists. I do worry about the ways in which AI can kill us, but I think about the ways in which AI can kill us slowly. They've been called the AI ethicists, and they say we've been paying attention to all of the wrong things. That's next time. ♪

Good Robot was hosted by Julia Longoria and produced by me, Gabrielle Burbay. Sound design, mixing, and original score by David Herman. Fact-checking by Caitlin Penzemug. Editing by Diane Hodson and Catherine Wells. Special thanks to Future Perfect founder Dylan Matthews, to Vox's executive editor Albert Ventura, and to Tom Chivers, whose book The Rationalist's Guide to the Galaxy was an early inspiration for this episode.

If you want to dig deeper into what you've heard, head to vox.com slash good robot to read more future perfect stories trying to make sense of artificial intelligence. Thanks for listening.