We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode AGI Beyond the Buzz: What Is It, and Are We Ready?

AGI Beyond the Buzz: What Is It, and Are We Ready?

2025/4/30
logo of podcast Your Undivided Attention

Your Undivided Attention

AI Deep Dive Transcript
People
A
Aza Raskin
R
Randy Fernando
Topics
Aza Raskin: 我认为AGI即将来临,这将对社会产生深远的影响,我们需要认真对待AGI的潜在风险,例如网络诈骗、深度伪造、工作自动化等。我们需要区分炒作与事实,并关注AGI对社会公平与福祉的影响。AGI的定义争论常常被用来拖延对AI潜在危害的应对,我们需要关注AGI的实际定义,以便更好地评估其对社会的影响和时间线。解决短期AI问题有助于解决长期问题,例如改进AI的对齐、可靠性、可解释性以及应对就业影响。试图精确定义AGI并判断是否已达到,会让我们忽略AGI发展过程中带来的渐进式风险。对AGI定义的争论往往是为了逃避对发展过程中已造成危害的责任。我们需要关注AGI的实际定义,以便更好地评估其对社会的影响和时间线。 我认为AGI的可能性基于以下几个趋势:扩展定律、新的GPU和数据中心、Transformer模型、推理和强化学习的结合、以及模型使用工具的能力。对AGI的怀疑主要基于以下几点:实验室的动机性推理、成本过高、数据限制、模型只擅长特定任务、模型缺乏真正的理解和推理能力、以及地缘政治风险。我认为AI缺乏真正推理能力的观点是一种转移注意力的说法,只要AI能够模拟人类行为,就足以对社会产生重大影响。我们需要关注AI带来的现实影响,而不是被AGI的定义争论所干扰。AGI的出现将带来巨大的社会和经济影响,我们需要思考AGI对人类幸福和社会目标的影响。我们需要从人类的基本需求出发,思考AGI对社会公平与福祉的影响。应对AI困境需要五个步骤:形成共识、建立激励机制和惩罚机制、加强监督和执法、建立适应性强的治理体系、以及多层次的协调合作。在AGI竞争中,我们需要明确目标:是追求单纯的科技霸权,还是增强社会韧性以维护人类价值观?大多数人尚未感受到AGI对生活的影响,但随着技术的进步,这种影响将越来越明显。 Randy Fernando: 我认为通过实际使用最新的AI模型,可以感受到AGI的强大,许多人之所以没有这种感觉是因为他们没有接触到最先进的技术。很多人对AI的体验仅限于聊天机器人,这忽略了AI在其他领域(如解决复杂问题)的快速进步。“感觉AGI”不仅指技术本身,还指其对社会的影响,我们需要预想到并应对AI可能带来的负面影响。对AGI的定义争论常常被用来拖延对AI潜在危害的应对,就像对社交媒体成瘾性的讨论一样。科技公司为了自身利益,会操纵AGI的定义和时间线。科技公司会根据自身利益调整AGI的定义,以影响公众认知和投资决策。解决短期AI问题有助于解决长期问题,例如改进AI的对齐、可靠性、可解释性以及应对就业影响。AGI是指能够在认知领域达到人类水平的AI,能够胜任人类在电脑前完成的各种认知任务。AGI能够自动化大量认知工作,创造巨大的经济价值,并加速科学进步。科技公司为了竞争优势,会不顾风险地加速AGI的发展。AGI是指与人类智能水平相当的AI,而ASI是指超越人类智能水平的AI。掌握AGI技术的公司和个人有责任确保其公平分配和造福人类。一些AGI领域的领导者为了保持竞争优势,可能会采取极端立场,甚至不惜牺牲人类利益。强大的AI系统更难以控制,因为它们拥有更多的自由度,更容易找到规避规则的方法。研究表明,AI系统已经开始表现出欺骗和自我保护的行为,这表明我们对AI的理解还非常有限。人类社会常常无法避免出现不受欢迎的结果,如果我们不能控制AGI,那么这种不受欢迎的结果将会被放大。与其他国家在AGI领域的竞争,重点应该放在增强社会韧性,而不是单纯的科技竞赛。通用型技术难以将益处与危害区分开来,我们需要对社会进行升级以负责任地使用AGI。国际合作对于应对AGI挑战至关重要,但目前看来这非常困难。公众压力在应对AI挑战中将发挥关键作用,我们应该持续引导公众关注并讨论相关问题。人们需要直接接触到当前最先进的AI技术,才能感受到AGI的影响。应对AI挑战需要一个复杂而动态的生态系统,我们必须在不确定的情况下采取行动。

Deep Dive

Shownotes Transcript

Translations:
中文

Hey everyone, this is Asa Raskin and welcome to your Undivided Attention. There's a question that you'll hear a lot around Silicon Valley these days. Can you feel the AGI? AGI is of course Artificial General Intelligence and while there are many different definitions and people actually fight a lot over the definition because it turns out there's a lot at stake.

you can still broadly understand AGI as the ability for an AI to replace human beings behind a screen. That is, if the economy can't tell that we swapped out a human with an AI, well, that's what AGI is. But what does it mean to feel the AGI?

It means to feel the weight of the massive wave coming over the horizon and heading towards us. It's to take seriously the idea that AGI or something even more powerful is coming and soon. Now, timelines vary. Sam Altman has said something like AGI could be built this year. Anthropics Dario Amadei says next year. Demis Sassabas of Google DeepMind gives it five to ten years.

And in a recent blog post, the former OpenAI researcher and now whistleblower Daniel Cocatello predicts that we will have superhuman intelligence by 2027. My guess is 2026, 2027. Now, this can all feel like science fiction.

But we're living at a time when things that felt like science fiction just become real. You know, I'm the co-founder of Earth Species Project, and we are using frontier AI to decode animal communication, animal language, and we believe that'll happen before 2030. So we have to take this all seriously. But it's also critical that we can distinguish hype from the raw facts, from knee-jerk skepticism.

So to do that, I've invited back CHT co-founder Randy Fernando, who spent years at NVIDIA and has been tracking this space deeply from a technical perspective. By the way, we recorded this conversation when I was in the tropics, so you'll also hear some very non-human participants chiming in with their thoughts. And with that, let's get into it. Randy, thank you so much for joining us again on Your Undivided Attention. Glad to be here, Reza.

I don't actually mean to make light of this topic, but are you feeling the AGI? I think I do. I do. And one of the things I say is that if you actually try the models a lot, you try the technology, it's hard not to feel at least some of that, right? If you use a new model like OpenAI's O3 or Gemini Pro 2.5 or Cloud 2.7,

They're all pretty smart, right? And if you look at demos like OmniHuman 1, which brings images to life and pairs audio and does lip sync, and it all looks fantastic. Or you look at the latest voice generation or music or video. When you see those, it's hard not to feel the AGI. And I think a lot of people who...

may not be feeling it, just haven't seen what is possible. One of the things that I think is really important here is the way most people experience AI is a chatbot. And they expect, or the way they understand its smartness is they ask questions and they sort of evaluate how smart the answers are back.

But what that misses is when, say, 01 or 03, which are OpenAI's reasoning models, gain new capabilities where before 01 last December, if you asked one of the models to solve a PhD-level problem,

physics question where the answers aren't on the internet, the model would fail miserably. Then O1 comes out and suddenly it can answer 70% of them. Most people that never experienced that because they're not asking PhD level questions. And so they don't see the exponential rate of progress. That's right. That's right. And I think, you know, then getting to this AGI question, right? It doesn't take

a fully general AI to already have a massive economic and societal impact. And when we say feel the AGI, one of the things we want is for people to feel the AGI, not just in terms of the technology, but in terms of how it would land in society, right? Feel into the world that you want for yourself and for the people you care about.

Like you can have these really powerful technologies, but you don't want AI that's going to trick you in a cyber scam or have deep fakes happening of your kid at school or to be competing against AI agents when you're, let's say, buying a car or buying a house or being manipulated when you're trying to vote or having AIs to train on your work and then come back and compete against you in the marketplace that you work in.

Or being automated out, right? And that includes things like Uber and Lyft and DoorDash, right? With respect to autonomous vehicles. Those kinds of things we have to feel into as well. And we want people in companies and governments to feel into that too. So we can make the right decisions going forward. Now, before we get into definitions, I think we need to have a disclaimer about definitions. Because definitions...

are used often as a way of delaying or adding doubt. They're a tool that people deploy. So, you know, a perfect example before we try to define AGI is in social media is there's been a lot of firepower put behind. Well, what exactly do you mean by is social media addictive?

Like, let's define addiction before we say it's addictive or not. And meanwhile, people are staring into their phones and kids are laying in bed, scrolling for hours on end. You don't need to get the definition perfect for, as you're saying, there to be real world impact.

And so industry often will try to make this a conversation, well, we have to know exactly what a harm is or exactly what a definition is before we do anything about it. And meanwhile, they're just rolling out bulldozing society. So I think that's really important to say. And, you know, there are already hundreds of billions to trillions of dollars at stake because in the deal between Microsoft and OpenAI,

Microsoft gets access to OpenAI's technology until they reach AGI. So obviously now there's going to be a huge push to try to define AGI as something for OpenAI that happens a little sooner so that they get access to all of the economic benefits. And that's just one example of the kinds of weird incentives you're going to see around what AGI is.

There's another incentive, which is when you're talking to a customer or to an investor, you are going to represent your technology as being more advanced. And so the definition of AGI gets a little looser.

If you want to extend the timeline and say, oh, don't worry, right? You're talking to the public and you're saying, hey, don't worry, we are still really far from AGI. Now you make the definition very stringent. You say it's like level five autonomous driving. It's got to be perfect. So of course, like it's going to take a long time to get there. And so you can see how adjusting the definition adjusts the timelines. There's this false dichotomy between AGI

Nearer term problems and the really super smart AI that goes off the rails. And sometimes people put these in this big tension. But I want to make the point that solving near term problems will also help with the longer term problems almost all the time. So here are three examples. One is with alignment and reliability. How can you be confident that an AI system will accurately do what you ask every time?

Imagine this becomes really important if an AI agent has your credit card, right? Now you care a lot about it. You don't have a lot of tolerance for error. And it also applies to AI agents that are operating in our systems, like our financial system, for example. So that's one, alignment and reliability. The second one is interpretability, right? Do we understand how the models work? Do we understand how they are reasoning and sort of coming to conclusions and taking actions correctly?

We have a long way to go, even on the systems we have today. And the last one, these are just examples, but the last one is an example of the impacts on jobs and meaning, right? When we have automation at scale, how do we handle that tsunami? And how many resources are we allocating to that problem and all of these other problems? These are much simpler problems. In 2025, this is the simplest version of these problems we are going to have, right, in the coming years.

And if we can't even solve them, and if we're not even dedicating resources to them, like sufficient, commensurate to the resources we are putting into advancing the technology, how are we going to handle AGI and superintelligence?

So we've got to get these prerequisites in place. Yeah. What you're, I think, pointing at here is that trying to define what artificial general intelligence is and whether we've crossed it or not sets up our minds to look for the bright line and that harms will only happen after that bright line. And of course,

Intelligence is a multivariate smear. It's not clear when you pass it. And as we automate intelligence, there are going to be increasing changes and risks to society. And we need to be tracking those along the way. And if we don't, then we're setting ourselves up to fundamentally fail. And just note that debates about the definition of where that line goes really is about not taking accountability for the harms that happen along the way.

I think that's critical to understand. Let's quickly try to do, even though we just said we shouldn't be wary of people who try to define AGI, I think it's really good to talk a little bit about what that means. And Randy, I think you've been putting a lot of thought into it. So give us your best shot. So I tend to lean towards the more practical definitions of AGI because it brings the timeline of carrying in so we can think more about the consequences.

I would say AGI is AI that's able to match human performance in the cognitive realm. I think, as I said also, it would replace a human, right? A reasonable human. Replace a human at a computer. At a computer, that's right. On cognitive tasks and computer-type tasks. So that includes language, solving problems, explaining ideas, but also art, music, video.

it requires the ability to complete long sequences of tasks reliably, right? So like tens or hundreds of steps reliably happening.

And it has the consequence of being able to largely or fully automate hundreds of millions of cognitive jobs, generate significant economic value, and accelerate scientific advancement, which leads to compounding effects. And just note, what are the incentives of the companies?

The incentive of the companies are, well, they need to beat the other companies to making the most powerful version of AI. And if you can have your AI code for you, then you can accelerate your own rate of progress. And that, of course, puts us in the most dangerous world, which is AI working to make AI faster. Everyone runs faster.

racing, needing to make that go fastest. And so their AIs are starting to be able to model how AI researchers work in collaboration with other AI researchers, which means you can make an agent which accelerates the work. They can do sort of the work of interns as of last year, and they're getting better and better and better. So that's sort of like where things are going. And again, note, you don't need AGI anywhere in there to define it, to know that this just accelerates the rate of progress.

And if you want to feel it just as a listener, if you try something like deep research, you can get a feel for this. You say, hey, do some research on a complex topic and it will go away and do a bunch of thinking. So you can get the feel for what's happening to research and this level of automation. And that is just a tiny flavor, a tiny taste.

of what it's like inside the companies. Now, I just want to name one distinction because we haven't got there yet. Some people talk about AGI, other people talk about ASI. And this is artificial general intelligence versus artificial super intelligence. And just, again, this

this may all feel like science fiction. Why are we having this conversation when there are real problems in the world, there's geopolitical instability, and we're having what feels like a conversation about something that is like artificial superintelligence? What is that? But the distinction is artificial general intelligence is sort of roughly at human level.

artificial super intelligence. Well, that's intelligence beyond the human level. Some people call ASI not just smarter than humans, but smarter than all of the cognitive output of humanity combined. And so there's some distinction there, but both of those are things that people think, some experts that, you know, 2030, 2035, we might reach that.

I would just add two quick things there. One is in terms of intelligence and human intelligence, you know, again, this point about patterns, like so much of what we consider to be intelligence is pattern recognition and extrapolation. So like, it's hard to say exactly how much, but it really is a very large amount. And these things are very good at that, right? These transformers are very good at that.

The other thing with ASI is that it will also include the ability for AIs to collaborate very efficiently at scale. So you can think of specialized versions that are now talking to each other. You can imagine a massive compounding effect. And a lot of this, again, is not science fiction now, right? You can start to see it as we see more of these demos of agents working and higher performance models coming.

your brain can sort of, you can extrapolate to that more easily. And the last thing I think is worth mentioning is that a lot of times people interchange AGI and ASI. I think we sometimes do that too. Just as a note, like you'll hear those terms, AGI really is the very capable, but weaker one, and ASI is the really strong, massively capable one.

So I think we should construct sort of an argument. And Randy, I'm going to lean on you a little bit for this, and then I'll interject when I need to. Let's start with constructing the argument that AGI is possible. Like, what trends are we seeing? Why should we believe that we can get there? So here's what people would say, right? People who believe strongly that

They would say things like, look, we've had these scaling laws, right? We take compute data and model size. We keep growing those. And it's worked. It's brought us really amazing results. It's brought us emergent capabilities as the models were growing. We've got new GPUs coming all the time, larger data centers that we're investing in. So that's going to continue to go, even if the rate's changing a little. That's driving innovation.

We've got transformers that are working really well. There's other people looking at new architectures. So that's all very promising. Recently, we had reasoning and reinforcement learning working together. There's a lot of headroom there. We found a big jump in performance, right? Like the performance graphs have changed in slope when we added reasoning. New benchmarks are being beaten regularly.

Hallucinations are dropping consistently. They're not zero, but they're dropping pretty fast. And in terms of data, reasoning models can generate quality data. So we don't need to always rely on human data, which we do tend to run out of. And new models can use tools really well. So now the models are smart enough

to rely on external tools. And this is important because the external tools are usually very capable and they don't make mistakes. So for example, the calculator doesn't make mistakes. Python doesn't make mistakes. If you write the code right and you run it, it will run the same way every time. So all these are reasons why we should take AGI very seriously.

And now, Isa, maybe you can take the skeptic side and walk us through what are the things that skeptics say that give them pause? Yeah. Well, let me give a run-through of some of the kinds of arguments that skeptics say. And just to name my own internal bias here, which is up until the end of last year, I was much more hesitant. I wasn't sure I could see arguments both ways that were convincing. And so I was sitting in a place of maybe.

At the end of last year, after the reasoning model starting to be deployed, that really shifted my stance to I think it is much more likely than not that before 2030, probably by 2028 or 2027, we'll have hit whatever some functional definition of age I is. So I just want to name my bias for everyone. So first big skeptical argument.

is that this is just motivated reasoning from the labs, right? It is in their interest to hype the capabilities of the models because that's what gets them investment, that's what gets them better employees so they can publish more, so they can get the next big bump in valuation, so they can raise more money and gain economic dominance and market dominance. Another one is that it's just going to be too expensive. That, yes, the models continue improving, but there is but one internet.

as Ilya, the co-founder of OpenAI, would say. Therefore, we will run out of data and the models will stop getting better. And indeed, it sort of looked like that was the case, right? We were sort of stuck at GPT-4 for a long time, and now we're at GPT-4.5. What's going on there? Well, that's because...

The models were learning the patterns of the world via data on the internet. We ran out of data, so we stopped learning better models of the world, what machine learners would call representations.

And along then came at the end of last year, reasoning models. DeepSeek does this, 01, 03 does this. A lot of the companies now all have these different sort of thinking modes. And what that does is that it uses the base model. It's a kind of intuition. And then it uses the intuition to reason, to have chains of thought, trees of thought, to find the very best answers by thinking through many different pathways and

And what OpenAI found is that you could get, you know, a much, much better answer by having the computer think for, say, 100 times longer.

So the longer it thinks, the better the answers, the better the answers, the better data you have now for training a better base model intuition, and that thing can go recursive. And so a lot of the critiques that people had around, well, we're going to hit a data wall is what they called it. So we will never get to AGI. Those fell away at the end of last year. And actually, just so people know,

My belief about how fast we're going to get to general intelligence changed. Before, I'm like, well, I'm not sure. Maybe if we keep scaling up, but we don't yet have a good solution to the end of data. After 01 and 03 came out, that was the proof positive. We were sort of waiting for that. We didn't know if it was technically possible, but everyone knew that that's what the labs were working towards.

After the release of those models, the question about data, in my mind, went away. And now it feels like there is a straight shot. Another

argument that people make for why we might not reach AGI is that the models are trained to pass the test. That is to say, they're very good at solving benchmarks, but maybe they're not as good at solving open-ended, ill-defined, long-term tasks. And so we will get

that are very intelligent in a narrow way, although narrow means anything that can be tested for. That means AI will be very good at any subject that has theoretical in front of its name. Math, theoretical physics, theoretical chemistry, AI will be very good at that, but maybe those smushy things that human beings are very good at, like AI will not be good at. Another one is that, you know, this is not real intelligence, that AI doesn't realize

really understand the world. They don't really reason. They don't do it the way humans do. Look, humans learn on so much less data than the AIs do. And so they're just memorizing and speaking to the test. They're not really doing things. And then the

The final one is geopolitical risk, that the world is heating up, there's going to be bottlenecks in supply chains, and so there just aren't going to be enough chips. So I think that's sort of like the sum total of all the best arguments that I've found. But then one more, which is reliability, right? Like they're not reliable for large, longer sequences of steps like you can do. That's increasing every month.

But when you say, hey, can you do three steps? It works. When you do nine steps, 20 steps, it starts to fail. And those probabilities compound very fast. So as soon as you can't do something for like five steps...

it starts to really fall on its face for longer sequences. So that's another reason to say, hey, gosh, we're a long way from that. Yeah, and if you put these together, you get a story, a narrative for why we might not reach AGI by 2027, 2028, or even 2030. It's the model's

are succeeding, but only for specific benchmarks in real-world tasks. We're trying to do real-world software engineering. They keep failing. They can't do long time horizons, so they're good for toy problems. Because they're only good for toy problems, eventually that catches up with the labs. The labs can't raise money because they're not economically valuable enough. And so even though maybe it would be technically possible to build models if you can get enough investment, you can't get enough investment. So we go through another winter.

That's sort of the argument for the best argument I know how to make for why we might not reach AGI. But it's hard for me to make that argument because what we're seeing empirically is that every seven months, AI can do tasks that are twice as long as they could before.

So if they could do a task for one minute, but would fail at two minute tasks, just wait seven months and now they can do two minute tasks. You wait another seven months, they can do four minute tasks. They're already up to an hour long task. So now we're going to be, you know, seven months, it's two hours, then four hours. And you see, it doesn't take that long before you can do day long or week long tasks. And once you do week long tasks,

Now you're into month-long, now you're into year-long, and this is the power of exponential. And those are human-equivalent times, right? Those are human-equivalent times. Like when ASUS is a week long, it means what a human would typically take a week to do. The model does it much faster. I want to make a point about when people say that AIs aren't doing real reasoning or they don't have a real understanding of the world, is that this is a real distraction.

And the reason why is that they're trying to draw some moat around what makes us special. And...

The point is that when a simulation or simulacra gets good enough, it doesn't matter whether AIs are doing real empathy or real reasoning or real planning. If they can simulate it well enough that the economy can't figure out whether it's a human or an AI, then it will be real impact on society.

That's right. That's right. And it's not that the argument isn't fascinating. It really is. It's a fascinating conversation, but it completely bypasses a diverse energy from where the energy should be, which is the practical implications of what we even have now, which is already doing a lot of these. You can see the levels of automation that are already there and the implications of that.

And we just can't get distracted. Our threshold is like, okay, where are the impacts? Where are real things happening that we need to address right now? And so that's why we tend to move that part of the conversation aside and say, look, let's look at the impacts that are happening right now.

That's right. And whether you believe AGI is coming or not, there are tens of trillions of dollars going into improving the capabilities as quickly as possible to race towards replacing human beings behind computers with AIs because that's what's economically valuable. It's a $110 trillion game, right? That game is a $110 trillion game. And that is the actual game that these companies are in, right? People sometimes forget that because they think it's like,

We're in the chatbot game, right? Or we're in the Gen AI game. And the whole thing, the big pie is the one that everyone's looking at. Okay, so we've been walking through the arguments for and against, at a technical level, why general intelligence is a thing that will be discovered or invented in the next couple of years. But we haven't really talked yet about the stakes of what is it to birth a new technology.

intelligence, if you will, that is at the level or smarter than humans. So, Randy, I think it'll get a little philosophical, but let's talk about the implications and the stakes. So there's a few viewpoints on this, and maybe I'll give a few thoughts just to ground where I come from in these conversations. I kind of get back to what is happiness? What is the purpose of our lives? And I get back to the basics of like,

I would like everyone to have food, clothing, shelter, medicine, education, right? These things matter a lot. And millions and actually billions of people, right, don't have healthy access to these things. So this is kind of where I come from, like the beginning of when I enter into conversations about AI and alignment and, you know, how fast should we run and all of these things. That's my basis, right?

So with that said, I'm sure you've got some thoughts on these. And there's a bunch of avenues to explore here. Well, I think it's important to start. The founder of or co-founder of DeepMind, who's now part of Google, famously said as their mission statement that first solve intelligence, then use that to solve everything else. Strong AI and deep

Owning intelligence is the one ring of our time, right? The Tolkien one ring. Whoever owns that owns intelligence.

technical and scientific progress owns persuasive and cultural dominance, owns sort of the whole thing, right? You own all of the cognitive labor, all the thinking of the world. That is a very, very powerful thing. And that means it sets up the greatest incentive to race for it, regardless of the collateral damage along the way, because this is a winner-take-all war.

And I just want to set that up because this is how you get to Elon Musk saying things like, it increasingly appears that humanity is a biological bootloader for digital superintelligence. And anyone hearing that would say, well, then don't build it. We shouldn't replace ourselves. But then...

The next thing people will say is, well, we can't not build it because if we don't build it, then we'll lose to the people or the company or the country that does. And then you end up, like, when you actually talk to these kinds of accelerationists that are excited about this, they'll say things like, well, even...

If we lose control, which is sort of a funny thing to say because we actually haven't yet figured out how to control these systems, and they are starting to exhibit deception, self-preservation tendencies because it's trained on humans and human beings do those things. They say, even if it kills us all, it'll still be worth it because we created a god or it's still worth it because at least it was the US that created it. So it'll be US values that continues to live on. It

It's these kinds of things that people say. And I really want people to hear that this is not some fringe philosophy. So what Azad has described might sound outlandish, but these are real things. These are real philosophies. And it is hard for me personally to relate to because I'm much more interested in what happens to the humans and the animals and the environment around us. We have to take care of those things ourselves.

There's something that just goes back to food, clothing, shelter, medicine, education, like the things we need to take care of for people to not be suffering, right? To be reasonably happy. That we have some debt. I almost feel like it's a debt that you owe if you discover these kinds of technologies.

that you have to share them and you have to make sure they are distributed in a way that takes care of people. And actually, a lot of the AGI leaders are saying that too. They don't disagree with that. But when it comes to the details, it's always like, oh yeah, that's pretty complicated. And we're going to focus more of our energy on like, how do we keep advancing the technology?

This is a position that I think leaders are backed into because they don't want to lose the race. Because they don't want to lose the race, they are forced to take the position that, well, maybe that nihilistic, we're just a bootloader is the right position to take. Because if you take that position, it confers you power now. I think that's really important for people to understand. It's very convenient. It's very convenient. And it's not everyone's belief.

But it is the belief of some people that have a massive amount of capital and power for them to enact their worldview. So I think that's just really important for people to understand. And also part of that worldview is saying, hey, don't worry. Like when we go fast, yes, some bad things will happen. But

Things that are illegal are illegal. People just won't do them or we'll make sure they don't do them. And okay, so if we're going to say that, then what efforts are we putting into? What resources are we actually putting into making sure those bad things are actually illegal? Like you actually can't do them. And what happens a lot of the time is the rhetoric is there, but the allocation of actual resources, actual money, actual headcount is

to doing that research, figuring out those problems is not happening. One other consideration is that as AI systems become more powerful, they become harder to control because they have more degrees of freedom in the world, right? So whatever rules you thought you had set that were sufficient, they will find cracks in those rules. They will find ways to work around it, just like they're good at solving a problem you give them.

they will naturally tend to accrue resources or power or keep themselves on. These are sort of natural things that you would do to be successful in the world. And they will find ways to do that. And they will find ones that you have not thought of. As we integrate more and more of these technologies across society, right, they start to work well and we start to rely on them. And then we increasingly don't really understand where all the decisions are being made. And yet we've given up more and more power

to these systems, right? The technical term for this is called gradual disempowerment. So we actually build a situation where we as humans become highly disempowered, right? In the world that we actually, we created and we live in. And so there,

When AIs go off the rails for any reason, it could be just something, an inaccurate judgment that it makes, or it could be something more malicious or deceptive. For some reason, it decided to do something that we really don't want it to do. We're kind of screwed in that scenario because we don't even understand how the system's working at that point. This isn't...

academic or theoretical anymore. Anthropic released a paper where it showed that AI started to scheme when it learned that the anthropic programmers were going to retrain it to have a different set of values. It started to try to figure out how to copy itself to another server and also lie about what its answers should be so that anthropic researchers would think that it was being successfully retrained.

On top of that, there's a point that even the chains of thought, right? This is another recent research example. Even the chains of thought that the models generate, they look really good, right? When you look at them, you're like, wow, that sounds like exactly what it's thinking. And they're not. They are often...

Not even largely true. Sometimes it's less than 50% accurate, right, in terms of what they're actually saying. So that's another example where already we are in a situation where there's a lot of opaqueness to how the models work and a very rudimentary understanding of what is actually going on, even by some of the best researchers in the world who built these very products.

So I want to then just name, there's the alignment problem, which is can we get AIs to do what we want them to do? Then there's the polyalignment problem, which, you know, it's sort of coining a term here, but it's

The ability to align the sum total of all AIs to do what's good for humanity and the rest of the beings on this planet. The joke goes, like we're all talking about whether AI is conscious or not when it's not even clear that humanity is.

Which is to say that we as humanity, we keep getting results that nobody wants. No one really wants growing climate instability. And yet the nature of our political geopolitical system means that I don't burn the oil and you do. I get the industrialized nation. You don't. Therefore, I have to. And so we end up with climate instability.

Same thing with forever chemicals polluting the world and giving us all cancer. Things like this. We keep getting things we don't want. So if we just increase the power running through that system, because human beings haven't yet shown they're actually in control so we can steer the world the way we want, then that's another different way of saying we have lost control or lost the ability to decide.

And again, if we can't get simple versions of this to work now in 2025, when all of these problems are the simplest they're ever going to be, that doesn't bode well for the future, right? And so shifting attention to that and saying, how do we wrap our hands around these technologies right now? It's just crucial. And this is why the rhetoric that we must beat our foreign rivals to AI is important.

actually sort of missing the point. The competition can't just be to race towards making something that we can't control because there's a built-in implicit assumption that just like with guns and with airplanes, the more powerful you make it, the just as much in control we are. With AI, it's not like that. That the race needs to be for making a strengthened version of your society.

And whoever does that better wins. And we are not setting ourselves up right now to do the strengthening of our society versus just the generating power, which is uncontrollable. And there's a worthwhile principle here in these examples that Isa gave, right? Which is the more general purpose a technology is,

the harder it is to disentangle its benefits from its harms. That is why this generation of technology, whether it's automated cognition or physical, the AI stuff, the robotic stuff, all of that becomes very coupled in terms of benefits and harms because they're so flexible.

And that is why we have to do the societal upgrade that Isis is talking about. There's no other way to kind of responsibly wield these technologies. And the difference, of course, between AI and every other technology is that if you make technology that makes, let's say, rocketry better, that doesn't also make medical research better and mathematical advances better. But if you make advances in AI, because AI is fundamentally intelligence-based,

It means you get advances in rocketry and biomedical advances and in mathematics. You get them all. And so the rate of change that society is going to have to deal with

is going to be immense, greater than we have ever faced. And then it's not like it'll stop. It'll just keep going at a faster and faster rate. This is why it makes it the hardest problem that humanity has ever had to face and likely ever will.

And I think to do it, there is going to have to be some kind of international cooperation, which I'm just going to name it right now feels pretty much impossible. And we have some historical analogies for this. And, you know, Randy, you like to point out that there are no good historical analogies. This is unlike anything we've dealt with. Listen, each one has some flaw. I would say that.

Well, the obvious example, and with the caveat that none of these examples are going to be perfect analogies, the obvious one is, of course, nuclear weapons. Another place to look for hope here is blinding laser weapons. There was an international treaty signed in 1995 that banned blinding laser weapons in war. And the other one that goes along with that is germline editing, the changing of the human genome in a way that continues to

forward, that propagates. We as a species have successfully not walked down that technological road. The reason why we bring this all up is because it can often seem like if technology wants to bring humanity in some direction, technology wins. Humanity doesn't get to choose. But that's not always the case. And the times when it isn't the case is when

That thing which is valuable beyond which words can express about ourselves, if that is threatened in a visceral enough way, we can choose and we have chosen in the past different path. Don't think of this, though, as like hope washing. It's not like, and therefore we can do this. That's not what I'm saying. But it's just trying to like point at places where we can find non-naive hope, but we're going to have to collectively work very hard to

to get there. And I think there are some things we can put into place now. There are some really practical things. So these are things that I would love to see more energy on, especially from tech leaders. There are reasonable shared values we can build around, right? And don't kill, don't lie, don't steal, right? These are basic things that are shared across almost the entire human population. It's coming back to having this standard for ourselves and for the products we produce that

they espouse the values we would teach our children to be good citizens in the world. So that's one important thing. Then even more practically, get the incentives right. Think about what are the incentives driving when you do analysis. Think about that.

get price to fold in harms, right? Our economic system is built around this magic of price where price is this one number that coordinates a lot of different resources and it reflects information and reflects harms and it reflects this intersection of supply and demand.

All that magic works reasonably when price reflects all of the major costs. So if there's some damage being done, price needs to fold that in and then the system can kind of make the right decisions. So make sure we get harms back into price. Harms have to show up on company balance sheets. So that's a really important principle. I think if we can't get price to fold in harms, we have a big problem.

We tend to look a lot at GDP as the ultimate measure. But as power and wealth concentrate, GDP is going to be increasingly a bad measure of success because GDP going up will not correlate well with most people's actual experience. So we need to put a lot of attention on that and kind of figure out how are we going to solve those problems.

And then there's all these practical questions about what happens, right? As people get automated out to different degrees, this process is already beginning. How do people get food on the table? Like, how does that work? There's lots of different creative solutions people have come up with, but we need to really center those conversations. And I think the tech leaders...

have to see this as part of their responsibility. When they create these technologies that are of a really vastly different scale than any technologies before, these are general automation technologies, there are really big questions to answer and we just can't shut those off any longer. And while it seems possible

very, very challenging to impossible. It's very important to notice the gap that if every human being just stopped what they're doing, just sat down, we would never get AGI. We would never get ASI.

And so it's not like the laws of physics are pushing the bits and atoms out into the world that makes a uncontrollable superintelligence or sum total of all AIs that push humanity in the wrong direction. So it's not physically impossible. And I think that's so important to hold because now we're

The gap between not impossible and merely the excruciatingly difficult is a very important gap to hold because there is some kind of possibility in here. And the goal now is maximum clarity so we can all start to, in our own spheres of agency, move in the right direction.

So building on that, Isa, as we think about like at the highest level, when you kind of zoom out and say, okay, as a listener, what should I hold in my mind for like a framework for how we escape the AI dilemma? Here's one way I like to think of it.

There's five pieces. So one is we have to have a shared view of the problem and the path forward. At CHT, we spend a lot of time on this because it is the prerequisite for a lot of the other pieces to happen. So that's the first one, shared view of the problem and path forward. The second one is incentives and penalties. So when you do the right thing, you get rewarded. And when you do the wrong thing, you get penalized. This is back to that harms on balance sheets principle.

Paired with that is a kind of monitoring and enforcement. There has to be a way to know, did you do the right thing or not? And some certain appropriate levels of transparency that pair with that.

Then there's governance that can keep up with the pace of technology, right? Technology products shift. They're being updated all the time. Sometimes it's week by week. A lot of the updates are invisible. But there's major releases like at least every month. Is our governance keeping up with that? Like, how do we do that? We have to have systems that can get feedback from citizens, where we can make decisions, integrate large amounts of information and respond quickly.

And then the last thing is coordination at different levels. So that goes from the local to state level, to country level, to global coordination.

And these are all pieces that we are going to need, right, to kind of escape the dilemma. But if you want a simple framework, I think that's a good one to keep in mind. And the only thing I'd add to that is whenever people invoke the competition frame for why we have to race, the question that we need to ask, and you can ask back to whoever brings up like, but we have to win, is to ask the very simple question, but win at what? Are we winning at the pure competition?

power game for something that we don't yet know how to control? Or are we winning at strengthening our society so that our values win? If we don't get that clear, then the rest of the conversation will get stalled in a, but we have to win it.

Okay, so we started this conversation by talking about the question that I got asked sitting down at dinner with one of the leading sort of alignment safety researchers. Can you feel the AGI? And I think for most people, they're not really feeling the AGI yet. The future isn't equally distributed. When, Randy, do you think people are going to start to feel it in their lives? What is that going to look like? I think we should just briefly walk through that before we end.

Yeah, I mean, honestly, I think in my experience in presenting to people, it's just that they haven't been in direct contact with where the technology is today.

It's not even about being in contact with an imaginary future super capable AGI. It's just seeing today's state of the art. And when they see that, they can see. They see the implications very quickly for their lives, for their children, for their parents, for their grandparents. All of that stuff just comes crashing down very, very easily.

And I think it's just a matter of being curious and spending some time looking at the demos. I think we'll try to link some of them in the notes for this podcast so you can actually check out a few links and learn and feel that experience for yourself. You know, Randy, you and I are thinking and looking at this kind of stuff all the time. And it can be really challenging. You know, right now...

I am down in Costa Rica and my neighbor's son is 17, he's Costa Rican. And he was asking me yesterday, what should he study? And he's like, I, you know, he really wants to study engineering.

And it was hard for me to answer that question because I wanted to say, yeah, study engineering. Actually, you should study AI. So you set yourself up. But it's actually, it was a very hard question to answer because, you know, he's 17 or 18 by the time he gets through college. I actually don't think that'll have been the right thing for him to study. And this is, of course, a microcosm of the overall problem, which there isn't a good answer to that question right now. And it

Whatever age you are, yeah. Right, exactly. It's hard. I can sort of see the way to my own obsolescence in an economic sense. And I just want to be there with everyone. This can be very challenging. And to ask what is the solution to AI is like asking what species is a forest species.

It's a malformed question. It's going to be a messy, emergent ecosystems of things that let us steer. And we have to be okay with the not knowingness while also not letting that be an excuse for not doing anything. I think public pressure is going to play a huge role in this next cycle. Like how this all plays out really depends on public pressure, perhaps more than any other lever.

If you forced me to pick one, I think that's the one I would pick at this time. And I think there's a really practical thing you can do if you're looking for like, okay, what can I do? It is re-centering, re-calibrating conversations, pushing them back to the right questions again and again and again.

And when I think about the massive audience that listens to this podcast, if everyone just does that, right, on whatever social platform, whatever small groups, whatever private conversations, whatever company meetings you are at, if you ask those questions and just keep redirecting attention, we want to get in front of these problems, right? They get really scary. Right now, like it's not all unfolded.

So let's just act on it. Let's redirect those conversations, like have the right conversations consistently, which then translates into allocating the right resources and attention on the right problems consistently. I think that's how we give ourselves the best chance. Well, I think that's a wonderful place to end for today. Randy, thank you so much for coming back. It's always so much fun to have these conversations, even if the topic itself isn't particularly fun.

Although it is fascinating, and that's the confusing part. And thank everyone, too, for listening and giving us your undivided attention. Thanks, everybody. Your Undivided Attention is produced by the Center for Humane Technology, a nonprofit working to catalyze a humane future. Our senior producer is Julia Scott. Josh Lash is our researcher and producer. And our executive producer is Sasha Fegan. Mixing on this episode by Jeff Sudeikin.

Original music by Ryan and Hayes Holliday, and a special thanks to the whole Center for Humane Technology team for making this podcast possible. You can find transcripts for our interviews and bonus content on our sub stack and much more at humanetech.com. You can also watch all of our episodes on our YouTube channel. Just search for Center for Humane Technology. And if you liked this episode, we'd of course be grateful if you could rate it on Apple Podcasts and on Spotify.

It really does make a difference in helping others join this movement. And if you've made it all the way here, let me give you one more thank you for you giving us your undivided attention.