We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode The Simulation Within with Karl Friston

The Simulation Within with Karl Friston

2024/10/18
logo of podcast StarTalk Radio

StarTalk Radio

AI Deep Dive AI Chapters Transcript
People
G
Gary O'Reilly
K
Karl Friston
N
Neil deGrasse Tyson
以主持《宇宙:时空之旅》和《星谈》等科学节目而闻名的美国天体物理学家和科学传播者。
Topics
Neil deGrasse Tyson: 本集探讨了感知的根源、人工智能的智能性以及自由能原理。他认为物理学是所有学科的基础,神经科学也不例外。他提出了许多问题,例如大脑如何理解外部世界,以及如何将大脑的工作机制应用于机器。 Chuck Nice: 他与嘉宾Karl Friston讨论了主动推理和贝叶斯推理的概念,以及它们在认知神经科学中的作用。他还对人工智能的“幻觉”现象提出了疑问。 Gary O'Reilly: 他与嘉宾讨论了人工智能的智能性,以及人类与人工智能之间的区别。他还对人工智能的道德问题提出了担忧。 Karl Friston: 他介绍了自由能原理,这是一种描述事物自我组织方式的原则和方法。他解释了主动推理和贝叶斯推理的概念,以及它们在理解大脑运作和精神疾病中的作用。他还讨论了人工智能的局限性,以及如何将自由能原理应用于构建更可持续、更节能的人工智能系统。他认为,当前的大型语言模型虽然能够生成流畅的语言,但这并不意味着它们具有真正的理解力,并且它们在能源消耗方面也存在问题。他认为,更小、更节能的生物启发式方法更可取。 Neil deGrasse Tyson: 他强调了物理学在理解大脑运作中的重要性,并对人工智能的未来发展方向提出了思考。他认为,我们正从信息时代进入智能时代,而智能时代的一个关键问题是判断当前的生成式AI是否具有代理性。他还对人工智能获得意识并对人类做出不利于人类的决定的可能性表示担忧。

Deep Dive

Chapters
The episode begins with an introduction to the concept of the brain as a simulator of reality, drawing parallels between physics and neuroscience. The discussion introduces theoretical neuroscientist Karl Friston and his work on the free energy principle.
  • Physics principles can be borrowed by neuroscientists to understand brain function.
  • Karl Friston is a leading neuroscientist known for his work on neuroimaging and the free energy principle.
  • The free energy principle is a method to describe self-organization and adaptive behavior.

Shownotes Transcript

Translations:
中文

Discovering a better way to money? That's everyone's jam. No matter what goals you've got queued up, a Northwestern Mutual advisor can help uncover opportunities that others might overlook. Let's get started at NM.com, the Northwestern Mutual Life Insurance Company, Milwaukee, Wisconsin.

New markdowns up to 70% off are at Nordstrom Rack Stores now. From fresh kicks to perfect gifts, there's always a find. Levi's, Paige, Frame, all the best denim. Why do I rack? Because check out these boots. I always score at rack. Head to your Nordstrom Rack Store to find amazing deals on all things winter. Great brands, great prices. That's why you rack.

So guys, I was delighted to learn all the ways that principles of physics could be borrowed by neuroscientists to try to understand how our brain works. Because at the end of the day, there's physics and everything else is just opinion. If you say so yourself. I love how you get to say that.

But it's the physics of intelligence. The physics of neuroscience is really what that is about. As you've educated us, the physics is in everything. Yes. You just don't think of it as being in neuroscience. Yeah, because we've compartmentalized what people do as professional scientists

have their own textbook and their own journals and their own departments at universities, at the end of the day, we are one. It's all physics, people. At the end, that's the lesson we learn. Coming up, all that and more on StarTalk Special Edition. Welcome to StarTalk, your place in the universe where science and pop culture collide. StarTalk begins right now.

This is StarTalk Special Edition. Neil deGrasse Tyson here, your personal astrophysicist. And if this is Special Edition, you know it means we've got not only Chuck Nice. Chuck, how you doing, man? Hey, buddy. Always good to have you there as my co-host. And we also have Gary O'Reilly, former soccer pro, sports commentator. What's up, the crowd? Cheering him on. Yeah, that's the crowd at Tottenham. Tottenham, yeah. Crystal Palace, they're all...

Anytime you mention my name in this room, there's a crowd effect. Gary, you know, with Special Edition, what you've helped do with this branch of StarTalk is focus on the human condition and every way that matters to it. The mind, body, soul. It would include AI, mechanical augmentation to who and what we are. Robotics. So this fits right in.

to that theme. So take us where you need for this episode. In the age of AI and machine learning, we as a society and naturally as StarTalk are asking all sorts of questions about the human brain, how it works, and how we can apply it to machines.

One of these big questions being perception. How do you get a blob of neurons? I think that's a technical term. Yes, technical for sure. In your skull to understand the world outside. Our guest, Carl Friston, is one of the world's leading neuroscientists and an authority on neuroimaging, theoretical neuroscience, and the architect of the free energy principle.

Using physics-inspired statistical methods to model neuroimaging data, that's one of his big successes, he's also sought after by the people in the machine learning universe. Now, just to give you a little background on Carl, a neuroscientist and theoretician at University College London, where he is a professor, studied physics and psychology at Cambridge University in England.

and inventor of the statistical parametric mapping used around the world and neuroimaging plus many other fascinating things. He is the owner of a seriously impressive array of

of honors and awards which we do not have time to get into. And he speaks Brit. Yeah, so he's evened this out. There's no more picking on the Brit because there's only one of them. Okay. Carl Fritz, then welcome to StarTalk. Well, thank you very much for having me. I should just point out I can speak American as well. Please don't. Carl, that takes a certain level of illiteracy that I'm sure...

That you don't possess. Yeah, please don't stoop to our level. So let's start off with something. Is it a field or is it a principle or is it an idea that you pioneered, which is in our notes known as the free energy principle? I come to this as a physicist and there's a lot of sort of physics-y words that are floating or orbiting your work. And

And so in physics, they're very precisely defined. And I need to know how you are using these terms and in what way they apply. So let's just start off. What is the free energy principle? Well, as it says on the tin, it is a principle. And, um,

In the spirit of physics, it is therefore a method. So it's just like Hamilton's principle of least action. So it's just a prescription, a formal mathematical prescription of the way that things behave.

that you can then use to either simulate or reproduce or indeed explain the behavior of things. So you might apply the principle of least action, for example, to describe the motion of a football. The free energy principle has a special domain of application. It talks about the self-organization of things.

things where things can be particles, they can be people, they can be populations. So it's a method really of describing things that self-organize themselves into characteristic states. Wow.

Very cool.

So a ball rolls off a table onto the ground. It doesn't roll off the ground onto the table. So it seeks the minimum place. And my favorite of these is the box of morning breakfast cereal.

And it will always say some settling of contents may have occurred. Yeah. And you open up and it's like two-thirds full. Yeah, two-thirds of powder. Two-thirds of crushed shards of cornflakes. So it's finding sort of the lowest place in the Earth's gravitational potential. So why the need for this new term?

Well, it's an old term. I guess, again, pursuing the American theme, you can trace this kind of free energy back to Richard Feynman, probably his PhD thesis. So he was trying to deal with the problem of describing the behavior of small particles and invented this kind of free energy as a proxy that enabled him to evaluate the probability that a particle would take this path or that path.

So exactly the same maths now has been transplanted and applied not to the movement of particles but to what we refer to as belief updating. So it's lovely you should introduce this notion of nature finding its preferred state that can be described as rolling downhill to those free energy minima. This is exactly the ambition behind the free energy principle, but the preferred states here are

are states of beliefs or representations about a world in which something, say you or I, exist. So this is the point of contact with machine learning and artificial intelligence. So the free energy is not a thermodynamic free energy.

It is a free energy that scores a probability of your explanation for the world in your head being the right kind of explanation. And you can now think about our existence, the way that we make sense of the world and our behavior, the way that we sample that world.

as effectively falling downhill, settling towards the bottom, but in an extremely itinerant way, in a wandering way, as we sort of go through our daily lives at different temporal scales. It can all be described effectively as coagulating at the bottom of the cereal packet in our preferred states.

Wow. So you're, again, I don't want to put words in your mouth that don't belong there. This is just my attempt to interpret and understand what you just described. You didn't yet mention neurons, which are the carriers of all of this, or the transmitters of all of this thoughts and memories and interpretations of the world. So when you talk about the pathways that an understanding of the world takes shape,

Do those pathways track the nearly semi-infinite connectivity of neurons in our brains? So you're finding what the neuron will naturally do in the face of one stimulus versus another? That's absolutely right. In fact, technically, you can describe in neuronal dynamics the trajectory or the path of neurons

nerve cells firing exactly as performing a gradient descent on this variational free energy. So that is literally true, but I think more intuitively the idea is, in fact the idea you've just expressed, which is you can trace back possibly to the early days of cybernetics in terms of the good regulator theorem. The idea here is that to be well adapted to your environment, you have to be a model of that environment.

In other words, to interface and interact with your world through your sensations, you have to have a model of the causal structure in that world. And that causal structure is thought to be literally embedded in the connectivity among your neurons within your brain. So my favorite example of this is,

would be the distinction between where something is and what something is. So in our universe, a certain object can be in different positions. So if you told me what something is, I wouldn't know where it was.

Likewise, if you told me where something was, I wouldn't know what it was. That statistical separation, if you like, is literally installed in our anatomy. So, you know, there are two screens at the back of the brain, one dealing with where things are and one stream of connectivity dealing with what things are. However, we are pliable enough, though. And of course, I'm not pushing back. I'm just trying to further understand.

We're pliable enough, though, that if you were to say, go get me the thing, you

Okay. And then you give me very specific coordinates of the thing. I would not have to know what the thing is and I would be able to find it, even if there are other things that are there. Yep. And that speaks to something which is quite remarkable about ourselves, that we actually have a model of our lived world that has this sort of geometry that can be navigated because our presupposes that you've got a model of yourself, the

moving in a world and you know the way that your body works. I'm tempted here to bring in groins, but I don't know why. Chuck injured his groin a few days ago. That's why he's been talking about it. We've all heard about it since. Carl,

I hear the term active inference, and then I hear the term Bayesian active inference. Let's start with active inference. What is it? How does it play a part in cognitive neuroscience? Active inference, I think most simply put, would be an application of this free energy principle we're talking about. So it's a description or applying the maths to understand how we think

behave in a sentient way. So active inference is meant to emphasize that perception read as unconscious inference in the spirit of Helmholtz

depends upon the data that we actively solicit from the environment. So what I see depends upon where I am currently looking. So this speaks to the notion of active sensing. You went a little fast. I'm sorry, man. I'm trying to keep up here. But you went a little fast there, Carl. You talked about perception, perception,

being an inference that is somehow tied to the subconscious. But when you're, can you just do that again, please? And just to be clear, he's speaking slowly. Exactly. So it's not that he's going fast. No. Is that you are not keeping up. Well, listen, I don't have a problem. Okay. I have no problem not keeping up, which is why I have never been left behind, by the way. I have no problem keeping up because I go, wait a minute. So anyway, could you just like break that down a little bit for me?

Sure. I was trying to speak at a New York pace. My apologies. I'll revert to London. Okay. So let's start at the beginning. Sense-making, perception. How do we make sense of the world? We are locked inside. Our brains are locked inside a skull. It's dark in there. You can't see other than what information is conveyed by your eyes or by your ears or by your skin, your sensory organs.

So you have to make sense of this unstructured data coming in from your sensory organs, your sensory epithelia.

How might you do that? The answer to that, or one answer to that, can be traced back to the days of Plato through Kant and Helmholtz. So Helmholtz brought up this notion of unconscious inference. It sounds very glorious, but very, very simply it says that if inside your head you've got a model of how your sensations were caused,

then you can use this model to generate a prediction of what you would sense if this was the right cause, if you got the right hypothesis. And if what you predict matches what you actually sense, then you can confirm your hypothesis.

So this is where inference gets into the game. It's very much like a scientist who has to use scientific instruments, say microscopes or telescopes, in order to acquire the right kind of data to test her hypotheses about the structure of the universe, about the state of affairs out there as measured by her instruments.

So this can be described, this sort of hypothesis testing, putting your fantasies, your hypotheses, your beliefs about the state of affairs outside your skull to test by sampling data and testing hypotheses. This is just inference. So this is where inference gets into the game. These are micro steps en route to establishing an objective reality.

And there are people for whom their model does not match a prediction they might make for the world outside of them. And they would be living in some delusional, some world that you cannot otherwise agree to what is objectively true. And that would then be an objective measure of insanity or some other neurological disconnect. Really though? I mean...

Is it really? Well, if you project your own fantastical world into reality and you know it doesn't sit, but it's what you want, then that's a dysfunction. You're not working with, you're working against. But we live in a time now where that fantastical dysfunction...

actually has a place and talk to James Cameron for just a little bit and you'll see that that fast fantastical dysfunction has was a world-building creation that we see now as a series of movies so is it really so you know aberrant that it's you know a dysfunction or is it just different

What I think he's trying to create artistically rather than impose upon. Yeah, so Carl, if everyone always received the world objectively, would there be room for art at all? Ooh, that was a good question. Yep, really. Well done, sir. I'm going to say I think I was the inspiration for that question. Yes, Chuck inspired that question. So there's a role for each side of this question.

the perceptive reality correct no absolutely so just to pick up on a couple of those themes but that last point was i think quite key it is certainly the case of one application of one use of active inference is to understand psychiatric disorders so you're absolutely right when people a model of their lived world is not quite apt for the situation in which they find themselves

Say something changes, say you lose a loved one, so your world changes, so your predictions and the way that you sort of navigate through your day, either socially or physically, is now changed. So your model is no longer fit for purpose for this world.

As Chuck was saying before, the brain is incredibly plastic and adaptive. So what you can do is you can use the mislatch between what you predict is going to happen and what you actually sense to update your model of the world. And before I was saying that this is a model that...

would be able to generate predictions of what you would see under a particular hypothesis or fantasy. And just to make a link back to AI, this is generative AI. It's intelligent forecasting prediction under a generative model that is entailed exactly by the connectivity that we were talking about before in the brain. And it's the free energy principle manifesting when you readjust to the changes. And it's finding the new roots.

that are presumably the more accurate

your understanding of your world, the lower is that free energy state or is it higher or lower? What is it? It's lower. Yeah, that is absolutely right. So actually, technically, you know, if you go into the cognitive neurosciences, you'll find a big move in the past 10 years towards this notion of predictive processing and predictive coding, which again just rests upon this reimagining

mean that we are our brains are constructive organs generating from the inside out predictions of the sensorium and then the mismatch is now a prediction error that prediction error is then used to drive the neural dynamics that then allow for this revising updating my beliefs sort of such that my predictions now are

accurate and therefore the prediction error is minimized. The key thing is, to answer your question technically, the gradients of the free energy that drive you downhill just are the prediction errors. So when you've minimized your free energy, you've squashed the prediction errors. Absolutely. Excellent. You're not going to roll uphill unless there's some other change to your environment. Get away with friends to the laid-back Maryland coast.

Thank you.

This episode is brought to you by Opel, the first over-the-counter daily birth control pill available in the U.S. Opel is FDA approved, full prescription strength, and estrogen free. Plus, there's no prescription needed. Finally, the days of needing a prescription for birth control are over. Opel is available online and at most major retailers. Take control of your health and reproductive journey with Opel. Birth control in your control. Use code birthcontrol for 25% off your first month of Opel at opel.com.

We all want to see our loved ones living their best lives, but it's hard when they're struggling with drugs and alcohol. For nearly 70 years, Karen has made it possible for so many to imagine a life beyond addiction. We combine advanced neuroscience with life-changing care. Visit CARON.org. Karen, where the science of treatment meets the heart of care. Now in network with most insurances. That's CARON.org.

I'm Kais from Bangladesh and I support StarTalk on Patreon. This is StarTalk with Neil deGrasse Tyson. So if we think back to early mankind and the predictability. So I'm walking along, I see a lion in the long grass. What do I start to predict? If I run up a tree high enough, that lion won't get me. But if I run along the ground, the lion's probably going to get me. Is this kind of evolutionary that we've borne for survival?

Yes. Or have I misinterpreted this completely? No, no, I think that's an excellent point. Well, let's just think about what it means to be able to predict exactly what you would sense in a given situation and thereby predict also what's going to happen next.

If you can do that with your environment and you've reached the bottom of the serial packet and you've minimized your free energy, minimized your prediction errors, you now can fit the world. You can model the world in an accurate way. That just is adaptive fitness.

So if you look at this process now as unfolding over evolutionary time, you can now read the variational free energy or its negative as adaptive fitness. So that tells you immediately that evolution itself is one of these free energy minimizing processes. It is also, if you like, testing hypotheses about the kind of denizens of its environment.

the kind of creatures that will be a good fit for this particular environment. So you can actually read natural selection as, in statistics, would be known as Bayesian model selection. So you are effectively inheriting inferences or learning

transgenerationally in a way that's minimizing your free energy, minimizing your prediction errors. So things that get eaten by lions don't have the ability to propagate themselves through to the next generation so that everything ends up at the bottom of the cereal packet avoiding lions because those are the only things that can be there because the other ones didn't minimize their free energy.

Yeah, unless, Gary, you made babies before you said, I wonder if that's a lion in the bushes. But if they've got my genes, then there's a lion with their name on it. That's exactly right. I want to share with you one observation, Carl, and then I want to hand back to Gary because I know he wants to get all in the AI side of this. I remembered one of the books by Douglas Hofstadter. It might have been Gödel, Escher, Bach, or...

or he had a few more that were brilliant explorations into the mind and body. In the end of one of his books, he had, was an appendix, I don't remember, a conversation with Einstein's brain. And I said to myself, this is stupid. What does this even mean? And then he went in and described the fact that imagine Einstein's brain could be preserved at the moment he died. And all the neurosynaptic elements are still in place.

And it's just sitting there in a jar. And you ask a question. And the question goes into his ears, gets transmitted into the sounds that trigger neurosynaptic firings. It just moves through the brain. And then Einstein then speaks an answer. And the way that set up was interesting.

It was like, yeah, I can picture this sometime in the distant future. Now, maybe the modern version of that is you upload your consciousness and then you're asking your brain in a jar, but it's not biological at that point. It's in silicon. But what I'm asking is the information going into Einstein's brain in that thought experiment presumably is,

trigger his thoughts and then his need to answer that question because it was posed as a question. Could you just comment on that exercise, the exercise of probing a brain that's sitting there waiting for you to ask it a question? I mean, it's a very specific and interesting example of the kind of predictive processing that we are capable of because we're talking about language and communication here.

And just note the way that you set up that question provides a lovely segue into large language models. But note also that it's not the kind of embodied intelligence that we were talking about in relation to active inference because the brain is in a body, the brain is embodied. Most of what the brain is actually in charge of is moving the body.

or secreting. In fact, those are the only two ways you can change the universe. You can either move a muscle or secrete something. There is no other way that you can affect the universe. So this means that you have to deploy your body in a way to sample the right kind of information that makes your model as apt or as adaptive as possible. So Chuck, did you hear what he said? It means you cannot bend the spoon with your brain.

Right. Tell that to Yuri Geller. Right. Just to clarify. Okay. So what I was trying to hint at, because I suspect it's going to come up in later conversation, that there's, I think, a difference between a brain and a vata, a large language model, that is the embodiment of lots of knowledge. So one can imagine, say, a large language model being a little bit like Einstein's brain, but Einstein plus 100 words.

possibly a million other people and the history of everything that has been written. You can probe by asking it questions. And in fact, there are people whose entire career is now prompt engineers. AI prompts. It's funny, the people who...

program AI, then leave that job to become prompt. The people who are responsible for creating the best prompts to get the most information back out of AI. So it's a pretty fascinating industry that they've created their own feedback loop that benefits them. And now you can start to argue, where is the intelligence? Is it in the prompt engineer?

As a scientist, I would say that's where the intelligence is. That's where the sort of sensing behavior is. It's asking the questions, not producing the answers. That's the easy bit.

It's certainly asking, queering the world in the right way. And just notice, what are we all doing? What is your job? Is it asking the right questions? Carl, can I ask you this, please? Could active inference cause us to miss things that do happen? And secondly, does déjà vu fit into this? Yes and yes. In a sense, active inference...

is really about missing things that are measurable or observable in the right kind of way. So another sort of key thing about natural intelligence and be a good scientist

Just to point out that sort of noting the discovering infrared, that's an act of creation. That is art. So where did that come from? From somebody's model about the structure of electromagnetic radiation. So I think...

Just to pick up on a point we missed earlier on, creativity and insight is an emergent property of this kind of question and answer in an effort to improve our models of our particular world. Coming back to missing stuff, it always fascinates me that the way that we can move depends upon ignoring the fact we're not moving.

So I'm talking now about a phenomena in cognitive science called sensory attenuation. And this is the rather paradoxical or at least counterintuitive phenomena

that in order to initiate a movement, we have to ignore and switch off and suppress any sensory evidence that we're not currently moving. And my favorite example of this is moving your eyes. So if I asked you to sort of track my finger as I moved it across the screen and you moved your eyes very, very quickly,

While your eyes are moving, you're actually not seeing the optic flow that's being produced because you are engaging something called saccadic suppression. And this is a reflection of the brain very cleverly knowing that that particular optic flow that I have induced is fake news.

So the ability to ignore fake news is absolutely essential for a good navigation and movement of our world. Is it fake or just irrelevant to the moment? If it's the New York Times, it's definitely fake. Fake news. But it's not so much fake. It's just not relevant to the task at hand. Isn't that a different...

It's a subtle one for the simplicity of the conversation. And then I'm reading fake as irrelevant, imprecise. So it's like it's unusable. So your brain is just throwing it out, basically. Like, don't don't nothing to see here. So get rid of that. So, Neil, this is this is in your backyard rather more than mine. But isn't this where the matrix pretext kind of fits in that?

our perception might differ from what's actually out there and then perception can be manipulated or recreated? Well, I think Carl's descendants will just put us all in a jar. The way he's talking. Carl...

Cole, what does your laboratory look like? Full of jars. Well, yes. Well, there are several pods, and we have one waiting for you. Yeah, in the film The Matrix, of course, which came out in 1999, about 25 years, a quarter century ago, which is hard to believe. What? It was very candid sense that your brain's reality is the reality you think of and understand, and it is not receiving external input.

All that your brain is constructing is detached from what's exterior to it.

And if you've had enough lived experience, or maybe in that future that they're describing, the brain can be implanted with memory. It reminds me, what's that movie that Arnold Schwarzenegger's in about Mars? Total Recall. Total Recall, thank you. Get your ass to Mars. Instead of paying thousands of dollars to go on vacation, they would just implant the memories of a vacation in you. And bypassing the sensory conduits

into your brain. Of course, these are movies and they're stories and it's science fiction. How science fiction-y is it really? Well, I certainly think that the philosophy behind, I think, probably both Total Recall, but particularly The Matrix,

I think that's very real and very current. Just going back to our understanding people with psychiatric disorders or perhaps people who have odd views, world views, to understand

that the way that you make sense of the world can be very different from the way I make sense of the world, dependent on my history and my predispositions and my prize, what I have learned thus far, and also the information that I select to attend to. So just pursuing this theme of

ignoring 99% of all the sensations. For example, Chuck, are you thinking about your groin at the moment? I would guarantee you're not, and yet it is generating sensory impulses from the nerve endings.

But you, at this point in time, were not selecting that. So the capacity to select is, I think, a fundamental part of intelligence and agency. Because to select means that you are not attending to or selecting 99% of the things that you could select. So I think the notion of selection

is a hallmark of truly intelligent behavior. Are you analogizing that to large language models in the sense that it could give you gibberish, it could find crap anywhere in the world that's online, but because you prompted it precisely, it is going to find only the information necessary and ignore everything else?

Yes and no, but that's a really good example. So the yes part is that the characteristic bit of architecture that makes large language models work, certainly those that are implemented using transformer architectures, are something called attention heads.

So it is exactly the same mechanism, the same basic mechanics that we were talking about in terms of attentional selection that makes transformers work. So they select the recent past in order to predict the next word. That's why they work, to selectively pick out something in the past, ignore everything else.

to make them work. When you talk about that probability in an LLM, that probability is a mathematical equation that happens for like every single letter that's coming out of that model. So it is literally just giving you the best probability of what is going to come next.

Okay. Whereas when we perceive things, we do so from a worldview. So for an LLM, if you show it a picture of a ball with a red stripe that's next to a house, okay, and say, that's a ball, and then show it a picture of a ball in the hands of a little girl who's bouncing it,

It's going to say, all right, that might be a ball. That may not be a ball. Whereas if you show even a two-year-old child, this is a ball, and then take that ball and place it in any circumstance, the baby will look at it and go, thaw, thaw. So there is a difference in the kind of intelligence that we're talking about here.

Yeah, I think that's spot on. That's absolutely right. And that's why I said yes and no. So that kind of fluency that you see in large language models is very compelling and it's very easy to give the illusion that these things have some understanding or some intelligence, but they don't have the right kind of generative model understanding.

underneath to be able to generalize and recognize a ball in different contexts the way that we do. Well, it would if it was set up correctly. And that setup is no different from you looking at reading the scene. I mean, a police officer does that busting into a room. You know, who's the perpetrator, who's not, before you shoot. There's an instantaneous awareness factor that you have to draw from your exterior stimuli. And so...

Because, you know, I'm reminded of here, Carl, I saw one of these New Yorker-style cartoons where there are two dolphins swimming in one of these water, you know, parks, right? And so they're in captivity. But the two dolphins are swimming, and one says to the other, of the person walking along the pool's edge, those humans, they face each other and make noises, but it's not clear they're actually communicating. It's not clear.

And so who are we to say that the AI large language model is not actually intelligent if you cannot otherwise tell the difference? Who cares how it generates what it is? If it gets the result that you seek, you're going to say, oh, well, we're intelligent and it's not? How much of that is just human ego speaking? Well, I'm sure it is human ego speaking, but in a technical sense. I think...

Okay, there's a loophole you're saying because I'm not going to say that bees are not intelligent when they do their waggle dance telling other bees where the honey is. And I'm not going to say termites are not intelligent when they build something a thousand times bigger than they are when they make termite mounds and they all cooperate. I'm fatigued by humans'

Trying to say how special we are relative to everything else in the world that has a brain when they do stuff We can't let me ask you then. So what's the common theme between the termite and the bee and And the policeman reading the scene. What do they all have in common all of those three things move and

Whereas a large language model doesn't. Doesn't. So that brings us back to this action, the active part of active inference. So the note of the question about large language models and attention was that large language models are just given everything.

They're given all the data. There is no requirement upon them to select which data are going to be most useful to learn from. And therefore, they don't have to build expressive, fit-for-purpose world models or generative models. Whereas your daughter, the two-year-old daughter playing with the beach ball, would have to, by moving and selectively reading the scene, by moving her eyes,

by observing her body, by observing balls in different contexts, build a much deeper, appropriate world or geriatric model that would enable her to recognize the ball in this context and that context and ultimately tell her father, I'm playing with a ball. Get away with friends to the laid-back Maryland coast, where you can catch up while casting off and hang ten while hanging out.

Where a day on board is never boring and full throttle is half the fun. Where you can sink a putt, raise a glass, and there's always room for one more round. Ocean City, Maryland. Somewhere to smile about. Book your trip at ococean.com.

This episode is brought to you by Opel, the first over-the-counter daily birth control pill available in the U.S. Opel is FDA approved, full prescription strength, and estrogen free. Plus, there's no prescription needed. Finally, the days of needing a prescription for birth control are over. Opel is available online and at most major retailers. Take control of your health and reproductive journey with Opel. Birth control in your control. Use code birthcontrol for 25% off your first month of Opel at opel.com.

My dad works in B2B marketing. He came by my school for career day and said he was a big ROAS man. Then he told everyone how much he loved calculating his return on ad spend. My friend's still laughing me to this day. Not everyone gets B2B, but with LinkedIn, you'll be able to reach people who do. Get $100 credit on your next ad campaign. Go to linkedin.com slash results to claim your credit. That's linkedin.com slash results. Terms and conditions apply. LinkedIn.com slash results.

LinkedIn, the place to be, to be.

So we had a great show with Brett Kagan, who mentioned your free energy principle. And in his work, creating computer chips out of neurons, what people call organoid intelligence, what he was calling synthetic biological intelligence. And that's in our archives. In our recent archives, actually. Recent archives, yeah. Do you think the answer to AGI is a biological solution, a mechanical solution, or a mixture of both? And remind people what AGI is.

Artificial general intelligence. I know that's what the words stand for, but what is it? You're not asking me for the answer. Don't ask me either. Seriously, I've been told off for even using that acronym anymore because it's so ill-defined and people have very different readings of it. So open AI has a very specific meaning for it. If you talk to other theoreticians, they would represent it.

I think what people are searching for is natural intelligence. It's natural.

Chuck, just coming back to your previous theme, notice we're talking about behaving systems, systems that act and move and can select and do their own data mining in a smart way as opposed to just ingesting all the data. So what I think people mean when they talk about superintelligence or generalized AI or artificial intelligence, they just mean natural intelligence.

They really mean us. It's our brain. Our brain, if you want to know what AGI is, it's our brain. If it was actually our brain, it would be natural stupidity. Well, that too. Our brain without the stupidity. That's really what it is. So back in December 22, you dropped a white paper titled Designing Ecosystems of Intelligence from First Principles.

Now, is this a roadmap for the next 10 years or beyond or to the Terminator, ultimate destination? And then somewhere along the line, you discussed the thinking behind a move from AI to IA, and IA standing for Intelligent Agents.

which seems a lot like moving towards the architecture for sentient behavior. Have I misread this in any way? No, you've read that perfectly. So that white paper was written with colleagues in industry, particularly versus AI, exactly as a kind of roadmap that those people who were committed to a future of artificial intelligence

that was more sustainable, that was explicitly committed to a move to natural intelligence and all the biomimetic moves that you would want to make, including implementations on neuromorphic hardware, quantum computation of photonics, all those efficient approaches that

would be sustainable in the sense of climate change, for example. But also speaking to Chuck's notion about efficiency, efficiency is also, if you like, bait into natural intelligence in the sense that if you can describe

intelligent behavior as this falling downhill, pursuing free energy gradients, minimizing free energy, getting to the bottom of the serial packet. You're doing this via a path of least action. That is the most efficient way of doing it. Not only informationally,

but also in terms of the amount of electricity you use and the carbon footprint you leave behind. So from the point of view of sustainability, it's important we get this right. And so part of the theme of that white paper was saying there is another direction of travel. Away from large language models, large is in the title. It's seductive, but it's also very dangerous. It shouldn't be large. It should be the size of a bee. So to do it biologically...

You should be able to do it much more efficiently. And of course, the meme here is that our brains work on 20 watts, not 20 kilowatts. And we do more than any large language model. We have low energy intelligence. We do. Efficient. I guess that's a way to say it. I've seen you quoted, Carl, as saying that we are coming out of the age of information.

and moving into the age of intelligence. If that's the case, what is the age of intelligence going to look like? Or have we already discussed that? Well, I think we're at its inception now, just in virtue of all the wonderful things that are happening around us and the things that we are talking about. We're asking some of the very big questions about what is happening and what will happen over the next decade.

I think part of the answer to that lies in your previous nod to the switch between AI and IA. So IA brings agency into play. So one deep question would be, is current generative AI an example of

agentic? Is it an agent? Is a large language model an agent? And if not, then it can't be intelligent and certainly can't have generalized intelligence. So what is definitive of being an agent? I put that out there as a question, half expecting a joke. I've got Agent Smith in my head, if anyone can take that and run with it. There you go. It's right about now where you hear people commenting on the morality of a decision.

and whether a decision is good for civilization or not. And everybody's afraid of AI achieving consciousness and just declaring that the world will be better off without humans. And I think we're afraid of that because we know it's true. Yeah, I was going to say, we've already come to that conclusion. That's the problem. Okay, Carl, is consciousness the same as self-awareness?

Yeah, there are lots of people who you could answer that question of and get a better answer. I would say the purpose of this conversation, probably not, no. I think to be conscious, certainly to be sentient and to behave in a sentient kind of way would not necessarily imply that you knew you were a self. I'm pretty sure that a bee doesn't have self-awareness, but it still has sentience. It still has experiences and has plans,

and communicates and behaves in an intelligent way.

And you could also argue that certain humans don't have self-awareness of a fully developed sort. I'm talking about very severe psychiatric conditions. So I think self-awareness is a gift of a particular, very elaborate, very deep generative model that not only entertains the consequences of my actions, but

but also entertains the fantasy or hypothesis that I am an agent and that I am self and can be self-reflective in a sort of metacognitive sense. So I think I'd differentiate between self-aware and simply being capable of sentient behavior. Wow, that is great. Let me play skeptic here for a moment.

mild skeptic. You've described, you've accounted for human decision-making and behavior with a model that connects our sensory, the sensory conduits between what's exterior to our brain and what we do with that information as it enters our brain. And you've applied this free energy gradient that this information follows. That sounds good. It all sounds fine.

I'm not going to argue with that, but how does it benefit us to think of things that way? Or is it just an after the fact pastiche on top of what we already knew was going on, but now you put fancier words behind it? Is there a predictive value to this model? Or is the predictivity in your reach because when you assume that's true, you can actually make it happen in the AI marketplace?

Yeah, I think that's the key thing. So, I mean, when I'm asked that question, or indeed when I asked that question of myself, I sort of applied to things like Hamilton's principle of least action.

Why is that useful? Well, it becomes very useful when you're actually sort of building things. It becomes very useful when you're simulating things. It becomes useful when something does not comply with Hampton's principle of least action. So just to unpack those directions that travel in terms of applying the free energy principle, that means that you can write down

the equations of motion, and now you can simulate self-organization that has this natural kind of intelligence, this natural kind of sentient behavior. You can simulate it in a robot, in an artifact, in a terminator should you want to, although strictly speaking, that would not be compliant with the energy principle. But you can also simulate it in silico and make digital twins of people and objects

choices and decision-making and sense-making. And once you can simulate, you can now use that as an observation model for real artifacts and start to phenotype, say, people with addiction or, say, people who are very creative or, say, people who had schizophrenia.

So if you can cast aberrant inference or false inference, believing things are present when they're not, or vice versa, as an inference problem, and you know what the principles of sense-making and inference are, and you can model that in a computer, you can now get a stamp it in which you can now

not only phenotype by adjusting the model to match somebody's observed behavior, but now you can go and apply synthetic drugs or do brain surgery in silico. So there are lots of practical applications of knowing how things work. Well, when I say things work, how things behave. That presumes that your model is correct. For example, just a few decades ago, it was presumed, and I think no longer so, that our brain functioned via neural nets.

neural networks where it's a decision tree and you slide down the tree to make an ever more refined decision. On that assumption, we then mirrored that in our software

to invoke neural net decision-making in my field, in astrophysics. How do we decide what galaxy is interesting to study versus others in the millions that are in the data set? You just put it all into a neural net that has parameters that select for features that we might, in the end of that effort,

determined to be interesting. We still invoke that, but I think that's no longer the model for how the brain works. But it doesn't matter. It's still helpful to us. You're right. And honestly, that is now how AI is organized around the new way that we see the brain working. Yeah. And why is the brain the model of what should be emulated?

I mean, the human physiological system is rife with baggage, evolutionary baggage. Much of it is of no utility to us today except sitting there available to be hijacked by advertisers or others who will take advantage of some feature we had 30,000 years ago when it mattered for our survival. And today, it's just dangling there waiting to be exploited. So...

A straight answer to your question. The free energy principle is really a description or a recipe for self-organization of things that possess a set of preferred or characteristic states, coming right back to where we started, which is the bottom of the cereal packet. If that's where I live, if I want to be there, that's where I'm comfortable, then I can give you a calculus that will, for any given situation,

prescribe the dynamics and the behavior and the sense-making and the choices to get you to that point. It is not a prescription for what is the best place to be or what the best embodied form of that being should be. It's saying that if you exist and you want to exist in a sustainable way,

where it could be a speech, or it could be a meme. In a given environment, yes, in a given setting. Yeah, it's all about the relationship. That's a really key point. So the variational free energy that we've been talking about, the prediction error, is a measure of the way that something couples to its universe or to its world. It's not a statement about a thing in isolation. It's the fit.

It's, you know, again, if you just take the notion of prediction error, there's something that's predicting and there's something being predicted. So it's all relational. It's all observational. It's a measure of adaptive fitness. That's an important clarification. Yes. Carl, could you give us a few sentences on Bayesian inference? That's a new word to many people who even claim to know some statistics. That's a, it's a way of using, um,

what you already know to be true to help you decide what's going to happen next. Are there any more subtleties to a Bayesian inference than that? I think what you just said captures the key point. It's all about updating. So it's a way of describing

inference by which people just mean estimating the best explanation probabilistically, a process of inference that is ongoing. So sometimes this is called Bayesian belief updating, updating one's belief in the face of new data. And how do you do that update in a mathematically optimal way? You simply take

the new evidence, the new data. You combine it using Bayes' rule with your prior beliefs established before you saw those new data to give you a belief afterwards, sometimes called a posterior belief. Because otherwise you would just come up with a hypothesis assuming you don't know anything about the system and that's not always the fastest way to

get the answer. Yes, you could argue that it isn't important. You can't do it like that. It has to be a process. It has to be a path through some beneath space. You're always updating, whether it's at an evolutionary scale or whether it's during this conversation. You can't start from scratch. And you're using the word belief the way

Here, stateside, we might use the word what's supported by evidence. So it's not that I believe something is true. Often the word belief is just, well, I believe in Jesus or Jesus is my savior or Muhammad. So belief is I'll believe that no matter what you tell me because that's my belief. And my belief is protected constitutionally.

On those grounds, when you move scientifically through data and more data comes to support it, then I will ascribe confidence in the result.

is measured by the evidence that supports it. So it's an evidentiary supported belief. Yeah. I guess if we have to say belief, what is the strength of your belief? It is measured by the strength of the evidence behind it. Yeah, that's how we have to say that. So Gary, do you have any last questions before we got to land this plane? Yeah, I do. Because if I think about us as humans, we have...

Sadly, some of us have psychotic episodes, schizophrenia. If someone has hallucinations, they have a neurological problem that's going on inside their mind. Yet we are told that AI can have hallucinations.

I don't know. Does AI have mental illness? AI just learned to lie. That's all. You know, you ask it a question, it doesn't know the answer, and it's just like, all right, well, how about this? That's what we do in school, right? You don't know the answer. You make something up, it might be right. Right, exactly. What's the answer? Ah, rockets? Oh, yeah.

Okay. Yeah, I was speaking to Gary Marcus in Davlos a few months ago, and he was telling me he invented the word or applied the word hallucination in his context. And it became word of the year, I think, in some circles.

And I think he regrets it now because the spirit in which he was using it was technically very divorced from the way that people hallucinate. And I think it's a really important question that, you know, theoreticians and neuroscientists have to think about in terms of understanding false inference in a brain. And just to pick up on...

Neil's point, when we talk about beliefs, we're talking about sub-personal, non-propositional Bayesian beliefs that you wouldn't be able to articulate. These are the way that the brain encodes probabilistically the causes of its sensations. And of course, if you get that inference process wrong, you're going to be subject to inferring things are there when they're not,

which is basically hallucinations and delusions, or inferring things are not there when they are. And this also happens to some of us in terms of neglect syndromes, dissociative syndromes, hysterical syndromes. These can be devastating conditions where you've just got the inference wrong. So understand the mechanics of this failed inference. I think, for example, hallucination is absolutely crucial. It usually tracks back to what we were talking about before in terms of the ability to select things

versus ignore different parts of the data. So if you've lost the ability to ignore stuff, then very often you preclude an ability to make sense of it because you're always attending to the surface structure of sensations. Take, for example, severe autism. You may not get past the bombardment of emotions

sensory input in all modalities of all parts of the scene on all parts of your sense. It's all alive, right? It's all alive guys. I think we've got to call it quits there. Carl, this has been highly illuminating. Yes, good stuff. And what's interesting is as much as you've accomplished thus far, we all deep down know it's only just the beginning and who knows where the next year, much less five years will take this.

It'd be interesting to check back in with you and see what you're making in your basement. With the Brit, Neil, it's garage. Oh, garage. You guys don't have basements. The basements is more the garage. We go out there and create lots of wonderful things. Exactly. Exactly.

Okay, Professor Carl, thanks for joining us. Well, thank you very much for the conversation, the jokes particularly, those are the most painful ones I've ever done. Thanks for joining us from London. Thank you. Time shifted from us here stateside. Again, we're delighted that you could share your expertise with us in this StarTalk special edition.

All right, Chuck, always good to have you, man. Always a pleasure. All right, Gary. Pleasure, Neil. Thank you. I'm Neil deGrasse Tyson. You're a personal astrophysicist, as always bidding you to keep looking up. Get away with friends to the laid back Maryland coast where you can catch up while casting off and hang 10 while hanging out where a day on board is never boring and full throttle is half the fun.

where you can sink a putt, raise a glass, and there's always room for one more round. Ocean City, Maryland. Somewhere to smile about. Book your trip at ococean.com.