We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Walter Sinnott Armstrong on AI and Morality

Walter Sinnott Armstrong on AI and Morality

2024/6/14
logo of podcast Philosophy Bites

Philosophy Bites

AI Deep Dive AI Insights AI Chapters Transcript
People
W
Walter Sinnott Armstrong
Topics
Walter Sinnott Armstrong: 人工智能的定义应广泛涵盖机器学习过程,即机器设定目标并学习改进实现目标的方法。构建人工智能道德体系面临挑战,单一道德体系(如功利主义或康德主义)难以满足所有人的需求。从下往上的方法,即从互联网数据中学习人类道德行为和价值观,存在吸收人类偏见的风险,且难以解释人工智能的决策过程。因此,最佳方法是结合自上而下和自下而上的方法,即混合方法。该方法需要识别道德相关的特征,并通过冲突情境来收集人们的意见,从而构建预测模型。该模型允许一定程度的地域差异,但必须设置限制以避免不公正,例如种族歧视。即使人工智能给出了道德决策,个人仍然可以表达不同意见并试图说服他人。在专业领域,专家的意见应该比普通人的意见更有权重,因为他们拥有更多相关知识。但如果专家和普通人都拥有充分的信息,那么他们的意见应该具有同等权重。人工智能在道德决策中可以作为辅助工具,帮助医生做出更明智的决定,而不是直接取代医生的判断。人工智能可以为其道德决策提供理由,方便医生进行审查和讨论,从而避免道德错误。许多道德错误是可以避免的,人工智能可以帮助人们识别并纠正这些错误。将人工智能应用于医疗道德决策还需要时间进行完善,预计在未来十年内可能实现广泛应用。人工智能在道德决策方面的应用范围很广,例如在医疗、招聘和军事等领域都有潜力。 David Edmonds: (主要为引导性问题,未形成独立的论述)

Deep Dive

Key Insights

How does Walter Sinnott Armstrong define artificial intelligence?

Artificial intelligence is broadly defined as occurring whenever a machine learns something, as learning involves intelligence. It often involves the machine being given a goal and learning new and better means to achieve that goal.

What is the challenge of programming AI with human morality?

The challenge lies in choosing which moral principles to program into the AI, as different ethical systems like utilitarianism and Kantian ethics conflict. There is no consensus on which moral system should dictate the AI's decisions.

What is the hybrid approach to introducing ethics into AI?

The hybrid approach combines top-down principles with bottom-up data collection. It involves asking people which moral features matter in a situation, refining those features, and building conflicts to train the AI to predict human moral judgments.

How does AI handle moral dilemmas like kidney allocation?

AI collects features that matter to people, such as age, dependents, or criminal records, and predicts which patient should receive a kidney based on these factors. It can also confirm or challenge a doctor's decision, aiding in the decision-making process.

What are the limitations of using AI for moral decision-making?

AI can inherit human biases from data, and its decision-making process is often a 'black box,' making it difficult to understand the reasoning behind its conclusions. Additionally, local values and expertise must be considered to ensure fairness.

How far are we from using AI for ethical decisions in hospitals?

While AI is already used in some kidney transplant centers for medical efficiency, integrating moral considerations is still in development. Walter Sinnott Armstrong estimates it could take about 10 years for such systems to be refined and widely adopted.

What are potential applications of ethics in AI beyond healthcare?

AI can be applied to dementia care, hiring decisions to ensure fairness regarding gender and race, and even military operations. These applications aim to introduce moral considerations into various decision-making processes.

Chapters
This chapter explores the challenges of programming human morality into AI, discussing the limitations of top-down and bottom-up approaches. It introduces the core problem of defining AI and integrating ethics into its decision-making processes, highlighting the complexities and biases involved.
  • Defining AI broadly as machines that learn and achieve goals.
  • Challenges of top-down (pre-programmed ethics) and bottom-up (learning from internet data) approaches.
  • The issue of human biases in AI algorithms.
  • The "black box" problem of understanding AI decision-making processes.

Shownotes Transcript

Translations:
中文

This is Philosophy Bites with me, David Edmonds. And me, Nigel Warburton. If you enjoy Philosophy Bites, please support us. We're unfunded and all donations will be gratefully received. For more details, go to www.philosophybites.com. Can AI, artificial intelligence, help us make practical ethical decisions?

The philosopher Walter Sinnott Armstrong thinks so, and he's been working with a data scientist and a computer scientist to try to build a system that will be of use to doctors faced with ethical dilemmas. Walter Sinnott Armstrong, welcome to Philosophy Bites. Thank you so much for having me. It's a joy.

We're going to discuss today how human morality can be introduced into AI. But I want to start with a very basic question because people seem to define AI in all sorts of different ways. What's your definition of artificial intelligence?

I think artificial intelligence should be defined very broadly. It occurs whenever a machine learns something, because learning involves intelligence. And in particular, often in AI systems, the machine is given a certain goal, and it learns new and better means to achieve that goal. That's when artificial intelligence occurs.

So it involves learning. A crucial component of AI is that the machine or the algorithm learns as it proceeds. Exactly. And also that it has a goal and it tries out different means to that goal, tests which means are working best, and then finds new and better ways to achieve those goals. So I want to program my AI with human morality. It sounds easy enough. If I'm a utilitarian, I just program it to maximize efficiency.

or minimise suffering. If I'm a Kantian, I programme Kantian ethics into the machine. What's wrong with that? Well, what's wrong with that is the Kantians won't like the utilitarians and the utilitarians won't like the Kantians. Who made you the dictator who got to tell everybody else which moral system should be built into the machine? The problem of choosing which principles to build into the machine is basic and difficult.

So if we can't do it that way, how else can we introduce ethics into AI? Well, some people say, let's scour the internet and look for all of the posts that people have put on the internet far and wide. Look at their values from what they buy. Look at the language they use in their emails and online chat rooms and so on. And then use the AI system to further

find out what a human would say. So instead of being top down, as it were, defining what principles we want, that's bottom up. We scour the internet, see how people actually behave or think morally, and build an algorithm based on those judgments and values that we find out there.

Exactly. And so if you look at something like ChatGPT, it predicts what a human would say in response to the question.

But the answer that it gives might not be right. It might not be justified. Sure, humans would say that, but that just shows the weakness of humans. I can see that because, for example, if you hoover up human judgment, you're also hoovering up all human biases, which may include being racist and sexist and all the other flaws that we have ethically.

The other problem is that you don't know why it's giving that answer. It gives that answer, but what are the reasons that are really driving it? A lot of these deep learning techniques that are used now are basically black boxes and you don't know what's going on behind the scenes. So top-down doesn't work, bottom-up doesn't seem to work. What's the alternative?

Well, both, and make them work together. We call it a hybrid. But the idea is that you need some principles, but you don't want to impose your own principles. So you ask people, what are the features that really matter to you in a moral situation?

which aspects of the agent or the victim in a particular action matter morally to you. And then we refine those features and put them together in conflicts. Let me get concrete.

Suppose that there's one kidney that's available because there was a car accident and the victim was a kidney donor, an organ donor. And there are two people who need the kidney. We want to ask which features of the patient matter to which of the two patients should get the kidney.

And people will say things like, well, how long they've been on a waiting list or how long they're going to live if they get the kidney or how many children they have at home or whether they have a criminal record, things like that. And we collect those features that matter to people and a few of our own, recheck them to make sure people agree that they matter, and then build conflicts. Well, what if patient A is 31 years old and patient B is 50 years old?

Well then some people would say we ought to favor patient A. But what if patient A has no children but patient B has two children that depend on them? Then you can help three people by helping the one with children. So you ask people enough of these conflicts and then the AI system can predict what they'll say in a brand new set that they've never seen before. When you say what they'll predict...

You mean what a consensus will predict, because you and I may differ. We may have all the same facts and we may still differ about who we think should get the kidney when there's only one kidney going. You may be in the majority. I may be in the minority. Does that mean we go with your choice?

No. In some of these medical situations, I think you should go with what the local values are. Now, the values in North Carolina, where I live, might be very different than the values in London, where you live. And so it's fine if one hospital has slightly different values than the other. It's not fine if in North Carolina they say blacks should never get the kidney. So there are limits.

But I think it's fine to have some variation that local communities don't always have exactly the same rules.

How do you spot what are the relevant limits? Why is race irrelevant, but some other considerations acceptable so that in North Carolina they can come to one conclusion, in London they can come to another? Well, I'd have to give you all the arguments for our racism is bad. And, you know, it's interesting, we talked about Kantians and utilitarians. They agree on that, that it's bad.

And when we do our surveys, we find that over 95% agree that race should not matter. And the ones that say it should matter often say, well, you should favor blacks because they've had so many other problems before that.

So we get an algorithm that produces certain results for North Carolina, but I live in North Carolina and I still don't agree with the result that the algorithm has thrown up. Where does that leave me? Well, that leaves you trying to convince other people. But that's going to be happening regardless of whether it's an AI that picks the patient or a doctor that picks the patient. These are difficult situations and there's never going to be an answer that satisfies everyone completely.

Should you weight it according to expertise? I imagine that when it comes to how kidneys are distributed, that the person on the street may have a very different view.

to the kidney surgeon, and it seems sensible to allow the kidney surgeon two votes or three votes compared to the one vote of the ordinary person on the street. Sure, that makes sense for the simple reason that the surgeon knows a lot more facts about kidney exchanges. When I went into this, I had no idea how many hours a week you spent on dialysis when you're waiting for a kidney transplant.

I thought it was maybe two or three. Turns out it can be six hours a day, six days a week. And that makes a big difference to whether you can take care of your kids and whether you can keep your job and so on. So if people are not aware of those facts and they would make different judgments if they knew those facts, then you want to favor the judgments of the people who know the facts.

And that is often going to be the experts in the hospital system. But we can also look to the general public and correct their mistakes by informing them, by seeing whether they stick with the judgments. When they're given better information, they better understand what the situation is. Then I don't see any reason to favor the expert over the person in the general population if they're both equally well-informed.

How in practice is the AI going to help in this particular case? So we feed on all the data and the AI will throw up a result which we hope everybody can then live by. Yes, we hope that everyone can live by it, but also it can confirm. So if, for example, a doctor says, I think it should go to Joe, the one kidney that's available should go to the patient Joe.

Doctors are not always completely confident. And if the AI says, yes, it should go to Joe, then that's confirmation and helps the doctor. But if the doctor says it should go to Joe and the AI says, no, it should go to Sally, now we realize maybe we need to look at this more carefully and ask some other people to input. So this system doesn't necessarily dictate the final answer, but it can be helpful in the process.

Do you think what would actually happen is that we would subcontract the decision-making to the AI? Because we see that if we're driving a car and our sat-nav tells us to go left and we would normally turn right, we quickly learn if we turn right, there's a big traffic jam and we should have followed what the AI said. I mean, in practice, aren't doctors going to look at what the AI determines and just go along with it?

So they might or might not. I don't know what doctors you know, but the doctors I know don't tend to just go along with things. I think what they're going to ask is why? Why is the AI picking patient A instead of patient B? Now if you did it bottom up and just searched the internet, you wouldn't be able to answer that question adequately. But notice that we have these features that are morally relevant features.

And the AI will be able to say, because of this feature and this feature. So it can give the doctor a reason. Now, at that point, the doctor can say, I don't think you're putting enough weight on these other reasons and can argue with it. And then we can go to a higher authority, an ethics board of some sort in the hospital. But the doctor might say, oh, you're right. I didn't think of that.

Because I think a lot of immorality or moral mistakes that occur in our society are things that people would recognize if it were pointed out to them. And so in those cases, at least, the doctor might say, this machine helped me avoid making an error.

There are far fewer kidneys around than people who need them. How far are we away from a new system where this AI can facilitate doctors making decisions about how to allocate kidneys? So AI is already used in some kidney transplant centers.

but not for moral purposes. And one reason we picked kidneys was we knew it was already being used to allocate kidneys in the most efficient way so as to serve the greater medical reasons. And now we're saying, well, if you can do that for the medical reasons, then put in the moral reasons as well.

How far are we from getting a hospital to actually use this system? Look, I don't want them to use the system right now. We've got our little system and it works pretty well, but we need a lot more refinements on it. We're talking 10 years in the future. But 10 years from now, I wouldn't be surprised.

And how widespread is the application of ethics in AI? We've talked particularly about kidneys. How far can this go? It can go a long way.

We're already looking at dementia and patient preferences in late-stage dementia and using data from the patient to predict what they would want. We're talking to people in the business school about using it for hiring in certain ways that will introduce moral considerations, fairness with regard to gender and race, for example, into hiring decisions.

We're in contact with the military at West Point in the U.S., but also the Special Operations Forces. And they're all interested in using this kind of technology, not because they think it's going to work, but because they think it's worth trying. Walter St. Armstrong, thank you very much indeed. Thank you very much for having me.

For more Philosophy Bites, go to www.philosophybites.com. You can also find details there of Philosophy Bites books and how to support us.