We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode AI and Jobs: What Do We Really Know?

AI and Jobs: What Do We Really Know?

2025/5/16
logo of podcast The New Bazaar

The New Bazaar

AI Deep Dive AI Chapters Transcript
Topics
Cardiff Garcia: 我认为人工智能在未来5到10年内将能够完成我目前工作中的许多子任务,例如写作和编辑。因此,我个人需要不断提升自己的技能,利用这些工具来改进我的工作,而不是仅仅沿用旧方法。我看到人工智能正在迅速发展,并且在某些领域已经超越了人类的能力。因此,我需要适应这种变化,并找到新的方式来为我的工作增加价值。 Nathan Goldschlag: 我认为人工智能目前最适合作为集思广益的辅助工具。它擅长编码,可以帮助我节省大量时间。然而,它无法取代人类的判断力和研究能力。我认为未来成功的关键在于利用这些工具来提升我们的工作效率。人工智能可以自动化某些任务,但将这些任务串联起来并解决实际问题仍然需要人类的智慧。我认为原始智力不再像过去那样有价值,我们需要培养其他互补技能。人工智能在企业中的应用率仍然较低,但它正在迅速增长,尤其是在大型和年轻的企业中。虽然人工智能可以自动化某些任务,但它也创造了对新技能和熟练工人的需求。总的来说,人工智能对就业的影响尚不确定,但我们需要关注使用率、劳动力参与率和职业分布的变化。

Deep Dive

Shownotes Transcript

Translations:
中文

Hi, I'm Cardiff Garcia, and this is The New Bazaar. Coming up on today's show. There's plenty of examples of technologies that were unexpected, or the implementation of technologies went in directions that you wouldn't have predicted up front. Nathan Goldslag on artificial intelligence and its effects on your job. Hey everyone, Cardiff here. Artificial intelligence, AI.

Is it going to help you do your job better so that you can produce more of whatever it is that you produce, make a lot of money, have a happy and fulfilling life with your brilliant new AI assistant? Or rather than help you do your job, is AI just going to straight up do your job? No need anymore.

For those of us who do knowledge work, you know, laptop jobs, thinky work, because we just will not be able to compete anymore. And there goes our usefulness to society, our employability, our self-worth, our dignity. Or will this all amount to something else? Something in between those two extreme examples or maybe something that we simply cannot anticipate now.

Today's guest is Nathan Goldschlag. Nathan is an economist, and until recently, he was at the U.S. Census Bureau's Center for Economic Studies, where his title was Principal Economist, and where, among other things, he led research on the impact of technology, including AI, on the economy. And by the way, any list of the world's best economists on this topic has got to include him, or else your list is totally worthless. Sorry.

And happily, he also was an easy guest to book because he's now the research director at the Economic Innovation Group, which happens to make this podcast and also happens to be where I work. So he's now my colleague. What I've learned working alongside him is that he has just a profoundly deep understanding of the available data on the effects of AI on the economy. And you can see it in his own work, his own research.

But it also makes him uniquely qualified to filter out a lot of the nonsense out there about this topic. And believe me, there is a lot of nonsense out there about this topic. And qualified, therefore, to find the outside analysis that really is worth paying attention to. So what we do in today's episode is we not only go through Nathan's own research, his findings about AI and work, we also go through some of the work done by others, right?

researchers, essayists, sub-stackers, and we go through all of their writings and analysis and meanderings that we think is worth sharing with our listeners. And of course, we'll also link to all that work on our website at eig.org. I couldn't wait to get Nathan on the show for this chat. Here it is.

Nathan, your debut on The New Bazaar. How are you? How are you feeling? This is great. I'm super excited to be here. I think some of the work I've done at Census gives me a good foundation to think about these questions, so I'm super excited to share some of the things that I've learned along the way. It sure does. And also because you're not at Census anymore, you can really say what's on your mind. I like talking to people who just left the government. That's right. That's right. The cuffs are off. Unshackled. I love it. Do

Do you think a lot about AI and its effects on your own job? Yeah. I mean, well, in some sense, I think that the current iteration of LLMs are super useful in the work that I do. I don't consider it to be able to do the work, you know, as EIG just hired me. LLMs, like ChatGPT, the large language models that you basically ask a question and they tell you anything. Right. It can't do my job, thankfully. Yet? Well...

Yet, yet, maybe. Maybe ever? Hopefully never. And so, you know, I think those technologies right now, they're super useful as like an aid to brainstorming. An aid to brainstorming. Yeah, I think about this all the time in the context of my own job. And when I look at you and what you do, you do a lot of things. One of them is you manage people, LLMs, chat GPT, can't do that.

right? You oversee people. You sort of have to use your judgment and your accumulated wisdom from the years in which you've been an economist and manage other people. You prioritize what it is that the researchers and economists who work for you are going to do. So there's a managerial component to it, but you also are an economist yourself. You do a lot of original research of your own. And how does AI help with that? So it helps you brainstorm. Does it help you do things like the

the coding? Does it help you, I don't know, check to see if you've made mistakes, things along those lines? Yeah, absolutely. So it's really helpful in coding. You don't have to hold all of that stuff in your head anymore. You can sort of just let it go and let the machine remember for you. That happens a lot. You sort of say, oh, how do I do this task? And you can just ask it and it'll give you a bunch of pieces that you have to string together oftentimes.

For the work that I do, I think that it's all about research taste, right? And so it's hard for these systems to have the ability to know what questions to ask.

And that's what makes it really useful for a brainstorming tool because you can sort of bounce ideas off of it. And you see the things that it's giving back and can string the more important and interesting pieces together. And that's the value add that I thankfully still have. I know a lot of people who use LLMs as an aid in the coding part of their jobs, but they all also tell me that they have to check the code.

that the chat GPT and the LLMs give back to them. So it can help you save a lot of time if you say, hey, give me this code. I want to do the following things. And I'm sure that

these tools will get way better. But then they do say like, you have to do, you have to be very careful after you've made that request and it gives you back the code and it has saved you the time. You have to then spend some time making sure that it's right. Is that your experience? That's right. It just gets you a lot of the way there, right? It'll give you something that still has some bugs and it forgot to do some things. But again, it saves you a huge amount of time and get you most of the way there with the code that you're trying to write with way less effort. Yeah. Some of the, some of the new AIs are also doing things like

releasing podcasts with fake AI podcast hosts that actually talk to each other. I think Google released a new program that does this last year. I think about AI in my own job all the time. I sort of have a feeling that a lot of the individual subcomponents of my current work

will be doable by an AI within, I don't know, five or 10 years, right? Like the talking part, clearly what you and I are doing right now on this podcast, like it's already started. And in terms of helping people write better, because I do a lot of editing, I do a lot of writing. I don't know. It seems like a lot of that actually is automatable. And

at the rate at which these tools are getting better, because they are improving all the time, it seems like a lot of the parts of my current job will be doable by these tools later on. And so I will personally have to evolve what it is that I bring to the job as well, and maybe use these tools to make me better and not just sort of keep doing things the old way. Do you think that's, as an economist who studies work and labor, do you think that's

That's good advice? No, I think that's right. And I think that in the future, it's going to, you know, to be successful, you're going to be have to, you know, be leveraging these tools in order to make them make you a better at what you do. Any one piece of the things that you do. Yeah, that could be automatable. But the whole thing, you got to remember, there's an agency problem of stringing a bunch of different things together. And so each individual task automatically

taking on its own, maybe the AI can do that piece. It could also help you do that piece. So it's a compliment as well as a substitute. But stringing all those things together is really, I think, the value add that you're going to retain. But at the same time, the tasks that you do are going to change what tasks you focus your time on and the allocation of your time across tasks is going to change. And it might create new tasks as well. Right. So you might be doing things that you didn't do before because these tools are opening up new doors. Do you think...

that raw intelligence will no longer be as valuable as it once was. Leave aside the issue of taste, which you mentioned. I like that point. The issue of judgment, wisdom, what to do with the output of raw intelligence, just like sheer cognitive power, the ability to remember stuff, the ability to make certain connections, the ability to

Just seems like with the elevation of these tools, like that's just not going to be as impressive as it has been for the last few decades. No, that's right. But it's also the case that in the in the data, the returns to IQ for wages aren't nearly as high as you might have thought.

So it's not as if the labor market is actually-

right? There's going to be diminishing returns to that. So there's going to be complementary skills and assets that humans are bringing to the table that make that equation work. Yeah, better to pivot to those as soon as possible. Right, right. Let's talk about your research because you've published a number of research papers about AI and technology generally and their effects on the labor market. Let me start with this. What do we know about

how many or the percentage of businesses that are actually using AI as of right now are the most recent available data? So this is actually something that I've been spending a lot of time on. And it's a lot of the work that I had done at Census was focused on these questions.

There's a really great team of economists at census that are thinking ahead of the curve on these things. Oftentimes for the federal statistical system, the technology is fully diffused by the time they ask the questions. Here in this case, we actually started asking them early. We actually have a picture of what this looks like over time. I think the bottom line is it's lower than you might have expected. In most recent data, something like 9%

firms say that they're using AI. Pre-GPT, that was maybe 3%. So it's increasing over time. But if you ask businesses, do you use AI in the production of goods and services? Only about 9% of them say yes. Yeah, that's really interesting because that does sound low given the prominence of the debate about AI.

In the culture, in the media, it's obviously something that you and I are very interested in. That's like lower than people think. Do you think that it's going to increase quickly though? How does this compare to before the introduction of

Those LLMs, these large language models, ChatGPT and so forth, because that was just a couple of years ago. So I wonder if now the uptake is going to be faster than it was before. Yeah, the great thing about some of the work that Census is doing is that it's on these really high frequency surveys. So the data that I'm referencing is actually from May. And so these are really new numbers. May of this year? This year. Okay. Right. So just recently. Yeah, exactly. Yeah.

These data show a trend. They're increasing. They've gone up from something like 4% a year and a half ago to 9% today. It's also the case that the firms that adopt these technologies tend to be much larger. Early on, a couple of years ago, that 3% use rate turns into something like 12% on an employment-weighted basis, meaning the number of employees that are at

businesses using these technologies. That's interesting. Right? They tend to be much larger, so the quote-unquote exposure to them is higher than the firm level use rate. That's really interesting. I also imagine that different kinds of businesses show a lot of variation in how much they use it, right? That's right. There's a strong relationship between size and the use of these technologies. It tends to be larger firms that use them.

But it also tends to be younger firms. So within different size groups, there's a negative relationship between use and age. So it tends to be larger, younger firms. And one of the things that we've seen more recently in the post-GPT era

period, is sort of a J shape in the size distribution, meaning that the very smallest firms have a little bit higher use rate than sort of the mid-sized, and then it goes back up and the largest firms have a very high rate. We think that might be suggestive of the fact that these new versions of AI are really amenable to the production technologies of small firms.

If you have a firm of only a couple people, you might be able to use these technology. You didn't actually hire a marketing person before. Now you can just use ChatGPT to do something that looks a lot like a marketing person. So there might be higher returns among the smallest businesses. That's really interesting. I would also imagine that if you're a small business and you're making the decision of whether to start hiring more people, increase the scale,

you'd be careful with that decision. Because if you're a small business, hiring another person is a really big challenge.

So you might just be more experimental with these technologies first, see if you can save the money of not going from five people to six people, which is, you know, that's a 20% increase in the size of your business versus a really large company, which might just be like, hey, we have some more work to do, like just hire somebody else to do it. You'd be maybe more willing to test out this technology if it might save you money if you're small. That's right. That sort of feeds into the, you know, what are the employment effects of these technologies? Yeah.

One of the nice things about these surveys is we can ask firms, you say, did you increase or decrease your employment based on your use of AI? Like it's a self-reported causal estimate. And when we do that, most firms, the overwhelming majority of firms say that the ones that use AI say that AI did not have an impact on their employment. And in terms of increase and decrease, it's about 50-50. Like among the small percentage that say that AI did change their employment, half say up and half say down.

I would also imagine that some companies would want to hire people to help them use AI. So it's not just that AI might...

automate away the work of some employees. There might also be some employees you'd bring on who happen to be experts in the technology and you'd bring them on as well, right? That's right. We have estimates of the skill composition. So we ask firms whether or not they're hiring more skilled workers because they use AI and the answer tends to be yes more than no. And so it's also true that when firms are implementing these technologies, it increases their demand for skill. They need

new types of skills and they need new skilled workers to implement those technologies and reap the benefits. Yeah, that's interesting. And I can understand, as we just said, why the smallest businesses would be willing to experiment with AI and

Bigger businesses may not be as thoughtful about trying out at the very beginning the thing, but the very biggest companies you said also tend to use it a lot more. And I wonder if in their case, it's just because they have the money to invest, to upgrade their systems or whatever, to accommodate AI. Because some companies do need to change the way they do things to use a new technology. Isn't that right? No, that's right. And there's certainly going to be returns to scale. One of the things that you see this in is with data. So these large businesses, they tend to have lots of data.

And that data, you can use AI to increase your returns of the adoption based upon that data. You can sort of customize these tools and implement them in specialized ways that leverage the information within the company. Yeah, I like that. And then if companies aren't using AI to automate away worker tasks, or maybe some are, some aren't, but it doesn't look like it's super prevalent just yet,

What do they say they are using it for? So the most common response is to improve products and improve processes. The third most common response is to automate the work done by employees. But among firms that say that they're using AI, about 27% say

say that they're using it to automate tasks, but only 5% say that they're changing their employment. So it's more often the case that firms are using AI to automate subsets of the things that different types of workers do, not necessarily do all of the things that those workers did. Big difference between, let's say, a company that mainly does, I don't know, internet stuff, information technology also, versus companies that

maybe do more stuff, I don't know, in the physical world, in the tangible world, I would imagine that the adoption rate of AI there also varies quite a bit. Oh, it varies a lot by sector. So it can get as high as 25% in the information sector, but something like 2% in the accommodation and food services sector. So one of the most common responses to why a firm isn't using AI is that it's not applicable. A lot of businesses still don't see how it can improve their production.

Did anything about all this research surprise you as you were doing it? You put out the surveys or you looked at the data, you and your colleagues, and you saw it and you were just like, huh.

That's interesting. Didn't know that was coming. So I think that overall, the low use rates were a surprise. There was sort of like within a week, ChatGPT had 100 million users or something. There's amazing statistics about diffusion. Like where do they work? Where are they? And so based on the headlines, you'd think that it's sort of everywhere.

And based upon the way that economists are talking about it and sort of modeling out the relationship between AI and labor markets, you'd also expect to see lots of displacement, lots of movement in employment among adopting firms. And that just doesn't seem to be the case yet. And we're well into the post-GPT period, right? These technologies have been around for a little bit now, but even in the latest data from earlier this year, it doesn't look like

these companies are changing their employment a lot, even though they are expected to, you know, something like 35% of them in the next six months are going to be using it to automate tasks. So I think the automation of specific types of tasks that workers are doing is going to go up. But I think, you know, taken on this face, these low percentages of use and the not rapid displacement of workers is kind of surprising. This might be too granular for the data that's available, but

I was talking to a friend of mine who's an executive at an investment firm recently. And one of the things he was saying was that they haven't figured out how to use AI yet for the main thing that his firm does.

What they use AI for are things like customer service or investor relations, these more kind of peripheral tasks where, yes, they can automate away some of the work that the workers do, or they can use AI to enhance the workers in those parts of the company. But it was only now that they were starting to figure out ways to use AI in the main thing they do, which is essentially financial analysis and investing. Right.

And I'm wondering also the extent to which a lot of the companies that respond to the survey are also saying, yeah, we use AI, but right now, at least, it's for things that are vital. They're important to what we do, but they're not the main thing that we produce. You know what I mean? No, I think that's actually points to – it's an interesting point. I think it actually points to one of the challenges in measuring the use of AI among businesses is

And it's actually about whether or not, if you think about our low percentage of uptake of AI, that might reflect the fact that asking them whether or not they use it in the production of goods and services is kind of a high bar. There might be lots of incidental use within the firm. Some of it might be embedded in products and services that it purchases, cloud services, or your email has lots of AI baked into it. But you wouldn't consider that part of your production of goods and services necessarily.

You might be using it to summarize different pieces of information, memos coming in or going out, summarizing email chains or whatever. Those incidental uses might not be picked up in these percentages. So it's true, the diffusion might actually be much, much higher, but

the specific type of diffusion we're looking at of the production of goods and services, which, by the way, is where we think the most TFP changes are going to be coming from. TFP, that's total factor productivity. That's right. The most essentially technological change that helps a company become more efficient, innovative, better, more productive.

Yeah, when we think about the implementations of AI that are going to reap the most economic benefits and make us better at producing things, it's going to be in the production of goods and services. It's not necessarily going to be that the autocomplete in your email can be exactly right. Much though I appreciate it, right? Perfect emojis at the end of every sentence, you know, for everybody. That's not necessarily going to move the needle on productivity.

You found something really interesting that you mentioned to me in a recent conversation about the difference between software developers and programmers. So give us sort of the context here in which you were looking at this, and then tell us what you found and what's significant about it. So this actually feeds back into the discussion earlier about tasks versus an agency and thinking about the role that we have in relation to these technologies.

So one of the often cited statistics when we think about the impact of AI on labor markets is the displacement of software programmers. And one of the really interesting things that we can see in the data from the Bureau of Labor Statistics, where we track the number of software programmers, is that it's actually split between changes in the software programmers and then software developers.

So software developers, who are much more numerous, they're not declining nearly as much as software programmers. Tell us first what the difference is between software developers and software programmers. So the software programmer, their primary tasks are related to creating, executing, and testing new scripts and programs. Whereas the software developer is much more deeply engaged in the why the program is being written and the requirements surrounding it and stringing together multiple pieces of the business technologies.

to solve these questions. Is it fair to say that what software programmers do is a little bit more routine oriented than software? That's exactly right. Okay. That's exactly right. And so when the finding, when you had this finding that software programmers in percentage terms are suffering in the labor market more than software developers, right?

There's a possible reason for it, but you also found that there's a lot fewer software programmers in the first place than software developers. Is that right? That's right. Software programmers between 2013 and 2014 were down something like, in counts, they were down something like 8%, 9%. Okay. It's a lot, by the way. It is. That's nothing to sneeze at. And the software developers was down less than a quarter of a

of a point. Oh, okay. So basically flat, but at the same time, software programmers is only something like 120,000 of them in 2023, but there's about 1.6 million software developers. So when we're thinking about like these, you know, stories about the displacement of software programmers and developers, you really got to split the different types of occupations by what types of tasks they're doing. And it's absolutely true that the more routine tasks are,

like writing and testing scripts, are going to be much more affected by the LLMs we have right now than something that engages more with business processes and thinking about why programs are being written. Something that's been really interesting about this whole conversation to me, and which I've brought up on the show before, is that it was only around the mid-2010s that this famous phrase started going around, learn to code. Everybody was supposed to learn to code because that's the job of the future.

And now there's a lot of people saying, well, wait a minute, coding is one of the tasks that's most vulnerable to now being done by artificial intelligence. And it's almost like you get whiplash from this because the responsible thing 10 years ago was to learn to code. And less than a decade or about a decade later, don't.

Don't learn to code. Learn to do something else is now what everybody's saying because of the rise of these AIs. And I think what you're finding about developers versus programmers does is it adds a little bit of nuance to that story, which is that there's still coding-related tasks, coding-related jobs that you can do, even if...

the LLMs themselves are now actually writing a lot of the code. Yeah, so that would have been terrible advice if you were going to take the very narrow version of that and just learn to write and execute scripts. Right. But if at the same time you learned how to think about business analysis and the writing of requirements for these different programs, thinking about how it integrates and solves problems for the business,

Like those types of skills, which again, more broadly, when you think about how we're going to integrate with these technologies, you should be thinking about broadening the tasks and changing composition of the tasks that we're doing. And so this is sort of a nice microcosm of that. So if you're super narrowly focused and you're just writing programs, that's going to be really susceptible. But if you're engaged in these complementary tasks, tasks that are

very highly agentic, meaning you've got to integrate lots of different pieces and lots of logical pieces of a chain, those are going to be less susceptible.

There have been a lot of articles written recently about how younger people are not getting hired at maybe the same rate that you would have expected them to, specifically college educated younger people in their early or mid twenties. And that the unemployment rate for these folks is now a little bit higher than the unemployment rate for the whole economy. Although it's quite low in both cases. And again, I'm talking only about recent college grads.

Do you make anything of that? Like, is it possible that that's related to the fact that like a lot of businesses are trying out these new technologies, at least to replace some of the more entry level work at, I don't know, banks or law firms where they would have previously hired these high skilled young folks to come in and do that work?

It's a super interesting trend. I don't know that I have a very clear answer for you, but I will say that it started to turn actually after the great financial crisis. When you think about the decline in that ratio of the all employment rate to the recent grad employment rate, it turns really, really early in terms of the AI timelines. The types of AI that we had in the 2010s were not nearly as capable as the ones that we have now.

right? And so it's not as if the decline in that rate was reflecting those specific types of technologies. Now, you know, I think there's probably really complex structural reasons for this. There's some fun things you can think about, like whether or not the signaling value of education is declining. Elite overproduction hypothesis, we just have too many of these people. You know, you can think about skill mismatch between the types of

skills that recent grads are coming out with and the skills that firms demand. There's a ton of different explanations, but I don't know that AI really fits the bill because it doesn't fit the time series trends. I was listening to a podcast at Tyler Cowen, an economist that you and I both follow, and he was talking to another AI expert and

And one of the questions that Tyler had was, what's the worst age to be right now as AI keeps getting better and better in terms of just being able to incorporate it into your life and into your work?

Tyler himself suggested that the worst age was 40 because you're along in your career enough that you might think that retraining is like too much to take on, but you have enough of your career left that actually would be very useful for you to be able to adopt.

this new technology and use it in your day-to-day life. But I'm kind of curious, Nathan, to know what you think about that question. I would say there's actually, maybe there's sort of a dumbbell pattern to this. 40 might be really bad if, you know, that age group is less amenable to changes in their work patterns in reflection of these technologies.

But everybody's going to have a tough time because it's just moving so fast. So that's one thing to keep in mind. But another age that might not be great is my kids, right? So they're going to be entering a school system that's going to be thinking about adopting these technologies. And it

It's not necessarily well thought out how that's going to work in the educational system. So being really young right now and being exposed to this, what's probably going to end up being a giant social experiment of including AI technologies in the public education system, and then also being in the workforce later in your career and less able to change. You'll be in an education system that's trying to keep up in real time with how it works and might not be able to. Right. Yeah. Fascinating. So it's like...

People in their early 40s and five-year-olds. Everybody else is going to be okay. I want to do a kind of long-term speculation here with you because we've been talking about what we know to this point. But

You are open to this story changing, right? Like, are there things we should be following over time to see if, for example, the capabilities of these AIs go from complementary technologies to people with good taste and with good judgment, and it takes away some of the sort of raw intelligence or the routine tasks to something that exists more in the creative realm, right?

We have to be open to that story, right? We should be open to many varied futures. And I'm just kind of curious to know, what should we be following to see if there is a big change in the trajectory of the adoption of AI and its effects on the labor market? So I think there's a couple of things I would be looking for. I'd be looking for changes in those use rates at the firm level. Those are going to be important. They might not tell you the whole story, as we discussed. They might only be capturing a certain type of implementation of these technologies within firms.

Person-level use rates, I think those are these sort of studies that look at ask people whether or not they're using the technology. That's going to be important. Labor force participation rates and shifting occupational employment distributions, right? So the Bureau of Labor Statistics puts out these really nice matrices that tell you how many people are working in different occupations over time. And so you can sort of string those together and sort of see how the occupation mix is responding to the implementation of these technologies.

And I think the one takeaway I would suggest you take from this is that it's very highly uncertain, right? It's very unclear how this is going to impact the trajectory of these labor markets going forward. And part of that's going to reflect the fact that

the implementation of these technologies may create new tasks. It may create new jobs, right? And this is sort of the history of technological change over centuries is once these big general purpose technologies hit, they totally reorder society and labor markets, right? And so they create new jobs that you wouldn't necessarily have thought would have existed before. Yeah, a couple of years ago, I was talking to an economist named Avi Goldfarb on this podcast on The New Bazaar. And

I asked him about this and I said, hey, man, let's say that my specific tasks are in fact automated way. So things like talking on a podcast knowledgeably about the economy, being able to, I don't know, assess a lot of information at once, some management skills, things like that. But

But I am left with one still useful quality or a few maybe, which is just people skills, you know, just charisma, the ability to make people laugh, whatever. Will that save me?

And he said, well, maybe, but the thing to remember is that a lot of other people who do what you do now will also suddenly be competing in that same space. So you're gonna have a lot of other people who are also automated out of their jobs. And so you think you've got this angle because you've got this one thing that hasn't been automated away, but you also have a lot more competition from other people who are now going to also be trying to do the same thing that you're doing because precisely for the same reason, right?

And I thought, well, that's disheartening, isn't it? Yeah, that's right. And I think it's an interesting point that the returns to different types of skills are going to change in response to these technologies. But again, I would point to the fact that new jobs and new tasks have been created throughout the history of technological change. And so in some sense, that's going to be the thing that saves the day, is that as the supply of social labor increases,

the set of ways that firms can engage with that is going to increase as well. It might also just be super random stuff that we haven't imagined yet, precisely because we can't anticipate it. You know, I sometimes think about chess, uh,

Because, I don't know, as of three decades ago, the machines still couldn't beat humans. And now there's no human that can beat even like a mid-tier chess machine. And yet people still are interested in human versus human chess. Precisely because they don't want to see a machine do that. They want to see two human beings competing in a certain way. Similar to with, I don't know, sports. You could come up with a machine that never missed a match.

you know, from 30 feet away about, you know, if you're shooting baskets or whatever, but we watch like great athletes compete against each other because we want to see the human element itself. But there's something about that. That's also more about like,

our entertainment, our amusement, our presenting ourselves to each other. It's a little bit different from what the labor market looks like right now and has looked like where we're sort of producing things that we buy and sell. Instead, we'd be buying and selling our ability to

capture each other's attention. And when I look around, I sort of see what looks like maybe the beginnings of that. I'm being very speculative here to be clear, but you see this in like influencer stuff. You see it in the survival of chess itself and in other places. And I don't, I don't have anything in particular to ask about this. I just find it interesting. I'm curious to know what you think. Yeah. I think, you know, that's everything's sort of downstream of culture, right? So the implementation of these technologies is going to be within a social and a cultural context. Um,

It makes perfect sense to me that people are going to put greater value on human-generated content, right? And so it might be the case that the AIs can make music or whatever it is, but I might still pay more in order to know that there's a human on the other side of that engagement. Yeah. I now want to do something with you where we go around the horn of

some other research that other scholars and other writers have done because you gave me this list. It's really interesting of stuff that you found worth engaging with. Doesn't mean you agree with everything on this list, but you think this is all very much worth engaging with. So we're just going to go through them one by one. And then we are going to, yeah, we're going to share it with our listeners in the show notes. So

First up, some economists who go by the names Bick, Blandin, and Deming, they found something that on the surface at least appears to contradict your earlier findings about the low usage rates at businesses, but you have a way to reconcile them. And I think a lot of people are going to have this question too, which is that your numbers sound low to them.

So reconcile what Bick, Blandon, and Deming found versus what you found, starting with telling us what they actually found in the first place. Yeah. So they ran sort of a current population style person level survey, and they find that about 40% of the working age population reported using AI and about 23% of the employed individuals used it in the past week or two at work.

And so those are obviously much higher than the firm level use rates. And I think that this points to an interesting problem for the measurement agenda and thinking about how do we measure the diffusion of these technologies. And it goes back to this sort of incidental, it could be the case that these are sort of incidental uses. And so, you know, workers would report, yeah, I used AI because I was, you know, bouncing around.

my text off of it for typos or whatever else. But when the firm is thinking about is the technology being used in the production of goods and services, they might say no. And so I think to reconcile these two things, we need to think about what it is that these individuals are using the technology for and how does that interact with the firm's production technology.

Here's another paper that you said was very cool and very interesting. And it was titled Answering the Call of Automation from James Feigenbaum and Daniel Gross. It introduces the concept of so-so automation. So tell us first what that is. What is so-so automation and why it matters? So I love this paper there. So the so-so automation, they're actually referencing some work by Doronis Moglu and Pasquale Restrepo.

who talk about so-so automation technologies as one that reduce the demand for labor because they're displacing workers, but they don't create a big enough productivity shock

to create compensating labor demand in other places of the economy. You only lose workers in the economy and you don't gain them in some other job. Right. So yeah, if the technology displaces some workers but allows you to scale up significantly because of the productivity gains, then you might actually on net end up with more workers, right? Labor demand on net might increase. Mm-hmm.

But a so-so automation technology is one where you're sort of displacing workers, but there's not really a productivity effect. You're sort of taking those tasks and doing them not any better, not any worse, maybe a little worse. Okay. But it doesn't actually affect your scale, your ability to scale, or your labor demand.

Interesting. And so this paper answering the call of automation from James Feigenbaum and Daniel Gross, it looked at one of those so-so technologies. What was it and what was their finding? So they were looking at an automation shock of the introduction of automated telephone switches. And so in the...

1930s, one of the most common occupations for women in the US was a telephone switch operator. You see this in the movies sometimes. That's right. You call and somebody at a switchboard somewhere picks up the phone and says, yeah, I'll connect you to James Street 554 or whatever. That's right. And so AT&T in the 1930s introduced new technologies that automated that process.

And, you know, in short, what they found was that the incumbent workers, the workers, the older women that were working in those occupations, they did suffer losses in wages, decreased wages, less likely to be working in the future.

But the younger cohorts of women that came after them, they weren't permanently scarred by these technologies, which suggests that local labor markets were able to re-equilibrate and adjust to the technologies, and firms were able to find new jobs and new ways to engage with those young workers without permanently reducing wages.

It's a really fascinating paper, and I think it introduces this question of whether or not the differential effect on incumbent workers and the process of how technologies are ingested by markets and how they adjust to it, and how those adjustments can be quite costly for some people and some subsets of workers.

But it doesn't necessarily have to be costly for everyone. Or costly in net terms for the economy. That's right. So some people lose out in this case, those incumbent older workers who lost out on those jobs of being a switchboard operator.

And overall, the economy still adapted, even though you didn't get that huge productivity burst from this technology. And so it leaves you with the question of, okay, so when there is a period of adjustment like that, do you do anything? We'll talk about policy later, but I wanted to make sure that that was the setup for this paper, which is very interesting. That's great. Okay. Next up is actually a website, AI-Polite.

It's by a series of people, Daniel Cocotahilo, if I'm pronouncing that right, Scott Alexander, and three other folks. They basically go through some of the more like boomy, not doomy, but boomy predictions of AI and how it's going to accelerate economic growth. Is that right? Yeah, there's this really interesting dichotomy in like the Silicon Valley AI types. You know, there's the boomers and the doomers.

uh you know the doomers are talking about ai safety and whether or not we're going to be turned into paper clips and then there's sort of the you know the boomers that think it's you know kind of we're going to get agi or artificial general intelligence or even artificial super intelligence asi within the next couple years and so this website what does that mean because it sounds like we're just giving adjectives here that

Yeah, no, that's what it is.

is kind of how they're thinking about it. So essentially, almost quasi-infinite economic growth is what you're saying. Well, yeah. So there's a lot of folks in this world that think, oh, in this sort of milieu, that think, oh, well, we might get 10%, 30% economic growth per year. And the group that wrote this, put down this website, to their credit, they sort of very nicely map out what that would look like at each stage, month by month,

between here and there. And I think it's, you know, it's, you know, for us in the economics community, thinking about innovation and technology change from like the economic modeling perspective, it seems kind of wacky and out there. But I think it's a really nice thing to keep on your radar as like, you know, smart people are thinking about it in this way. And it's something that I think it's, you know, worth engaging with.

Yeah. I was going to say also for our listeners, for context here, economic growth for the last quarter century has averaged about 2% a year, real GDP growth in the US. So going to 10% or 30% is just a massive, massive increase. Economic growth on an order that we have never, not just not seen, but not imagined previously. So it'd be amazing. I sort of think it is worth, like you said, keeping possibilities like this in your back pocket somewhere.

Yeah, that's right. Even if it's super low probability, like, hey, you know, it's a low probability things happen all the time. So in these uncertain times, it's it's worth keeping a wide lens. Right. And then up next is a conversation between the podcaster Dorkesh Patel and our friend again, Tyler Cowen.

in which Tyler essentially is the counterpoint to those guys who were saying that the economy is going to boom unlike we've ever seen. Tyler's saying that actually there are big obstacles between our world as it exists now and that world. Yeah, I think this was a really fun conversation that they had. And Dworkesh, to his credit,

He sort of embodies the uncertainty here, and he even admits on his shows, he sort of flip-flops like from week to week, you know, whether or not he's a boomer or a doomer. And so he was trying to bring that argument forward and say, okay, what if, you know, why won't we get, you know, these extra normal growth rates, 30% growth rates? And part of what he was getting at is that these economic models that we write down in macroeconomics, they often turn on population growth.

So, you know, the change in the GDP growth is tied to population growth in some of these models. And so with AI, if you can sort of copy them millions of times, you get an effective population growth that's quite large. And so shouldn't that flow through to economic growth? And so, you know, Tyler, I think, rightly points to the fact that there's a lot of frictions in the system. There's a lot of bottlenecks. You know, there's a lot of things, both social, cultural regulation issues.

what have you, that are going to slow down the diffusion of those technologies. And that's consistent with the history of technology change too, right? So when you got electrification of manufacturing plants, the whole plant had to be reordered. It's not going to be an overnight process to implement that. Stick with that example though, because that's a great example. And it's one of my favorites, which is

that in the late 1800s, with the advent of electricity, everybody's like, oh, this is going to change everything right away. And at first, it did change some things meaningfully. So if you used electricity to change one part of what a factory does, you could see the improvement. It was a nice change on the margin. But what was actually needed was a totally fundamental rethink of what a factory does in the first place and how it is designed. And that actually took decades.

So you had this incredible technology and the potential to radically accelerate economic growth. But first, you had to get the human element right, which was to get people to try the technology in this radically transformational way instead of just using it to improve the things that we already knew how to do.

That's right. There's, I mean, there's a lot of experimentation that needs to happen. You know, people need to think about how these technologies can be incorporated into their production processes. That's not an easy task. And with AI, it's, you know, with how fast it's moving, that makes it all the more difficult to sort of think about how to integrate these effectively. And that would put us more into that world of it's going to take some time to diffuse. But there's also lots of regulatory and other barriers that you'll run into.

Yeah, and by the way, that brings us to our next example of something that people should read, which is an article from Timothy Lee who writes Understanding AI, which is one of my favorite Substack newsletters. Everybody should definitely subscribe to it. He wrote a piece titled Driverless Trucks Are Coming and Unions Aren't

happy about it. And I know you had a really positive reaction to this one as well. You tell us the story because you love this story. I love this story. It sort of embodies all of the most interesting bits of the bottleneck story. So in the case of driverless trucks, the companies that were creating these new technologies were running up against regulations about the reflective triangles that could put out when a truck is disabled on the side of the road.

Which, by the way, are required by law. In other words, if you drive a big truck, and trucking is a big occupation in the US, if you drive a big truck and you pull over to the side of the road, you have to...

walk out of the truck and you have to put down one of these reflexive triangles on the street so that people are warned that there's a big truck that has stopped and pulled over to the side. Okay, that's the setup. Keep going. That's right. That's right. And so the companies that were producing these driverless trucks asked for a waiver on these regulations and they offered an alternative, which was a bunch of lights on the side of the truck or whatever. And they tried to prove their case that these lights were as good or better.

One of the funny little bits of that is like, well, the triangles might get blown over by the wind and the lights won't. But the bottom line is that they did not get these waivers. And so now what are they doing? They're thinking like, well, how do we automate the process of getting these little triangles out? Maybe we have a little robot that comes out from under the truck and put it out there. Or we could pay somebody to drive behind the truck, I guess, and then they'll be the triangle putter.

This is one of those things where it's actually better understood as a proxy fight, right? The people that are against the implementation of these technologies, they'll find these little places where you can create

create a bottleneck and accentuate a bottleneck. They use existing regulations to help them. That's right. They use the law to help them find or introduce those bottlenecks so that it slows down the automation. And you see this in a lot of examples. There was the big fight over automation at the ports and the extent to which the unions are just –

really aggressively against automation there because they're worried about workers losing their jobs. But the consequence is that there's no productivity growth at those ports, you know, and so there's bottlenecks everywhere. And even though the technology exists to make it all work better.

Yeah, and again, going back to the Feigenbaum and Gross paper, they're right to think that these automation technologies could hurt the incumbent workers, right? So that is going to be the case, but that doesn't necessarily mean that we should be freezing the economy in amber. That brings us to our next example of something that you think our listeners should check out, which is a paper called Generative AI at Work. It's by Eric Brynjolfsson and a couple of other scholars as well.

And tell us about this one. So this is a really neat one, and I think we need more of it. It's a paper that's looking at a relatively narrow example of AI use and implementation, but you can start to learn some lessons that may generalize. And so they looked at the staggered rollout of generative AI among customer support agents, about 5,000 customer support agents.

What they found was that it increased worker productivity. But really interestingly, the least experienced workers got the biggest boost, right? So it actually decreased inequality among workers. So the return to those workers that had the least tenure of having this generative AI help them in prompting them to do their work, they saw the biggest boost.

I want to close with a discussion of what, if anything, we should do about AI. All right. One of the topics that you like to delve into from time to time is the issue of should we try to steer this technology or not steer the technology? So give us a flavor of what you mean by that and where you come down. Yeah. So I think this is one of those cases where

It's easy to understand why people are worried about the displacement effects of these technologies. They could be quite large. But I think there's a couple reasons why it's not necessarily the best idea to go in and try to fine tune the system of innovation on the front end.

One is that these technologies are really nascent. There's a lot of uncertainty about how they'll change over time. And the history of technological change is, you know, there's plenty of examples of technologies that were unexpected or the implementation of technologies went in directions that you wouldn't have predicted upfront. And so when we start trying to, you know, put our thumb on the scale of the upstream innovation process,

We might end up cutting something off. We might say, oh, look, this thing looks like a so-so automation technology, and so we either try to regulate it or stop it in some other way. But that thing might actually, if you let it bloom through the system, might actually generate some labor demand boosting technology later on.

There's also parts of the economy that I think we would love to have even so-so automation technologies. So there's this wonderful idea in economics that some of my GMU folks have written about is Balmall's cost disease.

And so it's, you know, the idea of something like when there's a sector that has seen really slow productivity growth relative to other sectors, the labor costs in the low productivity growth sectors will go up, even if that sector has low productivity growth because their outside options are better.

Right. So they could have the workers. Yeah. The workers in that low productivity industry, they could go work in the high productivity industry, which can afford to pay the high wages, right? So the only way that a low productivity sector can afford to pay those workers to join is if they raise their wages.

the price of their product. So you get inflation in those sectors. And so, you know, that's right. And so in those cases, you might actually say, well, no, I mean, I'll take the so-so stuff. I'll take those technologies that displace labor without boosting productivity because it puts much needed downward pressure on prices. Prices. Yeah. Very interesting. The examples that you mentioned where

where we didn't expect the technology to go a certain way, but it did and it was amazing. Tell us one or two of those. So there's this wonderful one with the laser. So the gentleman who invented the laser famously said, "It's a problem in search of a solution." Now it's embedded in lots of communication technologies and many other things.

The inventors of the early versions of touchscreens, right? So they were, I think, originally invented for air traffic control systems. And what did we get? We got Angry Birds, right? So there's lots of ways that innovation unfolds that are really surprising. To be clear, you're being complimentary about Angry Birds, right? Well, more or less. More or less, kind of.

Not about Angry Birds specifically, but about the other uses of that technology.

It's combinatorial, like the things build upon one another. One innovation that comes in builds on others and you combine the new things that you have in new and complex ways. It's complex, it's combinatorial, and it's emergent. And so for all those reasons, it's really hard to predict which direction it's going to go. This happens, I think, in medicine a lot too, where you start taking medicine for one thing and you don't realize it has these other effects afterwards.

In science, I think this happens quite often. Wasn't penicillin sort of an accidental discovery too? That's right. Petri dish, right? Petri dish. Sitting on the windowsill or something like that? Yeah, something like that. Yeah, it's fascinating. And I think if I'm understanding your point correctly, you're saying that if we over-regulate AI,

at its onset, when it's really starting to become popular, that there's a risk that we end up missing out on the fascinating new directions that it could go in and all the gains from that. That's right. And I mean, it's also the case that more to the point is that we just don't know. We don't know

how the general equilibrium effects are going to play out. There's lots of predictions about the labor market effects and the displacement of workers. We don't know how firms are going to react. We don't know how labor markets are going to adjust. We don't know how the innovation ecosystem itself is going to change and adapt.

to the changes from these new technologies. So it's just very, very hard to know what direction it's going to go. So if you start putting pressure on the upstream innovation system, you might be cutting off things that would really benefit humanity. Yeah.

Yeah. And a lot of times I'll read a lot of things about how societally or policymakers should respond to the rise of AI. And precisely because of that uncertainty you just mentioned, it seems like we always end up with...

a policy that you can have a debate about whether it's good or bad completely regardless of whether or not it's a response to AI. So you mentioned that example earlier of the workers who lost their jobs as switchboard operators where the economy on net still benefited from the introduction of this new technology, but there was a subset of workers who did suffer. And so maybe you put in place the right safety net to help those workers, but

That's not really a conversation about technology anymore or about AI. You could make that argument about any worker who loses their job for any reason or is suffering for whatever reason. Like you could have these debates about how to organize society and how to put in place a good political system that's responsive to people being out of work or suffering particular labor market effects completely separate from

Yeah, given all this uncertainty, I think you want to focus on policies that are going to be a good idea either way. Right?

Right. So, you know, there's maybe some physical infrastructure change, things that you could be doing. You could be hardening infrastructure. You could be increasing energy transmission, all these other, you know, physical infrastructure things. Maybe those things complement the use of AI, but they might be good, good ideas in and of themselves.

And you could say the same thing of the labor market impacts. So you might come up with policies that you might want to have, whether or not AI turns out to be labor displacing or not. So you want to focus on those things that are going to be a good idea either way. Why? Because we have no idea how it's going to shake out. Nathan Goldslug, did you have fun on your first post-government podcast? Yeah, this is a little different than what I used to do in the data mines of census, but this was a lot of fun. Good. Well, it was great to have you.

We should go back to work now. Yeah. And that's our show for today. You can find links to not only Nathan's excellent work, but also the work of many other scholars that we reference in today's episode. Those other studies, reports, essays in the show notes for this episode, which you can find at EIG.org.

The New Bazaar is a production of the Economic Innovation Group. Adrian Lilly is our sound engineer and our music is by Scott Lane and DJ Harrison of Subfloor Studio. Special shout out to Ruben Francis and Steven Leverton for production help. Please follow or subscribe to The New Bazaar on your app of choice. If you enjoyed today's show, leave us a review or tell a friend.

If you want to get in touch, I'm on Twitter is at Cardiff Garcia, or you can email us at the new bizarre at EIG dot O-R-G. All one word, the new bizarre at EIG dot O-R-G. And we'll see you next episode.