We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Bonus: The DeepSeek Reckoning in Silicon Valley

Bonus: The DeepSeek Reckoning in Silicon Valley

2025/1/27
logo of podcast Big Technology Podcast

Big Technology Podcast

AI Deep Dive AI Chapters Transcript
People
A
Alex
通过在《Mac Geek Gab》播客中分享有用的技术提示,特别是关于Apple产品的版本控制。
M
M.G. Siegler
Topics
Alex: DeepSeek R1模型的成本远低于OpenAI的模型,仅为其3%-5%,且性能相当,这在数学测试等基准测试中得到了体现。其开源特性使得任何人都可以下载和运行该模型,这将对AI产业产生深远的影响,尤其对依赖高昂计算成本的企业造成冲击。 DeepSeek R1的出现,使得市场对AI模型的成本和定价策略进行重新评估,一些初创公司已经开始尝试使用DeepSeek R1模型来替代其他模型,以降低成本。这可能会迫使OpenAI等公司调整其定价策略,并可能导致市场价格战。 DeepSeek R1的低成本可能与训练方法和使用的GPU有关,但其性能的提升是真实存在的。这使得长期以来在硅谷流行的“规模假设”(即增加计算能力、数据和训练时间就能提高模型性能)受到质疑。 DeepSeek R1的出现,对AI产业的投资和估值产生了重大影响,并可能导致投资模式的转变。如果AI模型的成本大幅降低,那么之前基于补贴的AI行业投资模式是否还能持续?这将对英伟达等公司造成冲击,因为他们的收入很大一部分依赖于AI公司的巨额投资。 即使AI模型的成本降为零,也需要有实际的应用来创造经济价值。目前AI技术仍缺乏能够创造显著经济价值的实际应用,这需要进一步探索。 M.G. Siegler: DeepSeek R1的出现对AI行业的影响巨大,市场反应强烈,对AI行业的影响程度堪比里氏8级地震。其低成本特性导致市场对AI模型成本和定价策略的重新评估,一些公司已经开始尝试使用DeepSeek R1模型,但仍需进一步观察其效果。 DeepSeek R1的技术创新在于其模型蒸馏技术和强化学习方法,这使得其能够将大型模型压缩成更小的模型,并降低成本提高效率。这与美国对中国的芯片限制有关,迫使DeepSeek采用不同的AI模型训练方法,这在当前美国的AI发展环境下是难以实现的。 DeepSeek的出现可能使长期以来在硅谷流行的“规模假设”失效,这将对英伟达等公司的巨额投资产生影响。DeepSeek的成功,也可能标志着AI发展进入一个新的阶段,但并不意味着创新就此结束。 DeepSeek对不同科技公司造成的影响不同,英伟达受到的冲击最大,而Meta由于其开源理念,可能相对受益。各大公司需要重新评估其AI投资策略,并寻找新的商业模式。 DeepSeek的出现可能导致AI行业投资模式的转变,一些公司可能需要减少对计算能力的投资,而更多地关注AI应用的开发。目前,AI技术仍处于早期阶段,需要进一步探索才能创造出显著的经济价值。

Deep Dive

Shownotes Transcript

Translations:
中文

It's time for a bonus episode exclusively about DeepSeq R1.

as the Chinese open source AI model roils markets and threatens to upend the generative AI industry. That's coming up right after this. From LinkedIn News, I'm Leah Smart, host of Every Day Better, an award-winning podcast dedicated to personal development. Join me every week for captivating stories and research to find more fulfillment in your work and personal life. Listen to Every Day Better on the LinkedIn Podcast Network, Apple Podcasts, or wherever you get your podcasts.

Welcome to Big Technology Podcast. We're doing a bonus edition today exclusively on DeepSeek, what it means for the AI industry, what it means for markets. We're going to touch on technology. We're going to touch on business.

and so thrilled that you're here for a bonus episode with us. We're joined today by MG Siegler. He's a writer and investor. He writes Spyglass. You can find it at spyglass.org. It's a great newsletter. It's a must read for me. And he has a great piece out called AI Finds a Way. Has Deep Seek changed the AI game or just some equations? MG, great to see you. Welcome to the show.

Great to see you, Alex. Thanks for having me back. And sorry for my crazy winter beard. It is very cold and rainy right now in London, so I'm not ready for spring yet. Hey, it fits the season. I was just out in London to interview Demis from DeepMind. That's right. I listened to that. That was very good. Yeah, and very timely now. Yes, I can confirm. The sun does not shine in that city this time of year.

So first of all, I want to talk a lot about, I mean, only about DeepSeek and DeepSeek R1 and what it means for the AI industry right now. We are just about, the markets will open on this show, so I'll have a sense as to what it's going to do today. But it's looking pretty bad, especially for NVIDIA and some others. As we get going, I just want to thank all the podcast listeners who pointed me to DeepSeek.

Because we had some comments that came in over the past few weeks, I was able to ask Demis about it. I was able to get it in as the lead story on Friday's show. So thank you. I appreciate all of you for pointing me towards DeepSeek. So let me just talk a little bit because we didn't touch on this Friday and we're going to definitely fill some holes that were left on the Friday show. We talked a little bit about how much it costs to train this model, but not necessarily about the benchmarks it hit and about the cost it costs to use this thing. So first of all, it's an open source model.

It's much smaller than any of OpenAI's model. Yet, on the AIME mathematics test, it scored 79.8% compared to OpenAI's O1 scoring 79.2%. So it's OpenAI's best model on that. It scored 97.3% on the math test.

500 and it beat open AI, which scored 96.4%. Look, these are lots of different benchmark tests, but you can tell that just by these numbers, it holds its own. And now the most remarkable part about this, it costs 55 cents per million token inputs and $2.19 per million token outputs.

Just to give you a sense, OpenAI costs $15 per million input tokens and $60 per million output tokens. That's 3.5% of the cost that it costs to run OpenAI's O1 models. And you can do it again. It's open source. You could download onto your computer and run it.

So basically what DeepSeek R1 has done in a nutshell, and then we'll turn it over to MG, is it has created models that are as performant as the state of the art, right? It's ranked number three in the chatbot arena at 3.5%, 3 to 5% of the cost. And that has huge implications for the technology, for the business, and we're going to get into those. So MG, first question for you, MG.

If there was an AI Richter scale, right, assessing how big of an earthquake this is, what would you give this development? So, I mean, it depends on what, I guess, level you're sort of measuring the magnitudes, right? Because as you noted, the markets will open and that's going to be

Right now, last I looked in pre-market trading, NVIDIA was down, I think, 10% to 11%. And that's the biggest hit right now. Microsoft and a bunch of others are in the 3% range. So from a pure market perspective, it seems like it's...

Let's call it an eight. It's not going to totally destroy the stock market right now, but it's going to be rough, it seems like today. From a bunch of other perspectives, I think it's probably...

a little bit less of a shake in these earlier days. And I think that's because everyone's still even now sussing out what exactly this means for all different sorts of things. You noted how much cheaper it is

um, to run than say open AI's models. And, and, you know, over the weekend, just reading all of these sort of reports about the model and how many, um, individual startups are even just changing, swapping out right already, um, because it's so much cheaper, um, to do what they're doing right now by swapping in, um, deep seeks models. And so what does that do immediately? Like, you know, do we have to have price cuts immediately? And, and, you know, I think, um,

You could sort of see OpenAI doing some stuff. I think Sam Altman tweeted, you know, maybe on Friday, like about how they were like bundling, rejiggering some of the bundles, right, that they have, like what's in the free offering and stuff. And it sort of feels like we're going to see more of that.

You know, as a response, obviously, to some of this. But then, you know, there was a big report, I think, in the information about Meta's response to this in particular, which seemed pretty interesting in that, like, you know, it's all hands on deck, certainly. And there's like all these different teams warring up. It's war room time.

They don't do war room training.

But anyway, going back to the original question on the on the Richter scale, you know overall I think a lot of people are still figuring this out but right now the market thing is gonna be the most acute one because that's obviously gonna open and I think it's gonna be pretty hard for you know this day at least and then

I think I read some of the early analyst reports on this and they're all over the place, right? Like there's some folks who are saying like, oh, this is awful for NVIDIA. Some folks are saying, you know, this is not a big deal. This actually could be good in the longer run for NVIDIA in ways. And, you know, and then from big tech on down, what the ramifications are there.

And you mentioned that some startups are already swapping in DeepSeq R1 for the models they're using right now. How widespread do you think that is? Are any of the startups that you speak with saying, okay, well, to hell with OpenAI or to hell with Lama, time to put DeepSeq in? Or is this just beginning? Because it's, again, something that dropped last week. Yeah.

Yeah, I think this is just beginning. I think, you know, people will experiment with it, right. Just to see like how much you could, you know, get while you swapping them out, given the price differentiation you were talking about, but also there's downsides, of course, like people have noted sort of the, you know, the, um, censorship within China and, and of certain terms. And so, um,

I don't think everyone is quite certain what's in there. It's an open source in that it's open weight, but it's not clear exactly everything that's going on in there right now. And so I do think that

if this proves out say if if deep sea can release another iteration of the model and it still is on the same sort of you know footing um i think that then you'll start to see more startups uh potentially taking it really seriously i think now it's just a wait and see approach for sure and just people trying out to see if it is in fact as good as they say because i think you know part of this like my initial gut reaction you know deep seek obviously as you noted had been around for you know

basically since December and didn't really get all of the mass of pylon until sort of Friday, right? When R1 came out. And in part, it's like, you know, I've just, I don't know why my mind was drawn to this, but it's sort of like when they were talking about the room temperature conductor, right? Like, and everyone was talking about, oh my God, like there's this huge breakthrough that's happened and this is going to revolutionize everything. And then it turns out, oh, you know,

Maybe there was some funny business in that claim, and maybe it wasn't all it was cracked up to be. And of course, that turned out to be the case. And so I'm not saying, obviously, that's not the case with DeepSeek. It seems like now this R1 release has legitimized it. And as you note, on leaderboards and whatnot, people have been testing this. And again, the startups are part of that pressure test.

Right. And so the funny business, just to get this out of the way, the funny business might be on the training side. Like we think that they trained it for much less money. We think that they trained it with...

inferior GPUs that have been sort of the only things they can get their hands on due to export controls. We're not 100% sure if that's the case. But I think the bottom line here is that this is an open source model. It has been replicated. I mean, it has been downloaded to people's computers and used as effective as it is. And I think that the thing is the methods and the cost savings and the performance, that's all real. So even if, you know, basically...

all of Silicon Valley without those export controls couldn't do this or didn't do this. And maybe it's because they had a different method and we'll get into that. But the fact is that there's no putting the genie back in the bottle right now, which is that this company has created something that can rival OpenAI's performance at 3% of the cost. That's the big thing.

I also just think the overall mentality is one of the more interesting sort of earthquakes, to use your phrasing of it, that's happening right now. It's like...

And I think Steven Sanofsky summarizes well. He wrote a very long tweet thread, as he is wont to do, but then he also published it on his newsletter as well. But he goes into the history and he obviously has a lot of good historical context from Microsoft days on forward about what is going on here. But it's also...

think important to to talk through like how the constraints that were put in place by the US because of the you know everything going on with with chip constraints and and sort of forcing

AI companies not to export to China, you know, led to sort of this very interesting cauldron that I think could only happen in a place like China right now because they're so constrained. Whereas in the US, like it's still the period of abundance, right, with AI and everyone's going after the scaling.

And it's and it's it's just not something they were going to focus on trying. You know, they're making the smaller models and making the mini versions of the models. And those are great. And we're seeing that. But China, you know, the folks working in China had to do this this way. And I just think it's something you couldn't have seen in hindsight arise out of the U.S. in our current environment. Right. OK, so I want to talk quickly about the technology very quickly about the technology and then get into some of the more business side applications here. So

M.G., could you tell us just at a really high level what DeepSeek has done to be able to get these results? Because, you know, it's one thing to say, OK, they were able to do it on worse chips with a smaller amount of data. But I think just it's important to very briefly highlight just the technical technological innovation here.

Yeah, I mean, so and I'm not a I won't be a technical expert on this, but from my understanding, it's basically, you know, obviously, as you know, it started the DeepSeek project started out of a hedge fund that was focused on quant trading, you know, in China. And they had acquired a bunch of NVIDIA chips. I think there were H100s, you know, before all the import restrictions came in and

And basically they had those servers up and running and presumably they were running a bunch of different models, including some of open AIs, but including also a bunch of the llama stuff that Meta has been working on. And they've just used the process of distillation to effectively

bring those bigger versions of the sort of state-of-the-art models and distill them down into smaller models, which eventually led to this R1, the equivalent of O1 on OpenAI's side.

And again, for a fraction of the cost, fraction of the compute, and a fraction of the size for these to be able to run. And that latter part seems like it's sort of being under-discussed right now, but is important because, yeah, all of these models have

constraints about how you can run them like on your personal machines, right? Because, you know, they're going to require so much RAM and so and so much memory to be able to do that. And if you can get them down to really small sizes, which again, the the bigger U.S. companies have been doing with these mini models, but they're sort of taking this bifurcated approach, whereas

Now we're getting to the point with this R1 model where it seems like it can run on pretty much a lot of different type of hardware, which again, they need to do in China because of the restrictions that they have there.

Right. And there's also a methodology change here, which is that they've gone from effectively self-supervised learning, which is what has been used to train all of the LLMs, all the big LLMs to this point, to pure reinforcement learning, where the models tend to figure out what the right answer is on their own, which is just fascinating. Yeah. And it seemed like the, you know, sort of the American powers that be maybe felt like

We weren't ready for that yet to happen, right? Like that was always the hope that we get to those points and that we still were in the scaling point again where you need someone in the loop to be able to check and make sure all these things are working. And China, this Chinese company, because of some of the restrictions that we just talked about, just went for it and it's proving itself.

Right. And just to harp on one more technical issue before moving on, the distillation of models to me is fascinating. That they could take any big model and distill it using this form of training and distillation.

effectively be able to replicate its performance. So I could take like a llama model with just 70 billion parameters and distill it and then all of a sudden run it with this reasoning reinforcement learning style approach and it's cheaper, more efficient. It's just...

I mean, again, like I think the entire world is still trying to wrap their head around this and they'll be more on this feed to talk about exactly how impressive this is. But to me in the early innings of this, that is astonishing. Yeah. And I mean, it again at a high level, it makes sense.

It's just it's incredible how it's happened, because like, do you need all of the world's knowledge, you know, in every single model for every single use case? Of course not. Like that's going to be overkill for almost everything that you're going to do. And so does it point to a world where, yeah, we sort of lead towards more of these specialized models that are distilled? And obviously that's been happening. But this this one is still, you know, a model that can effectively do most everything distilled down from from those bigger ones.

So there's one sort of big question that I think needs to be asked here, which is there's been this all Silicon Valley, and you point to this in your piece, all Silicon Valley has been operating on effectively the scaling hypothesis, which is that you add more compute. We talk about it all the time on the show, add more compute, add more data, add more power, add more training time effectively to these models, and you will improve.

And now what DeepSeek has shown is that you can actually do all this without that. And so I'm curious if you think that this invalidates the scaling hypothesis because, and it might seem kind of like a, you know, obscure thing, but it's very important because this sort of sets up the whole business conversation, which is if the scaling hypothesis is invalidated, then all that multi-trillion dollar investment in NVIDIA, you know,

Nvidia CPUs or GPUs, my bad, becomes sort of thrown into question. So what happens to the scaling hypothesis from here?

And it's fascinating timing too, right? Because this is at the same time that everyone has now talked about sort of the quote unquote AI wall being hit, right? And even Demis, you know, when you talk to him, he noted that he doesn't necessarily believe in, you know, a wall being hit, but he did acknowledge that things are slowing and it'll just take longer to get more, you know,

juice out of the squeeze as it were, right? And so that's sort of the natural evolution that's been happening and everyone is now pointing to it or at least acknowledging that some aspect of that is real. And now at the same time this comes along and it calls into sort of more question.

There's one other element that sort of, I think, is related to this, which was the big news story last week as well. The Project Stargate, OpenAI and NVIDIA and Oracle all coming together. And one of the more interesting elements of that was the fact that Microsoft is effectively...

pushing off the compute costs to Oracle and some of the other players in that situation. And, you know, there's all sorts of reasons, you know, potentially why they're doing that, obviously, given the interesting relationship between OpenAI and Microsoft. But at the very highest level, again, if they're thinking that

you know, our capex is going to be, we've already stated it's going to be 80 billion for the year. We don't want to add another several billion, you know, for this particular project. And why would they do that? In part, probably because they're not necessarily sure that it makes sense to pay the billions upon billions to OpenAI to keep trying to scale on the frontier models. And this is, you know, sort of in line with what DeepSeek just did.

Right. Yeah. It's interesting. We also talking about more, uh, Andreessen Horowitz who sat out opening eyes last round and we were wondering on the Friday show, maybe they saw this coming. And it is interesting. I mean, you, you put it pretty, pretty, uh, perfectly in your story. Uh, you say, um,

And big tech companies are now the largest and most well capitalized in the world, which means that they have effectively all the money that they can put towards scaling. And the hammer met the nail. But there's no point hammering the nail after it's already been put into place.

And that's the point that can't be predicted, but is obvious once it's done. The question is if deep seek just pointed to the nail already hammered effectively, did they just solve this? Uh, it's sort of like going up, um, the, the scaling question, uh, in a similar way, an analog for the same, the same thing. Right. And, and going back to the history of compute, like, right, all these, you know, the powers that be tend to spend at the time tend to spend a ton of, of capital, um,

on the build out of of whatever the new technology happens to be. And, you know, there's obviously we all benefit from it in the long run, but in the short run, you know, this this segues into, I guess, you know, what's what's potentially going on with Wall Street and what it means for these larger companies with regard to the spend. Yeah. And I just want to ask your the question that you put in your newsletter just to you directly. Did they just point to the nail? Like, is it done?

I mean, again, I don't want to, you know, caveat this out, but I do feel like it's the exact question that everyone is sort of going to be scrambling to answer over this next week. And I think that it's not going to be as black and white as that for sure. But I do think if I had to guess at a high level, I

I do think that there's some element to, yes, the nail is already sort of driven into the board and we're moving on to what the next steps are. That's not to say it's over and there's no innovation from here,

But I think all of these things are in a way related, like that we've just been talking about. And the fact that they're all coming together at the same time, I don't think is a coincidence. I think it's because like, yeah, we're at the point where we now need to move on to the sort of the next phase of the AI revolution, as it were.

Yeah, and let's get into the business. And I'm smiling here because you're making me think of we have Reid Hoffman on the show on Wednesday, and I interviewed him before R1 came out. And the first half of the conversation is just talking about all the billions of dollars that have been spent and when they're going to get an ROI. And I mean, I'm still going to run the conversation, but there's going to be some context in there. Yeah, it's interesting knowing...

Yeah. After the fact. But and it's also I think Sanofsky brought this up, too. And I was sort of looking into this more last week. You saw it was a smaller news item, but both Microsoft and Google had altered the way that they're basically bundling together AI within, you know, either the 365 suite and within the Google ecosystem.

Google suite of apps because they're clearly still trying to figure out how exactly you you make money off of all this spend and what the right model is and and how you spur on usage of it and This just comes in and throws a grenade and you know into that that equation again And this gets us to like some of like the real thorny business questions. So just to kick this off I took a look at what all the

Big tech companies were doing pre-market. So this will obviously change across the day, but I imagine they'll stay directionally kind of the same. NVIDIA down 10%, Microsoft down 4%, Google down 3%, Meta down 2.6%, S&P down 2%. So this is all based off of this deep seek reckoning or this deep seek

And let me just put the sort of question to you, I think about as pointedly as I can, which is that the AI industry up until this point, like all the numbers we're seeing within Wall Street, the trillion dollar market caps, the billions of investment, the billions of

that have been raised by companies like OpenAI and Anthropic from companies like Microsoft and Amazon, right? So this is basically the whole game here.

they have effectively been what's been driving the numbers. And the question is, can we, you know, basically Wall Street has been following that and saying we expect them to get a return based on those numbers. And in fact, a lot of this AI spend was just a wealth transfer, I would say, from like meta advertising to Lama, from Google search revenue to Gemini, from Microsoft Azure to OpenShift.

open AI. So what happens here? Because, you know, basically, if they, if a lot of the AI industry has been driven based off of subsidies coming from other businesses, and doesn't need that type of spend anymore, like does the party end?

So I think it's different for each company. Probably Microsoft and Google are closest aligned in terms of where they net out. And it's sort of interesting, the numbers you just rattled off with where the stocks are at. That feels just like a very...

clear picture from Wall Street what they think now, right? Like they think NVIDIA is going to get hit fast because in this doomsday scenario, because obviously they're the beneficiary from all of those companies, all those other companies.

that you mentioned, big tech is pouring as much money as possible as they can. They can't get enough chips fast enough into NVIDIA. And if they pause that, that obviously is bad news for NVIDIA in the short term. Again, I think there's longer term stuff that's different for NVIDIA, which we can talk about. But to just hit on the rest of this question right now, I think

that Microsoft and Google, which are, as we just mentioned, you know, are trying to sort of figure out the right models for how to charge for AI. I think that this puts them in a really tricky situation if the underlying economics just totally changed overnight of what AI's, yeah, underlying economic model should be. And so they were, you know,

moving around different pieces, trying to get to the right end state so that, yeah, they could ultimately prove to Wall Street, like, look, we're adding X amount on top of what we were already doing revenue-wise, thanks to AI.

And a little bit, there's a little bit of weird obfuscation stuff going on there, right? It's like, well, it's bundled in now to 365. And so, you know, we don't necessarily need to tell you exactly what the uplift is, but you can just, you know, assume that it's a part of this because it's all baked in and AI is like, you know, the new internet and blah, blah, blah.

And so, you know, there's ways that they can finesse the messaging around that. But that, you know, to your exact question, I do think that there's varying degrees of being worried, certainly within Google and Microsoft. Meta is more interesting because their open source philosophy, open weight philosophy and model

is so similar to what DeepSeek has done, right? And so the problem there, in my mind at least, is again, they're spending whatever Zuckerberg just threw out, $65 million or whatnot, he said, at the end of last week, that they're going to spend on CapEx.

Why are they spending that amount now if DeepSea can do it for pennies on the dollar, if not even less than that? And so what does that mean for their world? So in my view, high level, I think that Meta is probably in a bit better position than the other ones just because at the end of the day, they do want like, you know, their whole philosophy is to open sources, not for money.

altruistic reasons, but because they know that it's historically helped them help their business, you know, to open source these things. The question of if it's not them open sourcing it becomes pretty complicated if someone else's, you know, you have to use someone else's models, but they can pull back spend. It feels like a little bit easier than the other folks can. On the other end of the spectrum, OpenAI, like they're, you know, the entire business is

is sort of built around being at the frontier and they've done a great job with that. They're a little bit different than Google and Microsoft in my mind just because they've done a good job getting mindshare both in terms of brand and product, right? Like JetTBT is number two in the App Store right now behind DeepSeek, you know, for a reason. People are interested. It's a brand and they know it. And so

What does it look like, though, if they're not the ones sort of powering the models? I don't think that they would give up and go with DeepSeek's model necessarily. But what does it mean if they're not sort of the only one or the main frontier model maker providing that? So there's all sorts of interesting offshoots and ramifications of that.

So, MG, there's like two views right now in terms of like what could happen with all this spending. Right. One is Silicon Valley will continue to spend these billions and they might get, you know, incrementally better performance and stay slightly ahead of the open sources of the world, deep seeks of the world that can just emulate their models. The other side of it is that they continue to spend and then they basically hit AGI or

You know what I'm saying? If the performance increases that we've seen with such little... Sorry, if the performance increases that we've seen with such efficient use of capital from DeepSeat can be emulated, then imagine what you could do

With 100 times the amount of spend. So the models are about to become much more powerful and all these fantasies that people have about what they can do, many of which Demis and I spoke about last week, all of a sudden become feasible because the capital is there. So which side of this do you lay on?

you know, aside from sort of open AI, which obviously is again tied with Microsoft and now Oracle. But besides them, the rest of these are public companies and Wall Street, you know, like it or not, they have a say sort of over what they're going to do. Like,

If they're going to get hammered, and this is something I've sort of been harping on for a while, not because I think that they were doing the wrong thing necessarily with the spend, but it's just obvious that like it always comes back around, right? Where it's like I equated it last year to when all the movie studios during COVID and TV studios closed.

were just bulking up on streaming, right? And just spending as much money as possible as they could in order to build up their streaming services. And Wall Street loved it at that time because, you know, Disney and everyone else was just gaining millions and millions of subscribers. And it seems like they had a path to take on Netflix. And, you know, this was the future of the industry. It's still, by the way, the future of the industry. But Wall Street then all of a sudden turned on all that spend and decided, like, you need to cut, like,

spend X amount. You need to, you know, unfortunately cut the employee base and basically just become way more efficient while doing the same high level thing. And it was, you know, always obvious that at some point they were going to do that to the tech companies as well with regard to AI spend. And so, again, they can all have the right mentality about

Like this is the future and say the right things that this is the future and this spend is important. And I don't disagree with any of that, but still they have to answer to Wall Street. It's, you know, to some degree, maybe Zuckerberg less so because he, you know, controls the controls the company so strongly. But like certainly Microsoft and Google, to a lesser extent, are going to have to answer for a lot of that.

spend. And this is the first real, real test. Meta had some of it, right? Like there was some backlash last year around their spend and certainly dating back to the VR and AR and XR spend. And so they had to answer for some of that and Zuckerberg did, right? And he got rewarded for it

after the fact. And that's like the game they're playing here. They know that if they cut spend because Wall Street doesn't like to see all the AI spend, they'll get rewarded in the form of the stock going up and then all the ramifications from that. And so it's natural that that is going to play out that way. And so

I think the narrative then shifts to other levels of not necessarily obfuscation, but other ways of framing it. It's like, OK, we agree that we shouldn't spend tens of billions of dollars on NVIDIA server farms, but we need to build out our in-person AI robotics arms in order to keep these models and keep sort of the next phase going as we march towards AGI and yada yada.

So markets just opened and NVIDIA opens up down 11%. So still above $3 trillion. So it's not like the AI revolution is over, but down 11%. So just a cool couple hundred billion dollars shaved off the market cap in a morning. Let me talk to you a little bit about what these companies are saying back to Wall Street or actually talking to Wall Street.

about to allow them to keep spending. So Satya Nadella is doing his tweets. He says, second about Jevon's paradox. He calls, he says, Jevon's paradox strikes again. As AI gets more efficient and accessible, we will see its use skyrocket, turning it into a commodity we just can't get enough of.

Let's say that happens. I don't know if you saw this. Last night, Gary Tan, the president of YC, tweeted the same thing. And so I'm like, is this coordinated? Or I mean, it is. There must be a group text going on. Yeah, there's a group text maybe going on where it's like, this is the answer. And it's not like a totally BS answer to it. But there's much more nuance and context that's sort of required to get to that being the excuse for this.

So let me just basically talk about the elephant in the room that's been hanging over this full conversation and will be sort of like the spoken or unspoken part of this discussion as it goes forward this week, which is that let's say the cost of intelligence goes down to zero, right? So that's what everybody is basically aiming for. It's one of OpenAI's stated goals to make intelligence, you know, close to free as possible. They don't really make a lot of money selling off their API or they even maybe might lose, right?

We need to see AI applications. Like we need to see an economy that takes use of this technology that is so impressive, right? Like you look at the chain of thought even in deep seek and you're just like, how is a computer, you know, quote unquote, thinking through this stuff, right?

But the economy needs to take hold of this powerful technology and make use of it and put it into play for really meaningful economic use, whether or not the deep seek thing existed, right? Like billions of dollars of economic or trillions of dollars of economic value needed to be created from this generative AI moment.

And what do we have now? We have OpenAI who has ChatGPT with 300 million users, which is okay, but still losing billions a year to run that thing. Maybe they'll be able to be more efficient and make those couple billion a year from it. We have some enterprises...

putting this into play uh but everyone every enterprise i speak with um there's a couple use cases cool use cases here or there but mostly what you see is proof of concepts and many of those proof of concepts aren't going out the door so don't we need to see one way or the other um ai applications uh whether that's standalone or integrated within business software that start to prove the real value of this technology that we just haven't seen to date

I mean, the answer is yes, of course. The reality, though, is, you know, maybe this confluence of events right now is going to help that because it's sort of just as forcing a fundamental rethinking of a lot of what, you know, we've just been experiencing.

not going through the motions, but we've been on this path, right. To scaling as we were talking about. And that, you know, even right. Like Sam Altman has said, like they see line of sight now to AGI, they just have to, to, you know, just dot, dot, dot underpants and then profit. Uh,

from there. Better get there now. But they say they have line of sight, right? To know what they need to do and it's just a matter of execution and sort of, you know, getting everything aligned in order to do that. And if this moment with DeepSeek being the, you know, the biggest catalyst thus far of it, if it doesn't cause the entire industry to sort of rethink that and at the same time, to your point, like, you know, asking about

Do does that sort of drive us to move on from, yeah, just like this, this nonstop scaling of frontier models? That is awesome technology, but unclear how it works. And from a practical standpoint, do we start to yet distill this, you know, for lack of better phrase down to actual products? And, you know, when I when I think about that,

That leads back to like whenever it was six months ago, seven months ago when Apple did their Apple intelligence stuff, which you and I talked about. Right. And it's like everyone jumps on Apple. And there was another news cycle, I think, this past week because, you know, Siri can't can't correctly answer who won previous Super Bowls. Yeah, the Gruber post was amazing. Yeah.

But Apple's mentality from the get go with launching Apple intelligence has clearly been we need the we

for lack of a better phrase, don't necessarily care so much about, yeah, the frontier of the vanguard of this technology. We care about the day-to-day usage of it, right? And, you know, they have a few things that are sort of front-end facing that haven't really worked that try to use AI like the emoji creator and things like that. But most of it is just baking it into their products. And that's what we've seen too with obviously what we talked about with Microsoft and Google. They all have, you know, their own like

Some have video generation, some have some other of their own standalone products. For the most part, they're just going to be baked in. But to what we were talking about earlier, none of that is really the promise it felt like of what this larger movement was going to be and everyone's waiting for

you know, not necessarily AGI right now, but they just want some other forward facing, user facing version of AI that can be good. And ChatGPT has been the closest that we've gotten to that. And maybe some of these video products, you know, end up being the next phase of that.

But I think that you're right, that ultimately you have to get to something that comes of this, that really sort of moves all sorts of needles. And again, I wonder if this...

News cycle and and just pause now doesn't lead to more of that. I hope that that's the case Yeah, and I would say the Apple intelligence is almost the perfect example of the problem that I'm pointing toward which is that we have this technology That's so promising and yet even Apple cannot implement it Successfully correctly and that might I mean obviously it says something about Apple, but it might say something about the technology as well Yeah, and you know as with everything like I

With everything in technology, I think about dating back to my reporting days and whatnot. It's just like having seen so much in a few different cycles now,

Are we too early still, right? Like everyone has been talking about and believing that like this is the moment where this is like really happening and this is great. But I do think that if you took a step back, you might wonder if we're not still doing this too early, you know, and trying and all of these companies are not raising way too much money when the timing is just not right for us.

exactly what you're trying to ask the question about. Like, how do you turn these into products and how do you ultimately turn this into a business that returns the capital that was spent on it? Now, no company would admit that right now, but hindsight will only prove one way or another whether that's the case. And I think everyone still remains super optimistic that now is the right time and you want to keep your foot on the gas. But again, this deep seek stuff sort of causes a pause and a natural re-examination of just how much money to spend and

what you should be focused on. Let me ask you to put your investor hat on for a moment. Are there startups out there that would exist today that don't exist because

effectively buying compute from the apis or running llama is cost prohibitive but they would exist if intelligence was zero and that's effectively what deep seek is gonna put to the test yeah that's really interesting i don't i don't have i don't want to just try to come up with something off off the top of my head not that i know of but i do think at a high level that your question is a really interesting one and if

If this is going to be truly transformational, DeepSeek as a whole, it will lead to something like that, right? Like a bunch of companies coming out and not just yet because it's not just the technical aspect. It's not just driving down costs because that seems like it's sort of going to happen as a result of that, which is great. But does this actually yield new companies that couldn't have existed beforehand that

And I don't know, like, I can't think of any off the top of my head. But that's also why I'm not a startup founder. And, you know, hopefully there are startups out there that are that are going to latch on to this. But something tells me that the answer is no. And the reason is, is because investors have been dying to throw money at AI companies and have been willing to lose a lot of money if the idea is promising enough. And.

I don't know. We haven't, we haven't seen a wave of AI startups hit. At least there have been many, uh, but you know, they're, they're not like, it's not like the, um, you know, the beginning of the mobile era where there was like a new consumer startup every day. It just isn't happening that way. In fact, most of the action is enterprise. One other just wrinkle and layer of that. Um,

which i feel like has been overshadowed in in all of the recent news but you know we talked about and talked about a lot last year but as the regulatory regime is changing now um if

If M&A sort of doesn't pick up with regard to exactly the type of companies you're talking about, right? Like they have great teams. They're working with, you know, this technology and they clearly know how to do things with it, but they haven't gotten the product right. They haven't gotten the business right. And so they're scooped up by the, you know, the Metas, the Googles, the Microsofts, the open eyes of the world. And that's,

you know, in and of itself won't be that interesting other than those companies getting good talent, perhaps. But if it just reignites sort of, you know, a passion within really early stage startup founders to keep reaccelerate sort of going after new problems, right? Like I do feel like there was a bit of a chilling effect the past year because M&A had basically been shut off.

that sort of kept people staying at Google and staying at Meta and staying at OpenAI, not forming new startups as they might have in years past because they knew that there was the potential. Obviously, pie in the sky, they want to build a big company, but there was also the potential, frankly, right, to like, you know, sell, build something that's big enough to sell for multi hundreds of millions of dollars, if not billions of dollars to some of these other companies. And so

You know, that might come into play with some of this. All right, let's put a bow in this conversation. You say the real problem is that it won't be so simple to simply pull back spend beyond a lot of it already being committed. There's obviously still a very real risk that deep seek is just a blip on the radar and not the bomb that blows up everything. What are we looking at over the next couple of months when it comes to the aftermath of this earthquake to go back to our original question?

And so that's just a call out to the obvious thing that everyone likes to overreact, obviously, to big news stories and big news cycles. And again, as we've been talking about, this is legitimate, but how legitimate is it? So we'll even see potentially play out over the course of today in the stock market, like

do they start to get nerves calmed a bit by, yeah, this talk of like, well, actually this isn't so bad for NVIDIA because while it hurts their, their immediate, it could potentially hurt their immediate, um,

Money coming in the door in the longer run, you know, it's it's again, Javon's paradox stuff where it's like, yeah, it's it's going to raise raise all boats as as this just permeates everything. And so they need chips and yada yada. And so that could help.

But yeah, I mean, I think that it won't be so easy also for, as I noted, for all these companies to pull back spend because they've already committed to buying X number of H200 chips. And then soon enough, we'll get the next iteration, you know, announced down the road. And so all these supercomputer manufacturers

mega clusters of data centers that are being built right now. They're just not going to put the brakes on all of that because there's a risk they're all playing in the same game, right? And if one of them pauses, maybe they get a short-term Wall Street pat on the back.

but if they're wrong that's like catastrophic and that's you know it's like a fire firing the ceo type offense um you know if if this is just uh you know even a blip on the radar obviously undersells it a bit but if this is not ultimately like a real fundamental sea change situation and is more just like a a step on on the road um

they might still want to keep their foot on the gas. Yeah, it's going to be very interesting to watch. The website is spyglass.org. The piece AIF finds a way. Joined, of course, by MG Seigler. MG, great to see you again. Thanks for coming on the show. Thanks for having me on. All right, everybody. Thank you for listening. We'll be back on Wednesday with my interview with Reid Hoffman. Obviously a little different now, but maybe...

As MG puts it out, maybe we shouldn't be overreacting too much. So looking forward to speaking with you then, and we'll see you next time on Big Technology Podcast.