We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
People
M
Marc Andreessen
联合创始人和风险投资家,专注于人工智能和技术领域的投资。
N
NLW
知名播客主持人和分析师,专注于加密货币和宏观经济分析。
Topics
NLW: DeepSeek的AI模型对美国在人工智能领域的领先地位提出了挑战,尤其是在开源与闭源的辩论中。我观察到,DeepSeek的应用程序因其高质量的推理模型而受欢迎。虽然OpenAI最初是开源的,但后来出于安全考虑改变了政策,停止完全分享研究成果。然而,许多人质疑OpenAI的闭源策略,认为这更多是出于竞争而非安全考虑。DeepSeek的出现改变了关于开源与闭源AI的讨论,促使我们重新思考美国的AI发展策略。 Eric Schmidt: (通过AI生成) DeepSeek的R1推理模型在逻辑任务上与OpenAI的GPT-4相当,且成本更低,其整个模型集合都是开源的,允许任何人复制和构建。中国公司正在成为事实上的开源领导者,而大多数美国公司则保持封闭,这是一个奇怪的转变。为了保持竞争力,美国必须支持开源生态系统的发展。DeepSeek的成功质疑了预训练的重要性,并展示了如何通过强化学习以更低的成本进行推理训练。我认为DeepSeek的发布标志着一个转折点,美国创新的前进道路不仅包括加强开源开发,还包括鼓励分享培训方法和增加对人工智能研发的投资。 Marc Andreessen: (通过NLW引用) 我认为对开源AI的恐惧是为了监管和扼杀开源AI。关于AI的秘密已经泄露,开源实现将大量涌现。每个人都知道如何编写transformer,如何进行RLHF,如何使用强化学习进行推理。除了DeepSeek和LLAMA之外,还将有成千上万的开源实现。这无法被放回盒子里。 Sam Altman: (通过NLW引用) 我认为OpenAI可能在开源问题上站错了队,需要新的开源策略。

Deep Dive

Chapters
The emergence of DeepSeek, a Chinese AI company, has challenged the US's dominance in AI. DeepSeek's open-source models are surprisingly competitive, raising questions about the US's closed-source approach and its implications for future AI development. The discussion covers the cost-effectiveness and performance of DeepSeek's models compared to those of OpenAI.
  • DeepSeek's AI models challenge US AI dominance.
  • DeepSeek's open-source approach contrasts with the closed-source models of US companies.
  • Cost-effectiveness and performance of DeepSeek's models are key discussion points.

Shownotes Transcript

Translations:
中文

Thank you.

Hello, friends. It has now been a few weeks since the Black Swan event that was the launch of DeepSeek. DeepSeek, of course, is a Chinese company spun out of a hedge fund, no less, whose AI models have recently totally challenged expectations and thoughts around just how far ahead the U.S. actually is.

Now, we've had a lot of chance to talk about different aspects of the DeepSeek story. How part of the reason that their app has been so popular is that whereas OpenAI was giving people the subpar models in their free chat GPT app, DeepSeek was actually giving a reasoning model right there.

There were also some UI innovations based on the way that it exposed its thought during its reasoning. And of course, the biggest part of the debate has been around distillation techniques and how they were able to get this much performance with so little money. And or on the flip side, disbelief that that actually happened. But the part of the conversation that I want to come back to, which I think is potentially the most significant when it comes to shifting the industry, is the idea of the implications for how the United States in specific thinks about open source versus closed source AI.

Now, this has been an interesting and ongoing debate. OpenAI was, of course, called OpenAI, but over the years started to shift its policy. It stopped sharing its research in full, and obviously none of the big models have been open source for some time. A lot of the justification for this was about safety and about it being dangerous to share things like model weights with the wider world because of all the bad people out there who might use them for nefarious purposes. There

There, of course, have been also tons of counterarguments, alongside a lot of assessment of the pragmatics of trying to keep things closed source, and more than a fair bit of skepticism that safety concerns are actually the reason as opposed to competitiveness concerns. I think complicating this fact is that one of the loudest people who has been contra to OpenAI's strategy is Elon Musk, who obviously has a very big axe to grind over there, and who, as Sam Altman has pointed out, hasn't been open sourcing the main Grok models either.

In any case, it feels very much like DeepSeek has shifted the nature of the conversation. We're going to read a quick piece, or rather turn it over to AI to read a quick piece, by former Google CEO Eric Schmidt called Will China's Open Source AI End U.S. Supremacy in the Field? With the advent of DeepSeek, the balance of power between the two nations appears to be shifting. So I'm going to throw it over to an 11 Labs version of myself, and then we will come back and keep discussing. It has become almost a cliche to say that the artificial intelligence landscape is changing fast.

But in recent days, even those on the cutting edge of AI research were taken by surprise by a Chinese company. Last week, the AI company DeepSeek released its R1 reasoning model, which is on a par with OpenAI's O1 and much better than the chat GPT models across a variety of logic tasks, including math and coding.

The cost of running it is also much lower, only about 2% of what OpenAI charges. And on Monday, DeepSeek released Janus Pro, a model small enough to run on your laptop that can generate synthetic images, which it claims outperform OpenAI's DALI 3. DeepSeek's speed of AI innovation is taking the world by storm. What's even more remarkable is that DeepSeek's entire collection of models is open source, which in this case means they have open weights that anyone can reproduce and build on top of.

It's a peculiar moment when a Chinese company becomes the de facto open source leader, while most major American firms, with the exception of Meta, continue to keep their methodologies tightly under wraps. In fact, this is a growing trend for Chinese AI companies, from startups such as Minimax to tech giants such as Alibaba, that are giving developers worldwide free access to their AI models.

Until now, closed-source models such as OpenAI's O3 and Anthropic's Claude 3 Opus were considered the industry standards with the most advanced capabilities. And they were built in the United States. Open-source and Chinese models were thought to be months behind. But DeepSeek's R1 and Janus Pro show just how quickly the tides of technological supremacy can turn. The introduction of these models has roiled stock markets and caused U.S. tech stocks to plunge.

The balance of power now appears to be shifting along two key axes, one between the United States and China, and another between closed and open-source models. Defenders of closed-source models are betting that they can preserve their capability gap by protecting their model weights and training methodologies. Open-source advocates, on the other hand, argue that transparency, allowing others to build on their work, can enable these systems to rapidly catch up with larger closed models.

If the open-source thesis is correct, this would turn the AI ecosystem on its head. Open-source models are generally cheaper to use, so when two equally capable models are available, one open, one closed, the open-source model is likely to gain wider adoption, giving it a strategic advantage. The United States already has the best closed models in the world. To remain competitive, we must also support the development of a vibrant open-source ecosystem.

The race between open and closed-source AI, as well as between the United States and China, does not yet have a clear winner. But there is clearly mounting pressure on America's big tech players if DeepSeat can compete with them using far fewer resources.

Export controls were aimed at choking off China's access to the most advanced computer chips, impeding its ability to keep pace. But in fact, the relative dearth of high-performing chips in China might have pushed the nation's companies and researchers to be more efficient and led them to uncover new methodologies that significantly reduced training costs. For example, DeepSeq demonstrated that large-model training could be made more efficient by bypassing the traditional supervised fine-tuning stage.

They even created R10, a model that omits this step in AI training, to challenge the research community's assumptions about fine-tuning's indispensability. DeepSeek's success has also called into question the importance of pre-training, which involves training ever-larger models that predict the next word based on vast amounts of text. This process requires enormous upfront investment in graphics processing units, GPUs, and data. So much data that OpenAI co-founder Ilya Sutskever recently noted we might soon exhaust all the data available on the internet.

But there is another emerging way to improve models' performance. Introduced with OpenAI's O1 model in December, this approach enables models to perform reasoning through self-reflection, similar to how humans reason, using intermediate steps and self-correction to reach a final answer. The training recipe for this approach had previously been closely guarded by OpenAI. DeepSeek blew the lid off that by publishing a paper detailing how it works, allowing others to implement the process.

DeepSeq even demonstrated that you can do this much more cost-effectively by taking a publicly available base model, such as Meta's LLAMA3, and teaching it to reason through reinforcement learning, a trial-and-error process with human-devised feedback and rewards.

Over time, the models seem to spontaneously learn how to reason, backtrack when they hit dead ends, and explore novel approaches. This method eliminates the need to expensively pre-train a new base model, and its implications for AI innovation are profound. Traditionally, even the top-funded university labs have struggled to contribute to AI research due to computing and data limitations. With DeepSeq's breakthrough, the moat surrounding large, well-funded companies might be shrinking.

It is unlikely that American frontier model companies will change their business models anytime soon, nor is it immediately clear that they should. Open and closed competition will most likely find a natural equilibrium, with a range of different offerings and price points for different users. But DeepSeek's release marks a turning point.

The path forward for American innovation involves not just ramping up open-source development, but also encouraging the sharing of training methodologies and increasing investment in AI research and development. Exemplified by the White House's recent announcement of the Stargate Project, which aims to spend $500 billion on AI infrastructure over the next four years, America's competitive edge has long relied on open science and collaboration across industry, academia, and government.

we should embrace the possibility that open science might once again fuel American dynamism in the age of AI. Today's episode is brought to you by Vanta. Trust isn't just earned, it's demanded. Whether you're a startup founder navigating your first audit or a seasoned security professional scaling your GRC program, proving your commitment to security has never been more critical or more complex. That's where Vanta comes in.

Businesses use Vanta to establish trust by automating compliance needs across over 35 frameworks like SOC 2 and ISO 27001. Centralized security workflows complete questionnaires up to 5x faster and proactively manage vendor risk. Vanta can help you start or scale up your security program by connecting you with auditors and experts to conduct your audit and set up your security program quickly. Plus, with automation and AI throughout the platform, Vanta gives you time back so you can focus on building your company.

Join over 9,000 global companies like Atlassian, Quora, and Factory who use Vantage to manage risk and prove security in real time.

If there is one thing that's clear about AI in 2025, it's that the agents are coming. Vertical agents by industry, horizontal agent platforms, agent-based platforms.

agents per function. If you are running a large enterprise, you will be experimenting with agents next year. And given how new this is, all of us are going to be back in pilot mode.

That's why Superintelligent is offering a new product for the beginning of this year. It's an agent readiness and opportunity audit. Over the course of a couple quick weeks, we dig in with your team to understand what type of agents make sense for you to test, what type of infrastructure support you need to be ready, and to ultimately come away with a set of actionable recommendations that get you prepared to figure out how agents can transform your business.

If you are interested in the agent readiness and opportunity audit, reach out directly to me, nlw at bsuper.ai. Put the word agent in the subject line so I know what you're talking about. And let's have you be a leader in the most dynamic part of the AI market. Hello, AI Daily Brief listeners. Taking a quick break to share some very interesting findings from KPMG's latest AI Quarterly Pulse Survey.

Did you know that 67% of business leaders expect AI to fundamentally transform their businesses within the next two years? And yet, it's not all smooth sailing. The biggest challenges that they face include things like data quality, risk management, and employee adoption. KPMG is at the forefront of helping organizations navigate these hurdles. They're not just talking about AI, they're leading the charge with practical solutions and real-world applications.

For instance, over half of the organizations surveyed are exploring AI agents to handle tasks like administrative duties and call center operations. So if you're looking to stay ahead in the AI game, keep an eye on KPMG. They're not just a part of the conversation, they're helping shape it. Learn more about how KPMG is driving AI innovation at kpmg.com slash US. All right, back to real NLW here. In an almost throwaway line, Schmidt gets at, I think, what has captured so many people's attention with this whole story.

That line is, it's a peculiar moment when a Chinese company becomes the de facto open source leader, while most major American firms, with the exception of Meta, continue to keep their methodologies tightly under wraps.

This is a strange turn of events. It doesn't seem like what should be. And of course, DeepSeek is not free from influence of China in the way you'd expect. It will not engage with certain politically sensitive questions, which is why, of course, many people who have chosen to engage with DeepSeek have done so in versions that are powered by the API that can get around some of those restrictions.

When it comes to the big labs, some have doubled down on their arguments against open source. Czarnik shared an interview with Anthropic CEO Dario Amadei, who said that AI safety evaluations conducted on DeepSeek showed it was the worst performing model they'd ever tested at generating potentially dangerous information. They said it had absolutely no blocks whatsoever against generating this information. Now, Amadei pointed out that he doesn't think that these models are actually dangerous in any way, but that we're on these exponential curves, and so that is a safety consideration.

Then again, some people rejected that position entirely, with Marc Andreessen writing, "'Fear-mongering for regulatory capture and to kneecap open-source AI. The existential threat of open-source AI is the big AI cartel.'" He actually even then quote-tweeted himself, saying, "'The reality is the secrets are out. Everyone knows how to code a transformer, how to RLHF, how to use reinforcement learning for reasoning. There will be thousands of open-source implementations in addition to DeepSeek and LLAMA. There is no putting this back in the box.'"

And whether it's just an acceptance of inevitable reality, a change of opinion on the safety, or competitive pressure, it's not just Andreessen saying this. Sam Altman in a Reddit AMA said that he thought that OpenAI might have been on the wrong side of history with this and needs a new open source strategy. And then just yesterday, news started to come out that suggested that Baidu had announced that it would be open sourcing its future Ernie models. After, as Interconnected Capital's Kevin Hsu put it, being one of the staunchest closed source model makers.

So it definitely feels to me like there are shifting sands here. I think the next big question is going to be seeing how it all plays out in the policy sphere, as that could really shape this discourse as well. For now, though, that is going to do it for today's AI Daily Brief, Long Reads edition. Till next time, peace.