We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Chetan Puttagunta and Modest Proposal - Capital, Compute & AI Scaling - [Invest Like the Best, EP.400]

Chetan Puttagunta and Modest Proposal - Capital, Compute & AI Scaling - [Invest Like the Best, EP.400]

2024/12/6
logo of podcast Invest Like the Best with Patrick O'Shaughnessy

Invest Like the Best with Patrick O'Shaughnessy

AI Deep Dive AI Insights AI Chapters Transcript
People
C
Chetan Puttagunta
M
Modest Proposal
Topics
Chetan Puttagunta: 大型语言模型(LLM)的扩展已达到瓶颈,正从预训练转向测试时计算。预训练依赖于海量数据,但数据资源日渐枯竭,合成数据也无法有效解决这一问题。测试时计算则通过迭代优化,在推理阶段进行扩展,但其扩展性仍面临挑战,算法、数据和硬件的改进依然重要。小型团队利用开源模型,以较低的成本快速达到先进模型的性能水平,这改变了对模型层投资的看法。 随着模型扩展模式的变化,对模型公司的投资策略也需要调整。Meta的Llama模型及其开源策略,对模型层的标准化和生态系统的发展起到了重要作用。投资重点转向AI应用层,因为其能创造以前无法实现的应用,并能快速实现商业成功,且模型的稳定性也为应用层发展提供了有利条件。 在新的推理范式下,应用开发者可以专注于优化自身工具,以提高推理效率,并构建模型公司不太可能构建的技术和工具。 大型基础模型厂商的战略定位和未来发展存在不确定性。OpenAI凭借ChatGPT获得了巨大的消费者市场份额,但其未来发展取决于能否在免费产品竞争中保持优势。Anthropic拥有强大的技术实力,但缺乏市场份额和明确的战略。Google拥有强大的技术实力和消费者市场,但其能否在新的竞争环境中保持优势仍存在不确定性。 Modest Proposal: 人工智能已渗透到更广泛的市场,对整体市场估值有重大影响。人工智能投资的预期结果分布已发生变化,从预训练转向推理时间。从预训练转向推理时间计算,使支出与收入更匹配,提高效率,并需要重新思考网络架构的设计。 大型科技公司需要重新评估其在人工智能领域的战略定位和投资策略。从预训练转向推理时间计算,使支出与收入更匹配,提高效率。这将影响网络设计、能源利用和电网设计等方面。 目前市场对这一转变的认识不足,需要重新评估各种情景路径。AGI(通用人工智能)可能即将到来,其定义是能够胜任各种经济上有价值工作的自主系统。 大型基础模型厂商的战略定位和未来发展存在不确定性。OpenAI的成功很大程度上依赖于其在消费者市场的领先地位,但其未来发展取决于能否保持这一优势。Anthropic和XAI面临着来自开源模型和大型科技公司的竞争压力。 对AGI和ASI(人工超级智能)的讨论和思考,需要考虑技术进步、经济效益和伦理风险等多方面因素。

Deep Dive

Key Insights

Why are leading AI labs hitting scaling limits and transitioning from pre-training to test-time compute?

Leading AI labs have hit scaling limits because synthetic data, generated by large language models, is not enabling further scaling in pre-training. In the pre-training world, scaling compute led to better models, but now labs are shifting to test-time compute, where models are asked to explore potential solutions and verify them iteratively.

Why is the shift to test-time compute beneficial for the AI industry?

The shift to test-time compute better aligns revenue generation with expenditures, making financial forecasting more manageable. It also shifts the focus from massive upfront CapEx to more efficient, usage-based inferencing, which can be more cost-effective and scalable.

Why are small teams and open source models becoming more competitive in AI development?

Small teams and open source models are becoming more competitive because the shift to test-time compute reduces the need for massive pre-training compute and large datasets. Open source models like LAMA provide a solid foundation that small teams can fine-tune and optimize for specific use cases with minimal capital.

Why is OpenAI's strategic positioning considered both strong and vulnerable?

OpenAI's strategic positioning is strong due to its brand and consumer mindshare, but it is vulnerable because it may struggle to compete with free models from Google and Meta. If pre-training scaling is plateauing, OpenAI's future success may depend on its ability to innovate and maintain consumer lock-in.

Why is Google's AI strategy both disruptive and defensive?

Google's AI strategy is disruptive because it leverages its existing strengths in deep learning and self-play, but it is also defensive as it tries to maintain its dominance in search and enterprise. The challenge is whether its AI innovations can replicate the success of its current business model.

Why are private market valuations for AI companies so high, and what signals do they reflect?

Private market valuations for AI companies are high due to the excitement around AI applications and the dramatic drop in compute costs. While these companies show promise, the high valuations also reflect the potential for intense competition and the rapid pace of innovation.

Why is test-time compute leading to more efficient and predictable infrastructure needs?

Test-time compute leads to more efficient and predictable infrastructure needs because it aligns with the bursty nature of inference tasks, rather than the constant, high-utilization needs of pre-training. This allows for better resource management and cost savings, making it easier for hyperscalers to forecast and meet demand.

Why is the concept of recursive self-improvement significant in the path to ASI (Artificial Super Intelligence)?

Recursive self-improvement is significant in the path to ASI because it suggests that AI systems could enhance their own capabilities without human intervention. Examples like AlphaGo and poker bots show that algorithms can sometimes operate outside the bounds of their initial training, which is a key aspect of achieving super intelligence.

Why is the agglomeration of AI innovation in Silicon Valley important?

The agglomeration of AI innovation in Silicon Valley is important because it fosters multidisciplinary collaboration and the rapid synthesis of ideas. Human network effects and the concentration of talent in one place can lead to significant breakthroughs and advancements in AI technology.

Shownotes Transcript

My guests today are Chetan Puttagunta) and Modest Proposal). Chetan is a General Partner at venture firm Benchmark, while Modest Proposal is an anonymous guest who manages a large pool of capital in the public markets. Both are good friends and frequent guests on the show, but this is the first time they have appeared together. And the timing couldn’t be better - we might be witnessing a pivotal shift in AI development as leading labs hit scaling limits and transition from pre-training to test-time compute. Together, we explore how this change could democratize AI development while reshaping the investment landscape across both public and private markets. Please enjoy this discussion with Chetan Puttagunta and Modest Proposal.

My guests today For the full show notes, transcript, and links to mentioned content, check out the episode page here.)

-----

This episode is brought to you by** Ramp**). Ramp’s mission is to help companies manage their spend in a way that reduces expenses and frees up time for teams to work on more valuable projects. Ramp is the fastest growing FinTech company in history and it’s backed by more of my favorite past guests (at least 16 of them!) than probably any other company I’m aware of. It’s also notable that many best-in-class businesses use Ramp—companies like Airbnb, Anduril, and Shopify, as well as investors like Sequoia Capital and Vista Equity. They use Ramp to manage their spending, automate tedious financial processes, and reinvest saved dollars and hours into growth. At Colossus and Positive Sum, we use Ramp for exactly the same reason. Go to** Ramp.com/invest**)** to sign up for free and get a $250 welcome bonus.**

This episode is brought to you by** Alphasense**). AlphaSense has completely transformed the research process with cutting-edge AI technology and a vast collection of top-tier, reliable business content. Imagine completing your research five to ten times faster with search that delivers the most relevant results, helping you make high-conviction decisions with confidence. AlphaSense provides access to over 300 million premium documents, including company filings, earnings reports, press releases, and more from public and private companies. Invest Like the Best listeners can get a free trial now at** Alpha-Sense.com/Invest**)** and experience firsthand how AlphaSense and Tegas help you make smarter decisions faster.**

-----

Invest Like the Best is a property of Colossus, LLC. For more episodes of Invest Like the Best, visit** joincolossus.com/episodes**)**. **

Follow us on Twitter:** @patrick_oshag**)** |**** @JoinColossus**)

Editing and post-production work for this episode was provided by The Podcast Consultant (https://thepodcastconsultant.com)).

Show Notes:

(00:00:00) Welcome to Invest Like the Best

(00:05:30) Introduction to LLM Scaling Challenges

(00:07:25) Synthetic Data and Test Time Compute

(00:08:53) Implications of Test Time Compute

(00:11:19) Public Tech Companies and AI Investments

(00:16:58) Small Teams and Open Source Models

(00:29:02) Strategic Positioning of Major AI Players

(00:35:49) AGI and Future Prospects

(00:46:50) AI Application Layer and Investment Opportunities

(00:54:18) The Paradigm Shift in AI Reasoning

(00:55:34) Investing in AI-Powered Solutions

(00:58:46) Economic Impacts of AI Advancements

(01:00:19) The Future of AI and Model Stability

(01:02:52) Private Market Valuations and Compute Costs

(01:05:05) Infrastructure and Utilization in AI

(01:12:50) The Role of Hyperscalers and GPUs

(01:18:02) The Evolution of AI Applications

(01:27:56) Philosophical Questions on AGI and ASI

(01:34:31) The Importance of Innovation Hubs