Leading AI labs have hit scaling limits because synthetic data, generated by large language models, is not enabling further scaling in pre-training. In the pre-training world, scaling compute led to better models, but now labs are shifting to test-time compute, where models are asked to explore potential solutions and verify them iteratively.
The shift to test-time compute better aligns revenue generation with expenditures, making financial forecasting more manageable. It also shifts the focus from massive upfront CapEx to more efficient, usage-based inferencing, which can be more cost-effective and scalable.
Small teams and open source models are becoming more competitive because the shift to test-time compute reduces the need for massive pre-training compute and large datasets. Open source models like LAMA provide a solid foundation that small teams can fine-tune and optimize for specific use cases with minimal capital.
OpenAI's strategic positioning is strong due to its brand and consumer mindshare, but it is vulnerable because it may struggle to compete with free models from Google and Meta. If pre-training scaling is plateauing, OpenAI's future success may depend on its ability to innovate and maintain consumer lock-in.
Google's AI strategy is disruptive because it leverages its existing strengths in deep learning and self-play, but it is also defensive as it tries to maintain its dominance in search and enterprise. The challenge is whether its AI innovations can replicate the success of its current business model.
Private market valuations for AI companies are high due to the excitement around AI applications and the dramatic drop in compute costs. While these companies show promise, the high valuations also reflect the potential for intense competition and the rapid pace of innovation.
Test-time compute leads to more efficient and predictable infrastructure needs because it aligns with the bursty nature of inference tasks, rather than the constant, high-utilization needs of pre-training. This allows for better resource management and cost savings, making it easier for hyperscalers to forecast and meet demand.
Recursive self-improvement is significant in the path to ASI because it suggests that AI systems could enhance their own capabilities without human intervention. Examples like AlphaGo and poker bots show that algorithms can sometimes operate outside the bounds of their initial training, which is a key aspect of achieving super intelligence.
The agglomeration of AI innovation in Silicon Valley is important because it fosters multidisciplinary collaboration and the rapid synthesis of ideas. Human network effects and the concentration of talent in one place can lead to significant breakthroughs and advancements in AI technology.
My guests today are Chetan Puttagunta) and Modest Proposal). Chetan is a General Partner at venture firm Benchmark, while Modest Proposal is an anonymous guest who manages a large pool of capital in the public markets. Both are good friends and frequent guests on the show, but this is the first time they have appeared together. And the timing couldn’t be better - we might be witnessing a pivotal shift in AI development as leading labs hit scaling limits and transition from pre-training to test-time compute. Together, we explore how this change could democratize AI development while reshaping the investment landscape across both public and private markets. Please enjoy this discussion with Chetan Puttagunta and Modest Proposal.
My guests today For the full show notes, transcript, and links to mentioned content, check out the episode page here.)
-----
This episode is brought to you by** Ramp**). Ramp’s mission is to help companies manage their spend in a way that reduces expenses and frees up time for teams to work on more valuable projects. Ramp is the fastest growing FinTech company in history and it’s backed by more of my favorite past guests (at least 16 of them!) than probably any other company I’m aware of. It’s also notable that many best-in-class businesses use Ramp—companies like Airbnb, Anduril, and Shopify, as well as investors like Sequoia Capital and Vista Equity. They use Ramp to manage their spending, automate tedious financial processes, and reinvest saved dollars and hours into growth. At Colossus and Positive Sum, we use Ramp for exactly the same reason. Go to** Ramp.com/invest**)** to sign up for free and get a $250 welcome bonus.**
–
This episode is brought to you by** Alphasense**). AlphaSense has completely transformed the research process with cutting-edge AI technology and a vast collection of top-tier, reliable business content. Imagine completing your research five to ten times faster with search that delivers the most relevant results, helping you make high-conviction decisions with confidence. AlphaSense provides access to over 300 million premium documents, including company filings, earnings reports, press releases, and more from public and private companies. Invest Like the Best listeners can get a free trial now at** Alpha-Sense.com/Invest**)** and experience firsthand how AlphaSense and Tegas help you make smarter decisions faster.**
-----
Invest Like the Best is a property of Colossus, LLC. For more episodes of Invest Like the Best, visit** joincolossus.com/episodes**)**. **
Follow us on Twitter:** @patrick_oshag**)** |**** @JoinColossus**)
Editing and post-production work for this episode was provided by The Podcast Consultant (https://thepodcastconsultant.com)).
Show Notes:
(00:00:00) Welcome to Invest Like the Best
(00:05:30) Introduction to LLM Scaling Challenges
(00:07:25) Synthetic Data and Test Time Compute
(00:08:53) Implications of Test Time Compute
(00:11:19) Public Tech Companies and AI Investments
(00:16:58) Small Teams and Open Source Models
(00:29:02) Strategic Positioning of Major AI Players
(00:35:49) AGI and Future Prospects
(00:46:50) AI Application Layer and Investment Opportunities
(00:54:18) The Paradigm Shift in AI Reasoning
(00:55:34) Investing in AI-Powered Solutions
(00:58:46) Economic Impacts of AI Advancements
(01:00:19) The Future of AI and Model Stability
(01:02:52) Private Market Valuations and Compute Costs
(01:05:05) Infrastructure and Utilization in AI
(01:12:50) The Role of Hyperscalers and GPUs
(01:18:02) The Evolution of AI Applications
(01:27:56) Philosophical Questions on AGI and ASI
(01:34:31) The Importance of Innovation Hubs