By 2025, the geopolitical focus in AI will shift from a U.S.-China arms race to a competition over which country's AI systems become the global standard. The U.S. and China will vie for dominance in exporting adaptable AI technologies worldwide, with many countries caught in the middle as 'geopolitical swing states.' The U.S. aims to ensure Western AI technology dominates globally to counter Chinese expansionist initiatives like the Belt and Road Initiative.
U.S. leadership in AI is critical for national security, particularly in potential conflicts like over Taiwan, where superior AI could provide a decisive advantage. Additionally, U.S. dominance ensures that democratic values, free speech, and open conversation are embedded in the global AI infrastructure, serving as a cultural export that aligns with American ideals.
China prioritizes rapid deployment of AI for military and surveillance purposes, leveraging its lack of concern for personal data privacy. However, its innovation ecosystem has weakened due to government policies, forcing reliance on U.S. open-source models like Meta's LLaMA. Despite this, China excels at catching up to U.S. advancements, as seen with DeepSeq's replication of OpenAI's reasoning models.
By 2025, AI agents will be actively used in military logistics, data processing, and decision-making. They will optimize complex systems, process vast amounts of battlefield data, and enhance drone autonomy. This shift will increase the lethality and effectiveness of military operations, raising concerns about the ethical implications of autonomous warfare.
By 2025, AI agents will begin handling end-to-end workflows for consumers, such as travel planning, calendaring, and personal project management. These agents will operate in the background, automating utility-based tasks and freeing users from repetitive activities. The challenge lies in creating intuitive user interfaces that move beyond the current chat-based paradigm.
By 2025, the focus in AI development will shift from solely scaling computational power to equally prioritizing data scaling. High-quality, complex data, such as multimodal and frontier data, will be essential for advancing AI capabilities. Hybrid data approaches, combining synthetic data with human expertise, will become crucial to avoid model degradation and ensure progress.
Current AI models struggle with multi-step reasoning and reliability. Future advancements will focus on improving their ability to handle complex, multi-turn tasks and eventually make autonomous hypotheses and discoveries. However, human expertise will remain essential for guiding models and ensuring accuracy, creating a symbiotic relationship between humans and AI.
Quantum computing, though still in its early stages, has the potential to significantly accelerate AI's ability to conduct scientific research in fields like biology, chemistry, and fusion. By 2025, quantum computing could enable AI to solve complex problems in natural sciences, leading to breakthroughs in areas that are currently difficult to model or understand.
Scale AI founder and CEO Alexander Wang joins us to predict where AI is heading in 2025, looking at everything from geopolitics to AI agents. That's coming up right after this.
From LinkedIn News, I'm Jessi Hempel, host of the Hello Monday podcast. Start your week with the Hello Monday podcast. We'll navigate career pivots. We'll learn where happiness fits in. Listen to Hello Monday with me, Jessi Hempel, on the LinkedIn Podcast Network or wherever you get your podcasts.
Struggling to keep up with customers? With AgentForce and Salesforce Data Cloud, deploy AI agents that know your customers and act on their own. That's because Data Cloud brings all your data to AgentForce, no matter where it lives. Get started at salesforce.com slash data.
Welcome to Big Technology Podcast, a show for cool-headed, nuanced conversation of the tech world and beyond. So thrilled about the show that we're bringing to you today because we have Alexander Wang here. He's the founder and CEO of Scale.ai, that company. It's worth $14 billion. It raised a billion dollars this year. It creates data that powers LLMs from OpenAI, Meta, and other big companies. And it also provides technical solutions to businesses and the U.S. government, which helps them build technology.
and deploy ai so alex is working with all the big companies really there in the heart of what they're doing and including you know not just companies but the us government and we're definitely going to touch on that so alex great to have you here thanks so much for coming on the show thanks for having me super excited to be chatting today
Yes. And we're going to get into plenty of your predictions. And I just want to kick off the one that I find the most interesting, which is that you see some geopolitical shifts coming up in the next year in the world of AI. Why don't you lead with that one? So I think the one of the big questions of AI has for the past decade has always been the U.S. versus China arms race. And
I think the question that's often asked is, which of the US or China is going to come out ahead on AI technology? Certainly, it's been a pretty tight race at various points over the past decade as we look at technology to technology. With autonomous vehicles, it was very close. Then now with military use cases of AI, it was very close. Then now with generative AI and large language models, it's once again quite close.
that the new admin will come in and help accelerate things to enable the US to compete more aggressively with China and ultimately come out ahead on the technology. My prediction really is that we're going to be talking a lot more about not only which of the two superpower wins, but which one has AI systems that are going to be adaptable and exportable worldwide. Which country
is going to have the AI technology that becomes sort of the infrastructure and the foundation of the world's AI systems. And there's a lot of countries that are kind of caught in the middle. Most of the globe is sort of caught in the middle between US and China.
And there's always these questions where I think both the US and China ask them, hey, you have to pick a side when it comes to which technology you're going to rely on. And so we like to call these geopolitical swing states or many, many countries which are sort of--
You know, they could go either way. They could go to Western and U.S. technologies or they could go to Chinese technologies. I think one of the best examples of this was in the past year, the Biden admin posed to the UAE, hey, which way are we?
are you going to go in terms of AI technology? You could either go into the sort of Huawei China stack, or you could go into the Microsoft United States technology stack for AI, and they ultimately pick the US stack. But I think this is going to be one of the under the line battles that really defines the course of the next few decades of geopolitics. I don't think we can really afford another Chinese expansionist
expansionary expedition like the Belt and Road Initiative or Huawei's technology being exported very broadly, we need to ensure that Western AI technology is dominant globally.
So basically what you're positing is that there's a series of AI models that U.S. companies like OpenAI, Google, Amazon, Meta are building. And then there's a series of models that Chinese companies like Huawei are building. And they're going to be in competition with each other in the globe. And it's important that the U.S. wins or the Western version wins because we also have Mistral in France. Why is that important?
There's two sides of this. I think first, there's the tactical question of, okay, which one is more powerful, US AI versus Chinese AI? And this is very relevant for national security. I mean, I think that if you believe that there's some potential of some kind of
conflict over Taiwan or other some kind of other like hot conflict between the US and China, then we really the United States needs to ensure that we have the best possible technology to ensure that we would prevail in any kind of hot conflict, that that democracy would prevail and that ultimately that we're able to sort of continue ensuring our way of life. That's sort of the better. Having the better chat GPT isn't going to make you victorious in a conflict over Taiwan.
Certainly, it will not be the only factor. But the history of war is a history of military technology. And time and time again, you know, you see when there's new technologies and new technological paradigms that come to warfare, it has the ability to fundamentally change.
shift the tides you know we saw that most recently in Ukraine with drone warfare becoming all of a sudden the major Paradigm by the way that I think that the drone warfare in Ukraine is becoming more and more enhanced by gender of AI and more advanced autonomy so that's definitely one thread that is continuing um but before you move on where would you say the US and China are in terms of
competitiveness on AI technology and especially, uh, not, not even broader, but like, especially about the way that they apply it in war. So if you look at just the raw technology, the U S is, is ahead, but China is, is, is moving, is fast following, you know, and we like to break it down across three dimensions. So AI really boils down to three pillars. It boils down to, um,
Algorithms, computational power, and data. Algorithms are the kinds that folks at OpenAI or Google or other companies build. Computational power comes down to chips and GPUs, the kind that NVIDIA
produces out of TSMC's factories or TSMC's fabs in Taiwan. And then lastly is data, which is maybe the least focused on of the three pillars, but certainly just as important for the performance of these AI systems.
If we're to rack and stack versus China, we're ahead on algorithms, we're ahead on computational power, thankfully due to a lot of the export controls that the Commerce Department has put in place. And then on data, it's a little bit of a jump ball. The conventional wisdom is that China is actually probably going to be ahead on data in the long run because they don't care as much about personal liberties and protecting personal data in the same way that we do in the West.
And so right now, the US is ahead. That being said, the sort of deployment of AI to military, you know, it's hard to track exactly. The PLA doesn't tell us exactly what they're doing. People Liberation Army out of China. They don't tell us exactly what they're up to. But I certainly am worried that they're moving faster than we are in the US. And this has been the sort of
pre-existing precedent when it comes to China's use of AI technology for national security or military use cases. The best example of this is in the past decade, they rolled out facial recognition technology widespread across the whole country for things like Uyghur suppression or global surveillance of their citizen base. They did that
incredibly quickly, much faster than any comparable technology scale up in the United States. So, my expectation is that they will actually deploy AI to their military faster than the U.S., even though the U.S. is ahead on the core technology.
Okay, so that's the military point. So basically, you're going to want the Western countries to be stronger than China. And AI makes a big difference there. So it's important for the AI industries to be stronger, because if you're not stronger than you're, there's a liability, especially as this stuff gets put into production on the battlefield with things like drones and computer vision, I guess, applied on top of satellite imagery, to figure out where people are stationed in the middle of hot conflicts. But
And there's a more subtle point, which is that it actually not only does it matter for hot conflict, for war, etc. It also matters just in terms of, OK, which technology becomes the commercially or economically speaking, the global standard. Right. And this is your second point here.
Yeah, exactly. And because in the US, you know, we benefit as a country from being the global standard in a number of areas. You know, we are the global standard for currency. That is something that's incredibly beneficial to our economy and to everything that we do. You know, certainly our our
our search. Um, uh, so Google and a lot of our technology companies are the global standards. So for search and for, um, for social media, uh, many of these are the sort of like global standards. We benefit a lot from these being the global standards. And I think when it comes to AI, you know, it's a very interesting technology because not only is it a sort of technological utility, but it's also a cultural technology. Ultimately, if you, if a lot of people within, uh,
on the globe are talking to AIs to understand what to think or how to feel about certain things, then ensuring that the AI substrate that gets exported around the world is one that is democratic in nature, that is sort of believes in the ideas of sort of free speech and sort of
you know, open conversation about whatever topic is necessary. You know, that's a, that's a really powerful cultural export that we can have from the United States that will over time, I think, fulfill a lot of America's vision of ensuring that we have, you know, freedom and liberty for all. So I think it's one of these things that is unbelievably important, even beyond the sort of hot military implications. It's one that's important just for, uh,
culturally ensuring that the United States is able to export our ideals.
So you're saying there's a soft power issue here as well? Yes, exactly. I want to ask you about China's development of AI, because I always hear two contradictory things about how China's progressing with AI. The first is that they have the government that's willing to put all the resources that they can into building the compute power to train and run models. And they don't care about data privacy, so they have all the data that they need.
Right. And then the algorithms are, you know, they're basically all published in that Google paper. You know, you can tweak them a little bit, but basically they have the algorithms. So they should be the lead. And then you look at what's actually going on on the ground, which is that and you correct me if I'm wrong. Right now, China is using a lot of American models, open source models. In fact, Meta's model, the Lama model, which is an open source model that they have developed and released, is
we know for a fact has been used in applications by the Chinese military. So explain this one to me. How has China been able to effectively, you know, put all these resources toward the problem, but still has to rely on American open source technology to build the things that they want to build? Well, there's probably two major things. I mean, one, one, you
undeniable trend over the past, let's call it five years, has been the sort of
the collapse of the Chinese startup sector. And this is really driven by policies from the CCP to significantly, you know, they killed certain startup industries. They really like hampered the entire innovation ecosystem. And you see it in the numbers, the sort of amount of capital flowing into the Chinese innovation ecosystem has fallen off cliff pretty precipitously. So why did they do that?
Before you move on, why did they do that? I know they also somewhat disappeared Jack Ma, right? Like they had Chinese tech icons that have sort of gone away. Was it that the tech industry was growing so large it threatened the government or what? It could be the possible logic there.
Yeah, I do think that was – that's the sort of fundamental risk. I mean I think that if the government, if the CCP has a desire to ensure that they consolidate all the power, either they have to nationalize the tech firms or they have to ensure that they stay weak. And so – and there were some other – yeah, there were some other –
Totally. And I think a lot of this hinges on I think they do really see the world differently from the way that we do. I think we, you know, in the West, it seems totally insane. But I think in certain doctrines or with certain ideals, I think it can make total sense, right? But-
But there is a death of the Chinese innovation ecosystem. So a lot of what they have to do in AI is just catch up and copy what we've been up to, which they have been pretty successful at. So for example, the...
OpenAI released O1 and released the O1 preview a number of months ago. This is its reasoning model. Yeah, this is OpenAI's advanced reasoning model, which is great at scientific reasoning and mathematical reasoning and reasoning in code, etc. The very first replication of that model and of that paradigm of model actually came out of China from a lab called DeepSeq, the DeepSeq R1 model. They certainly are extremely good at catching up. Now, there is a
very real hamper in a lot of their progress too, which is the chip export controls. And this has been an incredible effort, I think, from the US Department of Commerce and the Biden administration in general to sort of hamper the ability of the Chinese AI ecosystem to build foundation models of the similar size, scale, and magnitude as the ones we have in the US because they
They have not been able to get access to the cutting edge NVIDIA GPUs that we have in the States. And so, you know, whether or not you think that's good or bad policy, it has hampered the progress of Chinese AI development, which enables us to stay ahead.
So let's circle back to your prediction that you talked about how US and China will be head to head, trying to get their vision for AI adopted across the globe. So that's your prediction of like, what's going to happen? Who do you think is going to win there?
I think that the trend right now is currently very positively in the direction of the United States or of the West, broadly speaking. We have the most powerful models. We also have, I think, the most compelling value proposition in terms of our models are going to keep getting better and yes, maybe the Chinese ones catch up over time, but
We are the innovation ecosystem. We are going to be the ones who innovate far ahead of the adversaries. That being said, I think that there's-- on the flip side, you have to look at what's the total package that the CCP or China might be able to offer. In the Belt and Road Initiative, it was through this total package of technology plus infrastructure build-outs plus debt.
that sort of managed to move a lot of folks over to their side. And so, I think we need to watch it closely to make sure that we always have a compelling total value proposition. I do think, you know, one sort of sub-prediction that I have too, which is important to mention here, is that, you know, the technology is moving so quickly that I do think that
2025 will be the year where we start to see several militaries around the world start utilizing AI agents in active warfighting environments to great effect. I think you're going to start seeing this in some of the hot wars that we have going, as well as some sort of militaries
advanced militaries who aren't at war start utilizing AI agents. And so I think that the temperature, so to speak, on AI deployment to military is going to go up pretty dramatically over the course of the next year. Yeah, I just wrote a post on Big Technology about how AI is going to be an enterprise thing for a while, right? Like companies, B2B software companies, not exactly the most exciting stuff in the tech world.
Is it going to be where this stuff is adopted because it solves a problem for them where they have loads of information, they can't organize it, they can't share it, they can't act on it. And generative AI in particular is quite good at handling that. And then you think about, well, where else could this be of use if it's not going to be for regular people? Right. Like we're not we don't have an iPhone right now, but we have like plenty of companies working in software and the military is just like the perfect example of where it could apply because of all of the information and the logistics issues.
Yeah, exactly. And I think that this is, you're hitting on the core point, which I think is some, is often glossed over. I think when people think about the military and think about a war, they often think about the literal battlefield and the sort of actions on top of the battlefield. But, you know, 80% of the effort that goes into any warfighting effort or any military is all of the logistical coordination that goes into, you know, battlegrounds.
the manufacturing of weapons or the manufacturing of various supplies, the logistics and sort of delivery of all the supplies to a battlefield, the decision-making process, the sort of data processing of all the information that's coming in. And so most of what happens actually looks, to your point, a lot like an enterprise. The stakes are just dramatically higher. Yes. Yeah, military today is all about logistics. It's like the firing of the guns is like the last thing that happens, but...
Exactly. It's a logistics game. And so just to, you know, drill down a little bit on one of those sub predictions that you made. So how do AI agents help in that case? So, you know, there's there's probably two core areas where I think agents are going to have
immediate value. One is in, to reference your point on enterprises, it's in processing huge amounts of data. Right now, most militaries already have more information coming in the door than they have the ability to process. There's terabytes and terabytes of data that come in, whether it's data from the battlefield, data from their partners and allies, data from satellite networks, data from other
data collection formats, and they need to process that into insight that actually can help them make real decisions about what they should be doing differently. The first is just this huge problem of massive data ingest into real decision making. That's
And that sort of general problem set fits a lot of sub areas, whether it's in logistics or intelligence or military operation planning or whatever it might be. The second area where I see it--
having very very real impact is just in in fundamentally coordination and optimization of complex systems and this is this is really where the I think the logistics or the manufacturing cases are very clear where these are incredibly complex processes with lots and lots of moving parts and
It's hard for humans to get your hands around those processes and really optimize them effectively. Whereas AI systems can ingest far more information about the processes than otherwise, can run simulations on their own around what are various configurations that might operate better, and they can sort of self-optimize those processes to perform better. And then there's, I think, the sort of third area, which are more
sort of speculative or sci-fi, which is the use of AI agents more actively in drone autonomy or a lot of the autonomous missions that are being run right now. And I think this is an area of active experimentation for a lot of militaries. But I think if you start to see that happen, then you will have more autonomous drones that are able to be more and more lethal, more and more effective. And that
that's going to be a cat and mouse game in and of itself, a real race. That scares the shit out of me. Are you comfortable with that? I think it's no, I think, I think ultimately we're going to need to have
global conversations and global coordination around to what degree we actually want a lot of this uh a lot of AI agents to be used actually on the battlefield um that being said there are there are hot wars going on right now where uh militaries and countries are desperate and I think they'll do whatever they need to in the near term to get the uh to get the leg up yeah it's one of those things that I feel like once it leaves the station it ain't coming back and when we talk about agents it's basically like
AI applications that make decisions on their own. If we end up having that, you know, deployed in war, it's just going to, once somebody does it,
It's just everyone is going to do it. It's like the opposite of usually a stored destruction with nukes, I think, where that's like, oh, like, you know, if we do this and the world is over, whereas with like agents deciding what to bomb, where to bomb, how to attack, as long as they don't have access to nukes, it's really tough for that to go back in the barn, because if you don't use it, you're going to be destroyed.
Yeah. I think the good news is that if you take nukes as an example, what has happened with nukes is we've built incredibly advanced technology, technology that has the ability to, frankly, be world-ending, but that has actually led to more peace than without it because you have this deterrent threat of the utilization of nukes. And so, my hope certainly is that
While AI's application into military is something that is...
very concerning and potentially extremely powerful, it is the sort of same overall effect, which is to ultimately deter more conflict than create it. I hope you're right and I'm wrong. And we did have Palmer Luckey on the show a couple months ago, and he talked about countries don't start wars that they believe they're going to lose. And so maybe that adds to that. I mean, that's certainly been the case with nuclear. All right. I want to get into your second prediction. We already have brought up AI agents, but I think we should go a little bit deeper because, you know,
I think people hear about AI agents and they say, is that supposed to be something on my computer that's going to like book me travel, book me tables at restaurants, look things up for me, do my expense reports if I need them to do that. Or, you know, basically agents that act on behalf of the individual. We haven't really seen those yet. We've seen some examples of companies and militaries using these things. And the average person doesn't get a chance to touch that. But you think it's going to change?
Yeah, I do think that 20, yeah, I think that 2020, uh, 2025 is really going to be the year where we start to see some, uh, kind of very basic primordial AI agents really start working in the, in the consumer realm and creating, uh,
sort of real consumer adoption. Another way that I think about this is we'll see something like a chat GPT moment in 2025 for AI agents, which is you'll see a product that starts resonating, even though to technologists it may not seem
like all that or may not seem like that big of a leap relative to what we had before. I think a lot of that is going to come from probably two main threads. First, obviously, the model is continuing to improve and getting more reliable and getting down that curve. The second is really evolving in the UI and experience of what an agent does. Right now, we're so stuck as a
I think tech industry still on the sort of like chat paradigm and, you know, having everything be a chat with one of these models. And I think that's a constrictive paradigm to enable agents to actually really start working. And to me, what it really means for an agent to start working is, you know,
me as a user or consumers in general start actually outsourcing some real workflows to the agent that they would have had to do otherwise. And so we'll start to just sort of like fully trust the agent to do full end-to-end workflows. You know, maybe it'll be something around travel. Maybe it'll be something around calendaring. Maybe it'll be something even around just like, you know,
producing presentations or managing your workflow. But we'll start to really offload some of the meaningful chunks of our work to the agents. And there will be something that really starts to take off. I don't know if it's going to be one of the big labs or it'll be a new startup that comes up with it because I think so much of it will come from kind of like
experimenting and the natural innovation ecosystem working out. But what we see is that the models and their capabilities are certainly strong enough to enable a pretty incredible experience. There's all this talk about whether or not we're hitting a wall or whatnot, but the models are really, really powerful and we should see something big here.
Okay, so just walk me through like what that experience might look like. You know, we don't have to stick with this, like it doesn't have to necessarily be the use case. But since you've imagined the idea that AI agents could end up helping us in 2025, like what are some experiences that are in the realm of feasible for someone?
First, let's walk through what's an ideal AI agent. An ideal AI agent is one that I think is observing and naturally in all the core flows of information and core flows of context,
that you are in digitally. So it's in all your Slack threads, it's in all your email threads, it reads your JIRA or all of your tools to understand everything that's going on in your work life. And then it helps to sort of organize all that information to start taking certain actions. And so like one agent that I think
would be super beneficial and one that I think is in the realm of feasible is something that starts to take a hand at responding to a lot of your emails, flagging when it needs you for like
additional context or information to be able to address your emails, can sort of summarize a lot of your emails for you naturally. And so something that just turns the experience of doing email from, hey, I'm like having to respond piece by piece to every single email to leveling you up to being, hey, this is like
all of the overall work streams and workflows and how do you want to engage at a high level on top of those workflows but this is this is a business use case and I'm curious if you think that like how everyday people might end up using AI agents or is that just still a ways off like maybe not in 2025 uh everyone works you know so uh give me an example outside of the work context
Yeah, I think one that's more personal. I mean, I think similarly, I think in everyone's personal lives, you're also...
juggling and navigating a whole set of various priorities. I'm planning a trip with my friends over here and I need to get gifts for my family and figure out what they want for Christmas. And then I need to, I have all of these sort of personal projects, which are still sort of like sitting there. And so I think in the same way, helping you sort of like level up on
on top of all of the projects that you're navigating and sort of like help you sort of coordinate between all of them more naturally. I think that's something that we're going to start seeing. Now, I don't know the perfect way that that happens, right? I don't, I think that the product experience is so important.
so important as a part of this and having a product experience which where you don't expect it to be perfect but you expect it to be pretty good I think that's like 99% of the challenge and that's why we haven't seen it yet despite the fact that the models already can do a lot of this stuff pretty well
My 2025 prediction is that guys use AI agents to use dating apps for them. And some get found out and some don't. And we're going to see some stories about how like some guy like set it on autopilot and ended up, you know, lining up more dates than he could ever hope for. Yeah. Yeah. Well, hopefully that's already happening.
Hopefully there'll be good dates. Yeah, I don't know. What are you seeing? I know you had Benioff on the podcast a little bit ago. What are you seeing as the things that seem to make sense from an AI agent's perspective? Well, I think that Mark Benioff, the Salesforce CEO, when he came on, talked pretty convincingly that we'll have...
AI agents at work. And again, this is like the work or the enterprise use case, because work has all this data. And there are all these tasks that we do all throughout the day at work that are just arduous and really quite, you know, quite annoying, preparing reports, making dashboards, going to meetings we don't need to be in, pulling out highlights from those meetings, sending them to our bosses, telling our bosses, you know, in the Salesforce instance, for instance, like, you
how each conversation went and what our expected pipeline is to close that corridor and all this stuff can be used for AI. I think it can be used with AI. I think it's really interesting in the medical use case. I was just speaking with GE Healthcare about how they've now put in dashboards for doctors sort of summaries of cancer patients' medical histories, which run thousands of pages.
And the doctors never had a chance to read the whole history. And now the generative AI is summarizing it and going out and finding available treatments for them and notifying them when they miss tests. And I think this is also an example that Benioff gave about the healthcare example where that can actually be proactive in solving
scaling medical advice and medical treatment in a way that you'd never hear from like your doctor after you showed up to an appointment. And now can they create an agent that just kind of keeps you on your plan, you know, in terms of like follow-up stuff that you need to do. On the consumer side, like for everybody else, that's kind of where I wonder because all of our internet has been designed to effectively combat bots. But if we have agents that work on our behalf,
on the internet, like travel sites, dating sites, social media sites. I'm very curious, like whether they're going to come up against these bot protection systems. Like, are they going to do CAPTCHAs on our behalf? Are they going to get the text messages and fill in those numbers so they're able to log into different systems? Because again, the whole internet has been built to defend against these things. So I'm curious what you think. I mean,
Is this vision of, you know, personal agents that act on our behalf to do things like book travel, keep up with our health, take action on Internet services for us? Is it even a feasible thing to do, given all of the protections to sort of guard against them up until this moment? We will have to sort of fundamentally reformat how the Internet works to be able to support it. And I think that like.
We're going to need, in some senses, there will be two webs. There will be the web that humans use when they need to navigate stuff on their own, and then there will be the web that agents use, which is sort of
under the surface and something that humans will never see, but allows them to sort of conduct actions on our behalf more efficiently and easily. And that I think will be in the long run what ends up happening. And my honest take is I think that to the degree that most of us, there's sort of like two kinds of usages of the internet today. There's sort of a
consumption, which is where we're seeking out content and we're curious about things. And then there's utility-based usage. And I think the sort of addressable market, so to speak, for the agents is all the utility work. Everything where I'm using the internet just to get something done
I want that to happen faster, easier, better. I would rather have to not have to do that actively at all. Let's say it's like booking an appointment and looking up a particular piece of information or figuring out how to fill out my tax return or whatever it might be. That stuff should all be handled by the agents. We're still going to have to do a lot of consumption of content just as part of what we like to do.
Yeah, I think it's a really good point. I mean, I think ultimately, I think agents are going to start
in an area that they'll feel pretty, it'll feel like a toy, just like with any technology. So maybe, you know, we'll all start with like a language learning agent or we'll start with a cooking aid agent or it'll just be something that feels pretty innocuous, but then we'll start to realize we can really rely on it. And then,
and also relying on it for a lot more. And that's kind of what happened, I think, with ChatGPT. Initially, it was sort of, we realized, you know, it was kind of a toy. And then people started doing a lot of homework with it. People started to code with it. And then now people do all sorts of stuff with ChatGPT and other chatbots. That'll be the thread. Let me ask you this question before we move off of agents. Do you think it's ethical for me to like have my AI agent, which can type and talk,
go out and email and call a bunch of humans on our behalf, people working, you know, let's say in customer service or I don't know if I'm applying to schools and they're trying to find out like information about like whether I qualify and what I need to submit. I mean, these processes, maybe they've been designed as arduous to sort of filter out the people who aren't willing to do the work to sort of get in or pass that application threshold.
So it's in some way it's combating these guardrails that companies and institutions have set up for us. On the other hand, it could end up wasting a lot of people's time. Like I really am anticipating like no agent policies from like certain schools or institutions being like, if you're going to reach out to us, it has to be a person versus an agent. What do you think?
You know, I saw this thing on Reddit. There was this post of how an
An admissions officer, she sort of created all these ways in which they could track whether or not an essay was AI-generated or not. And they were very detailed things. It was very specific. They were listed maybe 20 or so criteria that they looked for. And I think that, to your point, it was kind of heartbreaking to see because that means that let's
If a student used an AI to generate an essay, they have to spend way more time just figuring out whether or not it was AI generated to sift through all the noise. I think you're totally right. I think we're going to need...
Almost in the same way that there will be like an internet for humans and an internet for agents there will be processes for humans processes for agents and and a lot of a lot of things that are high intent or very expensive or Otherwise special in some way are going to be reserved for humans only and and it'll sort of be the the sort of like more transactional stuff that that can be handed off to agents in in mass
That's right. I mean, in some ways I'm looking forward to this future. On the other hand, I do sort of think like the more we talk about it, how much AI will take care of for us. I do sort of feel like we're cannonballing our way towards that WALL-E future where we're all fat and drinking big sodas and having Roombas take us around the world. It's, yeah, I think, I think, yeah.
ease and convenience, which definitely are the directions that technology has taken us. You know, clearly there should be limits at some point, but but if we if they exist, we don't know where they are. Exactly. And this idea of like removing friction in some ways, it's made the world great. In other ways, it sort of changes the brain chemistry of people where like we don't expect to go through hard things and
And when we do, we lose our minds. And that's why you end up seeing the YouTube videos and the videos on acts of people in the airport, because we've removed so much friction and companies have competed on the base of customer experience to the point where now if something goes wrong, we're fragile. And we think that, you know, we deserve better. And there is something to be said for friction toughens people up a little bit. Totally. Yeah.
All right, we're here with Alexander Wing, CEO and co-founder of Scale AI, $14 billion company that works with others to help generate AI data for them and also help them scale their AI solutions. We're going to talk a little bit more about Alex's third prediction when we come back right after this. I'm Jesse Hempel, host of Hello Monday. In my 20s, I knew what I wanted for my career.
But from where I am now, in the middle of my life, nothing feels as certain. Work's changing. We're changing. And there's no guidebook for how to make sense of any of it. So every Monday, I bring you conversations with people who are thinking deeply about work and where it fits into our lives. We talk about making career pivots, about purpose and how to discern it, about where happiness fits into the mix, and how to ask for more money.
Come join us in the Hello Monday community. Let's figure out the future together. Listen to Hello Monday with Jesse Hempel wherever you get your podcasts.
Struggling to meet the increasing demands of your customers? With AgentForce and Salesforce Data Cloud, you can deploy AI agents that free up your team's time to focus more on building customer relationships and less on repetitive, low-value tasks. That's because Data Cloud brings all your customer data to AgentForce, no matter where it lives, resulting in agents that deeply understand your customer and act without assistance. This is what AI was meant to be. Get started at salesforce.com slash data.
And we're back here on Big Technology Podcast with Alexander Wang, the CEO and co-founder of Scale AI. So, Alex, I want to ask you about this interesting shift that we're seeing, right? So up until this point, we've talked entirely about AI models on the basis of how many GPUs or chips they're trained on, right? It used to be that you could train a model on like 16 chips, right? By the way, they're not cheap, like $20,000 to $40,000 each, right?
Then I went to a thousand and now towards the end of the year, we started hearing crazy numbers like a hundred thousand, two hundred thousand. I was just at Amazon's re:Invent conference in Vegas and Matt Garman, the CEO of AWS, told me that they're going to train the next anthropic model on hundreds of thousands of GPUs, GPUs or GPU equivalents. And then I was like, oh, that's a lot. And as he's saying that, Elon Musk came out and was like, well,
We are going to train the next XAI model in Memphis on a million GPUs. So I think we're really hitting, like, maybe we're hitting the limit, I don't know, of what you can do with chips. And so you believe that we're going to shift this conversation beyond chips in terms of what makes the most powerful model. So I will tee you up for prediction number three.
Yeah, and so much of the dialogue, to your point, over the past few years has really been around GPUs and computational power. And I think what's going to happen in 2025 is we're going to-- we're going to only be focused on who can create newer, better chips or bigger data centers with more chips, but also who can create newer and better data.
And one of the things that I think we're going to see is a focus of the focus shift from just computational power to computational power plus data being sort of considered nearly equally. You know, data really is...
at its core, the raw material for intelligence. The conversations around data are going to be really interesting. One of the big topics that's been bounced around for the past few months has been, are we hitting a wall? Have we hit the data wall?
Are we hitting a wall on progress overall? And I think the interesting thing that's been happening is, you know, this has come from an approach of scale up computational power at all costs. If we just scale up the number of GPUs and create huge, bigger and bigger GPUs,
You know data centers of GPUs without creating more and more data to train these models on then we're gonna hit issues and we're gonna hit walls and barriers where we we stop seeing the level of progress that we expect out of the models, so One of the big things that we see especially in our work with a lot of the frontier labs is you know It is true. They're scaling up the GPU clusters. They're scaling up the number of chips
that's still a very aggressive path for them. But the in parallel conversation is, how do we scale up data? And there's two sides of that. One is obviously scaling up the volumes, but also scaling up the complexity. So they're seeing the need to go towards more of,
what we call frontier data. So go towards advanced reasoning capabilities, agentic data to support the agents that we were just talking about, advanced multimodal data. We just saw today, for example, that OpenAI released Sora. And so the needs for video data and more complex combinations of video, text, audio, imagery, et cetera, all together is going to be really, really interesting going into the next year. And so...
I think one of the lessons that's really played out more recently with the models is that you can't just scale GPUs and expect to get the same levels of progress. You need to have a strategy by which you're going to scale up
all three of the pillars. You need a strategy to scale up the computing, you need a strategy to scale up data, you need a strategy to continue improving the models. And it's only through the sort of concert of all three of those things that you're going to be able to keep pushing the boundaries and barriers on AI progress.
But I'm curious what you think. I mean, you've talked to all these CEOs. What are they talking about? I mean, this is exactly the thing that they're talking about. We had Aiden Gomez from Cohere in a couple of weeks ago, and he basically said that this has sort of been the path of training the models, whereas in the early days, you could effectively bring anybody off the street, take down anything they had to say, and it would be new information for the models.
And then you started to have to bring in grad students to talk about their... Because that general knowledge base was built. So then you bring in grad students to talk about their area of discipline. Then you go to the PhDs. And then he goes, where do we go next? Because we have all this general knowledge. And now we have all the specialized knowledge that we've used to train these models on. And by the way, it's just amazing the way that they've improved and been able to sort of handle some complexity. It's really crazy. And so the question is like, where to go next? And I...
I think that's what you guys are working on now. And I'd be curious to hear what the process is like on your end for generating more data for these models to train on. Yeah, so it's exactly what you just mentioned. Like a lot of what we're focused on is how do we bring in expertise and really this sort of expertise from every field you might imagine, from medicine to law to science
math to physics to computer science to even knowing about really advanced systems of various kinds or being a great accountant or whatever field you might imagine, getting the sort of
What is all the arcane knowledge? What is all of the really specific, deep knowledge that exists in each of these areas and pull that into large-scale data sets that we can use to help train these models to keep improving in a lot of these areas? A lot of the effort for us has been something that we call hybrid data. How do we
So one of the things that we've seen over the past year in particular is that synthetic data has not worked as well as I think everybody had hoped. Pure synthetic data, just using
generate from the models, try to train future models that can sometimes cause real issues for the models. Um, and so one of the things that we've been really pushing forward is this idea of hybrid data. So you have synthetic data, but you use, um, human experts to mix in with the synthetic data to ensure that you're producing data. That's really, really, um,
accurate and high quality and it won't cause issues, but also you're able to do it very efficiently and at large scale. So you also have those PhDs that will sit down and kind of write what they know or dictate what they know. And then you feed that into the models. Yeah, exactly. And a lot of times it's even more targeted than that. You know, you run the model until you realize the model's making mistakes over and over again. And then you know you've hit sort of a limit of its knowledge or limit of its capability and you have a PhD...
sort of come in and help, you know, set the model up on the right track, so to speak. What's the limit then in terms of where are we going to get to? Because if we, let's say we have all these specialized fields input their knowledge, does that eventually make like AI complete if it just kind of knows everything about every subject or does it have to hit like a new benchmark to really show that it has this like next level intelligence? Like does it have to start making discoveries of its own? What do you think the benchmark should be?
Yeah, I think, well...
To me, I think there's clearly many more levels of improvement. Now, it's testing, okay, can it do each of these things right once or how-- the first track was just reliability. Getting these models from doing something right once in five times to right 99.99% of the time. That requires a lot of development just to get to that increased level of reliability of the systems.
And then to your point, it's really about how can the model start taking more and more actions in a row? You know, one of the things that really, um,
is true in all the models today is that they're not that good at taking multi-step actions. Whenever it has to take a few hops, whenever it has to chain a few things together, it'll invariably make mistakes along the way. And so the next level of improving reliability is really enabling the models to do more and more multi-turn, more and more multi-step reasoning to be able to enable them to sort of
do more and more complex tasks. And then the last piece as we go is, and this is the key to where you're going, is like, eventually, it'll be able to start making its own hypotheses, running those tests on its own, and sort of ultimately making its own sort of
discoveries or realizations or conduct its own research. Even then, it's still going to get stuck sometimes and still going to need a human PhD to come in and help it just in the same way that a PhD student these days still needs an advisor to still give it the right nudge.
And so I don't think the sort of like the symbiosis, so to speak, between the humans and the AI will ever go away. Like I think we'll always be able to sort of
will always be very important in helping the models get on the right track and ensure that they always are continuing to improve. But we're going to see the models sort of level up in terms of what is the degree to which they're able to be autonomous and the degree to which they're able to operate on their own. And on the multi-step thing, right, taking a bunch of different steps, I heard something interesting from Moody's last week, and I want to run it by you, where they said basically,
They've created 35 individual agents. So let's say they want to evaluate something for their portfolio, like a company for their portfolio. They'll have one that will look at one agent will look at the financial data. Another agent will look at the, let's say, weather risks. Another agent will look at the location that they're based in. Another one will look at the industry.
And they have 35 different variables or whatever it is. And then they have all of them come back and they deliver their results to this compiler agent, which evaluates all of it and then runs the results by voting agents, which ask, OK, is this reliable or not?
I walked away from that impressed by the idea, but also like kind of my reporter brain went off and was like, I don't know if this is real or not. So I'm curious what you think. Is that a possible solution? And how feasible is that in terms of a way to get into these multi-step processes?
So that's a very – in my work, it's a very sort of regimented way to try to enable the system to do multi-step reasoning. Because ideally, what you want the model to do is to – just like how a human does, be able to sort of go through and figure out what are the bits and pieces it needs to know as it goes along and be able to do so on its own dynamically without having to sort of like –
pre predetermined and preset this entire regimen for the models to need to go through. Yeah. Um, so then you're saying that might be something that a model can do entirely on its own. That's pretty cool. I think in the future, like we're, we're gonna, the models will improve to be able to get there, you know? And then I think the real, um, on the multi-step side and the multi-step reasoning point, I do think that the, um, there's a lot of blockers because, um,
This is the kind of thing that humans learn how to do kind of from a lot of trial and error and experimentation. Like we'll try to do a complex task and then we'll realize we'll learn that, oh, we actually missed. You know, let's say you try to bake a cake for the first time, you know, a reasonably complex endeavor. And then you realize you missed A, B, C and D. And then the next time around, you'll be like, OK, I'm definitely going to remember. I just had a pan of flour that came out of the oven.
Where did I go wrong? But yeah, exactly. I mean, we learn a lot through trial and error. And right now, the models are early in their process of doing the same thing, of going through and being able to do these sort of dynamic processes where they learn through trial and error and they are able to continually learn from their mistakes. That's where we need to get to.
Okay, great. I know we have just a couple of minutes left. So let me throw a couple quick hits at you and then we can head out. First of all,
I'm just curious. We talked a lot about how data is going to matter a lot, but I can't get my mind off the fact that Elon's going to try to build this million GPU super cluster. What's your prediction for what that spits out? I honestly think right now at where we are in AI development today, we are more bottlenecked by data than we are compute. So I will just have an incremental improvement then with something like that.
Yeah, I think that I think the real step changes come from data. Okay. So just a quick follow up to that. If we if we end up like I just saw there was news from Google today about this breakthrough they had in quantum computing, which we'll probably cover more on the Friday show. If we have working quantum computers, which can process data much faster, what do you think that does for AI? I really think so. I had the opportunity to tour Google's quantum computer.
facility earlier this year. It's very impressive. I think quantum computing is on kind of like the way AI was back in 2018. It's on a few scaling laws where you can definitely sort of squint and see that in five to 10 years, this is going to be a really, really impactful technology. And ultimately, I think what it's going to enable is it's going to speed up AI's ability to
do scientific discovery. And so whether it's, you know, I think a lot of the use cases that excite people are in biology or chemistry or fusion or a lot of these very chaotic and difficult to understand, you know, natural sciences. I think that's where quantum computing has the ability to be pretty transformational fundamentally. And I think AI will be able to use it as a tool to be able to enable it to do incredible research in those fields. That's great.
That's crazy. Okay, so all right, last one for you. We're in the middle of like this race where it seems like every week, the foundational model companies put out a new development, whether that's open AI, or that's anthropic or even XAI, Google, Amazon just released a set of new models last week. So who do you think is in the lead at the end of 2025?
Ooh, that's hard to say. I mean, I think that one thing that we see today with the models is that because all the benchmarks that were used today are...
what's called saturated, i.e., in other words, all the models do really well at the benchmarks. It's really hard to discern actually which models are fully on top versus not on top. There's a lot of argument, for example, on the internet, at least in the Twitter feeds that I see, in terms of whether
whether Cloud is better or O1 is better, and there's all these comparisons between the two of them. One of the things that I think we're going to need in 2025 are much, much harder benchmarks and much, much harder evaluations that are going to be able to help us
figure out, you know, separate the wheat from the chaff a little bit. I don't know who's going to be in the lead, but I do think that we need much better measurement to actually be able to discern between all of these incredible models that labs are pushing out right now.
Okay. All right. We'll take it. No, no prediction on who's going to be the best, but a definite interesting perspective on evaluations. Alex, great to meet you. Thank you for coming on the show. I think these predictions have been fascinating. Definitely stretched my mind in areas that I wasn't thinking about. So thank you. And we hope to have you back sometime soon. Yeah, this was a lot of fun. Thanks for having me. Thanks for being here. All right, everybody. Thank you so much for listening. We'll be back on Friday with Ranjan breaking down the news. We will see you then on Big Technology Podcast.