We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Copilot, Agent Mode, and the New World of Dev Tools with GitHub’s CEO Thomas Dohmke

Copilot, Agent Mode, and the New World of Dev Tools with GitHub’s CEO Thomas Dohmke

2025/3/13
logo of podcast No Priors: Artificial Intelligence | Technology | Startups

No Priors: Artificial Intelligence | Technology | Startups

AI Deep Dive AI Chapters Transcript
People
T
Thomas Dohmke
Topics
Thomas Dohmke: 我最近发布了 Copilot 的 Agent 模式,它允许开发者与 AI 协同工作,完成更复杂的任务,例如实现新功能或安装软件包。开发者仍然是掌控全局的人,Agent 只是辅助工具。未来,Copilot 将进化成团队成员,能够自主完成任务,例如创建和提交代码。但这需要模型在推理能力上的提升以及用户界面的改进。目前,Copilot 的发展受限于模型的推理能力和用户界面的设计。Copilot Agent 的成功取决于其可预测性、可控性、可验证性和容忍度。 AI 代理的根本挑战在于将大型构思分解成可执行的小任务,这与人类开发者面临的挑战相同。AI 代理在处理明确定义的任务(例如修复bug、添加功能)方面将取得显著进展,但在处理模糊或复杂的任务方面仍有局限性。Copilot 的开发周期包含 AI 工程、模型评估、A/B 测试等多个环节,并根据市场变化不断调整。赢得开发者市场关键在于关注开发者需求,持续改进产品,并应对快速变化的市场竞争。Copilot 的使用数据表明,AI 正在生成越来越多的代码,其效率远超预期。 GitHub 未来将重点关注代码审查、云端开发环境以及 AI 驱动的安全漏洞修复等领域。AI 生成代码的普及将改变软件测试、技术债务管理等方面的工作方式,但不会完全取代人类开发者。AI 将改变软件开发团队的工作方式,尤其是在需求定义和任务分解方面。未来软件开发工具生态系统将呈现多样化,开发者将根据自身需求选择不同的工具和平台。Copilot 的商业成功体现在其广泛的应用范围和显著的生产力提升。AI 代理的定价模式将基于计算资源消耗,而不是简单地替代人类开发者的成本。开源模型将推动 AI 领域的创新,并促进 AI 技术的普及。 Sarah: (问题和讨论) Elad: (问题和讨论)

Deep Dive

Shownotes Transcript

Translations:
中文

Hi, listeners, and welcome back to No Priors. Today, we're joined by Thomas Dunk, the CEO of GitHub, a platform used by over 150 million developers worldwide to collaborate and build software. As CEO, Thomas has overseen the development of tools like GitHub Copilot. Before becoming CEO, he helped shape GitHub's product strategy and powered global expansion and previously worked at Microsoft. In this episode, we'll talk about the future of software development, the role of AI encoding, open source, and product plans for Copilot.

Thomas, welcome to KnowPriors. Maybe we can start with the meat of it. What is happening with Copilot and the new releases at GitHub recently? We're heading straight into it. We're really excited about making Copilot more agentic

A few days ago, we announced the agent mode in Copilot and VS Code. So instead of just chatting with Copilot and getting responses and then copy and pasting the code into the editor or using autocompletion, the original Copilot feature, you can now work with an agent and it helps you to implement a feature. And when it needs to install a package, it shows you the command line, terminal, commando, and you can just say, okay, run this. You're still in charge, right? So that's the crucial part of Copilot.

these agents that we have available today that as the human you still as a human developer you still need to be in the loop but we also showed you know a teaser of what's about to come in 2025 we call this a project paravan you know because it's like a jedi in a paravan you gotta have patience and you gotta you know learn how to use the force but we think you know in 2025 we get into a place where you can assign a github issue a well-defined github issue to a co-pilot and then it

starts creating a draft pull request and it outlines the plan and then it works through its plan. And you can, similar to how you observe a coworker, you can see how it commits changes into the pull request and you can review this and provide feedback to Copilot. And so Copilot basically graduates from a pair programmer to a peer programmer that becomes a member of your team. The obstacles to that right now are some new model advancements,

Is it just building out some other core technology? Is it just the UI? What is keeping that from happening right now? Yeah, I think the first thing is the model, the full O3 model that's not available yet, but OpenAI showed as part of the ship miss right before the holidays.

we're going to see, you know, improved reasoning. And I think it's as the models get better in reasoning, we're going to get closer to 100% of this vBench, which is that benchmark out of 12 repos, open source Python repos, a team in Princeton identified 2,200 or so issue-pull request pairs. Effectively, all the models and agents are measured against. So that's number one, you know, the model and the agent combination. I think the second piece is just figuring out what's the right user interface flow, right?

If you think about the workflow of a developer, right, you have an issue that somebody else filed for you, you know, user or product manager, or something that you filed yourself. Now, how do you know whether you should assign Copilot to this, the agent to it, or whether you need to refine the issue to be more specific, right? It's complicated.

crucial that the agent is predictable, that you know that this is a task that the agent can solve. If not, then you need to steer it. So, steerability is the next thing you need to either extend the definition or the agent needs to come back to you and ask you additional questions. And then at the end of the process, you want to verify the outcome. And so, in our demo, that's where we are thinking the right flow here is actually that the agent works in a pull request, like similar to a human developer with lots of commits, and then you can roll back those commits or

or check them out in your VS Code. We saw that with some of the agents that are available is that do I, as a developer, actually tolerate the agent? Like, is it actually saving my time or is it wasting my time? And the more often you see it wasting your time and just burning compute cycles, the less likely you're going to use it again. And so if you're predictable, steerable, verifiable and tolerable, if you get to that, if

for all four criterias to a certain level, I think we're going to see a wide adoption of agents. How far away do you think these agents are from being

Sort of the median program or equivalent. And then how much longer do you think it takes to get to sort of superhuman? You know, I thought about this this morning, right? Like what if regardless of what agent you're thinking of a travel agent or a coding agent, or maybe it's an agent that designs your house, the fundamental challenge is actually the same as you have as a human being.

developer, right? Like you have this big idea in your head and you can sketch it on a whiteboard, but then you want to start coding and you have to take this big idea and break it down into small chunks of work. It's

I think that's the part where we're far away from agents actually being good enough to take a very rough idea and break it down to small pieces without you as developer or as architect or even when planning your travel, constantly getting questions back of what decisions you want to make, what database for cloud. Imagine you give the agent a task saying, build GitHub or

build a mobile app of something like it will just be not specific enough, right? So that's the systems thinking that I think the media and developer will not

be replaced by agent. And the flip side of that is a lot of, you know, what developers do is just picking up issues and fixing bugs and finding where to fix the bugs, adding a feature that comes from a customer, and then you have to navigate the code base and figure out what files you have to modify. And I think there we are going to see dramatic progress over the year. We actually, you know, when we recorded the demo for the Padawan project, we actually had one of our product managers use an issue and the agent create the progress themselves, right? And so,

a PM that usually doesn't code and doesn't write code in the code base was able to use the agent to create a real pull request that was then reviewed by the developer and merged into the code base. So in some ways, we're already there. In other ways, we need to get to the point where you trust it enough that you're using it day in, day out.

I'm sure you guys were doing a bunch of dogfooding before releasing Agent Mode and Hadawan as well. Maybe if we just sort of zoom out from that from the eval phase, like, can you describe what the overall like development cycle is for Copilot today? Like how you do planning and make decisions about what to try?

and how you improve it? The industry calls now AI engineering, which is, you know, we're extending the full stack of backend and frontend development with AI development. And so how do we use a new version of a model, a new model, as we have now the Model Picker and Copilot, we're constantly dealing with multiple models from multiple vendors. How do we integrate that into our stack? We have, you know, an applied science team that runs evaluations and

We have a team that builds out these benchmarks that the applied science team uses to compare models with each other, but also the teams that build features like code review agents or the SWE agents or agent mode users to validate their work as part of their test suite.

So it's no longer just the data scientist and the engineer. Those roles have more and more overlap, and they're collaborating day in, day out. We do a lot of experimentation with A/B testing where we flight new versions or new fine-tuned versions of a model after the offline test and an online test, first with GitHub and Microsoft employees and then with sets of the population.

And then overall, obviously we have a roadmap of features that we want to build in a long backlog, not just for Copilot, but all up for GitHub. GitHub is turning 18 this year, I think. It's 18 years since the founders in late 2007 started working on it and then it launched in early 2008. And Microsoft turns 50 actually April 4th.

And so we have a long backlog of customer feedback where we're using Copilot to build those features in agent mode now to accelerate our feature delivery. But at the same time, the market is moving so fast. And whether we're meeting with OpenAI or with Entropic or with Google, we learn about new model versions and then our roadmap changes from one day to another.

I'm sure you guys are seeing that as well. The market is moving so fast. We're literally sitting on an exponential curve of innovation that is hard to keep up and you can't really plan more than a month or two ahead of time. Why do you think about competition being on that exponential curve? I think it is wild to think that SWE agents, as you describe them, didn't exist as an idea. A year ago, we now have

a market full of folks experimenting with these products. How do you think about winning the, like GitHub is obviously a very dominant force overall as its co-pilot, but how do you think about, you know, winning the developer over and what they care about in that, you know, changing and competitive market?

The way we think about winning is that, you know, we care deeply about developers. And that's always been, you know, the heart of GitHub is that we put developers first and that we are developers that are building products for developers. You know, we have the saying at GitHub is that we're building GitHub with GitHub on GitHub using GitHub. Right.

And so everything that we do in the company, including our legal terms and our HR policies and product management, sales, sales enablement, all these functions are in GitHub issues and GitHub discussions and GitHub repos. So I think that's number one that we focus.

deeply care about our own product and we're using it for everything day in, day out. You know, the first thing I do in the morning is open the GitHub app on my mobile phone and then Slack as a lot of our, you know, operations company chat runs through Slack. Number two is, you know, you mentioned competition. I mean, it's obviously like I've never seen anything like that in the developer space. It's the most exciting time, I think, for developer tools. And, you know, I've been a developer for

over 30 years, it's amazing to see the innovation, you know, the news that is coming out every day. And I think that energy, you know, that is in the market and innovation driven, both on the open source side and on the closed source side, right? Let's not forget, you know, that it's not,

one-sided. As much as there's innovation on proprietary models and software, there is an equal amount of innovation in open source and on GitHub. And that energy obviously gravitates into us. I'm a big Formula One fan. It's good when there's competition because the races are so much more fun to watch if there's multiple teams that can win the championship. And I think the same we feel about the competition. It gives us

motivation every single day when we wake up to do better, to move faster and to ultimately win with the best product in the market. You have such like rich data about how people are actually using Copilot. What is surprising you even from the last week or so since Agent Mode was released? The thing that always surprised us from the early days was how much code Copilot is writing. You know, some of the folks from Microsoft and GitHub on your podcast in the past and

in the early days, you know, soon after we launched Copilot Preview, it already wrote like 25% of the code. And I remember that meeting where we looked at this in the product review and I said, that must be a mistake in the telemetry. Go back and validate that. It can't be true that it's writing 25% of the code because it was just auto-completion and, you know, as

as cool as that was at the same time, you know, it still made a lot of mistakes in the early days. But it quickly dawned on us that A, the number is true and B, that's just the learned behavior of software developers, right? Like you're typing something, you're always reaching the point where you need to look something up. So you go to your browser and you find code on Stackable Flow or on Reddit or blogs or on GitHub, and then you copy and pasting that and you're modifying it anyway afterwards, right? Like that's,

The inner loop is always this kind of like you write something, you try it out with the compiler and the debugger, and then you keep modifying until you make it work. And that number, you know, then quickly rose to around 50%, depending on the programming language.

If you look now with these agents, it's hard to measure that because when you can literally go into agent mode and say, I want to build a snake game in Python and it writes all the code for you. It writes multiple files so the denominator becomes zero. It's like infinite percentage because the only thing you wrote was a prompt and the 15 minute demo from two years ago is a one minute demo now and

I think that's, you know, still surprising in many ways that we are already so far ahead on that curve. And then the opposite is also true, right? You can get it into a place where, you know, it just keeps rewriting the same file or deletes the whole file because it gets stuck somehow in the logic. And so it grounds us also in the reality is like we're not close to

an agent just autonomously parsing through all my GitHub issues and fixing all my backlog for me. The only thing I'm really doing is just validating and becoming the code review human for the software development agent, right? So we are in this, you know, we're swinging between the excitement of how much it can already do and the reality where it gets stuck in very simple scenarios where it's like you're trying to kind of like figure out the prompt of telling it,

just do the one thing and then you just go into the file and change the whatever the background color yourself. That makes sense. Outside of a lot of the agentic efforts that you all are doing, and obviously I think that's amongst the most interesting stuff that's happening right now. What are other big areas you want GitHub to evolve over the coming few quarters? Like what are there other big thrusts or is it all kind of it's all in on AI and that should be the focus of the company?

So far, we only talked about the generic SPI agent where you can assign an issue and it generates the pull request. But if you actually look in the developer life, the day-to-day in most companies, that's maybe two or three hours of your day that you're actually writing code

and then you're spending an equal amount of time of reviewing code of your coworkers. And while we don't believe that goes away from a pure security and trust perspective, you always want to have another human in the loop before you merge code into production. At the same time, we believe code review agents and code review is a big topic where AI can help you, especially when you work with a distributed team in different time zones, where you don't want to wait for the folks on the West Coast

to wake up, to get an initial loop of feedback. So I think code review is a big topic for us. And again, you know, the AI part is one piece to that, but the user interface is equally important. Like, right, like if ideally you get feedback and then you can work with the code review agent on that feedback to loop because you know, won't always get exactly the right feedback to just click, accept, accept, accept. You have to have a user interface, you know, cloud environment where you can just open this. If you always have to, you know,

clone the repo on your local machine and install all the dependencies, switch to a different branch, you're still, you know, having way too much boilerplate work, right? So moving to a cloud environment where you can just, you know, try out the changes that came from code review and can modify them to make them work and have that, you know, fast

outer loop in that same realm of security vulnerabilities, which is A, you want your code scanning not only find vulnerabilities, but also fix them. An even simpler version of that is linter errors, like code formatting and those kind of things hopefully all go away and just the AI fixes all that instead of you going through 100 linter warnings telling you where to put the spaces in the parentheses.

But also, if you look in any decent-sized software project, it has outer dependencies. It has lots of known software vulnerabilities, hopefully non-high risk and a lot of them low risk or where somebody decided that's not actually crucial to fix right now because the code is not reachable or we have other priorities. Having AI to burn down that security backlog will make both the open source ecosystem and a lot of commercial software projects so much better because it brings that security

effort down that every engineering manager swings back and forth between the tech debt, the legacy code, the security, accessibility, European regulation, whatever, right? And the innovation backlog. And there isn't really like a balance between the two. It's just like, what is the most urgent issue, the biggest fire drill? Is it your sales team telling you if we don't get that one feature, we can't sell the product? Or is it the security team telling you, you've got to fix that one issue, otherwise we're going to

flag you up to the management chain, right? And so that's, I think, is the AI side of things. But similarly, GitHub as a platform needs to involve, to support and have all the primitives for these agents and the AI to work in tandem with the human. Do you think there are problems that people are not addressing yet that emerge from

this transition, how software development is done, right? Like, so, for example, you know, you feel like we're somewhere between crossing the tipping point of the majority of code being generated this year to maybe like all of the code in like some cases or some tasks. How does that change like testing or, you know, the way we should look at technical debt or any of that? To be clear, I don't think all of the code is written by AI. I think the way

this will work is that we have two layers. We have the machine language layer, which is Python or Ruby or Rust. Those are effectively abstractions of the chipset, the machine instruction set. And that's the last layer that's deterministic. Like programming language inherently does exactly what I want it to do. And then human language is inherently non-deterministic. The three of us can say the same sentence and mean a different thing.

And so while we will use human language to describe a lot of the features and behaviors that we're going to build, we will still have the programming language layer below that, that we are going back and forth as engineers to figure out is the code that was written by AI actually the correct one? Is it the one that aligns with my cost profile, if you will, as an example, right? Like at the end of the day, we're still all running businesses that have to have

positive profit margins. I think we're going to, as engineers, have both of these layers and we're heading into a world of more human language and less programming language. But at the same time, we are in a world where lots of financial services institutions still run COBOL code on mainframes. And we are very far away from just taking that code that's 30, 40 years old and just

hunting an agent that transforms that magically into a cloud application, right? Like, I think that's coming, but it's like self-driving cars are coming as well. But we don't know when that cutover point actually happens where you can have a car without a steering wheel and it drives you everywhere, you know, within the country you live in, right? Like,

it works for Waymo in San Francisco and it doesn't work for Waymo all the way down to SFO to San Jose yet, right? And so the scope will increase, but we are far away from, I think, solving all the tech debt and all the legacy code that exists. And so we are still

I think for like a decade or so, at least going to have software developers that work in lots of old school, you know, PHP code and COBOL code and all that stuff. While at the extreme other end of the spectrum with web development and AI, you're going to be able to

And we're already there. Just look at a 10-year-old, give them a tool like Copilot or Replicit, Bolt, you name it, and have them type a couple of prompts and have them explore how that works and how they can, similar to stable diffusion mid-journey, render software themselves and iterate on that.

You yourself lead a large team of software engineers. As you said, you have more human language and instruction versus machine language. Does it change what you look for or what you want to develop in your own team? Well, what you're looking at right now is I think this, how do you describe actually a problem specific enough that an agent can pick it up, right? Like basically the planning and tracking side of software development, the issue, right? That's often the biggest challenge, right?

that you have as soon as you have a decent team size. Or like 10 person startup has no problem and most of 10 person startups don't have a product manager. The founder is the product manager and the rest is just building the stuff. And if you have a problem to solve, you have very short communication paths.

If you have a thousand engineers, their biggest problem is what do you want to build? How do you build it? What did you actually mean when you've rolled up this thing? And if you look into that space, there isn't much AI helping you yet. We have been in early phases ourselves with that, with Copilot Workspace.

where we have a spec and a brainstorming agent that basically looks at what you wrote in a GitHub issue, compares it with a code base, and describes to you the before and after in human language. And then you can, similar to a Notion doc, just modify that and basically add stuff to the specification.

So I think that's going to be a whole set of agentic behavior that we're going to bring into the product management space. Similar for designers, right? Like today, a lot of designs are hand-drawn in Figma. I think tomorrow you're going to, as a designer, type effectively the same specification as a product manager and you have an AI to render the code for

the wireframes and then apply grounding out of your design system to make it look like your product. And so those disciplines get closer to each other and a product manager will be able to, if they're good in writing a specification,

create the whole change set and the designer will be able to take over part of the product management role and the engineer gets closer to these other roles as if they're good in describing the feature can take over that part as well. So I think that's where a lot of innovation is going to happen in rethinking how the traditional disciplines in the software engineering team are evolving in the coming years as we have more and more of these agents available and they're actually good at what they do. As you think about these different agents and these different use cases,

Do you think it's going to be the same company or product that provides all three? Do you think it's going to be one interface? Is it going to be a different interface? I'm sort of curious how you think about the actual flow in terms of very different users in some sense, although with some overlapping either responsibilities or goals.

Yeah. And what are the set of tools that they interact with? And is it a singular tool? Is it many? Is it one company? Is it many? Where does it launch out of? Like, how do you think about all that stuff? One of our strongest beliefs at GitHub is developer choice. And, you know, imagine a GitHub as a platform where you had only JavaScript libraries available or only, you know, React available to you. And we would tell you that's the only open source library you need to build an application, right? Like,

there would be a set of users using React, using GitHub because they love React and the rest would go somewhere else because some other platform would offer them all these other open source components. Right? In AI, I think we're going to see the same thing. We're going to see a stack or universe of, um,

companies that offer different parts of the software development lifecycle. And developers pick the one that they like the most, that they have experience with, and that convicted of the future. A lot of that is part of a belief system. Programming languages in many ways are very similar. And then if you look at the discussion between developers, you get the feeling they're very different.

to each other, right? Like the end of the day, they're all compiling down to an instruction set that runs on your Apple M4 chip or your Intel CPU or AMD or Nvidia or whatever. Right? So I think we are going to have a stack of different tools and there's going to be companies that offer, you know, all the tools. Well, not all of them because you're never going to have all of the developer tools out of one hand anyway.

Think about GitHub, we are a big platform, but then you still have an editor and an operating system and a container solution and a cloud that doesn't come from GitLab. HashiCorp, Terraform, All World as an example, or Vercel and Next.js as another example. Go into any random company in the Bay Area and they're all going to have a different stack of tools that they have combined because they believe that's the best stack for them at this point. So I think in this AI world, we're going to see the same thing. You're going to have

choice of different agents. You're already there where you have choice of different models and some believe the cloud model is better, others believe the open-eyes model is better. You know, the reality is somewhere in the middle and different scenarios are better with different models. So I think the same will be true in this agentic future that we're heading into. Is that true given the generalizability that we're seeing? In other words, if you were to remove X percent of the models and you just got stuck with one of the ones you mentioned,

Up to a point, you'd still be extremely happy given the relative capabilities we had four or five years ago. In other words, it's a little bit of like we have so many great options and some things are better than others. But fundamentally, any one of these things would be spectacular.

by any sort of baseline metric. It depends on what end state we're talking about, right? Like if the singularity is coming, then none of that matters. Five years from now, five years. You know, we started Copilot almost five years ago, June 2020. And that was what, GPT-3 at that point? GPT-3 was really the early experiments. And then we got this model that then eventually became Codex, which was this code-specific, you know,

version of the model. And today that, you know, no longer really exists, right? Like today, everybody is sets on top of one of these more powerful base models. Yeah. Yeah. And that's kind of my point is to some extent, the generalizability started to take over. And so I'm just a little bit curious how you think about generalizability versus specialization and five-year time horizon for agents. I can see that happening at the model layer.

But it's again, like predicting a little bit of, you know, when do we truly have self-driving cars? And, you know, I had a Tesla for 10 years with self-driving and autopilot in one form or another, and it still cannot make the left turn into my neighborhood. I can see that future happening, but I don't know when that is and when the models are basically just all, you know, about equal.

But I think for software developers, the lowest level only matters until there's differentiation at the higher layer of the stack. I think programming language or open source libraries are great examples for that because if you zoom out enough, they're all the same. At the end of the day, whether you're building an app with Swift or Kotlin or React Native, what does it matter? That's just the intricacies of software development and the belief system that we have

And so I think the differentiation is going to come from both where the developer gets the best experience and doing their day-to-day. Like, where can I start my morning, pick up something I want to work on, explore my creativity and get the job done with the least amount of frustration and the highest amount of ROI in terms of what can I ship in

Like software development, you know, over the last 30 years has always, or actually the last 50 years, if you go back all the way to the 1970s, when, you know, microcomputers came and all of a sudden you no longer had to share a mainframe with others, was always about how can I take all my grand ideas that are way bigger than what I can actually achieve as an individual. How can I, you know, get that done faster? I don't think we are at the top of that exponential curve. I think there's still a lot to come.

The other question you could ask is, when do a CEO of GitHub get to the point where my backlog is empty? And I just don't believe that that point is ever coming. Yeah, there's a super related, interesting question to what you're saying, which is, for how long are humans making decisions on what agents to use? Because if you look at it, there are certain roles, a lot of the ones that you mentioned, developers, designers, etc.,

that have traditionally tended to be a little bit trend-based. It's almost mimetic what certain developers will use sometimes. And obviously there's like the dramatically superior products and there's sort of clear choices around certain tooling. And sometimes it just feels like it's kind of cool and so people are using it. Same with programming languages, right? So it's almost an interesting question when the human component of decision-making goes out the window

are the decisions that are made radically different because you're getting rid of trendiness. You know, you're not going to use Go, you're just going to use Python or whatever. If I look at, you know, my team, how often as a CEO do I have to check in with them to see if what they're building is actually what I thought, uh,

I want them to build when I gave them the task, right? So it's number one, the human that takes over a task and feature an epic, whatever, still has a loop with other team members to kind of like ensure what they're building is actually the right thing. I don't see a world where we can be specific enough when we give the agent work that it can just do it all by themselves, unless, you know, the unit size is very, very small, right?

The other side of that question, I think, is when do we get to the point where all software is personal software? And in fact, I no longer install an app from the App Store. I just use a natural language interface to build all the apps myself. And so I have this completely personal software on my personal computer, my smartphone, instead of off-the-shelf software that is the same for all ourselves, where the user interface effectively is

is completely personalized. And, you know, we have science fiction movies or action movies like Iron Man, right? Like where Java is completely personalized to Tony Stark.

And so I think that future, that will happen in the next five years for sure. It's just the question is how good this job is going to be and can I just tell it Spring Break is coming up, same hotel, same family, you know, and it books me the trip and the only question I have to confirm is do I do the $5,000 trip? One other thing that's been striking about GitHub and Copilot and everything else is the actual business success of all of it, right? And I think it's been quite striking on the earnings calls

more recently that have been done.

What can you share in terms of business and financial metrics and the impact that Copilot and GitHub more generally are having for Microsoft? Not a lot beyond what's in the earnings call. I'm trying to remember. I think the last number we shared was a few quarters ago, 77,000 organizations using Copilot. And back then, the number of paid users was 1.8 million paid users. We haven't shared an updated number since, so I can't share that latest number, but

I think what's really interesting from these earnings calls, if you look at the number of logos that Satya has called out, it's across the whole spectrum of industries. It's not just cool startups. It's not just financial services institution. It's really every industry that has adopted Copilot. And I don't think there has been a developer tool that has been adopted that

with such a velocity across the whole spectrum of software development in any company size and in any industry. You know, if you think about it, $20 compared to the salary of an average software developer in the United States is like, what, 0.1%, if at all.

And then we're talking about, you know, 25, 28% productivity gains on the end-to-end, 55% or higher on the coding task. But as we said earlier, right, developers do more than just coding. That's an incredible ROI on the dollar spent.

And I think that's what is driving this adoption curve. And then any companies now, a software company, they all have the same problem described earlier. They have long backlogs and way too much work. And every time one of the managers goes to their team and asks them, how long does it take to implement a feature? It becomes the Jim Kirk, Scotty joke that, how long does it take to repair the wall drive? And you get an estimate that's

outrageously long and then it becomes a negotiation where the captain sets the deadline instead of the engineer actually estimating what's possible. And I think that's where, you know, a lot of the business success of Copa is coming from. All the people writing software are frustrated how long it takes, not because they don't think the engineers are good, but because of the complexity of building software. How much do you think this pricing changes? And I know it's just speculation at this point.

when you're actually replacing people. And I know in a lot of industries, it could be legal, it could be accounting, it could be coding. People say, well, eventually this will shift to value-based pricing because eventually instead of just paying 20 bucks a month to make a person more productive, you're actually replacing a person who costs 50 or 100 or $200,000 a year or whatever it is, depending on what their role is. Also just in different disciplines. So I'm just sort of curious how you think about, is this eventually a rent-a-programmer and it's priced like a programmer?

Does it all get commoditized and eventually something that would normally cost $100,000, $200,000, $300,000 a year costs $3,000 a year? How do you think about where this market goes? I think it's going to be compute-based or some unit that's a derivative of compute as a metric. So it's going to be cheap. It's going to be cheap in the same way that your dishwasher in your kitchen is not a derivative of what a person would cost you when doing your dishes every single day.

But I think the buyer persona is not going to be willing to pay for a machine, you know, whether that's a dishwasher or an agent, a price that's an equivalent of human developer. And I think that's actually the correct approach.

mindset because I don't believe that the AI agent is actually replacing the developer. The creative part is still coming from the software developer, the systems thinking. Predicting the future always has the fun part of that. I'm coming back on the podcast in a year or two and you're telling me how wrong I was about my predictions.

But I think there's a lot of decisions that are made in software development that a human has to make. What database, what cloud, a lot of that are a function of the business and how it operates. Which cloud you're using is not necessarily a question of how much the cloud costs. It's a strategic decision of the CTO or the engineering leadership team.

And more and more we see companies using more than one cloud because they don't want to have a dependency on just one single supplier. In the same way that any random car manufacturer has multiple suppliers for airbags because they don't want to be stuck with their factory line when airbags are not deliverable from that one supplier.

And so I think, you know, the agents, the price points will certainly go up as these agents become more powerful. You know, we see that with OpenAI, the highest tier now costs $200 for deep research and the O1 Pro model. And people see the value in that and

I think, you know, two years ago, if we had predicted that, we wouldn't have believed it. You're willing to pay $200 a month for a chat agent because the flip side of that often is in software that people feel like a $5 subscription for a mobile app is a lot of money. And you can just see that when you look into the reviews of apps that move from a one-time payment, you know, to a subscription model of how many people...

don't like that model because they feel like software is something that you buy once, like a CD, and then you own it. Definitely, they're going to price increases that will be based on the value that you're getting out of it. Because, you know, the other side of that is that

Human developers are expensive because there's limited supply. Agents will have infinite supply that will only be limited by the amount of compute capacity GPU is available in data centers. Speaking of that unlock of supply, like we've been talking about, like, what is the pricing of the code generation? I think there's also a question of just like what happens to the value of software at all.

Like everybody's been talking about Javon's paradox for a while. I don't want to ask about that, but maybe something more specific. You're from East Germany. You remember the Trabant car? I do. I had one. Oh, well, my parents had one. Oh, OK. Right. So you can you can tell me what it's actually like. But good for you guys, because it was the OK car, but it was the default car that ended up having this like 10 year waiting list because of the supply constraint with the rest of the world. And then as soon as the wall came down, you know, the demand completely collapses.

Yeah.

Are there types of software that you think collapses in value when AI takes away some of the scarcity of engineering? You know, the Trabant, the waitlist was actually, I think, 17 years in the late 80s. OK, 17, not 10. Yeah. That word, by the way, still exists today. It exists in supercars, right? Like often you can buy a supercar, you know, the top end Porsche 911 R3 or whatever, and

And then the resale price is higher than the new price because you can't get one to go to a dealer because at the dealer, you have to buy like 100 Porsches first before you get a slot for that exclusive top of the line Porsche or Ferrari is the same thing.

And so the Trabant, actually, the one that my dad owned, he sold, I think in '84, '85 to a neighbor at a higher price than we bought it because you could shortcut the 17-year wait to get a car. And often parents had a subscription, quote unquote subscription, like signed up their kids already for a car when the kids were still young. So you could actually get one by the time you reached adulthood and could do a driver's license.

And so, you know, I think we're going to see, coming to your software question, right? Like we're going to see it going both ways, right? Like if you think about Copilot, Copilot, you know, costs for businesses $20 per user per month. That's actually almost exactly the same price as you pay for GitHub Enterprise.

which is $21 per user per month, right? And so for storing all your repositories, managing all your issues, your whole software development lifecycle was $21 per user per month. And many used to perceive that as a lot of money for DevOps, right?

And then we came with co-pilot auto-completion and that was $20 a month. And so all of a sudden that sub-feature of the software development lifecycle, auto-completion, costs $20. And that goes back to Elad's question. If there's the value where you get the ROI and you get 25% productivity increases, yeah, you're willing to pay more for something that probably five years ago, if I told you auto-completion is going to be that standalone feature at Ruin by AI, that costs $20.

more than the average selling price for all of GitHub, you would have said, well, that sounds unlikely. I think we're going to see deflation of software prices. And so I think it's a mix of both. Some things we won't pay for it anymore. Nobody pays for the operating system anymore. And then at the same time, you pay way more than ever for your Netflix subscription and for your Office subscription and

and all those kind of things. So I think both of these things will be two at the same time. And it's all about how much value do you get for your business paying for that solution, whether it's doing it yourself or using something that you manage yourself or install on your own server. GitHub is foundational infrastructure for open source. So I'm sure you have like general opinions about what's happening in the open source ecosystem. Today, you can use Cloud and OpenAI in Copilot.

And Gemini, but not necessarily open source models right now. Correct. So in Copilot, we have Claw, Gemini, and then OpenAI. And OpenAI has different models. I was just processing this in my head. Wait, there's more than three models, but it's the 4.0 model and the 0.1 and the 0.3 mini model. In GitHub models, which is our model catalog, we have open source or open weights model like Lama.

as an example, and then all kinds of other models like Mistral, Cohere, Microsoft's 5.4 model. And the model catalog, you know, while it's a separate feature within GitHub, you can add models in Copilot because Copilot has extensions and you can actually reach from Copilot into the model catalog. And so if you want to just run quickly inference against 5.4, you can do that by using the add models extension in Copilot. So that way, if you have more models in the catalog,

the ones that are packaged into CoPilot. I know that. What do you think is the relevance of open source versus the proprietary model APIs for developers in the future? The biggest thing I think is that open source is going to drive innovation. And we saw that with DeepSeek earlier this year, or actually a couple of weeks ago, it's not that long ago even. Long year, yeah. It feels like already half a year has been passed instead of just a month and a half.

But I think open source is going to drive innovation. You know, we saw that with image models like stable diffusion and now there's the Flux model from a startup actually not too far from my home base in Germany in the Black Forest in Freiburg. Black Forest Labs is actually the company behind Flux and

And so we're going to see innovation, I think, on open source models that drive the other vendors and this back and forth between the open source ecosystem and the propriety, closed source companies will, I think, accelerate the whole space. DeepSeek is the most prominent example right now. We can look into this. The paper is open. The models are open. Some of them are like, you know,

open source under the MIT license. Others are like open weights and you can look at the weights and then the code run it as open source, but the weights itself are under some proprietary license and governed by Chinese law and whatnot. And I think that is going to drive innovation and it's going to open up that space and it democratizes access because if you just want to play with a model, you don't have to run

inference against the commercial API. You can just try it out yourself on your local machine and play with this. And if you think about kids and students and research, that opens up a huge space. And that's ultimately what has always been part of our DNA at GitHub. Was that a satisfying answer, Sarah? Yeah, yeah. I think the most satisfying answer is like somebody wins.

Right. But I think that's a very hard thing to predict right now. What has one iPhone or Android, Windows or Linux or Mac OS for that matter? I think we like to think about these binary battles in the tech industry. And the reality is that's not actually how that works and certainly not

in the developer space, right? Like React hasn't won. And there's always going to be the next thing, you know? Before React, there was jQuery or whatever library you preferred. I think there's going to be a next programming language after Python and TypeScript and Rust. And Rust in itself, you know, wasn't really a thing

five years ago. And so there's going to be more languages that are probably closer to human language and to be more specific about the natural language layer in AI and the programming language layer that converts down to the CPU or GPU. And so I think there's no winning. There's always just, you're playing the infinite game. It's like Minecraft.

Software is like Minecraft and there is no winning in Minecraft. You can win little battles and they're isolated to a certain sub-challenge or whatever, a quest. But ultimately, we're building a bigger and bigger world of software and there's always going to be a next big thing. That's a funny analogy. If I think about any individual developer, like

There's something people have been saying to me. Developers of a particular ilk, right? Really strong technical people who are more experienced, not all of them, but like more experienced gremlins, systems developers, often people very attached to Rust.

And they'll say basically, like, they're worried about the next generation of developers building the taste and understanding of architectural choices and the tradeoffs and corner cases of how a particular implementation can fail, given some shape of data, given their experiences of the actual implementation. Right.

Right. And so they're they're worried, you know, obviously the right thing to do for anybody who wants to win that next level of Minecraft in 2025 is like use AI aggressively, learn to use it. But like, does that concern from this segment of I'm sure you've heard it. Does that concern like resonate with you at all? Like, can you foster the requisite depth of understanding of engineering at an abstract level when we're not writing the code or is it like a silly, silly concern?

I wouldn't call it silly because obviously, you know, there's some truth to that, right? It's easy, you know, to cheat at a programming exercise or advent of code and those kind of things. As these AI models get better, these competitions of who's the best hacker or coder are going to have to move to a whole different level where you assume that the developer is using AI to solve the challenges because otherwise it's going to be way too easy.

If you're thinking about the next generation of developers, you know, maybe not 2025, but 2035. Like, look, you know, and you mentioned, you know, me growing up in East Germany and then the wall fell and I bought a Commodore 64, but they had no internet. And so I bought books and magazines and that was it, right? Like there was no forum I could go to and ask questions. I could go to, I went to computer club.

every Wednesday or so until nobody there had anything to say anymore that I didn't know already, right? If you take that and compare it to today, the kids of today and those that want to learn coding have an infinite amount of knowledge available to them. And you know what? Also an infinite amount of patience because Copilot doesn't run out of patience.

Parents do. I am one. And so it's incredibly democratizing to have AI available if you want to learn coding. Your parents don't have to have any technical background. All you really need is an internet connection on your mobile phone.

and, and, and one of these co-pilots or, or chat GPTs or what, whatever you prefer. And you can start asking coding questions and you can ask about Boolean logic and about systems thinking, and you can go infinitely deep on, on any of those questions and, and, and, and, you know, traverse to, to other topics as you like. Right. And so I think, you know, we are going to see, you know, a new generation of, um,

humans that, you know, grow up with the technology for them, it's just natural to, to leverage their, their personal assistant, their personal set of agents. You know, I recently called it the orchestra of agents and you're the conductor of that orchestra of agents and they know how to do that. And so they can achieve in the same amount of time, so much more than we could, you know, in the last, in the last 30 years. And I think that's incredibly exciting because like, again, like we,

find me a developer that doesn't have this big idea of that computer game or software system or feature that they always wanted to build and don't have the time for. Like my engineers talk much more about being overcommitted and burned out and not having enough time for all the things I'm asking for and the customer is asking for and the security team is asking for. And so I think that's just where we're heading and how this is going to be developed.

Super exciting. Both actually in open source as well, right? It was open source sustainability is another big topic that we probably spent another hour on and in any kind of software that people want to build. I definitely agree with that, you know, excitement and optimism. I think about my three kids and it's like,

like what they would be able to learn at what pace with the AI resources that people will have. And I'm incredibly jealous. I'm like, I could be much better as an engineer, so much faster with, as you said, the infinite patience and understanding of today's models. By the way, I was very lucky. My parents are both engineers, right? But it's a very human dynamic where I'd ask a question and my dad would be like, it's logic, Sarah. I'm like, oh no. Yeah.

Can I ask you, you know, maybe a more personal question to close? Like East Berlin, you know, you have this unique experience of this really rapid technological change after reunification. Do you think that informs at all how you think about like the speed of the current AI transition and how like users and human beings will react to it?

I always wanted to believe that a lot of my life has been defined by that one moment of change in 1989. And I remember the night when the wall fell or when it was announced that the wall would be opened. And it was a Thursday night and then Friday was normal school. Saturday was still school as well, a half day in school. And I think I was one of four kids that showed up in my class and then they sent us home and

And we actually crossed over to West Berlin. I think the thing that is important for that generation of kids that live through that change is that

They can no longer return to their childhood. You know, home is gone. Like, you know, there isn't like that store in the corner that's the same as it was like 40 years ago. And the schools are gone. The system is gone. The traditions, all that resolved into, you know, that new world. And so it's a bit like when you're moving from one country to another, which then I, you know, did 10 years ago as well to move when Microsoft

bought my company. Once you have done that step in your life, you gained a whole new perspective on things. And I think that's, you know, unification in 1990 and then, you know, through the steps of my life, including, you know, to become the GitHub CEO through, you know, random decisions or at the time that felt random,

This is how I got here and this is how I look forward and I'm optimistic about the future while recognizing my past and taking some of those experiences when I talk with you guys and reflect on what it was like in the 90s to program on a Commodore 64 before and after the internet, right? Before and after open source, before and after the cloud, before and after mobile. And now we have before and after AI and there's no looking back.

the future will be that we have AI for almost everything we do in our lives, if we want to. You know, you can still always throw your cell phone into the corner and enjoy your day without the internet. This has been great, Thomas. Thanks so much for having the conversation. Thank you so much for having me. It was great to connect, sir. I appreciate the time and everything else.

Find us on Twitter at NoPriorsPod. Subscribe to our YouTube channel if you want to see our faces. Follow the show on Apple Podcasts, Spotify, or wherever you listen. That way you get a new episode every week. And sign up for emails or find transcripts for every episode at no-priors.com.