Today on the AI Daily Brief, why AI skeptics are nuts. The AI Daily Brief is a daily podcast and video about the most important news and discussions in AI.
Hello, friends. Welcome back to the AI Daily Brief. Quick announcements. First of all, thanks to today's sponsors, KPMG, Blitzy, and Vanta. As always, if you are looking for an ad-free version of the show, go to patreon.com slash ai daily brief. Ad-free starts at just $3 a month. And today, of course, is a long reads episode of the show. An essay has been exploding around the AI Twitter sphere for the last week or so called quite plainly, my AI skeptic friends are all nuts.
Now, this is focused on AI coding specifically, but as you'll see, some of it is more broadly applicable. We are going to real human read some excerpts, and then we'll discuss it a little bit. The piece is by Thomas Potacek, and Thomas does want to make clear in an important caveat, quote, I'm only discussing the implications of LLMs for software development. For art, music, and writing, I got nothing. I'm inclined to believe the skeptics in those fields. I just don't believe them about mine.
Thomas writes, a heartfelt provocation about AI-assisted programming. Tech execs are mandating LLM adoption. That's bad strategy, but I get where they're coming from.
Some of the smartest people I know share a bone-deep belief that AI is a fad, the next generation of NFT mania. I've been reluctant to push back on them because, well, they're smarter than me, but their arguments are unserious and worth confronting. Extraordinarily talented people are doing work that LLMs already do better out of spite. All progress on LLMs could halt today, and LLMs would remain the second most important thing to happen over the course of my career.
Thomas then talks about how he has been shipping software since the mid-90s through a whole series of different languages, leading ultimately to the conclusion, however you define serious developer, I qualify.
In the first section, Level Setting, Thomas writes: "First, we need to get on the same page. If you were trying and failing to use an LLM for code six months ago, you're not doing what most serious LLM-assisted coders are doing. People coding with LLMs today use agents. Agents get to poke around your codebase on their own. They author files directly, they run tools, they compile code, run tests, and iterate on the results.
They also pull an arbitrary code from the tree or from other trees online into their context windows, run standard Unix tools to navigate the tree and extract information, interact with Git, run existing tooling, and make essentially arbitrary tool calls that you set up through MCP. If you're making requests on a chat GPT page and then pasting the resulting broken code into your editor, you're not doing what the AI boosters are doing. No wonder you're talking past each other.
Now Thomas moves into the positive case. Apologies for the explanation for those of you watching, but for those of you who are just listening, he shares an image of four quadrants. The quadrants are 1. Fun and important, 2. Tedious and important, 3. Tedious and pointless, and 4. Fun and pointless. Thomas continues, "LLMs can write a large fraction of all the tedious code you'll ever need to write. And most code on most projects is tedious. LLMs drastically reduce the number of things you'll ever need to Google.
They look things up themselves. Most importantly, they don't get tired. They're immune to inertia. Think of anything you wanted to build but didn't. You tried to hone in on some first steps. If you'd been in the limerit phase of a new programming language, you'd have started writing. But you weren't, so you put it off for a day, for a year, or your whole career. I can feel my blood pressure rising thinking of all the bookkeeping and googling and dependency drama of a new project. An LLM can be instructed to just figure all that crap out.
Often it will drop you precisely at that golden moment where things almost work, and development means tweaking code and immediately seeing things work better. That dopamine hit is why I code.
There's a downside. Sometimes gnarly stuff needs doing, but you don't want to do it, so you refactor unit tests, soothing yourself with the lie that you're doing real work. But an LLM can be told to go refactor all of your unit tests. An agent can occupy itself for hours putzing with your tests in a VM and come back later with a PR. If you listen to me, you'll know that. You'll feel worse yak shaving. You'll end up doing real work.
but you have no idea what the code is. And by the way, Thomas here is using these section headers to reflect the common complaints he hears. So again, this one is, but you have no idea what the code is. Thomas writes, are you a vibe coding YouTuber? Can you not read code? If so, astute point. Otherwise, what the F is wrong with you? You've always been responsible for what you merge to main. You were five years ago and you are tomorrow, whether or not you use an LLM. If you build something with an LLM that people will depend on, read the code. In fact, you'll probably do more than that.
You'll spend 5-10 minutes knocking it back into your own style. LLMs are showing signs of adapting to local idiom, but we're not there yet. People complain about LLM-generated code being probabilistic. No, it isn't. It's code. It's not YAC output. It's knowable. The LLM might be stochastic, but the LLM doesn't matter. What matters is whether you can make sense of the result and whether your guardrails hold.
Reading other people's code is part of the job. If you can't metabolize the boring repetitive code an LLM generates, skills issue. How are you handling the chaos human developers turn out on a deadline? For the last month or so, Gemini 2.5 has been my go-to. Almost nothing it spits out for me merges without edits. I'm sure there's a skill to getting a state-of-the-art model to one-shot a feature plus merge, but I don't care. I like moving the code around and chuckling to myself while I delete all the stupid comments. I have to read the code line by line anyways.
But hallucination! If hallucination matters to you, your programming language has let you down. Agents lint. They compile and run tests. If their LLM invents a new function signature, the agent sees the error. They feed it back to the LLM, which says, oh right, I totally made that up, and then tries again. You'll only notice this happening if you watch the chain of thought log your agent generates. Don't. This is why I like Zed's agent mode. It begs you to tab away and let it work, and pings you with a desktop notification when it's done.
I'm sure there are still environments where hallucination matters. But hallucination is the first thing developers bring up when someone suggests using LLMs, despite it being more or less a solved problem. But the code is crappy, like that of a junior developer.
Does an intern cost $20 a month? Because that's what Cursor.ai costs. Part of being a senior developer is making less able coders productive, be they fleshly or algebraic. Using agents well is both a skill and an engineering project all its own, of prompts, indices, and especially tooling. LLMs only produce crappy code if you let them. Maybe the current confusion is about who's doing what work. Today, LLMs do a lot of typing, googling, test cases, and edit-compile-test-debug cycles. But
But even the most clawed poison serious developers in the world still own curation, judgment, guidance, and direction. Also, let's stop kidding ourselves about how good our human first cuts really are. But it's bad at Rust. A lot of LLM skepticism probably isn't really about LLMs. It's projection. People say LLMs can't code, when what they really mean is LLMs can't write Rust. Fair enough. But people select languages in part based on how well LLMs work with them. So Rust people should get on that.
But the craft!
Do you like fine Japanese woodworking? All hand tools and sashimono joinery? Me too. Do it on your own time. I have a basic wood shop in my basement. I could get a lot of satisfaction from building a table. And if that table is a workbench or a grill table, sure, I'll build it. But if I need, like, a table for people to sit at in my office, I buy a friggin' table. Professional software developers are in the business of solving practical problems for people who code. We are not in our day jobs artisans.
Steve Jobs was wrong. We do not need to carve the unseen feet in the sculpture. Nobody cares if the logic board traces are pleasingly routed. If anything we build endures, it won't be because the codebase was beautiful.
Besides, that's not really what happens. If you're taking time carefully golfing functions down into graceful, fluent, minimal functional expressions, alarm bells should ring. You're yak shaving. The real world has depleted your focus. You're not building yourself soothing. Which, wait for it, is something LLMs are good for. They devour schlep and clear a path to the important stuff where your judgment and values really matter. But the mediocrity...
As a mid-late career coder, I've come to appreciate mediocrity. You should be so lucky as to have it flowing almost effortlessly from a tap. We all write mediocre code. Mediocre code, often fine. Not all code is equally important. Some code should be mediocre. Maximum effort on a random unit test? You're doing something wrong. Your team lead should correct you. Developers love to preen about code. They worry LLMs lower the ceiling for quality. Maybe, but they also raise the floor.
Gemini's floor is higher than my own. My code looks nice, but it's not as thorough. LLM code is repetitive, but mine includes dumb contortions where I got too clever trying to dry things up. And LLMs aren't mediocre on every access. They almost certainly have a bigger bag of algorithmic tricks than you do. But I'm getting ahead of myself. It doesn't matter. If truly mediocre code is all we ever get from an LLM, that's still huge. It's that much less mediocre code humans have to write.
But it'll never be AGI. I don't give a crap. Smart practitioners get wound up by the AIVC hype cycle. I can't blame them. But it's not an argument. Things either work or they don't, no matter what Jensen Huang has to say about it. This episode is brought to you by Blitzy, the enterprise autonomous software development platform with infinite code context.
Blitze is used alongside your favorite coding copilot as your batch software development platform for the enterprise seeking dramatic development acceleration on large-scale codebases. While traditional copilots help with line-by-line completions, Blitze works ahead of the IDE by first documenting your entire codebase, then deploying over 3,000 coordinated AI agents in parallel to batch build millions of lines of high-quality code. The scale difference is staggering. Copilots might give you a few hundred lines of code in seconds,
But Blitzy can generate up to 3 million lines of thoroughly vetted code. If your enterprise is looking to accelerate software development, contact us at blitzy.com to book a custom demo or press get started to begin using the product right away. Today's episode is brought to you by Plum. If you're building agentic workflows for clients or colleagues, it's time to take another look at Plum. Plum is where AI experts create, deploy, manage, and monetize complex automations.
With features like one-click updates that reach all your subscribers, user-level variables for personalization, and the ability to protect your prompts and workflow IP, it's the best place to grow your AI automation practice. Serve twice the clients in half the time with Plum. Sign up today at useplum.com. That's U-S-E-P-L-U-M-B dot com forward slash N-L-W. Today's episode is brought to you by Vanta. In today's business landscape, businesses can't just claim security, they have to prove it.
Achieving compliance with a framework like SOC 2, ISO 27001, HIPAA, GDPR, and more is how businesses can demonstrate strong security practices.
The problem is that navigating security and compliance is time-consuming and complicated. It can take months of work and use up valuable time and resources. Vanta makes it easy and faster by automating compliance across 35+ frameworks. It gets you audit-ready in weeks instead of months and saves you up to 85% of associated costs. In fact, a recent IDC whitepaper found that Vanta customers achieved $535,000 per year in benefits, and the platform pays for itself in just three months.
The proof is in the numbers. More than 10,000 global companies trust Vanta. For a limited time, listeners get $1,000 off at vanta.com slash nlw. That's v-a-n-t-a dot com slash nlw for $1,000 off.
Today's episode is brought to you by Agency, an open source collective for interagent collaboration. Agents are, of course, the most important theme of the moment right now, not only on this show, but I think for businesses everywhere. And part of that is the expanded scope of what agents are starting to be able to do. While single agents can handle specific tasks, the real power comes when specialized agents collaborate to solve complex problems. However,
Right now, there is no standardized infrastructure for these agents to discover, communicate with, and work alongside one another. That's where Agency, spelled A-G-N-T-C-Y, comes in. Agency is an open-source collective building the Internet of Agents, a global collaboration layer where AI agents can work together. It will connect systems across vendors and frameworks, solving the biggest problems of discovery, interoperability, and scalability for enterprises.
With contributors like Cisco, Crew.ai, Langchain, and MongoDB, Agency is breaking down silos and building the future of interoperable AI. Shape the future of enterprise innovation. Visit agency.org to explore use cases now. That's A-G-N-T-C-Y dot org.
But they take our jobs. So does open source. We used to pay good money for databases. We're a field premised on automating other people's jobs away. Productivity gains, say the economists. You get what that means, right? Fewer people doing the same stuff. Talk to a travel agent lately, or a floor broker, or a record store clerk, or a darkroom tech. When this argument comes up, libertarian-leaning VCs start to chant. Lamplighters, creative destruction, new kinds of work. Maybe. But I'm not hypnotized. I have no effing clue whether we're going to be better off after LLMs.
Things could get a lot worse for us. LLMs might really displace many software developers. That's not a high horse we get to ride. Our jobs are just as much in tech's line of fire as everyone else's have been for the last three decades. We're not East Coast dock workers. We won't stop progress on our own.
But the plagiarism. Artificial intelligence is profoundly, and probably unfairly, threatening to visual artists in ways that it might be hard to appreciate if you don't work in the arts. We imagine artists spending their working hours pushing the limits of expression. But the media and artists isn't producing gallery pieces. They produce on brief, churning out competent illustrations and compositions for magazine covers, museum displays, motion graphics, and game assets. LLMs easily, alarmingly, clear industry-quality bars. Gallingly, one of the things they're best at is churning out just-good-enough facsimiles of human creative work.
I have family in visual arts. I can't talk to them about LLMs. I don't blame them. They're probably not wrong.
Meanwhile, software developers spot code fragments seemingly lifted from public repositories on GitHub and lose their crap. What about the licensing? If you're a lawyer, I defer. But if you're a software developer playing this card, cut me a little slack as I ask you to shove this concern up your butt. No profession has demonstrated more contempt for intellectual property. The media and dev think Star Wars and Daft Punk are a public commons. The great cultural product of developers has been opposing any protection that might inconvenience a monetizable media-sharing site. When they fail at policy, they route around it with coercion.
They stand up global-scale piracy networks to sneer at anyone who so much as tries to preserve a new release window for a TV show. It's all special pleading anyways. LLMs digest code further than you do. If you don't believe a typeface designer can stake a moral claim on the terminals and counters of a letter form, you sure as hell can't be possessive about a red-black tree. Positive case redux. When I started writing a couple days ago, I wrote a section to level-set the state-of-the-art of LLM-assisted Programming.
A bluefish filet has a longer shelf life than an LLM take. In the time it took you to read this, everything changed. Kids today don't just use agents. They use asynchronous agents. They wake up, free associate 13 different things for their LLMs to work on, make coffee, fill out a TPS report, drive to the Mars Cheese Castle, and then check their notifications. They've got 13 PRs to review, 3 get tossed and reprompted, 5 of them get the same feedback a junior dev gets, and 5 get merged.
I'm sipping rocket fuel right now, a friend tells me. The folks on my team who aren't embracing AI, it's like they're standing still. He's not BSing me. He doesn't work in SFBA. He's got no reason to lie. There's plenty of things I can't trust an LLM with. No LLM has access to prod here, but I've been first responder on an incident and fed 4.0, not 0.4 mini, 4.0, logged transcripts, and watched it in seconds spot LVM metadata corruption issues on a host we've been complaining about for months."
Am I better than an LLM agent at interrogating open search logs and honeycomb traces? No. No, I am not. To the consternation of many of my friends, I'm not a radical or a futurist. I'm a statist. I believe in the haphazard perseverance of complex systems of institutions of reversions to the mean. I write Go and Python code. I'm not a Kool-Aid drinker. But something real is happening. My smartest friends are blowing it off. Maybe I persuade you. Maybe I don't. Probably I don't. But we need to be done making space for bad arguments.
but I'm tired of hearing about it. And here I rejoin your company. All day, every day, a sizable chunk of the front page of Hacker News is allocated to LLMs. Incremental model updates, startups doing things with LLMs, LLM tutorials, screeds against LLMs. It's annoying, but AI is also incredibly, a word I use advisedly, important. It's getting the same kind of attention that smartphones got in 2008, and not as much as the internet got.
That seems about right. I think this is going to get clearer over the next year. The cool kid haughtiness about stochastic parrots and vibe coding can't survive much more contact with reality. I'm snarking about these people, but I mean what I said. They're smarter than me. And when they get over this affectation, they're going to make coding agents profoundly more effective than they are today. And that is the piece.
I think what's super valuable about this on a high level is obviously if you are an engineer who has been avoiding these things, which I can't imagine if you're listening to this show is you, but maybe someone sends this to you. This is a person who is not like me, an AI podcaster or an AI entrepreneur. It's someone who has simply found undeniably that this new set of tools are fundamentally and inarguably transformative. It's valuable because he's not making moral judgments or value judgments.
He's not even jumping on board with the techno-optimism that I and many of you probably hold dear about how this all shakes out on the other side. Instead, he is just saying that these tools are too powerful to ignore. I think that's a fairly decent starting point for any version of any skeptic in this space.
The second thing that I think is extremely important and valuable about this is that it embodies in a huge way the shift that we are living through right now that has, in fact, I believe, right from under our feet switched without us in some cases even noticing it from the assistant era to the agent era.
As Thomas points out, this is not just some basic coding assistance. This is actually using coding agents like teammates, junior devs, pair programmers. It's setting a bunch of prompts, going off and doing other things, and coming back and interacting with what's been produced.
It's background agents. It's asynchronous agents. This is increasingly the modality that we are going to see for all of our interaction with AI. If you are a general consumer and you've used deep research, you've had a little taste of this. You tell whichever LLM you're using to deep research what you want. It asks a set of clarifying questions, and then it tells you to buzz off while it does its work.
5, 10, 15 minutes later, whatever it is, you come back and it either got you what you need or you run it again. But it happened while you were doing other things. And that in itself totally changes what the actual capability set here is.
A third thing, which is a subtext, which I think is extremely important, is that there are entire categories of people for whom these debates are irrelevant because they just simply jumped on this train and ran all the way. And I think this is particularly generational. Young people are not going to putz around discussing the ethical mores of AI. They're just going to use it to out-compete everyone who's not using it.
Period. Full stop. End of story. And the amount of value they produce, at the speed at which they produce it, will be nigh impossible for any sort of policy-imposed guardrails to slow down in any meaningful way. Now, of course, that does not mean that there aren't parts of AI that should be subject to
Two questions of ethics and values and the society we want to live in. But as those conversations happen, there are many, many use cases that will simply march on, changing completely everything that they touch in their wake.
And when it comes to influencing the shape of AI, for those standing on the outside vaingloriously saying whatever combination of things that Thomas mentioned and all the other arguments we hear in other areas, I will leave you with the sentiment, if not the exact quote, from the end of SLC Punk after the main character has realized that being a burnout in Salt Lake City is going to do nothing for the world, and that if he actually wants to make a difference, taking his opportunity to go to Harvard and study law is probably a better bet than
we can do a hell of a lot more damage in the system than outside of it. For now, that's going to do it for today's AI Daily Brief. Until next time, peace.