Loading...
Loading...

A tour of how “vibe coding” actually looks in early 2026, from autonomous agent swarms writing millions of lines of code to solo builders running always-on AI employees on cheap hardware; the episode breaks down why techniques like the Ralph Wiggum loop matter, how Clawdbot changes what autonomy feels like in practice, and why the real shift isn’t new models—it’s removing humans as the bottleneck in building and shipping software. In the headlines: Davos wrestles with AI jobs and productivity, OpenAI pushes hard on enterprise, and global leaders argue over whether AI is creation, destruction, or both.
Brought to you by:
KPMG – Discover how AI is transforming possibility into reality. Tune into the new KPMG 'You Can with AI' podcast and unlock insights that will inform smarter decisions inside your enterprise. Listen now and start shaping your future with every episode. https://www.kpmg.us/AIpodcasts
Zencoder - From vibe coding to AI-first engineering - http://zencoder.ai/zenflow
Optimizely Opal - The agent orchestration platform build for marketers - https://www.optimizely.com/theaidailybrief
AssemblyAI - The best way to build Voice AI apps - https://www.assemblyai.com/brief
Section - Build an AI workforce at scale - https://www.sectionai.com/
LandfallIP - AI to Navigate the Patent Process - https://landfallip.com/
Robots & Pencils - Cloud-native AI solutions that power results https://robotsandpencils.com/
The Agent Readiness Audit from Superintelligent - Go to https://besuper.ai/ to request your company's agent readiness score.
The AI Daily Brief helps you understand the most important news and discussions in AI. Subscribe to the podcast version of The AI Daily Brief wherever you listen: https://pod.link/1680633614
Interested in sponsoring the show? [email protected]
Today on the AI Daily Brief, how the pros are vibe-coding in 2026, and before that in
the headlines, the last word on AI from Davos.
The AI Daily Brief is a daily podcast and video about the most important news and discussions
in AI.
Alright friends, quick announcements before we dive in.
First of all, thank you to today's sponsors Zencoder, robots and pencils, and super intelligent.
To get an ad-free version of the show, go to patreon.com, slash AI Daily Brief, or you can
subscribe on Apple Podcasts, and to learn anything else about the show, including how to sponsor
it, how to try to come get me to Jabra at you in person, or any of our various initiatives
like the forthcoming AIDB Intel, you can find all of that information at AIDailyBreathe.ai.
Now one more thing before we dive in, if you live anywhere from basically Texas to Maine,
you are either in the midst of or just gotten out of, one of the wildest winter storms we've
had in some time.
Where I am not only has school been canceled for Monday already, but we are actually dealing
with a complete 36 hour travel ban.
With up to two feet of snow anticipated, I am not counting on the power still being on,
and so over the sake of you guys not having to miss an episode, and me not being stressed
out by not being able to produce one, I'm actually recording this one on Saturday before
it all hits.
Still, there's a pretty good chance that with the chatter this weekend, especially
the main episode would have been Monday's main anyway, but that's the story, without
any further ado, let's dive in.
Welcome back to the AIDailyBreathe headlines edition, all the daily AI news you need in
around 5 minutes.
Given that we are recording this one a little bit early, our main topic is actually a bit
of a catch up on last week.
The World Economic Forum of course happened in Davos, all throughout last week, we covered
a couple of the big conversations, the AGI timeline conversation from Demisisabis and
Dario Amade, among other things, but overall, what was the vibe there?
And I will say before I get into that, that I sometimes don't even want to cover this
type of news, because I think that more or less, for those of you who are just trying
to understand what AI is going to mean for you, how it's going to impact your career,
your company, your job, ignoring basically everything that happens in the types of conversations
that go on at a place like Davos, ignoring all the conversation around markets, and infrastructure
buildouts and bubbles.
You'd basically be better off taking all of that time that you would spend thinking about
what people were jabbering about, and instead taking that time to just go figure out how
to build with these tools.
Yet, of course, we live in the world that we live in, and like it or not, the conversations
that happen in Davos are a useful reflection on what global leaders think about this moment,
and so give us insight into the context in which this industry and this technology is
going to operate.
One side of the conversation was the voices coming from the tech industry.
Reuters summed up that voice as jobs, jobs, jobs, the AI mantra in Davos as fears take
a back seat.
Now, that is a specific reference to Nvidia's Jensen Huang, who basically made the argument
that the amount of demand for chips, the infrastructure layer that needs to be built, the energy
infrastructure that needs to be built to service it, is all a big moment of job creation.
And indeed, I think it is the case that fairly uniquely relative to other moments of creative
destruction, even the transitional moment has the potential for a lot of creation as well.
I think Jensen is right to identify that there is a lot more skilled labor outside of
knowledge work that needs to be developed for this transition.
In other places, tech leaders talked about the productivity benefits that they were seeing.
Cisco talked about projects that had been too tedious to even contemplate before that
could now be done in a couple of weeks.
IBM's Chief Commercial Officer Rob Thomas said that AI was at the ROI stage.
He told Reuters, you can truly start to automate tasks and business processes.
TechCrunch said that even though we anticipated AI being a big topic of conversation, the extent
to which it shaped the event, with even the physical surrounding being dominated by tech
companies and pavilions was notable.
And yet, of course, if the technology folks were excited, concerns about AI-related job
displacement were on the agenda as well.
Christy Hoffman, the General Secretary of the 20-million-member strong, unique global
union, said AI is being sold as a productivity tool, which often means doing more with fewer
workers.
International Monetary Fund Managing Director, Christy Lena Georgieva, called AI as
tsunami hitting the labor market, with the potential to transform or eliminate 60% of
jobs in advanced economies and 40% globally.
Now I remember a study from a couple of years ago from one of the big global institutions,
IMF or World Bank or one of them, that basically had those numbers, so I assume that's what
she's talking about.
Providing some bright spot, she thought that as high-skilled workers see their wages rise
because of AI, they would likely consume more in ways that benefited the local service
economy.
She said one in 10 jobs is already enhanced by AI, and the people in these jobs are paid
better.
When they're paid better, they spend more money in the local economy.
They spend more money in restaurants here or there.
Demand for low-skilled jobs goes up, and actually total employment seems to slightly increase
because of it.
Now, for those who might be skeptical of this or seem like it feels relatively polyannish,
there have been studies that have shown that, for example, in San Francisco for each new
local tech job, 4.4 jobs for positions like retail clerks, cooks, teachers, and dentists
is also created.
At the same time, the IMF still has some big concerns.
The two that stood out is stagnating middle-class wages, especially for jobs that are not enhanced
by AI and increasing barriers to youth employment, as AI takes over the entry-level tasks.
Now behind the scenes in Davos, there was also a lot of jockeying for position.
The information wrote a piece all about how some Davos meetings were part of what seems
to be a larger strategy for open AI to get more aggressive about its enterprise recruitment.
Now this effort was not strictly restricted to Davos.
In fact, last week in San Francisco, Sam Altman hosted an extended business dinner with
Disney CEO Bob Eiger and other corporate execs.
The information writes that the gathering was intended to preview a new open AI offering
aimed at large companies that they could not determine what that offering was.
All that was happening, while open AI, COO, Brad Lightcap, and new chief revenue officer
Denise Dresser were schmoozing over in Davos.
Clearly the company is trying to message that they are, in fact, not behind when it comes
to enterprise.
In a Davos session, open AI, CFO, Sarah Fryer, said that by the end of the year, approximately
50% of their business will come from enterprise customers.
In Sam Altman tweeted that they had added more than a billion in ARR over the last month
just from their API business.
Very clearly trying to shift the narrative, he says, people mostly think of us as Chad
CBT, but the API team is doing amazing work.
So what is this all add up to?
It's kind of hard to tell.
One of the reasons we may not be able to have quite a strong sense of what the general
sentiment around AI was is just that there were, of course, other more geopolitical conversations
that made even the AI conversation take a back seat.
I think if anything, Jamie Dimon's crisp realism that no one can put their head in the
sand, that AI is not a force that is likely to be stopped, but that there could be challenges
for how fast it's going to cause change in society that we may have to address, might
be a fairly good representation of the median.
Hopefully it's kind of notable to me just how little the momentum cares.
Going back to my initial point, if you mostly are interested in AI when it comes to how
it's going to impact your life, let's just say you can safely switch from this headline
section to what might be a much more pertinent main episode.
If you're using AI to code, ask yourself, are you building software or are you just
playing prompt roulette?
We know that unstructured prompting works at first, but eventually it leads to AI slop
and technical debt.
Enter Zenflow.
Zenflow takes you from vibe coding to AI first engineering.
It's the first AI orchestration layer that brings discipline to the chaos.
It transforms freeform prompting into spec driven workflows and multi agent verification,
where agents actually cross check each other to prevent drift.
You can even command a fleet of parallel agents to implement features in fix bugs simultaneously.
You've seen teams accelerate delivery 2x to 10x.
Stop gambling with prompts.
Start orchestrating your AI.
Turn raw speed into reliable production grade output at Zenflow.free.
Most companies don't struggle with ideas.
They struggle with turning them into real AI systems that deliver value.
Robots and pencils is a company built to close that gap.
They design and deliver intelligent cloud native systems powered by generative and agentic
AI with focus, speed, and clear outcomes.
Designs and pencils works in small, high-impact pods, engineers, strategists, designers,
and applied AI specialists working together to move from idea to production without unnecessary
friction.
Powered by RoboWorks, their agentic acceleration platform, teams deliver meaningful results
including initial launches in as little as 45 days depending on scope.
If your organization is ready to move faster, reduce complexity, and turn AI ambition into
real results, robots and pencils is built for that moment.
Start the conversation at robotsandpensils.com slash AI Daily Brief.
Robotsandpensils.com slash AI Daily Brief.
Robots and pencils, impact at velocity.
Today's episode is brought to you by my company's Super Intelligent.
In 2026, one of the key themes in Enterprise AI, if not the key theme, is going to be
how good is the infrastructure into which you are putting AI in agents.
Superintelligence agent readiness audits are specifically designed to help you figure
out one, where and how AI in agents can maximize business impact for you, and two, what
you need to do to set up your organization to be best able to leverage those new gains.
If you want to truly take advantage of how AI in agents can not only enhance productivity,
but actually fundamentally change outcomes in measurable ways in your business this year,
go to bsupert.ai.
Welcome back to the AI Daily Brief.
Today we are doing a little bit of a catch up on the terms that you might have heard in
passing, especially if you've been anywhere near AI Twitter slash X over the past couple
of weeks.
There are a few things that might sound like absolute Greek to you, but which combine
tell the story of how vibe coding, which I really mean AI in agent coding, are evolving
early into this year.
Entrepreneur and content creator Riley Brown recently tweeted, cool cloud stuff, emotion
skill, Claudebot, CLAWD, Agent SDK, Ralph, and co-work.
Now if you are thinking, I don't know what any of those things mean, don't worry, you
are not alone and we're going to get into much of it today.
The context of all of this is the big shift in perception over the last couple of weeks,
which has been pretty well chronicled in episodes throughout this month.
It wasn't that we got a new model or anything like that, it's that everyone went home
for the holidays, had just a little bit of downtime to start playing around, started
working on some personal or professional projects with Opus 4.5 or Claude code or 5.2
codecs or some combination thereof, and realized that what we could do with agent coding
was much, much farther than they might have thought.
This was reinforced a couple of weeks later, when Anthropic dropped Claude code work,
which is sort of like Claude code for the rest of us, and revealed that it had been
written 100% by Claude code in just about 10 days.
Now if you want even more of a primer, I'd suggest one of my previous episodes, why
everybody is obsessed with Claude code, Claude code is Claude code for everybody else,
or most recently and probably most importantly, why code AGI is functional AGI, and it's
here.
So that's the setup, and we just keep getting evidence of how much things have shifted.
Cursor CEO Michael Truel posted about a week and a half ago, we built a browser with
GPT 5.2 in cursor, it ran uninterrupted for one week, it's 3 million plus lines of code
across thousands of files.
The rendering engine is from scratch and rust with HTML parsing, CSS cascade, layout, text
shaping paint, and a custom JSVM.
It kind of works, it still has issues and is of course very far from WebKit and Chromium
parody, but we were astonished that simple websites render quickly and largely correctly.
And to be clear, this was an experiment in autonomy.
While at first blush people thought it was one agent writing 3 million lines of code,
it wasn't, it was actually hundreds of concurrent agents.
Cursor wrote it up in a blog post called Scaling Long Running Autonomous Coding, and it's
very clear that Cursor is interested in pushing this frontier.
They wrote, we've been experimenting with running coding agents autonomously for weeks,
our goal is to understand how far we can push the frontier of agent decoding for projects
that typically take human teams months to complete.
And indeed, if you want to take a step back and just try to understand, psychologically,
where the Vanguard of AI and agent decoders are right now, it is really all about pushing
the boundaries on autonomy, breaking out in other words of being the bottleneck where
without your consistent prompting the AI isn't doing anything.
The leading agent decoders are in the midst of trying to build systems that work all the
time with extremely minimal input from them.
They want nothing less than armies of agents that work while they sleep.
And that army idea is operative.
In that same cursor blog they write, today's agents work well for focused tasks but are
slow for complex projects.
The natural next step is to run multiple agents in parallel, but figuring out how to coordinate
them is challenging.
Initially, Cursor gave their coding agents equal status, and as they put it, let themselves
coordinate through a shared file.
Each agent would check what others were doing, claim a task, and update its status.
Ultimately, however, this failed, the locking mechanism they implemented to prevent two
agents from grabbing the same task ended up becoming a bottleneck.
As they put it, 20 agents would slow down to the effective throughput of two or three
with most time spent waiting.
They tried a second strategy where agents could read state freely, but rights would fail
if the state had changed since they last read it.
In other words, they couldn't make different updates to the same code at the same time in
an attempt to avoid conflicts.
However, Cursor wrote this didn't work either.
Quote, as they put it, with no hierarchy agents became risk averse, they avoided difficult
tasks and made small safe changes instead.
No agent took responsibility for hard problems or end-to-end implementation.
This led to work churning for long periods of time without progress.
The next approach they took was to separate roles.
Instead of a flat structure, they created a pipeline where a subset of agents called
planners would continuously explore the codebase and create tasks, and workers would pick up
those tasks and focus entirely on completing them.
The workers they wrote don't coordinate with other workers or worry about the big picture.
They just grind on their assigned task until it's done, then push their changes.
At the end of each cycle, a judge agent determined whether to continue, then the next iteration
would start fresh.
This they said solved most of our coordination problems and let us scale to very large projects
without any single agent getting tunnel vision.
Now this is the point at which they instituted the ambitious goal of building a web browser
from scratch.
Now as we heard at the beginning this worked but not without a lot of challenges.
They write our current system works but were nowhere near optimal.
Planners should wake up when their tasks complete to plan the next step.
Agents occasionally run for far too long.
We still need periodic fresh starts to combat drift and tunnel vision.
With the core question, can we scale autonomous coding by throwing more agents at a problem
has a more optimistic answer than we expected?
Hundreds of agents can work together on a single codebase for weeks making real progress
on ambitious projects.
Now one of the things that struck me as interesting when I was reading this was the way
the day they described their planners and workers system.
Swicks shared this section of the blog post and nailed it when he wrote, cursor independently
invented the Ralph Wigham loop to solve the problems they were seeing with parallel agent
orchestration.
So this gets us to Ralph Wigham.
One of the weirder of these names even if the concept itself isn't overly complicated.
The concept was coined by developer Jeffrey Huntley actually all the way back last July.
He wrote a blog post called Ralph Wigham as a software engineer and as he put it in his
introductory blog post in its purest form Ralph is a bash loop.
So you might ask what the heck is a bash loop?
First of all bash is short for Born Again Shell which is a command line interpreter that
basically means it's the program sitting between a person in the operating system when
they're working through a terminal.
It reads the commands you type, it understands scripts, it runs programs, and it handles
things like variables and loops.
A bash loop then is the way to tell a bash shell do this thing over and over until I say
stop or until a condition is met.
It's a way to automate repetitive command line tasks instead of copy paste in commands.
Simplifying it even more, it's a written instruction that tells the computer to repeat
the same task over and over automatically.
So let's use some analogies that aren't about coding.
Imagine you leave a sticky note for an assistant that says for each folder on my desk open
it check what's inside then move on to the next one.
You didn't list every folder, you didn't do the work yourself, you just described the
pattern once.
That's an example of this type of loop.
Another analogy would be a checklist with a rule.
Instead of a bullet list that says rename file A, rename file B, rename file C, you say
rename every file in this folder the same way.
The key idea is that this type of loop tells the computer what to repeat and when to stop.
So the idea of Ralph as applied to AI coding was described by developer Ryan Carson and
a post on X.
He writes, everyone is raving about Ralph.
What is it?
Ralph is an autonomous AI coding loop that ships features while you sleep.
Each iteration is a fresh context window, memory persists via get history and text files.
Now he gets into exactly what this loop looks like from a technical perspective, but
the startup ideas podcast with Greg Eisenberg had Ryan on to explain it even more simply.
And here's how they summed it up.
Step one, write a detailed PRD.
That's a product requirements document, which is a document that defines the purpose features,
functionality and behavior of any new project or feature.
It's going to define why the product is being built, what success looks like, detailed
requirements of what it should do, things like that.
Now after you write that detailed PRD, you're going to convert it to extremely small discreet
atomic to use their words, user stories.
Step three is that for each of those atomic units who add clear acceptance criteria,
step four is looping your AI agent through each story.
In step five, it logs learning so it doesn't repeat mistakes.
Step six, the person who initiated the Ralph loop wakes up, tests it, and fixes the
edge cases.
Basically, the idea is to break down a complex project into very discreet, smaller units
that the coding agent can take on one by one, testing and looping until it's finished
and moving on to the next.
Now people are still experimenting with this and figuring out the limits of the methodology,
but the excitement on the other side is captured once again by Ryan in a post called
how to grow your startup while you sleep.
And that really is the thing that people are so excited about.
The idea of shifting to a paradigm where we got agents just working for us in the background,
meaningfully advancing the goals that we have.
And yet, over the last week and especially weekend, the discussion has shifted from Ralph
to something called Claudebot, where the corresponding interests believe it or not,
in Mac minis.
Viral memes abound like this one from Flavio, Mom, how do we get so rich?
Your father bought a Mac mini to run Claudebot in 2026.
So what the hell is Claudebot?
If you want to follow along at home, you can find this at CLAD.bot, which describes
Claudebot as the AI that actually does things.
Clear's your inbox, sends emails, manages your calendar, checks you in for flights, all
from WhatsApp, telegram, or any chat app you already use.
It's basically a system that allows people to turn Claude code into an actual personal
assistant.
A post on starryhope.com reads, at its core, Claudebot is an open-source AI agent that
runs on your own hardware.
Unlike Chatchuby T or Claude's web interfaces, which process everything on remote servers,
Claudebot operates locally with a gateway that connects AI models to the apps and services
you already use.
It can talk to you through WhatsApp, Telegram, Discord, Slack, Signal, and even iMessage.
But the real magic is what it can do once it's running.
Given the right permissions, Claudebot can browse the web, execute terminal commands, write
and run scripts, manage your email, check your calendar, and interact with any software
on your machine.
Perhaps the most compelling feature is that Claudebot is self-improving.
Tell it you want a new capability, and it can often write its own skill or plug-in
to make it happen.
One user wanted access to university course assignments.
He asked Claudebot to build a skill for it, Claudebot did and then started using it on
its own.
Now, some are a little skeptical.
Former Nvidia engineer, Boyan Tungu, said, I'm as excited as the next guy about the possibilities
of Claudebot running on a cluster of small local mini-computers, but 99% of all use cases
that I've seen so far concern the corporate BS jobs and tasks, summarizing email, posting
on Slack, adding meetings to a calendar that shouldn't exist at all.
This is not what has people excited though.
Matt Elias and responded, saying, yeah, those uses are a waste of its potential in my
opinion.
And that would know, because he went viral when he posted a picture of a Mac mini about
a week ago and said, hired my first employee today.
He followed up writing, yeah, this was 1000% worth it.
Separate Claude, that's the C-L-A-U-D-E version, plus Claude, the C-L-A-W-D, managing
Claude code and code accessions I can kick off anywhere, autonomously running tests
on my app and capturing errors through a sentry webhook then resolving them and opening
PRs.
Basically, Matt has this setup to be working around the clock on a new agent that he's
building to automate agency-level content creation.
On Saturday morning, Matt posted, nothing like waking up to a report from Claudebot about
everything that went wrong in my app yesterday and what it already did to fix it.
A couple hours later, Matt was still going.
He wrote,
Build a customer success in support workflow for Claudebot now too, analyzes transcripts
from the day, emails customers with bad experiences apologizing and asking for any other feedback,
adds their feedback to the daily report for our next morning brainstorm.
Basically, he's got a digital employee that lives in a Mac Mini, uses Claude code Opus
4.5 and Codex 5.2, and which he communicates with via Telegram.
This is the type of capability that has people so excited right now.
There were so many people in fact talking about putting Claudebot on Mac Minis that they
actually tweeted a PSA, you do not need to buy a Mac Mini to run Claudebot.
That dusty laptop in your closet works.
Your gaming PC you feel guilty about works, a $5 a month VPS works, a Raspberry Pi held
together with hope probably works.
Entrepreneur and investor Dave Morin wrote,
At this point, I don't even know what to call Claudebot.
After a few weeks in with it, this is the first time I felt like I am living in the future
since the launch of ChatGPT.
Now if all of this has your head spinning and it just seems technically inaccessible, you're
not alone.
Jasmine Sun actually wrote a post called Claude code psychosis that talks about some of
the ways that Claude code is still inaccessible for people.
It's a nice counterweight because you can sometimes feel insane for being intimidated for something
like the command line.
I think the accessibility of these programs is going to change really, really quickly though.
Not only do you have anthropic themselves releasing Claude code work, which while not
there yet is meant to be a new type of interface for non-coding Claude code tasks, there are
also other tools like Conductor that are replacing the terminal interface with a GUI.
Natalyze and accidentally caused some controversy on Dan Shipper and every's vibe code camp.
When he said the CLI is the Stone Age from two months ago, GUIs are back.
He followed it up and said, I did not realize how controversial this would be.
If you're still using Claude and Codex in the terminal you're missing out, you should
absolutely be in Conductor.
Other people agree.
Notions Brian Lovin said that on an average day he's spending 5% of his time in Figma,
15% in Cursor and Claude code, 20% in Ghosty and 60% in Conductor.
Lenny Ritchitsky asked his followers what the most underhyped AI tools were, and Conductor
came in second behind only Whisperflow, which is the one that I mentioned here a bunch
of times.
Speaking of vibe code camp, if you want to take everything I've talked about here and
really start to go deep, like I said, Dan Shipper and the team at every recently did
an eight hour live stream with tons of really great vibe coders talking about all the different
things that they do.
I'll include a link to the live stream as well as a summary app that someone built with
all the takeaways from all the different people.
Summing up really quickly, if you want to know in a very short statement how things are
shifting this year and how the most successful vibe coders are trying to evolve, it's all
about extending and expanding the autonomy of the agents that are doing the coding.
It's about removing themselves as a bottleneck and seeing how much can happen in the background
when they're doing other work or even when they're sleeping.
Anyways, hopefully now some of these terms don't seem quite so crazy and inaccessible.
I'm sure we'll continue to come back to them for now.
That is going to do it for today's AI Daily Brief.
Appreciate you listening or watching as always, and until next time, peace.
The AI Daily Brief: Artificial Intelligence News and Analysis
