Loading...
Loading...

Did Anthropic outship literally everyone else this week? 🚢
And chances are, you missed it.
Even though some of Google's new AI features this week will change how the world searches, Claude's new Computer Use could shape how the world works in the future.
(Or, whatever comes next that Computer Use paves the way for.)
Unless you spend countless hours each day tracking LLM changelogs, you for sure missed a handful of updates that might change everything about how you work.
Join us for Feature Fridays, as we quickly recap the new AI drops you can't miss.
Newsletter: Sign up for our free daily newsletter
More on this Episode: Episode Page
Join the discussion on LinkedIn: Thoughts on this? Join the convo on LinkedIn and connect with other AI leaders.
Upcoming Episodes: Check out the upcoming Everyday AI Livestream lineup
Website: YourEverydayAI.com
Email The Show: [email protected]
Connect with Jordan on LinkedIn
Topics Covered in This Episode:
Timestamps:
00:00 "Weekly AI Updates Overview"
03:29 "Google vs. Suno in AI Music"
07:15 "Custom Music for Creators"
11:09 "Copilot's New Productivity Features"
14:53 "Google Translate: Multilingual Audio Modes"
17:24 "Google Gemini Audio Advancements"
22:02 "ChatGPT's Future Super App"
25:36 "ChatGPT Voice Mode Insights"
27:40 "Genspark: Efficient Multimodal AI Agent"
33:11 AI Agents: Next Big Interface
35:38 "Agentic Work Revolution Explained"
37:27 "Everyday AI: Subscribe & Connect"
Keywords:
Anthropic updates, computer use tool, agentic layer, Google music AI, Lyria 3 Pro, Suno 5.5
Send Everyday AI and Jordan a text message. (We can't reply back unless you leave contact info)
Start Here ▶️
Not sure where to start when it comes to AI? Start with our Start Here Series. You can listen to the first drop -- Episode 691 -- or get free access to our Inner Cricle community and all episodes: StartHereSeries.com
Also, here's a link to the entire series on a Spotify playlist.
This is the Everyday AI Show, the everyday podcast where we simplify AI and bring its power to your fingertips.
Listen daily for practical advice to boost your career, business, and everyday life.
Each week there's usually a dozen or so smaller AI updates that slip under the radar, yet can completely change how we can interface with AI.
This week was no different. We recently started this new Friday Features series here on Everyday AI to better showcase smaller AI updates in fresh features from all the biggest AI companies.
But to be honest, this week we could have literally just focused on Anthropics updates.
They had nearly 10 noteworthy new AI features, though their biggest one, their new computer use tool, went absolutely viral.
And I think it's the first mainstream look at the future agentic layer.
And Anthropic wasn't the only company going completely crazy this week with useful AI updates as Google dropped new features across music, real-time voice AI,
and even translation that will go over all of those today.
Oh, and there's new AI powered ways to search the web, better ways to chat with your files on chat, and a new fairly capable live voice agent.
All of these new AI features waiting to potentially disrupt your old manual workflow.
Did you miss any of these? Yeah, chances are you did. Don't worry, stick with me for the next 20-ish minutes,
and you're going to be up to speed and the smartest person in AI in your department or company.
So here's what we're going to be going over on today's show and what you'll learn.
Well, first you'll learn why the new Gen Spark super agent might be worth paying closer attention to with some of its new real-time capabilities.
You're going to learn a fast and freeway to schedule work from your phone with one of the biggest names in the tech.
And you'll learn why Anthropic just showed us the future of AI even though right now it's a little bit buggy.
All right, you ready to get into it? Let's get featuring on this Friday.
Welcome to every day AI. My name is Jordan Wilson.
If you're new here, things for you, it's your daily live stream podcast, free daily news that are helping every day, business leaders like you and me,
keep up with these bevy of AI updates. I tell you what matters, what doesn't.
You take that information and grow your company and career.
So if that sounds like what you're trying to do, sweet starts here, but for the real cheat code, that's our website, your everyday AI.com.
There you can not only sign up for our free daily newsletter where we recap each day's podcast and everything else going on in the AI world.
But also you can go watch the video version of every single episode as well. So sometimes these, you know, Friday shows a little bit more visual.
It should be fine for our podcast audience. You should be fine even if you just out walking your dog or on the treadmill or whatever.
I'll keep you going and make sure to go check out today's newsletter as well.
All right, let's get into it live stream. Let me know. Can you see my screen? Hopefully you can. All right.
First, Luria 3 pro from Google, Google got super musical this week. And you know what?
It might be one of those instances where it's like, okay, are they competing with Suno here? Does Google even need to compete with Suno? Probably not.
But does the new Luria 3 pro get there? Not quite. You know, because also this week Suno released their newest version, Suno 5.5, which is extremely impressive.
But I think for the most part, now even if you were a casual AI music user, well, now you have access to the new Luria 3 pro if you have a paid Google account.
So even if you were like, ah, you know, this Suno looks great or you know, you do looks great.
But I don't want to pay for it. Well, now you just have access to it with Google's new model. So let's go over what's new.
The biggest thing with Luria 3 pro because Luria 3 has been out for a couple of weeks. But Luria 3 pro takes the quality and the duration up.
The biggest thing is previously Luria 3 only gave you 30 seconds, which is like, what's the use, right?
Yes, it just came out like five weeks ago, but you know, I tested it around. I'm like, what can you do with 30 seconds, right?
There's nothing, you know, unless you need an intro for a podcast or something, right?
30 seconds isn't doing too much. At least if you want to actually, you know, use the music in a meaningful way, but now you can go up to three minutes.
Some other big updates kind of under the hood. Google says that Luria 3 pro better understands song structure and you can prompt for intros versus choruses, bridges, etc.
So you have a little bit more granular control. At least via your prompting over how the kind of music turns out, you know, one thing I didn't see that I wanted to do some testing on that I did, which it actually worked pretty well.
You know, I said, okay, three minutes is great. But what if you want something longer, right? One thing I love, I love low five.
I always listen to low five when I do work, just helps me kind of space out.
And you know, all these AI music generators, they make probably better music at their core than these low five stations, but you know, I need like an hour of low five to really lock in.
And I'm like, okay, I wonder if I can upload a Luria 3 music piece of you know, an MP3 into Google Gemini and have it pick up from where it left off.
So Google didn't say this, but it's actually really good at that. So if you have like a very random use case, it actually did a good job, right?
I generated something with Luria 3 pro. I uploaded it and said, hey, can you essentially pick off where the end of this song left off, right?
So a lot of these, you know, I know that Google is great with anything multi like multi modality.
And they have kind of first frame last frame for video. So I'm like, let me do the equivalent of this for music and it actually worked pretty well.
So here's how you can access Luria 3. So if you do have a paid account on Google Gemini, you already have access and you can access it within the Gemini app.
So if you're on the pro, the $20 a month, well, this makes it easy. It's 20 songs a day. If you're on the ultra, you get 50 songs a day.
You can also access it via vertex AI. So if you are wanting to use this in production, you have it that route inside Google's AI studio via the Gemini API, Google vids and producer AI.
If you are a free Gemini user, yeah, you're not going to get access to it right now.
So here's what I think it's useful for and maybe who can find it valuable. I think it's useful. It just moves AI music from kind of these novelty clips to actually usable content, right?
You know, something I used to do a lot way back in the day, like 15 to 20 years ago, right? I made a lot of videos.
And sometimes you would just spend a lot of time looking for the right, you know, music to transition from one clip to another. So if you're a content creator needing royalty free custom music, this is great, right?
You can kind of dictate it with the lyrics that you want, all of those things, you know, and just the new structural awareness means that outputs just sound like real songs and not just random loops, right?
The 30 second clips weren't too helpful or useful in my opinion, but now with three minutes, they are. And like I said, use my little hack that I tried, let me know if it works well for you.
And then you can actually maybe even string something a little bit more together. Also, I think this is good for enterprise video teams that are using Google vids.
Google vids is actually pretty slept on. I think it's really good. And also just for developers, you know, building creative tools via the Gemini API.
Also, it is important to know that all outputs are watermarked with Google's synth ID. So you can tell that it is a I generated and Google says it is trained on licensed permissible permissible data from YouTube and Google partners.
All right. So let's get going to our next one. And this is from Microsoft co pilot. All right. So on Wednesdays, FYI, and maybe I'll quickly explain kind of our weekly rhythm here. And hopefully you guys are liking it. I'm kind of enjoying the flow myself.
But on Mondays, we bring you the AI news that matters. But generally with the way that AI has just kind of come to take over the enterprise and the business world.
A lot of times we're not talking a ton about new features unless they're big AI model updates. Right. So on Mondays, we go over the AI news that matters on Wednesdays, we go hands on an in depth with one thing we do live demos, you know, really getting under the hood.
And then Fridays here, it's kind of the in between. So on Wednesday, we did go over the new co pilot tasks, which I was very impressed with. But the new feature here for feature Friday is co pilot task rolling out to mobile.
So you will have to update, you know, on iOS, your co pilot app. And then you should be good to go. The good thing here as well is it's free. One thing that's confusing. And I went over this a little bit more in depth.
So let me see what what episode that was our co pilot tasks episode. If you want to go listen to that, it's 741. Just keep in mind, this is for co pilot on the web in the co pilot app. This isn't for Microsoft 365 co pilot. So not the enterprise version, although I do think and hope that they'll be rolling this out.
And I did, I have been chatting with the head of this project over at Microsoft. They've already shipped out some of the features or some of the bugs that I found. So you know, if there is something in co pilot task that you would want.
You know, let me know. I'll see if I can get the Microsoft team to cook it up. But a little bit more about co pilot task, you know, and this will explain maybe why it's super helpful on the web.
And I think we've seen a lot of this over the past couple of weeks, specifically with everything that in traffic has been shipping, right, which is just bringing kind of that full, you know, a genetic capability, but in a remote way, right.
So now you can take advantage of the, you know, everything that co pilot has to offer but via mobile, but specifically when it comes to a genetic orchestration, right, because that's really what co pilot tasks is.
So this is Microsoft's new agent feature that executes multi step tasks in the background using its own cloud computer and browser.
So it is a research preview. And like I said, it is free to use now. And the coolest thing that I think aside from yes, now co pilot task art in a mobile app, which is super helpful.
I like being able to text co pilot tasks, right, simple things, you know, hey, what are the most important emails I have right now, you know, you know, being able to go do some competitive research and create a deck for me, right, being able to text something like that for me is really cool and then being able to schedule these things as well.
So right now this is available in a research preview, all right, but it is free. And now with this rolling out to the mobile app, the co pilot app on iOS, I think it's really, really helpful.
So like I said, I did go into this a little bit more in the dedicated Microsoft tasks show, but you know why it's useful. I mean, come on, this is like a true agentic powerhouse. I was actually kind of shocked how good this is just because for me, you know, co pilot has never been something that I'm like the first to rush to, right, unless I really need to do something inside Excel inside PowerPoint for a certain reason, right, because then they do have some of the new integrations with both cloths.
And with the chat GPT that I think are really, really good, but for the most part, I'm not, you know, rushing to use co pilot as number one or number two in my stack, if I'm being honest, but with this one, I am just because it did a really good job and it was fast, right, and go check out that entire episode, right, 741.
So one thing I really liked about co pilot tasks is you can even edit the slides that it creates, right, and the same way that, you know, Google has kind of an annotate feature in some of their different products, you can literally, it can go and do a bunch of research for you on its own.
And so, you know, you can schedule something every day or, you know, once a week, right, and connect it to your data as well, that's the other big thing, right, so connecting it to, you know, your Google drive or your Gmail or your, you know, your outlook email, right, if there's something that you do routinely every week that lives inside of those apps, right, you can schedule it, it can grab that, it can create you new documents, right, you can browse the web, it has access to its own browser and now it's also accessible via the app.
It's a pretty big one there that I think a lot of people are going to find useful.
AI moves too fast to follow, but you're expected to keep up. Otherwise, your career or company might lag behind while AI native competitors leap ahead, but you don't have 10 hours a day to understand it all.
That's what I do for you, but after 700 plus episodes of every day AI, the most common questions I get is, where do I start?
That's why we created the start here series, an ongoing podcast series of more than a dozen episodes you can listen to in order. It covers the AI basics for beginners and sharpens the skills of AI champions pushing their companies forward.
In the ongoing series, we explain complex trends in simple language that you can turn into action. There's three ways to jump in.
Number one, go scroll back to the first one in episode 691.
Number two, tap the link in your show notes at any time for the start here series, or you can just go to start here series dot com, which also gives you free access to our inner circle community, where you can connect with other business leaders doing the same.
The start here series will slow down the pace of AI so you can get ahead.
All right, next. This is a smaller one, but I think this is one where technically, like billions of people are going to find, I think, just immense value out of this.
This is because Google translate live is finally available in headphones. All right, stick with me. I know this sounds small. So essentially this has been a feature that had previously been rolled out to Android, but is now available on the Google or sorry inside the Google translate app on iOS.
And here's why I think this is really important. Well, first just kind of let me tell you what it does. So the new live translate with headphones feature is available on iOS and essentially any headphones that you have, you know, whether they're wired or Bluetooth right now they can kind of be a real time personal interpreter.
Right. So some of the newer, you know, AirPods as an example have this feature kind of built in, but they're pretty expensive, right. So now it's literally any pair of headphones. If you're using Google translate can translate live. Right. So this is I think a, you know, one of those things. Yes, there's plenty of business use cases. Obviously, right. If you do any international business, if you're traveling internationally, I think this is really going to help, especially if you don't speak the language.
You know, obviously on the personal side, this is huge. Right. And so this is powered by Google Gemini's speech to speech translation. And it also preserves tone cadence and emphasis of the speaker. And the best thing is, well, it's free. Right. And it does work with any headphones. You don't have to have, you know, any AI powered headphones, any of the new Google or Apple headphones literally anything as long as you have access to data on your phone.
And you have the latest version of Google translate. That's all that is needed. So there's three different modes. So there's a listening mode. That's real time translation in your ears.
There's conversation. So you can hear the translation and the headphones and others hear it out loud. And then there's text only that is just the on screen text translation.
So right now it supports more than 70 languages. And well, there's no more, you know, I've had to do this before in certain instances. There's no more kind of passing a phone, you know, back and forth. You can just wear the headphones now and listen.
So obviously, if you're visiting different countries, this is super helpful, you know, families with members that speak different languages or just business professionals in multi-lingual meetings. This one, I think, is going to be very helpful.
All right, our next new feature. This one also from Google. You might be saying, wait, Gemini 3.1 Flash isn't this an old model? Well, yes, but no, because it's brand new because this is now Gemini 3.1 Flash live.
So here is what's new and this does change the way that you search. So just stick around for that for a second here. So this was just released yesterday. This is Google's new audio to audio model.
And it powers the new Gemini live experience and search live. All right, and search live. So this is just a way to essentially, well, it's the way you search the web with live video.
Right. So this is also expanding globally to 200 plus countries. And this was previously in US only. So here's who has access. Well, literally anyone.
Anyone can, if you're using AI mode in Google search, you will be able to benefit from this. Obviously developers, you know, you can use this on the Gemini live API or in Google AI studio or enterprises, you know, in the Gemini enterprise for customer experience.
So here's why it's useful. And I think who will find it valuable? Well, I think it's just follows conversation way better, according to Google, right. And it also holds context for two times as long as the previous model.
So it's better at filtering it filtering out background noise, right traffic TV, also being able to recognize kind of tonal nuances. So it can adjust responses when users sound frustrated or confused right now 90 plus languages.
So this is obviously developers. If you're building voice agents, customer service agents, this is huge. Right. So the flash version. Obviously a little cheaper than the normal Gemini 31 pro.
You know, which is multi motor by default. So this is specifically made for audio to audio. And I think that where this is really going to get popular aside from being able to use it on the web is just with developers being able to create voice agents. I think this is one area where Google Gemini models are really good.
Also, I think enterprise contact centers are going to find a lot of utility out of this and just international users who couldn't use search live, right. This is only available in the US.
But now anyone can use it. I think there's a little kind of video demo might have been on a different page, right. But if you haven't used search live before it is literally right if you see something out in the world and you're like, what the heck is this.
Well, it's a live version of search. So you can click the live button. And it will see your camera and you can talk to it as well. So if you see something strange or if there's a sign again kind of, you know, going back to the multilingual aspect, right.
You're traveling and you're like, okay, like am I in the right spot. You know, it will be able to see your camera. You can talk to it, right. But this is where it's really helpful. It's just this faster inference, the faster model. And, you know, being able to, you know, work across different languages is obviously helpful as well.
So right now, also anything that's generated does have the synth ID, just FYI and developers, if you are migrating from 2.5 flash, this does use the thinking or make sure to use the thinking level instead of thinking budget.
Some things for developers, it had changed a little bit under the hood. So pretty impressive model. So here's kind of what Google said about it on their announcement post. They said today we're advancing a Gemini's real time dialogue capabilities with Gemini 3.1 flash live our highest quality audio and voice model. Yet it delivers this speed and natural rhythm needed for the next generation of voice first AI offering a more intuitive experience for developers enterprises and everyday users.
And you can access it in the Gemini live API Google AI studio or the, like I said, the Gemini Enterprise for customer experience and it's available in search live in Gemini live.
So another thing, right, the Gemini live aspect is helpful as well. Right. So if you are someone that inside, you know, Gemini as an example, or you're someone that loves to use, you know, Chatchy BT voice mode, you know, make sure to check out this new version inside Gemini with Gemini live.
All right. Next, this one's going to seem small, but Chatchy BT has 900 million weekly active users. This is a small thing that a lot of people are going to find helpful, but you probably didn't see anything about it because it's not even really technically a feature. It's just a new way that open a eyes essentially storing files inside Chatchy BT, but it's important. So stick around.
So stick around. So the Chatchy BT library has officially rolled out inside of Chatchy BT. This is a new persistent file storage hub in Chatchy BT's sidebar. So it automatically saves every file you upload or that Chatchy BT creates. So that's documents, spreadsheets, presentations, etc.
And here's the cool thing. Files persist across conversations until you manually delete them. So who has access to this right now? Well, sorry, it's not actually 900 million, but 900 million people have the ability to go use this if you want to because it is only available right now to paint subscribers.
So you do have to be on the Chatchy BT plus pro or business plan. It's not available yet on the Chatchy BT free or go plans. Although I do feel that eventually open AI will roll this out because on the free plan, you're right. I think that's where they're really going to benefit from being able to serve better ads.
And if you can, well, access people's uploaded files, you can serve them way better at. So my guess is eventually this will roll out to free consumers when we get the new super app from Chatchy BT, right. If you listen to this show yesterday episode 742 going over open AI's most chaotic week ever.
You talk about that open AI is eventually going to be rolling out a kind of single app to rule them all right, they got rid of Sora, but they're essentially going to be rolling their Chatchy BT codecs and atlas all in one app.
You see this new library feature being very useful in the future super app, but I mean when it comes to just utility on how you can use it today, I mean here's here's the biggest thing.
You can ask Chatchy BT about any files that you've saved right and that's the biggest thing and that's something I'm going to immediately find helpful.
There's a lot of different ways that you can organize your files, right, you can use projects, you can use GPTs, you can upload a file to an individual chat to an individual conversation, right.
But sometimes I'm like even thinking in my head, especially when I'm on the go, I'm like, oh, you know, I know I had chat GPT go, you know, I'm usually using GPT 54 pro for this, I know I had to do a deep dive on, you know, this big spreadsheet.
But where did I save that was that in the project was it in a GPT was it in an individual chat.
Now you just if you've uploaded the file, if you know what it's called or if you have an idea, it doesn't matter, you can just ask about it.
So this is great, especially for people like me who try their hardest to stay very organized in Chatchy BT, but sometimes you don't for whatever reason.
This is this is pretty big, I think, and I think this eliminates one of the biggest friction points of chat GPT, which is uploading the same files across different conversations or just losing place, right.
So for these artifacts that chat GPT does produce that's big as well. So yes, this is helpful for the files that you upload, but also what if, right, people, right, you need to be using chat GPT for this y'all.
You need to be using chat GPT to make you spreadsheets to make you PowerPoint presentations, right.
So now when chat GPT makes you documents, it all goes in the library as well, and you can reference those from anywhere.
So again, this might be one of the smallest updates on paper, but if you are a power chat GPT user, this is actually a pretty big update.
All right, moving on next. So Jen Spark, one of the super agents, right, obviously a growing category, you know, since open claw has really popularized the, you know, super agent or the, you know, all in one agent.
But Jen Spark in Manus, as an example, had been around for a while. And I think that Jen Spark just got a new update that a lot of people are going to like, right. So when we talk about real time voice and access to your data.
That's huge, right. That's one thing for me personally.
I don't know, chat GPT's voice mode hasn't really done it. I do have to give the new version of the Gemini 31, a little bit more testing since it literally just came out.
But I might actually, I don't even know if I have a paid Jen Spark account. It might be one of the only platforms I don't have a paid account for. I pay for literally all of them.
But this one, you know, they just released a demo video this week, essentially, you know, this woman who, you know, was using Jen Spark real time voice on her drive, you know, she's having it check her calendar, move things around, send emails, you know, check her, you know, her teams, right.
So anything that Jen Spark connects to, at least when this works. And if it does work correctly, again, I think I've tried all these other features on today's show except this one FYI.
So I don't have any personal experience with this one, but in the demos that I've looked at, it looks pretty good and pretty helpful. But, you know, the true, I guess feature here.
I think this could shine in areas, right. I think Chatchy PT voice mode is probably the ones that, you know, most of us might use or maybe, you know, many of us are just still kind of disappointed with the state of voice AI, although I do think it's getting better and better.
But one thing that I've always realized is sometimes at least Chatchy PT's voice mode, you know, you speak called advanced voice mode. It's so running a super old model.
It's running a version. I believe still of the GPT 40 series. So not the smartest model. And in my testing, I've done a lot of testing in the past.
It usually had problem accessing your files. So that's with Jen Spark real time voice. I do think that's helpful in being able to access the different connectors if you did give it access to.
So this isn't just transcription, right. This is voice commands that you can trigger, you know, multi step agent workflows across your different apps, you know, so you can just speak tasks and get them executed, manage your calendar, check your email, right.
This is one thing. When I saw this demo, I'm like, oh, wait, I might actually really benefit from this. So it is also kind of available via their speakly app. So Jen spark has kind of a dictation or, you know, real time agent voice app called speakly, but it does connect via Jen spark as well.
So you do get a some use case on the premium model, but not a whole lot. But if you are on the plus account or the pro accounts, I think you're going to get a lot more access out of this.
So here's why it's useful. Well, you can just speak a complex task and Jen sparks multi modal agent can break it down and execute it, right. And it claims four times more efficiency over keyboard inputs for complexity.
For complex tasks. And like I said, it does connect to the same agent ecosystem that handles, you know, your research email contents and even phone calls.
So I do think who's going to find this useful if you're someone that's on the go, right. Give it out. Give it a try on the on the on the freemium plan. But if you're someone that's on the go, it's going to be helpful.
If you've tried, you know, other kind of AI apps or AI agents that connect to your, you know, all these other tools that you use your Slack, your calendar, your email, and it struggles for whatever reason.
I think this is at least I'm not saying this is going to, you know, be the tool that solves your problem, but it's at least a contender that could potentially help.
All right. And last but definitely not least, this is the one that went mega viral. Yes. Anthropic released so many new features this week. All right.
Some of that didn't even make our list, right. They have the new work tools that are available on mobile. They have the new claw to code channels, right. A very open claw feature being able to talk to cloud code via telegram discord. I message.
They have their new co work projects. Like I said, Anthropic literally released like 10 pretty big updates this week, but the biggest one by far. And I'll say this.
This might be the most like viral AI update ever. That wasn't a new big model. Right. That's because at least by Twitter vanity metrics. This thing got like over 75 million views. Everyone had eyeballs on this. So this isn't one you missed.
All right. But this is their new computer use, right. Unfortunately, they didn't really give this a good name because at first they were kind of tying it to dispatch, which is the new way to kind of control the desktop version of clawed co work with your phone, right.
Dispatch. We covered this on the show last week. Really cool. Right. It's a way that you can now use your phone, right. The iOS app inside clawed.
And you can control your desktop computer via dispatch, right. So before the computer use, you couldn't actually control the computer, but with computer use, you can.
I'm talking about mouse clicks, you know, keyboard inputs, not just, you know, a web browsing agent. So this is completely different, right. What we've seen.
A lot with the, the current kind of a genetic tool use stack. It's virtual browsers, right.
Like a lot of these, you know, chat GPT agent mode as an example, it has a virtual machine and it works in your browser.
Computer use literally is computer use clawed can see and used your computer, right. I did a lot of testing on this. And you know, let me know.
Let me know if we should do this for Wednesday's show. I'm thinking maybe, all right. But if you're listening on the podcast or on the live stream, just say computer use, all right. I always have an arbitrary random number in my head. And if we hit that, I'm like, yeah, we'll do it.
So I did do some testing on it. It's super helpful. I mean, here's one of the reasons. So we did technically already have a lot of this functionality, whether it was via co work.
You know, being able to run certain things in the terminal. But now, like some of the testing I did, well, I was just having it launch and use other apps, right. I was kind of having fun with that.
So with the new computer use, I had, you know, clawed launch Atlas and you know, that's open a eyes, a genetic browser. And I was having it do some, I think I was having it use perplexity inside of Atlas and stuff like that.
Right. So I think there's some huge benefits. But first, let me just tell you a little bit more kind of about how it works. So it can now control your entire max. So that means open apps, navigate browsers, click buttons, fill spreadsheets, anything that requires a mouse click, right, which is something that even there.
Chrome extension would sometimes struggle with, right, because it was using screenshots and computer vision. So now this is a completely different ballgame. This is a new kind of layer.
So it does use connectors first FYI. So if you are asking something about slack as an example, it's first going to check a slack connector if you have it. But then it'll fall back to screen control.
If a connector doesn't exist, or if it can't find something in that connector. So right now it works in both clawed co work, but also clawed code as well.
So this is confusing because it says on their kind of blog post here, it says let clawed use your computer and co work. And then originally they're just tying it to dispatch, but you don't need to use dispatch to use it.
I'm actually kind of quite shocked. Maybe anthropic didn't understand how quite viral this would go to give it a proper name. So I think at first people were calling it computer computer control via dispatch. And then they were saying it's co work computer control. And I think now most people are just calling it computer use.
So yeah, this is really, really helpful. So you do have to be on a paid plan right now. And it is only available for Mac. And it is a research preview. It is buggy in my limited experience so far. I've been using it for a couple of days since it was announced. But when it works, right when it works going back to how I started the show.
I think this is the not the end interface of AI agents, but I think this is the next big interface of AI agents, because I think eventually, you know, especially like voice commanding agents, they're not going to need our computer per se.
But I think we may, you know, need a couple of years until agents just have kind of stronger and more reliable protocols, you know, versus just into a MCPs, right.
Eventually, I think agents will just be able to be able to nicely trade site everything and they won't need to necessarily use our desktop as an operating layer. But for now, I think 100%.
And the other thing, if this is used in enterprise, and that's a big if right, because obviously there's a lot of new dangers with this on the prompt injection side, right. But this could be something very useful, right for companies that are using old legacy software, right, that just doesn't work well with any kind of AI. Well, now you can just keep this open, give it a pretty difficult task and it can technically operate that legacy, that legacy software.
So who else is going to find value, why else is it useful, right. So like I said, this can still pair with dispatch. So that's the cool thing, you can literally control your entire desktop computer from the clawed app on iOS, which is really cool.
And I think this kind of fills the long tail, right. It covers the apps that will never have a dedicated connector. And you can, you know, if you're running late to something, right, I was doing this a lot. I was just out on, I don't know walk.
Right. And I'm like, oh, you know, for me, it's not like I needed something, but I was just kind of testing it in that scenario. Oh, like, I know I downloaded this file on my computer. I'm not sure where it is, right.
Like I said, a lot of these things were available before, if you had things set up correctly via co work, and it could access certain elements of your computer, but now it can literally control anything on your computer as well.
So I mean, just who's going to find this valuable, if all of us, right, literally anyone, if you're on Mac right now, if you have an iPhone, this is huge.
So now much workers who are just juggling a lot of apps, especially those apps that don't have dedicated integrations. That's the big thing. Yeah, you can use this to showcase certain things and it's fun in the school, but if there's a dedicated connector integration, you're just wasting your time by doing it this way, because it is still just using computer vision and clicking around.
Now, but for all of those things that don't have dedicated apps within quad, this is huge, right. So anyone that just wants to delegate desk work from their phone, while they're on the go, or just power users who are already using co work or dispatch.
Like I said, this, I think, signifies the next step in a genetic work. So that's a wrap for today's show. I hope this was helpful. Are you liking the new Friday features? It's a, well, I say it's a quicker show, but I geeked out and accidentally talked for 36 minutes, but hopefully this is a good way.
You know, it's so hard to keep up with everything's going on. Yeah, we do the AI new show on Monday, but like I said, a lot of that starts to see over into just big enterprise politics, society, culture, all these other things. So I think Fridays are shows.
And let me know if you find it helpful that you know are like, hey, you probably missed this. Here's why it's helpful. Here's how to go find it useful and to get you going on your way. So if this was helpful, if you are on the podcast, please do me a favor.
Leave us a rating if you could, right? If this is helpful, go tell other people about it, right? Subscribe to the podcast. If you're listening on Apple podcast, Spotify, we'd really appreciate that. Then go to your everyday AI.com. Sign up for the free daily newsletter. Thank you for tuning in. We hope to see you back next week and every day for more everyday AI. Thanks.
I'll see you next time.

Everyday AI Podcast – An AI and ChatGPT Podcast

Everyday AI Podcast – An AI and ChatGPT Podcast

Everyday AI Podcast – An AI and ChatGPT Podcast
