Everyday AI Podcast – An AI and ChatGPT Podcast

Ep 753: Anthropic Goes Full OpenClaw, Meta Muse Spark Drops, Google Gets Notebooks and More . 7 New AI Features You can’t afford To Miss

40 min
Apr 10, 20269 days ago
Listen to Episode
Summary

This episode covers seven major AI feature releases from Anthropic, Meta, Google, Microsoft, and Zapier, with emphasis on Zapier's new SDK enabling AI agents to access 9,000+ apps, Anthropic's managed agents platform, and Google's notebook integration in Gemini. The host highlights practical applications for business professionals while noting that infrastructure-heavy features may see slower adoption than consumer-facing tools.

Insights
  • Zapier's SDK represents a major breakthrough for AI agents by eliminating authentication friction across thousands of applications, making complex multi-app workflows accessible to non-technical users through natural language
  • Meta's Muse Spark demonstrates competitive capability despite years of silence, but ranks below top-tier models (Gemini 3.1 Pro, Claude Opus, GPT-4o), suggesting the AI model hierarchy remains stable despite new entrants
  • Google's notebook integration strategy creates a unique ecosystem advantage by syncing Gemini conversations to Notebook LM, potentially increasing user lock-in through seamless cross-product workflows
  • Infrastructure-focused releases (Anthropic managed agents, Microsoft foundry models) may see slower adoption than front-end consumer tools, as users prefer integrated experiences over backend complexity
  • Free tier access to advanced capabilities (Google Vids with Veo 3.1, Muse Spark on Meta.ai) signals a shift toward freemium models to drive adoption before monetization
Trends
AI agents moving from isolated tools to integrated ecosystem players with multi-app connectivityFreemium access to advanced AI models becoming standard competitive strategy across major platformsPersistent memory and knowledge management becoming core agent features rather than optional add-onsMultimodal reasoning (text, image, video, audio) consolidating into single models rather than specialized toolsEnterprise infrastructure abstraction: managed hosting and scaling becoming platform-provided rather than developer-builtCross-platform synchronization (Gemini-Notebook LM, Zapier SDK) creating ecosystem lock-in strategiesOpen-source alternatives (Open Claw) matching proprietary feature parity within weeks of commercial releasesVideo generation commoditization: Veo 3.1 free tier signals maturation of text-to-video technologyNatural language infrastructure setup reducing technical barriers for non-developers to deploy agentsCompetitive model consolidation: top-tier performance clustering around 4-5 dominant models (GPT-4o, Claude Opus, Gemini 3.1 Pro, Muse Spark)
Companies
Zapier
Launched SDK enabling AI agents to access 9,000+ apps and 30,000 actions without OAuth management; host considers it ...
Anthropic
Released managed agents platform for hosting and scaling Claude-powered agents; added video/music generation, persist...
Google
Integrated Notebook LM into Gemini with notebook feature syncing across platforms; added free Veo 3.1 video generatio...
Meta
Launched Muse Spark, first model from Meta Superintelligence Lab led by Alexander Wang; natively multimodal with text...
Microsoft
Released MAI Transcribe One, MAI Voice One, and MAI Image Two models; Transcribe One has lowest word error rate (3.8%...
Open Claw
Added built-in video/music generation, persistent memory wiki, experimental dreaming mode; most popular open-source s...
OpenAI
Mentioned for GPT-4o ranking in top-tier models; host notes rumors of upcoming image model; compared against for agen...
Claude
Anthropic's AI model used as default in managed agents; supports coding agents (Claude Code) with growing MCP integra...
Runway
Video generation provider integrated into Open Claw for text-to-video capabilities
Suno
Music generation provider integrated into Open Claw for text-to-music capabilities
Scale AI
Acquired by Meta for $14 billion; CEO Alexander Wang now leads Meta Superintelligence Lab developing Muse Spark
Obsidian
Note-taking app with compatible format supported by Open Claw's new memory wiki feature
Cursor
AI coding agent with programmatic access to Zapier SDK's 9,000+ apps via new integration
Codex
OpenAI's coding agent with access to Zapier SDK's full ecosystem of apps and actions
Notebook LM
Google's grounded research tool with near-zero hallucination rate; now syncs with Gemini notebooks for unified workflow
People
Jordan Wilson
Hosts daily AI podcast; introduced new Friday Features segment covering 7 AI features weekly
Wade Foster
Called Zapier SDK 'the most powerful thing they've launched in years' in video announcement
Alexander Wang
Former Scale AI CEO now leads Meta's new superintelligence lab developing Muse Spark model
Mustafa Suleyman
Leads Microsoft's new in-house AI team that released MAI Transcribe One, Voice One, and Image Two models
Quotes
"Zapier opened its SDK to everyone, including giving coding agents like cursor Claude in, uh, Claude code in codex programmatic access to Zapier's full ecosystem of 9,000 apps, uh, 30,000 actions and raw API access to 3000 apps."
Jordan Wilson~5:00
"This one I think is pretty big. So who has access right now? I actually know, I will point out, I watched the video, uh, with Wade Foster, their CEO, and he called it the most powerful thing that they've launched in years, uh, which I think was pretty telling anyways."
Jordan Wilson~8:30
"Anthropic is crushing it in 2026. They are winning the year."
Jordan Wilson~45:00
"I've always said, you know, when you're talking about buying your building, right, I'm like, well, you should technically be buying and using these systems on the front end because the front ends are becoming more and more powerful."
Jordan Wilson~50:00
"If you're not using notebook, LM literally every single day, I don't understand why not, unless you're not able to, because of your job, but I mean, use it for your personal life."
Jordan Wilson~28:00
Full Transcript
This is the Everyday AI Show, the everyday podcast where we simplify AI and bring its power to your fingertips. Listen daily for practical advice to boost your career, business, and everyday life. Anthropic continues to add new features that look and function kind of like Open Claw, the same platform they pretty much shut out last week. Google just dropped a notebook feature that's kind of like notebook LM, but kind of not at all. But it also adds new features to both platforms. We got new models from Meta Microsoft. We got AI video updates and a ton more. But the most meaningful updated AI feature that you may have missed this week comes from a company we haven't talked about in a while, Zapier. And I think that one will be big. All right, let's get into it. If you're new here, welcome to Everyday AI. But on today's show, if you stick with me for the next 20-ish minutes, here's what you're going to learn on our new Friday Features segment. You're going to learn how Anthropic is going full on Open Claw after banning its models from being used by Open Claw. If Meta's new Muse Spark model is worth the billions they've invested in, you'll learn what Google's new notebook feature in Gemini unlocks. And last but not least, you'll see why I think Zapier's new SDK might be your favorite tool, even if you're not technical. All right, welcome to Everyday AI. My name's Jordan Wilson. If you're new here, we do this every day. This is your unedited, unscripted guide to keeping up and getting ahead with AI. So if you haven't already, please make sure to go to youreverydayai.com. Sign up for the free daily newsletter. We're going to be recapping today's show and all the other AI news that you need to know to stay ahead. All right, so there's so much that's happening in the world of AI. Y'all, I have a daily AI podcast from Monday to Friday, at least. And I realized after like 700 some episodes that still 80% of all the actual useful stuff that we were all using, I wasn't covering. So that's what this new Friday features is all about. So on Mondays, we do the AI news that matters. On Wednesday, we go hands on, you know, with a new model or release. Last week, we did the Earth this week. Wow, 10 flies. This week, we did Google Gemma for the local model. So make sure to go check that out. And then on Fridays, well, we're going to fill in the gap in between because I realized that between the Monday AI news and the Wednesday deep dive, most of the things that we talk about aren't getting covered. And then on Tuesday and Thursday, we do different shows, rotate, sometimes interviews. All right, but let's get into it. There's a ton to cover this week. So, uh, live stream audience of bringing up my, uh, my screen here podcast audience. You can always see the video version at your everyday AI.com. Go click on the episodes, not only today's, but literally, uh, you know, 750 plus, uh, videos, podcasts, you can read it. It's all in the newsletter for free. It's a free generative AI university. All right, first, let's start with the one that I think is on the one that was, this is honestly probably the one I was most excited for, uh, even bigger than the anthropic agents one that we're going to get to in a little bit. But first, let's talk about the Zapier SDK. Uh, so here's the way that Zapier explain it. They say, let your agent connect to anything authenticated, govern access to the full Zapier catalog in code on behalf of your users. No OAuth flows, no token management. Zap Zapier handles the keys. So here's what that means. Zapier opened its SDK to everyone, including giving coding agents like cursor Claude in, uh, Claude code in codex programmatic access to Zapier's full ecosystem of 9,000 apps, uh, 30,000 actions and raw API access to 3000 apps. So here's the good thing. If that, if you heard that and you're like, well, what the heck does that even mean? Uh, you know, OAuths and raw API, you're like, wait, what? This isn't the show for non-technical business leaders. Yes, it is. Uh, so even if you don't know what most of that means, well, you can just use natural language to set it all up. So you will need to use a coding agent to obviously tap into this SDK. Uh, so that's just a software development kit, which used to be kind of technical, but in the age of AI and large language models and natural language processing, it's not actually very hard at all. But what this means is, let's say as an example, if you use Claude code, or if you use a open AI's codex, right? Uh, up until, well, before this Zapier, uh, kind of announcement, uh, what you could connect to was extremely limited. Uh, right. Obviously, um, Anthropics Claude code has a growing list of MCPs, uh, with codex. They do have some plugins, not a ton, but now essentially you have Zapier. So you have everything, right? So your agent or different agents that you use can now access your, everything from your Gmail to your, uh, CRM to Slack, to your project management tool, like whatever you're working on. So, uh, I haven't started to set this one up yet because it literally just came out like, like 48 hours ago. Um, but this one I think is pretty big. So who has access right now? I actually know, I will point out, I watched the video, uh, with Wade Foster, their CEO, and he called it the most powerful thing that they've launched in years, uh, which I think was pretty telling anyways. So this is an open beta right now and it's free during early access with no billing charges. So, uh, that's pretty cool. And then enterprise and teams plans, uh, are off by default and they do require manual opt in, uh, or contacting Zapier. So if you have a personal plan, this will be a little bit easier, uh, to enact right away. So here's why it's useful. It just eliminates the single biggest friction point for AI agents and that's managing all the authentication, uh, right? Across, you know, thousands of apps. You're probably not going to be using this across thousands of apps, but you'll be using it maybe across a dozen or dozens of apps, right? It's difficult to, uh, manage all of that, but now Zapier just kind of does that. And that's where, uh, Zapier's MCP gives agents a curated menu of pre-built actions and then the SDK lets agents write loops, handle edge cases and chain complex logic across apps. And you're existing. The cool thing is your existing Zapier connections work immediately. So you don't have to re-auth anything. So, uh, once you do authenticate, uh, the Zapier SDK in your coding Asian of choice, anything that you've previously connected, it works right away. Right. So I'm like, even as I'm saying this, I'm like, Oh frick, I should've prioritized this over other things because, you know, with Zapier, you know, I have our WordPress website set in there. Uh, I have our be hide email newsletter. I have our, our circle community, right? It's all there. So now I can instantly, uh, start working with those things inside of Claude code and inside of codex are the two that I use the most. Uh, so here's who I think it's going to find it valuable. Well, developers building AI agents, that's going to be extremely valuable. You know, automation experts. So if you are someone or, you know, anyone in marketing that has used, uh, Zapier for years, you're going to find this extremely helpful. I also think solopreneurs, entrepreneurs who maybe don't have the time, uh, to usually do all of these things. Well, now it made it a lot easier, uh, with this from Zapier. All right, let's go to our next one. Uh, doing some a little different open claw updates. Right. And Hey, let me know live stream audience, um, podcasts, uh, you know, if you're listening on Spotify, if I should make the Friday feature, right? If we're going to do this permanently. Should I just always like, we've been doing essentially seven things. I think that's easy, right? It's not too many. It's enough, but should one of them be open claw every week? Um, if so, right? Cause open claw ships updates like every day. Um, so let me know, just say open claw in the comments. Uh, you think I have a number, right? I actually do this. I always have a number and I say between LinkedIn and Spotify. If we get enough comments on this, uh, we'll do it. If not, we won't. So if you want to always include, uh, the weekly open claw update, let me know, but I will do it this week regardless, because I think it's pretty big because now in open claw, we got to build in video generation, built in music generation. We got some new memory features and an experimental kind of dreaming mode. All right. So well, if you don't know open claw at all, uh, I don't know, you must literally be a lobster sleeping under the sand somewhere. So open claw is the world's technically the most popular piece of open source software ever, um, is an autonomous agent that acts on your behalf. Right. Uh, you can talk with it, communicate with it via various channels. Uh, but let's get into the actual new update. So, uh, we got three major releases. Uh, so one is built in video and music generation. So these are obviously all things that you pay for, right? You have to, uh, connect, you know, your API keys unless you're using an open model. So if you do have a, you know, super fat, like max studio, like I do, you'll be able to do some of these things for free or your agent will be able to do them for free. Um, but otherwise you will be using API. So you'll be paying for your usage. Keep that in mind. Uh, so the new things we got the, um, built in video and music generation just from text, right, video, music and editing tools, and then the persistent memory, uh, kind of like a wiki knowledge system. So the video generation, uh, the agent can create videos and music tracks directly using configured providers like runway, uh, Google's Lyria, et cetera, without even leaving the chat, uh, then music and video editing agents can trim, mix, splice and refine existing video. That's really cool. And music files through natural language commands. So, uh, that's pretty sweet, right? Turning open claw into a creative editing assistant. Uh, you know, this, this is the first versions, right? I'm not going to try that one, uh, you know, until it's gone through some updates. You know, some of these things also keep this in mind. Uh, the open claw team is amazing. Right. There's a reason why it's literally the most popular piece of open source software, software ever. Uh, but for some of these things, uh, when they first are released, they're usually rough around the edges. That's how it goes. Right. And then the open source community, you know, finds fixes, uh, you know, they ship them out pretty fast, but you know, if, if, if you need something like this for production, for like a work project, don't think that you're going to get anything, you know, usable or great, but it is important to know what they're working on and what they're shipping. All right. So that is the, uh, the music and the, uh, video editing. And then there's the memory wiki. So, uh, this kind of replaces the old fuzzy recall system with a more structured, uh, kind of persistent knowledge base that works a little bit more. Like Wikipedia, uh, and it does support obsidian compatible X for. So if you do use the obsidian app, uh, which I know a lot of people do, uh, that will be helpful as well. And then last but not least, there's a new dreaming ex fear, uh, experimental mode. So this is opt in. And this is essentially now where the Asian processes conversations during idle time, uh, through light, deep in REM style, uh, phases to consolidate short-term memories into permanent knowledge. Uh, so pretty, pretty cool. Uh, then it kind of has a diary timeline user interface. So obviously who has access to these? Well, everyone, right? So as long as you're, uh, updated your open claw to the latest version, you will have all of these, uh, features. And, uh, the cool thing is obviously they, you know, added GBT five four support. Um, recently Gemini clawed only via the API because obviously in throbbing, uh, cut off access, uh, to using your, um, clawed paint paid plan, uh, via open claw, which we talked about in the show last week. So I mean, here's why it's useful and who will find it valuable. Well, uh, I don't think some of these things are going to be super useful now, but they will be very useful soon. So I think specifically if you work in, uh, anything creative, right? The fact that you now have, uh, video in music generation is pretty big, you know, runway support, uh, Suno, you know, comfy UI, whatever you use. And then the same thing with, uh, Google's, uh, Liria. Uh, so if you work in any creative industry, I think it's going to be good to just start, um, you know, toying around with this. I'm sure it's going to get much, much better. But then the memory in the, uh, the memory, the memory wiki solves a lot of, uh, you know, the common problems that people have been having with open claw, even if you have your, uh, you know, your soul and be file set up, your heartbeat set up, everything correctly. Still the memory piece is pretty tricky. So the new memory wiki, uh, is the next attempt to solve that. So who's going to find this valuable? Like I said, content creators, uh, great. If you just want, you know, quick and dirty, uh, you know, videos, uh, developers building AI, as you work flows, obviously, and then power users and tinkerers, uh, who are kind of using open claw as a primary assistant, right? And you want it to remember more, uh, the new memory features are their attempt to do just that. All right. Next, speaking of remembering things, we have a new. AI moves too fast to follow, but you're expected to keep up. Otherwise your career or company might lag behind while AI native competitors leap ahead, but you don't have 10 hours a day to understand it all. That's what I do for you. But after 700 plus episodes of everyday AI, the most common questions I get is, where do I start? That's why we created the start here series, an ongoing podcast series of more than a dozen episodes you can listen to in order. It covers the AI basics for beginners and sharpens the skills of AI champions pushing their companies forward. In the ongoing series, we explain complex trends in simple language that you can turn into action. There's three ways to jump in. Number one, go scroll back to the first one in episode 691. Number two, tap the link in your show notes at any time for the start here series, or you can just go to start here series.com, which also gives you free access to our inner circle community, where you can connect with other business leaders doing the same. The start here series will slow down the pace of AI so you can get ahead. Notebook integration in Gemini. So it's technically a notebook, LM integration, but not the same as the one that they added at the end of 2025. All right. So let me explain like what that means. So what they added at the end of 2025 is you could add your notebook. LM kind of notebooks as a source inside of Gemini. So this is a little different. The new notebooks feature is more of kind of like a folder, but not the folders that you would normally use inside of like chat, GBT or. Claude, these are more like a notebook, LM folder, but that you can use in Gemini and notebook, LM, and then it sinks. All right. Let me kind of read the official word here from Google. So it says notebooks and Gemini help you organize your chats and projects in the Gemini app with our air, with our AI powered research partner, notebook, LM for easier learning and working. All right. So here's kind of the, the gist of what it does. So it says with notebooks, you can keep conversations about a topic organized in one place, just click new notebook on the side panel of the Gemini app to get started. You can move past chats into notebooks. So that's, that is good. You can organize your past chats into notebooks and then give Gemini custom instructions and add irrelevant files like documents and PDFs to give Gemini more context. So in that regard, it is like standard, you know, chat, GBT projects and Claude projects, but here's where it's not at all. Because all of those notebooks technically sink to notebook. L.M. which is really cool. I probably, which is crazy to think outside of Google, Gemini's deeper research because Gemini is everywhere now, right? I use AI studio a ton and I use notebook L.M. a ridiculous amount. It might be one of my most used tools, aside from like GBT five four pro. Um, and I am using like most people Claude more and more, uh, you know, with all these new features that come out, but I actually am not using the, I'm not using the normal Gemini app to chat as much as I used to be, right? I'm using a lot more for deep research. I think Google's deep research is still goaded. It's so good. Um, but now with this, I'm like, okay, now I'm going to have to start doing more of my chats, uh, inside, uh, Gemini, right? Uh, because this feature to sync all of those, put all of those in a notebook and have that be synced to notebook L.M. Right. You have all of these sources that you bring in a notebook L.M. If you're a power user or maybe if you don't know, uh, notebook L.M. is a very unique product from Google. It is powered by Gemini, uh, but it's grounded. So that means the hallucination rate is essentially zero, right? We're all other large language models. You know, you're constantly having to be vigilant against, um, hallucinations, right? So this is so cool now that your chats inside Gemini can become your source material for notebook L.M. Just by adding it to a notebook. Uh, so that's really cool. So like I said, you do have that persistent project space where you can organize chats, upload documents and PDFs and give Gemini custom instructions for those ongoing products and the notebooks sync automatically, which is really cool. So if you add a file, you know, in notebook L.M. That file will be added in Gemini, uh, which is so, so cool. I, I'm pretty sure, uh, I messaged one of the, uh, notebook L.M. leads of something like this, like, uh, a year ago, they did add one of my features that I requested, uh, finally, um, getting on the, on the mobile app. Uh, you know, plus and minus, um, you know, plus and minus 10 seconds. Uh, I asked for 15, they did 10, but you know, can't complain. So who has access? So right now this is rolling out to paid subscribers first. All right. So if you're on a free plan, you know, maybe this will roll out eventually. I'm guessing it probably, well, I could be wrong. Cause I'm guessing this is going to be a more of a compute intensive feature, uh, especially if it's popular. So it is rolling out to, um, paid users on the web this week, uh, and then mobile access and European expansion in the following weeks and free users are said they'll get access after all the paid rollout completes. So no clue when that is, because sometimes it does take a little bit longer, uh, for all of the paid customers in Europe to get access. So sometimes those, you know, free users getting access gets delayed a little bit. So here's why this useful. And I think who will find it valuable? Uh, first of all, literally anyone is going to find this valuable, right? I don't care if you're a researcher, student, general knowledge worker. If you're not using notebook, LM literally every single day, I don't understand why not, unless you're not able to, because of your job, but I mean, use it for your personal life. If you have a personal Gmail account, even the free version of notebook, LM is can't miss. It's so good. Uh, right. But here's why I think that this new updates useful. It solves that scattered conversations problem, right? Obviously, if you've used, uh, projects in chat, GBT, or if you've used projects in Claude, especially if you've used like project memory inside of chat, GBT, you'll understand how useful it is, uh, to essentially that way you're not wasting all of that back and forth, right? So I think sometimes when we're working with large language models, you know, ultimately we just use whatever that last piece is, right? It's the last thing that you copy and paste, right? Or maybe you're going through a lot of iterations of something just to get a couple of facts, but what about the other 70% the other 80% the other 90% could all be really good stuff. So now this lets you automatically use that inside of notebook. LM that's probably the way I'll be taking advantage of it most is, uh, you know, moving from Gemini to notebook. LM not vice versa necessarily because I am a power notebook. I'm a user, uh, but regardless, this is pretty freaking sweet. All right. We have a couple more. All right. Next we have Muse spark meta meta after billions of dollars. Quite literally, right? They just had a single 14 billion dollar acquisition, um, of a scale AI, essentially an aqua hire, and then they reportedly were spending hundreds of millions on individual researchers. And then over the past year, they've done nothing with llama a year of silence in 2025 and 2026, like a decade. I think people forgot meta existed, but they came back with what I will say to their credit is a fairly impressive model in Muse spark. So here is Muse spark. This is their new model, but the first model from the meta, the new, uh, meta superintelligence lab or MSL led by former scale AI CEO Alexander Wang. And it is a natively multimodal reasoning model with text, image and tool use. But unlike the previous llama models, this is a complete rebuild. So this is not an open source model that you can download and fork locally. Right now though, it is free. So free, but not open source. All right. So it uses a multi-agent architecture where parallel reasoning sub agents tackle part of a problem simultaneously reducing latency on those complex tasks. Uh, so who has access right now? Like I said, it's free to use on meta.ai. Uh, and in the meta AI app across Facebook, Instagram, and what's app. There is also a thinking mode that they call contemplative. And then, uh, there is a private API preview right now, but it's not public. It's only open to select partners. So I mean, why it's useful. I mean, you get it, right? I don't know if you're, if you're a heavy meta Facebook, what's app user and you like AI, that's cool. You're going to have a much better AI assistant to chat with. Will this be a model that's good enough for businesses to replace their stack? I don't know. Probably, probably not. Right. It's actually really good. Right. The benchmarks for technically being a first model, like this might sound crazy. Because yes, it is technically their first model, but it's not right. They had lava before, but this is a complete rebuild. So if you give it the benefit of the doubt of being a complete rebuild, it's actually one of the best first models ever, but is it a top three model right now? No, it's not because it's still behind Gemini three one pro opus four six and chat, GBT is GBT five four. However, I have been using it. It's really good. It's good at coding. It's ability to write is pretty good. It's obviously a huge step ahead from llama. So, like I said, if I think you're going to find value out of this, if you are already using Metas products on a daily basis, or if for whatever reason, the certain benchmarks really just kind of hit what you're, you know, what you're trying to get out of it. Like I said, we actually, and I will share it, I'll reshare it in today's newsletter. I put together a chart comparing Muse Spark with its biggest competitors. And then also just for fun, I put the mythos model from Anthropoc that no one's going to get access to for the most part. So yeah, if you want to see how it ranks on the benchmarks, I'll have that in the newsletter. All right, next. Meta wasn't the only company with new models today or this week. Microsoft also did. Yeah, this one, I kind of came out right at the end of last week. I'd already, you know, planned the show out. So this one is like eight days old. All right. But stuff happens fast, right? So this is from Microsoft. I'm just going to read their little blurb here. So they said introducing MAI Transcribe One alongside MAI Voice One and MAI Image Two, what they call world-class quality at lightning speeds now available at the most competitive prices. And it is available now in the Microsoft Foundry and the MAI Playground. So I will say if you want to, you know, compare everything on the image side, right? It's I think it's a top three family of models. It's not a top three AI model, image model, but you know, for their first technical model, it's fairly impressive on the image side. And then on the transcribe or the voice side, it actually got some pretty good benchmarks here. So it was the lowest for mean word error rate, right? So when transcribing, it actually is the top model for that. So if you are building something with voice or transcription, it might be a model worth looking at, but some more details. So this is the first big series of releases from Microsoft's new in-house AI team now led by Mustafa Suleyman. And they released these three proprietary foundational models. So MAI transcribe one claims the lowest word error rate 3.8% across 25 languages, beating the likes of whisper Gemini flash in 11 lives scribe. Then you have MAI voice one. I don't know why the MAI is just hard to say. We had a MAI voice one that can generate 60 seconds of natural audio in under one second on a single GPU. So that's text to speech model or text to audio model very, very fast. And then the MAI image to debuted as top free on the arena AI leaderboard. So these are all available in right now, right? So you can access them on the Microsoft foundry or the MAI playground. So if you're company for whatever reason, right? If they're not a big Microsoft shop, maybe you're a Google Gmail shop, or you just don't have access to the Microsoft foundry, which is kind of there. Formerly the Azure AI foundry. Well, you can just go sign up for the MAI playground, right? That's the way I've been kind of playing around these models in the same way that open AI, Claude, etc. They have backends essentially for developers. That sounds daunting. Trust me, it's not. You just sign in with your credentials. You usually have to connect a credit card because then at that point you are charged, even if you're just playing around with it in the sandbox, you are charged. So here's why it's useful. I think if you have already built or your company has already built transcription or voice products, it's worth looking at it. Just because at least I do think on the transcribe and voice side, it's fairly impressive. Or maybe for whatever reasons, maybe your company has wanted to use AI images, but you are locked into Microsoft. In this case, this is their first image model that I think is definitely worth using. So who's going to find this valuable? I think enterprise developers who are building voice agents, if you need accurate, fast and affordable transcription at scale, that's one great use cases. Or marketing teams who need high quality image generation and you only have access to Microsoft. I obviously wouldn't choose this, at least not now, over the Nano Banana, over the GPT 1.5 image. We talked earlier this week that there's rumors, OpenAI is going to be coming out with their next image model. So it's definitely not the best, but I have seen some of the examples so far and it looks fairly impressive. All right, we have two more quick ones for you here. So our next one, Google Vids Update. Stick with me here. Why does this matter? Well, you probably know Google VO, right? And if you're on a paid plan, then you have access to Google's extremely impressive AI video generator in VO 3.1. Now it's free. Well, the version of it. And if you use Google Vids. So here's what's new. So Google Vids now includes free AI video generation powered by VO 3.1 for anyone with a Google account. All right, you do only get 10 clips per month. But y'all, if you don't remember, like when VO 3 came out or even VO 2, right? But when VO 3 came out, it's like everyone's like, oh my gosh, I'll pay a million dollars, you know, for this because at the time it was so good and it was, you know, so far ahead of, you know, the original Sora and so far ahead of all the other, you know, Chinese models. Now that gap has been closed. But the fact that now we're saying that you can get VO 3.1 for free, it's pretty impressive. So let's talk a little bit more about how this works. Well, one of the big features of Google Vids, I did do a dedicated show on that a couple of months ago, but it has AI avatars that can be directed via text prompts, placing custom scenes in dress to match your brand and interact with products and props, which is really, really cool. It also has a new screen recorder and sorry, a new Chrome screen recorder and extension to direct YouTube publishing as well. So here's who has access. Like I said, free, free VO 3.1 free. How many times can I say that? So up to 10 clips a month, even if you have a non paid account. Also, you have custom music and Lyria 3 and AI avatars, but that is for paid subscribers only. And then ultra subscribers gets 1000 video video generations per month. That's obviously on the 200 plus dollar plan. So here's why it's useful. And I think who's going to find it valuable. Well, it's useful because it creates really good VO 3.1. It generates eight second video clips from text prompts or photos at no cost. Right. So this is one of those. I think if you've been looking at some some ugly visuals on your company's website for many years, which be honest, this is probably most of us. Y'all live in it up. Or if you have this old, you know, corporate training video and maybe you can't replace all of it, you could at least splice it up and breathe some life into it with VO 3.1. All right. And then our last feature, which may be the one that's most talked about, although at least for me personally, I think I'm most excited about the agents SDK. But here we go. We got managed agents from Anthropic. All right. So more or less, if it sounds like every week that Anthropic is releasing one or two things that sound very much open claw. Well, that's because this is very open claw. All right. So this is technically a little bit more technical than using clawed.ai on the front end because you do have to use Anthropic's council, right? So that's their kind of back end, but it's super easy. I did mess around with this a little bit. The good thing is, is you can do it all in the natural language. You can do it all just by chatting with Claude and it will walk you through how to set it up, you know, authenticating things, etc. But here's essentially what's new. So they launched the Anthropic launch the Claude manage agents right now in public beta. And it's a platform that handles hosting, scaling, monitoring and failure recovery for AI agents. So teams can focus on agent logic instead of infrastructure. So you just define agents in natural language or you can use the animal. Then you can set guard brails and then Anthropic runs them. All right. So this isn't like you have to set up your own infrastructure or you don't even need to necessarily run these on a website or, you know, set up a local instance on your computer or anything like that. It is all running on Anthropic's instance. So this has built in orchestration that handles tool calling decisions, context management, and error recovery with session tracing. Right. So the easiest way to explain this think opus 4.6 is a genetic by default. Right. And I think that we've all seen I've said this many times, right. Anthropic is crushing it in 2026. They are winning the year. And one of the things that's most impressive is as an example, you know, adding skills, you know, inside of Claude, adding all of these different MCPs inside of Claude, right, interactive ones as well, these more interactive apps. Now imagine kind of similarly to Zapier being able to control that harness a little bit more. Right. Because if you didn't have, you know, as an example, a connector, you know, for whatever program that you want to use, or if it wasn't available as an MCP, yes, you could go set up a custom MCP server, but this is just a different way to set it up by doing it just in natural language. And then the thing that I love is that they're kind of ancient builder walks you through it, and it'll do the authentication and everything for you. Right. It'll walk you through it. It'll give you options, right. It'll say, Hey, you can do it via, you know, OAuth where you can just click the screen, or you can do it via API key as an example. So it's going to walk you through it. It's going to give you these different options. So who has access right now? It is a public beta. So anyone, all right, for API users and subscribers. So like I said, you do have to be on the back end. The cool thing is, well, you technically don't need to even be a paid Claude user. You do though, obviously have to connect a credit card, go into the back end. I think the minimum, you know, put in like five bucks, right? You can like prepay five bucks to go like mess around with this, right? So here's why it's useful, because you can abstract months of infrastructure work, right? The sandboxing, the state management, credential handling, permission systems, all these things that typically delay agent projects, and just let anthropic manage that piece, right? So anthropic claims 10 times faster time to production compared to building the agent infrastructure yourself. All right. So here's who's going to find it valuable. I think enterprise engineering teams who want to deploy Claude powered agents without building infrastructure from scratch, you know, project managers and non-developers who can define agents in natural language, or companies that are already using Claude's API who want manage scaling and monitoring for their agent workloads. So we will see how, you know, how popular that this is from anthropic. On the surface, it seems like it will be fairly popular, although it is a little bit more technical. But then at the same point, I thought that OpenAI's their version of this with their agent builder, it seems like that never really took off. And it's very similar to what Claude just released here in managed agents. So I could be wrong. We'll see maybe most of the momentum goes into things that you use on the front end, which is something I've been saying for many years, right? I've always said, you know, when you're talking about buying your building, right, I'm like, well, you should technically be buying and using these systems on the front end because the front ends are becoming more and more powerful. And I do think, you know, even as we see OpenAI eventually go to the super app, I think then it'll become apparent to people why using these things on the front end is probably going to become more prominent. Because, well, number one is it's going to share and keep your context, right? So yes, I think things like the managed agents from Anthropic are really good. I don't know if it's going to take off as an example, the way that Claude co-work has, the way that Claude code has, because when you do it there, right, you have it all kind of under the front end, quote unquote, front end system, whereas sometimes moving on the back end, yes, obviously for developers, that's not who I'm speaking to, you all can see the promise of this. But for everyone else, I don't know if this is going to take off, although I do encourage you to go play around with it just like I did. I think you're going to find plenty of use cases. All right, so that is a wrap. There are the seven features that I don't think you can afford to miss. So whether you are a Zapier power user, or if you want a managed agent inside of Anthropic, or if you've really been wanting some more content creator capabilities inside of OpenClaw, this week brought a whole lot and a whole lot from Google and Microsoft as well. So I hope this was helpful. If so, let me know about it. If you could leave me a rating on the podcast. I really appreciate it. Make sure to follow and subscribe to the show. Then go to your everydayai.com, sign up for the free daily newsletter. Thank you for tuning in. Hope to see you back tomorrow and every day for more. Everyday AI.