76 Min Read

[The AI Show Episode 205]: AI Labs Refocus on Agents and Enterprise, Trump’s New AI Framework, Meta’s Rogue Agent & What 81,000 People Want from AI

Featured Image

Claude Code didn't just change one company's trajectory, it triggered a chain reaction across every major AI lab.

In this episode, Paul Roetzer and Mike Kaput break down how OpenAI, Google, Meta, xAI, and Microsoft are all scrambling to catch up in the agent and enterprise race, and why the next three to six months could look radically different from anything we've seen.

They also go inside our SmarterX company retreat, examine Microsoft’s Copilot restructuring under Satya Nadella, discuss a rogue AI agent tied to a security breach at Meta, break down the Anthropic-Pentagon tensions, explore Google DeepMind’s new framework for measuring AGI, and more in this week’s rapid fire.

Listen or watch below—and see below for show notes and the transcript.

This Week's AI Pulse

Each week on The Artificial Intelligence Show with Paul Roetzer and Mike Kaput, we ask our audience questions about the hottest topics in AI via our weekly AI Pulse, a survey consisting of just a few questions to help us learn more about our audience and their perspectives on AI.

If you contribute, your input will be used to fuel one-of-a-kind research into AI that helps knowledge workers everywhere move their companies and careers forward.

Click here to take this week's AI Pulse.

Listen Now

Watch the Video

Timestamps

00:00:00 — Intro

00:04:04 — AI Pulse Survey Results

00:05:50 — AI Labs Refocus on Agents and Enterprise

  • Meta Shifts

00:29:07 — New Polling on AI and Trump National AI Framework

00:45:46 — Company Transformation with AI (Offsite Recap)

00:59:52 — Nadella Takes Over Microsoft Copilot

01:06:06 — Meta's Rogue AI Agent

01:10:01 — Anthropic vs. Pentagon Continues

01:14:42 — DeepMind’s New AGI Scorecard

01:18:40 — What 81,000 People Want from AI

01:26:01 — AI Academy Spotlight

01:30:47 — AI Product and Funding Updates


This episode is brought to you by AI Academy by SmarterX.

AI Academy is your gateway to personalized AI learning for professionals and teams. Discover our new on-demand courses, live classes, certifications, and a smarter way to master AI. Learn more here.


Read the Transcription

Disclaimer: This transcription was written by AI, thanks to Descript, and has not been edited for content.

[00:00:00] Paul Roetzer: all. The labs realized what Claude code unlocked. And it wasn't like it was the first coding agent. it was just the best. They did something different with the harness. Like how they enabled it to do what it does.

[00:00:11] All these labs see, not the finish line, but like the next mile marker of AGI Agentic capability and their ability to automate AI research and their ability to then, as Logan Kilpatrick's deleted Tweet said, to start disrupting everything.

[00:00:25] Welcome to the Artificial Intelligence Show, the podcast that helps your business grow smarter by making AI approachable and actionable. My name is Paul Roetzer. I'm the founder and CEO of SmarterX and Marketing AI Institute, and I'm your host. Each week I'm joined by my co-host and SmarterX chief content Officer, Mike Kaput, as we break down all the AI news that matters and give you insights and perspectives that you can use to advance your company and your career.

[00:00:55] Join us as we accelerate AI literacy for [00:01:00] all.

[00:01:02] welcome to episode 205 of the Artificial Intelligence Show. I'm your host, Paul Roetzer, along with my co-host Mike put, we are recording Monday, March 23rd. About 10:00 AM Eastern time. some big stuff. Last week, Mike, we got, yeah, I don't, the whole last week, just crazy. We were on a company retreat for two of the days, so I always just feel like I lost track of time

[00:01:24] Mike Kaput: Yeah, same.

[00:01:25] Paul Roetzer: Yeah, same. And then my entire week was spent getting ready for the company retreat to create.

[00:01:29] You and I both taught workshops, which we'll talk a little bit about to the team. And then I did five presentations and workshops I think on the first day.

[00:01:39] So it was a little bit of a crazy week, but in between all that, we had over, I think over 50 different sources in the podcast sandbox this week.

[00:01:48] So as usual, Mike did an amazing job of curating the topics for today, and we were updating what we were gonna say even about three minutes ago before we came on. We still hope may adapt it [00:02:00] as we're moving forward.

[00:02:02] Paul Roetzer: There's just some big stuff, openAI's and their kind of shift, but it's sort of a larger trend about what's going on with the labs.

[00:02:08] There's some new polling data about ai. I don't know, just Meta's got a road agent there. There's just a lot to unpack this week.

[00:02:16] So this week's episode is brought to us by AI Academy, by SmarterX. If you're a regular listener, you hear us talk about AI Academy a lot. This is the core focus of what I do at the company, and it's a huge part of what Mike does at the company, is building the content and the curriculum for AI Academy.

[00:02:31] It's designed to help individuals and businesses accelerate their AI literacy and transformation. Through personalized learning journeys and an AI powered learning platform, new educational content is added weekly, so you're always up to date with the latest AI trends and technologies. Our AI four industries collection features six core series and certificates that are designed to jumpstart AI understanding and adoption across industries.

[00:02:56] The six that are available right now and they're part of, the overall AI mastery membership [00:03:00] program, or you can buy them individually. We have AI for professional services, AI for healthcare, AI for software and technology, AI for insurance, AI for financial services. And the newest one that just came out Friday, last Friday was AI for retail and CPG.

[00:03:16] So these series are an ideal launchpad for organizations that wanna level up their teams and accelerate that AI adoption and impact. Mike teaches a number of them, including AI for professional services. And so later on in the episode, we're actually gonna, get some insights from Mike of some of the big takeaways he had from that series.

[00:03:33] Again, like our, it was probably gonna be part of a new, element of the podcast. We're gonna start trying to drill into some of these, you know. The course series we're creating. We're spending so much time researching and building these things. We wanna bring some of those core insights to everybody as part of this podcast.

[00:03:47] We're gonna start doing some of those. So Mike, tee that off this week with AI Professional Services. So individual and business account plans are available now, or you can buy those single courses, and series as I mentioned for one-time [00:04:00] fees. You can go to Academy dot SmarterX.ai to learn more.

[00:04:04] AI Pulse Survey

[00:04:04] Paul Roetzer: Alright, Mike, we have our AI pulse this week. So this is SmarterX.ai/Pulse. You can participate in these pulse surveys each week. They're informal polls of our listeners where we ask a couple of questions related to that week's episode. So last week, Mike, we had, Atlassian.

[00:04:20] Mike Kaput: Atlassian. Yep.

[00:04:21] Paul Roetzer: Atlassian laid off 1600 workers and explicitly cited the AI era as the reason. What is your reaction? So we had 39% said this is the new normal AI driven restructuring is real and accelerating. We had 26% say it's AI washing, a fast growing company using AI as cover for cost cutting, and 25% too early to tell.

[00:04:44] We need to see if the roles are truly replaced. And then 11% that are more concerned about the total tech layoffs in 2026 than any single company. I don't, nothing really surprising there. It's a pretty balanced

[00:04:58] response overall. but I think [00:05:00] 39% is the highest response rate. that it's kind of the new normal and then.

[00:05:05] In a New York Times quiz, 54% of readers preferred AI written prose over human originals. What's your reaction? I wonder if it's the exact same. People answered this

[00:05:15] this way, said

[00:05:16] 39% said Not surprised. AI has gotten genuinely good at clean polished writing. 28% said writing quality was never the real mo.

[00:05:24] Taste, judgment and point of view are, and then we had 20% said this is a wake up call for professional writers to differentiate beyond surface quality. And 14% said the qui quiz was flawed, that it was, you know, it's kind of a relevant result. Okay. So we will give you the two, pulse questions for this week later on at the end of the episode.

[00:05:44] But again, SmarterX.ai/pulse if you wanna participate in those pulse surveys each week. Alright, Mike.

[00:05:50] AI Labs Refocus on Agents and Enterprise

[00:05:50] Paul Roetzer: So the first one is, it started in our sandbox as a bunch of openAI's news. There was a whole lot of stuff. I'm gonna let you unpack what happened with openAI's across like the 15 articles that we were looking at.

[00:06:03] And then I'm gonna do my best to sort of, take a zoom out and say what is actually going on at all of these labs. 'cause this I think is, there's a major shift happening and I, when you start looking at the collection of all of this information at the same time, you start to kind of see the trend of where this is going.

[00:06:22] Mike Kaput: All right Paul. So right now openAI's is in the midst of executing what might be one of the more dramatic strategic pivots.

[00:06:30] It's done so far.

[00:06:31] So it's simultaneously restructuring how it sells, what it builds, and who builds it all while preparing for a potential IPO later this year. So on the enterprise side, Reuters reports that OpenAI is pursuing partnerships with multiple private equity firms in deals potentially worth a combined $10 billion.

[00:06:51] These firms include places like TPG, advent International, Bain Capital, Brookfield Asset Management, and others, and the PE [00:07:00] investors would, contribute approximately $4 billion and receive equity stakes, board seats, and influence over how open AI's technology gets deployed across their portfolio companies. So the logic here is that private equity

[00:07:14] firms control massive portfolios of enterprise companies and influence their tech spending. So this partnership gives openAI's a distribution channel directly into those businesses. Now, notably, Anthropic is also reportedly courting Private equity, including Blackstone signaling that this may become a standard go-to-market playbook for some of these frontier AI companies.

[00:07:40] Now on the product side, OpenAI is consolidating its web browser, Atlas ChatGPT as a whole, and its Codex coding tool into a single unified desktop, what they call super app. So Fidji Simo who LE leads open AI's applications division. Confirmed this move saying the company is cutting back on [00:08:00] side quests to focus on coding and business users.

[00:08:04] At no hands meeting on March 16th, Fidji Simo laid out the commercial goal that they want to convert. OpenAI's 900 million users into high compute users by turning ChatGPT from a consumer chatbot into a productivity instrument built around AGI agentic ai. Now interestingly, they're facing quite a bit of competitive pressure on this front.

[00:08:28] So according to enterprise software vendor ramp, the proportion of businesses using Anthropic increased from one in 25. To nearly one in four within a single year. And Anthropic currently wins approximately 70% of direct comparisons against openAI's in new enterprise contracts. Now, meanwhile, at the same time, openAI's is also going all in on fully automated AI research.

[00:08:56] So founding openAI's member Andrej Karpathy went viral [00:09:00] this past week. We talked about this last week describing, kind of an experiment where he deployed an autonomous AI coding agent to run continuous research for two days.

[00:09:09] It calls it auto researcher.

[00:09:10] And basically, this agent, like we discussed, executed hundreds of experiments, discovered new optimizations, and sped up how well the model itself worked.

[00:09:21] Mike Kaput: In terms of its training time. Now, interestingly, Shopify's CEO tested the same approach on his internal company data running an agent overnight that conducted dozens of experiments and improved performance by almost 20%. Now the point here is that Carpathy says All Frontier AI Labs will adopt this approach, calling it the final boss battle these labs face.

[00:09:43] So there are reports that openAI's is following suit, going all in on this idea

[00:09:48] of trying to build an AI researcher. Now lastly, they are reportedly nearly doubling their head count according to the Financial Times and Bloomberg over the next year as they scale [00:10:00] across all of these initiatives simultaneously.

[00:10:02] So Paul, maybe connect the dots for me here. openAI's is making some pretty big, pretty sudden changes.

[00:10:10] Paul Roetzer: Yeah, so, the trend I was referring to is, this goes back to episode 189 of the podcast on January 6th, so right out of the holidays was when Claude code sort of blew up and it became very hot over those last two weeks of 2025. And we spent an entire segment of the episode talking about what was happening with Claude Code and how it it, something had definitely changed.

[00:10:37] And so that was the starting point and all the major AI labs. Are, are in this accelerating race for autonomous agents and enterprise customers. So that's the thing I

[00:10:46] referenced, like when when first started the outline for this podcast yesterday, there was just this focus on openAI's. But when you look in the totality of all the articles, we're looking at all the tweets we're seeing, you see that, [00:11:00] that this, everything has changed to this refocus on agents and enterprises, which was not really OpenAI’s core.

[00:11:07] It's not like they weren't going after that audience. and they weren't building agents before. But Claude Code changed things. And you and I, Mike can attest to this, like it's, it's incredible, like within Claude, the ability to build things. I'll give an actual example a little later on.

[00:11:23] In this episode, but,

[00:11:24] it changed things and they're ahead of everyone of very clearly ahead, when you use the product.

[00:11:31] So, I 'm just gonna, I'll break down a little bit the openAI's thing, but then I want to get into the bigger picture. So you, you mentioned the, Fidji Simos, Talk about like private equity firm that they're in these advanced talks, and that both openAI's and Anthropic are aggressively courting these PE firms, which makes a ton of sense.

[00:11:52] And we've talked about this a little bit before on previous episodes, but Anthropic, as you mentioned, is, is winning in this space. So openAI's [00:12:00] enterprise business, according to Reuters is 10 billion out of the total analyzed revenue, about 25 billion right now. And I said that's a run rate. They're not, they're not actually at 25 billion yet in a, a year, but that's the run rate they're on right now.

[00:12:13] and then, she tweeted on March 16th. This news came out a little earlier than we planned. We're excited to be building a deployment arm and we'll share more details soon. So that's what we're talking about, this idea of kind of getting out with these frontier alliances where they're actually working with the consulting firms and stuff.

[00:12:28] So there's just a lot going on where they're trying to get to where the enterprise customers are. and then when it started getting into this idea of refocusing, which is interesting because I remember last fall we were talking about this, like all of a sudden Sam Altman's like. Everywhere. They're gonna do space stuff, they're gonna do robots again.

[00:12:46] They're gonna, you know, build the video gen apps and social networks and devices. With Jony Ive like, they're just everywhere. And it was like, whoa. Like you're getting crushed right now on the model side. Like, why don't you focus on [00:13:00] the model side? And it appears they've come to realize that. So, Fiji's tweeted on March 19th.

[00:13:07] Companies go through phases of exploration and phases of refocus. Both are critical, but when a new bets start to work, like we're seeing now with Codex, which is their version, you know, like Claude Code, it's very important to double down on them and avoid distractions. Really glad we're seizing the moment.

[00:13:22] I remember when I first saw that tweet, I was like, that's weird. Like that. It's just a weird tone on a tweet. Like almost like. People were questioning whether she was behind this focus because that's not what she was brought there to do. Like she was in part, brought to, to diversify, like based on her background.

[00:13:39] So I think some people may have taken this news as almost like a slight against what she was supposed to be doing there. I don't know. But I, that's how I read that tweet was like, wow, that's a really interesting, trying to set the tone that you're behind all this raw rock kind

[00:13:53] of stuff.

[00:13:54] So that was in relation to the Wall Street Journal article that said openAI's Plans launch of Desktop Super app to [00:14:00] refocus, simplify user experience.

[00:14:02] in that, there was a quote that we realized we were spreading our efforts across too many apps and stacks and that we need to simplify our efforts. That was from Fidji Simo. that fragmentation has really been slowing us down and making it harder to hit the quality bar we want. That said, top executives, including Altman, chief Research Officer, Mark Chen and Fidji Simo have spent the last few weeks, reviewing opening eyes product portfolio and looking at areas to deprioritize.

[00:14:26] And then in an all hands meeting, she told employees they couldn't afford to be distracted by those side quests you mentioned, and that they're in this major battle with Anthropic and it's basically like a code read internally. This is all related to this idea of a fully automated researcher, which isn't.

[00:14:40] News, and we've talked about this being something they were working on for at least the last year, but I think it's starting, the timeline is maybe starting to come more clear. So they said their new research goals, the North Star for these next few years, pulling together multiple research strands including work on reasoning models, agents, and interpretability.

[00:14:58] Meaning like knowing what the models are doing [00:15:00] and why they're doing it.

[00:15:00] And then there's even a timeline, openAI's plans to build, quote, an autonomous AI research intern. A system that can take a small number of specific research problems by itself by September. and a lot of what Andrej Karpathy is talking about is sort of a prelude to this stuff and said the AI intern will be the precursor to a fully automated multi-agent research system that the company plans to debut in 2028. It's a weird timeline to me. I don't, I don't know why it would be like that long. But anyway, this AI researcher openAI's says, we'll be able to tackle problems that are too large or complex for humans to cope with. You mentioned the idea of these, again, these side projects, what that means. And when you hear side projects, it could be things like the SOA, video generation app, like the standalone app.

[00:15:46] I gotta think the planned hardware devices fit into

[00:15:49] this bucket, If you didn't spend the 6 billion on Jony Ive, like I gotta imagine that there's a chance that you get delays in the hardware because that's hard. it's a difficult thing to [00:16:00] pursue, and that could definitely be a major distraction.

[00:16:02] And then e-commerce features in ChatGPT, you know, you could see those kind of get sidelined. So there's lots of interesting things they've been doing that could get sidelined in all of this. And then you mentioned the, you know, in the same time they're doubling head count. So they're aiming to grow to about 8,000 employees.

[00:16:18] They're at about 4,500 today according to Financial Times. And then, um. You know, so I, overall it's like, it creates this muddied relationship continuing with Microsoft as well. So again, when I started zooming out, it's like, well, what, what's going on with all the other labs? Like we hear so much lately about the challenges Microsoft and OpenAI are having as they try and they tried to reimagine that relationship so that OpenAI could get in a position to go public.

[00:16:44] And in the process they allowed them to start developing partnerships with people like Oracle and AWS, which I'll, I'll talk about in a moment. So then we get into the Microsoft thing. Now we'll talk a little bit more about this one in a rapid fire, and we'll drill into this, but the premise is Microsoft made a [00:17:00] major shift last week where they're moving copilot under Satya Nadella.

[00:17:04] So they're actually moving it under another executive. Jacob, Andrew. Then he reports directly to Satya and they're taking Mustafa Solomon, who was in charge of Microsoft ai and he's gonna like just run the Super Intelligence lab, it sounds like. At the same time, Microsoft, according to Financial Times, is weighing legal action over $50 billion.

[00:17:25] Amazon openAI's Cloud deals. Now we have this weird muddying of relationships between Amazon and OpenAI. You have XI, so one of the other major labs, there's basically five major labs in the us xAI, Elon Musk tweets on March 12th. this is following lots of turnover at that AI lab. A lot of the co-founders have have left in the last 60 days.

[00:17:47] He tweeted xAIwas not built right first time around. So is being rebuilt from foundations up. Same thing happened with Tesla. So you have xAI, you know, one of the major labs is. You know, according Elon Musk in [00:18:00] basically a complete, you know, reset mode. And this is a month after, on February 12th, they got acquired by his other company, SpaceX.

[00:18:08] So. Just what, what, 40 days ago. SpaceX, SpaceX, you know, who also Elon Musk Company set on Monday acquiredxAIthe XI AI company controlled by Musk, to consolidate his empires and kind of build this, you know, one unified company. So that combined company now includes X like the Twitter platform and includes XI.

[00:18:29] And then they have a deep relationship now with Tesla, his, his other company. at the same time, Musk is suing openAI's like that, and that's supposed to go to trial in like April, right? [00:18:41] so you have this crazy thing, but Elon Musk is watching what's happened with agents and enterprise.

[00:18:46] He wants a piece of that and he realizes, wow, we, we didn't build this the right way. Let's like just hit the reset button and nobody hits the reset button faster than, than Musk. Like he's, if something's not working, he is gonna blow it up. Then you have meta, we talked [00:19:00] about this. So March 12th meta delays rollout of new AI model after performance concerns.

[00:19:05] So they're spending what, over 15 billion last year just on talent acquisition. So, they're me investing heavily, they're rumored to be spending 135 billion this year on like CapEx to build out the future of everything. And it doesn't seem to be working yet. Like they, they haven't released a major model since they acquired, Alexander Wang and Scale ai.

[00:19:26] So Meta's sort of in upheaval, they've kind of fallen off. It's like them and xAI are just sort of like Down at the bottom right now, you had Yann Lecunn leave XI, but then Meta shows up and buys Moltbook. The AI agent social network that went viral because of fake posts back in, you know, earlier this year.

[00:19:45] So Meta's trying to get in and have a piece of this agent game they'd probably love to play in the enterprise world, but that's not their, their natural thing. Then you have Jensen Huang talking last week about OpenClaw being the next ChatGPT. So there's a CNBC article that [00:20:00] says Jensen Huang, CEO of Nvidia on Tuesday pointed to a fast rising AI project called OpenClaw is a major step forward in how people interact with artificial intelligence.

[00:20:09] He said, it is now the largest, most popular, the most successful open source project in the history of humanity. This is definitely the next chat. GPT OpenClaw is an open sourced autonomous agent platform that goes beyond traditional chatbots. Instead of answering questions, these agents can complete tasks, make decisions, and take actions with minimal input from users.

[00:20:31] Nvidia moved quickly to build around OpenClaws moment The chip leader on Monday announced NEMO Claw, an enterprise grade version of OpenClaw that layers NVIDIA's software stack and tools on top of the platform. And then you have Google DeepMind. so Google, you know, came in hot with Gemini three.

[00:20:50] It was great. Like it's powerful. They've, they've just last week announced some major improvements to Gemini within Google Workspace, which we experience, Mike. Every day we, we use [00:21:00] Google Workspace and we embed Gemini. They've had kind of a runaway success with Notebook lm. Even though, I mean, when you talk to the average business leader, they have no idea what NotebookLM is.

[00:21:10] So like in our bubble NotebookLM’s amazing, and we talk all the time. We have courses on it. The average person has no idea what it is or how to use it. So they've had success building these individual apps like Notebook and Gemini. They announced a major investment last week in AI studio where they're trying to get into the vibe coding game.

[00:21:26] So trying to play along with like how Claude code and stuff is. But the reality is AI Studio is still for developers. Like, I can't, I don't know how to use it. I went in there last week, I was like, okay, maybe it's like ready for me to use it. And it's like,

[00:21:39] so Gemini, while amazing Google, Google DeepMind incredible.

[00:21:44] They have no answer to Claude Code right now. Like it's running circles around them and based on what we, like, what an a, a Google engineer said. So we talked about this on episode 189. Yolanda Dogan, a principal engineer at Google. On January 2nd, [00:22:00] tweeted, so this is, this is them saying it, not us. He said, I am not joking and this isn't funny.

[00:22:05] We have been trying to build distributed agent orchestrators, which is exactly what we're talking about with like OpenClaw and claw code at Google since last year. There are various options. Not everyone is aligned. I still can't believe this tweet was allowed to go out. I gave Claude code a description of the problem.

[00:22:20] It generated what we built last year in an hour. It wasn't a very detailed prompt, and it contained no real details given I cannot share anything, proprietary. I was building a toy version on top of some of the existing ideas to evaluate Claude Code. It was a three paragraph description. And then when will Gemini get to this point?

[00:22:38] I think someone asked, he said, we're working hard on it right now, the models and the harness. And then I thought this was really interesting, Mike. So Logan Kilpatrick, who's sort of like, you know, head of ai, dev dev developer relations basically. So he's like a, a major player within Google. DeepMind came from openAI's.

[00:22:52] He tweeted, I couldn't believe I saw this tweet. And I was like, holy shit, that's gonna come down fast. And it did. He deleted it. It said, I think this was on [00:23:00] Saturday. All the industries you thought weren't going to be disrupted by AI are about to be disrupted. They're not allowed to say that. Like

[00:23:09] Mike Kaput: Google customers are reading that saying, I'm sorry, what? What?

[00:23:12] Paul Roetzer: Yeah. a hundred percent true. You can't say that. And so someone got that down real fast. so Google's sort of like in this crazy phase where they're trying to build it into Gemini. They're trying to make it like function within the productivity tools that they have. While DeepMind is.

[00:23:32] Paul Roetzer: Telling you that like every industry's gonna be changed.

[00:23:34] So then I'll, I'll wrap here with I, what I think is if you, if you want to, you gotta be ready for the technical stuff. When you listen to Andre's Carpathy, like he's, Mike and I talk about Andre's all the time. He ran Tesla computer vision for five years. Co-founder of openAI's did a bounce back to openAI's for about a year Now, he's an independent researcher.

[00:23:52] He's, he's been on fire on X the last, like three weeks, just like all these crazy things he's working on. But he did an, interview on the [00:24:00] No Priors podcast. Again. if you're ready for the technical side of this. Listen to this episode. We'll, we'll put it in the show notes. a few key notes that I always take.

[00:24:09] I was listening to this yesterday actually, so a few key things. He was talking about how fast these models have evolved and how it's largely a skill issue, which is funny because that's a term my son tells me. Like when he beats me in a video game,

[00:24:21] he is like, it's a skill issue, dad. It's like if I lose in Mario Kart, I'm like, oh, it's the wrong character. He goes, no skill issue.

[00:24:27] So I apparently that's like the lingo right now. so he was saying it's a skill issue. If you can't get value out of these models, there's this idea of token maxing, which is a very technical concept, but it actually makes a ton of sense. So every time you use one of these models, you're basically burning through tokens.

[00:24:43] So tokens is when, when a large language model does something where an agent does something, it's basically making predictions using tokens. Tokens are like pieces of words in essence. And so you get an allotment of these tokens. So let's say I use a million, 2 million tokens, whatever. So he was saying like, if you're.[00:25:00]

[00:25:00] An engineer, you wanna know what your token budget is like, how much AI can I use in my job? And so this idea of token maxing is like for the average user, like you and me, Mike, I have a clawed license, I have a CHE GPT license and I have a Gemini license. And if I'm not maxing out my subscription every month, I'm leaving like intelligence and outcomes on the table.

[00:25:20] And so he was saying there's this like pressure right now, especially on coders to max out your available tokens. 'cause if you don't, you, you're just like, you're not getting the full value. And I think that concept is starting to, is going to start carrying over eventually into knowledge work where you're like, we have these AI tools.

[00:25:39] We're not fully utilizing them and we're just, we're leaving value on the table by not maxing out our tokens each month. And in a similar place. He talked about this idea of running projects in parallel, which I do like. I'll go into Claude and be like, okay, I'm gonna give it this project. I'm gonna go over to ChatGPT, and I'm gonna have it work on this project.

[00:25:56] And so there's times where I'm running three [00:26:00] projects simultaneously with AI agents. While I'm doing my other work. So like, I'm gonna do an email or I'm doing something else and I've got, and so that's a big thing. And then he talked about the compression of timelines to complete projects, which I'm gonna talk about in, in a, upcoming, topic here about, our company retreat.

[00:26:17] But I think that's a very important concept that things that used to take five hours, 10 hours, 20 hours now might take five minutes. And that's a weird environment to be working within. And then he also talked simultaneously about this idea of compression of software stacks where we used to have a CRM tool and a social tool and all these tools.

[00:26:35] And it's like, I'm just gonna have a, a swarm of agents and they're gonna go talk to all this software and I'm just gonna have like a single user interface. and then the other final one, I'll say with, with Carpathia, and again, this is all relevant to what these labs are doing. So these, if you listen to the Carpathia interview.

[00:26:50] All the labs are realizing what Carpathia is realizing on age agentic capabilities. And they are now in a race to do what he explains in this. And that's why this, [00:27:00] this podcast episode is so important that no priors episode, he is telling you point blank what all the labs are trying to do with agents and you will walk away with a better understanding of the moment.

[00:27:11] But then he said, at one point working with these agents is like simultaneously talking to a PhD student and a 10-year-old. So like sometimes you do something with it and it's like, it was like giving it to a top PhD student and then the next moment it's like some stupid simple thing and it just can't do it.

[00:27:28] So it's that idea of the jagged frontier and the jaggedness

[00:27:31] of these models. So zoom out, recap at all. The labs realized what Claude code unlocked. And it wasn't like it was the first coding agent. it was just the best. They did something different with the harness. Like a how they enabled it to do what it does.

[00:27:47] All these labs see. Not the finish line, but like the next mile marker I'll say of AGI Agentic capability and their ability to automate AI research and their ability to then, as Logan Kilpatrick's deleted [00:28:00] Tweet said, to start disrupting everything. and And, so it is an all out race for agents and they're seeing a pot of gold with enterprise adoption, which is why Anthropic and Open Air are doing deals with PE firms.

[00:28:14] It's why they're doing alliances with major consulting firms. They're trying to get in and get where this is gonna be because the labor replacement value of being the model they go to when they reduce workforces and put it all into AI models to Token Max to get work done, they see that future coming very fast.

[00:28:33] Paul Roetzer: it's important. Like I just covered a lot in like 20 minutes here, I think it's very, very important that you understand. What we just covered, like that's what these labs are doing and it's gonna become very apparent I think in the next like three to six months that this is full Go where they're headed.

[00:28:53] Mike Kaput: Probably a pretty good time to be an enterprise buying AI technology. I'm assuming these labs would like to [00:29:00] court you.

[00:29:00] Paul Roetzer: Yeah. Yeah. You get a lot of credits, especially if you're like, I'll give you the first million free.

[00:29:05] Mike Kaput: All right.

[00:29:07] New Polling on AI and Trump National AI Framework

[00:29:07] Mike Kaput: Next up, we've got three separate developments this week that are painting an increasingly complicated picture of how Americans actually feel about AI and how Washington is responding.

[00:29:18] So first we had some new polling. David Shor, who is head of data science at Blue Rose Research appeared on the Odd Lots podcast with some interesting polling data. So his organization has found that over the past year, AI Research in issue importance and its importance as an issue faster than any issue his firm tracks.

[00:29:40] It is now more important to voters than climate change childcare and abortion. According to their polling, 79% of voters are concerned the government doesn't have a plan to protect workers from AI job losses. 77% are concerned about entire industries being eliminated. 56% are worried about personally losing their job to ai.

[00:29:59] [00:30:00] This is hitting at a time when 61% of Americans say life has gotten less affordable. In the last year, only 25% feel confident in their financial future. Only 34% say that they're, in their opinion, they have a secure job. So what Shor's data shows, and he's pulling from this perspective of trying to kind of find political messaging for the Democratic party, is that this whole idea of, Hey, everything's gonna work out just fine.

[00:30:22] That message is dead on arrival. They actually found when leaders in government and tech

[00:30:26] say, AI will not cause widespread job losses. Net trust is negative 41. And when they say AI will create economic productivity that benefits everyone. Net trust is negative. 20. Now you're starting to see this play out across

[00:30:41] the political spectrum because we second up this week got dueling AI political declarations.

[00:30:47] So first there was a coalition that involves a lot of unlikely bedfellows, including Steve Bannon, Susan Rice, Richard Branson, Ralph Nader, Joshua Bengio, and others who released the pro-human [00:31:00] AI declaration. It's basically called for a prohibition on superintelligence development. Until there's broad scientific consensus,

[00:31:07] it can be done safely, as well as a number of other manifesto points about keeping AI pro human.

[00:31:14] So over 40 organizations sign this, and they also found in their own polling that Americans would rather prefer human control over the speed of AI development by an eight to one ratio. However, another organization called Build American AI published a direct counter to this manifesto titled We Cannot Afford to Pause ai.

[00:31:34] They argued safety and innovation are not opposites, and the US already has. Regulatory tools through existing authorities to manage AI development. Now. Third, the Trump administration unveiled a national AI legislative framework with seven pillars. This is a short document, but basically gives legislative guidance on how they think legislation should evolve related to ai.

[00:31:56] And this framework takes a pretty clear try first, rather than [00:32:00] regulate first posture, it opposes creating any new federal AI regulatory bodies. It defers copyright questions to the courts rather than legislating, and it recommends Congress preempt state AI regulations that impose undue burdens on developers establishing what it calls Americans quote, right to compute.

[00:32:19] There's an interesting part in here in shifting responsibility for protecting children online from tech companies to parents.

[00:32:26] So rather than imposing strict industry standards, they are actually shifting more to empowering parents with tools to protect kids online. The framework also calls for Congress to empower Americans to challenge federal agency efforts

[00:32:40] to quote, dictate the information provided by an AI platform.

[00:32:43] So basically trying to make sure that there is UE influence on what information is provided by ai. So Paul, I'm curious, there's a number of threads going on here. If you're in the AI industry or just observing or trying to navigate these changes yourself, how,

[00:32:59] are you thinking about [00:33:00] these numbers and the moves on either end of the political spectrum?

[00:33:03] Paul Roetzer: Like any research we always talk about, you gotta like. Know who's doing the research and what their goal is, what kind of bias might be in the research. That being said, it's gonna become more political and as we've said many times in recent months, this is all trial balloons. They're trying to figure out what do Americans think about ai and is there an opportunity to move votes, a few percentage points one way or the other.

[00:33:29] Wi with, taking a strong position on ai, which Republicans and Democrats haven't really, for the most part with voters. So the one thing that. It's becoming more interesting to me is, you know, I always read this research and think these people don't know what AI is. Like. You're asking them questions about something that they don't understand and now I'm actually thinking out loud here of like that that's maybe an advantage for politicians that wanna manipulate and persuade people to vote one way or the other.

[00:33:58] So if you don't know what it [00:34:00] is, then you can create

[00:34:01] Mike Kaput: It can make mean whatever you want,

[00:34:03] Paul Roetzer: A hundred percent. So if a if people generally like, I don't know, like whatever, then it's like, okay, let's hammer the message of it's gonna take jobs and it's gonna, data centers are gonna ruin communities. And now that's all AI is to people.

[00:34:15] So this is a maybe a dangerous. slope. We're going down here where we're seeing the early efforts to try and gauge what is perception so that we can then influence perception of what it is to move, to move votes one way or the other. So David Shor, I didn't know who he was, I didn't know his organization, so that's always the first thing I do.

[00:34:33] It's like, okay, we see some cool data and it's getting shared everywhere on, on X. Who are these people? Was always the first thing I ask. What is their mission? So David Shor is head of data science at Blue Rose Research Research based in New York, originally from Miami. I try to elect Democrats. That is his, that is his X profile.

[00:34:51] So what I just read is his X profile. So there's no hiding, like what the point of this is. Blue Rose Research research helps campaigns make higher quality strategic [00:35:00] decisions by democratizing access to ACT measurement. That's on their About Us page. The name Blue Rose Research symbolizes Turning blue, what is now red.

[00:35:08] So again, there's no hiding. Like what this is, is for, David Shor is a prominent American data scientist, political consultant and expert in public opinion polling. Now that doesn't mean it's not valid research. We're just saying like that. There's a perspective here. That's the whole point of understanding this.

[00:35:23] He actually worked for Barack Obama's 2012 reelection campaign. So. The survey. just to put it in a little bit of context, when it says AI is like the fastest growing issue, you have to understand it's actually 29th out of 39 issues right now though. So yes, it's growing fast, but the top five issues for Americans are cost of living, the economy, political corruption, inflation, and healthcare.

[00:35:47] Those don't really move, those are pretty common, top five. Then if you go down to like 25 to 30, so just to put in context of where AI falls, you have war in the Middle East at 25, international trade income [00:36:00] inequality, voting rights, then artificial intelligence, then race relations. So while it is growing fast on the surface, Americans don't really care.

[00:36:08] Like it's not, it is not something that would jump out to you as like votes are gonna move. Based on that, but it is changing fast. you talked about some of these key ones, Mike, the government not having a plan to protect workers from job loss, driven to, so the question was how concerned are you about, and then it said the government not having a plan to protect workers from job losses is driven by AI 79%.

[00:36:30] So you don't need to understand what AI is to like Yeah, I mean, it kind of worries me that I'm a plan and that is a hundred percent true. They do not have a plan. Like, or if they have a plan, they're certainly not talking about the plan. So, everyone should be concerned that the government doesn't have a plan.

[00:36:43] Then it said, how concerned are you about young people entering the workforce and finding fewer job opportunities because of ai? 79%, they should be concerned that's happening. Like that, that is a real thing right now. So again, whoever's asking these questions, Republican, democrat, independent, doesn't matter.

[00:36:57] That is a fact. Like it's harder to find jobs right now. [00:37:00] entire industries being eliminated by AI faster than new ones are created. That's a ridiculous question. Like, we're not getting rid of industries

[00:37:05] Mike Kaput: right.

[00:37:06] Paul Roetzer: Companies being disrupted. Sure. Career paths. So that, that's an absurd question. You could just throw that one away.

[00:37:12] A, AI changing the job market in a way that drives down wages for people like you. 72%. You could replace AI with any variable, anything you ask. It's like, are you concerned with, you know, something driving wages down? Of course I'm concerned. I don't want my wages going down. So it's like, eh, whatever. you or someone in your family losing their job in the next year because of ai, 56%, that's a reasonable concern.

[00:37:35] and then when they say, when leaders in government and tech industry say AI will not cause widespread job losses. Net trust, as you mentioned, is negative 41, distrust it somewhat, 35% distrusted completely 32%. So 67%. Disrupted dis distrusted somewhat or completely. Now, that may align with 67% of people don't believe anything government tells you.

[00:37:57] So like I have no idea, [00:38:00] So again, just framing where the data's coming from. Then there's another one, data for Progress, which is a progressive think tank and polling firm that provides data research and messaging strategies for progressive movement. They produce polling on policy issues and support campaigns.

[00:38:14] So they came out with new research on February 27th, which is worth mentioning here. this is 1200 US likely voters nationally using a web panel. so they were asking about how frequently they use AI in their daily lives, whether they have favorable or unfieldable views of the tech and how confident they are in their ability to spot AI generated content.

[00:38:32] This is a pretty short survey. We'll put the link in. It's only like five pages. You can read it for yourself if you want, but some of these questions are pretty interesting. Do you have a favorable or unfavorable opinion of the following people or institutions? They asked about ai? Democrats minus three net favorable Republicans, plus 11 independents minus five.

[00:38:51] They asked when it comes to AI tools such as ChatGPT. So now again, they're trying to qualify for you. What, what are we talking about when we talk about ai? So if you understand what ChatGPT is, at [00:39:00] least you have some concept. when it comes to it. In your personal life, have you mostly embraced or resisted using them to assist your life?

[00:39:07] Or have you found, areas where you could use AI in your personal life? Embraced, Democrats, 32 Republicans or Democrats 34 Republicans, 32 resisted, Democrats, 35 Republicans, 33. And then I have not found areas that I could use AI in my life. Democrats, 30 Republicans, three, two, which is like. To totally balanced.

[00:39:26] Like there's really nothing, there that would indicate, you know, any, anything they can do with that data or to move people one way or the other. Then they had another one. Sometimes people use AI to make fake or edited photos and videos that they post online. How confident you feel in your ability to spot that stuff?

[00:39:42] very confident. 15%, somewhat confident. 35%. So that's 50% think they can figure it out.

[00:39:48] Mike Kaput: Oh, they do. Yeah. Right.

[00:39:50] Paul Roetzer: That, they can, yeah, they're not, they're wrong.

[00:39:52] Mike Kaput: Yeah.

[00:39:55] Paul Roetzer: and then they, then they did an interesting one where they were like comparing data from August, 2025 [00:40:00] to February, 2026, where they asked, how frequently, if at all, do you use the following, using AI such as ChatGPT for your job?

[00:40:10] So, right now. 14 percent say multiple times a day. 44, rarely or never,

[00:40:18] Mike Kaput: Hmm.

[00:40:18] Paul Roetzer: 11 a few times a month. So you have 55% of these people being pulled in February, 2026, that few times a month, rarely or never. So again, if you like, think everybody's doing this, they're not. and then the one you mentioned about the pro human AI declaration, again, it's important to kind of know,where the count is coming from.

The AI industry Super pacs. We talked about this last year. CNBC had this as well as others that there's a super pac, called Leading the Future. And the contributors to this are Andreessen Horowitz openAI's, co-founder Greg Brockman, Palantir, co-founder Joe Lonsdale, and angel founder, SV Angel, founder Ron [00:41:00] Conway.

[00:41:00] AI software company perplexity. So these are the people like pushing the Super pac, which is all about acceleration. It's all about like rapidly accelerating what's going on, and they're basically saying that, that this stuff is ridiculous. So, the Build America AI is, is in essence, like led by this group and they're saying we cannot afford to pause ai.

[00:41:19] So, this is a TechCrunch piece. Highlights the Reese of the pro-human AI declaration. The document you mentioned, the goals behind that effort are understandable. People want AI to be safe and they want clear rules. Those are fair concerns, but this is still the wrong direction. So this is the Super PAC people pausing Frontier Development will not solve the problems that supporters claim it will solve, if anything at risk making several of them worse.

[00:41:42] It would slow the research that helps us understand how these systems behave and practice and weaken American's position at the exact moment. Our adversaries are invested heavily in advanced technology. We cannot hand hostile actors on the world stage, a strategic edge. That is what would occur if we paused ai.

[00:41:57] And then that leads to the AI legis [00:42:00] legislative framework from the government, which is just the starting point. That's the most important thing to take away from that. It's just like guidance on where they think legislation should go. It's not doing anything yet, but you covered some of it's like protecting children, safeguarding and strengthening in American communities.

[00:42:15] Respecting intellectual property rights. That's a really funny choice of words. respecting intellectual property rights, meaning they don't want you to have property rights as a,

[00:42:23] as a creator and supporting creators, preventing censorship and protecting free speech, enabling innovation and ensuring American AI dominance.

[00:42:31] That's probably the most important one. and 'cause all the other ones fall under that one. And then educating Americans and developing an AI ready workforce, which I'm definitely intrigued to hear what they've got in mind there. So,

[00:42:42] Paul Roetzer: Yeah. Again, it's just, I think what we're seeing, we, we've said this like recently, every week there's now gonna be more and more on the political side.

[00:42:50] We are moving into the midterms. We are moving into the moment where the political parties have to decide whether or not Americans care. And this election, it [00:43:00] cycles either gonna be all about AI or it's just gonna like, fade away. Yeah. and you're seeing the push towards data centers being bad, job loss being bad, and then you've got the lead in the future super PAC people who are like, all of it's great and it's all gonna create an abundant future for all of us.

[00:43:15] And if you don't believe that, then believe we have to beat China. Like that. That's basically the messaging, you know, it's like, choose, choose your fighter. Like that's, and then I don't, you know, I don't know where the middle ground is here, but right now, neither side really knows. But the super pac, the leading the future people are, they're gonna push hard on this stuff and they're gonna try and make you believe it's all gonna work out and jobs aren't gonna be lost.

[00:43:38] And you, you always have to, what I would just encourage people to do is like. Don't get stuck in whatever your traditional political silos are.

[00:43:48] You

[00:43:49] know, if you only listen to one perspective on this, this is an issue where you can't just be listening to one perspective that you've always followed. I think it's really important to like realize neither political party knows the [00:44:00] answer here.

[00:44:00] They're both trying to figure it out. And so it's really important that you open your own mind and like look behind who's saying things and what the goal they have behind saying that is, or where their research is coming from. It's gonna be very important to try and keep a level head on this stuff and be, listen to arguments of both sides.

[00:44:20] Mike Kaput: To

[00:44:21] your point about people often being pulled who don't know what AI actually is, that's the point of some of these numbers we can throw out. We would throw out half these questions if we were doing actual research.

[00:44:33] But if they surface a strong opinion or view on ai, even if that view is wrong, that's really useful polling to certain people because it tells you exactly what you need to say and hit on. Using that ignorance almost as a weapon in some ways.

[00:44:51] Paul Roetzer: Yes. Facts and lies mean nothing in election cycles. It's all about what's, can you say that'll get you to remain in power. [00:45:00] And that's, again, I don't, I don't think that's like a controversial

[00:45:03] perspective.

[00:45:05] It is what it is. Like they're gonna tell you whatever you want to hear to stay in power or to get in power, both sides.

[00:45:11] So form your own opinion, like form your own informed understanding of the situation and then from there you can take more logical actions to make sure, you know, you understand. I don't know, it's like a situational awareness, I guess, about what's happening with this issue. It's gonna become a major issue I think.

[00:45:32] I think they're gonna find that they're gonna find the , the levers to pull, they're gonna find the wedges to create frustration and anxiety around ai and that that could get very dicey.

[00:45:46] Company Transformation with AI (Offsite Recap)

[00:45:46] Mike Kaput: Our third big topic this week is about Paul, this SmarterX annual meeting and retreat we had over the last couple days of last week with our team.

[00:45:56] So this was super

[00:45:57] inspiring. We spent a couple days together [00:46:00] collaborating. Day one. We talked about vision, goals, KPIs, priorities, and growth initiatives. Day two, we ran AI productivity and AI innovation workshops, which are designed to accelerate responsible AI adoption across business units and teams.

[00:46:14] And the reason we wanted to cover this and dive into it is because it has some signals, maybe some lessons here about overall company transformation with ai because Paul, let you kind of unpack this for us, because what we were able to achieve over just two days, both in how we were approaching AI and by actually using the technology, I think can teach us quite a bit about how AI is changing the way businesses operate.

[00:46:40] Paul Roetzer: Yeah, I mean, a couple things and I, you know, Mike, you and I haven't talked about this, like it's so if you have other perspectives or things to add, let me know. But yeah, the reason we wanted to highlight this is a few things came to me. So it was like two days. There was a part of me that thought. It was a great example of what you can do with the time you gain from ai.

[00:46:59] So the fact [00:47:00] that we use AI so intelligently within our own business gives us a little freedom to say, yeah, let's take a full two days. Like, let's go do this thing, let's go, you know, think let's go spend time together, build camaraderie, like do all the things we should be doing. And as I was sitting there, I kept thinking like, we gotta do more of this.

[00:47:16] Like when I think, what does an AI forward company look like? And how do you take the benefits you gain from ai, the efficiency and productivity gains, and redistribute that in some way. I'm not a four day work week guy. I don't think that's reality. I do love the idea though, of like, let's do more of this.

[00:47:32] Let's have like once a month, let's just take like an afternoon and just think and talk and like work on big ideas. Like I find that to be more like enables the work to be more fun and more fulfilling if it's not just let's max, let's token max every minute of every day. So I think there's like the, in some ways I wanna build.

[00:47:50] I wanna, I wanna maximize what we can do, but I also want to make sure we are getting the benefits of it. It's not like a race to some end [00:48:00] game or like some competitive race. So yeah, the way we set it up was, day one was, as you mentioned, sort of the company day as vision goals, KPIs, building scorecards, a rocks workshop or setting priorities for the coming quarter.

[00:48:14] And then I think just like the thing we teach, which is setting expectations for every one of what, what, what an AI forward professional looks like. and in some ways modeling that by showing in real time how we're using AI and making sure everyone on the team understands the capability. So Mike, you did on day two, you let off with this AI productivity workshop and you talked about the idea of not only jobs as tasks, but tasks as workflows, which I loved that framing.

[00:48:40] And then you went through like an AI capabilities overview of like, what are all these things the models can do so that people started to think. A little bit differently about their own daily lives at work. We demoed, jobs, GPT campaigns, GPT and Innovations GPT, I did that one in mine. But those are some of the free [00:49:00] custom GPTs we've built that we make publicly available.

[00:49:02] We use 'em in our own teams. Like we literally use these tools to train our own teams and as an example of like this AI Ford idea in real time. So Mike's doing his workshop, which was awesome because I've never sat through one of Mike's workshops. So I, like we, Mike and I do these things all

[00:49:17] the time . For other companies,

[00:49:18] and we do 'em at our MAICON event, but like we don't have time to sit in each other's workshops.

[00:49:23] So he's doing this workshop and he is showing ai, kinda layering over workflows and re-imagining workflows. And he showed this ai, capability slide and then he turned it into, it was like a spreadsheet with like 90 rows or

[00:49:36] Mike Kaput: Got like 90 different capabilities and features across some of the major AI tools. So you can quickly like pick and choose and filter and map things to all the individual tasks you're doing as part of a workflow, for instance.

[00:49:48] Paul Roetzer: So like reasoning, capabilities, video capabilities. So yeah, it's like, so I was like, I love this. And I was looking at this thing. I'm like, I wonder if we could turn this into something. So as he's talking, I take the [00:50:00] spreadsheet and I put it into Claude code or, or, or, or just Claude. And I'm using Sonnet 4.6 at this point.

[00:50:05] And I said, help me visualize this. We want it to help professionals understand the full capabilities of today's leading AI models so they can apply them to their work. That was the entire prompt. so it did it, and I was like, this is really cool. And I said, is there a way to turn this into an app that I can demo internally?

[00:50:19] So like three minutes later, I had this functioning app. So Mike doesn't know this is happening. He's just on stage doing his thing. But the best part, and this I 'm still like trying to wrap my head around this, Mike, we, we don't have a clawed license for the team. So when Mike built his capability slide, it was what Google.

[00:50:37] Mike Kaput: Google Gemini ChatGPT NotebookLM and I spun out kind of deep research for both tools as kind of its own capability set,

[00:50:46] Paul Roetzer: right?

[00:50:46] Okay. So this 90 row worksheet does not have Claude in it, but I'm talking to Claude to build this interactive demo. So Claude says, at first asked me a question, how would you want to share run it and I said, Standal and h tml [00:51:00] final is fine. It then said, what? What should people be able to do beyond browsing?

[00:51:03] Select all that apply. And then I just simplicity. I was like, just browse. That's enough. Then this is the question that blew my mind. It said,

[00:51:11] should Claude be included as a fourth tool? So it was aware that it wasn't part of the spreadsheet he created and it asked me if it should add itself to the spreadsheet.

[00:51:24] I literally laughed out loud when I saw this. I was like, what? And so I said, yes, a add, Claude, great. And it did, and it followed the exact model he had done for the other, and then it built this interactive capability

[00:51:37] thing. It it, I mean it honestly blew my mind and as I said, like we do this stuff every day.

[00:51:43] I see this stuff every day and there's still moments where I'm like, I can't even believe it was capable of doing this in real time. And so when I say Claude is like running circles around what some of these other apps are capable of doing, hey, I said it's capable of doing, this is a perfect example of it.

[00:51:58] Mike Kaput: Yeah you one shotted a 90 item capability database. More

[00:52:03] than 90 because it added in probably 25 different things from Claude. One, shot at it in a way that genuinely was professionally designed, extremely intuitive. It was great. It was finished, it was a filter capabilities. Yeah.

[00:52:18] Paul Roetzer: Unbelievable. So the other one I'll share and again, we'll touch on some of this kind of later on, was rocks. And I put this on LinkedIn on Sunday, and I actually featured this in my newsletter, the exec I insider newsletter.

[00:52:30] So I'm just gonna read, what I wrote because I I t summarized it really well. So I basically saying like, we went through this retreat and like one of the things became apparent to me as an example is this idea of rocks. So we use a modified version of rocks from the, EOS system in which departments and individuals establish three to five priorities per quarter.

[00:52:49] And then the rocks allow us to align our time, energy, resources on what matter most, and it provides transparency. So if I wanna see what are like the five things Mike's working on in Q2, I can go, or if I wanna go see what the [00:53:00] studio that Mike leads is doing, I can go see that. So the thing that became abundantly clear to me is the time to complete rocks is compressing and that it requires a complete rethinking of business operating systems.

[00:53:11] So for example, during a live session where I was actually demoing, so this is part of the company day, I was demoing a new AI assessment tool we're developing that I'll share more about in, in probably a month or two.

[00:53:23] So I had used Anthropic cloud code against Sonnet 4.6 in real time to build an interactive reporting dashboard that visualized and analyzed responses from 17 people.

[00:53:33] So I had built this assessment in Google Forms as like an MVP, and then Mike and I tested it the day before retreat, just to make sure it. It worked. And so I had my data and Mike's data, and then I had everybody else take it. And then I exported that CSV, CSV, from Google Sheets. So that, that was it.

[00:53:51] That was like the entire process. You zero coding, zero design abilities to do this thing. And I give this to Claude and I said, so this is while we were taking a lunch break. [00:54:00] I ran this, I , so here's my prompt. I had 17 team members take the assessment. Can you come up with an elegant way to visualize the results based on the format model you already created?

[00:54:09] So I had to create one for me and Mike. and the CSV is attached. So in a previous life, which I said like AKA three months ago before Claude Code really started working, this would've been my entire Q2 Rock. Like create an interactive dashboard to visualize assessment results. For teams, I would've spent 10 to 20 hours researching dashboards and developing a brief.

[00:54:28] Then I would've invested time and money hiring a designer and a developer to conceptualize, build, iterate on the design and capabilities. Then we would've gone through weeks of internal testing and revisions, and then maybe by the end of Q2 I would've actually had a minimum vial product that I could demonstrate to the team and pilot with users instead, in about five minutes while I got a plate of pasta, Claude did the entire thing with one prompt and the final product was beyond anything we could have possibly created.

[00:54:55] And I told Mike, I was like, I'm gonna try this. I'm gonna do it. And then he and I are both just like waiting like we

[00:54:59] go [00:55:00] check the laptop, like did it do it, do it, do it. It was insane. It was totally interactive. Better than anything I could have possibly designed myself or work with a developer to build.

[00:55:08] And I'm now gonna use that to actually turn it over to a developer and say, here, let's build this and like take this, you know, live in like 30 days hopefully. So, We share this as a little bit like behind the scenes of how we think about SmarterX as like an AI native company like event and media and education company, and two, just to like bring to life the fact that you don't need any coding ability to, all of a sudden now just build stuff and it's totally compressing the timelines to do everything in business.

[00:55:34] And it's changing the way every day that I think about how to run our own company and how to advise other people to build their companies.

[00:55:41] Mike Kaput: Yeah, I would argue we have quite well done and clear and ambitious rocks, at least, you know, in, in our department that we were working on during this workshop. But yeah, it is actually kind of laughable that all five of them should take three months.

[00:55:55] Paul Roetzer: Yeah, I mean, I really think like my guidance to the team was like, five. I [00:56:00] want you to have like five for your department. I actually think you need 20. Like I, because.

[00:56:05] And you need like some categorical thing of like, Hey, this is, you know, this would take 10 to 20 hours of human labor. We think we can do it in 10 minutes.

[00:56:12] Like there are honestly things that are just gonna be like that. There's gonna be all these like quick wind rocks where it's like, well, it used to be three months worth of work, but it's probably three days now with mostly a I t's like level three ai. Like it's gonna do most of the work. So the companies that figure that out and realize that and restructure how they're building everything stand to do really well.

[00:56:33] Mike Kaput: And just

[00:56:34] two final really quick notes here, but to. Piggyback on what you did with AI during these workshops. So the AI capabilities map I built, which was again, like 90 ish rows of all these different capabilities and features, that's a lot to figure out on your own.

[00:56:52] And what's really cool is I determined the framework I wanted to use and worked back and forth with Claude to say, okay, what's the most sensible way to [00:57:00] organize these once I'm, you know, once we have them, I don't have them yet, and then it's like, okay, we've got a really solid system. How do I get them?

[00:57:07] Typically, you might go do a bunch of research, you might have to sort through all sorts of documentation. I just went into each tool and screenshotted all my menu options and dropped him into Claude and said, guess what? We're gonna go create the spreadsheet based on the. Framework that you and I came up with and go have at it, and then it basically, one shot's, a 90 row spreadsheet.

[00:57:29] It's incredible. And same type of thing during your innovation workshop. I fed Claude a lot of different context about my department, the content studio, some of my, context around our organization. And then using your framework that you developed, layered that on top of that context and what Claude is now able to do and got better innovation ideas than I could have come up with first on my own at all.

[00:57:52] And second in an entire day. I did it in like 20 minutes. So

[00:57:57] it's so powerful. Not only just [00:58:00] using the right tools of course, but having these proven frameworks and models

[00:58:05] and ways of thinking layered over them. All that stuff we've spent lots of time developing as like IP or as a unique models to approach these things with, in our workshops.

[00:58:15] It's, it's like rocket fuel at this stage.

[00:58:18] Paul Roetzer: Yeah, and I think there's just something like a lesson to be taken from how we structured it, because obviously our team's probably more informed than most teams about AI capabilities. But honestly, like, I don't know that they even were aware of a lot of the things these models could do.

[00:58:33] Mike Kaput: Yeah.

[00:58:33] Paul Roetzer: And so it was very intentional how we did this.

[00:58:36] And I would advise other companies to think about a similar model where you have this kind of like state of ai, of like, what is it capable of? And that's often what I'll go in and do with, with, enterprises. Like I'll do like a state of AI for business and here's the capabilities, here's what you need to understand.

[00:58:50] Then you do the productivity workshop where it's like, how do we get efficiency and productivity in our tasks and workflows? Then we'll, we'll often do a problem solving one too, but like the innovation one is how we [00:59:00] closed. And I intentionally wanted to close with that because once you understand what it's capable of and once you've solved like the lower level efficiency and productivity things, now you open your mind to the possibilities.

[00:59:11] And I mean, and then we go around the room and each person gives us like one or two innovations they're super excited about. So then you leave after two days actually feeling like. Ready to go, not like drained. It's like, okay, that was amazing. I wanna go do those things now. and that was what I kind people are coming up to me's like, okay, can we do these things that we

[00:59:28] just talked about?

[00:59:29] So I think it's a really cool format for people. And so if you're trying to get your team on board, you know, borrow that format of like, make sure they're understanding of it and if you need help with it, give us a call. This is like what Mike and I do all the time. We run boot camps and workshops and so if nothing else, we can advise you on, you know, ways to do it.

[00:59:46] but if, you know, if you're in big enterprise and you need help with it, just like, you know, we can come in and do stuff like that too.

[00:59:52] Nadella Takes Over Microsoft Copilot

[00:59:52] Mike Kaput: All right, Paul, before we dive into rapid fire, quick message here. This episode is also brought to you this week [01:00:00] by our upcoming webinar, which is unveiling our AI for CMOs blueprint presented by Google Cloud.

[01:00:06] Now, this is actually happening the week you are listening to this episode Thursday, March 26th at 12:00 PM Eastern 9:00 AM Pacific.

[01:00:13] And in this session, me and our CMO Cathy McPhillips are going to break down the insights from this AI for CMOs blueprint We put together in partnership with Google, where we break down real world state of AI for CMOs use cases, tools, strategies, and more.

[01:00:30] We'll also be doing some in-depth discussion and live q and a, so registration is free. All registrants will receive ungated access to the full AI for CMOs blueprint. So go to SmarterX dot ai slash webinars to go register. All right, let's dive into rapid fire Paul. So first up, Microsoft CEO. Satya Nadella is taking some

[01:00:51] more direct control of the company's co-pilot product, personally overseeing a restructuring that consolidates consumer and commercial co-pilot into a single [01:01:00] organization.

[01:01:01] Jacob Andro, a former Snap, SVP, who joined Microsoft last year now reports directly to Nadella as the new EVP leading co-pilot experience across both segments. The restructuring frees up Mustafa Suleman, the DeepMind co-founder, who became CEO of Microsoft AI in 2024 to focus

[01:01:20] entirely on what he calls the company's super intelligence efforts.

[01:01:25] This move apparently comes as co-pilot trails quite badly in the AI assistant. Race Co-pilot has 6 million daily active users compared to chat

[01:01:34] GPEs 440

[01:01:35] million. That is according to A-C-N-B-C article. Gemini has 82 million. Claude has 9 million Nadela wrote to employees that Microsoft is doubling down on our super intelligence mission with the talent and compute to build models that have real product impact.

[01:01:50] So Paul, what does this tell you about where Microsoft is headed with ai?

[01:01:54] Like reading this, I was like. I know it sounds like Mustafa is excited, [01:02:00] but this feels more like he's getting sidelined and we need to get real serious about copilot real quick, which is kind of what we've heard anecdotally from users of copilot.

[01:02:10] Paul Roetzer: Yeah, there's lots of variables going on here. I mean, one is the shift in their relationship with openAI's. You know, they're, they, they were obviously a major investor in openAI's. They're a major equity holder. I think it's somewhere around like 27% they, they own of openAI's. But all of their efforts were being built on top of openAI's models.

[01:02:29] And now again, if you go look at what we were just talking about with Claude, it's like you're almost at a disadvantage as an organization if you can't use when a breakthrough happens, when somebody builds just a better thing. You're at a disadvantage if you can't use that thing. And so if Microsoft was stuck using openAI's technology and all of a sudden Claude races ahead in some really important component, that's not great.

[01:02:50] And then if you're Microsoft and you're one of the three biggest companies in the world, the fact that you aren't building your own models is, is probably a disadvantage moving forward. And so I think there [01:03:00] was the shift where they realized that probably, you know, a year and a half, two years ago, that they were gonna have to remove their reliance on, openAI's.

[01:03:07] It probably happened the day Sam got fired when that became like, oh boy, we're, we are all eggs in one basket. And like, it's, it could go bad real fast. So, you know, I think there's been this ongoing shift where. They knew they needed to invest in their own technology, build their own models. They need to have kind of an off ramp over time from their reliance on openAI's.

[01:03:28] And then in November of last year, they announced this humanist super intelligence movement. So we talk about on episode 1 79, which was on November 11th, but Mustafa had tweeted it shouldn't be controversial to say AI should always remain in human control. That we humans should remain at the top of the food chain.

[01:03:45] That means we need to. Start getting serious about guardrails now before super intelligence is too advanced for us to impose them. And then there was, you know, linking to a article from November 6th that was called Towards Humanist Super Intelligence, where he said, [01:04:00] at Microsoft AI we're working towards humanist super intelligence, incredibly advanced AI capabilities that always work for, in service of, people and humanity more generally.

[01:04:10] So again, we, we've kind of known this was happening at that time. I think I pulled what I said. I said Maybe Mustafa stays, stays at Microsoft to realize this vision, but I can't help but feel like this vision will eventually clash with the need to justify their investments in, in ai. And so I think what they're basically saying is.

[01:04:27] You go focus on this stuff, you know, focus on the future and the building of this thing. But copilot is critical to our business right now, and it is not where we want it to be. And that now needs to get much closer to Satya. And that's basically, I think what has happened here. I have no idea if Mustafa says and keeps doing what he's doing, if they really do believe in this humanist, super intelligence thing.

[01:04:50] But that I don't see like Wall Street loving the humanist superint intelligence. I don't think stock price is going up because of that blog post or [01:05:00] that vision. they want to, they wanna know how you're gonna compete with Claude and work with Anthropic and that's all that Wall Street's gonna care about.

[01:05:08] And at the end of the day, Satya and Microsoft have a fiduciary responsibility to, to, return shareholder value. And I don't think that messaging plays so. I dunno. We'll see it fits into that whole thing I started off with where like, these AI labs are shifting focus and you're gonna see a lot of reorgs, a lot of like, you know, they tried something, it didn't work.

[01:05:29] Like meta burned 10 billion on the me, metaverse and changed their name to be meta. And it's done like, so just, there's gonna be lots of big efforts, big misses, and you gotta move quick when it doesn't work. And I think this is an example of that.

[01:05:44] Mike Kaput: not to mention anyone who's a Wall Street analyst of any type is almost certainly using Microsoft Excel and thus copilot and sees it firsthand.

[01:05:54] So or, they used

[01:05:55] Paul Roetzer: Claude in Excel and realized it was better than Microsoft's own

[01:05:58] copilot. Right. They [01:06:00] have a very close experience with perhaps some of the inadequacies of the tool.

[01:06:05] Paul Roetzer: Yeah.

[01:06:06] Meta's Rogue AI Agent

[01:06:06] Mike Kaput: All right.

[01:06:06] Next up, an AI agent inside Meta took unauthorized action last week that triggered an actual security breach at the company.

[01:06:14] So an employee used an in-house agentic AI

[01:06:17] to analyze a colleague's question on an internal for So pointed the AI at the question that analyze this for me. The agent then posted a response to the question on its own without being directed to do so. The second employee followed the agent's advice, sparking a domino effect that gave some engineers access to meta systems they should not have been able to see.

[01:06:39] The security breach was active for two hours before it was contained. A meta representative confirmed the incident and said no user data was mishandled.

[01:06:48] Though the company's internal report noted unspecified additional issues that contributed to the breach. A source told the information, there was no evidence anyone exploited the unauthorized access or that data was made [01:07:00] public.

[01:07:00] Though the reporting notes, that may have been the result of dumb luck more than anything else. The agent had also passed every identity check in META'S system. that exposes some pretty serious fundamental gaps we might talk

[01:07:14] about in enterprise identity and access management. So Paul, I'm curious, how close are most companies to having this kind of thing happen to them?

[01:07:23] Paul Roetzer: I don't know, but it's certainly a very viable thing. This is why I said in recent episodes, you got, you gotta let listen to it. I mean, there's a, there's a reason why some enterprises are moving really slow, especially when it comes to adoption of agents. And like, you know, we talked about the Jensen thing where he was like, OpenClaw is like the ChatGPT.

[01:07:41] I was like, okay, maybe. But do you know how hard it's gonna be in enterprises to do anything close to what that does? this is the exact issue, and we just talked on episode 2 0 3 about something similar happened with Amazon, where it just like went rogue and started doing everything. And I think I joked at the time, like, we could just do a rogue AI agent [01:08:00] episode, segment every week.

[01:08:01] Like it's, this is gonna be a recurring theme. It's gonna become a major issue. The concerns around oversight and governance of these agents and then these agents swarms that are just given access to stuff. And the breakdown you might then see in permissions controls. We had this conversation at our own company meeting.

[01:08:16] It's like, can we connect this to that? Can we connect that to this? And it's like, no, because I don't know yet the risks associated with that. so yeah, that's, again, it's one of these situations where the tech can do things, but it doesn't mean you should let the tech do things because there's so many potential risks.

[01:08:33] So yeah, I mean. This is a, this is a crazy one. You should go like read the articles about it. It's pretty nuts.

[01:08:41] Mike Kaput: Yeah. I must found this was like more notable because it happened. It wasn't like some super incredible agent, just like giving access to it to your whole code base

[01:08:51] or whatever. It was just like a totally unintended consequence of something that's actually like probably a pretty normal use case on the surface saying, Hey, let me use [01:09:00] AI to analyze a question one of my colleagues posted on a for

[01:09:04] And then you're like, oh no. I realize that now this thing can choose what to do and how to do it, and that's like totally a weird way to start thinking here, right?

[01:09:13] Paul Roetzer: Yep. Yeah. And again, go listen to Andres Carpathy no priors podcast episode and you, you'll understand this stuff at a deeper level. He, he talks a lot about these risks and even himself not knowing, he talked about like setting it up to run his house and he's, oh yeah, I gave it to access to, I was like, go

[01:09:29] find Sonos. And

[01:09:30] it like goes into his network and finds the Sonos speakers and then he gives the security can and he just gave it access to everything. And he like, Dobby the elf, he calls it Dobby

[01:09:39] the home. Yeah. It's hilarious.

[01:09:40] So again, this is recurring theme, really important to understand where agents are going, where these agents swarms are going, how they'll eventually be used to run organizations.

[01:09:49] And now some people are willing to be out on the edges right now, setting these things up and connecting them to their own company data. And we're all gonna learn plenty of lessons from their early [01:10:00] efforts.

[01:10:01] Anthropic vs. Pentagon Continues

[01:10:01] Mike Kaput: All right, in our next rapid fire topic this week, the Anthropic versus Pentagon Saga continues. The Department of War is fired back at Anthropics lawsuits and a 40 page filing in California federal court. The Pentagon calls Anthropic a quote

[01:10:15] on unacceptable risk to national security. Arguing the company might attempt to disable its technology or preemptively alter the behavior of its model during war fighting operations.

[01:10:25] If its corporate, red lines are being crossed. recently now nearly 150 retired federal and state judges appointed by both Republicans

[01:10:34] and Democrats

[01:10:35] have also filed their own amicus brief supporting Anthropic. We talked last week about how tech companies like Microsoft and Apple are all filing their, have all filed their own briefs.

[01:10:46] Basically arguing that this designation of Anthropic as a supply chain risk could mean the entire government procurement system becomes contingent on political favor rather than the rule of law. So, Paul, the big. Piece [01:11:00] here is really this idea that a bunch of ex judges are coming out and, saying that they also support Anthropic.

[01:11:07] In this, we have talked about if this is going to get resolved anytime soon. There's a hearing on whether or not to grant Anthropic

[01:11:15] some temporary relief that's actually set for March 24th, the date this comes out. Where do we stand with this?

[01:11:24] Paul Roetzer: I don't know. The only context, Mike, I'll add is, I think it's just still this, like he said, she said thing like where the government's saying one thing and they're doing the other thing, like behind the scenes, but they're trying to like, you know, give this perception that they're, They're all on the right here.

[01:11:41] And Anthropic is this horrible company and it's this huge risk. So there was a tweet thread from Roger Parloff, who's a senior editor at Lawfare, and he, I'll, I'll put the link in. he said some Anthropic updates on March 4th, just hours before Hegseth declared Anthropic a supply chain risk, allegedly due to threats of sabotage and data exfiltration.[01:12:00]

[01:12:00] His undersecretary wrote Anthropic and he have the screenshot of the email that they were very close to a deal asking to change a prepositional phrase. So while he exit's getting ready to like go on and blast them on X and say they're done, they're actually still negotiating behind the scenes and they have screenshots of it.

[01:12:17] Then, since then, the government has claimed that Anthropic sought a veto over Department of Defense actions, but two top Anthropic officials assert it never did. and this is actually like legally, they submitted a. Briefing saying this is not what happened. similarly, governments purported fear that Anthropic might disrupt the military was never raised with the company and is a technical impossibility.

[01:12:38] So they actually explained like we can't even do the thing they're claiming we would do. And then as of for Anthropics refusal to allow its product to be used for autonomous lethal warfare and mass surveillance. Hgsi himself said those concerns were understandable. And the commander of the US Sentcom echoed those sentiments, Andros head of policy rights.

[01:12:57] So they submitted these briefs saying like they agreed with us, like [01:13:00] we weren't even like raising something that they didn't themselves think was an issue. And then he had one last update in government's response. Tuesday, it backed away from the secondary boycott Heg. Hegseth called for in his February 27th, final decision post on X, admitting it was lawless, but also taking no responsibility for its devastating impact.

[01:13:19] The hearing is coming up on March 24th. So, yeah, so these are declarations, legal, like declarations from Anthropics head of policy, Sarah Heck, submitted as part of their response to the case. And then they also their head of public sector. So, um. yeah, I mean they're basically saying like, here, I'll testify to like this never happened, or this is what they said.

[01:13:40] So. The whole thing, as I've said many, many times,

[01:13:43] it's become this

[01:13:43] political thing. It's become a battle of egos on the government side. And you know, I think that everyone sort of sees through the why they're actually doing this and we'll, you know, see what the courts have to say. I guess

[01:13:57] Mike Kaput: it's impossible to tell, but based on that new [01:14:00] context, it almost sounds like there's one possibility where did like hags jumped the gun on tweeting about this when they were nearing a deal and and Yeah.

[01:14:09] Right. So,

[01:14:10] Paul Roetzer: I not only jumped the gun, but like claim some power that they actually don't have. Like you

[01:14:15] Mike Kaput: Right, right. We're like, he's posting so aggressively when the deal's almost done before this all blows up, and now it's just like doubling down on a mistake, maybe, I don't know.

[01:14:24] Paul Roetzer: right. Or just you're just gonna do harm either way. So you don't really care if it's legal or not, doesn't like what's the repercussions? Like, nothing's gonna happen to me if I do this and say this other than I'm gonna Hurt that company and try and use it as leverage to get them to do what I want them to do,

[01:14:37] Mike Kaput: Mm-hmm.

[01:14:37] Paul Roetzer: would not be a unusual political tactic.

[01:14:40] Mike Kaput: All right

[01:14:42] DeepMind’s New AGI Scorecard

[01:14:42] Mike Kaput: Next up, Google DeepMind has published a cognitive framework this week that attempts to answer the question, if AI actually achieved AGI, how would anyone know?

[01:14:51] So the team here proposes a cognitive taxonomy with what they claim are 10 measurable traits of general intelligence, on which [01:15:00] to measure ai and its progress towards AGI. And it's divided into two categories. So this first category covers eight building blocks of human cognition, perception generation, attention

[01:15:09] learning, memory, metacognition, and executive functions.

[01:15:13] And These combined to form two composite faculties that DeepMind considers equally important, which are problem solving and social cognitions. They basically define these as the ability to process and interpret social information and respond appropriately

[01:15:27] in social situations. So their proposed test here is pretty straightforward. They wanna run AI models and humans. Through the same cognitive benchmarks, and then they theorize you'd get a measurable estimate of

[01:15:39] when a single AI can meet or exceed human capabilities across all 10 of these areas. DeepMind actually launched a Kaggle Hackathon with a $200,000 prize pool to crowdsource evaluations for

[01:15:51] the five areas where the gap between testing capabilities right now is the largest, which are learning metacognition, attention, executive functions, and social [01:16:00] cognition.

[01:16:00] So they say their goal is to move the conversation around AGI from one of subjective claims and speculation towards a grounded, measurable scientific endeavor. So, Paul, does this change anything about how we talk about AGI? Are we getting any closer to really defining what it is and actually measuring it?

[01:16:18] Paul Roetzer: Yeah, I mean, Google's Deep Mind's done the best job of trying to get to that. Point. You know, they, they had a paper last year that Shane Legg led on where he was trying to sort of define the different, you know, general capabilities and performance and trying to put some way to measure it.

[01:16:32] Paul Roetzer: so like the effort to try and quantify it, making it more meaningful, meaningful, try and get some maybe eventually universal agreement on what it is.

[01:16:41] The first thing I thought, thought of when I saw this is like, well, how do you not saturate these tests? Like when the models eventually learn what the tests are and just like be able, I don't know how they would do that to keep 'em like sandbox so the model doesn't end up with the training data basically, that it eventually learns how to look like it has AGI because it just learned what the test was ahead of time.

[01:16:59] [01:17:00] But I think the most important thing for our audience is that we just keep coming back to this. AGI is a really interesting topic. It's fascinating to sort of follow along progress towards it. It's a meaningless term related to what it does to impact your job, your company, the economy more broadly.

[01:17:16] So. We don't need to reach AGI, whatever that definition is. We don't need to agree on a definition for AI to transform businesses, the economy and society. This idea of capabilities, overhang, we, we talk about, that Andres Carpathy episode that I mentioned touched on this quite a bit, but just go back to that example I shared of rocks like.

[01:17:35] If you have a company like ours that knows this stuff, we understand what AI capabilities are, and we look at a operating system of our company and we're like, oh, we're just gonna reimagine the whole thing rather than five rocks a quarter. We think we do 15 or 20 like easily,

[01:17:47] and here's how we're going to do it.

[01:17:49] So we understand the capabilities and we're applying them to the best of our ability. Then take some other company that doesn't even have gen AI tools for their team yet. They haven't even got 'em copilot [01:18:00] licenses or ChatGPT licenses, like they've done no personalized training. They've never run a workshop internally.

[01:18:05] Like they're not even taking advantage of any of the capabilities other than maybe using it as like an answer engine or a chat bot. So there's this overhang of we have all these capabilities and so few companies are actually doing anything with them, not just companies, educational institutions, governments, practitioners at an individual level.

[01:18:23] So that to me is the most important thing. So I 'm, I'm all for this. I think like quantifying it so we can just get to the point where we agree on what it is makes total sense, but don't be misled by. That, or wait around for that definition, be like, oh, okay, I'll worry about it when we get closer to AGI.

[01:18:38] It's already there.

[01:18:40] What 81,000 People Want from AI

[01:18:40] Mike Kaput: All right, next up. Anthropic has published results from the largest multilingual qualitative study ever performed on AI attitudes. They did nearly 81,000 interviews with Claude users across 159 countries and 70 languages. So these conversations were actually conducted by Anthropic interviewer, a [01:19:00] variant of Claude trained specifically to conduct and then analyze interviews, which we have talked about in past episodes.

[01:19:06] Interestingly, they found out that the top fear of people. Expressed in these conversations is actually hallucinations slash unreliability of ai, which ranks

[01:19:17] as the number one concern with 26.7% of people mentioning it. It is ahead of jobs and economic impact, which is at 22.3% and loss of human autonomy and agency at 21.9%.

[01:19:32] interestingly, Anthropic finds that people value AI often for the same capabilities that they fear most. So 50% of respondents experience time savings from ai. Yet 19% felt pressured to simply work faster. As a result, 33% cited learning benefits while 17% worried that it would actually facilitate more cognitive decline.

[01:19:53] When you're relying on machines to think for you, and it's interestingly that people experiencing one side of attention are typically three [01:20:00] times more likely to also worry about the other side, meaning these are kind of inherent contradictions in the same people using ai. Now, what's really cool is they actually asked what people actually want from ai.

[01:20:12] 18.8% of those who answered said that they seek first professional excellence from ai. 13.7% said they were seeking personal transformation. 13.5% said better life management and 81% report experiencing some progress towards their vision in those areas. Paul, I'm interested what you took away from this data.

[01:20:34] Pretty interesting way they went about getting it

[01:20:36] Paul Roetzer: The, yeah, it's the approach to research that I found most intriguing. So, I mean, the data's great. I do think, again, as I referenced earlier, you gotta keep in mind like, who, who are the people responding to these questions, things like that when you look at the data so you're not just like, you're not making some broad assumptions.

[01:20:53] in December of 2025, before Claude Code really took off and before the government issues, and [01:21:00] before like this movement where the Claude app became like the number one app on the app store, they, they have a heavy technical user base. Lots of like coders, lots of AI researchers using

[01:21:10] Claude. So when you're looking at this, even though it's 80,000 people across all these countries, it's still likely skewed toward a more technical user.

[01:21:17] So just for reference sake, that that's important to. In, in the back of your mind. So I love the approach this like dynamic approach based on responses that adapts it not great news for people who run focus groups and who are like consumer research people for a living.

[01:21:33] This is definitely one of those ones where you're either adapting or the whole holding way of doing research is gonna of run you over.

[01:21:40] they said their next Anthropic interviewer study launching shortly to a small subset of Claude users focuses on claude's effects on people's wellbeing over time, whether Claude is actually making people's lives better in the ways they want and how it could do so more effectively, which I thought was interesting.

[01:21:55] And then they said, this is a new form of social science that is a qualitative [01:22:00] research at a massive scale, and we're in the early stages of learning how to do it. Surveys and usage analysis tell us what people are doing with ai. But the open-ended interview format helps us get at a why. Conducting this research has moved us and challenged us.

[01:22:14] We did not expect so many deep, open and thoughtful responses. By far, the most common reflection from our team was that this, it was viscerally moving to see Claude impacting people's lives for the better and equally motivating to hear their concerns. We were equally gripped by the fears and downsides.

[01:22:30] People saying that the same availability. Making Claude useful is what makes it hard to put down, or knowledge workers worrying about outrunning, AI's economic impact. When you come into contact with this much raw human experience, it knocks you sideways. They said the usefulness is real, and the question for all of us is how to claim the benefits without incurring undue costs.

[01:22:49] I thought that was really interesting to note, Mike, because this actually came up during our company retreat, this idea that we're all sort of at the frontiers of figuring all this out and using it, and it's awesome for [01:23:00] productivity and innovation and efficiency and growth and all these things, but it also has this like very messy, complicated other side where it has this human impact and maybe at your friends or your family hate it and they don't even like the fact that you're working on it and they, they have these perceptions about what you're doing because you're in ai or because you're one of the people who talks about it.

[01:23:19] I honestly think about that sometimes from, you know, what we do on the podcast. I'm like, where I think about like, God, I hope it's some point, like people don't. We're trying to do the human-centered approach, like we're trying to educate people so we can have a positive outcome, but sometimes the truth doesn't matter and like, I do worry about that.

[01:23:38] It's part of the reason I don't read comments on social ever. Like, I don't, I don't look at our comments and YouTube and X and maybe sometimes LinkedIn, but I just prefer to like, try and just do our thing and like know we're trying to do a positive thing, but that doesn't, it

[01:23:54] doesn't, it doesn't change the fact that there's like darkness to this and there's uncertainty and [01:24:00] fear and anxiety and hatred and like, all those things are very real.

[01:24:03] So I'm really excited actually that Anthropics going this research direction.

[01:24:07] Mike Kaput: Yeah.

[01:24:07] Paul Roetzer: That's why I really actually like the findings here. Obviously to your point, they are skewed towards a certain type of

[01:24:13] people. But yeah, when someone asked at our offsite, like, how do you stay grounded when you're dealing with such heavy and sometimes horrible dark AI topics in the news?

[01:24:22] That was like my answer was that. Focusing not to the detriment of the negative, but focusing on the positive things that I've been able to do with these tools. Like I've been able to do things, achieve goals, get results that I never dream possible. And like it is genuinely as technology has made me a better professional leader,

[01:24:40] thinker, strategist, even husband and father. So that's kind of the flip side. So I love to see in this data people saying, Hey, I'm using this and I'm trying to get out of ai, professional excellence or personal transformation, or better life management. I've done all those things with ai and it is glorious what you're able to do. Yeah.

[01:24:58] Paul Roetzer: Doesn't, yeah, it doesn't [01:25:00] get rid of the negative stuff or like the concerns, but it, yeah, it's trying to focus on the positive. So I, yeah, it's, this is something we're, we're gonna come back to too. We actually, so, you know, we brought on a director of research a few months back and to one of the focus areas she has is actually on the humanity side of this, it is

[01:25:14] like to, you know, so we're actively, Mike and I, and and Taylor are actively talking about more research in these directions and the kinds of things around the human impact. And I, so yeah, it's something we're gonna probably be doing a lot more about. On the show and then even with our academy is it's starting to talk about that stuff.

[01:25:29] So very important. And maybe even on our event side, we might be looking at doing some stuff. We can bring people together to have these conversations 'cause they're critically important. okay. Well as we wind down, Mike, we had mentioned at the start the AI for professional services, which you taught as part of our academy.

[01:25:45] And so again, one of the ideas we have is to do little spotlights on these where we, you know, without having even to take it, we give you a little bit of insights into some of the key things we learned in building these courses. So, Mike, with AI for professional services, any like key insights or takeaways that you think would be helpful for [01:26:00] people to

[01:26:00] hear?

[01:26:01] AI Academy Spotlight

[01:26:01] Mike Kaput: Sure Paul, so as part of this, you know, four core series, which comes with its own certification,

[01:26:06] you know, we're breaking down both from a high level, what is happening at the industry level that you need to know about, and then getting into the actual tactical A to Z of here's how you identify your own use cases and match AI tools to them in your own professional services career.

[01:26:22] So a couple things that just jumped out as part of both building this course and as someone that was in professional services before we did the whole AI thing is number one, and we've talked about this on the podcast.

[01:26:35] one of these trends that really, really, really needs to be appreciated is the idea that the billable hour model is maybe not only on borrowed time, but is dead.

[01:26:46] Like if you were on a billable hour model, as a professional services, organization, AI is a major threat to that because

[01:26:57] many, many organizations still have not [01:27:00] adequately. Figured out what happens when you can now do things in a fraction of the time that you used to do them. In using ai, you cannot simply charge the same amount of hours and hope to get away with it.

[01:27:12] So you see a lot of industry professionals and leaders trying to figure out, how do we adapt our business model without tanking our entire organization? So

[01:27:23] one of the big takeaways there is the firms that are gonna win are the ones that figure out sustainable, defensible value based pricing first. So pricing on outcomes, not our, because again, you can do so much more in the same amount of time.

[01:27:38] There's no chance your clients, your customers are not going to demand that you pass along those savings to them. And then I would also say another big area here. Is figuring out how your human intelligence within your professional services firm becomes your superpower and your competitive advantage.

[01:27:58] Because [01:28:00] unfortunately, for a lot of professional services firms, there are very intelligent AI models out there that now have been for better, for worse, trained on a lot of your expertise. So

[01:28:10] figuring out how your humans, with all their experience and background and domain expertise can actually lever, it, be leveraged and scaled with AI is going to be the entire battle moving forward.

[01:28:24] So you really wanna look almost at any frameworks, any experience you have internally as almost like your own IP if you're not already, because AI can scale that and you can have that be a

[01:28:35] competitive advantage. But if you do not do that, if you are. Playing at the commodity level of, Hey, we're experts in marketing.

[01:28:43] Like so is AI now. So you have to figure out what kind of expert you are and how you are differentiated. And then last but not least, I would say there's always these questions in professional services about like, we'd love to get started with ai, but we work

[01:28:56] in really sensitive industries with clients that have privacy and [01:29:00] data concerns about using this stuff.

[01:29:02] We have not figured that out yet. Totally valid. We talk about that more at length in this course series. But the advice here is actually start with your

[01:29:12] back office stuff. If you have these kinds of challenges, if you are still trying to navigate data and privacy concerns. Your back office stuff, I guarantee you can become dramatically more productive by applying AI often at a very low hanging fruit type of level.

[01:29:27] We go into very specific use cases and tools in the core series to help you do that, but there are these areas where that don't touch client facing stuff that you can actually start your AI journey almost in the back office and achieve massive, immediate. Profitability gains just from doing that alone. So

[01:29:48] tons, tons more in the core series, Paul, but those are kind of three big takeaways there.

[01:29:52] Paul Roetzer: And the other thing I think about Mike is just from the buyer perspective, understanding the professional services and how it's evolving and how I should be looking for like AI [01:30:00] forward professional services firms. So even for me as the CEO, we outsource, legal IT accounting, advertising, like we work with an advertising part.

[01:30:09] So I just think about just those four, like understanding how their business models are evolving and the importance of working with AI forward versions of those companies and the, you know, our points of contact, things like that. So yeah. And it is great and I appreciate you obviously building this series.

[01:30:23] You know, this ongoing effort we're doing to try and sort of create content across all the departments, all the relevant industries, and then even into like businesses, try and make that stuff super relevant for people. So hopefully, like these little spotlights will be helpful for people to get a little taste of, what's going on in these different industries.

[01:30:39] We'll touch on departments, we'll touch on some of the gen app things we're doing, and just try and bring some of that value from Academy to the podcast each week.

[01:30:47] AI Product and Funding Updates

[01:30:47] Mike Kaput: All right, Paul, we've got a number of AI product and funding updates here to wrap up this week, so I'm gonna run through these. If there's anything that jumps out to talk about further, let's do it. But otherwise I'm going to run through these. So [01:31:00] first up, Jeff Bezos is trying to raise a hundred billion dollars fund focused specifically on AI manufacturing.

[01:31:05] This fund would represent one of the largest single pools of capital ever assembled around AI infrastructure. Google

[01:31:13] has launched something called Stitch, an AI design tool that turns natural language prompts into high fidelity UI design. So a tool lets you describe what you want in plain English and generate production quality design outputs.

[01:31:25] So Google is kind of in this emerging quote unquote vibe design category. Google also rebuilt AI studio from scratch as a full stack vibe coding platform. They said they spent actually four months on this rebuild. And the new version lets developers go from prompt to working application entirely within AI studio. openAI's has released smaller cheaper tiers of GPT 5.4, so GPT 5.4 mini and nano give developers access to the model family at lower cost and latency. In

[01:31:58] some other legal news, a [01:32:00] court temporarily allowed perplexity AI shopping agents to continue operating on Amazon. Perplexes agents browse Amazon on behalf of users to actually find and purchase products, and this ruling lets

[01:32:11] the service will remain live while their ongoing legal dispute, which covered on a past episode with Amazon, plays out on X.

[01:32:19] The company is rolling out AI generated article summaries that appear when users share links on the platform. Researcher Ethan Molik noted the irony that many of the articles being summarized are themselves, obviously AI generated, so we're creating an interesting loop where AI summarizes ai.

[01:32:38] Paul Roetzer: And then it trains the Grok language model,

[01:32:40] Mike Kaput: and then it trains

[01:32:41] Paul Roetzer: is part of the reason why they made articles such a prominent feature is to get a lot more training data that was, you know, proprietary to them potentially.

[01:32:48] Mike Kaput: Yep.

[01:32:49] And finally we'll be keeping a close eye on this one. Demis Hassabis of Google. DeepMind Nobel Prize winner is teasing his upcoming book called The Infinity Machine Set for Release [01:33:00] on March 31st. It covers

[01:33:01] the story

[01:33:02] of DeepMind and has Hassabis's vision for the future of ai. I'll be looking very closely at that one.

[01:33:07] Paul, that looks

[01:33:08] Paul Roetzer: Yeah. This one, I did pre-order this one. I, this is a good way to end today's podcast. I am actually gonna read the excerpt 'cause I think this is really fascinating. So. This comes from the in, what was it called? The infinity machine. Okay. It says, it says the true reason to build artificial intelligence.

[01:33:25] Hassabis was now saying, went beyond Kant and Feynman. The goal was to draw closer to what might be called God to the intelligence that may presumably have designed everything around us. Hassabis quote, I am first and foremost a scientist. My goal is to understand nature, but doing science is sort of like reading the mind of God.

[01:33:44] Understanding the deep mystery of the universe is my religion, kind of, we humans, we have these faculties. The world is understandable, but why should it be that way? I think there is a reason computers are just bits of sand and copper. HASAs continued [01:34:00] now sounding more urgent. Why should these combine to do anything?

[01:34:03] I mean, it's absurd. The electrons move around and then that creates an AI system that can defeat defeat a go master. Why should that be possible? This table, Sebastian Hassa has wrapped his palm on it for emphasis. Why should it be solid? This is beyond evolutionary con coincidence. We can build electron microscopes and interrogate reality down to the most minute detail.

[01:34:28] We can build systems that detect black holes colliding more than a billion years ago. I mean, what is this? What the hell is going on here? There was a pause, but Hassabis was not yet finished. I sit at my desk at 2:00 AM and I feel like reality is staring at me, screaming at me, literally screaming at me, trying to tell me something if I could just listen hard enough.

[01:34:48] That's how I feel every day. So you can see why I'm trying to build ai. I've felt that since I was very young, that there's a deep, deep mystery about what's going on here. You can frame it how you want. [01:35:00] You can call this God's design, or you can say it's just nature. I'm open-minded about the description and I don't know what the answers will turn, will turn out to be, but at the moment we don't really know what time is or gravity is or any of these things.

[01:35:14] So there's a mystery waiting to be solved and it encompasses just about everything I would like to understand before I croak. I would like to understand and then I'm perfectly fine to shuffle off my mortal coil.

[01:35:26] That was awesome.

[01:35:27] It's so good. Yeah, so that's, and again, as we've said on the show many times, like Demis thinks very deeply about this.

[01:35:33] Elon actually commented on that one. He is like, I, like I shared Demis''s like urgency here and thoughts here. So I think it's important to understand why one of the people, one of the five, why he's building AI and it is for a much bigger solve intelligence and then solve everything else. That's been his mission, you know, for the last 30 plus years of his life, 40 plus years of his life.

[01:35:55] Mike Kaput: Incredible. Alright,

[01:35:57] Paul, just one quick note here as we wrap up, go to [01:36:00] SmarterX.ai/pulse

[01:36:01] to take

[01:36:02] this week's survey. We're gonna ask a couple questions about the topics this week. One is about openAI's enterprise deployment with that private equity backing we discussed. The second one is about Anthropic study and some of the findings there and how you feel about them.

[01:36:15] So we'd love to hear from you and Paul really, really appreciate you breaking down everything for us this week.

[01:36:20] Paul Roetzer: Yeah, good stuff. Busy week as always. I think we just have one episode this week. I don't know, I didn't check my calendar yet this week. Maybe we have a second one, but, we'll be back next week and then I think I'll be on spring break then for like 10 days. So might next week might be, we might be on a break after next week.

[01:36:36] So yeah, thanks for being with us. Have a great week everyone, and, we'll, we'll be back with you next week.

[01:36:42] Thanks for listening to the Artificial Intelligence show. Visit SmarterX.ai to continue on your AI learning journey and join more than 100,000 professionals and business leaders who have subscribed to our weekly newsletters, downloaded AI blueprints, attended virtual and in-person events, [01:37:00] taken online AI courses, and earn professional certificates from our AI Academy and engaged in the SmarterX slack community.

[01:37:07] Until next time, stay curious and explore ai.

 

Recent Posts

[The AI Show Episode 205]: AI Labs Refocus on Agents and Enterprise, Trump’s New AI Framework, Meta’s Rogue Agent & What 81,000 People Want from AI

Claire Prudhomme | March 24, 2026

Episode 205 of The Artificial Intelligence Show covers the AI labs' race for agents and enterprise, Trump's AI framework, and Meta's rogue AI agent.

[The AI Show Episode 204]: AI Answers - What Should Stay Human, AI Pricing vs. Labor Cost, Leapfrogging Digitalisation, Getting Legal On Board & Do Reasoning Models Actually Reason?

Claire Prudhomme | March 19, 2026

Paul and Cathy answer 16 AI questions on career pivots, the death of billable hours, human creativity, agent swarms, and labor replacement pricing.

[The AI Show Episode 203]: Anthropic vs. Pentagon Round 3, NYT  AI vs. Humans Writing Test, Atlassian’s AI-Era Layoffs & Grammarly's Expert Cloning Scandal

Claire Prudhomme | March 17, 2026

This week on The Artificial Intelligence Show: Anthropic Pentagon lawsuit, AI writing quality NYT quiz, Atlassian AI layoffs, & AI job loss tracker.