AI is Losing Ground: Futurist Brian Solis on Why AI Adoption is Failing
56 min
•Mar 30, 202619 days agoSummary
Brian Solis, futurist and head of global innovation at ServiceNow, discusses why AI adoption is stalling in enterprises despite hype. He argues that businesses are automating existing processes rather than reimagining workflows, and that successful AI transformation requires vision, leadership, and cultural change—not just technology implementation.
Insights
- Enterprise AI maturity scores dropped from 44/100 (2024) to 35/100 (2025) due to necessary focus on governance, security, and compliance—a regression for the right reasons, not failure
- 95% of companies aren't realizing ROI from AI because they're automating digitized processes without reducing costs or creating new value; true transformation requires end-to-end workflow reimagination
- AI business reinvention requires IT and HR collaboration to manage AI agents as software assets, breaking roles into tasks and reskilling workers for higher-value work rather than eliminating them
- Organizations lack vision for AI transformation; most copy tactics without understanding why, similar to how 86% claimed digital transformation investment but only 25% could articulate the purpose
- A new role—Chief Workflow Officer—is needed to audit, assess, and theorize end-to-end workflows before architecture, bringing together IT, HR, and business stakeholders to reimagine processes
Trends
AI adoption regression in enterprises driven by governance and compliance requirements, not capability gapsShift from cost-reduction focus to value-creation focus in AI implementations; companies exploring new revenue streams rather than headcount reductionEmergence of AI agents as managed software assets requiring IT-HR collaboration and organizational restructuringCEO and board-level recognition that AI transformation is complex, slow, and requires enterprise-grade capabilities—end of 'sprinkle AI' wishful thinkingCultural and psychological safety becoming critical success factors for AI innovation, not just structural or technical changesBottom-up innovation models (advisory committees, incubation programs) outperforming top-down innovation centers in AI explorationOrganizational restructuring: IT and HR reporting lines converging; technology moving closer to CEO in forward-looking companiesWorkflow reimagination identified as #1 attribute for AI ROI realization, driving new organizational roles and processesAI mindset shift from 'how do we save money' to 'why do we do this and how might we reimagine it for AI'Narrative correction: enterprise software companies better positioned than AI natives for sustainable enterprise transformation due to governance, compliance, and integration capabilities
Topics
AI Adoption and Maturity ModelsAI Business Reinvention vs. DigitizationWorkflow Reimagination and End-to-End TransformationAI Governance, Security, and ComplianceIT and HR Collaboration for AI Agent ManagementOrganizational Culture and Psychological SafetyVision and Leadership in AI TransformationAI Agents as Software AssetsChief Workflow Officer RoleROI Realization in AI ImplementationsAI Mindset and Critical ThinkingDigital Transformation Lessons Applied to AIAI Native vs. Enterprise Software NarrativeInnovation Centers and Bottom-Up InnovationCapability Overhang and Power User Gap
Companies
ServiceNow
Brian Solis is head of global innovation; company published AI Index research showing maturity regression
OpenAI
Research on capability overhang and power user adoption gaps; platform studied for AI usage patterns
IKEA
Deployed AI chatbot Billy; example of leadership vision converting automation savings into new business (interior des...
JP Morgan
CEO Jamie Dimon articulated vision to become 'AI mega bank by 2030' with detailed transformation roadmap
Moderna
Moved IT under CHRO, signaling organizational restructuring for AI agent management and HR-IT collaboration
Nike
Technology rolling up to CEO level, positioning AI as business enabler rather than IT function
Puma
Technology reporting to CEO, reflecting organizational shift to treat AI as strategic business function
Orica
Australian mining company with AI advisory committee including external stakeholders; identified as AI pace setter
Amazon
Prime service emerged organically from organization, not C-suite; example of bottom-up innovation success
Google
Research found psychological safety in high-performing teams; example of culture-driven innovation
Zappos
Example of company with innovation-focused culture and brand; Solis worked on culture initiatives
Southwest Airlines
Historical example of company where culture was a brand differentiator
Virgin America
Example of culture-driven brand; Solis worked on flight attendant training and culture implementation
Walmart
Successfully scaled incubation and acquisition models for innovation
Meta
Referenced as Silicon Valley innovation destination that executives visit
Capgemini
Popularized digital transformation term; partnered with Solis on innovation research
Ultimeter Group
Firm where Solis was principal analyst studying emerging disruptive technology
McKinsey
Quantum Black entity published 'State of AI' research identifying end-to-end workflow reimagination as top ROI attribute
Tata Group
Indian conglomerate with culture requiring 10 hours/week for learning or improvement; innovation culture example
Gaping Void
Company dedicated to creating cultures of innovation and transformation using culture science
People
Brian Solis
Futurist and thought leader discussing AI adoption failures, business reinvention, and organizational transformation
Geoff Nielson
Podcast host conducting interview and asking critical questions about AI adoption and enterprise transformation
Bill McDermott
Coined term 'AI business reinvention'; vision-setting leader for ServiceNow's AI strategy
Dave Wright
Co-authored research with Solis on IT as HR of AI agents; Jensen Huang CES 2025 analysis
Jamie Dimon
Articulated vision to become AI mega bank by 2030; example of leadership vision in AI transformation
Jensen Huang
CES 2025 statement that 'IT will become the HR of AI agents' sparked research and industry discussion
Kevin Roose
Documented AI native entrepreneurs and agents in San Francisco; cited for firsthand observations
Rachel Sandell
Leading AI advisory committee at mining company; example of forward-thinking AI governance
Sir Ken Robinson
Champion for creativity in schools and work; quoted on necessity of being wrong to create original value
Quotes
"Disruption is doing new things that make the old things obsolete. Another way to think about disruption is doing things, whether consciously or subconsciously, in ways that change behavior."
Brian Solis
"The real disruption is one that I wouldn't say is named or realized yet that businesses do not know what they do not know."
Brian Solis
"95% of companies weren't realizing ROI because we are automating what was digitized. So it's hard to take ROI out of that if you're not reducing costs."
Brian Solis
"If you are not prepared to be wrong, you will never come up with anything original. And if you can't come up with anything original, you will not create new value."
Brian Solis
"The number one attribute for ROI was when a business was able to reimagine a workflow end to end with AI."
Brian Solis
Full Transcript
Hey everyone, I'm super excited to be sitting down with Brian Solis. Brian is a globally recognized futurist and thought leader in business innovation, currently serving as the head of global innovation at ServiceNow. What I love about Brian is that from innovation to digital transformation to AI business reinvention, he's been the leading authority on how disruptive technology is impacting society and work for a generation. What really sets him apart though is that he gets beyond the ivory tower thinking that some futurists fall into and is actually written and executed to playbooks to reinvent the way companies do business. I want to ask him how big the risk is that AI native companies torpedo today's leading businesses. What does the enterprise of the future actually look like? And what do we need to do and avoid to build it? Let's find out. Brian, thanks so much for being here today. Super excited to have you. One of the things I wanted to talk about right off the bat is, you know, AI is obviously disrupting, you know, consumer patterns, it's disrupting businesses, it's disrupting the markets. And one of the things that I've seen in my experience with the clients that I'm advising and that we're advising is that businesses by and large seem to be losing ground in this adoption of AI. And I wanted to get your perspective in terms of what's going right, what's going wrong, and what's kind of the state of the nation when it comes to how businesses are attempting to adopt AI and better use it in their business practices. Well, Jeff, let's just kick it off with a big, huge question. I don't want to save the big ones for the end. You got to get into it. Where do you want to start? I mean, look, the keyword you used was disruption. You talked about consumer, then you talked about business. That is the right word to talk about this. So let's have some fun. So disruption, the way I talk about disruption, I like to give it sort of a tangible meaning so that it frames the rest of the conversation. The way that I define disruption is doing new things that make the old things obsolete. Another way to think about disruption is doing things, whether consciously or subconsciously, in ways that change behavior, so whether you realize it or not, change patterns, change thinking because of that adoption. So the reason I give that additional sort of context is because on a personal level, AI is absolutely disrupting how people think or not think. And it is certainly stoking or even in some cases sparking new biases, meaning like AI sycophancy, which is something that a lot of people are experiencing, where for those who aren't following that term, it's where AI is constantly giving you accolades and compliments. And it starts to create subconsciously in many cases this false sense of confidence and validation that can lead into all kinds of crazy cases that the news media has covered. And then also just like what I call cognitive Darwinism, but is referred to as AI atrophy, where the more you, for example, high school students having to do their homework for them to save time, what they're doing is giving the thinking to AI and their force of reducing their capacity for critical thinking, creativity, etc. And then you carry that all the way into business and you have what open AI calls capability overhang, which is described as the capacity for AI to do X, yet most people only use it for Y. And in their research, they found on their platform, for example, that they experience power users who use seven X more capabilities of open AI than the rest of the pack. So that's a pretty large delta. And if you look at Kevin Ruse, who's a famous New York Times tech writer, he describes that in detail of what he sees firsthand in San Francisco with all of these AI native entrepreneurs, the agents that they surround themselves with, or we could look at the news and everybody who's using Claude Baudet, Mac Minnis and all the agents that they have doing their work for them. So anyway, it's a long way of saying that the real disruption is one that I wouldn't say is is is named or realized yet that businesses do not know what they do not know. And that disruption is is is to borrow the open AI's term, the overhang, that at some point, competition and disruption is coming for them in ways that they're not seen because they're focused on skills, upskilling, fluency, literacy, but not necessarily against that vision of what do we want to become because of AI versus how should we use it. So all of that makes sense to me and it you know, it frames it up really nicely in terms of the use cases right now, the art of the possible, maybe not even with future tech, but with what's available right now. Something caught my attention while I was, you know, doing some research on UBrien, which is this AI index that that you and ServiceNow put out recently, and that in the past year, there's been a regression, it sounds like, if I'm reading it correctly, in the business AI index, which means that businesses feel less able or ready to capitalize on AI than they did a year ago. That is, you know, startling and maybe a little bit surprising. What what do you make of that? First of all, do I have that right? And what do you make about that? And then maybe maybe more importantly, what do we do about it? Well, first of all, thank you for doing your homework. And also thanks for for plug in ServiceNow. Yes, I'm very proud of our AI index research. And one of the reasons why is because it's the analyst in me, where, for those who don't know, I used to be a principal analyst at a firm called Ultimeter Group, and we studied emerging disruptive technology, and made sense of it at a time where most analyst firms were just studying the base, you know, like the stuff they have to cover. And for example, generative AI would have been something we were ahead of and and and discussing at a high level. And so it's like right up our alley. And so that when I say the analyst in me, when generative AI hit, the natural first thing that came to my mind was to do what I had done with digital transformation to see if I could start to document stages of how organizations might adopt it, and then how that adoption starts to transform the organization internally. And so what we had come up with was a maturity model of five notable stages of progress for generative AI between the most advanced and the those like everybody else because it's it's changes hard. So anyway, that maturity model became the foundation for what later became the AI index where we were able to survey against it around the world, since we had these these models to understand where people were in their journey. So with that setup, what we had found was that in 2025, the average score for AI maturity with 100 being the most mature was 35 out of 100. In 2024, the average score was 44 out of 100. So it was a it was a drop of nine points. And what we really sought out to understand was well, what what happened in that year, where companies might regress. And in that one year alone, we saw rapid evolutions of AI models, or you know, frontier models. And then we also saw AI agents start to become a thing. And also the concept then of an agentic enterprise, so it was just so much so fast that companies had to take a step back and say, Wow, AI natives all over it, because that's the foundation that's our DNA. But for other organizations, where things like governance becomes a huge factor, trust security risk, these things required organizations to better understand at a foundational level, how do we need to reorganize for this stuff, so that our organization isn't caught off guard, as we go further down this path. So it was it was a regression, but for the right reasons. If you work in IT, InfoTech Research Group is a name you need to know. No matter what your needs are, InfoTech has you covered AI strategy, covered disaster recovery, covered vendor negotiation, covered InfoTech supports you with the best practice research and a team of analysts standing by ready to help you tackle your toughest challenges. Check it out at the link below and don't forget to like and subscribe. It's notable to me as well that you said the score in 2025 was 35 out of 100, this wasn't a regression from like 95 to 86. This is like we're deep in the red, we're talking about like an F on like a grade point scale. Right? I mean, one way to put it. I mean, I'm struggling to look, I'm familiar with the five point scales and it's very typically the way they're constructed is it's pretty tough to get like a five out of five on them. But we're seeing a playing field where there's just not a lot of organizations, except for the AI native ones that are really excelling here. And so maybe to just back up for a second before we talk about the agentic organization is how big is the concern level, how big is the risk that AI native firms are going to swoop in and eat the lunch of a lot of these, if I can call them more legacy or established firms that are not implementing these technologies in a material way? There's certainly in the absence of understanding how things actually work and function and why there's certainly a narrative out there that says here come AI natives, they're coming for all of the legacy players. But in reality, there are just like a maturity model, there are layers. And so if we, if let's just take a step back to kind of set the conversation for everybody. If we look at a model and you had used the grade F for 35 out of 100, the way that I would want to think about it is if you're learning how to ride a bike or if you use the traditional analogy of crawl, walk, run, you know, you're grading, you know that ultimately someone is that running is the ultimate achievement, so that's the hundred. And so you're grading sort of someone's progress. And it becomes an issue if that progress becomes slower than the rest of your peers. And that's how we look at it. But the fact that there's progress and so quick in terms of adoption. So just as a reminder, generative AI hit in 2022. So with that context set, if we look at why that regression happened in terms of the score, it was because companies have to protect themselves, whether it's a matter of regulatory compliance, whether it's compliance in general, whether it's reporting, whether it's security, like these things matter at enterprise grade. If you're an AI native startup, you're you're not thinking about those things, you know, if you're if you're worried about auditing, if you're a publicly traded company, these are things that I think we we just take for granted because we don't know what we don't know. So with all of that said, enterprise software gets an A in those things for a reason because it has to you can't run your business on this stuff. If those things weren't covered. Now, where where the conversation gets interesting is now the nuance of what are we going to do with AI? It is one thing at an individual level. It is another thing at an agentic level, because now what you're talking about is not just automating how work flows, we're actually empowering software to take on tasks that might normally be done by a human employee. And so this brings up much bigger conversations than just the capacity of software. Now we're looking at human resources and IT working together to better understand now how software collaborates with people. And it's not just a matter of okay, this company is doing really cool things. Let's put it in there. It is now a bigger conversation of breaking roles down into tasks, understanding where those tasks are actually holding people back from their own professional progress, and then not only dividing in concrete, but starting to understand what can we do with that freed up time and those freed up resources to achieve what we didn't know we could do or what wasn't possible yesterday. And then also how do we do it in a secure way, in a governed way that does not expose the company to unnecessary risk or compliance issues. So these are really big conversations. And I don't know that they're sexy headlines, but they're reality. Well, and I completely agree with that. And certainly for a lot of our audience who is business leader and technology leaders, it's a really critical conversation to have. And maybe not one that's being sufficiently understood or is taking a little bit longer to understand. And one of the things that's been on my mind, and it feels like we're starting to have like I come to Jesus moment around it, is in 2023 and 2024, what I had heard is at the CEO and board level, there was sort of this wishful thinking around AI that it would be easy and you could just sprinkle AI into your business and suddenly it would be transformed, which I have to laugh at because I have a background working with IT and CIOs. And I've never met an enterprise grade tech project that was easy or took substantially less time and less cost than people expected. And it feels like we're finally starting to acknowledge at that enterprise level that as you said, oh, we need all these enterprise grade capabilities around risk, around governance, around integration. So before we talk about the how and what exactly that looks like, one of the challenges here is, you know, when you compare this to riding a bicycle or, you know, some of the technology implementations of the past, one of the things that strikes me as being unique about AI is there isn't necessarily a defined target state because the technology is changing. And in some ways, the technology is, you know, it's a means to an end versus you're doing it for this specific reason. And you talk about the agentic enterprise. And so, you know, at least in 2026, what end state or what capabilities should organizations be building toward? And what's sort of the right way or the wrong way to pick your North Star here? I love these questions because essentially what you're getting to, at least the theme that I'm picking up on throughout your questions is this idea of what Bill, our CEO would say, Bill McDermott for those who don't know, is what he would call AI business reinvention. That's essentially what we're talking about here. And or it's what we could be talking about here. And I say that because as someone who, for those who don't know, I wrote some of the original research around digital transformation, with all credit due to Capgemini, they were the ones who popularized the term in the early 2000s. My research was some of the first to look at not the IT implications of digital transformation, but the business opportunities for it. So, for example, the working hypothesis would be like if we believe mobile and digital and social media will become prominent mainstream, human level, transformative platforms, then how might a business reimagine parts of itself or a whole of it, you know, as a whole, in order to do business natively for a digital customer or a digital employee. And so it framed the conversation under the assumption that digital was going to transform the outside, therefore need to transform the inside. And I would work with the same assumption that AI is going to do the same thing, hence the maturity model that was originally created. So I say that because one of the things that I had studied in digital transformation was I think I remember a stat from one report that 86% of organizations around the world reported that they were actively investing in digital transformation, but only 25% of them could tell you why. And the reality is that like AI and like digital, back in digital transformation, I used to say we weren't transforming with digital, we were digitizing the existing business. So modernizing whatever terms you want to use, things that might even date back to analog processes and systems. So with AI, you could say that some of the same patterns are already starting to exhibit or are exhibiting. So for example, the famous MIT report last year that found that 95% of companies weren't realizing ROI, it is because like the digitization we are automating what was digitized. So it's hard to take ROI out of that if you're not reducing costs, like for example, whether that's headcount, whether that's software, whatever it is, you have to you have to realize something back from it. So one of the one of the interesting conversations that our team has is we're looking at how can AI save costs and resources by automating existing stuff. And then because it is, it is a frontier technology, how can we use it to do what wasn't possible yesterday to create new value? So it's this automation, augmentation, this iteration, innovation, conversation. And it's one I believe that you could you could make the argument that with digital transformation, it wasn't as profound, but with AI it is. And so now the conversation is a matter of vision. It's a long way of saying that this is vision and leadership, not just traditional IT, looking at existing infrastructures to take out costs and to increase efficiencies. It's part of it for sure. But I think that this is a bigger opportunity for technology leaders within the organization to say, Hey, we're not just here working in the back end, and we're not just here to take out costs. We are a growth engine opportunity now, and we need to work closer to business leaders in order to define what we can become with AI, not just use it as a tool or a cost takeout mechanism. I completely buy that. And the vision piece is really interesting to me, especially through the lens, Brian, that you just set up of digital transformation, because looking at digital transformation, and I imagine your experience was similar to mine, there were a lot of organizations that were deep in digital transformation, but shallow on the vision around digital transformation. It's a lot easier to say you're doing it than to actually have that compelling business behind it and be reinventing the business in a really big and valuable way. From your time in that space, what are the lessons learned that organizations should be thinking about? If they're really serious about what you and Bill was calling AI business reinvention, where does that vision come from? And what do you have to get right if you're going to move beyond potshots and small fixes to something actually truly transformational? I love this question. And we, as we were building out the initial model for the maturity assessment, vision and leadership was there from the beginning, and it was not there because I had seen it in AI implementations at the time. It was there because I had not seen it in the years of digital transformation. And so I knew that that would be a pillar at some point. So I'll give you an example. It's a couple years old, but it's still, I guess it's a great example to show you, even though it's a couple years old, it is how rare vision is right now. So Ikea famously deployed an AI chatbot named Billy after its most popular bookcase. And like everybody, an AI chatbot is meant to take on level one customer service engagements and try to solve every instance as best as possible and minimize elevation towards a human agent. And it was so successful, and forgive me if I get this statistic wrong, but it was something like this compelling like 57% effective in level one engagement. So it resolved those cases without human escalation. And most companies would then say, all right, wow, 57%. That's a whole lot of human agents we're just not going to need anymore. And the ROI in that implementation would come from reducing those heads. And that is a popular narrative and fear with AI today. But this is where technology leadership comes in in partnering with the business by saying, I don't know if this was the conversation that was happened, but I'll I'll act it as if it were. Well, what were the 43% that couldn't be resolved by Billy? That's that's a pretty staggering number of cases it couldn't resolve. Let's study what those were. And what they had realized in that research was that those been many cases, not all of the cases, but a notable amount of cases were inquiries about interior design in commercial and and residential applications. And the long story short is that they spun up an interior design consultancy within IKEA, reskilled those agents to now become interior designers and apply to service fee for that. And ended up in the first year, I think that it was generating one billion euro net new revenue. And so that is an example of leadership. I don't know that it was vision because it was more of a reactive thing, but it was a it was a leadership opportunity to grow the business. On the vision side, you have someone like Jamie Diamond at JP Morgan. And he famously and his executives famously put out a vision like a lighthouse last year that says we are going to become the AI mega bank by 2030 or whatever what whatever it was that they said, and they detailed how they see themselves getting there. Now that's vision. Now, you know, someone will say, Well, it comes down to execution, of course it does. But we don't see a lot of vision as to where we're going to transform or what we're going to become differently as a result of AI. And so when you take the IKEA, the JP Morgan examples, what we're looking at now is this synergy. Maybe the wrong word, sorry to use a buzz term, but this relationship between technology and now business leadership, because together, we're going to be able to now rethink our approach to technology. That's that's really interesting to me. And it's exactly where I want to focus, especially because it's an it's an area that traditionally has been fraught with tension, with distrust, with, you know, issues with execution. And so, you know, you made sort of a, you know, a throwaway comment, Brian earlier about if organizations are going to reinvent themselves in this space, we need it and HR much more fully involved than they've been in the past. What does that look like to you? Like just sort of broadly, what what's the what's the process here? And what are the roles for, you know, IT versus HR versus, you know, the CEO or executive leadership versus any other, you know, group that needs to come to the table here? You're, you're not giving me softballs, are you? This, we're, all right, now we're getting into the construct of business reinvention and actually what that looks like. So let's let's bring it back to leadership first. The reason why leadership is such a vision and leadership are such important pillars in all of this is because we're essentially navigating uncharted territory. And and I say that literally, I don't take it lightly that there is no playbook for this, though if you Google AI playbook, they are countless playbooks. I just mean that there isn't a blueprint for what the business of the future can become with AI, even if it's a legacy business or an incumbent. So that means that every business, depending on leadership vision and culture of how that organization makes decisions, takes risks, executes, etc. balances iteration innovation balances the known with unknown, everybody's going to attack this differently. So and there's nothing wrong with that, as long as we're making progress with this. But I'll give you an example. My, my boss, his name is Dave Wright, he's the Chief Innovation Officer at Service Now and together, we wrote some research around the concept of when Jensen Wang in at CES in 2025. So this is this is already old in the world of AI, but it was still provocative and still meaningful today. Jensen said something along the lines of IT will become the HR of AI agents. And a statement like that from someone like that is going to be scrutinized. And it was it made headlines all around the world. But what I had what we had noticed was that there was no real dissection of that statement, no real critical thinking around it, like, well, what does he mean? What does that look like? Why would he say that? And what we had found in the in the paper that we were, we were, and this is publicly available, by the way, so if you if you want it, you can, you can just email Brian to Brian Solis, and I'll send it to you. What we had found was that it requires for the first time, let's just take it from an IT perspective, and then we'll look at at the broader collaboration around your question. An AI agent is let's pretend it's not let's let's pretend it's take its term AI agent because that has human implications in its name. Intelligent software that can accomplish tasks on its own and theoretically learn from that execution in order to optimize itself every time it runs those tasks. If you put agents together, then they're handing off tasks to one another to then together drive towards an outcome. Now, they don't just magically appear, even though you can download them from your place or buy them from your place of choice, they have to be tuned to the specific way your organization works, the way where your data is is archived and stored, all the things that a human being, for example, would need to understand before it's allowed to go off and start doing things. So there's training with HR, there's onboarding. Even before that, there's identification that this is a resource we need. So in HR, we might hire for it, we might then skill and train for it, we might then deploy, manage and assess, etc. So in the in the AI agent model, someone's got to do that. And so that's that's what we we believe Jensen meant by that. But that just that doesn't necessarily just happen because IT realizes that there are tasks that can be automated. The bigger idea was that we explored is then how does IT work with HR to then assess what those tasks are within someone's role, and then build out the framework from there, the model from there, the management from there. What we learned after we published that or what we had, what we what we realized as we were continuing this research was that essentially, like in the HR world, how we manage people, how they're in an org chart, etc. How we have performance reviews, and those are documented. AI agents are essentially software, and software is an asset. And there is software asset management in the IT world. And so agents also then become managed as assets. So this Venn diagram really starts to take shape for these forward looking companies that are thinking about it beyond just the simple automation route. So then last but not least, you you have now some real world examples of where this is starting to take shape. So Moderna quite famously moved IT under the CHRO. And now that's a one of one. And she's highly, she's a highly technical CHRO. Let's let's make that clear. But it is an example of what is possible. So as a futurist, that is a signal, I want to understand the signal, what could this look like over time, you have in Nike and Puma's case, technology rolling up to the CEO now, because they see AI as a business technology as a business enabler. So my 2026 prediction was that in reality, regardless of how this stuff plays out, individually at an organization, what will be consistent is this collaboration between HR and IT, especially as agents start to become more sophisticated in their ability to do more than one task. It's an interesting prediction. And it makes sense to me that they would be working together. And I love the model of IT thinking of agents as assets and running that. And the two groups collaborating together in a big way. It seems like it could be fraught with risk without the appropriate sort of front end to that entire process, right? Like someone needs to be answering the question about how do we do the things that we do here, right? Because like, you have to, you have to decide, does this require an agent or intelligent software? Or does this require a human? And, you know, like, how do you triage, like, now I'm thinking about business architecture. Like, how do you triage that? And who answers the how do we do the things that we do here? That's, so what I love about your question is that it is exactly the reason why there is no true playbook. Because that is one of many questions that need to be answered, or even thought of, right? What what we're seeing right now, one of the reasons why maturity isn't just accelerating through the roof is because one, we have to be cautious. But number two, we don't know what we don't know. And when as human beings, when we don't know what we don't know, we lean on what we do know, which is years of proven expertise, experience against failures, successes, etc. And so this is why we see early adoption going against the things we do know. Your question begs, what do we not know? And what should we be thinking about? And so we call this an AI mindset, we publish the paper on this too, it's asking the right sets of questions in the right context. So in one way, it's how do we save money? How do we scale? How do we make this more efficient, etc. And the other one is why do we do it this way in the first place? How might we reimagine it for a world with AI when when this workflow was designed at a time where that didn't exist? And so when you balance both of those mindsets, you now start to unlock new opportunities, you now start to create the blueprint for whatever your playbook is going to be, as you start to think about and solve for these things. So one of the things that Dave and I have been thinking about in depth is was inspired by Quantum Black, which is a McKinsey entity, they published a paper last year called the state of AI, and there was one line that really struck me, because I'm in the workflow business, which is of all they had come up with 25 attributes that studied where businesses were realizing any kind of returns on AI investments. And they said the number one attribute was when a business was able to reimagine a workflow end to end with AI. So essentially, forcing all the tough questions, forcing all of the new things, but in an in a contained environment. And so what we realized, and this comes back to your question, sorry, these answers are so long, because your questions are so deep. Who's going to figure that out? Or who's going to think about that? So right now, it's dependent on the company, if anyone is asking those questions or thinking about it. But we what we had realized is, if we believe then that the number one attribute for ROI is going to be end to end workflow reimagination, then someone's got to own that. And that role doesn't exist. So we playfully came up with the term the chief workflow officer, that is someone's job now, who is going to audit and assess. And let's say theorize before we architect, then what that architecture could look like, and then bring the right people in, whether it's it HR, etc. to then reimagine now what that workflow could look like. That that is the extent, at least initially, to which we can now start to reimagine or reinvent a business. I've never heard a chief workflow officer before. And I like it. And I'm chuckling to myself, because you're in the workflow business, obviously. So of course, of course, there's a chief workflow officer. But aside from that, I do like the idea. And I do see I do see merit to it. But I wanted to, you know, I wanted to tease out something there, and get your take on it. And again, you may you may have an inclination based on the business that you're in. But there's there's sort of this implication that the way that this is designed is top down, or it's centrally managed, you have, you know, an office of business workflows, revision coming from the CEO. To what degree is should it also be bottom up or emergent from different parts of the organization where people put up their hand and say, I think we can make this workflow better, whether they're in the workflow, whether they're managing the workflow, or does this really truly need to be top down to succeed? I'm going to answer that with the answer that no one likes, which is it depends. It really does. So I'm someone who studied also innovation. So my whole career, I've worked with startups, I've worked with, I've helped establish some of the world's first actual not r&d, but actual innovation centers. I wrote research on how those innovation centers were and weren't successful. And what was consistent across that was when bottom up was a factor in order for gaining success, not just in bottom up within the organization, but also from the outside in. So this is where you see things like what Walmart has successfully scaled, which is incubation, acquisition, separation of acquisition between the mothership, and you know, a lot of these lessons were learned the hard way. So for example, in our research already, there is, what every company starts to do as they mature is they create a center of excellence, if you will, that goes by different names. But what we're looking at then is bringing together all of the key stakeholders now to ensure that they're doing the right things at every step, governance speed, one of them, enablement training, etc. be another one, a place where questions can be asked and sorted and answered is another. But there is a company out of Australia called Orica, they're a mining company. And Rachel Sandell is the person sort of leading that AI initiative over there, they have something beyond a COA, they have what they call an AI advisory committee, which is staffed by, of course, some internal stakeholders, but also external stakeholders. And it is for the reason you ask it is to make sure that it is a place where the right ideas or questions can come through and be vetted and also explored without the usual politics or bureaucracy of just everyday business. And in that case, they are what we call an AI pay setter, they are so much further ahead than everybody else, for the very reasons that you're looking at, because now they're exploring things that they might not have considered, because those ideas are coming from new places. It is actually just to add an aside, it is how if anyone wants to geek out, it's how Amazon Prime came to life. It did not come from the C-suite, it came from organically from the organization. Interesting. And I tend to be in general a pretty big advocate of that. And like you, I've got a background in innovation, and I've not just internally, I've done internal corporate innovation, I've done advisory with other firms in innovation. And one of the things that I'm a little bit skeptical of, but I'm curious in your perspective, because it seems like some of these models are flirting with it, is having a monolithic innovation organization within the broader organization. And that may or may not be how you're structured right now. But one of the challenges I've found with those models is that you're sufficiently divorced from the actual workflows, from the business itself, that it can create either relationship friction with the people that are running it, or just feeling a sense of distance from the actual workflows that you're trying to ultimately influence. Has that been your experience, or how do you best conceptualize of innovation that works versus that ends up in kind of a failed state that doesn't have the desired impact? I mean, it's all of the things. You know better than anybody that sometimes this stuff works, and sometimes this stuff doesn't work, and innovation is one of those loaded terms that I think you'll appreciate this. Everybody wants it, everybody talks about it. Certainly, you know, since the 90s and early 2000s, innovation, Silicon Valley, I mean, all of these things become, I don't know, everybody goes, I used to call it the Disneyland tour, everybody goes to Silicon Valley to visit Google or visit Metta, or we even host executives in our innovation center in Santa Clara, where they just want to touch Silicon Valley-ness, you know, I don't know if it's in the water or if there's it's in the air, because innovation is that, I don't want to call it a North Star, but it is like that thing that should be a North Star. But what it really comes down to is leadership, vision, and culture. So, culture was not something I set out to study, but it became something I had to study in order to understand how to make a culture ready for innovation. And the way I describe innovation is not just the allure of Silicon Valley, it's just really specific. Innovation is doing something new that creates new value, whereas iteration is doing what we did yesterday better, faster, cheaper. And so you need a balance of both of those things. A culture of innovation is one that allows for, and you use the right words, failure, taking risks, counterintuitively investing in failure and the acceleration of failure, because that term is so loaded, and that term is so it's become like stigmatized that you can't take risk for the fear of failure, so you do the right thing, or you do what you think is the right thing by iterating on a proven model, and if that fails, you still have something to fall back on. But one of my favorite thinkers of all time, his name is Sir Ken Robinson, and he was a champion for continuing creativity in schools and continuing creativity outside of school, like in our work, for example, where things like creativity aren't as celebrated, risk isn't celebrated as much as say when we were kids, we would just do whatever, we would think whatever, we would explore whatever because curiosity was an essential skill for us. And what he has said was, if you are not prepared to be wrong, you will never come up with anything original. And if you can't come up with anything original, you will not create new value, was essentially what he was saying. So culture then becomes essentially what people would understand as organizational psychology as essentially safety. Is it safe for me to have this to ask this question? Is it safe for me to propose this idea? And do I have a support system around me that allows for me to further that conversation? So some examples would be, I remember when one of the earliest ones that I found was the Tata group out of India. They culturally said, you have to spend 10 hours of your 40 hour week, if we believe that 40 hours was all they were, or we work, 10 hours either learning a new skill or thinking about how to improve something. But then everything above that managers, performance reviews, managerial reviews of themselves, etc. Everything then had to be systematized to ensure that those ideas were explored, vetted, deployed, or cascaded, or for the next thing to come through. So I say all of that to say that culture and organizational or psychological safety are things that ensure that innovation becomes a thing. Last example I'll give you is Google, who is I think we could both agree that is an innovative company. They studied why their high performers outperformed all the other high performers. And what they had found, they thought it was school, they thought it was leadership. And of course, those are probably factors. But what they had found was that those teams felt psychologically safer than the rest of the teams in order to do these things, to explore these things, to test and iterate and fail on these things. And so one of my favorite companies out there is called Gaping Void and they are dedicated to creating cultures of innovation and cultures of transformation, all vetted in what they call culture science, because it is a must in addition to all of these things that we're talking about. So the conversation around innovation very quickly pivoted to innovation being much more of a cultural force than necessarily how you structurally design the organization, which I really like and it makes a lot of sense to me. One of the challenges, and this is, I'll preface this by saying this is going to be a difficult question. One of the challenges a lot of organizations and a lot of organizational leaders face is they don't have that culture of innovation. It's not in their DNA and maybe their executives go to the Disney world of Silicon Valley, but then when they come back, they say, everybody needs to innovate. Oh, but by the way, that's subservient to the other 100 things you're doing for business as usual, and they functionally deprioritize it. And so if you were call it a technology leader or a non-executive business leader in an organization like this that suddenly has a mandate to bring in AI, which we've already said has no playbook, right? There's no like, follow these three things and tada, you'll have AI and it'll reinvent your business. Is it a fool's errand? Is a culture of innovation a necessary precondition to being successful with AI? And if you don't have it, should you give up or is there some other road you should take on that AI journey? Oh, you're right that it's not an easy question. And I don't think it should be because if it were an easy question to answer, then everybody would be innovative, right? So the way I would think about it is it's this way. So even before AI, when I used to publish these studies on monitoring and answering the questions you're asking just in general, how can corporations be more innovative? I had partnered with Capgemini at the time. I think we wrote like three or four or five reports on this. They became the first reports that really started to answer this. And there was no one answer. What we had found was it was highly dependent on the culture of the organization. And in those cultures, though, there would be different models that would work. So think Innovation Center, think External Innovation Center, think Incubation and Ventures programs. There were there were so many different models, but the successful ones always came down to leadership, and then the culture that that leader set for. And so I want I want to just give cultures also a loaded term. So what it isn't is a vision statement, a mission statement, value statements, that is not what a culture is. I think a lot of organizations sort of mistake those as culture theater. The the culture is how someone might define in any part of the organization, not just what we do, but where we're going and why and that everybody agrees that this is the right way to go. And that as a result, there has been sanctioned behaviors. And those behaviors become norms. And then those norms are what's measured celebrated, etc. And so in a more innovative organization, asking questions, supporting those ideas, that's just a norm. And those norms have to be established. And that's these are things we don't we don't we don't talk about. So for AI to do what what we know it is at least possible to do in terms of AI business reinvention, and there has to be an articulation of this. And it doesn't have to be like this is what it's going to do. This is what it's going to look like. It could even be like I don't know. But what I do know is that these AI natives coming out of Silicon Valley are completely reinventing their business with AI to agent ratios already. We don't have that. Nor are we close to that. So what I as a leader, I'm committing to is giving you the resources and the safety nets, and and whatever you need in order to start answering those questions of what do what does good look like with AI, what does great look like with AI, and I don't know the answers, but we're going to create a space of which then we can explore what comes back with that and do something with it. And then the rest of the organization then has to now bring that to life, the norms, the values, the measures etc. Like that's big stuff. That's not easy. That's not overnight. This is why I'm sure if you talk to Jamie Dimon, and you ask him, Hey, so when you said you're going to be a 2030 AI mega bank, how are you doing that besides the AI investments? Like what are you doing differently with your leadership? How are you? What are you doing differently with human resources? What are you doing differently about performance reviews? What are you doing differently about the word failure? How are you allowing people to explore and experiment without repercussions? How are you promoting or incentivizing around failure? These are real questions. And so if you think about like the companies where culture was a brand like Zappos or for many years Southwest Airlines or one of my favorite examples, because I got to work on it, well, I worked on Zappos too, which was Virgin America, well, it existed. These are companies that had leadership that could answer all of those questions. I can even give you specific examples of like with Virgin America, how flight attendants were trained and how that came to life in the airplane environment. But most companies don't have these types of conversations. What is one narrative around AI that you're hearing a lot in the media these days that you think is complete bullshit and people should completely throw out the window? I think I would speak for the entire enterprise software industry and saying that one of those narratives is that AI natives are going to eat enterprise software. I think that is not only an incorrect narrative, it is doing it to service to executives who are really trying to navigate uncertainty and complexity. And so it would, it is, it is something that is, it is what it is. But at the end of the day, it only incentivizes us and me and my work personally to be the clarity, to be the voice, to be the lighthouse of thinking through these tough questions and also thinking through the questions that aren't being asked to have more meaningful conversations at scale. And if I can, if I can find a PR platform for that to counter those narratives, at least in the public spotlight, that would be even more helpful. But if it's a message that I want to send to everybody is that you have people actually thinking deeply about this and what it means to you so that you can safely, securely, with all the right reasons, no matter how slow or fast it is, transform in the right way. I love that. And, you know, I'm inclined to agree with you as well. And another time we'll have to have a much more fulsome discussion on that. But for now, Ryan, I wanted to say a big thanks for coming on the program. I really appreciate your insights. Oh, Jeff, thank you. I mean, obviously we could keep going, but hopefully this isn't the last time. So thank you for this opportunity. If you work in IT, InfoTech Research Group is a name you need to know. No matter what your needs are, InfoTech has you covered AI strategy covered disaster recovery covered vendor negotiation covered InfoTech supports you with the best practice research and a team of analysts standing by ready to help you tackle your toughest challenges. Check it out at the link below and don't forget to like and subscribe. Thank you.