"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis

Bringing AI to Data: Agent Design, Text-2-SQL, RAG, & more, w- Snowflake VP of AI Baris Gultekin

99 min
Jan 14, 20263 months ago
Listen to Episode
Summary

Baris Gultekin, VP of AI at Snowflake, discusses how enterprises are deploying AI at scale by bringing AI to data rather than moving sensitive data to external model providers. The conversation covers the massive unlock of unstructured data, the current state of text-to-SQL and RAG systems, and how reasoning models have made natural language data analysis reliable enough for business users.

Insights
  • 80-90% of enterprise data is unstructured and was previously trapped in PDFs and documents, but AI is now making this queryable for the first time
  • Text-to-SQL has reached production quality with reasoning models, enabling business users to directly query data without going through analysts
  • The 'bring AI to data' approach solves data residency and governance concerns by running inference within the enterprise security boundary
  • Enterprise AI adoption is shifting from experimentation to ROI-focused deployment with strict security and governance requirements
  • Product development is fundamentally changing due to AI coding assistants, enabling rapid prototyping and validation before building full products
Trends
Democratization of data access through natural language interfacesShift from specialized AI models to frontier models for most use casesOpen standards development for semantic model interchangeIncreasing competition reducing vendor lock-in across AI platformsAgent-to-agent communication protocols emerging for enterprise workflowsHorizontal AI applications winning over vertical specialized solutionsRapid acceleration in software development cycles due to AI coding assistantsGrowing importance of data governance and security in AI deploymentsTransition from structured workflows to more autonomous AI agentsEnterprise preference for data residency over cloud-based AI services
Quotes
"80 to 90% of all data is unstructured data. And you know, because there weren't a lot of easy ways to process this, it was not necessarily seen as the most usable data and it is now very usable"
Baris Gultekin
"Our customers do not want to replicate data in multiple places. Instead they want to bring AI to run next to data"
Baris Gultekin
"The quality is at a place where you can now deploy them very broadly. So very high quality and very useful because the structured data is quite useful"
Baris Gultekin
"I think the way we build products has to change given the coding assistance. How you deploy quickly, how you test things quickly is changing because of how capable these coding assistants are"
Baris Gultekin
"The faster enterprises adopt AI, the more benefit they're going to get and the more intuition that they're going to get that changes the trajectories of these businesses"
Baris Gultekin
Full Transcript
3 Speakers
Speaker A

Hello and welcome back to the Cognitive Revolution.

0:00

Speaker B

Before we get started today, a quick final reminder. If you dream of a career in AI safety research, the deadline to apply to MAT's Summer 2026 program is January 18th. Listen to my recent episode with Matt's executive director Ryan Kidd for all of the reasons that you should consider applying and then get started@ Matsprogram.org d today my guest is Baris Gultikin, Vice President of AI at Snowflake, the cloud based data platform that now describes itself as the AI Data Cloud. Barish came to Snowflake along with Snowflake's current CEO Sridhar Ramaswamy as part of Neva, an AI powered web and personal knowledge base search engine that Snowflake acquired in May 2023. Since then, he's been working at the intersection of frontier AI capabilities and hard enterprise realities, deploying these systems in environments where security, governance and reliability are strict requirements. As you'll hear, Snowflake's core philosophy is to bring AI to the data rather than sending sensitive data out to model providers. And in this episode we unpack exactly what that looks like in practice. We cover a ton of ground, including the massive ongoing unlock of unstructured data, which is making the 80 to 90% of enterprise information that was previously trapped in PDFs and other documents queryable for.

0:03

Speaker A

The very first time the current state of both text to SQL and RAG.

1:33

Speaker B

Systems and why reasoning models have finally made natural language data analysis reliable enough for business users the trade offs between using frontier models versus smaller specialized models and when to use structured workflows versus letting models choose their own adventures how data residency requirements are shaping partnerships between.

1:38

Speaker A

Cloud providers, model labs and platforms like.

1:59

Speaker B

Snowflake how AI coding assistants are changing the discipline of product management by enabling rapid prototyping of working features where Bearish sees value accruing in the AI stack and his prediction that horizontal applications will win out over narrower vertical solutions and finally, why Barish takes the over on my timeline for autonomous drop in knowledge workers and what he believes will have to happen first if you want to understand how large enterprise companies are deploying AI today and what's really working as they mature from the early experimentation phase to the ROI at scale phase, taking all of the operational complexities and security and governance concerns into account. I think this conversation will be perfect for you, and with that I hope you enjoy this deep dive into enterprise AI adoption and the future of data intelligence. With Barysh Gultigan, Vice President of AI.

2:02

Speaker A

At Snowflake, Baris Gultigan, Vice President of AI at Snowflake. Welcome to the cognitive revolution.

2:59

Speaker C

Thank you, Nathan. Thanks for having me.

3:08

Speaker A

I'm excited for this conversation. There's going to be a lot to learn, I think people. We have a very diverse audience. The number one profile is AI engineer. And within that profile people work at a lot of different kinds of organizations, from solo entrepreneurs and consultants to startups to enterprises. So some people will certainly know Snowflake and will work at organizations that are customers of Snowflake. Others probably have heard of it and don't really know how to too much of the backstory. So maybe for starters, just kind of give us the real quick Snowflake 101 and then I'd love to go into how AGI Pilled is Snowflake today.

3:11

Speaker C

Sure. So Snowflake is a data platform. We call ourselves an AI data cloud. So what that means is our customers bring a lot of their data onto Snowflake so that they can secure it, govern it and analyze large amounts of data for various insights, dashboards and the like. And from an AI perspective, because there's a lot of gravity to data, our customers do not want to replicate data in multiple places. Instead they want to bring AI to run next to data. So that's a very high level overview. And AGI pilled is an interesting phrase for us. We're quite practical, we serve large enterprises and the goal is to get to high quality AI agents that create positive ROI for customers quickly. And that can happen today and it is happening today. Super excited about where things are.

3:47

Speaker A

I definitely want to come back to the bring AI to data strategy that you guys have in a few minutes, but to just double click a little bit on the before and after because obviously Snowflake has been around for a while before certainly anything like the AIs that we have now were available. So what were people doing before with Snowflake and what are the new AI use cases that have been unlocked over, say the last. I don't know when you would start the clock. Right. Do you start the clock at ChatGPT or do we need it? Was that like not quite strong enough to actually make things work? But yeah, there's a lot of different dimensions. Maybe let's start with before and after.

4:37

Speaker C

Sure, so I'll start with the before. So a lot of our customers have been using Snowflake mostly for structured data initially and this is where they'll bring the data in and then they'll run large scale analysis either to have insights to power bi dashboards for instance, or to do, to do various analytics to understand their business. The value of the platform is to be able to bring data from all the different places to break the data silos so that you could run analysis across large amounts of data. So what is happening with AI is there is a big unlock, of course, of unstructured data. And this plays out in two ways. One is if you have thousands, hundreds of thousands of documents, for instance, you can now extract structure from these documents and then you can analyze them. If you have contracts, for instance, and if you want to say, hey, what are the contracts that talk about this specific thing? How many of them do I have? Or what are the contracts that are expiring soon in this category, being able to run these analysis super easily is now very feasible. Then there's of course being able to do a lot of again kind of large scale analytics work, but with, with a lot of ease. So a lot of the pipelines that used to be built around, I'm going to go classify my data, I'll extract information from. It is now very, very simple to do. So the way we've done it is we've again brought AI to directly work with the engine for analytics so you can do things like classification, extraction of that data very, very easily. And increasingly, of course, there is a lot of interest in bringing natural language interface to all a company's data so that you could just talk to your data, you can democratize access to all that data for the organization.

5:14

Speaker A

Could you give a sense for sort of the balance of structured, unstructured or then I guess you're also saying that like structured data is getting structured through the process of basically AI retro annotation. My sense is that I think of, oh gosh, there's so many vector databases. The exact name of the one where the founder told me this is slipping my mind. But one of the interesting things that I've understood to be happening in general with business data is that structured data was kind of like the tip of the iceberg in many organizations where it was the most usable kind, but it was actually like a relatively small amount of the data. It was Chroma, it was Anton from ChromaDB who said that like most of the data that was going into Chromadba had never been in a database before at all. It was just lying around in various places. So have you seen like a sort of great unlocking of people, like dumping more and more data into Snowflake? Because now they have ways to make it useful where it just previously wasn't even worth it.

7:10

Speaker C

Yeah, we're absolutely seeing this. You know, 80 to 90% of all data is unstructured data. And you know, because there weren't a lot of easy ways to process this, it was not necessarily seen as the most usable data and it is now very usable, both from a kind of extract and then bring structure to it perspective as well as just talk to all of that data, find the right information using Vector DBS for instance, and then build agents and chat experiences off of it. So we're seeing this and it is again playing out in both ways. Both more and more data is getting structured so that you could run analytics on it down the road as well as you can just use all of this data in conjunction with the structured data that you have. So for instance, if you want to build anything, let's say a wealth management agent, so you still need to be able to look up what the stocks are doing in a structured way. But you also have all of the equities research that's in PDFs that you'd like to be able to use. So being able to combine both structured and unstructured is incredibly important for real world use cases.

8:11

Speaker A

Where would you say we are on text to SQL today? It's been a while since I've done a show on text to sequel. There have been probably two episodes on this theme historically and I guess there's, last I checked there was like a range of opinions where some people were like, yeah, it's just not really there. Other people were, is there? But you have to do a lot of work to make sure that you have a good semantic understanding. Because a lot of the SQL databases that come in are like, there's multiple columns and there's tribal knowledge on teams that like we don't use that column anymore, we haven't deleted it, but we don't use it anymore and it's superseded by this. Or there's all these little nuances that kind of live in people's heads. And so some people have said, oh, the models can't do it. Others have said the models can do it if they have enough of that kind of context. What does a process look like today and how good does it get if a new customer. I want to start to enable these talk to my data with a sort of text to SQL kind of strategy. What does that look like now?

9:26

Speaker C

Yeah, I mean you called out, right? It's been traditionally very difficult for models to get text to SQL. Right. And there are various reasons for it. First of all, if you ask, what's my revenue? The answer is there's only one answer. So the margin of error is very low and the expectations of quality is incredibly high. And the reason it's been really difficult for these models is because you need a lot of semantics to be able to figure out where to get the data from. So first of all, what is the definition of revenue? What is the definition of profit can change and then how is it modeled in the data side can be tricky, can change. When we're talking about real world scenarios, we're talking about thousands and thousands of tables that have hundreds of thousands of columns in them to be able to go and reason about. So it's been traditionally very difficult. What I'll say has happened in the last six months to a year is with the reasoning models getting substantially better and increasingly being able to bring the semantics relatively easily onto the platform, we've had pretty substantial gains in quality. So we now have our product to do. This, for instance, is Snowflake Intelligence. And we're seeing tremendous demand for you doing text to SQL. And the quality is at a place where you can now deploy them very broadly. So very high quality and very useful because the structured data is quite useful.

10:22

Speaker A

And when you say deploy broadly, you mean to users who are not data analysts.

11:44

Speaker C

That's right. So for us, for instance, this product I mentioned, Snowflake Intelligence, is our agent platform and it's being used by business users. And it is the fastest growing product that we have on Snowflake because we were now making large amounts of data easily accessible to business users to ask questions and get insight very quickly. In the past, they would have to go to an analyst who's familiar with the data for them to kind of build some analysis and then get back to them a week later. Now they can just directly ask questions.

11:50

Speaker A

What does that process of bringing semantics onto the platform look like? I can imagine a sort of big setup one time where you go out and interview the people that have set these things up and should know. I can imagine at runtime you might have to come back with questions and say, hey, I've got multiple columns that are ambiguous here. I did one episode also with a company you may know called Illumex, where they had a really interesting strategy that was around basically building what they considered to be the sort of canonical, like abstract ideal form of an enterprise in each major vertical that they served. And then they built their query engine off of that ideal. And then the mapping process was like, okay, now how does your actual Real world enterprise deviate from this ideal, let's map all that out. But then we know that once we've done that mapping, the logic on top of it will be trusted. What mix of strategies are you using? What would you find to be effective?

12:21

Speaker C

So first of all, again, reasoning models are now at a point where AI can help substantially in building out a semantic model. And the inputs to that are both, of course, the data that's in the system, all of the metadata that is in the system, the names of the tables and columns, as well as the data underneath them. But we have built a series of connectors to things like BI dashboards that also have a lot of semantics in them. That's super useful in building out semantic understanding for the organization. Also things like queries people have been running in the past, these are all kind of hints for these agents to go help our customers build semantic models. Snowflake has recently also announced what we're calling open Semantic interchange, which is trying to create an open standard for sharing that semantic model across the different platforms so that we can more easily create these common semantic understandings for AI to act on.

13:21

Speaker A

Okay, that's interesting. How does that, can you unpack that for me a little bit? Like, how does that work? What do I do as an enterprise if I want to adopt that standard?

14:31

Speaker C

Yeah, so it's still early. So we're working with companies like Tableau Omni, BBI platforms as well as others technology providers to create an exchange format so that if you create a semantic model in one platform, you could just use it in a different platform. So there's active development now with all the parties in the, in the open semantic interchange to define what that interface is so that we could support an open interchange of the semantic model, essentially. So what that would look like is a customer can go to Snowflake, for instance, they can go build out their semantic model, and then they could reuse that semantic model in another place that supports the open interchange, for one thing.

14:40

Speaker A

I wish that would come to electronic medical records sooner rather than later. I feel like yikes has been my recent experience there. I wonder how that going back to the first question about how AGI pilled various organizations are, I wonder how do you see that from a competitive dynamic standpoint? Because I think one of the things that's most interesting, where the dice are in the air, so to speak, it feels to me right now in the software market is like, sure seems like the pace of software development is increasing dramatically. People are, I think mostly it's outliers or just plain bluster at this point to say that people are like deleting systems of record and rolling their own in house. But you've at least got that talk is out there. And then presumably everybody is, geez, if I can maybe already see a major acceleration in my software development or if I can project a year or two in the future and I can see a clear path to a major and I vibe coded three AI apps for family members for Christmas presents this year and the acceleration at that level is certainly very real, then it seems like everybody is going to be incentivized to like try to go take sort of some conceptual territory from companies that maybe used to be partners, used to be complements. It seems like it's headed more toward competition. So all these companies that you named, right, historically you specialize in one thing, they did a little bit different thing. They work nicely together. You got a lot of customers in common. Great. But if I'm them, or maybe if I'm you, I might start to worry at this point. Geez, if I'm tableau, should I be afraid of Snowflake going to come after me with something that sort of replaces what we do and do I want to be partnering with them on these standards, or do I have to fear that whoever is the, whoever has their hooks deepest into the customer can box out and colonize these additional niches? Now that's a pretty AGI pilled point of view. Maybe you think that's just we're, I'm getting ahead of myself there. But what do you think?

15:34

Speaker C

Not at all. Not at all. I, I actually really love what's happening and what's happening is the, the silos are coming down. This is all great for customers, for consumers, right? With all these open standards, essentially the beneficiaries are our customers. There is no lock in anymore. And I think that's great. That's great for competition, that's great for innovation, that's great for customers. And we're seeing this play out across the board, right? Anywhere in AI, the differentiation is coming down. That means everyone is doing more and more things to create more and more value, which ultimately is great for the industry and great for consumers and customers. So I'm loving what's happening. As you called out, the walls are coming down. The lock in is no longer there. And that makes product development really important. That makes the speed of execution really important. And ultimately it's all about creating more and more value.

17:26

Speaker A

But does that take us to a place where companies that used to be friends are trending toward frenemies because it seems like there's only so many ideas. It just seems so obvious to me in so many places that a big platform like Snowflake would be like, sure, we could do what Tableau does, especially now that we can get so much more stuff shipped on a quarterly basis.

18:22

Speaker C

I would say the pie is growing, so I don't think it's a fixed pie that the people are trying to protect. So the types of things you can do is growing and that is super exciting. I also don't think that everyone can do everything. Ultimately, where each company focuses is closer to their area of expertise, to their differentiation. So I don't necessarily see that everyone is going to do everything, but I also do believe there is a lot of competition, but there is also growing pie, which is. Which is exciting.

18:43

Speaker A

Yeah, okay.

19:11

Speaker C

Hey.

19:13

Speaker A

We'll continue our interview in a moment after a word from our sponsors.

19:13

Speaker B

You're a developer who wants to innovate. Instead you're stuck fixing bottlenecks and fighting legacy code. MongoDB can help. It's a flexible, unified platform that's built for developers by developers. MongoDB is acid compliant enterprise ready with the capabilities you need to ship AI apps fast. That's why so many of the Fortune 500 trust MongoDB with their most critical workloads. Ready to think outside rows and columns? Start building@mongodb.com build that's mongodb.com build. Your IT team wastes half their day on repetitive tickets, password resets, access requests, onboarding all pulling them away from meaningful work. With Servol, you can cut Help desk tickets by more than 50% while legacy players are bolting AI onto decades old systems. Servl allows your IT team to describe what they need in plain English and then writes automations in seconds.

19:17

Speaker A

As someone who does AI consulting for.

20:19

Speaker B

A number of different companies, I've seen firsthand how painful and costly manual provisioning can be. It often takes a week or more before I can start actual work. If only the companies I work with were using servl, I'd be productive from day one. SERVL powers the fastest growing companies in the world like Perplexity, Verkada, Merkor and Clay. And Servil guarantees 50% help desk automation by week four of your free pilot. So get your team out of the help desk and back to the work they enjoy. Book your free pilot@servl.com cognitive that's S E-R-V-A-L.com cognitive let's go back more toward.

20:21

Speaker A

The technical side for a minute. We kind of went deep on text to SQL. Let's do the same thing for rag. So we've got all these unstructured vast amounts of data out there. They're getting loaded into platforms, they're getting metadata synthetically created by AIs coming through and just processing them suite by suite. How well is that working and what is actually key to making it work? We've been through eras of chunking. Strategy is really important. Or I've done episodes on graph databases and entity recognition and figuring out various ways to traverse the entity graphics that. And obviously it's going to be quite distinct for each enterprise with all the different entities that they're going to have that nobody else has. What is really driving results in that RAG paradigm today?

21:03

Speaker C

Yeah, at Snowflake we've been actually very fortunate. We acquired a company that I came with called Niva, which was a web scale search engine and all right, awesome. So we brought that technology into Snowflake to build out our search and RAG solutions. And there basically what determines quality is the quality of the embedding model that you're using. Of course there is more and more sophisticated chunking strategies of what you're indexing and then there is other layers like the hybrid search and the re ranker that you build on top of it, and so forth. Increasingly, a core part of it is also to be able to understand complex documents. PDFs are messy. You have images, you have tables, you have multiple columns in a page, and so forth. So being able to handle all of this, extract information really accurately, figuring out which model embedding model to use, whether you should use a multimodal embedding model or a text embedding model, and so forth. So all of those are incredibly important. But increasingly we're getting to a point where you can automate many of these things and then reduce the complexity. So that a lot of what we used to require practitioners to do can be relatively automated at this point. And now you're getting to a point where more interesting opportunities get unlocked. So for a company like Snowflake, for instance, being able to do what we're calling analytical agentic document analytics is something that is possible to do. So what I mean by that is let's say that you have thousands of PDFs and there is information in it. Let's say you have quarterly results over the last 10 years. So being able to say what's the average revenue over the last 10 years and if that is in multiple different documents. Being able to extract all of that and then do analytics on it is now possible. So overall I think RAG is both getting increasingly higher in quality and also simpler to build and increasingly more and more powerful to handle some of the new agentic use cases.

21:54

Speaker A

Would it be a fair distillation of what you've said there that you're trending more toward more powerful models? Like a project that I've been involved with recently is built around understanding, often scanned on like a physical scanner forms that are associated with the sale of a car from either a dealer person or person to person. These things of course have to get filed with the state and reviewed and they're super messy and whatever. So working a little bit with a company that's using AI to automate that in that context, I've really seen a pretty substantial simplification where 18 months ago it was like you might need your specialist embedding model here and your kind of table extractor model there and all this kind of deep specialization. Often not super large models, but like really dialed in on these use cases. And now I would say today Claude 45 opus or Gemini 3 mostly just solve the problem off the shelf in terms of understanding those documents at a higher cost, certainly inference wise, but definitely a lot lower cost in terms of AI engineering time. Am I right to say you're seeing the same trend? The less specialized models, there are different use cases.

24:13

Speaker C

If you are going to process in some cases hundreds of millions of documents, you're not going to use CLAUDE to do that. Instead you want to use a specific embedding model to embed certain aspects. You want to extract the so that you could reuse it later and so forth. But if you're talking about one or two documents, of course these large language models can't handle them really right now. So I still do believe there are different use cases. And those use cases call for different tactics, different models.

25:31

Speaker A

In terms of why you wouldn't send millions of documents through claude, is it just about inference cost or is there.

25:58

Speaker C

Some other it's cost and throughput. Oh, just to give you an. Yeah, exactly. So how long would it take for you to process that many documents is a challenge. So at Snowflake we have a document extraction model that we've built and fine tuned. It is in multiple orders of magnitude smaller than these large language models. That means it's substantially cheaper and much faster to go and process if the task is specific. I'm going to go extract information and, and extract these specific Fields, it is faster and cheaper to do versus using these very large models which are super capable, but again will be limited in terms of how fast they can do this. And of course the cost is another issue.

26:04

Speaker A

So that's interesting. We're sketching out a little bit of a Pareto frontier, so to speak here where we have on the simplest but most expensive at inference time potentially also like rate limit issues, we have our clods and other kind of frontier models. You're in the middle with a snowflake specialist model that is much smaller, does just what it does, but is still like something that is amortized off over a whole bunch of enterprise customers that you have. Is there. What are you seeing in terms of the other end of that spectrum? Like the. Is there still value in an individual enterprise trying to create its own super specialized model for some of these tasks or is that. Does that curve stop at the snowflake scale model?

26:50

Speaker C

Yeah, it's a good question. So first of all, we partner very closely with all the large language model labs out there and they have incredible capable models that we use every day. There are some cases where our customers would want something very specific and this is a case when a customer has large amounts of data and the use case is something that the model has not seen before and then they have strict either throughput requirements or cost requirements. Those are the cases where a custom model that is usually based on some of the other large language models out there makes sense. So we work with these customers to build custom models for them. But in most cases a well tuned rag solution text to SQL solution with the data that they already have with a large language model that's the frontier model is usually the go to scenario.

27:43

Speaker A

I'm halfway through doing an AMA episode. One of the questions I got was is fine tuning really dead? What do you think? So it sounds like you're saying it's not quite dead, but it seems like it's specialized. It's on the decline in your.

28:44

Speaker C

I wouldn't say it's on the decline. I think it is really well suited for certain types of things. And maybe the best example is actually what Cursor recently did right at their scale. It does make a lot of sense for them to have a custom model that is doing their autocomplete for instance. So being able to figure out in which situations you need a custom model versus not is something that is evolving. Starting with the large language models makes a lot of sense. And then over time as you have more and more data and if you have Specific needs. If you have either specific needs because of data or because of cost or throughput, that's when specialized models come into picture.

28:58

Speaker A

You mentioned these partnerships that you have with the frontier companies. Before getting to that, would you like to shout out or highlight any particular open source models that are your go tos we hear a lot about? Obviously the Chinese ecosystem is like continuing to open source a lot more than the American ecosystem at this point. I don't know if you guys feel like comfortable using Chinese models in your stack. I get very different answers on that when I ask that question. But what are the models that you guys go to today when you're like, okay, we're gonna explore some new custom direction either for all of our customers or even just for one customer. What are the handful of models that you go to as starting points to begin that journey?

29:45

Speaker C

Yes. So for Snowflake, we have a platform where we offer a series of models and our customers choose which model they'd like to use. And then there are certain products where the model is just part of the product and not necessarily a specific choice. For the models that we offer, there's of course all The Frontier models, OpenAI models, anthropic Gemini, as well as models from Meta Mistral and others. Some of these models are open source, others are proprietary. We also have Deepseek as a model that we provide for customers. And then in certain arrangements where customers are looking to build custom models, some of them are open to use using model weights from these models from China, others aren't. But it really depends on the customers.

30:29

Speaker A

Does that break down along industry lines or is it more of just an idiosyncratic gut feel on the part of the customer as to what they're comfortable with.

31:13

Speaker C

I think it's the latter, actually. It's not necessarily a industry specific thing. We have customers who are in technology, for instance, who say yes sometimes or who will absolutely not touch in some of these models.

31:21

Speaker A

For other customers, do you have a sense of how much they are leaving on the table? Are they leaving much on the table by cutting off the Chinese model option?

31:36

Speaker C

Models like Quin are incredibly powerful. If they'd like to start with models like that and then fine tune it, you can get very capable models, but you also have other alternatives. So it really depends on, you know, the internal policies of these customers to decide which route to go. I'd say it's such a competitive space that I don't think there is, you know, one model that dominates, dominates it all, whether that is in proprietary world or open source world. So there are a lot of choices out there.

31:50

Speaker A

Gotcha. Okay. Hey, we'll continue our interview in a moment after a word from our sponsors.

32:23

Speaker B

If you're listening to this podcast, you're probably thinking seriously about where AI is headed and maybe about how you can actually contribute to making it go well. I want to tell you about an opportunity that could become a pivot point in your career and a springboard for you to make a positive difference. A program that I've been so impressed by that I've supported it with a personal donation. I'm talking about MATS, a 12 week research program that connects talented researchers with top mentors working on AI alignment, interpretability, security and governance. These are researchers at Anthropic, Google, DeepMind, OpenAI, the AI Security Institute, Redwood Research Meter, the AI Futures Project, Apollo Research, Govai, RAND and other leading organizations. The track record here is remarkable. Mats has accelerated over 450 researchers, with 80% of alumni now working in AI safety and security. 10% have co founded AI safety initiatives, including Apollo Research, whose Co founder and CEO made the 2025 Time 100 AI list. MATCE fellows have co authored over 120 publications with more than 7,000 citations and helped develop major research agendas like activation, engineering, developmental interpretability and evaluating situational awareness. The program is fully funded, a $15,000 stipend, $12,000 compute budget, housing, catered meals, travel and office space in Berkeley or London. Everything you need to focus entirely on research for three months with the chance to extend up to a year. Applications open December 16th and close January 18th. If reducing risks from advanced AI is something you care about, you should apply. For more information, check out matsprogram.org TCR that's matsprogram.org TCR or see the link in our show notes. The worst thing about automation is how often it breaks. You build a structured workflow, carefully map every field from step to step, and it works in testing. But when real data hits or something unexpected happens, the whole thing fails. What started as a time saver is now a fire you have to put out. Tasklet is different. It's an AI agent that runs 24.

32:28

Speaker A

7.

34:39

Speaker B

Just describe what you want in plain English. Send a daily briefing, triage support emails, or update your CRM. And whatever it is, Tasklit figures out how to make it happen. Tasklit connects to more than 3,000 business tools out of the box, plus any API or MCP server. It can even use a computer to handle anything that can't be done programmatically, unlike ChatGPT Tasklet actually does the work for you, and unlike traditional automation software, it just works. No flowcharts, no tedious setup, no knowledge silos where only one person understands how it works. Listen to my full interview with Tasklet founder and CEO Andrew Lee. Try Tasklet for free at Tasklet AI and use code COGREV to get 50% off your first month of any paid plan. That's code COGREVASKLET AI.

34:40

Speaker A

So on these partnerships, we've got announcements recently of partnerships with Anthropic and also with Google for the Gemini models. I believe there's also one, although I think it was not so recently announced with OpenAI. I didn't see, I didn't catch anything with respect to XAI and grok. Providing all the latest and greatest stuff to customers is at the heart of that strategy. But tell me more about kind of some of the nuances of the partnership. Is there a XAI relationship? If not, why not? And does it have anything to do with them putting women in bikinis all over the place? And then I definitely want to get into the how are we bringing these models to data? Because that sort of is a bit of a narrative violation relative to what you typically hear is we can't use that because we'd have to send the data to them and we're not comfortable with that. So I'm very interested in unpacking how you are reversing that and bringing the models to the data on the Snowflake platform.

35:30

Speaker C

Yeah, absolutely. Actually, let me start there because that's incredibly important for us. So when we started the journey two and a half years ago or so, we heard loud and clear that our customers do not want to move their data out of the Snowflake security boundary. Instead, AI needs to come next to data, and that gives them a lot of advantages. You can just respect all of the security that you've established. You respect a lot of the governance on the data so that you're not replicating this data. The attack vectors shrink in terms of securing all of this information. So what we have done is, thanks to our relationships, we've built, we are bringing, we're bringing essentially inference to run inside the Snowflake security boundary. So that's accomplished through these partnerships, through the connections, as well as a lot of the legal guarantees around the data. So essentially these models become sub processors. There is no state that's saved in any of these models. So that's super helpful for our customers who are very sensitive, many of them in regulated industries. So when they're using any of these models, they know that the data still stays inside the Snowflake security boundary.

36:20

Speaker A

So does that mean then that the model weights have to come inside that boundary? And like, how is that happening? Because obviously the ic.

37:27

Speaker C

Yeah, the IP still belongs to. Yes, absolutely. And the IP is owned by the model providers. The inference is run by the cloud providers in their stack. The difference is we have a series of guarantees to ensure data residency, to ensure there is no state that's left. So all of those are through the relationships and the deals that we're doing with these model providers as well as the cloud providers.

37:36

Speaker A

So the cloud providers are key in this because they are certainly able to provide the inference and they're also providing the underlying, like, physical infrastructure that Snowflake is built on top of. And so it's because both of those things are true that we can draw the right dotted line security boundary, both of these things. Is there more that I should understand about this? Because one thing that has, of course, I don't know what, I don't know, but it seems like the more you move weights around to different clouds and stuff, the more risk you. As a frontier model developer here, I'm thinking OpenAI anthropic. Google obviously runs their own clouds to a very large extent, although everything's showing up everywhere. One of the fascinating things about this whole moment has been how many alliances or at least partnerships we've seen between big tech companies that previously were very much at odds with each other. So with all the models showing up everywhere, I'm like, how has it been that none of these have really leaked? It seems like there's so many people that work at these platform companies that if access isn't like really well figured out, I don't know, it just seems like something would leak at some point. But we haven't really seen that. We haven't seen the weights of a frontier model leak at all, as far as I know. And then people will speculate, maybe some state actor might have stolen them and not put it, not told anybody about it. But we haven't seen fundamental breakdowns. So how should we understand how that is happening to seemingly such a high degree? What role does trusted execution environments play? What role do other kind of measures play? Like, how are we. My general working heuristic is like, everybody's hacked, everybody's pwned, like nothing is secure. And yet at the same time we seem to not be having catastrophic leaks. So how can you help a simple person like me understand how we're achieving that.

38:06

Speaker C

Yeah, I mean, as you called out, these are very sensitive, important IPs that belong to the model providers and then secured by the cloud providers. They have a very strict series of requirements and set up to ensure that access is limited because they are the ones that are running the inference and setting up the environment. They've set it up in a way that is airtight. I don't have a lot to say beyond it. I think they absolutely take security very seriously. We work with them, we understand how important it is. You talked about all the different risks that are out there that they need to protect against. So this is something that both cloud providers and model providers, as far as I can see, are taking very seriously. And as we work closely with them for Snowflake, of course, security is at the heart of what we do. So we set up our own environment in a way that is, that has all of the security considerations in mind. I'll just say this is an incredibly important area and there is a of lot, but there's definitely focus in this area across. Across all the parties.

39:54

Speaker A

How much would you say of this kind of security has gone to the level of provable guarantees or cryptographically secured, as opposed to more roles and access controls and things where there's still a more fundamentally like, human element? I just did an episode not long ago with a couple experts in formal methods, including a guy who's a VP at Amazon, who's pioneered a lot of their use of formal methods to derive a lot of these security guarantees. But I'm not clear on how much of this is resting on that. Kind of like, we have proven that this is secure versus how much is like we have a process that we feel good about and we want you to trust.

40:55

Speaker C

So this is not my area of expertise, so I don't have a lot of depth. But purely from talking to both the cloud providers and the model providers, when you start looking at what are all of the attack vectors and whether there is what is possible, it doesn't seem like this is a human factor is an issue. The way the systems are set up is inherently very secure. That said, in security, of course, you can never say, hey, this is completely airtight and it can never be penetrated. But as I said, security is taken very seriously. And I don't think it's a human factor necessarily. The way the systems are set up is such that the execution environment doesn't have access to. You cannot do a lot with it other than just run inference through it. So but by design, access to the weights are limited.

41:40

Speaker A

Yeah, okay, cool. Thank you. I'm always trying to get a little bit better read on that particular corner of the world and it's not one that is as freely and openly as some of us curious minds might like. Going back to the models though themselves, what's your read right now on this is another thing where I think people have very different intuitions. Are the models going to be commoditized or are they going to be sufficiently differentiated as to maintain pricing power as we continue to go into the future? I guess there's that there's like, how do you help customers decide which model to use for a given case? Do you have an evals platform built in or do you help them do evals? How do you help them think about being like keeping agile so they can switch? Obviously new models are coming out all the time, so there could be something better or faster or cheaper that you could upgrade to. But you have to know with some confidence that you're going to be upgrading with good reason. There's a whole ball of wax there. Take your time in melting it.

42:28

Speaker C

But yeah, super interesting. I mean I, as you mentioned, the differences between these models is not large. Each model keeps getting better and then there is great healthy competition out there between the model providers, which again benefits companies like ours, our customers and so forth. For us because we're providing the choice to customers. The second part of the question is also really important, which is how do we help our customers choose which model is the best fit for their needs? There's a couple of considerations. One is for many customers, they do not want to leave again because of data residency requirements. If they are for instance an Amazon shop and today OpenAI is available through Azure or through directly OpenAI, that becomes a consideration. So some customers are okay with their data leaving that Amazon cloud boundary, others aren't. So that's one decision point. The second one is of course from a quality perspective. Many customers will go run evals side by side to decide which model is best suited for their needs. What's interesting is some of these models are cheaper, faster, but when you add reasoning on top of it, the equation changes. So. So certain models are very good at certain things. Again, just to call out Cloud is incredibly good at coding and continues to be a great model for that. So we help our customers in assessing which models to use for their needs.

43:26

Speaker A

Does that extend to how much of that is a service, a consultancy type relationship and how much is productized at this point or productization?

45:06

Speaker C

Yeah, it is more productized than a service on our product. You can easily choose models, you can easily do side by side comparisons, you can run evals and for many customers actually it's not necessarily for the first reason. Not all the models are available in their environment anyway because as a company they've decided that only these models are approved or only this environment is approved.

45:17

Speaker A

For me how often do you see people switching? This is even at my so I started a company which I used to be the CEO of. I'm no longer and we're only 40 people, we're doing like $10 million a year in revenue. So we're not an enterprise, we can fly a little faster and looser than and we're also not in a regulated industry. We basically do video content creation for local and increasingly like mid sized businesses. And I feel like we should be changing models more often than we do. Honestly I feel like the leapfrog effect is it's just happening so often and if I were to grade our own performance I'd be like eh B like we're definitely better than most, but I wish we were even a little more on top of eking out the latest and greatest performance from the latest and greatest models. But it's hard, it is hard to resolve. Sometimes you've got even just human inter rater disagreement which is tough to overcome. And how does that play out at the larger scale? Do you see people like oh we got a new Claude When Claude 4, 6 hits like how many people move to it in a week? In a month and a quarter?

45:37

Speaker C

Yeah, I think it really depends on the use case. In most cases we don't really see a lot of switching happening because the prompts get optimized for a certain model and you get high quality because you've optimized it for a certain model. And it's not as easy without further optimization to switch. And because the deltas between the models aren't much and they keep improving on a regular basis, the needs for switching is also not that much. As you were describing this, I was actually thinking about Google versus Bing. At some point Bing got to a good enough quality but there was the habit of continuing to use Google, the familiarity of the interface and so forth that switching wasn't as necessary. I don't think we're there yet necessarily for models there's still a lot happening, a lot of innovation happening. And also for certain use cases like if this is a one off I'm going to go run something and then do side by side comparison then you go pick the model that works best for you. But if you've already been investing in an application and you have thousands of lines of systems and prompts that you've built, then there's a cost to switching, in which case the gains have to be large enough to justify that cost.

46:39

Speaker A

So I'm no Ben Thompson, but it seems like from your perspective, you would want to commoditize your compliments and would want to do everything you could to reduce those switching costs. Right? Like one of the virtues of being on the Snowflake platform would be you've got all the things, but not only ideally you have all the things, but also you can. The interface is presumably a lot more unified than it would be if you were going directly to the model providers. And I imagine there's a bunch of different things you could do over time. You've got like things like DSPY out there that you can say, sure, this is my one prompt with this model, but maybe if I throw it into dspy, I can auto evolve my prompt to be more optimized for some other model, what have you. Is this like a goal? Do you think of it as like a success metric that you like, would help people be very fluid in switching from model to model as you go into the future?

47:54

Speaker C

I mean, that's not how we think about it. For us, it really does boil down to how do we bring the most value for customers quickly. So choice is an important factor there. So we'd like to offer choice. And customers make model choices for a variety of reasons. As I said, if some of them have only approved a certain model, they have their own AI governance boards where they decide which model to use and so forth. But for us, we start with the data at the core. So ultimately anything that you do is as good as the data that you provide to it. So a lot of the optimizations for us are can we do a phenomenal job at the retrieval layer? And then can we make sure that all of these models are optimized to the fullest extent so that any customer that's choosing one or the other for the variety of reasons I called get the best quality data agent, if you will, that they're building with us?

48:45

Speaker A

Okay, that's really interesting. What do you think that implies for the competitive dynamics between model providers? One takeaway you might have from that is whoever has the best model at any given time wins. Of course there are these other constraints, but leaving those aside for the moment, if I'm a customer that has no binding constraints and I can pick Whatever frontier model I want, it seems like whoever has the best model at any given moment in time wins that business and then actually stands to keep that business, even if that business that might not be that whole enterprise's business, but that particular use case you're saying is stickier than you might think. Switching costs are higher than they intuitively seem. And so having the best performance at the time it's initially evaluated is actually like pretty important.

49:44

Speaker C

So I think the model quality is incredibly important. But increasingly we're moving up the stack so that the product also becomes incredibly important. So when you look at from a consumer perspective, ChatGPT as a product starts having its own kind of stickiness because you start using it and you get accustomed to using it. Similarly, on coding side, cloud code has its own again, benefits. You'll start writing your instructions, your prompts to optimize for that workflow. So I think we're just moving up the stack. The model quality is absolutely central, but as the quality kind of keeps up across model providers, the next level of differentiation happens, happens at the application later.

50:34

Speaker A

So that's a perfect transition to talk about agents and what you guys are doing with agents. The way I structure my own thinking about agents is on a spectrum from on the one end, your Claude code style, choose your own adventure. I just give you the goal essentially and you, the agent, break it down and search around, grep around and figure out how to get there. And then on the other extreme is like potentially a totally linear structured workflow where we're going to run a series of prompts one after another. The Claude code is undeniably awesome interface, but I often feel like people are a little bit too drawn to that. And I sometimes say that's a don't try this at home sort of project. Like by all means, go use Claude code, but don't think at your business you should be spinning up a Claude code, choose your own adventure thing probably for most cases. I advise people like even still today, more structured is probably going to get you more of what you want faster in a way that everybody feels good about at the end of the project. What distribution are you seeing across that spectrum?

51:14

Speaker C

Yeah, I think it really depends on the Persona who are using these tools. So I'm huge fan of cloud code and coding assistants make a big difference, unlock great capabilities, and clearly very helpful for AI developers, builders. If you are a business user who's just asking questions like what was my the usage of this product over the last week as a product manager, for instance, I want A structured way to do this. I want an agent that is already optimized for that use case that has access to the underlying data. I do not want a cloud code interface for this. I want something that I know will be high quality and that's optimized. So that's kind of how I think about it really depends on the Persona that you're building for.

52:20

Speaker A

So for the talk to data product surfaces that you guys expose, how where would you say you tend to fall on that spectrum? Is it a you're going to use these tools in this order or is the model kind of choosing which tools to use at any given time?

53:02

Speaker C

So we have a product that we built for business users. So this is Snowflake Intelligence, where you can build a series of assistants, for instance, for the whole company. We built a sales assistant and we've deployed to 5,000 sellers. That product is, is a. Think about it as like a ChatGPT interface on top of all of the company's data so that you can ask questions like what are my upcoming renewals? How is my big book of business doing? And so forth. And you can get answers for it. So for that, clearly you want a highly optimized set of agents for those set of use cases. And then these are business users using it and they need to trust the answers that they're getting. Then we have a set of products that we're building for, for data engineers and for analysts to, to build data pipelines to analyze data that is more coding assistant, if you will. So we have our own coding agent that's integrated in that platform where they're just either analyzing data or writing code. So that is. That of course is a lot more flexible and it's also not tuned for a very specific set of use cases.

53:16

Speaker A

How do you think about the question of like one big agent that might be long running versus the other kind of big pattern is your sort of initial agent that then routes tasks to sub agents. Back when OpenAI came out with their agents SDK, they had this notion of the handoff as a really central idea. And I was never quite clear on were they doing that because they thought that was the best way to maximize performance or was it more a nod to. We think at these enterprises that are going to use this thing, there's going to be different teams responsible for different areas and we want to be able to modularize the work for human reasons as opposed to like for AI performance reasons. With Claude, on the other hand, and we have a sponsor of the podcast called Tasklet which is a maximalist when it comes to just let Claude cook basically is their philosophy. Give it everything it needs, let it make all the choices, let it run for as long as it can run, give it feedback. But like it's one long one agent that kind of does it all in one long session. Of course I'm sure you could say there's different use cases deserve different paradigms, but what do you see working the most in practice today?

54:23

Speaker C

Yeah, even the Claude code case, you have skills that are being developed. Right. So you're still modularizing the different kinds of things you want Claude to do and then giving, giving instructions for cloud to go do these things. I think the way you called out is what I'm seeing, which is in especially large enterprises you have different teams building different agents. You also have different agent platforms that are being used. So for instance, if I'm using Salesforce to manage all of my CRM, maybe I'm going to go build my sales related experiences with an agent there, but I still want that agent to talk to this other agent I'm building for something else. So being able to do that agent handoff and coordination is emerging. I wouldn't say this is necessarily super top of mind for everyone. I think still customers are focused on let me get this one agent right and working well before I start thinking about multiple agents kind of coordinating with one another. But that's starting to become increasingly important for customers. One of the biggest considerations is they do not want to be locked in to a certain platform. So they still want to be able to make sure that the again open standards are supported so that agents can talk to one another, agents can use the tools that you built for one agent by another agent and so forth. So MCP A2A. These are imported protocols that our customers expect to be supported.

55:38

Speaker A

I was just going to ask about A2A. Are you seeing traction with that still early?

57:09

Speaker C

We don't yet support it. We're starting to hear our customers don't necessarily ask for a 2A specifically. They do ask for ensuring that some kind of agent to agent communication is possible to do.

57:14

Speaker A

Is there any standard or protocol or platform that is bridging those? The Salesforce Continent and the various other continents of agents.

57:27

Speaker C

Today we either see a bit of a hack where these different solutions are used as tools through mcp. So still the Orchestrator uses them as tools and then manages them from an agent handoff perspective. Other than A two, I haven't really seen anything else.

57:36

Speaker A

Yeah, that's interesting. One of the things I find very funny about this whole thing is that just it seems to me like a fundamental property of intelligence is that you can everything. One of my. This is an overstatement, I don't mean it literally, but one. One of my refrains is like everything is isomorphic to everything else, meaning you can always squish and rearrange and play hide the intelligence and you can have a smart MCP that's actually an agent. And how you actually classify these things seems to be much more of a choice and much less a requirement imposed on us by nature because just the nature of intelligence itself is so flexible, fungible, subdividable, whatever.

57:50

Speaker C

Exactly. No, couldn't agree more.

58:33

Speaker A

One of the big things, huge theme Right. Of the communications that I've seen from Snowflake in preparing for this is the importance of trust. So I'd love to hear your thoughts on what are the levels of what are the dimensions and what are the levels that we have to hit in order for an enterprise to trust an AI process?

58:35

Speaker C

Yeah, just to reiterate, for us, trust is incredibly important. If I were to call out two important tenants, one is super ease of use. So how easy it is to build out these solutions and to use them. And of course, trust is at the core of everything. And trust spans multiple different dimensions. Right. You know, trusting from a security perspective, then from a governance perspective, then you have the quality layer on top of it and then there is kind of evaluations and then monitoring and so forth. So it's a full stack. So the way we think about this is by running AI next to data, a lot of the core governance that is put on the data is by design respected in our system. So what that means is let's say you have, you know, sensitive data that's only visible by the HR team. If you go build an agent, the person who asks the question can only get the answer that they're eligible to see and nothing else. This is super obvious and important. But because we have these types of very granular access controls from the ground up as part of the core data platform, building agents that respect that becomes much easier to do. Then you have governance at various layers and of course next level is evaluations of these. So a lot of the trust is in. Are you able to build high quality retrieval of context to pass to the agent? Is the agent orchestrator doing a great job figuring out which tool to use, which trajectory to use to answer the question? So evaluation is a core part of the platform and then ongoing monitoring, getting feedback and Then and that cycle of improving the quality, from a user perspective, the way that that trust manifests is when a user asks the question, we have UI elements that says, hey, this question has an answer that was verified by an owner. So again, bringing that trust element into the user experience is another tenet to our philosophy.

58:58

Speaker A

So did I catch correctly at the data governance level that the shorthand rule is like the agent can only access the same data that the user can access. And so in theory that could mean like multiple users could come to the same agent and have different experiences because the agent has different data access based on the user that's using it at the time.

1:01:04

Speaker C

Exactly. And this is exactly what our customers are asking for. And that's relatively easy to build on our platform. So for instance, the example I gave with our own kind of sales assistant, if a salesperson comes in and says what is my book of business? Summarize it. You should get an answer that is only your list of customers assigned to you versus another salesperson. As an HR person, if I ask, or as a manager, for instance, using an HR bot, if I say what's the salary of this person? I should only be able to see the salary of the person that I have access to to seeing versus somebody else. And underlying it is the same agent and it's the access controls that govern what I'm able to see.

1:01:30

Speaker A

Yeah, interesting. On the sort of performance reliability side, my experience has often been, and sometimes it's for good reason, certainly in like the self driving car realm, there's a certain logic to saying we don't just want these things to be like roughly human level, we want them to be like, clearly a step up before we're going to adopt them. Society wide good news. It seems like we're getting there. What do people have in mind as the intuitive standard of performance? Is it like they want these agents to be perfect? Is it that they want them to be like at the level of the human that used to do the job? Is there some heuristic in the middle that you think people often land on?

1:02:17

Speaker C

Yeah, super interesting concept, right? The more natural the interface is, the more human like intelligence we expect. Intuitively, if I'm talking to the agent versus typing, I think talking has much higher expectations. For instance, versus I'm just typing. I know I'm typing to a computer. So the expectations become a little less high. I don't think that adoption of this technology requires human like intelligence because even for the specific things that these models and these applications do, well, that is such High value, that we're seeing huge adoption. As you all know, we're seeing huge adoption of AI already. And it keeps getting better. And it keeps getting better at a super rapid phase, rapid pace. Yeah. I'm very excited about where the technology is.

1:02:57

Speaker A

Before going into your expectations for the year ahead, what are you seeing in terms of guardrails? Obviously one big pattern that I think is very natural to you guys is sourcing answers back to the document or the sort of authoritative place from which it came. Beyond that, though, we've got this whole constellation of different patterns, right. In terms of you can filter inputs for appropriateness, you can filter outputs, you can log things and post process logs, you can. AWS has, I think, a really interesting new service called Automated Reasoning Checks where you can put a policy in. They convert with a language model, your natural language policy into a set of rules and values and then they do like literal formal methods to ensure that at runtime, like the agent or whatever, the system, whatever it gave you back, that it actually passes those like formal reasoning checks that were derived originally from a natural language policy. That's like pretty interesting and pretty cutting edge from what I've seen. But I think in most places, my sense is like the frontier model. Companies are doing a ton of this stuff. Anthropic has pushed this to probably farther than anyone when it comes to preventing you from using Claude to do certain things in the biosphere. But are people at the enterprise level actually doing much of it or are they just saying, this thing seems to work. We've got an email set, it passes and we'll go with that?

1:03:50

Speaker C

I think the sophistication is increasing and usually companies start with products that are more internally focused. While it's important, the bar is a little lower than something that's externally focused. At Snowflake, we offer products to check for guardrails, doing things like checking for hate speech, violence and other violations. You can detect them and flag them and not have the model respond. But we, of course, also benefit from all of the great work that, as you called out, companies like Anthropic do on their own models as a baseline. The other thing that is also super interesting is as the models keep getting better, their adherence to instructions, of course, keeps getting better. So some of these also get codified as instructions to the agent. So not only do this and that, but also here is the policy that you need to comply with, and that tends to work quite well as well.

1:05:17

Speaker A

Have you seen anything in the interpretability realm being used for practical guardrail Monitoring kind of purposes in enterprises. So far.

1:06:11

Speaker C

We'Re seeing evaluations become really important. Right. So a lot of, you know, what companies tend to do is they'll go create their own eval sets, but also use LLMs as judges across various different dimensions to go score what's happening and then continue to monitor it on, on an ongoing basis. And as agents become more and more complex, you know, it's a. It's a pretty new area to understand. Is the agent, you know, taking the right route? Should I be optimizing it? Understanding where things go wrong becomes really interesting. So from a. So that, that's what I see not necessarily in the core consumer experience, but in the kind of developer experience where you're seeing, you know, what the model is doing, what the agent is doing through evaluations and monitoring. Monitoring, yeah.

1:06:26

Speaker A

Okay. So you mentioned being excited about the year ahead. You also mentioned voice experiences. It seems like we're at a moment, like literally right now where I don't know if people just had extra time over the holiday break or whatever to get into Claude code for the first time for many people, but it seems like the discourse has really shifted and expectations have really shifted. In just the last 30 days, people have said, oh my God, like, the coding experience now is like, it's not just vibe coding and eventually hitting a wall and giving up. Like, you can actually really make this work. And then the next big thing that people are saying over and over again is the same thing that's happened to coding over the last however many months is coming to a great many domains of knowledge work over the next year. So do you buy that hype and what do you think that looks like? Are we all going to be agent managers or are we all going to be like, talking verbally to agents while getting lots more exercise than we used to? What is the 2026+ Vision for Success?

1:07:15

Speaker C

Yeah, I'd love to see the world where I don't need to do anything and I can just go get more exercise. I do actually see the opposite. We're able to do a lot more and we end up doing a lot more. Especially in AI where everyone is sprinting. There's more work and more productivity out there. So what I'm seeing is agents are absolutely getting more and more capable. Coding agents, I think, as you called out, have passed this threshold where they're a lot more. Because they're a lot more capable, a lot more people are using them. I think it changes how products are developed. It changes jobs like product management, for instance. I've been in product management for 20 years and the way we build products has to change given the coding assistance. How you deploy quickly, how you test things quickly is changing because of how capable these coding assistants are. And that is a combination of different things. Right. One is the agent can do a lot of things can do coding well, but also the reasoning capability of the agent is increasing. The tool use becomes incredibly powerful. So you can apply that to other domains. If you have like now figuring out which tool to use you it effectively and then reasoning and figuring out the next steps is very, that allows you to build very capable agents across the board. I don't know, this is not a 2026 projection or anything, but I'm absolutely seeing clearly increasing capabilities with agents and also increasing use of them for production work.

1:08:20

Speaker A

Do you have a sense of how the progress in AI coding assistance or agents has changed how work is happening at Snowflake? Are you like instrumenting that or measuring? Obviously lines of code would be like too primitive, but features shipped or burned down points per cycle. Is there a way that you can begin to quantify the impact it is?

1:09:45

Speaker C

Actually there's the impact piece, but there's also the philosophy that is changing. You know, how we build products is changing. So if that requires change of behavior, usually my go to is I have, let's say there's this feature to be built. I'll think about a ui. I'll go build this UI and go make it happen. Whereas with a coding assistant, if my users are also living in coding assistants, maybe that is as simple as let me just go build a skill for this thing and then quickly test it out. I can just write the skill in a day, put it out in front you of of my customers and then have them use it, give them feedback. And only when I know exactly what the shape of the product is, I can go and kind of solidify it in more of consumer experience. So I think again product management is changing, product building philosophy is changing because of these coding assistants.

1:10:11

Speaker A

Yeah, the working prototype is the coin of the realm these days for sure. One of the big predictions that I have heard a lot recently, sometimes with a remarkable level of specificity, including from some anthropic people, is that we should see the first drop in knowledge worker products offered this year. Specifically folks have said Q2 of this year and what that means to them is basically a new employee that's and ultimately at heart is an AI, but it will have a very similar surface to a remote worker on your team. It'll have a name, it'll have all the same accounts, or at least you'll be able to give it all the same accounts that you can give to a human employee, which means they'll be on Slack and they'll be accessible via email and they'll be all over the place, can probably join calls. And the expectation is that this will be good enough in Q2 of this year that people will start to get value from it and this will be a new product category, I guess. First, do you buy that that can happen that soon? And second, how many of your customers do you think will be eager to try something like that out when it drops?

1:11:02

Speaker C

I do see that it's a natural progression. Today the agents that are being built are either automating certain processes from a productivity perspective, or they are more like copilots that I can ask questions to and then get responses versus kind of these autonomous intelligence entities, if you will. When exactly that will happen, I think really depends on how scoped can you get them. I don't think we're at a point where you can just create another colleague that can just do anything and everything, but if you can just very easily scope the task, then I think that like that absolutely is possible. Yes. I don't know. I do see it happening. Whether this is Q2 or not, not so sure yet. And again, yeah, but plus, as a data platform company, I will call out the importance of ultimately all of these capabilities come down to for any given company, the differentiation is their data and then the access to that data, being able to figure out and retrieve the right types of data to be able to answer that question. And then increasingly the tools that are given to these agents to take action. I think that changes industries, right? That changes how we think of data. That changes how we think of making AI, making the data AI ready as well as making the tools AI ready so that more and more kind of capable agents can be built.

1:12:15

Speaker A

Do you see changes to data itself? I guess one that we've talked about already is just retroactively going back and applying structured, unstructured data, creating metadata, so on and so forth in the wild. One big change that we're seeing to data is like the web itself is increasingly comprised of AI generated data. And so that's a weird feedback loop that we accidentally created. Are there any other perhaps surprising patterns in data within enterprises that you're seeing as a result of AI coming onto the scene? Or is it maybe just still too early for something like that to be.

1:13:51

Speaker C

I am seeing two things. One, is I'm seeing the access getting a lot easier. So that democratization of access to data and access to insights is a big shift. The other thing that I'm seeing is the value that our customers get from data is increasing because you're able to very easily glean those insights by just describing what you want in natural language and then getting it. So the value you get from data is increasing, how that just opens up new and new opportunities. You start using the data in ways that you haven't thought of before. One interesting study, one of our customers as SMP, they analyzed earnings calls to understand when CEOs are responding to analyst questions in either directly or indirectly, or if a question was already answered in opening remarks or not. And using that as alpha to determine which stock to buy. So stuff like that becomes very easy to build and then new use cases open up.

1:14:33

Speaker A

Yeah, that's an interesting metric. I've seen a bunch of stuff recently over the last, even just the last week. It was really the Venezuela moment, where all of a sudden people were like bragging about how they had created these AIs that monitor the prediction market platforms and we're looking for early signal and trying to capitalize on that. That is going to be a really interesting phenomenon. How about in terms of just actually let me go slightly different direction, then I'll do maybe a little lightning round to close us out. I have this sense that right now we're in this kind of expansionary phase. I'm no astronomer, but my experience with platforms in the past, and definitely experience with the meta platform formerly known as Facebook, where it was like they came on the scene, they opened up a ton of stuff, everybody could tap into all these data and social connections. And there was for a moment, there was like an incredible flourishing of a ton of different ideas. And then after that supernova came the black hole and it was like, actually we're going to close all this stuff back down. And a lot of that value that entrepreneurs created on the edges, experimenting with different ideas, the things that really mattered mostly ended up getting sucked back into the platform. And there wasn't nearly as much value created on the margins as it seemed like there was going to be. And I, I might be just over indexing on this experience of having lived through this pattern once before, but I feel like the AI moment is set up for that to happen to a lot of people again. Where we, for example, just this week, right, ChatGPT launches medical version, which is great. I think that's going to be awesome for a Lot ton of people and the fact that I can now just connect my ChatGPT into an EMR instead of having to go recently copy and paste or find some other third party thing to do. That kind of connector work for me, like the consumer surplus of that is going to be amazing. That's like my strongest belief is we will see high consumer surplus. But for the businesses it seems like it does create a very tricky balance where you're like, I want to go do a bunch of cool stuff, but how do I know which of these things will be durable over time? How do I know I'm not just doing like R and D for the next generation of the mega platforms that are ultimately going to eat my lunch? How do you guys think about like where you want to place your bets, but what is going to get absorbed into the models versus what you know, only you can do over a longer period of time?

1:15:33

Speaker C

Yeah, I think we are in a fortunate place because ultimately data is an incredibly important asset for all companies. That is what kind of defines and differentiates them and that is not getting commoditized anytime soon. So as a data platform we sit in that layer in between the application and the model, if you will. So the way we think about this is how do we help our customers build very high quality products that are catered towards their use cases, which is all powered by their data. And whether any of that can be subsumed by these other platforms, I am sure the shapes of products will continue to evolve. I think we're still in the very early innings of huge transformation. But I also believe intelligence is a Once these models are out there and there is enough competition and we're seeing enough competition, the dynamics seem to be playing in such a way that is all pro customers and consumers versus these mega platforms. So I do believe the competition will keep things in check and the opportunity is so massive that that growing pie will also create lots of new opportunities.

1:17:56

Speaker A

Obviously people have to have someplace to keep their data. One tell me what's wrong with this theory. If I were to be a skeptic or if I took the perspective of the snowflake bear for a second, a recent experience I had was my company's been a customer of Intercom for a number of years and I was trying to do just some basic analysis of recent tickets. So I and they didn't have the dashboard to do what I wanted to do. So I went to their docs and the docs were 100 pages of docs, right? It's full featured platform at this point. So a lot of docs. So the first thing I did was told a web agent, hey, go compile all these docs. And it literally went page by page, copied them all into a Google Doc, like in a browser. I ended up with some 600 pages of text. Then I took that to Gemini and said, okay, hey, there's a lot of repetition in here, but can you streamline that down to what I really need to know? And so it did that fit in them, even 600 pages, whatever, like, fit in the million tokens. So now I've got my consolidated single view of the docs. Then I go to a coding agent and I say, hey, here's what I really want to do, is export all my data from Intercom. And that also ended up being just one prompt to work. And then it exported all my data from Intercom and then it was able to do the analysis I wanted to do. But then the sort of eureka moment was like, wow, it's never been easier to unplug from Intercom if I want to take my data somewhere else. They didn't really anticipate it being that easy when they created all these APIs. So what. What prevents the data platforms of today from running into trouble there in the past, presumably, like somebody were to say, hey, I'm not happy with you, or I want a better price, whatever, you had some leverage of just, what are you going to do? You're going to pull out all your data? And I'm not saying that would be easy to say because I know you guys handle vastly more data than I have in Intercom, but it does also strike me that, like, it's become a lot easier to move things around. It's become a lot easier to understand what it is, especially once you've gone and done all this metadata layering on. So what are the moats? What are the sticking points? Has it changed or will it change?

1:19:07

Speaker C

It is going the direction that you're calling out, which is, I think, great for, again, customers and consumers. So today Snowflake supports open file formats for storing data. We support Iceberg, which is a file format. What that means is you do not have to have your data locked in somewhere. You can just put it in a. You can keep it in a managed place that's managed by you or by us, and then you can use Snowflake as an engine to go process your data. So we are absolutely embracing and supporting the ability for, for our customers to use these open file formats to not necessarily feel like they're locked into one platform. And I think that's great. That's great for customers, that's great for innovation. Ultimately, customers will end up using the product that is going to give them the best performance, best cost for the things that they'd like to do. And we're absolutely embracing that.

1:21:06

Speaker A

Does that translate to increased pressure on you and your team? Like, it would seem like maybe one way to think about that would be like in the past, if somebody wasn't happy for a year, maybe they would start to think about a switch where now it might be if they're not happy for a quarter. Does it shrink the timeline where you have to deliver?

1:22:10

Speaker C

I love it. It's great for product teams. Right. Ultimately, we're all driven by creating value for customers, building great products and, and we want to do that as fast as possible and competition allows. It's a great incentive in the system to go keep things in check and then have you deliver. So I don't think things change for my team. We already do feel pressure, not necessarily because of competition, because of the opportunity. The opportunity is massive and there's never been a great time to be a product manager. So you're easily able to build awesome products very, very quickly and then you're sprinting. So it's incredibly satisfying to build these great products and then you also reap the benefits by seeing how these are getting used in the market. So I love the competition. I also love the pace of innovation in the industry.

1:22:30

Speaker A

Does that lead you to a point of view? Big picture on this is a classic question, and again, it's striking to me how very informed and technically sophisticated people have very different answers. Where does the value. How do you think about the breakdown of where value accrues? Obviously we've got infrastructure, whether that's like chip creators or owners models, application layer on top. If you had to assign those three layers, relative value capture from the sort of AI opportunity, how do you think that breaks down?

1:23:25

Speaker C

Yeah, I think maybe the way I think about it is the middle will erode and the sides will continue expanding. So far we've been seeing a lot of value accruing to chip makers, Nvidia as well as the model providers, and then increasingly application developers who are able to go build very quickly unique businesses on top of these capabilities. Cursor comes to mind as an example. So I absolutely do see the value continuing to accrue at the infrastructure layer as well as at the application layer. And traditionally there's always been this kind of middle layer that's kind of Facilitating and connecting those two things. Because of the capabilities of these models, that middle layer may not be as valuable or as important anymore.

1:23:58

Speaker A

And that middle layer is the models, is that right?

1:24:51

Speaker C

No, no, no. I do believe the middle layer is all the companies that are kind of creating custom business logic for certain applications. So that business logic, as you called out, for instance, if you want to just build your own extractor, you can just wipe, code it over a weekend and go do it, versus a company that goes and builds it for you. So that layer isn't as important. So models are, in my opinion, will continue to accrue a ton of value.

1:24:54

Speaker A

Yeah. Okay, so to try to play that back to you, it sounds like you think all three of the layers that I described will do fine, but at the application, traditional businesses will change. Yeah, you're going more horizontal platform and less, you know, like relatively less excited about vertical. Because the fact, I mean so many SaaS applications like essentially exist to encode business logic or best practices or whatever. And we just probably don't need dedicated teams building out those kinds of things when we can just have agents kind of do it on the fly as. As needed.

1:25:26

Speaker C

That's what I'm guessing over time. Yeah.

1:26:03

Speaker A

Yeah. Okay, cool. One other big kind of question that I've asked a lot of people a lot of times and I think you're the perfect person to touch on it. So you, of course you know that databricks acquired this company called MosaicML not too long ago, maybe two years ago now. And what Mosaic was doing I thought was really interesting, which was starting with open source models, working with particular customers to do continued pre training on data sets, which I assume were very often internal data sets, like the sort of data sets that might sit in a snowflake. I was really surprised. I spoke to Ali Gazi, the CEO of Databricks at an event not too long ago and he said we killed that product. So they basically turned Mosaic into a in house research unit. But that product of offering this continued pre training to try to create a model that like really knows your business inside and out, basically they don't offer it anymore. I was very surprised by this because I think, man, if I had a, if I'm GE or if I'm 3M or any number of hundred year old millions of employees over the generations, companies that have this incredible history and so much data that's accrued that nobody really understands at the company these days. If I could have a model that could have similar command of that information that doesn't that only exists in my company and just nobody else outside has ever had access to. As the foundation models today generally have world knowledge, I would think that would be insanely valuable for a lot of enterprises. And yet we don't seem to be seeing to my knowledge many instances of whatever 3M GPT or GE GPT, Pfizer GPT. Like why don't we see that? Do you have a point of view?

1:26:05

Speaker C

I do. So this is kind of similar to how up until recently when you'd ask a question on ChatGPT you'll say hey, my information cutoff is whatever a year ago and I can only answer questions up to that point. And then web search as a tool came in and now all of these platforms would use web search to give you the most up to date information so that their world information can be it is more about the intelligence to figure out when to use the tool to retrieve the information and then make sense of it and then give it back to you versus having been trained, pre trained with all that information upfront, to me that pattern is exactly what's playing out right? So in the enterprise world you have a lot of information and then your text to SQL and RAG solutions can bring that information in for the agent, for the platform to reason with and then give information. The nice thing about that is it is substantially cheaper, the model keeps getting better as the underlying premier model keeps improving and it's also relatively easily tunable, you can update it, you can change things and so forth. So that means for me majority of businesses would continue to benefit from this architecture rather than codifying all of that information in the weights of the model, they'll just use the information and then use tools to retrieve parts of the information that are relevant. The exception to that is what we discussed earlier, which is if there are certain tasks that require either high throughput, low cost, if you have a lot of data and in an area that the model has not seen before, then it might make sense to go create custom models for those specific tasks. So I do believe there's going to be increasingly need for kind of task specific small models in large corporations or when you have we have that need, but still the majority of the use cases will be more retrieval oriented.

1:27:58

Speaker A

So I think that's a great first pass answer. If I think though even just about my own ability to search through my own stuff, right? My own Gmail, my own Google Docs. One of the intuitions I have pretty strongly is if I were to give you full access to my Gmail and give you full access to my Google Docs, you couldn't search through it nearly as well as I can. And that's despite the fact that you're clearly smarter than me. So I'm like, there seems to be something about the fact that I have had this like free training on this corpus that allows me to even just search through it a lot better. Because if nothing else, like one of the intuitions is I know when I found what I'm looking for.

1:30:09

Speaker B

Right?

1:30:52

Speaker A

You might not know. You could certainly you could do a hundred searches in my Google Drive and never be quite confident you got the absolute best document for whatever the question is. Whereas if it's my Google Drive and I created all those documents when I hit that document, that is the one that's yes, this is the one now. Yes, I remember this now. This is the one. So I have that sort of confidence that I've got to the answer. If nothing else I think which is strikes me as like really hard. And I've seen this when I try to give my own just whatever give Claude access to search my drive. Like it also struggles in that way. It doesn't know know how many times the search or if it's found the right thing or sometimes it's satisfied too easily, whatever. So I still feel like there's something there where you could expect that a model that really had a sort of more in the weights familiarity with the data could do a better job of navigating it. And then maybe it just comes down to upgrade cycles are terrible for this kind of thing. And yeah, as you said, like you want like keep taking advantage of better and better models. This potentially Rakesh has obviously influenced the discourse recently with kind of focus on continual learning. So maybe you need either a new architecture that's more suited to that or some sort of new training paradigm that would be more suited to it. I guess maybe one way to phrase this is if that were to flip. If you imagine a world a year from now where it's no longer the case that the best approach is take the best models and like tune, leave them as they are, but tune them through the search as you described, and it instead becomes one of these things where they actually do have this deeper familiarity with all the enterprise's data, what do you think would what would have changed to flip us from one paradigm to the other?

1:30:52

Speaker C

I was trying to think about how humans do this. We'd go into a environment that we don't know and then we'll go do a bunch of searches and then we'll read to create, to create some kind of knowledge. And then as you build out that knowledge, there's intuition that comes with it, so you don't need to keep referring back to it. And then somehow that turns into intuition. Right. Right now I think what these models are doing is the first part. I'll go pick the information as the context windows of these models also keep getting better. I can stuff more and more information in these models and then get an answer. What intuition is not really understood. So I don't really know how that changes the dynamics. What would change? If a model is trained with your data, you clearly need much less data to steer it to a certain direction. You'll have much more consistency in the responses. It still feels, I don't think you can ever get away from feeding it information, up to date information and so forth. But what I would imagine happen is first of all, that model that you want to do certain tasks doesn't have to write a poem in French. So you'll benefit from using the weights more efficiently for the task that you want to do. And therefore, perhaps again, you may not need as large a model. So you get benefits from more optimizations to reduce the cost, increase the speed and so forth.

1:32:32

Speaker A

Yeah, I think that many small models paradigm is also one that I'm pretty bullish on for quite a few reasons. One being just I think we stand a lot better chance of staying in control of the meta if we have a lot of like narrow AIs doing their jobs as opposed to, you know, a relatively smaller number of giant AIs like running things for us, the pull of that is obviously pretty strong. But the. I do worry that we're racing into having such general AIs that can do sort of anything before we've really thought through what the ultimate consequences of that are going to be. And the narrowness, safety through narrowness and maintaining control through narrowness, I think is an underdeveloped paradigm.

1:33:51

Speaker C

I fully agree that that's again, just going back to human analogy. That's how we operate as well. There are certain parts of the brain that are specialized to do certain tasks. And yeah, so I can absolutely see.

1:34:32

Speaker A

That this has been amazing. Couple of quick closing questions. What are you watching right now in terms of horizon scanning for surprises? Is there, is there a capability threshold that's on your mind? Or like what? Because obviously nobody can keep up with the AI news these days. Right. So everybody has to pick and choose what are the areas that you're watching and maybe another Dimension of is what are the metrics that you're watching? Are you looking at ARC AGI scores? Are you looking at GDP val? Are you looking at the MIRI task length chart? Are there other. What do you trust to give you the highest signal on what is actually important in the latest things that are coming out?

1:34:42

Speaker C

I actually don't watch the public benchmarks as closely. We do have internal series of benchmarks that watch very closely in terms of quality and latency for the tasks that we're optimizing for, which is of course built on top of the models that we get from the model providers. Whenever there's a new model that's about to be released, we'll go run out our tests, figure out what's improving, what are the gaps, and that I watch very closely. In terms of technology trends. One thing that is maybe unique to Snowflake, of course, is we have a lot of tabular data and that technology so far has been all about text to SQL and semantic models. So watch that space quite, quite carefully and there are some new trends happening in that space. These tabular foundation models are interesting. Being able to quickly build forecasting models and so forth are now possible. So those are other trends that I watch as well.

1:35:20

Speaker A

Yeah, that's an interesting one. There's a few public forecasting like benchmarks and competitions. I think those are really interesting too at the point where the models are better able to predict the future than our best superforecasters or even aggregations of superforecasters that will feel, I think, like a very meaningful shift in, in what's going on in the world. Any contrarian takes any. Anything you think, like the. Anything you want to correct that you think the audience at large might be misled or misconceiving right now.

1:36:11

Speaker C

I mean, we touched upon the one that. That's very top of mind for me right now. I think, you know, the way we build products is changed. I don't think it's contrarian, but I don't think it's happening fast enough. We're at a point where how we build products needs to radically change. And that means change of behavior because we've been trained to build products one way. So to me, that's the biggest one. In a world where these coding agents are such kind of capable platforms, how do you build new products? And in my mind it is all about starting with that first and then validating things quickly before you build the product in the first place.

1:36:41

Speaker A

Yeah, I think that's, I've been doing that with My mom, I made it, made the custom travel app, travel planning app for the holidays. And it has been, yeah, it's inverting that process. Right. I made a version, she has it and now I sat down with her this morning over coffee and I'm like, what do you want this thing to do that it can't do? And she's I don't want to ask you to do more on this. I'm like, mom, it's honestly so easy at this point. If you can articulate what you're missing, there's a pretty good chance we can get Claude code to check. Just make it and you can have it from one session to the next. My last question, then I'll give you the final word. What advice do you have for enterprise leaders in general? Obviously, you guys are much closer to the core. For all the executives and product owners at the companies that you serve, what do you think they should better appreciate or what can they learn from your experience?

1:37:24

Speaker C

There are some enterprises that are still quite careful about adopting AI and at this point it is, it is so powerful that there's a race. So the faster enterprises adopt AI, the more benefit they're going to get and the more intuition that they're going to get that changes the trajectories of these businesses. So to me, it's incredibly important to intuitively understand and natively use AI because it is going to change industries. And then underlying that is all about. Many of the hesitations tend to be about getting the data ready for AI from our perspective. So that means investing in that core foundation to essentially get the data AI ready for AI to use. So that means breaking down silos, getting the data accessible, locked in for certain use cases. That becomes a core enabler to build on top of.

1:38:10

Speaker A

Makes sense. We've covered a lot of ground. I really appreciate your time and jumping through all these topics with me. Anything else that we didn't touch on that you would want to leave people with?

1:39:14

Speaker C

No. Maybe the thing to call out is we talked about a lot of great capabilities as well as trends. One thing that is sometimes not necessarily appreciated is how easy it is to use AI and how easy it needs to be to use AI for adoption. So that's an area for Snowflake that's super cool. So as we build products, making it very easy to deploy high quality AI at scale is something that we strive towards. To me, from a design principle perspective, that is key as well.

1:39:22

Speaker A

Yeah, couldn't agree more. Baris Gultigan, Vice President of AI at Snowflake thank you for being part of the Cognitive Revolution.

1:39:52

Speaker C

Thank you Nathan. Thanks for having me.

1:39:58

Speaker B

If you're finding value in the show, we'd appreciate it if you'd take a moment to share it with friends, post online, write a review on Apple Podcasts or Spotify, or just leave us a comment on YouTube. Of course, we always welcome your feedback, guest and topic suggestions and sponsorship inquiries either via our website Cognitiverevolution AI or by DMing me on your favorite social network. The Cognitive Revolution is part of the Turpentine Network, a network of podcasts which is now part of a 16Z where experts talk technology, business, economics, geopolitics, culture and more. We're produced by AI Podcasting. If you're looking for podcast production help for everything from the moment you stop recording to the moment your audience starts listening, check them out and see my endorsement@aipodcast.ing. and thank you to everyone who listens for being part of the Cognitive Revolution.

1:40:01