Inside AI’s $10B+ Capital Flywheel — Martin Casado & Sarah Wang of a16z
Martin Casado and Sarah Wang from a16z discuss the unprecedented capital flywheel in AI, where companies can raise massive rounds, convert dollars directly into compute and capabilities, then use breakthrough results to raise even larger subsequent rounds. They explore how traditional venture/growth investing boundaries are blurring, the potential for foundation models to outspend entire ecosystems built on top of them, and whether the industry is heading toward oligopoly or fragmentation.
- AI companies can convert capital directly into capabilities through compute, creating a unique flywheel where breakthrough results enable 3x larger subsequent rounds
- Foundation models may be able to raise more capital than the aggregate of all companies building on top of them, potentially consuming entire ecosystems
- The line between venture and growth investing is blurring as AI deals require large capital but early-stage analysis and business development
- Traditional software companies are being overlooked despite offering solid returns, as investors chase extreme growth rates
- The talent war premium has created a new baseline where even non-flashy AI roles command tens of millions in compensation
"For the first time you can actually trace dollars to outcomes, right? Provided that scaling laws are holding and capabilities are actually moving forward."
"A model company can raise money and drop a model in a year and it's better. Right. And it does it with a team of 20 people or 10 people."
"If you can raise three times more every subsequent round, they probably can raise more money than the entire app ecosystem that's built on top of it."
"I've never seen the perception of the truth be further from the truth industry wide ever."
"There's no dark GPUs. Right. I mean, and so circular or not, I mean if someone invests in a company that, you know, they'll actually use the GPUs."
Hey everyone, welcome to the Late in Space podcast. Live from a 16Z. This is Alessio, founder of Kernel Labs and I'm joined by Zwicks, editor of Late in Space.
0:00
Hey, hey, hey. And we're so glad to be on with you guys. Also a top AI podcast. Martin Casado and Sarah Wang, welcome.
0:08
Very happy to be here and welcome.
0:16
Yes, we love this office. We love what you've done with the place. The new logo is everywhere now. It's still getting, takes a while to get used to, but it reminds me of like sort of a callback to a more ambitious age which I think
0:17
is kind of definitely makes a statement. Yeah, not quite sure what that statement is, but it makes a statement.
0:31
Martin, I go back with you to Netlify and you know you created software defined networking and all that stuff. People can read up on your background. Sarah, newer to you, you sort of started working together on AI infrastructure stuff.
0:37
That's right, yeah, seven, seven years ago now.
0:51
Best growth investor in the entire industry.
0:53
Oh, say more.
0:55
Hands down, Sarah is Sarah. I mean when it comes to AI companies, Sarah I think has done the most kind of aggressive investment thesis around AI models. Right. So she worked with Noam Shazir, Mira, Ilia, Fei Fei and so just these frontier kind of like large AI models. I think Sarah's been the broadest investor. Is that fair?
0:56
No, well, I was going to say I think it's a really interesting tag team actually just because a lot of these big CDs deals, not only are they raising a lot of money, it's still a tech founder bet, which obviously is inherently early stage, but the resources, I was going to say the resources. One they just grow really quickly, but then two, the resources that they need day one are kind of growth scale. So I think the hybrid tag team that we have is quite effective, I think.
1:20
What is growth these days? You know, you don't wake up if it's less than a billion or like
1:46
it's actually, it's actually very like, like. No, it's a very interesting time in investing because like you know, take like the character route. Right. Tends to be like pre monetization but the dollars are large enough that you need to have a larger fund. And the analysis, you know, because you've got lots of users, because this stuff has such high demand, requires, you know, more of a number sophistication and so most of these deals, whether it's us or other firms on these large model companies are like this hybrid between venture and growth.
1:52
Yeah, total and I think you know, stuff like BD for example. You wouldn't usually need BD when you were seed stage trying to get biz dev. Biz dev, exactly. But like now, I'm not familiar, what
2:18
does biz dev mean for a venture fund? Because I know what biz dev means for a. Yeah.
2:28
So a good example is, I mean we talk about buying compute, but there's a huge negotiation involved there in terms of, okay, do you get equity for the compute? What sort of partner are you looking at? Is there a go to market arm to that? And these are just things on this scale hundreds of millions maybe six months into the inception of a company. You just wouldn't have to negotiate these deals before.
2:32
Yeah, these large rounds are very complex now. Like in the past, if you did a series A or a series B, like whatever, you're writing a 20 to a $60 million check and you call it a day. Now you normally have financial investors and strategic investors and then the strategic portion always still goes with like these kind of large compute contracts which can take months to do. And so it's a very different ties this. I've been doing this for 10 years. It's the. I've never seen anything like this.
2:54
Yeah. Do you have worries about the circular funding from some of these strategics?
3:19
Listen, as long as the demand is there, like the demand is there. Like the problem with the Internet is the demand wasn't there.
3:24
Exactly. All right, this is like the, the whole pyramid scheme bubble thing where like as long as you mark to market on like the notional value of like these deals, fine. But like once it starts to chip away, it really.
3:29
Well, no, as long as there's demand, I mean, you know, listen, this is like a lot of these sound bites have already become kind of cliches, but they're worth saying it, right? Like during the Internet days, like we were raising money to put fiber in the ground that wasn't used. That's a problem. Right? Because now you actually have a supply overhang. And even in the time of the Internet, the supply and bandwidth overhang, even as massive as it was, as massive as the crash was, only lasted about four years. But we don't have a supply overhang. There's no dark GPUs. Right. I mean, and so circular or not, I mean if someone invests in a company that, you know, they'll actually use the GPUs. And on the other side of it is the, is the actual customer. So I think it's a different time.
3:40
I think the other piece Maybe just to add onto this, and I'm going to quote Martin in front of him, but this is probably also a unique time in that for the first time you can actually trace dollars to outcomes, right? Provided that scaling laws are holding and capabilities are actually moving forward. Because if you can put translate dollars into capabilities, a capability improvement, there's demand there, to Martin's point. But if that somehow breaks, you know, obviously that's an important assumption in this whole thing to make it work. But you know, instead of investing dollars in sales and marketing, you're investing into R and D to get to the capability, you know, increase. And that's sort of been the demand driver because once there's an unlock there, people are willing to pay for it.
4:25
Is there any difference in how you build the portfolio now that some of your growth companies are like the infrastructure of the early stage companies, like, you know, OpenAI is now the same size as some of the cloud providers were early on. What does that look like? How much information can you feed off each other between the two?
5:06
There's so many lines that are being crossed right now or blurred. So we already talked about venture and growth. Another one that's being blurred is between infrastructure and apps. So what is a model company? It's clearly infrastructure because it's doing kind of core R and D. It's a horizontal platform, but it's also an app because touches the users directly. And then of course, you know, the growth of these is just so high. And so I actually think you're just starting to see a new financing strategy emerge. And you know, we've had to adapt as a result of that. And so there's been a lot of changes. You're right that these companies become platform companies very quickly. You've got ecosystem built out, so none of this is necessarily new, but the timescales in which it's happened is pretty phenomenal. And the where we'd normally cut lines before is blurred a little bit. But that said, I mean, a lot of it also just does feel like things that we've seen in the past, like cloud build out and the Internet bailed out as well.
5:24
Yeah, yeah, I think it's interesting. I don't know if you guys would agree with this, but it feels like the emerging strategy is. And this builds off of your other question. You raise money for Compute, you pour that, or you pour the money into Compute, you get some sort of breakthrough, you funnel the breakthrough into your vertically integrated application. That could be ChatGPT, that could be cloud code, you know, whatever it is you massively gain share and get users, maybe you're even subsidizing at that point. Depending on your strategy, you raise money at the peak momentum and then you repeat, rinse and repeat. And that wasn't true even two years ago, I think. And so it's sort of to your just tying it to fundraising strategy. Right. There's a hiring strategy all of these are tied to. I think the lines are blurring even more today where everyone is. But of course these companies all have API businesses and so there are these frenemy lines that are getting blurred in that a lot of, I mean they have billions of dollars of API revenue. Right. And so there are customers there, but they're competing on the app layer.
6:24
Yeah.
7:24
So this is a really, really important point. So I would say for sure, venture and growth, that line is blurry. App and infrastructure, that line is blurry. But I don't think that changes our practice so much. But like where they're very open questions. Does this layer in the same way compute traditionally has, like during the cloud is like, you know, like whatever, somebody wins one layer, but then another whole set of companies wins another layer. But that might not be the case here. It may be the case that you actually can't verticalize on the token string. Like you can't build an app. Like it necessarily goes down just because there are no abstractions. So those are kind of the bigger existential questions we ask. Another thing that is very different this time than in the history of computer science is, is in the past, if you raised money, then you basically had to wait for engineering to catch up, which famously doesn't scale like the mythical mammoth. I take a very long time, but like that's not the case here. Like a model company can raise money and drop a model in a, in a year and it's better. Right. And it does it with a team of 20 people or 10 people. So this type of like money entering a company and then producing something that has demand and growth right away and using that to raise more money is a very different capital flywheel than we've ever seen before. And I think every. Everybody's trying to understand what the consequences are. So I think it's less about like big companies and growth and this and more about these more systemic questions that we actually don't have answers to.
7:24
Yeah, like at Kernel Labs, one of our ideas is like if you had unlimited money to spend productively to turn tokens into products, like the whole early stage market is very different because today you're investing X amount of capital to win a deal because of price structure and whatnot. And you're kind of pot committing to a certain strategy for a certain amount of time. But if you could like iteratively spin out companies and products and just throw. I want to spend $1 million of inference today and get a product out tomorrow. Yeah, like we should get to the point where like the friction of like token to product is so low that you can do this and then you can change the, the early stage venture model to be much more iterative. And then every round is like either 100k of inference or like 100 million from a 16C. There's no, there's no like $8 million year round anymore.
8:49
But, but, but, but there's a, there's a, the, an industry structural question that we don't know the answer to, which involves the frontier models, which is, let's take Anthropic. Let's say anthropic has a state of the art model that has some large percentage of market share. And let's say that, you know, a company is building smaller models that, you know, use the bigger model in the background. You open 4.5, but they add value on top of that. Now if Anthropic can raise three times more every subsequent round, they probably can raise more money than the entire app ecosystem that's built on top of it. And if that's the case, they can expand beyond everything built on top of it. It's like, imagine like a star that's just kind of expanding. So there could be a systemic, there could be a systemic situation where the soda models can raise so much money that they can outpay anybody that builds on top of them, which would be something I don't think we've ever seen before just because we were so bottlenecked on engineering. And it's a very open question.
9:38
Yeah, it's almost like bitter lesson applied to the startup industry.
10:41
100%. Yeah. It literally becomes an issue of like raise capital, turn that directly into growth, use that to raise three times more. And if you can keep doing that, you literally can outspend any company that's built. Not any company. You can outspend the aggregate of companies on top of you and therefore you'll necessarily take their share, which is crazy.
10:45
Would you say that kind of happens a character? Is that the sort of postmortem on what happened?
11:02
No. Yeah.
11:10
Because I think, I mean the actual
11:12
postmortem is he wanted to go back
11:14
to Duke, but like that's Another difference. We should actually talk about this.
11:15
Go for it. Take it.
11:23
I was going to say, I think the character thing raises actually a different issue, which actually the Frontier Labs will face as well. So we'll see how they handle it. But so we invested in character in January 2023, which feels like eons ago. I mean, three years ago feels like lifetimes ago, but. And then they did the IP licensing deal with Google in August 2024. And so, you know, at the time, Gnome, you know, he's talked publicly about this, right? He wanted to. Google wouldn't let him put out products in the world. That's obviously changed drastically. But he went to go do that. But he had a product attached. The goal was, I mean, it's Noam Shazir. He wanted to get to AGI. That was always his personal goal. But, you know, I think through collecting data, right? And this sort of very human use case that the Character product originally was and still is, was one of the vehicles to do that. I think the real reason that, you know, if you think about the stress that any company feels before you ultimately go in one way or the other, is sort of this AGI versus product. And I think a lot of the big, I think, you know, OpenAI is feeling that anthropic. If they haven't, you know, felt it, certainly given the success of their products, they may start to feel that soon. And they're real. I think there's real trade offs, right? It's like how many, when you think about GPUs, that's a limited resource, where do you allocate the GPUs? Is it toward the product? Is it toward new research? Right. Is it long term research? Is it toward near to midterm research? And so in a case where you're resource constrained, of course there's this fundraising game you can play, right? But the market was very different back in 2023 too. I think the best researchers in the world have this dilemma of, okay, I want to go all in on AGI, but it's the product usage revenue flywheel that keeps the revenue in the house to power all the GPUs to get to AGI. And so it does make, you know, I think it sets up an interesting dilemma for any startup that has trouble raising up until that level. Right? And certainly if you don't have that progress, you can't continue this flight, you know, fundraising flywheel.
11:24
I would say that because, because we're keeping track of all of the things that are different, right? Like, you know, Venture growth and app infra. And one of the ones is definitely the personalities of the founders. It's just very different this time. You know, been a very. Been doing this for a decade and I've been doing startups for 20 years. And so, I mean, a lot of people start this to do AGI and we've never had like a unified North Star that I recall. In the same way, like people built companies to start companies in the past. Like, that was what it was like. I would create an Internet company, I would create an infrastructure company. Like, it's kind of more engineering builders. And this is kind of a different, you know, mentality. And some companies have harnessed that incredibly well because their direction is so obviously on the path to what somebody would consider AGI, but others have not. And so like, there is always this tension with personnel. And so I think we're seeing more kind of founder movement.
13:32
Yeah.
14:27
You know, as a fraction of founders than we've ever seen. I mean, maybe since, like, I don't know, the time of like Shockley and the trade jurisdicte or something like that, way back at the beginning of the industry. I mean, it's a very, very unusual type of personnel.
14:27
Totally. And it, I think it's exacerbated by the fact that talent wars, I mean, every industry has talent wars, but not at this magnitude. Very rarely can you see someone get poached for $5 billion. That's hard to compete with. And then secondly, if you're a founder in AI, you could fart and it would be on the front page of, you know, the information these days. And so there's sort of this fishbowl effect that I think adds to the deep anxiety that these AI founders are feeling.
14:39
Yes.
15:07
I mean, just briefly comment on the founder, the sort of talent wars thing. I feel like 2025 was just like a blip. Like, I don't know if we'll see that again because Meta built the team. Like, I don't know if I think. I think they're kind of done and like, who's going to pay more than Meta? I don't know.
15:07
I agree. So it feels this way to me too. It's like it was like basically Zuckerberg kind of came out swinging and then now he's kind of back to building. Yeah, yeah.
15:23
You know, you got to like pay up to like assemble team to rush the job, whatever. But then now, now you like, you made your choices and now they got a ship.
15:31
Right.
15:37
Like, I mean, the other side of that is like, you Know, like we're actually in the job hiring market. We've got 600 people here. I hire all the time. I've got three open racks if anybody's interested that's listening to this investor. Yeah. On the team, like on the investing side of the team. Like, and a lot of the people we talk to have acting, you know, active offers for 10 million a year or something like that. And like, you know, and we pay really, really well. And just to see what's out on the market is really, is really remarkable. And so I would just say it's actually. So you're right. Like the really flashy one, like I will get someone for, you know, a billion dollars, but like the inflated trickles down. Yeah, it's still very active today.
15:37
I mean, yeah, you could be an L5 and get an offer in the tens of millions.
16:17
Yeah, easily.
16:22
So I think you're right that it felt like a blip. I hope you're right. But I think it's been the steady state.
16:23
Is everything got pulled up.
16:29
Yeah, yeah, exactly.
16:30
For sure. Yeah.
16:31
Yeah. And I think that's breaking the early stage founder math too. I think before a lot of people were like, well, maybe I should just go be a founder instead of like getting paid, yeah, 800k a million at Google. But if I'm getting paid 5, 6 million, that's different.
16:32
But on, but on the other hand, there's more strategic money than we've ever seen historically. Right. And so the economics, the calculus on the economics is very different in a number of ways. And it's caused a ton of change in confusion in the market. Some very positive, some negative. Like, so for example, the other side of the, the co founder, like acquisition, you know, Mark Zuckerberg. Poaching someone for a lot of money is like, we're actually seeing historic amount of M and A for basically acqua hires. Right. That you like, you know, really good outcomes from a venture perspective that are effective. Aqua hires. Right. So I would say it's probably net positive from the investment standpoint. Even though it seems from the headlines to be very disruptive in a negative way.
16:44
Yeah, let's talk maybe about what's not being invested in. Like maybe some interesting ideas that you will see more people build or it seems in a way, you know, as YC has gotten more popular, X has gotten more popular, there's a startup school path that a lot of founders take and they know what's hot in the VC circles and they know what gets funded and there's maybe not as Much risk appetite for things outside of that. I'm curious if you feel like that's true and what are maybe some of the areas that you think are under discussed?
17:33
I mean, I actually think that we've taken our eye off the ball in a lot of like just traditional, you know, software companies. So like, I mean, you know, I think right now there's almost a barbell. Like you're like the hot thing on X, you're deep tech, right? But I, you know, I feel like there's just kind of a long list of like good, good companies that will be around for a long time in very large markets. So you're building a database, you know, say you're building, you know, kind of monitoring or logging or tooling or whatever. There's some good companies out there right now, but like they have a really hard time getting the attention of investors and it's almost become a meme, right? Which is like if you're not basically growing from 0 to 100 in a year, you're not interesting. Which is just the silliest thing to say. I mean, think of yourself as like an individual person. Like, like your personal money, right? So your personal money, will you put it in the stock market at 7% or you put it in this company growing in a very large market, of course you can put in the company 5x. So it's just like we say these stupid things, like if you're not going from 0 to 100, but like those, like who knows what the margins of those are when clearly these are good investments for anybody, right? Like our LPs want whatever 3x net over, you know, the life cycle of a fund, right? So a company in a big market going 5x is a great investment. We'd, everybody would be happy with these returns, but we've got this kind of mania on these, these strong growths. And so I would say that that's probably the most underinvested sector right now.
18:06
Boring software, boring enterprise software.
19:29
Just traditional, really good.
19:31
No AI here.
19:33
Well, the AI of course is pulling them into use cases. But that's not what they are. They're not on the token path, right? Let's just say that like they're software, but they're not on the token path. Like these are like, they're great investments from any definition, except for like random VC on Twitter saying VC on X saying like it's not growing fast enough. Look at that.
19:35
Maybe I'll answer a slightly different question. Question, but adjacent to what you asked, which is maybe an area that we're not investing right now, that I think is a question. And we're spending a lot of time in regardless of whether we pull the trigger or not. And it would probably be on the hardware side actually. Right. And the robotics. Right. Which is, it's. I don't want to say that it's not getting funding because it's clearly it's, it's sort of non consensus to almost not invest in robotics at this point. But we spent a lot of time in that space and I think for us we just haven't seen the ChatGPT moment happen on the hardware side. And the funding going into it feels like it's already taking that for granted.
19:52
Yeah, yeah.
20:30
But we also went through the drone,
20:31
you know, there's a zipline right out there.
20:33
Was that AV era. And like one of the takeaways is when it comes to hardware, most companies will end up verticalizing. Like if you're, if you're investing in a robot company for an ad for agriculture, you're investing in an ag company because that's the competition and that's surprising. That's supply chain. And if you're doing it for mining, that's mining. And so the ad team does a lot of that type of stuff because they're actually set up to diligence that type of work. But for like horizontal technology investing, there's very little when it comes to robots just because it's so fit for, for purpose. And so we kind of like to look at software solutions or horizontal solutions, like applied intuition, clearly from the AV wave. Deep math, clearly from the AV wave. I would say scale AI was actually a horizontal one for, you know, for robotics early on. So that sort of thing, we're very, very interested. But the actual like robot interacting with the world is probably better for a different team.
20:36
Yeah, yeah. I'm curious who these teams are supposed to be that invest in them. I feel like everybody's like, yeah, robotics, it's important and like people should invest in it. But then when you look at like the numbers like the capital requirements early on versus like the moment of okay, this is actually going to work, let's keep investing. That seems really hard to predict in a way that it's not.
21:28
I mean CO2, Khosla, GC, I mean these are all invested in hardware companies. Is this, you know, and listen, I mean it could work this time for sure. Right? I mean if Elon's doing it, he's like, just the fact that Elon's doing it means that there's going to be a lot of capital and a lot of attempts for a long period of time. So that alone maybe suggests that we should just be investing in robotics. Just because you have this North Star who's Elon with a humanoid and that's going to like basically will into being an industry. But we've just historically found like we're a huge believer that this is going to happen. We just don't feel like we're in a good position to diligence these things. Because again, robotics companies tend to be vertical. You really have to understand the market they're being sold into. Like that's like that competitive equilibria with a human being is what's important. It's not like the core tech and like we're kind of more horizontal cortex type investors. This is Sarah and I. The ad team is different. They can actually do these types of things.
21:49
Just to clarify, AD stands for American Dynamism. All right.
22:42
Yeah, yeah.
22:46
I actually, I do have a related question that first of all I want to acknowledge also. Just on the chip side. Yeah. I recall a podcast that, where you were on, I think it was the Async podcast about two or three years ago where you, where you suddenly said something which really stuck in my head about how at some point, at some point kind of scale, it makes sense to build a custom ASIC for per run.
22:46
Yes. It's crazy. Yeah.
23:07
I think you estimated 500 billion something.
23:09
No, no, no. A billion dollar training run of one billion dollar training run. It makes sense to actually do a custom ASIC if you can do it in time. The question now is timeline.
23:12
Yeah.
23:19
Not money. Because just, just, just rough math. If it's a billion dollar training run, then the inference for that model has to be over a billion, otherwise it won't be solvent. So let's assume it's if you could save 20%, which you save much more than that with an ASIC. 20%, that's $200 million. You can tape out a chip for $200 million. Right. So now you can literally like justify economically, not time warning wise. That's a different issue. An ASIC per model, which is because
23:19
that's how much we leave on the table every single time we do like generic Nvidia.
23:44
Exactly, exactly. No, it's actually much more than that. You could probably get, you know, a factor of two which would be $500 million.
23:49
Typical MFU would be like 50. And that's good.
23:54
Exactly. Yeah, 100.
23:57
So yeah. And I just want to acknowledge like here we are in 2025 and opening eyes confirming like Broadcom and all the other custom silicon deals, which is incredible. I think that speaking about ad, there's a really interesting tie in that obviously you guys are hit on which is sort of like America first movement or sort of re. Industrialize here and move TSMC here, if that's possible. How much overlap is there from AD to I guess growth and investing in particularly US AI companies that are strongly bounded by their compute. Yeah, yeah.
23:58
So I mean, I would view, I would view AD as more as a market segmentation than like a mission. Right. So the market segmentation is it has kind of regulatory compliance issues or government, you know, sale or deals with like hardware. I mean they're just set up to, to, to, to. To diligence those types of companies. So it's more of a market segmentation thing. I would say the entire firm, you know, which has been. Since it's been incepted, you know, has geographical biases. Right. I mean, for the longest time we're like, you know, Bay Area is going to be like where the majority of the dollars go.
24:32
Yeah.
25:03
And, and listen there. There's actually a lot of compounding effects for having a geographic bias. Right. You know, everybody's in the same place. You've got an ecosystem, you're there, you've got presence, you've got a network. And I mean, I would say the Bay Area is very much back. You know, like I, I remember during pre covet, like it was like almost crypto, had kind of pulled startups.
25:03
Miami.
25:21
Yeah, yeah. New York. Because, you know, because it's so close to finance came up like Los Angeles had a moment because it was so close to consumer. But now it's kind of come back here. And so I would say, you know, we tend to be very Bay Area focused historically, even though of course we invest all over the world. And then I would say, like if you take the ring out, you know, one more it's going to be the U.S. of course, because we know very well. And then One Ring More is going to be getting us and its allies and. Yeah. And it goes from there. Yeah, Sorry.
25:22
No, no, I agree. I think from a. But I think from the intern that's sort of like where the companies are headquartered. Maybe your questions on supply chain and customer base. I would say our customers are. Our companies are fairly international from that perspective. Like they're selling globally. Right. They have global supply chains in some cases.
25:46
I would say also the stickiness is very different.
26:02
Yeah.
26:05
Historically between venture and growth like there's so much company building and venture. So much so like hiring the next pm, introducing the customer, like all of that stuff. Like, of course we're just going to be stronger where we have our network and we've been doing business for 20. I've been in the Bay Area for 25 years, so clearly I'm just more effective here than I would be somewhere else. Where I think, I think for some of the later stage rounds, the companies don't need that much help. They're already kind of pretty mature historically, so like they can kind of be everywhere, so there's kind of less of that stickiness. This is definitely in the AI time. I mean, Sarah is now the chief of staff of like half the AI companies in the Bay Area right now. She's like, ops, Ninja biz, dev biz ops.
26:05
Are you finding much AI automation in your work? Like what is your stack?
26:49
Oh, in my personal stack.
26:53
I mean because like, by the way, the reason for this is it's triggering. Yeah. Like I'm hiring ops people. A lot of founders I know are also hiring ops people and I'm just, you know, it's opportunity. Since you're also like basically helping out with ops with a lot of companies, what are people doing these days? Because it's still very manual as far as I can tell.
26:54
Hmm.
27:13
Yeah. I think the things that we help with are pretty network based in that it's sort of like, hey, how do I shortcut this process? Well, let's connect you the right person. So there's not quite an AI workflow for that. I will say as a growth investor cloud coworker is pretty interesting. Like for the first time you can actually get one shot data analysis. Right. Which you know, if you're going to do a customer database, analyze a cohort retention. Right. That's just stuff that you had to do by hand before. And our team the other. It was like midnight and the three of us were playing with Claude Cowork. We gave it a raw file. Boom. Perfectly accurate. We checked the numbers. It was amazing. That was my like aha moment. That sounds so boring. But you know, that's, that's the kind of thing that a growth investor is like you know, slaving away on late at night. Done in a few seconds.
27:13
Yeah, you got to wonder what the whole like Anthropic Labs, which is like their new sort of products studio, what would that be worth as an independent startup? You know, like a lot.
28:03
Yeah, true. Yeah. You gotta hand it to them. They've been Executing incredibly well.
28:14
Yeah, I mean, to me, like, you know, anthropic, like building on cloud code. I think it makes sense to me. The real pedal to the metal, whatever the phrase is, is when they start coming after consumer with. Against OpenAI and like, that is like red alert at OpenAI.
28:19
I think they've been pretty clear they're enterprise focused.
28:35
They have been. But like, here's like publicly, it's enterprise focused, it's coding.
28:37
Right.
28:42
And then, and, but here's cloud cowork and, and here's like, well, they, apparently they're running Instagram ads for cloud AI on, you know, for, for people.
28:42
I get them all the time.
28:51
Right. And so like, it's kind of like this, the disruption thing of, you know, OpenAI has been doing, consumer been doing, just pursuing general intelligence in every modality. And here's anthropic. They only focus on this thing, but now they're sort of undercutting and doing the whole innovators dilemma thing on everything else. It's very interesting.
28:52
Yeah. There's a very open question. So for me, there's like, do you know that meme where there's like the guy in the path and there's like a path this way? There's a path this way.
29:12
Which way? Western man?
29:20
Yeah, yeah. And for me, like, like all the entire industry kind of like hinges on like two potential futures. So in, in one potential future, the market is infinitely large. There's perverse economies of scale because as soon as you put a model out there, like, it kind of sublimates and all the other models catch up and like, it's just like software's being rewritten and fractured all over the place and there's tons of upside and it just grows. And then there's another path which is like, well, maybe these models actually generalize really well and all you have to do is train them with three times more money. That's all you have to do. And it'll just consume everything beyond it. And if that's the case, like, you end up with basically an oligopoly for everything. Like, you know, because they're perfectly general and like. So this would be like the, the AGI path would be like, these are perfectly general. They can do everything. And this one is like, this is actually normal software. The universe is complicated. You've got. And nobody knows the answer. My belief is if you actually look at the numbers of these companies. So generally, if you look at the numbers of these companies, if you look at like the amount they're making and how much they they spent training the last model, they're gross margin positive. You're like, oh, that's really working. But if you look at like the current training that they're doing for the next model, their gross margin negative. So part of me thinks that a lot of them are kind of borrowing against the future and that's going to have to slow down, that's going to catch up to them at some point in time, but we don't really know.
29:21
Yeah.
30:47
Does that make sense? Like, I mean it could be, it could be the case that the only reason this is working is because they can raise that next round and they can train that next model because these models have such a short life. And so at some point in time, like, you know, they won't be able to raise that next round for the next model and then things will kind of converge and fragment again. But right now it's not totally, I
30:47
think the other, by the way, just a meta point. I think the other lesson from the last three years is, and we talk about this all the time because we're on this Twitter X bubble, but if you go back to let's say March 2024, that period, it felt like, I think an open source model with a benchmark leading capability was sort of launching on a daily basis at that point. And so that's one period. Suddenly it's sort of like open source takes over the world. There's going to be a plethora. It's not oligopoly. If you rewind time. Even before that, GPT4 was number one for nine months, 10 months. It's a long time. Right. And of course now we're in this era where it feels like an oligopoly, maybe some very steady state shifts and it could look like this in the future too, but it's so hard to call. And I think the thing that keeps us up at night in a good way and bad way is that the capability progress is actually not slowing down. And so until that happens. Right. Like you don't know what's going to look like.
31:05
But I would say for sure it's not converged. For sure the systemic capital flows have not converged. Meaning right now it's still borrowing against the future to subsidize growth currently, which you can do that for a period of time, but at the end, at some point the market will rationalize that and just nobody knows what that will look like or the drop in price of compute. We'll save them, who knows?
32:09
Yeah, I think the models need to asymptote to specific tasks. You know, it's like, okay, now opus 4.5 might be AGI at some specific task. And now you can like depreciate the model over a longer time. I think now right now there's like no old model.
32:34
No, but, but let me just change that mental. That's. That used to be my mental model. Let me just change it a little bit. If you can raise three times, if you raise more than the accurate of anybody that uses your models, that doesn't even matter. It doesn't even matter. See what I'm saying? Like, yeah, so I have an API business. My API business is 60% margin margin or 70% margin or 80% margins. It's a high margin business. So I know what everybody is using. If I can raise more money than the aggregate of everybody that's using it, I will consume them whether I'm AGI or not. And I will know if they're using it because they're using it. And like, unlike in the past where engineering stops me from doing that, it's just very straightforward. You just train. So I also thought it was kind of like you must ask AGI. General. General. General. But I think there's also just a possibility of the capital market will just give them the ammunition to just go after everybody on top of them.
32:47
I do wonder though, to your point, if there's a certain task that getting marginally better isn't actually that much better. We've asymptoted to, we can call it AGI or whatever. Actually, Ali Goatse talks about this. We're already at AGI for a lot of functions in the enterprise. For those tasks, you probably could build very specific companies that focus on just getting as much value out of that task that isn't coming from the model itself. There's probably a rich enterprise business to be built there. I mean, could be wrong on that. But there's a lot of interesting examples. So, right. If you're looking at the legal profession or whatnot, and maybe that's not a great one because the models are getting better on that front too. But just something where it's a bit saturated, then the value comes from services, it comes from implementation. Right. It comes from all these things that actually make it useful to the end customer.
33:36
Sorry. One more thing I think is under discussed in all of this is like to what extent every task is AGI component complete.
34:24
Right.
34:31
I code every day. It's so fun.
34:32
That's a poor question. Yeah.
34:35
And like when I'M talking to these models. It's not just code. I mean, it's everything. Right? Like I, you know, like, it's, it's healthcare.
34:36
I mean, it's legal, but it's every.
34:44
It's exactly that. Like, I mean, it's everything. Like, I'm asking these models to. Yeah. To understand compliance. I'm asking these models to go search the web. I'm asking these models to talk about things I know in the history. Like, let's having a full conversation with. With me while I engineer. And so it could be the case that the most AGI complete. I'm not an AGI guy. I think that's. But the most AGI complete model will always win independent of the task. And we don't know the answer to that one either. But it seems to me that, listen, Codex, in my experience is for sure better than Opus 4.5 for coding. It finds the hardest bugs that I work in with, you know, the smartest developers I know work on it. It's great. But I think Opus 4.5 is actually very. It's got a great bedside manner. And it really, it really matters if you're building something very complex because, like, it really, you know, like you're a partner and a brainstorming partner for somebody. And I think we don't discuss enough how every task kind of has that quality. And what does that mean to like, capital investment and like frontier models and sub models? Like, what happened to all the special coding models? Like, none of them worked right.
34:46
Sort of. Some of them didn't even get released.
35:51
Magic.dev There's a whole.
35:53
There's a whole host. We saw a bunch of them. And like, there's this whole theory that, like, there could be a. And I think one of the conclusions is, like, there's no such thing as a coding model. You know, like, that's not a thing. Like you're talking to another human being. And it's good at coding, but, like, it's got to be good at everything.
35:55
Minor disagree. Only because I'm pretty. Have pretty high confidence that basically OpenAI will always release a GPT5 and a GPT5 codecs. Like, that's the code. The way I call it is one for Riz and one for Tiz. And then like someone internally opened it was like, yeah, that's a good way to frame it.
36:10
That's so funny.
36:32
But maybe it collapses down to reason. Tiz and that's it. It's not like 100 dimensions, it's two dimensions. Like, exactly beside manor versus coding, I
36:33
think for anybody listening to this or for AM for you, when you're like coding or using these models for something like that, actually just be aware of how much of the interaction has nothing to do with coding and it just turns out to be a large portion of it. And so I think the best Soto ish model is going to remain very important no matter what the task is.
36:45
Yeah, speaking of coding, I'm going to be cheeky and ask what actually are you coding? Because obviously you could code anything and you're obviously a busy investor and a manager of a giant team. What are you going?
37:06
I help Fei Fei at World Labs. It's one of the investments and. And they're building a foundation model that creates 3D scenes.
37:18
Yeah, we had it under pod. Yeah, yeah, yeah.
37:27
And so these 3D scenes are Gaussian splats, just by the way that kind of AI works. And so like you can reconstruct a scene better with, with, with radiance fields than with meshes because like they don't really have topology. So so they produce these beautiful 3D rendered scenes that are Gaussian splats. But the actual industry support for Gaussian splats isn't great. It's just never. It's always been meshes and things like unreal use meshes. And so I work on a open source library called Spark JS, which is a JavaScript rendering layer ready for Gaussian splats. And it's just because you need that support. And right now there's kind of a 3js moment that's all meshes. And so it's become kind of the, the default in 3js ecosystem. As part of that to kind of exercise the library, I just build a whole bunch of cool demos. So if you see me on X, you see like all my demos and all the world building, but all of that is just to exercise this library that I work on. Because it's actually a very tough algorithmics problem to actually scale a library that much. And just so you know, this is ancient history now, but 30 years ago I paid for undergrad working on game engines in college in the late 90s. So I've got actually a back and it's very old background. I actually have a background in this and so a lot of it's fun, you know, but the whole goal is just for this rendering library to.
37:29
Are you one of the most active contributors their GitHub Spark JS.
38:47
Yeah, there's only two of us, so yes. So by the way, the primary developer is a guy named Andreas Sundquist, who's an absolute genius. He and I did our PhDs together, and so we set it for comps and quality. It was almost like hanging out with an old friend. And so. So he's the core. Core guy. I did mostly kind of, you know, venture fund.
38:51
It's amazing. Like, five years ago, you would not have done any of this. And it brought you back the active.
39:14
The actual energy was so high because you have to learn all the framework bullshit. And I fucking used to hate that. And so, like, now I don't have to deal with that. I can, like, focus on the algorithmics and I can focus on the scaling.
39:19
Yeah, yeah. And then I'll observe one irony, and then I'll ask a serious investor question, which is like, the irony is Fei. Fei actually doesn't believe the LM can lead us to spatial intelligence. And here you are using LLMs to, like, help, like, achieve spatial intelligence. I just see. I see some, like, disconnect in there.
39:29
Yeah, yeah. So I think. I think, you know, I think. I think what she would say is LLMs are great to help with coding.
39:45
Yes.
39:51
But, like, that's very different than a model that actually, like, provides.
39:51
They'll never have the spatial intelligence.
39:55
Listen, our brains clearly. Listen, our brains clearly have both. Our brains clearly have a language reasoning section, and they clearly have a spatial reasoning section. I mean, it's just these are two pretty independent problems.
39:57
Okay. I would say that the one data point I recently had against it is the DeepMind, IMO, gold. Typically, the typical answer is that this is where you start going down the neuro symbolic path. Right? Like one sort of very sort of abstract reasoning thing and one formal thing. And that's what DeepMind had in 2024 with Alpha Proof, alpha geometry. And now they just use DeepThink and just extended thinking tokens. And it's one model and it's. And it's in all of them.
40:07
Yeah, yeah, yeah.
40:36
And so that was my indication of, like, maybe you don't need a separate system.
40:37
Yeah. So. So let me step back. I mean, at the end of the day. At the end of the day, these things are like nodes in a graph with weights on them. Right. You know, like, it can be models. You need to sell it down. But let me just talk about the two different substrates. Let's let me put you in a dark room, like, totally black room, and then let me just describe how you exit it. Like, to your left, there's a table, like Duck below this thing. Right. I mean, like, the chances that you're gonna, like, not run into something are very low. Now let me, like, turn on the light and you actually see and you can do distance and you know, how far something away is and like, where it is or whatever. Then you can do it. Right. Like, language is not the right primitives to describe the universe because it's not exact enough. So that's all Fei. Fei is talking about when it comes to, like, spatial reasoning is like, you actually know that this is three feet far. Like, that far away. It is curved. You have to understand, you know, like the actual movement through space. Yeah. So I do. I listen. I do think at the end of these models are definitely converging as far as models. But there's. There's. There's different representations of problems you're solving. One is language, which, you know, that would be like describing to somebody, like, what to do, and the other one is actually just showing them and the spatial reasoning is just showing them.
40:42
Yeah, yeah, yeah, Right, Got it, Got it. The. The investor question was on. On World Labs is, well, like, how do I value something like this?
41:55
What.
42:04
What. What work does. Do you do? I'm just like, Fei. Fei is awesome. Justin's awesome. And you know, the other two co founder. Co founders. But like, the. The. The tech. Everyone's building cool tech. But, like, what's the value of the tech? And this is the fundamental question.
42:04
Let me. Let me just. For, like, let me just maybe give you a rough sketch on the diffusion models. I actually love to hear Sarah because I'm a venture person, you know, so like, venture is always like kind of wild west.
42:16
You paid a dream. And she has to like, actually be marked to reality.
42:25
Exactly. So I'm going to say the venture can be like, okay, you little kid. Yeah. So, like, so. So these diffusion models literally create something for. For almost nothing. And something that the world has found to be very valuable in the past in real markets. Right. Like, like a 2D image. I mean, that's been an entire market. People value them. It takes a human being a long time to create it. Right. I mean, to create a. You know, to turn me into whatever. Like an image would cost 100 bucks in an hour. The inference cost is a hundredth of a penny. Right? So we've seen this with speech in very successful companies. We've seen this with 2D image. We've seen this with movies. Right? Now think about 3D scene. I mean, when's Grand Theft Auto coming out? Six. It's been 10 years. I mean, how, like how much would it cost to reproduce this room in 3D? If you hire somebody on fiber, like in any sort of quality, probably 4,000 to $10,000. And then if you had a professional who bought you $30,000. So if you could generate the exact same thing from a 2D image. And we know that these are used, they're using Unreal and they're using Blender, they're using movies and they're using video games and they're using all. So if you could do that for, you know, less than a dollar, that's four or five orders of magnitude cheaper. So you're bringing the marginal cost of something that's useful down by three orders of magnitude, which historically have created very large companies. So that would be like the venture kind of strategic dreaming map.
42:29
Yeah. And for listeners, you can do this yourself on your own phone with like, like the marble. But also there's many Nerf apps where you just go on your iPhone and do this.
43:50
Yeah, yeah, yeah. And in the case of marble though, what you do is you literally give it. So most nerf apps you like kind of run around and take a whole bunch of pictures and then you kind of reconstruct it. Yeah, Things like marble, just that the whole generative 3D space will just take a 2D image and it'll reconstruct all
43:59
the, like, meaning it has to fill in.
44:16
Yeah, like the back of the table, under the table, like the images it doesn't see. The generator stuff is very different than reconstruction. That like it feels, fills in the things that you can't see. Yeah. Okay, so, all right, so now.
44:18
No, no, I mean, I love that
44:27
now, the adult perspective.
44:28
Well, no, I was going to say these are very much a tag team. So we started this pod with that premise and I think this is a perfect question to even build on that further because it truly is. I mean, we're tag teaming all of these together, but I think every investment fundamentally starts with the same, maybe the same two premises. One is at this point in time, we actually believe that there are n of one founders for their particular craft and they have to be demonstrated in their prior careers. Right. So we're not investing in every, you know, now the term is Neolab, but every foundation model, any company, any founder is trying to build a foundation model. We're not, contrary to popular opinion, we're not invested in all of them. Right. We have a very specific thesis.
44:30
I don't think people say that about you. No, they don't. They don't.
45:09
They say that we're big, we're at everything. But, you know, if you think about Ilya, right, he's at ssi. He's sort of been behind almost every foundational breakthrough for the last 15 years. If you think about, you know, the Thinking Machines team, right, Amira and John, right? John is the godfather of reinforcement learning. And so I go through this because, you know, if you think about, for each of the bets that we've made, it goes back to one, to a very specific thesis about that person, the team they've assembled, and what they've done in a prior life. And I think obviously we talked about talent wars. We do think at this particular moment in time, there are particular people that can move needles. Clearly other companies believe that too, otherwise they wouldn't be willing to pay such crazy prices for single individuals. So that's one, and then two. We don't think it's a zero sum game. If that were true, OpenAI or actually just DeepMind would be number one in everything, right? There's clear value to specialization. It's like 11 labs. There have been so many audio models that have hit the market, they're still freaking number one, right? And so if you think about. And they've created a ton of value for their customers, for their investors, you know, for their team. And so if you think about those two put together, right, that's sort of the foundation of our thesis. When we back these foundation model companies, of course, the valuations, they sound astronomical. When you think about current revenue, the numbers one, I would say that's the market out there, because they are raising larger dollars. They have compute needs that's 80% of a round that they typically raise or typically of a round that they raise. But I think the thing that gets us excited about backing them is that the revenue growth has typically followed the capability breakthrough. It sort of ties back to that question of the cyclical nature, like, are you just funding it and you raise more funding? When there's a real capability breakthrough, the demand is there. And so the revenue growth is much faster than we've ever seen once it's turned on. There's a company, I can't share the name, but their product went GA in a few weeks. Tens of millions of revenue, right? We have SaaS.
45:12
I've seen this myself.
47:23
Yes, absolutely. We have SaaS. Companies that have been in business for seven years, and they get to the same level seven years later and the growth is eking to whatever it is. And by the way, great companies not at all diminishing what they've accomplished. But the fact is to get to that revenue growth that quickly. It's not just the two companies that people talk about. It's really a lot of these. Every domain has a specialist and we think if you can win that, you become very large very quickly. And that's actually played out in the numbers.
47:24
Yeah, our viewers are going to. So first of all thank you for that overall take. I think like it's important to hear you guys perspective because the rest of us are just kind of looking at headlines and not knowing how to make sense of any of this. We can't mention like our listeners will roast us if we, if we mention thinky and not discuss what happened. I mean obviously founder split happens but like I guess is the thesis unchanged is like you know, like what's going on in thinky.
47:57
Yeah, we're more excited than ever about them. They have some things that we're not going to do. Breaking news on a pod. You know, obviously they should share themselves but they've, you know, I think when you bring a team of that caliber together, there's special things that happen And I think 2026 is going to be a big year for them. Obviously, you know, some of the themes that we talked about before, even with just the media news storm, like the whole something happens and then it's everywhere instantly. You know, I think that's a tough situation for any company to be in. But to come out of that stronger than ever. I think that you know, we're more bullish about thinke than you know, even before and obviously and the story is tinker.
48:26
It's custom models. RL. Yeah. Is that, is that what, is that what we're aiming for?
49:16
Yeah. And a bunch of stuff we, we can't talk about here. Yeah, yeah, absolutely. But no, that team is cooking and you know, I think they'll, they'll be just fine from, they'll recover from the events in January.
49:22
Yeah.
49:33
I will say this is the furthest. So we have a very privileged position on the boards of these companies. And like I will say I've never seen the perception of the truth be further from the truth industry wide ever. Like I guarantee you for any of these gossipy things, I guarantee you it's way off, way, way like the general sentiment and like, and what happens is like we've got this crazy game of telephone right now where there's always like seeds of truth. But it gets so warped by the time like we hear all the time rumors about Stuff that we're directly involved in. Like, we're literally on the board. You know, like, we're the one that did the thing, and by the time it gets to us, it's gotten so warp. So twisted. I think this is, like, everybody's excited. There's a lot of focus. The shot and fried is so high that people just kind of will into being things that didn't exist. So I'm not, you know. You know, I don't want to comment specifically on the thinking machines, but, like,
49:34
it's an important message to the general audience.
50:31
I will tell you if you hear something on X, like, the chances that it's, you know, it is accurately representing, but it's saying to is very, very, very low.
50:33
Yeah.
50:42
I have never lost so much faith in the anon counts on Twitter that just seem very confident in what they're saying. Could it be further from the truth? I had a couple days stretch where I was like, oh, my God, Twitter is mind poison, and I love it.
50:43
But we talk to each other all the time because we actually know because we're there. Like, we're there seeing these things. And, like, you know, Sarah will, like, text me, like, whatever. Like, it's like, ridiculous. So for us, it's like, it's like this ridiculous. But the problem is we realize that things. That things start taking on a life of their own, and then people assume that they're real and everything. And so I think it's very tough for founders because, you know, it's tough enough fighting the real battle. You know, they're fighting phantoms too. And so, you know, you know, more and more, we're just like, I got this from the Cursor guys, which I really appreciate. Michael Trull, he's like, listen, heads down, focus on the business.
50:56
Andy absolutely crushed it. Yeah.
51:34
Yeah. And I think that's right. All founders should do it right now because the noise is so high.
51:36
Yeah, no, that team's been back to business for weeks. The thinky team, so. Yeah.
51:40
Yeah. Well, thank you for indulging in that. It is just the hot topic of the moment. Got to address the elephant in the room.
51:44
Cursor.
51:50
Right. Obviously, you guys are big investors. 2025, I would say, is Cursor's year. I mean, maybe decade, but just like, I think, you know, just going back to the discussion about how AGI would just kind of consume everything. Cursor is like the one, like, the kind of the shining example of, like, here's how you build application layer that's a wrapper but an extremely damn good one. And I guess just the general analysis, I guess, of cursor's development and what it means for everyone. Is there a cursor in every industry to be built?
51:51
Yeah. So the interesting thing about cursors, they actually, for a small fraction of the cost, 100 the cost or less, developed an almost soda model, which for a period of time was the most popular coding model in the world, which is really crazy to think about. I think they're just kind of doing it in reverse. Right. So there's two approaches. You start with a foundation model and then you verticalize up, or you start with the app and all of the product data and you go down and they're the ones that are doing that. I think any company that's doing an app has to ask the margin question, which is like, how do I extract margin on the tokens that are going through? Like everybody has to be on the token path and everybody has to ask that question. And I've just thought they'd been, been incredibly thoughtful about it. And one reason is if you ask, you know, Michael, what type of company they are, a developer company for professional developers. That's what they are. They're a dev tools company. They're just focused on coding. And that's a huge, I mean, even if you didn't do AI, that's a ma. You know, they, they, they, they acquired graphite. I mean, like, you know, listen, we were investors in GitHub, like we know how big this market is, so that's a massive market even without becoming a model company. But they've also been quite successful in doing their own models. And so I think it just shows you that if you are focused, you have a large use case, there's a huge opportunity not only to get the application but to start building your own models. Are these going to be the only models people use? Of course not. But they are in a great position to serve great models and they've demonstrated that.
52:24
Yeah, my sort of thesis, which we're not going to have to go into here, is actually, I think what I've been calling Agent labs, which are people who build on top of all the other models will probably have a better time with the margins because they price against the end user hours spent or like human labor, whereas models get commodity price per token. And so margin wise, we know inference economics for model labs, but Agent Labs, the difference is the delta between token intelligence, which keeps going down, and human costs, which keep going up. And so the margin should be higher.
53:51
They should be the caveat to that is if the models go first party, right?
54:32
Yeah.
54:38
What they can do is they can
54:39
which is the composer dream.
54:40
Yeah. They can subsidize themselves. The models. They can subsidize themselves. They can subsidize themselves and then they can charge the third party more. And it's a very delicate dance because you're kind of competing with your own customers. And so we've seen this historically we saw this with the cloud with EC2. So this is not unusual still. We saw this with the operating system. It's not unusual but it's playing out very, very quickly.
54:41
Yeah. Thank you for joining us. That's all the time we have today.
55:04
Such a pleasure.
55:06
You're welcome back anytime.
55:07
And thank you for being so open and also like just leading the industry in so many areas. It's really inspiring to see.
55:09
Thank you so much.
55:16
Thank you for having us.
55:16
Great.
55:17
Thank you.
55:17