AI transcript
0:00:10 In the Industrial Revolution, having oil was important, and now having data centers is important.
0:00:14 These models aren’t just compute infrastructure, they’re cultural infrastructure.
0:00:19 It’s not just sort of defining the culture, but sort of controlling the information space.
0:00:28 So if a model is trained by a country that’s adversarial to you, that’s actually very hard to eval or benchmark when the models are released.
0:00:29 This is a massive vulnerability.
0:00:33 Is that the new age of LLM diplomacy that we’re entering here?
0:00:37 Do we build? Do we partner? What do we do?
0:00:44 Today, we’re diving into a conversation that’s just as much about geopolitics as it is about technology.
0:00:50 This week, the Kingdom of Saudi Arabia announced plans to build its own AI hyperscaler called Humane.
0:00:57 But they’re not calling it a cloud provider, they’re calling it an AI factory, and that language alone suggests a shift.
0:01:03 For decades, cloud infrastructure has been concentrated in two places, the U.S. and China.
0:01:06 But with the rise of AI, that model is breaking down.
0:01:10 Nations no longer want to outsource their most strategic compute.
0:01:15 They’re building sovereign AI infrastructure, factories for cultural and computational independence.
0:01:27 To unpack what this means for the global AI stack, national sovereignty, and the new digital power dynamics, I’m joined by Anjane Miha and Guido Appenzeller.
0:01:37 We talk about what it takes to become an AI hypercenter, why governments are spending billions to control inference pipelines, and whether we’re entering a new Marshall Plan moment for AI.
0:01:39 Let’s get into it.
0:01:55 As a reminder, the content here is for informational purposes only, should not be taken as legal, business, tax, or investment advice, or be used to evaluate any investment or security, and is not directed at any investors or potential investors in any A16Z fund.
0:02:01 Please note that A16Z and its affiliates may also maintain investments in the companies discussed in this podcast.
0:02:08 For more details, including a link to our investments, please see A16Z.com forward slash disclosures.
0:02:18 Anj Guido, we want to talk about sovereign AI, AI and geopolitics, and let’s start with the news.
0:02:22 Our partner Ben is in the Middle East right now to participate in his own way.
0:02:24 What happened, and why is it its own point?
0:02:31 What happened is the kingdom announced that they’re going to build their own local hyperscaler or AI platform called Humane.
0:02:47 And I think why it’s notable is that as opposed to the status quo of the cloud era, they’re viewing the AI era as one where they’d like the vast majority of AI workloads to run locally.
0:02:56 If you think about the last 20 years, the way the cloud evolved was that the vast majority of cloud infrastructure basically existed in two places, right?
0:02:57 China and the U.S.
0:03:01 And the U.S. ended up being the home for the vast majority of cloud providers to the rest of the world.
0:03:05 That doesn’t seem to be the way AI is playing out.
0:03:13 Because we have a number of frontier nations who are basically raising their hands and saying, we’d like infrastructure independence.
0:03:28 The idea being that we’d like our own infrastructure that runs our own models, that can decide where we have the autonomy to build the future of AI independent of any other nation, which is quite a big shift.
0:03:43 And I think the headline numbers are somewhere in the range of 100 to 250 billion worth of cluster build-out that they’ve announced, of which about 500 megawatt seems to be the atomic unit of these clusters that they’re building.
0:03:52 So a number of countries, with the kingdom being the one that’s most recent, have been announcing what we could think of as sovereign AI clusters.
0:03:55 And that’s a pretty dramatic shift from the pre-AI era.
0:03:57 I don’t know if you’d agree with that.
0:03:58 I think it’s spot on.
0:04:04 I think many sort of geopolitical regions are reflecting back what happened in previous big tech cycles.
0:04:17 And wherever the technology is built and whoever controls the underlying assets has a tremendous amount of power of shaping regulation, shaping how this technology is being used, and also puts themselves in a position then for the next wave that comes out of that.
0:04:23 And, you know, it was the Industrial Revolution, having oil was important, and now having data centers is important.
0:04:26 And so I think it’s a very exciting development.
0:04:36 Yeah, in fact, you can often tell why something is important to somebody by the semantics that folks used to communicate a new infrastructure project.
0:04:44 In this case, if you look at how the cluster build-ups are being referenced, they’re being called AI factories.
0:04:47 They’re not being called AI data centers.
0:04:49 They’re being called AI factories.
0:04:50 And I think there’s two ways to respond to that.
0:04:53 One train of thought would be, hey, that’s just branding.
0:04:56 That’s just, you know, the marketing people doing their thing.
0:05:01 And under the hood, this is really just data centers with slightly different components.
0:05:11 But everybody in the world, in every industry is looking for a way to be relevant in the age of AI, and this is the compute infrastructure world’s way of doing that.
0:05:15 An opposing view would be, actually, no, this is not just marketing.
0:05:22 If you look under the hood and if you x-ray the data center itself, very little of it is the same as was the case 20 years ago.
0:05:26 The big difference in active components being GPUs, right?
0:05:30 About 20 years ago, what would you say the average number of GPUs were in a, what percentage of, right?
0:05:31 Pretty much, yeah.
0:05:32 Right.
0:05:33 It’s a very recent phenomenon.
0:05:43 And today, I think if you look at the average 500 megawatt data center, and you looked at what percentage of the CAPEX that was required to build that data center, or operated, rather, went to GPUs.
0:05:44 Massive, yeah.
0:05:46 That’s a huge shift.
0:05:49 I think we’re also seeing a specialization, right?
0:05:56 The kind of data center you built for a classic CPU-centric workload and what you built for a high-density AI data center look very different.
0:05:58 You need a good cooling to the rack.
0:05:59 You need very different energy supply.
0:06:01 You want it close to a power plant.
0:06:03 You want to lock in that energy supply early on.
0:06:12 And then we’re also seeing a change, I think, in the consumer behavior, where classically you want a very full stack that has lots of services that helps the enterprise build all these things.
0:06:18 We’re seeing more and more enterprises that are actually comfortable with just building on top of a simple Kubernetes abstraction or something.
0:06:19 Right, right.
0:06:25 And basically, you know, cherry pick a couple of snowflake or database type services on the side that help them complement that.
0:06:26 So I think there’s a new world.
0:06:35 And so that’s certainly true that you could kind of look at the technical components in an AI factory are completely different from a traditional data center.
0:06:37 And then there’s, what does it do?
0:06:56 And historically, a lot of the workloads that traditional data centers were doing were running one cloud-hosted workloads for enterprises or developers, whoever it might be, where most of that, the data sets and the workloads were actually not particularly opinionated.
0:07:03 And when I say opinionated, I mean, they’re not necessarily subject to a ton of cultural oversight.
0:07:03 Yeah.
0:07:27 But for the better part of the 2000s, until the rise of GDPR, CCP, and so on, we lived in an era of centralization, where having most of your cloud infrastructure in Northern Virginia was preferable for most of the world’s developers’ enterprises because it gave them economies of scale.
0:07:34 That started to change, of course, with GDPR, CCP, the rise of data privacy laws, because then you had region-by-region compliance.
0:07:46 And that made the rise of something like Cloudflare critical, where Cloudflare has this idea of distributed infrastructure where you can tie the workload policies to wherever the user is.
0:07:50 But by and large, that was critical for, especially for the rise of social media workloads.
0:07:55 But the vast majority of enterprise workloads didn’t need decentralized serving.
0:08:00 What’s different about AI seems to be that these models aren’t just compute infrastructure.
0:08:02 They’re cultural infrastructure.
0:08:07 They’re trained on data that has a ton of embedded values and cultural norms in them.
0:08:09 And then more importantly, that’s the training step.
0:08:21 And then when you have inference, which is when the models are running, you have all these post-training steps you add that steer the models to say something or not, to refuse the user or not.
0:08:32 And that last mile is where things over the last, I would say, year have made it more and more clear that countries want the ability to control what the factories produce or not within their jurisdiction.
0:08:35 Whereas that urgency didn’t quite exist as much.
0:08:39 Because of the cultural factors or because of certain independence or resilience?
0:08:41 It’s a good question.
0:08:45 My sense is there’s two things going on, but you should chime in if you think I’m being incomplete.
0:08:55 One, I think, would be the rise of the capabilities in these models being now well beyond what we’d consider sort of early toy stage of a technology.
0:09:03 I think our partner Chris Dixon has a great line, which is that many of the world’s most important technologies start out looking like toys.
0:09:10 And four years ago, when the scaling loss paper was published and GPT-3 was published, most people looked at it and said, okay, that’s cool.
0:09:12 Sure, it can produce the next word.
0:09:13 It’s a nice party trick.
0:09:14 It’s a nice party trick, right?
0:09:15 It’s a stochastic parrot.
0:09:23 And now you have foundation models literally running in defense, in healthcare, in financial services industries.
0:09:29 ChatGPD has about 500 million monthly active users making real decisions in their daily lives.
0:09:41 I think using these AI models, there was a paper that was recently published by Google that showed the efficacy of Gemini, their foundation model, at solving medical questions.
0:09:53 And one of the most interesting things you can see when you look at the usage, the types of prompts that people are using models for, relative to two years ago, three years ago, it was a lot of helping write my essay.
0:10:10 It’s turned into coding and helped me solve a whole host of medical problems or personal life-related questions and so on, where it’s clear now that these capabilities can be used, one, to drive mission-critical industries like defense, healthcare, and so on.
0:10:14 And also then influence a number of your citizens’ lives.
0:10:32 And so I think that makes a lot of governments go, wait a minute, if we are dependent on some other country for the underlying technology that our military, our defense, our healthcare, our financial services, and our daily citizens’ lives are driven on, that seems like a critical point of failure in our sovereignty.
0:10:37 So that’s one, it’s just that models have gotten good, and they seem to be good at a bunch of important things.
0:10:52 The second is, I think, an increasing belief that if you don’t have control over the model’s production pipeline, then you’re doomed or destined to use models that reflect other people’s cultural values.
0:11:03 We had a pretty in-depth debate about this with DeepSeq, right, where the question was, is DeepSeq fundamentally more biased or not than open-source models trained in the US?
0:11:18 And I think there’s early evidence to say that you can actually see, certainly in the post-trained DeepSeq, that there’s just a number of topics and types of tasks that it’s been told to avoid and answer differently from a model like LAMA.
0:11:22 So that’s the cultural piece. I think there’s a critical sort of national capability piece, and then there’s the cultural piece.
0:11:33 And I think both are combining to create this sort of huge rise in the demand for, you could call it sovereign AI, which is the idea that you want control over what the models can, can’t do.
0:11:35 Or you could call it infrastructure independence. I think everyone’s got a different word for it.
0:11:38 You could call it our local AI factory ecosystem.
0:11:42 But I think that all these terms are trying to get at the same thing, which is, we’ve got to control our own stack.
0:11:43 Yeah.
0:11:49 I think I would make it even stronger. I think it’s not just self-defining the culture, but self-controlling the information space to some degree.
0:11:54 I mean, today, we’re starting to see how, in many cases, models are replacing search.
0:11:59 They no longer go to Google, they don’t go to ChatGPT, and that comes back with an answer.
0:12:08 If there’s a historical fact, and say, in the Chinese model, it does not show up, in the U.S. model, it does show up, that is the reality that people grow up with.
0:12:14 And if you write an essay in school, in the future, many of us’ essays will be graded by an LLM.
0:12:15 Right.
0:12:25 So, in fact, in school, something that may be truthful, right, may be graded as wrong, because whoever controlled the model decided that should not be part of the trading policy.
0:12:30 So, it has a very profound effect on public opinion and, you know, on values.
0:12:30 Right.
0:12:34 The downstream use is an interesting one, because it’s very hard to measure for.
0:12:44 And certainly, relative to two years ago, when the vast majority of products and applications, like the ones Guido’s talking about, were basically pretty simple models, right?
0:12:47 Well, at the time, they were considered pretty complex, but the frontier changes so fast.
0:12:54 Today, we’d look back at a model like GPT-4 that was largely just a next-word prediction model and say, that’s pretty rudimentary, right?
0:12:59 Because if you x-rayed an app like ChatGPT, sure, on the surface, it looks like nothing much has changed, right?
0:13:00 It’s still a chat box.
0:13:03 You type in what you need, and it outspits an answer relative to two years ago.
0:13:10 But under the hood, there’s been this insane evolution where there’s four or five different systems interacting with each other, right?
0:13:19 You’ve got a reasoning model that can produce a chain of thought to think through what it should do next, including then doing what we call tool usage, right?
0:13:20 Calling out to third-party tools.
0:13:35 And then you have the idea that these models can start to self-learn, to go through a loop of taking your input and reasoning about what it needs to do, calling an action, and then evaluating its output, and then updating that loop.
0:13:39 That starts to look more, people use the word agent, right, to call it that.
0:13:43 But the idea is that it’s going from being a pretty simple model to being a system.
0:13:48 And it’s very hard to measure where the adversarial cracks are in the system.
0:14:05 So if a model is trained by a country that’s adversarial to you, to, when you’re writing code, open up a port, or what we’d call a call-home attack, right, where it’s transmitting telemetry, that’s actually very hard to eval or benchmark when the models are released, right?
0:14:10 Because these models are often tested in very academic or static settings.
0:14:14 And so when DeepSea came out, it was just such a great model.
0:14:18 It was a phenomenal piece of engineering that suddenly everybody was using it everywhere.
0:14:33 And a number of CIOs and CTOs got pretty nervous because they were like, wait a minute, if the model is being used in this agentic fashion, and I don’t have visibility on what it’s doing adversarily until it’s too late, this is a massive vulnerability.
0:14:44 And so I think the adversarial threat, as the systems go from being models to agents, is causing a lot of governments to go, well, we’d rather have the whole thing running locally in a way that we can lock down.
0:14:50 Again, it comes back to a sort of independence and a supply chain question.
0:14:53 And is your expectation that this is going to play out?
0:14:55 And to what extent is it going to play out?
0:14:58 On the cloud, as we’ve mentioned, there’s a Chinese internet and a sort of Western rest-of-the-world internet.
0:15:01 How widespread is this sovereign AI thing going to go?
0:15:02 Yeah.
0:15:09 I’m going to borrow an analogy that Guido used, which is that in the Industrial Revolution, you could look at where resources flowed, right?
0:15:17 I think you should talk about how viewing it from the lens of oil reserves, you know, can kind of dictate which countries can and can’t participate in the Industrial Revolution.
0:15:18 Go ahead.
0:15:23 So if you look at the Industrial Revolution, so oil was the foundation of a lot of the technologies, right?
0:15:25 You needed oil reserves in order to participate.
0:15:28 And I think it’ll be a little bit the same thing, right?
0:15:38 If you want to build industry in a particular country, if you want to be able to export things, if you want to be able to drive development, and if you want to harness the power that comes with that, you need the corresponding reserves.
0:15:49 And I mean, I think AI data centers are a little bit like these oil reserves, with the big difference being you can actually construct them themselves if you have the necessary investment dollars and the willpower to do it.
0:15:55 But I think they will be the foundations for building all the layers on top that ultimately, I think, determine who wins this race.
0:16:05 And in my mind, the countries that invest in building up the AI factories, or in this sense, the oil reserves, to borrow Guido’s analogy, I think of them as one body of countries.
0:16:07 Let’s call them hyper centers, right?
0:16:13 The idea is they’re centers that have enough compute capabilities to compete at the frontier and run their own sovereign models, sovereign infrastructure.
0:16:18 And then there’s everybody else who just doesn’t have the resources to do that.
0:16:25 And if you look at after the Industrial Revolution, you could argue the next major technology revolution was the advent of modern finance.
0:16:32 The Bretton Woods and IMF regime, where modern finance said, we’re going to all use this one measure of value called the dollar.
0:16:42 And you were either in a country that produced the dollars, like America, or you were in a country that produced a lot of goods that acquired dollars, like China.
0:16:47 And then if you weren’t in one of those two, you really had to figure out whether you aligned with one of these trade blocks or not.
0:16:58 And what happened is, you had countries like Singapore, Luxembourg, Ireland, and Switzerland, who realized, well, we just don’t have the resources to build out our own reserve system.
0:17:02 And there’s not that much by way of local production that we can do to acquire dollars, we can’t really trade.
0:17:06 So we’ve got to find a way to insert ourselves in the flow, right?
0:17:20 And so Singapore, of course, famously became the entry point for dollar flows into Asia, because they invested a ton in rule of law and a great tax regime and sort of stable government and low corruption and all of that.
0:17:24 Switzerland did something similar for European investments and European capital flows.
0:17:32 So I think what we’re watching right now is that build out where there’s US and China, which clearly have enough compute to be hypercenters.
0:17:36 And then you’ve got folks like the Kingdom of Saudi Arabia saying, we want to be a hypercenter.
0:17:41 And if that means we’ve got to trade our oil to acquire large numbers of NVIDIA chips, we’ll do that right now.
0:17:50 And I think in that bucket, there’s probably the Kingdom of Saudi Arabia, there’s Qatar, there’s Kuwait, there’s Japan, Europe, clearly.
0:17:54 And then I think the question is, everybody else, what do they do?
0:17:58 And it’s not clear to me what you have to do to become the Singapore of AI.
0:18:06 And maybe the Singapore of AI ends up being Singapore, because actually now they have an enormous sovereign wealth fund as a result of participating in modern capital flows.
0:18:11 But I think a bunch of other countries are sitting around wondering, is this the time where we actually buy?
0:18:14 Do we build? Do we partner? What do we do?
0:18:19 Yeah. And talk more about the implications behind what this means.
0:18:21 Is this something that the US should be excited about?
0:18:23 What does this mean as we think about foreign policies?
0:18:26 Are there now winners across the board and all these local requirements?
0:18:28 Why don’t you talk about some of the big implications here?
0:18:29 Do you want to take a stab?
0:18:32 I think every big structural revolution is both a threat and opportunity.
0:18:36 I think the United States and AI right now has the world leadership.
0:18:36 Yeah.
0:18:38 That’s an opportunity.
0:18:41 Hanging on to it won’t be easy, as it is in every tech revolution.
0:18:44 Don’t we want people to be dependent on us in the same way that they were in the cloud revolution?
0:18:47 Or do we benefit somehow from it being more decentralized?
0:18:49 The world is not one place.
0:18:52 So I think complete centralization won’t happen.
0:18:53 I think the leader is good.
0:18:57 Having strong allies that also have their technology is also very valuable.
0:19:00 So it’s probably a balance of those that we’re looking for.
0:19:00 Yeah.
0:19:05 To put a finer point on your last note there is that you could think about a balance.
0:19:07 Like we’re clearly in an unstable equilibrium right now.
0:19:08 Yeah.
0:19:14 And so Guido’s right that the arc of humanity and history is such that things will shake
0:19:15 out until there’s a stable equilibrium.
0:19:18 And so what is the stable equilibrium?
0:19:21 And I think one way to reason about it is you could look at historical analogies.
0:19:26 So post-World War II, when Europe was completely decimated, there was a group of really enterprising
0:19:31 folks in the private sector and the public sector who got together and said, hey, we can either
0:19:37 choose to turn our backs on Europe and adopt a posture of isolationism where we mostly focus
0:19:40 on a post-war American-only agenda.
0:19:48 Or we can try to adopt a policy where we know that if we don’t help out our allies, somebody
0:19:48 else will.
0:19:53 And so they came up with this idea called the Marshall Plan, right?
0:19:57 Where a number of leading enterprises in the U.S. got together like GE and General Motors
0:20:04 and literally subsidized the massive reconstruction of Europe that helped a lot of European economies
0:20:06 quickly get back on their feet.
0:20:10 And at the time, there was a ton of criticism of the Marshall Plan because it was viewed almost
0:20:13 as a net export of capital and resources.
0:20:19 But what it did end up doing is then solidified this unbelievable trade corridor between the
0:20:24 U.S. and Europe for the next 50 years, which really kept China out of that equation for the
0:20:25 70 years, yeah.
0:20:26 70 years, really.
0:20:28 And so I think we have a choice.
0:20:31 Either approach it the way we would the Marshall Plan for AI, right?
0:20:36 And say, well, a stable equilibrium is certainly not one where we just turn our back on a bunch
0:20:41 of allies because China definitely has enough of the compute resources to try to export great
0:20:42 models like DeepSeek to the rest of the world.
0:20:45 So what do we want our allies on, DeepSeek or Lama?
0:20:48 That’s what it comes down to at the model level of the stack, right?
0:20:55 And I think that the realities that a number of countries are not waiting around to find out.
0:21:00 That’s why you have efforts like Mistral in the EU, right?
0:21:07 Where they are being approached by a ton of not just European nations, but a ton of other
0:21:13 allies of Europe to say, hey, can you help us figure out how to build out our own sovereign
0:21:13 AI?
0:21:19 And so I think we’re about to see basically the single biggest build out of AI infrastructure
0:21:24 ever because most of the purchase orders and the capital is being provided by governments
0:21:27 because they realize this is a critical national need.
0:21:32 And their stable equilibrium is certainly not to depend on somebody else or depend on an uncertain
0:21:32 ally.
0:21:37 And so the ones that certainly have the ability to fund their own sovereign infrastructure are
0:21:37 rushing to do it right now.
0:21:42 And what does that mean for the sort of nationalization debate or how you see that playing out?
0:21:47 Leopold, Ashton Brenner, formerly of OpenAI, in his famous sort of report, talked about how,
0:21:52 hey, if this thing becomes as critical to national security as we think it will be, at some point,
0:21:54 governments aren’t just going to let private companies run it.
0:21:57 They’re going to want to have a much more integrated approach with it.
0:21:59 Where do you stand with the likelihood of that?
0:22:03 And what does that mean just in terms of the feasibility of regulation in a world where it’s
0:22:04 much more decentralized?
0:22:06 And we already have this with DeepSeek.
0:22:09 I mean, that already changed the game in terms of we’re in an arms race and you can’t control
0:22:09 everything.
0:22:11 We’re in the open source conversation as well.
0:22:12 We’re backing some of these players.
0:22:15 What are your thoughts on where the solid net’s out?
0:22:17 I think I have probably a strong opinion on that.
0:22:18 I mean, I grew up in Germany, right?
0:22:20 So benefiting from the Marshall Plan.
0:22:25 And also sort of seeing how that pulled away Western Germany towards the United States
0:22:29 and eventually Eastern Germany also towards the United States when everybody realized the
0:22:30 impact of that.
0:22:35 One lesson I took away from that is that I think any kind of centralized planned approach
0:22:35 does not work.
0:22:37 Eastern Germany versus Western Germany was a nice A-B test.
0:22:40 You know, central planning versus a free market economy, what works better, right?
0:22:42 And I think the results speak for themselves.
0:22:49 So I think basically having the government drive all of AI strategy, you’re in a Manhattan-style
0:22:52 project, Apollo project, pick your favorite successful project there.
0:22:53 I can’t see that working.
0:22:57 You probably need a highly dynamic ecosystem of a large number of companies competing.
0:23:02 There’s some areas, I think, where the government can have a hugely positive effect, right?
0:23:04 On the research side, we’ve seen it again and again.
0:23:09 Funding, fundamental research, which is not quite applied enough yet for enterprises to pick
0:23:14 up, right, is very valuable, I think it can help in terms of setting good regulation.
0:23:17 Bad regulation can easily torpedo AI, as we’ve seen.
0:23:21 And so I think there’s a strong will for government to lead this and to direct this.
0:23:26 There’s no master plan at the end of the day that you can make that basically has all the
0:23:27 details that has to come from the market.
0:23:30 I don’t agree with the Ashenbrenner point of view.
0:23:36 I agree strongly with Guido that the history of centralized planning at the frontier of technology
0:23:43 is not great, barring a few situations that were essentially brief sprints of war, right?
0:23:47 And arguably even the Manhattan Project, which is the analogy I think he uses in his piece,
0:23:49 we now know that there were leaks.
0:23:53 It was literally a cordoned off facility in Los Alamos or whatever, and they were still spies.
0:23:54 Right.
0:23:58 And so if you’re approaching this from the lens of the models are what are the equivalent
0:24:03 of nukes, and we’ve got to regulate the development of these by locking up our smartest researchers
0:24:08 in some facility in Los Alamos, and that’s what’s going to prevent the best models from getting
0:24:09 exported.
0:24:12 I think that’s great fiction, a very interesting novel.
0:24:12 Yeah.
0:24:17 But for anyone who has ever had the both pleasure and displeasure of working in any
0:24:19 large government system, it’s a pipe dream.
0:24:24 The good and the bad news is that, in a sense, it doesn’t really matter where the model weights
0:24:24 are.
0:24:27 It matters where the infrastructure that runs the models are.
0:24:29 In a sense, inference is almost more important.
0:24:33 And I think a year ago, we were in a pretty rough spot, I would say, with the arc of regulation
0:24:38 where there were a number of proposals in the United States to try to regulate the research
0:24:40 and development of models versus the misuse of the models.
0:24:44 I think that, luckily, we have moved on from that.
0:24:50 Where we are now is, unfortunately, still a state-level batch work on whack-a-mole of regulation.
0:24:51 It’s not consistent.
0:24:57 Hopefully, I think we’ve got a number of positive signals from early administration executive orders
0:25:00 that they’ve put out that hopefully means there will be unified regulation around AI.
0:25:10 But I don’t think that the answer is going to be one single lab that has one God model that
0:25:13 then the country protects as if it’s a nuclear bomb.
0:25:18 I think we are now in a state where, partially because of the build-out of AI factories that
0:25:23 we’ve discussed, a number of countries have the capabilities to train frontier models.
0:25:27 And a number of them are quite willing to export them openly.
0:25:29 China being a leading one.
0:25:34 DeepSeq has forced people to update their priors, where just a year before DeepSeq came out,
0:25:39 there were a number of tech leaders in Washington testifying that China was five to six years
0:25:41 behind the U.S. with confidence, on the record.
0:25:45 And then DeepSeq comes out 26 days after OpenAI puts out the frontier.
0:25:47 I mean, it just shattered all of those arguments.
0:25:50 So the calculus has changed.
0:25:55 And I think it means that the only way to win is build the best technology and out-export anybody else.
0:25:58 Then if the question is, whose math is the world using?
0:25:59 We’d love for it to be American math.
0:25:59 Right.
0:26:08 My view is that we are much better off embracing the ability for other countries to serve their own models.
0:26:13 And ideally, the best product wins, which is the best models just come from the U.S. and our allies.
0:26:14 Yeah.
0:26:17 Is that the new age of LLM diplomacy that we’re entering here?
0:26:21 Actually, Ben had a great talking point to this at FII Riyadh last year.
0:26:26 And he said something to the effect of, because these models, like we discussed earlier, are cultural infrastructure,
0:26:31 you don’t want to be colonized in the digital era, in cyberspace.
0:26:34 And I think that’s pretty spot on.
0:26:38 Instead of colonization, what we have is now, I think, foundation model diplomacy.
0:26:40 That’s a good way to put it.
0:26:40 I think that’s right.
0:26:47 It suits the U.S.’s relative skill sets, which is innovation and working with our allies relative to China, which has been a bit more closed off as a country.
0:26:56 I want to talk about the bull case for open source companies like Mistral in a world where some of these bigger players are open sourcing more, becoming more interested in that.
0:26:57 So there’s a couple.
0:27:15 And we’ve talked about this increasingly in a world where two years ago, I think when we led the investment of Mistral, we had a fairly clear hypothesis for how open source wins in an arc where foundation models end up looking more and more like traditional compute infrastructure, storage, networking, et cetera.
0:27:30 Because closed source usually, if you look at databases or operating systems, windows, closed source usually leads the way in terms of opening up new use cases, often captures a ton of value, certainly from consumers.
0:27:36 But when the enterprise starts really adopting that technology, they usually want cheaper, faster, and more control.
0:27:44 And in the world of AI, you can’t get the kind of control most enterprises want without having access to the weights.
0:27:50 And at the time, the only real comparable model to the frontier closed source was Lama.
0:27:52 And then the creators of Lama left to start Mistral.
0:27:54 So it was a pretty natural decision.
0:28:01 I think since then, there’s a different thing that’s turned up, which is the idea of sovereign AI infrastructure that’s not just models.
0:28:02 It’s everything else down and up.
0:28:10 And I think something we’ve been debating as well, does that mean the ideal provider of cloud infrastructure is also the provider of the best open source models?
0:28:20 Additionally, cloud infrastructure is a pretty well-dominated category owned by incumbents whose core business was either in telecom or in commerce, like Amazon.
0:28:24 And it seems like now that’s changing.
0:28:31 I think you put it more eloquently than I did, which is if you ask the wrong guys to design the data center, they’re going to design the wrong data center.
0:28:33 But I’m paraphrasing here.
0:28:34 No, I think it’s exactly right.
0:28:43 I mean, each of the last big technological waves, if you look at the PC revolution or the internet boom, right, we developed essentially a new building block for systems, right?
0:28:46 The CPU or the database or the network.
0:28:52 I think now we’re the process of building yet another building block, which is the model or AI, whatever it may be called in the end.
0:28:52 Right.
0:28:54 So it’s a fourth pillar in a sense.
0:28:56 Compute network storage has become a compute network storage model.
0:29:01 And in that kind of world, a cloud needs to provide all four.
0:29:02 Right.
0:29:03 And so I think you’re exactly right.
0:29:06 This is just part of the infrastructure layer that in the future you’ll build all the software systems.
0:29:07 Right.
0:29:10 I think one way to think about that is there’s two frontiers.
0:29:12 There’s the capabilities frontier.
0:29:14 And then there’s the Pareto efficiency frontier.
0:29:18 The capability frontier is usually dominated by closed source.
0:29:34 And then the Pareto efficiency frontier, because of all the goodness of open source ecosystem flywheel effects, right, where in this case you put out your model and the entire ecosystem of developers can distill it, fine tune it, ship better runtime improvements to the model, quantize it and so on.
0:29:40 That makes that family of technology much more efficient to run than the closed source version.
0:29:49 The second is more secure, because you have the whole world red team in your model versus just this limited group of people inside your company that if you’re a closed source provider.
0:29:53 So the business case is basically cheaper, faster, more efficient, more controllable.
0:29:56 It’s pretty strong for the raw model abstraction.
0:30:03 Then if you ask, okay, well, does the model provider have the right to win?
0:30:08 Is there a business case below the model stack at the data center, the chip level, at the cluster level?
0:30:10 And is there a right to win above?
0:30:14 Let’s start with the topmost part of the stack, which increasingly people would call agents.
0:30:28 A less sexy version would be to call it a fully end-to-end automated workflow, right, where today you have, if you take the world’s largest shipping company, the Mercs of the world or the CMA CGMs, right?
0:30:33 These are massive logistics and transportation companies that have fairly complex workflows.
0:30:47 And if you think about the power of these models being turned into an AI agent, the work required to customize that agent for one of these mission-critical industries is quite hard today.
0:30:54 An area where we’re seeing a ton of progress is reinforcement learning, where if you craft the right reward model, the agent gets much better at accomplishing that task.
0:30:56 Well, it turns out crafting the right reward model is really hard.
0:31:02 Even for sophisticated teams like OpenAI, I mean, they’ve literally rolled back an update to ChatGPT, I think, three days ago.
0:31:05 They called it the sycophancy update, where they crafted the wrong reward model.
0:31:11 And so a traditional legacy industry company has no clue how to do this.
0:31:25 And the question is, would they rather invest that energy to customize a closed-source model or an open-source model, where if the closed-source provider, for whatever reason, goes down, shuts shop, which happens, raises prices, and so on.
0:31:26 Steals their customers.
0:31:27 Yeah, steals their customers.
0:31:29 We’re essentially hosed.
0:31:36 And the natural arc of that as well for the agent layer seems to be to go to a deployment partner who has an underlying open-source base.
0:31:41 I think the cloud infrastructure, the sovereign AI layer, is a bit up for grabs.
0:31:44 And that might be a good topic for our next pod.
0:31:46 Yeah, absolutely.
0:31:47 Well, let’s wrap on that.
0:31:48 Anish Guido, thank you so much.
0:31:49 It’s been great.
0:31:49 Thank you.
0:31:54 Thanks for listening to the A16Z podcast.
0:31:59 If you enjoyed the episode, let us know by leaving a review at ratethispodcast.com slash A16Z.
0:32:02 We’ve got more great conversations coming your way.
0:32:03 See you next time.
What happens when AI stops being just infrastructure—and becomes a matter of national identity and global power?
In this episode, a16z’s Anjney Midha and Guido Appenzeller explore the rise of sovereign AI—the idea that countries must own their own AI models, data centers, and value systems.
From Saudi Arabia’s $100B+ AI ambitions to the cultural stakes of model alignment, we examine:
- Why nations are building local “AI factories” instead of relying on U.S. cloud providers
- How foundation models are becoming instruments of soft power
- What the DeepSeek release tells us about China’s AI strategy
- Whether the world needs a “Marshall Plan for AI”
- And how open-source models could reshape the balance of power
AI isn’t just a technology anymore – it’s geopolitical infrastructure. This conversation maps the new battleground.
Resources:
Find Anj on X: https://x.com/AnjneyMidha
Find Guido on X: https://x.com/appenz
Stay Updated:
Let us know what you think: https://ratethispodcast.com/a16z
Find a16z on Twitter: https://twitter.com/a16z
Find a16z on LinkedIn: https://www.linkedin.com/company/a16z
Subscribe on your favorite podcast app: https://a16z.simplecast.com/
Follow our host: Follow our host: https://x.com/eriktorenberg
Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.