The 80-Year Bet: Why Naveen Rao Is Rebuilding the Computer from Scratch

0
0
AI transcript
0:00:03 I think AI is the next evolution of humanity.
0:00:04 I think it takes us to a new level,
0:00:08 allows us to collaborate and understand the world in much deeper ways.
0:00:10 Naveen Rao, here’s an expert in AI.
0:00:13 Naveen Rao, probably one of the smartest guys in this domain.
0:00:15 He sees things well before anybody else sees them.
0:00:18 You had a lot of success doing Nirvana Mosaic and Databricks.
0:00:20 Why start a new chip company now?
0:00:21 First off, not a chip company per se.
0:00:25 Most of what we’re doing is really kind of looking at first principles
0:00:27 of how learning works, that physical system.
0:00:29 NVIDIA, TSMC, Google.
0:00:32 Are these potential allies for unconventional?
0:00:33 Are these competitors?
0:00:35 I think TSMC is absolutely going to be a partner.
0:00:37 Google kind of has everything internally.
0:00:39 And NVIDIA, of course, they’ve built the platform
0:00:40 that everyone programs on today.
0:00:43 So are we going to be at odds with NVIDIA going forward?
0:00:45 I don’t know. We’ll see what the world looks like.
0:00:46 But there could be a world where we collaborate.
0:00:49 Has anyone called you crazy yet for doing this?
0:00:50 Oh yeah, plenty of people.
0:00:54 A squirrel brain runs on a tenth of a walk.
0:00:59 Our AI data centers now consume 4% of the entire U.S. power grid,
0:01:02 and we need 400 more gigawatts in the next decade just to keep up.
0:01:05 Naveen Rao thinks the problem isn’t power generation.
0:01:08 It’s that we’ve been building the wrong kind of computer for 80 years.
0:01:11 Naveen sold his last AI chip company to Intel,
0:01:14 and now he’s back with a bet most people call crazy.
0:01:17 Analog computing, purpose built for intelligence.
0:01:21 In this conversation, A16Z’s Matt Bordenstein sits down with Naveen
0:01:24 to discuss why now is the time for this unconventional bet.
0:01:32 Our guest today is Naveen Rao, co-founder and CEO of Unconventional AI,
0:01:34 which is an AI chip startup.
0:01:36 Prior to that, Naveen was at Databricks as head of AI
0:01:39 and co-founder of two successful companies,
0:01:42 Mosaic in the cloud computing world and Nirvana,
0:01:44 doing AI chip accelerators before it was cool.
0:01:46 We’re here reporting from NeurIPS,
0:01:49 and great to have you on the podcast, Naveen. Welcome.
0:01:50 Thanks. Thanks for having me.
0:01:52 So you were kind of at the vanguard
0:01:56 thinking about what the proper hardware is for running AI workloads.
0:01:59 Absolutely. I mean, you know, it’s like when you have a hammer,
0:02:00 everything’s a nail, I suppose.
0:02:02 But the early part of my career was really about
0:02:06 how do I take certain algorithms and capabilities
0:02:07 and shrink them, make them faster,
0:02:10 put them into form factors that make those use cases proliferate,
0:02:14 like wireless technology or video compression.
0:02:17 You couldn’t do video compression in real time on a laptop back then.
0:02:19 There wasn’t enough computing power.
0:02:21 So you actually needed to build hardware to do those kind of things.
0:02:23 So early part of my career was all about that.
0:02:25 And then I went back to academia,
0:02:26 I did a PhD in neuroscience.
0:02:28 And so you still kind of look at it like,
0:02:30 hey, can I make something better that’s more efficient?
0:02:33 And so you sold Nirvana to Intel
0:02:36 and then founded Mosaic, which is a cloud company.
0:02:38 It’s interesting to sort of cross domains like that,
0:02:41 I think, to be able to look at hardware and software.
0:02:43 I would sort of argue Mosaic was really a software company.
0:02:44 How did you make that decision?
0:02:46 And why do you think you have these diverse interests?
0:02:49 Well, I think I was, I don’t know,
0:02:51 I guess you would call it an OG kind of full stack.
0:02:54 Now, full stack engineer means something different than it meant back then.
0:02:59 I think back then it meant someone who understands potentially devices like silicon,
0:03:04 how to do logic design, computer architecture, low-level software,
0:03:07 maybe OS-level software, and then application.
0:03:08 That was a full stack engineer.
0:03:10 And I actually had touched all those topics.
0:03:13 So to me, it’s very natural to kind of think across these boundaries.
0:03:16 To me, like, software and hardware is not really a natural boundary.
0:03:19 It’s just where we decide to draw the line and say,
0:03:21 okay, this is something I make configurable or I don’t.
0:03:24 And it’s like, where is the world going to consume something?
0:03:25 Where is the problem?
0:03:28 And then right size and figure out the solution to go and hit it.
0:03:31 Now full stack means, I know, JavaScript and Python.
0:03:32 That’s right.
0:03:35 So you’ve had a lot of success doing both of those things and at Databricks.
0:03:37 Why start a new chip company now?
0:03:38 It is kind of crazy.
0:03:41 It’s one of these things like, actually, it was first off there.
0:03:42 It’s not a chip company per se.
0:03:45 Most of what we’re doing is, at the beginning, is theory.
0:03:50 And we’re really kind of looking at first principles of how learning works in a physical system.
0:03:55 And the reason to go back and do this is just purely out of passion.
0:03:58 I think we can change how a computer is built.
0:04:01 We’ve been building largely the same kind of computer for 80 years.
0:04:03 We went digital back in the 1940s.
0:04:10 And in undergrad, in the 1990s, when I learned about, you know, the dynamics of the brain,
0:04:14 like the brain’s 20 watts of energy and the kind of computations that can happen inside
0:04:15 of brain and neural systems.
0:04:17 I was just blown away then.
0:04:19 I’m still blown away by it.
0:04:23 And I think we haven’t really scratched the surface of how we can get close to that.
0:04:26 Biology is exquisitely efficient.
0:04:27 It’s very fast.
0:04:30 It right-sizes itself to the application at hand.
0:04:35 When you’re chilling out, you don’t use much energy, but you’re still aware of other threats
0:04:36 and things like this.
0:04:39 And once a threat happens, like everything turns on, it’s very dynamic.
0:04:41 And we really haven’t built systems like this.
0:04:46 And I’ve been in the industry long enough to know that we have to have an incentive to
0:04:47 build things.
0:04:50 You can’t just say, hey, I want to build this cool thing, therefore I go build it.
0:04:51 Maybe in academia, you can do that.
0:04:53 But in sort of the real world, I can’t.
0:04:57 And now it’s exciting because those concepts are super relevant.
0:05:03 We’re at a point in time where computing is bound by energy at the global level, which just
0:05:05 was never true in all of humanity.
0:05:11 And so for those of us who aren’t experts, can you describe the difference between digital
0:05:12 and analog computing systems?
0:05:17 And like, why do you think the architecture has evolved the way it has sort of more digital
0:05:18 focus over decades, as you said?
0:05:25 Yeah, I mean, very simply, digital computers implement numerics and numerics with some sort
0:05:26 of estimation, right?
0:05:30 I mean, in a digital computer, a number is represented by a fixed number of bits.
0:05:33 And that has some precision error and things like this.
0:05:35 It’s just a way we implement the system.
0:05:40 If you make it enough bits, like 64 bits, you can largely say that maybe the error is
0:05:41 small, you don’t have to think about it.
0:05:47 And so the digital computer is capable of simulating anything that you can express as numbers and
0:05:47 arithmetic.
0:05:49 So it became a very general machine.
0:05:52 I can literally simulate any physical process.
0:05:55 All of physics, we try to do computational physics, right?
0:05:59 I have an equation, I can then write numeric solvers that sort of deal with those imprecisions
0:06:01 in the number of bits.
0:06:04 And so this became, obviously, computer science, the entire field now.
0:06:11 And we went that direction actually very early on because we couldn’t scale up computation.
0:06:14 It’s actually kind of an interesting conversation if you look from back then, not that it was
0:06:15 there, of course.
0:06:18 But if you look at the papers and things, they actually look very similar to today in terms
0:06:19 of scaling up GPUs.
0:06:21 Analog computers were actually some of the first computers.
0:06:23 And they worked really well.
0:06:28 They were very efficient, but they couldn’t be scaled up because of manufacturing variability.
0:06:30 So someone said, oh, okay, you know what?
0:06:35 I can actually say I can make a vacuum tube behave as a higher low, very reliably.
0:06:39 I can’t characterize the in-between very well, but I can say it’s higher low.
0:06:42 And so that was kind of where we went to digital abstraction.
0:06:43 And then we could scale up.
0:06:47 ENIAC, which was built in 1945, had 18,000 vacuum tubes.
0:06:47 Wow.
0:06:51 So 18,000 is kind of similar to how many GPUs people use now, right, for large-scale
0:06:52 training.
0:06:54 So scaling things up is always a hard problem.
0:06:57 And once you figure out how to do it, it makes a paradigm happen.
0:06:59 And I think that’s why we went to digital.
0:07:04 But analog still is inherently more efficient because it’s actually analogous computing is
0:07:05 the way to think about it.
0:07:09 Like, can I build a physical system that is similar to the quantity I’m trying to express
0:07:10 or compute over?
0:07:15 You’re effectively using the physics of the underlying medium to do the computation.
0:07:18 And so in digital computers, we have transistors.
0:07:22 Just to make it sort of concrete, what kind of substrates are you talking about for analog
0:07:22 computers?
0:07:25 Yeah, I mean, analog computers can do lots of different things.
0:07:29 There’s wind tunnels are a great example of an analog computer in the sense where I have
0:07:33 a race car, a track, or an airplane, and I want to understand how the wind moves around
0:07:34 it.
0:07:38 And you can, in theory, solve those things computationally.
0:07:41 The problem is you’re always going to be off.
0:07:43 It’s very hard to know what the real system is going to look like.
0:07:47 And doing things with computational fluid dynamics accurately is pretty hard.
0:07:49 So people still build wind tunnels.
0:07:50 That’s actually modeling that.
0:07:51 That’s an analog computer.
0:07:55 I think we still have lots of reasons to build these analogous type computers.
0:08:02 Now, in the situation we’re talking about, we can actually build circuits in silicon to recapitulate
0:08:03 behaviors of neural networks.
0:08:08 So what we’re doing today is more specified than what we were doing 80 years ago, in a
0:08:13 sense, is that then we were trying to automate generic calculation, which was used to calculate
0:08:14 artillery trajectories.
0:08:20 It was used to calculate finances, maybe some physics problems like going into space, things
0:08:20 like that.
0:08:26 Those require determinism and specificity around these numbers and these computations.
0:08:28 Intelligence is a different beast.
0:08:32 You can build it out of numbers, but is it naturally built out of numbers?
0:08:33 I don’t know.
0:08:36 A neural network is actually a stochastic machine.
0:08:43 And so why are we using the substrate that is highly precise and deterministic for something
0:08:45 that’s actually stochastic and distributed in nature?
0:08:51 So we believe we can find the right isomorphism in electrical circuits that can subserve intelligence.
0:08:54 That’s a pretty wild idea, isn’t it?
0:08:59 Maybe unpack it one level deeper, because I totally agree with you.
0:09:03 Computers for decades have been sort of the complement to human intelligence, right?
0:09:07 It’s like, hey, my brain isn’t really great at computing an orbital trajectory.
0:09:07 That’s right.
0:09:09 And I don’t want to burn up on reentry.
0:09:12 So like a computer can help us with this incredible degree of precision.
0:09:15 We’re now kind of going the opposite direction, right?
0:09:20 We’re actually trying to encode more sort of fuzziness into computer systems.
0:09:25 So go maybe just a little bit deeper on this idea of an analog and why intelligence is a good fit for analog systems.
0:09:29 Well, I mean, the best examples we have of intelligent systems’ nature of brains.
0:09:33 And it’s often been said, you know, human brains run on 20 watts of energy.
0:09:33 That is true.
0:09:37 But if you look at mammalian brains generally, they’re actually extremely efficient.
0:09:39 Like a squirrel or a cat, it’s like a tenth of a watt.
0:09:42 And so there’s something there that we’re still missing.
0:09:51 And not to say that we understand all of it, but part of what I think we’re missing is we have lots of abstractions in a computer that are quite lossy.
0:09:55 In a brain, the neural network dynamics are implemented physically.
0:09:57 So there is no abstraction.
0:09:59 Intelligence is the physics.
0:10:00 They’re one and the same.
0:10:04 There’s no, you know, OS and, you know, some sort of API and this and that.
0:10:11 So there’s some visual stimulus, for instance, that directly activates an actual neural network and produces some somatic response, that sort of thing.
0:10:11 Exactly.
0:10:19 And those things are mediated by chemical diffusion and, you know, the physical properties of the neuron, the physics itself.
0:10:26 So I think absolutely it’s possible to build something that’s much more efficient by using physics in an analogous way.
0:10:27 That is 100% true.
0:10:33 Can we do it and build products out of it is really the question we’re asking here at Unconventional.
0:10:39 And is part of the idea that now is the right time because AI is both a huge and a unique workload?
0:10:41 Yeah, absolutely.
0:10:42 You know, it’s interesting.
0:10:43 So just maybe some stats here.
0:10:47 Like, the U.S. is about 50% of the world’s data center capacity.
0:10:53 And today we put about 4% of the energy grid, of the U.S. energy grid, into those data centers.
0:11:00 And this past year, 2025, was the first time we started to see news articles about brownouts in the Southwest during the summer.
0:11:06 And, you know, just imagine what happens when this goes to 8%, 10% of the energy grid.
0:11:08 It’s not going to be a good place that we’re in.
0:11:10 So can we build more power?
0:11:11 Absolutely, we should.
0:11:16 Building power generation is very hard, expensive, and it’s infrastructure.
0:11:18 Like, it takes time.
0:11:22 You can’t, you can only bring online so many kilowatts or gigawatts per year.
0:11:25 So it’s something on the order of four per year.
0:11:32 By some estimates, we need 400 gigawatts additional capacity over the next 10 years to power the demand for AI.
0:11:33 Wow.
0:11:34 So we have a huge shortfall.
0:11:36 And so we really just need to rethink this.
0:11:45 The, you know, 15-year-old sci-fi nerd in me says, like, wow, we’re mobilizing, you know, species-scale resources to, like, invent the future.
0:11:46 We are.
0:11:47 And then there’s the practical.
0:11:57 It’s like, even if we add 400 gigawatts of production capacity, our, you know, 1970s-era transmission grid is probably going to melt under the load.
0:12:00 So, yeah, so there’s very serious sort of infrastructure hurdles to this, I think.
0:12:03 Like, it’s hard to get a lot of humans to act together, right?
0:12:04 It’s just a reality.
0:12:07 That’s what has to happen to solve these problems.
0:12:09 What trade-offs do you think this entails?
0:12:13 You know, sort of the path you’re pursuing versus the mainstream digital path, no?
0:12:17 Yeah, I actually don’t see it as, you know, it’s digital or analog.
0:12:18 This doesn’t work like that.
0:12:22 I think there are certain types of workloads that are amenable to these analog approaches,
0:12:27 especially the ones that are, that can be expressed as dynamical systems.
0:12:28 Dynamics mean time.
0:12:30 They have time associated with them.
0:12:32 In the real world, every physical process has time.
0:12:37 And in the computing world, like the numeric computing world, we actually don’t have that concept.
0:12:39 You simulate time with numbers.
0:12:43 Actually, simulating time is very useful in certain problems.
0:12:50 So, I think we should still build those things and we should still have those capabilities for the problems that we need to solve that way.
0:12:54 But for these problems where, you know, like you said, it’s a bit fuzzier.
0:12:58 I’m trying to retrieve and summarize across multiple inputs.
0:13:00 That’s actually what brains do really well, right?
0:13:07 They can, they can take in tons of data and sort of formulate a model of how those things interact.
0:13:10 And sometimes those models can be actually extremely accurate.
0:13:17 Like, look at an athlete, you know, you know, Alex Honnold climbed El Capitan, right?
0:13:19 Like, just think about the precision that’s required.
0:13:22 It still scares me every time I see it.
0:13:22 It’s insane, right?
0:13:22 Yeah.
0:13:26 And if he slips, like just, like he’s off by a millimeter in some places.
0:13:26 That’s wild.
0:13:28 He dies, right?
0:13:31 And that’s true for like every top level athlete.
0:13:33 They’re someone who’s, you know, at the Olympics.
0:13:34 Yes.
0:13:38 Steph Curry, you know, the story is he set up a special tracking system so he can make
0:13:41 sure the ball was sitting in the middle of the rim, not just going through.
0:13:46 So the level of precision these guys hit with a neural network that’s noisy is actually quite
0:13:47 high.
0:13:51 So neural systems can actually do a lot of precision under certain circumstances.
0:13:57 But what’s interesting about these situations is Steph Curry, when he shoots a ball, is never
0:13:59 going to shoot it under ideal circumstances in a game.
0:14:02 Always it’s a unique input.
0:14:05 And there’s a lot of different input variants coming at you.
0:14:08 Like where the other players are, precisely where you’re standing.
0:14:10 Maybe your shoes are different.
0:14:11 Maybe the surface is a little different.
0:14:14 Like maybe the ball is tackier or your hands are sweaty.
0:14:17 Like there’s so many inputs and we kind of put them all together.
0:14:19 integrate them and still have very accurate behavior.
0:14:22 So brains are exceptionally good at this.
0:14:26 And, you know, that’s a set of problems that is actually very useful to solve.
0:14:28 And now we’re approaching those problems.
0:14:34 But it doesn’t mean we don’t still use computational substrates to do actual computation.
0:14:36 This is kind of an intelligence substrate.
0:14:44 And so what types of AI models or data modalities do you expect your hardware will be well suited
0:14:44 for?
0:14:47 Yeah, so we’re obviously starting with.
0:14:48 The state of the art today.
0:14:50 Like transformers, diffusion models.
0:14:51 They work.
0:14:53 They do really good stuff.
0:14:54 So we shouldn’t throw that out.
0:14:57 And diffusion models and flow models are actually,
0:15:02 energy-based models are actually pretty interesting because they inherently have dynamics as part of them.
0:15:05 They’re literally written as an ordinary differential equation.
0:15:18 So that makes it such that, hey, can I map those dynamics onto the dynamics of a physical system in some way that’s either fixed or has some principled way of evolving?
0:15:23 And then can I basically use that physical system to implement that thing and do it very efficiently with physics?
0:15:25 So that’s kind of the nature of what we’re doing.
0:15:30 And we will be releasing some open source and things around this to let people play around.
0:15:40 But, you know, transformers are really, they’re a big innovation because they made the constructs of a GPU work extremely well.
0:15:48 And it doesn’t mean it’s wrong, but I don’t think, there’s nothing natural, there’s no natural law about the parameter of a transformer.
0:15:53 Transformers parameter is a function of the nonlinearities and the way a whole thing is set up with attention.
0:15:59 There’s going to be some kind of mapping between transformer parameter spaces and these other parameter spaces.
0:16:06 And it’s, transformers are, I think, have kind of used lots of parameters to accomplish what they do.
0:16:14 I have to ask, just since you mentioned energy-based models and Jan LeCun has been, you know, writing quite a lot about this.
0:16:23 Do you think pursuing these sorts of paths that you’re talking about gets us closer on the path to AGI, whatever AGI means?
0:16:25 Honestly, I do.
0:16:28 The reason I feel that way, and again, this is hand-wavy.
0:16:29 I’m going to be really honest.
0:16:31 That’s why I’m putting quotes around AGI.
0:16:33 I think that the discussion is necessarily hand-wavy.
0:16:35 It’s got to be because we just don’t know.
0:16:43 But my intuition says that anything where the basis is dynamic, which has time and causality as part of it,
0:16:45 will be a better basis than something that’s not.
0:16:48 So we’ve largely tried to remove that.
0:16:51 And, you know, a lot of times you can write math down.
0:16:52 It’s reversible in time and things like that.
0:16:56 But the physical world tends not to be, at least the way we perceive it.
0:17:02 And so can we build out of elements of the physical world that are, you know, do have time evolution?
0:17:05 I think that’s the right basis to build something that understands causation.
0:17:13 So I do think we’ll have something that is better and will give us something closer to what we really think is intelligence.
0:17:16 Because, yes, we have intelligence in these machines.
0:17:19 I don’t think they’re anywhere close to AGI because, I mean, they still make stupid errors.
0:17:21 They’re very useful tools.
0:17:24 But they’re not, it’s not like working with a person, right?
0:17:24 Yeah, totally.
0:17:27 That’s actually really interesting.
0:17:37 So the sort of thing that’s missing in AI behavior, which I think a lot of us see that there’s something missing but can’t quite put a name to it.
0:17:41 It sounds like you’re arguing part of that is sort of a real sense of causality.
0:17:41 Yeah.
0:17:50 And that training in a more dynamic sort of regime may impart this kind of like apparent understanding of causality better than what we have now.
0:17:51 Yeah.
0:17:52 And again, head wavy, but yes.
0:17:56 I mean, look, you have kids, little kids, and you see them.
0:18:00 I mean, children kind of innately understand causality in some ways.
0:18:02 Like, you know, this happened, then that happened.
0:18:05 And yes, I know you can say like it’s reinforcement learning or whatever.
0:18:06 That’s some part of it.
0:18:10 But there’s something innate that we understand causality.
0:18:12 And in fact, that’s how we move our limbs and all of that.
0:18:15 I know if I send a certain command to my arm, it’ll do something.
0:18:22 So I think there’s something innate about the way our brains are wired and built out of primitives that do understand causation.
0:18:30 Put unconventional in the context of the broader industry for me, like NVIDIA, TSMC, Google.
0:18:34 Are these, you know, potential allies for unconventional?
0:18:35 Are these competitors?
0:18:35 How do you think about it?
0:18:36 Yeah.
0:18:45 I mean, a couple of things that we set out to do when we built, we were starting this company was see if we can find a paradigm that’s analogous to intelligence within five years.
0:18:50 And then at the five-year mark, we should be able to build something that’s scalable from a manufacturing standpoint.
0:18:55 So, you know, you can think about building a computer out of many different things.
0:19:01 But if it’s not scalable from a manufacturing standpoint, we can’t intercept this global energy problem.
0:19:05 So we need to have somebody to say, okay, go build 10 million of these things, right?
0:19:11 So I think TSMC is absolutely going to be a partner forward, you know, met with them recently.
0:19:17 And, you know, we want to work closely with them to make sure we get what we need, get fast turnaround times to prototype and all of that.
0:19:24 Google, NVIDIA, Microsoft, all these guys are, you know, at the forefront of where the application space is.
0:19:36 Obviously, Google kind of has everything internally, and I think they’re working on sort of lower risk, but, you know, continual improvements for their hardware.
0:19:37 With TPUs, you mean?
0:19:38 With TPUs, yeah.
0:19:42 From what I can see, you know, just publicly is, it makes total sense, right?
0:19:44 They have a business to run, and they’re trying to make their margins better.
0:19:49 And, you know, how can I do that with all the tools I have, you know, in front of me?
0:19:55 NVIDIA, of course, you know, they’ve built the platform that everyone programs on today.
0:19:59 So, is it, are we going to be at odds with NVIDIA going forward?
0:20:00 I don’t know.
0:20:01 We’ll see what the world looks like.
0:20:06 But, I mean, we are trying to build a better substrate than Matrix Multiply.
0:20:10 There could be a world where we collaborate on such solutions.
0:20:13 And, you know, we’re open to all of these things.
0:20:18 Where do you personally get the motivation to get up in the morning and build this company?
0:20:21 I mean, you’ve had a lot of success in your career this year, you’re a startup.
0:20:24 What, you know, what’s exciting about this to you?
0:20:26 I don’t know.
0:20:27 It’s a weird thing.
0:20:29 Like, if you haven’t worked in hardware, it’s hard.
0:20:33 I’ve been fortunate to work in hardware and software.
0:20:38 And, you know, I love writing a bunch of software and then hitting a pile and seeing it work.
0:20:39 That’s a good dopamine hit.
0:20:45 But, man, when you work on a piece of hardware and you turn that thing on, that’s a big dopamine hit.
0:20:49 That’s like, it’s just like celebration, jumping, you know, jumping up in the air, high-fiving.
0:20:50 It’s a different thing.
0:20:53 And, I don’t know, you sort of live for these moments, you know.
0:20:59 Like, when I was at Intel, like, I was one of the only execs who would go to the lab and the first chip would come back.
0:21:00 I’m like, I want to see what happens.
0:21:03 Sometimes you turn it on.
0:21:05 You see the little puff of smoke come out.
0:21:06 You’re like, uh-oh.
0:21:09 That’s not good.
0:21:09 But you want to be there.
0:21:10 It’s going to be part of the moment.
0:21:12 But I think that’s part of it.
0:21:22 I think, for me personally, like, we have this opportunity now that we can really change the world of computing and make AI ubiquitous.
0:21:24 I’m the opposite of an AI doomer.
0:21:28 I think AI is the next evolution of humanity.
0:21:29 I think it takes us to a new level.
0:21:34 It allows us to collaborate, understand each other, and understand the world in much deeper ways.
0:21:35 Totally agree.
0:21:39 So, every technology has negatives, but the positives, to me, so far outweigh it.
0:21:43 And the only way we’re going to get to ubiquity is we have to change the computer.
0:21:49 The current paradigm, as good as it is and as far as it’s taken us, is not going to take us to that level.
0:21:51 I think that’s such a great way to say it.
0:21:57 AI actually can help us understand each other better, help us understand ourselves better, understand the natural world better.
0:22:04 I don’t think it’s at all what some of the doomers think of, you know, replacing sort of human experience.
0:22:05 That’s a short-term thing.
0:22:07 I mean, there will be bumps along the way, right?
0:22:08 Technology does that.
0:22:12 That’s what happens when you’ve seen too many sci-fi movies.
0:22:13 That’s right.
0:22:14 But what’s Star Trek?
0:22:14 Yeah.
0:22:15 Yeah, totally.
0:22:16 Totally.
0:22:16 It’s great.
0:22:20 This is a really big swing, right?
0:22:22 Like, this is a very ambitious company.
0:22:27 What gives you confidence that it’s going to work or, you know, has a reasonable shot of working?
0:22:29 There’s a number of data points.
0:22:32 Of course, like I said, the brains are existence-proof.
0:22:38 But there’s also 40-plus years of academic research, which is showing a lot of promise here.
0:22:45 People have built different devices, albeit not in the latest technology with professional engineering teams,
0:22:48 but they have built proofs of concept that actually show some of these things work.
0:22:55 We’ve also, from a theory standpoint, both from neuroscience and just pure dynamical systems and math theory,
0:22:59 do start to understand how these systems can work.
0:23:03 So, I think we now have pieces at different parts of the stack that show,
0:23:06 hey, if I can combine these things the right way, I can build this.
0:23:11 And, you know, that’s what great engineering is all about, is like, you know,
0:23:15 exploiting this thing that someone else built for something else, exploiting that thing.
0:23:21 Engineers are kind of like the opposite of theorists.
0:23:23 It’s like, well, all right, that thing doesn’t quite fit.
0:23:24 Sand it down and make it fit, right?
0:23:27 So, it’s like, we got to do a little bit of that right now, and then we can build something.
0:23:28 Put it all together.
0:23:28 That’s awesome.
0:23:31 Has anyone called you crazy yet for doing this?
0:23:32 Oh, yeah.
0:23:32 Plenty of people.
0:23:33 That’s fine.
0:23:34 Is it like everybody?
0:23:36 Well, I’m used to this at this point.
0:23:38 You know, my family, I would call crazy.
0:23:43 I was called crazy going back to grad school years ago when I had a very good career in tech.
0:23:45 So, it’s fine.
0:23:48 I think that’s, you need crazy people to go out and explore.
0:23:52 I mean, if you think about humanity out of Africa, all that, the crazy people went out.
0:23:54 We would be lost without crazy.
0:23:55 You need some crazy in there.
0:23:56 So, it’s okay.
0:23:56 I’m fine with that.
0:24:03 And so, what kind of people are you looking to bring on to the team of a very ambitious goal?
0:24:05 Who should be interested in joining you?
0:24:14 Yeah, I mean, I think some of the traditional, traditional-ish, when I say traditional, I mean, over the last five years, this field of AI systems has evolved.
0:24:19 Like, people who are really good at taking algorithms and mapping them very effectively to physical substrates.
0:24:30 Those folks who understand energy-based models, flow models, gradient descent in different ways, you know, this kind of thing is what we need there.
0:24:41 We need theorists who can think about different ways of building coupled systems, how I can characterize the richness of dynamical systems and relating that to neural networks.
0:24:43 So, there is a theory aspect of this.
0:24:46 Then, there’s folks who are, like, kind of at the system architecture level.
0:24:48 It’s like, all right, here’s what the theory says.
0:24:49 This is what I can really build.
0:24:50 How do I bridge that gap?
0:24:56 And then, there’s the people actually physically building this stuff, like analog circuit people, actually digital circuit people, too.
0:24:58 We’re going to have a mixed signal here.
0:25:00 So, that’s the whole stack.
0:25:06 The stack is, it’s hard because these are all things that no one’s really pushed to that level.
0:25:15 Like, when we build this chip, our first prototype, it’s going to be probably one of the larger, maybe the largest analog chip people have ever built, which is kind of weird.
0:25:17 First time you do something, things don’t usually work the way you think they do.
0:25:23 So, you can get in on that Cerebrus-Jensen game where they’re each pulling the biggest possible wafer out of an oven.
0:25:24 Something like that.
0:25:25 Yeah, yeah, yeah.
0:25:25 Exactly, right.
0:25:28 Put a few vacuum tubes on top for effect.
0:25:30 Yeah, I need blinking lights.
0:25:31 Yeah, exactly.
0:25:33 We’re not going to have cool heat sinks.
0:25:33 It’s going to be super cool.
0:25:34 It’s going to be cold.
0:25:36 You don’t need big heat sinks, you know.
0:25:39 So, I hope they make something that looks interesting here.
0:25:50 This is a funny time for top AI people, right, where you have sort of the option, if you want to start a company, there’s a lot of venture capitalists who probably would fund you.
0:25:57 If you want to get a cushy job at a big company, you can get a very cushy job and kind of do some interesting things.
0:26:08 Or, you know, people can join a startup like Unconventional that, you know, has a lot of the nice aspects people look for in sort of AI careers and are taking like super sort of big swings.
0:26:11 I’m just sort of curious, you’ve been on all sides of this.
0:26:15 Like, do you have any advice for, you know, younger people starting out in their careers?
0:26:17 Or how do you think about this?
0:26:23 I think you get such a breath of working in a startup that at the beginning of your career, that will pay dividends later on.
0:26:29 Because, like I said, like the reason I can think across the stack is because I did all those things very early in my career, you know.
0:26:31 I built hardware, I built software, I built applications.
0:26:36 And in big companies, it’s not Amos’ fault, it’s just the way it is.
0:26:38 Like, you get hired to do a thing.
0:26:42 And you do that thing over and over again, and you’re really good at doing that thing.
0:26:43 And that’s fine.
0:26:46 You need people who are really good at doing specific things.
0:26:56 But if you want to be prepared for change in the future, being really good at one thing is probably less valuable than being very good at, but slightly good at a lot of things.
0:26:57 Yeah, that’s interesting.
0:27:02 Is it fair to say on conventional sort of a practical research lab, is that kind of the culture you’re going for?
0:27:03 Absolutely, yeah.
0:27:05 I mean, the first few years, it really is open-ended.
0:27:06 I don’t want to close doors.
0:27:08 Like, I’m really specific about this.
0:27:12 Like, I always try to bring the conversation back because those people are like, oh, that’s going to be hard to manufacture.
0:27:13 I was like, stop.
0:27:14 Don’t think about that.
0:27:15 Will it work?
0:27:18 First, come up with existence proofs.
0:27:20 Then we go back and try to engineer it.
0:27:22 And, you know, all the trade-offs therein.
0:27:26 But if you make those trade-offs up front, you don’t go into a good place.
0:27:31 So, yes, we’re really thinking wide open, but with an eye on the future, we are building a product.
0:27:41 And to your point, it takes not only people with diverse skill sets, but people with kind of high agency to try new things and learn new things and kind of integrate across the stack.
0:27:52 Yeah, I mean, I think what I’ve done really well across the companies I’ve built has been going after hard problems, which kind of lends itself to smart people wanting to come in and try to solve them.
0:27:53 They see a challenge.
0:27:55 It’s like, here’s the amount of risk of climate.
0:28:04 But then giving them agency, and I sort of look at it like, what decisions can I make as a leader to increase agency of the org overall?
0:28:11 Like, me making top style decisions may be globally better for the company in the short term.
0:28:17 But I think long term, we will do better if more people have agency and can try more things out.
0:28:23 So, personally, I like to find ways to get out of the way when I see people who are very passionate about trying something.
0:28:25 It’s like, okay, well, you really want to do this.
0:28:26 That makes sense.
0:28:27 Go for it, you know?
0:28:28 And then you own it.
0:28:30 You own both the good and the bad, right?
0:28:31 And that’s agency to me.
0:28:33 It’s like, you guys are like, okay, I fucked up.
0:28:34 Now, this wasn’t the right question.
0:28:35 That’s okay, too.
0:28:37 But give people the room to do that, you know?
0:28:39 Anything else you want to say before we wrap up?
0:28:47 I mean, I think this is like an opportunity to do something that is generationally will be felt.
0:28:55 You know, to me, that’s what gets me up in the morning is, you know, you can go work on a product and make a tweak, and people will use it.
0:28:55 That’s great.
0:28:58 But, like, in five years, many times people forget those things.
0:29:04 But if we are successful here, the world will not forget this for a very long time, right?
0:29:05 This will be written in history books.
0:29:07 And so I feel like those opportunities are rare.
0:29:12 Thanks for listening to this episode of the A16Z podcast.
0:29:20 If you liked this episode, be sure to like, comment, subscribe, leave us a rating or a review, and share it with your friends and family.
0:29:24 For more episodes, go to YouTube, Apple Podcasts, and Spotify.
0:29:30 Follow us on X at A16Z and subscribe to our Substack at A16Z.substack.com.
0:29:33 Thanks again for listening, and I’ll see you in the next episode.
0:29:48 As a reminder, the content here is for informational purposes only, should not be taken as legal business, tax, or investment advice, or be used to evaluate any investment or security, and is not directed at any investors or potential investors in any A16Z fund.
0:29:53 Please note that A16Z and its affiliates may also maintain investments in the companies discussed in this podcast.
0:30:00 For more details, including a link to our investments, please see A16Z.com forward slash disclosures.

Naveen Rao is cofounder and CEO of Unconventional AI, an AI chip startup building analog computing systems designed specifically for intelligence. Previously, Naveen led AI at Databricks and founded two successful companies: Mosaic (cloud computing) and Nervana (AI accelerators, acquired by Intel). 

In this episode, a16z’s Matt Bornstein sits down with Naveen at NeurIPS to discuss why 80 years of digital computing may be the wrong substrate for AI, how the brain runs on 20 watts while data centers consume 4% of the US energy grid, the physics of causality and what it might mean for AGI, and why now is the moment to take this unconventional bet.

 

Stay Updated:

If you enjoyed this episode, please be sure to like, subscribe, and share with your friends.

Follow Naveen on X: https://x.com/NaveenGRao

Follow Matt on X: https://x.com/BornsteinMatt

Follow a16z on X: https://twitter.com/a16z

Follow a16z on LinkedIn:https://www.linkedin.com/company/a16z

Follow the a16z Podcast on Spotify: https://open.spotify.com/show/5bC65RDvs3oxnLyqqvkUYX

Follow the a16z Podcast on Apple Podcasts: https://podcasts.apple.com/us/podcast/a16z-podcast/id842818711

Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details, please see http://a16z.com/disclosures.

Stay Updated:

Find a16z on X

Find a16z on LinkedIn

Listen to the a16z Show on Spotify

Listen to the a16z Show on Apple Podcasts

Follow our host: https://twitter.com/eriktorenberg

 

Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.

Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.

Leave a Reply

a16z Podcasta16z Podcast
Let's Evolve Together
Logo