Author: The AI Podcast

  • Imbue CEO Kanjun Qiu on Transforming AI Agents Into Personal Collaborators – Ep. 239

    AI transcript
    0:00:10 [MUSIC]
    0:00:13 >> Hello, and welcome to the NVIDIA AI podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:17 One of the big transformations being
    0:00:20 enabled by AI is the way we create software.
    0:00:21 From coding co-pilots to
    0:00:23 in-development systems built to translate
    0:00:26 plain language requests into fully functional applications,
    0:00:28 generative AI is fueling a new wave
    0:00:31 of tools to help us create software faster.
    0:00:33 Our guest today is Kanjun Kew.
    0:00:36 Kanjun is co-founder and CEO of Imbue,
    0:00:38 a three-and-a-half-year-old company somewhere,
    0:00:40 and they’re founded in 2021,
    0:00:43 that is building AI agents that work with us to
    0:00:46 translate our ideas into code and bring them to life.
    0:00:48 There’s a lot more to it than that,
    0:00:51 but why hear it from me when you can hear it directly from Kanjun.
    0:00:53 Kanjun Kew, welcome.
    0:00:56 Thank you so much for taking the time to join the NVIDIA AI podcast.
    0:00:58 >> Thank you, Noah. It’s great to be here.
    0:01:01 Let’s talk about software development and AI,
    0:01:03 and Imbue, and all of that good stuff.
    0:01:05 But maybe let’s start with agents.
    0:01:09 Agents are, I don’t want to use the word “hot”
    0:01:11 because I don’t want it to sound fluffy, right?
    0:01:13 But agents are a thing right now.
    0:01:15 We’ve had some guests on recently talking about
    0:01:17 agents in different contexts,
    0:01:21 and Imbue’s approach to agents is something worth delving into.
    0:01:24 Maybe we can start there. What are AI agents?
    0:01:26 Why do we need them? Why isn’t Imbue working on them?
    0:01:28 >> Yeah. Agents are really popular right now.
    0:01:32 So Imbue was founded in early 2021,
    0:01:33 and at that time,
    0:01:37 our goal was to figure out how to make and work on general AI agents.
    0:01:40 At that time, people thought we were totally crazy.
    0:01:41 Like, what are agents?
    0:01:43 Everyone’s working on AGI.
    0:01:46 AGI is going to rule everything.
    0:01:48 But what we were really interested in is,
    0:01:53 how can we have systems that we as people can mold,
    0:01:56 shape to what we’re wanting?
    0:01:59 As opposed to, oh, this external intelligence that knows all the answers.
    0:02:02 And so we started as a research lab at that time
    0:02:05 because the technology was certainly not good enough
    0:02:08 to build general agents that could reliably do anything
    0:02:09 that we wanted them to do.
    0:02:11 And, you know, in the very beginning,
    0:02:14 we thought of agents in a similar way
    0:02:16 to how a lot of people think about agents today.
    0:02:19 As these kind of, you know, think about what an agent is.
    0:02:21 Often people think about kind of an autonomous,
    0:02:23 personal assistant type thing.
    0:02:25 You ask it to do something, it does it.
    0:02:27 For you, it comes back to you.
    0:02:29 Now everyone has their own personal assistant.
    0:02:32 And actually, a lot of our learning has been
    0:02:36 that that’s a really tricky user experience.
    0:02:39 And we experience it ourselves with human agents,
    0:02:41 which is that often when I delegate something,
    0:02:43 it comes back, it’s not quite what I wanted.
    0:02:46 And now I have to negotiate how to get what I wanted.
    0:02:48 I was listening, I told you before we record,
    0:02:51 I was listening to a little bit of your fireside chat
    0:02:54 with Brian Cutton-Zera from GTC this year,
    0:02:55 which listeners go check that out.
    0:02:57 It’s a great listen.
    0:02:59 And you were talking about the difficulty,
    0:03:01 and I can relate to this so much,
    0:03:04 the difficulty inherent in delegating work
    0:03:06 to someone else, right?
    0:03:08 And to your point, thinking of it as humans,
    0:03:09 you have to break the problem down,
    0:03:11 you have to sort of figure out,
    0:03:14 well, how much do I tell them exactly what to do?
    0:03:14 – Yeah.
    0:03:15 – Yeah, all that, yeah.
    0:03:17 – What context does it need ahead of time?
    0:03:19 What instruction should I give?
    0:03:21 Delegation is actually a really tricky paradigm
    0:03:23 because it actually puts all the onus
    0:03:24 on the person who’s delegating
    0:03:26 to define the problem, define the scope.
    0:03:29 Of course, the person who’s being delegated to the agent
    0:03:31 might come back with some questions and stuff like that,
    0:03:34 but it’s a very tricky thing to trust.
    0:03:35 So what we’ve learned over the years
    0:03:38 working on general agents is we’ve actually started
    0:03:40 to think about agents in a very different way.
    0:03:43 And this is both from a kind of pragmatic business,
    0:03:46 like user perspective and also from a mission perspective.
    0:03:49 The way that we think about agents is,
    0:03:51 if you think about what an agent is,
    0:03:53 what this personal assistant is doing,
    0:03:56 what it is is it’s kind of this intermediary layer
    0:03:58 between you and your computer.
    0:04:00 And you’re kind of telling it stuff
    0:04:02 or interfacing with it in some way,
    0:04:05 whether it’s a UI or a natural language,
    0:04:07 and it’s interfacing with your computer.
    0:04:10 And the most effective way to interface with your computer
    0:04:11 is by writing code.
    0:04:13 That’s what your computer is made of.
    0:04:15 And there are really kind of two types of code.
    0:04:17 There’s like everything is hard-coded,
    0:04:19 which is the default for software today.
    0:04:20 Everything is hard-coded.
    0:04:22 So now your agent can only do a subset of things.
    0:04:25 Or now with language models, you can generate code.
    0:04:27 You can write code that does new stuff
    0:04:28 that’s not hard-coded.
    0:04:30 And now you have an agent that’s much more general.
    0:04:31 It’s able to do sets of tasks
    0:04:34 that I didn’t program into it ahead of time,
    0:04:35 but now it’s able to do.
    0:04:37 And so the way that we think about what agents are
    0:04:40 is actually as this intermediary layer
    0:04:41 between you and your computer,
    0:04:43 and it’s essentially a layer of abstraction
    0:04:48 on top of programming that allows us as regular people
    0:04:50 to be able to program our computers
    0:04:53 without even thinking about what we’re doing as programming.
    0:04:54 And so we think of our mission
    0:04:58 as essentially trying to reinvent the personal computer
    0:05:02 and really deeply empower everyone to be able to create
    0:05:04 in this computing medium of the future,
    0:05:06 because this digital medium
    0:05:08 is becoming more important in our lives.
    0:05:11 And what we want is actually not what I want,
    0:05:12 at least personally,
    0:05:14 is actually not a super centralized assistant
    0:05:16 that someone else has decided
    0:05:19 is able to do this, integrate with that.
    0:05:23 What I actually want is something that I can make and own
    0:05:26 that is mine, and that does what I wanted to do,
    0:05:27 kind of serves me.
    0:05:29 And today we’re kind of in a world
    0:05:31 where like all of our software is rented,
    0:05:32 it serves other people.
    0:05:33 So that’s kind of what we think of what agents are
    0:05:35 is this layer of abstraction on top of programming
    0:05:37 makes it so that it’s very intuitive
    0:05:39 for everyone to program.
    0:05:41 And that actually requires quite a bit of invention.
    0:05:43 So can get into that in historical context.
    0:05:46 – Yeah, well, the way you were,
    0:05:47 we’re describing it,
    0:05:50 I know you weren’t describing the sort of AI,
    0:05:52 the assistant, the agent layer,
    0:05:54 all the A-words, AI, assistant, agent,
    0:05:55 you weren’t describing the agent layer
    0:05:57 as replacement for a user interface.
    0:05:58 So you mentioned, you know, UI,
    0:06:00 but that’s kind of what I was thinking of,
    0:06:02 right, that abstractive layer that’s kind of in between.
    0:06:06 So how is imbue approaching it sort of on the ground
    0:06:09 and working with you or primarily with enterprise clients?
    0:06:12 – No, we primarily with, I would say it’s prosumer.
    0:06:14 So people who are,
    0:06:16 so the way to think about what we’re doing is,
    0:06:19 instead of thinking of agents as automation systems,
    0:06:20 right now we’re in kind of the agent,
    0:06:22 automation, agent paradigm,
    0:06:24 we think of agents as collaborative systems.
    0:06:27 So how do we enable a system that empowers me
    0:06:30 and helps me do more of what I want to do and work with?
    0:06:31 I can work with it.
    0:06:33 And so in the beginning,
    0:06:36 we’re actually, you know, enabling you to write code.
    0:06:38 Right now, these models are not that good at writing code.
    0:06:40 As they write code, you actually have to go check it.
    0:06:42 So you have to understand how to write code to go,
    0:06:44 to see, oh, did the agent do a good job?
    0:06:47 But over time, as these models get better at writing code,
    0:06:49 now you don’t have to check it anymore.
    0:06:50 And so in the beginning,
    0:06:52 we start with software engineers as the prime,
    0:06:54 or we call them software builders.
    0:06:55 So you don’t have to be an engineer.
    0:06:56 I’m a software builder.
    0:06:58 I’m not a software engineer.
    0:06:59 We start with software builders who can’t.
    0:07:01 – I’m a prototype builder then.
    0:07:02 I wouldn’t go as far as software.
    0:07:03 – Okay.
    0:07:06 So you could probably be a soon to be user.
    0:07:08 Once we get to a place where you don’t have to read
    0:07:10 and write, touch the code so much.
    0:07:12 Right now, we’re targeting software builders
    0:07:14 who can read and touch the code and be like,
    0:07:16 okay, well, it’s not quite right.
    0:07:17 We want to adjust it in this way.
    0:07:20 And over time, as the models get better,
    0:07:22 and you don’t have to be so low level in the code,
    0:07:24 now more and more types of creatives,
    0:07:27 types of builders can use these systems
    0:07:28 to really mold and shape their computer
    0:07:30 to what they want it to be.
    0:07:32 – And what level of complexity,
    0:07:35 what level of depth is the actual software
    0:07:38 that users are building within view?
    0:07:40 There’s the issue of accuracy, obviously,
    0:07:42 as you were saying they’re not,
    0:07:45 none of the models are creating 100% perfect code yet.
    0:07:48 But I also wonder how complex can these things get?
    0:07:50 And that kind of brings us to talking about reasoning.
    0:07:51 We don’t have to get there yet,
    0:07:54 but kind of edging towards that conversation.
    0:07:55 – Yeah, I think one of actually,
    0:07:58 our biggest learnings has been that if as a user,
    0:08:01 I have a code base that is fairly modular
    0:08:02 and I’ve kind of broken things down,
    0:08:04 then the model is actually pretty good at dealing
    0:08:06 with a very complex system,
    0:08:08 because it doesn’t have to load that much stuff
    0:08:10 into its head and it doesn’t have to cross check
    0:08:12 all of these dependencies.
    0:08:14 So just like with humans building software,
    0:08:16 where you don’t want a ton of dependencies,
    0:08:19 also, if you have a slightly more isolated system,
    0:08:20 it’ll do a better job.
    0:08:23 Similarly, there’s a lot of kind of user expertise
    0:08:25 in using this kind of product.
    0:08:27 So our product, it feels very collaborative.
    0:08:28 It’s almost like a document editor
    0:08:31 and kind of like interfacing and interacting
    0:08:32 with an agent in that way.
    0:08:35 And so as a user, you can basically,
    0:08:37 we learn to give it tasks,
    0:08:39 it’s more likely to succeed out.
    0:08:41 And we learn to structure our work
    0:08:43 so that we can delegate it to agents.
    0:08:45 And we’ve seen this with other AI tools as well,
    0:08:47 like Co-Pilot, our team definitely writes code
    0:08:48 in a slightly different way,
    0:08:50 so that Co-Pilot can work well for them.
    0:08:51 – Right, right.
    0:08:52 – To your question of complexity,
    0:08:54 it like really depends on the user,
    0:08:58 like some of us can make it work with really complex things.
    0:08:59 – Yeah, yeah.
    0:09:02 Where are you seeing agents being used the most
    0:09:04 or perhaps it’s more that they’re having
    0:09:06 a dramatic impact where they are being used
    0:09:10 and how does that translate into businesses,
    0:09:13 remaining competitive, having a competitive edge.
    0:09:15 Folks that I’ve talked to and I keep talking about,
    0:09:17 2022 last year,
    0:09:19 we’re kind of the year of the models coming out
    0:09:22 and mainstream taking notice of JNAI.
    0:09:23 And perhaps this year has been the year
    0:09:25 where people are trying to figure out,
    0:09:28 what apps do I build to leverage these things,
    0:09:30 as a software building business
    0:09:33 or as a business that does other things
    0:09:34 that wants to leverage this.
    0:09:37 So where is in view seeing impact being made
    0:09:40 or even looking to areas in the near future?
    0:09:42 – You know, the interesting thing about agents
    0:09:45 is it’s such an ill-defined term right now.
    0:09:47 And people are calling all sorts
    0:09:49 of very trivial things agents and that’s fine.
    0:09:51 But I think there’s a spectrum of kind of agent,
    0:09:53 usefulness, effectiveness.
    0:09:56 There’s like a system that scrapes the web
    0:09:57 and then aggregates the data
    0:09:58 and pulls the data out in some way.
    0:10:02 This is kind of like basically just a JNAI model.
    0:10:04 Like, you know, it’s kind of very similar to chatGPT,
    0:10:05 but you like put something on top of it
    0:10:08 to like dump the output into a different system.
    0:10:09 You could call that an agent.
    0:10:10 So some people call that an agent.
    0:10:13 We see that kind of thing being implemented
    0:10:14 in all sorts of places.
    0:10:16 But I think the much more exciting thing
    0:10:19 when it comes to agents is these more general agents
    0:10:22 that enable people to kind of start doing things
    0:10:25 that they previously didn’t even imagine that they could do.
    0:10:28 You know, I think like some of the really simple examples
    0:10:33 right now are for us, like some researcher or scientist,
    0:10:35 a biologist has a ton of data that they need to process
    0:10:37 and they’re not a software engineer,
    0:10:38 but they’re technical enough
    0:10:40 that they can kind of like pull in the data
    0:10:41 and then get something out of it,
    0:10:43 get something that helps us out of it.
    0:10:45 If they’re able to use something like this
    0:10:48 that lets them work at this slightly higher level
    0:10:50 or, you know, kind of over time,
    0:10:54 that a very exciting thing is as we start to build
    0:10:58 the tools that we need, like an example is my grandmother
    0:11:00 gets a bunch of scam calls in Chinese
    0:11:02 and, but all of her calls are in Chinese.
    0:11:04 And if I want to build a piece of software
    0:11:06 that filters out her scam calls from her other calls,
    0:11:08 like this is very hard right now,
    0:11:11 even for me as someone who knows how to build software.
    0:11:13 And it’s such a niche market.
    0:11:16 Like no one else is going to build that software for her.
    0:11:18 We’ve tried to find software like that in the US,
    0:11:19 doesn’t really exist.
    0:11:21 And so, exactly.
    0:11:24 So right now we’re in this world where software is built by–
    0:11:25 – Not to interrupt you.
    0:11:28 If it exists in the US for English language spam,
    0:11:30 it doesn’t work that well either for my–
    0:11:32 – Exactly, exactly.
    0:11:34 So, you know, right now we’re in this world
    0:11:37 where other people build software for us.
    0:11:39 We have to rely on other people to build software for us.
    0:11:41 And it’s actually really strange.
    0:11:44 Like we don’t really own our digital environments
    0:11:45 in that way.
    0:11:46 Like everything is kind of built by someone else
    0:11:49 because it’s too hard for us to like build or own things.
    0:11:52 And I think there is a future in which like I could actually
    0:11:54 pretty easily build something for my grandmother
    0:11:57 or for my community or for my group of friends
    0:12:01 or for my church to manage registrations or whatever it is.
    0:12:04 And that can be really tailored to my particular use case
    0:12:06 and to me, my community, my friends.
    0:12:09 And so, I think the really exciting thing about
    0:12:12 this future is like all of this like bespoke software
    0:12:14 as opposed to today where we have this
    0:12:17 kind of centralized software.
    0:12:19 It’s almost like people don’t often think
    0:12:20 of their digital environment in this way,
    0:12:23 but the digital environment is like the physical environment.
    0:12:26 And today it’s as if we all live in corporate housing.
    0:12:29 – I used to be so excited that I could listen
    0:12:32 to any music I wanted for, you know, 10 bucks a month.
    0:12:33 And now I’m thinking like,
    0:12:34 “But I don’t own any of it.”
    0:12:36 They could take it away for a minute or a second.
    0:12:39 – Yeah, and honestly, I think a lot of the kind
    0:12:41 of frustration people have about big tech, about technology
    0:12:44 is that we don’t feel like, and I don’t feel like
    0:12:46 I have control over these things
    0:12:48 that are a huge part of my life.
    0:12:51 And so that’s what at MBU, what we wanna do
    0:12:54 is give that control and power back to the people.
    0:12:57 And we do that by creating these tools and systems
    0:12:59 that collaborate with you to help you
    0:13:01 be able to create stuff for yourself.
    0:13:03 – So how hard is it to build these things
    0:13:06 for folks who use what, you know, as you mentioned,
    0:13:10 there are different, many different, you know,
    0:13:12 voices, individuals, companies,
    0:13:14 talking about the agent’s agentic AI,
    0:13:16 and a lot of them are defining it,
    0:13:18 talking about it at least slightly different.
    0:13:20 I’m sure taking, you know, different approaches,
    0:13:22 kind of under the hood.
    0:13:23 What are the challenges?
    0:13:25 What are the things that, you know,
    0:13:28 we can get a little bit technical here as you like.
    0:13:30 Some of the things, some of the problems that, you know,
    0:13:32 you and your teams are solving to make it easier
    0:13:35 for the rest of us to translate our ideas into software.
    0:13:37 – Yeah, so some problems I would say
    0:13:39 are kind of universal across all of these different types
    0:13:40 of people who are building agents.
    0:13:43 And some problems are unique to us
    0:13:44 and what we’re trying to do.
    0:13:47 So I would say, you know, most people are building agents
    0:13:49 in this kind of like workflow automation paradigm
    0:13:50 I mentioned earlier.
    0:13:52 And so for that paradigm,
    0:13:54 robustness, reliability is really important.
    0:13:57 Like, okay, you know, I built a thing
    0:13:59 that responds to customer service tickets.
    0:14:00 But if it 3% of the time,
    0:14:02 it says something really terrible to the user,
    0:14:04 like this is not a usable agent.
    0:14:07 For us, reliability and robustness is important,
    0:14:09 but it’s actually a little bit less important.
    0:14:12 As it gets better, the user experience just gets better.
    0:14:15 As a user, I don’t have to check stuff as much.
    0:14:18 But even if it’s not the best, it’s still okay.
    0:14:19 Like I can still use it as the user
    0:14:21 and I’ll like fix the bug that, you know,
    0:14:22 the model produced and that’s okay.
    0:14:24 So a lot of what we think of is kind of like,
    0:14:25 how do we get agents to meet
    0:14:28 both the model capabilities and also users
    0:14:29 where they are today.
    0:14:31 – So that expectation is built in
    0:14:35 that we’re not at the stage yet where it’s error-free.
    0:14:37 And as a user, you need to know that.
    0:14:39 And it’s not just like, okay, you have to accept that,
    0:14:41 but it’s actually like your experience
    0:14:43 is gonna wind up being better, right?
    0:14:44 ‘Cause you know you’re part in it.
    0:14:46 And it’s a, again, as you said,
    0:14:47 it’s not sending it off to do something
    0:14:50 and, you know, giving us the final result
    0:14:51 we had no part in.
    0:14:52 – Yeah, exactly.
    0:14:54 I think, you know, people often think of agents
    0:14:55 as a research problem, but we think of it
    0:14:58 as both a research problem and a user experience problem.
    0:15:00 And this user experience part
    0:15:02 is really about setting the right expectations
    0:15:04 so that with the experience,
    0:15:06 so that it’s like, I’m not expecting it to go off
    0:15:08 on its own for 10 minutes or 10 hours or 10 days
    0:15:10 and come back with something magically correct.
    0:15:12 Instead, I’m kind of iteratively working with it
    0:15:14 and seeing, oh, it’s kind of like clay, you know,
    0:15:16 I’m molding it, shaping it, shaping the output.
    0:15:19 I think the workflow automation agents,
    0:15:20 some of these agents kind of,
    0:15:23 they’re a little bit more, the bar is higher
    0:15:25 for how accurate they have to be
    0:15:28 because what we have found is that as a user,
    0:15:30 the longer it takes to come back with an answer,
    0:15:32 the more I expect the answer to be correct
    0:15:34 and I’ll get frustrated if it takes a really long time.
    0:15:37 So we’re very much on the, like, highly interactive,
    0:15:39 don’t take a super long time to come back with something,
    0:15:42 be an agent that really works with the user side.
    0:15:44 – Thinking about the idea of moving from,
    0:15:47 I’m expecting from a computer, right?
    0:15:50 The utmost inaccuracy, my calculator always says
    0:15:51 two plus two is four.
    0:15:54 Kind of moving from that to just a different frame of mind
    0:15:58 as an end user saying, okay, we’re prioritizing, you know,
    0:15:59 I don’t want to put words in your mouth,
    0:16:01 but kind of the speed and experience
    0:16:04 and you know, going in, it’s not gonna get them all right.
    0:16:05 Is that something that, you know,
    0:16:08 because this is the nature of AI and gen AI
    0:16:10 and people are kind of used to that,
    0:16:12 by now people are accepting of
    0:16:14 or is there still kind of a, I don’t know,
    0:16:16 maybe I’m just old, is there still kind of a mental hurdle
    0:16:19 to getting past the right, that expectation?
    0:16:21 – Yeah, so one of our core philosophies
    0:16:23 is that we need to meet people where they are
    0:16:25 in terms of what mental models we have
    0:16:26 in our heads as people.
    0:16:29 And so actually a good historical analogy
    0:16:31 is a back before the personal computer,
    0:16:33 people are really excited about the super computer.
    0:16:35 And the first personal computer first came out,
    0:16:36 everyone made fun of it.
    0:16:38 They were like, this is a hobbyist toy.
    0:16:40 And the super computer, you know,
    0:16:42 you accessed it with a terminal,
    0:16:44 you were time sharing on these super computers.
    0:16:45 It was not especially usable.
    0:16:48 And so a very small set of people were able to use it.
    0:16:51 But as time went on, a small group of people at Xerox Park
    0:16:53 actually invented a lot of these primitives
    0:16:56 that led the personal computer to be usable by people.
    0:16:58 They invented the desktop, files, folders.
    0:17:00 These are like concepts that we understood
    0:17:02 as humans at that time.
    0:17:04 And so for us, you know,
    0:17:06 part of actually building a good user experience
    0:17:08 around agents requires invention.
    0:17:11 It requires inventing concepts that match kind of like,
    0:17:13 are able to map this technology
    0:17:15 to what we currently understand as humans.
    0:17:16 So earlier I was saying,
    0:17:18 it’s kind of like a document editor right now,
    0:17:19 you know, our current product experience.
    0:17:22 And it may not ultimately be that way,
    0:17:25 but a document is something that I as a person today
    0:17:27 understand how to edit and work with.
    0:17:30 And it’s almost like an interactive editor
    0:17:31 that helps me accomplish my tasks.
    0:17:34 And to your question of how users receive it,
    0:17:36 one thing that’s really interesting that we observe
    0:17:40 is software builders, the feedback so far has been,
    0:17:42 wow, like this is really interesting.
    0:17:44 It lets me work at this higher level.
    0:17:47 I don’t have to dig so far into the code all the time.
    0:17:50 I can kind of stay and thinking at this higher level.
    0:17:53 And it’s actually able to go down into the code
    0:17:56 and surface to me, like do the task and then I check it.
    0:17:58 And that’s pretty cool.
    0:17:59 It like lets me move a lot faster.
    0:18:01 And that’s really interesting.
    0:18:03 You know, that for us, that’s kind of the primary thing.
    0:18:06 Like I want people to be able to work
    0:18:11 at our human problem level of abstraction with software
    0:18:14 instead of having to get so deep into the weeds.
    0:18:15 Yeah, yeah.
    0:18:19 No, I can relate from the standpoint
    0:18:23 of learning how to use these tools when I’m writing,
    0:18:27 when I’m not on the podcast, asking people like you questions.
    0:18:29 You know, a lot of my work is writing.
    0:18:32 And if I’m working from a document, a transcript,
    0:18:34 source materials, it’s that same thing
    0:18:38 when I can use the tool to just kind of surface to me.
    0:18:40 You know, did anywhere in the document
    0:18:45 or in the transcript at CanJune talk about pineapple on pizza?
    0:18:48 You know, and just being able to surface that back, right?
    0:18:50 It saves all the time of going through the document.
    0:18:51 I don’t need the exact words.
    0:18:53 I don’t need it 100%.
    0:18:54 Then we’ll get to that later,
    0:18:56 but I can just kind of go back and check.
    0:18:57 Like, oh, right.
    0:19:00 She said she is, isn’t a fan of Pepperoni.
    0:19:00 You know, yeah.
    0:19:03 And it’s, it’s, it’s incredibly helpful.
    0:19:04 You know, it’s not just a saving time,
    0:19:05 but I think you said it really well.
    0:19:08 It allows you to stay at that level of thinking.
    0:19:09 Exactly. Yeah.
    0:19:11 I think our core, the thing I really care about
    0:19:14 is helping people be able to be creative in this way.
    0:19:17 And often there’s such a barrier between our ideas
    0:19:18 and the execution of those ideas
    0:19:20 that we never get to do a lot of things
    0:19:22 that we really want to do with our lives.
    0:19:24 And so I think the like true power of computing
    0:19:26 as a medium hasn’t really been unleashed yet
    0:19:28 and we want to unleash it.
    0:19:30 And what that looks like is that people are able
    0:19:31 to kind of like take their ideas.
    0:19:33 Like you can take your ideas about writing this piece
    0:19:35 or aggregating all of the pieces you’ve written
    0:19:37 and being able to draw insights out of them
    0:19:40 in order to create your book, like take these ideas
    0:19:42 and actually be able to work with them at this higher level
    0:19:44 so that you’re not always bogged down in the weeds.
    0:19:46 And I think the true power of AI,
    0:19:48 the true power of computers, like that’s what it enables.
    0:19:50 And we’re not there yet, but we can get there.
    0:19:53 And it’s not just about automation or business automation,
    0:19:54 business workflow automation.
    0:19:55 Right, right.
    0:19:58 Now, what was the, in your conversation with Brian from GTC,
    0:19:59 what was it that he said?
    0:20:01 I’ve got all these ideas and I sit down to code
    0:20:04 and I’m like import, what do I want to import, right?
    0:20:06 And you’re, I think that was great
    0:20:07 ’cause you’re derailed immediately
    0:20:10 and I can relate in the work that I do that, you know, yeah.
    0:20:11 Yeah, 100%.
    0:20:14 Yeah, one of our users said, wow,
    0:20:16 I never realized how much context switching I do
    0:20:18 when I’m writing code from high to low level.
    0:20:21 Same with when you’re writing normal stuff.
    0:20:23 I’m seeking with CanJune Q.
    0:20:27 CanJune is co-founder and CEO of Imbue
    0:20:31 and we have been talking about Imbue’s work on AI agents
    0:20:35 that help people code and really fascinating approach
    0:20:36 that, you know, as we’ve been talking about,
    0:20:39 I think goes beyond just expressing in code,
    0:20:42 but code being the way that we interface with computers
    0:20:44 and get them to do the things we want them to do.
    0:20:47 Well, I want to ask you about AI models and reasoning.
    0:20:51 And then I also want to ask you sort of about scale
    0:20:56 and what goes into, you know, building an agent for yourself
    0:20:59 and then what goes into building agents and multiple agents
    0:21:02 and agents that collaborate for users at scale.
    0:21:04 Is there an order we should go in?
    0:21:06 Should we talk about reasoning first?
    0:21:08 Is there a relation?
    0:21:08 That’s interesting.
    0:21:10 Let’s talk about scale first.
    0:21:11 Okay, cool.
    0:21:14 Yeah, so one way that people think about scale with agents
    0:21:17 is a lot of agents interacting with each other
    0:21:18 and kind of what that looks like.
    0:21:20 And some people do that by giving different agents
    0:21:23 different prompts so they have different personalities
    0:21:23 and things like that.
    0:21:28 And honestly, I think that’s a little bit, it’s interesting.
    0:21:32 It’s a little bit limited because we already have agents today.
    0:21:34 All software is agentic.
    0:21:37 The whole point of what an agent is is it something
    0:21:40 that takes action and it uses your computer
    0:21:42 to kind of like execute something.
    0:21:44 And so almost all software is executing something.
    0:21:47 It’s kind of like changing the state of your computer,
    0:21:49 website, data, et cetera.
    0:21:51 Now, the difference between most software
    0:21:54 and like AI agents, what we call AI agents today
    0:21:56 is that AI agents can process stuff
    0:21:58 in a way that’s like not fully deterministic.
    0:22:03 But even so, we still had AI agents in Facebook news feed,
    0:22:05 for example, recommendation engine is an agent
    0:22:09 that non-deterministically decides what to show you.
    0:22:11 So we’ve had agents since forever, since we had software.
    0:22:13 And so, you know, kind of the way I think about scale
    0:22:15 and agents is actually about,
    0:22:17 is the same as the way I think about scale of software.
    0:22:19 So in the future, in the next 10 years,
    0:22:21 I think there’s gonna be this explosion of software
    0:22:24 and the software is gonna be slightly less hard coded
    0:22:24 than before.
    0:22:27 It’s gonna be able to work with more empty US input.
    0:22:29 It might be more interactive.
    0:22:31 Hopefully a lot of people are going to be able
    0:22:33 to create it if we succeed.
    0:22:37 And so now we end up with this giant ecosystem,
    0:22:40 like world of software that’s far beyond what we have today.
    0:22:43 And in that world, what happens is now you have a lot
    0:22:46 of different automated systems interacting with each other.
    0:22:49 And this is actually, could be super exciting.
    0:22:52 Every person could have their own automated systems
    0:22:54 that do things for themselves, for their lives.
    0:22:57 They have, you know, I’m supported by the software
    0:22:59 that surrounds me, as opposed to today
    0:23:01 maybe being bothered by it.
    0:23:02 (laughing)
    0:23:05 – I was listening to you and I was thinking of how to phrase,
    0:23:07 how to try to phrase the question,
    0:23:10 getting back to your point about, you know, sort of,
    0:23:13 I was thinking of it as sort of one size fits all software,
    0:23:15 you know, that’s deterministic and it does what it does
    0:23:18 versus, you know, me being able to create
    0:23:20 and reshape things as we go.
    0:23:21 And you answered it for me.
    0:23:22 So that’s that.
    0:23:23 – Oh, that’s great.
    0:23:26 I love this, I love this one size fits all software today
    0:23:28 versus future bespoke software.
    0:23:29 That’s a great term.
    0:23:30 – Are you worried at all about,
    0:23:33 I think, I don’t know if the term AI slop applies to code,
    0:23:38 but this idea of, you know, AI models creating texts
    0:23:41 that’s kind of meaningless or value-less,
    0:23:44 but it’s being automatically put out onto the web and stuff
    0:23:47 for a very sort of, you know, relatively ignorant point
    0:23:51 of view, the notion of model-generated software
    0:23:55 that can run itself being deployed out on the public web,
    0:23:57 you know, is a little scary to me,
    0:23:59 but also I’m sure there’s stuff out there,
    0:24:01 but how do you think about that?
    0:24:05 – Yeah, the way I think about kind of AI slop for software
    0:24:09 is automated systems that like infringe on us.
    0:24:12 So scam calling or spam calling is a good example
    0:24:14 of an automated system that infringes on us,
    0:24:16 or like apps that have lots of notifications
    0:24:19 or, you know, games that are meant to be extractive.
    0:24:21 Those are our systems that infringe on us.
    0:24:24 And, you know, I actually think the default path
    0:24:27 of where things are going with centralized AI systems
    0:24:30 and the kind of like returns the scale
    0:24:33 of improvements of the underlying models
    0:24:35 is that we will kind of, as humans,
    0:24:38 be a little bit disempowered and be holding
    0:24:41 to whoever controls the automated systems.
    0:24:43 It’s not necessary, you know, I’ve kind of told you
    0:24:45 about this beautiful vision of a future
    0:24:46 where everyone can create software
    0:24:47 and everyone creates bespoke software,
    0:24:49 and like that’s the future we want to build,
    0:24:51 but it’s not necessarily the future
    0:24:53 that is the default path.
    0:24:56 The default path could be that there’s a lot of,
    0:24:58 like even more software out there today
    0:24:59 that’s trying to get our attention
    0:25:00 and trying to extract from us.
    0:25:03 And what we need, I think, what I want
    0:25:05 is for people to create defenses
    0:25:07 and to kind of like get rid of that stuff
    0:25:08 and disrupt it.
    0:25:11 You know, hopefully when I can build my own software,
    0:25:13 I can actually disrupt a lot of the stuff
    0:25:14 that’s being built today,
    0:25:16 so that I have software that’s serving my own interests.
    0:25:18 I have agents that are serving my own interests
    0:25:20 and helping me do what I want to do.
    0:25:21 So to your question of AI slop,
    0:25:24 I think there’s definitely going to be people
    0:25:25 making more automated systems
    0:25:28 that are bots and bother people.
    0:25:30 And just like in cybersecurity,
    0:25:32 I think there’s an attack defense dynamic,
    0:25:34 and what we want is to enable people
    0:25:36 to create opposing systems for themselves
    0:25:38 that like help defend their environment and themselves
    0:25:41 and help kind of protect us to do what we want
    0:25:43 and liberalize we want.
    0:25:45 And hopefully there’s also kind of some,
    0:25:47 you know, there’s already some regulatory aspect of this
    0:25:50 that exists and, you know, there hopefully will be more
    0:25:51 in response to what we see.
    0:25:53 So that effect is real.
    0:25:54 – Right, all right.
    0:25:59 There’s been a lot in the generative AI news cycles.
    0:26:01 There’s a thing, that’s a thing.
    0:26:04 This year in particular about models and reasoning
    0:26:06 and future models, models being trained
    0:26:08 that have reasoning capacity, that kind of thing.
    0:26:12 Recently open AI launched a new iteration of a model
    0:26:13 talking about reason capabilities.
    0:26:18 Is reasoning something that does or can happen in an LLM?
    0:26:21 Is it something that the agents bring to the table,
    0:26:22 so to speak?
    0:26:23 How do you think about reasoning?
    0:26:27 How does in view approach building reasoning capabilities
    0:26:28 into your products?
    0:26:29 – Yeah, that’s a great question.
    0:26:31 So yeah, reasoning is a buzzword right now
    0:26:34 and models definitely do reasoning.
    0:26:38 That’s, you know, and in the way that like underlying LLMs
    0:26:40 definitely do reasoning in the way that humans kind of,
    0:26:43 it’s not exactly how we do reasoning maybe,
    0:26:44 and it’s not always perfectly correct.
    0:26:47 And it’s often continuing to justify its own reasoning,
    0:26:48 although humans do that too.
    0:26:50 – Now he’s gonna say, that’s familiar.
    0:26:53 – So unclear how similar or different it is to humans,
    0:26:56 but the underlying LLM definitely does some reasoning.
    0:26:58 One key difference that we observe right now
    0:27:01 is that the underlying LLM isn’t necessarily
    0:27:05 as good at verifying if its own answer is correct.
    0:27:08 And as a person when I’m doing a task,
    0:27:09 I actually, we don’t notice this,
    0:27:11 but we’re always constantly checking like,
    0:27:12 is this right?
    0:27:12 Did I do that right?
    0:27:13 Is this what I expected?
    0:27:14 And there’s not that loop.
    0:27:17 So that loop is kind of added by the agentic piece.
    0:27:18 – Okay.
    0:27:20 – Yeah, we think a lot actually about our research direction
    0:27:24 as around this kind of verifying verification.
    0:27:24 Is it correct?
    0:27:25 Did I do it right?
    0:27:27 And if I have an agent that’s writing code for me,
    0:27:29 I do want it to check itself like,
    0:27:30 hey, did I do that right?
    0:27:31 I did.
    0:27:32 Oh, I didn’t.
    0:27:33 Let me fix this mistake and then come back to you.
    0:27:36 And so the better it is at verifying its own answers,
    0:27:39 the more like a better user experience it is.
    0:27:41 And so, when we talk about reasoning,
    0:27:43 we mostly talk about this kind of like verification
    0:27:44 and robustness.
    0:27:47 Is it able to verify what it’s doing?
    0:27:49 We’ve actually learned some pretty interesting things
    0:27:52 when working on verification around kind of,
    0:27:55 it turns out that, so in software development,
    0:27:56 when you write a function,
    0:27:58 you also often write software tests.
    0:28:00 You’re testing, okay, did the software
    0:28:01 have the behavior expected?
    0:28:03 And given really good tests,
    0:28:05 actually the underlying models are pretty good
    0:28:08 at treating the function or treating the piece of software.
    0:28:09 – Right.
    0:28:10 – But given the piece of software,
    0:28:12 the underlying models are not very good
    0:28:13 at treating good tests.
    0:28:16 Which is kind of interesting.
    0:28:17 – Yeah.
    0:28:18 – One.
    0:28:19 – Any idea why?
    0:28:21 – Yeah, one, you know, it’s partly because the models
    0:28:23 are probably not trained that much on this particular task
    0:28:24 of creating tests.
    0:28:27 Two, though, maybe it’s possible.
    0:28:29 We don’t know, we’re not 100% sure,
    0:28:31 but it’s possible that actually verifying
    0:28:34 if something is correct is a harder reasoning problem
    0:28:37 than kind of creating the thing in the first place.
    0:28:38 – Yeah.
    0:28:40 – So it kind of requires this like analysis and judgment.
    0:28:43 And so our research direction is primarily focused
    0:28:45 on verification, how do we get to models
    0:28:47 that actually are able to properly verify
    0:28:49 that the output is correct and kind of is
    0:28:50 what the user wanted.
    0:28:53 And we think of that as the hard problem
    0:28:54 in reasoning for agents.
    0:28:58 – At the same time, MBU has, did you pre-train a model?
    0:28:59 Did you build a foundation?
    0:29:01 You didn’t build the model from scratch,
    0:29:03 but it’s a 70 billion parameter model.
    0:29:04 – That’s right, we actually did pre-train
    0:29:06 a 70 billion parameter model from scratch.
    0:29:08 – It is from scratch, okay, one mistake.
    0:29:10 – Yeah, we actually learned a ton from that process.
    0:29:12 And one of the things we learned was like,
    0:29:14 actually we don’t know if we need to do tons
    0:29:15 of pre-training going into the future.
    0:29:16 We’ll see. – Yeah.
    0:29:20 – But we got a lot out of post-training on that model.
    0:29:23 And so for a lot of the verification work,
    0:29:26 we’re actually really interested in post-training,
    0:29:27 fine-tuning, reinforcement learning
    0:29:29 on top of the underlying models.
    0:29:31 – That seems like a good place to ask this question.
    0:29:33 What does the future look like?
    0:29:35 I almost want to leave it just there, but that’s not fair.
    0:29:37 What does the future of AI agents look like?
    0:29:38 What is MBU’s approach?
    0:29:41 Do you have a roadmap you can share any of?
    0:29:42 – Where is this heading?
    0:29:44 And I know that’s an impossible question to answer
    0:29:46 in many ways, but I’m also guessing
    0:29:48 you have something of a vision, so.
    0:29:49 – Yeah, that’s a great question.
    0:29:52 So I talked today about trying to enable everyone
    0:29:55 to build software, but really internally,
    0:29:57 the way we think about it is all software
    0:29:59 in the future will be agents, basically.
    0:30:01 I mean, not all software, but most software.
    0:30:03 It’ll be a little bit smarter,
    0:30:04 kind of just like living software.
    0:30:07 And what we want to do is enable everyone
    0:30:10 to be able to build agents so that in the world,
    0:30:12 we’re all able to build our own agents for ourselves
    0:30:13 or use each other’s copy,
    0:30:16 someone’s modify a little bit from myself.
    0:30:18 And that’s kind of what our product
    0:30:19 is meant to do over the long term.
    0:30:23 And so actually in our 70 billion parameter model,
    0:30:26 we released a set of blog posts that taught people
    0:30:29 how to set up infrastructure to train such models.
    0:30:32 We expect most people, most of us won’t train our own models,
    0:30:34 but it’s kind of part of this desire
    0:30:36 to democratize a lot of these capabilities.
    0:30:40 We also released a toolkit for people doing evaluations
    0:30:43 of their models and with clean data and everything.
    0:30:46 And so in terms of kind of what’s the future
    0:30:49 of building agents, my hope is that agent building,
    0:30:50 unlike software building, is not something
    0:30:53 that only a few people are able to do well.
    0:30:54 My hope is that there’s actually something
    0:30:56 that’s like widely democratized
    0:30:58 and where everyone is empowered
    0:31:00 to be able to create their own agents.
    0:31:05 And I think right now we have this very scary view
    0:31:07 of somebody else is going to create a system
    0:31:09 that automates my job.
    0:31:09 And that sucks.
    0:31:11 That’s like really disempowering.
    0:31:13 I don’t want that for my job.
    0:31:14 – Yep.
    0:31:17 – But the thing I love doing is automating parts
    0:31:18 of my own job.
    0:31:19 – Yeah.
    0:31:20 – I like, you know, love making it better
    0:31:21 in all of these different ways.
    0:31:23 And that’s what we want to enable people to be able to do.
    0:31:26 Like by giving you the tools to make your own agents,
    0:31:28 that means that you can make your own things
    0:31:29 that automate parts of your job.
    0:31:32 And now your job can be higher leverage and higher level.
    0:31:33 And now you can do a lot more.
    0:31:34 – Right.
    0:31:35 – And so we want to give kind of, you know,
    0:31:38 someone else automating my job is very disempowering to me,
    0:31:39 but someone giving me the tools
    0:31:41 so that I can make my own tools for myself.
    0:31:43 That’s very empowering for me.
    0:31:44 And I think this mentality shift
    0:31:46 is actually really important.
    0:31:47 – Amen.
    0:31:49 Ken June for folks listening
    0:31:51 who would like to learn more about imbue.
    0:31:54 You mentioned a blog as well.
    0:31:55 Where should they start online?
    0:31:58 Website, social media, podcast.
    0:31:59 There’s a podcast, I think.
    0:32:00 Where should they go?
    0:32:00 – Great question.
    0:32:03 So imbue.com is where we are on the internet.
    0:32:06 And you can follow our Twitter account imbue.ai.
    0:32:08 And we will have, you know,
    0:32:11 as we start to release the product a little bit more
    0:32:13 publicly, we’ll probably have announcements
    0:32:16 and things where you can start experimenting
    0:32:17 with what we’re doing.
    0:32:18 So please do follow us.
    0:32:20 There’s also a newsletter sign up
    0:32:22 where we send extremely rare emails.
    0:32:24 (laughing)
    0:32:27 Because we mostly focus on building.
    0:32:30 – You’re not an application trying to extract constantly?
    0:32:32 – No, no, not trying to get your attention,
    0:32:33 trying to make a useful product.
    0:32:34 – Good, good, good.
    0:32:35 Ken June, this is delightful.
    0:32:38 Thank you so much for taking the time to come on the pod.
    0:32:39 And best of luck with everything you’re doing.
    0:32:41 Maybe we can check in again down the road.
    0:32:42 – Definitely.
    0:32:43 Thank you, Noah.
    0:32:44 This is super fun.
    0:32:46 (upbeat music)
    0:32:49 (upbeat music)
    0:32:52 (upbeat music)
    0:32:54 (upbeat music)
    0:32:57 (upbeat music)
    0:32:59 (upbeat music)
    0:33:02 (upbeat music)
    0:33:04 (upbeat music)
    0:33:07 (upbeat music)
    0:33:10 (upbeat music)
    0:33:12 (upbeat music)
    0:33:15 (upbeat music)
    0:33:17 (upbeat music)
    0:33:20 (upbeat music)
    0:33:22 (upbeat music)
    0:33:25 (upbeat music)
    0:33:28 (upbeat music)
    0:33:30 (upbeat music)
    0:33:33 (upbeat music)
    0:33:43 [BLANK_AUDIO]
    Vietnamese translation content goes here.

    In this episode of the NVIDIA AI Podcast, Kanjun Qiu, CEO of Imbue, explores the emerging era where individuals can create and utilize their own AI agents. Drawing a parallel to the personal computer revolution of the late 1970s and 80s, Qiu discusses how modern AI systems are evolving to work collaboratively with users, enhancing their capabilities rather than just automating tasks.

  • How AI Can Help Boost Disability Inclusion – Ep. 238

    AI transcript
    0:00:10 [MUSIC]
    0:00:13 >> Hello, and welcome to the NVIDIA AI podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:19 A study released this past July of 2024 by
    0:00:22 the Special Olympics Global Center for Inclusion and Education,
    0:00:26 found that 64 percent of educators and 77 percent of
    0:00:29 parents of students with intellectual and developmental disabilities,
    0:00:31 IDD as they’re known,
    0:00:33 view artificial intelligence as
    0:00:36 a potentially powerful mechanism to promote more inclusive classrooms,
    0:00:41 and close educational gaps between students with and without IDD.
    0:00:45 But only 35 percent of educators surveyed believe that
    0:00:48 developers of AI currently account for
    0:00:50 the needs and priorities of students with IDD,
    0:00:55 pointing to the need for the creation of more disability inclusive tools.
    0:01:00 More recently, the G7 held its first ever meeting on disability in Umbria,
    0:01:05 Italy just a few weeks ago as we record this in October of 2024.
    0:01:07 The event aimed to bring greater attention and
    0:01:10 action to disability issues globally,
    0:01:14 with a major focus on assistive technology and AI.
    0:01:18 Our two guests today have been involved in working with AI and
    0:01:22 the potential of AI in special education for some time now.
    0:01:25 One of them was involved with a survey I mentioned,
    0:01:28 and the other led the US delegation to the G7 meeting.
    0:01:32 They’re here to talk about the potential for using AI in special education,
    0:01:36 and some of the things we need to do to best realize that potential.
    0:01:39 Sarah Minkara is the Special Advisor on
    0:01:42 International Disability Rights at the US Department of State.
    0:01:46 Special Advisor Minkara leads the United States comprehensive strategy
    0:01:49 to promote and protect the rights of persons with disabilities,
    0:01:52 internationally, and across our foreign policy.
    0:01:56 Dr. Timothy Schreiber is the Chairman of the Board of Special Olympics,
    0:02:01 where together with six million Special Olympics athletes in more than 200 countries,
    0:02:03 he works to promote health, education,
    0:02:06 and a more unified world through the joy of sport.
    0:02:12 Sarah, Tim, thank you so much for making the time to join the NVIDIA AI podcast.
    0:02:12 Welcome.
    0:02:14 Thank you.
    0:02:15 Thank you, thank you.
    0:02:18 I covered only just the tip of the iceberg of
    0:02:21 all the work that you both have been doing throughout
    0:02:26 your careers and your lives to further the cause of folks
    0:02:30 with special educational needs and disabilities.
    0:02:33 But I only scratched the surface in that intro,
    0:02:35 and I’d much rather, the audience would much rather hear from the both of you.
    0:02:39 So maybe we can start and Sarah, I’ll ask you to go first if you don’t mind.
    0:02:43 Just tell us a little bit about what you do,
    0:02:47 what you work on, and then we can roll into the whole AI technology part.
    0:02:49 But let’s just start with what you do.
    0:02:51 Definitely. So my team and I,
    0:02:55 we lead the US government foreign policy on disability.
    0:02:56 What does that really mean?
    0:03:00 As we enter the foreign policy conversations,
    0:03:03 dialogues where we focus on global peace, security, prosperity,
    0:03:07 where we focus on trade, economic development, AI,
    0:03:09 climate change, food insecurity,
    0:03:12 how do we ensure that disability is integrated,
    0:03:15 mainstream, and part of those dialogues and conversations?
    0:03:19 My office, we cover the globe when it comes to disability forum policy,
    0:03:22 but the work is still very much,
    0:03:24 there’s so much still that needs to be done.
    0:03:28 How do we ensure that once a society looks at disability and foreign policy,
    0:03:32 it’s not seen from a charity line, it’s not seen from a pity line,
    0:03:34 but it’s seen from a value-based lens?
    0:03:37 To really understand, when you include people with disability into your economy,
    0:03:39 it’s going to help your GDP up to 7%.
    0:03:43 When you bring people with disabilities into the system, society, and innovation,
    0:03:48 we will be able to bring so much more innovation and benefit to the technology world.
    0:03:51 You cannot achieve full recovery, reconstruction,
    0:03:55 that’s accessible for everyone if you don’t bring in the disability line.
    0:03:58 So our goal is, we need to ensure that the civility is at the table
    0:04:02 and that we’ll never be able to reach full global peace and prosperity and security
    0:04:06 if you don’t bring in 1.3 billion individuals in this world with a disability to the table.
    0:04:07 Excellent. Tim?
    0:04:10 The Special Olympics movement is just over 50 years old.
    0:04:16 We were started in the 1960s when the norm for a person with an intellectual disability,
    0:04:20 let’s say someone with Down syndrome or a comparable kind of challenge,
    0:04:22 the norm would have been an institutionalized life,
    0:04:27 often begun at birth and then for their whole lives relegated to living
    0:04:31 in a congregate care setting that often was included subhuman conditions.
    0:04:33 The reason I mention that is because the movement has,
    0:04:37 the world has changed a great deal in some ways for the better, but the movement hasn’t.
    0:04:42 Our movement has always been focused on how do we use the power of sport
    0:04:47 to heal these wounding attitudes, to heal these discriminatory attitudes,
    0:04:52 to heal this sense in which cultures tell people with intellectual and developmental challenges,
    0:04:55 you don’t matter, you don’t belong, you can’t fit in.
    0:05:00 So our goal is to, yeah, we play soccer, we run track meets and all these kinds of things,
    0:05:04 but we also try to put sports at the center of conversations in schools
    0:05:09 about how can young people, three, four, five, six, seven year olds,
    0:05:13 how can they learn together with their peers with intellectual disabilities,
    0:05:17 learn how to play together and all of a sudden you’re starting to learn how to learn together
    0:05:19 and then ultimately you learn how to live together.
    0:05:21 So we’re in 190 countries around the world.
    0:05:28 We do, you know, in an average year about 60,000 games a year at the community level,
    0:05:34 you know, single sport, basketball, games, bowling, tournaments, spring games, summer games,
    0:05:39 softball challenges, you name it, all over the world, villages, small towns.
    0:05:44 And in every one of them, it’s an invitation to see the potential of a person
    0:05:49 with an intellectual disability to fit in, but equally, as Sarah has just pointed out,
    0:05:53 to have the culture, the community see that without including everyone,
    0:05:59 you’re selling yourself short, you’re actually penalizing everyone by excluding anyone.
    0:06:00 Couldn’t agree more.
    0:06:05 Sarah, before we get started into talking about the summit and kind of the larger role
    0:06:10 of AI in education, could you, would you like to talk a little bit about your own experience
    0:06:16 as a woman with a disability and how that’s kind of impacted both your entrepreneurial
    0:06:18 and public service work?
    0:06:24 Yes, you know, disability is a big part of, of course, my identity, my lived experience.
    0:06:29 I became blind when I was seven, my sister also became blind, but she was seven years old.
    0:06:31 And because of two reasons.
    0:06:34 One, because of our laws and legislations and policies in our country here in the U.S.
    0:06:40 and because of my parents who did not allow society’s expectations or lack thereof
    0:06:44 on disability to ever enter the home, because that narrative, when it comes to disabilities,
    0:06:46 you can’t, you can’t, you can’t.
    0:06:48 And that’s a very common narrative when it comes to disability.
    0:06:54 Instead, my mom really pushed us to have ambition streams and really taught us,
    0:06:56 let’s work on breaking down barriers so we can reach your potential.
    0:07:01 But then growing up, I realized most kids with disabilities all over the world don’t have
    0:07:03 that same privilege and it shouldn’t be a privilege.
    0:07:09 It should be a reality, but don’t have that, you know, that, that reality of being seen,
    0:07:10 being heard and being valued.
    0:07:12 And I think that’s what Special Olympics is also doing.
    0:07:18 For me, that lived experience really then influenced my trajectory.
    0:07:21 I, yes, I was a math and econ major and undergrad, but I ended up actually then starting
    0:07:27 a nonprofit organization, working in countries like Lebanon to really work with youth
    0:07:32 with disabilities, so to creating an empowering system for them to be integrated and seen
    0:07:33 and heard and valued.
    0:07:38 And that led to then my, my company, and then that led to this role in government.
    0:07:42 So ultimately, my, my entire life has really stemmed from the fact that I had parents who
    0:07:46 really, who really gave me the opportunity to live my whole life, my full identity and
    0:07:50 embrace my identity as a blind person with pride.
    0:07:51 Makes such a difference, right?
    0:07:52 It’s a lottery.
    0:07:57 The parents we, we wind up with, but glad to hear yours have been supporting you in
    0:07:59 all the best ways it sounds like.
    0:08:05 One of the things I always talk to people about is how do we partner up with folks to
    0:08:08 break down barriers so people with disability can be fully integrated?
    0:08:09 Right?
    0:08:12 It’s almost like people always look at like, oh, we help people with disabilities.
    0:08:15 No, no, it’s actually we help society by breaking down barriers that people with
    0:08:18 disabilities can contribute and that becomes a benefit for everyone.
    0:08:21 And that’s an important thing that we always started to talk about.
    0:08:24 Policies can help break down barriers, which then will allow for society to
    0:08:26 benefit from our contribution.
    0:08:30 I, I appreciated to that point in kind of your, uh, your answer to my opening
    0:08:34 question about, tell us about, you know, who you are and the work that you do, that
    0:08:38 you framed it that way from the beginning, that look, there are all these people who
    0:08:40 you’re choosing to exclude.
    0:08:44 And if you would include them, your bottom line is going to go up, like for starters,
    0:08:48 you know, and yeah, so let’s talk about the role of technology.
    0:08:52 Let’s talk about the potential and some of the potential pitfalls and how we kind
    0:08:54 of work around them to use technology to help.
    0:08:57 We talked about our education way back in the day.
    0:09:04 I tried my hand at classroom teaching and, uh, I know firsthand how hard classroom
    0:09:08 teachers work, not because of me, because of the people I saw, you know, I found my
    0:09:12 other path, but the promise of technology and education has been around for as
    0:09:13 long as technology and education.
    0:09:14 Right.
    0:09:18 Uh, but with AI, there’s a lot of talk about personalized learning, dynamic
    0:09:19 learning, all of these sorts of things.
    0:09:24 And when we talk about AI kind of writ large and generative AI, there’s a lot
    0:09:28 of talk, especially more recently, which is great on bias on things like what kind
    0:09:30 of data was the model trained on?
    0:09:32 Where’s the data coming from?
    0:09:33 Who is it representative of?
    0:09:35 Who is it not representative of?
    0:09:37 All of these factors, right?
    0:09:43 Nearing in on the world of IDD, special education, what are the benefits?
    0:09:47 What are some of the risks that are presented by AI for persons with disabilities?
    0:09:51 Well, I can jump in and maybe summarize, at least at a high level, what we are
    0:09:55 seeing in the Special Olympics movement when we talk to people across the globe,
    0:09:58 whether it’s the Northern Hemisphere, the Southern Hemisphere, or different cultures.
    0:10:04 The first thing is that in schools, personalized instruction is fantastically
    0:10:06 helpful for people with intellectual and developmental disabilities.
    0:10:08 It’s a game changer.
    0:10:12 One of the primary obstacles to inclusion is the absence of the teacher’s
    0:10:15 capacity to devote the time for individualized instruction.
    0:10:19 She or he has a group of 20 or 30 or 40 or 50 children.
    0:10:24 So AI can be a game changer for personalizing the learning experience of
    0:10:25 children with intellectual disability.
    0:10:30 Number two, we know that when children with intellectual disabilities are included,
    0:10:31 everybody wins.
    0:10:35 Bullying goes down, graduation rates go up, test scores go up.
    0:10:39 Can you imagine if people think if you have a child with intellectual
    0:10:41 disability in my kid’s class, my kid’s not going to learn.
    0:10:42 It’s the opposite.
    0:10:43 Your kid’s going to learn more.
    0:10:45 So this will benefit all kids.
    0:10:49 The third thing we can say about AI is that we’re not at the table right now.
    0:10:52 We’re not at the table as people are building the models.
    0:10:56 We’re not at the table as people are thinking about the application of the
    0:10:58 models for making life better.
    0:11:03 We’re not at the table listening to the needs and gifts of people with
    0:11:07 intellectual challenges, their families, their friends, their brothers, their sisters.
    0:11:11 So we’ve got this tool with enormous potential.
    0:11:12 We can see it.
    0:11:13 Our parents can see it.
    0:11:16 Overwhelming numbers, 70, 80, high 80 percent.
    0:11:18 Yeah, we want this for our kids.
    0:11:22 Yes, we want our kids to learn it, but we’re not there now, which is why, if I
    0:11:26 can say this in a crass way now, I’m so excited to be talking to your audience
    0:11:31 because your audience and your community is on the inside.
    0:11:33 You’re behind the, you know, for us, you’re behind the curtain.
    0:11:35 You’re the guys doing the real work.
    0:11:37 You know, you’re making it all happen.
    0:11:40 And we just, all we want to say is, hey, look, we’re raising our hands over here.
    0:11:44 We want to help you, but we also want you to help us, right?
    0:11:46 And so what does that look like?
    0:11:48 Do you have a vision, the beginnings of visions?
    0:11:53 What’s the, um, yeah, I mean, we have, like, you know, we were talking the other day.
    0:11:55 Well, what’s our exactly description?
    0:11:57 We don’t have a strategy because it’s too early.
    0:12:01 Honestly, we don’t have, and maybe Special Olympics, we’re much more of a grassroots movement.
    0:12:03 So strategy sort of often emerges from the field.
    0:12:09 But look, I could see a time when an AI enabled Special Olympics coach can do
    0:12:10 what we can’t do today.
    0:12:14 We have three, four, three COVID, six million athletes.
    0:12:15 They don’t all have a coach.
    0:12:19 And there’s another six million on the sidelines because we can’t find them a coach.
    0:12:24 Their coaches don’t have the time to help them with nutrition and fitness, with hygiene
    0:12:28 and functional skills, with tracking performance and developing skills.
    0:12:35 And AI, the so-coach, could be that companion that doesn’t replace human contact, but
    0:12:38 supplements and compliments human contact.
    0:12:43 Think of that just in the context of physical activity, play, social engagement.
    0:12:48 But then think about that same kind of coach, if you will, in classrooms.
    0:12:55 Think about that coach who’s helping people manage behavior issues, emotion, self-regulation.
    0:12:58 How do I find time, like, when I need to seek help, how do I do it?
    0:13:00 The AI coach could be right there.
    0:13:03 Hey, you know, you’re struggling with stress, you’re not feeling included.
    0:13:04 How do I get help?
    0:13:10 The coach could almost be there to talk me through, to work me through these more
    0:13:11 social and emotional.
    0:13:16 I mean, the headline here is, without huge technological advances, from my
    0:13:20 understanding, AI could really make life better.
    0:13:23 I mean, I’m going to say that bluntly.
    0:13:23 Right.
    0:13:24 I know there are risks.
    0:13:25 We can talk about them.
    0:13:31 But AI can make life better for people who, for centuries upon centuries, have been
    0:13:34 excluded from the basic needs to make life better.
    0:13:39 In the case of AI, you know, it’s a huge umbrella term.
    0:13:39 It’s being thrown around.
    0:13:40 I’m working on it, et cetera, et cetera.
    0:13:44 We’re sort of in this stage that I’ve heard some folks refer to.
    0:13:47 Probably somebody who was on the show, I guess, in the show referred to it as kind
    0:13:53 of we’re in the applications phase where the models, you know, that made a big splash,
    0:13:57 GPT and Claude and the other ones released to the public and people started using them
    0:13:58 and they caught on and everything.
    0:14:02 And now people are figuring out how to build applications that leverage these
    0:14:06 models to do things, to improve lives.
    0:14:06 Right.
    0:14:10 And there’s a lot of focus, obviously, on business and industry and things like that.
    0:14:15 Tim, what you’re seeing is that the developers and the free market who’s
    0:14:21 building these apps on top of the models just aren’t serving your community?
    0:14:21 Or is it?
    0:14:23 I don’t want to make it a negative.
    0:14:29 I mean, Brad Smith tweeted out our column that we did on AI and at Special
    0:14:30 Olympics, the polling data.
    0:14:34 So, you know, there’s obviously an enormously powerful person that had
    0:14:38 one of the most powerful companies in the world, Microsoft, who’s talking about it.
    0:14:41 I don’t want to say that people aren’t interested.
    0:14:42 I want to say that we’re at a moment.
    0:14:49 Look, most revolutions, if you look back, have left our folks out for decades, decades
    0:14:51 upon decades. We don’t want that to happen this time.
    0:14:54 So we’re trying to get ahead of it here.
    0:14:58 Like we’re trying to invite, in a sense, we want to invite ourselves in.
    0:15:03 To the application development phase, because we have huge confidence that there
    0:15:08 are people of goodwill who would like to see the applications of AI include our
    0:15:13 community. But to your point, if they’re thinking about where’s the most lucrative
    0:15:15 application to build, it’s going to be in business.
    0:15:21 The most popular applications to build is going to affect 70, 80, 90% of the
    0:15:26 population. Maybe that’s not going to be us, but the most impactful applications to
    0:15:31 build. I defy anyone to say there is an application for AI that would be more
    0:15:34 life changing than the kinds of applications we could build for our
    0:15:35 community.
    0:15:39 So can we call this an open call to the audience if you’re listening, you’re a
    0:15:46 developer, you want to be a developer, you’re someone who project managers
    0:15:50 run whatever your skills are. This is an opportunity to get involved.
    0:15:56 I mean, thanks to Sarah, we’re being heard on this issue at the highest levels
    0:15:59 of government. We have ministers of education around the world who are
    0:16:05 asking for solutions, for strategies, for protection, for building the kind of
    0:16:10 ground rules and the codes of conduct that can make AI successful, protecting
    0:16:13 safety, privacy, security, these kinds of things, at the same time, responding to
    0:16:18 the needs. So, you know, Sarah has put us on the stage at the level of political
    0:16:22 leaders and many other stages too. I don’t mean to diminish it to just that.
    0:16:27 But the open call is, you know, we need a G7 meeting with the G7 of tech.
    0:16:34 You know, with Google and Microsoft and all these guys. We want that G7 meeting.
    0:16:40 And I want Sarah to give a keynote address at that G7, at that G7, the big
    0:16:44 seven or the big whatever it is, and invite them, you know, just say exactly
    0:16:47 what you just said. No, look, hey, gang, you know, we’re going to change the
    0:16:51 world. That’s not a question. We know that. Our parents know that. Parents of
    0:16:55 kids with special needs, they know it’s coming. We want to make it great, you
    0:16:59 know, we want to join you. And I have no doubt that in that audience, there
    0:17:04 will be people both philanthropically using their kind of passionate purpose
    0:17:09 mindset, but also from a business point of view, saying, you know, by the way,
    0:17:14 if you do this, if we built the state of the art special effects coach that
    0:17:18 helped with sports, with hygiene, with social needs and so on, guess who would
    0:17:24 benefit? Everybody. Everybody’s kids benefit. So anyway, it’s a potential
    0:17:25 real win-win.
    0:17:31 Sarah, do you, since Tim brought up the G7 and maybe this call, maybe this
    0:17:36 podcast will spark, you know, the big tech seven coming into the Special
    0:17:39 Olympics table. Sarah, do you want to talk about the G7 that you just
    0:17:44 attended and maybe dovetail and from whatever direction it makes sense for
    0:17:48 you to start, but get in a little bit to your work with the U.S. government and
    0:17:53 kind of what the U.S. government has been doing to drive development, drive
    0:17:58 access to assistive technologies for folks with disabilities?
    0:18:03 Definitely. I’ll actually take it first, a broader view, bring it back to G7,
    0:18:08 take it back to G7. So just, you know, to what Tim said is really important from
    0:18:14 our foreign policy kind of lens, you know, we have the G7, we have G20, we
    0:18:20 have Asia Pacific Economic Cooperation, we have ASEAN, we have the C5 plus one,
    0:18:23 so the Central Asian countries plus the U.S. These are all different multi-lot
    0:18:29 spaces that in every single of those spaces, there’s an AI discussion
    0:18:33 happening. Either an AI ministerial meeting, actually during the G7
    0:18:36 disability ministerial meeting, there was an AI one happening at the same time.
    0:18:43 Okay. And so just with that perspective, right, a lot of times these AI leaders,
    0:18:47 and I’ve been to a lot of different AI conferences, are not thinking about, as
    0:18:49 Tim said, they’re not thinking about the disability and accessibility.
    0:18:54 They’re not. Right. And this is where it’s really important on one hand for
    0:18:58 governments that are working on disability, this is a call for anyone
    0:19:02 that’s working on disability within government beyond, how do you insert
    0:19:06 yourself in those main AI spaces and say, hey, disability, disability and
    0:19:09 disability. I’ll give you an example. So ASEAN, the Southeast Asia countries
    0:19:13 right now, they’re negotiating the Digital Economic Framework Agreement.
    0:19:16 How do we make sure disability accessibility is part of that, right?
    0:19:20 It’s not, you know, so these are the things that we need to map out and
    0:19:24 understand where are the main AI spaces and how do we make sure we bring the
    0:19:28 disability community, we bring organizational forces with disabilities,
    0:19:32 we bring the experts to the table. That’s number one.
    0:19:37 Number two is, you know, to the point of Tim, it’s we bring value,
    0:19:40 we bring innovation when you do, when AI is not accessible, it further
    0:19:44 marginalizes us. And AI versus technology, yes, there’s AI versus technology,
    0:19:48 but there’s also AI in general that we’re billing, we need to make sure it’s
    0:19:53 accessible, right? Those two layers. And when you bring people to those spaces,
    0:19:56 like, you know, if you know Microsoft Teams, there’s the automatic captioning.
    0:20:00 Where did that come from? That came from an employee of Microsoft who’s deaf who
    0:20:03 needed to make sure it’s accessible for him. And that automatic caption was
    0:20:05 created because of him, and now everyone uses automatic captioning.
    0:20:10 Right. But then also AI is still built on a biased data, which actually has
    0:20:14 been hurting us in a lot of different spaces, including applying to jobs,
    0:20:17 for instance, when companies are using AI as an initial platform.
    0:20:22 So that’s the layers. These are the conversations we need to bring to the
    0:20:25 multi-lots spaces and the bilateral spaces to say, hey, you need to make
    0:20:29 sure we bring the disability community. That point to that, we also work on
    0:20:33 making sure we’re bringing upskilling the disability community into AI
    0:20:38 development, cybersecurity, technology development, programming. We should be
    0:20:43 part of that job sector. So to answer your question on the G7, this year we
    0:20:47 had the first ever G7 inclusion and stability ministerial meeting. It was
    0:20:52 on the protocol level. It was G7 ministers of disability and their
    0:20:56 delegations coming together where we talked about technology and AI. We
    0:21:00 talked about independent living. We talked about crisis and disability. It
    0:21:05 was amazing. Thanks to Italy, we had civil society, private sector and
    0:21:10 government. We need to continue elevating disability through that, but we also
    0:21:14 need to continue mainstreaming disability through the other G7 ministerial
    0:21:17 meetings that happened throughout the year. And I think that’s as important.
    0:21:21 Again, it goes back to the twin track approach. And I just want to end to one
    0:21:27 last point. AI could be a solution to so many different issues, whether it’s
    0:21:31 climate disaster, whether in terms of responding to crisis, whether it’s
    0:21:36 food insecurity for farmers and the technology that they use, whether it
    0:21:40 is economic and business development. We need to make sure AI is a solution for
    0:21:44 a lot of those faces. We need to make sure that disability lens is also part
    0:21:48 of integrating across the board through that. To your point, there’s so much
    0:21:53 that digital technologies of all sorts lend themselves to flexibility and
    0:21:57 lend themselves to being able to make the information accessible through
    0:22:04 whatever, whether it’s through audio or through images and text or going
    0:22:08 forward through robotics and other sort of more physical means. Please correct
    0:22:14 me where I’m wrong here, but kind of as I’m hearing the larger issue of across
    0:22:18 all of these different spaces in the world and things that people do in ways
    0:22:22 that whether it’s legislation or private enterprise, ways that things get
    0:22:29 done, folks with disabilities and bringing AI to the disability community is
    0:22:34 just something that needs to happen more. And I’m wondering if that’s, in
    0:22:38 your work, if that’s primarily a governmental function, is it kind of a
    0:22:43 public-private partnership? Is it, as with both things, sort of some of both
    0:22:47 government and then you have individuals like the person at Microsoft who, out
    0:22:52 of a personal need, arose this thing that’s helped so many people. I’m not
    0:22:55 asking so much where the responsibility lies, but just sort of how does this
    0:22:59 happen? How do people get more of a seat at the table?
    0:23:03 A really good question. I think it’s a combination of a few things. One, yeah,
    0:23:10 it’s a policy issue where how do we ensure that when companies are developing AI,
    0:23:14 it’s a standard that it is accessible for all, right? How do we get to that point
    0:23:19 where companies need to, you know, bring in that lens from the get-go? Because then
    0:23:22 that would translate, because I have so many conversations with startup
    0:23:26 companies and I was like, how’s your beta version going? And is it accessible?
    0:23:31 Not yet, not now, maybe later, but later, right? There’s also a narrative issue.
    0:23:36 We are an invisible population a lot of the time. We are a population that
    0:23:40 people don’t think about us. But many times I hear, oh, I’ve never thought
    0:23:44 about this, right? And this goes back to why having a conversation with you and
    0:23:47 with other folks in the media world and in the entertainment world that we bring
    0:23:52 and make more visibility on the intersection between AI and disability, AI disability,
    0:23:56 and other issues, like security, climate, what if the more we bring those forward,
    0:24:00 you know, the media world create more of the narrative, the more that people will
    0:24:03 think about it and bring it forward. I think it’s multi-layered, but I think
    0:24:05 policy is definitely an important part of the question.
    0:24:10 I’m speaking with Sarah Minkara and Dr. Timothy Schreiber. Sarah is the Special
    0:24:14 Advisor in International Disability Rights at the United States Department of
    0:24:19 State. And Tim is the Chairman of the Board of Special Olympics. And we’ve been
    0:24:24 talking about the role of the role that AI can play and will play and kind of more
    0:24:29 broadly the role of assistive technology and technologies in general and how
    0:24:33 people with learning disabilities and other disabilities have historically
    0:24:37 not have as much of a seat at the table as we’ve been calling it, just
    0:24:42 have not been as much of a part of the process as really they should be for a
    0:24:46 number of reasons benefiting all of us that Sarah and Tim have been talking
    0:24:52 about. As we were saying, Sarah has just come from the G7 meetings in Italy and
    0:24:56 the first-ever ministerial on inclusion and disability. So, of course, when we’re
    0:25:02 talking about the growing presence of AI and its ability to, for me, I think about
    0:25:06 its ability to hyper-personalize things and, you know, the sort of positive side of
    0:25:11 that, the sunny side is the ability to just really advance the way that we learn
    0:25:14 and fitting different modalities and skill levels and interests and needs and
    0:25:19 abilities and disabilities and all those things. Kind of a dark side of that is
    0:25:24 the endless scroll, right? I’m guilty of it as much as anybody where I’ll find
    0:25:27 myself reaching for advice, looking at something because it’s
    0:25:31 interesting to me, or at least that’s what I perceive it. And then however much
    0:25:36 time goes by and I feel sort of isolated and lonely and maybe not that great
    0:25:38 because I’ve just been kind of doom-scrolling for a while.
    0:25:42 You know, it’s something that Tim and Prepping for the podcast, I read some, I
    0:25:46 don’t know if there were your thoughts directly, but some articles and some
    0:25:49 things and I think you have some thoughts on that that the audience I’m
    0:25:53 sure will be interested to hear. I mean, I think first of all, the risk is real.
    0:25:57 We do live in a time when most young people in this country in the United
    0:26:02 States, at least, are either lonely or depressed. I mean, I say most, just over
    0:26:06 50% depending on how you count, that’s an enormously painful number.
    0:26:10 It’s shocking to think of 16, 18-year-olds, more than half of them
    0:26:14 feeling it’s fair about the future, negative
    0:26:17 impression about their own capacity, distrustful of others.
    0:26:21 That’s a recipe for disaster. So we got to take that issue seriously and for
    0:26:24 people with intellectual developmental disabilities, that’s been the
    0:26:29 norm for centuries, lonely, isolated, left out, right?
    0:26:33 So here’s the point. We need a gross well of innovation right now. I mean,
    0:26:37 this technology is coming out not when the printing press came out or when the
    0:26:40 combustion engine came out or when the personal computer came out.
    0:26:44 In all those eras, people with disabilities were invisible.
    0:26:48 Today, we’re on playing field. Look at Sarah’s role, look at her leadership.
    0:26:51 We’re on the playing field. We want to make this revolution different,
    0:26:56 right? We want to make it more thoughtful about human interactions,
    0:27:00 more aggressive on innovation, more attentive to the risks of
    0:27:05 isolation and despair and distrust that we see all around us.
    0:27:09 This technology, maybe we can talk about all in grandiose terms,
    0:27:13 all the things it can do, but if it doesn’t do this well,
    0:27:17 it will have failed humanity. This is my view. If we don’t feel
    0:27:21 distrust in the fear in our relationships and if we don’t use it to
    0:27:24 empower people on the margins to join the mainstream and empower
    0:27:28 people in the mainstream, open their hearts and minds to what they
    0:27:32 will learn from people who have been excluded, if we don’t do that,
    0:27:36 technology will have failed us. I don’t care how fast it writes a novel.
    0:27:40 I don’t care how fast it can solve a complex problem.
    0:27:45 It will fail us if it does not address this fundamental tension
    0:27:48 all over the world right now, which is how we heal
    0:27:52 this fear and this anxiety and this distrust between us.
    0:27:57 So, is it real? People are talking about warning labels on phones.
    0:28:03 That’s a real thing. Anybody who’s got a child who’s a teenager or older
    0:28:08 who has a device is worried. I mean, 98% of parents are worried.
    0:28:13 So that’s not a joke. That’s a real fear, but I think Sarah and I agree on this.
    0:28:16 I’m not waiting for the government to show us how to do this.
    0:28:21 I know the government can cooperate with the innovative
    0:28:27 energy, creative energy, the rate, the boundaries of what’s possible
    0:28:30 energy. That will come from the private sector. That will come from coders.
    0:28:33 That will come from engineers. That will come from people who are being honest.
    0:28:38 And then they will say, “Hey, Mr. Government, look what I can do for your kids.
    0:28:42 Look what I can do for healthcare. Look what I can do for community building.
    0:28:46 You know, join me.” But the creative energy has got to come
    0:28:50 from the business and the technology sector. I’m confident we can do it, but
    0:28:54 Sarah’s right. People like Sarah, me, and the millions of people with intellectual
    0:28:57 development or disabilities, they’re families. They’re going to have to be
    0:29:00 powerful self-advocates. They keep hitting the door and saying,
    0:29:06 “Wait a minute. Don’t leave without us. We have the key for a successful revolution here.”
    0:29:10 And to Tim’s point, thank you, Tim, for what you said. It’s so true.
    0:29:15 And if AI and technology is made without accessibility being in mind and it further
    0:29:20 marginalizes it, again, it’s failed. Here’s what really bothers me a lot
    0:29:25 in traveling the world. I still have communities that say they’re fine that
    0:29:28 people with disabilities are left behind. They’re left in their homes.
    0:29:32 They’re still shackling the people with disabilities. They’re still people in
    0:29:36 institutions. They’re still sterilization of people with disabilities.
    0:29:41 So this still is a reality. Sometimes I talk to companies that are talking
    0:29:45 about the challenges of people with disabilities and getting to employment and
    0:29:49 they say, “Well, they can just work from their home or they can just be whatever it is.”
    0:29:54 The solution is always let them just stay behind. Sorry, but that’s not right.
    0:29:59 And we should not accept that. And AI, if it’s made without accessibility,
    0:30:04 it’s going to further leave us behind. But again, if AI is made with disability in
    0:30:08 mind, it can allow us to be further included in society. Again, for me,
    0:30:14 I use technology on a not even more than hourly basis. Because of apps,
    0:30:18 I’m able to call Uber and I’m able to get an Uber on my own and I’m able to get
    0:30:23 to my location because of apps I’m able to use. But again, there’s a lot of apps
    0:30:27 and websites that are not accessible. So I’m not able to fly. It’s a simple thing
    0:30:31 is that it was taken from there from the beginning. There’s hotels now that are
    0:30:36 being fancy and there’s the touch screens of things. But again, I’m not able to
    0:30:40 turn on and turn off my life on my own without having someone helping me out.
    0:30:43 You can have any solution for that. But again, it’s because they don’t have us in mind.
    0:30:48 So my ask is always we need to get to a point where it’s a standard. We see the
    0:30:52 disability community, they’re brought to the table and sorry, but it’s not enough
    0:30:56 for people to say pull up a seat to the table. Is this space physically accessible?
    0:30:59 Is it accessible for deaf people? Is it accessible? Sorry, it’s not enough for you
    0:31:04 to say pull up a seat to the table. And then from there, I want to get to a point
    0:31:09 where companies demand this because they see the value that it brings to society.
    0:31:14 So this goes back to narrative change. No, it’s a fight that I think we all need
    0:31:18 to fight on behalf of one another. And I think you both articulate much better
    0:31:22 than I could, but that’s the feeling I’m coming away with. So, Sarah, Timothy,
    0:31:27 thank you so much for taking the time. For folks who are listening, who want to learn more,
    0:31:33 who want to donate their skills, who want to get involved, who have a crazy innovative idea
    0:31:39 that’s going to really change things. Sarah, I’ll start with you. Where can folks go online to,
    0:31:43 it doesn’t have to be online, but if there’s a website, to learn more about the work that you’re
    0:31:48 doing, about the work that’s going on in any of the governmental or other organizations
    0:31:52 you’re involved with. Where’s a good place for listeners to go? Definitely the State Department
    0:31:58 of Social Media, the DRL Sandals, the website, there’s media notes that come out on all of our
    0:32:02 travels. We’ve been to 48 countries, so all of our travels, there’s media notes in terms of
    0:32:05 the work we’re doing. If you’re in a certain country, the U.S. Embassy will always be
    0:32:09 in touch with the work that we’re doing. Fantastic. And Tim, for folks who want to
    0:32:14 learn more… I mean, look, the Special Olympics movement is in pretty much a problem. I’m guessing
    0:32:20 every country that a listener to this podcast is listening in. So… If you’re out there listening
    0:32:26 and the Special Olympics is not visible in your country, and please drop us an email, Tim.
    0:32:31 Yeah, if you’ve got a great idea, let me tell you this is a movement that is
    0:32:38 imposed of creative, innovative rule breakers. Our founders were rebels in pursuit of social
    0:32:42 and political change. That’s the only way they got what they got. That’s the way we got into
    0:32:47 190 countries. That’s the way we found a million volunteers. We’re a movement of volunteers.
    0:32:53 99% of our workforce is volunteers, so we welcome the innovative energy of citizens
    0:33:01 who want to make a difference. My personal email is tshriver@specialolympics.org.
    0:33:08 Write me if you’ve got an idea. This is a movement that is eager and, you know, as angry and as
    0:33:14 frustrating and as infuriating as it is so many times to be an advocate for or a person with an
    0:33:20 intellectual or developmental or a physical disability, as frustrating as it is. This
    0:33:26 is a hopeful community, and it’s a community that wants participation, and it’s a community that
    0:33:33 wants to bring a greater degree of fulfillment to the world. We’re a community that’s hungry
    0:33:39 for the chance, and everybody that I know of in this field says the same thing. Once I opened
    0:33:46 the door, once I saw, once I went, once I gave it a chance, changed my life. Very few people say,
    0:33:50 oh, I met Sarah Minkar. She convinced me to do something, and now I’m never going to do it again.
    0:33:56 Most of the people say, I met her. She told me to do this, and wow, things are so much better. Same
    0:34:00 thing is true in Special Olympics. So I hope the message people hear when they listen to this is,
    0:34:06 wow, there’s some folks out there really need me and want me and would welcome me. And I hope
    0:34:11 you’re hearing that. Be inclusive as a starting guy and get involved. I mean, you know, the simple
    0:34:16 thing when people go, I don’t have time for that. You know, one thing everyone can do right now,
    0:34:21 text HR at your company. If your company’s got six people, or if it’s got 6,000, or if it’s got
    0:34:27 600,000, text HR. Do we hire people with disabilities in our company? Just send out one text,
    0:34:32 and you watch. If the answer is no, you’re going to get a change in policy. And if the answer is
    0:34:35 yes, you’re going to learn about people in your company that can make a difference.
    0:34:39 There you go. Tim, Sarah, thank you again. Let’s do it again sometime down the road, and
    0:34:46 hopefully we’ll have positive, positive updates, momentum, cool new apps. The SO Coach?
    0:34:49 The SO Coach. Yeah. There you go. You heard it here first.
    0:34:55 Thank you both so much. Thank you. Thanks, everyone.
    0:34:59 [Music]
    0:35:03 [Music]
    0:35:15 [Music]
    0:35:23 [Music]
    0:35:35 [Music]
    0:35:43 [Music]
    0:35:53 [BLANK_AUDIO]

    In this episode of the NVIDIA AI Podcast, Sara Minkara, U.S. Special Advisor on International Disability Rights, and Timothy Shriver, Chairman of Special Olympics, discuss AI’s potential to enhance special education and disability inclusion. They emphasize the importance of including disability communities in AI development, as well as the cultural and social benefits of building an inclusive future.

  • NVIDIA’s Louis Stewart on How AI Is Shaping Workforce Development – Ep. 237

    AI transcript
    0:00:10 [MUSIC]
    0:00:16 Hello, and welcome to the Nvidia AI podcast. I’m your host, Noah Kravitz.
    0:00:21 We’re fortunate on the podcast to host guest after guest with amazing stories to tell about
    0:00:26 using AI to achieve scientific breakthroughs and transform industries. But in a time of such
    0:00:32 rapidly evolving technology, how do we all ensure that the next generation is equipped to work and
    0:00:37 thrive in an AI-powered world? Workforce and economic development have a vital role to play
    0:00:43 when it comes to all of society realizing the benefits of AI. Here to discuss why AI education
    0:00:48 is so important and how artificial intelligence is impacting workforce training and economic
    0:00:52 development is Lewis Stewart. Lewis is head of strategic initiatives for Nvidia’s global
    0:00:58 developer ecosystem, a role to which he brings over two decades of experience and expertise in
    0:01:03 innovation, entrepreneurship, and economic development. Prior to joining Nvidia, Lewis was
    0:01:08 the first chief innovation officer for the city of Sacramento, California, a position that capped
    0:01:14 off more than a decade of public service at both the state and city levels. Lewis, thank you so
    0:01:19 much for making the time to join us and welcome. Yeah, thanks, Noah. It’s an honor to be here.
    0:01:27 So should we start with setting the table, defining one or it might be two related terms.
    0:01:33 What does it mean when we talk about workforce and economic development in the context of AI
    0:01:40 and this AI era that we’re embarking into? Yeah, so great question. I think the challenge right
    0:01:48 now is everything gets conflated. It gets conflated into what is, you hear governments talking about
    0:01:56 gen AI, you hear governments talking about AI taking jobs, you hear a whole lot of generalizations,
    0:02:07 if you will. And so for what does it mean truly? AI is fueling a lot of change in all ecosystems
    0:02:12 right now. And so what that means is disrupting how traditional economic development is thought
    0:02:20 of. So how states, countries plan what happens next, how they stay competitive globally with
    0:02:28 each other. And workforce, AI and the speed at which it’s causing innovation or disrupting the
    0:02:36 world. On the workforce side, it has the potential to create one of the widest disparities out there.
    0:02:42 So training and getting folks to understand that it’s imperative for people to jump in
    0:02:49 right now and be part of the conversation is huge. And so economic and workforce development is,
    0:02:56 I think, at the crux of this next part of the conversation because the innovation and the
    0:03:03 research and everything surrounding that is driving change so rapidly. So to back up a step
    0:03:08 several years and then maybe we can work our way forward. When you were the Chief Innovation Officer
    0:03:14 for City of Sacramento, were you thinking about the same kinds of things, you know,
    0:03:18 sort of big picture that you’re thinking about now? Yeah, yeah. So I appreciate that question.
    0:03:26 So the answer is yes. Was it truly focused on AI? Absolutely not. Right. But it wasn’t focused on
    0:03:31 NVIDIA. No, I just knew NVIDIA was a player in the space, right? And so moving into that role in
    0:03:37 the City of Sacramento, I was bringing with me a lot of the knowledge from 10 years at the state
    0:03:43 working for two governors being the innovation guy for California and having conversations with
    0:03:50 global leaders that were coming to California to talk about innovation. And so working in Sacramento,
    0:03:56 what I understood is Sacramento was the capital. So all the legislators were here at least four
    0:04:04 days out of the week. Sacramento was trying to move into a place to be an innovation leader
    0:04:12 in the state and nationally. And I knew that I had a lot of knowledge behind that. So coming
    0:04:20 into that role, the other part that I knew is I knew the city because I grew up here. Okay. And I
    0:04:25 knew that there were areas of the city that would never see innovation unless it was brought to
    0:04:33 them. Right. So it was important to me to look at how do you help the city shape economic and
    0:04:39 workforce development policies, efforts, whatever, that included everybody. So that means if I was
    0:04:45 bringing autonomous cars to Sacramento, it was for the legislators, it was for the CHP,
    0:04:50 but it was also for the people so that people could understand what technology was coming,
    0:04:56 how it would help them, how they needed to be trained differently, whether it be mechanics,
    0:05:02 learning how to work on gas engine cars, now needed to understand the computers behind electric
    0:05:08 vehicles and autonomous cars. Or just moms and dads needing to understand that they could put
    0:05:13 their kids in an autonomous car going from home to school so they can stay at work and continue
    0:05:19 to work. And so there were so many nuances in those conversations that working in that role,
    0:05:25 I had great aspirations because I understood the innovation coming, but just not the speed at
    0:05:30 which things were going to change. And so you left that post and came to NVIDIA. How long ago now?
    0:05:36 It’s been four years, a little bit over four years. Okay. And so what’s kind of carried through and
    0:05:41 then to sort of jump ahead? Well, you can walk ahead as opposed to jump, if you like, through
    0:05:48 the barriers. But what is the role entailed day to day now? And I don’t know, how are you thinking
    0:05:53 about these things in the context of, I mean, these have been quite the four years in the AI
    0:06:01 world, right? Right, right. Well, so it’s really interesting, though, because coming here to NVIDIA
    0:06:07 four years ago, I still spoke the language of government. So everything I started talking
    0:06:12 about four years ago was workforce, workforce, workforce. But that at the time was not the
    0:06:18 language of NVIDIA. NVIDIA was talking about training and didn’t see the two as synonymous.
    0:06:25 Right. So I needed to learn how to figure out the way NVIDIA moved. But the North Star was still,
    0:06:31 how do you get underrepresented communities involved in understanding what was coming next?
    0:06:37 And so these last four years at NVIDIA has been incredible. We’re at the center of a lot of the
    0:06:44 changes that’s happening, which is fantastic. But it’s actually increased my feeling that it was
    0:06:52 100% necessary to really pivot hard into workforce development conversations throughout the country,
    0:06:59 but really throughout the world. And falling back on my government background to help legislators,
    0:07:08 to help state leaders, to help whoever needed to hear that this wave is coming. And working
    0:07:14 from fear doesn’t really work. And you need to understand that if you really, truly care about
    0:07:19 your citizens, it’s time to actually start working with companies like NVIDIA and others
    0:07:25 to train a workforce to get your higher education systems in line and thinking about
    0:07:31 what’s happening next so that as our students graduate, they’re employable. They understand
    0:07:39 and have had touches in AI. And this isn’t just an engineering or a research challenge. This is
    0:07:46 really a true workforce issue. We did an episode a little while back with Georgia Institute of
    0:07:52 Technology who had opened a makerspace. And one of the key things about that is that it was open
    0:07:57 to undergraduates, not just graduate students and CS. That’s an important step, right? But that still
    0:08:03 feels like a big step away from bringing the autonomous cars to all parts of the city to
    0:08:09 put it that way, right? So there was a federal call to build federal in the United States,
    0:08:14 because we’re talking globally, to build a sustainable workforce. What does that mean and
    0:08:19 how does that kind of drive some of the work you’re doing, some of the programs that either
    0:08:27 you’re involved with or seeing in the United States? Yeah, so that call for the need for a
    0:08:32 sustainable workforce really came out during the pandemic. And you saw all the shortages in the
    0:08:41 supply chain. And I’m sitting here at NVIDIA watching change happen even faster because
    0:08:47 of the technologies that we’re rolling out across industries. And you’re kind of observing the
    0:08:53 landscape. You’re looking at the schools. You’re looking at even my kids, and I’m worried about
    0:09:00 them. You’re looking at legislation. You’re looking at all of that because having an understanding
    0:09:04 of what it actually takes and what programs actually go into building a workforce,
    0:09:11 you see that there was a huge misalignment. So moving into this conversation with NVIDIA
    0:09:21 entailed, okay, let’s try to become a trusted advisor and a responsible steward of the technology
    0:09:28 when in the eyes of government, right? So help them understand that we’re a piece of the puzzle.
    0:09:37 We’re not the solution. But if you want a workforce that can thrive, help your economy,
    0:09:43 help your innovation ecosystem, we have stuff that can help, whether that be our deep learning
    0:09:51 Institute workshops or our startup ecosystem called inception or just the knowledge that we have
    0:09:57 around chip design, right? And so on one side, we do the chip design that’s manufactured something
    0:10:02 else. So that’s one part of the workforce that we need to worry about. But then when you go back
    0:10:06 to the economic development conversation, there is the workforce that’s generated out of the higher
    0:10:12 education system. Some of those higher education systems like HBCUs, you have first time students
    0:10:17 that they’re not first time students, first time college students, you know, and their families.
    0:10:24 And how are they being taught? How are they learning about all the changes in AI so that
    0:10:29 they can be more hireable in the industries around those universities or back at home where
    0:10:33 they’re from. That’s not just HBCUs, but that’s that’s been one of my focus areas over the past
    0:10:39 four years is trying to reach into the minority serving institution community. And so you have
    0:10:44 to look at the workforce that’s being trained. But then you have to look at the Ansley workforce,
    0:10:51 which is say you have a large bab that’s being built in a particular state and the union,
    0:10:56 and they tell you that there is not a workforce to do the job that they’re trying to do. But
    0:11:04 that fab makes our chips. So we actually need to care about that workforce as well,
    0:11:09 even though we don’t make the chips ourselves, right? So there’s the stuff that we can touch
    0:11:13 directly through our workshops, through our subject matter experts being engaged with the
    0:11:18 students and telling them what a day in the life of NVIDIA is like. There’s our industry partners
    0:11:22 that can walk alongside us with their training and their subject matter experts. But then there’s
    0:11:29 the fabs that we need to be watching as well to make sure that should we need to think about
    0:11:33 our supply chain differently, we can make sure that there’s a workforce to support that here in
    0:11:38 the States as well. Right, right, of course. I’m speaking with Lewis Stewart. Lewis is head of
    0:11:43 strategic initiatives for NVIDIA’s global developer ecosystem. And we’re talking about
    0:11:49 workforce and economic development. It’s been kind of a whirlwind couple of years, particularly
    0:11:55 with generative AI and capturing the public’s imagination, let alone the resources being
    0:12:02 invested. But it’s still early days in terms of kind of the long-term impact of AI on everything.
    0:12:08 But in terms of the workforce, when you’re talking about college level, undergraduate level, or that
    0:12:17 age of a student or learner, and you’re talking about AI education, is it first time using a chat
    0:12:24 bot? Is it industry-specific training for somebody who knows that they want to go into
    0:12:29 whatever field it might be? Is it kind of the whole gamut of things? What does that look like?
    0:12:34 That’s phenomenal question. So it’s really the whole gamut, right? So our efforts in the U.S.
    0:12:40 right now, state by state, we do an economic analysis of the landscape there to understand
    0:12:44 what industries are the most prevalent. The top five, top 10 industries that are most prevalent
    0:12:48 in that state. We look at the universities to make sure they’re teaching stuff that’s aligned
    0:12:53 with our efforts. And then we try to do a statewide initiative. So every state is going to look and
    0:12:57 feel a little bit different based on the industries that they have. Is that to jump in for a second?
    0:13:04 Is that with all 50 U.S. states or? Well, we hope to get to all 50 states right now.
    0:13:09 Yeah, right now we have an agreement with California and the community colleges in
    0:13:15 California. So that’s 116 community colleges. We’re currently in active conversations with
    0:13:21 five other states, and we figure that we’ll have more coming on over the next couple of years.
    0:13:26 Again, just because of the speed at which stuff is happening. But to dive a little bit deeper into
    0:13:33 that question, community colleges cannot be the lowest level that you start thinking about
    0:13:38 AI training and curriculum and reskilling and upskilling. As we travel the state,
    0:13:42 as we get into these conversations, states, and we get into these conversations,
    0:13:49 everybody’s asking about K-12. And NVIDIA doesn’t have a K-12 practice per se. So we try to do
    0:13:55 partnerships with dual enrollment opportunities where the community colleges reach back into
    0:14:00 high schools. We try to work with colleges that have high school to college pipelines
    0:14:05 in order to influence that. But for this conversation, if you really talk about workforce,
    0:14:12 you have to literally start at kindergarten level, like the folks in Gwinnett County out in Georgia
    0:14:20 have where they’ve done K-16 curriculum, getting kindergarteners thinking about AI and ethics and
    0:14:24 stuff like that. They’ve opened up a couple high schools. I don’t know if you have the details,
    0:14:29 but what are they having kindergarteners think about? So I don’t actually have the details
    0:14:33 right now. And I know it’s evolved, but what they presented to us two years ago was phenomenal,
    0:14:40 right? Just AI basics, like thinking about associations, thinking about what should AI do,
    0:14:48 how do you think about AI? So really, if young kids are already playing with AI on their phones,
    0:14:56 yeah, they should actually be thinking about it a little bit deeper. So they’re not just
    0:15:00 users of the technology, but they can actually start thinking about and seeing themselves as
    0:15:06 creators of the technology and being part of that evolution. So thinking about slightly older
    0:15:13 learners, what are some of the things that NVIDIA is offering to students, developers? Obviously,
    0:15:18 there’s the whole developer ecosystem. There’s the Deep Learning Institute. What are some of
    0:15:21 the educational offerings that training development offerings NVIDIA has?
    0:15:27 Yeah. So within NVIDIA, you mentioned that we have the Deep Learning Institute and a lot of the
    0:15:35 material in there, it’s self-paced learning. So folks can actually come in and take specific
    0:15:41 workshops, either two to eight hour workshops, where they’re actually manipulating GPUs, right? So
    0:15:47 while our trainings aren’t promising you a job or training you for a specific job,
    0:15:57 they are absolutely teaching you how to effectively and efficiently use the resource that is the GPU.
    0:16:01 So if you’re a researcher, you’re learning how to manipulate the GPU to help accelerate your
    0:16:08 research. If you’re an enterprise or an industry, it’s helping you figure out how to create solutions
    0:16:13 for your enterprise or your business. So you have the self-paced stuff. We have teacher or
    0:16:19 instructor-led courses, which are a little bit more in-depth. We have teaching kits that are
    0:16:24 available to instructors at universities. So you have to actually be an instructor and the
    0:16:32 teaching kits can be taken on their semester-long courses of study that can be taken as they are
    0:16:37 or you can download them and use them a la carte so you can incorporate them into your curriculum.
    0:16:42 And then we also have the ability for professors to become ambassadors. So then they’re actually
    0:16:49 certified by our master instructors to teach workshops, NVIDIA workshops. So those are all
    0:16:55 some of the tools and resources that I lean on when I walk into a conversation at a state or
    0:17:01 university and say, again, this is a piece of the puzzle because everybody’s not ready to walk
    0:17:06 into master’s PhD level learning like our deep learning institute. You have to be technically
    0:17:12 super technically proficient. And so we look for opportunities, especially at the university level
    0:17:16 where they’re already teaching Python, they’re already teaching Algebra 3, they’re teaching some
    0:17:23 of the basics so that you can discover there’s the aptitude already and you can get intro to AI
    0:17:27 courses so that people can scale up into what we do. But then we also on the other side are looking
    0:17:32 at the researchers, looking at the engineering talent to make sure that they have what they
    0:17:38 need so that when they graduate out, they have augmented AI skills to be better employable
    0:17:45 than the market. Fantastic. Outside of the tech industry, we talked a little bit about this, but
    0:17:52 on the show across the country, look around in the world and AI is transforming all kinds of
    0:18:00 industries and things that people do and create. And so if you’re thinking about less technical
    0:18:06 industries or somebody who, as you alluded to, isn’t going to go to that master’s PhD level
    0:18:17 of education, how do you think about, how do you work with partners thinking about AI education
    0:18:24 and workforce development for workers who were doing jobs that aren’t highly technical?
    0:18:31 Yeah, absolutely. So more of it, please understand that when we try to walk into these partnerships,
    0:18:37 one of the things that we put on the table is having our subject matter experts or other
    0:18:42 individuals with the company be accessible in these partnerships. So when you think about
    0:18:49 the non-technical side of the house, we try to reach into our marketing team and get some of
    0:18:55 our marketing experts to talk to business schools. We try to get our sales team to talk to business
    0:19:01 schools and whoever else. Just so, one, you can see that it’s not all super tech focused,
    0:19:07 and other job opportunities exist. But that also helps when you start talking about reskilling
    0:19:13 and upskilling. So if somebody is making a transition and it could just be a change in
    0:19:19 language, right? You may have the transferable skills, but not know how to speak the language
    0:19:24 of a tech company, kind of like when I first came in from government, right?
    0:19:28 Now you speak both languages, so you got the advantage.
    0:19:34 So yeah, and so it’s getting people into those, even just those conversations. And so for us,
    0:19:39 sure, we rely heavily, and we lean heavily on the research and engineering side, and that’s
    0:19:45 what drives our company. But as part of trying to do good, we need to make sure that people
    0:19:52 understand that there’s a spot for everybody in this space, right? And whether you’re going to
    0:19:58 start an AI startup, whether you’re trying to change jobs, whatever it is, and let’s figure out
    0:20:06 how we can help with that. What’s the feeling out there about how AI is impacting the workforce?
    0:20:13 I mean, that’s broad, and I’m not asking you to speak on behalf of anybody, right? But in the
    0:20:19 work you’re doing with folks and around the subject, I mean, it’s been, well, I don’t have
    0:20:24 to tell you that it’s always a hot subject, but particularly, again, in the past couple of years
    0:20:29 with generative AI and everything, it’s been, AI is going to take the jobs. No, AI is going to make
    0:20:35 us all actually more productive. No, wait, we’ve invested in AI, but it’s a long haul, so we’re
    0:20:40 not sure yet. What’s the feeling of the folks who you’re working with day in, day out?
    0:20:44 Yeah, it’s another phenomenal question. You’re just full of phenomenal questions today.
    0:20:50 The subject matter. Someone who knows what they’re talking about.
    0:20:57 So it’s interesting, because internally and externally, I have that part of the conversation
    0:21:01 a lot, because what we hear when we’re on the road is a lot of fear and trepidation, right?
    0:21:08 From legislators, there’s a lot of fear about the impact on workforce, but then they’re also
    0:21:14 trying to be responsive to this new technology, unlike other technologies that they feel that they
    0:21:21 let run loose and didn’t have control of, right? So there is a lot of trying to get folks comfortable,
    0:21:28 just even understanding what a GPU is and what’s possible. That’s oftentimes the starting point,
    0:21:36 even with high-ranking officials and really level-setting. With the community being honest,
    0:21:42 like, yes, there will be displacement, but most of this placement is going to come from people
    0:21:49 that actually are using AI versus AI itself as a thinking being that can go just take a job.
    0:21:56 And so trying to get folks to understand that all the innate things that are human, creativity,
    0:22:02 critical thinking, teamwork, all that stuff is really critical right now when you think about
    0:22:10 workforce and AI and what’s next in these next two to five years. And really helping students,
    0:22:17 as well as adults and companies, think about this productivity thing. Let’s go back to,
    0:22:24 hey, look, right now, human in the loop is the best thing. And how does this augment your skills?
    0:22:29 How does this augment productivity? How does this augment versus how does it replace?
    0:22:34 And once you can get people listening to what it is that you’re saying, again,
    0:22:38 kind of speaking the language that they speak, you can get past some of that.
    0:22:46 But understand everybody reads the headlines. So there’s a speed at which technology is changing
    0:22:54 things right now. The headlines are intended for a certain purpose. And you have to fight hard to
    0:22:58 counter that narrative that’s out there. And part of that is just by telling the truth.
    0:23:02 If you can look at the World Economic Forum, you can look at all the reports that come out.
    0:23:06 Yeah, there’s going to be displacement. There’s displacement with any big shift in innovation.
    0:23:12 But what industry does, what government does to help close those gaps is crucial.
    0:23:17 So we were talking about the states and how Alabama wants to be the data center of the South,
    0:23:22 and perhaps Mississippi has different ambitions. Is there kind of a wide variety of approaches
    0:23:26 states are taking when it comes to these things? Workforce development and thinking about
    0:23:35 economic development around data centers and chip fabs and all the other business things,
    0:23:40 avenues of business that AI and Accelerate Computing and all this stuff opens up.
    0:23:45 What’s it like sort of working with all these different states that must have different points
    0:23:51 of view? Yeah, like working from an NVIDIA perspective with these different states
    0:23:55 is actually refreshing compared to when I actually used to work for the state.
    0:24:04 All the states are uber competitive with each other. Some are more loud about what it is that
    0:24:10 they’re doing and what they’re known for versus others. They let everybody else think what they
    0:24:17 want to think, but they’re working on stuff that nobody knows about. So what’s been super refreshing
    0:24:23 about coming at it from an NVIDIA perspective is yes, every state is different. Every state has
    0:24:30 their idea about what they want to do when it comes to AI. At the end of the day, when we walk in,
    0:24:36 we’re talking about empowering the workforce, creating job opportunities, collaborative innovation,
    0:24:43 inclusive growth. When we speak those words, walking into a government, they tend to open up
    0:24:48 and let us know what it is they’re actually working on. So there’s some states, southern states,
    0:24:52 that we’ve actually been pleasantly surprised at how far along they are with their AI strategies.
    0:24:58 And then there’s other states where we walk in and we’re like, whoa, you guys, whoa.
    0:25:09 We need to think differently about how we even talk to you and what you’re actually ready for.
    0:25:17 When we successfully signed an MOU with California back in August, and it opened up a bunch of
    0:25:22 conversations throughout the other states, which is great. But understanding that signing the MOU
    0:25:28 is just the beginning of the conversation. In California, we’ve committed to working alongside
    0:25:34 them to try to train 100,000 people over the next three years. That’s going to take all the community
    0:25:40 colleges, 116 of those, that’s going to take the Cal State system, that’s going to take the UC system
    0:25:47 to get that done. That we have to work with the Cal HR system to help them understand how
    0:25:53 these AI skills are incorporated into job roles within the state for their state IT folks.
    0:25:58 So there’s a lot that there’s a lot of stuff that has happened in California. But then you go to a
    0:26:05 state like Mississippi and they have a platform where literally you actually just go to the
    0:26:11 platform and say, here’s what I’m doing. And it already tells you what you qualify for when it
    0:26:18 comes to a state workforce. Did we know that walking into Mississippi? Absolutely not. Did we
    0:26:23 know that the legislature was 100% bought into the AI strategy? No. So there’s, Mississippi is
    0:26:28 one of those states where you’re pleasantly surprised. Yeah, yeah, that’s amazing. But I know
    0:26:31 nothing about Mississippi to be clear. They just, I was trying to think of something. No, no, no,
    0:26:38 yeah. Near Alabama? Is it, they’re going to take that out. No, no, it’s fine. It’s fine. But what I
    0:26:44 told my team was it’s actually not surprising because when I was working for the state of
    0:26:50 California, a lot of the federal initiatives, we actually lost to Mississippi. So we lost the drone
    0:26:56 competitive opportunity. We lost the cyber competitive opportunity. And there were two
    0:27:02 schools in Mississippi. But what we’ve, what I found out is because everybody in the state has
    0:27:06 bought in. So as opposed to in California, we have Northern California versus Southern California.
    0:27:11 And you got to try to get the two together. And there’s just a lot of work there. Right. Mississippi,
    0:27:15 they’re like, yeah, we’re just all one team. Yeah, we’re doing. And yeah, let’s go get it.
    0:27:23 It helps. And so now discovering how they work versus what I understood
    0:27:31 when I was in the space, it’s actually awesome because it allows us to have a much deeper
    0:27:37 conversation where we’re not starting from what is a GPU. We’re like, oh, okay, how do we tap into
    0:27:44 what you guys are doing? How’d you guys build that platform? Pretty cool. Yeah. And how do we help
    0:27:51 your research get further? How do we, so every state is different. What we like to do is engage
    0:27:58 the governor’s office because that kind of gives cover for us. It also gives cover for the higher
    0:28:04 education systems to partner with us. And but then, you know, we absolutely look at, you know,
    0:28:10 where the opportunities are for alignment with NVIDIA to support sustainable development,
    0:28:14 to have real-world impact, and to help develop a future-ready workforce.
    0:28:20 Lou’s last question before we wrap up here. For a listener, maybe it’s a student, maybe it’s an
    0:28:27 educator or somebody working kind of in the education or local system who hears the message
    0:28:32 and wants to get on board, but doesn’t have the resources in front of them, kind of doesn’t know.
    0:28:37 The autonomous car hasn’t come to them just yet, but they’re listening, right? So they have you.
    0:28:44 What’s your advice to someone in those shoes? Is it just start using the tools? Is it research?
    0:28:51 I should stop trying to put answers in your mouth. No, no, no. Like, I am very clear what folks,
    0:28:56 when I’m on panels, when I’m doing keynotes, whatever, when I’m talking to my friends,
    0:29:02 this is not the time to be shy. And this is the time to jump in. Start understanding when I talk
    0:29:08 to small businesses. Start using tools like ChatGPT and see how it actually can transform your
    0:29:14 business. If you don’t know how to use it, then partner with the high school and bring in some
    0:29:18 high school interns and have them develop it. So then you’re building a workforce that way.
    0:29:25 But staying on the sidelines right now is not the best idea. So this is one of those where you say,
    0:29:31 come on in, the water’s fine. It’s just really deep, right? Because, you know, so, but it’s time to
    0:29:37 explore and get involved in the conversation. Absolutely. Lewis, for listeners who would
    0:29:42 like to learn more about some of the work that you’re doing, perhaps, you know, other places in
    0:29:46 NVIDIA related to workforce development and economic development and everything else we’re
    0:29:53 touching upon, where’s a good place for them to start? Yeah. So I work with the internal team. So
    0:29:57 every once in a while, there’s a blog post on the NVIDIA website, but really everybody can find me
    0:30:02 on LinkedIn. I try to stay pretty active on there. And if you don’t know the spell my name,
    0:30:08 just look up Meet Mr. Stewart and you’ll find me. Fantastic. Lewis, thank you again for making the
    0:30:13 time and best of luck on the work you’re doing. It’s kind of, I don’t know, it’s one of the core
    0:30:17 reasons I think that we’re all doing all this is to make a better world for everybody. Yeah,
    0:30:35 thank you. It’s been a pleasure and hopefully we get to chat again.
    0:30:44 So,
    0:31:09 , you.
    0:31:12 (upbeat music)
    0:31:20 [BLANK_AUDIO]

    In this episode of the NVIDIA’s AI Podcast, Louis Stewart, head of strategic initiatives for NVIDIA’s global developer ecosystem, discusses why workforce development is crucial for maximizing AI benefits. He emphasizes the importance of AI education, inclusivity, and public-private partnerships in preparing the global workforce for the future. Engaging with AI tools and understanding their impact on the workforce landscape is vital for ensuring these changes benefit everyone.

    https://blogs.nvidia.com/blog/workforce-development-ai/

  • How the Department of Energy Is Tapping AI to Transform Science, Industry and Government – Ep. 236

    AI transcript
    0:00:10 [MUSIC]
    0:00:13 >> Hello, and welcome to the NVIDIA AI podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:18 This past October, leaders from the public sector
    0:00:22 joined NVIDIA customers and partners at AI Summit in Washington, D.C.
    0:00:24 for three days of connection and discussion
    0:00:28 around all of the innovative and meaningful work being done with AI.
    0:00:32 One of the panels at the summit, AI for Science, Energy, and Security,
    0:00:36 focused on how artificial intelligence is transforming scientific discovery,
    0:00:38 economic growth, and national security,
    0:00:40 and how the US government can lead the way
    0:00:43 in developing safe and trustworthy AI
    0:00:45 to address national and global challenges.
    0:00:49 It’s a great discussion, and I encourage you to watch it via NVIDIA on demand.
    0:00:51 But even better than that,
    0:00:54 we’re fortunate enough to have the chance to delve a bit deeper into the role
    0:00:57 the US Department of Energy is playing in AI development.
    0:01:00 Spoiler alert, they’re doing a lot more than you might think.
    0:01:04 With us now is Halina Fu, Director of the Office of Critical and Emerging Technology,
    0:01:07 CET, at the US Department of Energy.
    0:01:11 Director Fu previously served as Director for Technology and National Security
    0:01:13 at the White House National Security Council
    0:01:17 as Director of International Science and Technology Cooperation
    0:01:20 and Trusted Research for the Office of Science at the Department of Energy
    0:01:24 and at the White House Office of Science and Technology Policy.
    0:01:27 Director Fu’s credentials extend much further than that,
    0:01:29 but I’ll leave it there so we can welcome her onto the show.
    0:01:34 Halina Fu, thank you so much for taking the time to join the NVIDIA AI podcast.
    0:01:36 Great to be here, Noah. Thanks for inviting me.
    0:01:41 So perhaps we can start with the CET and what your role is as Director.
    0:01:43 Tell us a bit about that.
    0:01:48 Sure. So the Office of Critical and Emerging Technologies, or CET,
    0:01:51 because we love a good acronym, is a new office at the department.
    0:01:54 It was launched in December of 2023.
    0:01:58 So we’re less than a year old, coming on that one-year anniversary.
    0:02:02 And our senior leadership at the department
    0:02:05 really wanted to have a central node within the department
    0:02:09 that was focused on leveraging capabilities and expertise
    0:02:12 across the department and at 17 national labs
    0:02:15 in specific areas of critical and emerging technologies.
    0:02:20 And our office focuses on four specific technology areas.
    0:02:24 AI, microelectronics, quantum information science,
    0:02:25 and biotechnology.
    0:02:29 And these four we really see as foundational technologies
    0:02:33 that enable so many other critical and emerging technologies.
    0:02:34 Absolutely.
    0:02:37 And so how does that fit into the broader scope?
    0:02:39 We were chatting for a second before we started recording.
    0:02:41 I was watching the panel you were on,
    0:02:44 and there were opening remarks from the Secretary of Energy as well.
    0:02:48 So again, it’s a great panel, and I learned a lot by watching it.
    0:02:49 And in particular, I was struck.
    0:02:53 I think there was a joke made on the panel about calling it Department of Everything.
    0:02:57 I was struck by how much the DOE broadly does.
    0:02:59 So maybe you could take a second, if you would,
    0:03:01 and kind of speak to a little bit to that.
    0:03:04 Yeah, I think that’s a joke that our Secretary made.
    0:03:08 But actually, many Secretaries of Energy have made this joke
    0:03:11 because as everyone comes into the department,
    0:03:15 they see the vast scope of what we actually do here at DOE.
    0:03:20 So as you said, many people think about DOE as the clean energy department.
    0:03:23 And we are, but we do do so much more.
    0:03:27 We fund basic scientific discovery and research.
    0:03:31 We steward scientific infrastructure all across the country.
    0:03:35 And these are some of the world’s most powerful x-rays, particle accelerators,
    0:03:38 colliders, sort of really, really big science.
    0:03:41 We advance energy research applications,
    0:03:46 which is the part that I think many people know and touch when it comes to DOE,
    0:03:51 as well as, you know, we’re the sector risk management agency for the electricity sector.
    0:03:54 But we also play a key role in the national security space
    0:03:57 through the National Nuclear Security Administration.
    0:04:00 And on top of that, we steward the 17 national labs,
    0:04:03 which are located all across the country.
    0:04:06 Many of them were birthed out of the Manhattan Project,
    0:04:12 but today they serve as that scientific infrastructure backbone of the United States
    0:04:17 and the capability not just for DOE, but for the broader U.S. government.
    0:04:21 And so I talked a little bit about that scientific infrastructure.
    0:04:24 DOE has 34 scientific user facilities.
    0:04:28 Now, these are places where scientists within the United States,
    0:04:32 from academia, from industry, as well as international scientists,
    0:04:37 they can access these facilities on a peer-reviewed basis, of course.
    0:04:41 These are the kinds of particle accelerators, molecular foundries,
    0:04:45 supercomputers, that the department stewards for the country.
    0:04:52 So when I think about, and I think a lot of the conversations that we’ve been having on the podcast,
    0:04:56 when I think about the relationship between AI-accelerated computing,
    0:05:00 these other advanced technologies and emerging technologies and energy,
    0:05:02 I sort of think of two different things.
    0:05:08 And I’m curious how you think about it, if you think about it in similar buckets or not.
    0:05:14 One is compute uses energy. AI, all this stuff, the bigger, the faster,
    0:05:16 the more powerful, the more energy. So there’s that.
    0:05:20 And there’s been talk lately about, is nuclear going to play a role?
    0:05:23 All these different things, how does clean energy or renewable energy,
    0:05:27 how does all of this relate to powering the GPs, the data centers, all that stuff?
    0:05:35 And then the other side of it is, how can all of the technology inform finding new sources of renewable energy,
    0:05:41 making it more efficient, all of that good stuff to make the whole process of creating and using energy
    0:05:46 better for everybody? I can’t even imagine how complex these things are when you think about them.
    0:05:50 But how do you describe the relationship between energy and specifically what you’re doing,
    0:05:55 CET and DOE, and the relationship to all of this technology stuff that we’re talking about?
    0:06:00 I think that’s a great question. And really, the Department of Energy is sitting at that intersection.
    0:06:08 This intersection of AI as applied to the energy sector and the energy availability for AI to power
    0:06:15 the data centers that these AI models are trained on. And so we are absolutely laser focused on both
    0:06:21 of these issues. And I think one really concrete example of where this all comes together really
    0:06:27 is in some of the partnerships that we’ve been able to develop with industry to drive energy
    0:06:34 efficiency of computing. This is the story behind the Exascale Computing Project and the initiative
    0:06:40 that DOE has been very, very involved in over the last decade. And the very beginnings of that
    0:06:48 was this presupposition that to actually get to Exascale Computing, you would need to make huge
    0:06:54 advances in energy efficiency. Can I stop you for a second and just ask you to define Exascale
    0:07:00 Computing for everyone or the project, I should say? Sure, sure. So this is something that was
    0:07:06 really done in deep partnership between the Office of Science at the Department of Energy
    0:07:12 and the National Nuclear Security Administration. Okay. So this is a really a committed partnership
    0:07:20 that was begun back in 2016 and running through 2024. And this was really about how we could build
    0:07:27 some of the most powerful and fastest supercomputers in the world, right? And so Exascale is really
    0:07:33 referring to how quickly how many floating point operations per second could actually be performed,
    0:07:38 right? And so that is one quintillion floating point operations per second or one exathlon.
    0:07:43 I realize I said define as in like, what did I mean? But really, I just wanted you to talk
    0:07:49 about the project, but that’s a great definition. So thank you. Back at that time, DOE was already
    0:07:55 investing in advanced supercomputing, but we recognized that to really push the boundaries
    0:08:01 of science, energy applications and national security, we would need even faster and more
    0:08:07 powerful supercomputers to do the kinds of exquisite modeling and simulation that would be needed for
    0:08:14 the nuclear stockpile, but also model the climate. And so there were so many different applications
    0:08:18 in that space. But really, at the time, it was an energy efficiency question. How could we build
    0:08:25 these kinds of huge, huge supercomputers? And how would we manage the power envelope and the cost
    0:08:31 associated with power in these supercomputers? And so way back during that time, the goals that
    0:08:38 were set for Exascale were not set by what was technically possible at the time, but were really
    0:08:44 set by just a number. How much we were willing to pay over the life of the system, and then just
    0:08:51 divided it by five. And that was the goal. And really, at the time, I think many folks thought
    0:08:58 we were crazy, but I think that’s really the magic of DOE. I don’t want to call it magic,
    0:09:03 it’s not magic, it’s just the ability of the department to partner deeply with industry,
    0:09:10 to accomplish things at scale. I think that is really how I would think about where DOE fits
    0:09:15 in the larger AI ecosystem. I think one other thing that would be really helpful to kind of
    0:09:22 weigh out. So I talked about AI and energy and energy for AI as one intersection where DOE sits.
    0:09:29 I think the other one that is not very widely known, but kind of goes to all of the different
    0:09:36 parts of DOE, is that we sit at this intersection of open science and classified science. So open
    0:09:44 science and national security. I’m imagining you walking a tightrope with the balance beam thing,
    0:09:50 just saying classified science kind of brings that area. There’s many national security applications
    0:09:56 of science, of course. And much of that really is dependent on advances in open science. And I
    0:10:05 think the fact of DOE as a mission-driven R&D agency that is able to bridge these two
    0:10:12 arenas and to do that at scale, I think scale, again, is an important demarcator there, is really
    0:10:17 what makes working at DOE so exciting. And frankly, as we bring it back to our Office of
    0:10:23 Critical Immersion Technologies, the four technology areas I mentioned, AI, microelectronics,
    0:10:29 quantum information science, and biotechnology, they’re all dual-use technologies. So I think really,
    0:10:35 as we think about both opportunities to advance these technologies and opportunities to manage
    0:10:42 the risks of those technologies, you really do need that duality of approach where the science
    0:10:47 is going and what the national security implications are going to be to address this effectively.
    0:10:55 Are there specific AI-related initiatives going on within DOE, within CET, that you can talk about?
    0:11:01 Sure, yes. And we’re very excited about the opportunity space of AIS, if you haven’t caught
    0:11:09 on already. We do have a very exciting proposed initiative called FAST, Frontiers in AI for
    0:11:18 Science, Security and Technology, which we see as a real opportunity space for a step change in U.S.
    0:11:25 government capability in AI. We all see the advancements that industry is making. We need to
    0:11:31 be able to harness those advancements for our mission space. And I think the recently issued
    0:11:38 national security memorandum on AI just last week really directs departments and agencies in the
    0:11:44 national security space to do just this. But we also think that there’s even more to do, right?
    0:11:52 So FAST is sort of organized in four general buckets, one around data, one around advancing
    0:11:59 compute, one on developing new kinds of models, and the fourth around the application space. So how
    0:12:05 do we actually use these models to solve things in the real world? Yes, there’s been, I actually was
    0:12:10 recording a podcast earlier this week with somebody and used this same line. So forgive me, listeners.
    0:12:17 But there’s kind of this sense that 2022, 2023, in the generative AI space in particular, we’re
    0:12:23 sort of the years that LLMs and the idea of models kind of exploded into the mainstream. And businesses
    0:12:28 and other organizations started to take notice and really think about what we need to do, invest,
    0:12:34 we need to do things. And then 2024 now has been the year where folks in the application space
    0:12:40 are really starting to hone in on, okay, how do we take all this power and translate it into
    0:12:45 something that is, you know, practical, useful, and easy for an end user to kind of pick up and go
    0:12:51 with. Is there a similar sort of timeline or are you kind of in the same sort of headspace
    0:12:57 with relative to building the applications and that focus now in your government work or is it
    0:13:02 a different, are you in a different spot? So I think that’s a really interesting question because
    0:13:09 when it comes to agency deployment of industry models, many of these are
    0:13:15 large language models, right? And so we see tremendous efficiency, potential efficiencies
    0:13:20 within the department on utilizing these. And we have some things that we’re beta testing now
    0:13:26 and excited about them. But what we’re talking about in FAST is really around how we harness
    0:13:34 the scientific data that DOE generates, which is some of the most unclassified and classified
    0:13:40 scientific experimental data in the world. You can’t even imagine, yeah. How we really advance
    0:13:47 computing in this space in a way that we were able to do through the exascale computing project,
    0:13:54 but really think further beyond the horizon. Yeah. And then the models themselves, we really think
    0:13:58 that much of industry obviously is focused on large language models. And there’s been a lot of
    0:14:04 exciting developments. And even those models are becoming much better at scientific reasoning.
    0:14:09 Really, we’re seeing tremendous advances in that space. And that really excites us because that
    0:14:15 means the time to discovery, whether it’s for in discovery science or in the energy applications
    0:14:21 or in national security, that time is going to be cut so significant. But we also think
    0:14:26 that there’s going to need to be new kinds of models, additional kinds of models that are not
    0:14:31 language-based, but are graph neural networks or physics-informed models
    0:14:37 that are the kinds of models that need to be able to learn from the kinds of data that we hold.
    0:14:44 And that can be then shown to translate into the physical world, which is why we have laboratories.
    0:14:50 So I think this is where we get so excited because we think the opportunities for partnership
    0:14:57 with industry are really enormous. And a place where there hasn’t yet been the same kind of
    0:15:03 attention and focus more broadly. Because frankly, these are things that are in the realm of
    0:15:07 science, right? Like government investment in science. That’s why there is that sort of
    0:15:12 public-private partnership opportunity. If I could just add one thing though.
    0:15:17 Please. Because there’s been a lot of attention when we talk about AI for science. There’s been
    0:15:23 a lot of attention, for example, on AlphaFold. And I think what it’s important to note there
    0:15:31 is AlphaFold’s statistical models, they’re trained on data from experimentally determined protein
    0:15:37 structures that were only made possible by the use of DOE’s unique large-scale light sources.
    0:15:45 And the protein structure data that was used and that is free and open to the public
    0:15:51 are stored on DOE-funded protein data bank, which gets funding from other U.S. government
    0:15:56 departments and agencies. But I think that’s an example of where there is that public-private
    0:16:02 dimension to even industry advancements. My guest today is Helena Fu. Helena is the
    0:16:07 director of the Office of Critical and Emerging Technologies at the U.S. Department of Energy.
    0:16:11 And as we’ve been talking about, the Department of Energy, the work that they’re doing, the work
    0:16:16 that director Fu’s team is doing as well, touches so many different parts of private life, public
    0:16:22 life, different parts of the government working with science and industry and everything else.
    0:16:28 So many different places. Let’s kind of land on a note of opportunity, and I’m not going to ask
    0:16:33 you to predict the future. But what are some of the spaces that you in particular, your teams,
    0:16:39 are really excited about when it comes to the applications of large language models, AI, any
    0:16:44 of the technology we’ve been talking about, or other things that we haven’t gotten to yet?
    0:16:50 Thanks for that question. I think I might answer in two timescales. The first is the near term.
    0:16:55 So we have a number of activities already underway that are really seeking to harness
    0:17:02 these powerful tools. One example here is really around smart grids and using AI. For example,
    0:17:08 our grid deployment office funded Portland General Electric to deploy utility data smart meters,
    0:17:14 which have GPUs inside of them to customers. And we think that there’s an opportunity also for
    0:17:20 local models that help disaggregate loads like electric vehicles, heat pumps, and that provide
    0:17:28 utilities insight while also protecting customer privacy. Our Office of Policy and Pacific Northwest
    0:17:34 National Lab are also leading this voltaic initiative, which is really around how we can harness
    0:17:39 large language models for permitting and how we find efficiencies in that process.
    0:17:43 Permitting as in granting permits to citizens to do things?
    0:17:51 Permitting process like environmental permitting reviews, NEPA. And in fact, one of the things
    0:17:58 that we were able to do just earlier this year was to take the entire corpus of NEPA data over the
    0:18:06 last decade or so and make that AI ready and available and is now open and available to
    0:18:12 the public for researchers to work on for industry to develop new tools around. We are working with
    0:18:18 other agencies in the Federal Permitting Council to see how they can potentially use this tool
    0:18:22 for their permitting processes. So that’s exciting. And that’s at the federal level,
    0:18:27 but the voltaic initiative is actually also focused on expanding to state and local
    0:18:32 permitting because there’s even more variants across state and local ordinances, for example,
    0:18:40 for EB charging or for any other kind of siting. Yeah, the mention of the grid and the smart grid
    0:18:46 meters with the GPUs and then you’d be able to compute on the fly. I mean, that hits home for me.
    0:18:51 I’m kind of the stereotypical Northern California. I’ve got an EV, the whole thing,
    0:18:55 but it really makes me think about all of the opportunities in the physical world
    0:19:01 that if the infrastructure is in place and you’ve got this data to work with, which obviously is
    0:19:06 huge and I can’t even imagine the troves of data that are around waiting to be processed and put
    0:19:13 to use, but you can deploy things at the edge like that to a customer’s, in my case, to where
    0:19:19 I live and a residential person can have that smart meter or can have the ability to shift
    0:19:25 the loads locally. And I can only imagine the possibilities that opens up for grid resilience
    0:19:32 and all of those things. That’s exactly right. And I think moving beyond the short term and into
    0:19:38 that medium and long-term time scale, we’re so excited about the opportunity of AI applied
    0:19:43 to materials discovery, which I think has applications across science, energy and national
    0:19:49 security, right? It’s enormous, yeah. We often talk about this example from PNNL and Microsoft
    0:19:54 to look at the universe of potential materials and then fabricate a battery that uses 70%
    0:20:01 less lithium. But I think that’s just only the tip of the story, right? Obviously, battery materials,
    0:20:07 carbon capture sorbents, hydrogen catalyst, there’s so much opportunity to discover new materials
    0:20:15 that will have both impacts on like, abundance and affordability of energy, but also for strategic
    0:20:22 technologies like critical minerals and hypersonics, right? There’s many, many applications across
    0:20:28 the spectrum. Similarly, in the physical sciences, right, there’s obviously a lot of work already
    0:20:34 underway on physics-informed models for climate. We need to be able to bridge climate and weather
    0:20:39 across multiple timescales. And that will have huge implications for disaster response and
    0:20:45 preparedness and just climate modeling more generally. Yes, yeah. As you said, that’s just
    0:20:51 kind of the tip of the iceberg. And the idea of 70% more efficient car batteries built from
    0:20:56 renewable energy is sort of mind-blowing to me anyway, but that really is just kind of the
    0:21:01 beginning of the story, right? We’re in the early innings here in general. Before we wrap up,
    0:21:07 any closing thoughts, people listening out there, perhaps there are organizational leaders,
    0:21:13 IT leaders, people in there, whether it’s private or public sector, who are really starting to work
    0:21:19 on, they understand the opportunity, they’ve used some AI tools, they’re thinking in that same way,
    0:21:24 right, that kind of short and kind of midterm timescales. Advice you would give to somebody who
    0:21:30 is, whether their organization is small or governmental scale, who’s trying to kind of
    0:21:37 lead that initiative to take advantage of AI in a thoughtful and sort of sustainably minded way.
    0:21:43 Any words of advice from your experience? Yeah, I am also the acting chief AI officer
    0:21:48 for the Department of Energy. As I think about how the Department of Energy is going to utilize
    0:21:54 AI and work very closely with our chief information office on this, we think really,
    0:22:01 and this goes for within DOE, also companies, as they think about adopting AI, we really need
    0:22:05 to innovate up that ladder of trust. So we really need to think about, I mean, when I think about
    0:22:10 this for DOE, I think about what are the immediate use cases? How do we make sure that they’re not
    0:22:16 rights and safety impacting? What are the processes we have in place to manage those risks?
    0:22:22 But I think the overall advice here really would be, this is a transformative tool.
    0:22:28 It is a tool in the tool bell. It’s very powerful. We should all figure out how to use this to the
    0:22:34 best of our ability to amplify and augment the work that we’re doing. So we are thinking about
    0:22:39 this day in, day out at DOE. We think that partnerships are going to be essential to our
    0:22:46 success when it comes to applying AI for our broad science, energy, and national security mission.
    0:22:53 Fantastic. For listeners who would like to learn more about any of the many things that you covered,
    0:22:56 what the Office of Critical and Emerging Technologies is doing, the Department of
    0:23:02 Energy is doing, all of that good work. Where are some places they can go online to get started?
    0:23:05 Yeah, well, the Office of Critical and Emerging Technologies has a website,
    0:23:10 and that is a place where we really look to point to all of the different and amazing
    0:23:15 activities happening across the department and the national laboratory. So that’s a really good
    0:23:21 resource to learn more about our Frontiers in AI for Science, Security, and Technology initiative.
    0:23:28 And also, TESPEs that are available at DOE, training opportunities at the national laboratories,
    0:23:35 foundation models that we’ve already developed in partnership with others, and so much more.
    0:23:40 Fantastic. Well, again, thank you so much for taking the time, Director Fu, to come on the podcast
    0:23:46 and kind of extend the conversation that you started at the AI Summit, and hopefully we can do it
    0:23:50 again down the line. In the meantime, I’m really excited to follow the work that you and your teams
    0:23:55 and the rest of the folks in the government are doing to make life better for all of us. So thank you.
    0:24:07 Thank you. Thanks so much.
    0:24:18 [Music]
    0:24:46 [Music]
    0:24:56 [BLANK_AUDIO]

    Helena Fu, director of the DOE’s Office of Critical and Emerging Technologies (CET) and DOE’s chief AI officer, discusses the latest groundbreaking efforts with AI that are transforming national security, infrastructure, and scientific discovery. With oversight of 17 national labs and 34 facilities, the DOE is at the forefront of AI research and development.

  • Zoom CTO Xuedong “XD” Huang on How AI Revolutionizes Productivity – Ep. 235

    AI transcript
    0:00:10 [MUSIC]
    0:00:13 >> Hello, and welcome to the Nvidia AI Podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:18 Zoom became a household name in 2020,
    0:00:19 as it rose to prominence as
    0:00:23 the go-to video conference platform during the COVID pandemic.
    0:00:26 Since then, the company has not only been refining their video technology,
    0:00:30 but also helping us all rethink the way we approach work in
    0:00:33 the era of digital communications and AI.
    0:00:35 At Zoom Topia this past October,
    0:00:39 Zoom took the wraps off of a number of new AI-first products and initiatives,
    0:00:42 all in service of the company’s mission to deliver
    0:00:45 an AI-first work platform for human connection.
    0:00:49 Here to discuss everything from Zoom’s approach to federated AI and
    0:00:55 AI agents to the future of how we all live and work with technology is Dr. XD Huang.
    0:00:58 XD is Zoom’s Chief Technology Officer and has
    0:01:01 a prolific background in artificial intelligence coming to Zoom from
    0:01:05 Microsoft where he founded the Speech Technology Group in 1993,
    0:01:10 and most recently served as Azure AI CTO and Technical Fellow.
    0:01:13 XD is an IEEE and ACM Fellow and
    0:01:16 an elected member of the National Academy of Engineering,
    0:01:19 and the American Academy of Arts and Sciences.
    0:01:22 Most importantly, he’s with us right now, so let’s get to it.
    0:01:27 XD Huang, welcome, and thank you so much for joining the NVIDIA AI podcast.
    0:01:29 Thank you, I’m glad to be here.
    0:01:33 So we’re recording this on the Friday immediately following Zoom Topia,
    0:01:37 where Zoom basically announced to the world that you’re going all in on AI.
    0:01:39 We want to hear all about the new stuff of course,
    0:01:43 but first maybe we can set the scene a little for the audience.
    0:01:48 Can you tell us broadly about Zoom’s approach to AI and AI in the workplace?
    0:01:52 Yes, I think this is really the most exciting time.
    0:01:57 I started working on AI since I was a graduate student.
    0:02:00 This has been over 40 years.
    0:02:06 Now Gen of AI really, really transformed how activity is going to be.
    0:02:08 So Zoom is in the forefront.
    0:02:13 We have provided amazing media conferencing for the whole world.
    0:02:18 It’s a household name that everyone understands what Zoom is about.
    0:02:21 – It’s a verb even at this point, right? – Yes, that’s right.
    0:02:28 So we’re now facing even more exciting opportunities in the world.
    0:02:32 So meeting is one of the most important business functions,
    0:02:39 but we want to expand that capability for people to work happily on Zoom platform.
    0:02:44 So Zoom workplace is going to take advantage of Gen of AI capability.
    0:02:49 We believe that Gen of AI is going to really provide an exciting opportunity.
    0:02:55 I reflected my own journey where I started writing my first master thesis
    0:03:00 in Beijing’s Qinghua University, the first paper I was using typewriter.
    0:03:03 I love the expensive liquid paper.
    0:03:09 In China, in Beijing’s Qinghua University, that was 1983-82.
    0:03:12 I remember liquid paper was expensive, it’s a luxury.
    0:03:17 By this type, any letter, I have to really use the liquid paper.
    0:03:19 I could rather interrupt that.
    0:03:26 And when I wrote my book, Smoke and Energy Processing with my colleague and Microsoft,
    0:03:30 I was fortunate we had the Microsoft Word.
    0:03:36 And even at that time, Word kind of accommodated 800 pages document.
    0:03:37 – Too big? – Too big.
    0:03:43 So we have to separate the file for each chapter,
    0:03:45 but Microsoft Word did a wonderful job.
    0:03:48 It’s hard for me to imagine without Microsoft Word,
    0:03:51 we have to use a typewriter to write that.
    0:03:56 How hard it’s going to be, because we have a lot of graphs, math, and references.
    0:03:59 – Right. – Time passed quickly.
    0:04:04 One of my colleagues in Microsoft wrote his book
    0:04:07 with the systems of GPT-4 charging.
    0:04:09 It’s just amazing.
    0:04:13 Not for the activity, it really pushed everything to getting level.
    0:04:15 So you can see that reflection journey.
    0:04:18 To stop you for a quick second,
    0:04:22 if you can think back to when you were writing on the typewriter,
    0:04:24 could you have imagined where we’d be now?
    0:04:27 You’ve been in this field for a long time, so perhaps you could.
    0:04:30 But I’m just curious, you know, if 30 years ago,
    0:04:34 where we’re sitting today, your colleague using GPT-4 to help write a book,
    0:04:36 is that something you thought about back then?
    0:04:40 Yes, I selected a speech recognition as my thesis,
    0:04:42 and the pages should be in the bookstale.
    0:04:45 At that time, I had only IBM PCS.
    0:04:47 I don’t know if you know what that means.
    0:04:52 – Yeah, I remember. – Plus, a few Apple II computers.
    0:04:56 – Yeah, I had a 2E growing up. – Yeah, that was all we had.
    0:05:00 – Right. – And I actually told myself,
    0:05:03 “If I could let the computer understand spoken language,
    0:05:06 I could retire.” – Right.
    0:05:09 – 40 years passed. I’m excited then ever.
    0:05:12 So I’m not retiring. The frontier is Bushla.
    0:05:15 – Yes. – It’s not about speech recognition.
    0:05:20 In Microsoft, we were the first to reach the humanity
    0:05:24 on the most difficult speech test switchboard in 2016.
    0:05:26 Most people didn’t believe we could have done that.
    0:05:27 Yes, we did.
    0:05:32 Now, China GPD really redefined and opened up
    0:05:34 the imagination for the whole world.
    0:05:36 – Right. – I think Oakland did a great job
    0:05:40 to really redefine the new frontier.
    0:05:42 – We shared a story in YZOOM. – Yes.
    0:05:45 You know, going to grab this opportunity
    0:05:48 to redefine productivity. – Yeah.
    0:05:52 Every era of computing created productivity lead.
    0:05:55 Microsoft revolutionized desktop computing,
    0:05:58 office, unquestionably,
    0:06:01 as the productivity leader for desktop computing.
    0:06:02 That’s why I shared with you.
    0:06:04 When I wrote the book,
    0:06:07 spoken language processing with my colleague and Microsoft,
    0:06:09 we loved Microsoft Word.
    0:06:13 When the work came together, Google took advantage of it.
    0:06:16 They enhanced productivity
    0:06:19 to support multiple people working on the same docking.
    0:06:23 As we know, Google Docs and the Sheet Slides,
    0:06:27 they all really supported the cross-pin collaboration.
    0:06:29 Right, the collaboration, yeah.
    0:06:32 That took notice from almost everyone.
    0:06:34 – Right. – One viewer
    0:06:37 is an incremental info, in my opinion.
    0:06:39 – How so? – Now, generally,
    0:06:41 we are all in the same leveling field.
    0:06:44 Whether it’s Microsoft, Google, or Zoom.
    0:06:47 Of course, Zoom has a unique advantage.
    0:06:48 You know, the most important business function
    0:06:51 to connect people on the meeting.
    0:06:52 We are the leader.
    0:06:55 But just having that meeting capability
    0:06:56 would be insufficient.
    0:06:58 If you think about the work,
    0:07:01 we have probably a few key functions.
    0:07:05 Like, one is to consume information.
    0:07:06 We need to learn.
    0:07:09 This is using the human body to read,
    0:07:13 to satisfy our own curiosity.
    0:07:15 Generally, I can help you to read
    0:07:20 500 or 800 page books, like one my colleague and I published.
    0:07:22 800 pages in one page.
    0:07:26 So, magically, YoungDai can redo this
    0:07:31 and create an amazing amount of learning for almost everyone.
    0:07:34 We just do not consume information.
    0:07:35 That’s one of the important functions.
    0:07:37 We also need to communicate.
    0:07:39 Young friends bring people along.
    0:07:43 So, YoungDai can help you to compose the draft
    0:07:44 because we understand your need.
    0:07:49 So, those two most important human fundamental capabilities
    0:07:51 to read, to write, to speak
    0:07:54 are a group to write and to speak on the same.
    0:07:57 Right. Consume information to communicate, yeah.
    0:08:01 They are going to be really, really fundamentally helped.
    0:08:05 So, magically, we can take that capability.
    0:08:08 We design productivity,
    0:08:12 not just both those capabilities to the existing software.
    0:08:15 That’s an opportunity to really, really possess.
    0:08:18 Right. The approach for us to work on
    0:08:22 productivity suite and the approach we are addressing AI,
    0:08:24 there are three key ways about the highlight.
    0:08:27 Now, talk about this in ZoonTopia.
    0:08:30 Let me just explain this in details.
    0:08:32 The first thing I want to really highlight is
    0:08:34 our federally-AI stack.
    0:08:38 So, we integrate the best from leading AI companies,
    0:08:42 like Floppy, OpenAI, Meta,
    0:08:45 (indistinct)
    0:08:47 across leading open source opportunities.
    0:08:51 We operate this with web search leaders,
    0:08:53 they get a new one for complexity.
    0:08:56 Okay. So, we federally, all of them together,
    0:09:01 in addition to our own proprietary small language model
    0:09:03 we are trained, we are developing,
    0:09:06 that is really reaching amazing capability.
    0:09:10 We appreciate that the small language model,
    0:09:11 because of the scaling law,
    0:09:14 need to really work together
    0:09:17 with this amazing cloud-based large language models.
    0:09:19 So, we have this unique approach
    0:09:22 who combine them together seamlessly behind
    0:09:27 to support the productivity of each individual state.
    0:09:31 What does the small language model do in the stack?
    0:09:32 How does it differ from what
    0:09:34 you’re tasking the large language models with?
    0:09:38 We are training the small language model,
    0:09:40 like everyone else, to train large language model.
    0:09:42 It’s just a whole other task.
    0:09:44 Okay. In addition to that,
    0:09:49 we’re also incorporating each individual’s unique contacts.
    0:09:52 So, we can really make it that personalized.
    0:09:56 In addition to consuming this massive amount of tokens,
    0:09:58 right, always somewhere.
    0:10:02 So, if I’m using Zoom’s AI features
    0:10:04 and I give permission, Zoom can basically
    0:10:07 just ingest all of my conversations,
    0:10:11 all the meetings I have, the voice conversations,
    0:10:14 the documents, the chats, all of that and use that all,
    0:10:18 as context for the generative AI going forward.
    0:10:21 This is the feature that is coming together
    0:10:23 through our AI companion.
    0:10:24 Oh, got it.
    0:10:27 AI companion, to put it all, is horizontal, generic,
    0:10:29 not personalized.
    0:10:29 Okay.
    0:10:31 It’s a custom AI companion
    0:10:35 that we will introduce later this next year.
    0:10:36 Next year, okay.
    0:10:38 We’ll actually incorporate the ability
    0:10:42 for anyone to customize and the personalized.
    0:10:43 Right.
    0:10:45 This is actually a very powerful opportunity
    0:10:49 for the small language model running on the devices
    0:10:53 to really augment what the large language models cannot offer
    0:10:56 because you don’t understand your personal needs,
    0:11:00 your learning pattern, your writing pattern, etc.
    0:11:02 So, I just want to really first highlight
    0:11:07 our federal AI stack is unique, very unique in the industry.
    0:11:09 Unlike many other productivity companies,
    0:11:10 they use only one.
    0:11:11 Right.
    0:11:13 And so for the audience who might not be familiar,
    0:11:16 federated AI, a federated stack,
    0:11:19 does that essentially just mean that the system can choose
    0:11:23 which LLM to prompt depending on the situation?
    0:11:25 Or what does federated mean, the way you’re using it?
    0:11:28 There are multiple ways to federate.
    0:11:30 The way to federate the large language models
    0:11:33 and the small language model is a new frontier.
    0:11:37 The way we are federating this is different from federated
    0:11:37 learning.
    0:11:38 Okay.
    0:11:40 That’s what you’re trying to really combine
    0:11:45 multiple models together to form this powerful capability
    0:11:48 that can preserve the client, obviously.
    0:11:51 What we’re doing is we can choose based on different workloads.
    0:11:56 Because AI complaining to collo is almost like a super agent.
    0:11:59 That is trying to understand with different modality,
    0:12:03 different memory, expand, etc.
    0:12:08 So, we can choose what model is best for different tasks.
    0:12:11 We can also combine different models together.
    0:12:14 And we can reflect like a chain of thought, we think.
    0:12:16 And we perform the same tasks based
    0:12:19 on what we have learned from a small language model,
    0:12:20 for example.
    0:12:23 So, if a small language model can perform a task
    0:12:25 with very little work, then we start there.
    0:12:26 It’s sufficient.
    0:12:28 So, it’s a very sophisticated system
    0:12:32 that can actually orchestrate multiple models together.
    0:12:36 This has been developed and pushed by Zoom AI talents.
    0:12:39 So, this is a very unique approach
    0:12:41 that set us apart from almost anyone else.
    0:12:42 Yeah.
    0:12:43 You use the word agent.
    0:12:47 And as we’re recording this, AI agents are…
    0:12:48 There’s a lot of buzz.
    0:12:50 I’m hearing a lot of buzz around the word agent
    0:12:54 and the concept of agentic AI, which isn’t new.
    0:12:56 But as it’s come to the fore lately,
    0:12:58 can you talk a little bit about what that means,
    0:13:01 what the idea of an AI agent is, kind of broadly,
    0:13:04 but then specifically to how Zoom is using it?
    0:13:06 And I want to come to that later.
    0:13:07 You want to come to that later?
    0:13:08 OK.
    0:13:11 Nick, why we approach AI differently
    0:13:14 with three key ways that are different
    0:13:16 from traditional approach, right?
    0:13:20 So, the first one, if you think about traditional
    0:13:22 or the typical suite, most of the companies
    0:13:26 are using one model, either OpenAI or Gemini,
    0:13:29 to augment what they do.
    0:13:33 They bolded their capability to the existing software.
    0:13:34 Right.
    0:13:37 So, on the back end, they are mostly using
    0:13:40 one very good generation model,
    0:13:42 either Gemini or OpenAI.
    0:13:43 So, approaches different.
    0:13:48 We massage OpenAI, anthropic, Gemini, matter,
    0:13:51 and our own smaller model together
    0:13:53 to offer a matched performance.
    0:13:55 So, that’s the number one I want to really highlight.
    0:13:58 Of course, we’re also integrating our partner,
    0:14:00 our Plexi, for the amazing web search results.
    0:14:01 For the web search.
    0:14:05 So, whether it’s web, question, or work, question,
    0:14:07 and in the future, personal question,
    0:14:08 we can massage them together.
    0:14:11 That’s what we are pushing to differentiate
    0:14:14 the Zoom AI through our favorite approach.
    0:14:15 That’s the number one.
    0:14:16 OK.
    0:14:20 Number two, our user experience is AI first.
    0:14:22 This is what I call AUI.
    0:14:27 We often swing optimized, graphically used interface,
    0:14:31 as defined by Xerox, many, many, many years ago.
    0:14:32 Yes.
    0:14:34 You can go back.
    0:14:38 It’s populated by Mac, Microsoft Windows.
    0:14:43 So, both Office, Google Docs, are examples
    0:14:46 of taking advantage of graphically used interface.
    0:14:49 So, that’s the way I understood.
    0:14:54 And challenging, we define conversational use interface.
    0:14:57 They reach 100 million users, amazing, fast.
    0:14:59 Right, faster than anyone.
    0:15:02 So, what the Zoom is doing is developing AUI.
    0:15:07 That will seamlessly combine GUI and the CUI together.
    0:15:10 What that means in Zoom workplace,
    0:15:14 AI companion to promote will be a persistent panel on the right.
    0:15:15 OK.
    0:15:18 And the traditional graphically used interface services,
    0:15:20 whether it’s scheduling a meeting
    0:15:22 or having a meeting with someone.
    0:15:24 Right, you’re under your meeting.
    0:15:25 Yeah.
    0:15:26 It’s on the left.
    0:15:27 OK.
    0:15:31 Information flows seamlessly between those two in the AUI.
    0:15:33 So, we are trying to take advantage of both
    0:15:37 conversational use interface and the screen optimize
    0:15:39 using the face seamlessly.
    0:15:44 The future AI emission is one where the technology
    0:15:47 intuitively adapts to your own needs.
    0:15:48 That’s getting more personal.
    0:15:50 That’s what we are coming with custom AI companion.
    0:15:51 Sure.
    0:15:56 When you say adapts, do you mean that the user interface changes
    0:15:59 or that it can create a conversational window
    0:16:01 sort of in context when you mean it?
    0:16:07 Or could AI potentially just redesign the UI on the fly
    0:16:08 to match what you’re doing?
    0:16:09 How do you envision that?
    0:16:11 AI has the vision.
    0:16:13 We’re leading the point.
    0:16:17 And with that, what information you want to consume,
    0:16:18 how you want to consume it.
    0:16:19 OK.
    0:16:20 I want to AUI explain a bit.
    0:16:24 It’s not because the interface has strategy
    0:16:26 but it’s defined today.
    0:16:28 I’ll just graph using the face.
    0:16:30 Like a Zoom meeting is defined today.
    0:16:31 Right.
    0:16:34 Also, combine seamlessly in the multimodal environment.
    0:16:38 We learn that based on your individual needs.
    0:16:41 But right now, we’re trying to combine those two categories
    0:16:42 into one.
    0:16:46 GUI, traditional, is so massive.
    0:16:48 Most of the world’s services and applications
    0:16:50 that are GUI optimized.
    0:16:50 Yeah.
    0:16:55 CharterGDT, conversational user interface, is a new category.
    0:16:58 And we just kind of have that to be the only one.
    0:17:01 We’re bringing those two together with information flowing
    0:17:04 across those two categories seamlessly.
    0:17:05 Right.
    0:17:07 And they’re trying to understand the user’s needs
    0:17:09 and adapt on the fly.
    0:17:09 OK.
    0:17:11 That is what AUI is going to be.
    0:17:11 Got it.
    0:17:14 I want to really call in this world AUI.
    0:17:16 So this is on record.
    0:17:19 This is the first time I’m telling you in detail
    0:17:21 what the future user interface is going to be.
    0:17:21 World premiere.
    0:17:22 I love it.
    0:17:22 Yeah.
    0:17:26 So that is the principle of the approach.
    0:17:31 Zoom is taking, embracing AI natively.
    0:17:32 That’s what will go AI first.
    0:17:35 You joined Zoom about a year and a half ago,
    0:17:36 a little less.
    0:17:36 Yes.
    0:17:39 When you can join, did you–
    0:17:43 I know Zoom has had AI functionality, AI companion,
    0:17:44 version one.
    0:17:48 And you can use third party apps for transcription
    0:17:49 and et cetera, et cetera.
    0:17:51 That’s been around for a little while.
    0:17:55 But when you joined, is this sort of you came in and thought,
    0:17:57 OK, let’s rebuild this from the ground up.
    0:17:58 AI centric.
    0:18:01 Was that sort of already happening when you joined?
    0:18:05 Just kind of wondering, as you stepped into the role,
    0:18:09 sort of what was envisioned and how much you’ve
    0:18:10 shaped things since then?
    0:18:14 So every year, Zoom’s CEO got this great vision.
    0:18:17 So Zoom has invested in AI before.
    0:18:18 Yes.
    0:18:19 Yeah.
    0:18:22 Since I came, I worked with Eric and the leadership team
    0:18:23 together.
    0:18:24 We defined AI first.
    0:18:25 OK.
    0:18:26 Great.
    0:18:29 Before I came, it was just adding AI,
    0:18:31 both AI, almost every other company.
    0:18:32 Right.
    0:18:32 Yeah.
    0:18:36 We have transformed that because of the consensus
    0:18:39 and pushing AI first to the platform.
    0:18:41 So what does AI first mean?
    0:18:43 That’s three things.
    0:18:45 The first is that AI technology back in.
    0:18:50 Both the edge, small-lock, small-language model,
    0:18:54 and build on the shoulders of great AI companies out there,
    0:18:59 whether it’s OpenAI or Anthropic or Meta or other open source
    0:19:01 companies like Mistra, et cetera.
    0:19:03 Just that there are a lot of them.
    0:19:06 It will be a mistake not to take advantage of all of them.
    0:19:07 Of course.
    0:19:07 Right.
    0:19:10 So it’s like we form a committee working
    0:19:12 to support our workloads.
    0:19:14 That’s always better than just using
    0:19:18 one single model trying to really perform the same task.
    0:19:18 Right.
    0:19:19 Two brains are better than one.
    0:19:20 Yeah.
    0:19:23 So you see how inclusive we are.
    0:19:26 On the models, we’re trying to combine parts of them together.
    0:19:28 And they’re using the face.
    0:19:31 Like some companies that try to use the face is the only way.
    0:19:34 Or graphics using the face is the only way.
    0:19:36 I’m going to add a button here and there.
    0:19:38 We are combining those two categories,
    0:19:42 using them face classes, into one
    0:19:44 that will adapt to your own needs,
    0:19:47 with information flow between those two categories
    0:19:51 seamlessly as a second important advance.
    0:19:54 I want to use AUI as the phrase.
    0:19:57 Longerize this design principle.
    0:19:59 So the third thing I want to talk about is,
    0:20:03 what is the work productivity suite?
    0:20:06 That’s in the general AI Europe.
    0:20:11 I would say it’s all about creating a true system of action.
    0:20:14 We exist, but we have a task to do.
    0:20:16 We take action.
    0:20:19 Of course, you can say you want to entertain people,
    0:20:22 but that’s not productivity suite.
    0:20:24 They allow for payment software.
    0:20:24 Sure.
    0:20:26 AI will do that.
    0:20:30 So when we say we are AI first work platform,
    0:20:33 this is about AI companion is designed
    0:20:35 to understand your workflow.
    0:20:38 Can learn from your pattern.
    0:20:39 Everyone got different workflow.
    0:20:44 Everyone got different selection of services, software.
    0:20:48 And we use AI to anticipate your personal needs,
    0:20:50 emphasize that, personal needs.
    0:20:54 And they can take action on your behalf with your permission,
    0:20:58 or with your code participating to make a better decision,
    0:21:01 more than what you can just do by yourself.
    0:21:06 Those are the soul and the spirit of AI first productivity.
    0:21:10 That’s very different from just to replace paper
    0:21:13 with word processing.
    0:21:18 Or just to support three people co-editing the same document.
    0:21:23 Or just about formatting this document with nice fonts.
    0:21:25 It’s about those three things.
    0:21:29 It’s about learning from your own pattern,
    0:21:32 anticipating your own personal needs,
    0:21:34 and take action on your behalf.
    0:21:37 Whether it’s tracking tasks or managing action items,
    0:21:40 it’s always one step ahead of you,
    0:21:44 ensuring that productivity flows seamlessly,
    0:21:49 effortlessly throughout the whole ecosystem,
    0:21:51 in workplace and the third part of the solution.
    0:21:52 Right.
    0:21:57 So if the AI companion understands my workflow
    0:22:00 and then can suggest to me actions to take,
    0:22:02 either now or going forward, is it
    0:22:06 a case of imagining the AI would say to me,
    0:22:09 hey, you should do these things in this order?
    0:22:13 Or will it actually call up an additional tool
    0:22:16 to help facilitate getting these things done?
    0:22:17 Like, how does that?
    0:22:19 Or how do you envision that working?
    0:22:25 Just envision AI companion can proactively inform you.
    0:22:29 You are not answering the question right in the meeting.
    0:22:31 Only you can see that, right?
    0:22:32 Right.
    0:22:33 Just imagine how powerful that is going to be.
    0:22:34 Right.
    0:22:36 We’re in real time, and I start to give the wrong answer on it.
    0:22:40 AI companion is always opening your ability
    0:22:43 to influence others, make others like you better.
    0:22:47 So this is just what I want to call another phrase.
    0:22:51 So I’ll talk about federated AI stack, unique.
    0:22:51 Yes.
    0:22:54 I talked about the use interface, that’s AUI.
    0:22:55 The AUI, yes.
    0:22:59 This is about action-oriented task flow.
    0:23:00 Action-oriented task flow, OK.
    0:23:02 This will flow through every corner
    0:23:05 for the whole life cycle of what you need to do.
    0:23:08 Because it’s almost like you have a very
    0:23:09 expensive ecosystem.
    0:23:10 Yes, right.
    0:23:13 The most important task that you need to pay attention to
    0:23:16 for the life cycle of the whole project,
    0:23:19 until you get that project done beautifully
    0:23:21 in a time-sensitive manner.
    0:23:25 And in a way, you delight your co-worker,
    0:23:29 your family members, for better human connection.
    0:23:34 This is the goal of Zoom’s AI-first work platform.
    0:23:36 It’s action-oriented information flow.
    0:23:39 If there’s something you don’t need to take action,
    0:23:42 we can still accumulate those tasks to confuse you.
    0:23:43 That’s OK.
    0:23:45 And you can decide.
    0:23:48 And you do not want to actually track those actions.
    0:23:50 We learn that pattern from you.
    0:23:50 Yeah.
    0:23:53 And we improve our ability to track.
    0:23:57 But if you select the old AI companion,
    0:23:58 tell me this action is important.
    0:24:00 You check that on that.
    0:24:01 AI companion will work harder.
    0:24:03 Keep your eyes open.
    0:24:06 So a week later, if you receive an e-mail
    0:24:08 that is relevant to the task you’re tracking,
    0:24:13 AI companion will work 24/7 to update what you need to do
    0:24:17 and to give your tips and to advise what you have to do better
    0:24:19 to accomplish that task.
    0:24:23 This is what I talk about, action or render information flow.
    0:24:23 Right.
    0:24:27 And so is this the AI companion of an example of an AI agent
    0:24:28 working on your behalf?
    0:24:29 Absolutely.
    0:24:34 AI companion to Verneau already brought agent-like capabilities,
    0:24:38 like in a meeting, will not just actually
    0:24:40 use speech recognition to understand
    0:24:42 what is being talked about.
    0:24:46 If you presented your slide, AI companion to Verneau today
    0:24:48 understand what is presented in the slide
    0:24:52 or what you wrote on the paper that you shared
    0:24:55 your swimming with that capability multimodal.
    0:25:00 Or if you showed your points in the side panel with chat,
    0:25:02 we take that into account as well.
    0:25:03 That’s amazing.
    0:25:06 It’s almost like an agent really participating in the meeting
    0:25:06 as you do.
    0:25:09 Then we present meeting recap.
    0:25:12 In that meeting recap, the most powerful way for us
    0:25:16 is identify next steps you need to pay attention to
    0:25:19 or your colleague need to pay attention to.
    0:25:21 Then next step is unique.
    0:25:23 We offer a larger quality.
    0:25:27 We worked on this so hard in the past year and a half
    0:25:30 to improve next steps to reduce hallucination,
    0:25:33 to assign the right task to the right person.
    0:25:36 We are roughly right now probably 80% accurate.
    0:25:36 OK.
    0:25:39 So we’re not done with– it’s not perfect.
    0:25:41 But 80% is really impressive.
    0:25:44 I was going to say, in my experiences with LLMs
    0:25:48 and hallucinations and accuracy, 80% sounds pretty good.
    0:25:50 So we do not stop here.
    0:25:54 So let’s say you have a meeting, you discuss.
    0:25:58 You are campaigning to identify the task.
    0:26:03 And that task we show up in the upcoming suite.
    0:26:04 This task panel.
    0:26:07 And a week later, through the whole lifecycle,
    0:26:09 that’s something you want to try.
    0:26:10 You want to try it.
    0:26:13 You receive a piece of email from Zoom
    0:26:16 and create an update that on your behalf.
    0:26:19 And if you want to read or write a book about that action item
    0:26:23 or status report for your colleague,
    0:26:25 that information flow into Zoom docs
    0:26:27 will drive the status report on your behalf.
    0:26:29 You’ll feel pretty happy with a few changes
    0:26:32 without doing everything using liquid paper
    0:26:34 or Microsoft Word to format everything.
    0:26:37 And you want to say, hey, change this into the form
    0:26:42 that I can present in that next status update meeting.
    0:26:42 Done.
    0:26:43 One comment.
    0:26:43 Yes.
    0:26:44 Right.
    0:26:48 And it’s not as beautiful as what a PowerPoint can
    0:26:50 do with beautiful picture.
    0:26:52 But the key points is very much like when
    0:26:56 I was a colleague in Maryland, when we presented information
    0:26:58 with black and white to tear off.
    0:27:03 And printed on the really transparency paper.
    0:27:08 And just really projected to talk about the key points.
    0:27:10 We still actually reflect it.
    0:27:11 It still works, yeah.
    0:27:15 The information still flows with that beautiful color
    0:27:16 black marble animation.
    0:27:18 So that’s the point I want to take.
    0:27:23 Zoom docs alone is actually performing most of the function
    0:27:25 because of the general AI.
    0:27:27 With AI company, you can instruct,
    0:27:30 you can summarize in the form of status report,
    0:27:33 who can publish this as a blog that’s
    0:27:36 more ready to be consumed for the public,
    0:27:39 or in the simple form of slides, that you
    0:27:41 can communicate your points in your next meeting.
    0:27:45 You do not need the last generation productivity suite,
    0:27:46 as we know.
    0:27:48 So that is the actual high-level three key points.
    0:27:50 Friendly AI stack.
    0:27:54 AI first, use interface, AUI, and action-oriented information
    0:27:55 flow for productivity.
    0:27:57 That is really the landmark.
    0:28:01 How AI first, productivity suite,
    0:28:04 or differential form, web centric, productivity suite,
    0:28:06 or desktop centric.
    0:28:09 Of course, both web centric and desktop centric
    0:28:12 can add AI capability.
    0:28:13 That’s bolded.
    0:28:13 That’s not–
    0:28:14 Sure.
    0:28:14 Yeah.
    0:28:16 What Zoom is defining this native.
    0:28:17 Right.
    0:28:19 Back-action-oriented information flow
    0:28:23 into every corner of the productivity suite.
    0:28:24 I made myself clear.
    0:28:25 Only three things.
    0:28:27 [LAUGHTER]
    0:28:29 Our guest today is XD Wang.
    0:28:32 XD is the CTO of Zoom, a position
    0:28:34 he’s held for going on a year and a half now.
    0:28:39 Before that, he was with Microsoft for quite some time.
    0:28:41 And really, it’s just a continuation
    0:28:44 of an illustrious career that started way back,
    0:28:46 as XD was talking about in the days of typewriters
    0:28:48 and liquid paper, as we both know.
    0:28:50 XD, I want to switch gears here.
    0:28:51 We have a few minutes left to talk.
    0:28:55 I want to look at things from the public standpoint,
    0:28:57 and specifically from business users
    0:29:00 and the types of customers who Zoom has been working with
    0:29:01 for a while now.
    0:29:06 When Zoom is talking to business customers about AI,
    0:29:09 about adopting Zoom’s products and all these wonderful things
    0:29:13 you’re building, but personally just about using generative AI
    0:29:16 and making the investment and spending the time to upskill
    0:29:20 workers and figure out, how are we using these things?
    0:29:24 How do you help your customers think about both adopting AI
    0:29:27 and also how to measure return on investment?
    0:29:30 There’s a lot of conversation we’ve had on the podcast
    0:29:33 and just generally in the world about being kind of,
    0:29:36 for exciting as the past couple of years have been,
    0:29:39 still being in the early days of figuring out,
    0:29:41 what can Gen AI do?
    0:29:42 How do we use it?
    0:29:45 How do we rethink things like productivity
    0:29:47 suites from the ground up with AI?
    0:29:50 So when you’re talking to the customer companies,
    0:29:52 how do you educate them about getting started
    0:29:54 and measuring performance?
    0:29:58 There are a few things, absolutely our customers love.
    0:30:03 First, Zoom workplace as a whole offers ease of use.
    0:30:07 That’s just unmatched with its tingle, docks, or chat.
    0:30:11 The second thing is really, Zoom AI campaigning to promote
    0:30:14 is offered under no additional cost.
    0:30:17 That’s just stunning for most of the customers.
    0:30:21 Because they get used to, you have to pay $30 a month
    0:30:22 for person, right?
    0:30:27 So Zoom offered this capability for the high-end customers
    0:30:29 who are going to offer a custom AI company
    0:30:32 where we charge $12 per person per month.
    0:30:33 So you can bring your own data,
    0:30:36 get your own pattern into the AI campaigning,
    0:30:39 campaigning where you can run that on your behalf
    0:30:41 with fine-tuning the customized capabilities.
    0:30:45 So Zoom offers this amazing horizontal,
    0:30:48 absolutely game-changing capability
    0:30:53 to make a Zoom workplace a very viable productivity candidate
    0:30:54 for just this.
    0:30:55 And for the high-end,
    0:30:58 we offer you unmatched custom AI campaigning
    0:31:01 before the info was $12 per person per month.
    0:31:04 It still offers the best in the TCO.
    0:31:08 So interviews, cost-defective, unmatched quality.
    0:31:11 That’s what we know of our customers.
    0:31:14 – And is kind of the, I don’t know, a goal or sort of the vision
    0:31:19 that customers who use Zoom come away with is that the AI,
    0:31:22 the companion is just over time going to learn more and more
    0:31:26 about how you work and what your workflows are like
    0:31:29 and how you sequence tasks and the people,
    0:31:31 the colleagues you’re working with,
    0:31:33 and the companion will just be there
    0:31:36 to help you think a couple of steps ahead,
    0:31:39 help you maximize your own efficiency,
    0:31:40 whatever the word is.
    0:31:43 Is that, ’cause that’s a different conversation
    0:31:45 than conversations I’ve had
    0:31:46 or I’ve read about or listened into
    0:31:48 where companies are saying,
    0:31:52 “Okay, we need to start with wrangling all our data
    0:31:54 “and then we need to figure out how to clean the data
    0:31:55 “and how to…”
    0:31:58 And it’s kind of this big, deep investment process
    0:32:00 where it sounds like with Zoom,
    0:32:02 it’s more like, “Hey, you’re already using it
    0:32:05 “for video calls and now we’re gonna give you
    0:32:09 “this groundbreaking change the way
    0:32:11 “that you do everything companion.”
    0:32:12 And it’s just kind of gonna be there
    0:32:15 and there’s not a lot you have to do as the user.
    0:32:18 – Yeah, so we offer the choice to our customers.
    0:32:21 If they have the comfort,
    0:32:25 yes, who have improved customized capability
    0:32:27 to suit their needs.
    0:32:29 If they don’t, they can decide
    0:32:31 how much data they want to share
    0:32:32 or whether they want to turn off
    0:32:36 for some sensitive meeting.
    0:32:41 That complexity is in the hands of the customer.
    0:32:43 They can control themselves directly.
    0:32:44 – So giving them the choice.
    0:32:46 – Yeah, on top of that,
    0:32:47 I want to really emphasize,
    0:32:50 Zoom never takes any custom data
    0:32:53 in the meeting to train all the way on model.
    0:32:56 – Let’s end on a kind of looking ahead note,
    0:32:57 if that’s all right.
    0:32:59 As you envision the next,
    0:33:01 I’m gonna say three years, you can change that.
    0:33:04 Two years, five years, whatever you think it is.
    0:33:06 Both in terms of Zoom’s mission
    0:33:08 and using AI and generative AI
    0:33:11 to help people do things smarter,
    0:33:13 better, faster in the workplace.
    0:33:15 And then more broadly,
    0:33:18 as Gen AI and other forms of AI,
    0:33:19 and machine learning and deep learning
    0:33:22 just continue to impact the world more.
    0:33:25 What are you most excited about
    0:33:27 in the short term?
    0:33:28 Again, three years, however long it is.
    0:33:30 What are you really excited about
    0:33:34 and see coming down the pike that may,
    0:33:36 I don’t know if it’s the next transformational moment
    0:33:38 or just kind of a trend
    0:33:39 that’s gonna really take fire
    0:33:40 and change the way we do things.
    0:33:42 What are you looking at too?
    0:33:43 – I’m really thinking this
    0:33:45 action of like the estimation flow
    0:33:47 to be your companion.
    0:33:48 – Yeah.
    0:33:50 – This is really just a game changer
    0:33:54 and all of us will not have enough time.
    0:33:57 So if a company can really help you
    0:33:59 to get job done quickly,
    0:34:02 you can have additional time to do whatever you want.
    0:34:04 That’s an opportunity capability,
    0:34:05 some entertainment for–
    0:34:06 – Whatever it is, yeah.
    0:34:08 – Yeah, and we’ll also bring,
    0:34:10 this will be a better place.
    0:34:11 That’s what Gen AI is.
    0:34:13 We’ll make you work happy,
    0:34:17 be happy and do whatever you want.
    0:34:18 – Be happy in between.
    0:34:19 You’re doing whatever you want.
    0:34:22 – So Delighted Customers is in a core mission.
    0:34:23 – Excellent.
    0:34:26 XD, for people who would like to learn more
    0:34:28 about what Zoom’s doing,
    0:34:30 announcements at ZoomTopia,
    0:34:31 perhaps some of the,
    0:34:33 I don’t know if there’s a technical blog
    0:34:36 for developers and people more technically inclined
    0:34:38 to learn more about how you’re approaching everything,
    0:34:41 federated AI and everything else we’ve discussed.
    0:34:44 Where’s a good place or some good places online
    0:34:46 for people to get started to learn more?
    0:34:50 – Yeah, you can check out the Zoom blog in Zoom.com.
    0:34:53 That’s actually probably the best place to learn.
    0:34:54 – Best place to start.
    0:34:55 – But even better wise,
    0:35:00 really start turning on AI company in Zoom workplace.
    0:35:03 Without using that, you do not know how powerful–
    0:35:06 – You gotta use it, yeah, absolutely, fantastic.
    0:35:08 XD, thank you so much for taking the time,
    0:35:09 particularly at the end of this,
    0:35:11 what I’m sure was a busy week, a crazy week for you.
    0:35:14 But congratulations on ZoomTopia,
    0:35:15 on the work you’ve done so far,
    0:35:19 and I for one am excited to use Companion 2.0,
    0:35:22 if I could have a panel on the side of my screen
    0:35:25 that’s always telling me that the next best thing I should do,
    0:35:27 that would be a game changer for me personally,
    0:35:28 so I’m excited to go ahead and start–
    0:35:30 – Absolutely, I haven’t even got an AI companion.
    0:35:33 My productivity software, seriously.
    0:35:35 – Yes, fantastic.
    0:35:36 Well, thank you again,
    0:35:38 and perhaps we can catch up somewhere down the line
    0:35:40 to see what’s going on at ZoomTopia next year.
    0:35:42 – Absolutely, thank you.
    0:35:43 I have to be here.
    0:35:46 (somber music)
    0:35:49 (inspirational music)
    0:35:52 (inspirational music)
    0:35:55 (inspirational music)
    0:35:58 (inspirational music)
    0:36:02 (inspirational music)
    0:36:05 (inspirational music)
    0:36:08 (inspirational music)
    0:36:11 (inspirational music)
    0:36:14 (inspirational music)
    0:36:17 (inspirational music)
    0:36:21 (inspirational music)
    0:36:24 (inspirational music)
    0:36:27 (inspirational music)
    0:36:30 (inspirational music)
    0:36:33 (upbeat music)
    0:36:43 [BLANK_AUDIO]

    Zoom, a company that helped change the way people work during the COVID-19 pandemic, is continuing to reimagine the future of work by transforming itself into an AI-first communications and productivity platform.

    In this episode of NVIDIA’s AI Podcast, Zoom CTO Xuedong (XD) Huang shares how the company is reshaping productivity with AI, including through its Zoom AI Companion 2.0, unveiled recently at the Zoomtopia conference.

    Designed to be a productivity partner, the AI companion is central to Zoom’s “federated AI” strategy, which focuses on integrating multiple large language models.

    Huang also introduces the concept of “AUI,” combining conversational AI and graphical user interfaces (GUIs) to streamline collaboration and supercharge business performance.

  • NVIDIA’s Josh Parker on How AI and Accelerated Computing Drive Sustainability – Ep. 234

    AI transcript
    0:00:10 [MUSIC]
    0:00:14 Hello, and welcome to the NVIDIA AI podcast.
    0:00:16 I’m your host, Noah Krebs.
    0:00:20 The emergence of generative AI into our collective consciousness has led to
    0:00:25 increase scrutiny around how AI works, particularly when it comes to energy consumption.
    0:00:29 The interest and focus on how much energy AI actually uses is, of course,
    0:00:33 important, our planet faces energy-related challenges ranging from
    0:00:36 grid infrastructure needs to the impact of climate change.
    0:00:42 But AI and accelerated computing have a big part to play in helping to solve these challenges
    0:00:46 and others related to sustainability and energy efficiency.
    0:00:49 Joining us today to talk about all of this is Joshua Parker,
    0:00:52 the Senior Director of Corporate Sustainability at NVIDIA.
    0:00:57 Josh brings a wealth of experience as a sustainability professional and engineer.
    0:01:02 Before his current role, he led Western Digital’s Corporate Sustainability function
    0:01:05 and managed ethics and compliance across the Asia-Pacific region.
    0:01:09 At NVIDIA, Josh is at the forefront of driving sustainable practices
    0:01:15 and leveraging AI to enhance energy efficiency and reduce environmental impact.
    0:01:19 Josh, welcome and thanks so much for taking the time to join the AI podcast.
    0:01:21 Thanks, Noah. Long time listener for some caller.
    0:01:28 Love it. I always dreamt of hosting an AM radio call-in show, so you’re inspiring me.
    0:01:33 So I’m going to just kind of open this up broadly to you to get us started.
    0:01:38 I kind of alluded a little bit to it in the intro, but computing uses energy.
    0:01:43 Everybody is talking about AI, obviously, and there’s, you know, with good reason,
    0:01:47 interest, scrutiny, focus on, well, how much energy is AI using?
    0:01:52 And if we start using more and more AI going forward, what’s the impact going to be on,
    0:01:57 you know, all of these energy-related things that we deal with on our planet?
    0:02:00 So let me ask you to start. How much energy does it really use?
    0:02:03 Is this a warranted discussion? What are the things that we should be
    0:02:11 thinking about and talking about and working on when it comes to energy and sustainability,
    0:02:16 and not just AI, but accelerated computing and the other advanced technology that goes with it?
    0:02:22 It’s definitely a reasonable question. And as someone who’s been in sustainability for a while,
    0:02:28 it’s something that we always talk about, climate and energy and emissions.
    0:02:32 Those are very big, urgent topics that we’re all thinking about in every context.
    0:02:38 So when you see something like AI that really bursts onto the scene, especially so rapidly,
    0:02:42 it’s a very legitimate question to ask, okay, what is this going to do to energy?
    0:02:46 And what is this going to do to emissions associated with that energy?
    0:02:48 So it’s the right question to ask.
    0:02:52 The answer, though, turns out to be pretty complicated, because number one,
    0:02:57 we’re in a period of rapid, rapid growth, and it’s hard to predict where we’re going to be
    0:03:03 in just a couple of years in terms of the expansion of AI. Where is it going to be used?
    0:03:06 How is it going to be used? What benefits do we get from it?
    0:03:12 And there are lots of nuances to that as well, including things like the hardware that it’s
    0:03:18 being built on. This accelerated computing platform itself is rapidly, rapidly evolving
    0:03:24 in ways that actually support sustainability. So it’s the energy efficiency gains that are
    0:03:29 being developed in that accelerated computing platform are really, really dramatic.
    0:03:35 So if you want to paint a really accurate picture, as accurate as we can get in terms of where we’re
    0:03:41 going with AI energy consumption and the emissions associated with that, you need to have a really
    0:03:48 complex, nuanced analysis to avoid coming to very inaccurate and potentially alarming conclusions.
    0:03:53 So let’s dig into that a little bit within the context of a half-hour podcast.
    0:03:56 Let’s talk about some of those nuances, and you mentioned the hardware,
    0:04:03 and so obviously GPUs, a big part of that. How is accelerated computing sustainable?
    0:04:11 Accelerated computing is a very tailored form of computing for the type of work required for AI.
    0:04:18 The accelerated computing platform on which modern AI is built takes the math that was
    0:04:26 previously being done on CPUs in a sequential order and basically uses these very, very efficient,
    0:04:33 very purpose-built GPUs to do them in parallel. So you do many, many more operations,
    0:04:40 and these GPUs are optimized to do that math, the matrix math that’s required for AI,
    0:04:45 really, really effectively and efficiently, and that’s what’s driven both the huge gains in
    0:04:52 performance and also the huge gains in efficiency in AI, and it’s really what has enabled AI to boom
    0:04:59 the way it has. The traditional CPU paradigm, CPU-only paradigm for trying to run this math
    0:05:06 just wasn’t scaling, and so we really need GPUs to unlock this exponential growth really in
    0:05:14 performance and efficiency. So if you compare CPU-only systems to accelerated computing systems,
    0:05:22 which have a mix of GPUs and CPUs, we’re seeing roughly a 20 times improvement in energy efficiency
    0:05:29 between CPU-only and accelerated computing platforms, and that’s across a mix of workloads.
    0:05:35 So it’s a very dramatic improvement in efficiency, and if you look just over time at accelerated
    0:05:40 computing itself, so compare accelerated computing platforms from a few years ago to ones that we
    0:05:49 have today, that change in efficiency is even more dramatic. So just eight years ago, if you compare
    0:05:57 the energy efficiency for AI inference from eight years ago until today, we’re 45,000 times more
    0:06:02 energy efficient for that inference step of AI that where you’re actually engaging with the models,
    0:06:11 right? And it’s really hard to understand that type of figure. One 45,000 of the energy required
    0:06:16 just eight years ago is what we’re using today. So building in that type of energy efficiency gain
    0:06:21 into your models for how much energy will we be using for AI in a couple of years is really,
    0:06:27 really critical because it’s such a dramatic change. Yeah, that’s a huge number. I don’t mean
    0:06:33 this as a joke, but the best way I can think of to ask it is, are the workloads now 45,000 times
    0:06:40 bigger or more energy intensive than they were eight years ago, or is really the efficiency
    0:06:47 outpacing all of this new attention on AI? So that ends up being a very complex question as well
    0:06:53 because you have to get into the realm of figuring out how many times do we need to train a model,
    0:06:59 and then versus how many times can I reuse it with inference? So you know, big models like
    0:07:04 Claude 3.5, ChatGPT40 and so forth, they’re trained, takes a lot of time to train them,
    0:07:09 but the inferencing when you’re actually engaging with the model, if it ends up being durable,
    0:07:16 then that inference step is very, very efficient. So it’s because we’re still in this inflection
    0:07:23 point where things are moving very rapidly, it’s hard to see how the compute requirements are scaling
    0:07:28 versus the energy efficiency. Certainly, they’re scaling. We continue to see bigger and bigger
    0:07:35 models being used and trained because companies are seeing huge benefits in doing that. But yeah,
    0:07:39 this is what makes it complicated is that the energy efficiency is ramping up very dramatically
    0:07:46 at the same time. Right. So along those lines, there’s been attention, well, there’s been attention
    0:07:52 on the stability and durability of power grids, national, regional, local, as long as they’ve
    0:07:58 existed, but certainly over the past five years, 10 years or so. But since AI has come into the
    0:08:05 public consciousness, there have been news stories and what have you about kind of the localized
    0:08:12 effects of, oh, this data center was built in wherever it was and it had this huge impact on
    0:08:16 the local power grid or people are concerned it might. Can you talk a little bit about the
    0:08:22 common concerns around AI’s energy consumption, particularly when it comes to the impact on
    0:08:29 local power grids, whether it’s in the area where a data center might be or other places
    0:08:33 where people are concerned that AI is impacting the local energy situation?
    0:08:38 The first thing to look at when you’re trying to put this in context and figure out what the
    0:08:45 local constraints might be on the grid is the fact that AI still accounts for a tiny, tiny fraction
    0:08:50 of overall energy consumption generally. If you look at it globally first and then we’ll
    0:08:56 get to the local issue, look at it globally. The International Energy Agency estimates that
    0:09:03 all of data centers, so not just AI, all of data centers account for about 2% of global energy
    0:09:10 consumption and AI is a small fraction of that 2% so far. So we’re looking at much less than 1%
    0:09:16 of total energy consumption currently used for AI-focused data centers. Now that is absolutely
    0:09:23 growing, we expect that to grow, but ultimately this is a very small piece of the pie still compared
    0:09:28 to everything else that we’re looking at. The second thing to consider is the fact that AI
    0:09:34 is mobile, especially when you think about AI training. So when you’re working to train these
    0:09:39 models for months at a time, potentially very large models, you need a lot of compute power,
    0:09:44 that training doesn’t have to happen near the internet edge, it doesn’t have to happen
    0:09:50 in a particular location. So there is more mobility built in to how AI works than in
    0:09:55 traditional data centers because you could potentially train your model in Siberia if it
    0:10:03 were more efficient to do that or in Norway or Dubai. Wherever you have access to reliable,
    0:10:08 clean energy, it would be easy to do your training there. And some companies, including
    0:10:13 some of our partners, have built business models around that, locating AI data centers
    0:10:19 and accelerated computing data centers close to where there is excess energy and renewable energy.
    0:10:26 So to get back to your original question, will we see local constraints and problems with the grid?
    0:10:32 I think for the most part, we’re able to avoid that because of those issues. AI is still relatively
    0:10:39 small and the companies who are deploying large data centers already know where there is excess
    0:10:45 energy, where there are potentially constraints. And of course, they’re trying to find locations for
    0:10:50 the AI data centers where it’s not going to become a problem and they’re going to be able to have
    0:10:58 good access to clean, reliable energy. So is the sort of pessimistic or it sounds like overblown,
    0:11:05 if data centers only comprise 2% of global energy usage and AI-specific data centers are
    0:11:13 only 1%, is the sort of proliferance of pessimistic stories around AI’s impact on the energy grid?
    0:11:18 Is that just kind of the dark side of a hype cycle that we’re used to and this is how it’s
    0:11:23 coming up with AI? I don’t want to say that those concerns are misplaced. Certainly, if you’re living
    0:11:28 in a community and you see an AI data center going up, you may have questions about what
    0:11:34 it’s going to do to your local grid. And we are because we’re in this period of very, very rapid
    0:11:40 and to some extent unexpected deployment of AI because ChantGPT really took the world by
    0:11:46 storm and by surprise two years ago. There is some churn right now, which you would expect
    0:11:52 when you have a new technology, a new industrial revolution that’s bursting on the world.
    0:11:58 There’s going to be a little bit of time where resources are not perfectly allocated.
    0:12:03 But what we’re seeing is we’re already working through that phase and the companies who are
    0:12:10 deploying the big AI data centers are finding ways to do that that are sustainable and that won’t
    0:12:15 threaten local grids. Even if in the near term, there are some constraints that we all need to
    0:12:20 work through. In the long term, even in the medium term, we’re very optimistic that these are
    0:12:25 solvable issues. Sort of to look at the bright side of things relative to AI,
    0:12:31 as with so many industries and so many problems that people are trying to solve in all walks of
    0:12:39 life, AI can be a help when it comes to optimizing grids and energy use and even perhaps trying to
    0:12:43 solve some of these climate challenges that we’re all facing. Can you talk a little bit about that
    0:12:51 and about how AI can or perhaps already is making a positive impact on our energy situation?
    0:12:59 Sure. There are two examples that I’ll focus on that speak to different aspects of sustainability.
    0:13:08 The first one is helping us adapt and mitigate the worst impacts of climate change. AI and
    0:13:14 accelerated computing in general are game changers when it comes to weather and climate
    0:13:21 modeling and simulation. NVIDIA has a platform called Earth 2 and we partner closely with
    0:13:26 national labs and non-governmental organizations and other organizations to develop
    0:13:35 systems where we can much, much more accurately forecast weather, model weather, and help mitigate
    0:13:41 the worst impacts of near-term weather and also look longer-term at climate so that we’ve got
    0:13:46 a better understanding of where we’re going with climate and can better prepare for and plan for
    0:13:51 that. The other piece of the puzzle is that accelerated computing and AI, both of them,
    0:13:59 have real-world applications that directly reduce energy and emissions. One example of that is
    0:14:08 we’ve transitioned the PANDAS library in Python. It’s one of the most well-used libraries for
    0:14:15 simulation and for high-performance computing. We’ve taken that, basically, and written libraries
    0:14:21 that will translate that code, code that’s written for that library, onto the accelerated computing
    0:14:29 GPU platform. And doing that, we’ve basically opened up for the world of researchers a way to
    0:14:37 run their simulations in that library without any code changes at 150 times speed and many,
    0:14:43 many times more energy efficiently. So, the application of this itself is going to end up
    0:14:48 reducing energy consumption and also reducing the associated emissions. Right. Now, it sounds
    0:14:54 like a virtuous cycle. So, to kind of dig into that for a second, people familiar with NVIDIA,
    0:14:58 longtime listeners to the podcast, perhaps, understand that NVIDIA is not just a hardware
    0:15:05 company. It’s hardware. It’s software. It’s all of the tools and everything in the stack to leverage
    0:15:09 the GPUs in all of these different systems, accelerated computing, AI, and what have you.
    0:15:14 Can you talk a little bit more, you mentioned earlier, some of the efficiency gains,
    0:15:20 but a little bit more about some of the efficiency improvements in AI training and inference,
    0:15:26 and then also NVIDIA’s role in developing more efficient models. You mentioned Earth 2 just a
    0:15:32 second ago, but some of NVIDIA’s other work in increasing the overall efficiency of the hardware,
    0:15:39 the software, and then these models themselves. Sure. If you start with inference, one data point
    0:15:46 that I’d like to share is that just in one generation of improvement, so if you look at
    0:15:52 one generation of NVIDIA hardware, our ampere platform, or sorry, our hopper platform, which is
    0:15:58 the one that we’re shipping in the highest volume right now, and you compare that to the Blackwell
    0:16:04 platform, which we’re releasing next. We’ll come out within the next several months. The Blackwell
    0:16:10 platform is 25 times more energy efficient or AI inference than hopper was. Just in the space of
    0:16:18 one change, one generation of NVIDIA hardware and software, it’s using 1/25 of the energy. That’s a
    0:16:25 96% reduction in energy use. There are performance gains associated as well.
    0:16:31 That’s right. Yeah, significant performance gains. I understated it, but performance gains are
    0:16:39 amazing, but that’s incredible. A 96% efficiency gain while also getting these enormous next
    0:16:45 generation performance gains as well. That’s right. You asked a little bit about how we’re
    0:16:53 getting there. That 25x improvement is through innovation in many spaces. It includes things like
    0:17:00 quantization, where we’re using lower precision math, basically finding ways to optimize the
    0:17:06 model on training and inference in ways that allow us to do it even more in parallel and do it more
    0:17:13 efficiently. It includes things like water cooling our data centers, so that we’re using less water
    0:17:17 and significantly less energy to keep the data centers cool. Of course, it includes things like
    0:17:25 better GPU design. All of these levers we’re pulling at the same time to drive those energy
    0:17:31 efficiency improvements. We expect that to continue because energy efficiency is something that we
    0:17:36 care about and our customers care about. It really helps enable more performance when we’re able to
    0:17:41 take out waste and to be able to do things more efficiently. It enhances our ability to drive more
    0:17:47 performance and to make the AI even more valuable than it was. Our guest today is Joshua Parker.
    0:17:53 Josh is the Senior Director of Corporate Sustainability at NVIDIA. We’ve been talking a little bit
    0:17:59 about energy and energy in the AI area, climate change, sustainability, all these incredibly
    0:18:05 important things that form the basis of our ability to live on Earth and how these things are being
    0:18:11 affected by all of these rapid advances in technology and obviously being fueled by the
    0:18:16 interest in AI, which AI has been around for a while as you well know listening to this show,
    0:18:20 but over the past couple of years, it’s really ramped up in intensity. Josh, you mentioned just
    0:18:27 a second ago, customers. Maybe we can dig in a little bit to some case studies, customer examples,
    0:18:34 real-world applications of AI improving energy efficiency. Sure. One that I love to talk about
    0:18:41 is with a partner of ours called Wistron. It’s a Taiwan-based electronics company. It has a lot
    0:18:46 of manufacturing that many people may not have heard of, but it’s a large sophisticated company.
    0:18:53 They took our Omniverse platform, which is a 3D modeling platform, and they modeled
    0:19:02 one of their buildings in Omniverse, and then they used AI to run emulations on that digital twin
    0:19:08 that they created in our Omniverse platform. We’re looking for ways to improve energy efficiency.
    0:19:15 After doing that, after using that digital twin, applying AI to run some emulations, they were
    0:19:21 able to increase the energy efficiency of that facility by 10%, which is a dramatic change,
    0:19:27 just based on a digital twin. In this case, it resulted in savings of 120,000 kilowatt hours per
    0:19:34 year. Fantastic. A word that I hear a lot that I’m not really sure what it means. I’ve got a
    0:19:41 working understanding is decarbonization. My understanding is that NVIDIA has been involved
    0:19:47 in some work optimizing processes for decarbonization in industry. I think you know a little bit
    0:19:51 about that, and I think it’s relevant to what we’re talking about. Could you dig into that a
    0:19:59 little bit as well? Decarbonization is a really broad term, and it makes sense to have a nuanced
    0:20:04 appreciation for everything that encompasses. Basically, I think best understanding is that
    0:20:11 describes our efforts to reduce greenhouse gas emissions, and an effort to try to mitigate
    0:20:17 the climate change that we’re seeing. It can apply to a lot of things, including things like
    0:20:22 carbon capture and storage, where we’re actually pulling carbon out of processes or out of the
    0:20:27 atmosphere and finding ways to store it. That’s an area actually where we have some
    0:20:35 good work being done and some partnerships with NVIDIA and Shell, for example, where we’re finding
    0:20:44 ways to use AI to greatly enhance carbon capture and storage technologies. That’s one example.
    0:20:51 Another example of decarbonization, and this goes directly to emissions, is we’ve also partnered
    0:20:59 with California, I believe in the San Diego area, to help firefighters there use AI to monitor
    0:21:06 weather risks that could lead to wildfires. In doing so, we’ve been able to expand their
    0:21:13 responsiveness, to improve the responsiveness of their firefighting efforts, and not only to
    0:21:19 potentially save lives and property, but also to significantly reduce emissions associated with
    0:21:25 those wildfires. That’s another example of decarbonization that we’re seeing. Then the third
    0:21:31 example I’ll give is NVIDIA itself. We are trying to decarbonize our own operations
    0:21:38 by transitioning the energy that we use from standard energy to renewable. This year, we’re
    0:21:45 going to be 100% renewable for our own operations, so we’re very excited to be transitioning over
    0:21:50 there. Quick show note to listeners if you’re interested. We did an episode previously about
    0:21:55 the use of AI in firefighting and combating wildfires in California. I would imagine it’s
    0:21:59 the same organization. Forgive me, it might not be, but definitely worth a listen. It’s a great
    0:22:07 episode. Josh, you mentioned a little while ago, data centers. We talked about being a citizen and
    0:22:11 seeing a data center come up. Of course, it’s good to have questions and concerns about how
    0:22:16 is that going to impact things. The design of the data centers themselves obviously plays a big
    0:22:23 part in how efficiently they do or don’t operate. You mentioned a little bit earlier water cooling
    0:22:31 as a technique that’s been effective in reducing or increasing energy efficiency, I should say.
    0:22:38 Can you talk a little bit more about the data center, how data centers relate to sustainability
    0:22:43 broadly and some of the innovations that have helped in that regard? Yes. The first thing you
    0:22:48 get to mention is to put data centers in context because it’s easy to think about data centers
    0:22:54 being really impactful in terms of sustainability. You see them, they’re large, you hear about them
    0:22:59 using all this energy, all this water, and so forth. It’s a legitimate question to ask,
    0:23:05 but ultimately, again, IEA estimates that all of data centers only account for 2%
    0:23:11 of global energy consumption right now. It’s much, much smaller than most of the other centers.
    0:23:17 Not to interrupt you, but I’ve obviously been working in this arena for a while now,
    0:23:22 but that figure really blew my mind. I was expecting something a little bit bigger than 2%.
    0:23:28 That makes sense because there’s so much attention on this. AI is very much in the zeitgeist right
    0:23:35 now. We’re talking about it. We see the rapid expansion. That’s one of the things where it’s
    0:23:42 important to put it in context. The innovation in data center design is one of those levers that
    0:23:49 I mentioned that we’re all pursuing to try to improve energy efficiency. As we’re transitioning
    0:23:55 to this new generation of products at NVIDIA to Blackwell, our reference design, our recommended
    0:24:02 design for the data centers for our new V200 chip is focused all on direct-to-chip liquid cooling,
    0:24:09 which is much more efficient, really unlocks better energy efficiency, of course, but also
    0:24:15 unlocks better performance because the cooling is more effective. We’re able to run the chips
    0:24:21 more optimally in ways that lead to better performance as well as to better energy efficiency.
    0:24:28 Paint the picture. Sorry to interrupt you again. When you talk about direct-to-chip cooling,
    0:24:32 what is that replacing? What’s the thing that it’s more efficient than?
    0:24:39 That’s in comparison to air cooling, where you have heat sinks and you’re using air flow. Direct-to-chip
    0:24:45 liquid cooling, we’re able to get liquid in closer to the silicon and to more effectively get heat
    0:24:52 away from that. One of the reasons why this is so effective and helpful with accelerated computing
    0:24:58 is that the compute density is so high. If you look at a modern AI data center and you see a rack,
    0:25:05 for example, of a modern AI data center, there’s as much compute power in that rack as there was
    0:25:12 in several racks, many racks of a traditional computing data center. The compute density is so
    0:25:18 high that it makes more sense to invest in the cooling because you’re getting so much more compute
    0:25:23 for that same single direct-to-chip cooling element that you’re using.
    0:25:30 Obviously, all of this is a work in progress, so to speak, that advances in AI aren’t slowing down
    0:25:37 anytime soon. You’re talking about the increases in efficiency and the increases in performance
    0:25:44 and all that from generation to generation. As you said, this is early days of the world
    0:25:50 leveraging AI to put it that way. As we like to do on the podcast as we move towards wrapping up,
    0:25:56 let’s talk about the future a little bit. Not to put you on the spot to make crystal ball predictions,
    0:26:03 but what are some of the things that are being explored as future directions for AI and energy
    0:26:10 efficiency and really supporting this growing energy demand, not just from AI and data centers,
    0:26:18 as you rightfully pointed out. Just in general, how is AI being explored to meet future energy
    0:26:25 demands across the globe? There are many ways, and most of them are yet to be discovered and talked
    0:26:33 about because we’re in such early days that it’s hard to know what opportunities are yet in front
    0:26:41 of us. Some examples are, for example, grid update. Updates to the grid in ways that will enable
    0:26:46 more renewable energy. When you have more and more renewable energy coming online,
    0:26:53 that is more cyclical than traditional energy. If you’re burning coal 24/7, it’s a steady stream
    0:27:00 of energy. If you’ve got wind or solar, it’s more variable. Also, with things like residential solar,
    0:27:07 you have times when you may be wanting to allow those residential solar panels to send energy
    0:27:13 onto the grid instead of pulling data for the house off the grid. All of those types of things
    0:27:19 benefit from modernization and modernization in a way where AI can play a significant role in
    0:27:26 helping to avoid waste and to create ways for the energy flow to be optimized. That’s one very near
    0:27:34 term area where we’re seeing progress. A partner of our Portland General Electric is using AI and
    0:27:39 using some of our products to do just that, to put smart meters around to help them manage the
    0:27:46 growth in renewable energy. There certainly is a perfect opportunity right now for us to do this,
    0:27:52 to engage in grid modernization because we have so much value to potentially unlock with AI.
    0:27:58 We’ve got these big companies who are really good at developing infrastructure who are motivated
    0:28:03 to help us modernize the grid and introduce more renewable energy and do that in a responsible
    0:28:11 way. It’s a perfect time for us to be focusing on this. There are also fantastic other sustainability
    0:28:18 related benefits from AI in terms of drug discovery for human welfare and materials
    0:28:24 discovery for things like electric vehicle batteries and batteries more generally. We’ve
    0:28:30 heard reports from Microsoft as well as from Google about discoveries they’ve made in material
    0:28:35 science that could potentially lead to much more efficient batteries in the future, which of course
    0:28:41 would not only save resources but also save energy as well. Right. It’s funny listening you talk about
    0:28:46 sustainability and it’s obviously related but it makes me think about or when you mentioned about
    0:28:52 residential solar and being able to send solar power back to the grid and thinking about the
    0:28:58 virtuous implications of that. It made me think about recycling for whatever reason and recycling
    0:29:04 being one of those things that individually if we all do it it’s forms of collective and then
    0:29:11 obviously if we’re moving from residential to industry and large corporations and factories
    0:29:18 and what have you the importance of recycling is sort of bigger obviously in these spots in a
    0:29:24 factory than in an individual house. I’m wondering about the importance of collective action when it
    0:29:30 comes to all of these things that you’ve been talking about with AI and sustainability and energy
    0:29:38 demand and then sort of dovetailing from that. What about the role of industry or even of governments
    0:29:45 in driving these kind of new and emerging best practices that will support sustainability for
    0:29:52 all of us? I think there’s a great role for governments policy makers to play here in terms
    0:29:59 of number one setting an example of how accelerating computing and AI can be used as tools for good
    0:30:04 and we’re seeing some great work by regulators especially in the United States but also in
    0:30:10 Europe and elsewhere where there’s a real appreciation of the potential benefits to society
    0:30:17 and to sustainability and adopting both accelerated computing and AI and using that for the public
    0:30:23 good. So I think that’s the first way in which I think there’s an opportunity here accelerating
    0:30:28 all these workloads transitioning them over to a sustainable platform and then using AI to try
    0:30:35 to benefit society in an environmental way and in a social way as well and then also to
    0:30:42 encourage that type of sustainable deployment in industry as well and make sure that we’re
    0:30:49 again modernizing the grid and creating the environment where we have a clear path towards
    0:30:56 sustainable deployment of AI because ultimately you know this is moving very very quickly. This
    0:31:01 fourth industrial revolution is we’re living through it. It’s very exciting and we don’t want
    0:31:08 anything to undermine our ability to capture the benefit from this so that we can try to mitigate
    0:31:14 climate change. We can develop new drugs. We can see all of the potential benefits from sustainability
    0:31:19 and we can have that at the same time with sustainable deployment of AI data centers if we’re careful.
    0:31:25 Absolutely. Josh, before we wrap I just want to mention the NVIDIA Deep Learning Institute is
    0:31:31 actually sponsoring this episode. I just want to give them a shout out and listeners out there who
    0:31:38 are interested obviously in AI but in AI and sustainability in particular head over to the
    0:31:44 NVIDIA Deep Learning Institute. There are DLI courses on AI and sustainability. You can learn
    0:31:49 so much more about what we’ve been talking about and go out and make an impact of your own which
    0:31:55 we obviously encourage everybody to do. Josh, kind of as we wrap up here and along those lines
    0:32:01 for listeners who would like to learn more about the work that you’re leading, the work that NVIDIA
    0:32:06 is doing on sustainability, energy efficiency, everything we’ve been talking about, maybe even
    0:32:11 some of the work that NVIDIA is doing with partners along these fronts. Where is a good place
    0:32:17 for a listener to go online to kind of start digging in deeper to this? We are publishing more
    0:32:23 and more content about the connection between auxiliary computing and AI and sustainability
    0:32:31 on our website. We have a sustainable computing sub-page on our public web page. We have some
    0:32:36 corporate blogs there, some white papers and so forth. Those were all really interesting and
    0:32:42 very readable I think in terms of giving you examples of how NVIDIA and our partners actually
    0:32:47 are doing so much good work in sustainability. We also of course publish an annual sustainability
    0:32:53 report if you’re interested in the corporate level view, what we’re doing in terms of energy
    0:32:57 efficiency in our systems and our own corporate commitments that’s in our annual sustainability
    0:33:02 report which is on our website as well. Fantastic. Closing thoughts, anything you want to leave
    0:33:08 the listeners to take with them or ponder when it comes to the future of AI, the future of
    0:33:15 sustainability, made better by AI, anything else we’ve touched on? I just offer a healthy dose
    0:33:22 of optimism. We’ve heard, we’ve heard I think an unhealthy dose of pessimism and skepticism about
    0:33:29 AI specifically in the realm of sustainability and again those are all legitimate questions but
    0:33:36 AI I currently believe is going to be the best tool that we’ve ever seen to help us achieve
    0:33:42 more sustainability and more sustainable outcomes. If we capture this, if we capture the moment and
    0:33:49 use AI for good and if we use this new auxiliary computing platform to drive better efficiencies
    0:33:54 then we’re going to see really dramatic and positive results over time as we do that more and
    0:33:59 more. That’s what we want and let’s end on the optimistic note. Josh, thank you so much for
    0:34:04 taking the time to come on and talk with us and it goes without saying but all the best of luck
    0:34:08 to you and your teams on the work you’re doing. It could be more important. Thanks Noah, really
    0:34:19 appreciate it.
    0:34:30 [Music]
    0:34:41 [Music]
    0:34:59 [Music]
    0:35:09 [BLANK_AUDIO]

    From improving energy efficiency to helping address climate challenges, AI and accelerated computing are becoming key tools in the push for sustainability. In this episode of NVIDIA’s AI Podcast, Joshua Parker, senior director of corporate sustainability, shared his perspective on how these technologies are contributing to a more sustainable future.

    https://blogs.nvidia.com/blog/ai-energy-efficiency/

  • SonicJobs CEO Mikhil Raja on Using AI Agents to Connect the Internet, starting with Jobs – Ep. 233

    AI transcript
    0:00:11 [MUSIC]
    0:00:13 >> Hello, and welcome to the NVIDIA AI podcast.
    0:00:15 I’m your host, Noah Craven.
    0:00:17 There’s been a lot of talk over the past year or
    0:00:22 two about whether or not AI will take jobs away from humans.
    0:00:23 Our guest today, however,
    0:00:27 is already using AI to connect more humans to more jobs,
    0:00:30 which is good for job seekers and good for employers.
    0:00:35 Companies in the United States alone spend $15 billion annually
    0:00:38 on clicks to advertise their job vacancies,
    0:00:43 but 95 percent of all job applications are abandoned before they’re completed.
    0:00:47 That’s a big problem for a lot of job seekers and a lot of employers.
    0:00:51 Mikhail Rajesh is the co-founder and CEO of Sonic Jobs,
    0:00:55 a startup and a member of NVIDIA’s Inception Program for startups,
    0:00:58 that’s making it easier for more candidates to apply to more jobs.
    0:01:01 Sonic Jobs uses a unique approach to
    0:01:03 agentic AI and web agents,
    0:01:06 which I’m excited to dig into with Mikhail now.
    0:01:08 So let’s welcome him on. Mikhail Rajesh,
    0:01:12 welcome and thank you so much for joining the NVIDIA AI podcast.
    0:01:13 >> Thanks, Noah, for having me.
    0:01:15 >> So there is a lot in there,
    0:01:17 in the intro teasing about Sonic Jobs.
    0:01:19 But since we have you,
    0:01:23 why don’t you tell the audience a bit about what Sonic Jobs is?
    0:01:26 >> Yeah. So companies in the US spend
    0:01:29 $15 billion on advertising their vacancies,
    0:01:34 and the way they do that has moved from per listing,
    0:01:37 which is how they used to advertise to per click,
    0:01:39 where it’s performance-based.
    0:01:45 The most important metric in that performance-based recruitment advertising space
    0:01:49 is the conversion from the paid click to the completed apply.
    0:01:51 As you touched on in your introduction,
    0:01:53 industry-wide today,
    0:01:57 only 5 percent of people complete the job application,
    0:02:01 which means 95 percent of them abandoned that application.
    0:02:02 >> That’s astounding.
    0:02:05 Only 5 percent of applications that get
    0:02:08 started are actually completed and submitted.
    0:02:09 I figured, yeah,
    0:02:12 I know applications get abandoned, but that is amazing.
    0:02:15 Forgive me, I just wanted to express that.
    0:02:19 >> Yeah, and the biggest reason for that is the redirection.
    0:02:22 So candidates today still go through
    0:02:27 the same 1.0 experience as you did in the 90s,
    0:02:29 where when you want to apply for the job,
    0:02:30 you start on the job platform,
    0:02:33 and then for each job that you want to apply for,
    0:02:36 you get redirected to the company site to apply.
    0:02:38 As soon as you’re redirected,
    0:02:40 you have a 70 percent bounce rate.
    0:02:45 So the majority, the biggest factor in that 95 percent is this redirection step,
    0:02:50 which causes the huge friction that we experience in the market.
    0:02:53 >> Right. So just so I understand from the job seekers perspective,
    0:02:55 I’m on whatever, I don’t even know,
    0:02:59 dice or LinkedIn or Monster indeed,
    0:03:00 whatever they are, the job boards.
    0:03:04 I see a job for podcaster at NVIDIA and I click,
    0:03:06 I’m making this all the way up, I don’t know, but I click,
    0:03:08 and so then the redirect is,
    0:03:11 it takes me from the job board to, in this case, NVIDIA,
    0:03:13 but wherever the employer is site,
    0:03:18 and you said you’re losing 70 percent of people just on the bounce.
    0:03:19 >> Exactly right. Yes.
    0:03:20 >> Okay.
    0:03:22 >> Historically, people have tried to build
    0:03:26 APIs to connect this whole ecosystem,
    0:03:29 but it’s too fragmented an ecosystem.
    0:03:31 So there’s hundreds of job platforms,
    0:03:33 some of which you touched on just now.
    0:03:39 There’s over 200,000 employers that advertise vacancies in America,
    0:03:41 and there’s over 10 million jobs,
    0:03:43 each of which will have different flows and
    0:03:47 different requirements for each application process,
    0:03:50 and so APIs haven’t worked,
    0:03:53 which is why you still have this experience that we touched on,
    0:03:56 which is actually still the 1.0 experience,
    0:04:00 web 1.0 experience where you start and then you’re redirected,
    0:04:01 and you have to complete the application.
    0:04:04 >> So the idea of the API would be,
    0:04:08 if I was on a job board where I’d already entered some info,
    0:04:10 or maybe they have a profile of me,
    0:04:11 and it’s got my basic info,
    0:04:16 and the API theoretically would take that info,
    0:04:20 and as I’m redirected to the employer site,
    0:04:21 it would carry some of that with me,
    0:04:24 so I don’t have to re-enter all the basic stuff
    0:04:26 that you always have to re-enter, got it.
    0:04:28 But those APIs don’t work.
    0:04:30 I’m seeing you nod, I’m thinking,
    0:04:32 well, the whole web is made out of spaghetti at this point,
    0:04:33 but I’m not a developer.
    0:04:35 So is that what’s going on with the APIs?
    0:04:38 >> Yeah, exactly. It’s not worked,
    0:04:40 and just to elaborate on your point,
    0:04:44 ideally, you would apply directly on the platform that you’re on.
    0:04:47 When you’re on booking.com,
    0:04:51 you don’t get redirected to Hilton if you want to book a room,
    0:04:53 you can just book on booking.com.
    0:04:56 So if you take that analogy here,
    0:04:59 you should be able to apply directly on the job platform,
    0:05:01 which is exactly what we’ve built.
    0:05:03 So instead of using APIs,
    0:05:05 what we use is AI agents,
    0:05:07 which I know you want to talk about,
    0:05:09 I’m happy to go deeper into that,
    0:05:13 where effectively we understand using
    0:05:15 computer vision and also the HTML,
    0:05:19 every single input field of the job application,
    0:05:23 so that when the candidates on the job site impresses apply,
    0:05:27 we ask all of the relevant questions directly on the site,
    0:05:30 and then the AI agent takes all of that data,
    0:05:32 and on behalf of the candidate,
    0:05:36 submits their data on to the job application flow
    0:05:38 to complete the application.
    0:05:39 And so instead of that 5%,
    0:05:41 which we touched on earlier,
    0:05:43 and that 95% wastage,
    0:05:45 we’ve got a 26% conversion,
    0:05:50 so five times better than a standard job application flow,
    0:05:52 which is fantastic.
    0:05:55 It is, five-fold increase.
    0:05:57 I want to divert from my notes here
    0:05:59 and ask you about that business end of things,
    0:06:01 because I’ll just ask you this question,
    0:06:03 then we’ll come back to the business stuff.
    0:06:05 But how long did it take you to achieve
    0:06:08 that five-fold increase in performance
    0:06:10 over the industry standard?
    0:06:13 Yeah, we’ve been building our technology for five years,
    0:06:19 and it combines what we call traditional AI with LLMs
    0:06:21 and kind of the new wave of AIs.
    0:06:23 Yeah, happy to talk more about that as well.
    0:06:25 Yeah, yeah, so pin in that bad question for me.
    0:06:26 So that’s a teaser for later.
    0:06:27 Stick around, audience.
    0:06:32 We’ll find out about how Sonic Johns has grown so quickly.
    0:06:33 But you mentioned computer vision,
    0:06:36 and you mentioned looking at the HTML.
    0:06:38 So maybe let’s dig in.
    0:06:40 I sort of want to ask you from two perspectives.
    0:06:41 From the user perspective,
    0:06:44 I just want to clarify kind of the sequence
    0:06:47 and where the agent comes in and starts working.
    0:06:51 And then I’m curious about how the computer vision works,
    0:06:52 and then also what’s happening
    0:06:54 at the use of HTML behind the scenes.
    0:06:57 Yeah, so from the user perspective,
    0:06:58 they stay on site.
    0:07:00 So you get this win-win scenario
    0:07:03 where the candidates answering all the questions,
    0:07:05 they stay on the site that they’re on.
    0:07:08 You know, the good analogy is the booking.com experience.
    0:07:12 You fill out all of the details on the platform itself.
    0:07:14 On the ground board to–
    0:07:16 Yeah, exactly, exactly.
    0:07:16 And then you’re done.
    0:07:18 You press supply, it’s all done.
    0:07:21 From the employer’s perspective
    0:07:24 and kind of how that all maps through,
    0:07:27 because we’re using AI agents
    0:07:31 and the input fields are all publicly facing
    0:07:34 from a job application flow perspective,
    0:07:37 they don’t need to do any development work.
    0:07:40 So we’re asking exactly the questions that are needed
    0:07:43 for that job apply flow that’s on their website,
    0:07:46 and they don’t need to kind of build an API
    0:07:48 or have tech resources, et cetera.
    0:07:50 And they only pay when they–
    0:07:52 The employer, et cetera, exactly.
    0:07:55 And they only pay when they receive the application.
    0:07:56 So they pay–
    0:07:57 Right, right, like that conversion–
    0:07:58 At the app–
    0:07:59 At the Alphabet, smile, yeah.
    0:08:00 Right.
    0:08:05 So is this CV, is your agent sort of looking
    0:08:08 at the job application web page
    0:08:12 and visually making sense of what’s being asked?
    0:08:13 Like, oh, this is the first name form,
    0:08:15 this is the last name form, this is the,
    0:08:20 why do you want to work at, you know, Acme Company SA field?
    0:08:22 Like, is that what the CV is doing?
    0:08:23 You said CV, I think–
    0:08:25 The computer vision element.
    0:08:27 It jumped out to me when you said computer vision before.
    0:08:27 So–
    0:08:30 Yeah, CV in the UK means–
    0:08:31 Oh, right, curriculum.
    0:08:32 Sorry, so I got–
    0:08:34 Of course, my bad.
    0:08:37 So, yeah, that’s exactly what the agent’s doing.
    0:08:41 The average job application form has six pages
    0:08:43 and 40 input fields.
    0:08:45 And so the agent is understanding
    0:08:46 every single input field.
    0:08:49 It can be radio buttons, drop downs, open questions,
    0:08:51 open text, conditional questions,
    0:08:54 all that sort of stuff is transforming that
    0:08:57 to kind of a context and to a structure
    0:08:59 and then asking those questions exactly right.
    0:09:02 And also inputting that data back into those questions.
    0:09:05 So maybe that can lead us into this talk
    0:09:07 about AI agents sort of broadly,
    0:09:09 because I think a lot of what I’m thinking about
    0:09:13 goes back to like a year, even a year and a half ago
    0:09:15 where there was buzz around auto GPT
    0:09:17 and some other open source projects.
    0:09:20 Talk to us about agents, how Sonic Jobs uses them,
    0:09:23 and maybe how that’s different from other approaches.
    0:09:26 Yeah, so you’re exactly right now.
    0:09:30 12 months ago, auto GPT, baby AGI.
    0:09:31 Baby AGI, yeah, yeah.
    0:09:32 Captured our imagination.
    0:09:36 It was like the agent that can do everything.
    0:09:41 What we’ve seen is that the architecture of the agent
    0:09:45 needs to combine very tightly the application layer
    0:09:47 and the reasoning layer.
    0:09:51 And that’s because accuracy and reliability
    0:09:53 are really important for companies.
    0:09:59 Like the challenge with baby AGI and auto GPT
    0:10:02 was that LLMs are not deterministic.
    0:10:05 And so they can’t do step one, step two, step three,
    0:10:08 step four in the order that you want them to
    0:10:12 whilst an enterprise client wants the output
    0:10:15 and the workflow to be structured in a particular manner,
    0:10:17 which is domain specific.
    0:10:19 You know, they’ll have specific steps to,
    0:10:22 and again, our domain is obviously job applications.
    0:10:27 And that means that it’s really important to combine
    0:10:29 in our experience, in our view,
    0:10:32 to combine what we think of as traditional AI,
    0:10:35 which is very high on accuracy
    0:10:39 and lower on, let’s say, generalizability.
    0:10:41 Combining that with the LLMs,
    0:10:44 which can actually generate data
    0:10:48 and use the memory of the successful training data
    0:10:52 to basically use that for future,
    0:10:54 use cases for future workflows, et cetera.
    0:10:56 And so you’ve kind of got this combination,
    0:10:59 which we think is really powerful
    0:11:02 and very domain specific.
    0:11:04 And that’s where our experience
    0:11:08 of kind of this vertical AI agent is that,
    0:11:13 it’s much more powerful for the specific B2B use cases
    0:11:17 versus this general agent that can kind of do everything,
    0:11:20 which actually ends up doing nothing
    0:11:22 rather than being able to do everything.
    0:11:24 So it’s kind of interesting.
    0:11:26 – This is me we’re talking about, I’m not a developer,
    0:11:28 but my own experiments with those early systems
    0:11:31 did a whole lot of nothing as it turned out,
    0:11:32 but that’s a separate take.
    0:11:35 So this approach to agents that you’re talking about,
    0:11:37 when Sonic Jobs, well, at whatever point
    0:11:39 in the Sonic Jobs story,
    0:11:44 did you sort of see this approach to agentic AI
    0:11:49 and kind of think, oh, this maybe is a good solution
    0:11:51 to this problem we’re seeing with the redirects
    0:11:55 and the other things in the recruiting industry,
    0:11:57 or did it come about another way?
    0:11:59 What was kind of the moment where you thought,
    0:12:01 oh, this could be the approach we use?
    0:12:03 – Yeah, it’s a really good question.
    0:12:07 I spent three years before Sonic Jobs at AutoTrader,
    0:12:10 which is a karma place.
    0:12:15 And I became obsessed with this idea
    0:12:18 that you could remove friction
    0:12:23 and have much better engagement by users.
    0:12:27 And particularly as mobile came in,
    0:12:31 what I saw in the job space was that
    0:12:33 this redirection point of friction
    0:12:37 was becoming a bigger and bigger issue.
    0:12:40 And we thought about it in the context
    0:12:44 of really creating this API-less API.
    0:12:48 And so we create kind of this version of an API
    0:12:51 where actually the company did no tech work,
    0:12:55 but the candidate could have a really seamless experience
    0:12:58 like they do on booking.com.
    0:13:02 And yeah, it’s taken us the majority of our existence
    0:13:03 to build what we’ve built.
    0:13:05 So to answer your question,
    0:13:09 we’re 28 people, 24 engineers,
    0:13:13 four commercial people including myself.
    0:13:16 And over the five years, we’ve spent over three years
    0:13:18 just building out the technology.
    0:13:22 And last year, we launched in the U.S.
    0:13:24 That’s gone super well.
    0:13:28 And we’re primarily a U.S. business now.
    0:13:30 But yeah, the engineering part of,
    0:13:32 and the agent part of our business
    0:13:34 is at the very heart of what we do.
    0:13:37 – Of course, we were talking before we started recording
    0:13:41 that you moved from England to the U.S.
    0:13:44 Is the rest of the company global in the U.S. and England
    0:13:47 where are the semi-jobs people?
    0:13:49 – Yeah, we’re a fully remote team.
    0:13:51 I should mention also my co-founder
    0:13:54 has a background in robotics and AI.
    0:13:57 So that’s how we all kind of came to this conclusion.
    0:14:00 And at the time we started, this didn’t even have a name.
    0:14:03 It’s amazing that everything’s called AI agents now.
    0:14:08 It’s a term and agents kind of represent lots of things now.
    0:14:11 Maybe we could call what we do AI web agents,
    0:14:13 specifically for the web.
    0:14:17 And we used to call it kind of AI plus RPA.
    0:14:19 No one had a clue what that meant.
    0:14:21 – What’s RPA?
    0:14:22 – Robotic process automation.
    0:14:26 So it’s kind of workflow automation,
    0:14:28 which is a little bit more manual.
    0:14:30 And so we were kind of combining the two.
    0:14:32 But yeah, the agent term has taken off
    0:14:34 and it’s come a long way in the last five years.
    0:14:36 – I want to ask you more about the agent thing.
    0:14:39 I want to go back for a second, though, to the company.
    0:14:42 You mentioned that the business has sort of taken off
    0:14:44 and shifted focus to the U.S.
    0:14:47 So you’re primarily working with U.S.-based employers?
    0:14:49 – Yeah, we work with the largest,
    0:14:51 some of the largest U.S. employers today.
    0:14:54 So Dish, CBS, Walgreens.
    0:14:59 And we’ve seen huge traction for our use case.
    0:15:02 – And then also mentioned the intro
    0:15:05 that you’re in the NVIDIA deception program.
    0:15:07 – Yep, NVIDIA has been awesome with us.
    0:15:09 And we’re excited.
    0:15:13 We’re building kind of deeper RAG infrastructure
    0:15:15 or with NVIDIA’s help.
    0:15:18 And it’s been, yeah, great collaboration.
    0:15:21 – What’s on the Sonic Jobs roadmap right now?
    0:15:23 Are you, is it just businesses boom in
    0:15:24 and you’re doing your thing?
    0:15:28 Is there a product roadmap that you’re oriented towards?
    0:15:30 – Yeah, I think it’s really important to emphasize
    0:15:34 how early we are in this journey on agents in particular.
    0:15:36 And it’s taken us five years,
    0:15:40 but we’re still, I think as the whole ecosystem,
    0:15:42 still at the very start of that journey.
    0:15:46 And so today our agent can interact
    0:15:50 with about 60,000 jobs in the U.S.
    0:15:54 So it understands the input fields from 60,000 jobs.
    0:15:57 The total market size just in the U.S. alone
    0:15:59 is 10 million jobs.
    0:16:02 And so there’s a long way to go in terms of,
    0:16:05 even with a specific domain,
    0:16:07 there’s a long way to go in terms of being able
    0:16:11 to understand every single input field, every single job.
    0:16:15 – Is it a matter of just continuing to chip away
    0:16:18 at a sort of very big unwieldly problem,
    0:16:20 which is what you described earlier
    0:16:24 about making sense of the multi-page application form
    0:16:26 or are there other technical things?
    0:16:28 We’re talking agents on the pod.
    0:16:31 So if there’s more interesting stuff you can talk about,
    0:16:32 I would love to hear.
    0:16:34 – Yeah, no, I love this, I love this.
    0:16:37 The real answer is we don’t fully know.
    0:16:41 There is a scale element to this
    0:16:46 where the more that you get successful applications
    0:16:49 going through, and we’ve got now millions
    0:16:52 of successful applications that have gone through our agent,
    0:16:57 the more the model learns and is able to adapt
    0:17:00 to the next job application flow that it sees,
    0:17:03 which has been very successful as a technique
    0:17:06 and continues to bear fruit for us,
    0:17:09 the more clients we work with, et cetera.
    0:17:13 There’s a reasoning, a general reasoning layer.
    0:17:16 So where you think about the kind of LLMs
    0:17:18 are not great as we talked about earlier,
    0:17:21 following steps, planning, et cetera.
    0:17:23 I’m confident that will get better as well,
    0:17:27 and that will enhance our kind of vertical approach
    0:17:30 within the application layer as well.
    0:17:32 I think it’ll be a combination of a few things
    0:17:35 that basically end up kind of being able
    0:17:40 to create a vertical AI agent across the entire domain.
    0:17:43 – Is it, you’re talking about the application layer
    0:17:44 and the reasoning layer.
    0:17:47 And if I understand it, or to the extent I understand it,
    0:17:49 it makes a lot of sense to me that you know,
    0:17:53 you’ve got whether it’s processing job applications
    0:17:58 or submitting, I should say, I guess,
    0:18:01 or another business enterprise task,
    0:18:02 whatever it might be.
    0:18:05 So I understand that sort of the verticalized,
    0:18:08 specific tasks that the agent is being developed to do.
    0:18:09 And then underneath that,
    0:18:12 you’ve got this reasoning layer you talked about,
    0:18:14 which is the LLM.
    0:18:14 – Yep.
    0:18:17 – And so when you’re talking about the reasoning layer
    0:18:20 needing to get better, correct me if I’m wrong,
    0:18:23 but my understanding is you’re not building foundational models
    0:18:27 from scratch, but you are doing a lot of work
    0:18:30 to kind of fine-tuning them the right way to say it.
    0:18:31 – Exactly right.
    0:18:34 – To get them to perform how you want them to
    0:18:36 in conjunction with your system.
    0:18:38 So all this is getting to my question,
    0:18:42 which is, is there a lot that a company like yours can do
    0:18:45 to kind of tweak the reasoning layer?
    0:18:49 Or is it mainly a matter of waiting for the,
    0:18:51 the various companies of the world
    0:18:53 who are releasing foundational models,
    0:18:56 is it kind of more like waiting for them
    0:19:00 to ship something better and then you can,
    0:19:02 or is it kind of a combo of both?
    0:19:06 – No, there’s a ton you can do on the fine-tuning side.
    0:19:09 And also what we call on the tooling side.
    0:19:13 So traditional AI now is referred to almost as tools
    0:19:17 and the LLMs can combine with these tools
    0:19:20 and the fine-tuning that you’ve put on top of the LLMs
    0:19:23 to create this kind of hybrid structure.
    0:19:25 And you touched on it and then it’s an important point
    0:19:30 to emphasize in domain specific workflows,
    0:19:34 you don’t want randomized outputs.
    0:19:38 You want predictable, accurate outputs.
    0:19:42 And so depending on your domain,
    0:19:46 you need to structure the architecture
    0:19:50 so that you might even choose to kind of hard code
    0:19:55 or create specific layers that do specific tasks
    0:19:57 and then specific layers that create,
    0:20:00 that do reasoning tasks or do error resolution tasks
    0:20:03 or do input detection tasks.
    0:20:07 And so our view, and you asked about kind of all developers,
    0:20:08 so I’ll maybe touch on that.
    0:20:13 Our view is that there are always going to be
    0:20:18 application layer architecture that’s going to be needed
    0:20:22 to create a vertical specific use case
    0:20:24 for a particular enterprise.
    0:20:27 And that’s going to be always really valuable.
    0:20:29 And again, you’re kind of riding on the wave
    0:20:32 of the LLMs becoming smarter because then you need to do,
    0:20:35 your value add can be more and more
    0:20:37 and you can scale faster and faster.
    0:20:39 But there’s a ton you can do,
    0:20:42 even as a small company like ourselves,
    0:20:45 to create that domain specificity,
    0:20:47 which is hugely valuable.
    0:20:51 – So your technology, and not that Sonic Jobs
    0:20:56 is focused on anything besides the talent acquisition industry,
    0:20:59 but is the approach you’re taking,
    0:21:01 could it serve as a framework,
    0:21:06 even just conceptually for kind of similar work
    0:21:07 in a different domain?
    0:21:10 – Yeah, I would answer your question in two ways.
    0:21:14 One is we think that as you get better
    0:21:16 at a specific domain and vertical,
    0:21:18 it actually gives you the springboard
    0:21:21 to potentially explore other domains,
    0:21:23 which we think is interesting in a way
    0:21:27 that we think is more valuable and more likely
    0:21:30 to be more successful than starting as a generalist
    0:21:32 and moving to being a generalist,
    0:21:33 actually starting at a vertical
    0:21:35 and then moving out to other virtuals
    0:21:38 we think could be more fruitful.
    0:21:41 The second thing is, and it’s a little bit of a,
    0:21:43 as I mentioned, I’m new to Silicon Valley,
    0:21:48 so I’ll say how I see it, obviously, as someone new.
    0:21:52 I think AI and AI agents today
    0:21:54 have been largely built by people
    0:21:57 who are excited about the technology
    0:21:59 and looking for a problem.
    0:22:02 I think there’s a lot of room,
    0:22:05 and I think the other group is the group of people
    0:22:08 that have a problem, whether that’s a B2C problem
    0:22:12 or a B2B problem, and then look at AI
    0:22:15 as creating a solution that could never be created before.
    0:22:19 I think the way we’ve done it is the latter
    0:22:22 in that we had a problem and we looked at AI
    0:22:23 to create a unique use case,
    0:22:26 and I think I would encourage anyone listening to this,
    0:22:29 and you mentioned that there’s a lot of engineers
    0:22:31 that listen to this.
    0:22:34 I would encourage anyone with a problem
    0:22:38 to know that it’s open and not too often.
    0:22:40 It can feel like a wall of garden AI
    0:22:42 coming in from the outside,
    0:22:44 particularly if you’re not from Silicon Valley.
    0:22:47 I would encourage people to lean into the problems
    0:22:48 that they have.
    0:22:52 – I mean, even at my sort of 50,000-foot level,
    0:22:54 what you said just kind of rang true.
    0:22:58 Simple as it sounds, that idea of LLMs in particular,
    0:23:01 almost by nature, are this semi-black box
    0:23:03 or just this emergent capabilities,
    0:23:05 all these words that are used to describe
    0:23:09 this kind of general process of figuring out what it can do,
    0:23:11 and so that approach of starting with,
    0:23:13 I don’t know, I personally, if you give me a blank sheet
    0:23:15 of paper, I have the hardest time coming up with words,
    0:23:18 but if you say, “Hey, here’s what we need,”
    0:23:20 so yeah, there’s kind of a similar thing
    0:23:22 resonated with me hearing you say that.
    0:23:25 All right, got one last question for you before we wrap up.
    0:23:27 It’s a little bit of a flip.
    0:23:30 We usually like to ask how the work you’re doing
    0:23:33 and AI’s impact on the work you’re doing
    0:23:35 is gonna affect things going forward,
    0:23:37 which we’ve talked about a little bit already,
    0:23:42 but since you’re in the hiring space, so to speak,
    0:23:44 and we’ve been talking about the inefficiency
    0:23:47 of the posting model, changing the conversions
    0:23:49 and technology like Sonic Jobs,
    0:23:50 increasing that conversion rate.
    0:23:53 Again, to put you on a spot here,
    0:23:55 what advice would you give to a company
    0:23:58 who’s trying to hire right now,
    0:24:01 depending on the industry, we’re hearing a lot about
    0:24:03 the economy and employment in certain industries,
    0:24:07 talent crunches and others, can’t find a qualified person.
    0:24:09 So kind of given that broad spectrum,
    0:24:12 are there any kind of high level pieces of advice
    0:24:14 you’d give to somebody out there
    0:24:16 who’s trying to advertise and hire for roles,
    0:24:17 but just having a hard time?
    0:24:21 – Yeah, I think the simplest piece of advice would be
    0:24:26 to go to a job platform, find your job,
    0:24:29 and apply for your job on the job platform.
    0:24:32 What most companies do is apply for their job
    0:24:37 or look at their apply flow on their own company site.
    0:24:38 – Oh, but not from the, yeah, okay.
    0:24:41 – But you’ve got to remember that most candidates
    0:24:43 start their journey on a job platform,
    0:24:47 whether that’s LinkedIn or Indeed or Sonic Jobs,
    0:24:49 and really think about that experience
    0:24:51 from the job seekers perspective,
    0:24:55 because if you do that, you’ll optimize your own conversion
    0:24:59 and have a better experience for your candidates,
    0:25:00 which is pretty cool.
    0:25:02 – For folks who are listening
    0:25:04 and would like to learn more about Sonic Jobs
    0:25:08 from whatever perspective, the technical perspective,
    0:25:12 maybe they’re an employer who’s looking for new ways to hire,
    0:25:13 or I guess maybe the job seekers
    0:25:15 can go look at job listings as well.
    0:25:18 Where would you send people to go online?
    0:25:20 What all is on the Sonic Jobs website?
    0:25:23 And then is there social media as well
    0:25:24 that listeners can follow?
    0:25:27 – Yeah, so if you’re an employer
    0:25:30 and you want to talk more, add me on LinkedIn,
    0:25:35 Michael Raja, if you’re a job seeker, AI engineer
    0:25:36 that think we’re doing something cool,
    0:25:39 we’re hiring, so please reach out to me.
    0:25:41 And if you just want to learn about our technology,
    0:25:46 we’ve created a page, Sonic Jobs/AI agent on our website
    0:25:49 where you can learn more about our technology.
    0:25:51 – Great, sidejobs.com.
    0:25:51 – Exactly.
    0:25:54 – We’re into the age that used to be
    0:25:56 way back when it was dot com,
    0:25:57 but people would use other ones
    0:25:58 if they couldn’t get the name.
    0:26:00 And then it was all dot com, org.
    0:26:03 Now we’re getting like the dot AI and the other search.
    0:26:04 I don’t remember to ask.
    0:26:07 So sonicjobs.com, fantastic.
    0:26:09 McKeil, I feel like we could talk Asians
    0:26:12 and talk the future of the web for longer,
    0:26:15 let alone the talent acquisition industry.
    0:26:17 But I think that this gives a great purview
    0:26:20 into what Sonic Jobs is doing, has been doing,
    0:26:24 and also kind of bringing back up the topic of agents,
    0:26:25 which is interesting to think about
    0:26:28 how things have changed in a short amount of time.
    0:26:29 But it feels like a long amount of time
    0:26:31 since Jena AI hit the scene.
    0:26:34 All the best to you and your whole team.
    0:26:35 Thank you for coming on.
    0:26:38 And yeah, any words in closing?
    0:26:40 – No, thanks very much for having me, really exciting.
    0:26:41 – Fantastic.
    0:26:44 (upbeat music)
    0:26:45 .
    0:26:47 (upbeat music)
    0:26:50 (upbeat music)
    0:26:53 (upbeat music)
    0:26:55 (upbeat music)
    0:26:58 (upbeat music)
    0:27:00 (upbeat music)
    0:27:03 (upbeat music)
    0:27:05 (upbeat music)
    0:27:08 (upbeat music)
    0:27:11 (upbeat music)
    0:27:13 (upbeat music)
    0:27:16 (upbeat music)
    0:27:18 (upbeat music)
    0:27:21 (upbeat music)
    0:27:24 (upbeat music)
    0:27:26 (upbeat music)
    0:27:29 (upbeat music)
    0:27:31 (gentle music)
    0:27:40 [BLANK_AUDIO]

    Companies in the US spend $15bn annually on talent acquisition. The most important metric in recruitment advertising is the conversion from the paid click on the job platform to the application the employer receives. Industry-wide, apply conversion is just 5%. Redirection of the candidate from the job platform to the company site is the biggest cause of abandonment; this step has a 70% bounce rate. In this episode of NVIDIA’s AI Podcast, host Noah Kravitz speaks with Mikhil Raja, Cofounder and CEO of SonicJobs, about how they have built AI Agents to enable candidates to complete applications directly on job platforms, without redirection, boosting completion rates to 26% from 5%. Raja delves deep into SonicJobs’ cutting-edge technology, which merges traditional AI with large language models (LLMs) to understand and interact with job application web flows. He also emphasizes the importance of fine-tuning foundational models to achieve more impactful and scalable innovations.

    SonicJobs is a member of the NVIDIA Inception program for startups.

  • Machina Labs’ Edward Mehr on Autonomous Blacksmith Bots and More – Ep. 232

    AI transcript
    0:00:10 [MUSIC]
    0:00:13 >> Hello, and welcome to the NVIDIA AI podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:17 One of the most fascinating,
    0:00:21 dynamic, and visceral applications of deep learning,
    0:00:23 machine learning, computer vision,
    0:00:25 all these terms that we now put under
    0:00:28 the AI umbrella is the world of robotics.
    0:00:32 Robotics has been a vibrant space for years and years now,
    0:00:35 but since the advent of large language models,
    0:00:36 the interest in robotics,
    0:00:38 the exposure it’s getting to the mainstream audience,
    0:00:41 and certainly some of the ideas about how we can
    0:00:45 communicate, train, and leverage robotics has expanded.
    0:00:48 Here today to talk to us about the world of robotics,
    0:00:52 specifically in the manufacturing industry is Ed Mer.
    0:00:55 Ed is co-founder and CEO of Makina Labs,
    0:00:58 and I am absolutely delighted to welcome Ed
    0:00:59 onto the podcast.
    0:01:01 This has been in the works for a little while now,
    0:01:04 so I’m excited to hear all about what Makina Labs is up to,
    0:01:06 and trust me, it’s some really, really cool stuff.
    0:01:08 Ed, thank you so much for taking the time to
    0:01:11 join the NVIDIA AI podcast, and welcome.
    0:01:13 >> Thank you. Great to be here.
    0:01:17 >> So let’s start at the beginning, if you will.
    0:01:20 Tell us a little bit about what Makina Labs does,
    0:01:23 maybe how it got started, and we’ll go from there.
    0:01:26 >> Sounds good. So at Makina Labs,
    0:01:29 we’re trying to, or we’re building the next generation
    0:01:31 of manufacturing floors, and the main enabler is
    0:01:33 artificial intelligence and robotics.
    0:01:36 The main challenge that we’re trying to solve is
    0:01:39 something that I learned in my past careers,
    0:01:42 and it is that every time you have to build a physical part,
    0:01:44 a part made out of the material,
    0:01:47 a design that’s made out of physical material,
    0:01:50 you pretty much have to build a factory around it.
    0:01:53 There’s a lot of machinery and equipment that you need to
    0:01:56 build that are specifically designed for that part,
    0:01:59 for that geometry, for that material that part is
    0:02:02 using that cannot easily change.
    0:02:04 Every time you want to change the design and material,
    0:02:06 they have to go change these machineries and
    0:02:08 these toolings that you have to deploy in your shop.
    0:02:10 So what we’re trying to do is,
    0:02:13 can we build a truly software-defined factory?
    0:02:15 A factory that can change its design,
    0:02:18 can change its operation without having to change
    0:02:20 its machinery, which is very expensive,
    0:02:22 it takes a long time.
    0:02:24 These challenges have been
    0:02:27 prevalent in manufacturing for the past 100 years,
    0:02:30 and I think right now we are at the point where
    0:02:32 both robotics and AI are mature enough that we
    0:02:34 can rethink this paradigm of how
    0:02:37 manufacturing floors and shop floors have been built.
    0:02:39 >> So when you mentioned learnings you picked up in
    0:02:42 your previous jobs,
    0:02:45 you were at SpaceX and also Relativity Space, is that right?
    0:02:46 >> Great.
    0:02:48 >> Were you working on manufacturing and
    0:02:50 sort of the everything surrounding
    0:02:52 the manufacturing process there as well?
    0:02:53 >> Yes.
    0:02:54 >> Got it.
    0:02:55 >> Okay. So I have
    0:02:57 little experience with manufacturing myself.
    0:02:59 I’ve been living in the digital world as
    0:03:01 a content creator for a while now,
    0:03:03 but it makes sense to me that if you want to
    0:03:04 create something physical,
    0:03:06 you have to build the physical tooling
    0:03:07 and such around it to create it,
    0:03:09 and obviously, tearing that down,
    0:03:11 reconfiguring would be an expensive process.
    0:03:13 So when you talk about the idea of
    0:03:16 a software-defined manufacturing floor,
    0:03:17 what does that mean?
    0:03:19 Because obviously, I would think,
    0:03:21 we’re not able to turn zeros and
    0:03:24 ones into physical output as such.
    0:03:26 So how does that work? What does that mean?
    0:03:27 >> Yeah, no, absolutely.
    0:03:29 So I have a little bit of
    0:03:31 a kind of an interesting, not interesting,
    0:03:33 maybe a more hybrid background.
    0:03:34 So my education-
    0:03:36 >> I bet it’s interesting, but we’ll see.
    0:03:37 >> Yeah. Appreciate it.
    0:03:41 My education was mostly focused around robotics and software.
    0:03:43 When I went to school,
    0:03:46 academically, did computer engineering,
    0:03:48 and then focused more specifically
    0:03:49 toward machine learning and
    0:03:51 protocol modeling and AI.
    0:03:53 I spend early days in my careers at
    0:03:55 companies like Google and Microsoft.
    0:03:59 But always had it been for manufacturing since I was a kid.
    0:04:02 I did a lot of welding and
    0:04:04 carpentry when I was in school.
    0:04:06 So I always wanted to be able to combine and bring
    0:04:10 this world of software and robotics into manufacturing.
    0:04:13 So when I went to SpaceX,
    0:04:16 that’s when I started to learn how tough it is
    0:04:17 to manufacture parts.
    0:04:19 So to illustrate that,
    0:04:21 I’ll give you an example from one of
    0:04:23 the projects we were working on in SpaceX around
    0:04:27 2010-2012 timeframe, which is Falcon 9.
    0:04:31 Once we decided on the diameter of the Falcon 9 rocket,
    0:04:33 a lot of tooling, a lot of equipment in
    0:04:37 the factory were configured for that diameter, right?
    0:04:38 Meaning that, for example,
    0:04:41 the tooling that you used to build
    0:04:45 the tanks for that rocket are all fixed at that diameter.
    0:04:47 The moment you want to go get a fatter rocket,
    0:04:49 then you want that because a fatter rocket means
    0:04:51 you can put more fuel in it,
    0:04:53 it can go to higher orbits,
    0:04:54 but it’s just not an easy change.
    0:04:56 It means hundreds of millions of dollars investment
    0:04:57 in the factory to change that.
    0:04:59 So actually, if you follow Falcon 9
    0:05:03 throughout its life or just Falcon family,
    0:05:05 you will see that the rocket kept getting
    0:05:08 taller but could never get larger in diameter.
    0:05:09 As a matter of fact,
    0:05:11 if you look at SpaceX in the past 20-something years
    0:05:13 that the company has existed,
    0:05:18 they have two rocket families that they have developed,
    0:05:19 actually one and a half.
    0:05:20 So there’s a Falcon family,
    0:05:22 and there is a Starship family,
    0:05:23 which is significantly larger rocket
    0:05:26 and still hasn’t become productionized yet.
    0:05:28 So one and a half, that’s why I call it one and a half.
    0:05:31 In order to build Starship,
    0:05:32 they had to go start from scratch,
    0:05:35 new facility, but a lot of new tooling out in Texas.
    0:05:37 So that’s what I mean when we say
    0:05:40 manufacturing is very hardware-specific.
    0:05:43 Now, there are technologies that have been
    0:05:45 developed in the past few decades
    0:05:49 that allows us to slightly move away from that.
    0:05:50 3D printing is one of them.
    0:05:53 That’s why I’ve moved from SpaceX to Relativity Space,
    0:05:55 and the goal at Relativity Space is,
    0:05:59 can we 3D print as much of the rocket as we can
    0:06:02 to get rid of this limitation,
    0:06:03 a limitation where we cannot change
    0:06:05 the design of the rocket if you want to.
    0:06:07 So over there, I was in charge of a team
    0:06:09 that was building a very large and low-metal 3D printing.
    0:06:13 You could print structures that were like 15, 20 feet wide
    0:06:15 and 20, 30 feet in length.
    0:06:17 And this was basically a robotic arm
    0:06:19 with a welder attached to it
    0:06:23 that was welding layer by layer,
    0:06:24 basically what 3D printing is,
    0:06:27 layer by layer a very complex structure
    0:06:28 that could be very large.
    0:06:30 So for the first time,
    0:06:33 techniques like 3D printing kind of give this promise of,
    0:06:35 now you can actually turn a design
    0:06:38 into something that incrementally can build a geometry,
    0:06:41 and it’s not tied by a specific tooling.
    0:06:42 The challenge with 3D printing
    0:06:45 was that there’s a lot of geometries
    0:06:47 that are still not accessible to it.
    0:06:49 Either physically not possible to print it,
    0:06:51 that’s a lot of physical challenges,
    0:06:55 or economically it’s just not feasible, right?
    0:06:58 And that’s kind of gave way toward the thinking around
    0:07:02 how do we build a little bit more overarching automation
    0:07:04 that can do different types of processes,
    0:07:05 different types of material,
    0:07:07 and it’s not a specific to just maybe 3D printing
    0:07:09 or machining or one process.
    0:07:11 It can do different types of manufacturing processes
    0:07:13 that need it in the shop floors
    0:07:15 without having to be tool specific.
    0:07:17 And that was the genesis of Machina.
    0:07:19 And the core idea is to answer your question,
    0:07:21 like how is this gonna be possible?
    0:07:26 The core idea is, we used to have this flexibility, right?
    0:07:27 If you go a couple of centuries back,
    0:07:30 manufacturing used to be arts and crafts,
    0:07:32 people actually manufactured things.
    0:07:34 And they could one day you would go to a blacksmith
    0:07:36 and you say, okay, build me a sword,
    0:07:39 and they will start from a raw material,
    0:07:41 use their hammer and apply it incrementally
    0:07:42 to that material.
    0:07:43 And they had a creative mind,
    0:07:45 so they could apply different set up steps
    0:07:48 and be creative with it to get it to a sword.
    0:07:50 And then the day after that, you could go say,
    0:07:51 well, can you build me a sheet?
    0:07:54 And they would start from a flat sheet of metal
    0:07:55 and then they apply the same hammer,
    0:07:56 but in a different way.
    0:07:58 And it will give you a shield, right?
    0:08:00 Out of a sheet metal.
    0:08:01 So they were super flexible.
    0:08:04 The challenge was they were not scalable.
    0:08:07 So with the current manufacturing paradigm,
    0:08:10 we kind of traded that flexibility for throughput.
    0:08:14 But I think now if we can replicate what a craftsman does
    0:08:15 in an automated fashion,
    0:08:18 and then we can maybe get the best of both worlds.
    0:08:20 We can get the flexibility that a craftsman had,
    0:08:21 but also scale it.
    0:08:23 And that’s basically what we’re doing at Machina.
    0:08:25 We’re building what we call a robo craftsman.
    0:08:27 And the core components is,
    0:08:29 can you keep the replicate the dexterity of the craftsman,
    0:08:31 which is the robotic comes into play?
    0:08:32 But more importantly,
    0:08:35 can you replicate what happens in the mind of the craftsman?
    0:08:38 You know, how do they come up with a set of procedures
    0:08:40 and steps using the similar, very simple tools
    0:08:43 to get to a part that is accurate?
    0:08:44 And then because it’s robotic system,
    0:08:45 you can easily scale it.
    0:08:48 – So I want to ask you to dig into that a little bit.
    0:08:50 But before I do,
    0:08:54 are there particular industries,
    0:08:57 size and types of projects or other parameters
    0:09:00 that define what Machina is working on,
    0:09:03 and then also what you are not working on,
    0:09:07 is it specific to, you know, are you building,
    0:09:10 are you building robotic blacksmiths for that matter?
    0:09:12 Or are they aviation industry, different industries?
    0:09:14 What’s sort of that end of the approach?
    0:09:17 And then maybe we can dig into some of the technical bits
    0:09:18 about how you’re doing it.
    0:09:22 – Yes, so the robotic craftsman cell that we build,
    0:09:24 this is a robotic system, we call it a cell.
    0:09:26 It’s actually industry-agnostic, right?
    0:09:28 You could build parts for aerospace,
    0:09:30 you can build part for automotive,
    0:09:31 and that’s the beauty of it, right?
    0:09:33 Like in the morning, you can do automotive parts,
    0:09:36 in the afternoon, you can do airspace parts.
    0:09:38 But in order to build a good business,
    0:09:39 we always want to start from somewhere
    0:09:41 that we are a very good fit,
    0:09:43 based on the current economic conditions.
    0:09:45 Develop it there,
    0:09:47 and then from there expand to other verticals.
    0:09:49 So the focus that we have today
    0:09:50 is mostly on airspace defense.
    0:09:52 We have some focus on automotive,
    0:09:56 but dominantly our business comes from airspace defense.
    0:09:58 And the reason for that is that, you know,
    0:10:00 the traditional factories are very good
    0:10:03 at producing the same thing over and over again.
    0:10:06 So we don’t want to necessarily start competing with them
    0:10:07 on making the same thing over and over again.
    0:10:11 We want to compete in areas where design is constantly changing,
    0:10:13 the parts are constantly changing,
    0:10:16 and what we call is a very high-mix environment.
    0:10:19 And that’s where airspace and defense is a very good fit.
    0:10:22 And they’re also traditionally good adopters of new technology,
    0:10:23 compared to maybe some of the other industries
    0:10:27 that have lower margins and cannot necessarily risk,
    0:10:29 you know, kind of testing a new technology.
    0:10:30 So airspace defense is where we start,
    0:10:32 because of kind of like market factors
    0:10:35 and voter market kind of considerations.
    0:10:37 But the goal is that this can apply
    0:10:39 to any vertical in the future.
    0:10:41 Yeah, it makes sense.
    0:10:43 I should ask, when was Muck and Affounded?
    0:10:46 Yeah, so we started the company in 2019.
    0:10:47 Okay.
    0:10:49 So the company is right now like four years old,
    0:10:50 four or five years old.
    0:10:53 But yeah, but the team comes from, you know,
    0:10:55 manufacturing background, you know,
    0:10:57 my co-founder Bob is a material scientist.
    0:11:00 He comes from aerospace and automotive in the past.
    0:11:02 Lots of space, six alumni,
    0:11:04 some of the folks from relativity space.
    0:11:07 So a lot of people who have been into agile manufacturing space
    0:11:08 and robotic space.
    0:11:11 And then we started in 2019.
    0:11:15 And then I think we have our first manufacturing cell in 2020.
    0:11:16 Cool.
    0:11:18 And in the interest of full transparency,
    0:11:20 this would be a good time for me to mention
    0:11:24 that NVIDIA is actually an investor in Muck and Affounded Labs.
    0:11:26 All right, so let’s dig into it a little bit
    0:11:30 from whatever angle and to whatever extent you’d like to add.
    0:11:32 I guess I have sort of two questions
    0:11:34 and you can pick the one that makes more sense.
    0:11:38 One is kind of how does it work in terms of
    0:11:40 what does your manufacturing floor look like
    0:11:42 or the best that you can describe that over the radio?
    0:11:44 So to speak and kind of, you know,
    0:11:47 how have you built the cells to be able to accommodate
    0:11:51 this mix of different manufacturing tasks?
    0:11:54 And then the other question I have, of course,
    0:11:59 is how has and is AI playing a role in informing, you know,
    0:12:03 and I’m hoping that we get into sort of the creative mind
    0:12:06 of the, what was the term, robo-manufacturer?
    0:12:08 No, robo-craftsman.
    0:12:09 Robo-craftsman, thank you.
    0:12:13 I knew it was a little warmer of a term than manufacturer.
    0:12:16 So where should we start to dig in a little more here?
    0:12:18 Yeah, no, I think starting from the kind of the layout
    0:12:20 of the floor has a good start.
    0:12:21 It’s because it is different
    0:12:23 than traditional manufacturing floors, right?
    0:12:25 We have a facility, we have a couple of facilities here
    0:12:29 in LA around 110,000 square feet now.
    0:12:31 And it’s very different than, you know,
    0:12:33 what you would see, for example,
    0:12:35 if you go to a automotive factory.
    0:12:37 The traditional paradigm of manufacturing, like I said,
    0:12:39 is very much has been focused
    0:12:41 on making the same thing over and over again.
    0:12:44 And the way we’ve been able to achieve that
    0:12:47 is definitely because of a invention
    0:12:50 that actually we did in America done by Henry Ford
    0:12:53 in early 19th century called assembly line, right?
    0:12:56 Where you have a product moving in a linear fashion
    0:13:00 in the factory and each station or along its way,
    0:13:02 people are just basically installing different things
    0:13:05 or doing different manufacturing operations on it.
    0:13:06 So if you go into a, you know,
    0:13:08 high volume factories of today,
    0:13:11 you will kind of see that assembly line flow
    0:13:13 where, you know, material comes in from one end,
    0:13:15 they start getting put together on the assembly line.
    0:13:18 Each operation is getting done at each station.
    0:13:20 And again, you get a, you get a call
    0:13:22 or whatever product that you get coming out of it.
    0:13:25 And for my sort of non-expert point of view,
    0:13:26 the thing that I always think of is
    0:13:29 each station is doing exactly the same thing
    0:13:30 or more or less exactly the same thing
    0:13:32 over and over and over again.
    0:13:33 – Yes, and that’s how we’ve been,
    0:13:36 we have been achieved to get very high throughput.
    0:13:38 But then the Achilles heel is that, you know,
    0:13:39 you cannot change.
    0:13:41 The moment you want to change it,
    0:13:44 basically you have to overhaul this whole factory, right?
    0:13:46 So that’s why it becomes very expensive.
    0:13:48 So our paradigm kind of like flips that a little bit
    0:13:49 on its head.
    0:13:52 And it says, okay, let’s have these, what we call cells.
    0:13:55 And these cells are enabled by robots
    0:13:57 and each robot can be programmed
    0:13:59 to do a different operation.
    0:14:02 And now we’re decoupling logistics also
    0:14:03 from the manufacturing.
    0:14:06 Meaning that now you have a facility
    0:14:08 that actually more looks like a data center
    0:14:11 for folks who are familiar with the compute world,
    0:14:13 where you have basically the same way in a data center,
    0:14:15 you have a whole bunch of computers
    0:14:16 and you can program these computers
    0:14:18 and these computers are connected together.
    0:14:19 You can program these computers
    0:14:22 to do different types of operations through software.
    0:14:24 That’s the same kind of paradigm we have in our shop floor.
    0:14:27 There is a ray of manufacturing cells.
    0:14:29 These cells are robotic cells
    0:14:31 and can be configured to do different operations
    0:14:33 and different parts.
    0:14:35 And then through a kind of centralized system,
    0:14:38 you can basically program what each cell does.
    0:14:41 Maybe one cell forms a hood for a car, another cell,
    0:14:44 maybe it’s welding it, another cell is trimming it.
    0:14:46 And the logistics is completely decoupled
    0:14:48 because we want to get this flexibility.
    0:14:50 Now there’s a whole bunch of other benefits as well, right?
    0:14:52 Like not only you get flexibility
    0:14:54 but you get kind of rolling updates.
    0:14:55 You can kind of like, you can bring down a cell
    0:14:58 and the whole manufacturing doesn’t need to get overhauled
    0:15:00 as it is in the assembly line, right?
    0:15:01 So there’s a whole bunch of other benefits
    0:15:02 that come into play.
    0:15:05 But yeah, our shop floors look very different
    0:15:07 than what you would see in a, let’s say a car factory today.
    0:15:09 When you’re talking about the robotic cells,
    0:15:12 the robots themselves, are they arms?
    0:15:15 Do they have different physical forms,
    0:15:17 depending on the task and how much does
    0:15:21 or doesn’t the physical form of the robot itself
    0:15:23 to call it, the RoboCraftsman.
    0:15:25 How much does that dictate?
    0:15:27 I’m going back to your space example
    0:15:29 about the diameter of the rock being fixed
    0:15:32 and imagining, well, there’s got to be some constraint
    0:15:37 on what one of these cells just physically can’t do.
    0:15:38 Yeah, yeah.
    0:15:41 So this has actually have been kind of like
    0:15:44 very big subject to debate within our company
    0:15:44 since early days, right?
    0:15:48 Like how do we make these cells as agile
    0:15:50 and as flexible as possible?
    0:15:53 So there’s I think two considerations, right?
    0:15:54 One is the size of the cell
    0:15:56 and the other one is what are the components?
    0:15:58 If you look at the industry today,
    0:16:00 there are multiple options we have had.
    0:16:03 Like, you know, we can go very simple automation.
    0:16:06 A lot of people can think of like gantries and systems
    0:16:09 that are kind of basically XYZ movements
    0:16:09 that you can have in them.
    0:16:10 So that’s very simple system.
    0:16:11 And then you can go all the way
    0:16:15 to a very complicated robotic system like humanoid, right?
    0:16:17 Where they have a lot of degrees of freedom
    0:16:18 and they can do a lot of things.
    0:16:21 But then the downside is as a system gets more complicated,
    0:16:23 it’s harder to manufacture.
    0:16:25 The accuracies maybe is not as good.
    0:16:28 That the amount of forces they can apply is not as good.
    0:16:31 Versus on one end, you have very simple system,
    0:16:34 can do a lot of accurate stuff on the gantry side,
    0:16:35 can apply a lot of force.
    0:16:36 And then you have on the other end,
    0:16:37 very complex system, let’s say humanoid,
    0:16:40 which companies like figure and others are trying to do.
    0:16:42 And you get a lot of flexibility,
    0:16:44 but maybe you lose precision, accuracy,
    0:16:46 and the amount of load they can apply.
    0:16:49 So we kind of chose to be somewhere in the middle.
    0:16:50 So we use robotic arms.
    0:16:52 We use industrial robotic arms
    0:16:53 to kind of have the benefit of both worlds.
    0:16:57 Get enough flexibility in our manufacturing selves,
    0:16:59 because we can kind of replicate almost the same dexterity
    0:17:03 as a human, I would say Saxex or seven axis robotic arms.
    0:17:04 But then at the same time,
    0:17:07 you’re using a kind of slightly commoditized system.
    0:17:09 So we don’t have to build a wheel from scratch.
    0:17:13 So robotic arms already have existed for a few decades.
    0:17:16 There are multiple vendors that are offering the same thing.
    0:17:18 So it’s more commoditized,
    0:17:20 but then it still gets us that benefit of,
    0:17:21 so it has flexibility,
    0:17:22 but it has the benefit of applying a lot of force.
    0:17:24 Still, it’s still very accurate.
    0:17:27 And we can kind of get rid of a lot of its limitations
    0:17:28 through our software stack.
    0:17:30 So use robotic arms,
    0:17:33 but ideally you might want to have multiple configuration
    0:17:35 of these robotic arms, right?
    0:17:36 The same way in a data center,
    0:17:40 you have one system that has X amount of RAM
    0:17:41 and this type of a CPU,
    0:17:43 another system might have much larger RAM,
    0:17:45 another system might have a GPU,
    0:17:46 but hopefully we can only have like
    0:17:49 few five, six different configurations.
    0:17:50 And then what we do is five, six different configurations,
    0:17:53 we basically can do all kinds of different types of parts.
    0:17:55 Right now, for example, we have very strong robots
    0:17:57 that can do very thicker material.
    0:18:00 We have thinner robots that are slightly more accurate,
    0:18:02 weaker robots that are slightly more accurate,
    0:18:04 but then they cannot apply as much force.
    0:18:06 So we have few different configurations,
    0:18:08 but together pretty much can get,
    0:18:10 do different types of geometries,
    0:18:14 maybe 80, 90% of geometries out there is accessible to us
    0:18:16 with different configuration, different size.
    0:18:17 – I’m speaking with Ed Mer.
    0:18:21 Ed is the co-founder and CEO of Machina Labs,
    0:18:24 who as Ed has been detailing,
    0:18:27 is revolutionizing the manufacturing industry
    0:18:30 by combining robotics with AI and a new approach
    0:18:35 to what the manufacturing floor itself looks like,
    0:18:37 how it can be reconfigured, what it’s all about.
    0:18:41 So let’s talk now about how you get the robotic cells
    0:18:43 to do what you want them to do.
    0:18:44 And maybe even past that,
    0:18:48 I’m curious when you talk about the creativity aspect
    0:18:52 and combining that sort of old world human world,
    0:18:55 there’s still artisans out there making things,
    0:18:57 putting that into this robotic process,
    0:18:58 what that’s all about.
    0:19:00 – Yeah, so ideally what is the system that you want?
    0:19:02 And you want a system that actually works
    0:19:04 like a robotic craftsman or a true craftsman,
    0:19:06 right, you can go to the system and say,
    0:19:08 “Hey, here’s a design.
    0:19:09 “Here’s a design of a part that I want.”
    0:19:12 Basically input design intent.
    0:19:13 And then the system can come up
    0:19:15 with how it will direct its kinematics,
    0:19:19 basically robotic arms, to do the right set of operation,
    0:19:22 pick up the right tool, apply the right way to the material,
    0:19:24 and then get you the right part into it, right?
    0:19:26 So that’s the ideal solution that we want.
    0:19:30 Today, those steps are done with a lot of experts, right?
    0:19:33 So there’s a expert designer, get that look at the cat,
    0:19:36 maybe it modifies the cat or the design intent
    0:19:38 to get to something that’s manufacturable.
    0:19:40 It goes back and forth maybe with the person
    0:19:42 who brought the design intent to figure out,
    0:19:43 okay, what are compromises you can make
    0:19:45 to make this manufacturable?
    0:19:48 Then there’s another expert that turns the finalized,
    0:19:51 the negotiated design into a set of steps
    0:19:53 that the robotic system can do,
    0:19:57 and then programs that robot to do those steps.
    0:20:01 And at each point in time, if the robot is not doing well,
    0:20:02 maybe we do a QC and say, okay,
    0:20:03 the robot is not doing the right thing,
    0:20:04 we have to iterate again,
    0:20:07 and then kind of update the robotic instructions,
    0:20:10 and then see if we can get to a better part.
    0:20:12 And then in the end, you have to do some kind of a QC.
    0:20:14 The part comes out, somebody looks at it and say,
    0:20:16 okay, does it actually toward the specs that we want?
    0:20:20 Now, these are the steps that are currently done by experts.
    0:20:22 So what we want is kind of combine all of those
    0:20:24 into an intelligent system.
    0:20:27 Now, what you see has been done with chat GPs and others,
    0:20:29 you would say, oh, okay, yeah, it seems like
    0:20:31 we are very close to be able to do these things.
    0:20:32 Like now we have these robotic system
    0:20:36 that I can ask it to do something that’s very complex,
    0:20:37 and it will give you an answer.
    0:20:40 There are news now that latest iteration of training
    0:20:44 of these models is as smart as maybe a PhD, right?
    0:20:45 So we have these very expert systems
    0:20:49 that can reason through with AI through these steps.
    0:20:52 The main challenge we are facing is that,
    0:20:54 these systems that you see like chat GPD and others
    0:20:56 have been trained on the data
    0:20:57 that has been publicly available, right?
    0:20:59 You can use internet to train them.
    0:21:02 With manufacturing data, the data has not
    0:21:04 been traditionally available.
    0:21:07 So we have that extra challenge is that, okay, yes,
    0:21:09 seems like the models are capable enough
    0:21:12 to be able to decipher and turn the design into steps
    0:21:14 that the robot needs to do,
    0:21:16 but we don’t have enough data or available data
    0:21:18 to train things on them.
    0:21:21 So the extra additional complexity for us was,
    0:21:24 we need to build systems that can generate data
    0:21:25 and incrementally get better.
    0:21:26 So we need to start with systems
    0:21:30 that use heuristic methods first, right?
    0:21:33 Non-model is necessarily to generate a lot of data.
    0:21:36 And then we use the data to kind of build models
    0:21:40 and incrementally improve its capabilities over time
    0:21:42 to get to a point where we have that robotic craftsman.
    0:21:44 You can literally put the design intent,
    0:21:46 input the design intent and comes out the other
    0:21:49 and actual part that these robotic systems are building.
    0:21:53 – So do you generate that data in a simulation?
    0:21:55 How did you go about building the data sets?
    0:21:56 – Very good question.
    0:21:58 So were we started?
    0:22:01 – It’s always the question and no matter who, right?
    0:22:02 Talking to you, talking to somebody
    0:22:05 about wildlife conservation and tracking animals,
    0:22:06 it’s always about the data.
    0:22:07 – Yes, yes.
    0:22:09 So here’s the challenge, right?
    0:22:11 Simulation is a good way to create the synthetic data, right?
    0:22:13 The challenge is even our simulations today
    0:22:16 are not fast enough for our specific processes.
    0:22:18 So if you look at, for example,
    0:22:20 in some of the other robotic tasks
    0:22:23 where you do manipulation, like you move stuff around,
    0:22:26 then you have simulations that are fast enough, right?
    0:22:28 They call it in these robotic gyms, right?
    0:22:29 Where the robots basically can do
    0:22:31 a whole bunch of different things
    0:22:34 and kind of explore the space and kind of get training data.
    0:22:36 Because the physics, the kinematic physics
    0:22:39 is actually pretty simple, right, to simulate.
    0:22:40 The challenge with our process
    0:22:44 is that we are simulating a physical phenomenon
    0:22:46 that’s much more complicated.
    0:22:48 You know, when we are forming these sheet of metals,
    0:22:49 we have to simulate the deformation.
    0:22:51 So plastic and elastic deformation,
    0:22:53 we need to simulate friction.
    0:22:56 We need to simulate material cohesion.
    0:22:57 There’s a whole bunch of things.
    0:22:58 We need to simulate tear in there.
    0:22:59 – Yeah, yeah.
    0:23:03 – So when we initially, we started to do some of this work,
    0:23:05 for example, like try to simulate it.
    0:23:06 We realized that actually,
    0:23:09 if we properly want to simulate the environment
    0:23:13 and our processes, it’s actually very expensive
    0:23:15 and it takes a long time.
    0:23:17 A part that would take for us to form with the robots
    0:23:19 for 15 minutes at the time
    0:23:22 on the differential equation-based simulations,
    0:23:24 we call it finite element analysis models.
    0:23:29 It would take one week on 27 cores of a CPU machine, right?
    0:23:30 So it’s like, okay, it’s cheaper maybe
    0:23:32 to just run it in real life.
    0:23:34 – So just capture the data, yeah.
    0:23:35 – Capture the data.
    0:23:37 But we took both approach.
    0:23:38 So we said, okay, let’s build systems
    0:23:40 that we can do a lot of trials in real life
    0:23:41 and make sure we build them in a way
    0:23:43 that we can capture the data,
    0:23:46 a state of this process every four milliseconds.
    0:23:47 That’s what we’re doing.
    0:23:49 So we captured terabytes and terabytes of data
    0:23:51 with that over the past four, five years.
    0:23:53 We’ve built thousands and thousands of parts.
    0:23:56 But also we are working on making faster simulations
    0:23:58 using GPUs, right?
    0:24:00 Can we actually expedite the speed of simulation
    0:24:03 so that we can also augment it with data
    0:24:05 that comes from a synthetic world?
    0:24:06 – So we have to do both.
    0:24:08 But yes, it has been a huge challenge.
    0:24:10 So these are the two challenges that are kind of ahead of us
    0:24:12 that maybe other companies,
    0:24:15 like the ones that are working on natural language models,
    0:24:18 I don’t have because internet is giving them,
    0:24:19 user-generated data for years.
    0:24:23 – Yeah, no, you mentioned QC and it’s stuck in my head.
    0:24:28 I wouldn’t want the robotic QC craftsman, so to speak,
    0:24:31 hallucinating based off of at least the stuff
    0:24:32 I read on the internet.
    0:24:32 Let’s put it that way.
    0:24:37 So you mentioned the data and the training
    0:24:40 and then the speed of everything
    0:24:41 being kind of two big hurdles
    0:24:44 that you’ve been working to overcome.
    0:24:47 – There are other either hurdles you’ve overcome
    0:24:51 or perhaps just moments along the way
    0:24:54 that were surprising or just kind of stand out
    0:24:57 as kind of like these big milestones
    0:25:01 in the development of everything Muckin’ The Labs is doing.
    0:25:03 And then sort of the follow-up to that
    0:25:06 is I guess beyond the training and the speed,
    0:25:07 which obviously huge.
    0:25:10 If you look ahead to the next couple of years
    0:25:13 or whatever the best timeframe is,
    0:25:14 maybe what are some of the things
    0:25:18 that you and other robotics and manufacturing companies
    0:25:20 are trying to clear to get to sort of the next stage
    0:25:22 of your evolution?
    0:25:24 – Yeah, so there’s an interesting challenge
    0:25:26 when you want to generate your data,
    0:25:27 which has been probably the biggest challenge
    0:25:28 for a robotic company.
    0:25:31 I remember there was a podcast from Ilya
    0:25:35 talking about why open AI dropped a robotic arm
    0:25:37 back in the day, right?
    0:25:39 And the main challenge is that, yes, you say,
    0:25:41 “Okay, I have to generate my own data.”
    0:25:43 But you’re generating data with a very complex
    0:25:44 physical system.
    0:25:49 These robots can break and they might need maintenance.
    0:25:52 So you almost have to figure out a way
    0:25:54 to be very good at operation, right?
    0:25:57 You need to be able to run a large fleet of robots
    0:26:00 very smoothly to generate data.
    0:26:01 And that actually has been one of the main challenges,
    0:26:04 like I said, even open AI had at the time,
    0:26:06 where we’re like, I don’t know if we are in it,
    0:26:08 like we don’t have the right expertise
    0:26:10 to create this operational rigor
    0:26:12 to operate these robotic cells.
    0:26:16 Now, they were funded by billions, actually,
    0:26:19 of kind of they could go and dedicate themselves to do this.
    0:26:22 We also don’t have the luxury of that.
    0:26:24 We need to create, we’re venture-funded business,
    0:26:26 obviously we have a good amount of funding,
    0:26:28 but we need to also generate results for our customers.
    0:26:31 So I think one of the biggest challenges for us was,
    0:26:34 in order to enable this AI-powered world,
    0:26:37 we need to actually become a very good operators
    0:26:39 and create value for the customers fast.
    0:26:41 So a lot of challenges that we had in early days
    0:26:45 was like, how do we create a robotic facility
    0:26:48 that can work around the clock and generate data
    0:26:50 and kind of resolving a lot of those challenges?
    0:26:53 So that’s been one big area of work for us.
    0:26:55 And the other piece, I think,
    0:26:58 kind of like maybe a little bit away from technical work
    0:27:03 is because you are operationally heavy company,
    0:27:05 you also need to deal with a lot of people.
    0:27:06 One challenge that people–
    0:27:07 How big is Makada?
    0:27:08 How many people?
    0:27:13 Mostly now 70 people, but across many skill sets, right?
    0:27:15 We have technicians all the way to robotic engineers,
    0:27:19 to AI engineers, to software engineers, material scientists.
    0:27:21 And if you traditionally want to build
    0:27:25 such a multidisciplinary team, you have to balloon up, right?
    0:27:27 You have to create all these different expertise
    0:27:29 that kind of work together.
    0:27:30 So we have to also figure out a way,
    0:27:31 how do we not balloon up
    0:27:36 while we operate a huge operational facility
    0:27:39 and then generate data while we also maintain
    0:27:42 a rigor and expertise in the fields like AI and others.
    0:27:44 So I think the biggest challenge for us
    0:27:46 has been kind of navigating those waters.
    0:27:51 As a technologist, you always underestimate
    0:27:54 how important the people factor is.
    0:27:57 And I think that’s something that I kind of,
    0:27:58 I kind of have learned over the time
    0:28:00 that it’s probably the most important factor, right?
    0:28:03 A lot of technological developments are already there,
    0:28:05 but having a team that is excited
    0:28:07 and can operate efficiently
    0:28:09 has been one of the biggest challenges here.
    0:28:11 Absolutely, yeah.
    0:28:13 So to put you on the spot here a little bit,
    0:28:17 how do you see manufacturing changing,
    0:28:19 because of AI and the things that we’re talking about?
    0:28:23 And as we go forward, I mean, five years, 10 years,
    0:28:25 20 years down the line,
    0:28:28 is all the manufacturing gonna be,
    0:28:32 look very different and have this sort of flexible,
    0:28:35 scalable, reconfigurable type of design
    0:28:39 that Makana has and is continuing to build.
    0:28:43 Are we gonna have to sort of knock down or gut
    0:28:45 the existing manufacturing facilities
    0:28:47 and rebuild them with this in mind?
    0:28:49 And then this may be out of your wheelhouse,
    0:28:51 so please feel free to demure on this,
    0:28:56 but do you see these types of things coming to,
    0:28:58 sort of consumer and household robots going forward
    0:29:03 with every a time where I’ll be able to have my home robot,
    0:29:06 build me a new desk for that, whatever it may be.
    0:29:09 So it’s an interesting time right now, right?
    0:29:12 We are in a situation where,
    0:29:15 I think there’s a lot of tailwinds
    0:29:16 for changing manufacturing.
    0:29:20 As a country, we develop a lot of previous manufacturing
    0:29:25 methods in late 19th century and 20th century.
    0:29:28 As we develop these technologies,
    0:29:30 but then over time,
    0:29:34 as the cost of labor went high in United States,
    0:29:37 we started kinda moving these things into other countries.
    0:29:40 So a lot of manufacturing moved out of United States,
    0:29:42 but now we’re at the point where we’re realizing that,
    0:29:44 oh maybe that was not such a great idea,
    0:29:47 because now we have dependency
    0:29:50 on these large centers of manufacturing
    0:29:53 that might not necessarily align with our values, right?
    0:29:55 There might be a little bit more authoritarian
    0:29:58 or they might just not wanna have the same interest
    0:29:59 as we do.
    0:30:00 So now we’re thinking about,
    0:30:04 okay, we need to maybe be a little bit more self-sufficient.
    0:30:07 The challenge is that the core technologies
    0:30:09 that are underlying these manufacturing techniques
    0:30:11 does not allow them to come back to the United States.
    0:30:13 We talked about a factory right now
    0:30:15 needs to be built for every part.
    0:30:17 So what does that mean?
    0:30:19 That means that that factory needs to be building
    0:30:21 a lot of the same part for the rest of the world
    0:30:23 before it’s economical.
    0:30:25 So it’s very hard to replicate that same factory
    0:30:28 in a smaller version and still be competitive
    0:30:29 in United States.
    0:30:31 The nature of the technology
    0:30:33 lends itself to centralization.
    0:30:34 So we’re kind of finding the nature
    0:30:35 of the technology at this point.
    0:30:36 It’s not just like with gusto,
    0:30:38 we can bring manufacturing back in the United States.
    0:30:40 Maybe some of it we can,
    0:30:43 but for the most part, unless the technology doesn’t change,
    0:30:46 I don’t think we can be able to bring manufacturing back.
    0:30:48 But that being said, like I said, there’s a lot of tailwind.
    0:30:49 People wanna bring it back
    0:30:53 because now it’s becoming a existential threat.
    0:30:55 So I think because of these tailwinds,
    0:30:57 there’s gonna be a lot of changes
    0:30:59 that gonna happen in manufacturing.
    0:31:01 And automation is one of those biggest things
    0:31:03 that’s gonna change, right?
    0:31:05 In order for manufacturing to come back to the United States,
    0:31:09 it’s not gonna look like what we had in ’60s and ’50s.
    0:31:11 It’s gonna be the new paradigm of manufacturing
    0:31:12 that requires less labor
    0:31:14 or maybe requires a more sophisticated labor
    0:31:17 with higher productivity than we had before.
    0:31:21 So automation, flexible manufacturing,
    0:31:23 easy to reconfigure shop floors
    0:31:26 are gonna be big part of bringing manufacturing
    0:31:27 back to the United States.
    0:31:30 You look at some of the towns in Midwest
    0:31:32 that used to be manufacturing towns,
    0:31:34 you go there as it goes downtown.
    0:31:36 The whole economy around that factory died
    0:31:38 because that factory,
    0:31:39 the product it was making became obsolete
    0:31:43 and it was not flexible enough to change, it died.
    0:31:45 So this will change.
    0:31:46 Now the question is how fast?
    0:31:48 That’s a harder one to predict
    0:31:52 because I think there’s a lot of geopolitical parameters there.
    0:31:54 Interesting times, as you said.
    0:31:55 You enter a conflict with China,
    0:31:57 that will change very fast.
    0:32:01 If you don’t, maybe it’s gonna be a little bit more paced,
    0:32:02 but yes, it’ll definitely change.
    0:32:04 Timeline is tougher to predict.
    0:32:08 Now, will we have these systems in our homes?
    0:32:11 I think that timeline is a slightly higher, longer.
    0:32:13 There are people predicting at some point,
    0:32:14 everybody’s gonna have a human or robots
    0:32:16 that’s gonna help them with things.
    0:32:18 I think the lower hanging fruit is getting robots
    0:32:21 to use AI and automation in factories
    0:32:23 because we already have figured out
    0:32:24 a lot of other stuff around these robots
    0:32:26 and we have a very good supply chain around them.
    0:32:28 But I do see a future where we’re gonna have
    0:32:32 human or robots in every home, everybody will own one.
    0:32:33 Maybe a little bit longer.
    0:32:35 It’s harder to predict anything beyond 10 years,
    0:32:39 but I would put that in more 20, 25 years out there.
    0:32:40 Fair enough.
    0:32:45 For folks listening who might be aspiring entrepreneurs
    0:32:48 in the manufacturing or robotics automation space
    0:32:52 or maybe studying and interested in robotics and automation,
    0:32:55 but not quite sure what path to take,
    0:32:57 what advice would you give either
    0:32:58 from sort of the technical side
    0:33:02 or you obviously have an entrepreneurial background
    0:33:03 and spirit yourself here.
    0:33:05 So like the point you made earlier
    0:33:07 about the people being at the core of it
    0:33:09 and sometimes that being the harder thing
    0:33:11 for technologists is a salient one,
    0:33:14 but what would you give us kind of advice?
    0:33:18 I think one thing that has changed since I came out of school
    0:33:22 is we need way more multidisciplinary people.
    0:33:25 When I came from a family that valued education a lot,
    0:33:29 my mom was a teacher, my dad has his PhD,
    0:33:32 so it was very focused on like go to school,
    0:33:35 get expert in one area and that’s gonna be your thing.
    0:33:38 I think that paradigm slightly is breaking
    0:33:42 where we need people that can connect different components
    0:33:44 more, we always gonna need experts,
    0:33:46 but we also need more people
    0:33:49 that maybe they know mechanical engineering,
    0:33:51 but they also know software, they also know robotic,
    0:33:53 they also maybe know a little bit of material science
    0:33:56 because we have these kind of breakthrough technologies
    0:33:59 like AI and we need people who can connect the dots
    0:34:02 between multiple disciplines and build something new.
    0:34:04 So that is becoming more and more important.
    0:34:06 So I would suggest for people
    0:34:09 who are going to school now, get exposure,
    0:34:11 maybe become an expert in one area,
    0:34:12 but get as much as exposure as you can
    0:34:15 in other fields of science and engineering,
    0:34:17 that makes you a much more attractive candidate
    0:34:20 for the companies of next generation companies, right?
    0:34:23 So that would be one main area of say advice I have
    0:34:25 for people who wanna be in the technical field.
    0:34:27 On the entrepreneurial side,
    0:34:32 I think it is becoming more and more important
    0:34:34 to build mission-driven companies.
    0:34:37 There used to be a time where there was a lot of
    0:34:39 kind of incremental improvement companies
    0:34:41 that you can build, that the system’s already there.
    0:34:43 You just wanna make some incremental improvement.
    0:34:46 I think with AI, now the shifts
    0:34:48 are gonna be more fundamental.
    0:34:49 There’s gonna be always incremental thing,
    0:34:51 but incremental things are just easier to build.
    0:34:53 So I don’t know if companies are necessarily
    0:34:55 gonna be well equipped, startups well equipped to do that.
    0:34:57 As a startup, you wanna kind of go after
    0:34:58 something more fundamental.
    0:35:01 And for that, you need more kind of mission-driven companies.
    0:35:04 So find something that you’re really excited about.
    0:35:07 You become a big believer that that change needs to happen
    0:35:08 and go after that.
    0:35:10 And I think there gonna be less companies
    0:35:12 that are just gonna optimize this and that
    0:35:14 because it’s just easier to do those things
    0:35:15 for the companies themselves,
    0:35:16 for the larger companies themselves.
    0:35:18 – Sounds like sage advice to me.
    0:35:21 Edmere, for folks listening who want to learn more
    0:35:24 about Machina Labs, your latest innovations,
    0:35:28 what you’re up to, perhaps more technical-oriented
    0:35:30 research papers and blogs and stuff like that,
    0:35:33 where can listeners go online to learn more?
    0:35:34 – Yeah, obviously our website,
    0:35:37 MachinaLabs.ai is a good place,
    0:35:39 but also we’re very active on social media.
    0:35:42 LinkedIn, I think we put both even technical
    0:35:45 and commercial milestones on it very often.
    0:35:47 Twitter as well, and Instagram.
    0:35:49 Yeah, so I think follow us on those channels.
    0:35:53 I actually, very openly, one of the main kind of philosophies
    0:35:56 we have on Machina is like build out into public.
    0:36:00 So we actually share even our technical challenges online
    0:36:03 and people respond to it and we are very open.
    0:36:05 So if you follow, I think even for the technical folks,
    0:36:07 you will see a lot of interesting kind of content
    0:36:09 that we put out there that they can contribute.
    0:36:10 – Fantastic.
    0:36:13 Ed, thank you so much for taking the time to chat with us.
    0:36:15 I could binge at your ear all day
    0:36:16 or ask you questions all day.
    0:36:17 I should say I’m listening to you talk about it.
    0:36:19 It’s fascinating, fascinating stuff.
    0:36:21 And as you said, interesting times we live in
    0:36:24 and it’s probably only gonna evolve faster than it has been.
    0:36:27 So I look forward to keeping an eye
    0:36:29 on what Machina Labs is up to
    0:36:31 and maybe we can talk again down the line.
    0:36:32 – I would love that, thanks though.
    0:36:33 – Thank you.
    0:36:35 (upbeat music)
    0:36:38 (upbeat music)
    0:36:41 (upbeat music)
    0:36:43 (upbeat music)
    0:36:46 (upbeat music)
    0:36:48 (upbeat music)
    0:36:51 (upbeat music)
    0:36:54 (upbeat music)
    0:36:56 (upbeat music)
    0:36:59 (upbeat music)
    0:37:01 (upbeat music)
    0:37:04 (upbeat music)
    0:37:06 (upbeat music)
    0:37:09 (upbeat music)
    0:37:12 (upbeat music)
    0:37:15 (upbeat music)
    0:37:18 (upbeat music)
    0:37:20 (upbeat music)
    0:37:23 (upbeat music)
    0:37:33 [BLANK_AUDIO]

    Edward Mehr works where AI meets the anvil. The company he cofounded, Machina Labs, blends the latest advancements in robotics and AI to form metal into countless shapes for use in defense, aerospace, and more. The company’s applications accelerate design and innovation, enabling rapid iteration and production in days instead of the months required by conventional processes. NVIDIA AI Podcast host Noah Kravitz speaks with Mehr, CEO of Machina Labs, on how the company uses AI to develop the first-ever robotic blacksmith. Its Robotic Craftsman platform integrates seven-axis robots that can shape, scan, trim and drill a wide range of materials — all capabilities made possible through AI.

  • Snowflake’s Baris Gultekin on Unlocking the Value of Data With Large Language Models – Ep. 231

    AI transcript
    0:00:10 [MUSIC]
    0:00:13 >> Hello, and welcome to the Nvidia AI podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:18 Today, I’m joined by Barasch Kultekin,
    0:00:20 the head of AI at Snowflake.
    0:00:22 At Snowflake, he’s driving the development of
    0:00:24 cutting-edge AI and ML products,
    0:00:27 including Snowflake Cortex AI,
    0:00:29 and ARTIC, their new foundational model.
    0:00:33 But Barasch has also had a remarkable journey through AI himself,
    0:00:35 having co-founded Google Now and
    0:00:38 led AI efforts for Google Assistant.
    0:00:40 So we’ve got a ton to talk about background,
    0:00:42 the present at Snowflake, and of course,
    0:00:44 we’re not going to let Barasch off the hook without asking
    0:00:49 him to opine about the future of AI and the hot topic today.
    0:00:51 What’s going to happen to all of us in the age of AI?
    0:00:53 So I can’t wait to get into it.
    0:00:55 Barasch, thank you so much for joining
    0:00:57 the Nvidia AI podcast and welcome.
    0:00:58 Thanks a lot, Noah.
    0:01:01 I sometimes wonder if the pods we tape on a Friday,
    0:01:03 let alone today, a Friday afternoon,
    0:01:05 have a little bit looser of a feel than some of the other ones.
    0:01:08 So we’ll have to do an exit survey at the end here.
    0:01:10 But I’m super excited to talk to you.
    0:01:12 We were talking offhand.
    0:01:13 I’ve been tracking Snowflake,
    0:01:18 especially over the past few years as interest in AI and data has exploded.
    0:01:21 A good friend of mine has been with Snowflake now for a few years,
    0:01:23 so I’ve been tracking her journey as well.
    0:01:25 So I’m thrilled to have you here, Barasch,
    0:01:28 and to learn more about Snowflake and your own journey.
    0:01:32 Maybe we’ll start with Snowflake and we can get into your background as we go.
    0:01:36 Can you start by telling the audience a little bit about what Snowflake is,
    0:01:39 how long the company’s been around, what you do,
    0:01:45 and obviously your role in the burgeoning AI explosion that we’re all part of?
    0:01:48 Of course. So Snowflake is the AI data cloud.
    0:01:51 We started the journey over a decade ago
    0:01:55 focusing on how do we make data a lot more easily accessible,
    0:02:00 how do we make data processing a lot more easily accessible to companies.
    0:02:02 As the data volumes were growing,
    0:02:07 there was a big innovation of separating the storage of data from compute
    0:02:12 that allowed a massive unlock in the data space.
    0:02:17 And since then, we’ve been evolving from a provider of a data warehouse
    0:02:21 to a data cloud where that data can be easily shared.
    0:02:23 And now we’re in the third phase of that journey
    0:02:29 where you can now unlock a lot more value of that data with AI for a lot more users.
    0:02:32 Right. Maybe just to set baselines,
    0:02:36 can you quickly just talk about terms like data warehouse, data lake,
    0:02:40 the AI data cloud, as you describe Snowflake,
    0:02:44 and what those terms mean and maybe how they’ve evolved a little bit?
    0:02:48 Of course, customers have a lot of structured data.
    0:02:52 They need to have a way to bring all of that data
    0:02:57 and then manage it, govern it, be able to run analysis on it.
    0:03:02 So data warehouse allows our customers to very efficiently run
    0:03:06 massive scale analysis across large volumes of data.
    0:03:07 Right.
    0:03:11 And that’s for structured data for tables.
    0:03:17 With data lake, you expand that to bringing in more unstructured data
    0:03:19 as structured data into the mix.
    0:03:20 Right.
    0:03:24 And so there’s issues at hand of the way that the data has stored itself,
    0:03:27 obviously, the physical media it’s stored on.
    0:03:31 But then, as with all things technology and certainly all things AI,
    0:03:34 machine learning, deep learning, the software plays a role,
    0:03:36 the interconnectivity plays a role.
    0:03:39 Obviously, if it’s a data cloud, there’s all of the…
    0:03:43 Everything that happens between your cloud and my workstation,
    0:03:44 wherever in the world I am.
    0:03:46 So there’s a lot going on.
    0:03:50 So maybe we can dive in a little bit to some of the breakthroughs
    0:03:54 that Snowflake has ushered in, is currently working on.
    0:03:56 And kind of how data plays a role,
    0:04:02 what the data cloud’s role is in the modern ALML stack, if you will.
    0:04:03 So let’s talk a little better,
    0:04:09 or if you would talk a little bit about how Snowflake works with enterprises
    0:04:12 in helping them unlock their data, as you put it.
    0:04:13 Yeah.
    0:04:17 So first of all, we’re super excited about what AI is capable of,
    0:04:18 what AI is bringing.
    0:04:20 And of course, when we think about AI,
    0:04:25 there is no AI strategy without a data strategy, is what they say.
    0:04:27 AI is fueled with data.
    0:04:32 So what our customers are interested in is they are trusting Snowflake with their data.
    0:04:36 How can they now get the most out of this data
    0:04:39 by bringing AI compute right next to where the data is?
    0:04:43 So believe deeply in being able to bring compute to data
    0:04:48 versus bringing massive amounts of data out into where the compute is.
    0:04:52 So the way we work with our customers is that we built a AI platform.
    0:04:58 And with this AI platform, our customers can run natural language analysis,
    0:05:02 can build chatbots, can talk to their data in natural language.
    0:05:06 And we make this platform available, and our customers are using it
    0:05:08 to build all sorts of AI applications.
    0:05:10 And so the customer stores their data with you,
    0:05:13 but then also the training, the inference,
    0:05:17 all the compute operations are done on your side as well.
    0:05:18 That’s right, exactly.
    0:05:21 So basically, our AI platform runs fully inside Snowflake.
    0:05:25 So for Snowflake, it’s really important to govern the data,
    0:05:28 as well as AI that’s running.
    0:05:30 So we run everything inside Snowflake.
    0:05:33 So maybe you can get into a little bit of some of the product offerings.
    0:05:36 Do we want to start with Snowflake Cortex?
    0:05:40 Yeah, so Snowflake Cortex is our managed service.
    0:05:44 It is our offering where we’re running a series of large language models
    0:05:45 inside Snowflake.
    0:05:49 Our customers have super easy access to these large language models.
    0:05:55 But the way we think about AI is we want to make AI easy, efficient, and trusted.
    0:05:57 With Cortex, it is incredibly easy
    0:06:00 because AI is running right next to where the data is.
    0:06:02 So our customers don’t have to build data pipelines
    0:06:06 to manage data in multiple places and govern it.
    0:06:08 We were running it in a secure environment
    0:06:11 and we have a series of very efficient models
    0:06:13 from our own to a series of partner models.
    0:06:14 We make that available.
    0:06:17 And then Cortex also makes it very easy for our customers
    0:06:21 to build, talk to my data experiences, if you will,
    0:06:26 build chatbots, both for documents as well as for structured data.
    0:06:30 What are some of the use cases that are kind of most popular, most seen?
    0:06:35 I’ve heard people talk about, we’re recording in late June of 2024.
    0:06:41 And I’ve heard people reference 2023 as being the year that generative AI
    0:06:45 took over all the headlines and everybody was talking a lot,
    0:06:47 but not really sure what to do.
    0:06:53 And 2024 maybe being the year that businesses start to actually
    0:06:57 develop applications or use applications others have developed,
    0:07:02 but really start to do things using leveraging AI on their own data
    0:07:07 to whatever it is, improve processes, try new ways of working,
    0:07:08 all that kind of stuff.
    0:07:12 What are you seeing on your end is the things that customers
    0:07:15 and enterprise customers, other snowflake customers,
    0:07:17 are interested in doing and then maybe on the flip side,
    0:07:21 some of the things that, I don’t know, they’re concerned about
    0:07:25 or don’t quite understand or still trying to kind of wrap
    0:07:26 their collective heads around.
    0:07:28 Yeah, so I agree with you.
    0:07:32 In 2023 was, I’d say, the year of proof of concepts.
    0:07:34 Right, well said.
    0:07:38 We got their hands on AI and then started building demos.
    0:07:43 And this year we’re starting to see these turn into real production use cases.
    0:07:46 I can give a couple of examples.
    0:07:50 We’re working with global pharmaceutical company Bayer
    0:07:54 in building an experience where Bayer’s internal teams
    0:07:56 and sales organizations, marketing organizations,
    0:07:59 can ask questions of their structured data.
    0:08:03 So Bayer believes that dashboards can only take them so far.
    0:08:05 Dashboards tend to be very rigid.
    0:08:08 And then first thing that happens when you see a dashboard
    0:08:11 is you have three questions, four questions where you want to drill in
    0:08:15 and figure out why something isn’t the way you want it to be or where.
    0:08:19 So now we give that power to not only the analysts,
    0:08:21 but to business users.
    0:08:24 So business users can ask questions of their data, drill in in natural language.
    0:08:26 And that’s super powerful.
    0:08:29 We’ve been working with a series of companies and like Bayer,
    0:08:34 they’re finding it very valuable to give democratized access to that kind of data.
    0:08:35 Right.
    0:08:39 Another interesting one is we’re working with kind of Siemens.
    0:08:41 They have a large research organization.
    0:08:45 They’ve just recently built a research chatbot
    0:08:50 that has 700,000 pages of research that’s now unlocked
    0:08:52 and available for this research organization.
    0:08:52 Wow.
    0:08:57 So, you know, instead of kind of figuring out how and where to get that data
    0:09:01 to continue your research now, the team feels a lot more productive.
    0:09:04 How many tokens is 700,000 pages?
    0:09:05 It’s a lot of tokens, you know.
    0:09:06 A lot of tokens.
    0:09:08 A lot of tokens.
    0:09:12 So when you’re working, and I’m sure the answer is different depending on the customer,
    0:09:17 but when you’re working with a customer to do something like take 700,000 pages of documentation
    0:09:20 and turn it into something that, you know, the average employee,
    0:09:24 the average user can just speak to using natural language.
    0:09:30 What’s the process like in terms of what you’re doing on the technical side?
    0:09:32 Are you fine tuning the model?
    0:09:35 Are you building kind of custom rag pipelines?
    0:09:39 And again, you know, I’m sure it’s different with different use cases.
    0:09:43 But what are some of the things that Snowflake does with a customer
    0:09:48 that, you know, they couldn’t get just by sort of brute force
    0:09:51 uploading these documents to a publicly available model?
    0:09:57 So when we actually surveyed our customers, as they think about going from, you know,
    0:10:01 these demos, proof of concepts to production, usually three big things emerge.
    0:10:06 They are concerned about quality hallucinations.
    0:10:06 Sure.
    0:10:10 The second one is they’re concerned about kind of security of their data,
    0:10:13 governance of that system, and finally the cost.
    0:10:16 So those are the top three things that always emerge.
    0:10:19 And then we try to address these three concerns head on.
    0:10:23 Cortex Search is a new product offering that we are just recently releasing.
    0:10:30 And we’ve tuned Cortex Search to be the highest quality in terms of a rag solution.
    0:10:32 So we implement a custom rag solution.
    0:10:36 We have our own embedding model, and we’ve built a hybrid search engine
    0:10:38 that can provide high quality.
    0:10:42 And we will tune the system so that it knows when not to answer questions,
    0:10:44 you know, reducing hallucinations.
    0:10:44 Got it.
    0:10:49 And hybrid search meaning combines rag functionality with internet search?
    0:10:50 Or–
    0:10:50 Exactly.
    0:10:54 Our hybrid search is basically combining vector search with the traditional keyword
    0:10:55 based tech search.
    0:10:56 Right, great.
    0:10:56 Catch you.
    0:10:57 Okay, cool.
    0:11:02 On the hallucination point, any interesting learnings or insights when it comes to,
    0:11:09 I would assume it’s more sophisticated than manually just line by line telling the model,
    0:11:11 you know, don’t say this, don’t say that.
    0:11:15 But how do you sort of coax a model into hallucinating less?
    0:11:15 Oh, yeah.
    0:11:18 So first of all, there is definitely model tuning.
    0:11:20 That’s important in this.
    0:11:26 But also, we just touched on the hybrid search element.
    0:11:30 The nice thing about hybrid search is it can give you meaningful information
    0:11:35 about whether the set of documents is relevant to the question.
    0:11:36 Okay.
    0:11:42 And, you know, usually, LLMs tend to hallucinate when they are not grounded on data.
    0:11:46 So the system can know that the match to that question is low.
    0:11:49 And rather than trying to answer the question, it should just reject it.
    0:11:49 Got it.
    0:11:53 The speaking of models, there are a bunch of Snowflake products.
    0:11:54 We don’t have time to get into all of them, obviously.
    0:11:58 And I’m going to leave room to bring up things that I didn’t ask about.
    0:12:00 But I do want to ask you about Arctic.
    0:12:05 So Arctic is an LLM that Snowflake built?
    0:12:06 Or how would you describe it?
    0:12:08 Yes, Arctic is our own language model.
    0:12:10 It’s actually a family of language models.
    0:12:17 You have the Arctic LLM as well as an embedding model and a document model.
    0:12:25 So the LLM is an open source, large language model that is quite unique in its architecture
    0:12:32 by combining both a mixture of experts model with a dense architecture.
    0:12:36 We were able to have a very efficient and high quality model.
    0:12:40 So we focused on what we’re calling enterprise intelligence,
    0:12:46 being able to follow instructions, being able to do well in coding and SQL.
    0:12:51 And then we were able to achieve the highest benchmarks in these categories
    0:12:54 amongst open source models while being incredibly efficient.
    0:12:59 We trained Arctic at about one eighth the cost of similar models.
    0:13:03 And that means when we train custom models, for instance, for our customers,
    0:13:07 we can be very, very cost effective while delivering very high quality.
    0:13:10 Understood if these are trade secrets, you don’t want to divulge.
    0:13:14 But how did you figure out how to train the model so much more efficiently?
    0:13:19 So we actually really pride ourselves in our openness.
    0:13:23 We’ve released cookbooks of our not only the model weights,
    0:13:27 but also the research insights as well as our kind of data recipes.
    0:13:28 Oh, fantastic.
    0:13:28 Okay.
    0:13:29 All of those are available.
    0:13:31 And we’ve shared some of these insights.
    0:13:40 It boiled down to having some of the best researchers that have pioneered MOE models,
    0:13:43 a mixture of experts models, you know, back in the day,
    0:13:48 along with some of the VLLM original team members.
    0:13:51 And all working together to develop this architecture.
    0:13:52 Very cool.
    0:13:55 Again, for folks who might not be familiar and unfamiliar,
    0:13:57 but I don’t fully understand how it works.
    0:14:01 Someone asked, what is a mixture of experts approach?
    0:14:02 What does that mean?
    0:14:03 Was it entail?
    0:14:05 Why is it different better than other approaches?
    0:14:10 There are two major architectures that we’re seeing.
    0:14:12 One is what’s called a dense model.
    0:14:17 In the dense model, all of the parameters are active and they’re being used
    0:14:18 when you’re doing inference.
    0:14:21 So also during training, all of these parameters are active.
    0:14:26 Mixture of experts model has larger set of parameters,
    0:14:28 but only a subset of them gets used.
    0:14:32 So you have different number of experts essentially that are,
    0:14:35 you know, that are getting activated to answer one question.
    0:14:35 Right.
    0:14:37 That tends to be very efficient.
    0:14:38 Right, got it.
    0:14:40 So you can hone in on the accuracy that you’re looking for,
    0:14:44 but then also it’s more efficient because you’re turning things on and off
    0:14:46 as you need them instead of just leaving all the lights on.
    0:14:49 Yeah, it’s efficient to train as well as efficient to run inference.
    0:14:53 So it tends to be cheaper because it has lower number of active parameters.
    0:14:53 Got it.
    0:14:58 Any other specific products, innovations that Snowflake has put out
    0:15:01 during your time that you’re particularly excited about?
    0:15:05 I am excited about our Cortex Analyst product that we just recently announced.
    0:15:15 As I was alluding to, there’s a lot of data gem that is locked in in very large amounts.
    0:15:22 Bringing, allowing more people to have easy access to that data is really important.
    0:15:27 So far, you know, data teams have to run kind of SQL analysis
    0:15:30 to get insights from these data sets.
    0:15:33 Large language models have been exciting to see,
    0:15:36 hey, can we turn language into SQL?
    0:15:38 And turns out that is a really, really difficult task
    0:15:42 because, you know, the world of data tends to be massive.
    0:15:47 You know, you have tens of thousands of tables, hundreds of thousands of columns,
    0:15:50 really complex abbreviations of column names and so forth.
    0:15:55 So we work really hard to have the world’s best text to SQL experience.
    0:15:56 And then we’ve achieved it.
    0:16:01 So we have the state of the art when it comes to text to SQL.
    0:16:05 And that now becomes available to business users
    0:16:07 who can now ask natural language questions.
    0:16:09 And then we turn that into SQL.
    0:16:11 We run that SQL and generate an answer.
    0:16:15 So something like, how is my revenue growing in this region for this product
    0:16:18 now becomes an easy question to ask for a business user.
    0:16:19 Right, fantastic.
    0:16:22 My guest today is Barish Gultekin.
    0:16:27 Barish is the head of AI at Snowflake, the AI data cloud.
    0:16:30 And we’ve been talking about, well, the role of data.
    0:16:32 We always talk about the role of data on this show
    0:16:33 because data is what fuels AI.
    0:16:39 But in particular, Snowflake’s approaches to everything
    0:16:43 from fine-tuning customer data, unlocking structured and unstructured data
    0:16:47 so that the customers, the developers, folks on Snowflake’s ends
    0:16:50 can turn the data into insights and applications.
    0:16:54 And then also some of the innovative approaches that Snowflake has taken
    0:16:58 to fine-tuning models, building models, converting text to SQL
    0:16:59 as we were just talking about.
    0:17:02 Let’s switch gears for a second, Barish, if we can,
    0:17:06 and talk about your background in AI at Google, even before Google.
    0:17:10 Have you always been a data nerd to put it that way?
    0:17:12 You’ve always been interested in data, computer science.
    0:17:14 Where did your journey start?
    0:17:19 Yeah, I have been, actually, before it was cool to say AI.
    0:17:24 I started this journey at Google a long time ago.
    0:17:32 And at some point around 2010, 2011, we started building Google now.
    0:17:35 And it was the traditional 20% project at Google,
    0:17:39 where we basically thought our phones should do better than what they do today.
    0:17:42 They should be able to give us the information we need when we need it.
    0:17:47 So it was this proactive assistant, and we built that product.
    0:17:52 Even though, at the time, the technology wasn’t quite there where it is now,
    0:17:55 we were able to give helpful information like,
    0:17:57 “Hey, there’s traffic on your commute,
    0:18:01 and you should just take this alternate route or your flight is delayed.”
    0:18:03 And all of that information felt magical
    0:18:07 because of bringing context with kind of prediction.
    0:18:08 Right, right, right.
    0:18:10 Even though it was a set of heuristics,
    0:18:12 it felt like, “Oh, there’s something intelligent here.”
    0:18:15 And that was the beginning, and I loved Indians.
    0:18:21 And then after that, I worked on Google Assistant.
    0:18:25 And Google Assistant is, again, exciting because it understands language,
    0:18:28 it can respond in natural language.
    0:18:31 Early on, it is just a series of use cases
    0:18:33 that are just coded one by one.
    0:18:34 Right, right.
    0:18:36 And now we’re at a point where, finally,
    0:18:38 computers can understand language,
    0:18:41 and you don’t have to kind of code each use case one at a time.
    0:18:41 Right, right, right.
    0:18:42 Exactly.
    0:18:46 When you’re working on all the things that you work on at this scale that you do now,
    0:18:49 or you did at Google, and now you do at Snowflake,
    0:18:54 how much do you trust the answers given by a generative AI model?
    0:18:58 And what sort of your own, I don’t know if it’s a workflow,
    0:18:59 so much as just kind of a mental,
    0:19:05 like, do you go back and verify results that you’re not sure about?
    0:19:07 Have you kind of gotten to, you know,
    0:19:11 do you have a feel for when something’s grounded versus hallucinated?
    0:19:12 And this is a little bit more of a,
    0:19:16 I don’t know, metaphysical question, perhaps, than the other stuff.
    0:19:19 But I’m just wondering, someone with as much experience
    0:19:22 and day-to-day working with this stuff as much as you do,
    0:19:26 what your kind of feel is for where the systems are right now?
    0:19:29 I try to see where we are.
    0:19:34 This generative ability, the creativity, if you will,
    0:19:36 a feature to a certain degree, right?
    0:19:40 So the types of use cases that are great are,
    0:19:43 when you ask the language models,
    0:19:45 to generate something, to generate content.
    0:19:48 And if my question is a factual question,
    0:19:51 then I know to be careful.
    0:19:54 But if it’s more of a help me brainstorm,
    0:19:57 let’s think about this, how do you say this differently?
    0:20:00 Those are the things where you’re leaning into the creativity,
    0:20:03 into the hallucination as a feature, if you will.
    0:20:04 Right, right, right.
    0:20:08 And so then how does that translate to enterprise customers you’re working with?
    0:20:10 I would imagine there’s a, you know,
    0:20:15 they sort of run the gamut from folks who are really excited to work with this,
    0:20:18 who maybe you have some customers who are a little more reticent,
    0:20:20 but feel like they should be,
    0:20:23 how do they relate to this whole notion of, you know,
    0:20:26 hallucinations being part of the deal?
    0:20:31 I think it’s incredibly important to know that,
    0:20:35 right now, where the technology is, we need to build systems.
    0:20:38 And these systems need to have grounding in them.
    0:20:42 So work hard to provide technology to help our customers,
    0:20:45 to make their systems, their products, their chatbots,
    0:20:48 a lot more grounded with the data that they provide.
    0:20:50 If an LLM is provided with grounding,
    0:20:54 if an LLM is provided with the data, it does not hallucinate, right?
    0:20:58 Only when there is lack of information, it then kind of makes it up sometimes.
    0:21:00 So we work hard on those solutions.
    0:21:01 We work with our customers.
    0:21:05 We also want to make sure our customers are able to evaluate these models.
    0:21:08 We’ve acquired a company called Truera just recently.
    0:21:14 Truera is a company that focuses on ML, LLM observability,
    0:21:19 being able to evaluate whether a chatbot that’s built is grounded,
    0:21:23 whether the quality is right, whether the cost is, you know, how they want it.
    0:21:26 So those are the technologies, tools that we’d like to offer to our customers.
    0:21:28 And we work closely with them.
    0:21:31 Right. And so along those lines, so that was an acquisition obviously,
    0:21:36 but Snowflake’s partnering, you mentioned that kind of your company’s openness and transparency.
    0:21:38 And there seems to be a spirit of that.
    0:21:45 And perhaps because everyone’s laser focused now on this, you know, frontier technology that
    0:21:50 inherently we’re all sort of figuring it out, whether as a user or developer to some degree.
    0:21:56 What’s the nature of some of the other partnerships or sort of what’s Snowflake’s
    0:22:00 role in working and partnering with some of the other tech giants and companies out there
    0:22:03 working at the leading edge of AI and ML?
    0:22:08 Yeah, we have very close partnerships with NVIDIA, with Meta,
    0:22:12 as well as Mistral and Raka, you know, the large language model providers
    0:22:14 who’ve invested in some of them as well.
    0:22:19 We basically see our platform as a way where we provide choice,
    0:22:24 but we work very closely with our partners in kind of helping us build specific solutions
    0:22:29 when it comes to kind of making sure that our RAG solutions are grounded,
    0:22:33 making sure that we have the world’s best-class text-to-sequel experience
    0:22:35 that requires partnerships.
    0:22:36 We work very closely with our partners.
    0:22:40 So in terms of openness, openness matters for many of our customers,
    0:22:45 understanding what kind of data was used to train a model is important.
    0:22:52 We also partner with some of our providers to have high-quality proprietary models as well.
    0:22:58 Snowflake, as I understand it, is a global company, has more than around 40 offices worldwide.
    0:22:58 That’s correct, yeah.
    0:23:00 Right, and how many data centers do you run?
    0:23:04 So we’re running on the three clouds, AWS.
    0:23:06 Okay, got it.
    0:23:06 ZP, yep.
    0:23:07 Right, right.
    0:23:11 So kind of looking at the present, but looking forward a little bit.
    0:23:12 I’m going to put you on the spot now, like I said I would.
    0:23:15 We touched on some of these things during the conversation,
    0:23:23 but are there major trends you’re seeing in business adoption and sort of real-world use cases
    0:23:30 that your customers, Snowflake’s customers are adopting now, or really are there trends
    0:23:36 in areas they’re really interested in exploring with the power of AI?
    0:23:42 And then kind of piggybacking on that, where do you see the industry headed?
    0:23:46 Sort of in broad strokes, if you like, over the next, say, three to five years.
    0:23:49 So we’re seeing a lot of exciting use cases.
    0:23:55 I’ve mentioned a couple of them, but our partners are, again, building production use cases.
    0:24:02 Some of them are our bread and butter, running large-scale analysis across their data inside
    0:24:02 Snowflake.
    0:24:08 So we’re seeing a lot of super simple, just using English language to be able to create
    0:24:13 categorization, extract information, and kind of make sense of a lot of data.
    0:24:21 For instance, one of our customers, Sigma, that’s a BI provider, are running analysis on
    0:24:26 sales logs from sales transcripts, sales calls.
    0:24:29 And if you come out understanding, why do we win?
    0:24:31 Why do we lose deals?
    0:24:37 So being able to run this now across a large data set of all sales calls in a period of time
    0:24:39 is as simple as writing English language.
    0:24:40 So that’s fascinating to me.
    0:24:41 That’s amazing, right?
    0:24:42 Yeah.
    0:24:47 And then, as I mentioned, of course, the bread and butter, high-quality chatbots,
    0:24:50 as well as being able to talk to your structured data for BI-type use cases.
    0:24:53 Those are the use cases that we’re seeing.
    0:24:58 What I’m seeing, of course, the world of AI evolves incredibly fast.
    0:25:01 Week over week, we get a new announcement, something new, exciting.
    0:25:02 I know.
    0:25:05 Three to five years, I should have said months or weeks even.
    0:25:06 That’s on me.
    0:25:08 Also, it feels like a year.
    0:25:14 Of course, the next big phase that is coming that’s already getting traction
    0:25:15 is the world of agents.
    0:25:22 So how many are we seeing the ability to answer questions by looking at documentation,
    0:25:23 but being able to take action?
    0:25:23 Right.
    0:25:28 And these agents are, this agentic systems are coming, this ability to reason,
    0:25:33 this ability to self-heal, the ability to take action for agents to talk to each other,
    0:25:37 collaborate, and that is the next evolution of the technology.
    0:25:37 Right.
    0:25:43 Are there any agentic frameworks on Snowflake that customers can access?
    0:25:44 So very soon.
    0:25:49 Right now, the agentic systems that we’ve built are behind the scenes.
    0:25:56 The Text2SQL BI experience uses a series of tools to deliver the product.
    0:25:56 Gotcha.
    0:25:58 So we will make that available to our customers.
    0:25:59 Right.
    0:25:59 Very cool.
    0:26:03 Looking back on your time at Snowflake, further back,
    0:26:06 I won’t prescribe time frame zero.
    0:26:07 I learned my lesson.
    0:26:13 Is there a particular story, moment, something that springs to mind as an important,
    0:26:20 perhaps unexpected learning that’s kind of really impacted how you view your work
    0:26:22 and the landscape today?
    0:26:27 Maybe a problem that the solution turned out to be something unexpected
    0:26:30 or something you thought was going to be hard and turned out to be simple.
    0:26:33 So I’ll give two examples.
    0:26:38 One is very early on, as we were building Cortex, we talked to a customer.
    0:26:41 This customer is a long-time Snowflake customer.
    0:26:46 They’ve built a pipeline to take their data out and then get it processed by an LLM running
    0:26:49 elsewhere and then get back.
    0:26:52 And of course, that pipeline took two months or so to build,
    0:26:56 and it was quite expensive to maintain and they were concerned about it.
    0:27:02 And our early prototype was able to replace the full thing with literally a single line of code.
    0:27:06 So we are on to something.
    0:27:12 When you bring compute, when you bring AI right next to where the data is,
    0:27:14 makes everything a lot simpler.
    0:27:17 And when it’s a lot simpler, it just unlocks a lot of usage.
    0:27:20 So I’m super excited about just ease of use, simplicity.
    0:27:27 The other example is just realizing how kind of demos are easy to build,
    0:27:28 but production systems are hard.
    0:27:36 You have, especially when it comes to working with structured data, generating SQL is difficult.
    0:27:42 So we work really hard on how do we build a system that together creates a very,
    0:27:43 very high quality response.
    0:27:49 When you’re essentially asking revenue questions, it’s not enough to be 80% accurate.
    0:27:53 So that’s another big important area that we focused on.
    0:27:55 All right, getting into the wrap up here.
    0:28:02 I always ask this question anyway, but I have a child who seems to be getting older every year,
    0:28:07 and now he’s in high school, interested in computers, computer science, physical science.
    0:28:12 What advice would you give to either a young person kind of looking out,
    0:28:16 maybe on the edge of graduating, a little older, maybe graduating college,
    0:28:21 or maybe somebody who’s older and is just interested in AI and sort of keeps hearing
    0:28:26 the things that we’ve been talking about, which is both that things are changing so fast,
    0:28:30 but also, there are things that we can do in the present moment,
    0:28:33 and still plenty of problems to be solved.
    0:28:35 So where should they go?
    0:28:38 Is studying computer science still a viable path?
    0:28:44 Is it better to just dive right into the work world and start working on,
    0:28:50 as you said, prototyping is one thing, building a production scale system is quite something else.
    0:28:56 What’s the advice that you give to young people or maybe older people looking to
    0:28:58 dive further into AI?
    0:29:05 So I think everyone has their own unique path, and everyone is drawn to something,
    0:29:09 and it’s important to be able to connect to what you’re drawn to,
    0:29:12 and it’ll be different for different people.
    0:29:16 But I just focus on that, just listening to that inner voice,
    0:29:20 which is hard to listen to sometimes, especially given there’s so much noise out there.
    0:29:24 But I will say, even though AI sounds intimidating,
    0:29:29 or there is this kind of artificial intelligence that sounds very complex,
    0:29:34 and it is complex when you start going down the rabbit hole and start doing your research,
    0:29:40 however, the use of the AI is going to unlock, it’s incredibly easy.
    0:29:45 All of these systems are now an API away, and they’re probably powerful.
    0:29:52 So I think creativity is going to determine all sorts of super interesting technologies
    0:29:53 to be built next.
    0:29:57 So I would say, don’t be intimidated with technology, just dive right in,
    0:30:02 and it’s incredibly easy to use, and really looking forward to what’s
    0:30:04 to come in the next two years or so.
    0:30:07 Love it, love the optimism, more audio only, which is a shame,
    0:30:11 because your face lit up, smile like I did when you were talking about that.
    0:30:16 Bearish, you alluded earlier to cookbooks and other resources that Snowflake makes available.
    0:30:19 Maybe we can divvy this up into two parts.
    0:30:23 Potential customers who want to learn more about what Snowflake does,
    0:30:26 what the offerings are, how to maybe engage with you.
    0:30:30 And then folks, practitioners working in AI wanting to learn more about,
    0:30:33 you know, what Snowflake’s been doing, research,
    0:30:37 some of the techniques we talked about, where can people go online to learn more?
    0:30:43 So our website, snowflake.com, if you are trying to figure out how do I use AI
    0:30:48 just in seconds and bring my data, analyze my data, we have a solution for you.
    0:30:49 Thanks.
    0:30:51 So snowflake.com is the place.
    0:30:52 Perfect.
    0:30:53 Bearish, thank you.
    0:30:54 This was great.
    0:30:58 As with many of these conversations of these days, I feel like this was kind of the warm-up,
    0:31:02 and we’ll have to get back in touch down the line to really dig into where things are headed.
    0:31:07 But the Snowflake story, you know, is a great one, and it seems like it’s just getting started.
    0:31:10 So congratulations on the work so far.
    0:31:12 All the best to you going forward.
    0:31:15 And, you know, look out for my unnamed friend I mentioned earlier,
    0:31:16 if you see them around campus.
    0:31:17 That sounds great.
    0:31:18 Thanks for having me.
    0:31:28 You know.
    0:31:38 [Music]
    0:31:48 [Music]
    0:32:08 [Music]
    0:32:18 [BLANK_AUDIO]

    Snowflake is using AI to help enterprises transform data into insights and applications. In this episode of NVIDIA’s AI Podcast, host Noah Kravitz and Baris Gultekin, head of AI at Snowflake, discuss how the company’s AI Data Cloud platform enables customers to access and manage data at scale. By separating the storage of data from compute, Snowflake has allowed organizations across the world to connect via cloud technology and work on a unified platform — eliminating data silos and streamlining collaborative workflows.

  • Recursion CEO Chris Gibson on Accelerating the Biopharmaceutical Industry With AI – Ep. 230

    AI transcript
    0:00:11 [MUSIC]
    0:00:13 >> Hello, and welcome to the NVIDIA AI podcast.
    0:00:15 I’m your host, Noah Kravitz.
    0:00:19 My guest today is CEO and co-founder of Recursion,
    0:00:22 one of the world’s leading biotech,
    0:00:25 or tech bio as they call it, companies in the world.
    0:00:29 Chris Gibson started Recursion based on the work he developed while working on
    0:00:32 his joint MD-PhD at the University of Utah,
    0:00:37 and today the company is dedicated to coding biology in the name of radically improving lives.
    0:00:41 Recursion is building one of the largest proprietary datasets in their field.
    0:00:44 They just took the wraps off one of the most powerful supercomputers in the world,
    0:00:48 and they’re one of the company’s leading the growing field of AI-powered drug discovery.
    0:00:52 Chris is here to tell us all about it and then some, so let’s dive right in.
    0:00:56 Chris Gibson, welcome, and thank you so much for joining the NVIDIA AI podcast.
    0:00:58 >> Thanks, Noah. I’m delighted to be here.
    0:01:01 As is often the case with these episodes,
    0:01:04 I try to pack everything I can into the intro to set the stage.
    0:01:07 You’ve done so much, you’re working on so much,
    0:01:08 but let me turn it over to you.
    0:01:12 Maybe we can start by you explaining to the audience what Recursion is all about.
    0:01:14 >> Yeah. I want to tell you about Recursion,
    0:01:17 but first I want to tell you about the problem that we’re solving.
    0:01:17 >> First.
    0:01:22 >> Everybody knows somebody who has a disease where there’s not a good treatment.
    0:01:24 You’ve got a relative who’s died of cancer,
    0:01:26 you have somebody who’s suffering from Alzheimer’s or Parkinson’s,
    0:01:29 or some of these other devastating diseases.
    0:01:33 Today, despite all of the incredible technology in the world,
    0:01:37 90 percent of drugs that go into clinical trials fail.
    0:01:40 Nine out of 10 drugs that our industry puts into clinical development
    0:01:42 fail before they get to patients.
    0:01:48 What that tells me is that biology is just this massive quagmire of complexity.
    0:01:53 It’s just dramatically complex to a level where despite all of these hundreds of
    0:01:57 thousands of scientists around the world and about $50 billion a year of R&D
    0:02:01 investment from the industry, we still aren’t that good at it.
    0:02:07 We imagined that perhaps there was a way that you could bring together new technologies
    0:02:10 to try and take a less biased approach to biology,
    0:02:12 to try and step back and say,
    0:02:16 instead of trying to understand how every gene interacts with every gene and
    0:02:21 every drug interacts with every gene and build it all in our heads,
    0:02:23 which is the way it’s been done traditionally.
    0:02:25 Could we actually take an industrial approach,
    0:02:28 build maps of biology using automation,
    0:02:33 using AI, and then leverage those to tell us where to go,
    0:02:37 to essentially have the algorithm tell us where to develop a medicine.
    0:02:40 That’s what we’ve been working on for the last 10 years.
    0:02:43 Now we’re what’s called a clinical stage biotech company.
    0:02:46 We’ve got five drugs that are in the clinic ourselves.
    0:02:49 We’ve got big partnerships with companies like Roche Genentech and
    0:02:53 Buyer to bring drugs into really hard areas of biology.
    0:02:56 We just built this incredible supercomputer with Nvidia,
    0:02:58 one of our partners as well.
    0:03:03 So we’re I think really one of the companies leading at the intersection of biology,
    0:03:06 but also of technology.
    0:03:08 A million things to get into.
    0:03:09 But before we go forward,
    0:03:11 I want to ask you use the word map.
    0:03:14 You talked about an industrial approach mapping biology,
    0:03:17 and that made me think of things like mapping the human genome.
    0:03:19 Is there a parallel there?
    0:03:21 Is it a similar approach or how does that work?
    0:03:23 Yeah, so traditionally in our industry,
    0:03:25 people have worked on one disease at a time.
    0:03:26 Okay.
    0:03:29 And so if we wanted to go after a disease together,
    0:03:31 we’d go read the literature,
    0:03:32 we’d build a team,
    0:03:37 and then we would build specific experiments for that one disease.
    0:03:37 Right.
    0:03:40 And we would generate data and five to 10 years later,
    0:03:42 maybe we would have a drug going into the clinic and
    0:03:44 you know what the success rate is there.
    0:03:45 It’s a 90% failure.
    0:03:47 There’s a different kind of approach,
    0:03:49 which is instead of working on one thing at a time,
    0:03:55 can we build vast data sets that span very large scales?
    0:03:58 And I think the human genome project is a great example.
    0:04:00 People said let’s map the entire human genome.
    0:04:03 And now today there’s tens of millions of genomes
    0:04:04 that have been that have been mapped
    0:04:06 and we can compare them and contrast them.
    0:04:08 We’re using some of those same data.
    0:04:12 And so we’ve taken this approach that to invest more initially
    0:04:15 to build really large, complex data sets
    0:04:16 that at the very beginning,
    0:04:18 you’re paying a lot to build this data set
    0:04:21 and you don’t yet have enough data to actually make any progress
    0:04:22 against any disease.
    0:04:23 But over time,
    0:04:26 you start to build these network effects where today,
    0:04:29 if we run an experiment in our automated laboratory
    0:04:30 and we get some result,
    0:04:34 we can compare it to over 250 million experiments
    0:04:35 that we’ve run over the past few years
    0:04:37 and all of that data is relatable.
    0:04:38 So instead of one disease at a time,
    0:04:40 instead of slices of biology,
    0:04:42 we’re actually building like a volume
    0:04:45 and we’re sparsely sampling this volume
    0:04:47 and then using AI to sort of fill in
    0:04:50 what we can predict about the rest of it.
    0:04:53 – Right, is that a unique approach in the industry?
    0:04:55 – I think it’s pretty unique in the industry.
    0:04:56 Yeah, there’s a handful of companies
    0:04:58 that are taking similar approaches.
    0:05:01 I’m unaware of any company that has generated a data set
    0:05:03 that is this broad.
    0:05:05 So we’ve knocked out with CRISPR-Cas9
    0:05:07 that some of your listeners may have heard of.
    0:05:11 It’s like a molecular scissors that lets us cut out genes.
    0:05:13 We’ve knocked out every gene in the human genome
    0:05:16 in multiple different human cell types.
    0:05:18 We’ve profiled millions of molecules
    0:05:21 and all of these data exists now in multiple layers.
    0:05:23 We call it omics.
    0:05:24 People may have heard of like genomics.
    0:05:27 Well, we’re building phenomics and transcriptomics
    0:05:29 and invivomics and proteomics
    0:05:31 and all of these omics data layers.
    0:05:32 And you can think of it as,
    0:05:34 go back to like the Google map days, right?
    0:05:36 Started out with maps from airplanes
    0:05:38 and we got kind of where all the streets were
    0:05:40 and then eventually you had street view
    0:05:42 and cars were driving around.
    0:05:43 We’re building all these same layers
    0:05:45 but instead of doing it in the physical world,
    0:05:49 we’re using a massive automated laboratory full of robots
    0:05:50 to do millions of experiments
    0:05:52 to figure out what the roads and streets
    0:05:54 and intersections are of biology and chemistry.
    0:05:56 And it’s a complex space.
    0:05:58 I mean, I think it’s about as complex a problem
    0:05:59 as one can work on.
    0:06:01 – I can only imagine.
    0:06:03 Maybe by the end of this conversation,
    0:06:06 I’ll know more about what I don’t know at the very least.
    0:06:09 So maybe we can start with the graduate work,
    0:06:11 the kind of seated recursion.
    0:06:12 And if this is the wrong approach,
    0:06:13 take a different one,
    0:06:16 but I’m imagining that might help me and the audience
    0:06:19 sort of to understand the problem,
    0:06:20 the scope of the problem
    0:06:23 and how and why you started with this approach.
    0:06:24 – Yeah, of course.
    0:06:26 I think this is a great way to kind of explain
    0:06:27 what we’re doing at recursion.
    0:06:30 So I joined the lab of a guy named Dean Lee.
    0:06:32 Dean’s actually now the president of Merck Research Lab.
    0:06:35 So he’s like the head scientist of Merck.
    0:06:37 Very physician scientist, brilliant guy.
    0:06:39 The lab was so diverse.
    0:06:43 There were physicians and engineers and geneticists
    0:06:45 and molecular cellular biologists.
    0:06:47 And we were working on all these cool, hard problems.
    0:06:49 And one of the things we were working on
    0:06:52 was this rare genetic disease
    0:06:54 called cerebral cavernous malformation.
    0:06:57 And I bet about 1% of your audience has heard of it
    0:07:01 because about 1% of the people in the world
    0:07:03 have this disease.
    0:07:06 So it’s like a rare disease, but it’s not that rare.
    0:07:10 It’s six times more prevalent than cystic fibrosis.
    0:07:12 But there’s no drug, there’s no treatment.
    0:07:13 And so because of that,
    0:07:15 people don’t know about the disease as much.
    0:07:18 And we were trying to figure out how this disease works.
    0:07:20 And we use traditional molecular
    0:07:21 and cellular biology approaches
    0:07:24 where I could hop on the whiteboard that’s behind me
    0:07:27 and I could draw protein X goes to protein Y
    0:07:28 goes to protein Z.
    0:07:30 And we think protein Z gets too high
    0:07:32 and that causes the disease.
    0:07:35 And after a decade of working on this,
    0:07:37 we think we figured it out.
    0:07:40 We’re sitting in lab and we think this protein called row A
    0:07:42 is what’s causing the disease.
    0:07:46 And we take a row A inhibitor and we put it in mice.
    0:07:47 And five months later, I remember sitting in lab
    0:07:50 meaning we unveiled the data.
    0:07:51 Oh, we changed the mice.
    0:07:53 We changed the mice in the wrong direction.
    0:07:54 They got worse.
    0:07:56 They got more of these lesions.
    0:07:58 And this is one of those problems with biology.
    0:08:02 It’s like as humans, we are reductionist problem solvers.
    0:08:04 We take a really complex system
    0:08:06 and we try and reduce it down to these core elements
    0:08:08 of protein A, protein B, protein C
    0:08:09 so that we can put it on a whiteboard
    0:08:11 or put it in a nature paper.
    0:08:12 The reality isn’t biology.
    0:08:14 There’s probably hundreds of interconnections,
    0:08:17 thousands of feedback loops that are all working together.
    0:08:19 And if you take a reductionist approach, I would argue
    0:08:23 that’s maybe why we’re failing 90% of the time in the clinic.
    0:08:26 That the way we need to actually explore biology
    0:08:28 is not to take a reductionist approach,
    0:08:30 but to up-level our understanding of biology,
    0:08:32 to understand the whole complex system
    0:08:35 and to build maps that truly embrace
    0:08:37 how every gene interacts with every other gene.
    0:08:38 And so that’s hard.
    0:08:41 I mean, we’ve been working at this for 10 years,
    0:08:44 but we took a very early version of this approach
    0:08:47 in Dean’s lab after that failure.
    0:08:50 We took microscopy images of human cells
    0:08:52 where we were modeling this disease
    0:08:54 and we took microscopy images of human cells
    0:08:56 that were healthy and we trained
    0:08:59 a basic machine learning classifier to recognize the two.
    0:09:01 And then we added thousands of drugs to the disease cells
    0:09:04 and we simply asked the machine learning classifier,
    0:09:06 do any of the disease cells look healthy again?
    0:09:07 Without any understanding
    0:09:09 of what else was happening in biology.
    0:09:13 And today, recursion is a few months away
    0:09:15 from reading out a phase two clinical trial
    0:09:19 against a drug that we discovered doing that work.
    0:09:23 It was a totally surprising way that this drug was working.
    0:09:25 It was not going after ROA or anything else.
    0:09:27 And so I guess the point of this is
    0:09:28 when you embrace the complexity of biology
    0:09:31 and let biology give you the answer,
    0:09:33 it can be surprising, it can challenge dogma.
    0:09:35 But if you’re willing to follow that,
    0:09:37 that our belief is that eventually
    0:09:39 with enough technology and investment,
    0:09:41 this could be a more sustainable industrial way.
    0:09:44 And so I finished my PhD,
    0:09:45 I took a leave of absence from medical school,
    0:09:47 subsequently dropped out
    0:09:49 because recursion ended up kind of taking off
    0:09:50 and started the company.
    0:09:52 And we’ve been building for the last 10 years
    0:09:55 at this interface of tech and bio.
    0:09:58 – So 10 years ago, or thereabouts,
    0:10:00 when you got these lab results back
    0:10:04 and things were going in the wrong direction,
    0:10:05 was there a sense either with you
    0:10:09 or colleagues in the lab or just colleagues in general,
    0:10:11 was there a sense of like,
    0:10:14 we know that there’s a different approach
    0:10:16 to embrace the complexity,
    0:10:18 but we just can’t do it right now
    0:10:20 ’cause it’s too big for the human brain
    0:10:24 or even dozens of the best human brains working in parallel.
    0:10:27 It’s just too much information to sift through.
    0:10:28 And did you know then,
    0:10:30 I mean, you said you used an ML classifier,
    0:10:32 but was it a thing of like,
    0:10:33 if only the tech was a little better
    0:10:35 or what was kind of the mind state then?
    0:10:37 – Yeah, I think it was,
    0:10:40 so we had this idea to use technology
    0:10:43 to try and take a less biased approach.
    0:10:45 We’re not the first people to have done this.
    0:10:46 There’s a handful of other people
    0:10:48 that were working on similar things at the time,
    0:10:50 but AI wasn’t really being,
    0:10:52 this is 2011-ish,
    0:10:55 like AI is not really being thrown around as a term,
    0:10:57 it’s kind of machine learning at the time.
    0:10:59 People aren’t doing a lot of work in neural nets.
    0:11:01 I mean, like ImageNet hasn’t even come out yet.
    0:11:03 And so we were like,
    0:11:06 among a very early wave and certainly in biology,
    0:11:09 among a deeply early wave of people saying,
    0:11:12 let’s use like computer vision to look at images.
    0:11:13 And the work had actually really been pioneered
    0:11:16 by a woman named Anne Carpenter at the Broad Institute.
    0:11:18 She built the software tool that we used
    0:11:19 that helped us go fast.
    0:11:21 And yeah, we wanted the software to be better,
    0:11:22 but at the same time,
    0:11:25 we were pushing up against the frontier
    0:11:27 at the time in biology.
    0:11:29 And so what we’ve done now at recursion
    0:11:32 is we’ve pioneered the industrialization of this approach.
    0:11:35 And what I joke about is if you visited our headquarters here
    0:11:37 and you saw a robotic laboratory,
    0:11:39 there’s like a sad and exciting fact.
    0:11:41 And that is that this robotic laboratory
    0:11:44 does the equivalent of all the experiments
    0:11:46 I did in my entire five years of my PhD,
    0:11:49 every 15 minutes on average.
    0:11:51 And so like that is both sad in some ways
    0:11:52 and kind of exciting in others.
    0:11:54 I can only imagine you’re humbled.
    0:11:55 And at the same time,
    0:11:57 you’re also like excited in the wave.
    0:11:58 Yeah, yeah, it’s amazing.
    0:12:01 So I don’t know if you want to jump all the way
    0:12:03 to the present or what the best way is to walk us through,
    0:12:06 but I want to get to what recursion is doing now.
    0:12:09 Yeah, I mean, I think we can jump to where we are today,
    0:12:11 which is we’ve taken this philosophy
    0:12:14 of creating virtuous cycles
    0:12:15 of what we call wet lab and dry lab.
    0:12:18 And what that means is empirical data generation.
    0:12:20 So a wet lab where we’re doing real experiments
    0:12:21 with human cells and a dry lab,
    0:12:23 which is our supercomputer system
    0:12:25 and all of our software tools and AI tools.
    0:12:28 And I think that what we’re doing here at recursion
    0:12:31 is analogous to so many technology companies.
    0:12:34 So Netflix is recording what you’re watching,
    0:12:35 when you’re watching it, when you turn it on,
    0:12:38 when you turn it off, who’s watching it in the household,
    0:12:40 which scenes you’re turning it off on.
    0:12:43 And they’re actually now then making predictions
    0:12:45 and sort of A/B testing and going back through this loop
    0:12:49 to the point today where Netflix is generating content
    0:12:51 based on an algorithmic suggestion
    0:12:54 of what’s gonna be popular for people.
    0:12:56 This is like the drug discovery version of that.
    0:12:59 We’re doing experiments, we’re breaking genes
    0:13:01 and adding compounds and combinations of the things
    0:13:02 that different human cell types.
    0:13:06 And our A/B experiment is we make a bunch of predictions
    0:13:08 about what genes and what drugs are connected to each other.
    0:13:10 And then the next week we go back
    0:13:11 and we test those predictions
    0:13:13 and create this kind of flywheel approach.
    0:13:16 The problem, of course, is that biology
    0:13:19 is just so, so, so complex.
    0:13:22 There’s this combinatorial explosion
    0:13:23 of what we always joke about.
    0:13:26 There’s about 21,000 human genes.
    0:13:28 And in biology, there’s this really cool thing
    0:13:30 called like synthetic lethality
    0:13:32 or like synthetic relationships
    0:13:36 where you can start to predict that two genes are related.
    0:13:37 If you break both those genes
    0:13:40 and you get an unexpectedly large or small effect,
    0:13:41 it tells you that they might be
    0:13:43 in some kind of feedback loop together.
    0:13:45 If we were gonna do this synthetic experiment
    0:13:49 of knocking out every gene with every other possible gene,
    0:13:52 it’s about 250 million experiments.
    0:13:52 It’s doable.
    0:13:54 We’ve done 250 million experiments.
    0:13:56 It’s taken us 10 years to get there,
    0:13:58 but we now do up to 2.2 million a week.
    0:14:00 So like this is feasible.
    0:14:02 But if I just said, what if we did gene by gene by gene?
    0:14:03 Three genes.
    0:14:06 Now you’re talking about trillions of experiments.
    0:14:07 And if you imagine doing four genes,
    0:14:10 like you instantly in biology to really explore this,
    0:14:12 you get to this combinatorial explosion
    0:14:13 where we can’t brute force it.
    0:14:15 We’re not gonna be able to empirically do it.
    0:14:17 And so this is the beauty of AI,
    0:14:20 just like in all of these other technology fields.
    0:14:23 If you can sparsely fill this massive volume,
    0:14:26 this matrix of genes and compounds and cell types
    0:14:30 and interactions, ML and AI tools are often,
    0:14:32 if the data is robust, really good
    0:14:33 at helping you fill in and predict
    0:14:35 what’s gonna happen in between.
    0:14:36 And I think that’s what we’re trying to build.
    0:14:39 That’s what that map is to us.
    0:14:40 – And so forgive me,
    0:14:42 this is a very sort of simplistic question,
    0:14:45 but is it a case of where you’re running experiments
    0:14:48 in the dry lab, doing it all in the AI,
    0:14:49 you know, the computerized system,
    0:14:52 and then sort of the ones that have the best chance
    0:14:54 of going somewhere or go into the wet lab?
    0:14:55 Is that kind of the basic?
    0:14:57 – Yeah, that’s exactly right.
    0:15:01 Yeah, and then what’s cool about these virtuous cycles
    0:15:04 is if the prediction we made from the dry lab
    0:15:07 goes to the wet lab and it works, it’s really robust.
    0:15:08 Awesome, okay, we’ve got a new program.
    0:15:10 Like let’s take it forward towards patients.
    0:15:12 If it doesn’t work,
    0:15:14 we’ve now that week generated a bunch of new data
    0:15:17 upon which we can retrain the model.
    0:15:20 And that new data is existing in a space
    0:15:22 where the model was making a poor prediction.
    0:15:24 And so by definition, week over week,
    0:15:26 as you keep kind of going through this virtuous cycle,
    0:15:28 despite the complexity of biology and chemistry,
    0:15:30 you start to make progress
    0:15:32 because there are areas of biology
    0:15:35 that are kind of like, you know, like the Midwest,
    0:15:38 like, you know, cornfields as far as you can see,
    0:15:42 it kind of looks reasonably similar 100 miles apart.
    0:15:44 And there’s other places in biology
    0:15:47 that are like, you know, the Rocky Mountains or, you know,
    0:15:49 whatever, and it’s like, you go 10 feet to the left
    0:15:50 or 10 feet to the right,
    0:15:52 dramatically different situation.
    0:15:55 And so we can help to hone in with this iterative approach
    0:15:58 around the parts of biology where we need more data
    0:16:00 and the parts where we need less.
    0:16:04 – Are there certain areas of biology that are,
    0:16:06 and again, I’m gonna be overly simplistic here,
    0:16:09 that are easier or harder to,
    0:16:12 and I don’t know if it’s to map out or to understand
    0:16:15 or to be able to, you know, sort of act on?
    0:16:17 – For sure, so there are areas of biology
    0:16:19 that we think are easier to start with.
    0:16:20 And there are areas of biology
    0:16:24 where we know exactly what is causing the disease.
    0:16:26 So for me, there’s three areas there.
    0:16:30 One is genetic diseases where we like cystic fibrosis.
    0:16:34 We know that mutations in the CFTR gene cause cystic fibrosis.
    0:16:37 Another good example are some cancers
    0:16:39 where we know that mutations in certain genes
    0:16:40 cause those cancers.
    0:16:42 And the third, infectious diseases.
    0:16:45 We know that like this virus or this bacteria
    0:16:46 causes this disease.
    0:16:48 What we like about those is that if we model those
    0:16:52 in cells in our lab, you know, we add the virus
    0:16:53 or we break the gene,
    0:16:56 we know that we’ve at least partially recapitulated
    0:16:58 something relevant about the biology.
    0:17:01 The harder areas are like in neuroscience.
    0:17:03 So take like Alzheimer’s.
    0:17:05 We don’t really know what causes Alzheimer’s.
    0:17:07 There’s debate, some people think they do,
    0:17:09 but most of the drugs against those targets have failed.
    0:17:12 So like we don’t really know what causes it.
    0:17:15 And so how do you go about trying to understand Alzheimer’s?
    0:17:17 And that’s actually neuroscience broadly
    0:17:20 is probably one of the areas that’s hardest
    0:17:22 because like we really don’t know
    0:17:24 what causes a lot of different neurologic diseases.
    0:17:28 And so we’ve actually partnered with Roche and Genentech,
    0:17:30 the, you know, one of the biggest biopharma companies
    0:17:32 there is, one of the most innovative
    0:17:36 in a decades long collaboration to go map the genome
    0:17:39 in neural specific cells.
    0:17:42 So we can start to uncover some of these answers.
    0:17:45 But those areas, those are the hard parts of biology.
    0:17:46 – Right, right.
    0:17:47 Should we talk supercomputers?
    0:17:49 – Yeah, let’s talk supercomputers.
    0:17:51 – All right, so I would imagine taking, you know,
    0:17:54 somewhat of a rogue isn’t the right word,
    0:17:56 but a different approach from the beginning, right?
    0:17:59 The unbiased sort of industrialized approach.
    0:18:02 And then now talking about, you know, mapping things out
    0:18:05 and there’s so much we had a guest on recently
    0:18:08 who talked about, you know, he had a great metaphor
    0:18:09 for the difference between what we know
    0:18:12 and what we don’t know about his field, right?
    0:18:14 And so similar thing, right?
    0:18:17 So I would imagine that the amount of data,
    0:18:19 the amount of compute, it’s like with any other problem,
    0:18:20 it just goes up and up and up.
    0:18:23 So over the years, maybe you can talk a little bit
    0:18:25 about kind of the technological path
    0:18:29 that led you to, we’re at BioHive 2 now.
    0:18:31 – All right, so maybe you can walk us through that a little bit.
    0:18:34 – Yeah, I mean, we started out on my laptop
    0:18:38 and then we moved to a server we built in-house called Golgi,
    0:18:41 which we eventually mourned when we deprecated that.
    0:18:44 But today we’ve got about 50 petabytes
    0:18:47 of proprietary biological data.
    0:18:48 And these data are a little bit different
    0:18:51 from sort of like text and other things, some of it’s text.
    0:18:54 But a lot of it is images, they’re very memory rich,
    0:18:57 memory intensive, you know, they’re like big images.
    0:19:00 And so some of the traditional cloud-based approaches
    0:19:03 don’t quite work for us because you end up starving the GPU
    0:19:05 from just a memory aside.
    0:19:10 And so we in 2020 decided we needed to build a supercomputer
    0:19:13 in order to make the best use of this giant dataset
    0:19:15 we’ve accumulated and built.
    0:19:17 And so we built a supercomputer called BioHive 1.
    0:19:21 It was at the time on the top 500 list when it came out,
    0:19:24 I think like 85th or 58th or something like that.
    0:19:27 And what we found was that the scaling laws apply,
    0:19:30 the bitter lesson holds in biology.
    0:19:33 And that is that more data and more compute
    0:19:35 both give you better outcomes.
    0:19:37 And we’re generating more data every day.
    0:19:40 And so we’re, I think in biology,
    0:19:44 probably data is the bigger bottleneck than compute.
    0:19:48 You know, there’s not the same arms race as in other fields
    0:19:51 because there’s not yet enough relatable data
    0:19:54 like we’re building for everybody to be doing this.
    0:19:55 We are building the data,
    0:19:57 but we also knew we needed more compute.
    0:20:00 And so just maybe six months ago,
    0:20:03 we announced actually NVIDIA invested in us last summer.
    0:20:06 And then as we continued building our relationship with them,
    0:20:09 we ultimately decided to build BioHive 2,
    0:20:12 which just came out top 500 list.
    0:20:14 It’s number 35.
    0:20:18 It’s about 23 petaflops and it’s 504 H100s.
    0:20:23 And now we just deprecated BioHive 1 with our 300 plus A100s
    0:20:26 and we’re moving those over to the new facility.
    0:20:31 And so we’ll have this Frankenstein A100 H100 giant supercomputer,
    0:20:33 at least in our field, little recursion.
    0:20:38 This biotech, tech bio startup in Salt Lake City now owns and operates
    0:20:41 the fastest supercomputer in the world of any biopharma company,
    0:20:44 you know, from Pfizer, you know, all the way down.
    0:20:46 -Amazing. -It’s pretty cool.
    0:20:48 -Yeah, yeah, no, it’s amazing.
    0:20:50 We, you know, over the years of doing the podcast,
    0:20:56 this phrase democratization of tools has always been almost like a mantra.
    0:21:00 And, you know, for a while it was thinking about like a lawyer,
    0:21:04 his computer and his apartment office has a GPU
    0:21:07 and he realized he could use it to do some machine learning stuff.
    0:21:09 And, you know, similar thing, right?
    0:21:11 You guys are a much bigger scale than that.
    0:21:15 But as you said, compared to the longstanding giants of the industry,
    0:21:17 you know, for you to be able to build that is amazing.
    0:21:20 Do you have a good story or a lesson learned
    0:21:26 or something sort of surprising, specific to building BioHive 2?
    0:21:29 And that process, that was something that, you know,
    0:21:32 a hurdle you had to overcome or something you were expecting
    0:21:34 was going to go one way and it turned out to be something different
    0:21:36 or, you know, maybe just an anecdote?
    0:21:39 -Yeah, I mean, building BioHive 2 was a tremendous effort.
    0:21:43 Multiple teams came coming together, the recursion team, the NVIDIA team,
    0:21:48 other suppliers teams, we ended up wanting to get this done
    0:21:51 in time for submission to the most recent top 500 list.
    0:21:53 -Okay. -And, you know, as is the case right now
    0:21:56 with the popularity of GPUs and the importance of AI,
    0:21:59 we got our GPUs, we got our cluster,
    0:22:03 but we were missing racks and cables, certain racks and certain cables,
    0:22:08 and those ended up coming in 15 days before the top 500 submission.
    0:22:12 And so, our team was sleeping on cots in the data center
    0:22:16 and the NVIDIA team joined us, like the amount of effort that they put in.
    0:22:22 We got that thing up and running, burned in and benchmarked in 15 days
    0:22:23 from getting all the materials.
    0:22:29 It ended up being, it’s the first on-slab H100 setup, as far as I’m aware,
    0:22:32 and it ended up having really, really good performance.
    0:22:33 -Excellent.
    0:22:35 My guest today is Chris Gibson.
    0:22:39 Chris is co-founder and CEO of Recursion,
    0:22:42 one of the world’s foremost biotech.
    0:22:44 TechBio, I want to ask you about that in a second,
    0:22:46 companies in the world based out of Utah,
    0:22:48 and as Chris was just detailing,
    0:22:53 they are now the proud owners and maintainers.
    0:22:56 Was it the 35th ranked?
    0:22:58 -Yeah, 35th. -Super pure in the world, BioHive2.
    0:23:03 TechBio, I assume, it’s just kind of a flip on the term biotech
    0:23:07 that the tech is kind of leading the way forward?
    0:23:08 -Yeah, that’s right.
    0:23:10 I mean, the term biotech was coined in the ’80s
    0:23:14 to talk about using human proteins as drugs.
    0:23:16 -Okay. -And at the time, that was tech.
    0:23:18 And it was Genentech that really started this off.
    0:23:20 And so credit to them, they were first.
    0:23:23 They coined their industry biotech.
    0:23:25 We would have called it biotech, but that was taken.
    0:23:28 And so somebody in our field called it TechBio,
    0:23:30 and the tech is kind of at the forward side.
    0:23:34 And it’s a good term to describe the few hundred companies now
    0:23:37 that are more digitally native, that are building in this space,
    0:23:40 where technology is really at the foundation of what they build.
    0:23:41 -Got it.
    0:23:44 Also in prepping for this conversation,
    0:23:48 I noticed the name BioHive being used.
    0:23:50 It’s the name of a public-private partnership
    0:23:52 that I think you also have a hand in.
    0:23:54 So maybe you can talk a little bit about that.
    0:23:56 As I was listening to you earlier,
    0:23:59 I was thinking about some of the recent podcasts we’ve done.
    0:24:02 We’ve done one or two others in the drug discovery space
    0:24:06 and a bunch of others in the sciences.
    0:24:08 And more and more, we’re hearing about,
    0:24:10 I think the technology is a big part of it,
    0:24:13 obviously, in making an open-source collaboration
    0:24:16 and just other collaboration with large amounts of data,
    0:24:20 easier to do across the world.
    0:24:23 So I’m interested both in the specifics of BioHive
    0:24:27 and then also just kind of your take on the current state of,
    0:24:31 you know, science companies, biotech companies,
    0:24:33 working in sort of a capitalist environment
    0:24:37 where you’re trying to, you know, compete, you are competing.
    0:24:38 But then at the same time,
    0:24:39 given the work that you’re doing,
    0:24:41 sharing is obviously quite important.
    0:24:43 So maybe you can start with BioHive
    0:24:46 and kind of get into just what collaboration is like
    0:24:47 in the field today.
    0:24:50 -Yeah, I think this new generation of tech bio companies,
    0:24:54 we really do believe that we will all do better
    0:24:57 if we all are collaborating at least
    0:24:59 at certain stages of the process.
    0:25:01 Because, you know, like at the end of the day,
    0:25:03 if there’s a disease that doesn’t have a treatment,
    0:25:05 we all still have work to do.
    0:25:08 And so we believe a lot in investing
    0:25:09 in the communities in which we work.
    0:25:12 We acquired a company last year called Valence
    0:25:16 that now basically is building out a community
    0:25:19 for people to host, you know, different foundation models
    0:25:22 and other sorts of things that are important across biology.
    0:25:23 They’re hosting data sets.
    0:25:27 They’re really investing in building that community.
    0:25:30 And we also believe that’s true in the locations we work.
    0:25:35 And BioHive is what we call the kind of ground-level
    0:25:38 organization that’s helping to brand, bring together
    0:25:41 and build the life science ecosystem in Utah.
    0:25:43 We also do investments in our other, you know,
    0:25:47 we’ve got offices in Toronto, Montreal, London,
    0:25:49 and Melpitas, California.
    0:25:50 And so we also make investments in those places
    0:25:53 because we feel like we’re on like a 30 year…
    0:25:56 I mean, we look at NVIDIA 30 years in,
    0:25:59 and we feel like we’re 10 years into our 30-year journey
    0:26:00 to kind of prove out this vision
    0:26:03 that we think could be so impactful for the world.
    0:26:07 And we know if you’re on that kind of trajectory
    0:26:08 over that kind of timeline,
    0:26:11 you’ve got to build community around you as you go.
    0:26:13 And so these are important investments for us.
    0:26:16 Open-sourcing data sets has been a critical piece
    0:26:18 of how we’ve not only helped build the community
    0:26:21 but also attracted talent over the years.
    0:26:24 And so I think we’re going to continue pushing the industry,
    0:26:26 the pharma and biotech industry,
    0:26:29 to be a little bit more open to these kinds of approaches.
    0:26:31 – Excellent, I love to hear that.
    0:26:34 There’s data, there’s compute, there’s hardware.
    0:26:37 There’s also software, the tools that make it all run.
    0:26:41 I wanted to ask you about a recursion tool called Lowe
    0:26:42 that I heard a little bit about.
    0:26:44 My understanding is it kind of helps orchestrate
    0:26:47 the workflows that uses GenAI in some way.
    0:26:48 What’s that all about?
    0:26:51 – Yeah, so we’ve been building all these software tools.
    0:26:53 Some of them are leveraging neural nets
    0:26:54 and other sorts of things.
    0:26:56 And we’ve now got dozens of these tools
    0:26:58 and it’s become complicated enough
    0:27:00 that if you’re a scientist at recursion,
    0:27:03 you can’t keep up with all the versions of all the tools.
    0:27:05 It’s just, you look on your iPhone,
    0:27:07 there’s tens of thousands of apps.
    0:27:09 It’s hard to know how to use all of them.
    0:27:11 Same kind of thing is happening here in science.
    0:27:15 And so what we did was build an LLM and tuned it
    0:27:18 to actually interact with the APIs for all of these tools
    0:27:22 and to have a sense of when to use different tools
    0:27:23 based on natural language.
    0:27:25 So I can go in and I can say,
    0:27:29 “Give me five novel targets in non-small cell lung cancer.”
    0:27:31 And I just type that in and hit enter.
    0:27:33 And then the LLM knows which of the software tools
    0:27:36 we’ve built at recursion can go look at our data,
    0:27:39 at public data, can look for like arbitrage
    0:27:41 between those data sets and it can just surface back to you
    0:27:44 some insights about what targets you might want to go after.
    0:27:46 And then you can say, “Design a drug
    0:27:48 that would inhibit one of these targets.”
    0:27:51 And it’ll use Gen AI and the protein structure
    0:27:55 that AlphaFold or others have predicted for that target.
    0:27:58 And then it’ll help design a molecule that can bind
    0:27:59 into the binding sites of that target
    0:28:01 and we would predict inhibit it.
    0:28:02 And you can do all of this with natural language.
    0:28:06 And then you can say, “Design and execute an experiment
    0:28:08 to validate this interaction.”
    0:28:12 And then it can order the chemical from our suppliers.
    0:28:13 It can design an experiment
    0:28:15 that we can run in our automated wet lab.
    0:28:16 From a security perspective,
    0:28:18 we make a human approve that experiment
    0:28:19 because we want to make sure
    0:28:21 that we’re not running any bad experiments
    0:28:22 that could do something bad.
    0:28:24 And also you don’t want to accidentally run
    0:28:28 like a $6 million experiment because nobody approved it.
    0:28:29 So we have a human in the loop.
    0:28:31 But if somebody approves that, you can go run the experiment.
    0:28:34 And so this for us is, I think this is a lot like
    0:28:37 the late ’70s and early ’80s in personal computing
    0:28:39 where you moved from like the Apple One,
    0:28:41 where you had to be this expert user,
    0:28:43 you had to know how to solder and all this stuff,
    0:28:46 to then with the LISA, the GUI,
    0:28:47 you could actually start
    0:28:49 to have this democratization of these tools.
    0:28:51 And we think the same thing is happening now,
    0:28:53 but instead of a graphical user interface,
    0:28:56 it’s like a discovery user interface.
    0:28:58 And low is our take on this.
    0:29:00 There’s a big pharma company, GSK,
    0:29:02 that’s building one of these.
    0:29:04 There’s a couple other startups building these.
    0:29:06 But ultimately we think these kinds of tools
    0:29:09 are gonna mean that even if you don’t have 30 years
    0:29:12 of experience in chemistry in the biopharm industry,
    0:29:13 you’re like fresh out of school,
    0:29:16 you’ll still be able to make bigger contributions faster
    0:29:17 with this kind of approach.
    0:29:18 That’s fantastic.
    0:29:21 And you kind of just spoke to my next question,
    0:29:23 but I’ll ask in any way.
    0:29:25 Jensen said something in,
    0:29:28 recently it was a couple of months ago, I think now,
    0:29:30 in an interview where somebody asked about,
    0:29:32 the future of the computer science field.
    0:29:35 And he said something along the lines of,
    0:29:39 my advice is go study a field
    0:29:40 that you’re interested in,
    0:29:42 develop your domain expertise,
    0:29:45 because we’re already on this path
    0:29:48 and we’re getting to a spot where you’re not,
    0:29:51 and I’m paraphrasing you to use these exact words,
    0:29:53 you’re not necessarily gonna have to learn
    0:29:57 to write Python scripts or learn R to wrangle data.
    0:29:59 You’re gonna use natural language prompts.
    0:30:01 And then your domain expertise
    0:30:03 is really what’s gonna become valuable
    0:30:07 ’cause you’ll be able to work with the AI systems,
    0:30:08 vet their output, et cetera, et cetera.
    0:30:12 It sounds like that’s kind of where recursion’s at
    0:30:14 to some extent anyway.
    0:30:16 What’s your take kind of on that trajectory
    0:30:19 and maybe just expound a little on what you just said
    0:30:23 about even new graduates being able to contribute more
    0:30:26 to the field because of the natural language prompting?
    0:30:27 Yeah, I know, I think he’s exactly right.
    0:30:29 I think he was on stage with us
    0:30:31 at an event we hosted in January
    0:30:32 where he said something similar.
    0:30:33 Okay, yeah.
    0:30:35 And what we talked about was how important
    0:30:37 a classical education will be
    0:30:39 in learning how to interpret problems,
    0:30:40 identify the right problem,
    0:30:43 and then how to ask and answer questions
    0:30:44 about that problem.
    0:30:46 That’s what’ll matter, it won’t be coding
    0:30:47 because everything will be natural language.
    0:30:48 So I agree with him.
    0:30:50 And so what we’re looking for now
    0:30:51 are people who are really good
    0:30:53 at operating at the interface.
    0:30:56 We actually don’t need somebody that’s memorized
    0:30:58 the entire molecular cellular biology textbook.
    0:31:00 Just like we don’t need doctors anymore
    0:31:04 to memorize every single possible disease,
    0:31:05 we are gonna have these tools
    0:31:07 that mean you can type in,
    0:31:08 here’s the symptoms the patient has
    0:31:10 or here’s what I’m seeing in the data,
    0:31:11 and then those tools are gonna help
    0:31:13 pull out all of that deep information
    0:31:15 that nobody should have to memorize.
    0:31:16 And what’s gonna be critical is somebody
    0:31:18 who can take that information
    0:31:19 and say, okay, what do I do with this?
    0:31:20 Like what’s the next step?
    0:31:22 What’s the killer experiment that I can go run?
    0:31:24 Or like, how would I treat this patient?
    0:31:26 And so yeah, we’re gonna move to a place
    0:31:27 where people who are good
    0:31:30 at integrating lots of different ideas and data,
    0:31:33 those are the people that we’re looking for.
    0:31:36 We’re looking for not just the biology PhD,
    0:31:38 but the biology PhD who loves using
    0:31:39 all the different new AI tools
    0:31:41 or the computational biologists.
    0:31:43 People were really working at those interfaces.
    0:31:44 – Right, right.
    0:31:46 So I think you mentioned at the top,
    0:31:49 recursion has created some drugs that are in trials now.
    0:31:52 – Yeah, we’ve got five programs that are in clinical trials
    0:31:54 and a few more on the way to the clinic as well.
    0:31:57 – Okay, so I’m asking that kind of as context
    0:32:01 for what the future holds, both for recursion
    0:32:03 and then also sort of for the industry
    0:32:05 for the field more broadly.
    0:32:07 Is the plan on recursion’s end
    0:32:10 to keep developing more drugs,
    0:32:12 getting more actual solutions to put it that way
    0:32:14 out into trials?
    0:32:16 What’s sort of the, I don’t know,
    0:32:19 the MO for recursion over the coming few years?
    0:32:20 – It’s a great question.
    0:32:23 I mean, at the end of the day, what matters
    0:32:25 is you get a medicine to a patient
    0:32:26 and it makes the patient healthy again, right?
    0:32:28 So somebody’s gotta develop a product.
    0:32:30 When we started recursion,
    0:32:31 we believed we would build all these tools
    0:32:34 to help people identify what products to build
    0:32:36 and that they would then go build the products.
    0:32:38 What we were surprised by, what we got wrong
    0:32:40 was the reticence to this industry,
    0:32:42 to these new technology tools.
    0:32:43 And I think it’s a combination
    0:32:47 of the regulatory environment with the FDA and EMA.
    0:32:49 It’s a combination of that with just the general
    0:32:51 conservative nature of an industry that spends,
    0:32:56 it’s $2.6 billion of invested R&D
    0:32:59 per new drug approval every year in our industry.
    0:33:02 So like, and remember, that cost is that high
    0:33:05 because 90% of the drugs that people take in the clinic fail.
    0:33:06 So like, it doesn’t cost that much
    0:33:08 to develop the drug that succeeds.
    0:33:11 It’s just that you’ve gotta, you know,
    0:33:12 spend all the money on all the failures.
    0:33:14 And so the industry’s been conservative
    0:33:15 and we kind of felt like
    0:33:18 they’re not really taking this up as fast as we would hope.
    0:33:21 Do we want to let these drugs we believe in die?
    0:33:22 – Right.
    0:33:23 – Or do we want to let them see the light of day?
    0:33:25 And so ultimately, the biggest pivot we’ve made
    0:33:27 as a company is we now have our own drugs
    0:33:28 and that’s required more capital.
    0:33:30 We’ve had to build new parts of our team
    0:33:32 and kind of build the culture in that direction
    0:33:35 towards clinical development and interacting with the FDA.
    0:33:38 But sometimes you got to go vertical, right?
    0:33:39 Sometimes you got to go vertical
    0:33:41 if you really believe in something.
    0:33:44 And, you know, a great example of this is like Tesla
    0:33:46 building the supercharging network.
    0:33:49 Like, I guarantee you they didn’t want to build a supercharging.
    0:33:51 They’re like, of course gas stations are gonna see
    0:33:52 in the future that people will use electronics
    0:33:54 and they’ll put in superchargers.
    0:33:56 No, nobody was read as this.
    0:33:58 So they had to build it and that kind of sucked.
    0:34:00 I’m sure we’ve got to build it too.
    0:34:01 But at the end of the day,
    0:34:05 if we can get medicines to patients, we’ll be happy.
    0:34:07 – So before we wrap up here, Chris,
    0:34:09 you mentioned, you know, a couple of times
    0:34:11 and just now sort of the shift in the field
    0:34:14 and kind of the reticence, you know, dating back
    0:34:16 to when you were first doing your graduate work,
    0:34:18 even up until now and the reason you had to go vertical
    0:34:21 and start actually producing the drugs yourself.
    0:34:23 People are resistant to change.
    0:34:26 Technology takes time to get used to sort of culturally,
    0:34:30 globally we’re almost certainly in the opening stages
    0:34:34 of this big industrial revolution, you know,
    0:34:36 spurred by automation and AI and everything.
    0:34:39 Along that sort of cultural front,
    0:34:42 have there been kind of big, I mean, is that a big thing?
    0:34:46 Is that kind of the, you know, the data, the compute,
    0:34:48 all of that stuff, but kind of culturally are you,
    0:34:51 do you feel like you’re building kind of a different way
    0:34:54 to do things and are there bumps in the road along,
    0:34:55 you know, that come with that?
    0:34:58 – Yeah, absolutely, I would say that the data
    0:34:59 that compute those are moats for us,
    0:35:01 but our biggest moat is our culture.
    0:35:04 And we’ve had to take software engineers, data scientists,
    0:35:06 biologists, chemists, drug developers
    0:35:09 and supporting functions and have all of these different
    0:35:13 folks working in these deep technical domains come together
    0:35:15 and it’s basically meant that we’ve had to create
    0:35:17 a new language, a new language where all of these people
    0:35:20 can build in the same environment,
    0:35:22 both literally and figuratively.
    0:35:25 And so I think that we had a C level exec
    0:35:27 from a big pharma company here a few years ago.
    0:35:29 And on his way out the door, he said,
    0:35:31 we could never replicate this culture.
    0:35:33 That’s your biggest competitive advantage.
    0:35:35 Everything else with enough money we could do,
    0:35:36 but that we couldn’t do.
    0:35:38 And I think that’s probably true.
    0:35:41 And it’s, this is why sometimes those up and coming
    0:35:44 young companies out, end up out innovating, you know,
    0:35:46 some of the bigger ones is that it’s about the culture
    0:35:47 as much as anything else.
    0:35:48 – No, absolutely.
    0:35:50 How big is recursion now, how many people?
    0:35:52 – We’re about 550 people now.
    0:35:54 – Wow, amazing.
    0:35:57 I mean, it’s relatively small to what you’re doing.
    0:35:59 I mean, maybe it’s very, very small to what you’re doing,
    0:36:01 but that’s just amazing.
    0:36:02 For folks who would like to learn more,
    0:36:04 we covered a lot of ground, we could do an hour
    0:36:07 just on how you built BioHive too, obviously.
    0:36:10 But for folks who want more of the recursion story,
    0:36:13 want to dive into some more of what you’re working
    0:36:16 on the specifics, maybe even are looking for, you know,
    0:36:18 a new role, a great culture to join.
    0:36:20 Where can they go to find out more?
    0:36:21 – Yeah, go to recursion.com.
    0:36:23 You can find out about careers.
    0:36:25 You can read all of our papers.
    0:36:29 And you can also go to rxrx.ai,
    0:36:32 which is where we have lots of large data sets.
    0:36:33 If you’re a data scientist and you want to play
    0:36:36 with some biology, you can go download them there.
    0:36:37 – Fantastic.
    0:36:40 Chris Gibson, this was a pleasure.
    0:36:41 I learned a lot that’s going to leave me
    0:36:43 with many more questions as I process it.
    0:36:46 So we may have to do this again,
    0:36:48 but appreciate you taking the time to come on the podcast,
    0:36:49 tell everybody about the work you’re doing.
    0:36:53 And, you know, it feels like an understatement
    0:36:56 and a cliche to say, but you’re working on solutions
    0:36:58 that can literally change people’s lives
    0:36:59 through better health.
    0:37:02 So all the best of luck to you and your teams.
    0:37:04 – Thanks Noah, really appreciate it.
    0:37:06 (upbeat music)
    0:37:09 (upbeat music)
    0:37:10 .
    0:37:12 (upbeat music)
    0:37:15 (upbeat music)
    0:37:17 (upbeat music)
    0:37:20 (upbeat music)
    0:37:23 (upbeat music)
    0:37:25 (upbeat music)
    0:37:28 (upbeat music)
    0:37:30 (upbeat music)
    0:37:33 (upbeat music)
    0:37:36 (upbeat music)
    0:37:38 (upbeat music)
    0:37:41 (upbeat music)
    0:37:43 (upbeat music)
    0:37:46 (upbeat music)
    0:37:48 (upbeat music)
    0:37:51 (upbeat music)
    0:37:53 you
    0:38:03 [BLANK_AUDIO]

    Techbio is a field combining data, technology and biology to enhance scientific processes — and AI has the potential to supercharge the biopharmaceutical industry further. In this episode of NVIDIA’s AI Podcast, host Noah Kravitz speaks with Chris Gibson, cofounder and CEO of Recursion, about how the company uses AI and machine learning to accelerate drug discovery and development at scale. Tune in to hear Gibson discuss how AI is transforming the biopharmaceutical industry by increasing efficiency and lowering discovery costs.