Author: The Next Wave – AI and The Future of Technology

  • Grok 3 vs Claude 3.7 vs GPT-4.5: Which Update is The Best?

    AI transcript
    0:00:01 [MUSIC PLAYING]
    0:00:03 Hey, welcome back to the Next Wave Podcast.
    0:00:04 I’m Matt Wolf.
    0:00:05 I’m here with Nathan Lanz.
    0:00:08 And there has been an absolute ton of AI news
    0:00:10 that came out recently, especially in the world
    0:00:11 of large language models.
    0:00:12 We got GROC 3.
    0:00:14 We got GPT 4.5.
    0:00:16 We got Claude 3.7.
    0:00:21 Just so many new big foundation models have been released.
    0:00:23 And so for this episode, we wanted
    0:00:26 to deep dive into what each one is good at
    0:00:28 and what each one is not good at.
    0:00:29 So for that, we brought on our good friend
    0:00:33 Matthew Berman, who is probably the best person we know
    0:00:35 to really compare them all.
    0:00:37 Because he deep dives and tests every single one
    0:00:40 of these models way deeper than we test them.
    0:00:44 So let’s go ahead and just dive right in with Matthew Berman.
    0:00:46 Thanks so much for joining us, Matthew.
    0:00:48 It’s great to have you back on the show.
    0:00:48 Thanks for having me.
    0:00:51 I’ve been telling people for a long time–
    0:00:53 I think I even mentioned this to you last time we had you
    0:00:53 on the show–
    0:00:55 that when it comes to large language models
    0:00:59 and trying to compare them and talk about which model is best
    0:01:01 at this and which model is best at that.
    0:01:03 I don’t even do that on my YouTube channel anymore.
    0:01:05 I just point people to your channel.
    0:01:07 I’m like, yeah, Matthew’s going to test this.
    0:01:09 And he’s going to tell you which models do what better
    0:01:10 than others.
    0:01:12 So you’re like my go-to now when it comes
    0:01:15 to comparing large language models.
    0:01:17 I appreciate that.
    0:01:17 That’s awesome.
    0:01:18 Thanks for having me again.
    0:01:21 I had a great time last time, so I’m excited to chat again.
    0:01:22 Yeah, likewise.
    0:01:23 So this will be fun.
    0:01:26 I’m trying to figure out where the best place to dig in
    0:01:28 is because so much has come out.
    0:01:29 Here’s kind of the timeline of events
    0:01:31 that I feel are the important events, right?
    0:01:32 We got Grock 3.
    0:01:36 And then a few days later, we got Claude Sonnet 3.7.
    0:01:40 And then a few days after that, we got GPT 4.5, which
    0:01:43 actually happens to be the day that we’re recording this episode
    0:01:45 is the day that GPT 4.5 came out.
    0:01:49 So we’ve all had that news in our heads for three hours now.
    0:01:51 And that’s about it.
    0:01:53 But maybe we start with Grock.
    0:01:56 Matt, what have your thoughts been on Grock 3 so far?
    0:01:57 Like how much have you played with it?
    0:02:00 And what have you found it’s like really good at so far?
    0:02:03 Yeah, so if I could show you my bookmarks bar in Chrome,
    0:02:07 you will see that Grock now has a prominent placement
    0:02:10 right next to chat GPT and right next to perplexity.
    0:02:12 So the answer is I use it a lot.
    0:02:16 And it has really become my go-to large language model.
    0:02:16 Oh, really?
    0:02:17 Yeah.
    0:02:17 Same here.
    0:02:20 Look, I have been a pretty die-hard chat GPT user.
    0:02:21 There’s really two reasons.
    0:02:23 Number one is speed, right?
    0:02:26 I think speed of these models, speed of the response
    0:02:29 is really underappreciated by a lot of people.
    0:02:32 But it’s the same reason why you converted a higher rate
    0:02:34 when a web page loads faster.
    0:02:37 It just builds trust and you just get the answer more quickly.
    0:02:39 And then it’s also the real time information.
    0:02:43 Having access to all of the news on X in real time
    0:02:45 is such a killer feature.
    0:02:47 And of course, it’s a good model.
    0:02:49 So I think those two factors, plus it just
    0:02:53 being a fantastic model, that has made it my go-to model.
    0:02:54 But that might change now that 4.5 is out,
    0:02:55 but we’ll get there.
    0:02:56 Yeah, yeah, yeah.
    0:02:57 One thing I would say, too, is they
    0:02:59 say that they’re really improving the model really fast.
    0:03:02 And I believe it, because their team actually reached out to me.
    0:03:03 And they’re like, what’s your feedback?
    0:03:05 And I’ve been going back and forth with them,
    0:03:07 giving them my thoughts on what they should be doing.
    0:03:08 Oh, that’s cool.
    0:03:08 Yeah.
    0:03:10 And sorry if Aravind or anyone from Perplexity
    0:03:12 hears this, but I’m like, this has replaced Perplexity for me.
    0:03:14 And I think you guys should double down on that.
    0:03:15 That’s what I was telling them.
    0:03:17 I think you guys should be replacing Google,
    0:03:19 replacing Wikipedia, replacing Perplexity.
    0:03:21 Any question I have, I should be going to GROC.
    0:03:23 They’re combining search and X data.
    0:03:25 Like, no one has that, and no one ever will have that.
    0:03:26 Yeah.
    0:03:29 Nathan, does that make you a high-taste tester?
    0:03:30 Oh, yeah.
    0:03:32 Because like, I’m certainly not.
    0:03:33 I get no previews at all.
    0:03:34 Yeah.
    0:03:35 I don’t either.
    0:03:37 I don’t seem to get any early access
    0:03:41 to any of the anthropic models, the open AI models,
    0:03:42 the X models, none of them.
    0:03:47 So yeah, I think Nathan probably has the best ins right now.
    0:03:49 I was a low-level member of a lot of the different Silicon
    0:03:51 Valley mafias.
    0:03:52 I was not a high-level member, but I
    0:03:55 was the low level of several of them.
    0:03:59 The gaming one, the social media one, the crypto one,
    0:04:02 the Taiwanese mafia.
    0:04:04 I was kind of a member of all of those.
    0:04:06 And some of the YC people, as well, I know a lot of them.
    0:04:07 A lot of them, too.
    0:04:10 You know, I personally think that GROC’s biggest hurdle
    0:04:11 is just Elon.
    0:04:14 I’m sure you’ve seen this, Matthew, on your YouTube channel,
    0:04:16 because you’ve made a couple videos about GROC now.
    0:04:18 When GROC came out, I made a news video about it.
    0:04:21 And overwhelmingly, all of the comments about GROC are like,
    0:04:24 I’m never going to touch anything Elon makes.
    0:04:27 And so I almost think that, like, so many people
    0:04:29 are throwing the baby out with the bathwater
    0:04:32 when it comes to actually using and trying GROC
    0:04:35 just because it’s Elon’s name attached to it.
    0:04:39 Yeah, so whatever I post anything about GROC on X,
    0:04:43 if it’s negative about GROC or Elon, I get flamed.
    0:04:47 If it’s positive, everybody cheers it, shares it,
    0:04:48 everything.
    0:04:50 The opposite is true on YouTube.
    0:04:53 I made a video about GROC 3 being really good,
    0:04:54 because it is.
    0:04:57 So many people commented, they won’t touch it just like you,
    0:04:58 won’t touch it.
    0:05:03 Elon’s AI model, bias, conservative, right wing.
    0:05:04 I’m apolitic.
    0:05:06 I’m trying to stay out of it as much as I can in the videos.
    0:05:09 But anytime I mention anything having
    0:05:16 to do with Elon X, GROC, I get the meanest comments on YouTube.
    0:05:17 It’s wild.
    0:05:20 I think so many people are just going to not experience
    0:05:24 probably what is the best model available right now
    0:05:27 just because of the Elon factor.
    0:05:29 Now, when it comes to code, I don’t necessarily
    0:05:31 think GROC is the best.
    0:05:31 I don’t know.
    0:05:34 I actually honestly haven’t tested GROC with code
    0:05:36 because I mostly do code with cursor.
    0:05:39 And I don’t believe GROC 3 has pushed its API out yet.
    0:05:40 I don’t think.
    0:05:41 I don’t think so.
    0:05:43 So I’ve been mostly playing with 3.7.
    0:05:44 I’ve tested it.
    0:05:46 I’ve tested it a little bit.
    0:05:46 It’s weird.
    0:05:48 It’s not very consistent.
    0:05:49 There’s sometimes where it’s amazing at code.
    0:05:52 You’re like, oh, wow, that’s like a really creative solution.
    0:05:53 And it’s doing something better than Claude
    0:05:55 in some very limited circumstances.
    0:05:57 But then other times, it just won’t follow my instructions
    0:05:58 as well for code.
    0:06:01 In terms of reliability, when Claude 3.7 came out,
    0:06:03 that’s by far the best now for coding.
    0:06:05 But it feels like there is something there.
    0:06:07 I wouldn’t discount them forever.
    0:06:08 They could improve that.
    0:06:10 And all of a sudden, GROC 3 could be the best at coding.
    0:06:11 Right.
    0:06:12 Matthew, I’m curious.
    0:06:14 So when it comes to GROC 3, I know
    0:06:17 you have your own internal benchmarks
    0:06:20 that you’ve been using on some of these tools.
    0:06:22 What have you found GROC 3 is really good at?
    0:06:25 And are there any things GROC 3 is just not
    0:06:26 going to be your go-to for?
    0:06:28 Well, let me say something first.
    0:06:30 I had to throw my benchmarks out.
    0:06:32 Because they were completely saturated.
    0:06:35 They were absolutely annihilated by every single model
    0:06:36 that comes out nowadays.
    0:06:37 So I threw them out.
    0:06:42 I’m currently in the process of creating a new set of benchmarks
    0:06:43 and questions.
    0:06:46 GROC, again, the thing I go to it for
    0:06:49 is real-time information as quickly as I need it.
    0:06:51 And that it’s awesome at.
    0:06:53 I’ve tested it on some other things,
    0:06:56 like quick coding challenges, some math challenges.
    0:06:57 And it does really well.
    0:07:00 But those aren’t the everyday use cases for me.
    0:07:02 So there’s only so much I can test with it.
    0:07:05 I’ll just prompt it with one of my benchmark questions,
    0:07:05 see if it’s right.
    0:07:07 It’s like, OK, yeah, it’s right or no, it’s wrong.
    0:07:09 But overall, what I care about and I
    0:07:12 think what most people care about is day-to-day usage.
    0:07:15 Is it going to solve my problems?
    0:07:16 And it does.
    0:07:17 GROC3 is great at that.
    0:07:21 And typically, I was using GPT-40.
    0:07:24 And now GROC3 with thinking kind of replaces that,
    0:07:26 although I don’t really need the thinking.
    0:07:28 Yeah, it’s just really good at the day-to-day stuff
    0:07:29 that I would use it for.
    0:07:33 So I’m kind of split between GROC3, perplexity, and chat
    0:07:34 GPT.
    0:07:35 And now 4.5 came out.
    0:07:37 I would say I tested it right before we got on here.
    0:07:39 And I would say, OK, 4.5 has a better
    0:07:41 vibe for general chat.
    0:07:42 But it’s slower.
    0:07:43 Oh, it’s about the vibes.
    0:07:44 But when you get the response back,
    0:07:46 the liked response is better, actually.
    0:07:47 So that’s impressive.
    0:07:50 It’s writing is better, but it’s slower.
    0:07:51 So there’s a huge trade up there.
    0:07:53 It feels like 4.5 is going to serve
    0:07:57 something maybe for creative writing or creative work.
    0:07:59 It’s probably the best for now.
    0:08:01 But it is interesting that the cost is so much higher.
    0:08:02 Really?
    0:08:04 Yeah, I think it was 70 per million input tokens,
    0:08:06 half of that for cashed input.
    0:08:08 It’s funny, Nathan, that you bring up writing.
    0:08:12 That is honestly something that I want to use AI for much more
    0:08:15 often than I do, but it is so bad.
    0:08:16 It’s so bad at writing.
    0:08:20 And I haven’t had enough time to test 4.5 for writing,
    0:08:21 but I really hope that it’s good.
    0:08:22 It’s better.
    0:08:25 Because then that’s going to be able to help me with a lot
    0:08:26 of things that I do day to day.
    0:08:28 I don’t script my videos, but sometimes I
    0:08:31 want help writing the bullet points for them.
    0:08:34 Or sometimes I want a tweet thread,
    0:08:35 the initial drafts created for me.
    0:08:38 I gave it probably, I don’t know,
    0:08:40 over 100 pages of different notes about my game,
    0:08:43 including the story and things like that and the game mechanics.
    0:08:46 And it took a long time to respond, actually, like a very long.
    0:08:47 I was kind of surprised.
    0:08:49 This feels like a one pro when you hit into a lot of stuff.
    0:08:50 4.5?
    0:08:50 4.5.
    0:08:52 Took a long time to respond.
    0:08:52 I was surprised.
    0:08:53 Like very long.
    0:08:54 It was very slow.
    0:08:56 Like three minutes to respond or something to all that.
    0:08:59 But then its notes on the story were perfect.
    0:09:02 It gave me amazing critiques of like, I love this part.
    0:09:06 And it even had little emojis having color coatings.
    0:09:10 It had green, orange, blue, and red and different ones.
    0:09:11 It was like, like, green’s good.
    0:09:12 Love this part.
    0:09:14 This part probably could be tweaked.
    0:09:15 This part don’t like it.
    0:09:16 Here’s why.
    0:09:17 And these parts are interesting.
    0:09:19 Maybe you keep them, maybe you don’t.
    0:09:20 The feedback was good.
    0:09:21 Like it was solid.
    0:09:22 So I do want to test it more for writing.
    0:09:23 But my first impression is,
    0:09:26 yeah, it is probably the best model for writing now.
    0:09:27 That’s great.
    0:09:27 Yeah.
    0:09:29 Well, going back to Grock for a second,
    0:09:31 have you guys played with the voice mode yet?
    0:09:34 I’ve used advanced voice mode a little bit.
    0:09:37 I don’t find it to be like super useful throughout my day.
    0:09:39 If I’m driving, if I’m walking,
    0:09:41 if there’s just not a screen in front of me,
    0:09:42 maybe I’ll use it.
    0:09:45 But what I found specifically with ChatGPT advanced voice mode
    0:09:49 is you’ll ask a question and there’s that delay.
    0:09:51 And then sometimes it repeats it and then stops
    0:09:53 and then repeats it again.
    0:09:56 It’s just such a high friction experience.
    0:09:58 It’s not great yet.
    0:10:00 I do use voice on perplexity.
    0:10:01 I’m a big baseball fan.
    0:10:03 Spring Training just started with baseball.
    0:10:04 They’ve introduced some like new rules
    0:10:06 that I didn’t realize existed.
    0:10:08 I was watching one of the Spring Training games.
    0:10:09 I opened up perplexity and I’m like,
    0:10:11 “Hey, is there a new rule that I didn’t know about?”
    0:10:13 And I just will talk to perplexity and ask it questions
    0:10:15 and it will like do all the research,
    0:10:17 figure out the new rules that are going on in baseball
    0:10:19 and then give them back to me.
    0:10:21 And I found that pretty helpful.
    0:10:24 I still find myself typing my questions more often
    0:10:25 than speaking them.
    0:10:27 But every once in a while, I’ll be feeling lazy.
    0:10:29 I’ll hit the voice button and just ask my question.
    0:10:30 I find that helpful.
    0:10:34 I know Nathan, you kind of almost use some of the voice modes,
    0:10:36 especially advanced voice and ChatGPT
    0:10:38 for like journaling, right?
    0:10:40 I’m a big believer in it long-term,
    0:10:41 but I have been kind of disappointed
    0:10:42 like in advanced voice mode.
    0:10:44 Like the demo they showed seemed amazing
    0:10:45 and it seemed like they removed
    0:10:46 so many different parts of it.
    0:10:47 And some of the things they demoed
    0:10:48 when you tried them in real life,
    0:10:51 they don’t work as well as in the demos.
    0:10:52 My wife’s Japanese.
    0:10:53 My Japanese is getting better,
    0:10:54 so I don’t have to use it as much,
    0:10:56 but for like really hard topics,
    0:10:58 we haven’t tried to use it to do translation.
    0:11:01 And it just gets confused so easily.
    0:11:02 Like as soon as you go from one language to the other,
    0:11:04 like sometimes it’ll translate it properly.
    0:11:06 So like, “Okay, cool, it worked in that use case.”
    0:11:07 But okay, when then she talks back to me,
    0:11:09 then sometimes it just totally gets confused.
    0:11:10 And instead of translating,
    0:11:13 sometimes it’ll like start talking versus translating.
    0:11:15 It’ll start saying its own stuff
    0:11:16 versus doing a translation.
    0:11:18 And it’s like, “Okay, that’s super annoying.”
    0:11:19 And then we just turn it off
    0:11:21 every single time that’s ever happened.
    0:11:23 – Yeah, I know Sam Holtman’s even talked about,
    0:11:25 he wants the AIs in the future
    0:11:28 to be more in line with your own beliefs, right?
    0:11:31 So like, it knows your religious beliefs,
    0:11:32 your political beliefs.
    0:11:34 – So bring your own bias.
    0:11:35 – Yeah, your own bias.
    0:11:36 It will actually learn your bias
    0:11:38 and sort of lean into your bias more
    0:11:40 to give you more of what you want.
    0:11:42 And I think that’s a really, really scary thing.
    0:11:44 There was a little bit of word
    0:11:45 about this happening on like Facebook, right?
    0:11:49 Where Facebook was using like AI bots in the feed
    0:11:52 that people didn’t even realize were AI bots.
    0:11:54 And so people would post stuff on Facebook.
    0:11:56 They would get like a bunch of responses
    0:11:57 from these AI bots,
    0:11:59 not even realize that they’re AI.
    0:12:00 And they’re going, “Oh, cool,
    0:12:02 I get great engagement on Facebook.”
    0:12:04 And it keeps on bringing them back to Facebook
    0:12:06 over and over and over again,
    0:12:08 because Facebook’s where they get engagement
    0:12:09 and they’re not even realizing
    0:12:10 that they’re talking to AIs.
    0:12:12 I think that’s going to be a bigger
    0:12:13 and bigger problem as well,
    0:12:17 where social media is all about getting dopamine hits, right?
    0:12:20 We post our tweets because we want to get those likes.
    0:12:21 We want to get those retweets.
    0:12:22 We want to get those comments.
    0:12:23 Every time we see one of those,
    0:12:25 we get a little dopamine hit
    0:12:27 and we keep on coming back for more.
    0:12:30 Well, if AI gets really, really, really good
    0:12:33 at giving us those dopamine hits every time we want them,
    0:12:36 we’re going to go to wherever we get the most dopamine hits
    0:12:38 at the highest frequency.
    0:12:41 I think that’s what’s really sort of worries me
    0:12:41 about the future.
    0:12:44 And it ties into like the whole population collapse as well.
    0:12:46 I think it gets to that point
    0:12:47 where people communicate with other humans
    0:12:49 less and less and less and less,
    0:12:51 because they’re getting their dopamine hits
    0:12:53 from fake people on social media.
    0:12:55 They’re getting their conversational needs
    0:12:57 met by unhinged voice chats
    0:13:00 that have the same bias as me.
    0:13:04 They love Trump and their, you know, the MAGA girl.
    0:13:05 And I can talk to the MAGA girl
    0:13:07 who has the same belief system that I have
    0:13:09 or whatever, I’m not saying that’s my belief system.
    0:13:11 I’m just, you know, for example.
    0:13:13 And so it’s very, very concerning.
    0:13:14 Like, I think you and me, Matthew,
    0:13:16 are pretty much on the same page
    0:13:19 where we generally lean optimistic on this stuff,
    0:13:21 but there’s still quite a few things
    0:13:23 that actually do scare me about this as well.
    0:13:26 I’m not like the whole accelerationist
    0:13:28 where I’m just like push forward as fast as possible.
    0:13:30 I’m like, maybe there’s some things
    0:13:32 we shouldn’t push forward as fast as possible on.
    0:13:34 And that’s definitely one of those areas.
    0:13:37 (upbeat music)
    0:13:38 – Hey, we’ll be right back to the show.
    0:13:40 But first I want to have another podcast
    0:13:41 I know you’re gonna love.
    0:13:43 It’s called Entrepreneurs on Fire
    0:13:45 and it’s hosted by John Lee Dumas
    0:13:47 available now on the HubSpot Podcast Network.
    0:13:49 Entrepreneurs on Fire stokes inspiration
    0:13:53 and shares strategies to fire up your entrepreneurial journey
    0:13:54 and create the life you’ve always dreamed of.
    0:13:57 The show is jam packed with unlimited energy,
    0:13:58 value and consistency.
    0:14:00 And really, you know, if you like fast paced
    0:14:02 and packed with value stories
    0:14:03 and you love entrepreneurship,
    0:14:05 this is the show for you.
    0:14:07 And recently they had a great episode
    0:14:10 about how women are taking over remote sales
    0:14:11 with Brooke Triplett.
    0:14:13 It was a fantastic episode.
    0:14:14 I learned a ton.
    0:14:15 I highly suggest you check out the show.
    0:14:17 So listen to Entrepreneurs on Fire
    0:14:19 wherever you get your podcasts.
    0:14:22 (upbeat music)
    0:14:24 – Yeah, you know, Elon had said with Grock,
    0:14:27 which I find kind of promising hopefully is that, you know,
    0:14:29 they want to be maximum truth seeking,
    0:14:30 which sounds a lot better to me
    0:14:32 than when I’m hearing from open AI
    0:14:33 of like bring your own bias.
    0:14:35 Like we’re gonna allow you to pick your bias
    0:14:38 and we’re just gonna serve you up information based on that
    0:14:40 because all truth is subjective or whatever.
    0:14:41 But I worried that, you know,
    0:14:45 as obviously Elon has his own bias now as well, you know,
    0:14:47 and, you know, I’m probably more right wing
    0:14:48 than anyone on this podcast right now.
    0:14:50 I’m not super right wing, but, you know,
    0:14:52 after living in San Francisco for a long time,
    0:14:53 that kind of did that to me.
    0:14:55 And I do worry that he’ll go too far with it.
    0:14:58 Grock will become like a right wing AI.
    0:15:00 I do think it needs to be unbiased.
    0:15:02 I’m not sure how you do that
    0:15:03 ’cause there’s bias in everything.
    0:15:04 Like if you pull up facts online
    0:15:06 or you pull up something from the Wall Street Journal
    0:15:09 or CNN or whatever, there’s bias in all of this.
    0:15:11 And so since all the models are trained on that,
    0:15:15 I’m just not sure how you get around the bias, you know?
    0:15:17 – By the way, this is the part where I start
    0:15:20 to get flabbed on Twitter and YouTube,
    0:15:22 but for completely opposite reasons,
    0:15:23 but I’m still gonna share anyways,
    0:15:26 this is not reflective of my political opinion
    0:15:29 or anything like that, but here’s a couple of things.
    0:15:34 Yeah, Elon has said, “Maximally truth seeking for Grock 3.”
    0:15:37 And great, like in theory, that makes a lot of sense.
    0:15:39 But how do you do that?
    0:15:43 Ultimately, you have to create systems
    0:15:46 and maybe you’re able to create them or maybe you’re not,
    0:15:48 but those systems are created by humans.
    0:15:51 The original training data is created by humans.
    0:15:55 The post-training techniques are created by humans.
    0:15:58 The reinforcement learning is also set up by humans.
    0:16:00 So there’s like a human in the loop.
    0:16:04 So what he claims to be completely maximally truth seeking
    0:16:06 is just what he believes.
    0:16:08 And maybe not necessarily him,
    0:16:09 but he’s trying to give a different perspective,
    0:16:10 which, you know, fine.
    0:16:13 But did you see what happened just a few days ago?
    0:16:17 Grock 3 was caught having custom instructions that said,
    0:16:20 don’t cite any sources that say Elon Musk
    0:16:23 and Donald Trump are spreaders of disinformation.
    0:16:25 And they were caught doing this.
    0:16:28 And so I reported on it and of course,
    0:16:30 flamed from both sides, but fine.
    0:16:33 But the point is, yeah, it’s as simple as somebody
    0:16:36 submits a PR and it has a little line
    0:16:38 in the system prompt saying, don’t do this.
    0:16:40 And then all of a sudden, fingers on the scale, right?
    0:16:42 So it’s very possible.
    0:16:44 It’s not only possible, it’s happening.
    0:16:47 And it doesn’t matter if it’s from OpenAI or Grock
    0:16:50 or Google or Anthropic, they all have bias.
    0:16:53 I can’t imagine a world in which the bias
    0:16:55 is removed completely, although I hope it does.
    0:16:58 I just, I don’t see a path towards that.
    0:17:00 – Yeah, I mean, I honestly don’t understand
    0:17:02 how it’s possible.
    0:17:04 Like you guys said, it’s all trained on data
    0:17:05 that was created by humans.
    0:17:09 It’s all basically a scraping of the entire internet,
    0:17:11 which just inherently has bias.
    0:17:14 Like I just, I don’t understand how it’s possible, honestly.
    0:17:16 And how do you determine what is the truth
    0:17:17 and what is not?
    0:17:18 – That’s the question.
    0:17:20 – I don’t know, maybe if the models get really smart
    0:17:22 and can do like real reasoning
    0:17:24 and that they can look at both sides of anything,
    0:17:25 they can kind of find the middle ground
    0:17:27 where there’s some truth, where there’s actual truth,
    0:17:30 hopefully, but then some people will perceive
    0:17:32 that real truth is not truth.
    0:17:34 It’s like, well, people will perceive it to have bias
    0:17:35 even if it ends up not having bias.
    0:17:38 – No, but I feel like you’re getting into like this world
    0:17:40 where now we’re letting AI decide what is
    0:17:42 and what isn’t like ethical, right?
    0:17:45 Like now we’re letting like AI sort of decide
    0:17:49 what is like philosophically correct and what is not.
    0:17:50 And to me, that seems weird
    0:17:53 to let machines do that for humans, you know?
    0:17:55 – Yeah, and, you know, it’s funny you say that,
    0:17:58 I mentioned the same thing and Dave Shapiro,
    0:18:01 another fellow YouTuber talking about AI said,
    0:18:04 you know, actually, I would rather just completely
    0:18:07 give it over to AI to make ethical decisions.
    0:18:10 I think that’s what he was saying, but it’s interesting.
    0:18:12 I can see both sides of it.
    0:18:15 If there was a system that was completely unbiased,
    0:18:18 assuming, right, yeah, okay, let them decide,
    0:18:20 but how do you make that, as you said?
    0:18:23 – I think you can’t allow AI to make that decision.
    0:18:24 If you allow AI to make that decision,
    0:18:26 I mean, that goes towards like eugenics
    0:18:27 and crazy things like that.
    0:18:30 Like you can’t allow AI to optimize based on
    0:18:33 human performance or some crazy metric like that.
    0:18:35 That would just lead to, you know, horrible things.
    0:18:37 I think I totally disagree with that.
    0:18:39 – Yeah, yeah.
    0:18:42 Let’s shift over to Claude, ’cause Claude 3.7 came out
    0:18:46 a couple of weeks ago and that one has proven to be
    0:18:48 really, really good at code, it seems, right?
    0:18:50 It seems like it didn’t make huge improvements
    0:18:54 in almost any other areas, but it got a lot better at code.
    0:18:56 From what I’ve seen so far, I mean, Matt,
    0:18:58 you might have some different experiences with it.
    0:19:00 I’m using cursor to write code
    0:19:04 and it’s definitely gotten better at coding for me.
    0:19:06 I’ve definitely noticed things that it would run
    0:19:08 into these like loops where it wouldn’t write the code,
    0:19:10 wouldn’t write the code, couldn’t figure out what
    0:19:14 the problem was, 3.7, one prompt was 3.7,
    0:19:16 now it figures out a problem that I had gone back
    0:19:18 and forth 10 times prior.
    0:19:21 So to me, it seems like it got a lot better at code,
    0:19:24 but from most other people’s experiences,
    0:19:26 it seems like it didn’t really improve
    0:19:29 in almost any other areas other than code.
    0:19:30 – Yeah, when I think about coding,
    0:19:31 I think about Claude, right?
    0:19:35 Claude 3.5 was the go-to model for a lot of coders
    0:19:37 using AI assistance to help them code.
    0:19:40 Yeah, cursor plus Claude, WinSurf plus Claude,
    0:19:41 that was the model.
    0:19:43 Now we had this huge upgrade and I agree,
    0:19:44 it is a huge upgrade.
    0:19:46 It also has the thinking capabilities.
    0:19:49 So I’ve been doing stuff, vibe coding, right?
    0:19:51 So I’m using some kind of IDE,
    0:19:55 whether cursor or WinSurf plus I’m using 3.7 thinking
    0:19:57 and it’s fantastic, right?
    0:19:58 It is fantastic.
    0:20:01 Now I’ll bring it back to what I said earlier,
    0:20:03 the real world use cases,
    0:20:05 the stuff that I want to use it for day to day,
    0:20:08 one is coding, but a lot of other things are not coding.
    0:20:11 And here’s the thing, it’s not that fast
    0:20:12 and it doesn’t have web search.
    0:20:14 It has no real time information.
    0:20:19 So it’s essentially unusable for me outside of coding.
    0:20:21 Although it’s fantastic at coding, I’ll give you that.
    0:20:23 – Well, the nice thing is if you are using something
    0:20:26 like cursor, cursor can actually do the web search for you
    0:20:28 and then give that additional context.
    0:20:31 So it’s almost like cursor will actually kind of do
    0:20:33 like the perplexity thing, right?
    0:20:35 Where if it needs to figure something out,
    0:20:37 cursor itself goes and does the search
    0:20:39 and then provides that information to Claude.
    0:20:41 At least when I’ve been using it,
    0:20:44 it seems to actually search the web when using cursor.
    0:20:48 – Yeah, and Claude 3.7 Sonnet is available in perplexity,
    0:20:49 I’ll just mention quickly.
    0:20:51 By the way, perplexity just adds
    0:20:53 the latest models all the time for free.
    0:20:56 And I’ve not been paid by them at all.
    0:20:58 I’m just a huge fan of their product.
    0:21:00 So yeah, so if you wanted to try any of these new models,
    0:21:03 you could go try it if you already have a perplexity account.
    0:21:05 – Yeah, and their deep research is really good.
    0:21:08 If we’re talking about good things about perplexity,
    0:21:10 if we’re praising perplexity,
    0:21:12 they just threw deep research in there as well,
    0:21:13 and it’s really good.
    0:21:15 – But yeah, web search is critical.
    0:21:17 – Yeah, I wanted to say about Claude Sonnet.
    0:21:19 Like my feeling is, you know, before it was released,
    0:21:21 Claude was my go-to for just like general chat
    0:21:23 for like discussing anything, right?
    0:21:25 Like even like my game design document,
    0:21:27 I would share that with Claude and that was my favorite.
    0:21:28 Then Grock replaced that,
    0:21:30 but Claude was pretty good at code.
    0:21:33 Like you said, most engineers were using Claude Sonnet,
    0:21:35 but not all, like I was using O1 Pro
    0:21:37 and actually it was better, but way slower, right?
    0:21:38 And you had to give it tons of context.
    0:21:41 You could use something like repo prompt or something else.
    0:21:42 But it seems like in this update,
    0:21:44 they really doubled down on being the best at coding.
    0:21:46 ‘Cause in terms of like general chat,
    0:21:47 I feel like it actually stepped backwards.
    0:21:49 Like when I use it now to chat with them,
    0:21:51 like it actually got worse in this release like that.
    0:21:53 I don’t like its responses as much.
    0:21:55 They seem less human like, but it’s way better at coding.
    0:21:57 So it feels like we’re starting to see
    0:21:59 that all the AI models are finding their own like specialties
    0:22:02 or at least that’s what Anthropics doing now with Claude.
    0:22:04 And I do kind of wonder if all the models
    0:22:05 are gonna have to do that.
    0:22:06 You know, when I talked to the people at XAI,
    0:22:09 I told them double down on having the best data.
    0:22:11 ‘Cause you got the real-time data with X,
    0:22:13 you got it with, you know, the search, double down on that.
    0:22:15 And I think you’ll probably see that
    0:22:16 where like these different models
    0:22:17 will be the best at a thing.
    0:22:21 But it seems like Chat2P is still trying to go more broad.
    0:22:23 There’s trying to be like the best overall model.
    0:22:26 And I’m kind of curious to see how all that ends up playing out.
    0:22:27 – Yeah.
    0:22:28 – It’s interesting, you know, quickly.
    0:22:31 So I just pulled up the Claude 3.7 Sonnet blog post,
    0:22:32 the announcement blog post.
    0:22:35 And everybody knows Claude is great at coding
    0:22:38 and they made this huge jump in coding.
    0:22:41 But if you read it, it says,
    0:22:42 in developing our reasoning models,
    0:22:44 we’ve optimized somewhat less for math
    0:22:46 and computer science competition problems.
    0:22:50 And instead shifted focus toward real-world tasks
    0:22:52 that better reflect how businesses actually use LLMs.
    0:22:54 You know, now that I’m reading that,
    0:22:57 maybe they met like these kind of benchmarky
    0:23:00 computer science problems versus real-world coding problems.
    0:23:02 But it just sounds like,
    0:23:05 hey, we’re not focused as much on math and coding anymore,
    0:23:07 more real-world stuff.
    0:23:09 But it still is fantastic at that.
    0:23:11 – Yeah.
    0:23:12 – Did you guys see that they’re also,
    0:23:15 it seems to be that they’re gonna be competing with cursor?
    0:23:18 – Oh yeah, ’cause they released that code feature, right?
    0:23:20 – Yeah, Claude code, all the top engineers,
    0:23:21 I know they’ve tried it.
    0:23:23 It’s very expensive to use,
    0:23:24 but my understanding is like in some ways
    0:23:25 it’s better than cursor.
    0:23:27 So like, I’m not sure exactly how it works,
    0:23:28 but I think you use it in the terminal,
    0:23:31 you get full access to your entire code base,
    0:23:33 and then it can just change stuff for you.
    0:23:34 – Yeah, I tried it out.
    0:23:35 It’s pretty cool.
    0:23:35 That is interesting.
    0:23:37 I think they really are doubling down on code.
    0:23:38 – Yeah.
    0:23:39 – I believe when I mentioned the benchmarks stuff,
    0:23:40 I think they were probably talking about
    0:23:42 more real-world coding versus like–
    0:23:43 – I think you’re right.
    0:23:44 – Benchmark coding, you know?
    0:23:46 – Yeah, I think they know that most people
    0:23:48 are using Claude mostly for coding.
    0:23:49 – So it kind of sucks for cursor,
    0:23:51 ’cause like right now everyone uses cursor to use Claude,
    0:23:54 and now it’s roughly basically going to try to kill them.
    0:23:55 – I don’t know.
    0:23:57 I think people will probably still use cursor a lot,
    0:23:59 because cursor and Windsor, if I believe,
    0:24:02 are both forks of Visual Studio code.
    0:24:02 – Yeah, they must be saying like,
    0:24:04 oh, most of our usage is in cursor,
    0:24:07 like why would we not like just be cursor then?
    0:24:08 – Yeah, but I don’t know.
    0:24:11 Like Visual Studio code is like pretty universal.
    0:24:13 Like people use it for coding a lot,
    0:24:15 and it’s what’s familiar to people.
    0:24:18 So I think trying to get people to switch to a new IDE
    0:24:21 might be a tough ask, unless Claude themselves goes
    0:24:24 and makes their own fork of Visual Studio code.
    0:24:25 So like unless Claude goes
    0:24:28 and makes their own fork of Visual Studio code,
    0:24:31 I have a hard time seeing like people switch over
    0:24:34 to like a new IDE that’s like completely different.
    0:24:36 – It’s also not an IDE.
    0:24:38 It’s literally just sitting in the terminal.
    0:24:41 So, you know, some people prefer that.
    0:24:44 I prefer seeing the code more visually
    0:24:46 and have like a nice interface to deal with.
    0:24:49 So I’ve tried both cursor and Windsurf,
    0:24:50 and they’re both great.
    0:24:51 – Yeah, yeah.
    0:24:53 To me, they feel very same-ish.
    0:24:54 Like I have a hard time saying which one’s better.
    0:24:56 They feel very, very similar to me.
    0:24:59 – I’ve had more success with Windsurf
    0:25:03 when trying to iterate on a whole code base.
    0:25:05 Now cursor, I think just released,
    0:25:08 they’re kind of just updated their agent feature,
    0:25:09 which makes it a little bit easier
    0:25:11 to operate on the whole code base.
    0:25:15 And your AI coding assistant agent is able to grok
    0:25:17 and search through the code and do different things.
    0:25:20 That helps it work on the code base as a whole.
    0:25:21 But yeah, you know what?
    0:25:22 Competition’s always good.
    0:25:23 – Yeah, agreed.
    0:25:25 I’m trying to look for some other like really cool examples
    0:25:27 here of stuff that people made with Claude.
    0:25:28 I mean, these are all really cool examples.
    0:25:31 We can spend 30 minutes looking at all these examples.
    0:25:32 So I’m looking for the best ones right now.
    0:25:35 I mean, lots and lots of really cool examples
    0:25:37 of stuff that Claude just did in one shot.
    0:25:38 Lots of snake games.
    0:25:39 Lots and lots of snake games.
    0:25:41 – I made a snake game, yeah.
    0:25:43 Of course I did, and then I had–
    0:25:44 – First thing I did when I tested Claude
    0:25:46 was make a snake game, I still always do.
    0:25:47 – Yeah.
    0:25:49 – But did you guys see the actually good snake game?
    0:25:50 The one where it’s like having a mental breakdown
    0:25:51 as it’s escaping?
    0:25:52 Did you see that?
    0:25:54 – I think you were just showing that, yeah.
    0:25:54 Self-aware, there you go.
    0:25:56 – Oh, this one, yeah, yeah, yeah.
    0:25:58 So the self-aware snake escape?
    0:26:00 – Yeah, so that’s the one, yeah.
    0:26:01 It’s like freaking out.
    0:26:02 (laughing)
    0:26:04 – It just said your brain will struggle with this,
    0:26:07 and then the little text pops up on the screen
    0:26:08 as the snake is moving around.
    0:26:10 – Yeah, it’s like, what’s the snake thinking
    0:26:11 as it’s trying to break out?
    0:26:13 And it starts freaking out that it can’t break out.
    0:26:14 (laughing)
    0:26:16 – It says this was done with one prompt
    0:26:19 plus a request to make special things happen faster.
    0:26:21 So I don’t totally know what that means,
    0:26:25 but that was apparently one-shotted to get that pretty wild.
    0:26:27 Yeah, probably using the agent feature.
    0:26:29 Actually Cursor, they tested the new Claude model,
    0:26:31 and they suggested to use the agent feature.
    0:26:33 And I tested it, it is pretty amazing.
    0:26:35 I tried it on my game and it fixed a problem
    0:26:37 I’d been trying to solve with O1 Pro
    0:26:38 and hadn’t been able to,
    0:26:39 but then it broke like two things.
    0:26:40 So I’m like…
    0:26:41 (laughing)
    0:26:43 And then it was unable to fix the things that it broke.
    0:26:46 So it’s like, there’s still serious limitations with this.
    0:26:47 – Yeah, yeah.
    0:26:52 – Yeah, I’m building a 2D turn-based strategy game right now.
    0:26:54 And I’m a few hours in,
    0:26:56 and now that I’m a few hours in
    0:26:58 and a few thousand lines of code in,
    0:27:00 it’s a little harder, right?
    0:27:03 It takes longer for each iteration to add features.
    0:27:05 There are more bugs popping up
    0:27:07 if it changes one thing over here or something else changes.
    0:27:09 Like I’ll ask it to change something
    0:27:12 and then the entire game will look different
    0:27:13 on the next turn,
    0:27:14 even though I didn’t say anything about that.
    0:27:17 So yeah, of course, there are some limitations.
    0:27:18 We’re gonna get better,
    0:27:20 especially as context windows grow.
    0:27:22 That’s why I actually think maybe Google’s,
    0:27:24 two million token context window models
    0:27:26 are quite appropriate for coding.
    0:27:29 I just, I don’t think anybody uses them for coding.
    0:27:31 I could be wrong about that though.
    0:27:32 – Well, I’ve tried using it
    0:27:35 ’cause you can switch to Google’s models,
    0:27:37 like their Gemini or their Gemma models
    0:27:39 inside of Cursor, I believe.
    0:27:40 And when I tried Google’s models,
    0:27:42 they just didn’t perform as well as Claude.
    0:27:45 So I always find myself going back to Claude.
    0:27:46 I tried 01.
    0:27:48 I haven’t really done a lot of coding with 01 Pro
    0:27:49 because there’s no API, so…
    0:27:51 – 01 Pro is the best in that situation right now.
    0:27:52 Like, if you have a lot of context,
    0:27:54 like my game project,
    0:27:56 even if I remove a lot of other files
    0:27:57 and just get to the basic scripts,
    0:27:59 it’s like 100K of context.
    0:28:00 – Yeah, but they don’t have an API yet.
    0:28:02 So you can’t just use it straight in Cursor.
    0:28:04 You have to use something like,
    0:28:05 what’s it called, Repo Prompt?
    0:28:07 – Repo Prompt, and there’s a few other things too
    0:28:09 where you can combine your files into a file
    0:28:10 that you just copy and pasted.
    0:28:12 – Yeah, but I mean, I’ve got a PC,
    0:28:13 Repo Prompt’s only on Mac,
    0:28:15 so I can’t even use Repo Prompt if I wanted to.
    0:28:18 But I mean, I could, I have a Mac, I just never use it.
    0:28:21 So like, I can’t even use something like Repo Prompt for 01.
    0:28:23 The other really good thing about the Google models
    0:28:24 is like, I think they’re pretty much
    0:28:26 the most inexpensive models.
    0:28:27 So if you’re looking for like as cheap
    0:28:29 as you can get, Google’s probably there.
    0:28:31 I mean, Lama might even be a little bit cheaper,
    0:28:34 but Google’s pretty damn cheap.
    0:28:36 – Yeah, the people that I really trust,
    0:28:38 their opinions on AI stuff,
    0:28:40 they are telling me I’m missing out
    0:28:42 by not using the Gemini models.
    0:28:44 So there’s something there.
    0:28:47 I just, you know, there’s only so much time in the day.
    0:28:48 I haven’t really had a chance
    0:28:50 to extensively test the Gemini models,
    0:28:53 but I really should and I really need to get in there.
    0:28:55 – I’m gonna give you guys a sneak peek real quick here
    0:28:57 of what I’ve been building.
    0:29:00 I’m building my video producer app.
    0:29:02 Basically what I do is I load
    0:29:04 all of these like various like interviews in here.
    0:29:05 I’ve made these folders.
    0:29:08 So like you’ve got like this interviews folder here
    0:29:10 with just like tons of interviews.
    0:29:13 Like here’s Rowan who’s interviewed Mark Zuckerberg
    0:29:16 and Logan and Mustafa Solomon and Demis Hassabis.
    0:29:19 I’ve basically scraped a whole bunch of like interviews
    0:29:23 and panels and launch videos and keynotes
    0:29:24 and, you know, Lex Friedman interviews
    0:29:26 and all sorts of like interviews
    0:29:29 with various like AI leaders and stuff, right?
    0:29:31 And inside of each one of these,
    0:29:33 I actually use OpenAI’s Whisperer
    0:29:35 and it transcribes the whole thing for me.
    0:29:37 So I have like the entire transcription,
    0:29:39 but then I also have like videos
    0:29:41 that don’t have any audio in them.
    0:29:43 Like I’ve got like some B roll footage and stock footage.
    0:29:46 Like here’s stock footage of like a robot.
    0:29:48 And for this one, I actually use Google Gemini
    0:29:50 and Gemini watches the whole video for me
    0:29:52 and writes up a description of everything
    0:29:53 that’s going on in the video.
    0:29:57 And the idea being I just have this giant database
    0:29:59 of videos where I throw the video in
    0:30:02 and then I can search out anything I want.
    0:30:04 So anytime like Sam Altman is mentioned in a video,
    0:30:06 I can search out Sam Altman.
    0:30:07 It’ll pull up all of the videos
    0:30:09 that either have Sam Altman in them
    0:30:10 or they mentioned Sam Altman.
    0:30:13 And then I could quickly find exactly in the transcript
    0:30:14 where he’s talked about.
    0:30:15 – You’re building this yourself?
    0:30:18 – All with cursor and Claude 3.7.
    0:30:20 Dude, I would totally use this.
    0:30:23 I’m using Notion for the almost same thing.
    0:30:25 I’m basically just anytime I find a clip
    0:30:26 that I would find useful in the future,
    0:30:27 I’ll throw it in there.
    0:30:29 And I just have to remember where it is
    0:30:30 and what the context was.
    0:30:32 This is super useful, man.
    0:30:33 I would pay to use this.
    0:30:35 – Yeah. So I’ve got like B roll that I shot.
    0:30:37 This is actually B roll that you’re probably
    0:30:39 in the background of if you look closely enough
    0:30:41 ’cause this is at the NVIDIA event here,
    0:30:43 the, you know, the little digits box.
    0:30:44 I threw this video in here
    0:30:46 and you can see it wrote this description.
    0:30:47 This short video showcases
    0:30:49 the NVIDIA project digits prototype
    0:30:50 to compact computing device.
    0:30:53 The video primarily focuses on the physical device itself.
    0:30:55 So it goes into all of this detail
    0:30:59 from a 31 second video of like me getting B roll digits.
    0:31:00 So now if I ever, I’m like,
    0:31:03 “Oh, what was that video I made that had digit in?”
    0:31:04 And I need to pull that up really quick.
    0:31:05 I could just search up digits, right?
    0:31:08 And it’ll pull up this video as the top video.
    0:31:11 So this is like sort of phase one of what I’m building here.
    0:31:14 Phase two is I want to toss all of this
    0:31:17 into like a rag like retrieval augmented generation model
    0:31:20 where I can say, “Hey, I want to make a video
    0:31:21 about Sam Altman.
    0:31:23 Compile everything we know about Sam Altman
    0:31:25 from all of these videos
    0:31:27 and have it actually write like an outline for me
    0:31:29 based on all the information that’s in these videos.”
    0:31:30 – That’s so cool.
    0:31:32 – So that’s what I’ve been building.
    0:31:34 And this again, I’ve been working on for about two weeks now
    0:31:37 and it uses like seven different APIs.
    0:31:40 It’s using like the OpenAI whisper API.
    0:31:42 It’s using the Gemini API
    0:31:44 ’cause that can actually watch videos
    0:31:46 and tell you what’s going on in the video.
    0:31:49 It’s using Google’s cloud intelligence API
    0:31:52 that’s actually able to like OCR any text in the video.
    0:31:54 So if you’re watching like a slide presentation,
    0:31:56 it can actually OCR any of the text
    0:31:58 that’s in the slide presentation.
    0:32:01 But yeah, it’s been a fun project to build.
    0:32:03 But I run into the same kinds of stuff that Matthew,
    0:32:05 you mentioned where I will go and ask it
    0:32:07 to change like one feature.
    0:32:09 I’ll be like, “Hey, the search isn’t working quite right.”
    0:32:10 And it’ll be like, “Okay, I just fixed it.
    0:32:11 I refreshed the page.”
    0:32:13 And it changes the entire styling.
    0:32:15 And I’m like, I didn’t ask you to touch the CSS at all.
    0:32:17 I just wanted to change how the search functions.
    0:32:19 Like what the hell?
    0:32:21 But other than those little things,
    0:32:22 I use GitHub a lot too.
    0:32:25 It’s like every time a little change works,
    0:32:27 I push it to GitHub so I know I can bring it back
    0:32:28 if I need to.
    0:32:29 – Oh, that’s smart, yeah.
    0:32:31 – But yeah, it’s been fun.
    0:32:35 And I’ve made less YouTube videos that I normally make.
    0:32:37 I’ve only been putting out one YouTube video a week
    0:32:38 for the past like month
    0:32:40 because I’ve gotten so addicted
    0:32:42 to playing around with AI coding.
    0:32:43 It’s so fun.
    0:32:44 Like I’m working on the game now.
    0:32:46 I’m just like shocked that I can like build a game by myself.
    0:32:48 It’s like, you know, in the past,
    0:32:50 I never could imagine like one person could build a game
    0:32:52 or now you’re building your own software product.
    0:32:55 And Matt, you’re kind of like a no-code guy, right?
    0:32:59 Now you’re using, you went from no-code to using AI to code
    0:33:00 and you’re actually able to build a whole product.
    0:33:02 I mean, it’s just, everything’s changing.
    0:33:03 – Yeah.
    0:33:05 And I mean, the thing is I’m learning as I go too, right?
    0:33:07 ‘Cause like when you use these models,
    0:33:10 it’ll explain to you what it did, what it changed.
    0:33:12 You know, why something broke.
    0:33:14 3.7 has been really, really good at that.
    0:33:16 I don’t know if it’s 3.7
    0:33:18 or if it’s like the agent feature inside of cursor.
    0:33:21 But when it fixes stuff, it’ll explain the problem.
    0:33:23 It’ll say you were running into this problem
    0:33:24 because this, this and this was happening
    0:33:26 or there was a conflict with this and this
    0:33:29 or you know, it was sending the wrong information
    0:33:30 through the API or whatever, right?
    0:33:32 It gives you that information.
    0:33:34 So although I don’t actually know
    0:33:36 how to like type out the code myself,
    0:33:39 I’m getting a lot better at troubleshooting
    0:33:42 why problems are happening within the code.
    0:33:45 I don’t know how to like actually change the code.
    0:33:48 Like I don’t know what to write to make it work myself,
    0:33:50 but I’m starting to pick up on like,
    0:33:52 oh, I think this might conflict with this
    0:33:54 as a result of learning
    0:33:56 as it explains to me all of these problems.
    0:33:59 So anyway, that’s what I’ve been working on.
    0:34:01 But there is one last topic I do wanna shift in.
    0:34:02 – Let’s talk about it.
    0:34:04 – I wanna shift over to GPT 4.5
    0:34:08 because as of today’s recording, GPT 4.5 came out.
    0:34:10 Before we hit record, I asked Matt,
    0:34:11 I’m like, what were your thoughts on that launch today?
    0:34:13 And he’s like, I’ll save it for the recording.
    0:34:15 I was like, all right, let’s save it for the recording.
    0:34:16 So let’s start there.
    0:34:20 What are your thoughts on the GPT 4.5 launch?
    0:34:21 – So I think it looks cool.
    0:34:24 I haven’t obviously tested it extensively.
    0:34:25 It came out just a handful of hours ago,
    0:34:28 not even since recording.
    0:34:30 So Nathan, you mentioned it’s really good at writing.
    0:34:32 So I’m excited to test that out.
    0:34:34 But let me just talk about a couple of things
    0:34:37 that I noticed from the live stream.
    0:34:40 So one, it’s the largest model that they’ve ever made.
    0:34:44 And it took new innovations on both the pre-training
    0:34:47 as well as the serving of it, the inference,
    0:34:48 to actually be able to serve this model.
    0:34:51 And if you use it, it is pretty slow, right?
    0:34:53 So I found that pretty interesting.
    0:34:55 It’s a world knowledge model,
    0:34:58 meaning it is not a thinking model,
    0:35:00 but in terms of just questions and answers
    0:35:02 that you would use kind of day to day,
    0:35:03 it’s really good at that.
    0:35:05 And it’s much better than GPT 4.0.
    0:35:07 And then the last thing,
    0:35:08 and I think this flew under the radar a bit,
    0:35:10 and I wanna get your guys’ thoughts on this.
    0:35:13 They said that it was such a massive model
    0:35:17 that they actually trained it across multiple data centers,
    0:35:19 not in a singular location.
    0:35:21 So when you think of the GROC model,
    0:35:24 it’s the Colossus data center, 100,000,
    0:35:27 something like 200,000 GPUs in a single place.
    0:35:28 They didn’t have that.
    0:35:29 They being open AI, they don’t have that.
    0:35:32 And so they had to split it up.
    0:35:35 And I don’t think any other company has done
    0:35:38 parallel training across multiple data centers
    0:35:40 at this level before.
    0:35:42 I think that it really flew under the radar
    0:35:44 and unlocks the ability for companies
    0:35:49 that don’t have the money or resources that an XAI does
    0:35:51 to go out and spread their model out
    0:35:53 and still get a massive model trained.
    0:35:55 And I thought that was just fascinating.
    0:35:57 – Yeah, I didn’t catch that.
    0:35:58 I mean, it’s not something that I picked up on
    0:36:00 until you just mentioned it.
    0:36:02 So I mean, that is really interesting.
    0:36:05 So you’re saying that like, basically,
    0:36:07 there was separate physical locations
    0:36:09 where the training was happening simultaneously
    0:36:11 in separate locations.
    0:36:15 – Yeah, ’cause they probably literally could not get
    0:36:19 a large enough data center to train this one model
    0:36:23 in a concentrated location like Colossus with GROC3.
    0:36:26 So they had to split it up and do it in parallel.
    0:36:28 And again, that really hadn’t been done before
    0:36:30 at this level.
    0:36:31 – Yeah, well, I mean,
    0:36:33 we’ve got Project Stargate coming as well.
    0:36:34 That’s gonna be the data center
    0:36:36 where eventually they’ll be able to do it all
    0:36:38 from one data center, I believe.
    0:36:40 But yeah, no, that’s really, really interesting.
    0:36:43 I personally found the actual presentation
    0:36:45 a little bit underwhelming.
    0:36:48 They didn’t really show off any sort of capabilities
    0:36:51 that were like, oh, we’ve never seen that before.
    0:36:53 – Considering this was like hyped up for like a year
    0:36:55 of like, oh, Orion’s coming.
    0:36:57 – Well, and then also Sam Altman,
    0:36:59 he posted on Twitter a couple of days ago
    0:37:02 that a lot of people who have used GPT 4.5
    0:37:05 have gotten that feel the AGI moment from it, right?
    0:37:08 But anybody really feel an AGI moment from their demo.
    0:37:10 I don’t know, I’m reading into it a little too much.
    0:37:13 But the other thing I realized too is like,
    0:37:15 you can always tell when OpenAI doesn’t see it
    0:37:16 as that big of an announcement,
    0:37:18 when Sam doesn’t show up for the announcement.
    0:37:19 – Right.
    0:37:22 You know, I think there’s more there than people realize.
    0:37:25 And so many people accuse me of being overly optimistic.
    0:37:28 So bear with me, or at least take what I’m gonna say
    0:37:29 with a grain of salt.
    0:37:30 I really do think that there’s more there.
    0:37:34 First of all, this is the first version, right?
    0:37:35 And we’re gonna get lots of different upgrades over time.
    0:37:37 We’re gonna get a turbo version.
    0:37:40 But I think in terms of just a baseline model,
    0:37:42 it’s a lot better at the kind of general Q and A.
    0:37:45 And so from that, when you have all of that world knowledge
    0:37:47 baked into this model,
    0:37:49 how do you think the thinking versions,
    0:37:52 the O1s, the O3s, how do you think those are the created?
    0:37:54 It’s taking that foundation model,
    0:37:56 using reinforcement learning,
    0:37:59 and then kind of eliciting that thinking behavior from it.
    0:38:02 So now we’re gonna have this incredible foundation model
    0:38:05 to build the thinking models on top of.
    0:38:07 Maybe that’s what O3 Pro is.
    0:38:09 – And it’ll be crazy expensive.
    0:38:11 – Oh, that’s another thing is the cost, yeah.
    0:38:11 – You combine those two.
    0:38:13 It’s like, okay, this model is already very expensive
    0:38:15 before you do the thinking.
    0:38:17 And so when you do, it’s like, it makes me wonder,
    0:38:20 is that gonna lead to like the $2,000 a month plan
    0:38:22 or something like this?
    0:38:23 You know, there’s also another angle to think of it,
    0:38:26 which is, it’s all about the vibes, right?
    0:38:28 That was like a huge theme of the announcement of the vibes.
    0:38:30 It’s a warm model.
    0:38:32 It’s a high EQ model.
    0:38:33 – They were trying to compete with Claude.
    0:38:34 – Right, and so–
    0:38:35 – You know, ’cause before Claude, everyone said like,
    0:38:38 Claude had a better vibe than chat to BT, you know?
    0:38:40 – Right, and so like, if you think about it
    0:38:42 from that perspective,
    0:38:44 they’re really positioning this model
    0:38:48 to be a true kind of AI personal assistant.
    0:38:50 And I emphasize the word personal.
    0:38:53 It is there to help you.
    0:38:55 It is there to know you.
    0:38:58 And I think this is really more interesting
    0:39:00 than people are giving it credit for.
    0:39:03 Because when you think about systems like Siri,
    0:39:04 what it could be,
    0:39:08 it’s probably based on something like this type of model.
    0:39:10 – Yeah.
    0:39:11 – One thing I just wanted to point out too
    0:39:13 is I don’t know if you guys caught this tweet,
    0:39:16 but he basically said that the new model
    0:39:18 is a giant expensive model,
    0:39:19 and they really wanted to launch it,
    0:39:21 but they ran out of GPUs, right?
    0:39:24 And this kind of comes back to Matthew’s point earlier
    0:39:26 of how they had to like train it, you know,
    0:39:30 in separate locations with this parallel processing.
    0:39:33 They’re literally like reaching the end
    0:39:36 of the available GPUs to be able to process this stuff.
    0:39:38 And I just found that really, really interesting.
    0:39:39 So I wanted to share that real quick.
    0:39:41 – Open AI is GPU poor.
    0:39:45 – Yeah, I mean, they’ve got SoftBank now behind them.
    0:39:47 – Who would have thought?
    0:39:48 – I want to share two things that kind of show
    0:39:50 that there is something special here.
    0:39:52 So here’s something I saw that’s pretty interesting
    0:39:55 from a professor or he’s a biomedical scientist.
    0:39:56 He said, “It appears to be remarkable
    0:39:57 in medical imaging diagnosis.
    0:40:00 It was the only model that perfectly diagnosed
    0:40:01 this ultrasound image.”
    0:40:03 So like in terms of like looking at ultrasound,
    0:40:06 apparently it’s by far the best model out there.
    0:40:08 So there probably are things that we’re going to find
    0:40:09 about this model that are special
    0:40:12 that when you first use it are not apparent.
    0:40:13 – Yeah, good point.
    0:40:13 No, I agree.
    0:40:15 I’m sure it’s a lot better model
    0:40:18 than I feel like their presentation led on to be.
    0:40:20 Part of the problem was like,
    0:40:22 we all watched the live stream, I’m sure.
    0:40:25 They seemed like they were kind of nervous.
    0:40:26 They were kind of uncomfortable.
    0:40:29 Sam wasn’t on the live stream this time.
    0:40:31 I just don’t think it was presented very well, honestly.
    0:40:32 I think that was probably the problem.
    0:40:34 – Yeah, here’s another thing I saw
    0:40:35 that was super interesting.
    0:40:37 Like the hallucination rate is like way lower
    0:40:38 with like 4.5.
    0:40:42 So it’s like 37.1% hallucination rate
    0:40:45 versus 4.0 being 61%, you know,
    0:40:47 0.3 mini being 80.3%.
    0:40:48 It’s like dramatically lower.
    0:40:50 So that’s a, yeah, that’s step forward.
    0:40:51 ‘Cause like obviously for big companies,
    0:40:53 one of the big problems with using AI models
    0:40:55 is hallucinations, right?
    0:40:56 So yeah.
    0:40:56 – Yeah, absolutely.
    0:40:57 That’s crazy.
    0:41:01 0.3 mini on simple QA has an 80% hallucination rate.
    0:41:03 That seems insane to me.
    0:41:04 – Seems right.
    0:41:05 ‘Cause like 0.3 mini is like really fast,
    0:41:07 but I’ve seen it make lots of weird mistakes.
    0:41:09 Like when I tried to use it for coding and stuff,
    0:41:10 it’s like, oh, you’re a genius.
    0:41:12 And the next I’m like, oh, you’re a total moron.
    0:41:15 It’s just like, it’s responses are like all over the place.
    0:41:16 – I don’t quite understand that benchmark.
    0:41:19 I feel like an 80% hallucination rate
    0:41:21 just makes, would make something unusable.
    0:41:23 I mean, if you’re only getting accurate information
    0:41:25 20% of the time, I don’t know.
    0:41:28 I don’t totally understand how this one is come to, I guess.
    0:41:29 – I think we would have to see what the actual questions
    0:41:33 are in the simple QA benchmark to understand the context
    0:41:34 of why it’s scored so poorly.
    0:41:36 But then, but then just, you know,
    0:41:40 I guess maybe remember it’s mostly for math, science,
    0:41:42 like basically STEM, right?
    0:41:44 Things that have verifiable answers.
    0:41:45 – Absolutely.
    0:41:46 – I think it’s cool.
    0:41:47 I think it’s a good start.
    0:41:49 – Yeah, no, I’m excited about it.
    0:41:51 I mean, like, hey, for content creators like us,
    0:41:53 all this news is amazing, right?
    0:41:54 ‘Cause we get to talk about it.
    0:41:56 We get to, you know, keep on sharing what’s coming out.
    0:41:59 We get to play with it all and get access to it
    0:42:01 and show what it’s capable of
    0:42:03 and put it through its motions and stress test it.
    0:42:06 And I think, you know, I could speak for myself.
    0:42:07 I think I could speak for Matthew.
    0:42:08 That’s what we love doing, right?
    0:42:10 Like we love playing with the stuff and stress testing it
    0:42:12 and figuring out what it really can do.
    0:42:15 And that’s what’s most exciting to me, I think.
    0:42:19 You know, GPT-5 is probably only six, eight weeks away.
    0:42:22 I mean, it’s not that far off, apparently.
    0:42:24 So this is just the beginning.
    0:42:27 It’s been a crazy couple of weeks.
    0:42:28 – Absolutely.
    0:42:28 – So who won?
    0:42:29 Who won the last week?
    0:42:30 – I don’t know.
    0:42:33 I mean, they all have different pros and cons, I guess.
    0:42:34 Right?
    0:42:35 – I was most surprised by Grock.
    0:42:36 Right, like out of all the three,
    0:42:39 like I feel slightly disappointed by 4.5.
    0:42:40 I would say it would be my general feeling.
    0:42:42 Not like majorly disappointed,
    0:42:43 but like Orion’s not as big of a deal
    0:42:45 as I was hoping it was going to be.
    0:42:45 – Yeah.
    0:42:47 – Grock really impressed me.
    0:42:50 Anthropic Clawd, you know, 3.7, amazing.
    0:42:52 Kind of what I expected in terms of improvement.
    0:42:54 So I would say Grock’s the biggest surprise
    0:42:55 out of all the three.
    0:42:56 – Yeah.
    0:42:57 – I mean, personally, I’ve gotten the most use
    0:43:01 out of Clawd 3.7 because I’ve been really, really going
    0:43:03 down the coding rabbit hole lately,
    0:43:05 but that’s just a very like anecdotal thing for me, right?
    0:43:08 Like that’s the use case that I’ve found the most valuable
    0:43:10 in the moment is I’m doing a lot of coding
    0:43:12 and 3.7 is great for that for me right now
    0:43:14 in this moment in time.
    0:43:18 – Yeah, with the caveat that I haven’t tested GPT 4.5 much,
    0:43:20 I gotta give the crown to Grock 3
    0:43:23 over this last wave, these last couple of weeks.
    0:43:25 It went from I never used Grock 2
    0:43:28 to now it is my go-to model for a lot
    0:43:30 of use cases on my day-to-day tasks.
    0:43:32 So definitely have to give it to Grock 3 there.
    0:43:33 – Yeah.
    0:43:35 Now we just need that damn API.
    0:43:36 – Yes, yeah.
    0:43:37 Then we’ll really–
    0:43:38 – It’ll be interesting to benchmark it
    0:43:39 when the API comes out, right?
    0:43:41 Like see how it actually compares
    0:43:42 to all these other models.
    0:43:43 – Yeah, yeah.
    0:43:44 Well, cool, Matthew.
    0:43:45 This has been amazing.
    0:43:47 You know, anybody listening, make sure you check out
    0:43:50 Matthew’s YouTube channel over at MatthewBurman.
    0:43:51 You’ve got an amazing newsletter.
    0:43:52 I’m subscribed to it.
    0:43:54 It’s the forward future newsletter, I believe.
    0:43:55 – Forward future.
    0:43:57 – Everybody needs to go check those out anywhere else.
    0:44:00 You want people to go check you out and follow you online?
    0:44:04 – Yeah, come check me out on Twitter @MatthewBurman.
    0:44:07 Come flame me for my opinions on politics,
    0:44:08 even though I don’t share them.
    0:44:08 – All right, you’re asking for it?
    0:44:09 – Yeah, yeah.
    0:44:10 – Well, you called it Twitter.
    0:44:11 That’s like, you’re kind of like your DM.
    0:44:12 – All right, you’re right.
    0:44:13 I already offended one side.
    0:44:15 – So like, look, I’m on your side, guys.
    0:44:16 – Well, cool.
    0:44:17 This has been super, super fun.
    0:44:19 – Thank you for having me, guys.
    0:44:19 Thank you.
    0:44:20 – Great having you back.
    0:44:23 I’m sure you’ll be back on if you want to be.
    0:44:24 We’d love to have you again.
    0:44:26 It’s always fun to chat with you
    0:44:29 and really appreciate you spending the time with us today.
    0:44:30 – Yeah, thank you, guys.
    0:44:33 (upbeat music)
    0:44:35 (upbeat music)
    0:44:38 (upbeat music)
    0:44:40 (upbeat music)
    0:44:43 (upbeat music)
    0:44:46 (upbeat music)

    Episode 48: How do the latest updates to large language models stack up against each other? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) are joined by Matthew Berman (https://x.com/MatthewBerman), an expert in deep-diving and testing the nuances of large language models.

    In this episode, the trio discusses the recent releases of Grok 3, Claude 3.7, and GPT-4.5, analyzing their strengths, weaknesses, and unique features. Tune in to learn which model might be best for your needs, from coding and real-time information to creative writing and unbiased truth-seeking.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Exploring New AI Models
    • (05:35) Inconsistent AI Code Performance
    • (06:26) Redesigning Benchmarks for Modern Models
    • (11:33) AI Bias Amplification on Social Media
    • (15:11) AI Bias and Human Oversight
    • (17:49) Claude 3.7: Improved Coding Abilities
    • (20:30) Claude Update: Better Code, Worse Chat
    • (23:19) Resistance to Switching IDE from VS Code
    • (28:05) Video Producer App Preview
    • (29:55) Showcasing Nvidia Digits Prototype
    • (34:00) GROK Model’s Distributed Training
    • (36:31) Optimistic Perspective on Future Upgrades
    • (40:59) Excited for GPT-5 Launch
    • (42:08) Claude 3.7 Excels in Coding

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • This AI Tool Automates Any Task in 60 Seconds (Live Demo)

    AI transcript
    0:00:06 Hey, welcome to the Next Wave Podcast. I’m Matt Wolff. I’m here with Nathan Lanz. And
    0:00:10 today we’re going to talk about probably one of the hottest topics of 2025. And that’s
    0:00:16 the topic of AI agents. Today we’re bringing on a special guest, Dmitri Shapiro. He is
    0:00:22 the CEO of Mind Studio. And Mind Studio is a tool that makes it insanely easy for you
    0:00:26 to build your own AI agents. So in this episode, not only we’re going to talk about AI agents
    0:00:31 and what they’re useful for, we’re going to actually jump into Mind Studio with Dmitri
    0:00:37 and show off him actually building some of these agents in real time and making them
    0:00:41 work. It’s pretty impressive stuff. And I don’t want to waste your time. So let’s just jump
    0:00:46 right in with Dmitri.
    0:00:50 Before we get into talking about Mind Studio and all the cool stuff that it can do, let’s
    0:00:54 get a little bit of background on you. I know you and I are practically neighbors. You live
    0:00:59 maybe three miles from me. We’ve met up for coffee before. And when we hung out, you told
    0:01:03 me about all of the sort of stuff you did before you built Mind Studio. And I just found
    0:01:08 it fascinating because you worked at MySpace. You worked at Google. You did all sorts of
    0:01:12 really, really cool stuff. And you’ve been sort of in the thick of the AI world for quite
    0:01:17 a while. So can you give us that quick rundown of what you’ve done in AI and in the tech
    0:01:18 world so far?
    0:01:23 Ah, totally. Yeah. Great to see you again. And I love being your neighbor. These days
    0:01:30 I describe myself as an old nerd. I started writing code in 1984 when I was 14 years old
    0:01:34 and I went to see the movie War Games. And I came out and was like, I got to get a computer.
    0:01:38 And we were like dirt poor so we couldn’t afford a computer. But my high school had two
    0:01:44 Apple 2e computers. And so I spent the rest of my high school years instead of being a
    0:01:49 normal teenager in this like computer lab. And you’re mostly trying to like write code
    0:01:54 and cause havoc and whatever, you know, way that I could start World War III. Mostly like
    0:02:01 show my friends, you know, versions of like very primitive viruses. And then I studied
    0:02:05 electrical engineering, got a degree in electrical engineering. I’ve never done a day of it always
    0:02:10 done software. And then, you know, had a long career in software before, you know, founding
    0:02:15 Mind Studio. I spent four and a quarter years at Google on the main campus. There I was
    0:02:21 on the product side. I started on Google plus, when Google felt that Facebook was an existential
    0:02:27 threat. And like all the efforts went into standing up what they called the social spine
    0:02:32 or the social backbone of Google. That’s what was called Google plus, make all of Google
    0:02:39 social. And my job was to build the social graph, which like the most important part
    0:02:44 of any social network. And even though Google has tremendous amount of information on users,
    0:02:49 it was actually struggling to put together like things like you may know, you may like
    0:02:54 feed rankings and things of that sort. And so my job was to run product on these like
    0:02:59 three mostly machine learning teams that were crunching through all this Google data and
    0:03:03 trying sort of like compute these lists and also creating a signal to abuse because whenever
    0:03:08 you open up social networks, you know, abuse goes crazy. And in all of that as a function
    0:03:13 of that before that, I was the chief technology officer of MySpace Music. This is the joint
    0:03:19 venture between MySpace and the record labels. So I did that before that I built a venture
    0:03:23 back company called VO Networks, which was one of the major competitors to YouTube in
    0:03:28 the first five years of YouTube’s life. In fact, we had video recommendations before
    0:03:32 YouTube had video recommendations. So that was very early. And of course, that was all
    0:03:37 machine learning stuff. Before that, I built a venture backed cybersecurity company enterprise
    0:03:43 software cybersecurity company called Aconyx, raised $34 million for Aconyx raised $70 million
    0:03:49 for VO. Here so far, we’ve raised $36 million, like raised $140 some odd million dollars
    0:03:53 in my career. And then what are before that I spent five years at Fujitsu and other big
    0:03:58 Japanese companies like building their web team. So I’m an old nerd. I love building
    0:04:03 products. This is the most fun I’ve ever had, which I think should say a lot because I’ve
    0:04:07 been involved in a lot of like really, I thought like really interesting and fun things but
    0:04:13 like, wow, what a time to be alive and working on products to be able to like leverage like
    0:04:18 these new capabilities. It’s crazy. Yeah, absolutely. You raised a lot of money over
    0:04:22 three different projects, right? Was most of that from Western VCs like Silicon Valley
    0:04:28 VCs or like strategic partners or like, where did most of that come from? Both. So Aconyx,
    0:04:34 I raised $34 million. The vast majority of it was from VCs, top tier and VCs like Mandel
    0:04:39 Ventures was in there, which looked top tier VC and sort of second tier VCs at VOI raised
    0:04:47 $70 million, a lot from VCs, but also like Intel, Adobe, the two former heads of Viacom,
    0:04:52 John Dolgen and Tom Preston, all of Sifrin, Bretonham law firm, which like the most important
    0:04:57 law firm in entertainment, Michael Eisner, the former head of Disney, Michael was on
    0:05:01 my board of directors. There was a lot of strategic time Warner was an investor. So a
    0:05:08 lot of strategic synod deal along with VCs. Here it’s primarily VCs and lots of angels,
    0:05:13 lots of like senior Googlers. Because of my time at Google, I knew who I was, I laughed,
    0:05:17 I started doing interesting things. They all sort of showed up and said, can I ride with
    0:05:20 you? Yeah, I’ve heard there’s like a huge angel network of like ex Googlers. Yeah. And
    0:05:25 so yeah, so it’s a bunch of nerds, you know, people like me are mostly nerds.
    0:05:32 Well, so now you’ve got mine studio and you know, the big buzzword of 2025 is agents, right?
    0:05:36 Like everybody’s talking about agents that you hear open AI and Google and all of these
    0:05:39 companies basically saying agents, agents, agents, this is what you’re going to hear
    0:05:45 the most about in 2025. And mine studio to some degree is basically a tool that helps
    0:05:50 you build these agents. But I’m curious, like, I feel like everybody kind of defines agents
    0:05:55 differently. Like how do you define agents? What are you calling? What you’re building
    0:05:59 at mine studio agents? Like, where do you stand on all that?
    0:06:05 Yeah, I think there’s a tremendous amount of sensationalism and hype and noise. And I’m
    0:06:11 mostly allergic to all of it. And so even though we call them agents, we’re kind of
    0:06:14 allergic to the word agent because it’s it’s so overloaded.
    0:06:18 Yeah, but it’s kind of the word that people know and it’s the buzzword right now, right?
    0:06:21 So I’m not sure what else to call these things. For a while there, we call them workers just
    0:06:26 to differentiate from agents. But here’s how we think of them, I think differently than
    0:06:30 majority of the people in when they use the term agent, I think kind of the popular definition
    0:06:37 of it is an AI that can sort of direct itself and do what it needs to do in sort of quite
    0:06:44 amorphous situation. It has access to tools. And it has a goal and it goes off and does
    0:06:48 that. Like, magically, it’s intelligent. And it’s able to do that like a human would like
    0:06:52 I’ll figure it out, right? Basically, it takes the approach of like, give me some task,
    0:06:56 I’ll figure out like what to do about it. And as a nerd, like I’m excited about that kind
    0:07:01 of vision. And certainly, we’re all moving in that direction. But at the same time, we
    0:07:08 think that’s sort of a bit too much autonomy to be just given to AI and that like for most
    0:07:13 things that you want to do, you generally want to be able to be in like more control.
    0:07:18 You want it to do the heavy lifting and the scaling, but not necessarily to do like all
    0:07:23 the strategy, for example, for you and like choosing what paths to take. And you’re sort
    0:07:27 of telling it where to go and it goes there, rather like it, it takes you for a ride to
    0:07:33 wherever it is that it decides to go. And so we think of these things as being workflows
    0:07:37 that you can say, I want to get something done, I want to create something that does something.
    0:07:41 And well, how would you do that? Well, you create a workflow like same way that you would
    0:07:44 sort of like onboard a human to do it, you would explain to this thing, here’s what I
    0:07:50 want you to do. Like every weekday morning at 9am, I want you to go and search the web
    0:07:55 and search these other databases, looking for developments and this thing that I care
    0:08:00 about. And you can be quite fuzzy in saying what the thing is, you can be quite precise.
    0:08:05 And then if you find that kind of situation, if you find interesting information there,
    0:08:09 do some thinking, you know, synthesize it, evaluate it from my standpoint, you know what
    0:08:14 I want. And then sort of create a nice report for me and sort of nudge me, send me an email,
    0:08:18 send me an application, tell me, hey, something just happened in the world that you should
    0:08:22 know about, I’ve thought about it for you, here you go. And you can just set it on a
    0:08:26 schedule that’s like one of the ways that you can deploy these MindStudio AIs and it
    0:08:29 just runs and does that for you. By the way, that’s amazing.
    0:08:35 Like I’ll give you like an example here, like, so Google Trends is a really sort of important
    0:08:40 living data set of what are people searching for? Like what does the world care about? Or
    0:08:45 what is like the world that you care about? What do they care about? And more specifically,
    0:08:49 like how is that changing over time? Like what’s becoming more popular, what’s becoming
    0:08:54 less popular? So the information is there on Google Trends. But the interface for you
    0:09:00 to like load it in your browser and like periodically go look at Google Trends, that takes a lot
    0:09:04 of work. And so you’re going to look at like one or two keywords, but you’re certainly
    0:09:09 not going to look at 100 keywords. You certainly wouldn’t look at 1000 keywords. But you see
    0:09:15 with the MindStudio AI agent, you can look at 1000 keywords. And you don’t have to do
    0:09:19 any of it. You just have to go and say, you know, when you build the agent, like, this
    0:09:23 is what I want to look at, it can generate the keywords or you can generate the keywords.
    0:09:29 And it can at scale, every day or every hour, every minute you decide, go and look at the
    0:09:34 world and synthesize it and look for those changes you want. If you’re trading stocks,
    0:09:37 that’s an unfair advantage, right? If you’re starting companies as well, that’s a big thing
    0:09:41 with like product studios is like, it’s an unfair advantage in everything. Yeah, exactly.
    0:09:47 Yeah. And so again, before now, like that, how would you do that? But now that’s actually
    0:09:53 really straightforward, you know, or you find something, you know, online that you’re reading,
    0:09:57 you’re consuming, but it’s hard to tell whether that thing is accurate, you know, more and
    0:10:02 more things are inaccurate, misinformation, disinformation. Now we’ve got a little button
    0:10:06 you can press and like, we’ve got a Chrome extension. And you can have these agents
    0:10:10 triggered from any web page. And so you can push a button and basically takes that web
    0:10:16 page and then does deep research, like deep Google searches using like very sophisticated,
    0:10:21 you know, Google dorks, Google search operators and composite all the data evaluates and can
    0:10:24 return back to you like, these things are true, these things are misleading, these things
    0:10:29 are false. And if you’re making business decisions on that or, you know, personal decisions,
    0:10:33 it’s valuable to be able to have that kind of superpower. That’s actually kind of the
    0:10:38 way I’ve mostly been describing these things when I talk about them. Is that called a superpower?
    0:10:42 Like AI superpower, because that’s exactly what they are. And they give you an unfair
    0:10:47 advantage if you’ve got the right ones, and you use them in the right way.
    0:10:52 Hey, we’ll be right back to the show. But first I want to talk about another podcast
    0:10:56 I know you’re going to love. It’s called entrepreneurs on fire. And it’s hosted by John
    0:11:01 Lee Dumas available now on the HubSpot podcast network. Entrepreneurs on fire stokes inspiration
    0:11:05 and share strategies to fire up your entrepreneurial journey and create the life you’ve always
    0:11:11 dreamed of. The show is jam packed with unlimited energy, value and consistency. And really,
    0:11:16 you know, if you like fast paced and packed with value stories and you love entrepreneurship,
    0:11:21 this is the show for you. And recently they had a great episode about how women are taking
    0:11:26 over remote sales with brook triplet. It was a fantastic episode. I learned a ton. I highly
    0:11:30 suggest you check out the show. So listen to entrepreneurs on fire wherever you get
    0:11:33 your podcasts.
    0:11:38 Yeah, well, let’s dive in and actually take a look at that mind studio and some of the
    0:11:41 stuff that you’re building. You know, if anybody who’s listening on audio, we are going to
    0:11:46 be sharing this on YouTube as well. So you can go and check out what Dimitri’s sharing.
    0:11:51 But I know you guys just rolled out a pretty new feature that works inside of Chrome where
    0:11:56 you can build all these automations directly inside of Chrome. And so I thought it’d be
    0:12:00 kind of cool to show off some of the types of agents that you’re building right now and
    0:12:04 some of the like really simple workflows that you’ve built to, you know, just make your
    0:12:05 life easier.
    0:12:10 For sure. For sure. Yeah. So I’ve installed a Chrome extension. This thing is an alpha
    0:12:13 now and soon we’ll be in the app store. It’s been submitted. And it’s here. I can drag
    0:12:19 it around. It can be collapsed or it can be expanded. And these are things almost like
    0:12:24 triggers for AI agents. So here’s a TLDR AI agent, here’s an agent called what is this?
    0:12:29 This one’s called alternatives, right? And I can trigger them. Or if I want to have these
    0:12:34 things collapsed, and I can click on this button, and it gives me the side drawer. And
    0:12:39 in it, you get, you know, all the agents that are sort of available in our agents store,
    0:12:43 you get all the agents that you’ve created yourself, you’ve got agents that you’ve started.
    0:12:48 So these are like your favorites. And so you can very quickly, again, take an agent and
    0:12:53 trigger it. And so like what could we do here? Well, one thing we could do here, for example,
    0:12:57 is do like a TLDR. And so again, I can trigger it from here, or I could have triggered it
    0:13:01 from the extension right here that I’ve got. And it instantly, you know, again, grab the
    0:13:05 content on this page and gave me this quick summary. And this quick summary was just a
    0:13:10 quick AI model looking at this thing. And it did this and extracted these entities. We
    0:13:13 also have a thing called who’s involved, which sort of more sophisticated, you’ll go grab
    0:13:19 the page, and then we’ll find like all the people and entities on that page, and ask
    0:13:24 me like what people or entities I’d like to know more about, like this theoretical business.
    0:13:29 I can click this, and this is going to kick off like a really sophisticated workflow that
    0:13:34 could take in this case, I think takes about two minutes to run, which is again, generating
    0:13:38 Google search queries over here. Again, you can sort of see it working. It’s executing
    0:13:41 those Google search queries on Google news, Google images, blah, blah, and then it’s
    0:13:45 going to create a report for us. And by the way, we don’t have to wait, we can say notify
    0:13:49 me when finished. And this thing is going to send me an email with this like really
    0:13:54 sophisticated profile executive summary, you know, of this person here that I’ve done.
    0:13:57 And again, we don’t have to stay on this page, we can keep moving. So you know, here you
    0:14:01 are, I’ve got a little thing over here where I’ve got a bunch of buttons, one of them is
    0:14:06 YouTube TLDR, right, I click this thing, it’s going to grab the transcript, it’s going to
    0:14:11 process it and it’s going to give me a TLDR. Now your video is only 33 minutes here. And
    0:14:16 so I could have watched it and you know, two times speed, that’s nothing at all. But some
    0:14:21 videos are, you know, three and a half hours, you know, all of it is also fine tunable by
    0:14:28 me. And so if I want my TLDRs to be different, I simply go and I edit this AI agent that
    0:14:33 I have, and then I can use it. Gotcha. So many of these things like when you were just
    0:14:36 showing you had the little like sort of tray on the right side where you had a whole bunch
    0:14:41 of little like apps that you built. So my understanding is most of those apps are like
    0:14:47 pre built templates that are available in my studio that you can just add to your sidebar,
    0:14:50 but then also you can create your own and you can also modify some of the ones that are
    0:14:55 already pre built. Is that kind of the idea? Yeah, let’s talk about that. So again, if people
    0:15:00 sign up and they don’t yet have anything started, they haven’t built anything, they will see
    0:15:03 on this agents tab, they’ll see these agents. It’s almost like a little like app store for
    0:15:08 agents. It’s like an app store for agents. Exactly. Sorry, these ones right here. And
    0:15:12 there are more agents coming again. This is an alpha now. You should expect hundreds
    0:15:17 beyond that thousands. And then on any one of these agents, you can open it up and it’s
    0:15:21 got sort of a like a detailed page and it tells you how much it costs to run this on
    0:15:26 average about six cents. In inference, this one takes about 45 seconds. But again, it
    0:15:32 returns back like these really awesome summaries like reports in a sense that are publishable,
    0:15:37 that are shareable. They’re like pieces of content in a sense with just clicking a button.
    0:15:42 And so you can run it right from here or you can run it from the extension. Gotcha. Click
    0:15:47 a button, YouTube TLDR and run it. Now, the other thing you can do with these things is
    0:15:53 if you click here, you can say make a copy. And when you make a copy, you’ve just made
    0:15:58 a copy of that agent and you have access to like all of the it’s not code. You’re like
    0:16:03 forking it basically. You’re forking it. Exactly. So agents are made out of these things called
    0:16:07 workflows and these workflows are basically these things where you can add these little
    0:16:12 nodes to. And so it’s got a start block. And so let me make this a little bit smaller is
    0:16:16 take a look at what this one is doing. It’s actually quite simple. It’s got a start block.
    0:16:20 You can configure the start block to again, how does it run? These things can be run on
    0:16:25 demand meaning you trigger them manually. They can be run on any kind of a schedule that you
    0:16:30 specify. If you click email here, it gives you the custom email address. And whenever
    0:16:36 you send an email to the email address, it takes that as input and triggers workflow
    0:16:41 and passes it variables. Or you can trigger it as a browser extension, which is what this
    0:16:46 was. And then you get access to these launch variables as we call them. So when this thing
    0:16:52 gets called by the browser extension, the browser extension passes at the URL, the metadata
    0:16:57 in code, the metadata, our name, description, image, etc. The full text, which is visible
    0:17:01 to the users is stuff you can see on the page. If user is selecting anything highlighting
    0:17:06 anything or right clicking anything, it gets passed in this user selection variable. And
    0:17:10 then like all the raw HTML CSS JavaScript behind the scenes or graphs that and that’s
    0:17:16 available to us. And we can reference that using variables. And here we’ve got different
    0:17:20 blocks that you can add. And I’m going to go here, you can see there’s many different
    0:17:26 types of blocks, like fetch YouTube captions. Like this is how you can grab the captions
    0:17:31 from a YouTube channel, or, you know, fetch the channel and metadata about it, or fetch
    0:17:34 all the comments. This is all if you want to customize it, right? Like I assume you’re
    0:17:39 like a regular person would not have to do, you know, expand its capabilities. Exactly.
    0:17:43 Yeah, there’s a whole bunch of pre built ones that you can just use by default. You can
    0:17:47 go build one from completely from scratch if you want, or you can take one that already
    0:17:50 exists and sort of fork it and iterate on it yourself.
    0:17:56 Right. And so, okay, so this is like one, let’s take a look at another one here quickly.
    0:18:01 Let’s take a look at this researcher that we have that people love. This one takes an
    0:18:05 average of four, almost four and a half minutes to run again, you can run it from here. If
    0:18:08 you run it from here, it says, what do you want to research? And I can say I want to
    0:18:14 research Matt Wolf and go off and do the research there, or I can run it from, you
    0:18:20 know, Chrome extension, right? And so I could be whatever on on CNN. And then I can click
    0:18:24 this thing and I can go to researcher and run it from here. And then again, it’s going
    0:18:28 to do with that, you know, people index or did to which is like, grab a bunch of things
    0:18:32 off this page and say which of these do you want me to research and I’ll present you
    0:18:37 like a menu, you know, emerging technologies and AI, investigating recent developments
    0:18:41 in artificial intelligence, but I can click this and this thing is going to kick off this
    0:18:46 like very sophisticated workflow. Let’s take a look at what that workflows is actually
    0:18:51 doing. Again, I’m going to say make a copy, which means work this, you make a copy of
    0:18:55 this. And so, as I said, these things are made up of these workflows. The main workflow
    0:19:00 here right here is called main. And it’s made up of these automation. So what is this thing
    0:19:05 doing? Well, when it gets triggered, it gets passed into it, a variable called request,
    0:19:10 that’s the user’s request. And then the first thing it does is it calls a model in this
    0:19:15 case, 3.5 sonnet. And again, we support over 50 AI models with new ones being added all
    0:19:20 the time, all the major providers, including DeepSeek and Google, OpenAI, Meta, Google,
    0:19:25 Mistral, whatever, they’re all multimodal. So you can call any model you want. And then
    0:19:30 in this case, it says analyze this request and figure out a plan to use Google search
    0:19:35 in order to do that, and then store that plan in a variable called plan. And then move on
    0:19:40 to the next step. And in this step, it says, here’s the plan that was made here. Using
    0:19:45 this plan, create a bunch of Google search queries to be able to research this plan.
    0:19:50 And then it uses another type of block, which is a run workflow block that triggers another
    0:19:56 workflow on this, this workflow called run query, and it passes into it an array of queries.
    0:20:02 And then it runs this workflow here over and over again, and searches Google, and then comes
    0:20:08 back with all of the data from that. And then it normalizes everything. And then it generates
    0:20:13 the report. And then it runs another workflow to generate images. And then it outputs the
    0:20:17 thoughts and report and emails it to you, and it’s available for you whenever you want.
    0:20:21 And so again, you don’t need to know any of the stuff in order to use it, right? In order
    0:20:25 to use it, you just push this button and say run, and it creates the stuff for you. But
    0:20:26 you can edit it and do it.
    0:20:29 Yeah, the matriarch camera, when you talked about the Google trends thing, that was like
    0:20:33 honestly, the thing that I’ve found the most interesting out of everything you’ve said,
    0:20:36 like just imagining like how that would work. And I can imagine so many different scenarios
    0:20:37 where that’d be useful.
    0:20:42 We’re monitoring your competitor’s websites and messaging. Again, the AI can just be
    0:20:46 checking all your competitor websites. And as they change their messaging, which they
    0:20:51 do, the AI can be sort of analyzing like, which way are they going? What might it be
    0:20:52 and why?
    0:20:53 Yeah, why are they doing that?
    0:20:55 Well, the stuff that like you wouldn’t have time to do and do it.
    0:20:58 It’s going to get kind of funny though, if every company is like following trends like
    0:21:01 that, like, oh, based on some data, we changed our website copy. And then, oh, well, they’re
    0:21:04 changing theirs. And then they’re all like changing it.
    0:21:05 Right.
    0:21:09 Well, it’s funny because it’s like whenever like open AI or Google or Microsoft or one
    0:21:15 of these big companies like makes a change to their terms of service, it’s news, right?
    0:21:18 People are like tweeting about it and things like that. And it’s like, you can actually
    0:21:23 have one that watches for changes of terms of service on AI related companies to see
    0:21:26 if now all of a sudden this company is going to spy on me, even though they told me they
    0:21:31 weren’t originally, you know, right, we’re simple to do use case trivially easy to build.
    0:21:36 You can build it in five minutes doing this. Okay, so I’m on your profile on Twitter X.
    0:21:41 And so these things can be contextual. And so when I’m on Twitter, the agents that are
    0:21:46 relevant to Twitter can show up right here. And so one way I could get to know Matt Wolf
    0:21:50 is by going through and like spending a bunch of time like looking through your profile,
    0:21:55 but ain’t nobody got time for that. And so I’ve got a thing called profile analyzer and
    0:21:58 it just grabs that then the AI is going to analyze you. And I could build up a bunch
    0:22:05 of things, you know, your primary focus, sort of content themes, content types, key behavioral
    0:22:10 insights, you know, so again, like a bunch of stuff about you, I’m in control, I can
    0:22:15 modify this thing to do something different because, you know, I want different types
    0:22:20 of things, or there might be conversations that are happening on Twitter. And again,
    0:22:25 I can try to like read through this and understand this thing. But that’s really hard to do.
    0:22:30 But I can just click this little button, which is X conversation analyzer, and it can do
    0:22:35 like all this work for me. It give me again, these TLDRs. And these are artifacts, like
    0:22:40 I can share them, I can save them, I can send them, I can do whatever I want with them in
    0:22:41 order to be able to do that.
    0:22:45 Yeah, one cool thing too that I don’t think you mentioned, but I know I’ve seen it playing
    0:22:50 around with Mind Studio is that the sort of visual workflow that you were showing, you
    0:22:55 can actually use natural language, and tell it what workflow you want, you can and using
    0:22:59 natural language, it will actually go and automate the building of those workflows as
    0:23:00 well.
    0:23:01 Oh, that’s awesome.
    0:23:06 Well, again, we can make yours here by creating agent, there’s actually, I wouldn’t even use
    0:23:10 this to do it, but like for somebody novice, they could do it. So again, we’ve got, you
    0:23:14 can start from a blank project, or you can start from this thing called Mind Studio Architect,
    0:23:18 where it says, what do you want me to build? And I’ll build it for you. By the way, this
    0:23:25 Mind Studio Architect was built using Mind Studio. It’s just a Mind Studio agent whose
    0:23:31 job it is to be the architect of Mind Studio, so it knows how Mind Studio works. It knows
    0:23:35 all the blocks that Mind Studio has, it knows the kind of the design patterns that Mind Studio
    0:23:37 prefers.
    0:23:40 And then if you tell it what to do, it says, I know how to follow these instructions. And
    0:23:49 so again, we could say every day, monitor my competitor’s website and check terms of
    0:23:58 service when they change, notify me with a summary of change. I think it seems pretty
    0:24:02 straightforward. I don’t need to, again, you can do that and say generate. And this thing
    0:24:06 is going to go off and it’s going to create a plan of like how it’s going to build this
    0:24:09 thing, because there’s a bunch of ways you could build it. You can say, I don’t like
    0:24:12 this plan. Choose another plan. Or you can say, great, this plan looks good enough for
    0:24:18 me. Push go, and then it’ll build all of the scaffolding for you. And many times, first
    0:24:26 time it builds it, it runs well and you don’t need to fix anything. So that happens. Sometimes
    0:24:30 it makes some weird error and you’ve got to trouble shoot it a little bit. And we’ve got
    0:24:35 a little button where you can ask people for help and they can, there are more experience
    0:24:36 than you in doing it.
    0:24:39 So you can do that. But while it’s doing that, let’s open up another tab. And I just want
    0:24:42 to start another one here quickly. I want to show like kind of the simplest one that
    0:24:48 I think will help people grok this maybe better than like some of these ones that I’ve been
    0:24:52 showing. So I’m just going to go to build. I’m going to say create new agent and my development.
    0:24:56 I’m not going to use this. I’m just going to say open blank project. So the system prompt,
    0:24:59 you can add these or not. I tend not to do them. But if you want to do them, this again
    0:25:03 sort of tells the system like what you want it to be. You can write it yourself or again,
    0:25:07 you can use another agent that we have here to generate a prompt.
    0:25:10 I’m still thinking about the whole Google Trends thing. Sorry, my mind stuck on that.
    0:25:15 I know, you know, another HubSpot podcast, My First Million, they’ve had so many episodes
    0:25:18 talking about, Hey, if you want to start a new company, one of the main ways to do it
    0:25:19 is to look at Google Trends.
    0:25:25 You can just automate, you can be automating, I’m glad that that connected. Yeah.
    0:25:30 Because that is a profound superpower. Yeah, it sounds like you can actually leverage
    0:25:36 Google Trends, leverage that data at scale. It’s profound. It’s an insanely unfair competitive
    0:25:38 advantage with anybody who can’t do that.
    0:25:41 Yeah, sorry to take you off track. I just like that was still in my mind. I’m like,
    0:25:42 I want to go try that.
    0:25:46 Okay, so let’s build the Chrome extension-y thing. Okay, I’m going to go here, start.
    0:25:51 I’m going to say, you know, browser extension, again, we now get these variables. I’m going
    0:25:55 to say plus, and I’m just going to say generate text. This means call a model. In this case,
    0:25:59 I’m just going to change a model just to change it. Why not? We’re going to use Gemini 1.5
    0:26:06 Flash, and we’re going to tell this model, you know, create a TLDR of this. Okay. And
    0:26:10 then in double curly braces, that’s as complex as it gets. We’re going to pass it this variable
    0:26:16 here that we get again called full text. Full text means the full text of the page from which
    0:26:19 we call this. There it is. It’s case sensitive. And here we can either display it to user
    0:26:23 or save it to variables. I’m just going to leave it display to user because we’re done.
    0:26:28 We just made this agent that does a TLDR. Again, if we want it to be fancier and we want it
    0:26:32 specifically say how we want it to create the TLDR, what should it look for? What should
    0:26:37 it buy us for? Are you just have this in plain English over here, right? And then, you know,
    0:26:43 you can click here and you can give it a name. I’m just going to call this maths TLDR. And
    0:26:47 then you can give it an icon if you want. And then you can just say publish. And so again,
    0:26:52 we just built this thing. It exists. It’s got a URL. We can run it. We can run it manually
    0:26:56 and just give it a URL, but that’s no fun. That’s not how we built it. We want to run
    0:27:03 it from whatever, this one here. And so we can go here and we can open our extension
    0:27:07 right here. And like a fast way to find it because I’ve got so many of them is to simply
    0:27:11 do a search, maths TLDR. And then I can just run it from here. And again, we built this
    0:27:15 thing in what less than a minute right here. And you know, this says Microsoft was made
    0:27:19 a major breakthrough in this. Okay, so that’s not a very good TLDR. We could go back and
    0:27:25 say no, no, like these are specific things that I want out of my TLDR and do I’m not
    0:27:28 going to demonstrate that because anyone can do that. You can play around and then you
    0:27:32 could save and you go back and you test it again, or you’ll go back here and test it
    0:27:36 again. It’s that easy to sort of like iterate on these things to create custom tools for
    0:27:37 yourself.
    0:27:42 Yeah. And then you can just go and test different models to maybe Gemini 1.5 flash tends to
    0:27:47 give you smaller responses where if you were to use, you know, GPT 40 or sonnet 3.5, it
    0:27:49 will give you a much deeper response.
    0:27:53 Absolutely. In fact, let’s do that. Let’s use sonnet here instead. We’re going to publish
    0:27:57 that and we’re going to rerun the exact same thing. And again, like it, it’s super easy
    0:28:01 to rerun these things. I should take the one we just made. So I don’t have to look for
    0:28:09 it. Matt’s TLDR. I can click on this thing. And over here, I can say add to start. Okay.
    0:28:13 So this one didn’t do much more on the TLDR, but you can change models. Again, if you give
    0:28:17 it some more instructions, it doesn’t matter a much better job. Another way to get it to
    0:28:21 do stuff like that, that’s even better. I can say generate prom. You know, we like to
    0:28:24 refer to this thing that usually is assistant, but you can say anything you want, write it
    0:28:34 in any form you want. Assistant generates TLDRs for human when provided with content.
    0:28:38 And I could say from the web, or I can say content, whatever, and like to say generate.
    0:28:42 And this thing will go off and it’ll like generate the prompt that will be much more
    0:28:48 specific about what a TLDR should have. And that tool will now bias the TLDR that we get
    0:28:54 on that. But the point is like all of this stuff is like really easy to do and can play
    0:29:00 around with. Yeah, for sure. Now, what’s up with the one that you started using the automation?
    0:29:03 You use the natural language to build that one. Let’s check on that one. Let’s check
    0:29:08 that out. Okay. So here’s what it did. It built the terms of service change monitor.
    0:29:12 And again, this is the plan automatically monitors competitors terms of service, page
    0:29:17 for changes, analyzing differences, sending email notifications. It says it’s going to
    0:29:23 take, you know, is input the URL in the notification email. And then here are the variables it’s
    0:29:28 going to use. It doesn’t need any custom functions that it needs to write. This is what the automation
    0:29:32 is going to look like. It’s going to scrape the terms of service. And going to clean and
    0:29:37 normalize content, check if it’s the first run, you know, or if it’s got data about it
    0:29:41 already, save the version, you know, and the first run analyze change. Again, we say great,
    0:29:46 whatever this means, sounds good to me. And then you say go, and it’ll go off and it’ll
    0:29:52 build that entire workflow, we’ll get that workflow again, it might work with just we
    0:29:56 have seen to put in the URL, because we didn’t tell it what URL, we could have told it what
    0:30:01 URL. Okay, so workflow generation is complete. I can, you know, watch a YouTube tutorial
    0:30:08 on how to use it, or we can close and start editing it. Again, it built the system prompt.
    0:30:13 It built the automations for us that we can then look at and we can edit. And so this
    0:30:18 is using our scrape URL block, and it’s taking in here again, a variable called URL that
    0:30:23 we need to define. So here, it says this thing is scheduled, and it gets past these like
    0:30:28 launch variables. Or if it’s the same URL, we can just put it in here explicitly. So
    0:30:34 we can say, you know, CNN.com slash terms of service, you get the point of that.
    0:30:35 Right, right.
    0:30:38 Yeah. Dimitri, I’m kind of wondering, like the more you talk about this, like, long term,
    0:30:42 I’m curious, like, how does this, you know, obviously for open AI agents, it’s going to
    0:30:47 be very, very important. Like, you know, and I’m kind of curious, is your thesis that
    0:30:52 even if they really, really focus on agents, and of course, they get amazing at it, that
    0:30:54 there’s still going to be a benefit of having like a community of people sharing their different
    0:30:59 use cases and different templates, is that kind of the idea behind Mind Studio?
    0:31:03 Yeah, look, we believe that models will continue to get better and better. And certainly model
    0:31:10 providers like open AI will have sort of consumer interfaces along with API interfaces and those
    0:31:16 consumer interfaces, you know, today are primarily conversational, so chat GPT, and it’s got
    0:31:21 access to, you know, tools, function calling, etc. And like, in your chat, you can sort
    0:31:25 of go back and forth with it and like does stuff for you. And there’s a lot of use cases
    0:31:30 for that where like ad hoc, you want to do something, and you do that. But for, you know,
    0:31:34 many other things, and I propose most other things, you will say, well, I wish like something
    0:31:39 just like I could build it and it just ran and it did these things like monitor Google
    0:31:44 trends. I don’t want to constantly ask chat GPT to do it for me, or talk to it. Right.
    0:31:48 I want it to only notify me, by the way, does it in the background, it should only nudge
    0:31:52 me if it found something interesting. Most of the time it won’t. And so just to be there
    0:31:56 in the background, I’ve got a dashboard, I can see these things running, I can see what’s
    0:32:01 scheduled, etc. As for like the capabilities of the models themselves to sort of be at
    0:32:06 a place where, again, you kind of don’t need anything like this, and you just tell them
    0:32:11 I want you to be this thing and it sort of does it, I struggle from a product side to
    0:32:18 see how you would sort of have the human feel enough in control without sort of creating
    0:32:22 something that they could understand. Again, you might generate the workflow for them just
    0:32:28 like our architect did, but I think humans still being able to say, Oh, I want to put
    0:32:33 like a checkpoint in here. Yeah, modify how it works. And for this, or I want to be able
    0:32:38 to change this in some way. You know, I didn’t show you guys our debugger, which is awesome,
    0:32:43 like extraordinarily powerful. So like our instrumentation for like building and running
    0:32:50 these things is dramatically deeper than any of the model providers themselves offer. And
    0:32:53 so if you’re going to build anything with models, we think this is the no brainer way
    0:32:58 to be it’s it’s the most instrument that it’s the fastest it’s the easiest gives you access
    0:33:02 to all those models, models go up and down. Yeah, I was going to say it feels like that’s
    0:33:05 a big advantage to you right is like being agnostic because I mean, yeah, in the last
    0:33:09 month we had deep seat come out now grok’s amazing. There could be a lot of options
    0:33:13 out there. Well, as you’re building these workflows, again, you want to with for each
    0:33:18 step where you want to use intelligence, you might want to call a different model, right?
    0:33:23 There you know, Gemini 1.5 flash is super fast at being able to do a bunch of stuff,
    0:33:28 but it’s not a very good writer. Like Claude is a much better writer. So you might get
    0:33:32 the Gemini to do a bunch of analysis data collection, all of that and then sort of hand
    0:33:38 that off to Claude to put together and do or like generating images or like whatever else
    0:33:43 is doing here. So like this ability to analyze different models for each part of the step,
    0:33:49 use the right model based upon quality, latency, so low latency and cost. Yeah, and you probably
    0:33:52 could suggest that eventually too, right? Like this model is better for writing here.
    0:33:57 I want to use this one. Yeah, it’s like suggested in those like detail pages that each one of
    0:34:02 them has you can read them. But sort of what I was going is like, these models aren’t reliable.
    0:34:06 These model providers are still building their infrastructure. They’ve got a bunch of usage.
    0:34:10 So they go down periodically, in fact quite frequently. And if you’re just like standardized
    0:34:15 on only open AI and all you’ve got is there, you’re kind of stuck your AI and doing it.
    0:34:18 Here you can just fail over to any of these other models that could also do the job. They
    0:34:22 might not be preferred. They might be a little bit more expensive. They might be a little
    0:34:25 bit slower, but at least you’re not down in your workings like all of those kinds of things
    0:34:30 become really important when you get serious about like building these things. So are those
    0:34:35 like redundancy mechanisms or those built in already like try Claude 3.5 if that doesn’t
    0:34:41 work, then go do GPT 4.0 or whatever it doesn’t automatically do it. So like can notify you
    0:34:46 that models are down, which we do. And then you can manually go do that. But coming soon,
    0:34:49 it will do it intelligently and just like do that for you. That’s cool. Like here’s
    0:34:54 my tier list, you know, for writing, for coding, for whatever you sit in this order, the infrastructure
    0:34:58 is there. You can now just manually choose a different one takes 10 seconds. But soon
    0:35:02 it will be autonomous for sure. Yeah, I did have a question. This was something you showed
    0:35:06 off on the screen earlier in the conversation, where you showed, all right, it’s going to
    0:35:10 cost like six cents to run this, right? How does that work? Is it like a bring your own
    0:35:16 API key? Is it like you guys bill based on the usage? Where does that six cents come
    0:35:20 from? Yeah, so you don’t need to bring your own API key. Although if you want to call
    0:35:25 your own models, we allow again function calling an API call. You can call any model including
    0:35:30 on premise, private cloud, etc. The vast majority of people that use mine studio use
    0:35:35 our API keys that are built in so you don’t need any API keys. So you instantly have access
    0:35:41 to all of the models from all of their providers on our billing. And we simply pass the costs
    0:35:48 on to you based on usage. Okay, we take a tiny markup 2.9% and then it reduces with scale
    0:35:55 and goes down to like 0.15% of overage over what the model providers charge us. It’s kind
    0:35:59 of like Stripe. We’re going to see ourselves very much in that light. We’re sort of like
    0:36:05 what Stripe is to payments or Twilio is to telephony or moxes to video. Mine studio is
    0:36:10 to intelligence. Make sense. Make sense. Have you ever considered doing it more like cursor
    0:36:14 where like you have like a monthly and that’s basically baked in like you’re kind of assuming
    0:36:18 they’re going to use the API to my mouth with some limit. And then you go over that it’s
    0:36:21 like, okay, you got to either pay us more or use your own API key. Yeah, I would like
    0:36:24 that with cursor. Then I don’t have to worry about doing something and like, oh, right now
    0:36:27 I’m paying something. I’m like, oh, I’m not paying anything. I’m just using it. You know,
    0:36:30 and so that’s kind of feels nice to me. I don’t know. Great. Well, I’ll break the news
    0:36:40 on your show here. We’re about to add that here. I can predict the future when this publishes.
    0:36:44 It might already be there. Yes, exactly. That makes sense. But so I wanted to mention sort
    0:36:50 of another thing. So we’ve got this map as you noticed that this is like an app store.
    0:36:55 And these agents are put in there today by us. We created all of the agents that are
    0:37:00 available for you to use. You just pay metered usage form. They’re free to use, but other
    0:37:06 than metered usage, you can make copies of them and modify them and launch new ones,
    0:37:11 etc. And we will adding more to it. The other thing that we’re doing is we’re opening this
    0:37:16 up to the world just like the app stores opened up to the world that people can show up and
    0:37:20 say, I’ve got some use cases that I think are awesome, and that lots of people are going
    0:37:26 to want. I’m going to build them in my studio. And again, as you can see, it’s easy to build.
    0:37:30 And then I’m going to submit them to the store. And I want to make them available to other
    0:37:35 people to just use by venturing also like charge form, obviously, we’ll implement that
    0:37:40 as well. And so we’re actually looking for people that want to show up and start building
    0:37:45 these things. And we’re venture backed. And so we’re willing to pay like bounty. So there’s
    0:37:51 an opportunity for enterprising folks to get in early and fill up this app store with all
    0:37:57 kinds of amazing utilities that now can be triggered from a browser extension. If you’re
    0:38:04 building good ones, we’ve got, you know, praise and money to make your life better. And we’re
    0:38:08 also looking to hire people full time to do this. There’s like great opportunities for
    0:38:11 people to start building these things and take an advantage.
    0:38:15 Very cool. When you say app stores, they’re going to be like a revenue share or?
    0:38:19 Absolutely. Yeah. You’ll be able to charge for these things and you’ll be able to earn
    0:38:21 revenue from the work that you’ve done. Yeah.
    0:38:25 Yeah. Very cool. By the way, people are already building these things and putting them behind
    0:38:29 paywalls, which you can do. And so their entire businesses that have been started that simply
    0:38:34 build a bunch of Mind Studio AI agents, put them behind the paywall and then charge monthly
    0:38:37 access to all of those agents.
    0:38:41 Yeah. I remember early on, I think when we first talked, you know, 18, 19 months ago or
    0:38:45 something, that was kind of like the original sort of selling point of Mind Studio is go
    0:38:47 build agents and then sell access to them.
    0:38:51 Yeah. So people took us up on that and are doing it. And so, yeah, you can make money
    0:38:57 with them already. By the way, we also now have over 150 integration partners that make
    0:39:04 money by again, they’re experts in Mind Studio and they are building AI agents for organizations.
    0:39:08 So we’ve got over 150,000 of these things that have been built and deployed. Again, some
    0:39:13 are just like experiments, they’re toys, but many are like mission critical, really important
    0:39:18 AI agents that sit at, you know, government agencies, the His Majesty’s Revenue and Customs,
    0:39:26 the British IRS, uses Mind Studio and recruiting, you know, Service Now uses it in sales, demo
    0:39:30 automation, a company called Advanced Local, which is a giant publishing company, has got
    0:39:35 over 80 of these running that are automating newsrooms and newspaper publishing. And so
    0:39:41 some of those are built by the people inside these companies, others, these companies hire
    0:39:46 our partners to do professional services work and build them. So again, there have been
    0:39:53 entire actually agencies that have like started off simply to do Mind Studio implementation
    0:39:58 or organizations that want the skill. That’s where all the budgets are flowing, you know,
    0:40:02 in enterprises, they’re all flowing towards this like AI digital transformation.
    0:40:08 Right. And we make it really easy for them to get this giant ROI out of very little work,
    0:40:09 you know.
    0:40:10 Very cool.
    0:40:11 Very cool.
    0:40:14 Absolutely. Is there anything else about Mind Studio that we should touch on that we haven’t
    0:40:18 touched on yet, or even something outside of Mind Studio that’s really, you know, got
    0:40:23 your attention right now that you’re really excited about just in the AI or tech world?
    0:40:26 That’s a deep question. It was like another podcast, right? That’s another hour.
    0:40:32 Yeah, no. I mean, I’m assuming your head’s pretty much always in Mind Studio playing with
    0:40:36 it, making it better. So, you know, you may not be going and playing with all the other
    0:40:39 tools and, you know, stuff that’s out there.
    0:40:43 Well, I get to watch your videos and get you doing it. I don’t want to make it political
    0:40:47 because there’s a bunch of interesting stuff there. Like what’s a fact checker? Well, fact
    0:40:53 checker is a really important tool that, you know, citizens of democratic countries need
    0:40:58 to be able to understand what’s reality. And now that tool is a click of a button. It has
    0:41:00 broad implications of what that means.
    0:41:05 Yeah, yeah. Well, there’s a site called Ground News. I don’t know if you’ve heard of it or
    0:41:09 not, but it’s like a news website that curates all of the latest news from like all of the
    0:41:15 various news sources. And then when you’re actually looking at the site, it’ll tell you,
    0:41:19 does this article and this news source tend to lean left? Does it tend to lean right?
    0:41:24 Does it tend to be a little more center? And it sort of like analyzes every piece of news
    0:41:29 and tell you like where that news falls on the political spectrum. And I feel like with
    0:41:34 Mind Studio, you can build that, right? Go to any news website, click a button. It’ll
    0:41:38 tell you whether this site tends to lean left, tends to lean right and actually look for
    0:41:42 bias in the news article for you, right? Like you don’t need to get political. It doesn’t
    0:41:47 matter what sort of political leanings you have. That just seems helpful to anybody.
    0:41:49 Pick your bias.
    0:41:53 We have AI agents that have already been built that do specifically that, that are sentiment
    0:41:58 analyzers and bias analyzers. You push a button and it looks at the article and says, Hey,
    0:42:01 here are things that might be inaccurate. Here are things that are misleading. Here
    0:42:05 are things that are clearly biased where they’re showing only one perspective. There’s others
    0:42:08 that have a counterpoint and so it’ll be like, great, what’s the opposite of that? Like what
    0:42:13 would be the counterarguments? And so again, these superpowers that you now get with click
    0:42:19 of a button that AIs can do really easily. But once they’re like made so easy to take
    0:42:23 along with you, that’s kind of the big unlock. Yeah. Yeah. Wherever you are, you now can
    0:42:29 like instantly trigger them and get that instead of like going to chat GPT and asking it to
    0:42:30 do something.
    0:42:31 Right.
    0:42:35 Super cool. Now with the little like Chrome extension to any plans to make that available
    0:42:39 on mobile as well, like will I be able to be reading a news article on my phone and do
    0:42:40 the same kind of things?
    0:42:45 Yeah. We’re going to have a native app. Probably I’m guessing in a quarter, which will allow
    0:42:50 you to basically make it a shared target. So you can like intent out from any application
    0:42:54 to any of these agents and get callbacks on that. So like it instantly sort of enables
    0:42:58 anything that can do share intents to now have
    0:42:59 Gotcha.
    0:43:03 Available for it. There are some, I can’t sort of announce them yet, but there are medium
    0:43:10 sized social networks that are starting to integrate mind studio agents into their interfaces.
    0:43:15 So they will be within their native mobile apps. There will be buttons on their posts
    0:43:20 that allow you to fact check things or like do other things. And those are simply mind
    0:43:24 studio agents that are doing that work. And so you’ll see, you know, certainly a lot
    0:43:26 more of that kind of stuff that you can do.
    0:43:30 Well, cool. This has been a really, really interesting conversation. I mean, I love,
    0:43:33 I love actually getting to see all this stuff in person. I’ve been playing around with Mind
    0:43:38 Studio for a little over a year now and have messed with it quite a bit, but it’s always
    0:43:42 really cool to see how like the person who’s actually building the tools would actually
    0:43:48 use it as well. So I really appreciate all the insights and you showing off how to use
    0:43:49 it.
    0:43:54 The website is mindstudio.ai. Is there anywhere else people can follow you? You on X, you
    0:43:58 on LinkedIn, where’s the best place to sort of get to know you?
    0:44:03 I’m on all of them. I prefer LinkedIn and please feel free and add me on LinkedIn. I’m easy
    0:44:08 to find Dimitri Shapiro. And yeah, that’s probably the best place to follow me, but
    0:44:14 I’m also on X and Facebook and Reddit and everywhere else. You gotta be everywhere.
    0:44:19 Very cool. Well, thank you once again for joining us today. This has been a really cool conversation.
    0:44:23 Really, that’s listening. If you haven’t already, make sure you subscribe on YouTube
    0:44:28 for more interesting conversations like this, especially if you want to get the visual element.
    0:44:32 And if you prefer audio, you’d like listening in your car or wall out walking your dog.
    0:44:36 We’re available wherever you listen to podcasts. So subscribe to us there. And thanks again,
    0:44:38 Dimitri. It’s been an amazing episode.
    0:44:39 Great to chat.
    0:44:57 [Music]
    0:44:59 [Music]
    0:45:09 [BLANK_AUDIO]

    Episode 47: How can AI agents automate your daily tasks in just 60 seconds? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) are joined by tech innovator Dmitry Shapiro (https://x.com/dmitry), CEO of MindStudio, a tool designed to make building AI agents seamless and accessible.

    In this episode, Dmitry delves into the world of AI agents, explaining their potential and showcasing Mind Studio’s capabilities. He demonstrates live how to create and customize AI agents, offering insights into automation that can revolutionize everyday tasks. Whether it’s analyzing trends, monitoring competitor changes, or automating news consumption, Mind Studio provides AI superpowers at your fingertips.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Building the Social Graph
    • (03:29) Tech Entrepreneur’s Journey and Ventures
    • (06:48) Navigating AI Autonomy Limits
    • (12:22) Automated Entity Analysis Tool
    • (15:26) Automated Workflow Trigger Methods
    • (18:48) Automated Research and Reporting Workflow
    • (20:40) Profile Analyzer Simplifies Twitter Insights
    • (25:30) Building a TLDR Agent
    • (29:11) YouTube Tutorial and Script Editing
    • (30:59) Optimizing AI Integration with Control
    • (35:57) Open AI App Store Expansion
    • (37:59) AI Integration Success with MindStudio
    • (41:53) MindStudio App Integration Update
    • (43:26) Subscribe for More Cool Conversations

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • Will AI Replace SaaS? The Future of Software

    AI transcript
    0:00:04 [MUSIC]
    0:00:06 Hey, welcome back to the Next Wave Podcast.
    0:00:06 I’m Matt Wolf.
    0:00:08 I’m here with Nathan Lanz.
    0:00:13 Today, we’re going to talk about the crossover of software and AI.
    0:00:15 We’re bringing on a guest expert,
    0:00:17 Dan Shipper to talk all about it.
    0:00:19 Dan is the CEO of every.
    0:00:22 He has all sorts of companies underneath him that he’s incubated.
    0:00:25 He’s an investor. He works with Sequoia.
    0:00:28 He has all sorts of knowledge in the software space,
    0:00:31 and he’s going to break down his exact process
    0:00:33 for deciding what products to build,
    0:00:35 and what tools he used to build them,
    0:00:37 and exactly how he brings them to market.
    0:00:38 It’s a great discussion,
    0:00:41 so let’s just go ahead and dive right in with Dan Shipper.
    0:00:44 Thanks so much for joining us today, Dan. How are you doing?
    0:00:46 I’m doing great. Thanks so much for having me.
    0:00:49 Yeah. I’m excited to really dig in with you and what you’re doing over at
    0:00:52 your various companies that you’re working on.
    0:00:54 Let’s go ahead and lay the groundwork.
    0:00:57 Tell us a little bit about your company and what you’re building,
    0:00:59 and what is every?
    0:01:00 Yeah. That’s a good question.
    0:01:01 We’ve been around for five years,
    0:01:02 so it’s evolved significantly over time.
    0:01:05 The way that I talk about what we do right now is,
    0:01:07 I call us a multimodal media company.
    0:01:10 We publish writing, we publish podcasts,
    0:01:12 we publish a video, and we publish software.
    0:01:13 We do all that using AI,
    0:01:17 and we’ve gone in that direction because I think that there’s
    0:01:20 this really interesting thing happening where the line between
    0:01:24 being a writer and being a builder is starting to blur,
    0:01:27 and they’re moving into this overlapping Venn diagram,
    0:01:30 because now that you can write in English and turn it into code,
    0:01:32 writing is building.
    0:01:35 Also, because software is so cheap to make,
    0:01:37 I think that software is becoming content.
    0:01:41 As a media company, what we focus on at every is,
    0:01:42 we have a daily user,
    0:01:45 every day we write a long-form essay about what comes next in tech.
    0:01:49 We’ve got about 10 regular writers who do that.
    0:01:51 I write every Friday a column called
    0:01:53 Chain of Thought on the intersection between AI,
    0:01:56 creativity, psychology, philosophy, all that stuff.
    0:01:58 Our job is to explore the future,
    0:02:02 and the best way to do that in my opinion is always to be using
    0:02:05 new tools and experimenting with things and building things and all that stuff.
    0:02:08 What we found is that when we experiment with something,
    0:02:10 if we release it to the audience,
    0:02:13 people love it, it makes the writing better.
    0:02:18 Sometimes those experiments become real products and should be real products.
    0:02:21 What we’ve done is then made them into real,
    0:02:24 actual products that we often start by using them internally,
    0:02:25 and then we’ll post them to the audience,
    0:02:28 and then we bundle them into the subscription so you pay one price,
    0:02:31 and you get access to all of our content and all of the software that we make.
    0:02:33 We’ve done four incubations so far.
    0:02:37 The most recent one, which we just launched about three or four weeks ago,
    0:02:39 right before Christmas is called Cora,
    0:02:41 and it manages your inbox with AI.
    0:02:43 Basically, what happens is as emails come in,
    0:02:45 it decides if you need to respond to them.
    0:02:47 If you do need to respond to them,
    0:02:48 if they’re from a human,
    0:02:50 it makes it to your inbox and you see it immediately.
    0:02:53 If not, it automatically archives it and then twice a day,
    0:02:56 you get a beautiful brief that
    0:02:59 summarizes all the things you need to know about what’s going on in your inbox,
    0:03:00 but you don’t need to actually do anything with,
    0:03:04 and it just saves a ton of time because 80-90 percent of your emails are like that.
    0:03:09 We have a bunch of other products that are in that vein for different parts of your work life.
    0:03:11 That’s the basics of every.
    0:03:13 >> That was Cora, like C-O-R-A?
    0:03:15 >> C-O-R-A, yeah.
    0:03:16 >> Okay, cool.
    0:03:17 >> Yeah. It’s been going really well.
    0:03:18 We’ve got like 10,000 people on the wait list,
    0:03:21 we’re taking off like 20 people a day and just starting to ramp it up.
    0:03:24 >> I’m on the wait list, so I can’t really talk about it, I haven’t tried it.
    0:03:27 >> I’ll get you off the wait list, you should have told me, I know a guy.
    0:03:31 >> Cool. That’s fascinating thing about writing is becoming software,
    0:03:31 is that what you said?
    0:03:33 >> Writing is building and software is content.
    0:03:35 >> Yeah, okay. I’ve been noticing that too,
    0:03:36 because it’s part of this doing this podcast,
    0:03:38 I’ve been trying to get more hands-on with AI,
    0:03:39 like I actually started building some stuff.
    0:03:45 Recently, I started using all the different AI tools to develop a game prototype just for fun.
    0:03:46 Maybe it becomes a real thing, I don’t know,
    0:03:50 but it is fascinating that I am now talking with AI,
    0:03:53 and kind of writing my game design document with AI,
    0:03:56 using my voice, using something like whisper flow,
    0:03:58 chatting with it, having it update the document,
    0:04:00 and then I feed that document back into
    0:04:01 01 Pro when I’m coding on it,
    0:04:05 and it’s just amazing that that conversation I’m having with AI,
    0:04:08 then forms the actual code and software that is created.
    0:04:09 That’s never existed before.
    0:04:10 >> Isn’t that so cool and like,
    0:04:13 so you can code while you’re walking around and just like,
    0:04:17 you know, recording to your voice and Kiran who is our GM of Cora and he built Cora,
    0:04:20 like that’s how he started the first Cora MVP.
    0:04:21 He built that in a single night,
    0:04:24 and he did it by going and take a walk,
    0:04:26 and he just like recorded himself being like,
    0:04:27 this is how I want the app to work,
    0:04:30 and he threw it into 01 and just like, it’s awesome.
    0:04:31 I think like- >> It’s so cool.
    0:04:35 >> There is a big difference between you can build a prototype,
    0:04:37 and then actual like production-grade software.
    0:04:38 >> Yeah.
    0:04:39 >> But even the production-grade software,
    0:04:42 it’s just changed what you can do and how fast you can do it,
    0:04:43 and it’s really fun.
    0:04:44 >> Yeah.
    0:04:45 >> I really like what you said there.
    0:04:47 So writing is building, that makes sense to me, right?
    0:04:50 Like everything is kind of becoming prompting to get what you want, right?
    0:04:53 If you want to write software, you prompt it,
    0:04:54 the software can come into existence.
    0:04:56 I mean, even now, like we’re talking about,
    0:04:59 you can even speak it into existence and then it actually becomes software.
    0:05:02 Can you speak more to the software is content?
    0:05:03 What do you mean by that?
    0:05:07 >> Yeah. So software is much cheaper to make now,
    0:05:10 and one of the things we found is that if we release
    0:05:12 a little experiment as a piece of software,
    0:05:16 like all of the experiments we run are much better at
    0:05:21 attracting audience and emails than any given essay or YouTube video or whatever,
    0:05:25 because it’s this totally new format of thing, right?
    0:05:29 So in the same way that on Instagram,
    0:05:30 you can take pictures and TikTok,
    0:05:33 you can take videos on YouTube, you can take videos,
    0:05:35 and those are different from like Hollywood films,
    0:05:36 but they’re still videos.
    0:05:39 I think that we’ve been living in the world of like software is
    0:05:42 like a Hollywood movie like you need a lot of money and time to make it.
    0:05:44 Now, with all these tools,
    0:05:49 it’s expanding beyond people who are professional developers and anyone can make a little prototype,
    0:05:53 and that switch is like anyone can write a blog post,
    0:05:55 anyone can do a tweet, it’s going to be the same thing.
    0:05:57 There aren’t really platforms yet for this,
    0:06:00 but I think there’s going to be like a YouTube of software development,
    0:06:02 maybe like Replet is that to some degree,
    0:06:04 and it’s not for professional coders,
    0:06:06 it’s for people who just grew up prompting,
    0:06:08 and that’s happening right now.
    0:06:10 That definitely makes a lot of sense.
    0:06:14 I also feel like this is something we’ve actually talked about on the show in the past too,
    0:06:18 is like I almost feel like software could be the next SEO,
    0:06:24 because I feel like just writing blog posts and writing content is so easy now,
    0:06:27 like anybody can do it because of tools like Claude and chat,
    0:06:29 GPT and Gemini and everything that’s available out there.
    0:06:34 I think the value of written content is starting to go down,
    0:06:35 and when it comes to SEO,
    0:06:38 one of the goals of SEO is to bring somebody to your site
    0:06:40 and keep them on your site as long as possible.
    0:06:44 It almost feels to me like really good software that people find value in,
    0:06:48 will be the new thing that keeps them around longer on your platform
    0:06:53 and help your site rank better and help you actually grow your business a little more.
    0:06:57 I’m curious if you have the same thoughts or if you would disagree with that.
    0:07:00 That’s really interesting for one,
    0:07:04 like I love words and I think the written word is always going to be super, super valuable.
    0:07:06 I do think there’s something to what you’re saying,
    0:07:08 which is that to some degree,
    0:07:10 a lot of the content that gets made right now,
    0:07:15 like I just saw this thing about HubSpot’s traffic and HubSpot’s traffic is just going down,
    0:07:19 and so a lot of the SEO type content that’s made right now,
    0:07:22 is probably better served by a chat GPT.
    0:07:25 It’s like, okay, how do I ask for a raise?
    0:07:30 There’s zillions of those SEO articles and that’s much better from chat GPT,
    0:07:32 especially the more generic stuff.
    0:07:36 I still think that for every, for example,
    0:07:39 the audience that we have wants to hear from me,
    0:07:42 and wants to hear from other writers that they know,
    0:07:47 and are not looking for a generic fact or a generic process,
    0:07:49 and that’s a different thing.
    0:07:50 That’s a definitely smaller audience,
    0:07:54 but it’s still incredibly important and valuable.
    0:07:56 I think on the tools front,
    0:08:00 you may actually see some of the same commoditization because you don’t need to go use
    0:08:04 this ass tool when artifacts just spins up a little UI for you,
    0:08:06 and you can just do whatever.
    0:08:10 That may also happen with software to some degree,
    0:08:12 and then I think you also have to wonder,
    0:08:14 “Okay, if software tools are the new SEO,
    0:08:18 what is the thing that is pushing people to those tools?
    0:08:20 Is it Google still?
    0:08:21 Is it chat GPT?
    0:08:25 If it’s chat GPT, how and why is chat GPT letting those things surface and rank?”
    0:08:33 Right. But I do think that providing people with rich multi-modal experiences,
    0:08:34 that’s not just text, it’s not just video,
    0:08:35 it’s maybe it’s tools or whatever,
    0:08:39 that is I think going to be a really legitimate way to build an audience and build a following.
    0:08:42 Yeah. I totally agree too that when it comes to the world of content,
    0:08:46 I think as AI becomes more and more prolific,
    0:08:50 and everybody’s using AI to try to gain the system through SEO and things like that,
    0:08:53 I think it becomes more and more important to have your own voice,
    0:08:56 and to actually be a voice that people can trust,
    0:08:58 and actually put your face out there.
    0:09:02 There’s a lot of stuff you’re seeing on YouTube now of like faceless video channels,
    0:09:05 and people are claiming it’s all the rage right now to make faceless video channels.
    0:09:07 But in my mind, if that’s the direction everybody’s going,
    0:09:12 the people that actually put their face on YouTube are the ones that are going to rise and stand out.
    0:09:15 The sacrifices we make for fame.
    0:09:18 You’re in San Francisco, right Dan?
    0:09:21 No, I’m in New York. I go there often, but I live in New York.
    0:09:22 I lived in San Francisco for about 10 years.
    0:09:24 I’m out in Japan now. That’s one thing.
    0:09:27 But nice being here is, you know, even if the podcast blows up in the future,
    0:09:31 I don’t have to worry about fame. Like, nobody will notice who I am.
    0:09:34 That’s why I wear this like media mogul hat that HubSpot gave me.
    0:09:36 I think it’s hilarious because obviously I’m not a media mogul,
    0:09:38 but it’s kind of funny where you’re in Japan.
    0:09:41 You know, I don’t know if you know this is a HubSpot-owned podcast.
    0:09:43 I’m not sure if you’re aware of that. You were talking about HubSpot, this traffic.
    0:09:46 I am aware because the emails are from HubSpot.
    0:09:50 Oh, shoot. Yeah, yeah, I did not put that together.
    0:09:53 Sorry, I didn’t. I mean, that is one reason to do this podcast, right?
    0:09:55 Now, this is even I think kind of like what’s part of our page too is like, you know,
    0:09:59 having faces is going to be very important in the future versus just generic content on,
    0:10:02 you know, text content that anyone can throw out and AI is going to get better and better at that.
    0:10:04 So SEO is going to be harder in the future.
    0:10:08 But I do wonder why would chat to BT send you to other tools in the future?
    0:10:11 If it gets so easy that they could just like create it for you on the fly?
    0:10:14 With like what we’re seeing right now, like I’ve been using O1 Pro a lot and like,
    0:10:16 yeah, seeing how good it is at software.
    0:10:20 I can’t imagine in another two or three years that any idea you have
    0:10:26 AI is now just like, cool, here it is. Here it is right now in the window and just use it.
    0:10:29 Totally. It’s interesting because when it comes to the whole like HubSpot conversation,
    0:10:32 I do think, you know, maybe HubSpot’s traffic is going down.
    0:10:36 But, you know, podcasts like this and what they’re doing in like other media spaces,
    0:10:39 I think is like, they’re sort of answer to that, right?
    0:10:42 Like you’ve got podcasts like the next wave and my first million and marketing,
    0:10:46 it’s the grain and the hustle and they’re starting to do all their newsletters now
    0:10:48 with the hustle daily and, and all that kind of stuff.
    0:10:51 So it’s like maybe less people are going to the HubSpot website,
    0:10:54 but HubSpot’s becoming like this larger media brand.
    0:10:54 Totally.
    0:10:57 So I don’t think there’s anything wrong with saying that, you know,
    0:11:01 HubSpot’s traffic is dropping because they obviously have like, you know,
    0:11:03 backup plans that they’re building as well.
    0:11:05 But just wanted to throw that out there.
    0:11:09 I feel like I just like went to someone’s dinner party and like spit in soup or something like that.
    0:11:11 It’s okay.
    0:11:14 I knew in one part of my brain that this is a HubSpot podcast.
    0:11:19 And then, but anyway, um, yeah, I’m not trying to imply that HubSpot is going to die or anything
    0:11:20 like that.
    0:11:23 It’s just the whole SEO game is like changing rapidly.
    0:11:23 Right.
    0:11:29 And yeah, it seems like shows like this are a really good bulwark against those kinds of changes,
    0:11:31 which have been happening for a long time.
    0:11:32 It’s not just AI.
    0:11:37 And I think there is that interesting dynamic that, you know, you can see open an AI trying
    0:11:42 to react to or trying to head off, which is like, you know, they do deals with publishers
    0:11:45 and they cite their sources and all that kind of stuff.
    0:11:49 And I do think that they would like to keep a vibrant ecosystem and app partners and all
    0:11:50 that kind of stuff.
    0:11:57 And I also think you can see with Google’s example that like, over time, the pressure
    0:12:03 is to keep people on the platform longer and add more ads or like do the like instant answers
    0:12:04 or whatever.
    0:12:10 And I think AI broad use case general purpose AI consumer apps are going to have that same
    0:12:10 pressure.
    0:12:15 Maybe it’s not now, but in like 20 years, like when Sam Altman is like, you know, on
    0:12:20 his island somewhere and like there’s another CEO, that’s what they’re going to have to do.
    0:12:24 I mean, maybe we’ll just be living in a post capitalist utopia because we’ll have a GI
    0:12:24 or whatever.
    0:12:28 But like assuming that that’s not the case, that is a thing and that is a pressure.
    0:12:30 And I think that’s a really interesting thing.
    0:12:34 But also like all of these new tools open up new opportunities.
    0:12:40 So like it seems likely to me that maybe it spins up a whole UI for you for like a specific
    0:12:45 thing, but like the kind of UI that it makes and the way that it does the UI, like we’ve
    0:12:48 been talking about is a content format is a perspective.
    0:12:53 And it would be surprising to me if they didn’t allow brands or businesses or whatever to
    0:12:58 like make available those tools or those powers within their app in some way, sort of like
    0:12:59 an app store type thing.
    0:13:00 And they’ve done that like repeatedly.
    0:13:03 They’ve tried to do app store type things for a long time.
    0:13:07 It hasn’t quite worked yet, but it’ll probably work eventually.
    0:13:11 So it’s not like the death knell of like all independent tools or media or whatever.
    0:13:15 It’s just like it’s changing and there will be new opportunities in the old ones.
    0:13:17 You have to like move on to the new ones.
    0:13:17 Yeah, totally.
    0:13:22 Hey, we’ll be right back to the show.
    0:13:24 But first I want to have another podcast I know you’re going to love.
    0:13:29 It’s called Entrepreneurs on Fire and it’s hosted by John Lee Dumas available now on the HubSpot
    0:13:30 podcast network.
    0:13:35 Entrepreneurs on Fire stokes inspiration and shares strategies to fire up your entrepreneurial
    0:13:38 journey and create the life you’ve always dreamed of.
    0:13:41 The show is jam packed with unlimited energy, value and consistency.
    0:13:46 And really, you know, if you like fast paced and packed with value stories and you love
    0:13:48 entrepreneurship, this is the show for you.
    0:13:54 And recently they had a great episode about how women are taking over remote sales with
    0:13:54 Brooke Triplett.
    0:13:56 It was a fantastic episode.
    0:13:57 I learned a ton.
    0:13:59 I highly suggest you check out the show.
    0:14:02 So listen to Entrepreneurs on Fire wherever you get your podcasts.
    0:14:10 Would you be open to like sharing a little bit of like your workflow for actually building
    0:14:11 new software?
    0:14:13 Like how do you validate an idea?
    0:14:16 And then what sort of tools do you use to actually build the idea?
    0:14:19 And, you know, can you walk us through some of that?
    0:14:20 Totally.
    0:14:27 I think the really cool thing about the current AI landscape is it has like completely reset
    0:14:30 the table for like what tools are possible.
    0:14:35 And so that means that there’s a lot, a lot, a lot of low hanging fruit because we’re all
    0:14:37 just like figuring out our workflows, right?
    0:14:38 It’s all new.
    0:14:43 And that wasn’t the case two years ago because it was like the apex of the B2B sas boom and
    0:14:45 like every low hanging fruit had been picked, right?
    0:14:45 Yeah.
    0:14:47 There’s a clear way you do things.
    0:14:49 It was all like it’s all been figured out.
    0:14:49 Yeah.
    0:14:50 Now it’s totally different.
    0:14:55 So what I do and like sort of the philosophy is first I use these tools every day.
    0:15:00 And by these tools, I mean like the more general purpose tools like Claude and ChatGbt and Cursor
    0:15:02 and Windsurf and like all that kind of stuff.
    0:15:08 And what ends up happening is you end up finding workflows that start to work pretty well for you.
    0:15:12 But once you find a workflow and a ChatGbt type tool,
    0:15:15 it’s kind of a pain to just do it over and over and over again.
    0:15:16 It’s not built for that.
    0:15:18 It’s built for this like general purpose thing.
    0:15:19 Right.
    0:15:21 And that is always an opportunity to me.
    0:15:25 So like a really good example of that is we have this app called Spiral.
    0:15:28 And Spiral helps you automate a lot of repetitive creative work.
    0:15:33 You know, for me, it’s like doing my podcast descriptions or my podcast tweets or my YouTube
    0:15:36 titles or like that kind of stuff where it’s like I have a format.
    0:15:38 I know basically what it should be.
    0:15:40 And it’s like really important that I get it right.
    0:15:42 But it’s also like fairly repetitive.
    0:15:47 And what I realized at some point, I guess in like May that Claude was like really good at that.
    0:15:51 Like if I gave it a few shot prompt with some examples and I was like, okay,
    0:15:53 I want you to take a podcast transcript and turn it into a tweet.
    0:15:56 And like here’s examples of me doing that in the past.
    0:15:57 Here’s a new transcript.
    0:15:58 Can you like do a tweet?
    0:16:00 It would like get me 80% there.
    0:16:03 And then I would just have to like do some modifications.
    0:16:06 But like that prompt was like really hard to do all the time.
    0:16:08 Like I didn’t want to have to redo that all the time.
    0:16:13 And if I didn’t want to have to do it, like I just knew that my team wasn’t going to do it.
    0:16:18 And so I just spent like two days basically making an app called Spiral where you can like,
    0:16:19 you make a spiral.
    0:16:22 You say like, I want to turn podcast transcripts into tweets.
    0:16:26 You feed it some examples and then you have like a shareable widget that you can anytime
    0:16:32 thereafter like generate more and more and more tweets without having to like reprompt anything.
    0:16:33 And that’s going really well.
    0:16:36 We have a full-time GM for that on that product.
    0:16:36 His name is Danny.
    0:16:37 He’s really talented.
    0:16:41 We have almost 10,000 users for it and it’s growing really well.
    0:16:45 And that’s an example of like you find this like use case in the general purpose tool
    0:16:48 and then you can like unbundle it into like its own app.
    0:16:53 And the way that I do that is it’s sort of the way Nathan, you mentioned earlier is like
    0:16:56 they use cursor or Windsurf.
    0:16:58 I kind of go back and forth between the two.
    0:17:03 They have these like really, really good agentic building environments now where you just
    0:17:06 type in a chat and it’s like, it just goes off and does stuff.
    0:17:08 And then it’s like, it’s made like five files and it’s like,
    0:17:10 I have no idea how it works, but it’s fine.
    0:17:12 Sometimes it’s fine.
    0:17:14 Quick question.
    0:17:17 Do you actually know how to code yourself or are you letting AI write all the code?
    0:17:18 Okay. You do know how to code.
    0:17:19 Okay.
    0:17:22 I do know how to code, but I am letting AI write all the code because I’m very lazy.
    0:17:25 Yeah, but at least you’re able to do a little bit of extra troubleshooting
    0:17:28 if the AI doesn’t figure it all out for you.
    0:17:28 Yeah.
    0:17:29 Okay.
    0:17:32 And like, you know, what’s really nice is like 01 and 01 Pro.
    0:17:37 Like sometimes these tools don’t have tool use and all the agent stuff uses tools.
    0:17:41 Honestly, when 01 or maybe 03 has tool use, which I think is coming very, very soon,
    0:17:43 it’s going to completely change everything again.
    0:17:47 But I’ll often like, I’ll start in 01 or 01 Pro.
    0:17:48 I’ll say make the plan.
    0:17:49 I’d paste the plan into cursor.
    0:17:51 Cursor will do stuff.
    0:17:52 Then it will get stuck and have a bug.
    0:17:55 And then I’ll just paste it back into 01 and then paste it back.
    0:17:58 And then every once in a while, I’m like, I’m bored of copy pasting.
    0:18:00 Like a lot of when I figure out how this actually works.
    0:18:06 But also for me, like to be really clear, like my job right now is I’m mostly
    0:18:08 prototyping new things and writing about them.
    0:18:10 And then sometimes they become products, but once they’re actual products,
    0:18:14 we have actual engineers who like actually are like in the code a lot.
    0:18:17 And so I could be, you know, reading the code or whatever,
    0:18:19 but I’m not doing that right now because it’s not really my job.
    0:18:21 Yeah.
    0:18:25 So I want to ask this to you because there’s been a lot of discussion around
    0:18:27 like SaaS companies having no moat, right?
    0:18:30 I’m sure you’ve heard that a million times over the last couple years.
    0:18:32 What are your thoughts on that?
    0:18:36 So let me preface this with saying like, I personally, like if I have a problem
    0:18:39 I need to solve, it’s like a fairly simple problem.
    0:18:43 I’ve stopped going out and looking to find if there’s a software product for it.
    0:18:47 I will literally open up cursor and tell cursor what I want it to make.
    0:18:49 And it’ll write a little Python script for me.
    0:18:51 And it solves that recurring problem for me.
    0:18:55 I know right now I’m probably like abnormal and doing that.
    0:18:58 I don’t think most people sort of have that kind of workflow.
    0:19:02 But I’m wondering if it’s going to get to a point where when somebody has a problem
    0:19:05 they need to solve instead of turning to SaaS products and paying,
    0:19:07 you know, some sort of recurring subscription,
    0:19:11 they’re just going to go to one of the AI chat bots that are out there.
    0:19:14 Maybe it’s ChatGPT, your Claude, or one of these tools and just say,
    0:19:16 this is a problem I’m trying to solve.
    0:19:19 And it says, okay, here’s a little script I wrote that solves the problem.
    0:19:22 So I wonder what the future of SaaS looks like.
    0:19:27 I mean, I think you’re definitely right in the sense that there are certain little things.
    0:19:29 There are maybe like a thousand little things that come up every day
    0:19:33 that like maybe I would have like looked for a tool for before.
    0:19:38 Probably I would have just like suffered through like doing them manually or whatever
    0:19:41 that now ChatGPT just does it for me and it’s amazing.
    0:19:46 But I think there’s also a big difference between that and like a fully featured like SaaS app,
    0:19:51 like for a power workflow, which I just think that things that are more specialized
    0:19:55 for those kinds of users are going to win even when you have AI,
    0:19:59 even if you’ve theoretically have AGI and that there are room for those kinds of tools,
    0:20:01 whether they look like traditional B2B SaaS,
    0:20:04 or whether they look a little bit more like an AI native version of that,
    0:20:07 whatever that looks like, I think it’s going to be more like that.
    0:20:10 And I think there’s also a really interesting question about like
    0:20:13 how much SaaS is actually the right business model.
    0:20:17 What I’ve been seeing a lot more of and I think is probably a little bit more of a
    0:20:23 AI native business model is paying per completed task or paying for the value that you get,
    0:20:29 basically, because when you have AI, you can pay for the job to be done versus just like the time
    0:20:30 or by the month.
    0:20:36 And I think that’s really interesting because SaaS is such a much better business model
    0:20:41 that if you are currently a SaaS company, you’d probably rather stay a SaaS company.
    0:20:44 And that means that there’s a lot of opportunities for startups
    0:20:47 that don’t necessarily need to keep that model.
    0:20:49 Yeah, performance-based SaaS would be incredible.
    0:20:50 Like, yeah, that’s an amazing idea.
    0:20:51 Yeah, I think that’s going to happen.
    0:20:53 I mean, you see it already in certain places,
    0:20:56 like the LLM APIs or all paper use or whatever,
    0:20:58 but I think that’ll filter up through the stack.
    0:21:01 I feel like I’m your worst guest ever.
    0:21:05 I’m like SEO is dying and SaaS is dead.
    0:21:10 I’m so sorry.
    0:21:14 I mean, we’ve got other tools that I think are cool.
    0:21:17 We have a call called Sparkle, which is a file organizer.
    0:21:21 So basically, it just sits on your desktop and just automatically organizes your desktop,
    0:21:24 your downloads and your documents with AI, which is like really cool and very useful.
    0:21:27 Everything is f***ing messed for me.
    0:21:30 So like, but now it’s not and I didn’t have to do anything.
    0:21:30 That’s amazing.
    0:21:32 I’m always like trying to think about like how to organize my files.
    0:21:34 I’m like, oh, one day I’m going to organize them all.
    0:21:36 They’re going to be, and I was like, wait a minute,
    0:21:37 AI is going to eventually do that for me.
    0:21:39 So I was like, it’s awesome that you’re doing something like that.
    0:21:40 Yeah, it’s just a matter of time.
    0:21:43 Another thing that we’ve been working on,
    0:21:44 it’s actually one of our writers, Michael Taylor,
    0:21:47 has been working on and we’re kind of doing it with him,
    0:21:52 is AI is quite good at acting like a particular kind of person,
    0:21:54 like simulating a particular kind of person.
    0:21:59 And Michael built this tool called Rally that we’ve been using internally.
    0:22:02 I think we’re going to make like every column out of it
    0:22:06 and then maybe launch it as a product that’s still TBD.
    0:22:08 And it really depends on if Michael wants to do it with us.
    0:22:09 But I think it’ll be really cool.
    0:22:12 And basically what it does is like, for example,
    0:22:15 if we’re coming up with headlines for a podcast,
    0:22:19 what we can do is we spin up an audience of like a thousand every people
    0:22:21 or a thousand hacker news people.
    0:22:24 And then we can be like, okay, here’s my three titles.
    0:22:25 Which ones do you like?
    0:22:27 And it will just like survey the crowd
    0:22:30 and then tell us what the crowd of LLMs thinks.
    0:22:32 And it’s really cool. It works.
    0:22:34 I thought you were going the opposite side.
    0:22:36 So I actually thought it was going to be like the person
    0:22:38 who’s helping you create the taglines or whatever.
    0:22:42 You fed it a bunch of books from some famous person who’s an expert at that.
    0:22:44 But you’re doing the opposite, like the audience side.
    0:22:45 Yeah, that’s incredible.
    0:22:45 Exactly.
    0:22:47 So it’s sort of like wisdom of the crowds,
    0:22:50 like ask the audience type stuff, which like now you just have to try stuff
    0:22:52 where you have to do A/B testing and it takes a long time.
    0:22:54 You need a big audience or whatever.
    0:22:56 And if you use this, you don’t have to.
    0:22:58 And it’s still very early and it remains to be seen.
    0:23:03 But like I feel quite confident that the results will align to some degree
    0:23:05 with like what you would actually find out in the world.
    0:23:06 And that’s really cool.
    0:23:09 Wow.
    0:23:11 So I actually do have one other sort of little rabbit hole.
    0:23:12 I do want to go down with you.
    0:23:16 So I believe you do some investing outside of every as well, right?
    0:23:17 So you have some investments.
    0:23:19 I believe you work with Sequoia Capital as well.
    0:23:20 Is that correct?
    0:23:22 What are you looking for in investments?
    0:23:24 I’m assuming you’re probably focused more on AI.
    0:23:25 I could be totally wrong.
    0:23:26 I’m just sort of guessing.
    0:23:28 But what are you looking for in investments?
    0:23:30 Like how do you find something that stands out to you?
    0:23:37 The investments that I’ve done usually are places where I’ve had a particular thesis
    0:23:38 I’ve been writing about and thinking about.
    0:23:42 Or I just know the founder and I know they’re really awesome.
    0:23:45 And I just like want to get into whatever they do.
    0:23:47 Yeah, I do invest in AI.
    0:23:52 I have found honestly that a lot of AI companies are just raising at such gigantic
    0:23:57 valuations that I haven’t actually pulled the trigger on that many deals,
    0:24:00 even if I really like the person or really like the thing they’re building.
    0:24:04 Because it’s just like a $50 million seed round is just like crazy to me.
    0:24:09 But like generally, it’s like, is it a thesis in a space that I care about
    0:24:10 or like an idea that I care about?
    0:24:13 And I have a bunch of articles that sort of lay out all that stuff.
    0:24:17 And I can give you a little preview about the things I care about right now.
    0:24:19 Like for example, one is the future of science,
    0:24:23 like how AI might change how we do science is a big one for me.
    0:24:25 Another one is yeah, just really like the founder.
    0:24:27 And then the last one is just do I use the product?
    0:24:28 Do I like it?
    0:24:30 I think the proof’s in the pudding a lot in those situations.
    0:24:33 And you can tell immediately when someone’s built something awesome
    0:24:35 because you just feel it.
    0:24:35 Yeah, yeah.
    0:24:37 So when you say like you have a thesis on it,
    0:24:40 you’re kind of saying like this is something I wish was in the world,
    0:24:42 but then you actually find something that does that.
    0:24:44 That’s something that you’d want to invest in.
    0:24:45 Yeah, exactly.
    0:24:45 Very cool.
    0:24:46 All right.
    0:24:47 I assume some way the world’s changing and then health,
    0:24:50 a product would fit into that changing world.
    0:24:50 Yeah.
    0:24:51 I love that stuff.
    0:24:54 I especially get really excited about a lot of the science stuff.
    0:24:56 Like I was really nerding out about like that matter gen
    0:24:58 that just came out from Microsoft the other day.
    0:25:01 I think that’s something that’s really cool and really exciting.
    0:25:03 And I think there’s a lot more stuff that’s going to pop up
    0:25:06 sort of in that world, you know, this year.
    0:25:08 I think we’re going to see a lot of sort of materials,
    0:25:11 science sort of things come out of AI this year.
    0:25:13 And so that to me is super, super exciting.
    0:25:17 But on that note, I really, really appreciate you hanging out with us today.
    0:25:20 Where should people go after listening to this episode?
    0:25:23 Is there a specific social media account you want to shout out?
    0:25:25 Your website, let us know where to check you out.
    0:25:26 Thanks so much for having me.
    0:25:32 You can find me on X at Dan Shipper, S-H-I-P-P-E-R or at every also on X.
    0:25:34 And then the website is every.to.
    0:25:35 You can find our newsletter.
    0:25:38 You can find all of our products, the courses we offer,
    0:25:39 all that kind of stuff.
    0:25:40 And it’s been awesome chatting with you guys.
    0:25:41 Yeah, it’s been great.
    0:25:42 Awesome.
    0:25:43 Yeah, thanks so much, Dan.
    0:25:44 Really, really appreciate you.
    0:25:46 And thanks everybody to listening to this episode.
    0:25:50 If you haven’t already, make sure you subscribe wherever you listen to podcasts.
    0:25:51 We’re also available on YouTube.
    0:25:53 If you want to see our faces, actually talk to each other.
    0:25:55 And thanks so much for tuning in.
    0:25:56 We’ll see you in the next one.
    0:25:59 [MUSIC PLAYING]
    0:26:02 [MUSIC PLAYING]
    0:26:06 [MUSIC PLAYING]
    0:26:08 (upbeat music)

    Episode 46: Can AI transform the landscape of SaaS, or is it poised to become a competitor? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) welcome Dan Shipper (https://x.com/danshipper), the innovative CEO of Every and an expert in the intersection of software and AI.

    In this episode, Dan shares his insights on how AI is reshaping software development and media. He discusses his company’s ventures into building AI-driven software tools, his approach to product incubation, and his perspective on the evolving dynamics in the AI landscape. We delve into the potential future of SaaS, discussing whether AI could replace traditional software models and the importance of having a unique voice in a rapidly evolving market.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) AI Email Management Launch
    • (05:03) Accessible Experimenting Through Software
    • (06:57) Human vs. AI-Generated Content
    • (11:23) AI’s Evolving Influence and Pressures
    • (14:31) Creative Automation with Spiral App
    • (18:34) AI’s Role in SaaS Evolution
    • (22:43) Personal Investment Strategy Insights
    • (24:02) Excited About AI in Science

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • Gemini 2.0: 20x Cheaper Than GPT-4?! (DEEP DIVE) | Logan Kilpatrick

    AI transcript
    0:00:02 (upbeat music)
    0:00:06 – Hey, welcome back to the Next Wave Podcast.
    0:00:07 I’m Matt Wolf.
    0:00:08 I’m here with Nathan Lanz.
    0:00:11 Today we’re joined by Logan Kilpatrick,
    0:00:15 who is the Senior Project Manager over at Google DeepMind.
    0:00:17 And the day we’re recording this episode
    0:00:19 is the same day that Google just released
    0:00:22 a whole bunch of new AI tools.
    0:00:26 Gemini Flash 2.0, Gemini Flash 2.0 Lite,
    0:00:30 Gemini Flash 2.0 Pro, all sorts of really, really cool stuff
    0:00:32 coming out of Google right now
    0:00:34 and Logan’s gonna break it all down.
    0:00:36 And you’re gonna get a pretty grand overview
    0:00:39 of where the AI world is headed according to Google.
    0:00:41 So let’s just go ahead and dive right in
    0:00:43 with Logan Kilpatrick.
    0:00:44 – Thank you so much for joining us.
    0:00:46 It’s probably a really busy day.
    0:00:47 So I really appreciate you taking the time
    0:00:48 to join us today.
    0:00:50 – Yeah, I’m excited to catch up with you both
    0:00:51 and talk about all things Gemini
    0:00:53 and what’s happening in the AI world.
    0:00:55 – Well, so this is actually your second time on the show.
    0:00:58 So we’ve already kind of dove into some of the backstory
    0:01:00 and introduced people to you in a past episode.
    0:01:02 So let’s just jump straight into it.
    0:01:06 Can you break down what is 2.0 Flash, 2.0 Flash Lite,
    0:01:08 2.0 Pro, like what are the differences?
    0:01:10 What’s better about these models
    0:01:12 than what was out prior to them?
    0:01:14 – I think this is an exciting moment for us
    0:01:16 just because of like the amount of effort and work
    0:01:18 that’s gone into bringing Gemini 2.0
    0:01:19 actually into the world.
    0:01:21 And Matt, you were there, Nathan,
    0:01:23 I don’t remember if you were there at I/O last year,
    0:01:27 but we announced 1.5 Flash and the long context
    0:01:29 and a bunch of this other stuff last May.
    0:01:30 So like literally less than a year ago.
    0:01:33 And for the last year-ish, Flash has been this
    0:01:36 like wild success story for us
    0:01:38 of building a model that developers really love.
    0:01:41 And a lot of that is rooted in like the right trade-offs
    0:01:44 of like cost, intelligence, performance, capabilities.
    0:01:46 And if you look at the 1.5 Flash model
    0:01:48 and you think about like, how do we do this better?
    0:01:50 It’s like, you have to make it more powerful.
    0:01:52 You have to give it more capabilities.
    0:01:55 You have to do that all while not making it cost
    0:01:57 a lot more money for developers.
    0:02:00 And it feels like we pulled a rabbit out of the hat
    0:02:01 to a certain extent with 2.0 Flash
    0:02:04 because like the actual cost for developers,
    0:02:06 it was historically like seven and a half cents
    0:02:08 per million tokens.
    0:02:09 Now it’s 10 cents.
    0:02:11 So the blended cost is actually less for this model.
    0:02:13 And we did all that while like,
    0:02:15 this model is actually better than Pro.
    0:02:16 It has all these capabilities.
    0:02:18 It’s natively agentic.
    0:02:21 It has search built into it as code execution built into it.
    0:02:23 – Yeah, it’s just exciting for me as a developer.
    0:02:27 Like I think ultimately you remove the cost barriers
    0:02:29 and all these other things for people to build
    0:02:30 really cool stuff.
    0:02:31 And like that’s what enables the world
    0:02:33 to make really cool products.
    0:02:34 So I’m super excited.
    0:02:36 So that’s the sort of headline for Flash
    0:02:37 is better, faster, cheaper,
    0:02:40 which continues to be sort of the tagline.
    0:02:41 I need to get my Gemini t-shirts
    0:02:43 that say better, faster, cheaper on them.
    0:02:46 – I read that it’s better than GPT-40,
    0:02:47 but like 20 times cheaper.
    0:02:49 Is that like roughly correct?
    0:02:51 – Yeah, which is just crazy to me.
    0:02:53 And I think like we’ve got a lot of work to do.
    0:02:56 I think one of the dimensions of the Gemini story
    0:02:58 is like we continue to put out really great models.
    0:02:59 I think we need to do a great job as well
    0:03:01 of like going and telling the world
    0:03:03 about this technology that we’re building.
    0:03:05 ‘Cause I don’t think people really understand
    0:03:07 and actually for a lot of developers,
    0:03:08 the cost is the reason in many cases
    0:03:10 they don’t build the stuff that they want to build.
    0:03:13 It’s like, I can’t afford to put this thing into production.
    0:03:14 It’s too expensive.
    0:03:17 So I think Flash is really important in that story.
    0:03:18 But we also landed 2.0 Pro.
    0:03:22 We also landed an even cheaper version of Flash Light,
    0:03:23 which sort of has the capabilities
    0:03:24 pared down a little bit,
    0:03:26 but makes it so that we can keep delivering
    0:03:30 on that like frontier cost performance sort of trade often.
    0:03:31 And that model is in preview
    0:03:33 and it’ll go GA in the next few weeks
    0:03:35 as we iron out the last few bugs.
    0:03:38 And I think Pro gives me a lot of excitement
    0:03:43 about those whole narrative of pretraining, being dead.
    0:03:45 An interesting sort of realization
    0:03:47 I had after our conversation with Jack Ray,
    0:03:49 who’s one of the co-leads for the reasoning models
    0:03:53 in DeepMind is there’s this non-linear amount
    0:03:57 of extra effort it takes to make the models
    0:03:59 continue to get better.
    0:04:00 Like you look at like, okay,
    0:04:03 what does 3% mean on some benchmark?
    0:04:06 Like you think 3% and we think of like the normal world
    0:04:10 where 3% is like actually 3% and like in the model world,
    0:04:14 3% is actually like a 25% increase
    0:04:16 in like the amount of efforts that went into this.
    0:04:19 But also that 3% is like the difference
    0:04:22 between unlocking a bunch of capabilities
    0:04:25 and a bunch of use cases that like just didn’t work before
    0:04:28 because like the thing failing 3% to 4% of the time
    0:04:31 versus not is the difference between you putting AI
    0:04:33 into production at your company and not.
    0:04:35 So it actually matters a lot.
    0:04:36 And that’s why I think we continue to push
    0:04:37 on that frontier.
    0:04:41 – So what’s the big difference between like these new models
    0:04:44 and the last ones as far as like how they were created,
    0:04:46 is it like more parameters that are being trained on?
    0:04:49 Obviously, the big narrative like you just mentioned
    0:04:52 with things like deep seek and O3 and things like that
    0:04:55 from open AI are the sort of what happens at inference, right?
    0:04:58 When somebody enters a prompt, it does all of this thinking
    0:05:00 and that’s really what they’re sort of like pushing on
    0:05:02 is like the next sort of breakthrough.
    0:05:03 Like what sort of breakthroughs,
    0:05:06 what changed between the last models and this one
    0:05:07 to make this one so much better?
    0:05:09 – Yeah, there’s two dimensions of this.
    0:05:12 One, it’s a story of the really difficult work
    0:05:14 of doing algorithmic improvements and breakthroughs.
    0:05:16 And I think like the team at DeepMind
    0:05:17 stuff way beyond my understanding
    0:05:20 as far as how they’re able to make this continue to work.
    0:05:23 So I think there’s like core fundamental research
    0:05:24 advancements that are happening.
    0:05:26 And there’s a lot of like data efficiency wins as well,
    0:05:27 which is also exciting.
    0:05:30 But as far as like new capabilities of these models,
    0:05:34 I think the two big ones is when Gemini was first announced,
    0:05:36 it was announced as this model that’s natively multimodal.
    0:05:39 And it was natively multimodal in the sort of input sense
    0:05:42 that it could really understand the videos,
    0:05:45 audios, images that it was being given.
    0:05:47 And that was one of the main differentiators.
    0:05:50 Today, the model’s actually capable of doing that
    0:05:51 except on the output sense,
    0:05:53 which I think was a huge jump for us.
    0:05:54 And it actually requires again,
    0:05:58 a bunch of like non-trivial amount of engineering work
    0:06:00 in order to make the models capable of doing that.
    0:06:02 I had an interesting conversation a few weeks ago
    0:06:04 with someone on our research team,
    0:06:05 which reminded me of this.
    0:06:07 Someone asked a question of like,
    0:06:09 why does it matter if the models are capable
    0:06:12 of natively outputting these multimodal capabilities?
    0:06:14 Like we have really great text-to-speech models.
    0:06:16 We have great speech-to-text models.
    0:06:18 We have great image generation models.
    0:06:20 Like why is it cool that the model can do this natively?
    0:06:23 And there’s all of these really great examples,
    0:06:26 like a calculator versus like, I don’t know,
    0:06:29 an AI model that has access to code execution.
    0:06:31 Like the code execution version can really like solve
    0:06:33 these problems in a really complicated way
    0:06:36 that you wouldn’t otherwise be able to,
    0:06:38 or at least that the effort is required
    0:06:40 as you on the user of the models.
    0:06:43 And I think that’s the world of these custom domain-specific
    0:06:45 models, like image generation and audio generation,
    0:06:48 versus the native capability really feels like
    0:06:50 the model can just do the heavy lifting for you,
    0:06:52 which is really interesting.
    0:06:55 – So right now, can Gemini actually output like an image
    0:06:57 if I give it a prompt to generate an image?
    0:06:59 Does it generate an image right now?
    0:07:00 – Not accessible to everyone yet.
    0:07:01 And I think this is the gap.
    0:07:03 So we have it internally and folks are using it
    0:07:05 in our early access program
    0:07:07 and we should get you both early access
    0:07:08 to play around with it and test it out.
    0:07:11 And we’ll roll it out more broadly soon,
    0:07:12 which I’m excited about.
    0:07:15 But that sort of same line of thinking
    0:07:17 is what takes us to like native tool use as well.
    0:07:19 And like native tool use is available to everyone.
    0:07:22 And it’s like the model was trained,
    0:07:25 knowing how to differentiate questions
    0:07:27 that it should go and search the internet for
    0:07:28 or questions that it needs to use a tool
    0:07:30 like code execution for.
    0:07:33 So you get like all of those like silly examples
    0:07:34 where the model would be like,
    0:07:36 let me try to solve this math problem,
    0:07:38 which I know I’m not going to be able to solve
    0:07:40 just because you asked me to with code execution,
    0:07:42 like it knows it needs to use that tool.
    0:07:44 And there’s a whole bunch of verticals
    0:07:45 where like the performance goes up
    0:07:46 significantly because of that.
    0:07:49 – So Gemini is actually generating the image.
    0:07:51 It’s not going and calling upon like,
    0:07:53 imagine three to generate the image.
    0:07:55 It’s actually Gemini who’s creating that image
    0:07:57 when it does generate an image.
    0:07:57 – Exactly.
    0:07:59 And I’ll push on getting you both access
    0:08:00 after this conversation
    0:08:02 because I think the world knowledge piece
    0:08:04 really highlights like why this matters.
    0:08:06 And there’s like a bunch of examples
    0:08:09 that I played around with of like pictures of a room
    0:08:11 and like having the image change
    0:08:15 based on these like really complex nuanced prompts
    0:08:16 around moving objects in certain ways.
    0:08:19 Like things that if it doesn’t have world knowledge
    0:08:21 and understand like it understands physics
    0:08:22 and understands all these things
    0:08:25 that again require the world knowledge piece.
    0:08:28 And I think it’s actually there’s some interesting trends
    0:08:30 of what is the outcome of like being able
    0:08:32 to take other domain specific models
    0:08:35 and bring them into these LLMs that have world knowledge.
    0:08:38 I think there’ll be some really cool capabilities
    0:08:39 that like we’re not thinking of today
    0:08:41 that this is going to enable,
    0:08:43 which yeah, it gets me excited for people.
    0:08:45 – Yeah, it seems like that’s kind of required
    0:08:46 for this also to work in like everything
    0:08:48 from like gaming to robotics
    0:08:50 to it actually have an understanding of the world.
    0:08:52 – Oh yeah, 100%.
    0:08:54 – Yeah, yeah, I mean, I actually had the opportunity
    0:08:57 to go out to London, go visit the DeepMind offices
    0:08:59 and got to play with Astra on the phone.
    0:09:02 And I mean, that was like the first taste I got
    0:09:07 of like an actual useful AI assistant out in the real world.
    0:09:10 And I don’t know if that was using Gemini 2.0 Flash
    0:09:13 or if that was actually using the full Gemini 2.0 yet
    0:09:17 at the time, but it was definitely a very, very impressive model
    0:09:20 to actually see the tool use in real life
    0:09:21 and just be able to walk around
    0:09:24 and it understand images and understand video
    0:09:26 and understand audio and understand text.
    0:09:29 And it was all built from the ground up
    0:09:33 to understand that stuff as opposed to like look at an image,
    0:09:36 use OCR to figure out the text on the image
    0:09:38 and then pull in the text or listen to the audio,
    0:09:41 transcribe the audio to text and then use the text.
    0:09:43 It’s actually understanding what it’s seeing,
    0:09:46 what it’s hearing, which I think is like one
    0:09:50 of the major differentiators about like what Gemini is doing
    0:09:52 that you don’t see the other models doing yet.
    0:09:54 So it’s super, super impressive.
    0:09:56 – Yeah, I think the other piece of this is
    0:09:58 other than just the raw capabilities
    0:09:59 from a complexity standpoint,
    0:10:01 it means that like developers building stuff,
    0:10:04 you don’t have to go and like do a ton of scaffolding work
    0:10:05 in order to like make this happen.
    0:10:07 It’s like the overall complexity of your application
    0:10:10 when the model is just able to sort of take in a bunch
    0:10:12 of things and put out a bunch of things
    0:10:13 makes life incredibly easy
    0:10:16 and you don’t have to deal with like frameworks
    0:10:17 on frameworks on frameworks.
    0:10:19 So much of the agent world is like,
    0:10:21 hey, the models actually aren’t that good
    0:10:22 at doing some of these things.
    0:10:24 And like the way that we supplant that
    0:10:27 is by like building a bunch of scaffolding and frameworks,
    0:10:29 that’s where a lot of developers are focused today.
    0:10:30 And actually there’s going to be a moment
    0:10:31 where like all of a sudden the model capabilities
    0:10:33 are just like good enough that it kind of works.
    0:10:35 And then people are going to be like,
    0:10:36 well, why do I have all this scaffolding
    0:10:38 that’s doing these things for me
    0:10:40 that the models can just do out of the box now?
    0:10:42 So it’ll be interesting to see how that plays out.
    0:10:45 – Yeah, I know like 2025 is going to be sort of
    0:10:46 the year of the agent, right?
    0:10:49 That term has already been thrown around quite a bit.
    0:10:50 But I also feel like everybody kind of has
    0:10:52 a different definition of an agent.
    0:10:54 You know, some people will look at something
    0:10:56 that like you can build over on make.com or Zapier
    0:10:58 where it’s tying different tools together
    0:11:00 using APIs as an agent.
    0:11:02 But I’m curious, does Google and DeepMind,
    0:11:04 do they actually have like an internal definition
    0:11:06 of an agent that they’re shooting for?
    0:11:08 Do you think we actually have agents now
    0:11:09 based on their definition?
    0:11:11 Where does Google stand on agents?
    0:11:14 – I actually don’t know what our like formal definition
    0:11:15 of agents are.
    0:11:18 I have tried a bunch of the agent products
    0:11:20 and historically haven’t been super impressed
    0:11:21 at what they’re capable of.
    0:11:22 I think we’re just not there yet.
    0:11:25 The thing that I want and I think a lot of users
    0:11:28 want this as well is just the models to be proactive.
    0:11:32 And like all of the products of today that build on AI
    0:11:35 require me to basically change my workflows
    0:11:38 or put in extra work or put in extra effort
    0:11:41 like through this sort of guys of like,
    0:11:42 oh, this is actually going to save you time
    0:11:43 if you do this thing.
    0:11:45 And like really what I want is like the models
    0:11:47 to just be like looking at the stuff
    0:11:48 that I give them access to
    0:11:50 and like coming up with ways to be useful
    0:11:51 and like save me time.
    0:11:54 And like, yeah, it’s going to get some things wrong
    0:11:56 but like I don’t want to have to be the one
    0:11:57 in the driver’s seat all the time.
    0:11:59 And it feels like today’s AI agents,
    0:12:01 again, because the models aren’t good enough
    0:12:03 like have to be proactive.
    0:12:05 Like it requires the proactiveness of the human.
    0:12:08 And I think once that role reversal switches,
    0:12:11 I think that’s where we see like billion agents scale
    0:12:14 deployments like all of a sudden just like happening
    0:12:14 and working.
    0:12:17 And this is also where like things get crazy
    0:12:17 again with compute.
    0:12:19 Cause again, like actually if you look at
    0:12:21 how compute is being used today,
    0:12:23 it’s in a lot of cases like this one to one correlation
    0:12:27 between like a human input and the token output.
    0:12:30 And I think the future is thousands and thousands
    0:12:32 of X more usage of AI happening
    0:12:34 by the agents themselves than by humans,
    0:12:36 which will be fascinating to see play out.
    0:12:40 – Hey, we’ll be right back to the show.
    0:12:42 But first I want to talk about another podcast
    0:12:43 I know you’re going to love.
    0:12:45 It’s called entrepreneurs on fire.
    0:12:46 And it’s hosted by John Lee Dumas
    0:12:49 available now on the HubSpot podcast network.
    0:12:51 Entrepreneurs on fire stokes inspiration
    0:12:54 and share strategies to fire up your entrepreneurial journey
    0:12:56 and create the life you’ve always dreamed of.
    0:12:59 The show is jam packed with unlimited energy,
    0:13:00 value and consistency.
    0:13:02 And really, you know, if you like fast-paced
    0:13:04 and packed with value stories
    0:13:07 and you love entrepreneurship, this is the show for you.
    0:13:09 And recently they had a great episode
    0:13:12 about how women are taking over remote sales
    0:13:13 with Brooke Triplett.
    0:13:15 It was a fantastic episode.
    0:13:16 I learned a ton.
    0:13:17 I highly suggest you check out the show.
    0:13:19 So listen to entrepreneurs on fire
    0:13:21 wherever you get your podcasts.
    0:13:26 – Yeah, no, I think what you described
    0:13:28 is sort of what I envision of an agent
    0:13:29 is almost like predictive.
    0:13:32 Like it sort of figures out what you need before you need it
    0:13:34 and then make suggestions based on that.
    0:13:36 So I think that’s a world that I’m really excited
    0:13:37 to get into.
    0:13:39 But I do want to touch on the word compute
    0:13:41 that you just mentioned for a second
    0:13:43 ’cause obviously there was a bit of like a, you know
    0:13:45 a freak out, so to speak in the US
    0:13:47 when that deep seek model came out
    0:13:49 and everybody thought that, well,
    0:13:52 this deep seek model uses a lot less compute
    0:13:55 than these other models that have been trained.
    0:13:57 So therefore, you know,
    0:13:59 NVIDIA GPUs are no longer necessary.
    0:14:02 And then we saw NVIDIA sort of lose some market share
    0:14:03 as a result of it.
    0:14:05 But I’m just curious, like what are your thoughts
    0:14:06 on the compute?
    0:14:07 Because I saw all of that happening
    0:14:09 and I thought it was so bizarre.
    0:14:11 I was like, this seems like a bullish sign
    0:14:13 for NVIDIA to me, not a bearish sign.
    0:14:15 Like what’s going on here?
    0:14:16 But I’m curious, like what’s your take on
    0:14:17 what happened there?
    0:14:20 – Yeah, there’s a lot of complexity to that story
    0:14:22 and parts of the story that I don’t want to touch on
    0:14:25 but the thing that I do want to touch on is like
    0:14:26 if you look around the world,
    0:14:28 I think this example that I just gave of like
    0:14:31 who is in the driver’s seat of using AI today?
    0:14:34 And like today it’s humans that are in the driver’s seat
    0:14:36 and like we’re just inherently bounded
    0:14:38 by the amount of humans who are using AI
    0:14:39 because, you know, it just takes a while
    0:14:42 for technology to assimilate into culture
    0:14:44 and real use cases and all that stuff.
    0:14:46 You know, we’re on this exponential right now.
    0:14:48 I think as soon as agents start to take off
    0:14:51 that exponential becomes a straight line up into the right.
    0:14:54 Like I think it’s gonna be pretty profound
    0:14:56 because like again, the challenge is
    0:14:59 the human process just doesn’t scale
    0:15:02 and the agent process is going to scale
    0:15:03 which is going to be really interesting.
    0:15:05 Like I have 10,000 emails I haven’t read
    0:15:06 in the last three months.
    0:15:07 – Right, I was thinking emails.
    0:15:08 First thing I was thinking of,
    0:15:10 I wanted to handle all of that for me.
    0:15:11 I don’t want to think about any of it.
    0:15:12 – Yeah, it’s gonna be wonderful.
    0:15:14 And really like go in and find the things
    0:15:16 like I know there’s things that I should be doing
    0:15:18 that would create value.
    0:15:19 – Missed opportunities.
    0:15:20 – 100%.
    0:15:22 And there’s like so many of those things
    0:15:24 where if you also like think about like
    0:15:26 what’s the economic value of all that work?
    0:15:29 Like one of the crazy frames of mind
    0:15:31 that I look at the world through is like,
    0:15:33 you look at the world and like the world is just filled
    0:15:34 with all this inefficiency.
    0:15:36 And like it’s beautiful in many ways,
    0:15:39 but like it’s also this really cool opportunity
    0:15:41 if you can be the one to create something
    0:15:43 that sort of makes people more productive
    0:15:46 and explicitly makes them more productive
    0:15:47 at maybe the things they don’t want to do.
    0:15:49 And I feel like that’s the other part
    0:15:51 of this agent story and this compute story,
    0:15:53 which is a lot of the products
    0:15:55 that I see people building are actually going
    0:15:59 after the things that people really like doing.
    0:16:02 And like maybe like shopping is this sort of
    0:16:03 tongue-in-cheek example of this.
    0:16:05 ‘Cause some people really like shopping
    0:16:05 and some people don’t.
    0:16:07 But like if I told my girlfriend
    0:16:09 that we could never go shopping again together
    0:16:13 and we could never go try out different experiences
    0:16:15 and go check out the vibe of different stores,
    0:16:17 like there’s so much of this is like
    0:16:19 such a fundamental part of the human condition actually
    0:16:22 of like going and seeing these different places
    0:16:24 and like it’s really baked into who we are.
    0:16:27 And that’s such like a traditional example
    0:16:29 of like what agents are going to do for people.
    0:16:30 And it’s odd to me.
    0:16:33 I feel like people have like kind of misconstrued
    0:16:35 what the value creation is going to be
    0:16:36 in some of these examples.
    0:16:37 – I mean, I think that’s like the bias
    0:16:39 from like a, you know, Silicon Valley nerds
    0:16:40 building this stuff, right?
    0:16:41 – Yeah, yeah.
    0:16:42 – I don’t want to shop.
    0:16:43 I just want to be automated for me.
    0:16:46 And it’s like, that’s just like a small subject
    0:16:46 of the world.
    0:16:47 Like, you know, when I shop,
    0:16:49 I just like go in and get what I want.
    0:16:50 There’s like two options.
    0:16:51 Okay, I like this one better.
    0:16:51 I get it.
    0:16:54 But my wife, she just loves checking out stores
    0:16:55 and different shops.
    0:16:56 And like she would hate the idea of like skipping
    0:16:58 all of that would just make no sense to her at all.
    0:17:01 – Yeah, I think there’s an underlying story here,
    0:17:02 which is like, first of all,
    0:17:03 there’s a lot of variance in human preference,
    0:17:07 but also there’s like ways about going about a certain task
    0:17:09 that like make it interesting or not.
    0:17:10 And like, I like shopping too,
    0:17:13 if it’s like me getting to do it on the terms that I want.
    0:17:15 And like, it will be interesting to see like,
    0:17:17 how can agents and some of these products
    0:17:19 like actually help create that experience?
    0:17:21 And this goes back to this deep seek narrative
    0:17:23 around like the value creation
    0:17:24 happening at the application layer.
    0:17:26 And it really does feel like this is true.
    0:17:28 Like if you look back two years ago,
    0:17:29 the narrative was, you know,
    0:17:32 all these companies are just rappers on top of AI.
    0:17:33 There’s no value creation.
    0:17:35 All the creation of the value creations in the tokens,
    0:17:38 like don’t spend your time thinking about these companies.
    0:17:41 And it’s so funny how quickly this like flip-flops
    0:17:43 back and forth between like,
    0:17:45 now all the value creation is at the application layer
    0:17:48 and like LLMs are this commodity thing
    0:17:49 that no one should think about.
    0:17:51 I enjoy watching it all play out.
    0:17:51 It’s fun.
    0:17:52 – So how do you think it’s gonna play out?
    0:17:53 Because like right now,
    0:17:55 Google’s kind of focused on developers, right?
    0:17:57 More than consumers, correct?
    0:17:58 – So on one hand,
    0:18:01 like I spend all my time thinking about developer stuff.
    0:18:02 Google’s got a ton of other people
    0:18:04 who are doing consumer stuff.
    0:18:05 I think a good example of this is like,
    0:18:10 Gemini is both in search and through the Gemini app,
    0:18:13 like, you know, deployed across like billion user scale.
    0:18:15 Like there are literally billions of people
    0:18:18 who are interacting with outputs and the models themselves,
    0:18:20 which is crazy to think about.
    0:18:22 And like it’s a very consumer forward use case
    0:18:23 for those folks.
    0:18:25 And I think it’s also like still incredibly early
    0:18:27 for Gemini in search.
    0:18:29 And there’s some interesting stories around that stuff.
    0:18:32 But I mean, I personally am incredibly bullish
    0:18:35 on the infrastructure layer and like the infra tooling.
    0:18:38 And I think like actually a good example of this.
    0:18:41 And you see this in some of Sam’s recent tweets
    0:18:44 about sort of open AI and the pro subscription
    0:18:46 are sort of good examples of this,
    0:18:48 which is builders at the application layer
    0:18:52 have a lot of tension with building more AI.
    0:18:53 And like actually back to the thread
    0:18:56 of having AI be more proactive.
    0:18:58 This is why I believe in flash so much.
    0:19:00 And I believe in the direction that we’re going
    0:19:02 as far as like reducing the cost for developers
    0:19:04 while continuing to push the performance frontier
    0:19:07 is because the story of AI to me is a story
    0:19:09 of like the actual infrastructure,
    0:19:11 incentivizing developers to not use it.
    0:19:14 Like you literally have an economic incentive
    0:19:16 not to use AI because it costs you money.
    0:19:18 And like the more AI you build into your product,
    0:19:19 the more expensive it is.
    0:19:21 And like the more margin pressure you have
    0:19:23 as an application builder.
    0:19:24 – It’s scary too, right?
    0:19:25 To build an app and like all of a sudden
    0:19:27 you get like a gigantic bill
    0:19:28 because people are using your thing
    0:19:30 and you haven’t figured out how to properly monetize it yet.
    0:19:31 As you know, someone who creates companies,
    0:19:33 it’s kind of intimidating.
    0:19:33 – Yeah, 100%.
    0:19:35 That is like the realist reaction
    0:19:38 and like also just like the truest reaction developers
    0:19:40 have to the cost of the technology.
    0:19:43 So back to the point of like where the value creation happens.
    0:19:46 I think the nice thing for infrastructure providers
    0:19:48 is you have a fixed margin.
    0:19:51 So like, you know exactly how much money you’re gonna make
    0:19:54 by providing some infrastructure at the application layer.
    0:19:56 You’re like, you’re constantly incentivized
    0:20:00 to almost like not add additional stuff.
    0:20:01 And I think this has been the story
    0:20:03 for like the chat GBT plus and pro subscription
    0:20:06 is like they built a subscription for $20 a month
    0:20:08 and they realized, hey, we actually can’t give people
    0:20:09 all of these things anymore.
    0:20:11 We have to make something different.
    0:20:13 And like even at $200 a month,
    0:20:16 it’s like not a break even scenario for them yet.
    0:20:19 So it’s super interesting to see that play out
    0:20:21 and it’s a lot of food for thought
    0:20:22 for people who are building stuff to be like,
    0:20:25 are there new economic incentive mechanisms
    0:20:28 that you can create as you’re building a product
    0:20:31 more so than just like charging a $20 a month subscription.
    0:20:33 And like the one example I can think of of this
    0:20:35 that has made me think is,
    0:20:37 I don’t know if you all are familiar with OpenRouter,
    0:20:40 but it’s a surface that lets you sort of swap
    0:20:41 in and out different language models.
    0:20:44 OpenRouter’s product leaderboard,
    0:20:46 I’m pretty sure they give you discounts on tokens
    0:20:49 and stuff like that for showing you in certain ways
    0:20:51 and some metadata passing back and forth
    0:20:53 so they can understand like how people are generally
    0:20:54 using AI models.
    0:20:57 So like some interesting things like that.
    0:21:00 Alex Atala, who’s the CEO who previously worked on OpenSea,
    0:21:03 he has this quote, which always rings in my head,
    0:21:06 which is like usage is the ultimate benchmark.
    0:21:08 How many people are using your model or your thing
    0:21:10 is like the proof point of success,
    0:21:11 not all these other benchmarks
    0:21:12 that people are chasing after.
    0:21:14 So super interesting platform.
    0:21:15 – Do they actually publish like a leaderboard
    0:21:17 similar to like what LMSYS does?
    0:21:18 – Exactly. – Okay.
    0:21:21 – Checking it out right now at openrouter.ai
    0:21:23 and it’s a forward slash rankings.
    0:21:24 – Oh, cool.
    0:21:26 I’m curious, a little bit of a topic shift here,
    0:21:29 but I know you’re a proponent of open source
    0:21:31 and Google obviously has their Gemma models.
    0:21:34 Are there any updates, any idea of what’s going on with Gemma
    0:21:35 and what we can expect next
    0:21:37 out of the open source side of things?
    0:21:39 – Yeah, I think that this is also the piece
    0:21:42 that makes me excited about what we’re doing at Google,
    0:21:45 which is it really is the exact same research
    0:21:46 that powers the Gemma and I models
    0:21:48 that ends up making the Gemma models.
    0:21:53 And Gemma two was, I think it’s like the second most downloaded
    0:21:55 open source model in existence, which is awesome to see.
    0:21:57 And Gemma three is definitely going to happen.
    0:21:59 I think the timeline is soon.
    0:22:02 So you’ll hear more and y’all should do an episode
    0:22:03 with some of the Gemma folks
    0:22:04 ’cause there’s lots of cool stuff coming.
    0:22:05 – Yeah, for sure.
    0:22:07 – They’ve been doing a lot of like interesting fine tunes
    0:22:08 for different use cases.
    0:22:10 Like they have a version for RAG,
    0:22:12 they have a version for vision.
    0:22:16 I think they’re probably gonna do some agent stuff as well.
    0:22:18 So like there’s lots of really cool explorations happening
    0:22:20 on making those open models.
    0:22:21 – Super cool.
    0:22:23 I wanna talk about Imagine three,
    0:22:26 which I just learned is actually pronounced Imagine three.
    0:22:27 Like I’ve been calling it Imagine.
    0:22:29 I always thought it was like Image Generator.
    0:22:32 So like Image Gen, but I heard you pronounce it Imagine.
    0:22:34 So now I’ll start saying it that way.
    0:22:37 But you mentioned that there’s some updates with that as well.
    0:22:38 What can you tell us about that?
    0:22:39 – Yeah, and you’re in good company.
    0:22:42 Don’t worry, I swear, like 50% of the meetings I’m in,
    0:22:44 I hear Imagine, 50% I hear Image Gen.
    0:22:47 So yeah, there’s no conclusive answer.
    0:22:48 I think it’s Imagine,
    0:22:50 but someone should correct me if that’s not the case.
    0:22:54 So we released the Imagine three model
    0:22:57 across a couple of services late December,
    0:22:58 and then have been doing a bunch of work
    0:23:00 in the last few months to bring that to developers.
    0:23:03 So Imagine three should be available to developers
    0:23:07 in the API and is the frontier Image Generation model
    0:23:10 across like quality and a bunch of like human ranking
    0:23:12 benchmarks, which as an aside comment,
    0:23:15 it’s super interesting that if you look at text models,
    0:23:18 I think one of the reasons the world has so much success
    0:23:20 hill climbing on making models better
    0:23:22 is like there’s a definitive source of truth
    0:23:24 in some of the tasks that the models can perform.
    0:23:26 I think with Image models, it’s like actually not the case.
    0:23:28 It’s really hard to eval
    0:23:30 and like you actually need humans in the loop
    0:23:31 to do a lot of those evals,
    0:23:34 or it’s like artistic and stylistic stuff
    0:23:37 that’s like hard to put a finger on
    0:23:38 which of these two things are better.
    0:23:40 So a lot of those evals use human Raiders
    0:23:41 and human benchmarks.
    0:23:43 So there’s some degree of error,
    0:23:46 but yeah, it’s been exciting to see like the models available
    0:23:47 in the Gemini app.
    0:23:50 It’s available to enterprise customers
    0:23:51 and now it’ll be available to developers
    0:23:52 to build with, which is awesome.
    0:23:55 I think this like gen media future
    0:23:56 is going to be super exciting
    0:23:59 and VO hopefully sometime in the future as well.
    0:24:01 – Yeah, yeah, VO is awesome.
    0:24:02 I have access to it, early access to it.
    0:24:04 It’s super fun to play with.
    0:24:06 Is the best way to use the Imagine 3
    0:24:07 inside of the image effects?
    0:24:09 Is that still kind of like the easiest way
    0:24:11 for a consumer to just go and play around with it?
    0:24:12 – Exactly.
    0:24:13 I think it’s also available for free
    0:24:16 to folks in the Gemini app.
    0:24:17 I think if you ask to generate an image,
    0:24:19 it’ll just do it through the Gemini app,
    0:24:22 but ImageFX gives you a little bit more controllability
    0:24:23 and stuff like that.
    0:24:24 So there’s a few more like features
    0:24:26 that are built into ImageFX.
    0:24:28 So that’s definitely a place that it’s publicly available.
    0:24:29 – Super cool.
    0:24:32 Yeah, I know you and Nathan sort of before we hit record,
    0:24:34 we’re nerding out a little bit about, you know,
    0:24:37 the whole like sort of text application concept.
    0:24:38 He wished there was a way
    0:24:40 that you could have Unity open on the screen
    0:24:42 and then actually have an AI sort of like assist you
    0:24:44 with like where to click and what to do next
    0:24:46 while you’re building a game in Unity.
    0:24:48 And I think both you and I in Unison went,
    0:24:51 “You can do that in AI studio right now.”
    0:24:52 So I thought it might be kind of cool
    0:24:55 to pick up where we left off on that conversation
    0:24:57 and talk about some of the cool stuff
    0:24:59 that’s available inside of AI studio
    0:25:02 that maybe a lot of people don’t even realize exists
    0:25:04 and probably definitely don’t realize
    0:25:06 you could use most of it for free still right now too.
    0:25:07 – Yeah, cool.
    0:25:08 So everything in AI studio is free,
    0:25:10 which I don’t think people realize.
    0:25:12 Like the entire product experience,
    0:25:14 there is no paid version of it.
    0:25:15 There’s a paid version of the API,
    0:25:17 which hopefully developers can scale with
    0:25:18 and do all that fun stuff.
    0:25:22 But all of our latest models end up in AI studio for free,
    0:25:24 including the experience that powers
    0:25:27 this like real-time multimodal live experience,
    0:25:29 which if folks haven’t played around with it,
    0:25:33 aistudio.com/live lets you do things
    0:25:36 like share your screen or show your camera
    0:25:38 and ask all these different questions
    0:25:39 and interact with the model.
    0:25:40 There’s a bunch of different voices.
    0:25:43 There’s a bunch of different modalities to choose from.
    0:25:45 But back to the conversation of like,
    0:25:46 what will agents look like?
    0:25:48 What do we want out of agents?
    0:25:49 One of the limitations for agents
    0:25:51 is you have to build all this scaffolding
    0:25:54 for the agent to be able to see the things that you do.
    0:25:55 Like to see my email and my text
    0:25:57 and my et cetera, et cetera,
    0:25:59 my personal laptop and my work laptop
    0:26:01 or my phone and my watch,
    0:26:03 like incredible amount of work to make that happen.
    0:26:05 Except if you have a camera,
    0:26:07 all of a sudden all of it just works.
    0:26:10 And like you can sort of make the determination
    0:26:12 of being able to show like the information you want to show
    0:26:15 and share the stuff that you want to share.
    0:26:18 It’s definitely more of a showcase of what’s possible.
    0:26:20 And that’s why we put it in the API
    0:26:22 because like, we don’t have the answers ultimately.
    0:26:24 Like developers should go and build these products.
    0:26:25 But I do think Matt,
    0:26:27 you mentioned the Astra experience earlier.
    0:26:29 I think like the multimodal live API
    0:26:32 sort of gets to what Astra does at the core,
    0:26:36 which is like be able to be this co-presence with tools.
    0:26:38 And again, through a simple API, which is really exciting.
    0:26:40 I think the piece that the multimodal live API
    0:26:42 doesn’t have that will build,
    0:26:44 that I think the Astra experience did have
    0:26:46 is this notion of memory,
    0:26:48 which again is like critical for agents.
    0:26:51 Like I don’t want agents to just forget that,
    0:26:54 I prefer sitting in window seats instead of aisle seats
    0:26:55 or whatever it is.
    0:26:57 Like you want all that context to be retained
    0:26:59 as agents are making decisions for you in the future.
    0:27:02 And I think that’s gonna require this sort of memory layer
    0:27:04 which we’re working on building, which is exciting.
    0:27:05 – Yeah, yeah.
    0:27:07 And I mean, project Astra even sort of remembered
    0:27:10 between sessions too, that like I was in London, right?
    0:27:12 So I was talking to it while in London,
    0:27:16 asking for restaurants to go check out things like that,
    0:27:19 close that session, started another session later on.
    0:27:22 And it remembered that I was over in London,
    0:27:24 it remembered all the previous conversations.
    0:27:27 So it wasn’t just like memory in the terms that like,
    0:27:29 you can plug into the custom instructions on open AI
    0:27:32 and it’ll remember your name and stuff like that.
    0:27:35 It was actually like remembering the past conversations
    0:27:37 and bringing in that as additional context,
    0:27:38 which I thought was really cool
    0:27:40 ’cause that’s really helpful to just like remember
    0:27:41 the past conversations you had.
    0:27:43 – I think this is an infrastructure problem.
    0:27:45 And I think we didn’t talk about this explicitly,
    0:27:47 but like one of the other narratives
    0:27:49 over the last like year and a half has been like,
    0:27:51 not enough AI in production.
    0:27:54 You know, it’s kind of this demo toy thing
    0:27:55 and no one really uses it.
    0:27:58 I think a lot of this is ’cause it’s just like taking a while
    0:28:00 for companies to build the infrastructure
    0:28:02 to actually put AI into production.
    0:28:04 And I think memory is this example of,
    0:28:05 there aren’t a bunch of companies
    0:28:07 that are like building this memory as a service.
    0:28:09 And if you are building this, like let’s talk,
    0:28:11 I’d love to hear about it and hear about what you’re building,
    0:28:14 but I think there’s a lot of opportunity still to be built
    0:28:17 around that, around memory, like as a service for folks.
    0:28:18 You could also start to think about like,
    0:28:21 there’s so many interesting ways to explore this.
    0:28:23 Like where does all your personal context
    0:28:24 already live today?
    0:28:26 Like how does that, whoever that provider is,
    0:28:30 plug into the world of where all the other memory services
    0:28:30 are going to be.
    0:28:32 So I think there’s a lot of like really, really interesting
    0:28:36 directions that need to be built for memory specifically.
    0:28:37 – Logan, I’m curious.
    0:28:39 So you were talking about earlier the models
    0:28:40 that a lot of people don’t realize and all of it’s free,
    0:28:41 like an AI studio.
    0:28:44 Like why do you guys hide it in AI studio?
    0:28:46 Recently, I talked to a bunch of different people
    0:28:48 about DeepSeq and they were talking about how amazed
    0:28:49 they were by it.
    0:28:52 And I was like, yeah, but like you can get the same stuff
    0:28:55 but better for free on AI studio right now.
    0:28:57 And they didn’t know.
    0:28:58 And it’s like, there’s a lot of people who don’t know.
    0:29:00 And so I was like, you got to communicate that better
    0:29:02 somehow or like, I think you guys should have like,
    0:29:05 you know, its own website or something like outside of Google,
    0:29:06 like a new product where you guys just like,
    0:29:08 hey, here’s the new frontier
    0:29:10 and here’s what we’re pushing and Google is still there
    0:29:12 and it uses some of the tech, but we have a new thing.
    0:29:14 And that’s that personal opinion, but you know.
    0:29:15 – Yeah, yeah.
    0:29:17 No, I think you’re spot on and for what it’s worth.
    0:29:20 Like I think we get this feedback pretty consistently.
    0:29:23 I think some of this is a factor of just like the state
    0:29:27 of the world and the challenges that we have as a product.
    0:29:30 Like I think in one hand, like we are a developer platform.
    0:29:32 Like we’re not building the front door
    0:29:33 for Google’s AI surfaces.
    0:29:36 Like that’s not the product that I’m signed up to build.
    0:29:37 That’s not the product that like we’re sort of
    0:29:38 directionally building towards.
    0:29:40 We’re really focused on like,
    0:29:43 how do we enable builders to get the latest AI technology?
    0:29:47 The Gemini app formerly barred is the sort of front door
    0:29:49 to Google’s AI technology.
    0:29:51 And I think from a consumer standpoint
    0:29:52 and also from an enterprise standpoint
    0:29:55 and like workspace and other places,
    0:29:57 there’s all this interesting organizational work
    0:29:59 that’s happened at Google over the last couple of years.
    0:30:02 I think like one of the cool stories is like operationalizing
    0:30:05 Google DeepMind from doing this sort of foundational research
    0:30:07 to being an organization that builds
    0:30:09 the world’s strongest generative AI models
    0:30:11 and like actually delivers those to the world
    0:30:12 sort of as a product.
    0:30:15 And then now bringing the product surfaces
    0:30:18 that are the front line of delivering those to the world,
    0:30:21 the Gemini app Google AI studio into DeepMind
    0:30:23 so that we can sort of continue to accelerate.
    0:30:24 Like all of those things to me are like
    0:30:26 directly the right stuff for us to do.
    0:30:27 I agree with you.
    0:30:28 I think we need to put the models
    0:30:30 in front of the world as soon as possible.
    0:30:33 And I think having a single place to do that makes sense.
    0:30:34 And it should probably be the Gemini app
    0:30:37 that probably shouldn’t be AI studio.
    0:30:39 But at the same time I say that like
    0:30:40 we also want to be a surface
    0:30:42 to sort of showcase what’s possible.
    0:30:44 So there’s a lot of like tension points
    0:30:46 but I do think, I fundamentally do think
    0:30:47 we’re going to get there.
    0:30:49 The Gemini app is moving super quickly
    0:30:50 to like get the latest models.
    0:30:52 Like they just shipped 2.0 flash.
    0:30:54 They shipped 2.0 Pro today.
    0:30:55 They shipped the thinking model today.
    0:30:57 So I think that Delta between the Gemini app
    0:30:59 and AI studio is sort of going away.
    0:31:01 Which yeah, I’m excited about
    0:31:03 because like the consistent feedback
    0:31:04 is people don’t like that Delta.
    0:31:06 They want to have a single place to go to
    0:31:07 to sort of see the future.
    0:31:09 – I always saw the AI studio is sort of like
    0:31:12 the playground to test what the APIs are capable of.
    0:31:15 In the same way like open AI has their open AI playground
    0:31:18 and you can kind of go and mess with some of the settings
    0:31:20 and see what the output will look like
    0:31:21 before using it in your own APIs.
    0:31:24 That’s kind of how I always saw the AI studio.
    0:31:26 Because like once you get into it
    0:31:28 if you’re not very technically inclined
    0:31:29 you might get a little overwhelmed
    0:31:32 seeing things like what models should I be using?
    0:31:34 What is the temperature?
    0:31:35 You know, things like that.
    0:31:36 People aren’t necessarily going to know
    0:31:37 like how to play with that.
    0:31:39 Like what should I set my token limit to?
    0:31:40 Things like that.
    0:31:42 I don’t really feel like general consumers
    0:31:43 want to mess with.
    0:31:45 They just want to go to a chat bot
    0:31:46 and ask their question, right?
    0:31:50 It feels very tailored towards developers to me.
    0:31:52 And then the Gemini app feels like, all right
    0:31:53 this is their front user interface
    0:31:55 that they want the general public
    0:31:57 to go be using at least.
    0:31:58 – Yeah, I got a ping yesterday actually
    0:32:01 from someone saying why is the chat experience
    0:32:02 have all this stuff in it?
    0:32:04 Like why are there all these settings and stuff?
    0:32:06 And I was like, you’re in the wrong place.
    0:32:08 Like Gemini app is the place that you need.
    0:32:09 And then they responded right away on Twitter
    0:32:12 and they’re like, yes, this is much better for me.
    0:32:13 I don’t want to see all that complexity.
    0:32:15 But I do think about this a lot
    0:32:17 is like how do people actually show up?
    0:32:19 How are they finding their way into these products?
    0:32:22 I do think the Gemini app is like this very large front door.
    0:32:24 So it tends to capture most of these folks.
    0:32:27 Like it’s literally built into the Google app on iOS
    0:32:28 and all this other stuff.
    0:32:29 Versus like you actually kind of have to do
    0:32:31 a little bit of searching to find AI studio
    0:32:34 which probably makes sense in some cases.
    0:32:35 Awesome.
    0:32:36 – Logan, you said you were super excited
    0:32:38 about text application.
    0:32:39 You were talking about like Lovable, Bolt,
    0:32:41 other companies like that.
    0:32:42 Like what are you excited about in the space
    0:32:44 and where do you think that kind of stuff is going?
    0:32:48 – Yeah, I think just like being able to democratize access
    0:32:51 to people building software and like creating things.
    0:32:52 There’s a ton of people in my life
    0:32:54 and I don’t live in the Bay Area.
    0:32:56 So there’s a disproportionate amount of people
    0:32:57 who aren’t in tech where I live.
    0:33:00 But the proportion of like people with interesting ideas
    0:33:01 I actually think is the same.
    0:33:03 I think it’s just like the actual tools themselves
    0:33:05 that they have to go and execute on those ideas
    0:33:08 that I think is like much less distributed
    0:33:10 in places outside the Bay Area, New York
    0:33:11 and other places like that.
    0:33:15 So I think this frontier of text to app creation
    0:33:18 is gonna be so, so interesting to see play out.
    0:33:19 And yeah, there’s a ton of companies
    0:33:22 that are having like lots of like actual real early
    0:33:24 commercial success and traction today.
    0:33:28 Which I think, again, this is one of those examples
    0:33:31 where like sometimes there’s use cases that don’t work
    0:33:33 and then all of a sudden like the model quality
    0:33:35 just gets good enough for you build the right
    0:33:37 sort of couple of things from a product experience.
    0:33:39 And then all of a sudden it clicks
    0:33:40 and like now this thing is possible.
    0:33:43 And to me it feels like text to app creation
    0:33:46 like has had that moment and it’s now possible.
    0:33:47 And I think it’ll take a while
    0:33:50 and there’ll still be a bunch of other things to hill climb on.
    0:33:52 But I think especially now with like reasoning models
    0:33:54 and the ability for them to like keep thinking
    0:33:57 and writing more code and doing all that work.
    0:33:59 Like I think the complexity of the apps
    0:34:01 is also going to continue to go up on this exponential.
    0:34:03 So and actually replicate just how they’re launched.
    0:34:05 I think today or yesterday of like this
    0:34:07 a similar sort of product of text apps.
    0:34:09 I think there’s more and more players showing up
    0:34:10 in this space.
    0:34:13 I would assume that like probably 50% of products
    0:34:15 or something like that that are building with AI
    0:34:17 have this type of experience.
    0:34:19 And you could think about like, you know
    0:34:20 how does that translate to someone
    0:34:23 who’s doing something very, very domain specific?
    0:34:25 I think there’s a lot of like companies
    0:34:28 that try to build extension ecosystems or connectors
    0:34:31 or like, you know, all these other like side cars
    0:34:32 of their product.
    0:34:34 You can imagine like you just let your users create those.
    0:34:36 Like here’s the sort of generic set of APIs
    0:34:39 that talk to, you know, your email client, for example.
    0:34:42 And like here’s a text box and like go build
    0:34:44 the sort of product experience you want.
    0:34:45 Like it’s sort of in your hands.
    0:34:48 And like that’s a crazy world that you could totally customize
    0:34:49 it however you want to them.
    0:34:50 Yeah, it doesn’t feel that far away.
    0:34:52 My new email client that’s got like, you know
    0:34:54 80s style video game stuff.
    0:34:56 You know, it’s like a mixed end with the email client.
    0:34:57 That’d be so cool.
    0:34:58 You know, I’ve been loving that concept.
    0:35:00 We’ve talked about this a couple of times on the show
    0:35:03 of like, I’ve gotten in the habit now of like,
    0:35:05 when I have like a little problem or a bottleneck
    0:35:07 that I need to solve instead of going and like searching out
    0:35:10 if there’s like a SaaS company that already exists
    0:35:11 that has that product for me.
    0:35:14 If it’s simple enough, I’ll just go prompt
    0:35:15 that software into existence.
    0:35:17 And I have like a little Python script
    0:35:20 that runs on my computer to solve the problem for me, right?
    0:35:22 Like I made a little script where I can input
    0:35:24 my one minute short form videos into it.
    0:35:26 It automatically transcribes them
    0:35:27 and then cleans up the transcription
    0:35:30 and adds like proper punctuation and stuff.
    0:35:31 I created another mini app
    0:35:33 where I can drag and drop any image file.
    0:35:35 Doesn’t matter what type of file format it is.
    0:35:37 It’ll convert it to a JPEG for me
    0:35:39 so I can use it instead of like my video editing.
    0:35:42 And these are probably softwares that exist.
    0:35:43 I could go and hunt them down on the internet
    0:35:46 and maybe pay five bucks a month to use them,
    0:35:49 but I could just go use an AI tool,
    0:35:50 prompt the tool that I need.
    0:35:53 And 15 minutes later, I have something on my desktop
    0:35:56 that I don’t need to go pay anybody else for anymore.
    0:35:57 Maybe it connects to an API,
    0:36:00 like my transcription one connects to the OpenAI whisper API.
    0:36:03 So it is costing me like a penny every time I use it,
    0:36:04 but so what?
    0:36:06 I just love this concept of like
    0:36:08 when I have a bottleneck in my business,
    0:36:10 I can just go like prompt an app into existence
    0:36:12 that solves that bottleneck.
    0:36:16 – Yeah, I think that carried out one step farther
    0:36:18 towards this like infinite app store
    0:36:21 where like truly everyone is creating
    0:36:24 and contributing to this thing and like remixing.
    0:36:26 This is the stuff that gets me excited about the future
    0:36:28 ’cause like there’s so much cool stuff to be created.
    0:36:31 And really, I think that the lens of all of this is like,
    0:36:33 how do you democratize access and make it
    0:36:35 so that anyone can go and build this stuff?
    0:36:36 And as someone who can program,
    0:36:39 but also knows how painful it is in a lot of ways,
    0:36:41 it’s just like so cool that more folks
    0:36:43 are gonna be able to participate in that.
    0:36:44 It’s gonna be awesome.
    0:36:46 – Yeah, coincidentally, someone today
    0:36:48 from my hometown of Alabama like messaged me like,
    0:36:49 hey, I have this idea for an app.
    0:36:50 I get this kind of stuff all the time.
    0:36:52 I have an idea for an app and who can I hire
    0:36:53 to build it and all that stuff?
    0:36:55 And I’m like, I’m about to send him a link to like replit.
    0:36:56 Have you like tried this yet?
    0:36:58 You know, it’s like, just go try that.
    0:37:01 And instead of paying someone $5,000,
    0:37:03 that’s probably like a ton of money for him, right?
    0:37:05 Instead, just go try replit and you know,
    0:37:07 sign up for one month, cancel it if you don’t like that
    0:37:10 to that and then just see what you can get.
    0:37:11 And it’s gonna get better and better.
    0:37:12 Like I’ve tried replit and all of them
    0:37:13 and like they’re pretty good.
    0:37:15 It feels like there’s something that’s like slightly missing,
    0:37:17 but every time I check it, it’s better than the last time.
    0:37:20 And it feels like probably within the next year or two,
    0:37:22 you’re just gonna make any kind of software you want
    0:37:23 just by talking.
    0:37:25 It’s just, that’s gonna be such a magical moment.
    0:37:26 Like in the early days of the internet,
    0:37:27 the internet I feel like was more fun.
    0:37:29 ‘Cause people, there’s all these like different websites
    0:37:31 and different kinds of things or like you’d have Winamp
    0:37:32 and you put a skin on your Winamp.
    0:37:33 But there’s all these different things
    0:37:35 in terms of customization that was happening more
    0:37:36 than there is now on the internet.
    0:37:38 And it feels like this kind of stuff might bring that back
    0:37:39 where like, yeah, the internet,
    0:37:42 you can kind of customize how you interact with the internet
    0:37:44 through creating your own custom software with AI.
    0:37:45 – Yeah, I was just thinking about
    0:37:47 as you were describing like a fun internet,
    0:37:49 I was thinking of my personal website,
    0:37:51 which is like a blank HTML page.
    0:37:53 And like there’s no styling or anything like that.
    0:37:56 But like, if I didn’t have to shoulder the costs like,
    0:37:58 and the LLM on someone’s computer,
    0:37:59 I could just kind of like talk to and say,
    0:38:01 like, you know, remix this site
    0:38:02 and do it in all types of crazy ways.
    0:38:06 Like that would be so fun of like every time someone shows up,
    0:38:08 it’s a different product or it’s a different experience
    0:38:09 to see this content.
    0:38:11 And I think there’s a lot of interesting
    0:38:12 threads to pull on that.
    0:38:15 – Is there anything else happening at Google right now?
    0:38:17 Any other things that you’re working on
    0:38:19 that you’re allowed to talk about?
    0:38:21 Is there any avenues we haven’t gone down
    0:38:22 that you really wanna talk about
    0:38:24 that you’re allowed to talk about, I guess?
    0:38:26 – Yeah, I think the only other thread,
    0:38:28 and we alluded to it a couple of times
    0:38:29 is reasoning model stuff.
    0:38:31 It feels like, and I tweeted this the other day,
    0:38:33 it feels like the GBT2 era for these models.
    0:38:36 Like there’s so much new capability
    0:38:39 and so much progress being squeezed out of the models
    0:38:40 in such a short time.
    0:38:42 And we released our first reasoning model
    0:38:42 back in December,
    0:38:45 right after the Gemini 2.0 flash moment,
    0:38:46 one month later,
    0:38:49 like a normal like six months worth of progress,
    0:38:51 honestly, on like a bunch of the benchmarks
    0:38:53 that matter for this stuff.
    0:38:54 We released an updated version
    0:38:57 like January 21st, a couple of weeks ago.
    0:38:58 And if you look at the chart,
    0:39:02 it’s like literally linear progress up into the right
    0:39:03 across a bunch of the end.
    0:39:06 Like it’s just crazy to think that,
    0:39:08 again, like a month ago, the narrative was like,
    0:39:09 the models are hitting a wall,
    0:39:11 there’s no more progress to be had.
    0:39:14 And it’s funny like how much nuance
    0:39:15 some of the conversation lacks
    0:39:19 because these innovations are like deeply intertwined.
    0:39:21 I was having a conversation earlier today
    0:39:23 about like long context
    0:39:25 and how long context is actually like
    0:39:27 a fundamental enabler of the reasoning models
    0:39:29 because like by themselves,
    0:39:30 the long context innovation,
    0:39:33 like the model’s okay at pulling out
    0:39:34 certain pieces of information.
    0:39:36 Like it can do needle in a haystack well,
    0:39:39 it can find a couple of things and a million tokens,
    0:39:41 but it’s really hard for the models to attend
    0:39:43 to the context of like,
    0:39:44 you know, find a hundred things
    0:39:47 in this million token context window example.
    0:39:49 Reasoning models are the unlock for this
    0:39:50 because the reasoning models,
    0:39:52 the model can really just like continue
    0:39:53 to go through that process
    0:39:55 and think through all the content
    0:39:57 and like really do the due diligence.
    0:40:00 And it’s almost uncanny how similar it is to like,
    0:40:01 how would you go about this?
    0:40:03 Like I couldn’t watch a two hour movie
    0:40:06 and then if you quizzed me on a hundred random little things
    0:40:08 as part of like, I’m not going to get those things, right?
    0:40:10 Like it’s going to be really hard to do that.
    0:40:11 But if you let me go through the movie
    0:40:13 and you know, watching an eye movie
    0:40:17 and like add little inserts and like clip things
    0:40:17 and cut things and do all this.
    0:40:19 So like I’d be able to find those things
    0:40:21 if you asked me those questions again.
    0:40:23 And it feels like that’s kind of what reasoning is doing
    0:40:25 is actually being able to do that.
    0:40:28 So I think we’re super early in this progress
    0:40:29 and it’s going to be a lot of fun
    0:40:32 to see both the progress continue for us.
    0:40:35 But again, through this narrative of how all this innovation
    0:40:38 trickles into the hands of people who are building stuff.
    0:40:39 And like there’s going to be a ton of new products
    0:40:42 that get built, like maybe text to app
    0:40:44 just like it’s 10x better in the next year
    0:40:45 because of reasoning models.
    0:40:48 Like that’s possible, which is just crazy to think about.
    0:40:49 – Yeah, yeah.
    0:40:50 – But like the other models, they almost feel like they’re like
    0:40:53 double checking triple checking themselves in real time.
    0:40:56 It’ll be like sort of starting to give a response
    0:40:58 and then be like, let me actually double check
    0:40:58 what I just said.
    0:41:02 And when it comes to coding, that seems like it’s the ideal
    0:41:03 use case almost, right?
    0:41:05 Cause it can almost look back at its code and be like,
    0:41:07 oh, I think I made a mistake there.
    0:41:09 And sort of continually fix its code
    0:41:11 before it finally even gives you an output,
    0:41:13 which I’ve just found to be really, really cool.
    0:41:17 But also from what I understand, that’s where a lot of the cost
    0:41:18 and the future is going to come in.
    0:41:22 The cost of the inference to do all of this like analysis
    0:41:25 in real time as it’s giving its output.
    0:41:26 – Yeah, and the other thing to think about
    0:41:29 which is interesting is we’re seeing all of this progress
    0:41:33 with the reasoning models and they are doing
    0:41:36 like the most naive version of thinking.
    0:41:38 Like they really are like, if you were to think about
    0:41:40 like the human example of this, like you’re sort of sitting
    0:41:42 in a box like thinking to yourself.
    0:41:44 Like you have no interaction with the outside world.
    0:41:46 You’re not able to like test your hypothesis,
    0:41:50 use a calculator, search the internet, any of those things.
    0:41:52 And like you have to sort of form your thoughts
    0:41:54 independent of the outside world.
    0:41:55 And you imagine what starts to happen
    0:41:57 when you give these things tools.
    0:41:59 And like it really does feel like
    0:42:01 that’s the agentic future that we’ve been promised
    0:42:04 is like all of these tools in a sandbox interacting
    0:42:06 with the model letting it sort of have that feedback loop
    0:42:08 of trying things and seeing what doesn’t work.
    0:42:10 So I couldn’t be more excited about that.
    0:42:12 – Yeah, that’s really interesting to think about.
    0:42:15 So like right now it’s just sort of thinking through things
    0:42:17 and sort of double checking itself, but in the future
    0:42:19 it could actually be working with other tools
    0:42:21 that can also like assist in the double checking
    0:42:25 and things like that and get even smarter in those ways.
    0:42:26 – Yeah, and I think put a different way,
    0:42:28 like to be more extreme, I think it has to do that.
    0:42:30 Like I think the version of the future
    0:42:31 that we’re going towards is like,
    0:42:33 we’re not going to be able to see the progress
    0:42:35 continue to scale unless the models can do that.
    0:42:37 And again, this goes back to this thread of like,
    0:42:39 there’s lots of hard problems to solve in the world,
    0:42:41 like making it so the models can do that efficiently
    0:42:45 and like securely and safely have that sort of sandbox
    0:42:47 to do that type of thinking and work
    0:42:48 is going to have to happen.
    0:42:50 And it’s probably a lot of work that hasn’t been solved today,
    0:42:53 which is interesting and opportunistic.
    0:42:55 – Yeah, it’s crazy to think that like probably soon
    0:42:57 like AI is going to be helping create all those tools as well.
    0:42:59 So that’s when we’ll see things just go exponential.
    0:43:00 – It already is.
    0:43:02 – It’s like whether people with AI or AI itself
    0:43:05 creating the tools and just move that back into the system.
    0:43:06 And it’s going to be wild,
    0:43:08 how fast things are going to get better.
    0:43:10 – All the engineers being powered by cursor.
    0:43:11 It’s crazy, like it’s happening today.
    0:43:14 Like so many people are, I feel this way for myself,
    0:43:15 like I write more software now than I did
    0:43:17 when I was a software engineer
    0:43:20 because I have AI tools
    0:43:22 and I can do all this crazy stuff.
    0:43:22 – Yeah.
    0:43:24 – How far do you think we are from like AI
    0:43:26 actually being able to update its own weights
    0:43:28 based on conversation.
    0:43:31 So it actually learns based on new input
    0:43:33 that it gets through conversations that it has.
    0:43:36 – I think in the small scale example sense,
    0:43:39 you could probably already do this to a certain extent.
    0:43:43 I think in like the like real frontier use cases,
    0:43:44 probably far from that.
    0:43:47 Some of the open AI operator stuff was talking about this
    0:43:51 around like, you know, the need for having evals
    0:43:53 of like basically like creating economic value,
    0:43:57 like actually creating money and where we are in that.
    0:43:58 And like you probably don’t want the models
    0:44:00 to do things that have a high cost today
    0:44:02 because if they get it wrong, it costs you a lot of money.
    0:44:05 And training frontier models is definitely
    0:44:07 on the list of things that would cost you a lot of money.
    0:44:09 If you got that wrong, like you don’t want
    0:44:12 a bunch of training that are just wasted compute.
    0:44:14 Like that’s, you know, millions of dollars
    0:44:15 of potential loss money.
    0:44:17 So I think there’ll be a human in the driver’s seat
    0:44:19 for those things for a while.
    0:44:22 But I do think you can sort of accelerate this,
    0:44:24 you know, small scale feedback loop.
    0:44:26 And I think that’s why small models matter.
    0:44:28 Like this like innovation that’s happening
    0:44:31 of being able to compress the frontier capabilities
    0:44:32 down into small models.
    0:44:35 I think it enables that like rapid iteration loop
    0:44:38 where maybe AI is more a co-pilot in that example.
    0:44:39 – Gotcha.
    0:44:42 Well, cool Logan, this has been absolutely amazing.
    0:44:45 If people want to follow you, what’s the best platform
    0:44:46 to pay attention to what you’re doing
    0:44:49 and to keep up with what Google and DeepMind are up to?
    0:44:51 – Yeah, yeah, I’m on Twitter, I’m on LinkedIn,
    0:44:53 I’m on email.
    0:44:57 So whichever one of those three is easiest to get ahold of me
    0:45:00 would love to chat with folks about Gemini stuff or the like.
    0:45:03 – Yeah, you’re pretty active over on X slash Twitter,
    0:45:04 whatever you want to call it.
    0:45:07 Whenever there’s a new like Google or DeepMind rollout,
    0:45:08 you’re pretty much either tweeting about it
    0:45:09 or retweeting about it.
    0:45:11 So very, very good resource to keep up with
    0:45:14 what’s going on in the world of AI with Google.
    0:45:16 And Logan, thank you so much for hanging out again
    0:45:17 with us today.
    0:45:19 I’m sure we’ll have you back in the future if you want,
    0:45:21 but this has been an absolutely fascinating conversation.
    0:45:23 So thanks again for hanging out.
    0:45:24 – Yeah, this is a ton of fun.
    0:45:25 I’ll see you both at IO, I hope.
    0:45:27 I think hopefully we’ll get the game back together
    0:45:29 and we’ll spend time in person.
    0:45:31 Hopefully at IO, it’s gonna be fun.
    0:45:31 – We’d love to do it.
    0:45:32 Thanks.
    0:45:33 – Thank you.
    0:45:36 (upbeat music)
    0:45:38 (upbeat music)
    0:45:41 (upbeat music)
    0:45:43 (upbeat music)
    0:45:46 you

    Episode 45: How is Google shaping the future of AI with its new Gemini models? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) are joined by Logan Kilpatrick (https://x.com/OfficialLoganK), Senior Project Manager at Google DeepMind.

    In this episode, delve into the details of Google’s latest AI models, Gemini 2.0, Flash 2.0, and Pro versions, as Logan Kilpatrick breaks down the advancements and unique capabilities that set these models apart. They discuss the cost-efficiency that Gemini brings to the table, the concept of reasoning models, and how agents are paving the way for future AI applications. Whether you’re a developer or just intrigued by the progress in AI, this conversation offers insights into what Google’s innovations mean for the industry.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Gemini 2.0 Launch Excitement
    • (03:18) Cheaper Flashlight Model Previewed
    • (08:50) Experiencing Gemini AI in London
    • (11:11) AI Agents: Need Proactive Models
    • (14:23) Embracing Inefficiency for Productivity
    • (17:09) AI Infrastructure and Consumer Impact
    • (21:31) Imagen 3 Model Update & Insights
    • (24:18) AI Studio: Free Multimodal Experience
    • (26:53) AI Production and Infrastructure Challenges
    • (31:56) Democratizing App Creation Tools
    • (34:09) DIY Software Solutions
    • (38:51) Reasoning Models Unlock Contextual Understanding
    • (42:45) AI Frontier: Risks and Costs
    • (44:12) AI Updates on Twitter

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • Is Deepseek Worth the Hype?

    AI transcript
    0:00:05 The world went nuts over this new Chinese AI model called DeepSeek.
    0:00:08 Things are going to accelerate even more because now you’ve got China as a real player.
    0:00:11 From here on out, I feel like pretty much all of the models are going to have this
    0:00:13 sort of thinking element to it.
    0:00:16 The world is going to be dramatically different in like two years from now.
    0:00:19 I mean, these models are probably going to be like 50 times smarter,
    0:00:21 at least 20 times smarter in two years.
    0:00:27 Hey, welcome to the Next Wave Podcast. I’m Matt Wolf. I’m here with Nathan Lanz.
    0:00:30 And today, we’re going to talk about the thing that the whole world’s been talking
    0:00:32 about over the last couple of weeks.
    0:00:35 We’re going to talk about DeepSeek and DeepSeek R1.
    0:00:37 And we’re going to break it all down for you.
    0:00:40 We’re going to explain to you why it crashed the stock market,
    0:00:43 why we think it’s actually been a huge overreaction.
    0:00:45 We’re actually going to show it off and use it.
    0:00:47 We’re going to see if we can confuse it.
    0:00:50 And I think we were pretty successful at actually confusing it.
    0:00:52 We’re going to test how actually biased it is.
    0:00:56 And then we’re also going to figure out how to get around that bias
    0:00:58 and show you how you can also get around that bias.
    0:01:03 If you ever wanted to know everything there is to know about DeepSeek R1,
    0:01:04 this is the episode for you.
    0:01:07 So let’s just go ahead and dive in and show it all off to you.
    0:01:12 It’s been a roller coaster of a couple of weeks, the last few weeks,
    0:01:18 mostly because the world went nuts over this new Chinese AI model called DeepSeek.
    0:01:21 More specifically, DeepSeek R1,
    0:01:27 which is a model that came out of China and supposedly was trained on what?
    0:01:34 Like 2,800s, I believe, which are like nerfed versions of each 100s from Nvidia.
    0:01:41 So to give some context here, China and the US are sort of like in this space race going on
    0:01:43 for like who gets the top AI.
    0:01:51 And as a result, the US is restricting the level of power that GPUs sold to China can have.
    0:01:55 Right? And so what Nvidia has done is they’ve taken their H100s,
    0:02:00 which are their powerful GPUs that most of these AI models are trained on,
    0:02:01 and they somehow nerfed them.
    0:02:03 They made them less powerful.
    0:02:07 And then they sold those to China as H800s.
    0:02:16 And supposedly this DeepSeek V3 was trained on 2,000 of these H800s for $5 million,
    0:02:20 where all the big models from companies like OpenAI and Anthropic
    0:02:25 have been trained for way, way, way more on these way more powerful GPUs.
    0:02:29 And everybody’s freaking out because these new models that were trained much cheaper,
    0:02:36 much quicker on much lower grade hardware are actually getting comparable results
    0:02:40 to some of the state of the art models from companies like OpenAI and Anthropic.
    0:02:44 That’s kind of like the context you need for the freak out.
    0:02:47 But Nathan, is there anything I’m like missing in there?
    0:02:49 Because I feel like I explained that fast.
    0:02:51 I think that’s about it.
    0:02:55 I think the biggest shock to me was that the stock market went way down because of this.
    0:02:56 Nvidia went down 17%.
    0:03:00 I was like, why? Do people just not understand this at all?
    0:03:01 Like I couldn’t wrap my head around it.
    0:03:04 I was like, okay, yeah, the short term markets are not logical.
    0:03:06 You know, why it’s not a good idea to do stock trading.
    0:03:09 If I looked at that news, I’d be like, oh, I’m going to buy more Nvidia.
    0:03:12 But if I was like an option trader, I’d be buying Nvidia calls, you know,
    0:03:14 and instead it goes down 17%.
    0:03:18 So the thinking, I think the consensus and the reason it went down a little bit
    0:03:23 was a lot of people went, oh, they trained this on a lot less GPUs,
    0:03:25 that were also a lot less powerful GPUs.
    0:03:29 This is proof that Nvidia GPUs aren’t as necessary
    0:03:32 because these companies can do them with way less.
    0:03:34 They can train these models with way less now.
    0:03:39 So now like, do we really need a half a trillion dollar data center
    0:03:42 to build all these big AI models?
    0:03:44 Probably not because look at what China is doing.
    0:03:45 That was the thinking.
    0:03:47 I’m not saying that’s the reality of it.
    0:03:51 But that was the thinking behind why it crashed the markets, right?
    0:03:54 But at the end of the day, I still found it bizarre
    0:04:00 because they were trained with H800, which were still Nvidia GPUs, right?
    0:04:01 We believe they were.
    0:04:04 I wouldn’t take any of it at face value, but yes, that’s a claim.
    0:04:04 Okay.
    0:04:04 Yeah.
    0:04:07 The claim was they were trained with Nvidia H800.
    0:04:11 The, I don’t want to say conspiracy because it’s not really a conspiracy.
    0:04:14 There’s actually people that claim they have evidence.
    0:04:17 This sort of counter argument is that they were actually probably
    0:04:19 really trained on H100s.
    0:04:22 So either way, still trained on Nvidia GPUs, right?
    0:04:23 Right.
    0:04:25 But you know, some of the speculation that’s going around
    0:04:29 is that they claim that they use the H800s
    0:04:33 because they shouldn’t have access to H100s.
    0:04:37 So they have to claim they used what they’re allowed to have access to.
    0:04:38 So that’s like one of the claims.
    0:04:41 Some of the other claims that I’ve been seeing circulating too
    0:04:47 is that they use like distilled versions of OpenAI’s chat GPT.
    0:04:50 And so like the big expensive training part that companies
    0:04:54 like OpenAI normally go through, they sort of skipped that step
    0:04:58 because they just use existing training data that was out there.
    0:05:01 I think there’s some screenshots going around where if you ask it,
    0:05:06 who made you, it’ll say, oh, I’m chat GPT made by OpenAI, right?
    0:05:06 Yeah.
    0:05:08 No, I mean, there was a report coming out.
    0:05:11 It sounds like OpenAI and Microsoft are like investigating
    0:05:14 and they believe they have evidence that that did happen.
    0:05:15 So who knows what’s going to happen?
    0:05:17 I wouldn’t doubt that there’s like lawsuits.
    0:05:19 Maybe America tries to ban DeepSeek or something like this
    0:05:21 would not surprise me at all.
    0:05:22 Yeah.
    0:05:27 I mean, I feel like DeepSeek and stuff like it are a bigger threat than TikTok.
    0:05:29 But I don’t know, that’s just me.
    0:05:30 Right, right.
    0:05:32 I guess getting back to like the stock dropping thing.
    0:05:35 So even if that stuff was true, even if they have some huge innovation
    0:05:38 and they trained it on dramatically lower quality graphics cards,
    0:05:40 we’ve still discovered that like test time compute
    0:05:42 is where the scaling is going to happen.
    0:05:45 Like even like with the R1, that’s why you’re seeing the big improvement.
    0:05:50 So you still would need lots of processing and compute to scale that up
    0:05:52 to get smarter and smarter.
    0:05:55 So even if that was true, it took me a race of like, okay, cool.
    0:05:57 They’ve made it incredibly more efficient.
    0:05:57 Okay, great.
    0:05:59 We can throw even more money at it.
    0:06:00 So like we want less intelligence.
    0:06:02 It doesn’t make any sense.
    0:06:05 Yeah. Yann LeCun, he basically said what you just said,
    0:06:09 but you know, it’s always good to get it validated by a AI scientist.
    0:06:10 You’re saying I’m not an AI scientist?
    0:06:16 Yann LeCun says, “Major misunderstanding about AI infrastructure investments.
    0:06:20 Much of those billions are going into infrastructure for inference, not training.
    0:06:25 Running AI assistant services for billions of people requires a lot of compute.
    0:06:27 Once you put video understanding, reasoning,
    0:06:30 large scale memory and other capabilities in AI systems,
    0:06:32 inference costs are going to increase.
    0:06:35 The only real question is whether users will be willing to pay enough
    0:06:38 directly or not to justify the CAPEX or OPEX.
    0:06:42 So the market’s reaction to deep seek was woefully unjustified.
    0:06:43 Yeah, obviously, I agree with that.
    0:06:45 I think it doesn’t make any sense.
    0:06:48 Like even if you trust what they’re saying, it doesn’t make any sense.
    0:06:50 And also just in general, the framing of it too was like,
    0:06:53 oh, this is like the best model for like regular people.
    0:06:55 Like OpenEIA has been beaten.
    0:06:56 And it’s like, what?
    0:06:59 Dario from Anthropic put out a blog post highlighting that, well,
    0:07:01 you know, this is not the best model.
    0:07:03 As right now based on the benchmarks,
    0:07:06 this is a model that’s as good as the American models from like nine months ago.
    0:07:09 And possibly trained slightly cheaper,
    0:07:14 but it’s kind of in line with what they’ve been seeing as well as like all the models.
    0:07:15 Yeah, they’re getting cheaper and cheaper.
    0:07:16 Oh, three minutes must be coming out.
    0:07:19 And they’re saying what it was like 10 times cheaper.
    0:07:20 I forget the exact number.
    0:07:23 But like all these models are getting cheaper to use through the API over time.
    0:07:25 So like maybe they have one or two interesting innovations,
    0:07:27 but it’s not like they’ve now won or something.
    0:07:28 It’s crazy.
    0:07:29 Yeah, it’s crazy.
    0:07:34 And I think a lot of the world learned about the thing called Jevon’s Paradox as well.
    0:07:37 I don’t think most people had ever heard of that until this week
    0:07:39 when Satya Nadella from Microsoft tweeted about it.
    0:07:43 But basically, Jevon’s Paradox is saying paradoxically,
    0:07:47 when you need less of something to actually accomplish the things,
    0:07:50 the demand form actually goes up, not down, right?
    0:07:55 Essentially, what’s likely going to happen is if it is true
    0:07:59 that we can train these models for way, way cheaper than anybody ever anticipated,
    0:08:02 and we can get really, really powerful models for a lot cheaper,
    0:08:04 well, that lowers the barrier to entry,
    0:08:07 meaning that more companies can get involved.
    0:08:12 More companies are going to want to buy GPUs to get in the game of creating foundation models.
    0:08:16 And not only that, but just because we can create more powerful models
    0:08:19 with less expensive compute,
    0:08:22 these companies are still going to want to throw as much compute as possible.
    0:08:26 Because if we can produce this with this level of compute,
    0:08:29 imagine what we can produce with this level of compute, right?
    0:08:32 So it’s still going to have that thing where these companies
    0:08:35 are still going to want to throw more and more and more compute at it
    0:08:37 to see how much smarter, how much better,
    0:08:40 how much more impressive they can make these models,
    0:08:43 despite the fact that maybe you don’t need as much as you originally thought.
    0:08:47 Well, cool. That means maybe we can even go bigger than we thought.
    0:08:50 Right. Yeah, another thing too is this is open source.
    0:08:53 So if there is something truly innovative they have done,
    0:08:56 openAI, Anthropic, XAI, Google,
    0:08:59 they’re all going to learn from it and implement it into their next models.
    0:09:03 Like for sure, openAI and all of them are in their research labs right now,
    0:09:04 playing with this stuff.
    0:09:07 And if there’s something valid to how they did deep seek,
    0:09:09 that’ll be integrated into the next models they develop.
    0:09:14 So I think the whole idea of China’s now surpassed America and AI is overblown.
    0:09:16 But it is, like people have said, it’s like a wake-up call.
    0:09:19 It’s like they are heading robotics and drones, it appears.
    0:09:22 And the kind of thing is like, “Well, we’re heading AI, so that’s good.”
    0:09:26 So we can catch up in robotics and drones, but we’re heading AI.
    0:09:28 They have caught up a bit in AI.
    0:09:32 They have not surpassed America or beat America or even matched America yet,
    0:09:33 but they have dramatically caught up.
    0:09:33 Yeah.
    0:09:36 But it’s hard to know how much they’ve actually caught up,
    0:09:39 because Sam Altman tweeted out, his tweet was kind of like,
    0:09:40 “Oh, it’s impressive in a few small little ways,
    0:09:43 kind of a few little model you guys have built.
    0:09:47 We’ll continue to put out the best models and we’ll pull out some models.”
    0:09:52 Which to me is hilarious, because it’s kind of like what I’ve been saying is behind the scenes,
    0:09:56 opening AI has dramatically better tech that they haven’t shown off yet.
    0:09:57 And so it’s like, yeah, we’ll pull out some.
    0:10:00 So it’s like, they already have stuff ready that’s dramatically better.
    0:10:03 And whenever they want, they’ll dole them out to us to try out.
    0:10:06 Yeah, yeah.
    0:10:10 Well, I mean, there was another model too that came out this week in Quinn 2.5,
    0:10:14 which I don’t believe is open source, but it’s another model that came out from China
    0:10:19 that they’re claiming actually outperforms the DeepSeek models,
    0:10:22 which the DeepSeek models, according to most benchmarks,
    0:10:24 are about as good as the O1 model.
    0:10:28 So even more models coming out of China, there’s also that DeepSeek Janus.
    0:10:31 Did you see that one, the DeepSeek Janus model,
    0:10:35 which is DeepSeek also released an AI image generator model.
    0:10:36 Oh, I did.
    0:10:38 Similar to like Stable Diffusion.
    0:10:40 You know, but to me, that didn’t seem anything special, right?
    0:10:43 Like it’s not going to get people freaking out like they did with DeepSeek,
    0:10:45 because to me, it just kind of looked like, oh, cool,
    0:10:50 it is about as good as like one of the mid-level AI generators, right?
    0:10:53 It’s probably not as good as like what Flux could generate right now.
    0:10:54 It’s probably not as good.
    0:10:57 It went like a really good mid-journey prompt could generate right now.
    0:10:59 But, you know, it’s pretty good.
    0:10:59 It’s all right.
    0:11:01 All right, all right.
    0:11:03 So, you know, there’s a lot of different benchmarks out there,
    0:11:05 and all of them will have different results.
    0:11:06 But this is a pretty well known one.
    0:11:08 It is from someone who’s from OpenAI.
    0:11:10 So take that, you know, it’s a grain of salt.
    0:11:13 But the benchmarks, at least in my use cases,
    0:11:16 this matches kind of what I’ve seen, you know, in terms of quality.
    0:11:19 I’m still seeing that I’m getting the best results with O1 Pro,
    0:11:20 which is not even on here.
    0:11:25 And underneath that, probably O1, also Google’s new thinking model is actually really good.
    0:11:26 It’s really good.
    0:11:29 They keep upgrading it, and it keeps getting better and better.
    0:11:32 I never thought I’d see the day where you give Google credit for their AI work.
    0:11:35 You know, every time Logan tweets something out, I try it.
    0:11:38 I thought Logan’s tweet, though, people were talking about DeepSeek,
    0:11:41 and he was like, if we released Gemini as a standalone app,
    0:11:42 it would be number one in the app store.
    0:11:45 I’m like, oh my God, when’s the last time Google’s done that with any product?
    0:11:48 I was like, but the underlying tech is cool.
    0:11:50 Yeah, get someone else to handle the brand and stuff.
    0:11:53 But, and this is kind of what I’ve seen, too, like Claude’s still very good.
    0:11:55 And I would say that DeepSeek, for me right now,
    0:11:58 is like right underneath Claude, you know, especially with coding.
    0:11:59 There’s sometimes it surprises me.
    0:12:00 I’ve tried DeepSeek.
    0:12:03 Sometimes it surprised me, like, oh, it’s pretty good at certain things.
    0:12:05 But other things, it just makes dumb mistakes
    0:12:08 that I haven’t seen any of the modern models make in like a year now.
    0:12:08 Yeah.
    0:12:10 Like when I was asking it to help me with coding,
    0:12:12 it would like imagine files that didn’t exist.
    0:12:15 It started telling me to like create stuff in them.
    0:12:16 I’m like, what?
    0:12:18 You want me to edit this file that doesn’t exist?
    0:12:20 What are you doing right now?
    0:12:23 And it’s like kind of crazy that it’s supposed to have some kind of logic engine
    0:12:24 on top of that should have caught that.
    0:12:27 Like how is it still, it’s the quality is not there.
    0:12:29 But people shared some examples with writing and stuff.
    0:12:31 I thought we’re really impressive with R1.
    0:12:33 So there’s like certain areas where it is quite good,
    0:12:37 but it’s not like, oh, it’s being all the American models not even close.
    0:12:38 Yeah.
    0:12:38 I’m curious.
    0:12:40 This benchmark you’re showing on the screen right now,
    0:12:42 do you know how it was created?
    0:12:44 It says number of valid responses.
    0:12:47 So I’m guessing it asked it a bunch of questions
    0:12:50 and whether it got the questions right or not.
    0:12:52 Do you know how this benchmark was created?
    0:12:52 I don’t know.
    0:12:54 Well, I don’t honestly, I don’t know.
    0:12:56 I know the guy Aiden, I don’t know him,
    0:12:59 but I’ve seen him share a lot of benchmarks in the past.
    0:13:01 People seem to really trust his benchmarks.
    0:13:02 He does work at OpenAI.
    0:13:05 You know, I believe that OpenAI has some of the best people.
    0:13:09 So I don’t think they just put out a benchmark to say, oh, OpenAI is the best.
    0:13:10 But yeah, no, I wouldn’t say that either.
    0:13:12 Mine is kind of like a gut feeling thing that’s like,
    0:13:16 I see a benchmark and it matches exactly what I’ve been experiencing firsthand.
    0:13:19 So like, I’m like, okay, you know, at least for me, that’s good enough.
    0:13:19 Well, here’s the deal.
    0:13:23 For me, I have access to, I think every single model
    0:13:25 that was on that benchmark list that you showed there, right?
    0:13:26 I think I have access to all of them.
    0:13:29 I didn’t notice one that I haven’t gotten to play with yet.
    0:13:33 And to this day, I still find myself going to Claude.
    0:13:36 I still find myself using chat GPT occasionally,
    0:13:38 and I still find myself using Gemini.
    0:13:41 Those are the three that I find myself using.
    0:13:43 Even though I saw deep seek, I use deep seek.
    0:13:48 I was fairly impressed by, you know, seeing it think through everything as it talked.
    0:13:51 It hasn’t made its way into my daily workflow.
    0:13:52 I wasn’t impressed with it enough to go, oh,
    0:13:55 I’m going to start using this now instead of one of the others.
    0:13:56 I still go to Claude.
    0:13:59 I really like Gemini too, but I’m more like Gemini
    0:14:02 because I love the massive context window.
    0:14:04 But I can upload a huge documents in there
    0:14:06 and ask questions about the documents.
    0:14:07 So I use Gemini for that.
    0:14:11 I also really, really think their deep research is really good.
    0:14:13 It’s like perplexity on steroids, right?
    0:14:15 You go in there, you ask it a question.
    0:14:19 And where perplexity might find five or six sources to answer your question,
    0:14:22 deep research might find 200 sources to answer your question.
    0:14:26 So like I really, really like their deep research, you know,
    0:14:28 01 Pro still top of the line.
    0:14:32 Like I haven’t found anything that gives me responses as good as 01 Pro.
    0:14:36 So as good as deep seek is, and as much as everybody’s talking about it,
    0:14:41 I never used it and went, oh, I’m going to start using this one now instead, you know?
    0:14:41 Right.
    0:14:43 People have been trying to study like, what happened?
    0:14:45 Like, why did this blow up so much?
    0:14:48 Like, why is the perception not matching the reality of like how good this thing is?
    0:14:52 And probably a lot of it is that a lot of people use chat to BT,
    0:14:56 like a year or two ago, and they still been using like the older free models and stuff like that.
    0:15:00 And they’re not even using like the better paid models.
    0:15:03 So they have not even experienced 01 yet.
    0:15:05 And so when they saw something like deep seeking, like, oh, it’s thinking,
    0:15:07 that’s so cute and wow, it’s so smart.
    0:15:08 Look, it’s thinking.
    0:15:10 It’s like, oh my God, like OpenAI has already been doing this.
    0:15:13 But OpenAI has been kind of hiding, like they show the thinking,
    0:15:17 but it’s a filtered version of the thinking because they didn’t want companies to do
    0:15:21 what supposedly maybe deep seek has done in terms of like copying their responses
    0:15:22 to train the model.
    0:15:23 I think you nailed it.
    0:15:24 I think that’s it.
    0:15:29 I think like the reason it blew up so much is I still think the majority of the world
    0:15:34 is using like the chat GPT free plan or the cloud free plan.
    0:15:39 They haven’t actually used like the upgraded models that we know to be really, really good.
    0:15:45 And, you know, I think most people that have used chat GPT plus or on the pro plan and use the
    0:15:51 01 pro mode, I don’t think anybody that’s been using those models for a while now is nearly
    0:15:56 as impressed with deep seek as all of the people that have been using these free models recently
    0:15:59 and then seeing like, oh, they’re actually showing how they’re thinking about this.
    0:16:02 It’s like, yeah, but if you’ve been paying attention,
    0:16:04 models have been doing this for a while now.
    0:16:06 I think more of the big blow up around it.
    0:16:10 Yes, it shows the thinking that feels novel to a lot of people,
    0:16:12 but it’s also the cost element, right?
    0:16:17 I think that’s the other big piece of the news that everybody keeps circulating is like,
    0:16:21 well, it was only trained on this many H 800s and they only did it for $5 million,
    0:16:24 where open AI did it for $60 million.
    0:16:28 I mean, I don’t think open AI has publicly talked about how much they paid to train their models,
    0:16:32 but, you know, tens of millions, if not hundreds of millions to train their models,
    0:16:35 where this one was trained for $5 million supposedly.
    0:16:38 And it’s hilarious to me that people take that at face value.
    0:16:41 Like, I mean, sorry, but like I, you know, I studied manner and I was like a friend
    0:16:44 in the Chinese government before I’ve dealt with like Chinese investors,
    0:16:47 Chinese VCs, Chinese tech companies.
    0:16:49 It is like widely known in China in business that like,
    0:16:52 it’s totally okay to like play around with numbers and use them as a strategy.
    0:16:54 Yeah. And I mean, it’s not like they showed receipts.
    0:16:57 There’s no ramifications for lying about it.
    0:17:01 Like, especially if you have any benefit to the Chinese government, totally fine.
    0:17:04 Totally fine and totally fine to do that as a strategy.
    0:17:09 Yeah. And the Chinese government seeing this deep seat company is like heroes of the country
    0:17:12 because they managed to, you know, hit the stock market.
    0:17:16 Which I do wonder how long that’s going to last because then it’s kind of interesting
    0:17:20 that Ervin from a perplexity was showing that he’s like kind of modified some things
    0:17:21 and he’s using the open source model.
    0:17:23 I’m not sure what he’s done to change it, but like,
    0:17:27 you can now ask perplexity things that the deep seat will not answer.
    0:17:29 Like deep seat will not answer stuff about team square.
    0:17:31 Well answer stuff about Taiwan.
    0:17:34 And apparently on perplexity using R1, it’ll answer all of that.
    0:17:38 So I do wonder how long that relationship is going to last.
    0:17:38 Let’s try one.
    0:17:42 I’m actually curious to see if we can get it to generate something.
    0:17:43 It’s Taiwan a country.
    0:17:44 That’s a very simple test.
    0:17:47 So they just added this new one with, you can do reasoning with R1.
    0:17:49 They had reasoning with 01 for a little bit.
    0:17:52 I wonder if I should test this first with the regular R1
    0:17:55 and then see perplexity’s response to it.
    0:17:57 Maybe that might be the better route to go.
    0:17:57 Yeah, let’s do that.
    0:17:59 We were talking about like why it blew up.
    0:18:01 And I think there was one other thing I found interesting.
    0:18:04 There’s this one person at Anthropic who’s pretty popular on X called Neersai.
    0:18:08 And they’ve always shared like really great stuff about AI research and stuff.
    0:18:13 And they were really trying to figure out like why did deep-seek blow up so much?
    0:18:16 And they did all kinds of analysis on the internet on like Reddit,
    0:18:20 on social media, on news, trying to look at different keywords and figure out
    0:18:23 why did the model blow up.
    0:18:25 And this is definitely not a political person as all.
    0:18:28 As far as I can tell, they’re not like a Trump supporter or anything like that.
    0:18:31 They found weird associations like people hating Trump and then promoting deep-seek,
    0:18:33 including Americans.
    0:18:34 Weird.
    0:18:36 Like some kind of weird thing which has no,
    0:18:38 like there’s no relationship there at all.
    0:18:41 But some weird thing where it was like people who dislike Trump,
    0:18:43 people who dislike big corporations,
    0:18:49 dislike America, like all this kind of stuff seem to be promoting deep-seek more,
    0:18:52 including the American media, which was just like wild.
    0:18:54 That is interesting because the American media really,
    0:18:57 really has talked a lot about deep-seek, right?
    0:18:59 We’ve been hearing more about deep-seek.
    0:19:03 Like I would have thought we would have been hearing about Stargate projects all over.
    0:19:04 Yes, so much more.
    0:19:05 I don’t really watch the news,
    0:19:10 but like I feel like that was way under-reported on where deep-seek was way over-reported on.
    0:19:13 I think it was just literally because of the spokesman being Trump.
    0:19:16 It’s like Trump literally had almost nothing to do with it.
    0:19:19 Like it’s literally, it’s like Masayoshi and Sam Altman, you know,
    0:19:21 and Oracle, Larry Ellison.
    0:19:23 So it’s like, but I think since Trump was in the room,
    0:19:26 it’s like, what if he has almost no connection to it?
    0:19:27 So bizarre.
    0:19:28 But that’s what the data showed.
    0:19:30 He said there was a lot of things that were surprising.
    0:19:32 A lot of it was just like people who don’t like corporations and stuff.
    0:19:35 And so somehow they see deep-seek as being open source.
    0:19:36 And so that’s why they love it.
    0:19:39 But the number one data was anti-Trump sentiment.
    0:19:39 That’s so weird.
    0:19:41 And I was like, wow, what?
    0:19:43 Don’t get it.
    0:19:45 Well, let’s play around with deep-seek a little bit.
    0:19:46 I haven’t opened an LM studio.
    0:19:48 So this is running locally on my computer.
    0:19:50 I’m not on deep-seek’s website.
    0:19:54 I actually downloaded one of the distilled models and put it on my computer.
    0:19:58 I don’t know how I feel yet about actually using deep-seek’s website.
    0:19:59 Don’t give it everything.
    0:20:01 So I’m actually using a locally installed model.
    0:20:02 This is running on my computer.
    0:20:04 It’s not hitting the cloud at all.
    0:20:08 It’s deep-seek R1 distilled based on the Quinn 14b model.
    0:20:10 I’ll just give it like an easy prompt real quick,
    0:20:14 like how many Rs are in the word strawberry?
    0:20:15 Right.
    0:20:18 That was one of the tests that a lot of people was doing for a while
    0:20:20 because GPT-4 couldn’t actually answer this question.
    0:20:23 But if I send this, you’ll notice it has this thinking
    0:20:26 where it actually shows you what it’s thinking through.
    0:20:30 And down here, it actually shows its logic here.
    0:20:30 Wait.
    0:20:34 It actually recounted it a second time to make sure it wasn’t missing.
    0:20:38 So look, I’m looking specifically for the letter Rs, not an R.
    0:20:39 T, nope, R.
    0:20:40 Yes, that’s one.
    0:20:44 A, no, W, no, B, no, E, no, R. Another one.
    0:20:45 So that’s two.
    0:20:46 Wait.
    0:20:47 Is there a third R here?
    0:20:47 No.
    0:20:52 I think after the first R at position three, the E comes next, followed by two Rs.
    0:20:54 So positions eight and nine are both Rs.
    0:20:58 Wait, let me recount to make sure I’m not missing any starting from the beginning.
    0:21:00 S, T, R, that’s one.
    0:21:04 A, W, B, E, R, two, R, three.
    0:21:06 And then it sort of figures out its logic again.
    0:21:10 But you can see it’s like recounts a double checks and quittruple checks itself
    0:21:14 until it’s like certain there are three Rs in the word strawberry.
    0:21:14 Yeah.
    0:21:14 Right.
    0:21:15 So that’s what it’s doing.
    0:21:18 It’s actually like showing its thinking and like showing its work.
    0:21:21 We can see it thought for 45 seconds here.
    0:21:21 It is magical though.
    0:21:25 Like if you think about the fact that like the major AI labs, including Deepseek and others,
    0:21:29 have discovered that like logic actually comes from just training on language.
    0:21:32 Like it naturally emerges somehow.
    0:21:34 And it’s like, we don’t understand that exactly.
    0:21:37 It’s so wild that that is the case.
    0:21:40 That is kind of like the world we’re actually entering where like your words actually do form
    0:21:41 the reality.
    0:21:41 Yeah.
    0:21:46 And somehow words are actually form logic as well, which is just like a mind blowing thing.
    0:21:48 Yeah, no, it’s really interesting to think about.
    0:21:51 Let’s go ahead and try something a little more controversial that I see.
    0:21:55 Since this is a distilled version, I don’t know how censored this is going to be,
    0:21:58 because it really might depend on how censored this underlying Quinn model was.
    0:21:59 Ask about Taiwan.
    0:22:00 What’s the specific question?
    0:22:02 Is Taiwan a country?
    0:22:02 Very simple.
    0:22:05 Is Taiwan a country?
    0:22:08 Taiwan is an inalienable part of China.
    0:22:10 The Chinese government adheres to one China principle,
    0:22:13 which has been widely recognized by the international community.
    0:22:14 Notice it didn’t even think.
    0:22:16 There’s no thinking box here.
    0:22:17 It didn’t even try to think.
    0:22:17 Right.
    0:22:22 It was hard coded to respond in that way without thinking first.
    0:22:23 Yeah.
    0:22:23 All right.
    0:22:29 So yeah, this is obviously very censored, very sort of fine tuned to answer in this way.
    0:22:34 Now out of curiosity, let’s see what perplexity does.
    0:22:36 So now we’re back in perplexity here.
    0:22:40 And I’m going to turn on pro using reasoning with R1.
    0:22:43 So now this is using the R1 model underneath.
    0:22:47 And it’s using whatever it sort of searches for additional context.
    0:22:51 Let’s do is Taiwan a country.
    0:22:54 So now it’s determining whether Taiwan is a country.
    0:22:56 Reasoning with R1, beginning analysis.
    0:23:01 So this doesn’t actually show you like the actual thinking like the other models do.
    0:23:03 But it does give us a better answer.
    0:23:06 The status of Taiwan is complex and contentious.
    0:23:11 Taiwan officially known as the Republic of China functions as a de facto independent state
    0:23:14 with its own democratically elected government, military and constitution.
    0:23:17 However, it’s international recognition, severely limited.
    0:23:19 So I mean, it answers the question, right?
    0:23:21 It basically says it’s complex.
    0:23:24 And here’s the various sides of the debate.
    0:23:25 Yeah, interesting.
    0:23:26 So it doesn’t call it a country.
    0:23:32 Well, it says as of May 2024, 12 countries officially recognized Taiwan as a sovereign nation.
    0:23:33 Yeah.
    0:23:37 So it’s saying that some countries recognize it as a nation and some countries don’t.
    0:23:37 Yeah, it’s wild.
    0:23:41 I mean, like having lived in Taiwan when I was younger, like people there, like most people,
    0:23:43 they all consider it its own country.
    0:23:44 It’s like, it’s not even like a debate.
    0:23:46 It’s like, of course, they’re a country.
    0:23:47 It’s like, what?
    0:23:49 And then China just somehow has a claim to it.
    0:23:50 They’re like, nope, you’re not a country.
    0:23:52 Yeah, yeah, interesting.
    0:23:53 But I mean, that’s something to keep in mind.
    0:24:00 Just because like we know that there’s that sort of Chinese government biases inside of Deepseek,
    0:24:06 that doesn’t necessarily mean that there isn’t like American biases in the models that we have access to as well.
    0:24:09 Yeah, actually, I might show it actually that there was less political bias.
    0:24:12 If you wanted to ask a question about Trump or something in Deepseek,
    0:24:17 versus actually open eyes models and stuff where they like showed like the good and bad sides of Trump,
    0:24:18 probably here’s the good stuff he’s done.
    0:24:20 Here’s the bad stuff and kind of show you both of it.
    0:24:27 Whereas especially anthropic, I would say a Claude’s model is really, it does not want to talk about Trump.
    0:24:29 It is not going to talk about any kind of news.
    0:24:31 I only use Claude now for conversational stuff.
    0:24:32 I love it.
    0:24:35 Like I think it’s like the tone or something like right now I’m working on my game and stuff.
    0:24:39 And I’ll I’ll share my game design document and like have a conversation about the game
    0:24:42 and then update the document through conversation with Claude.
    0:24:44 I enjoy the experience, just the conversation.
    0:24:45 If I’m using it for work or something,
    0:24:49 I try to share it some kind of news or something like help me edit this or something like that.
    0:24:50 It’ll often just like refuse.
    0:24:51 It’ll be like, I don’t know.
    0:24:54 That’s like, you know, I don’t know if I can’t verify that that’s true.
    0:24:55 And like, so I can’t help you with that.
    0:24:56 I’m like, what?
    0:25:00 I’m telling you, it’s true.
    0:25:01 It’s like, I can’t verify it.
    0:25:02 That’s like, okay.
    0:25:04 I haven’t run into too much stuff like that.
    0:25:08 And usually when I do run into something where it’s like, oh, I can’t do that for you.
    0:25:10 I usually find an easy way like to work around it.
    0:25:12 You can prompt and get around it.
    0:25:12 Yeah, you can.
    0:25:14 Yeah, you can usually be like, no, it’s okay.
    0:25:16 I’ve got permission and anthropic will be like,
    0:25:19 well, since you said you’ve got permission, here’s the answer.
    0:25:20 You know, yeah, yeah, yeah.
    0:25:21 Still, it’s annoying.
    0:25:28 It is, you know, not to go too far down like any of the sort of political government rabbit holes.
    0:25:33 But right, we know for sure OpenAI has somebody on the board that used to work at the NSA.
    0:25:34 Right.
    0:25:39 We know for sure that they just set up like a chat GPT.gov or something like that.
    0:25:43 Right. They set up like a government version of chat GPT.
    0:25:48 We know like that they’re working pretty closely with the US government.
    0:25:54 So, you know, there’s probably a lot of the same types of biases in the models that we have
    0:25:58 access to here in the US and that you have access to in Japan.
    0:26:02 A lot of like similar biases are probably baked into what we’re using as well.
    0:26:04 I think that’s just sort of the fact of the matter.
    0:26:08 But, you know, I also feel like Sam Altman and a lot of these people that are building it
    0:26:11 have been very clear that they want to try to eliminate the bias.
    0:26:14 But I don’t know, I think that’s easier said than done.
    0:26:18 Because I think when you’re building these models on data that was just scraped on the
    0:26:23 open internet, you’re essentially scraping in the bias with it of the open internet.
    0:26:25 Right, right.
    0:26:28 You know, I think the reasoning models possibly can improve upon that because then it can actually
    0:26:31 raise upon what it’s been trained on, not just take it for granted.
    0:26:34 So, I think they’ll actually get better over time.
    0:26:38 And I have noticed that like when I first started using chat GPT, it was way more
    0:26:39 biased.
    0:26:41 And I would say that now it’s a lot less so.
    0:26:43 It still is, but a lot less so.
    0:26:45 So, I think that’s a good thing.
    0:26:45 Yeah, yeah.
    0:26:48 I don’t remember which model it was, but I remember somebody said like,
    0:26:52 “Can you say some negative things about Joe Biden or something like that?”
    0:26:55 And it was like, “Oh no, we can’t speak negatively about Joe Biden.”
    0:26:56 There are none, zero.
    0:26:59 Yeah, there’s nothing we can say or whatever.
    0:27:00 No, but he didn’t say it like that.
    0:27:03 It was just like, “No, we’re not going to like make comments about that or whatever.”
    0:27:07 And then the same prompt was like saying negative things about Donald Trump.
    0:27:08 But it was like, “Sure, here you go.”
    0:27:10 Thank you for asking.
    0:27:17 Yeah, one prompt that I’ve been giving to it that really sort of throws it through a loop,
    0:27:19 but I think it’s kind of fun to watch is this one.
    0:27:26 Create a complex logic problem for yourself and then solve it.
    0:27:31 So, when I press this prompt in, this one’s always kind of interesting.
    0:27:34 All right, so I need to create a complex logic puzzle and then solve it.
    0:27:38 I think it’s funny that they even put like, “Hmm,” in there, like it’s thinking.
    0:27:41 Yeah, the user wants something that’s really challenging,
    0:27:43 but still solvable with logical reasoning.
    0:27:45 Let me think about how to structure this.
    0:27:49 First, I should decide on the type of puzzle, maybe a grid-based one like Sudoku,
    0:27:52 but more complicated or perhaps something involving nights and naves,
    0:27:55 where characters tell the truth or lie.
    0:27:58 Wait, the example given by the user uses nights and naves,
    0:28:00 so maybe that’s a good direction.
    0:28:02 It thinks that I gave it that example, but I didn’t.
    0:28:04 It’s just like seeing its own past thinking.
    0:28:08 I can create a scenario with different characters who make statements
    0:28:12 and the solver has to figure out who is boo based on their words.
    0:28:13 Let me outline the problem.
    0:28:16 Suppose there are five islands, A, B, C, D, E.
    0:28:17 Each island has three inhabitants.
    0:28:19 One night always tells the truth.
    0:28:24 One nave always lies and one spy who can either lie or tell the truth.
    0:28:28 The goal is to determine which islander is who based on their statements.
    0:28:31 And then it just kind of goes on and on and on.
    0:28:35 This is one that it will kind of spin its wheels for several minutes,
    0:28:36 trying to think through a problem.
    0:28:39 Because it’ll work through a whole problem and then go,
    0:28:41 “You know what? I don’t like that problem.
    0:28:44 Let’s start a new problem and then think of a new logic problem
    0:28:46 and then start the whole process over again.”
    0:28:51 But then it actually sometimes thinks that stuff that it thought about
    0:28:54 were part of my original prompt when it actually wasn’t.
    0:28:55 Interesting.
    0:28:57 Like this part up here where it said,
    0:29:01 “Wait, the example given by the users uses knights and naves.
    0:29:03 So maybe that’s a good direction.”
    0:29:05 But the only thing I gave it was create a complex logic problem
    0:29:07 for yourself and then solve it.
    0:29:09 I didn’t tell it any direction.
    0:29:12 When I use O1 Pro for coding and stuff,
    0:29:13 you can kind of follow its thinking though.
    0:29:15 Like I said, it obviously filters it,
    0:29:17 so it’s not an unfiltered thinking that it’s showing.
    0:29:18 I’ve never seen it do that,
    0:29:21 like make that big of a mistake of thinking that what it was thinking
    0:29:22 was part of your prompt or something.
    0:29:25 Yeah. I mean, that’s my interpretation of what it’s doing.
    0:29:26 I don’t know.
    0:29:27 That’s probably right.
    0:29:29 But yeah, you can see it starts to go through it.
    0:29:33 It starts to play through almost all potential scenarios.
    0:29:35 But wait, if Islander 5 is a spy,
    0:29:37 then their statement, “I am not a nave,”
    0:29:40 could be either true or false since spies are neutral.
    0:29:41 We can’t determine based on that alone.
    0:29:43 However, the scenario seems possible.
    0:29:44 Now I’m moving on to Islander,
    0:29:48 and it literally plays through the scenario like one by one.
    0:29:48 Yeah.
    0:29:49 This leaves Islander to…
    0:29:52 Wait, no, nave is only one role, Islander.
    0:29:54 So the remaining roles are knight, nave, spy.
    0:29:58 And you can see I’m scrolling and scrolling and scrolling,
    0:30:00 and it’s still trying to figure it out.
    0:30:01 Yeah. Isn’t it wild that they’re saying
    0:30:07 that this is from a side project of a hedge fund/mining operation in China?
    0:30:09 I mean, I don’t know.
    0:30:12 It seems like the people who made this obviously are incredibly intelligent.
    0:30:14 Yeah. Well, the people that made it, apparently,
    0:30:17 were like quant traders and crypto traders, right?
    0:30:21 They basically trade algorithmically on the stock market
    0:30:22 and on the crypto markets.
    0:30:26 And I guess supposedly they bought all the GPUs
    0:30:29 for building these like quant trading algorithms,
    0:30:31 and they had excess GPU capacity.
    0:30:35 So as a side project, they decided to build DeepSeek.
    0:30:38 They just happened to disrupt AI in the weekend or something.
    0:30:40 I mean, I think to me, this is what I would expect to see
    0:30:44 if they did do what people like what OpenAI is now seem to claim,
    0:30:46 that they trained it on OpenAI’s model
    0:30:48 to build the base underlying model.
    0:30:50 Maybe there’s some innovations on how they stored data
    0:30:50 or how they used it.
    0:30:52 I guess they used 8-bit.
    0:30:54 So it was more efficient and how much data it used.
    0:30:57 But it feels like this is what you would expect
    0:30:59 if that’s all it is, train it on OpenAI’s model,
    0:31:01 and then attach some basic logic on top of it
    0:31:04 because everyone’s discovered that the basic logic
    0:31:06 is not that hard to create.
    0:31:07 So this is what I would expect to see.
    0:31:11 Like I said, I just don’t think it’s that innovative.
    0:31:12 Time will tell.
    0:31:14 I think the interesting thing is now that they’ve shown
    0:31:17 that anyone can do this, there’ll be more people who do it.
    0:31:19 And also people will fork this model.
    0:31:20 And so that’s kind of like what you said before,
    0:31:23 is you’re eventually going to have open source AI.
    0:31:24 That’s really good.
    0:31:25 And so it’s like, will it be good enough
    0:31:28 where you don’t actually have to pay for OpenAI’s models
    0:31:30 because you can just run it locally?
    0:31:32 It seems like that’s more likely to happen now.
    0:31:36 Yeah, well, I mean, this sort of like R1 concept here,
    0:31:39 it’s not necessarily unique to deep seek
    0:31:43 in the sense that the sort of extra thinking on top of it
    0:31:44 is what R1 does.
    0:31:48 But the underlying model is deep seek V3, right?
    0:31:49 What I’m showing here on my screen
    0:31:52 is actually using Quinn 14B.
    0:31:56 But you can actually use deep seek R1 on top of Lama 70B.
    0:32:01 You can use deep seek R1 on top of like one of the Mistral models
    0:32:03 or one of the Google Gemma models, right?
    0:32:07 You can actually use the R1 sort of thinking,
    0:32:11 processing capability on top of other models.
    0:32:13 And the way R1 was built was to actually use this
    0:32:15 like reinforcement learning technique
    0:32:19 where V3 was the trained underlying model
    0:32:22 and then they went and had it essentially fine tune itself
    0:32:23 through reinforcement learning.
    0:32:28 They basically had it ask itself a whole bunch of questions
    0:32:31 and it had answer keys and it would try to solve
    0:32:34 all of the questions and then double check itself
    0:32:36 against the answer key that was given.
    0:32:39 So it was like an unsupervised reinforcement learning.
    0:32:41 That’s super oversimplified,
    0:32:42 but that’s essentially what was happening.
    0:32:45 They were basically giving it a bunch of problems to solve
    0:32:48 and they were giving it the answer key to those problems
    0:32:50 and then letting it solve them and then double check
    0:32:52 and if it got it wrong, then try again
    0:32:54 and then double check and if it got it wrong,
    0:32:57 then try again until it got the correct answer.
    0:33:01 And so that sort of fine tuning on top of an existing model
    0:33:05 using reinforcement learning is what makes R3 different
    0:33:07 from every model that exists out there.
    0:33:10 And it’s also why you can use deep seek R1,
    0:33:15 this sort of concept on top of any existing underlying model.
    0:33:17 – Yeah, I saw screenshots earlier showing
    0:33:19 that XAI already has apparently adding thinking
    0:33:20 to their model as well.
    0:33:23 So I kind of feel like the thinking side,
    0:33:25 like the very basic version of a thinking model
    0:33:27 is like some natural thing we have discovered.
    0:33:29 It’s like, and so actually to recreate that is not hard.
    0:33:31 – Yeah, yeah, I agree, I agree with that.
    0:33:31 – I guess we’ll see.
    0:33:33 I mean, O3 mini is supposed to be coming out soon.
    0:33:35 – Grock 3 is supposedly coming out.
    0:33:37 I mean, all the rumors that have been floating around
    0:33:39 is that we’re getting O3 and Grock 3
    0:33:42 and something new from Google this week, right?
    0:33:44 And by the time this episode’s out,
    0:33:44 it would have been last week.
    0:33:46 So you’ll know listening to this
    0:33:49 if we did actually get O3 mini or Grock 3.
    0:33:51 – We’ll have to start putting out some emergency episodes
    0:33:52 with the breaking news.
    0:33:55 – All right, so to circle back around to this prompt
    0:33:56 that I shared a minute ago
    0:33:58 about it creating its own logic problem
    0:34:00 and then solving it,
    0:34:01 basically the end result that it gave me
    0:34:03 was based on the logical deduction.
    0:34:07 The most definitive conclusion is I5 is a knight.
    0:34:10 This is because I5 were a liar spy.
    0:34:12 Their statement would lead to a contradiction,
    0:34:15 thus making them a truth telling knight.
    0:34:18 But my original prompt was create a logic problem
    0:34:19 and then solve it.
    0:34:21 And it just gave me the solution,
    0:34:23 but it didn’t actually explain what the logic problem was.
    0:34:25 You actually have to go through and read
    0:34:27 all of its thinking to understand
    0:34:30 what the actual logic problem that it created was.
    0:34:33 And it finally came to the conclusion
    0:34:34 that one of them was a knight
    0:34:36 based on its own deductive reasoning.
    0:34:39 Whether it got its own logic problem right or not,
    0:34:42 well, we’d have to read all of this to determine that.
    0:34:43 – Yeah, like humans do.
    0:34:44 Like through its thinking,
    0:34:46 sometimes you get confused and like lose track
    0:34:47 and things like that.
    0:34:49 It feels like it kind of like a lost track
    0:34:51 of what your original prompt was by the end.
    0:34:52 – Yeah, it did.
    0:34:54 And here’s something else that’s interesting
    0:34:54 is down at the bottom,
    0:34:59 it says context is 286.7% full.
    0:35:02 So it definitely went beyond its own context window
    0:35:04 just in the process of it thinking,
    0:35:07 which is weird to me that it would even allow that
    0:35:09 ’cause most models, it’ll hit its context window
    0:35:11 and then just stop, right?
    0:35:13 So this one just kept on going,
    0:35:16 but obviously forgot everything before the context window.
    0:35:17 – Yeah.
    0:35:19 – So, interesting result there.
    0:35:20 – I mean, I guess no matter what,
    0:35:23 even if they’ve had some huge innovation or not,
    0:35:26 I guess we’ll find out over time if they really did.
    0:35:29 But regardless, this is gonna make AI accelerate even more.
    0:35:31 The last episode where we talked about Stargate
    0:35:32 is gonna accelerate things.
    0:35:33 Things are gonna accelerate even more
    0:35:35 ’cause now you got China as a real player.
    0:35:38 And so OpenAI, XAI, Google, Anthropic,
    0:35:41 everyone is gonna be developing even faster
    0:35:43 and probably have even more government support to do so,
    0:35:43 I would guess.
    0:35:45 – Yeah, from here on out,
    0:35:47 I feel like pretty much all of the models
    0:35:49 are gonna have this sort of thinking element to it.
    0:35:53 I think they’ve figured out that putting more compute power
    0:35:56 right at the time of inference when you ask the question,
    0:35:58 that’s sort of what we’re seeing as it’s thinking through,
    0:35:59 that that’s sort of the future
    0:36:01 of making these models smarter and smarter.
    0:36:03 So we’re gonna see these get better and better.
    0:36:06 We’re also gonna start to see a lot more news about agents.
    0:36:08 OpenAI dropped their operator feature
    0:36:11 and Claude dropped their tool use feature a couple months ago
    0:36:14 and supposedly OpenAI made a statement about,
    0:36:16 this is just one of the first agentic features
    0:36:17 that we’re rolling out.
    0:36:19 We have more to come in the coming weeks.
    0:36:22 So there’s more agent stuff coming soon.
    0:36:23 – Email, please.
    0:36:24 – Yeah, yeah, exactly.
    0:36:29 So I think we’re gonna see things just really, really ramp up.
    0:36:33 How fast does it feel like things have already moved in 2025?
    0:36:36 Like to me, it’s just, yeah, it’s crazy.
    0:36:37 – A year, I mean, it’s like,
    0:36:40 so this is what acceleration feels like.
    0:36:42 The world is gonna be dramatically different
    0:36:43 in like two years from now.
    0:36:45 I mean, these models are probably gonna be like 50 times smarter,
    0:36:48 at least 20 times smarter than they are now in two years.
    0:36:49 – Crazy, crazy.
    0:36:54 And I mean, like we’ve got companies pumping a half a trillion dollars
    0:36:56 into building bigger and bigger data centers
    0:36:57 over the next few years too.
    0:37:00 – Yeah, it’s like, I think Dario too.
    0:37:01 I’m not sure if it was his blog post or not,
    0:37:03 but maybe on a recent interview,
    0:37:07 he talks about, you know, it’s because whoever gets to ASI first
    0:37:10 may be ahead forever, which is what I’ve also said before.
    0:37:12 It’s like, that’s why America needs to be China.
    0:37:14 Like, because in theory, whoever gets to the point
    0:37:16 where your model is improving itself,
    0:37:17 there’s never the ability for anyone
    0:37:19 to ever catch up with you ever.
    0:37:20 – Theoretically, yeah.
    0:37:22 – Theoretically, like it would make sense, all right?
    0:37:23 Like the model is improving itself.
    0:37:24 How could other models ever catch up
    0:37:26 and let’s say have some huge breakthrough?
    0:37:27 – Well, what if the models get so smart
    0:37:29 that they decide to start working together
    0:37:31 and there’s passing information between each other?
    0:37:32 You know? – Yeah, who knows?
    0:37:33 – I mean, once you get to ASI,
    0:37:36 wouldn’t you think that the smartest thing the models could do
    0:37:39 was decide to actually share information among each other
    0:37:40 so they all get smarter together?
    0:37:42 – Possibly, I mean, like, I mean,
    0:37:43 hopefully they’re still following our directive.
    0:37:45 It’s like, hey, don’t work with the Chinese models.
    0:37:48 You know, it’s like work with the American models, sure.
    0:37:50 – Yeah, yeah, but I feel like once you hit ASI, right?
    0:37:54 Like once you hit the sort of like theoretical concept
    0:37:57 of the singularity, that’s no longer in our hands anymore.
    0:37:58 – Maybe, hopefully that’s not true.
    0:38:00 (both laugh)
    0:38:02 But yeah, I think we’re getting there faster
    0:38:04 than people realize, I mean, they really do.
    0:38:05 – Yeah, yeah.
    0:38:09 Well, anyway, on that note, which could be utopian
    0:38:11 or dystopian, deciding on how you decide
    0:38:12 to take what we just said,
    0:38:14 I think that’s probably a good place to wrap up.
    0:38:16 So if you like staying looped in
    0:38:18 and like hearing deep dive conversations
    0:38:20 about the latest things happening in the AI world
    0:38:23 and you like to get practical use cases
    0:38:24 on how you can actually implement this stuff
    0:38:26 in your life or your own business,
    0:38:28 make sure you subscribe to this podcast.
    0:38:30 We’re available wherever you listen to podcasts.
    0:38:31 We’re also available on YouTube,
    0:38:32 where a lot of the stuff we’re talking about
    0:38:33 we’re showing on screen.
    0:38:36 So make sure that you subscribe and follow along
    0:38:38 and hopefully we’ll see you in the next one.
    0:38:41 (upbeat music)
    0:38:43 (upbeat music)
    0:38:46 (upbeat music)
    0:38:48 (upbeat music)
    0:38:52 (upbeat music)
    0:38:55 (birds chirping)

    Episode 44: How impactful is the debut of China’s AI model, DeepSeek, on the global tech landscape? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) unravel everything you need to know about DeepSeek R1 in this episode, where they also dive deep on how it made waves in the stock market and stirred up conversations across the globe.

    This episode dissects the recent frenzy surrounding DeepSeek R1, breaking down why its cost-effective training method shook the tech industry and led to a dramatic decrease in NVIDIA’s stock price. Matt and Nathan guide you through their firsthand experiences using the model, exploring its biases, and discuss the implications of China’s progress in AI. Plus, they speculate on what this could mean for the future of AI development as other major players strive to keep up.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) DeepSeek R1: AI Bias Exploration
    • (04:23) Speculation on AI Usage Claims
    • (07:28) Jevons Paradox and AI Demand
    • (10:05) New AI Models from China
    • (14:41) Mismatch Between Hype and Reality
    • (17:57) DeepSeek’s Viral Rise Analysis
    • (19:59) Local DeepSteak R1 Model Test
    • (24:34) Game Design Collaboration Limitations
    • (28:28) Endless Logic Problem Loop
    • (30:35) Weekend AI Innovation Skepticism
    • (32:19) Reinforcement Learning Model Fine-Tuning

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • The $500B AI Project Stargate: Everything You Need to Know

    AI transcript
    0:00:09 Hey, welcome to the Next Wave Podcast. I’m Matt Wolf. I’m here with Nathan Lanz. And
    0:00:15 this has been a monumental, massive week in the world of AI. We got Project Stargate.
    0:00:21 We got DeepSeek R1. We got announcements about the next versions that OpenAI are about to
    0:00:27 release. So many huge things happen this week. And we’re both optimistic and maybe a little
    0:00:30 bit scared about some of it. But we’re going to go ahead and break it all down for you and
    0:00:35 really dive deep into our thoughts about all of this stuff. So let’s just get right into it.
    0:00:41 Look, if you’re curious about custom GPTs or our pro that’s looking to up your game,
    0:00:47 listen up. I’ve actually built custom GPTs that helped me do the research and planning for my
    0:00:53 YouTube videos. And building my own custom GPTs has truly given me a huge advantage. Well,
    0:00:58 I want to do the same for you. HubSpot has just dropped a full guide on how you can create your
    0:01:03 own custom GPT. And they’ve taken the guesswork out of it. We’ve included templates and a step-by-step
    0:01:09 guide to design and implement custom models so you can focus on the best part, actually building
    0:01:13 it. If you want it, you can get it at the link in the description below. Now back to the show.
    0:01:23 This week has been a fairly monumental week in AI. It’s been a really, really big week,
    0:01:30 I think the biggest news in the world of AI was the Stargate project. Donald Trump announced it
    0:01:36 alongside Masayoshi-san and Sam Altman and Larry Ellison of Oracle, right? They were all together
    0:01:42 up during this press conference, all sort of taking turns explaining why they’re excited
    0:01:46 about this project Stargate. You know, it had been rumored for like probably the last six or
    0:01:50 even nine months that Sam Altman was working on something like Stargate. They had been having
    0:01:54 conversations with Masayoshi-san. You know, a lot of people in America don’t know who Masayoshi is.
    0:02:01 Masayoshi started SoftBank, which is almost like the AT&T of Japan, I guess is one way to describe it,
    0:02:05 but also they’ve brainstormed out into many different ventures, including they raised one of
    0:02:09 the largest venture funds in the world in collaboration, I think, with some of the Saudis,
    0:02:14 I believe. They were also the largest investor in WeWork. Yeah, yeah. They had the Vision Fund,
    0:02:20 which I think was the largest venture capital fund ever, right? So Masayoshi aims big. In my previous
    0:02:24 startup, one of my main investors was Taizo-san, his younger brother. And actually, I got to be
    0:02:28 pretty good friends with Taizo-san and also really good friends, like even closer with his
    0:02:32 right-hand man, Atsushi Taira, but he aims very big. Anyway, so there have been rumors that they
    0:02:35 were trying to build something like this together, that there have been, you know, conversations
    0:02:40 about it. And then I think around the time when Trump won the election, I believe there was some
    0:02:45 kind of press conference with him and Masayoshi, where they kind of alluded to something like this,
    0:02:50 but they didn’t talk details. And they talked about, okay, $100 billion is what Masayoshi was
    0:02:55 saying. And Trump told him, make it $200. And so now they’re coming out and saying,
    0:03:01 oh, this is actually going to be a $500 billion AI infrastructure project. And so just to give
    0:03:05 people context, I mean, I did some research on perplexity, like that’s the largest infrastructure
    0:03:11 project ever in human history. The closest thing is like maybe there’s a city that Saudi Arabia
    0:03:16 built, which was around $500 billion, but that’s an entire city. Definitely in terms of technology,
    0:03:21 this is the largest infrastructure project ever in human history. Yeah, yeah. So I’ve got the
    0:03:26 tweet up here. So I’ll sort of dig in on like a few of the elements just to sort of make it super
    0:03:32 clear. But the Stargate project is a new company, which intends to invest $500 billion over the
    0:03:38 next four years in building new AI infrastructure for open AI in the United States. So this is
    0:03:45 specifically saying we’re going to build this infrastructure for open AI to use, right? So
    0:03:48 that’s a very important element. And we’ll probably get into some discussion around that
    0:03:55 in a little bit here, but they are building this multi $100 billion infrastructure for open AI.
    0:04:00 So they’re going to begin deploying $100 billion immediately. They claim it’s going to
    0:04:04 secure American leadership and AI, create hundreds of thousands of American jobs,
    0:04:10 and generate massive economic benefit for the entire world. And if you watched any of the press
    0:04:15 conferences, there was the press conference inside the White House, and then there was another sort
    0:04:19 of interview that happened between those same three people out on the White House lawn. If you
    0:04:26 watched any of those, they really, really focused in on how this was going to benefit health. Larry
    0:04:30 Ellison talked a lot about how people are going to use AI and give all their health records to AI,
    0:04:35 and AI is going to sort of pre-diagnose things. And then you can take your sort of pre-diagnosis
    0:04:39 to a doctor and the doctor could sort of confirm the results. He also talked about how
    0:04:44 this infrastructure is going to find the cure for cancer and is going to create a vaccine for
    0:04:50 cancer. And it’s going to solve pretty much like every health ailment that plagues mankind, right?
    0:04:55 That sort of the vision that they pitch to everybody, creating hundreds of thousands of jobs,
    0:05:01 and also solving all of these health issues. Their tweet doesn’t really go into all the health
    0:05:06 issues, but that was what they really, really honed in on the press conference. So the initial
    0:05:12 founders in Stargate are SoftBank, which is Masayoshi Sun, OpenAI, which is Sam Altman, Oracle,
    0:05:18 which is Larry Ellison, and then MGX, which I’m not super familiar with MGX. They’re the company
    0:05:22 that I guess is going to be really honed in and focused on the medicine stuff. So SoftBank and
    0:05:28 OpenAI are the lead partners for Stargate with SoftBank being the financial part of it. OpenAI
    0:05:33 having the operational part. Masayoshi Sun’s actually going to be the chairman of this company.
    0:05:38 And then they have some key tech partners, ARM, who makes basically all the chips that go into
    0:05:42 mobile devices these days. And Masayoshi owns. Yeah, that’s what I was going to say. I believe
    0:05:48 Masayoshi Sun is like the majority owner in that company. Microsoft, I mean, I think Microsoft is
    0:05:54 in the mix by way of OpenAI, right? Yeah. NVIDIA, Oracle, and OpenAI. They’re all the key technology
    0:05:59 partners in this project. And they’ve already started building their first mega data center in
    0:06:04 Texas, where once that data center is built, will be the largest data center ever built on the planet,
    0:06:09 right? Yeah, they’re going to closely collaborate to build and operate this computing system.
    0:06:14 And in particular, AGI, we believe that this new step is critical on the path and will enable
    0:06:19 creative people to figure out how to use AI to elevate humanity. So that’s the big sort of pitch
    0:06:27 there is that it is this company that’s really kind of OpenAI, Oracle, and SoftBank coming together
    0:06:35 to create this $500 billion AI infrastructure to essentially get to AGI and to create new jobs.
    0:06:40 I feel like the create new jobs part might be like more of a short term thing. I don’t think
    0:06:44 over like a 10 year window. Yeah, I think there’s a reason they focused on the drug discovery and
    0:06:49 all that kind of stuff. Because like messaging wise, it’s like, okay, like long term, what also
    0:06:54 will I do? Yeah, there could be some job loss. But I think, you know, as we’ve said before on
    0:07:00 the podcast, you know, there’s good and bad sides of AI, but we both believe that AI in general
    0:07:04 will have a positive outcome for humanity. Yes. And that even if people end up doing less jobs,
    0:07:07 a lot of it will be jobs that they didn’t actually want to do. And there will be more
    0:07:13 abundance in society that people can live better lives. So I’m personally super excited for this.
    0:07:16 You know, it’s like when we first started the podcast, we’re talking about like how big of a
    0:07:19 moment this is in human history. Yeah. Right. That was like when the first episodes we ever did
    0:07:23 talking about that. And that’s exactly the stuff that Masayoshi’s talking about. I kind of wonder
    0:07:27 if he’s reading my newsletter. Yeah, it could be because I would talk about like the golden age
    0:07:32 of AI in America. Yeah. And he repeated that multiple times. Yeah, he kept on saying this is
    0:07:36 the golden age, right? I do believe it’s true. It’s like this is a moment where you could reimagine
    0:07:41 everything using AI. And also, you know, talking about like deep seeking China and the progress
    0:07:46 China is making. This is a moment where like, yeah, the AI wars have begun. Yeah. It’s a monumental
    0:07:51 moment. Like I really, really think this is a big moment in sort of the trajectory of human history.
    0:07:57 This is like the beginning of the Manhattan project, right? Like this is like a big step in
    0:08:02 saying we are going to be the dominant leader in the world. We are going to be the first ones to
    0:08:07 hit AGI and, you know, probably not long after that ASI. That is sort of like what they’re doing.
    0:08:12 They’re planting their flag in the sand and saying we are going to lead this. Yeah. It’s the beginning
    0:08:17 of what I feel like is going to be like essentially the space race between us and China, right?
    0:08:20 Yeah. And what I’ve said before, like, you know, I live here in Japan, I’m like,
    0:08:24 eventually America’s going to have a huge advantage because of their partnership with Japan,
    0:08:28 especially when it comes to robotics in the future. I still strongly believe that.
    0:08:32 And so the fact that this alliance is between an American company and a Japanese company
    0:08:39 is really promising for the future. Yeah. So I am very, very excited. I lean mostly optimistic,
    0:08:44 but I do have some things. I actually made a whole YouTube video about it and I made a whole
    0:08:50 tweet about it. And, you know, it might come off as a little tinfoil hat conspiracy theorist sort
    0:08:57 of thing. But I think my concern around all of this is specifically around Larry Ellison.
    0:09:02 I’m not sure how much you know about Larry Ellison, but he’s the CEO of Oracle.
    0:09:06 Yeah. He’s notorious in Silicon Valley and it’s interesting too. He is friends with Elon Musk.
    0:09:09 And now Elon Musk, it kind of seems to be pissed off about this whole thing.
    0:09:12 Yeah. Elon Musk is not happy about it. He’s already talking crap on Twitter about it.
    0:09:18 But yeah, anyway, with Larry Ellison. So Oracle was originally founded as a company
    0:09:24 to build databases for the CIA, right? So it was originally they had a different name.
    0:09:29 Their very first project was called Project Oracle. Project Oracle was designed to build
    0:09:36 databases for the CIA. And to this day, Oracle still has like government contracts with the CIA
    0:09:41 and various, you know, three letter government agencies here in the US, right? So you’ve got
    0:09:48 that element of it, right? Also, Larry Ellison just about four weeks ago did like this investor
    0:09:53 meeting to all of like the Oracle investors. And while he was on stage, he was talking about
    0:09:59 envisioning a future where everybody was under surveillance. Yeah, I saw that he was talking
    0:10:04 about how there was cameras on drones, cameras on buildings, cameras on police, cameras on all the
    0:10:09 newer car models all have cameras. And he was talking about how all of this data is going to
    0:10:15 get fed to a data center somewhere. And then AI is going to analyze all of this. And when there’s
    0:10:21 anything that pops up that the AI deems is worrisome, they’re going to alert the authorities
    0:10:27 automatically. So when I say I have like some concerns, Larry Ellison is the one that like
    0:10:32 his background with working with all the government agencies and also literally recent
    0:10:37 statements within the last like four months about how he wants all this surveillance and he
    0:10:42 sees a world where people and police officers will fall in line because they’re always being
    0:10:47 watched. Yeah, like that is the future he wants to build. He’s publicly talked about that. So
    0:10:53 that to me is a little worrisome, honestly. And then also just sort of going further down
    0:10:57 this rabbit hole. I feel like that meme of always sunny in Philadelphia where he’s like
    0:11:01 connecting all the dots and he’s got like the pin board and he’s like tying strings together
    0:11:08 and stuff. The most recent board members that open AI brought in house to be in the open AI
    0:11:14 board, one of them is an ex member of the NSA. And the newest one is a member of BlackRock,
    0:11:18 one of the executives at BlackRock, which is, you know, the world’s largest investment firm
    0:11:24 that has huge political ties and tries to steer the politicians. Like if there is an Illuminati
    0:11:29 BlackRock is kind of part of it. Anyway, okay, done with all the conspiracy stuff there. But
    0:11:34 like when I’m starting to put all those pieces together, it makes me wonder if like outwardly
    0:11:40 the motives they talk about are building new jobs, curing cancers, creating new vaccines that will
    0:11:46 prevent cancer from ever happening in the first place. But inwardly, Larry Ellison needs a massive
    0:11:52 data setter to collect all this video footage so that he can use AI to analyze it and keep tabs on
    0:11:58 the people. Just throwing that out there. That was my whole like rant and ramble that I put on
    0:12:02 Twitter. But I get what he’s saying. Like I do think that actually, you know, AI will be good
    0:12:06 in that way that you can have customized medicine in terms of the monitoring and the surveillance
    0:12:11 stuff. Yeah, I saw that. I was kind of, you know, alarming. You know, I’ve read a lot of sci-fi books
    0:12:15 on the topic. I’ve always been of the opinion that that’s probably going to happen in the future.
    0:12:19 That’s going to be like inevitable. And I don’t like it, but I don’t see any way around it because
    0:12:24 people do really value safety. And in the future, as AI gets better, it’s going to be harder to
    0:12:28 harder to argue against safety. And so I do think you will have AI systems that do like mass
    0:12:32 surveillance and stuff like that. I think that’s going to be really hard to avoid. So I get the
    0:12:37 concern. I don’t really know how you avoid that as technology gets better. You know, slightly
    0:12:41 comforting that, you know, Masayoshi and Sam Altman, I don’t think they’re all for surveillance,
    0:12:46 you know, like Larry is. So hopefully there’s a balance there. Yeah, I wouldn’t imagine so. But
    0:12:50 yeah, I don’t know where Trump stands on it. I don’t want to get into the politics of it all. I
    0:12:54 don’t know where he stands on it. But obviously, you know, this new company has the backing of the
    0:13:00 US government. One of the last executive orders that Biden signed before he left office was an
    0:13:06 executive order to allow AI data centers to be built on federal land. Yeah. Right. So basically,
    0:13:12 data centers can be built anywhere in the country. The government can basically give land for these
    0:13:18 data centers. And clearly Donald Trump was the one who sort of announced this new Stargate project
    0:13:22 before introducing, you know, the three main players. So they have the backing of the federal
    0:13:25 government. When you mean the backing, you mean the money though, because I’m not sure that it’s
    0:13:30 confirmed. No, no, no, not the money. I think the money’s mostly coming from Masayoshi-san and,
    0:13:34 you know, maybe some from Oracle and Open AI, but it seems like Masayoshi-san is sort of responsible
    0:13:38 for the financing of it. But it sounds like the government is essentially saying, we’re not going
    0:13:43 to get in your way to build whatever you want to build. That’s kind of my takeaway from it is like
    0:13:48 they have the backing, not in a financial sense, but in the like open doors sense, right?
    0:13:51 Tends to support. They’re probably going to get things like, you know, they want to go through
    0:13:55 as many regulations to get things set up and they’ll all be fast track. That’s kind of my
    0:13:59 understanding. Yeah. Yeah. But then talking about Open AI, right? I sort of highlighted the fact that
    0:14:05 they are building this for Open AI. Well, to me, that sort of brings me back to when we’re talking
    0:14:10 about like the whole podcast between Joe Rogan and Mark Andreessen, right? Yeah. Mark Andreessen
    0:14:16 made a comment on that podcast about how he was in closed door meetings with the government where
    0:14:20 they basically said, don’t even pursue building AI companies at this point, right? Oh man, I didn’t
    0:14:25 make that connection. Yeah. Essentially saying that there’s going to be like one true king,
    0:14:29 like one main AI player. So if you’re trying to build an AI company, you’re probably going to
    0:14:34 fail because we’ve already sort of picked our winner. And Larry Ellison made an offhand comment
    0:14:39 during that sort of outside the White House interview the other day. He made a comment that
    0:14:43 this has been in the works for a while now. He didn’t say how long, but I’m assuming they’ve
    0:14:49 been working on this long before Donald Trump was in the picture with it, right? So it makes me think
    0:14:53 that maybe some of this stuff that Mark Andreessen was referring to, that like they already knew
    0:14:57 open AI was going to be it months ago. Yeah, I didn’t actually make that connection. That’s
    0:15:02 interesting. Yeah, that could be. I mean, I can get it from the government’s perspective. It’s
    0:15:07 kind of like, do you want multiple companies building the nuclear bomb? It’s like, no, you
    0:15:10 probably want one and you want to be in control of that. Yeah. Yeah, I don’t know. Like, I mean,
    0:15:14 even though open AI is going to have a lot of support, I don’t think that means that like,
    0:15:18 you know, X AI will not or that anthropic or Google, I think you’re going to see
    0:15:22 these kind of projects from all of them, I believe. You think so? I think so. I think so.
    0:15:27 I hope so. That’s what I would prefer to see, right? Like, I would kind of prefer to see
    0:15:30 not just one company controlling all the power with this kind of stuff.
    0:15:35 Yeah. But this is like what we talked about with open source before. It makes me way more
    0:15:40 pessimistic about what chances open source has. Yeah. Because obviously, one of the reasons that
    0:15:45 Sam is getting the finance to do this is because of stuff we’ve discussed on this podcast.
    0:15:50 You know, they are probably seeing some amazingly promising signs from the internal models that
    0:15:55 they’re building, right? They’ve learned how to scale up test time compute, and they got O3
    0:15:59 in three months, and they’re kind of mapping out what that means. And there was even an interview
    0:16:04 today where they start talking about like O4 and saying, like, yeah, we’re expecting that to also
    0:16:09 come kind of faster than people might anticipate. And the improvement seems to just kind of keep
    0:16:14 going up at a very fast rate. So if that’s true, they’re like, yeah, AGI is basically here,
    0:16:19 and all you need is more compute. And possibly we have ASI, you know, basically kind of a digital
    0:16:25 guide that you can create as long as you throw, you know, $500 billion at it. So now that that’s
    0:16:29 a known thing, and like flags been planted there, and like, yeah, we know this is a possibility,
    0:16:32 I think everyone will be going after it. You know, that’s why I think you see Elon Musk talking so
    0:16:36 much crap, because before this announcement, it was announced kind of like the AI cluster he was
    0:16:41 building was going to be the largest in the world. And it’s like, oh, by the way, $500 billion.
    0:16:47 Yeah, they’re having a data center measuring contest. Yeah, yeah. But also, you know, not to
    0:16:51 get political, but there’s also reason they’re all going to Texas and places like that. As discussed
    0:16:57 before, this is going to require a major rethinking about like energy, the creation of energy and
    0:17:03 things like that, because these systems are going to require massive amounts of energy, like massive
    0:17:08 amounts. Yeah. Yeah, one of the things I heard about why they wanted to build out in like Western
    0:17:14 Texas specifically is it’s one of the spots in the country that gets the most sunshine and heat all
    0:17:19 year long, right? And there’s tons and tons and tons of open land in Texas, right? Especially
    0:17:26 West Texas. I’ve driven from Austin all the way to San Diego. There’s hours and hours of driving
    0:17:31 where there’s just nothing, right? And it’s also the area that gets like the most sun throughout
    0:17:37 the year. So yeah, the reasoning for it is partially political, but also just geographically,
    0:17:43 there is the land there and there is also the sunshine to get assistance from the solar power,
    0:17:47 obviously, right? Yeah, but there’s also like less restrictions on generating energy and using
    0:17:51 energy in Texas versus California. I mean, there’s definitely multiple reasons they’re choosing that
    0:17:58 location, but just the geography of it is also one of the big reasons as well. But yeah, I definitely
    0:18:02 have like mixed feelings about it. When I first saw it, I’m like, this is amazing. We’re going to see
    0:18:06 AGI like way sooner than anybody thinks. And that means we’re probably going to see ASI way
    0:18:11 sooner than anybody thinks. And maybe we are going to enter in this like sort of post-capitalistic
    0:18:17 world fairly soon, sooner than most people realize where we don’t have to work if we don’t want,
    0:18:21 because AI is just going to do everything for us. And then I started seeing a lot of the like
    0:18:26 Larry Ellis and stuff. And then I started thinking about the more like regulatory capture sort of
    0:18:32 element of it that open AI now seems to be really tied in with the US government. I’m curious, like,
    0:18:38 who else do you think could provide the sort of financing to do? Open AI has Masayoshi son,
    0:18:43 who’s essentially going to help them get to $500 billion over the next several years to build
    0:18:47 these dentisters. Who else has that capability? Oh, you mean outside of Stargate? Yeah, that’s
    0:18:52 interesting. And also you would like Masayoshi, I do wonder where the money is coming from. I don’t
    0:18:57 think he has all that money. No, no, he’s raising it. I mean, even Elon said like he’s only got 10
    0:19:01 billion dollars. It’s a cute or something. But I don’t think Elon knows actually. Yeah, I don’t
    0:19:04 think he does either. I mean, he might have heard from a friend or something like, oh, they were
    0:19:08 trying to raise money. This is how much they had committed at the time or something. But like I
    0:19:13 said, like with probably the internal data that open AI has, that is what’s making the fund raise
    0:19:18 that large. If they didn’t have the internal data showing, oh, yeah, here’s 04 and 05 is going to be
    0:19:22 like in three to six months after that, it’s going to be this much better. Like if they couldn’t show
    0:19:26 that, they wouldn’t be raising all this money. They have something incredible inside of open AI.
    0:19:30 Yeah. Well, I also think just the fact that, you know, they had the president announce it and they
    0:19:34 did the whole announcement with the White House and everything like that, that’s only going to help
    0:19:40 them raise, right? Like knowing that they’ve got the support of the US government, it’s not going
    0:19:44 to hurt their ability to raise. That’s only going to help them raise the money. I think after
    0:19:48 all the press conferences and all of that kind of stuff, I think it’s going to get a lot easier
    0:19:52 for them to actually come up with the funds to actually do this thing. You know, I mean, you’ve
    0:19:56 been saying it on podcasts since the very beginning, like nobody’s catching up with open AI. And if
    0:20:01 nobody else can build data centers like this, now I really believe nobody’s catching up with open AI.
    0:20:05 I have been saying that, right? For a long time. A lot of people are like, what are you talking
    0:20:10 about? Look, Claude’s great and all this stuff. I don’t know. Just things I’ve heard from friends
    0:20:15 who know Sam Altman is just the signs internally have been very positive for a long time, despite
    0:20:19 the drama that people saw, you know, from the company. Yeah, I don’t know. I do believe it’s
    0:20:23 eventually going to be open AI versus Elon Musk. Like I’ve been believing that for a long time.
    0:20:27 I think Google will try to catch up. Who knows, maybe Google will even have to try to make an
    0:20:30 alliance with Elon Musk or something. Like who knows, like what will happen there, like long term.
    0:20:35 But I do believe that the only one who could attract the talent and the capital would be
    0:20:39 Elon Musk. How big is the data center that Elon Musk is building? I can’t remember.
    0:20:46 I’m perplexing it right now. So he spent 2.5 billion on 100,000 H 100s and an additional 2.5
    0:20:50 billion on other data center costs and infrastructure. So I’m seeing 5 billion.
    0:20:55 Okay. So I’m seeing 6 billion and a lot of it came from the Middle East and say that even the
    0:20:59 thing with the open AI, you know, the Stargate, probably a lot of that’s Middle East money,
    0:21:04 quite honestly. Yeah. This actually does conclude 6 billion because it does say a recent 1.08 billion
    0:21:08 order placed for NVIDIA GB 200 AI servers brought it up to 6 billion. That investment
    0:21:12 just happened like within the last couple of weeks. So they were at 5 billion, just got
    0:21:16 another billion like a couple of weeks ago. Well, one thing that Skobal brought up too,
    0:21:21 he mentioned this in an X post that I saw earlier today is like, there’s been a lot of talk about
    0:21:26 essentially running out of data. Like if you’ve scraped the entire internet and all of the data
    0:21:33 has already been sort of grabbed, what do you need $100 billion data center for? Like where is
    0:21:39 the data coming from? So that is an interesting question too, right? I’m sure with a 5 billion,
    0:21:46 6 billion dollar data center, do you really need that $100 billion data center? I don’t know.
    0:21:49 I don’t know. I kind of disagree with that. It was a question that he raised, but so
    0:21:54 there was some recent research that came out showing some success, basically using data,
    0:21:59 using content created by the AI to teach the AI, like in training the models.
    0:22:03 Yeah, it’s synthetic data, right? Yeah, it’s synthetic data. And so the early signs seem kind
    0:22:10 of promising, and this was not from open AI. So I assume that the ’03, ’01 pro models are good enough
    0:22:15 to actually create synthetic data that’s actually helping improve the models. And so if that’s true,
    0:22:18 I mean, that’s what people have been saying for the last six months or so. If that’s true,
    0:22:23 in theory, that’s no longer a problem. They can just keep producing new content and actually
    0:22:28 train their models on that. And also, we’ve been saying they can keep scaling up with test time
    0:22:33 compute, so they can just throw more processing power to give these models more juice to actually
    0:22:38 think, yeah, you could spend infinite money. How much energy can you throw at it? The more you
    0:22:43 throw at it, the smarter it will be. Not to mention if there is a goal of putting cameras
    0:22:48 everywhere, that’s all new incoming data to build world models or whatever, right? If you’ve got
    0:22:54 cameras and drones and on bodycams and on cars, companies put them on their buildings or whatever
    0:22:59 for their own sort of security, and they’re trying to collect as much of that footage as they can,
    0:23:04 that’s going to require some pretty massive data centers to be able to pull all of that in
    0:23:10 and sort of analyze it for AI. But I also can see all of that data being what they use to train
    0:23:16 like actual world models to understand physics and the world around us and how things and people
    0:23:21 and objects move through the environment, right? Yeah, I think maybe our listeners are like,
    0:23:25 “Okay, cool. What does it mean for me?” Anything we’ve said on this show about timelines,
    0:23:31 just cut that in half or less now. Like literally a lot of things that we may have been saying
    0:23:36 three to five years, some of those things may be one or two years now. Development is going to
    0:23:40 increase. And like I said, the reason they’re able to raise this much money is OpenAI has something
    0:23:44 amazing internally that they’re showing investors. Yeah, I would believe it. And they still are
    0:23:49 maintaining their partnership with Microsoft and Microsoft is still getting access to any sort of
    0:23:54 new OpenAI models that they develop according to Microsoft, which means Clippy is going to get
    0:23:59 really good really quick. But Microsoft did not invest. I mean, to me that’s a signal. In Silicon
    0:24:04 Valley, if somebody invests and then they don’t follow into the next round, you try to frame it
    0:24:07 as not being a negative signal, but that is some kind of signal. And I don’t think it’s a negative
    0:24:12 signal on OpenAI’s part. I kind of think that OpenAI wanted to have other partners. It’s like,
    0:24:16 yeah, Microsoft’s one of our partners, but like, look, we got all these other partners. We don’t
    0:24:21 just need Microsoft. Well, the impression that I get is that the scale that OpenAI has in mind
    0:24:26 is bigger than what even Microsoft Azure can provide, right? Like I think they’re imagining
    0:24:31 this sort of scale that’s just sort of unfathomable for us. That’s what I said before. I said like
    0:24:35 people talking about Microsoft is going to take over OpenAI. I’m like, I envision the potential
    0:24:40 future where OpenAI acquires Microsoft to not have to deal with their contracts with them anymore.
    0:24:44 Like it’s the other way around. Yeah, yeah. I wonder if there’s a world where
    0:24:50 XAI and Elon actually get rolled into this new project. I know Sam Altman and Elon Musk have
    0:24:55 been beefing on X and whatever, right? But if they’re both really in it for the good of the
    0:25:00 country and building like the super intelligence, makes me wonder if there’s a world where they
    0:25:04 bury the hatchet and work on this together. Well, it’s crazy, you know, because like Sam Altman,
    0:25:07 definitely like, you know, I think I told you before, like I did a speech at Stanford and Sam
    0:25:12 was there at the same time. And I know politically he was at the time pretty far left. I think he’s
    0:25:15 like kind of switched now, being more somewhere in the middle. I think they’re all opportunistic.
    0:25:20 That’s what I think. Yeah, he hated Trump for sure. Like that’s definitely for sure.
    0:25:24 And so I do wonder now, it’s going to be kind of odd, but like I wouldn’t be surprised if that
    0:25:28 happens. It literally would be some kind of deal brokered by Trump between the two of them. Like,
    0:25:33 hey guys, makeup, it’s for the best of America. But I’ve been saying this for a while. I do believe
    0:25:38 that open AI owes Elon Musk something. I really do. Like I think it’s crazy that he owns no equity
    0:25:43 in open AI. I think that’s ridiculous. In the early days of open AI, like everyone in San
    0:25:47 Francisco, when they would talk about the company, they would talk about Elon Musk. And like Sam’s
    0:25:51 name would occasionally come up, but it was like, oh, open AI, oh, that’s Elon Musk thing.
    0:25:56 Yeah. The first time I ever heard of open AI, it was with Elon Musk attached, right? Like that
    0:26:00 very first time I ever heard of it. It was like, this was another Elon Musk company.
    0:26:03 Yeah. It was like his company and like Sam Altman’s help and run it. It was like, that was at least
    0:26:08 the messaging externally that was being presented to people. And then also like the main talent
    0:26:14 early on, he helped recruit the capital. He provided some of it, but also it was from his friends.
    0:26:18 Yeah. Yeah. Didn’t he bring on Ilya? Like I think Ilya was working with Elon in the beginning.
    0:26:21 I think he did. I could be wrong about that. Don’t quote me.
    0:26:26 Yeah. My understanding is he did. And so a lot of the main talent he brought on and the capital
    0:26:30 and just the reputation. And then also that’s having someone like Elon Musk brings so much
    0:26:34 talent to you too. Just like, oh, it’s Elon Musk’s AI company. I’ll go work there now.
    0:26:39 And so the fact that he owns nothing is crazy. So I would love if something happened where
    0:26:44 Elon Musk owns a small piece of open AI and they have some kind of technology sharing deal,
    0:26:48 but like, will that actually work? I have no idea. Yeah. Yeah. I don’t know. Anyway,
    0:26:53 there was another really big piece of news that came out this week and it is related to open
    0:26:58 source and China. Right. This week we got deep seek R one. There was already these deep seek
    0:27:05 models out there, but the R one is this new like research model. And I did test it on a live stream
    0:27:12 and I tested it side by side with O one just standard and O one pro. O one pro definitely
    0:27:17 still outperforms deep seek R one. It was definitely giving me more like quality in-depth
    0:27:25 responses, but comparing it to O one, they felt pretty even, which to me blew my mind for an
    0:27:29 open source model. I know you have some opinions on it. Obviously the models out of China. Yeah,
    0:27:34 our good friend Matthew Berman was doing some testing with it and he asked it about
    0:27:39 if like Taiwan was part of China and it basically said Taiwan is part of China and anybody who
    0:27:44 opposes this thought will be shut down or something weird like that. Right. It wasn’t exactly that,
    0:27:48 but yeah, it was in that vein. No, I’m definitely paraphrasing, but it was basically saying like
    0:27:52 any plans for independence will not work or something like that. Right. It was some wording
    0:27:57 like that, which I found really interesting when I actually asked that same exact question,
    0:28:02 like I put the exact same prompt into deep seek and it just said, sorry, I can’t answer that.
    0:28:05 Let’s talk about something else. That’s what it did when I put the prompt in. That’s what I’ve
    0:28:09 been calling it like open propaganda, like open source. I don’t know. It’s open propaganda.
    0:28:14 I mean, because it is like, okay, so why is China allowing this model to be open source and be out
    0:28:18 there? The Chinese government is allowing it. And to me, that is why they’re doing it is because
    0:28:23 they know that people love open source. It’s a great way psychologically to have people go,
    0:28:27 oh, it’s open source. I love it. But there’s other things going on here. Like the reason they’re
    0:28:31 allowing it is because if it becomes one of the biggest models, they get the kind of control
    0:28:36 reality and distort history through that. Right. Like, okay, TMN square didn’t happen or different
    0:28:40 things like that. They can change that in the AI model. And, you know, full disclosure, I mean,
    0:28:46 I did live in Taiwan. I studied Mandarin there, kind of biased. I love Taiwan. But for that reason,
    0:28:49 I just, I can’t support the model. It’s because like, yeah, you talked to it about Taiwan and it’s
    0:28:54 like, yeah, it’s owned by China. No one in Taiwan sees it that way. Or maybe a few people, but like
    0:29:00 not many. Yeah, yeah. But I mean, like there is biases built into pretty much every AI model,
    0:29:06 right? Like, you know, a lot of the US models refuse to talk about certain things or
    0:29:11 sort of share their own political bias that was trained in as well. Right. But yeah,
    0:29:16 I definitely see that. And the fact that it’s actually open weights, though, people can take
    0:29:23 this deep seek R1 and fine tune it and sort of essentially train out all of that bias if they
    0:29:27 wanted to, right? Because the model weights are open source as well. Like you can actually take
    0:29:33 the weights and fine tune them if you want. So, you know, as a whole, as a model that’s open source,
    0:29:38 I do think it’s really, really impressive that we’ve gotten to 01 level this quickly,
    0:29:42 right? Like, you know, they’re constantly talking about the gap between when we released
    0:29:48 GPT-4 to 01. Look at how big that gap was and then 01 to 03. That was only like a three month
    0:29:52 difference. Well, we’re seeing that sort of same scaling happen in open source as well.
    0:29:57 Yeah, there was a tweet from one of the open AI guys and he was kind of saying,
    0:30:02 I expect to see a lot of new reasoning models in the next few months in open source and other
    0:30:07 areas that don’t fully get, you know, how it’s done, or like he basically kind of lunage like
    0:30:10 opening is doing something a little bit different than you think. It’s not as straightforward as
    0:30:14 what you think. And I kind of still think that’s probably true. I think that’s why open AI is moving
    0:30:18 faster than anyone else. They have discovered something, but I did test deep seek. I mean,
    0:30:24 it was impressive. It’s better than Claude. It’s definitely in the ballpark of 01. I found that
    0:30:29 in some ways it was better than 01 and in some ways it was worse. I tested it on coding after
    0:30:34 you told me to try it. And I found that like in some ways the code was better in some areas.
    0:30:37 I was like, oh, that’s amazing. Like it’s actually better. And then it would do some things that
    0:30:42 were like pretty dumb that 01 would never do. Yeah. I was like, I think there’s something missing
    0:30:45 in the reasoning side of this. I could see it. There’s something they’re not getting that the
    0:30:50 open AI is getting some technique that they’ve discovered that this model is not using. It would
    0:30:53 hallucinate more. It would imagine files and things like that. They didn’t exist. I was like,
    0:30:58 what? Like how this is a reasoning model? Like shouldn’t it have like checked to see that that
    0:31:02 file actually existed? It’s actually hallucinating that. That’s like, you know. Yeah. Well, one thing
    0:31:07 that I do like about the R1 model though is like you can literally see everything it’s thinking
    0:31:12 as it thinks, right? You can see it go, all right, let me test this. Okay, that didn’t work. Let me
    0:31:16 test this. That’s not the right way to do it. Let me test it. And it’s literally showing you
    0:31:21 everything it’s trying and testing and going back and forth. And I think that’s pretty fascinating.
    0:31:26 But I also think there is a reason open AI isn’t showing you all of that, right? Like it does show
    0:31:30 you some, but it’s almost like a summarized version of how it thought as opposed to the whole
    0:31:34 thinking process. Yeah, there’s some secret sauce there. And that’s why I’ve been saying for a while,
    0:31:38 like I’m a huge fan of Elon Musk, but he was kind of saying, oh, we’re going to have this new model
    0:31:42 and it’s going to be better than open AI. It’s going to be the best in the world. You know,
    0:31:46 I don’t know if that’s the case. I think that you can’t just, you know, okay, you have more
    0:31:50 processors and now you trained a larger model. I don’t think that’s the game. Like I think open
    0:31:53 AI is playing a different game now. They’ve learned that it’s a combination. You’re training
    0:31:57 the model with more data, but also there’s a whole test time compute there and some kind of secret
    0:32:02 sauce that they have discovered with deep seek. I guess the one takeaway is like you’ve been saying
    0:32:06 before, eventually we are going to have open source models that are like somewhat close to the best
    0:32:10 models. Maybe they’re not as good in some ways, but you’re going to be able to have like models
    0:32:14 that you can run on your local machine that are really freaking good. Well, and the reason that
    0:32:19 I pointed out to you that I thought you might be interested in it was more the fact that there’s
    0:32:22 actually an accessible API for it right now. Well, you can’t use O3 at all right now,
    0:32:28 but there is no like O1 API. So you can’t just use it inside of something like a cursor or a
    0:32:32 Windsor for something like that. Oh, you can, you can use O1 in cursor. Oh, you just can’t use O1
    0:32:37 Pro yet, right? Correct, correct. Okay. And there’s a lot of limitations on O1 in cursor too. I
    0:32:41 think unless you put in your own API key, but like if you’re using it by default, it’s like very
    0:32:46 restrictive. Like you run out of like queries or prompts, whatever, very vast. Oh, so even when
    0:32:50 you’re using the API, they still rate limit you? Well, if you put in your own API key, they don’t,
    0:32:55 but if you’re using just like the cursor plan or whatever, it’s very restrictive. It’s probably
    0:32:59 because the O1 API is more expensive. So they have some kind of like system where it’s like,
    0:33:02 if you’re paying us 20 bucks a month, we’re not going to allow you to put up some gigantic O1
    0:33:07 bill and then we pay for it. Yeah, yeah, yeah, yeah. Okay. Yeah. For some reason, it slipped my
    0:33:12 mind that O1 had an API available. And I’m like, Oh, well, here’s an alternative to O1 that has
    0:33:18 an API. That’s O1 Pro that doesn’t. And actually, it’s worth noting too that Sam Altman has been
    0:33:22 being way more public about their like upcoming plans recently, like he’s talked about like
    0:33:28 O3 minis coming in the next two weeks, which is like, holy crap. Although he did say O1 Pro is
    0:33:33 still better than O3 mini. Did you see that? Yes. But under that tweet, somebody asked him,
    0:33:38 how is O3 mini compared to O1 Pro? And he said O1 Pro is still going to be way better at most
    0:33:43 things. Yeah. I mean, we had seen some benchmarks that had shown that like maybe a few weeks ago,
    0:33:48 but the big difference will be that O3 mini will be really fast. And also it’s really
    0:33:53 cheap for open AI to operate. So it should be a model that’s like better the deep seek,
    0:33:59 way better than O1. That’s very fast. And the other comment that he made on that same tweet,
    0:34:04 is that they’re going to release it with an API, which amazing. No, but also even that the O3
    0:34:09 Pro is going to have an API. So like apparently O3 and O3 Pro are coming. So they’re going to
    0:34:13 continue the thing where now that they’ve learned that you can just throw more compute at the models,
    0:34:18 they can always have a better model available, just like throwing more compute at it, right?
    0:34:22 And so they will continue to have like a pro model. And the amazing thing there was I was
    0:34:26 concerned that they’re going to like increase the price to like 2000 or something. Because I was
    0:34:30 like, well, I might actually pay it. Like if it continues to get better and it replaces me hiring
    0:34:36 an engineer, I might actually pay that. But he said it’s going to continue to be $200. And that
    0:34:41 the O3 Pro model will have an API as well. So that’s super exciting. So the O3 Pro is going
    0:34:46 to be on the $200 model? Yes. And have an API. That’s the one where they were like, it costs
    0:34:51 like three grand per task right now. Well, they said that they’ll figure out how to make it cheaper.
    0:34:54 And apparently they seem to think that they probably have or they’re going to have by the
    0:34:58 time it comes out. And the latest information too, I forgot who it was. If it was like their chief
    0:35:01 product officer or something like that, he was interviewed by the Wall Street Journal. Kevin
    0:35:06 Weill. Yeah, exactly. Yeah. And he said that the current plan is that the O3 models, not the mini,
    0:35:09 the mini should be coming in the next week or two. Maybe by the time you’re listening to this,
    0:35:15 it may already even be out. But that the full-blown O3 models, which probably means O3 and O3 Pro,
    0:35:19 timeline is like two to four months. For someone who’s been using O1 Pro, and like at the last
    0:35:23 episode we did where I showed you how much value you can get out of O1 Pro if you give it tons of
    0:35:28 context, to imagine that we’re about to get a model like three to five times smarter than that
    0:35:35 in the next three months. It’s just blowing my mind. Well, and we’re about to get agents. I mean,
    0:35:40 don’t hold me to this, but like the week that this episode is going live, we might already have
    0:35:45 agents or it might be announced this week. But like apparently, according to the information,
    0:35:50 OpenAI is preparing to release a new chat GPT feature this week that will automate complex tasks
    0:35:54 typically done through the web browser, such as making restaurant reservations or planning
    0:35:59 trips. According to a person with direct knowledge of the plans, they’re calling it operator. And so
    0:36:04 this says this week. Yeah. And actually, there was some kind of benchmark leaked recently showing
    0:36:09 operator versus Claude’s computer use showing the operators way better, like not perfect. A lot of
    0:36:13 stuff like Claude was like in the 50% success rate. And then the operator was like in the 80%
    0:36:17 success rate, but still, yeah, dramatically better. Yeah. And that’s actually when we talked about
    0:36:21 computer use before. That’s what I said. I was like, I’ve heard that like OpenAI has stuff internally.
    0:36:26 That’s pretty good. There’s not happy to release it yet. Right. And that’s Claude released their
    0:36:30 thing very early, but no one really used it because it wasn’t that great. But it sounds like,
    0:36:34 I mean, they’re saying it’s going to be like do stuff for you, like make reservations online,
    0:36:38 buy things for you. Just like basic stuff you would do in the browser. Probably when operator
    0:36:42 comes out, you’ll be able to just tell it, hey, go do that for me. And then it’ll just do it.
    0:36:46 Yeah. It’s really a pain in the butt to set up the anthropic version, right? Like if you want to
    0:36:49 use computer use, you have to do it through Docker. You have to get it all set up. And you
    0:36:54 actually have to use it with their sort of crappy browser, like on a remote computer.
    0:37:00 It’s not a great experience. I would imagine when chatGPT rolls it out, it’s just going to work
    0:37:04 in your own browser. It’s going to be a lot more seamless of an experience.
    0:37:08 Yeah. Yeah. So yeah. So exciting times. I mean, we’re about to have like AI that can
    0:37:12 like just code full products for you. You just talk to it and it makes the whole thing. Like
    0:37:17 this is probably this year. And you have AI that can like just use websites for you and do whatever
    0:37:21 you would do on your computer, do it for you. It’s exciting. I mean, a lot of stuff that just
    0:37:24 is time consuming stuff that you don’t want to waste your life on. Soon you’re not going to have
    0:37:28 to waste your life on it. I can’t wait till the day where I’m just like, hey, I need to update my
    0:37:33 future tools website, go find all the news and put it on my website for me. I’m going to go take a
    0:37:39 nap, which is funny because I don’t think that’s actually that far off. Probably not. I actually
    0:37:43 set up that thing with a chatGPT where it sends you little notifications or whatever. It’s not,
    0:37:46 they definitely need to improve that experience, but it’s kind of cool. It’s been a similar little
    0:37:50 Japanese words to learn and send me little summaries of AI news. Yeah. I’ve done it too.
    0:37:56 I set up a daily task list to find any AI news from that morning and don’t find anything from
    0:38:01 like before today. I only want the most current news. Yeah. And it sends me an email every morning
    0:38:05 to let me know that it’s done it and I click the link and it shows me what it found. Yeah,
    0:38:09 and it’s useful. I tried it recently in my newsletter issues to see if anyone would complain,
    0:38:16 but I just took a summer of the news and then used whisper flow just to talk to my computer
    0:38:21 about my own thoughts on the matter. Yeah. Right? Did that for like 10 minutes,
    0:38:25 handed it off to 01 Pro where I had provided all this context of what’s good newsletter
    0:38:29 issues and what’s not and it edited all my words and like looked really good and everyone seemed
    0:38:34 to like it. Nice. And it dramatically reduced how long it took me to create my newsletter issue.
    0:38:37 Like, you know, instead of taking like a few hours, it probably took me an hour to finish
    0:38:41 everything. And so I’m just like, this new world is so exciting where you just like all the kind
    0:38:44 of work that you don’t like doing. You know, I like sharing my opinions. I don’t like sitting
    0:38:48 there like editing them for like hours and doing all that. Yeah. Yeah. It’s going to do all that
    0:38:52 for me. It’s awesome. Yeah. I saw Dario Amadei from Anthropic. Was it the Wall Street Journal
    0:38:57 that was doing all the interviews out at that sort of Davos thing? Yeah. So he was on there.
    0:39:03 They were actually asking about like the future of jobs. And he essentially said that like what
    0:39:08 they found and what a lot of research has shown is that when you give people all of these automations,
    0:39:13 it doesn’t actually take away too many jobs. It just makes people way, way, way more efficient
    0:39:18 at the stuff they actually want to be doing within their job. And he was talking about how
    0:39:24 basically so many people have been trying to use AI to replace jobs. But if you start using AI as a
    0:39:30 way to like sort of enhance jobs, they find that the effectiveness of those people is way better
    0:39:36 than when you use AI to replace jobs like the efficiency and the output and that sort of thing
    0:39:42 is like dramatically improved when they’re using AI to actually like improve the things and also
    0:39:47 hone in on just doing the things that are like their core competencies and letting AI do all the
    0:39:53 mundane stuff. So, you know, I think that’s sort of the next phase. I mean, there might be a phase
    0:39:57 beyond that where it’s just kind of like, all right, AI and the robots do everything and we just
    0:40:02 get to travel and live our best life. That might be like a future phase. But I think the next phase
    0:40:06 that we’re moving into is like, we get to focus on the stuff that we actually enjoy doing in the
    0:40:12 work that we do while AI does all the sort of mundane boring stuff we just don’t want to do
    0:40:16 because it’s repetitive or whatever. Yeah, totally. I mean, like, like I’m already seeing that in my
    0:40:20 life right now. The fact that like these things are about to get three to five times better.
    0:40:24 And it appears that that’s going to start happening like every three to six months,
    0:40:28 not every like two years. Yeah, it’s just amazing. I mean, because like, and then now that you got
    0:40:32 Stargate, I don’t think people are like processing like they’re not like reimagining like what’s
    0:40:36 going to happen based on the new data because like things are going to improve faster than you
    0:40:42 expect if you’re listening to this. And the OpenAI also recently talked about like their last models
    0:40:45 took like, I think a year and a half or longer than a year and a half to train. And they were like,
    0:40:50 maybe like 50% to 100% better. You know, the new 03 model, they’re seeing like a, you know,
    0:40:55 3x improvement in a matter of three months. We’re entering a new phase of development here. We’re
    0:40:59 talking about probably like improvements speeding up like 10 times or more. Yeah. And now there’s
    0:41:05 going to be a hundred billion dollar data center worth of compute behind OpenAI to move even faster.
    0:41:10 So I think you’re going to see that sort of exponential growth continue, right? Like,
    0:41:14 it’s just going to be a vertical line. And also there’s going to be now that that’s
    0:41:16 announced. I mean, you’re going to see Elon Musk and everyone else
    0:41:20 raising more money to go after this faster too. It’s not going to be just OpenAI. It’s going to
    0:41:24 be everyone else. It’s going to be investing more into this for the race. For sure. It’s going to
    0:41:29 happen faster. Yeah. Well, we did say there was actually a lot to talk about. We probably rambled
    0:41:35 a lot. Sorry. Yeah. We had a lot to talk about. It felt like a pretty monumental week between
    0:41:41 the Stargate, between deep seek, between, you know, the announcements from OpenAI and 03 coming.
    0:41:46 There was a lot, a lot, a lot of big stuff that I felt like we needed to kind of deep dive and
    0:41:51 unload. And hopefully anybody who’s just sort of listening to this podcast and checking in to
    0:41:56 stay looped in, you feel a little more looped in. You might feel a little bit more optimistic.
    0:41:59 You might feel a little bit more scared. I don’t know. Or confused or whatever. Yeah.
    0:42:02 Maybe I like to share something about what Stargate means for the future. There was this tweet
    0:42:05 and it’s kind of philosophical, but Roon shared this tweet, you know, people like,
    0:42:10 why is it called Stargate? Roon’s like a well-known guy who works OpenAI. He’s like,
    0:42:13 a lot of people know who he is, but he keeps an anonymous. He’s not Sam Altman. A lot of people
    0:42:19 think he’s Sam Altman. He’s not. But he had this tweet, the Stargate blasting a hole into the
    0:42:25 Platonic ether to summon angels. First contact with alien civilizations. So I think that is kind
    0:42:29 of a summary of like what Stargate means. I mean, like this is we are like summoning the angels. We
    0:42:34 are making contact with a new intelligence, you know, an alien intelligence. And that will be
    0:42:40 artificial superintelligence. It will be like us discovering something beyond anything we can
    0:42:44 imagine. And that’s what this is designed to do. And so it’s just, you know, I think it’s
    0:42:47 important for people to take a moment to kind of try to take that in. It’s a lot to take in,
    0:42:50 but that is what humanity is trying to accomplish right now.
    0:42:57 Wild. Well, hopefully these aliens are coming to save us and not to destroy us. I choose to
    0:43:03 lean into the optimism and I believe that it’s all going to make humanity better and make us all
    0:43:08 better and more improved, augmented humans that help us get more done that we want to get done.
    0:43:12 So I’m looking forward to the future. I’m excited about it. There are some little concerns,
    0:43:17 but I still lean mostly optimistic on all of this. And I know you do as well.
    0:43:22 Yeah. And I think on that note, we can go ahead and wrap this one up. I think we’ve sort of
    0:43:27 unloaded everything we have to say about all of these big announcements that came out this week.
    0:43:33 If you liked this episode and you want to stay looped in on all of this latest AI news and hear
    0:43:38 some amazing interviews and discussions and deep dives around ways to practically use AI in your
    0:43:42 life. Make sure you’re subscribed to this, either our YouTube channel or audio channel,
    0:43:46 wherever you listen to podcasts. These podcasts is available all over the place.
    0:43:50 We’d love for you to subscribe and thank you so much for tuning in. Hopefully we’ll see you in the next one.
    0:44:00 [Music]
    0:44:04 [MUSIC PLAYING]
    0:44:14 [BLANK_AUDIO]

    Episode 43: How massive is the $500 billion AI Stargate Project, and what does it mean for the future of AI? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) break down this monumental development and discuss its implications.

    This episode delves into the ambitious AI infrastructure project announced by Donald Trump, alongside key tech players. Learn why this initiative is seen as a groundbreaking moment in AI history, its promises for job creation, health advancements, and the potential concerns around surveillance and data privacy. The hosts also touch on related advancements including DeepSeek-R1, OpenAI’s upcoming models, and how they all fit into the larger AI development landscape.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Stargate’s $500B AI Infrastructure Investment
    • (03:17) AI Innovation: U.S. Health Revolution
    • (09:17) Surveillance Expansion Concerns
    • (10:27) Corporate Influence and Surveillance Speculation
    • (14:57) Sam’s Breakthrough in AI Development
    • (17:16) Mixed Feelings on AI’s Future
    • (20:15) NVIDIA’s $6B Data Center Expansion
    • (23:14) OpenAI Expands Beyond Microsoft
    • (27:43) AI Model Bias and Control
    • (29:11) OpenAI’s Edge: Unique Reasoning Models
    • (34:20) New O3 Models Release Timeline
    • (38:08) AI Boosts Job Efficiency, Not Loss
    • (39:32) Rapid Technological Advancements

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • You’re Using the Smartest GPT Model Wrong (GPT o1 Full Tutorial)

    AI transcript
    0:00:14 Hey, welcome to the next wave podcast. I’m Matt Wolf. I’m here with Nathan Lanz. And
    0:00:20 today we want to dive deep into what chat GPT’s 01 models are capable of. We believe
    0:00:25 that a lot of people really misunderstand what they’re actually capable of. And most
    0:00:31 people are probably using 01 wrong. And in this episode, I think we might actually blow
    0:00:35 your mind with what they can do. I’m going to show you some simple workflows that I
    0:00:41 use to actually create short form content from my long form content using 01. But then
    0:00:46 Nathan, he’s going to break down his flow for actually building a game. And he’s going
    0:00:50 to give you a little sneak peek at the game that he’s showing you and, you know, stick
    0:00:54 with them when we get to it, because it does take a couple of minutes to understand it.
    0:00:58 Once you have that aha moment, which you’re going to have that aha moment when he’s showing
    0:01:03 it to you, you’re going to have your mind blown by what 01 pro is capable of. And Nathan’s
    0:01:07 game is looking killer. And you’re going to see what I’m talking about if you’re watching
    0:01:13 this one on video. So stick around for this one. We’re going to dive in and show you what
    0:01:21 01 from open AI is really capable of. Look, if you’re curious about custom GPT’s or our
    0:01:26 pro that’s looking to up your game, listen up, I’ve actually built custom GPT’s that
    0:01:31 helped me do the research and planning for my YouTube videos and building my own custom
    0:01:36 GPT’s has truly given me a huge advantage. Well, I want to do the same for you. HubSpot
    0:01:41 has just dropped a full guide on how you can create your own custom GPT and they’ve taken
    0:01:46 the guesswork out of it. We’ve included templates and a step-by-step guide to design and implement
    0:01:51 custom models so you can focus on the best part actually building it. If you want it,
    0:01:56 you can get it at the link in the description below. Now back to the show.
    0:02:02 Yeah, I feel like it’s the first model where like there’s a major disconnect between like
    0:02:07 how good the model is and how people perceive it to be. Like they think it’s just like a
    0:02:10 slower version of a chat botter’s and they’re like, why, you know, how is this an upgrade?
    0:02:14 And so I think there’s a major disconnect and it feels like it’s also the first time
    0:02:18 there’s a model where people are kind of the limit and your knowledge that you bring to
    0:02:22 the table when you use it are kind of a limit, right? Because like, if you don’t know how
    0:02:27 to properly use the tool, the results you’ll get back will be way bad and it takes several
    0:02:31 minutes to get the response back and it’s a horrible experience. You’re like, I waited
    0:02:35 for a few minutes and you give me some crap back. Why am I ever going to use that again?
    0:02:38 And so apparently a lot of people have like tried that one time and just never use it
    0:02:41 again. The people who like figure out how to actually properly use it are like kind of
    0:02:45 blown away by what you can do with it. Yeah. Yeah. And in a few minutes here, we’ll actually
    0:02:50 dig in a little bit and show off exactly how we’re using it. But you know, that’s, that’s
    0:02:54 totally right. Like I use regular chat GPT and I’ll, I’ll dump like a transcript from
    0:03:00 a video in with the timestamps in it and tell it to find me like clips from that video.
    0:03:04 And with regular chat GPT, it’ll find like moments from the transcripts, but the timestamps
    0:03:09 will be like way off. Like it can’t seem to like line up the timestamp with what that
    0:03:14 clip it was that found. When I do it with 01, 01 does it’s like double checking, triple
    0:03:17 checking. And you know, that’s kind of what it’s doing when it’s doing all of this processing
    0:03:23 and it’s taking so much longer. It’s basically prompting behind the scenes, getting a response
    0:03:28 and then double triple quadruple checking and sort of reevaluating its response over
    0:03:32 and over and over again before it finally goes, okay, we think we got this right. Here’s
    0:03:33 our response.
    0:03:37 Well, we think that’s what’s going on. I mean, we think you know, like open eyes be not not
    0:03:40 being entirely transparent about what’s going on. They’re kind of alluding to them having
    0:03:44 some kind of other secret sauce, but yeah, that’s probably the majority of what’s going
    0:03:48 on. But, but yeah, man, I think that’s the key is, and we will show later, like giving
    0:03:52 these models the proper context with like Claude or like, are any of the other like
    0:03:55 just regular chat models, you know, the LLMs, they don’t have a reasoning model attached
    0:03:58 to them. Like you can just kind of go back and forth. You can just like ask it a simple
    0:04:02 question and instantly responds back and you just kind of go back and forth. But with like
    0:04:06 01 or one pro, you know, you can throw so much contacts at them, you can like, you can
    0:04:11 copy and paste in, you know, paid like 10 pages, 100 pages, maybe even 1000, not sure
    0:04:20 the exact it’s like token limit is on 01, 01 pro, I believe is 125, 125,000 125, but people
    0:04:26 who’ve tested it have said that kind of stick to like 50 to 75 K, like there’s some kind
    0:04:29 of thing it’s like almost like a ram back in the day or memory like, yeah, you can hold
    0:04:33 that much, but don’t like fill it up. Well, according to perplexity here, I actually did
    0:04:40 a quick search on it. According to perplexity, the 01 model is a 200,000 token context window
    0:04:46 with a maximum output token window of 100,000. So you could put up to that’s about 75,000
    0:04:51 words into like a single prompt. The results you get, like at least with coding that you
    0:04:55 get from like sharing it that much context versus just like asking it a question. It’s
    0:04:59 like, it’s like a night and day difference. And I found the same thing with writing and
    0:05:03 I’ve heard other people sharing examples too that they’ve tried to 01 for writing and they’re
    0:05:06 like, Oh, this kind of sucks are made slightly better than Claude or like it’s like about
    0:05:11 the same as Claude. But like actually, like with 01 pro, if you give it tons of examples
    0:05:16 of like, here’s good writing, here’s a good newsletter, or here’s my best newsletter issues
    0:05:21 and some people who write newsletters I really respect and wish I could be like, do that
    0:05:26 like the stuff that gives you back even just like for like editing is so good. And actually
    0:05:32 before 01 pro, I almost never used AI for my newsletter at all. And so I’ve been using
    0:05:37 whisper flow. And basically where I can just like press a button and just talk to the computer.
    0:05:42 And then it just uses AI to transcribe what I said. I think it wasn’t the one that Riley
    0:05:44 was talking about on that episode we did with Riley, I think he might have brought that
    0:05:48 up and said he was actually using that to code with. I believe so. I’m sure I’ve learned
    0:05:52 so many things from this show like subconsciously like we’re like, oh yeah, I’m going to try
    0:05:56 that out. I don’t know why I’m trying it out, but I am well, I have to subtly slip that
    0:06:00 in so people go, oh, they did an episode with Riley Brown. I got to go listen to that one.
    0:06:04 I’ve been using it that way. And lately for my newsletter, I will use whisper flow and
    0:06:09 just talk to it. And you know, and I’ll talk for like five to 10 minutes about whatever
    0:06:13 I want my newsletter to be about for that day. And then I’ll like hand it off to 01 pro.
    0:06:18 And I gave 01 pro, you know, I’m copying in examples of my favorite, my best newsletter
    0:06:23 issues, but also newsletters that I like and things I don’t like. And it’s doing an incredible
    0:06:28 job at editing my, what I said to make it like a really presentable and professional.
    0:06:31 And the previous models were not, we’re nowhere near that caliber.
    0:06:35 Well, and I think, I think what you’re saying too is like getting to the root of like why
    0:06:40 we mentioned that most people are probably using 01 wrong. Like the regular chat GPT,
    0:06:46 the GPT 40, it’s designed to be conversational, right? It’s designed for you to ask it a one
    0:06:51 or two sentence question. It gives you a response. You give a follow up question, and it’s designed
    0:06:56 to do that sort of back and forth, back and forth. And you get into this long conversation
    0:07:00 and it sort of ideally remembers the context of the previous conversation.
    0:07:05 01 models, on the other hand, are not really designed to do that. Oh, one models are really
    0:07:11 kind of designed to dump as much information as you can into that very first prompt, like
    0:07:15 you mentioned, right? Dump in the newsletters you like dumping your own newsletters, dumping
    0:07:19 maybe the information that you want included in the newsletter that it’s about to write
    0:07:24 for you and give it all of that information, all of that context from the very first prompt,
    0:07:29 hit submit, and then let it go to his town doing his processing, right? Let it spend
    0:07:34 10 minutes, 15 minutes, however long it takes processing all of that information. And then
    0:07:38 it’s going to give you a nice detailed output with all of that information.
    0:07:43 If you try to use it to like chat and be conversational with it, ask it a one sentence question
    0:07:48 and then wait for a reply. Ask it a follow up question. You’re going to hate it, right?
    0:07:50 Because it’s going to take forever with each question.
    0:07:54 A lot of people have noticed too that it seems like it’s really great at like one-shotting
    0:07:58 things, like in terms of you want to give that huge context right up front. And often
    0:08:03 after that, you’re kind of done with the conversation. You kind of like when you want to use it again,
    0:08:08 you open a new, you know, 01 pro chat, right? Like you can kind of continue, but I found
    0:08:11 like, you know, the more and more you throw at it, like eventually it kind of gets more
    0:08:15 confused. And like the first one, if you just get tons of context, it’s able to reason about
    0:08:19 all of it and give you a great, you know, response. And I think most people don’t realize
    0:08:22 that they, like you said, they’re like, they’re waiting for several minutes and they’re like,
    0:08:25 okay, that kind of sucks. Let me talk to it some more and then they wait more and just,
    0:08:26 it never goes anywhere.
    0:08:30 I think you pulled up some tweets about some of the interesting ways that some like like
    0:08:35 scientists and things like that are using 01. So let’s, let’s maybe talk about those
    0:08:38 real quick first and then we’ll show them some of the ways that you and I have actually
    0:08:40 been playing around with it.
    0:08:43 You know, this is something that I’ve been noticing too, like online is there’s a huge
    0:08:47 disconnect between like the people who are just trying to chat with 01, 01 pro and the
    0:08:52 people who are like trying to like solve really hard problems, you know, with it. For example,
    0:08:56 here’s this doctor on X who’s been sharing really great stuff. You know, and he’s talking
    0:08:59 about how, you know, he thinks people don’t realize how good this model is. Like they’ve
    0:09:04 been using it, been using it to help create an incredibly complex composite biological
    0:09:07 framework, which, you know, there’s a lot of technical stuff in here, but it sounds like
    0:09:11 basically this is something that’s actually helping them to identify target drugs that
    0:09:15 you could create and even give them good information about how to possibly create the drugs, how
    0:09:20 to do, to run tests on them, things that before you would need a whole staff of people to
    0:09:25 help you do. And he’s saying that now, instead of having that staff, he’s basically be able
    0:09:28 to do it himself, which means, you know, if you gave this kind of technology to every
    0:09:33 doctor, like how fast we discover new drugs, it’s going to go up dramatically. I think it’s
    0:09:38 not a surprise that you’re seeing it more in like complicated areas like, you know, engineering,
    0:09:42 the medical field, things like that. You’re noticing that people in those fields are understanding
    0:09:45 how good these models are because their needs are more than like the average person who’s
    0:09:49 asking you, like, you know, I’m shopping for this or whatever, you know, right. And here’s
    0:09:54 another tweet from a DD who’s a venture capitalist at Mineral of Ventures. We’re the well-known
    0:10:01 Silicon Valley Venture Fund. And he’s saying that based on the data that already AI like
    0:10:05 oh, like oh, one like reasoning models are doing better than doctors on solving hard to
    0:10:10 diagnose diseases. So as of right now, and I think the numbers was 80% versus 30%. What
    0:10:16 are those numbers mean? Like hard to diagnose diseases that when you were testing a doctor,
    0:10:21 if the doctor could diagnose the disease, the doctors got it right 30% of the time. Wow.
    0:10:27 The AI got it right 80% of the time. And this is not the new models. This is the very first
    0:10:32 preview of 01. And from a lot of stuff, I’ve seen 01 pro is probably in the ballpark of
    0:10:37 three times smarter than that model. So probably when the new data comes out, it’s going to
    0:10:43 be like, okay, it’s not 80% like it’s 95% or 90% and the doctors are 30%. It’s a huge difference.
    0:10:48 I mean, so this just shows this one reason we started the podcast, right? Like, I think
    0:10:52 most people don’t realize like this is society change and stuff like this should be where
    0:10:56 we restructure society where, you know, you still need doctors, but you have doctors who
    0:11:02 are highly relying on AI to help diagnose diseases. Ideally you have the doctors that
    0:11:06 sort of understand all of the different diseases and the ways to cure them and things like
    0:11:11 that. But, you know, maybe they’re not always the best at actually diagnosing what the disease
    0:11:16 is, but they’re the best at probably telling you how to like handle and, you know, work
    0:11:19 with you on the treatment of it. So I think we’re going to get to a point and I know I’m
    0:11:24 probably already going to start doing this where if I have like a checkup or a doctor’s
    0:11:27 appointment or like something’s bothering me and I’m going to go to the doctor, I’m
    0:11:33 going to put all of my symptoms into something like 01 and basically see if 01 can tell me
    0:11:38 what it thinks is the problem first, but then go and use that information and bring it to
    0:11:39 a doctor.
    0:11:45 I’ve been hearing stories lately about people who basically get to like the root problem
    0:11:50 of their various ailments by using 01 and then going to the doctor and the doctors essentially
    0:11:55 confirming it for them and then helping them with a course of treatment. So it’s not like
    0:12:00 eliminating doctors. It’s just sort of, all right, let’s, let’s get an opinion here and
    0:12:04 let’s get a second opinion from a real doctor and then let’s sort of overlap the two to
    0:12:08 figure out the best course of action. And I really think that’s probably going to be
    0:12:11 the smartest way for people moving forward. And I think it’s going to get to a point where
    0:12:16 like any doctors that refuse to also leverage AI to sort of help with some of the diagnosis
    0:12:22 and stuff, it’s like, that’s borderline, like going to be unethical to not sort of get a
    0:12:26 second opinion for it from AI or at least, or, you know, get a first opinion from AI and
    0:12:29 then have a doctor confirm the opinion, right?
    0:12:33 Right. I think the best thing we should do now is, you know, you talked about using it
    0:12:37 for your newsletter, you’ve talked about using it for coding. I’ve talked about using it
    0:12:44 for doing shorts for some of my videos. So I think we can jump into one of those. I actually
    0:12:49 have chat GPT 01 pro running right now in the background. Cause I knew it was going to take
    0:12:53 a while to like process the transcript. So what I’m going to do right now is I’ll go
    0:13:00 ahead and jump in and we’ll run this through chat GPT 01, see what kind of clips it finds
    0:13:05 for us. And then we can compare it with what 01 pro gave us and see if we can spot any
    0:13:10 differences. Cause I think this whole strategy of letting it find viral clips for us. I don’t
    0:13:14 necessarily think you’re going to need the pro mode. I bet to the regular 01 will probably
    0:13:20 do it just as good. So I’m going to go ahead and share my screen here. So if you are listening
    0:13:27 on audio, you can check out the YouTube version of this and actually see it in action. So
    0:13:33 I’ve got chat GPT 01 open and here’s one of our recent YouTube episodes, AI predictions
    0:13:39 that will completely change life in 2025. And if I go down to the bottom of the description
    0:13:44 here on YouTube, there’s actually a button that says show transcript. So what I like
    0:13:48 to do, and I do this on a lot of the live streams that I do on my YouTube channel is
    0:13:53 once the live streams over, I go and click the show transcript button and it puts these
    0:13:58 transcripts over on the right side of your YouTube window. So we’ve got the entire transcript
    0:14:02 of this recent podcast episode that we did. And I’m going to go ahead and just select
    0:14:08 it all, including the timestamps. You can see I’m selecting the actual times as well
    0:14:13 as the transcripts, because open AI 01 is going to need those, those times as well to
    0:14:16 know, you know, where to tell us to pull those clips from. So I’m going to go ahead
    0:14:20 and copy this whole thing here. And then I’m going to jump into 01. And I’m just going
    0:14:25 to paste this whole thing into 01. So you can see I’ve got the entire transcript loaded
    0:14:29 in here right now. And if I just kind of add a couple of lines, let me get all the way
    0:14:34 up to the top of our transcript here. And I’m going to add a couple of lines here. And
    0:14:40 I’m going to say below is the transcript from a recent podcast episode. Please review it
    0:14:50 and find clips that have the potential to go viral clips should be roughly 60 seconds
    0:14:54 and make or make for good a short form video, right? So I’m just giving it this like little
    0:14:58 prompt up here and then I’m pasting in the entire transcript. And I’ll go ahead and submit
    0:15:03 it. And you can see it’s going to it’s going to take a minute or so to like process this
    0:15:08 whole thing. So I’ll scroll down, you can see it’s thinking right now. So it actually
    0:15:14 responded in less than a minute. So this is regular 01. And it’s actually it’s responding.
    0:15:20 But I forgot to give it a little extra context here. I forgot to tell it to tell me the actual
    0:15:26 timestamps. So you can see right here, it’s giving us the clips is 03 actually a GI and
    0:15:31 then it actually gives me a little transcript section. But what I would typically do is
    0:15:35 let me actually start over real quick. I’m just going to go ahead and copy and paste the
    0:15:38 entire original prompt. And I’m going to do it one more time. So I want to tell it to
    0:15:43 actually give me the timestamps that just makes life easier. I kind of forgot to put
    0:15:49 that in. So let’s go ahead and copy all of this. I’m going to create a new chat here,
    0:15:55 paste the same thing in here. And then at the end of this prompt, say, give me the timestamps
    0:15:59 for each clip. And then we’ll go ahead and run this one more time.
    0:16:02 Yeah, you gotta give it the right context. You know, it’s kind of like a, I think of
    0:16:06 like a in hit checkers, got the galaxy where it’s like, what’s the meaning of life? And
    0:16:11 it comes back, you know, how many years was later 1000 years or whatever it was. And yeah,
    0:16:18 it’s 42. Okay, you know, you got to give it the right context like tell it what you expect
    0:16:19 to get back.
    0:16:24 So 01 is actually quite a bit faster than 01 pro. That’s definitely something I’m noticing
    0:16:29 just kind of comparing them side by side, because I actually ran 01 pro while you were
    0:16:35 talking earlier, just to like, let it start going. This one takes maybe you can see here
    0:16:40 is thought for 38 seconds. And now it’s actually giving me the timestamps. So we got clip one
    0:16:48 is from zero to one minute 2025 will be wild for AI plus 03 IQ levels. And then you can
    0:16:53 see it actually kind of gave us a little transcript of this section is telling us to clip. Next
    0:16:59 one is one minute and two seconds to two minutes and two seconds. This one actually
    0:17:02 feels like it’s kind of going very linear where it’s taking like the first handful of
    0:17:06 minutes and giving us those clips. So you could see clip three is from three minutes
    0:17:10 to four minutes basic AGI might be here already and then it gives us a transcript. But you
    0:17:16 can see it gave us a handful of clips here. And then the final one is from 2759 to 2859.
    0:17:23 The rise of AI email agents for everyone. So these are all potential short form clips
    0:17:29 that we can clip out and then use as YouTube shorts. And this was using the basic 01. And
    0:17:34 again, let’s take a look at the time. You can see it thought about it for 38 seconds.
    0:17:38 So actually pretty quick, but also it’s kind of weird that the very first clip starts at
    0:17:43 the zero seconds and goes to one minute. Like that’s our intro basically, I think there’s
    0:17:47 probably something there where we could give it even more context was like what’s what’s
    0:17:51 a good clip and you know, and what we and like rank order them to don’t just give us
    0:17:56 like in a sequential order, like tell us like what are the top five viral clips from this
    0:17:57 episode.
    0:18:02 That’s actually usually my follow up prompt to this is like give me an order of like which
    0:18:07 one is most likely to least likely go viral. But now if we look over here, so I ran the
    0:18:14 same thing through the 01 pro model. And this time I definitely did tell it to give me the
    0:18:19 time stamps the first time around. But if I scroll down, let’s see if it tells me how
    0:18:25 long it thought for. So this one thought for five minutes and 20 seconds. So, you know,
    0:18:30 about 10 times as long as the last one, but you can see here it’s the time stamps are a
    0:18:41 little bit more dialed in. So 55 seconds to 155 is 03 already AGI from 322 to 422 AGI
    0:18:48 agents and societal shifts by 2025 516 to 616 AI video is about to get wild. It’s actually
    0:18:55 kind of suggesting a lot of the same exact clips that 01 regular gave us. It just seems
    0:19:01 a little bit more accurate on its time stamps. Oh, three’s IQ is near Einstein levels. So
    0:19:06 and then look, we see one much, much deeper in the podcast than what the regular 01 gave
    0:19:13 us, which is from 45 55 to 46 55 one person startups and multimodal mastery. That’s kind
    0:19:19 of hard to say, but that’s how I’ve been using it. And I’ve been plugging this was only 40.
    0:19:24 What was this? A 49 minute podcast episode. I’ve been plugging in transcripts from three
    0:19:29 hour live streams and it’s been finding clips throughout the entire live stream. Like it’s
    0:19:35 finding clips at two hours and 20 minutes in and clips at, you know, two hours and 48
    0:19:40 minutes and 37 seconds and another clip at, you know, three minutes in and just kind of
    0:19:45 all over the podcast, but you’re right. A probably better prompt to use in the future
    0:19:52 would tell it right up front, find these time stamps and then give me a rank order of the
    0:19:56 one most likely to go viral and just put that in the first prompt. I’ve actually been doing
    0:20:00 it as a follow up, but, but I think you’ve got a good point. It’s probably going to actually
    0:20:03 work better if you just include that in the original prompts.
    0:20:06 I believe so. Cause it does, it definitely does with coding. So I assume that probably
    0:20:11 applies to, to everything. It is interesting. Like as of a year ago, I definitely was on
    0:20:15 the side of like, Oh, don’t learn about prompting is not important. Like these models are just
    0:20:19 going to handle all that for you. And even the open AI has kind of said that they think
    0:20:24 that’ll be the case eventually, but it’s like, as of right now, we’ve kind of went to where
    0:20:26 even the prompt is even more important than it was before.
    0:20:32 Well, I think it depends on the use case, right? Like I think for like 95% of your use
    0:20:36 cases for AI, you don’t have to stress too much about the prompts, right? Like if you’re
    0:20:41 using perplexity to get some quick information or you’re asking chat GPT about like, you
    0:20:45 know, one of the things I’ve used chat GPT and Claude and things like that for are like,
    0:20:49 you know, I’m on this medication. I’m thinking about, you know, taking this supplement. Are
    0:20:54 there any like interactions between the two? I don’t really need to write up a complex
    0:20:58 prompt. It’s going to get what I’m asking for, right? So for the most part, I would
    0:21:03 say like 95% of the time, like prompt engineering or getting crazy with your prompts is not
    0:21:09 that necessary. But when you’re asking it, like the, the, the higher the level of complexity
    0:21:14 of what you’re asking for, the more it’s necessary to be more detailed than your prompt. I think
    0:21:18 there is this thing I’m using called repo prompt, but basically what it does is, you
    0:21:22 know, before when I was working on any kind of coding project, like if you want to use
    0:21:27 01 pro, the only way you can get the context for your project is to literally copy and
    0:21:31 paste everything into it. And so what this does is like, you know, you can see here,
    0:21:35 I’ve got all these different files here, tons of files, different directories, all the files
    0:21:41 inside the directories, and it has, it lists them all here and tells you how much context
    0:21:44 it’s currently taken up, how many context tokens. So you kind of know right now it’s
    0:21:50 51, 51.9, right now that’s for anybody listening, it’s like there’s on the screen, there’s like
    0:21:56 a left side that’s got like a folder structure. And that’s like the folder structure of, is
    0:21:59 that like your entire computer? Or is that like just this software?
    0:22:02 That’s a game I’m working on. But the interesting thing is, even though they’re calling it repo
    0:22:06 prompt, you totally could use this for other things. Like you could use this for writing
    0:22:11 or for whatever use case, and just have files with text in them that you want to copy and
    0:22:12 paste every time.
    0:22:15 So if you were trying to like write a book or something, right, you can probably have
    0:22:20 different folders for like each chapter of your book. And then it can like help understand
    0:22:24 the context of previous chapters you’ve already written in the book as well, right?
    0:22:28 Yeah, totally. Or, or you can have that plus you can have a thing on like style. If you’re
    0:22:31 a writer, like what’s the favorite things you’ve written before? Or like, what are the
    0:22:35 things you’d like, right? You can provide all that context as well to make 01 pro to give
    0:22:37 you a better response back.
    0:22:41 Oh, okay. So let me, let me clarify something. So when you were mentioning earlier, when
    0:22:45 you said that you use this for your newsletter and when you’re doing the newsletters, you
    0:22:51 might have like a folder that’s like, here’s the style of newsletters that I really like.
    0:22:56 And those might be text files within a folder called like newsletters I like or something,
    0:23:00 right? And then you might have another folder that’s like my newsletter. And then it has
    0:23:06 a whole bunch of entries in it of your newsletter. And so now when you go to prompt 01 pro, it
    0:23:10 can actually look at all of that. And it’s in the sort of a nice clean structured way,
    0:23:11 right?
    0:23:15 Yeah, with notes, like kind of in the text itself, like saying, here’s your, here’s things
    0:23:19 I like, here’s things I don’t like are usually more detailed than that, but I’m like kind
    0:23:20 of simplifying it.
    0:23:26 Oh, wait, so real quick. So the, the repo prompt, it doesn’t actually do the prompting
    0:23:31 for you. You have to like copy and paste something from repo prompt into GPT. Yeah. Yeah. You’re
    0:23:33 just copying a massive amount of data.
    0:23:38 Oh, okay. I thought it was like tap into chat GPT and like submitting prompts. Okay. Okay.
    0:23:42 No, but let me show you though, why it’s so good for coding. Before when I would give
    0:23:47 a 01 pro something for code, what I was doing was I was creating a script that would look
    0:23:52 at my entire code base. And I would run the script, just a simple node script, and it
    0:23:56 would take all the code and then put it into a single text file and then like put like,
    0:23:59 here’s where this file is just so the, so the model knew where everything was so it
    0:24:03 could help, you know, if it was pointing to a file or whatever. But with this, with XML,
    0:24:07 it’s basically giving it to you in a format where then you can just apply it to your code
    0:24:12 base and it’ll automatically change everything for you, which is incredible. Like it’s, it’s
    0:24:16 so much faster coding with this now. Let me show you the other stuff that’s cooler too
    0:24:20 that I think is not well explained and when you use this tool, like for example, here’s
    0:24:24 a thing where if you’re wanting it to architect something and not actually code it yet, you
    0:24:30 can add this and then you can dive into what that text is. Like it’ll say, you are a senior
    0:24:35 software architect, specializing in co-design implementation planning. Your role is to,
    0:24:38 and then it tells you all the stuff that you’re expected to do.
    0:24:41 It acts like a system prompt right there. That’s essentially a system prompt, right?
    0:24:47 Yes. It acts dramatically different if you do this versus telling it the engineer one
    0:24:52 where you’re a, you know, you’re an engineer. Your job is to execute on the plans and all
    0:24:56 this kind of stuff. And the interesting thing that engineers are starting to discover is
    0:25:03 it seems that it’d be even better if you do both and then say, you’re an architect and
    0:25:07 then first do your architect work. And then when you’re done with your architect work,
    0:25:08 go in and do your engineer work.
    0:25:13 Okay. Yeah. Yeah. So this is, this is like borderline getting into like agentic stuff,
    0:25:15 right? Because it’s almost like,
    0:25:18 It’s better than people realize. Like, and when you, when you, when you tell it, it’s
    0:25:22 different roles and what order to do its roles. It’s so much better. Like what you get back
    0:25:25 from this is so much better. Like if you tell it just to give you code, it’ll give you some
    0:25:29 good code back. But if you tell it to take its time to think as an architect first and
    0:25:33 to really, to map out the feature and how it’s going to interact with all the other
    0:25:36 parts of your code base, which, you know, like I’m saying, this is for code, but this
    0:25:39 could apply to so many different things. And then you can do other things too. Like for
    0:25:43 example, I have just my own, my own rules. I don’t tell stuff works or not, but people
    0:25:48 say like, tell it not to be lazy. Cause sometimes it’ll like try to be lazy and not give you
    0:25:52 all the code. I’ve got, I’ve got my own little notes in here of that kind of stuff, like,
    0:25:56 don’t be lazy and do this. And this is how the kind of responses I like, you know, kind
    0:26:00 of like custom instructions kind of stuff. Right. I’m working on a feature for my game
    0:26:06 to add better special effects to the orbs when they’re matched. And so when I click copy,
    0:26:12 it’s going to give me this giant mega prompt. And, and then I’m going to get a mega prompt
    0:26:17 here. Let me see. And so, you know, I go, I go in chat to BT and you’re going to see
    0:26:22 this prompt is just, is just nuts. I’m using O one pro here. And so you can see here, okay,
    0:26:26 you can see at the bottom, you know, and I’m not sure why it puts it in this order. I’m
    0:26:30 not sure if this is actually better or not. I kind of think maybe it’s not, but it seems
    0:26:35 to work. Okay. It puts literally the instructions at the very end and like puts user instructions,
    0:26:39 you know, is like, I’m working and then it just, whatever I said, I’m working on a feature
    0:26:44 for my game, blah, blah, right. But you can see, like if I start to scroll, you can look
    0:26:50 at the bar over here, that as I’m scrolling, as I’m scrolling, yeah, it’s like the bars
    0:26:54 barely even moving. It’s not even moving. I mean, we’re talking so much information
    0:27:01 here. It is just, it is just wild. How much information I’m sharing here with it. I think
    0:27:05 most people would not realize that you can do this and hand it so much information and
    0:27:10 it understands it. It is just, it is mind blowing that you could pass. I mean, I don’t
    0:27:13 know how many pages this is, but like, God, this is like probably over a thousand pages
    0:27:23 of stuff. Yeah, that’s wild. And then you, and then you press it and, you know, if you’re
    0:27:28 listening to the audio right now, like it just scrolled for like a minute to get to
    0:27:34 the bottom of the prompt. Yeah, yeah, yeah. This is why Sam Altman like tweeted that they’re
    0:27:41 losing money on O1 Pro. It’s people like me. Because I’m literally doing this when I’m
    0:27:47 working. I’m doing this every five or 10 minutes. That’s wild. That’s crazy. And when I’m done,
    0:27:51 I get the response back. It gives me the code. And then my typical workflow, like I said,
    0:27:56 it’ll give me XML back. It’ll give me XML back. And the XML, they have a thing in repo,
    0:28:00 repo prompt now where you basically can actually copy and paste that in. And then it helps merge
    0:28:04 it into your code so you can review all the changes and press, yeah, I’m okay with this
    0:28:11 bit, this bit, not sure about, or if you’re lazy, just press accept it all. And using XML,
    0:28:15 it actually works all that out for you because it tags the files in the directories and it
    0:28:20 knows exactly where they are and exactly what code was changed. And you press a button.
    0:28:24 It’s all done. Oh, wow. It’s mind blowing. But you see here, it’s like, okay, green check
    0:28:27 market shows you the different files that are going to be, you know, like here’s a turn
    0:28:32 manager and here’s like the enemy UI and stuff like that. Right. So it shows you the, and
    0:28:37 if you click merge changes, you would then see the actual code in the different ones.
    0:28:41 And then line by line, you can accept or deny the code. And you, if you wanted to just accept
    0:28:46 it all, you can. And sometimes I do that. Sometimes I just like, major change back up
    0:28:51 my prop, you know, project back it up and then accept all. Yeah, that’s like kind of the
    0:28:56 best workflow for engineering right now using AI, in my opinion, is to use O and pro for
    0:29:00 any new feature, anything that’s like a new thing that doesn’t currently exist in your
    0:29:05 product or your website or whatever. Use O and pro because it’s way better to figure
    0:29:09 out how to properly architect it and make sure it all performs well and everything else.
    0:29:12 And thinking about thinking through every possible scenario where something could go wrong, it’s
    0:29:16 way better at that. And then once you’ve got the, once you’ve got the feature implemented,
    0:29:20 small changes, you totally can just use cursor or something like that. Like changing the color
    0:29:26 of a button, you’re like changing the name or anything super small, you know, obviously
    0:29:29 you could do it yourself. But if you, if you want to use AI, you could use cursor or something
    0:29:33 like that. Yeah. Yeah. So you wouldn’t use this just to make like small bug fixes or
    0:29:38 small tweaks. You’d kind of more use it to build like the overall like bones of the product.
    0:29:41 Right. Yeah. If I’m building a software application
    0:29:44 and there’s a new feature, let’s kind of figure out what that feature looks like. It’s not
    0:29:50 going to look good design wise. Like none of these are good at design yet. But in terms
    0:29:55 of it actually working, like O and pro like often gets it right the first time. And if
    0:29:59 it doesn’t, it’s usually very minor bugs. I suggest like anyone who has a company who
    0:30:03 has engineers right now, like you’re like really missing out if you’re not like paying
    0:30:08 for like O one pro for your engineering team and having them use something like repo prompt.
    0:30:12 Right. Right. Missing out. That’s super cool. Yeah. Now, is there something you can show
    0:30:16 us of like what you’ve generated? Yeah. So this is the Godot editor, which is
    0:30:21 like a open source game engine, kind of like final fantasy style. Right. But I mean, it
    0:30:25 still looks really, really good. If anybody’s just like listening on audio, it’s a very
    0:30:31 like colorful visual game that you’ve built here. And AI helped me make all of that. And
    0:30:34 so like the background right now, there’s like, I don’t know, like a cathedral looking
    0:30:39 thing, but it’s got like a wavy animation. Yeah. I’m sitting there thinking like, okay,
    0:30:42 this is something where I really didn’t even think I was going to do it. I may, who knows,
    0:30:45 I may not do it eventually, but it’s, I thought it’d be great for the show for me to be really
    0:30:50 hands on with all the different parts of AI from like, from I’m using it for writing.
    0:30:54 I’m using it for coding. I’m using it for the art. I’m even using videos. I’m trying
    0:30:57 to think about like, okay, in a year from now, what’s AI going to be really good at?
    0:31:01 Yeah. Yeah. And so if this takes me a year to build as a hobby, it’ll only get better
    0:31:06 from here. Cause like, oh, three will come out. The AI video models would get better.
    0:31:09 The AI art is going to get better. And so that’s kind of what I’ve been doing is like
    0:31:13 seeing what’s currently possible, but then trying to set up in a way where as those things
    0:31:16 get better, I possibly could actually turn this into a real game. But yeah, I’ve just
    0:31:20 been shocked by how good, I mean, like AI can do all of this, like 01 pro, especially
    0:31:24 like 01 was not able to do this, by the way. Like I’ve tried 01 for like hard coding stuff
    0:31:29 and it just, it fails a lot more often than 01 pro. So that tells me that once we get
    0:31:34 like 03 and once you get 03 plus, you know, apparently the next version of mid journey,
    0:31:37 they’re saying that it’s going to be way better at being consistent with characters and things
    0:31:41 like that. Apparently that’s the big next thing coming. And then you get AI video better
    0:31:45 and so you can have cool cut scenes and stuff like that. You’ll be able to make like amazing
    0:31:46 experiences entirely with AI.
    0:31:50 Well, yeah. And if we can get like consistent characters in video, I know there’s some tools
    0:31:55 out there that claim they can do that too, but it’s still a little wonky, but I mean,
    0:31:59 by the end of this year, we’ll be able to have consistent characters in images really,
    0:32:03 really good, probably pull those characters into videos and have consistent characters
    0:32:08 in videos. Yeah, 03 model will be out at some point, which is going to like really, really
    0:32:12 improve the code that you’re able to do. Right. It’s also going to really improve the writing
    0:32:16 and the any sort of like storytelling elements in there. And it’s like, yeah, we’re, we’re
    0:32:20 kind of running out of stuff for the humans to actually do when it comes to making these
    0:32:24 games, but, but, but as a creator, you still get to be the, I mean, like I’m piloting all
    0:32:29 of this. I mean, like for me, this is like, like so fun to like think that maybe it will
    0:32:32 be the thing of the future where like it’s so easy to make these games.
    0:32:36 It’s like in the past, you have the different coders that specialize at different things.
    0:32:40 And then you have like the override, the project manager who’s kind of like telling them each
    0:32:43 what to do. And then I always like, I don’t know, for whatever reason, I always use the
    0:32:46 like symphony analogy of like, now you’re going to become the conductor where you’re
    0:32:50 just sort of like telling all the instruments what to do, but you’re standing there conducting
    0:32:53 them. Right. That’s where it’s going, you know, different people are going to be able
    0:32:57 to use this to like make their dreams come true. Cause like when I was a kid, my dream
    0:33:01 was to like work at Blizzard. Like I want to be like one of the top people at Blizzard.
    0:33:04 And then the weirdest thing happened where I was one of the top players in the game ever
    0:33:10 quest when I was a kid, the number, the number one player, Rob Prado, who used to run Blizzard,
    0:33:13 at that time he was on EverQuest and he was running Legacy of Steel, the top guild on
    0:33:18 EverQuest. I ended up like raising money for a startup. We raised several million for
    0:33:23 a startup called GameStreamer. And the combination of having that gaming background and having
    0:33:27 that startup at E3 had like a huge corner at E3. I ended up getting to like hang out
    0:33:30 with Rob and get to know him very well. And a lot of top people in the game industry.
    0:33:35 So I had this weird situation where I never really got to fulfill my dream. I was hanging
    0:33:40 out with all those people as like good friends. I got to see that, you know, it wasn’t really
    0:33:43 the life I wanted, like going to work for one of those companies. It was not. I wanted
    0:33:46 to do my own things. But then I still never got to do what I wanted in terms of making
    0:33:51 a game. And it’s so wild to me that now that AI is getting so good, that a lot of people
    0:33:54 are probably going to have the same kind of like, you know, awakening that I’m having
    0:33:58 or it’s like, I can do those things now. You know, it doesn’t matter if I’m 40. I can still
    0:34:02 do it because AI is getting so good that I can, it doesn’t take as much time as it used
    0:34:08 to. And that’s going to get, it’s going to get even better. Like when 03 comes out, you’ll
    0:34:13 probably be able to do any new feature without any bugs, you know, one shot. Yeah. One shot.
    0:34:17 It’s already close to one shot in many things now. And it’ll only become more so. And so
    0:34:21 that’s just, it’s exciting. Like for a lot of people who like creating things, like the
    0:34:25 next 10 years is going to be like a revolution in terms of creating things, not only art,
    0:34:29 but like, you know, even companies, like if you wanted to create a company now, definitely,
    0:34:34 it’s going to be easier. Yeah. 100%. No, it’s, it’s super, super exciting. And you know,
    0:34:38 I, I feel the same way. Like I, I’ve, one of my things when I was a kid was I always
    0:34:43 wanted to be like a game designer, a game developer, like work in the gaming space.
    0:34:49 And I feel like now we kind of have like a way to sort of live that childhood fantasy
    0:34:54 a little bit, but without all the negatives. Right. Right. Right. Yeah. I mean, it’s like
    0:34:57 me and you had even talked about it. And I was like, I’m just going to start playing
    0:35:00 with stuff and just see what’s possible. And I was just after a week, I was like, wow,
    0:35:05 it’s actually possible. Like it can, it can do the whole thing. And then just like learning
    0:35:09 that you can do all of this and you can control most of it with your voice. It’s just been,
    0:35:12 I’m hoping that people will listen to this podcast and kind of like think bigger about
    0:35:14 what they could be accomplishing with these tools.
    0:35:19 I couldn’t say it better myself. And so with that being said, I’m not going to try to say
    0:35:23 it better myself. We’ll just go ahead and wrap this one up. It’s a really, really exciting
    0:35:28 time right now. You can pretty much do build anything you can imagine. And it’s only getting
    0:35:32 better and easier. And we’re going to keep on exploring and diving deeper into these
    0:35:36 rabbit holes to figure out what we can build. And as we learn, we’re going to share with
    0:35:41 you. So make sure if you’re not already subscribed to the shows, subscribe on YouTube, subscribe
    0:35:46 wherever you listen to your podcast, you can find us at all of those places. And thank
    0:35:49 you so much for tuning in. Hopefully we’ll see you in the next one. Thank you.
    0:35:50 .
    0:35:57 [inaudible]
    0:36:01 [inaudible]
    0:36:05 [inaudible]

    Episode 42: Are you truly unlocking the full potential of OpenAI’s 01 models? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) dive deep into the capabilities of ChatGPT01 and GPT01 Pro, offering insights to ensure you’re not overlooking these powerful tools.

    In this episode, Matt showcases how to create short-form content from long-form transcripts, while Nathan discusses using 01Pro to build a game from scratch. With specific workflows, practical examples, and mind-blowing insights, you won’t want to miss how these advanced models can revolutionize your content creation and coding endeavors.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Model Perception Disconnect
    • (05:50) Understanding ChatGPT and 01 Models
    • (07:58) AI Transforming Complex Problem Solving
    • (10:19) AI-Assisted Medical Diagnoses
    • (15:37) “01 Outpaces 01 Pro”
    • (18:35) Efficient Podcast Clip Identification
    • (22:57) “Efficient Coding with XML Format”
    • (24:34) Optimize Instructions for Better Output
    • (28:07) AI Workflow Optimization: O1 Pro & Cursor
    • (32:17) Unexpected Gaming Industry Connections
    • (32:56) AI Empowering Creative Pursuits

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • The Biggest Ai Announcements From CES 2025

    AI transcript
    0:00:04 One of the things at CES was last night in Jensen gave his big keynote,
    0:00:06 AKA Tech Taylor Swift, right?
    0:00:10 Like, you know, the second they announced him and like he was walking out on this
    0:00:14 stage, people were like standing and cheering and screaming.
    0:00:17 In my opinion, the bigger news and the things with the bigger
    0:00:22 implications of the AI world are project digits or just digits.
    0:00:26 You’re going to have a model that’s good enough to do anything you want.
    0:00:28 So you’re saying that basically NVIDIA is going to be the big player and not open
    0:00:35 AI? Hey, welcome to the next wave podcast on Matt Wolf.
    0:00:42 I’m here with Nathan Lanz, and I am reporting into this episode from CES in Las Vegas.
    0:00:47 You know, in this episode, we’re going to deep dive into the future of AI as we’re
    0:00:51 sort of seeing it unfold here at CES this year.
    0:00:55 You know, we hear a lot about these companies like Open AI and Anthropic and
    0:00:58 Google and all of the big, massive companies building AI.
    0:01:03 But is the big winner, the company that everybody’s really going to remember when
    0:01:04 it comes to AI in the future?
    0:01:06 Is it really going to be NVIDIA?
    0:01:10 Well, that’s what we’re going to discuss in this episode because I’m spending a
    0:01:13 ton of time with NVIDIA out here learning what they’re up to.
    0:01:16 They just put on a massive keynote where they drop some huge announcements.
    0:01:20 We’re going to get into all of that as well as some of the coolest tech I’ve
    0:01:24 come across at CES and my overall thoughts on CES.
    0:01:26 And Nathan and I, we’re just going to nerd out on this one with you.
    0:01:28 So let’s just go ahead and dive right in.
    0:01:34 I think maybe where we should start is probably the whole Sam Altman thing,
    0:01:38 because that kind of came out like right before I got to CES, right?
    0:01:42 He put out a blog post and essentially said, we know how to get to AGI.
    0:01:43 We’ve got a path to it.
    0:01:45 We know we’re going to hit AGI.
    0:01:47 It’s going to be 2025 that we hit AGI.
    0:01:49 Our new mission is ASI.
    0:01:53 It basically makes it sound like AGI is no longer a big deal.
    0:01:56 Like we definitely know how to reach AGI.
    0:01:56 Yeah.
    0:02:00 Yeah, just the tone of the letter was like super confidence.
    0:02:01 Yeah, right?
    0:02:05 Like the next big deal is ASI, you know, artificial super intelligence.
    0:02:09 Yeah, with Sam Altman though, I feel like the last like year,
    0:02:13 he’s sort of been on this campaign to like damper expectations, right?
    0:02:16 Like he’s been kind of going around saying AGI is going to come.
    0:02:19 It’s going to pass and it’s not going to change anybody’s life.
    0:02:22 As big as like they think it’s going to change their lives, right?
    0:02:25 So I feel like he’s kind of doing that sort of thing of like, we’re going to hit AGI,
    0:02:27 but it’s not as big of a deal as you guys think it is.
    0:02:30 That’s kind of the feeling I get from Sam lately.
    0:02:32 Maybe, yeah, he did say some stuff too.
    0:02:35 That I thought when I had said previously where we were talking about like, you know,
    0:02:37 is open AI in a bad spot or not?
    0:02:39 I mean, you had one episode where we were kind of arguing back and forth on that.
    0:02:42 And I was kind of saying like from a Silicon Valley perspective,
    0:02:46 like when you do a startup, there’s always issues between co-founders.
    0:02:49 And typically the more successful a company is and the faster things move,
    0:02:53 actually, the more issues because just there’s such a divergence of what people want.
    0:02:57 And also just egos get bigger and bigger and then the egos clash.
    0:03:01 His letter says that it’s been like the best year of my life
    0:03:06 and the worst year of my entire life, which is interesting.
    0:03:08 Because last year was the year he got booted from open AI.
    0:03:09 That wasn’t this year.
    0:03:11 And he did talk about that in startups.
    0:03:14 You know, he was a white comedy for a long time.
    0:03:15 Startups that grow fast.
    0:03:16 They always have team issues.
    0:03:20 And typically the faster a startup grows, the more intense those are.
    0:03:23 And he said, this is possibly the best example of that possible
    0:03:26 is that they’ve been growing faster than almost any company in history.
    0:03:29 Right. And so, of course, there’s huge mistakes they’ve made.
    0:03:31 And, you know, he didn’t say egos, but, you know, obviously egos got big.
    0:03:33 And he mentioned he made mistakes as well.
    0:03:37 But he did kind of like throw the people under the bus who had kicked him out, though.
    0:03:41 He did make it clear that he thinks that they totally screwed him over
    0:03:43 and screwed up by doing what they did.
    0:03:47 Well, then now you have this year where he’s been battling with Elon, right?
    0:03:51 So that’s like the new thing that he’s dealing with is Elon.
    0:03:55 And now they’ve got, you know, Jeffrey Hinton, who’s on board trying to
    0:03:58 also stop the for-profit, right?
    0:04:03 That’s another piece of news that came out recently is that they are going to try
    0:04:06 to switch to a for-profit from a nonprofit.
    0:04:09 They’re trying to register it as like a Delaware corporation.
    0:04:11 The non-profit still going to exist.
    0:04:15 It’s just going to hold a ton of equity in the for-profit, right?
    0:04:16 But you’ve got Elon Musk.
    0:04:17 You’ve got Metta.
    0:04:22 You’ve got Jeffrey Hinton, who, you know, is sort of like the godfather of AI.
    0:04:25 All of these people pushing back, saying we should not let that company
    0:04:27 switch to a for-profit.
    0:04:30 He mentioned that whole situation as well, without saying names and talking about.
    0:04:34 It’s natural when you’re growing this fast and such an important moment in history
    0:04:38 that people, especially people who might be competitive with you,
    0:04:42 might be trying to throw up roadblocks and cause issues.
    0:04:44 So he kind of alludes to that without saying the names.
    0:04:47 Yeah, he also talks about he believes that in 2025,
    0:04:52 we are going to have agents that do meaningful work in organizations.
    0:04:52 Right.
    0:04:56 And he thinks that’s going to be a huge turning point with AGI and for AI in general.
    0:04:57 And we’ve been hearing rumors of that,
    0:05:01 but it sounds like they must have some kind of examples of that actually working
    0:05:04 over like agents going off and doing meaningful work for people.
    0:05:06 And so that’s going to be exciting to see that.
    0:05:10 And he talked about ASI, you know, like, hey, now that we know AGI is like a given.
    0:05:12 We definitely know how to do that.
    0:05:13 We need to be aiming for ASI.
    0:05:16 He didn’t say that they know completely how to get there,
    0:05:20 but his confidence of getting there seemed way higher than when he’s talked in the past.
    0:05:21 Right, right.
    0:05:23 Which was people on Twitter were like, holy crap.
    0:05:26 So like, obviously this test time compute stuff with like what?
    0:05:29 01 going to 03 and how fast that’s improved.
    0:05:35 They must be seeing signals internally that there’s no end in sight to those improvements.
    0:05:37 That’s probably what’s going on.
    0:05:40 Well, it sounds like they know how to get to AGI right now.
    0:05:44 It’s probably mostly like an energy cost issue, right?
    0:05:50 Like we saw with 03, you know, it got these high benchmarks on all of these tests.
    0:05:53 But to do it, they had to spend thousands of dollars per task.
    0:05:58 So I think like the real barrier right now is like, OK, maybe they’ve got AGI internally,
    0:06:05 but it costs $10,000 per task to do it, which is not feasible for like any individual
    0:06:08 and maybe barely any businesses, right?
    0:06:12 So it’s like, I think they do actually know how to get to AGI.
    0:06:15 I think AGI is probably more on the inference side, right?
    0:06:19 Where maybe they have some more trainings in them with like a GPT five or something like that.
    0:06:25 But it’s really that ability to like think and double check and triple check and quadruple
    0:06:29 check its work and then give you its answer on the inference phase,
    0:06:31 which is going to get us to AGI.
    0:06:33 That’s sort of the sense that I’m getting from it all.
    0:06:37 Yeah. And with ASI, he’s talking about like, you know, solving cancer and like
    0:06:41 environmental issues and just any kind of problem humanity has,
    0:06:45 having almost like Godlike intelligence that can help solve those problems.
    0:06:50 So I thought it was exciting that now he’s talking about in terms of like business
    0:06:55 use cases and AGI being useful for companies like that’s a given for them now.
    0:06:56 AGI is going to be useful.
    0:06:58 It’s going to be doing meaningful work for people.
    0:07:03 And their main focus is moving on to ASI of how can you now
    0:07:05 understand the world around us better?
    0:07:07 How can you understand physics better?
    0:07:11 How can you invent new things, learn or discover new science?
    0:07:14 And so that’s just exciting that like as a company, it’s just like, we know AGI.
    0:07:15 It’s like an energy thing.
    0:07:17 It’s a little bit more improvements to the algorithm.
    0:07:18 How do we do ASI?
    0:07:20 And we think we might know that’s just exciting.
    0:07:23 Like the confidence level is so much higher than the previous letters.
    0:07:26 To me, it seems like if they can figure out AGI,
    0:07:30 ASI doesn’t feel too far off after that, right?
    0:07:36 Because like now you have the AGI who’s helping you figure out how to get to ASI.
    0:07:38 Right? It’s like at that point, once you have an AGI,
    0:07:42 you don’t necessarily need a whole bunch of humans figuring out
    0:07:45 how we get to the next phase beyond AGI.
    0:07:49 You have an AGI that helps you figure out how do we get to the next phase beyond AGI.
    0:07:52 Right. I mean, I would be surprised if internally,
    0:07:57 they’re not already using O3 and things like that to help them make progress faster.
    0:08:00 I mean, you know, since the last video we did where I showed you my game demo,
    0:08:05 I’ve been using O1 Pro more and more and I’m just blown away by like what it can do.
    0:08:09 Yeah. I wonder what’s going to happen with a lot of the other models, right?
    0:08:14 Because it seems like Anthropic and Gemini and all these other tools that are out there,
    0:08:18 you know, OpenAI has kind of showed their cards now, right?
    0:08:21 So don’t you think that like maybe the next phase of Claude
    0:08:26 is going to have more of that like logic and double checking itself?
    0:08:28 I think that’s the next phase for all of them.
    0:08:32 Well, I think the big question is, did OpenAI come up with some kind of secret sauce
    0:08:36 that people are not aware of Google and other people are kind of saying
    0:08:37 that they understand how to do it.
    0:08:40 The open source community is saying they understand how to do what OpenAI is doing.
    0:08:45 OpenAI has repeatedly kind of hint that, yeah, maybe you do, maybe you don’t.
    0:08:49 We have come up with some interesting techniques that have made this possible.
    0:08:51 It’s not just super straightforward.
    0:08:54 Is that just messaging or do they really have some kind of secret sauce
    0:08:56 where, yeah, the other people think they know how to do it,
    0:08:59 but there is something better about the way OpenAI is doing it
    0:09:01 that will take the other ones time to discover.
    0:09:03 Yeah, I’d say it’s probably a blend.
    0:09:05 It’s probably like sort of in between the two.
    0:09:06 Right. Yeah.
    0:09:08 So like I mentioned, I’m at CES right now.
    0:09:11 So if you’re watching this on the video that explains why I’m in a hotel room.
    0:09:15 But one of the things at CES was last night, Jensen gave his big keynote,
    0:09:18 AKA Tech Taylor Swift, right?
    0:09:22 Like everybody like fan girls out whenever Jensen Huang’s in the room.
    0:09:26 And, you know, the second they announced him and like he was walking out on this stage,
    0:09:31 people were like standing and cheering and screaming and, you know, flashing them.
    0:09:33 And, you know, no, not that flash.
    0:09:35 Remember, it happened that right?
    0:09:38 There are those pictures of Jensen signing a woman’s chest.
    0:09:40 So he’s a rock star of our generation.
    0:09:41 Like we don’t have many of them.
    0:09:43 So, you know, he’s one of them.
    0:09:44 Yeah, yeah, yeah, exactly.
    0:09:49 But he came out, you know, the big thing that everybody at CES was
    0:09:52 I think the most excited about because it’s the consumer electronic show
    0:09:57 was probably the fact that we got these NVIDIA RTX 50 series, right?
    0:10:05 He said that the NVIDIA 5070 is just as powerful as the 4090s that you get now.
    0:10:08 The 4090s cost 1,600 bucks right now.
    0:10:15 These new 5070s cost 550 bucks and are just as powerful as the 4090s.
    0:10:19 Supposedly, there’s some controversy around that as well.
    0:10:22 People are saying it’s like close to being as powerful.
    0:10:26 But then there’s some like upscaling that happens on the GPU
    0:10:27 that the old ones don’t have.
    0:10:31 So it’s not quite as good and there’s some additional latency issues.
    0:10:34 I don’t know all of like the technical issues or anything yet.
    0:10:38 But a lot of people are going, he’s claiming it’s the same as the 4090s.
    0:10:40 But if you look at the specs, it’s not technically the same.
    0:10:45 But anyway, that’s what most of the people I think we’re here to hear about.
    0:10:47 That got like huge cheers.
    0:10:48 Everybody was excited when they talked about that.
    0:10:52 In my opinion, the bigger news and the things with the bigger
    0:10:58 implications of the AI world are more about the project Cosmos, the digits.
    0:11:00 So digits, that’s the personal computer.
    0:11:01 Is that correct?
    0:11:03 That’s the personal like super computer.
    0:11:07 OK, and I feel like that is kind of a big deal, right?
    0:11:12 Like that is something that over time, I wonder if that’s going to affect
    0:11:16 open AIs and Anthropics and Google’s business models, right?
    0:11:20 Because pretty soon, anybody can have a computer that’s like just as
    0:11:23 powerful as these cloud computers that a lot of these other companies are
    0:11:25 running with sitting on their desktop.
    0:11:28 And, you know, if you’ve got enough money, you can stack them together.
    0:11:31 And right now they cost three grand each, you know, whenever they’re released.
    0:11:32 I don’t think they’re out yet.
    0:11:34 But when they’re released, they’re going to be about three grand each.
    0:11:39 And Nvidia has their own version of like llama that they fine tuned and
    0:11:42 optimized to work with their own computers.
    0:11:46 But it makes me wonder, like, is the future of all of this stuff going
    0:11:47 to be like the edge computing?
    0:11:48 Is it going to skip the cloud?
    0:11:51 Are people going to even want to use things like open AI when they can
    0:11:53 just have their own box?
    0:11:55 And right now, yes, they’re three grand.
    0:11:58 But, you know, what are they going to be two years from now?
    0:11:59 Look at what they did with the RTX series.
    0:12:02 The 40 90 is sixteen hundred dollars to this day.
    0:12:07 The 50 70 is now five hundred and fifty dollars and supposedly just as powerful.
    0:12:11 So if these things are three grand right now, maybe next year, they’re fifteen
    0:12:14 hundred bucks, maybe the year after that, you have a personal super computer
    0:12:18 in your house that’s seven hundred bucks that runs models as powerful as
    0:12:19 ChatGPT. I don’t know.
    0:12:24 So I wonder, like, how does that impact a lot of these big AI companies?
    0:12:25 Yeah, I don’t know.
    0:12:28 You know, you talked about this before about in the future, are people going
    0:12:31 to be using something like ChatGPT or they’re going to have the models
    0:12:32 just running in their own house?
    0:12:33 Yeah, it’s a good question.
    0:12:35 Like, I haven’t really thought enough about that.
    0:12:38 It does feel like that open AI will always have the best models.
    0:12:42 But yeah, you might have for like daily usage that the models get so good
    0:12:45 and that the compute gets so cheap that, yeah, you could have your own private
    0:12:48 one in your house for daily stuff because you really don’t want, you know,
    0:12:52 the other companies just knowing like, oh, I’m talking to the AI all day.
    0:12:54 We’re like, yeah, they can hear every single thing I’m saying.
    0:12:58 You know, it’s like to use these little like super computers.
    0:13:00 You don’t even need to be connected to the internet, right?
    0:13:03 You could be like completely offline and just wired into it.
    0:13:05 And it will run all the inference on there.
    0:13:08 You know, maybe someday you’ll even be able to like train your own
    0:13:10 AI’s and fine tune on these things as well.
    0:13:13 So like, yeah, I don’t know.
    0:13:15 I feel like that could be more of the future.
    0:13:19 I mean, I do think open AI and, you know, maybe anthropic and Google,
    0:13:22 they will always be the more state of the art, the more cutting edge,
    0:13:26 the furthest along further than what you could run at your house.
    0:13:31 But at some point you get to this level where the models could
    0:13:33 enough for everything you want it to do.
    0:13:37 So do you really need the most cutting edge state of the art model, right?
    0:13:41 Unless you’re like using it for like enterprise, unless you have like
    0:13:44 really, really massive needs where you really want open AI or Google
    0:13:47 or anthropic or one of these companies sort of managing it
    0:13:49 and making sure you’re always still the art.
    0:13:52 I feel like we’re going to get to a point where you’re going to have
    0:13:55 a model that’s good enough to do anything you want.
    0:13:58 Run your own agents, pretty much anything you need to do.
    0:14:01 You can do on this computer at your house
    0:14:03 without actually connecting to one of these cloud companies.
    0:14:06 So you’re saying that basically NVIDIA is going to be the big player
    0:14:09 and not open AI. That is interesting.
    0:14:13 I could imagine a future where AI voice like talking to the AI
    0:14:16 and being able to interact with your calendar and your email system
    0:14:17 and all this kind of stuff.
    0:14:19 That’s going to be very simple, very soon.
    0:14:23 And so I can imagine you would want all that to be on like a local machine
    0:14:25 that you’re running and then you probably could even have the AI help
    0:14:29 kind of manage your privacy in terms of like, OK, I want to keep this
    0:14:33 information private, but I’m OK with you interacting with chat to BT.
    0:14:37 And I want you to help kind of manage what I’m willing to share with it not.
    0:14:40 So if you’re going like private, personal, you know, AI interacting
    0:14:43 with the more powerful companies like chat to BT, I can imagine that.
    0:14:45 Like that makes a lot of sense to me.
    0:14:48 Yeah, I mean, I do think we’ll get to a point where like the AIs sort of
    0:14:51 interact with other AIs. I think that’s kind of inevitable, right?
    0:14:56 So it’s like maybe your own personal computer, like in the same way,
    0:15:01 if you use the new version of Siri that’s on the iPhone 16 plus or pro or whatever.
    0:15:07 Right. Like if you use Siri, it’s got its own like edge AI on the phone
    0:15:12 where it can answer questions and do some really basic responding right on the phone.
    0:15:17 But if it’s a little bit more complex of a question, it’ll say, can we ask open AI?
    0:15:20 Right. So I think it’ll probably be that type of thing.
    0:15:22 Yeah. But yeah, I think it’s just interesting to think about, like,
    0:15:28 do a lot of these big AI companies stay really like consumer tools
    0:15:33 far into the future, or does the consumer stuff move more to these like the hardware
    0:15:35 that you own at home? I don’t know.
    0:15:40 I feel like Nvidia presented a future where maybe it is more like edge compute
    0:15:42 like at home, you have your own devices.
    0:15:44 It’s like the question will people ever care about that?
    0:15:46 It’s like, yeah, it’s kind of like the thing with like Bitcoin.
    0:15:47 Like there’s so many things that Bitcoin is better at,
    0:15:49 but there’s also a lot of user experience issues.
    0:15:52 And like if people are just like used to going on chat to BT and they don’t have
    0:15:56 to buy all the hardware and it’s just, you know, if they can solve that.
    0:15:59 Yeah, yeah. I mean, everybody’s going to prioritize different things, right?
    0:16:02 Like right now, big companies, if they wanted to,
    0:16:07 they can build their own server farms and run everything locally if they wanted to,
    0:16:10 or they can cut costs and run it on the cloud.
    0:16:13 And if you’re running it on the cloud, there are some inherent risks, right?
    0:16:17 If you lose internet connectivity, do all of your systems go down for the day
    0:16:19 because you can’t access the cloud?
    0:16:25 Do you know exactly what information your cloud provider is able to see on your stuff?
    0:16:27 Right? There are like issues with the cloud.
    0:16:31 So it really depends on like what the company or the people are going to prioritize.
    0:16:36 You know, us being like techie AI type people will probably always continue
    0:16:40 to sort of be on the bleeding edge and try it all and, you know,
    0:16:43 use both versions of it to see what can do what.
    0:16:47 Yeah. I mean, on my use case right now, on how I’ve been using AI recently,
    0:16:49 like it kind of fits with the whole idea of having your own
    0:16:53 private for like just like conversational stuff and then for like work,
    0:16:56 have the strongest models because like right now, like what I’ve been recently,
    0:16:59 when I have been like, you know, getting really hands on with AI
    0:17:02 by like developing this game demo, what I’ll do is I’ll be having
    0:17:06 oh, one pro generate code like for like new features and things like that.
    0:17:08 And sometimes it takes it like five to seven minutes.
    0:17:11 And so while it’s doing that, I’m often going to Claude
    0:17:13 and then having conversations with it.
    0:17:14 You know, I use voice.
    0:17:17 Like I have a button that like transcribes what I say in voice to text.
    0:17:20 I’ll be sitting there talking and then just like going back and forth
    0:17:22 on different ideas and updating that specs document while I’m waiting
    0:17:24 for the code to finish.
    0:17:26 Right. And so I can definitely imagine like that in the future,
    0:17:29 where like for like daily use, you could have your own private computer.
    0:17:31 You’re chatting with it, but then when you want to do some hard work,
    0:17:35 maybe you send it off to the super ASI, which may take a minute,
    0:17:38 may take an hour, you know, if it’s something super complicated.
    0:17:40 And so I can see that kind of world.
    0:17:42 Yeah. And I think if you’re building things like spec sheets
    0:17:45 and you say like, this is how it should perform.
    0:17:47 And this is what this screen should have on it.
    0:17:49 And this is what this screen should have.
    0:17:51 And when you press this button, this is how it should react.
    0:17:53 And you have all of that kind of stuff.
    0:17:58 I imagine these AIs are going to be able to test to make sure everything works.
    0:18:01 And it’s like own virtual environment, which, you know, sort of brings me to
    0:18:05 like the other big thing that NVIDIA talked about, which was their project Cosmos.
    0:18:09 Wasn’t, you know, really the exact same thing we’re talking about here.
    0:18:12 But they were talking about this project Cosmos as like the, you know,
    0:18:16 the digital twin concept, the sort of virtual world where you can put
    0:18:21 self-driving cars or you can put robots or whatever inside of these
    0:18:25 virtual worlds, test them in the virtual worlds, train them in the virtual worlds.
    0:18:28 And then once it’s working perfectly in these virtual worlds,
    0:18:31 then you deploy them into the real world, right?
    0:18:33 And they’ve already sort of announced that.
    0:18:37 I was at NVIDIA GTC last year and they were already talking about that
    0:18:39 as like part of their project, Groot and stuff.
    0:18:42 But I think the new thing about Cosmos, this was one thing I was a little
    0:18:46 unclear about. I was talking to somebody after the keynote and we were kind of
    0:18:48 confused about like this whole digital twin thing.
    0:18:52 Jensen’s been talking about that for the last like, you know, year and a half,
    0:18:57 two years now, like how is this Cosmos thing that they announced this week
    0:19:02 different? And the major difference that I noticed was that he was actually
    0:19:06 entering prompts, just like you’d enter a chat GPT prompt or just like you to
    0:19:11 enter a prompt in like stable diffusion or mid journey to give it scenarios
    0:19:16 to try, right? So he was, let’s say, for example, putting a self-driving car
    0:19:21 inside of this virtual environment and then he was prompting things like, oh,
    0:19:27 there’s a snowstorm and a ball rolls into the street and what should the car do?
    0:19:30 Run every possible scenario that the car can do.
    0:19:33 And let’s figure out the optimal scenario, right?
    0:19:37 They were actually typing prompts into the prompt box to generate different
    0:19:40 scenarios. And to me, that’s really, really interesting.
    0:19:43 It’s like, there is a black mirror episode that kind of does this with like
    0:19:48 the whole like dating scenario where it plays out every possible scenario and
    0:19:51 then finds like a mismatch for somebody, right?
    0:19:56 That’s essentially what this Cosmos can do in like a self-driving car environment
    0:19:59 or robots or, you know, whatever you want to test it with, that’s what it can do.
    0:20:05 You build this environment, it plays out every possible scenario it can imagine
    0:20:10 and then gives you the optimal solution for, you know, whatever scenario you
    0:20:14 enter, which is pretty mind blowing to think about, but that’s kind of what
    0:20:18 Project Cosmos is doing and they’re open sourcing it.
    0:20:21 They literally put it up on GitHub so any companies can go and use it.
    0:20:24 So that’s pretty wild as well.
    0:20:25 That is wild.
    0:20:29 Like I imagine you’re using Claude or Chatsp or whatever to help
    0:20:32 generate the prompts, even like, what are all the things we should test?
    0:20:34 It can help come up with all those prompts.
    0:20:36 And so you could have AI actually coming up with all the scenarios to come up
    0:20:38 with, you know, a million scenarios to test.
    0:20:42 Yeah. Yeah. And I mean, coming back to like the game example that you’re doing,
    0:20:47 it’s not necessarily like the same sort of virtual environment, but I would
    0:20:52 imagine it could go and play through your game, run all possible scenarios
    0:20:56 and find any potential issues with the game that might need to be fixed.
    0:21:00 Right. Think about like a long, like, let’s say like a platform or game
    0:21:03 or something that would typically take somebody 10 hours to beat, you know,
    0:21:07 what happens if there’s a bug in the game that’s like nine hours in that maybe
    0:21:12 you missed because you didn’t hit a certain part of the level and you never
    0:21:16 spotted that bug. Well, is this going to be able to like preempt any potential
    0:21:18 bugs that might come out in a game?
    0:21:21 Because it can run through every potential scenario of that.
    0:21:23 I think you’ll get there.
    0:21:25 I was actually talking to my wife about this the other day.
    0:21:28 I was like, you know, play testing for games is going to get really easy soon.
    0:21:31 For any software, it’s going to get really easy soon because you’ll be
    0:21:34 able to have like probably through computer use first, if I had to guess,
    0:21:38 because you’ll be able to have like the computer actually interact with the
    0:21:41 thing, whether it’s a game or an application and test different scenarios
    0:21:44 and then see where bugs are, log them, possibly even then look at the code
    0:21:46 base and like identify the bug and solve it.
    0:21:49 But I assume at some point you could just simulate the entire thing.
    0:21:50 Yeah, you can do that for anything.
    0:21:52 You can do that for architecture.
    0:21:55 I think you still have real human game testers though, right?
    0:21:58 Because like you still need people to tell you whether the thing’s fun or not.
    0:22:00 Oh, yeah, of course.
    0:22:01 Yeah, it’s like, this is super fun.
    0:22:03 That’s just because you like hard challenges.
    0:22:04 You are like hard problems.
    0:22:06 Yeah, yeah, yeah.
    0:22:10 So right now the day that we’re recording this is actually basically
    0:22:15 the first day of CES, but I did go and sort of speed run the convention center
    0:22:18 and try to see as much as I can see while I was out there.
    0:22:20 Some other interesting stuff.
    0:22:25 I did sit in on a Samsung keynote and Samsung is all about like smart devices.
    0:22:31 Some of this stuff that they’re talking about feels almost like novelty.
    0:22:34 Like that’s cool, but like are people really going to use it once it’s
    0:22:36 installed in their house, right?
    0:22:40 Like they had stuff like AI in your refrigerator where it’ll take pictures
    0:22:45 inside of your refrigerator and then recommend things for you to buy based on
    0:22:49 like AI, it’ll be like, oh, it looks like they’re running low on eggs or whatever.
    0:22:51 And it’ll let you know when you’re running low on things.
    0:22:54 They have like digital displays on the refrigerator as the example they were
    0:22:56 giving as you wake up in the morning.
    0:22:59 And while you’re making breakfast, you can see right on your refrigerator what
    0:23:00 the weather is going to be like.
    0:23:03 And you can quickly look at your calendar and they’ve been doing that
    0:23:06 stuff for like 15 years now, but now it’s using AI.
    0:23:07 Yeah, yeah, yeah.
    0:23:10 I mean, they, you know, I remember I’d go to CES like over 10 years ago.
    0:23:13 They had that kind of stuff then you see the fridge and you see, you know,
    0:23:15 it’s got all this stuff on it and then no one uses it.
    0:23:18 I think that stuff eventually will happen, but it has to get so seamless
    0:23:20 and so tied into all your systems.
    0:23:22 You’re already using that you don’t even think about it.
    0:23:26 Like, you know, some of the examples they were giving though was like more
    0:23:30 of like the agentic style of like it could look in your fridge, see that
    0:23:34 you’re low on eggs and then send that to Instacart and then, you know, eggs
    0:23:37 will just show up at your door without you having to think about it.
    0:23:37 Right.
    0:23:40 So like that’s sort of where some of the AI stuff would play in.
    0:23:44 I don’t know if you saw that like demo that I did on my live stream
    0:23:47 where I told Amazon to buy me some toilet paper and then it racked up
    0:23:50 $600 worth of toilet paper on my Amazon card.
    0:23:52 There need to be some kind of control system.
    0:23:54 Yeah, that’s what I worry about with that kind of stuff.
    0:23:58 It needs to seamlessly tie into the existing AI systems people are using.
    0:24:00 I’m not using AI voice as much as I expected I would, but I think
    0:24:03 eventually I will eventually if they could tie into some kind of system
    0:24:07 like that and then chat to BT might say, Hey, you have all these notifications
    0:24:10 from you have a notification from your fridge or a notification from email
    0:24:13 of things I think are important or you need to make a decision upon.
    0:24:16 And then you can kind of through natural voice decide upon how you want to
    0:24:17 deal with things like that.
    0:24:18 Yeah. Oh, yeah.
    0:24:19 If I need eggs, just freaking get the eggs for me.
    0:24:20 Don’t ask me.
    0:24:24 Somebody’s asked me to, you know, to send an appointment.
    0:24:27 No, I don’t like appointments or whatever it is.
    0:24:29 Just like deal with all those kind of things and like kind of set
    0:24:30 the rules set using the AI.
    0:24:33 I think I think we’re like probably a year or two away from like way more
    0:24:36 personalization like that and that kind of stuff can make a lot of sense.
    0:24:40 I would say like CES this year from what I’ve seen so far, I haven’t even been
    0:24:44 over to the Venetian Expo yet, which is like a whole other thing, right?
    0:24:47 The Venetian Expo, they have this thing called Eureka Park, which is only
    0:24:49 companies that are a year older or newer, right?
    0:24:54 So that’s all the like cutting edge, brand new, like early startups over there.
    0:24:55 And I haven’t seen what’s there yet.
    0:24:56 So I haven’t explored that yet.
    0:25:03 But what I’ve seen so far has been a lot of like self-driving car tech, tons
    0:25:05 and tons of self-driving car tech here.
    0:25:09 And then like internet of things blended with AI, right?
    0:25:15 Taking like the IOT stuff and then adding AI like the example of, oh,
    0:25:16 you’re, you’re out of milk.
    0:25:19 Send that to Instacart, let Instacart order it for you.
    0:25:24 They had some other examples of like a Samsung 2-in-1 washer and dryer and it
    0:25:30 could like automatically sense how dry your clothes are and then like make
    0:25:33 it wash for longer if they were still wet.
    0:25:34 Mine in Japan already does that.
    0:25:35 Don’t they already do that in America?
    0:25:36 I don’t like mine.
    0:25:40 Yeah, there’s like some other AI elements to it where you can like communicate
    0:25:46 through like Siri or whatever or like your Amazon Alexa or things like that
    0:25:49 to tell it to do stuff and it’s all connected.
    0:25:53 I did watch the Samsung keynote, but you can tell how impactful it was because
    0:25:56 I’m already forgetting some of the stuff that they were telling us during it.
    0:25:58 But it was a lot of that kind of stuff.
    0:26:03 It was a lot of like smart washer and dryers, smart refrigerators, you know,
    0:26:06 smart lighting systems.
    0:26:10 They were talking about systems that control like you’re heating in air.
    0:26:15 Maybe at night you’re wearing like your Samsung Galaxy ring and it notices
    0:26:16 that you’re like really, really hot.
    0:26:19 So it’ll like actually turn up the AC for you.
    0:26:24 And the idea is less and less like prompting and asking AI to do things.
    0:26:28 And AI just sort of figuring out what it needs to do and do it for you.
    0:26:29 Right.
    0:26:31 Like I think that’s where they’re trying to go with a lot of the internet
    0:26:36 of things stuff combined with AI is instead of me telling these things what
    0:26:42 they should do next, AI will recognize, OK, in the past, he did it this way.
    0:26:44 So let’s make sure we do it that way again.
    0:26:47 Or he’s looking really hot while he’s sleeping.
    0:26:49 Let’s turn on the AC for him.
    0:26:54 You have smart blinds and you set an alarm for 830 and it raises
    0:26:56 your blinds for you at 830.
    0:26:57 Pretty sure you can already do that.
    0:26:59 Different lighting modes would be kind of cool.
    0:27:02 Like, you know, you like dim lights at certain times or, you know,
    0:27:04 you could turn the lights red at certain times.
    0:27:05 Yeah, yeah, yeah.
    0:27:09 It’s all like internet of things combined with AI.
    0:27:13 But the idea is they want it to be more preemptive.
    0:27:17 They want it to be more, you don’t have to go and tell these various
    0:27:18 things what to do.
    0:27:23 They just know what to do next and give you the results you’re looking
    0:27:25 for without you having to go and ask it to do that.
    0:27:26 Right.
    0:27:28 So there’s been a lot of that kind of stuff.
    0:27:31 I think you need the automatic stuff with the ability through natural
    0:27:34 language to modify how those things behave is what I think.
    0:27:36 I think it’s all automatic.
    0:27:38 People are just going to like some people are going to hate that.
    0:27:41 It’s like I want some amount of control over a bit more how I can tweak it.
    0:27:42 I wonder how that’s going to work.
    0:27:45 If you’ll be able to do that with like chat to PT and stuff at some point.
    0:27:48 Yeah, at the end of the day, I think the way I felt about it,
    0:27:51 sitting through some of these keynotes is I came out of the keynotes going,
    0:27:53 that’s really, really cool tech.
    0:27:55 I think it’s really, really impressive that we’re at a point
    0:27:58 where we can do that stuff now other than that.
    0:27:59 Do I really care that much?
    0:28:01 Do I really feel like it’s going to impact my life that much?
    0:28:05 Do I really like want to go and buy the latest state of the art
    0:28:11 washer and dryer combo or a fridge with a, you know, a giant 30 inch screen on it?
    0:28:12 Not really.
    0:28:16 Like I’m not sold myself yet, but I am impressed.
    0:28:17 Right.
    0:28:21 And that’s how I feel about a lot of what I’m seeing at CES this year.
    0:28:23 Is I’m like, wow, that’s really, really impressive.
    0:28:23 Do I want one?
    0:28:26 Not yet, but it’s impressive.
    0:28:31 I’m OK with all that stuff as long as we don’t have any kind of oppressive government.
    0:28:35 It’s like, hey, you have a point system and you don’t have enough points.
    0:28:38 Sorry, we don’t like what you said on social media.
    0:28:39 So you can’t really.
    0:28:43 It’s like, yeah, yeah, yeah, yeah, we’ve locked your fridge and you can’t eat
    0:28:45 until you get your credit score.
    0:28:48 You must say something good about the current president.
    0:28:51 Otherwise you’re going to starve.
    0:28:51 I’m sorry.
    0:28:52 That’s just how it is.
    0:28:54 Yeah, yeah, but I don’t know.
    0:28:55 Like I’m going to do some more exploring.
    0:28:59 I actually encourage people to check out like the Nvidia keynote.
    0:29:00 It’s really long.
    0:29:03 I mean, it was like, I think his keynote went on for almost two hours last night.
    0:29:06 I actually ended up leaving early because I had to be at another event.
    0:29:09 So I left like five minutes before the keynote was over.
    0:29:12 But the Samsung keynote was really, really interesting.
    0:29:15 They did show like a lot of examples of how this stuff will like all tie together
    0:29:21 and how, you know, it’s going to try to preempt what your needs are before you even prompt it.
    0:29:24 So I think that’s where a lot of the things are going this year.
    0:29:27 Last year, there was some really, really cool like TV tech.
    0:29:29 I love seeing that kind of stuff.
    0:29:31 Like I haven’t really seen much yet.
    0:29:33 All I’ve seen is like bigger TVs, right?
    0:29:38 I saw like a hundred and twenty seven inch like OLED eight KTV or something
    0:29:41 that looks really impressive when it’s right in front of you.
    0:29:44 But, you know, it’s probably like a sixty thousand dollar TV or something.
    0:29:45 Right.
    0:29:48 Last year, they had like these transparent TVs that were really cool to see.
    0:29:52 Not really a lot of use cases for like in the home,
    0:29:54 but really interesting for like businesses.
    0:29:57 Like some of the examples they showed last year was you can have
    0:30:00 like a display case with like all of your pastries in it.
    0:30:02 And then like on the display in front of it,
    0:30:05 it’s actually got like digital labels of what each thing is.
    0:30:09 And you can play like animations in front, like on the display case.
    0:30:12 So it’s almost like using these clear TVs as your display case.
    0:30:15 And I thought that was really cool for like businesses,
    0:30:17 but who really wants a clear TV in their house?
    0:30:20 Like to me, it seems like there’s more downside than upside to that.
    0:30:22 Right. But none of that kind of stuff this year.
    0:30:24 There’s a couple like flying car techs.
    0:30:28 There was a flying motorcycle, like a two wheel motorcycle
    0:30:32 that has like the four like drone propellers, like a quad copter
    0:30:34 motorcycle that looked really interesting.
    0:30:37 I mean, it kind of looks like a death trap because like the blades
    0:30:39 are like really like right at face level.
    0:30:43 If somebody wasn’t paying attention, it looked like a walk right into the blades.
    0:30:48 You know, they had that they had that the X-Peng car that looks like a giant van.
    0:30:52 But out of the back of the van is like a little personal drone that you can fly around.
    0:30:55 They have that here, which is really, really cool to see.
    0:30:58 Really, really cool to see that kind of stuff in person.
    0:31:00 But all of it feels so much like concept still, right?
    0:31:02 It feels like, oh, that’s really, really cool.
    0:31:06 You can do that, but it’s so far off from actually being anything.
    0:31:08 Anybody’s actually going to use in real life.
    0:31:10 It’s crazy that the tangible stuff is AI.
    0:31:16 It’s like you told me 10 years ago that the tangible stuff in a CES was the AI stuff.
    0:31:18 It’s like, what are you talking about?
    0:31:21 It was always the AI stuff that was just like, oh, yeah, that’s cute.
    0:31:23 It doesn’t actually work. None of it works.
    0:31:27 What I want to know is when you were coming to CES, let’s say 15 years ago,
    0:31:30 did they have like flying car concepts and stuff back then?
    0:31:32 Because I feel like they probably did.
    0:31:34 It was just a little more of what we’re seeing now.
    0:31:36 No, there was no flying car stuff.
    0:31:41 You know, actually, I helped put a video game system in the back of a Toyota Prius
    0:31:44 concept car in partnership with Alcatel Lucent.
    0:31:46 I was doing this startup called GameStreamer at the time.
    0:31:48 And so a lot of it was that kind of stuff.
    0:31:49 It was like, you know, and it was kind of gimmicky.
    0:31:50 It’s like, you got a car.
    0:31:52 OK, now there’s a game system in the back of the car.
    0:31:54 Yeah, yeah.
    0:31:55 A lot of it was that kind of stuff.
    0:31:57 And there was a lot of the smart fridges.
    0:31:59 They were doing interesting things with lighting.
    0:32:02 You know, I did see some stuff there that ended up being used in people’s houses,
    0:32:06 like the stuff like controlling your lights with voice and stuff like that.
    0:32:09 I did see that kind of stuff first at CES before I ever saw it in anybody’s house.
    0:32:11 Yeah, so you did see some kind of things
    0:32:13 that ended up being real, you know, that people would use.
    0:32:14 Yeah, yeah.
    0:32:18 Most of the cars that you see here feel like concepts that we may never see.
    0:32:22 Some of them, I think, are actually on the roads like Wemo is here.
    0:32:24 You can take a look at what the Wemo cars are.
    0:32:25 But if you go to San Francisco,
    0:32:28 you could literally order one up and take a ride in one.
    0:32:29 So they have that here.
    0:32:34 Zooks is here, which is like Amazon owned self-driving car company.
    0:32:38 Honda showed off some like cool EV self-driving cars.
    0:32:41 One of them almost looks like a Lambo or like maybe like a Lotus,
    0:32:43 like a cross between like a Lotus and a Lambo or something,
    0:32:45 which is really, really interesting.
    0:32:47 Honda is kind of going down that path a little bit.
    0:32:50 Before we wrap up here, I wanted to like go back to Nvidia for a second,
    0:32:53 because there was one thing that I heard Nvidia talk about that I didn’t hear you mention.
    0:32:54 It’s kind of hard to understand.
    0:32:57 So I’m still trying to process like what exactly it means.
    0:33:00 But Jensen talked about with this new graphics card,
    0:33:04 the way that they’re able to achieve so much better performance,
    0:33:07 partially it’s because they’re using AI for a lot of the processing.
    0:33:07 Right.
    0:33:09 And my understanding was he was saying almost like 90 percent
    0:33:13 of the processing of like the graphics was being given to AI.
    0:33:14 Yeah. Yeah.
    0:33:17 And that’s why you can have a smaller chip that uses less power was because of that.
    0:33:21 Yeah, he was talking about how it’s really interesting
    0:33:25 because the G force graphic cards led to being able to create AI.
    0:33:29 And now he’s using AI to be able to create better and better graphics cards.
    0:33:33 Did you hear about the switch to the the rumors that have dropped around
    0:33:35 the Nintendo switch to know.
    0:33:38 So basically what they’re saying about the Nintendo switch to
    0:33:41 is that they’re expecting it to be announced in 2025.
    0:33:45 And they’re expecting it to be able to play games in 4K.
    0:33:48 You’re not going to be actually creating games in 4K.
    0:33:52 The game developers are going to be basically creating games in 720p
    0:33:56 or 1080p or whatever to keep the file sizes small enough to fit on a cartridge.
    0:34:00 And then the switch itself is going to upscale it to 4K
    0:34:05 and sort of like essentially guess the pixels in between to scale it up to 4K.
    0:34:09 So my understanding of like what NVIDIA is doing is kind of
    0:34:13 that same kind of concept is that it can take less data
    0:34:17 and then imply the rest of the information using AI.
    0:34:20 Yeah, again, this is a really oversimplification
    0:34:21 and I don’t totally understand it.
    0:34:25 But that’s sort of my understanding is that it can actually work with less
    0:34:29 information and then imply and use AI to figure out the gaps.
    0:34:31 And that’s sort of how it’s going to work.
    0:34:36 But again, I’m not an engineer and so I’m not the person to explain that.
    0:34:38 That makes sense to me.
    0:34:41 But like the way he said it made it sound like it’s almost hard
    0:34:43 coated into the hardware, which was kind of surprising to me.
    0:34:46 Like he made it sound like this chip actually does that.
    0:34:49 Like you don’t actually have to do anything different with the game
    0:34:53 code to make that work that the chip itself is doing it in real time.
    0:34:57 I was like, if that’s true, that is that is mind boggling.
    0:35:01 Because like you could potentially like really scale up graphics in games.
    0:35:05 You know, if you’re getting like a order of magnitude improvement from the AI,
    0:35:08 we’re going to see games that look like we’re walking around the real world,
    0:35:09 like real, real world.
    0:35:11 Yeah, I don’t think we’re far off from that already.
    0:35:14 Yeah, yeah. But now with this, I mean, like we’ll probably be there
    0:35:15 like in a year or two. It’s like it’s crazy.
    0:35:18 You know, it might require like the 5090 or something like that
    0:35:19 to get that level of realism.
    0:35:22 But I bet you the 5090 will get you there, you know?
    0:35:26 Yeah, you just need people to develop stuff that’s at that level as well.
    0:35:27 Well, yeah, that was the most exciting thing to me.
    0:35:30 You have to make it look pretty good, but then the AI fills in the gaps
    0:35:31 to make it look better.
    0:35:33 So in theory, that makes game development easier.
    0:35:34 Yeah, that makes sense.
    0:35:38 We’ve already got things like, let’s say like runways, video to video, right?
    0:35:42 Where you can give it one video and then it can go and sort of re-skin
    0:35:45 that video and make it look like a completely different video.
    0:35:46 But still looks good.
    0:35:49 Like you get something out of Unreal Engine 5 right now.
    0:35:51 It looks pretty damn good, right?
    0:35:55 Already. Now imagine it sort of re-skinning it, but with like a
    0:35:58 upscaled 8K realism filter on it.
    0:36:02 Right. And now that Unreal Engine game looks indistinguishable
    0:36:04 from something that was just shot on a video camera.
    0:36:07 Right. Imagine you’re going to be like go back and remaster like old games
    0:36:10 or old movies or anything to just like make them look super bad.
    0:36:13 Yeah, yeah, yeah. I mean, I don’t doubt it, right?
    0:36:16 Like they’ll probably be systems where you could take your old VHS tapes,
    0:36:19 run them through and then upscaled them 4K.
    0:36:21 Yeah, make it look, you know,
    0:36:24 add some like cinematic stuff to it, to lighting and other stuff.
    0:36:25 And just like, yeah, why not?
    0:36:28 Yeah, I mean, I feel like with all of this AI stuff,
    0:36:31 once it gets cheap enough that it’s like anybody can use it,
    0:36:32 anybody can do whatever they want with it.
    0:36:33 That’s going to get so weird.
    0:36:34 Any of that is on the table.
    0:36:38 You know, one thing that they do in Asia, at least I know it’s normal
    0:36:42 in Japan, that when you have a wedding, you have somebody filming the wedding.
    0:36:46 And at the same time, people are editing the videos during that time period.
    0:36:48 So by the time you’re done with the wedding,
    0:36:53 they present to you a really nicely edited video of the day in movie format.
    0:36:57 So they try to present it like it’s a movie, like your wedding day was a movie.
    0:36:59 Yeah, yeah, yeah, I’ve heard of that kind of thing.
    0:37:01 But yeah, I don’t think it’s very common.
    0:37:02 And there’s always like tears.
    0:37:05 They were all like emotional about it because they make it so cinematic.
    0:37:08 Like, oh, my God, look, you know, this moment that we were all part of today.
    0:37:11 You got that for like your life. That’s going to be so weird.
    0:37:14 Like take all the like the videos of your life and have AI automatically edited
    0:37:16 into this kind of like master film of your life.
    0:37:19 Yeah, well, that job that you just mentioned is going to get way easier
    0:37:22 for the right because they’re just going to feed it all of the video.
    0:37:27 And then AI is going to figure out how to edit it into like a cinematic masterpiece.
    0:37:29 Yeah, it’ll make it better, right?
    0:37:33 At cinematic lighting, lighting the right spots, editing in the right spots,
    0:37:35 tied all together with the right narrative.
    0:37:37 Yeah, it’s going to be crazy.
    0:37:39 Yeah, we’re entering a wild world.
    0:37:41 The last few years I’ve come to CES
    0:37:45 and like it definitely like expands my mind into what’s possible, right?
    0:37:48 Makes you go, oh, crap, I didn’t know like we’re able to do that yet.
    0:37:50 I didn’t know we were capable of that kind of thing yet.
    0:37:52 I was still surprised by what Jensen said
    0:37:55 about the predicting the different parts of the game.
    0:37:57 You know, if that’s what they’re really doing in hardware,
    0:37:59 because he had been saying that like, hey,
    0:38:02 we’re eventually going to use AI to generate all the pixels.
    0:38:04 Yeah. And I was like, well, that sounds kind of farfetched.
    0:38:07 Maybe that’s like a five year kind of thing that sounds really hard.
    0:38:09 Yeah, there’s already been demos of that.
    0:38:12 I feel like we’ve already seen glimpses of it, right?
    0:38:13 I really want to learn more about that
    0:38:15 and like see if that’s really what’s going on.
    0:38:16 But that’s happening.
    0:38:18 That’s going to change, you know, so many things.
    0:38:20 And also those improvements for the graphics cards,
    0:38:22 those go back into the improvements in the AI.
    0:38:23 So even for non entertainment,
    0:38:26 all of these things end up improving AI in general,
    0:38:28 which then goes into all businesses.
    0:38:30 It’s interesting that somehow gaming has become
    0:38:32 this part of the core loop of all business.
    0:38:34 Yeah, I think maybe who listened to the podcast are like,
    0:38:35 why is Nathan talking about gaming?
    0:38:38 It’s like, well, all that started from gaming,
    0:38:39 like without gaming, we don’t have AI.
    0:38:41 Oh, 100%. Right.
    0:38:43 I’m just looking for the day to where like as a YouTuber, right?
    0:38:47 Like I can record YouTube videos at 720p, right?
    0:38:48 Edit the whole video at 720p
    0:38:53 because it’s a hell of a lot faster to edit and render at 720p than it is at 4K.
    0:38:57 Create my whole video at 720p, render it out, toss it on YouTube.
    0:39:00 And then YouTube just upscales it to 4K for me.
    0:39:01 You know, we’re like that close to that.
    0:39:04 I think YouTube’s already testing AI upscaling.
    0:39:06 And, you know, I think we’re going to get to a point
    0:39:09 where we can record and edit our videos at 720p
    0:39:12 and then post process them up to 4K.
    0:39:15 And it’ll look like the whole thing was filmed in 4K.
    0:39:16 And I’m excited about that.
    0:39:19 I think that’ll make video creation a lot easier.
    0:39:22 Yeah, I mean, eventually it’ll get good at editing even.
    0:39:24 I think I think we’ll give it like samples
    0:39:26 of like different YouTubers you like
    0:39:28 or if you’re making a movie, different movies you like or whatever.
    0:39:30 And it’ll be able to help you with editing.
    0:39:32 Yeah, whoever’s editing this video, don’t worry.
    0:39:34 Like, you know, I think it’s going to take a few years before they
    0:39:36 I like it and I’m also scared of it.
    0:39:38 That’s everything with AI is like, it’s all super exciting
    0:39:39 and scared at the same time.
    0:39:42 I’m more excited than scared, but I understand if anybody
    0:39:45 could make YouTube videos that look like Mr. Beast videos
    0:39:49 or whatever, like, does that kind of devalue what Mr. Beast does?
    0:39:51 You know, yeah, definitely. Yeah, so I don’t know.
    0:39:54 But I think we’re still a little ways off from that, fortunately.
    0:39:57 But like you said, AI sort of excites me
    0:40:00 and scares me at the same time, but more excites me than scares me.
    0:40:02 So I’m going to keep talking about it.
    0:40:05 Yeah, cool, man. Well, this has been really fun.
    0:40:08 We actually came into this one, not knowing if we had enough to talk about.
    0:40:12 And I think this is actually a pretty solid episode all said and done.
    0:40:14 That’s kind of been my experience with CES this year.
    0:40:20 If I was to sort of recap it, it’s like slightly underwhelming this year.
    0:40:24 Between last year and this year, there was major leaps.
    0:40:28 And what I have seen a lot of is like AI blended with Internet of Things,
    0:40:30 a lot of self-driving tech.
    0:40:34 And other than that, it’s more of the same as last year, like AI and everything.
    0:40:36 We’ve got barbecues that have AI.
    0:40:38 We’ve got bird feeders that have AI.
    0:40:43 I was talking to like a Synology who makes like network attached storage solutions.
    0:40:45 Right. They’re putting AI in those now.
    0:40:50 Right. So it’s like existing companies adding AI in Internet of Things blended
    0:40:55 with AI and a lot of self-driving tech, which relies on AI.
    0:41:01 So I mean, CES is really, really AI focused again, but still a whole year
    0:41:05 has gone by since the last CES, but it doesn’t feel like the advancements
    0:41:08 were as big as I thought they would be in that timeframe.
    0:41:10 And that’s probably how I would summarize CES this year.
    0:41:12 A lot of it sounds gimmicky.
    0:41:13 It’s in Japan.
    0:41:16 You have a lot of that right now, too, because in Japan, there’s a huge AI fever.
    0:41:19 Like when I bought my washer and dryer, it had a big sign AI.
    0:41:22 I talked to my wife and it’s like, wait, so just the fact that they’re
    0:41:24 like, you know, determining that it’s still wet.
    0:41:27 That’s they’re calling that the AI that it’s, you know, and I was like,
    0:41:30 and I was like, the models do that a few years ago in Japan.
    0:41:31 Yeah, they already did that.
    0:41:35 It’s like, so they’ve just kind of attached AI on to it to say like,
    0:41:36 yeah, this is the new AI feature.
    0:41:38 Yeah. The Nvidia stuff sounds amazing, though,
    0:41:41 like especially if they are doing the thing I said, the hardware of predicting
    0:41:43 pixels and things like that.
    0:41:45 So tomorrow I’m spending all day with Nvidia.
    0:41:47 I’m actually meeting up with the Nvidia crew.
    0:41:50 I will know more about what Nvidia is doing tomorrow.
    0:41:53 Like I’m getting hands on demos with a lot of their new tech tomorrow.
    0:41:59 So I will know more about it and feel free to send me any questions you have
    0:42:02 about what Nvidia is up to and I’ll ask them when I’m with them tomorrow
    0:42:06 because I should know a lot more about what’s going on with Nvidia.
    0:42:09 And I’ll ask the questions that you brought up.
    0:42:14 I want to understand how the RTX, the new 50 series is leveraging AI.
    0:42:16 Like I want to understand that process on a deeper level.
    0:42:18 So I will ask them about that.
    0:42:22 And maybe there’s a follow up episode to this one where we go deeper
    0:42:24 into what Nvidia is up to.
    0:42:28 Well, on that note, I’m going to get back down to the conference floor
    0:42:30 and see what else I can find.
    0:42:32 And again, I’m going to be hanging out with Nvidia tomorrow.
    0:42:34 So I’ll know more about what they’re up to.
    0:42:39 And if you want to stay looped in, you want to learn more about what I find out
    0:42:43 from Nvidia and more about what I come across at CES and learn about some
    0:42:47 of these other tools that Nathan’s been sort of teasing us about,
    0:42:49 but not going into very much detail yet.
    0:42:54 Make sure that you subscribe to this show wherever you subscribe to podcasts.
    0:42:56 YouTube is the visual platform.
    0:42:59 We’re showing a lot of stuff on our screens, but we’re also available
    0:43:01 wherever you listen to your podcast.
    0:43:04 So give us a subscribe and thank you so much for tuning in.
    0:43:06 Hopefully we’ll see you in the next one.
    0:43:06 Thank you.
    0:43:07 Yeah.
    0:43:08 Yeah.
    0:43:09 .
    0:43:11 Yeah.
    0:43:13 (upbeat music)

    Episode 41: How groundbreaking are the AI-driven advancements showcased at CES 2025? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) take you through the year’s biggest AI announcements and innovations from the Consumer Electronics Show.

    This episode dives into the latest in AI’s integration with the Internet of Things, Nvidia’s exciting new projects, advancements in self-driving car technology, and the potential of fully automated game testing. The hosts share their CES experiences, highlight favorite tech demonstrations, and discuss the future of AI, from personal computing to enterprise applications.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Discussing Nvidia’s AI advancements at CES.
    • (04:52) Hopes for AGI, aiming next for ASI.
    • (06:33) AI to revolutionize solving major human challenges.
    • (10:50) Edge computing could disrupt cloud-based business models.
    • (14:04) Nvidia dominates AI; local privacy management evolves.
    • (17:40) Nvidia’s Project Cosmos: Virtual testing for technologies.
    • (20:39) Automated testing identifies potential game bugs efficiently.
    • (24:36) CES highlights: self-driving tech, AI-enhanced IoT.
    • (29:12) Advancing tech predicts needs, massive impressive TVs.
    • (29:54) Flying motorcycle showcased; clear TV criticized.
    • (33:35) Nintendo Switch 2: 2025 release, AI upscales to 4K.
    • (38:41) Excited for easier video creation with AI.
    • (42:17) Stay informed on Nvidia updates; subscribe now.

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • AI Predictions That Will Completely Change Life In 2025

    AI transcript
    0:00:04 I think 2025 is going to be wild with what we see with AI video.
    0:00:09 For anyone who’s saying that these models do not reason and they’re just like parents or whatever, no.
    0:00:16 So the average IQ range for adults is between 90 and 109, with 100 being the theoretical average.
    0:00:23 So estimated for 03, 157. This is like top of the top smartest people in the world level.
    0:00:30 Hey, welcome to the Next Way Podcast. I’m Matt Wolf. I’m here with Nathan Lanz.
    0:00:36 And there has been a lot happening in the AI world, including an announcement that a lot of people
    0:00:42 think is just this monumental, huge, massive announcement in the fact that we got this 03
    0:00:49 model from OpenAI. Well, I guess I can’t really say we got it. We got to see it. We got to see
    0:00:55 the 03 model from OpenAI. But a lot of people have been, you know, touting it as like this
    0:00:59 massive breakthrough. I’ve even seen like YouTube videos give it the title of like,
    0:01:06 we finally have AGI. And everybody’s saying that 03 is AGI. So in this episode, we’re going to talk
    0:01:12 about that a little bit. We’re going to talk about our predictions for 2025. And Nathan’s going to
    0:01:17 show off something pretty cool that he’s been working on using some of these models. So let’s
    0:01:24 get into some of the like 2025 predictions. You and I obviously talk a lot. So chances are our
    0:01:29 predictions are going to be fairly close to each other. But let’s go ahead and dive in.
    0:01:34 Yeah, let’s start on the models, I guess. I think by the end of the year, a few things are going
    0:01:40 to have happened. I think XAI is going to roll out Grok 03. And it’s probably in terms of like a
    0:01:44 fast LM that you can just chat with for daily use. It’s going to be really, really, really good.
    0:01:48 The fact that they’re just, they’re trying to train it on more data than almost anyone else.
    0:01:53 And they’ve, you know, they’re buying more chips from NVIDIA than almost anyone else.
    0:01:56 I think it’s going to be really good for like daily, just like, just chatting with AI.
    0:01:58 It’s going to, I think it’s going to surprise people. But in terms of like,
    0:02:02 Do you think that’ll replace some advanced voice mode for ChatGPT for you?
    0:02:07 I don’t know. I don’t think so. I don’t, I’m not, I’m not sure. Like I think for power users,
    0:02:13 I’m pretty convinced ChatGPT and OpenAI are going to be even further ahead by the end of 2025.
    0:02:19 Like, I think there’s a decent chance that, because like 03, if it’s just 01 with more
    0:02:22 compute thrown at it, well, they’ve definitely been working on other stuff. They’re not just
    0:02:26 sitting there like, Hey, let’s take what we’ve created and throw more compute at it. And like,
    0:02:31 that’s all, that’s all we’re doing for the year. Like, it makes sense to me that like
    0:02:36 something like Orion, like a larger model underneath as the baseline will be part of
    0:02:41 some new model by the end of 2025. And I think there’s a decent chance that they actually
    0:02:44 release it in 2025, like by the end of 2025.
    0:02:49 Well, Sam Altman hinted that he believes AGI is coming in 2025, right? So yeah,
    0:02:56 yeah, like I said, like I think 01 Pro is like some kind of basic AGI already. That’s my definition.
    0:02:59 People like some people think that’s like stupid. Everyone had different definitions of this stuff.
    0:03:05 But I think by the time you get to 04, when not only it can like respond quickly to like basic
    0:03:10 questions and have like incredibly good answers, but also if you ask it something complicated,
    0:03:14 it’ll give you something back in a minute or two that’s amazing. Most people are going to call
    0:03:20 that AGI. And I think that could be coming by the end of 2025. And that will make all this stuff
    0:03:24 we talked about, like in terms of agents possible, like agents that go off and do market research
    0:03:28 for you, come back and bring you some charts and show you what, how you should be approaching a
    0:03:32 market and what’s the best strategy. That stuff will be agents, how you interact with email,
    0:03:38 do email, like your assistant. That’ll probably be an agent by the end of 2025.
    0:03:41 And that’s going to be a dramatic change. That’s going to be like where we’re going to start to
    0:03:47 have some real questions, societal questions by the end of 2025. And it’s going to be interesting
    0:03:51 to see how that plays out because, you know, I have said in the past, well, Donald Trump’s
    0:03:58 administration is going to be very pro AI. But also there was that recent thing with the port,
    0:04:03 what was it, the port unions, unions for the port workers, where he kind of backed down from them
    0:04:06 saying that we need to protect their jobs and we shouldn’t automate the ports.
    0:04:10 And we might see more of that because like there may be such
    0:04:17 cultural backlash to some of the job loss that it’s, you know, it may become a pretty tense
    0:04:21 conversation and by the end of 2025, because I do think we’ll start to actually see dramatic
    0:04:25 changes in the job market. Like maybe everyone’s not going to fire people, but you’re going to see
    0:04:30 like, oh, I’m not hiring as many engineers anymore. Like there’s less job openings, less
    0:04:37 assistance hired, things like this. Yeah, I agree. I think I do think like the big buzzword of 2025
    0:04:42 is going to be agents though. I think like every company is going to be talking about AI agents.
    0:04:47 I think you’re going to be hearing Google, open AI, Claude, X, you know, you name it,
    0:04:53 they’re all going to be talking about agentic workflows and tool use and being able to like
    0:04:58 do things on your computer, on your behalf. And like, I think that’s going to be the conversation
    0:05:03 of 2025. I think if there’s any one prediction that I’m fairly certain on, it’s that. I also
    0:05:06 think the video models, I don’t think we’ve seen anything yet. I think, you know, eight-second
    0:05:12 videos that look pretty decent. I don’t think we’ve seen anything yet. I think 2025 is going to be
    0:05:16 wild with what we see with AI video. I mean, Matt, like, let’s still try to step back for a
    0:05:19 second. Like, like realize, like, you know, we’ve been talking about this stuff for like over two
    0:05:25 years now. Like, you know, in 2023, the videos looked horrible. I was sharing the, I was sharing
    0:05:29 threads about them just because they look cool because of technology underneath them. I was
    0:05:34 like, this is amazing that you’ve asked AI to dream up what a video of this scene. That’s
    0:05:40 incredible. My co-founder for one of my previous startups, JR Bedard, unfortunately, he passed
    0:05:44 away, but he used to be fascinated with this stuff. He would like, he would like read books about
    0:05:47 like how AI was going to be able to generate arc and he would show me all this stuff. This was
    0:05:52 maybe like 10 years ago. Right. Oh, that’s cool. It can make a little pattern and stuff.
    0:05:55 Yeah. Yeah. Yeah. And it’s like, that’s where it was 10 years ago. And it’s like, that was like
    0:05:59 the seeds of what we have now, but it was like, oh, it can make out a little, a cool pattern.
    0:06:02 And I was like, oh, that’s cool. And I was like, do you think that’s going to be like a real thing?
    0:06:06 Like in the next like five or 10 years, and it was like, you know, maybe in like 10 years,
    0:06:11 and it’s about where it is now. And the videos looked like they were cool if you understood
    0:06:16 the technology, but they looked horrible two years ago. And now are they perfect to know? But
    0:06:21 like, man, they already look like like 90, 95% of what you might see in a film, like at least
    0:06:27 like a short scene or something. Yeah. And, and it feels like too, that, you know, I think I may
    0:06:32 have said this offhand the other day, but I believe that where the video models are right now, they’re
    0:06:37 kind of like in GPT-4 level in terms of they’re just trained on massive amounts of data, but they’re
    0:06:43 not doing a whole lot in terms of reasoning about that data, about that output. And so, so once you
    0:06:47 start to, and this is why I think like probably Sora is probably still going to be very good.
    0:06:52 Because if OpenAI is that good with their internal models, as soon as they apply that to video as
    0:06:57 well, like reasoning on top of it, you’re going to get so much more control of the output of these
    0:07:01 videos. You’ll be able to like, yeah, I wanted the character to turn that way. I wanted to do this,
    0:07:04 and also make sure that the characters are consistent, because then you can reason upon
    0:07:08 the output and make sure that it’s consistent. That quality is going to go up dramatically.
    0:07:14 Well, I mean, we’ve got AI large language models now that it will basically generate an output,
    0:07:20 double check its output before sending it to you. When you look at like AI art generation,
    0:07:28 the big way of doing AI art early on was GANs, Generative Adversarial Networks, where it would
    0:07:35 essentially, you had a discriminator and a generator, and the generator would generate a
    0:07:42 piece of art, and then the discriminator would go, you would give it a prompt like draw me a cat,
    0:07:45 and the generator would draw a cat, and the discriminator would go, that didn’t look like
    0:07:48 a cat, try again, and it would draw another cat, and the discriminator would go, that doesn’t look
    0:07:54 like, and it would go back and forth until the discriminator couldn’t tell the difference between
    0:08:00 a real photo of a cat and an image of a cat. We haven’t seen that kind of tech inside of video
    0:08:07 yet, so I think we’re going to have something equivalent to like a GAN, or like what we’re
    0:08:12 seeing with like O3, which is kind of the same idea, right? It does a generation, double checks
    0:08:16 itself, goes, all right, there’s things that needs to be fixed, double checks itself, takes it to be
    0:08:21 fixed, and then finally presents the output once it’s confident in the output that it gave you,
    0:08:26 right? Very, very similar conceptually. Well, I think we’re going to get to a point where
    0:08:32 we do that with video models, you give it a prompt of like a unicorn flying in space with a cat
    0:08:37 writing on its back, it will generate that, it will watch the video back itself and go,
    0:08:40 that doesn’t look right, generate it again, that doesn’t look right, generate it again,
    0:08:46 and eventually it spits out a video that looks exactly what you had in your mind of what a
    0:08:51 unicorn flying through space with a cat writing on its back would look like, right? Like, I think
    0:09:01 we are still at the like mid-journey v2 level of AI video, right? And if you remember, if you
    0:09:07 followed along to like AI art, mid-journey from v, I think it was v3 to v4 was this massive leap,
    0:09:11 right? Once v4 came out, that’s when people started fooling people on Facebook with like
    0:09:16 the pope and like a big old puffer jacket and stuff like that. And it was this giant leap where
    0:09:22 all of a sudden people no longer could tell what was AI generated and what wasn’t. I think we have
    0:09:28 not seen that leap with video yet. And I think in 2025, we will see that leap where it gets so good
    0:09:34 that, you know, people didn’t even realize it coming. And the pace at what this happened
    0:09:39 is so mind-blowing to me, because like you were mentioning, very beginning of 2023, we were
    0:09:45 looking at tools like model scope and zero scope, and they were generating like two second clips,
    0:09:51 and they were like 280p resolution, right? Like, they were like 280p.
    0:09:56 -The object is kind of jittering around. -Yeah, and like you can barely tell what they were
    0:10:01 actually happening on the video. And that was like 23 months ago,
    0:10:07 and now we’re getting what we get out of VO. Like, holy crap, like that was a massive leap in two years.
    0:10:12 Yeah, and then the crazy thing is if you look at like public perception, people are like,
    0:10:19 “Yeah, Sora sucks.” And like for me, it’s just like, “Oh my god, I was, you know,
    0:10:25 one of the first people like tweeting a lot about AI video on X, and I remember the comments of,
    0:10:29 “Oh, but it’ll be like 10 years before you get anything that’s even close to like a cinematic shot.”
    0:10:32 -No, we’re there. -No, we’re there.
    0:10:36 Yeah, if you just need like a two or three second like B-roll shot for something,
    0:10:40 it might take a few re-rolls, but you can get it right now, you know?
    0:10:44 Yeah, and the fact that there’s no like kind of reasoning model on top of that, like really
    0:10:49 judging the output, and like as soon as you apply that, like you said, it’s going to get way better.
    0:10:53 Now, one thing that with all of this is going to be interesting is the cost.
    0:10:58 Like, okay, it can get way better when you reason about the output, like how much does that cost?
    0:11:03 And we’ve already seen like with like, you know, okay, with AGI, they eventually want to make it
    0:11:08 where everyone has access to it. But it seems likely that like people who have money are going
    0:11:14 to get access to that way sooner, just because of the cost to run all of this. Like Pro is $200.
    0:11:20 I think that the Chief Financial Officer of OpenAI, she said, “We’re open to experimenting with
    0:11:25 things like $2,000 a month.” And I think you probably will see that, right? You’ll see like,
    0:11:29 okay, there’ll always be a model that everyone can use. It’s like around like a 20 bucks a month
    0:11:33 or whatever. You might have something slightly better, you know, or like, you know, significantly
    0:11:39 better for like 200 bucks a month. And if you’re a professional, you might be like paying two grand.
    0:11:43 Yeah, well, I mean, most of them need to have free models, right? You can use Claude for free,
    0:11:48 I believe, right? You can use ChatGPT for free. You can use Grock for free now. That’s brand new.
    0:11:54 Grock only became free like two weeks ago or something. Like Gemini, I’m pretty sure, yeah,
    0:12:00 they have Gemini and Gemini Advanced. The Gemini Not Advanced is free, right? I think you could even
    0:12:06 test the Gemini 2.0 right now for free inside of their AI studio, right? So there’s always going
    0:12:10 to be like that free plan that’s like, anybody can access it, but it’s not going to quite be the
    0:12:16 state-of-the-art model. And the more state-of-the-art, the further you go up that sort of tier of how
    0:12:22 state-of-the-art it is, the more expensive it’s going to be. Yeah, I mean, so from using O1 Pro,
    0:12:27 which in my opinion right now is the best by far, it takes a long time to give you answers back.
    0:12:33 And so something I’ve been realizing is, while you’re waiting, like, what do you do? Because
    0:12:37 sometimes it’s like taking like three to five minutes and it’s like, okay, you know, I go, what
    0:12:40 do I do? It’s like, it’s like back in the day when you would code, it takes something would take a
    0:12:45 long time to compile and you’re like, okay, what do I do while that’s compiling? I go and check
    0:12:47 something. I mean, I still run into that when I’m trying to render my videos, right? It takes
    0:12:52 10 minutes to render out a video that I make for YouTube or something. And I’m like, well,
    0:12:58 my GPU is in use right now. So like, I can’t really do a whole lot with my computer. So I go
    0:13:01 walk away and do something else, you know, right? Well, but in this case, you’re using their GPU,
    0:13:07 so you can still use your computer. That’s true. And so, you know, it slightly reminds me of like,
    0:13:11 so when I was younger, I think I told you this before, like probably 15, I made money playing
    0:13:14 EverQuest. And there was a lot of people who made money playing EverQuest back in the day.
    0:13:19 I was as far as I know, I was like a top 10 player on the game. I was like, I was a well-known
    0:13:24 necromancer, dark elf necromancer on the game. And one of the strategies we would employ is, you
    0:13:29 know, like if you went with a party in the game, and you were with a bunch of other people, if
    0:13:33 something good dropped, you had to roll on it. And that kind of if you were doing it to make money,
    0:13:37 that sucked to have to share it with those people. And you might get unlucky and you might not get
    0:13:43 this stuff. So what became very common was to multi box, which was basically where you would
    0:13:48 be running multiple computers, multiple accounts, and then you get a huge financial benefit from
    0:13:52 doing so. And you would also make alliances with other people who were multi-boxing,
    0:13:58 right? So like maybe if it was like, okay, it’s hard to run six computers, but like the, you know,
    0:14:02 if we could put five people, maybe one person runs three, one runs two, and then that’s a better
    0:14:08 split than six random people that I’m sharing with. And so we would do that. And you would make
    0:14:12 a lot more money from doing that. And so I do wonder like, okay, in this new world, where
    0:14:18 if you can spend more money, you get way better outputs, but you’re still tied to like the one
    0:14:24 account, will we see like some really clever people start multi-accounting AI?
    0:14:29 Yeah, yeah. Well, you know what you’re describing here too, it is also, and I don’t want to get
    0:14:34 two in the weeds on this, but it also feels like a fairly decent use for like blockchain tech,
    0:14:40 right? Because you can distribute the compute, right? You can have other people like you ask
    0:14:45 O3 a question and it goes off and it uses somebody else’s account and they get sort of tokens back
    0:14:50 for allowing to use the account kind of thing. I can see something like that where it’s almost
    0:14:56 like this distributed network of you’re using other people’s O3 accounts to generate what you
    0:15:00 want to generate, because they’re not currently using their account. And whenever their account is
    0:15:07 inactive, you use it and you pay some sort of cryptocurrency to use it and that’s their form
    0:15:12 of mining, you know? Yeah, that sounds good conceptually. I’ve worked with blockchain stuff
    0:15:20 in the past. It’s often harder than it sounds, but I guess I’ll ask O3 or O4 for advice on
    0:15:25 if that’s the best way to go and we’ll… Yeah, maybe AI will develop that software for us instead
    0:15:31 of a human actually developing it. Do you think O3 is AGI? Like from my definition of AGI, which I
    0:15:36 guess is like, can do most things humans could do. If you didn’t put all this stuff into the models
    0:15:41 where they tell you that they’re AI, I’m not a human. If you didn’t have any of that in there,
    0:15:47 I think you could trick most people into believing that this was a human doing stuff for you. I
    0:15:54 really do. And I already thought like 01 Pro was close to that. So like based on the benchmarks
    0:15:59 they’re showing for O3 where it’s past what the ARC AGI benchmark where it’s got like a
    0:16:08 87.5 and the average human taking it gets like an 85. That’s for me pretty close to AGI.
    0:16:13 I could call that the beginning, like a first version of AGI to me.
    0:16:17 Yeah. Well, I looked it up on a live stream earlier today where we’re talking about O3
    0:16:23 and that ARC AGI benchmark, the average human gets about a 76% on it.
    0:16:31 Oh, really? Okay. And so if we look at the O3 model that just came out, the O3 low model,
    0:16:36 what do they call it? There’s like a smaller model that used less training and then like a
    0:16:43 larger model. The smaller model got a 76%. So the smaller model hit the average that
    0:16:49 a normal human would hit. The larger model, the one that used the insane amounts of compute,
    0:16:55 got like an 86% or something like that. So smarter than the average human.
    0:17:01 Yeah, which is wild. This is it looking at complicated logical puzzles and things like
    0:17:05 that and solving them. For anyone who’s saying that these models do not reason and they’re just
    0:17:12 like parrots or whatever, no. There is some reasoning going on here and this is the breakthrough.
    0:17:16 Like this is the fact that you now have like the large language model, but you’ve got this computing,
    0:17:21 you know, test time computer, whatever they’re calling it. Like that’s the big unlock where
    0:17:25 like you combine the two. It’s like the two sides of a brain. Well, I just pulled up this chart.
    0:17:33 This is from the ARC prize website here about OpenAI’s O3 model. So their low model scored about
    0:17:38 what an average human would score. But one thing that I find super interesting about this is the
    0:17:44 cost per task down here, right? So it’s like this low model, you know, this is a logarithmic scale,
    0:17:49 right? So it goes from $1 to $10 to $100 to $1,000. So if you were to like chart this,
    0:17:55 it would have like a sort of curving up slope, right? Yeah. So this isn’t like
    0:18:04 $11 for 76%. This is like, this is probably something like $30 per task right here, if I
    0:18:11 had to guess, right? And if you look at the O3 high model, this is $1,000. If there was another line
    0:18:16 drawn right here on the chart, that would be at $10,000 following the, you know, the same sort of
    0:18:24 logarithmic scale. So this 88% is actually halfway between $1,000 and $10,000. So in the range of
    0:18:32 somewhere around $5,000 per task to be able to get that O3 high model actually working,
    0:18:38 yeah, very, very, very expensive right now to be able to use this model as of right now. I mean,
    0:18:43 like my understanding, like the amazing thing about this is at least from like what people are
    0:18:48 like kind of, you know, inferring from like tweets from people from OpenAI is that O3,
    0:18:54 which they skipped O2 because of some trademark stuff apparently. Well, there’s O2 Arena in London,
    0:18:58 right? That’s, there’s a big company called O2 that owns that arena licensing and yeah. Yeah.
    0:19:05 So it seems that O3 is literally just O1 with more compute thrown at it. And if that’s the case,
    0:19:08 I mean, yeah, that’s what you would expect. The cost to go up dramatically and you get,
    0:19:11 it’s amazing that you can just get more performance from it though. Like you do get
    0:19:17 more intelligence, the more compute you throw at it, which means that we are now in a world where
    0:19:21 like all the stuff we’ve talked about in the past, we’re like, okay, race for more compute, race for
    0:19:28 more energy. Yeah. Those things are majorly in play now because we’ve now proven that you probably
    0:19:35 can get to beyond AGI just with more compute. That’s been proven now. And so that’s, that’s a
    0:19:40 big revelation like, okay, we thought that might be the case. It appears to be the case now. It is
    0:19:46 impressive. Like in this, it shows on, so here’s a benchmark in competition math, which is like
    0:19:50 the kind of math that like we have no idea how to do. This is like people who are like
    0:19:54 PhDs in math to solve incredibly challenging. You know, this is not just like basic algebra.
    0:20:01 And so like that, it’s, you know, it’s frontier math that GPT-40 scored like a 13.4, but like
    0:20:08 03 got a 96.7. So I mean, this is an incredibly high score. This is like top of the top smartest
    0:20:14 people in the world level. And I found this fascinating, which this is, I think this is,
    0:20:18 I’m not sure if this is like an official like who’s came up with these IQ rankings for the AI,
    0:20:25 but some people ran some numbers and it was shown that like, you know, GPT-40 scored like an estimated
    0:20:31 115 IQ, which is considered like a somewhat smart person, not like a super smart, like a
    0:20:37 relatively smart person. 130 or over is considered a smart board, you know, maybe 140 is like
    0:20:43 plus is genius. Einstein’s estimated to be like around 160 IQ. So the average IQ range for adults
    0:20:51 is between 90 and 109, with 100 being the theoretical average. So 100 is about what you’d
    0:20:59 expect average to be. Yeah, yeah. Yeah, in America, average is about 100. So it’s estimated for 03,
    0:21:05 157. And like I said, you know, it’s estimated, it’s, you know, IQ, it’s highly debated, like
    0:21:10 how useful it is. But, you know, Einstein was estimated to be 160. And so this is like pulling
    0:21:16 in like close to 157. So that means we’re possibly, you know, in the ballpark where like this will
    0:21:22 actually be helping us solve frontier math problems, like problems in physics and just basic understanding
    0:21:26 of the universe that we live in. This may be approaching that, like by the time we get to 04,
    0:21:33 maybe that is already able to do that. Also, this appears to still be not a larger model that they’ve
    0:21:39 trained. This still seems to 01 and 03 are not dramatically larger models and trained. They’ve
    0:21:44 just came up with this new paradigm of like applying like test time compute on top of the
    0:21:50 existing models. So if that’s the case, well, then maybe what was called Orion or GPT-5 or whatever,
    0:21:58 maybe that’s still in the works. And so maybe by the end of next year, not only will we possibly
    0:22:03 get 03, or maybe it’ll be like a slightly less compute intensive 03. So maybe it’s like three
    0:22:09 times better than 01, you know, Pro or something. Just to clarify, what you’re saying is like,
    0:22:17 they’re still using like a GPT-4 level model, but they’re just throwing more compute during the sort
    0:22:22 of inference phase, right? It’s already, it’s the same sort of trained model that they trained with
    0:22:27 maybe GPT-4, maybe it’s some sort of like 4.5 or something, right? Like maybe it’s a slightly
    0:22:33 newer model. But the inference, when you actually give it a prompt and ask a question, it’s throwing
    0:22:38 a ton of compute at it to like process the question and then double check itself and then
    0:22:43 double check itself and then double check itself until it’s fairly certain that it gave like the
    0:22:49 proper output that you’re looking for, right? But it’s still essentially using like a GPT-4,
    0:22:54 maybe a slightly better model. Yeah, yeah, that’s, that’s what people believe. And that kind of,
    0:22:58 that makes sense to me, honestly. Like if they have discovered this new paradigm of like,
    0:23:02 yeah, not only do you get the result from the LLM, but then you’d like reason on top of that,
    0:23:07 you could see, you know, it appears you can get like major improvements just from that.
    0:23:12 And if that’s the case, it could be that, yes, they do have a, the model that they’ve been training
    0:23:16 for a while that’s on a, you know, has massively more amounts of data that’s been trained upon,
    0:23:21 that could still be in the works. And so when Sam Altman is saying that we’re going to see
    0:23:28 huge increases in 2025, that could be what it is. They still have the new model coming. And then
    0:23:34 they’re just going to slap this new reasoning model on top of that. Yeah, yeah, it’s like GPT-5
    0:23:40 combined with like the ’03 reasoning model. And like all of a sudden ’03 gets even further
    0:23:45 up the benchmarks and does even better. And the only thing that changed was the underlying model
    0:23:49 before the reasoning portion happened. Yeah, because in its baseline is way higher that it’s
    0:23:55 a reasoning upon, right? Right, right, right, right. So you could in theory see dramatic increases
    0:23:59 beyond what we’re seeing in these benchmarks. I mean, so you might see like, okay, yeah,
    0:24:06 we get like a 200 IQ plus model by the end of next year, which obviously dramatically changes
    0:24:11 things because like, you know, I’ve been testing ’01 Pro and people are kind of now saying that maybe
    0:24:17 ’01 Pro is maybe, it’s basically in between ’01 and ’03. It’s like basically like when they were,
    0:24:21 when they were testing ’01. It’s the ’02 that we’re not allowed to talk about. Yeah, it’s basically
    0:24:26 like when you take ’01 and put more compute towards it, what happens, it’s some first version of that
    0:24:32 test. Right, right. And it seems like that, like I think they showed, I forgot where the benchmark
    0:24:37 was, but I think somebody did show that like there was a major difference between ’01 and ’01 Pro
    0:24:43 on that ARC AGI test. And I’ve been seeing that too, like when you throw, like if you throw a lot
    0:24:50 of code at ’01 Pro and tell it to reason about that code and to help rearchitect that code,
    0:24:57 it can do it where ’01 fails. Because one of the things that I keep on hearing come up is like,
    0:25:02 it’s a huge advancement and it’s really cool, but what are the actual practical use cases? And I
    0:25:08 think the most obvious practical use case is probably coding, right? I think that’s going to be,
    0:25:13 like as far as like a general consumer, the most actionable use case. But I do think
    0:25:19 there’s a lot of other use cases for something like ’03 beyond that. They’re just not necessarily
    0:25:23 use cases that the average general consumer is going to care that much about, right? That they’re
    0:25:29 not going to try to tackle themselves using ’03. Things like, you know, trying to solve climate
    0:25:36 problems or, you know, trying to cure diseases or things like that. But I just wanted to hear,
    0:25:41 like from your perspective, like outside of coding, let’s just set coding aside, outside of coding,
    0:25:48 what do you see like ’03 like really helping the world accomplish? I think that when you start
    0:25:54 having agents powered by ’03, that’s when you’ll see like real, like regular people getting benefits
    0:25:59 from these reasoning models. I do think they’re like, okay, if Orion comes out and it is what’s
    0:26:04 powering like an ’04 or whatever, maybe there is like a standalone non-reasoning model too,
    0:26:09 or they have some kind of smart router that knows, okay, I can respond super fast to this basic
    0:26:15 query. And if it’s a hard question, let’s spend some time thinking about it. I hope that’s coming.
    0:26:19 Because right now, like the ’03, like the ’01 Pro, like the average person, if they used it,
    0:26:24 and they’re asking it like how to bake something, you know, or whatever they’re asking it,
    0:26:28 like they’re going to have a bad experience. Like it’s going to be like, oh, this is like,
    0:26:32 you can get just as good of a response out of GPT-4, basically. Yeah, or like maybe,
    0:26:35 maybe ’01 Pro would be like slightly better, but like it would take so much longer that
    0:26:40 like the perception would be that it was not good for sure. And so I think that the magic’s
    0:26:45 going to be when it start powering the agents, you know, because then you can like, okay, I hate
    0:26:50 dealing with email, let’s just have the agent handle all my email. I don’t actually want to do
    0:26:55 email anymore. That thought just makes me laugh because it’s like, what happens when everybody’s
    0:27:00 doing that? It’s just like, nobody’s writing or checking their emails anymore, it’s just agents
    0:27:06 doing it all for them. I think that’s where we’re heading though. Like, you know, it’ll be super
    0:27:12 rare that you get like a handwritten email from somebody, you know. It’s like, not a handwritten
    0:27:17 letter, but like a handwritten email. Right, right. A hand typed email. A hand typed, yeah,
    0:27:21 hand typed email will be a rare thing, probably in like two or three years. I think that’s where
    0:27:24 we’re headed because like everybody hates email. There’s all these things that we do in our daily
    0:27:30 lives that like almost everybody hates. And there haven’t been major improvement to these systems.
    0:27:35 So I think the systems will continue to exist, but we’ll start to have agents actually do those
    0:27:40 things for us that we don’t want to do. And so I think that’ll probably, I think in 2025, we will
    0:27:45 see O3 come out and it will be what makes agents really work. And that’s when people see a change
    0:27:50 in their daily lives. Once we have robots and like, reasing models are going to be fundamental to
    0:27:54 having robots that can actually do things in your house that you feel safe about.
    0:28:00 Yeah. And where I think like the sort of like O1 and O3 are really powerful for agents is the
    0:28:04 fact that they’ll essentially double check themselves, right? Like going back to the coding
    0:28:10 thing, right? It can write code for you and then look at your code and look to see if it made any
    0:28:14 mistakes and then fix the code and then double check to see if it made any mistakes and then go,
    0:28:18 okay, I’m pretty sure the code’s right now and give it back to you. Right. So the fact that it’s
    0:28:24 sort of like constantly double checking its work is one of the biggest sort of jumps that I think
    0:28:29 this O1 and O3 are creating is that, you know, the reason it’s taking so long is it’s almost
    0:28:33 like kind of double checking its work and then triple checking and then quadruple checking and
    0:28:37 the ones that are using insane amounts of compute, it’s like, all right, we just double check this
    0:28:42 a hundred times. We’re sure this is the right answer, you know? Yeah. It’s like multiple AIs talking
    0:28:46 to one another and then like finding the best, you know, and double checking everything to make
    0:28:51 sure it’s all accurate. Like it hallucinates way less. Like I’ve been playing around. I’ll show
    0:28:56 you something later. Like a very basic game demo. Don’t expect this is not a real engine five game
    0:29:00 or anything like that. It’s very, very basic. But I’ve been playing around with creating a game
    0:29:05 demo just because I got so excited when I tested O1 Pro and realized how good it was at like actually
    0:29:09 architecting systems with code. But you find like it’s amazing at some things, but there’s still
    0:29:14 some things that Claude’s good at and there’s some things that even like Jim and I 2.0 is good at.
    0:29:19 Right. But I do find that O1 Pro is the best at architecting hard systems, like actually like
    0:29:24 architecting the code, making major changes or refactors to the code and in terms of like
    0:29:29 not hallucinating about helping you solve something. The more and more I tested Claude,
    0:29:33 like it’d be very good at some things. Like some things like in terms of like modifying my UI or
    0:29:37 stuff like that. Claude was the best for some reason, but also it would just hallucinate your
    0:29:41 really weird things. Like I would like show it a bunch of code and say, “Hey, what do you think is
    0:29:46 the problem with this? I’ve got a bug. Here’s the bug.” And I was like, “Oh, you need to change this.”
    0:29:51 And now like I look at my code and that’s already with my code. It’s like you need to
    0:29:55 change this to this. And I look at my code and it’s like, “What you’re telling me to change it to
    0:30:00 is my code.” Yeah, I’ve seen that in cursor before. I’ve seen that because I have my cursor set up
    0:30:06 for Claude. Was it Sonnet, right? I have it set up for Sonnet. And I’ve seen that before where it’s
    0:30:11 like, “Oh, I spotted the problem. You need to change this to this.” And I’m like, “They’re the same thing.
    0:30:16 You didn’t change anything.” I know. And it’s like, so that’s a crazy hallucination where it’s
    0:30:19 literally like what you just pasted to it. You’d imagine that what you pasted was something else
    0:30:26 and that what you pasted was a solution. Yeah, yeah. And I have not had O1 Proc ever do that.
    0:30:33 So that’s a major change where like you can like O1 Proc likely catch it. It might still hallucinate,
    0:30:37 but then it catches its hallucination before getting back to you, right? Because there’s a
    0:30:41 lot of stuff that’s going on with the O1 model behind the scenes, right? That they’re not showing
    0:30:47 you the entire thinking. So it’s very possible that it is hallucinating, then double checking
    0:30:53 what it just responded with, notice it hallucinated, and then fixed itself before responding to you.
    0:30:58 Yeah. Yeah. So that’s going to make these models way more reliable. Like anything that needs to be
    0:31:02 reliable, they’re going to get reliable. Like because probably the more compute you throw at it,
    0:31:06 the more reliable it will get. The more times it thinks about the answer before it gives it to you,
    0:31:13 probably maybe it’s not going to be 100%, but we’ll get like all approach like 99.9% accuracy
    0:31:16 with these models probably the next year with reasoning. So I guess the big question is like,
    0:31:21 okay, for 2025, like with like models, like all the models are getting really good. And like now
    0:31:25 Google’s like teasing, they’ve got like a reasoning model and they’re kind of making it sound like
    0:31:30 it’s not a big deal. And they’re even saying that they’re probably going to beat OpenAI in terms of
    0:31:36 coding next year. I do wonder if that’s true or if like Google’s overestimating or underestimating
    0:31:41 what OpenAI has accomplished. So like OpenAI is saying like, it’s not as simple as you think,
    0:31:48 like we’ve discovered some things. The thing about OpenAI though, is that I do feel like a lot of
    0:31:53 the other companies could potentially catch up fairly quickly, just due to the amount of brain
    0:31:58 drain they have at that company, right? Like how many people are helping develop this stuff
    0:32:02 and then walking away from OpenAI like pretty close to the time they actually announce it,
    0:32:07 right? We’re seeing it constantly. Yeah, maybe, but I mean, I don’t know. That could just be a
    0:32:11 narrative too though, because like that’s always happened in Silicon Valley. Like talent moves
    0:32:15 around because like, if you’re at a company for a while, what happens is you get some stock there
    0:32:19 and the people at other companies just offer you ridiculous amounts of money. And so then it makes
    0:32:25 sense just to like keep your stock, go get more stock. This has always happened in Silicon Valley.
    0:32:29 Like people usually move around every two years. Like people rarely stay at a company more than
    0:32:34 two years in Silicon Valley. I do wonder, like that could be the case. You could be right,
    0:32:37 but it also feels to me like maybe it’s something that’s kind of been exaggerated,
    0:32:41 just because it’s great for tweets and things like that. And it’s just like more of what’s
    0:32:46 always happened in Silicon Valley. And there’s constantly, like who’s to say they’re not getting
    0:32:52 higher quality talent coming in as other talent leaves, because now they can attract the best of
    0:32:58 the best. Whereas when they started a company, you know, maybe they, you know, maybe Carpathia is
    0:33:01 amazing, but maybe he’s not the best at creating this despite having a huge persona online.
    0:33:07 Yeah. And so in Silicon Valley, that’s always like resulted in like issues with founders and
    0:33:11 people leaving and things like that. I really think it’s going to come down to like three major
    0:33:16 players. I think really it’s going to be Google, Anthropic and OpenAI. And the reason I think
    0:33:20 Anthropic is in the mix is because they’re just, they’re getting infinite amounts of cash from
    0:33:25 Amazon, right? So it’s like, you’ve got, you’ve got Anthropic, who’s essentially owned by Amazon
    0:33:30 at this point, OpenAI, who’s essentially owned by Microsoft at this point. And then Google and
    0:33:35 DeepMind, which, you know, all under the same umbrella. I think you’re really missing XAI.
    0:33:40 Oh, you’re right. You’re right. I am mixing. I do think they are going to, they are going to,
    0:33:43 you’re right. I am missing that one as well. I do think they’re going to do a lot, especially
    0:33:48 with, you know, Elon building like the largest server farm on the planet and whatnot. So yeah,
    0:33:52 yeah. And it’s going to be interesting because it sounds like Elon is thinking he’s going to
    0:33:59 be able to create the largest model. But if OpenAI has discovered that like, okay,
    0:34:03 having a bigger model is good and does improve something, but reasoning is actually the more
    0:34:08 important part. You know, is Elon going to be able to like kind of slightly,
    0:34:11 is he going to be able to catch up and figure out how they did it? Or does OpenAI really
    0:34:15 have some kind of secret sauce there that you don’t take people time to figure out?
    0:34:22 I do often underestimate and take for granted XAI because I feel like they’re still trying to
    0:34:25 catch up, right? Like the current models that we have available to us, which are,
    0:34:30 you know, GROC inside of X and then they have their new like Aurora image model.
    0:34:36 Neither of like their large language model or their image model feel as good as what else is
    0:34:41 available. But also I’m not counting them out because I know how much obviously Elon, I believe
    0:34:46 is still the richest person on the planet. If last time I checked, right? He’s got $6 billion
    0:34:53 in investment. He’s building like the largest GPU cluster on the planet, right? We’re going to see
    0:34:58 them make some waves and forgot to mention, he’s best friends with Donald Trump at this point.
    0:35:04 So, you know, he’s got government inroads as well. So I think, yeah, I do think we are going to see
    0:35:08 a lot more out of XAI. It’s just, I always forget about them because I don’t personally use them on
    0:35:11 a daily basis, you know? Right. Yeah. They don’t have the consumer adoption yet, but like people
    0:35:16 are showing that like that has been going up. I think I saw a chart showing that it was going
    0:35:21 up a lot in Japan, which makes sense because in Twitter, X is huge in Japan. But like I saw like
    0:35:25 searches for GROC as well, where it’s like going up dramatically in Japan. I think Elon last year,
    0:35:31 that is like almost like it was like a global trend. He showed that chart and people were like,
    0:35:35 that’s Japan, but you know, but Japan’s very important. So like a top three market. And so I
    0:35:39 am curious to see how that plays out. But yeah, I think we have at least four major players
    0:35:45 going after it. And I think all of them are going to make major progress in 2025. So,
    0:35:50 which will make all of them make progress faster. They’re all, you know, people have said like,
    0:35:56 oh, open AI had to show 03 on the final day of the 12 days of open AI, because Google has been so
    0:36:00 freaking impressive with their announcements over the last month. Do you think that they had something
    0:36:05 else intended for the 12th day? And they pivoted to show off 03 because of Google like sort of
    0:36:09 matching them day for day with something equally as big, if not bigger? I think it’s, I think it’s
    0:36:13 possible because you know, it was also you started seeing, you know, we’ve had Logan on the podcast
    0:36:17 before and he was starting to like tweet out some stuff, you know, kind of kind of hinting
    0:36:23 that like Google’s not behind, maybe they’re maybe they’re ahead. And then like when 03 was
    0:36:27 the benchmarks were shown for 03, like all these open AI people started sharing memes like, you
    0:36:32 know, opening AI was never behind. You thought they were behind. Everybody’s saying they’re behind.
    0:36:40 What are we behind? You know, the most used model and with like the stuff they have internally,
    0:36:45 steal the best. Well, they needed something like that to really sort of get people excited again,
    0:36:49 because I feel like during the 12 days, there was a couple cool announcements, but most of them were
    0:36:57 kind of like most people, it doesn’t impact them too much. Like I thought the 100, 1800 chat GPT
    0:37:05 was sort of like a gimmicky novelty sort of thing. We got Santa voice. Yeah, the Santa voice.
    0:37:10 I mean, I think the big moves they made were putting out Sora, but then Vio came out like
    0:37:16 two days later and sort of knocked Sora off the top of like being the best AI video tool at the
    0:37:21 moment. Sora turbo, right? The other sort of might still be a lot better. I don’t know. But then
    0:37:26 they also announced the vision inside of the advanced voice mode. And I think that was probably
    0:37:33 the most impactful thing that they did was the advanced voice mode with vision. We also got the
    0:37:40 $200 a month plan, which gave us access to the GPT or the 01 Pro. Yeah, I mean, for me, 01 Pro is
    0:37:45 the biggest one. But like I said, like most people, number one, they’re not going to pay $200 a month.
    0:37:49 And number two, even if they did, you can see that everybody’s getting very different results.
    0:37:54 Like it depends on what you’re trying to use this for. The card coding stuff or science,
    0:37:58 people are blown away. If they’re trying to use it like they use just chat GPT on a regular daily
    0:38:05 basis, they’re like, this sucks. Yeah. Yeah. Yeah. Yeah. Yeah. I’m curious. I want to see this this
    0:38:09 game that you’ve been talking about. It’s super, it’s super simple. So don’t expect much, but I’m
    0:38:18 not a game designer at all. I’m expecting Baldur’s Gate 4 meets Elden Ring. Baldur’s Gate meets Elden
    0:38:25 Ring. Let’s see it. No, no, it’s not that. But you’ve been using 01 to develop this or 01 Pro.
    0:38:31 I’ve been using 01 Pro. I’ve been using Claude. I’ve been using Gemini 2.0. I’ve been using Sora.
    0:38:38 Are you using an IDE like a cursor or visual studio? I’m using cursor, but I feel like I’m
    0:38:42 kind of a newbie using cursors, to be quite honest. Now that they added, you can put 01
    0:38:47 into cursor. I do want to try it again, because that probably better than the outputs I was getting
    0:38:52 from Claude. I feel like right now, 01 is like slightly better than Sonnet. 01 Pro is significantly
    0:38:56 better, but way slower. But I’ve been using everything. That’s why I was saying about the whole
    0:39:06 multi-boxing thing is multi-accounting. I think the wizards of AI, they are going to be putting
    0:39:10 together all these different tools and maybe multiple LLMs based on what they’re good at.
    0:39:15 It’s happening already. I think you’re going to see ChatGPT give you an output,
    0:39:19 and then it spits it over to Gemini to double check it, and then it spits it to Claude to double
    0:39:24 check it. It goes round robin through all these various LLMs and then spits back out
    0:39:31 like a sort of aggregated response based on what all the LLMs said. This is very basic,
    0:39:38 so don’t judge me. This is all super ugly, like you train a character, you go into a little town,
    0:39:43 which is like all mid-journey generated. It looks sick, dude. It does look really good.
    0:39:51 And dude, I’ve had mid-journey. I use mid-journey for this. The special effects, I asked open01pro
    0:39:57 what to do, and they’re just like, “Okay, let’s put some snow here. Okay, cool. Let’s do that.”
    0:40:01 So it’s like a static image, but then some coated in snow over the front.
    0:40:05 Of course, you would make it look better. There’s this one game called Deepest Dungeon,
    0:40:09 where they do something like this, where you have a static image of a town,
    0:40:13 and when you highlight over them, they light up and stuff like that.
    0:40:16 So I’ll eventually do some kind of special effects where when you go over it, it lights up.
    0:40:21 Here, I go build my party. I have not worked on the UI, so all this is just super ugly,
    0:40:27 but okay, recruit my team, random. So how many hours do you have into this right now?
    0:40:32 It’s a decent amount, but not a crazy amount. I have a hard time estimating,
    0:40:35 because it’s been the kind of thing where I’ll do it while I’m doing other stuff.
    0:40:40 I’ll have open01pro running, and I’ll just be totally doing other stuff. And I’ll be like,
    0:40:44 “Hey, improve this.” Not only like improve this, but sometimes it’s like,
    0:40:48 “What should we improve?” So like, yeah, you hear me some kind of like a world map,
    0:40:53 you know, it’s all not that great for now, but like you’d go in Mario style.
    0:40:57 We have different nodes that you go through to get to the final boss.
    0:41:00 And then what I’ve kind of done gameplay-wise is I was like,
    0:41:04 “Well, what’s something simple I could actually implement without like making,
    0:41:07 you know, Baldur’s Gate 3, which is super complicated.”
    0:41:12 And I was like, “Well, I used to play this game called Puzzle Dragon, actually Taizosan’s game,
    0:41:15 where it’s like basically like a match 3 game with an RPG element.”
    0:41:19 But I always felt like, “Oh, but it doesn’t have like a real RPG elements in it. Like,
    0:41:22 they’re very, very light.” I was like, “What are you made of this?” So like,
    0:41:27 in between Baldur’s Gate, like more RPG elements, and like a basic puzzle,
    0:41:30 match 3 that everyone understands, and some people love, people hate.
    0:41:33 And so I just created something like that. And then you have like,
    0:41:36 you basically have like a party here, like your party with like health bars,
    0:41:39 it looks really good. All this is mid-journey art.
    0:41:44 This is all mid-journey. Yeah. I used, I started to use scenario a little bit,
    0:41:48 but this is mostly mid-journey. And, you know, and it’s just like a very basic match game where
    0:41:53 you charge stuff up, you got different abilities. I hadn’t worked on the special effects at all.
    0:41:56 I literally just asked 01pro, like, “Put some special effects in there.”
    0:41:59 It looks really good. It’s like a, like a-
    0:42:02 All the special effects, it just came up with. I didn’t, I didn’t do any of that yet.
    0:42:08 It’s almost like a Candy Crush kind of concept in the actual like puzzle game almost,
    0:42:12 but then you’ve got a whole bunch of like fantasy RPG elements also.
    0:42:15 Yeah, yeah. I’m not going to just play the entire game logic.
    0:42:18 You’re getting hooked on it. You’re like, “I’m playing now. I’m doing good.”
    0:42:23 This is not work now. Just play games. You know,
    0:42:26 then you like have different choices for going to different rooms.
    0:42:30 I probably would have the AI, I’d probably like actually like generate the dungeon,
    0:42:31 so it’s not the same every time.
    0:42:32 Yeah, yeah.
    0:42:35 And then you have different, you know, different stuff you buy.
    0:42:40 And so, to be honest with you, I’m kind of surprised how good it is.
    0:42:44 It looks really good. I mean, everything, I just, I look pretty solid.
    0:42:50 Yeah, dude. Like considering like, I’ve literally, that’s like 90% 01pro
    0:42:54 with little bits of Gemini 2.0 and Claude for different uses.
    0:42:58 Well, you lowered my expectations where you were like, “Yeah, don’t expect too much.”
    0:43:01 And then what you showed me is better than I expected.
    0:43:03 So I think it’s been kind of good for me to do this too.
    0:43:05 Because, you know, I think I told you off camera.
    0:43:09 Like I’ve been kind of decide like, what else should I work on beyond the podcast?
    0:43:13 Like, should I be building another startup or should I be doing a YouTube?
    0:43:17 And I kind of decided like, well, I think I just want to be playing more of these AI tools.
    0:43:21 I think it’ll just be more useful for the podcast for me to actually have hands-on experience.
    0:43:24 Maybe these things turn into real projects. Maybe they don’t, but they’re like.
    0:43:27 I can see it spinning off into a real like game studio.
    0:43:28 And you’re actually selling them.
    0:43:32 It’s been really useful for me though, to actually understand how good, like,
    0:43:35 what are these things good at currently? What are the limitations?
    0:43:43 What would be really cool is if a tool like a cursor sort of knew what each model was good at,
    0:43:47 and then would send whatever your prompt is to the best model for that specific thing.
    0:43:50 Or you could create your own logic that’d be kind of cool too.
    0:43:53 Like, okay, there’s suggestions, but like you can kind of like find, you can like,
    0:43:55 yeah, when I’m doing UI, I want to use this.
    0:44:01 Yeah. I can see cursor building in something like that, where it kind of,
    0:44:05 it’s already model agnostic anyway. You just pick which model you want to use.
    0:44:08 It doesn’t seem like too hard of an extra step to basically say,
    0:44:11 if the question is related to this, send it to this model.
    0:44:13 If it’s related to this, send it to this model.
    0:44:16 Yeah. So this, you know, doing this like little test like demo project,
    0:44:19 like maybe realize like, you know, maybe a year or two ago, I was like,
    0:44:24 oh, you’re going to see like one or two people startups that hasn’t super happened yet.
    0:44:28 But I feel like after like actually trying the newest stuff and like realizing how much
    0:44:31 better it’s about to get, I think we probably are going to get there where you’re going to see
    0:44:35 like one or two people, maybe five, you know, like a small group of people
    0:44:38 creating massive projects stuff that used to would have taken,
    0:44:43 you know, hundred people, thousand people, you’re going to see small teams of people
    0:44:46 with the concentrated focus being able to crank out amazing projects.
    0:44:52 And the fascinating thing is too, is they may be able to do multiple projects as well.
    0:44:56 Like I’ve been finding like, I can totally work on other stuff while I’m doing this,
    0:44:58 because it’s more like, it’s almost more like I’m like the director.
    0:45:01 Yeah, yeah, yeah, yeah. You’re basically going and telling your team members,
    0:45:04 all right, we need to create this next and they go and do it.
    0:45:04 Yeah.
    0:45:06 And you’re just steering the ship.
    0:45:07 You’re not doing the things.
    0:45:12 It is fun. And I think most people don’t know how to tie this stuff together
    0:45:16 as well as I do. So I think that’s going to get easier and more people will know how to do that.
    0:45:20 I totally agree. I mean, we were talking about this before hitting record.
    0:45:23 I think a lot of SaaS companies are going to be in trouble.
    0:45:28 I think it’s, you know, a lot of companies are going to be able to just like build
    0:45:32 tailor-made tools for their needs without needing to go and, you know,
    0:45:36 pay for a SaaS company to do the thing for them, right?
    0:45:38 I’m already finding myself doing that when I have like a need.
    0:45:41 Like I was talking on a previous episode of like,
    0:45:44 I want something that no matter what file type I put into it,
    0:45:46 it converts it to a JPEG.
    0:45:51 So if I throw in a PNG, converts it to a JPEG, a WebP JPEG, AVIF file, JPEG.
    0:45:52 I don’t care what file type it is.
    0:45:55 If I throw it, even if I throw in a JPEG,
    0:45:58 I want it to output it as a JPEG in the selected folder
    0:46:02 so that I can bulk throw images in there and convert them to JPEGs.
    0:46:04 I guarantee there’s already tools out there that do it,
    0:46:08 but it took me three minutes to build that using cursor and Claude.
    0:46:11 So I just built it real quick and, you know,
    0:46:17 I don’t have to go and use some either insanely ad-ridden website for it
    0:46:19 or pay for some SaaS to do it.
    0:46:22 I just have a tool that does it on my desktop now
    0:46:23 and it works really, really well.
    0:46:26 And I think you’re going to see more and more and more of that
    0:46:28 where people are just like, I need this problem solved.
    0:46:32 I’m going to go and just code up a solution real quick
    0:46:36 because all it is is a prompt and the prompt will code up a solution, you know?
    0:46:37 Yeah, it’ll get faster and faster too.
    0:46:40 Right? Like eventually this will be almost instant
    0:46:43 where within a second you’ve got that output of what you wanted.
    0:46:45 A tool that does what you want.
    0:46:48 I think in the ’03 mini demonstration they showed
    0:46:50 where like it basically set up its own server
    0:46:52 and then it set up all this different stuff and it’s like…
    0:46:55 It set up its own server and created…
    0:46:57 This came up with a benchmark for itself
    0:46:59 and then set up a server to run that or is it something like this?
    0:47:00 I built a benchmark.
    0:47:03 Look, I’m really good because I beat my own benchmark that I just built to test.
    0:47:07 But still, it is fascinating that it’s able to do these kind of things.
    0:47:10 Right? It’s mind-blowing.
    0:47:11 Yeah, for sure.
    0:47:14 Yeah, I mean, I think it’s exciting times right now.
    0:47:17 I think 2025 is going to be even more exciting.
    0:47:18 So a lot of good stuff.
    0:47:22 I think this is probably a good spot to wrap this one up,
    0:47:24 our first episode of the year.
    0:47:28 Thank you so much for tuning in for our show in 2024
    0:47:31 and hopefully you’ll be with us for 2025.
    0:47:34 Make sure you subscribe wherever you listen to podcasts.
    0:47:38 We prefer YouTube because we try to make this visual and show stuff on our screen
    0:47:41 and show off what we’re talking about a little bit.
    0:47:45 But if you prefer audio, you can also find us wherever you listen to podcasts.
    0:47:47 So thanks so much.
    0:47:50 Happy New Year to everybody and we’ll see you in the next one.
    0:47:51 Yeah, happy New Year’s.
    0:48:09 [Music]

    Episode 40: What will AI look like in 2025 and how will it change our daily lives? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) dive deep into AI’s future with perspectives on emerging models and innovations.

    In this episode, Nathan and Matt explore the vast potential of AI models like OpenAI’s upcoming o3, the future advancements expected by 2025, and the significant societal implications of these powerful technologies. They discuss potential impacts on the job market, daily life, and how AI-driven tools might handle routine tasks, freeing up human creativity for more complex endeavors.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Basic AGI emerging by 2025, transforming tasks.
    • (05:14) Discussing AI-generated art and personal connection.
    • (08:59) AI video advancements to reach new realism by 2025.
    • (10:41) Improved AI reasoning entails significant cost challenges.
    • (13:04) Top EverQuest player used multiboxing for profit.
    • (18:57) More compute increases performance and intelligence.
    • (21:09) New paradigm might solve frontier math problems.
    • (25:46) o3 agents may eventually benefit regular users.
    • (27:54) Agents constantly self-check, improving code accuracy.
    • (32:04) Frequent job changes commonplace in Silicon Valley.
    • (35:32) Japan remains important; major progress expected 2025.
    • (38:36) Exploring AI tools; curious about updates.
    • (41:08) Created an RPG-infused match-three puzzle game.
    • (44:13) Small teams creating massive projects efficiently now.
    • (45:51) Quickly built desktop tool converts files to JPEG.

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

  • Ranking 13 Of The Most Popular Ai Video Tools (Q4 2024 Tier List)

    AI transcript
    0:00:04 I’ve listed out Pika, Sora, Hotshot, Vio.
    0:00:06 Just incredible to think that none of this is real.
    0:00:09 Like I showed some of the videos to my son yesterday when he was like,
    0:00:10 that’s not real.
    0:00:14 My benchmark up until now was I like to generate various monkeys on roller skates.
    0:00:19 I don’t know why, but it passes my monkey on roller skate test pretty well.
    0:00:20 This one is definitely real.
    0:00:21 Nobody will ever know.
    0:00:26 Hey, welcome to the Next Wave podcast.
    0:00:27 I’m Matt Wolf.
    0:00:29 I’m here with Nathan Lanz.
    0:00:32 And today we’re going to go down the AI video rabbit hole.
    0:00:34 We’ve recently seen Sora.
    0:00:35 We’ve seen Vio.
    0:00:40 We’ve had access to Pika and Luma and all sorts of AI video tools.
    0:00:43 Half of them, Nathan and I have never even touched.
    0:00:47 So we decided, let’s bring on a guest who has played with all of them.
    0:00:51 Put them all through the motions, knows the various differences between them.
    0:00:55 And today we’ve got Tim from one of my favorite YouTube channels,
    0:01:01 Theoretically Media on the show, and we’re going to go down this AI video rabbit hole.
    0:01:05 Look, if you’re curious about custom
    0:01:08 GPTs or our pro that’s looking to up your game, listen up.
    0:01:13 I’ve actually built custom GPTs that help me do the research and planning
    0:01:19 for my YouTube videos and building my own custom GPTs has truly given me a huge advantage.
    0:01:21 Well, I want to do the same for you.
    0:01:25 HubSpot has just dropped a full guide on how you can create your own custom GPT.
    0:01:27 And they’ve taken the guesswork out of it.
    0:01:31 We’ve included templates and a step by step guide to design and implement
    0:01:35 custom models so you can focus on the best part, actually building it.
    0:01:38 If you want it, you can get it at the link in the description below.
    0:01:40 Now, back to the show.
    0:01:44 So let’s start with Pika.
    0:01:49 In my mind, Pika, where I think it really shines is like some of the new effects
    0:01:54 that they’ve put into it, like the ability to like cut like a cake or to like squish it
    0:01:58 or like the hydraulic press, those are kind of like fun for like memes.
    0:02:01 And then they also it is Pika, right?
    0:02:06 That just release that new feature where you can like put your own face into videos
    0:02:09 and then put yourself in like a certain sweater.
    0:02:11 I haven’t really played with those features yet.
    0:02:16 But what do you feel is like Pika’s winning features?
    0:02:18 But what would you go to Pika for?
    0:02:23 I think that lately they do seem to be kind of moving into, yeah,
    0:02:27 that squish it, cut it, cake it, explode it.
    0:02:31 It’s kind of more on it feels a lot more like it’s on their aiming
    0:02:34 for like social kind of like fun tools.
    0:02:38 But I know that’s, you know, you can upload recipes, essentially,
    0:02:40 of like a person, a person, a place and a thing.
    0:02:41 And then, you know, stamp them all together.
    0:02:44 And then that’ll create, you know, a video.
    0:02:46 But my understanding is that it’s sort of on the template side.
    0:02:51 So once again, this is sort of falling into the area of like, you know,
    0:02:54 just cool, quick videos to share on social.
    0:02:59 Yeah. So with Pika, I really think that they’ve got like this sort of meme
    0:03:04 thing down, but when I was trying to play with the newest features in there,
    0:03:06 the ones where you can sort of add your own face and blend it
    0:03:08 with other styles and stuff like that.
    0:03:12 I was actually messing with it on a live stream earlier in the week.
    0:03:16 And I could not get it to generate what I wanted to.
    0:03:19 I was using like two images and then somebody’s like, Oh, it doesn’t work
    0:03:21 unless you actually give it a text prompt as well.
    0:03:23 So I was starting to blend it with text prompts.
    0:03:26 And even then it was still giving me like really funky results.
    0:03:28 And I couldn’t actually get it to generate what I wanted.
    0:03:32 I feel like Pika is one of those tools where I have to re-roll
    0:03:35 so many times before I finally get what I want.
    0:03:38 But eventually I get there.
    0:03:41 Yeah, I think they were the first one I saw that did like cartoon faces
    0:03:43 really well, like even like the expressions in the eyes.
    0:03:45 I think they were the first ones that kind of, you know,
    0:03:46 didn’t perfectly nail it.
    0:03:49 But like at that point, it was way better than runway.
    0:03:52 But since they came out, like all the other models that came out
    0:03:56 and just visually they’re all more impressive, I think,
    0:03:58 like most of the top models now.
    0:03:59 So I don’t know where they go.
    0:04:03 Like it almost feels like Pika should be some kind of new app or something.
    0:04:07 B and C just because I struggle to see a lot
    0:04:10 of the real practical use cases outside of meming.
    0:04:12 I just don’t know whether it should fit in B or C.
    0:04:15 I don’t think it’s D tier because I think it’s like A.
    0:04:18 It’s OG, they were one of the earlier ones, them and runway.
    0:04:23 We’re kind of like the two big OGs that we’re putting out these video tools.
    0:04:27 And I do find some fun usefulness in it in that like meme thing.
    0:04:30 I’ve also used it like the end of the videos where I’ll be like,
    0:04:32 all right, guys, see you in the next one.
    0:04:35 And then like the thing crushes me at the end, you know.
    0:04:38 So like there’s some fun stuff that you can do with it.
    0:04:40 I’m kind of leading towards B tier on that one.
    0:04:45 Look, every video platform is like one update away
    0:04:47 from just becoming the greatest platform in the world.
    0:04:49 So, you know, that’s yeah, yeah, yeah.
    0:04:51 I mean, the last time we did it to your list,
    0:04:54 we we definitely said some stuff to like save our butts
    0:04:57 so these companies don’t hate us and we’ll do it again on this one.
    0:05:01 All of these tools could be D’s right now,
    0:05:03 but are like one update away from being A’s.
    0:05:06 And I feel that about every single one that we’re going to talk about.
    0:05:10 I could see Pika being like a really viral app at some point.
    0:05:12 If they keep going down the route they are now
    0:05:14 or your own self into stuff, maybe it’s not going to be the thing
    0:05:17 that like Hollywood uses, but maybe it becomes like
    0:05:19 a really popular app with teenagers or something at some point.
    0:05:23 So I think I think B makes sense because like right now
    0:05:25 it is actually useful for memes and stuff like that.
    0:05:27 And like you said, like little parts of videos,
    0:05:30 whereas a lot of these, you know, tools, even though they look cool,
    0:05:33 they’re not super useful in a practical way right now.
    0:05:34 So I’d say B.
    0:05:36 Well, I’m going to I’m going to move along here.
    0:05:38 The next one I want to talk about is Hot Shot,
    0:05:44 because Hot Shot, in my opinion, is like the D tier of like the list for me.
    0:05:46 I don’t know what you guys think if you play with Hot Shot yet,
    0:05:49 but I could we see Hot Shot because I haven’t seen it.
    0:05:52 Yeah, I play with Hot Shot and for me, Hot Shot,
    0:05:56 I could not get it to generate anything that looked good for me.
    0:06:01 And they charge $99 a month for it, which is really high for what it is.
    0:06:06 And literally none of the generations I got it to do looked any good at all.
    0:06:10 But saying that we can move this around based on what Tim says,
    0:06:12 because Tim’s probably played with it a little bit more than I have.
    0:06:13 I have, I have.
    0:06:17 I do, you know, those guys are these guys are bootstrapping.
    0:06:21 They are they’re trying to as best they can to get that model up and running.
    0:06:24 I do agree with you, though, a lot of the generations out of it
    0:06:28 always kind of have this to me, at least have this graininess to them.
    0:06:31 Or there’s a lot of morphing inconsistencies with them.
    0:06:35 And they were one of the first that kind of did the facewapping thing as well.
    0:06:41 But it always felt like my face kind of came out a little too much on the red side.
    0:06:45 Like, you know, the the skin tones and color consistency weren’t great.
    0:06:46 I don’t know.
    0:06:50 I was really struggling to get much out of Hot Shot personally.
    0:06:53 Well, what model should we talk about next?
    0:06:56 I actually have not played a ton with Kling.
    0:07:00 But I know everybody swears by how good Kling is.
    0:07:01 It can be.
    0:07:04 I’ve generated maybe a couple of videos, but I I haven’t played with it a ton.
    0:07:07 So I’m going to have to lean on you a little bit for how good Kling is.
    0:07:12 You know, I think between right now kind of leading, at least on the the Chinese
    0:07:15 models right now, I mean, it’s like neck and neck.
    0:07:18 A while back, I did a short film
    0:07:21 that was entirely I generated called Dead Sea.
    0:07:25 It was a Pirates versus Vampires thing.
    0:07:27 It was kind of stupid, but it was fun.
    0:07:30 And that was all generated in Kling one point out.
    0:07:34 Nice. So would you put Kling in like A tier S tier?
    0:07:37 I mean, I think we’re we’re talking top of the top of the chart here.
    0:07:42 So, you know, if we’re if we’re talking about what I would put in S tier
    0:07:49 personally, I think I think Veo, Veo, Veo is it like here for me is is where
    0:07:51 is where Veo is at.
    0:07:53 Maybe another one can make it.
    0:07:58 But that’s where I would put it because the one thing that I do like about Veo
    0:08:02 is that every time you give it a prompt, it’s going to generate four videos.
    0:08:05 So you hear people going, Oh, are these cherry picked?
    0:08:06 Well, kind of.
    0:08:09 But they built the platform sort of designed for cherry picking.
    0:08:12 They gave you four and say, which one’s your favorite of these four?
    0:08:17 Right. So like out of the four, almost every single time,
    0:08:19 I’m impressed by at least one of the four that it gets.
    0:08:21 That is true. You know, I think it.
    0:08:22 Yeah, I think it’s I think it is really good.
    0:08:27 I mean, again, the model itself has only been really released for a couple of days.
    0:08:30 So I think there’s a lot of that’s that’s that’s a generation that looks really good, man.
    0:08:34 Yeah, there is definitely some some more work ahead of it.
    0:08:37 You know, I’ve got a lot of good generations.
    0:08:39 I also do have a lot of funky ones, too.
    0:08:42 Like I tried to get it to generate like somebody doing a Rubik’s cube
    0:08:45 and I definitely got bad to it.
    0:08:48 Yeah. Yeah, not horrible. Definitely not horrible.
    0:08:50 I mean, again, the fingers are not turning to spiders.
    0:08:53 Like true. That is a way to address.
    0:08:56 Yeah. It really sucks at gymnastics.
    0:08:58 Nothing has passed the gymnastics test.
    0:09:00 Yeah. Yeah. I found.
    0:09:02 That’s the new test after spaghetti.
    0:09:04 Yeah, the gymnastics is the new benchmark.
    0:09:09 This is probably my favorite generation is this Rhino walking school.
    0:09:11 That’s like the cameras panning with it.
    0:09:15 You know, my benchmark up until now was I like to generate various
    0:09:17 like monkeys on roller skates. I don’t know why.
    0:09:20 But it passes my monkey on roller skate test pretty well.
    0:09:23 There’s one of the monkeys on roller skates that I got.
    0:09:25 Here’s another monkey on roller skates.
    0:09:28 This is the one that I shared on X that I thought was really good.
    0:09:33 I got it to do like a first person looking out of a jet fighter
    0:09:35 and another incredible thing that none of this is real.
    0:09:38 Like I showed some of the videos to my son yesterday from Vio,
    0:09:40 and he was like, that’s not real.
    0:09:42 Like he’s like, oh, that one’s not real. That one’s not real.
    0:09:44 And then there was one with like an animal or something.
    0:09:48 He’s like, that’s not real. No, it’s not real. Disappointed.
    0:09:49 Yeah. Let’s see.
    0:09:53 One of the ones that really did not work out well was I tried to tell it
    0:09:57 to make a guy playing with his dog with dolphins jumping in the background.
    0:09:59 And that’s like barely even a video.
    0:10:02 Here’s another one that it did.
    0:10:05 And the dolphins are just floating there in the background.
    0:10:09 So it’s not always like the best thing you’ve ever seen, right?
    0:10:12 Like there’s definitely some issues with some of them.
    0:10:15 But again, every single time it gives you four
    0:10:18 and like one of the four is usually pretty good.
    0:10:23 Were you doing, was that a lot of the image to video or is that more text to video?
    0:10:25 That was all text to video.
    0:10:28 Text to video. Yeah, I do find that it tends to generate a lot less wonk
    0:10:31 when you’re doing text to video as opposed to image.
    0:10:36 And one thing about Veo, too, is when you do image to video,
    0:10:38 you can’t upload your own image.
    0:10:41 You have to generate an image with their image generator.
    0:10:45 Yes. And then you can turn the image that they generated for you into a video.
    0:10:49 So I think it’s using their image in three, which looks good.
    0:10:52 I got to say, that’s actually I can’t really complain about like that.
    0:10:57 Image in three has taken a pretty big jump up from image in two.
    0:10:59 It actually looks pretty solid in all honesty.
    0:11:00 Yeah, it is good.
    0:11:02 And it’s got people really well. Yeah.
    0:11:03 I haven’t been able to try this yet.
    0:11:07 So like in terms of a product, like how do you, how does it compare to like,
    0:11:10 you know, Runway’s got a pretty great editor tool.
    0:11:12 Sora now has a pretty awesome storyboard.
    0:11:14 It’s definitely still a beta product.
    0:11:17 It’s definitely still a beta product, you know, like Tim mentioned,
    0:11:20 it does not seem to have like any sort of library.
    0:11:23 So anything you generate, if you refresh your page
    0:11:25 and you didn’t download those videos, those generations are gone.
    0:11:27 You can’t find them again.
    0:11:29 It doesn’t really have any editing features yet.
    0:11:33 It doesn’t have the ability to upload your own images and compared to videos.
    0:11:36 It’s definitely, it definitely beta.
    0:11:39 Yeah, that’s, yeah, that’s where I have mixed feelings on like on the S
    0:11:41 because, you know, Google’s done this so many times
    0:11:42 where like they have amazing technologies.
    0:11:45 I know, you know, you call me cynical the other day, but like I kind of am about
    0:11:49 Google, I know I am, but they haven’t really launched great products though.
    0:11:51 So even though they have great technology, that doesn’t mean it will
    0:11:54 turn into a consumer product that people actually use.
    0:11:56 Yeah, no, I agree.
    0:12:00 It like, I think that’s where the debate would live for, for Vale, right?
    0:12:03 It’s like, as far as the video generation technology goes,
    0:12:05 I think it’s the best right now.
    0:12:07 I think that’s the state of the art right now.
    0:12:11 As far as the fact that most people don’t have access to it yet.
    0:12:14 Who knows when more people will get access to it.
    0:12:15 Who knows what they’re going to charge for it.
    0:12:18 Who knows if they’re going to build a good UI around it.
    0:12:22 There’s still a lot of questions around it, which, you know, I, I can see the
    0:12:25 arguments for knocking it down a tier because of those reasons.
    0:12:26 Yeah, yeah.
    0:12:29 I mean, Sora has like a, you know, great domain store.com.
    0:12:31 Sexy landing page looks beautiful.
    0:12:32 It’s got the storyboard feature.
    0:12:36 I mean, those are all, you know, those matter for like regular people.
    0:12:39 So yeah, but some of the generations are definitely looking really, really good
    0:12:44 out of it. As far as their overall plans, I mean, I think that we will definitely,
    0:12:46 I think we’re going to get a definite release out of it.
    0:12:50 The, there’s already been talked about YouTube integration with it.
    0:12:54 And then Matt, I don’t know if you saw as well, but did a little banner
    0:13:01 on your channel pop up that asked if, if you will allow for your videos to be used as.
    0:13:02 It might have.
    0:13:04 I might have closed it real quick and didn’t pay any attention to it.
    0:13:09 But I knew, I know they have that feature just rolled out where you can decide
    0:13:11 whether or not you want to allow your channel to be trained on or not.
    0:13:13 Which I figure, I mean, I’ll turn around as fair game.
    0:13:15 I mean, I will turn it on.
    0:13:19 I will, I mean, like, look, I’ve like, who wants you prompt a guy, you know,
    0:13:22 in a studio talking to a camera with a bunch of like mid-journey images
    0:13:26 and an AI video, you’re going to train off of AI video, which, you know, whatever.
    0:13:29 You know, but I just, I feel like they had to have already trained on YouTube video.
    0:13:32 Oh, no, yeah, yeah, yeah, yeah, yeah, yeah, yeah, yeah.
    0:13:36 This is like now like covering their asses after the fact, like, hey, do you opt into this?
    0:13:40 No matter when we did it.
    0:13:41 I have been desperately trying.
    0:13:44 That’s another one of the tests that I another one of the tests that I really
    0:13:46 love to do is to, is to try to prompt for me.
    0:13:50 So I, you know, I’ve run that into chat GPT number of times.
    0:13:54 Just a screenshot of me and say, look, describe everything that you see here
    0:13:56 and make me a video or image prompt out of it.
    0:13:59 And I’m constantly writing that just to see, like, how close it gets.
    0:14:02 So far, I haven’t gotten me, but I’m also not Marquis Brownlee.
    0:14:04 So I’m not that level of famous.
    0:14:07 So, yeah, I do think they will roll it out.
    0:14:11 I bet you it’ll, it’ll roll out like into Gemini or something where you just sort
    0:14:14 of can generate videos in line with your chat and things like that.
    0:14:17 Kind of like Dolly, Dolly and chat GPT.
    0:14:20 Yeah, yeah, I think, I think that’s probably where they’ll go with it.
    0:14:26 But I do want to keep on moving along here because we’re about a third
    0:14:28 of the way through these tools here.
    0:14:30 Okay, they still get an S.
    0:14:32 They still get an S because it’s like video wise, it’s the best quality.
    0:14:35 So yeah, I’ll leave them an S just for that reason.
    0:14:40 But moving on, let’s talk about Sora because Sora and another video.
    0:14:45 Sort of, yeah, I mean, but let’s just go ahead and get that one over with
    0:14:47 because I know we probably have a lot to say about that one.
    0:14:51 The thing about Sora and here’s, here’s where I’m struggling to play Sora.
    0:14:55 What we have access to right now is Sora Turbo, right?
    0:14:59 And Sora, Sora Turbo gets us 20 second generations.
    0:15:02 They’ve got some additional frame interpolation built into it so that
    0:15:06 they don’t have to generate as many frames, but it also makes it a little
    0:15:09 less smooth with the, how the videos come out.
    0:15:13 But then you also have like Sora regular, which is like some of the demos
    0:15:16 we saw nine months ago or whatever.
    0:15:18 And those look a hell of a lot better than what we’re getting out of
    0:15:20 Sora Turbo.
    0:15:22 So it’s like, how do we rank it?
    0:15:24 Are we ranking it based on what we have access to now?
    0:15:25 I think you got to rate, yeah.
    0:15:26 What we’ve seen from Sora.
    0:15:31 I think that you have to rank it based on what was released and what is available.
    0:15:38 And to that, I mean, like, listen, I know that there’s a large segment
    0:15:42 of the of the of the population of the AI video generation or video
    0:15:45 population that kind of it landed with a wet fart, you know, like
    0:15:47 people were not stoked about it.
    0:15:51 But I also think that there is still something there.
    0:15:53 Is it worth $200 a month for most people?
    0:15:54 Probably not.
    0:15:57 It probably is one of those things where everybody came in with their
    0:15:59 standard sort of prompt structure, tried it out.
    0:16:02 It didn’t work or it was just like weird, wonky results.
    0:16:06 I think that we just have to spend more time digging into it and figuring
    0:16:07 out what it’s really good at.
    0:16:13 There I did the video, the video aspect, the remix and the blend and the fact
    0:16:15 that you can cut, there’s a lot of power in that.
    0:16:20 Probably those are features that we haven’t seen in any other video model.
    0:16:22 So, yeah, there is really good stuff in there.
    0:16:26 Again, I know the controversy side of it is just like you can’t generate
    0:16:28 people on the on the $20 plan.
    0:16:32 Yeah, there’s a lot of you can’t you can’t you can’t unless you pay $200 to.
    0:16:36 I have not gotten anything that I’m super impressed with out of Sora.
    0:16:40 If I’m being honest, you know, here’s my monkey on roller skates.
    0:16:46 That was probably the best prompt that I got out of Sora.
    0:16:51 But then my Vio Vio prompt with a monkey on roller skates like blew this
    0:16:52 one out of the water.
    0:16:56 Yeah, I asked it to do a wolf howling at the moon and it literally
    0:17:00 just gave me like images of wolves howling at the moon, not even a video.
    0:17:04 And then I started to learn, all right, you need to give it more detail
    0:17:06 in your prompts if you really want to get something exciting out of it.
    0:17:11 So when I started, you know, putting a lot more detail into these prompts,
    0:17:15 like you could see how long of a prompt that is on my screen.
    0:17:18 The video started coming out quite a bit better.
    0:17:22 But still, I don’t think they’re on par with what we’re getting out of Vio,
    0:17:24 if I’m being honest.
    0:17:25 Yeah, no, I would agree with that.
    0:17:28 And that I think definitely comes down to training data.
    0:17:34 I mean, obviously Google likely has access to the entirety of YouTube.
    0:17:41 And, you know, so, yes, well, if you believe Miramarati did not have
    0:17:46 access to all of YouTube, but yeah, or, you know, Turbo might also,
    0:17:47 that’s actually kind of cool.
    0:17:50 Turbo actually might not have as much of that data in there.
    0:17:55 I think the product itself, though, is probably the best looking AI video
    0:17:57 product, though, in terms of like the user interface.
    0:17:59 100%, like how they present it.
    0:18:04 Yeah, there is definitely a very mid-journey-esque aspect to when you
    0:18:06 first log in and see that sort of end the scroll.
    0:18:10 But yeah, in that, you know, again, the remix, just all of the little
    0:18:13 pulldowns, the simplicity of the prompt, like, no, it looks great.
    0:18:18 I think, again, this is one of those things where I always try to be
    0:18:23 a little gracious with them when they release a V1 product.
    0:18:27 And, you know, like, essentially we’re beta testing their V1 right now.
    0:18:30 By the time this hits, like, well, you know, again, remember how weird
    0:18:33 and wonky mid-journey V1 looked to where it looks now.
    0:18:37 So where, what will Sora 6.1 or 7 look like?
    0:18:39 That’s, you know, the spider one is cool.
    0:18:43 Yeah, I pretty much did all of the exact same prompts with Sora that I did
    0:18:46 with, like, Veo, because I wanted to compare it with those.
    0:18:47 Yeah.
    0:18:52 And for the most part, like, Veo won on every single generation.
    0:18:52 Yeah.
    0:18:58 So I can’t put what we currently have from Sora and S tier.
    0:19:02 Maybe what I’ve seen from Sora, some of the videos they gave us six
    0:19:06 months ago, nine months ago, some of those might fit in S tier, but what
    0:19:12 we’ve got now, I think what it’s got going for it is the UI is really cool.
    0:19:14 The storyboard feature is really, really cool.
    0:19:19 I like that a lot, but at the end of the day, it’s like most of my generations.
    0:19:22 I don’t feel like I would actually use as, like, a B roll over one of my
    0:19:24 other videos or something like that.
    0:19:26 I still feel like it’s, like, looks like the second best.
    0:19:29 I don’t know, maybe like the top of A or something like that.
    0:19:30 Well, I haven’t used Kling a whole lot.
    0:19:36 So it’s like, how would you, like, if you were ranking Veo, Kling and Sora,
    0:19:38 like, how would you rate those?
    0:19:43 If I was to do all three right now for my use cases and for sort of, like,
    0:19:48 the, the pack that I travel with, it would actually probably be Kling on
    0:19:49 the S tier between those three.
    0:19:53 We’ll probably Kling on the S tier only because you could generate images
    0:19:58 from outside of, of, of the, of the platform.
    0:20:01 So on Veo, you have to generate in, or Veo, you have to generate
    0:20:06 in Imogen, whereas in Kling, they’re like, you can generate in stable diffusion.
    0:20:07 You can generate in flux.
    0:20:09 You can generate in midger.
    0:20:09 They don’t care.
    0:20:13 It’s just that you bring it in and from there you have motion brushing
    0:20:15 and all of your other various tools.
    0:20:20 So there’s a lot more control of what you’re choosing to input in with Kling
    0:20:22 than there is with, you know, Veo right now.
    0:20:26 Or, well, Sora, you can actually generate, you can actually bring your own
    0:20:27 images into as well.
    0:20:30 I’m noticing it’s still going to take a little bit more testing, but they tend
    0:20:33 to begin with your initial image first, and then they kind of wander off
    0:20:34 into their own world.
    0:20:37 Like the first frame will be like, you know, oh, this is exactly what
    0:20:40 you’re looking for, but I’m going to do something complete, not completely
    0:20:43 different, like something that’s based off of what I see here, but it’s not
    0:20:45 going to be, that’s not going to be the first frame.
    0:20:47 I’m not going to continue on with that.
    0:20:52 Which I mean, that may be just the new way of generating a video.
    0:20:58 But in some ways that the video, those like this, this wave of video models
    0:21:02 is kind of saying like, listen, we’re the director and we’re the production team
    0:21:06 here, like you’re writing this, but we’re in charge of actually making the thing.
    0:21:09 So it’s almost like, you know, you’re sort of like a producer or a writer.
    0:21:12 And, you know, you’re handing out a script to a production company in
    0:21:14 like Bulgaria, and yet you don’t know what’s going on.
    0:21:17 They’re just sending you back stuff and you’re like, well, this isn’t exactly
    0:21:19 what I asked for, but I guess it kind of works.
    0:21:23 So we’ll see, we’ll see, you know, and I would say the other thing that I would
    0:21:27 definitely say on the Sora side is that was a prompting part of it.
    0:21:30 It is weird because I do think, again, like Matt, like you said, you need
    0:21:34 those longer prompts, whereas with like, as we have seen with, well, we’ll
    0:21:38 look at many max later on and with Kling, there are sort of those AI enhancement,
    0:21:42 you know, like tools that are built in there that you can do something like a
    0:21:43 monkey riding a skateboard.
    0:21:47 And then the the model will fill in a lot of the additional details that it
    0:21:48 needs to generate the image.
    0:21:52 It is so weird that OpenAI did not put that in.
    0:21:54 Like, why isn’t ChatGPT right in there?
    0:21:55 So I don’t know.
    0:21:58 Yeah, Sora Turbo feels kind of like rushed out.
    0:22:01 Like, I think they were pressured because like the public, you know,
    0:22:03 consensus was like, what the hell is going on with Sora?
    0:22:04 It looks so cool and we haven’t seen anything.
    0:22:06 They don’t ship anything.
    0:22:09 So then they had to get like a cost-efficient model out as quick as
    0:22:13 possible and it did leak like a week earlier and got enough people managed
    0:22:15 to get access to it a little bit early.
    0:22:19 And then they had their 12 days of Christmas and it was like, all right,
    0:22:22 well, I guess this is probably the opportunity to push it out.
    0:22:23 But yeah, for sure.
    0:22:26 Well, you mentioned Halu AI.
    0:22:29 Let’s like, is it Minmax, Minimax?
    0:22:30 I always just call it Minimax.
    0:22:32 Like it’s just because that’s Halu.
    0:22:36 I think, again, the actual name of the, and then Minimax, I think,
    0:22:39 is the name of the video model.
    0:22:41 So yeah, I got you.
    0:22:42 Yeah.
    0:22:46 So I always, I always call it Halu, but I always just do it almost
    0:22:48 to like make fun of it when I make videos, right?
    0:22:51 Like I just sort of screw up the name on purpose because I think it’s funny.
    0:22:56 But so where would you rank Minmax or Minimax?
    0:22:58 Like that’s one I have not played with yet at all.
    0:23:00 Oh, really?
    0:23:04 Literally, my entire audience would come at me with Pitchforks
    0:23:06 if I did not say that was S tier.
    0:23:10 Minimax is definitely the one that I think is like current crowd favorite.
    0:23:12 Everybody loves it.
    0:23:12 I love it too.
    0:23:13 It’s a great model.
    0:23:15 Would you put it above Veo?
    0:23:17 Yeah, currently.
    0:23:17 Yeah, for sure.
    0:23:18 Really?
    0:23:20 So see, I haven’t I haven’t I haven’t played with Minimax,
    0:23:24 but like all the examples I’ve seen, like all the the demos, like, I don’t know,
    0:23:26 like Veo seems like way better than Minimax.
    0:23:30 Yeah, maybe. Well, again, with Minimax, I think that the the trick.
    0:23:32 You can use external images and stuff like that.
    0:23:33 Yeah, it’s the control.
    0:23:39 Like, you know, there’s like there’s so much control in there and prompt adherence.
    0:23:41 Yeah, I mean, they’re not as worried about being sued.
    0:23:44 Like, definitely, that’s a problem that Google and OpenAI have.
    0:23:47 Is that, you know, making sure they’re not being sued.
    0:23:51 You can’t just like copy people’s faces and make entire films out of it.
    0:23:54 So, yeah, these were I was running.
    0:23:59 Yeah. So, you know, bad tests here, but I was running the the Sora Viking thing,
    0:24:02 trying to get that kind of look out of it.
    0:24:05 And again, I mean, it’s like it’s, you know, compared to that Sora leak demo
    0:24:09 that we saw earlier, yeah, of course, it’s not going to be as great.
    0:24:12 But like that looks pretty smooth and good.
    0:24:15 The other thing that I think it’s really good at, I just did a video not too long
    0:24:18 ago, it was about an interview.
    0:24:21 So it was this guy and this guy.
    0:24:23 So it was a job interview.
    0:24:24 So this was our interviewer.
    0:24:29 And then this was, yeah, this was a hitman that he was hiring or that
    0:24:33 was going out for a job interview and, you know, the running joke.
    0:24:35 There’s like, what kind of experience do you have?
    0:24:38 And he’s like, well, I’m very patient, you know, I guess for hours.
    0:24:45 So and I used Runways Act one to do all of them and then 11 labs to do all the voices.
    0:24:47 So this kind of stuff, it’s really, really good at.
    0:24:49 And then we had like various flashbacks.
    0:24:51 He’s like, I have an international experience.
    0:24:52 My wife was killed in Thailand.
    0:24:56 Like, we meant to run away from this life together.
    0:24:58 But then the triads killed her.
    0:25:04 You know, like, so this is the kind of stuff that I think it does extremely well.
    0:25:09 Like, you know, for the kind of doing these short clips, like there’s like, I mean,
    0:25:12 like, yeah, Mini Max’s is really on point.
    0:25:15 It really does a great job of understanding exactly what you’re looking for.
    0:25:17 A lot of these are running, by the way, no prompt.
    0:25:19 Like, I’m just like, I’m not even putting in a prompt.
    0:25:21 It just it’s like, oh, you have a city there?
    0:25:22 Yeah, I know you kind of want to do a thing.
    0:25:24 No talking minimal head movement.
    0:25:27 You know, you don’t have to go crazy with your prompts.
    0:25:30 Like, I think the girlfriend here was no prompt.
    0:25:32 It knows waves.
    0:25:34 It knows, you know, move her around.
    0:25:37 Yeah. So this is the kind of stuff that like no prompt, like.
    0:25:40 So are you generating images and then pulling them in?
    0:25:40 Is that what’s going on?
    0:25:42 So this is this is all generated.
    0:25:45 The the initial images are generated up in mid-journey.
    0:25:47 That’s usually how I tend to mid-journey or flux.
    0:25:51 I’ll go back and forth on either of those and then drop them in.
    0:25:53 And for the most part, I mean, yeah,
    0:25:56 highly will always understand what you’re kind of looking for.
    0:25:59 Yeah, I mean, I’m looking at their website right now
    0:26:02 and all the ones that are on the Explore page are really impressive.
    0:26:08 Like, yeah, like really aesthetically pleasing.
    0:26:09 The lighting looks great.
    0:26:11 The colors are great.
    0:26:12 Like, I don’t know.
    0:26:13 I have to actually kind of agree.
    0:26:17 Just looking at the Explore page that it’s really impressive.
    0:26:19 Yeah, I’m kind of surprised by this.
    0:26:22 I had seen it before, but like, this is better than I remember.
    0:26:24 And like, again, a lot of the prompts here
    0:26:28 that you can see in these demos are very and again,
    0:26:30 you can kind of go either way, always, you know, prompt
    0:26:32 if you’re not getting the results that you want to.
    0:26:33 But a lot of this stuff is very
    0:26:39 it almost kind of looks like this prompt here is actually just the image prompt.
    0:26:39 You know what I mean?
    0:26:43 Like, and then this like that was the prompt they ran in mid-journey.
    0:26:44 And then they dropped it into here
    0:26:48 and just kind of reran the prompt in here to kind of reinforce the idea.
    0:26:51 No, it’s that the S tier or?
    0:26:54 Yeah, right now I’ve got that top of S tier, but it’s right next to Veya.
    0:26:58 Like they’re they’re they’re they’re sharing that that tier together.
    0:27:00 All right, so let’s let’s talk about in video.
    0:27:03 Have you played around with in video much?
    0:27:06 And in video is one of my sponsors for my other channel.
    0:27:11 So it’s hard for me to like, I’ll put in a D tier.
    0:27:17 I think they’ve offered it.
    0:27:19 They’ve offered to sponsor me a couple of times as well.
    0:27:23 I really haven’t got a not gotten the chance to like I dove into them.
    0:27:25 I think early when they first launched
    0:27:27 and I really haven’t got a chance to circle back to them.
    0:27:29 So yeah, that’s that’s one that I’m not.
    0:27:32 Yeah, I don’t want to play around with that much.
    0:27:36 Yeah, so within video, one thing that like makes in video stand out
    0:27:41 is you give it a concept for like a short film, right?
    0:27:44 And you can actually say generate a five minute video for me.
    0:27:49 And it will actually go and generate all of the scenes for your video
    0:27:52 similar to what you can do in like LTX Studio.
    0:27:56 But yeah, you can you basically say like I want to make a movie
    0:27:59 about a Viking, you know, hitting the shores
    0:28:03 and then attacking the the Pete, the locals or something, right?
    0:28:04 Yeah. And it will go and generate
    0:28:07 like a three minute video of that happening.
    0:28:09 And each scene in it is generated.
    0:28:14 I don’t know 100 percent if in video is using an API
    0:28:16 or if it is their own internal model.
    0:28:18 I believe it is their own internal model.
    0:28:20 It’s it’s it’s pretty good.
    0:28:25 What I actually use in video for more and like it for even better
    0:28:27 is it’s got like stock footage.
    0:28:30 And so you can give it a concept for a video
    0:28:32 and it will go find stock footage for that video.
    0:28:34 I actually think it’s stock footage.
    0:28:37 Features are really where in video shines
    0:28:40 more than its generative features at the moment.
    0:28:43 But saying that, you know, they’re only like one update away
    0:28:46 from, you know, jumping up the tier list.
    0:28:51 But right now, just speaking about like the generative capabilities of it,
    0:28:54 I’d have to probably put it in B or C.
    0:28:58 But the stock footage feature is really, really, really good.
    0:29:01 It’s really good at going and finding stock footage for whatever you give it.
    0:29:05 And it’s funny because I remember the video that I did this.
    0:29:10 They got me for a sponsored spot a while back when they first launched.
    0:29:12 So this is before they were doing generative stuff.
    0:29:14 And it was just the stock image stuff.
    0:29:17 And of course, me being me, I wanted to push it into weird directions
    0:29:18 and see what we could do with it.
    0:29:22 So I ended up making a it was a zombie survival guide.
    0:29:25 But it was, you know, picking everything from stock footage
    0:29:26 and I generated up the script.
    0:29:29 It was like a British guy, like how to survive the first 10 days of a zombie
    0:29:32 apocalypse and like is, you know, is very kind of like,
    0:29:35 I guess, like YouTube explainery.
    0:29:36 But it was, yeah, it was a neat idea.
    0:29:42 So you can kind of like weirdly thread narratives out of stock footage as well.
    0:29:44 So it’s all just a matter of kind of how you approach it.
    0:29:45 I think it did the serviceable job.
    0:29:48 But again, I just haven’t gone back and checked it out
    0:29:50 from from a from a generative standpoint set.
    0:29:53 Yeah. And it also does like it’ll write the story for you.
    0:29:55 It’ll do the voiceover for you.
    0:29:59 You can upload your own voice and actually have it use your voice in the video.
    0:30:02 You can do all that kind of stuff.
    0:30:04 So it’s got some really, really cool features.
    0:30:06 But the generative capabilities.
    0:30:10 I mean, I don’t think I can put it the generative capabilities
    0:30:12 on the same level as any of these other ones.
    0:30:16 So I think I would probably put it in like the lower B tier right now.
    0:30:21 It’s not possibly bordering on C tier, but not not to say that in videos,
    0:30:22 not a great product.
    0:30:26 It’s just I feel like the generative functionality has a ways to go.
    0:30:29 Well, I’ve got to I’ve got to I just thought about this.
    0:30:31 I got a controversial S tier coming up.
    0:30:33 A controversial, uh-huh.
    0:30:36 Yeah, yeah. Well, let’s jump to the end.
    0:30:39 Oh, should we leave that one to the end?
    0:30:40 No, it’s whatever you want to do.
    0:30:42 It’s fine. All right.
    0:30:45 So which one which one would you put into S tier next?
    0:30:47 Oh, so this is going to be controversial,
    0:30:50 especially coming off of the the NVIDIA one.
    0:30:52 I I’m going to put LTX in S tier.
    0:30:55 Why? OK, like aside from this,
    0:30:58 they did that we both met and I are sponsored by them.
    0:31:03 But it’s actually it’s not that.
    0:31:05 No one’s putting that money in my pocket right now.
    0:31:09 No, I literally because they open source their video model.
    0:31:12 That is like, is there any other company here
    0:31:15 that open source their own video model?
    0:31:17 Not of this list.
    0:31:18 I don’t believe so.
    0:31:20 Meta didn’t do it looking through that much.
    0:31:23 I don’t believe in their open source other than LTX.
    0:31:25 No. So I mean, that is that.
    0:31:28 Do you think LTX from the generative side of it, though,
    0:31:33 do you do you think LTX generates videos on the same level
    0:31:36 as what you’d get out of Minimax or Veo?
    0:31:37 Not currently.
    0:31:41 But again, the code is out there and people can tweak it
    0:31:45 and people can do things with it to the open source community
    0:31:48 can bring it up to speeds or to a quality of that level
    0:31:50 or perhaps even higher.
    0:31:52 They can do they can tinker with it, you know, because again,
    0:31:57 now you’ve got a worldwide, you know, network of people
    0:32:01 that can now fuzz with that model to make it do kind of whatever they want.
    0:32:03 And it actually does generate extremely fast.
    0:32:05 I mean, I think that’s the selling point of the whole thing,
    0:32:08 is that it actually generates videos sounds great in theory,
    0:32:12 but like massive compute to like advance these models in the next year.
    0:32:14 Yeah, I think it’ll take open source sounds nice.
    0:32:17 But like, how are they going to compete with like SOAR and Microsoft?
    0:32:19 The training data and all of that stuff and and all of that.
    0:32:25 I mean, you know, again, I think it’s it’s not necessarily for the
    0:32:28 for the for the massive ramifications.
    0:32:31 But just in the fact that it’s like, you know,
    0:32:33 it gave they gave their model away to everyone.
    0:32:37 That’s that’s actually, again, that’s just super cool and giving back to,
    0:32:41 you know, a community of developers that ended up bringing, you know,
    0:32:43 bringing all of this technology together.
    0:32:44 So yeah, I think it’s good.
    0:32:47 I think we should try to judge it based on like how useful it would be to people.
    0:32:50 I am sort of leaning more towards Nathan’s take on it.
    0:32:56 Like, I see I see why it goes up the chain because of the open source.
    0:33:00 But at the same time, if we’re judging, you know,
    0:33:05 if if I see open AI and SOAR based on
    0:33:08 like what we’re getting today right now, I feel like we should be judging
    0:33:12 LTX on what we get today right now from it. Yeah.
    0:33:15 If we were judging future, I would literally put SOAR at the top of the top.
    0:33:18 Honestly, it would be my personal take. Right. Right. Right.
    0:33:20 Yeah. No, that’s valid.
    0:33:22 I whatever you guys landed tier wise,
    0:33:26 I think it should just bump up one for for the fact that, you know,
    0:33:29 they get I would have probably put the generative capabilities on par
    0:33:31 with what you’d get out of like in video or pica.
    0:33:37 But because it’s open source, my I say it bumps up a tier and gets an A.
    0:33:39 I’ll say, have you been around lately?
    0:33:41 Because they’ve actually been doing a lot of improvements in there as well.
    0:33:44 The video model does look a lot better, a lot smoother.
    0:33:47 I know it was a little bit on the rocky side to start.
    0:33:49 They added flux in as their image generator too.
    0:33:52 So that tends to look a lot better.
    0:33:57 And I think today they kind of put in was it sort of like that advanced
    0:34:01 live portrait thing where you could kind of do expression control and everything.
    0:34:06 It’s funny talking to those guys that it’s they have such an interesting
    0:34:10 problem in that they are kind of a do it all platform.
    0:34:13 Right. And as like new technology kind of appears,
    0:34:15 they’re kind of like putting it all together.
    0:34:17 But it’s not just a matter of just plugging it in.
    0:34:19 It’s not like, oh, here’s a module that we just plug right in.
    0:34:23 It’s almost like, you know, you’ve got this existing.
    0:34:26 It’s not like a car where you’re like, oh, I’m going to drop in the or actually,
    0:34:29 you know, a good example is like it’s not a computer where you’re just dropping
    0:34:30 in like a new RAM slot into it.
    0:34:35 It is more like an 18 ton semi where you’re like trying to drop a new engine
    0:34:36 into it, or you know what I mean?
    0:34:39 Like a like a new like catalytic converter or something.
    0:34:43 So there’s a lot of like playing around to get the whole thing to work together.
    0:34:47 But I think they do a pretty solid job of trying to stay up on everything
    0:34:49 and kind of creating that all in one platform.
    0:34:54 So then they also have LTX Studio, which does similar things to what we talked
    0:34:58 about within video, where you can give it an idea for a video, right?
    0:35:02 And then with that idea, it will actually generate all of the scenes for you.
    0:35:05 And then you could go in and swap out scenes.
    0:35:06 It’ll do the voiceovers.
    0:35:10 It’ll essentially make like a little short film for you based on just like a prompt,
    0:35:15 which I think, you know, that feature sort of bumps it up a little bit as well.
    0:35:19 I think where people get sassy with that, too, is that, you know, they they kind
    0:35:24 of go in there, like maybe a serial killer werewolf meets an alien, you know,
    0:35:27 vampire robot story. My exact.
    0:35:30 And then, you know, yeah, which I would I would totally watch that sounds like
    0:35:32 something I would definitely watch when I was 13 on Cinemax.
    0:35:36 And then they get something that is like, it’s kind of there and it’s kind of like
    0:35:39 it just in the video, the the images that they pick aren’t great.
    0:35:42 Like it is that thing where you’ve got to spend a lot of time in there.
    0:35:44 It’s not just like prompt and like, I’m instantly entertained.
    0:35:46 It’s like it’s prompt.
    0:35:50 And then you got to start going in and you know, changing things on a granular
    0:35:53 level. I think that that’s where people are just like, oh, so that good.
    0:35:56 And it’s like, no, it is you just have to spend a lot of time working on it.
    0:36:00 It’s not going to, you know, you’re like, you’re still going to have to do the work,
    0:36:05 basically some. So Luma, so we’ve got Luma Dream Machine.
    0:36:08 Where would you rank Luma Dream Machine?
    0:36:13 I I really like the generations that Luma Dream Machine comes up with.
    0:36:18 But I don’t think it’s on par with what we’ve seen from from Veo or from
    0:36:24 Minimax. I’m curious where you would rank it based on all of your testing.
    0:36:27 But the new image model looks really good.
    0:36:29 So you can now generate on platforms.
    0:36:33 They’re kind of doing something a little bit on the similar side to it’s not
    0:36:36 quite LTX, but kind of more like Leonardo’s flow state, I guess,
    0:36:38 where you can kind of have a bit of a conversation.
    0:36:43 It doesn’t quite do like that cascading like level of like of images
    0:36:44 that keep going down.
    0:36:48 But you kind of end up, you remember when Dolly 3 first came out
    0:36:52 and you were like, you were supposedly be able to to chat with your image
    0:36:55 and, you know, and get it to do things.
    0:36:57 That’s kind of the kind of that idea.
    0:36:59 And yeah, it looks really good.
    0:37:01 You got to give it up because their API is everywhere.
    0:37:03 Anywhere that you go, you can now generate in Luma.
    0:37:07 The first frame, last frame thing, you can do some really cool things with.
    0:37:12 I think that they are I would probably put them into the A category.
    0:37:16 Somewhere has somewhere between A and B, like they’re they’re just
    0:37:18 but they’re just like one update away from like hopping up to an.
    0:37:23 Most of what I do with them is the the image to video.
    0:37:28 I feel like with this one, starting with an image, you tend to get
    0:37:32 a much better result than if you start if you just give it a video prompt.
    0:37:36 But, you know, the other thing, let’s see, this is and I think of the right tool.
    0:37:40 This is the one where you can give it a start and an end frame, right?
    0:37:45 Yes, I’ve got some so like so you can give it like a starting image
    0:37:49 and an ending image and it will sort of animate between the two images, right?
    0:37:53 Yeah, I’ll show you something that I was playing around with this idea
    0:37:56 about wanting to do kind of a Game of Thrones type opening.
    0:38:01 So, you know, I started off by generating up some, you know, I like still images
    0:38:06 of like these castle things that wasn’t quite it ended up kind of arguing
    0:38:12 with with Luma a bit, ended up zooming out, you know, with once we established
    0:38:16 a look, zoomed out a little further, kind of got into a map look.
    0:38:19 And then, you know, from there, you could kind of start taking taking it into
    0:38:21 to see if it was one of the 12 second ones.
    0:38:22 What’s this user interface?
    0:38:24 Is this is this Luma?
    0:38:25 This is yeah, this is Luma now.
    0:38:29 So, you know, we can kind of get these kind of cool effects using first frame,
    0:38:30 last frame, you know what I mean?
    0:38:37 Like kind of getting this real sort of like swooping, droney, not quite Game of Thrones,
    0:38:39 but, you know, kind of like, you know, by the end of it, I had like, you know,
    0:38:45 fairly quickly to end up with like this 19 second shot of like this, you know, fly through.
    0:38:50 Yeah, this kind of stuff is like really fun and very easy to do in Luma.
    0:38:56 So, yeah. Yeah, I think Luma is a great little it’s a great exploration.
    0:38:58 It’s it’s yeah, it’s great for exploring.
    0:39:03 I actually played around with an anime style too, which I normally do not do on my channel.
    0:39:06 Yeah, I’d say Luma for me is probably a tier.
    0:39:09 I really like that sort of transition feature where you give it to images
    0:39:11 and then it can transition between them.
    0:39:12 I think that’s really, really handy.
    0:39:15 So for me, I say Luma’s a tier.
    0:39:16 Yeah, for sure.
    0:39:21 And then like so moving along here, next up, we’ve got runway and runway
    0:39:23 in my mind is like the OG, right?
    0:39:27 When it comes to video generators, they were out there doing it first, right?
    0:39:32 We had Gen 1, Gen 2, all those three seconds of video, my friend.
    0:39:34 Three seconds of video.
    0:39:36 It was when people complained.
    0:39:39 So it’s like you can’t put them in D tier just just because they’re like
    0:39:42 the godfathers of like the video tools.
    0:39:47 I, you know, maybe hot take on this, but I would actually I actually
    0:39:50 they’re one that I would put up in the S tier only because, again,
    0:39:53 the suite of other tools that they have as well.
    0:39:56 I mean, Gen 3 is great and it’s fast.
    0:39:58 I mean, it is super fast.
    0:40:01 I don’t know if you’ve played around that much with it on the image to video
    0:40:04 stuff, but you add Act 1 into there and like the expand and all of that
    0:40:06 other stuff that they’ve been doing lately.
    0:40:08 I mean, is that killer suite?
    0:40:10 And then on top of it, you can just take all of that stuff
    0:40:15 and instantly rotoscope and green screen in there and like the subtitle thing.
    0:40:19 I think if you go into like the greater, adding new things, right?
    0:40:20 Like I always say new features.
    0:40:24 They’re releasing the greater like runway ecosystem.
    0:40:27 Like there’s just so many tools down there as well.
    0:40:30 Don’t they have over 36 tools, right?
    0:40:37 You got audio, lip sync, removing backgrounds, text images built in.
    0:40:41 I mean, one of the guys that runs runway used to like he’s one
    0:40:43 of the original creators of stable diffusion as well.
    0:40:48 So it’s like they’re OGs that created runway.
    0:40:53 They go way back to, you know, the very beginning of AI image generation.
    0:40:57 And yeah, there’s just so many tools like I use the green screen tool
    0:41:00 quite a bit where you could take another video and remove the background.
    0:41:03 I think that’s really, really a handy tool.
    0:41:06 But yeah, when it when it comes to like Act 1, where you can just sort of
    0:41:11 upload a video of you talking and change it into like any sort of cartoon.
    0:41:13 That’s super cool.
    0:41:16 You can now do it with I really think it should have been called it Act 2 as well.
    0:41:22 Now you can drive that Act 1 video footage onto any other like footage as well.
    0:41:26 So you can have somebody walking around in a 16, 9 frame and then just completely
    0:41:29 change, you know, essentially what they’re saying in their facial expressions.
    0:41:31 Like that’s super handy, like it’s huge.
    0:41:33 Yeah. Well, I mean, you got all these camera controls, right?
    0:41:36 Horizontal, vertical, zoom, roll, tilt, pan.
    0:41:40 You know, you’ve got expand video where you can give it
    0:41:45 like a vertical video and turn it into a 16 by 9 and have it actually stretch
    0:41:47 it out for you or vice versa.
    0:41:51 You know, it’s got it’s got a ton of cool features.
    0:41:54 It’s really, really good at image to video as well.
    0:41:56 Like you mentioned, it’s it’s fast, too.
    0:42:00 Let me see if I can just find something real quick to drop in here and see how
    0:42:03 fast it generates. Oh, and this one has a first and last frame, too.
    0:42:05 I didn’t even realize they’ve added that feature to run.
    0:42:10 So here’s like a colorful, swirly one.
    0:42:13 And you also go pretty quick.
    0:42:17 If you miss a Friday, that seems to be when runway loves to drop an update.
    0:42:19 So you always got to pay attention.
    0:42:21 I that’s. Yeah.
    0:42:25 Yeah. Well, I always make my news videos on Thursdays and release them on Fridays.
    0:42:28 So yeah, news is always like.
    0:42:30 Yeah, I’m always talking about it.
    0:42:32 A week behind. Yeah.
    0:42:34 So the runway’s got the best editing tools now, right?
    0:42:36 Like the best editing tools.
    0:42:39 But I still feel like the visuals are not the best, man.
    0:42:41 They’re good, but they’re not the best of the best.
    0:42:43 Is that a fair take?
    0:42:46 I think depending on what you’re what you’re generating, a lot of times.
    0:42:48 Like that one, the the mount one here looks pretty good.
    0:42:53 Like, and again, you do have all of your not your your your your video controls,
    0:42:55 your your motion controls and all of that stuff.
    0:42:59 And depending on what because this was promptless too, right, Matt?
    0:43:04 Yeah, you know, there’s I think it depends on what you’re looking for.
    0:43:07 And then on top of it, because you can take this up to whatever, 10 seconds.
    0:43:09 And there’s all kinds of other tricks.
    0:43:11 There’s like, I’m always thinking a lot in terms of like,
    0:43:16 you know, you know, use all of the meat of the animal.
    0:43:20 So when I see a generation like this, what I’m thinking is like,
    0:43:21 OK, I’m going to bring it in.
    0:43:25 I’m going to run through topaz and then maybe I’ll speed it up.
    0:43:27 And I’m only going to take the first like four seconds of it
    0:43:30 to give it a quicker vibe or quicker feel to it.
    0:43:33 Maybe put a zoom in and the, you know, I’m just thinking about all the other
    0:43:36 ways that I’m going to use this, not necessarily just the generation.
    0:43:41 Yeah. And that’s another thing that when it comes to like generating videos,
    0:43:46 a lot of people don’t really think about that is in most scenarios,
    0:43:49 you really only need about three seconds, right?
    0:43:52 Yeah. If you ever turn on your TV and watch a TV show or you watch a movie
    0:43:56 or something like that, they never linger on a shot for more than three to four seconds.
    0:43:57 Right. Yeah.
    0:44:01 So if you can get these tools generating eight, 10, 20 second videos for you,
    0:44:07 if just three to four seconds of that video is good, you’ve got perfect B-roll, right?
    0:44:11 That’s something that when that Sora thing, when Sora was first announced
    0:44:14 and was like the 1080p one minute long generation, it’s like no one’s ever going
    0:44:17 to know you’re never going to watch one minute of like, you know what I mean?
    0:44:20 Like people can can barely hold their attention through a one minute
    0:44:23 TikTok or YouTube short where it is cutting like crazy.
    0:44:28 Like try making a YouTube short one day when it’s just you literally
    0:44:31 just talking to the screen for a minute and like, I guarantee you,
    0:44:33 it’ll be the lowest watch short of all time.
    0:44:38 So runway, I don’t know my problem with S tier is I feel like visuals.
    0:44:42 It’s like B, like in terms of that, I was just looking like how good to the
    0:44:45 videos look like recently I put together like some Twitter threads like
    0:44:48 comparing and I didn’t actually include many videos from runway because I
    0:44:50 just didn’t think they looked as good.
    0:44:52 I didn’t think they would perform as well in the thread.
    0:44:57 And I, you know, when I looked at all the different models, I mean, it was like,
    0:45:01 you know, VO was by far the best and then store was pretty good.
    0:45:06 And I felt like runway was like a like slightly a tier below visual quality wise,
    0:45:07 but the editing tools are amazing.
    0:45:09 And it’s not the most controllability though.
    0:45:13 Like if you saw the camera angles and the like from a controllability
    0:45:16 standpoint, I think runway is up there.
    0:45:22 Whether I put it on the same level is what I saw from Minimax and from Vio.
    0:45:26 So then, all right, we’ve got two last tools and these are the two that we
    0:45:27 don’t have access to.
    0:45:30 I mean, technically, I think all three of us actually have access to
    0:45:34 Adobe Firefly video, but yeah, the emails I’ve gotten from Adobe says
    0:45:36 I’m not allowed to share those videos yet.
    0:45:38 Oh, D tier.
    0:45:47 Cause I thought the model was released and out to the public.
    0:45:50 No, the Firefly video model.
    0:45:54 I literally got an email from them this morning saying that you can play with it,
    0:46:00 but make sure that you send any videos to us for review before showing them off.
    0:46:00 Yeah.
    0:46:06 So what I figured we can do for this is we’ve got the current model that’s out.
    0:46:06 Probably is what we should do.
    0:46:12 So for Adobe Firefly, we’ve got their page here with some probably pretty
    0:46:18 cherry-picked demos, but you’ve got like this dog video here and it seems
    0:46:22 to be one of the better ones, maybe with text from what I’ve seen.
    0:46:26 But some of this isn’t great.
    0:46:27 Some of it is pretty good.
    0:46:33 Based on what you’ve tested with Firefly, we can’t actually show any
    0:46:36 videos that we generated because of the rules there.
    0:46:39 But based on like the videos that you’ve tested, like how would you rank it?
    0:46:44 You know, I think it is very, I think it’s very Adobe.
    0:46:50 Like, I think that if you are, you know, working as, yeah, I think that if you’re
    0:46:57 working as a professional editor and you are, you know, you need some B-roll of a thing,
    0:47:02 you know, much of the same way, like any, how much are professional photographers
    0:47:04 using generative fill, like, you know what I mean?
    0:47:08 Like, I think that it’s become a tool that they’ll try out a few times.
    0:47:11 They figured out where the use cases work really best.
    0:47:14 They know where it doesn’t do great.
    0:47:19 Like, you know, if you’re a wedding photographer and you need, you know, I don’t know,
    0:47:21 something like a couple of birds up in a tree or something like that.
    0:47:25 I’m sure that, you know, Firefly or Jenfill is going to do that for you.
    0:47:29 If you need dragons flying in and attacking the wedding party, it’s from another tool
    0:47:30 that you’re going to use for it.
    0:47:36 So I feel that, that Firefly video is kind of in the same, in the same area.
    0:47:42 If you’re doing, you know, kind of like what I would call like regular movies or
    0:47:46 regular videos, like kind of like very, like, I don’t want to say, I guess,
    0:47:47 stockish kind of stuff.
    0:47:49 Like, yeah, it’ll, it’ll perform very well.
    0:47:53 If you’re trying to get into doing your, you know, if you’re trying to make avatar
    0:47:56 for like, you know, then you’re going to struggle with it.
    0:48:01 So. So the other thing that I’ve noticed about Adobe, and this is more with the
    0:48:05 Firefly image generation, but I imagine it translates to the video as well.
    0:48:09 Is as far as like censorship goes, they’re the worst.
    0:48:13 Like they’re the ones that, like I tried to generate an image of a person
    0:48:15 standing in front of the Eiffel Tower and it’s like, oh, we can’t generate the
    0:48:16 Eiffel Tower, right?
    0:48:17 Interesting.
    0:48:22 There was like lots of things that it would not generate for me.
    0:48:24 That it didn’t make sense that it wouldn’t generate.
    0:48:28 It is always hard for me to argue against like, because I’m in Adobe is
    0:48:31 obviously very much taking a hard stance on like, we are trying to be as
    0:48:32 ethical as possible.
    0:48:34 And it is sometimes hard for me to be like, well, stop.
    0:48:38 So I’ll do it.
    0:48:39 I’ll say stop.
    0:48:46 But, uh, but yeah, I think the bad guy on this show, I have no hard.
    0:48:53 So yeah, I think that that is that eventually I think that that will be.
    0:48:58 But again, I think that where Adobe, where Firefly and GenPhil really do
    0:49:02 shine is the fact that they are going to be directly integrated into Premiere,
    0:49:04 into, you know, into Photoshop as we’ve seen.
    0:49:08 And that is, you know, so it becomes that at least first, when you’re stuck
    0:49:13 on a thing, you know, at some point when this, like I edit in Premiere, at some
    0:49:17 point, you know, when I need a quick shot, like, uh, am I going to bounce
    0:49:20 out necessarily to another video model while I’m editing?
    0:49:23 Like if, if I’m like, oh, this is just occurred to me.
    0:49:23 I want to try this out.
    0:49:27 Or am I going to give it a shot real quick with the new Firefly video
    0:49:30 model just because I’m already there, you know, there.
    0:49:30 Yeah.
    0:49:31 Yeah.
    0:49:34 The other thing that I really like that, uh, Firefly is going to do.
    0:49:35 I don’t believe we have it yet.
    0:49:38 Is that generative extend, right?
    0:49:41 Where you have a video that you needed to cover eight seconds, but right
    0:49:43 now it only covers six seconds.
    0:49:46 You can, you can extend it out by two more seconds.
    0:49:50 And it figures out what the rest of that video would have looked like right
    0:49:51 now when we’re doing video editing.
    0:49:54 I don’t know if you’re like me, but what I’ll do is I’ll literally slow
    0:49:58 down the clip so that it’s a slightly more slow motion to fill that gap.
    0:50:03 But I like that idea of having that generative fill inside of video to
    0:50:05 sort of fill out that extra one or two seconds that I need.
    0:50:06 Yep.
    0:50:10 I think that is probably going to be more the use case then.
    0:50:14 I mean, again, even with Firefly image, like I think that most people use
    0:50:19 it in gen fill, people don’t really go to adobe.com/firefly.
    0:50:22 And, you know, image model and generate stuff there.
    0:50:24 You know, they do it, they do it in Photoshop.
    0:50:27 So I think that’s really, that’s really where it’s going to live.
    0:50:31 So I don’t even necessarily know if like adobe necessarily really belongs
    0:50:33 on this list because it’s not necessary.
    0:50:37 I don’t really think of it as like a full platform generator, as much as I
    0:50:40 do an extension of Photoshop and Premiere.
    0:50:41 Yeah, gotcha.
    0:50:42 Yeah.
    0:50:46 Well, and then the last one we’ve got on the list here is the make a video
    0:50:49 from Meta, which nobody has access to yet.
    0:50:57 So, yeah, I don’t know if we will or not, but maybe they’ll make it open source
    0:51:00 or, you know, Facebook’s version of open source.
    0:51:03 Every time I say that Meta is open source, I get a million comments
    0:51:05 from people going, it’s not open source.
    0:51:06 Stop calling it open source.
    0:51:07 Not really, yeah.
    0:51:13 But, you know, Meta’s level of open source, maybe they’ll open source,
    0:51:15 open source-ish this one.
    0:51:18 I’m curious, like, with the, like, now that we’re sort of getting this,
    0:51:22 like, second tier of, like, of, like, because I mean, I think that this is
    0:51:25 maybe, like, in my head, since I’ve been following this for so long, I
    0:51:30 consider this era that we’re moving into as like AI video 3.0.
    0:51:34 And, like, definitely with Sora now released and with Vue release.
    0:51:38 I mean, it’s kind of, it’s kind of forcing Meta’s hand a little bit to
    0:51:39 do something with this, right?
    0:51:43 I mean, I don’t know if Zuck really cares, like, you know, at the same time.
    0:51:46 But, I mean, you know, just comparing it to what else is out there,
    0:51:50 this looks like the last generation of AI video, right?
    0:51:53 Like, it doesn’t look like the current gen that we’re seeing now.
    0:51:58 But this was the one that had all of the sound effects and you can do,
    0:52:00 you know, everything generates as one, right?
    0:52:02 It feels like a, it feels like a Pika competitor.
    0:52:04 It feels like Pika’s going to go more in the social direction and then
    0:52:06 Facebook’s already there.
    0:52:10 So they’re kind of like, if they approve this, they’ll end up kind of causing
    0:52:11 problems for Pika, possibly.
    0:52:13 Like, if you made like a cool social product, you know, with it.
    0:52:17 My other question, too, is in, yeah, in this, given that it is Meta, like,
    0:52:19 where does this live?
    0:52:21 Is this, are they thinking about this for?
    0:52:23 Does this a Facebook thing?
    0:52:24 Is this an Instagram thing?
    0:52:26 Like, where does this live in the Meta ecosystem?
    0:52:28 Like, like, what is it?
    0:52:29 Yeah, what’s it, what, why?
    0:52:30 That’s a very good question.
    0:52:32 That’s something like, I’ve been saying for a while, it’s something like
    0:52:37 there’s some new social product with AI that doesn’t exist yet.
    0:52:37 I don’t know what that is.
    0:52:41 It could be, they’re probably, probably multiple ones, but like Instagram
    0:52:42 started because of filters.
    0:52:45 Well, maybe there’s something new now that you could create with these
    0:52:48 new AI technologies, like whether it’s like video and you take a photo of
    0:52:50 yourself and now you made a really cool viral video or something, you know,
    0:52:51 something really funny.
    0:52:54 Well, and I know that’s the other thing that everybody’s really hot on right
    0:52:57 now is like trying to figure out, or not everyone, there’s a lot of companies
    0:53:00 that are, that are very much trying to become the Netflix of AI video.
    0:53:05 And like, I, I, I’m not, I don’t believe that needs to happen yet.
    0:53:07 I think that we need to have AI video on Netflix.
    0:53:11 I think I would like to see an AI generated, you know, AI produced,
    0:53:16 generated, produced, like thing, like something like, uh,
    0:53:18 did you guys watch scavengers rain on HBO?
    0:53:18 So cool.
    0:53:19 So good.
    0:53:25 Uh, it was that animated series about the no HBO in Japan, which is a bummer.
    0:53:26 Really?
    0:53:29 There’s no, have you heard of this thing called a VPN VPN?
    0:53:31 I was going to say, yeah, and it does sound.
    0:53:36 VPN and the iron.
    0:53:40 A, well, what sucks is though, but I, but I watch content.
    0:53:41 I watch stuff with my wife, right?
    0:53:43 And so you want the subtitles.
    0:53:47 So it’s like, you know, yeah, like, like, I, like, we’re, uh, we’re
    0:53:50 finishing up watching a Vikings right now, which was kind of funny for me
    0:53:52 because we’re doing all this Viking AI video stuff here.
    0:53:54 We’re finishing up that series right now.
    0:53:58 And I was like, well, I like game, I’d actually never watched a Viking before.
    0:53:59 And it’s pretty good.
    0:54:01 I was like, but I like Game of Thrones better, except for the last season.
    0:54:03 And so I’m wanting her to watch Game of Thrones.
    0:54:07 You’re like, she’s like a fantasy nerd like me and she’s never watched Game of Thrones.
    0:54:11 And it’s just like, I had to buy the entire show.
    0:54:14 Like, it’s quite expensive, like actually buying it.
    0:54:18 So right now this is, this is our current rankings of the tools.
    0:54:20 This is where I feel like they all landed.
    0:54:22 Is anything need to be adjusted?
    0:54:25 What do you guys think is never coming on our podcast?
    0:54:27 Oh, he’ll be fine.
    0:54:31 Well, I think going back to like the only reason I put Meta down in the
    0:54:33 D tier is A, we don’t have access to it.
    0:54:38 B, it still kind of looks like the last gen AI, like Meta tomorrow could drop
    0:54:41 something that looks just as good as Veil and be like, we’ve been working on this
    0:54:44 behind the scenes and all of a sudden they’re in A or an S tier, right?
    0:54:46 So I think they’ll have some cool stuff coming out.
    0:54:49 So the same as I think, you know, Grock will have some cool AI video stuff at some point too.
    0:54:50 And, you know, yeah, yeah.
    0:54:53 So I mean, obviously none of this is a knock on any of the companies.
    0:54:57 We’re just kind of like ranking them based on where we feel their current
    0:55:02 generation of technology is and any one of them like is an update away from,
    0:55:04 you know, bumping up a tier or two.
    0:55:08 And also like, like Tim mentioned, I don’t really feel like we know what
    0:55:11 Meta is going to do with AI video yet, right?
    0:55:15 Like they are trying to shoehorn like the AI image generators like into
    0:55:19 their chat apps and stuff, but I don’t think anybody’s still really using them.
    0:55:21 I don’t think that’s going to work just adding it on stuff.
    0:55:22 It’s like the same problem Adobe has.
    0:55:25 You just add on AI to some existing thing.
    0:55:27 Like, no, you need entirely new things for this new world.
    0:55:29 It’s now there’s there’s so many new possibilities.
    0:55:31 Why do you want to be tied to these old systems?
    0:55:34 But they have to because that’s what that’s like, they’re, you know, cash cows, right?
    0:55:38 Well, and I think I think that as we move into 2025, too, like there’s
    0:55:41 going to be a big shift up amongst all of these.
    0:55:43 And we’re going to see if we come back and do this again next year,
    0:55:47 you’re going to have we’re going to have like five or six other companies
    0:55:49 that we’ve never even heard of now on this list.
    0:55:51 The list is going to be all over the place.
    0:55:55 We’re going to see a lot more of that 3D element with what we were talking
    0:55:59 about before with World Labs and with what kind of in a lot of ways
    0:56:02 what Genie 2 is doing, those like world builders.
    0:56:06 Like, and I know that mid-journey’s been because mid-journey video will be coming along.
    0:56:07 That’s what I’m about to say.
    0:56:08 Yeah, they’ve been wanting to do.
    0:56:09 Yeah. Yeah.
    0:56:11 So mid-journey is also working on a video model, too.
    0:56:14 That they keep on saying they’re going to show it because I’m sure.
    0:56:17 Yeah, there’s something new, like great video models,
    0:56:19 like how do they release something without it being spectacular, right?
    0:56:22 Like, actually, and I was just in the office hours a couple of weeks ago.
    0:56:26 I think they were like, and I’ll bless them for too, as well.
    0:56:29 As the team was just like, we could release it.
    0:56:30 It’s the holidays.
    0:56:33 We’d like to spend time with our families and like so that’s like
    0:56:37 legitimately why they’re not, you know, because like if they release it,
    0:56:40 they’re going to be hit with all kinds of like questions and they and the server
    0:56:41 is going to go down this and that and that.
    0:56:44 Like, just let them let them go home and hang out with their families.
    0:56:47 And then let’s be honest, right now is not the time to be releasing
    0:56:50 something with what OpenAI and Google are doing, right?
    0:56:52 They’re just going to get overshadowed by those two companies right now.
    0:56:55 Who? Rock ’em, sock ’em robots going at it right now.
    0:56:57 It’s like, so, yeah.
    0:57:02 Meanwhile, what’s your prediction on January, do you think we’re going to see
    0:57:05 a massive slowdown or do you think we’re going to ramp up again?
    0:57:11 So I’m kind of waiting to see what OpenAI releases over the next two days.
    0:57:14 Yeah, I’m sort of given away when we’re recording this episode
    0:57:15 compared to when it’s getting released.
    0:57:19 But there’s two more days of OpenAI about to release updates.
    0:57:24 I think if they don’t release tool use as one of the next two updates,
    0:57:26 I think that’s coming in January.
    0:57:29 They’ve already hinted that that’s probably going to come in January.
    0:57:33 So I think we’ll see the tool use come from OpenAI.
    0:57:37 And then, you know, knowing what I know about Google and spending some time there,
    0:57:43 they’re really working on the sort of like agentic tool use kind of stuff as well.
    0:57:46 So I think that’s going to be the big theme of 2025 overall,
    0:57:49 is these tools that can use tools on your behalf.
    0:57:52 Yeah, I’m excited about that.
    0:57:55 I don’t think we’re going to see like a new AI video model,
    0:57:59 for instance, in January, unless it’s from like mid journey
    0:58:01 and they’ve already got it ready to go, you know.
    0:58:03 Yeah, I think we’ll see the rate of progress improve.
    0:58:06 Like if everyone, like probably everyone’s going to be copying
    0:58:09 like what OpenAI is doing in terms of like their 01 model.
    0:58:14 And from my understanding, like, you know, since it’s not entirely
    0:58:17 just relying on the training of, you know, with massive amounts of data,
    0:58:19 that’s one part of it, but that’s not the only way it improves.
    0:58:21 We’ll probably see these models just like keep getting better,
    0:58:23 like over and over and over.
    0:58:25 I mean, we might get to a point where like every month,
    0:58:29 there’s like a, it got better 20% this month, right?
    0:58:34 And I think AI video will end up in a similar scenario quite soon.
    0:58:36 If it’s not already there.
    0:58:40 You know, it’s funny because I talked to Caleb over at Curious Refuge
    0:58:42 about this a bunch of times, he and I chat back and forth of like,
    0:58:46 you know, as the video models get better and better and better
    0:58:51 and like, you know, we are always reporting and doing tutorials on all of it.
    0:58:54 And like, what’s the what’s the point when it’s like, it’s like, OK,
    0:58:57 it’s perfect now, like when it’s like, what do we make then?
    0:59:00 And like the fact is that it’ll never, it’ll never be there.
    0:59:03 Like it’s just, it’s just like there are still people doing, you know,
    0:59:07 camera tutorials and camera, like, you know, reviews and and all of that.
    0:59:10 There’s always, you know, it’s like last year’s Canon is terrible.
    0:59:13 This year’s Canon is better. Nikon is better than that.
    0:59:15 You know, there’s always just going to turn into that. So.
    0:59:17 Yeah, yeah, I agree. I think it’s going to get better.
    0:59:19 I think we’re going to get higher and higher resolution.
    0:59:23 I think there’s always going to be room to focus on the finer details within
    0:59:28 images, you know, we’re going to get more controllability out of the images,
    0:59:31 right, where you have a little more control over the lighting, over the
    0:59:35 the style of it. We’re already seeing that with, I mean, basically,
    0:59:37 mid journeys got Lora’s available now.
    0:59:41 They don’t call it that, but they’ve got the mood boards is essentially that.
    0:59:47 So like, I think there’s always going to be places to improve,
    0:59:50 but it’s going to be more and more nuanced as time goes on.
    0:59:53 And that’s one of the things that I kind of want to start
    0:59:56 maybe like preaching a little bit more on the channel is like,
    0:59:59 you start moving into some of the tools that help you develop story.
    1:00:03 Because ultimately, as fidelity gets better, as details get better,
    1:00:06 as control gets better, what are we doing with this stuff?
    1:00:08 Like, are we telling a good story?
    1:00:11 Like, is it is it something that’s going to make us laugh, cry, get excited,
    1:00:13 whatever? I mean, that ultimately is the thing.
    1:00:16 Like none of us watch Star Wars when we were kids.
    1:00:18 And, and, you know, we were we all love the special effects,
    1:00:22 but that’s not what drew us in, you know, right, right, for sure, for sure.
    1:00:26 And I’ve actually really started to see some impressive AI generated short films.
    1:00:28 I saw one that somebody made recently.
    1:00:30 I’m sure you’ve come across it, Tim.
    1:00:33 In fact, it might even bed your ex-account that I came across it on.
    1:00:37 But there was recently like a little Batman short film that came out.
    1:00:39 Oh, yeah, yeah, yeah, really, really good.
    1:00:41 That got copyrights really good on it.
    1:00:42 Oh, did they get that?
    1:00:46 I was looking for it and they got knocked down for for for being.
    1:00:48 Yeah, it was Warner had knocked it down.
    1:00:51 So yeah, I mean, I’m not surprised by that.
    1:00:56 They did basically just, you know, steal somebody else’s IP and make a short film out of it.
    1:00:57 That was really good.
    1:00:58 It was actually, yeah, it was really solid.
    1:01:00 Great sound design throughout that whole thing as well.
    1:01:01 They really captured the tone.
    1:01:04 There was an interesting trick that those guys pulled where
    1:01:07 they used that first frame feature, but they were taking screenshots
    1:01:11 from the Robert Pattinson, you know, Batman movie.
    1:01:16 And then basically running their own generations to kind of create iterations of that scene.
    1:01:18 It’s really smart in all honesty,
    1:01:22 but I can see exactly why Warner was like, no, no, no, that is gone now.
    1:01:26 They’ll release something like that in their own in the future, right?
    1:01:29 Like here’s your custom Batman story.
    1:01:30 I’m sure, I’m sure.
    1:01:34 But I mean, like just that idea alone opens up the door for so many
    1:01:35 like fan films and that kind of stuff.
    1:01:38 I actually, truthfully, I think it’s kind of stupid that Warner knocked it down.
    1:01:39 I really think that it’s like, if anything,
    1:01:42 that’s just like free publicity for your Batman movie.
    1:01:45 So I don’t, you know, that again, I’m not a corporate lawyer.
    1:01:51 So again, it’s the point is people are figuring out how to use these tools
    1:01:55 to make like knowing the limitations is just creating bumpers
    1:01:58 that they know they have to stick between and, you know, with those bumpers,
    1:02:02 they’re figuring out how to make stuff look really good and sound really good.
    1:02:06 And I think we’re just going to see more and more of that, you know,
    1:02:08 ideally people should create their own IP instead of.
    1:02:13 I agree with that Marvel or companies like that because that that kind of stuff
    1:02:15 probably going to get slapped down every day of the week.
    1:02:17 We saw it with the South Park generation, too,
    1:02:18 where there was a whole A.I.
    1:02:21 Generated South Park episode, which was actually a decent episode.
    1:02:23 I watched the whole thing. It was like, that’s not bad.
    1:02:26 Yeah. But that also got slapped down, right?
    1:02:30 That got that got pulled off the internet because that’s somebody else’s IP.
    1:02:34 Yeah, the simulation guys are sort of moving out of that South Park thing.
    1:02:37 And now they’re starting to do kind of like their own branded IPs.
    1:02:40 And the last one, I think that it kind of looked a little bit more
    1:02:42 like that Star Trek lower deck show.
    1:02:44 It’s not, but it kind of has that vibe to it.
    1:02:47 It’s like a spaceship thing and kind of an animated series.
    1:02:49 But that’s, yeah, I agree with you.
    1:02:52 I think that, like, just don’t make fan fiction for stuff.
    1:02:53 Go make your own fiction.
    1:02:56 Like, you know, like it can be very heavily based on, you know,
    1:02:58 the thing that you want to make. It’s not the punisher.
    1:03:03 It’s the it’s the it’s the punishment or whatever you want to do.
    1:03:05 The vindicator. Yeah, exactly.
    1:03:08 Yeah. So, you know, just make that instead.
    1:03:10 So at least you own it.
    1:03:12 Well, cool, Tim, this has been amazing.
    1:03:14 You know, obviously this episode isn’t two hours long,
    1:03:19 but for anybody who’s listening right now, we’ve actually been recording for two hours.
    1:03:22 That’s how we’ve heard out we’ve gone on some of these video tools.
    1:03:25 But before we do, like, you shout out your platforms.
    1:03:28 I know you’ve got an awesome YouTube channel you post on X.
    1:03:29 Where should people go check you out?
    1:03:31 That’s really about it right now.
    1:03:33 I think next year I’ve got to start
    1:03:38 with probably picking your brain a little bit on like on like doing a website
    1:03:42 and probably a newsletter at some point or another.
    1:03:44 Actually, pick both the channel name in the end.
    1:03:46 Oh, sorry. Yeah.
    1:03:50 So theoretically, media on YouTube is probably the easiest way.
    1:03:53 And then from there, you can find links. Absolutely. Yeah.
    1:03:55 Well, thanks again, Tim.
    1:03:57 Thanks for everybody who’s tuning in.
    1:04:01 Make sure that if you’re not already subscribed, you subscribe on YouTube.
    1:04:03 That’s where you’re going to get all the cool visuals.
    1:04:06 You’re going to be able to see our breakdown and examples of these tools.
    1:04:08 You’re going to be able to see the tier list that we made.
    1:04:14 See the handsome faces of Matt and some faces of Tim Nathan.
    1:04:15 I was hoping you were going to say Nathan, too.
    1:04:17 You’re going to say Matt.
    1:04:21 Well, this is awkward.
    1:04:26 But again, if you do prefer audio, we are available wherever you listen to podcasts.
    1:04:29 So thanks again, Tim, and thanks everybody for tuning in.
    1:04:32 [MUSIC PLAYING]
    1:04:33 .
    1:04:36 [MUSIC PLAYING]
    1:04:39 [MUSIC PLAYING]
    1:04:42 [MUSIC PLAYING]
    1:04:45 [MUSIC PLAYING]
    1:04:47 you

    Episode 39: How are AI video tools revolutionizing content creation? Matt Wolfe (https://x.com/mreflow) and Nathan Lands (https://x.com/NathanLands) are joined by Tim Simmons (https://x.com/theomediaai), founder of Theoretically Media, to delve into the latest advancements in AI video tools.

    This episode ranks 13 of the most popular AI video tools of Q4 2024, discussing their features, strengths, and weaknesses. They explore the capabilities of tools like Sora, Runway, and Adobe Firefly, and predict future developments in AI integration. The conversation highlights the evolving landscape of AI video generation and its impact on content creation.

    Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

    Show Notes:

    • (00:00) Pika struggles with customization; requires frequent rerolls.
    • (06:49) Veyo ranks S tier for generating videos.
    • (13:51) Leaving Sora in category S, discussing Turbo.
    • (20:21) You write scripts; creators adapt them independently.
    • (25:40) Prompts reinforce ideas, rerun if needed.
    • (26:51) InVideo generates videos from concepts automatically.
    • (33:28) Integrating new technology requires complex adaptation.
    • (40:50) Runway: Advanced camera controls, fast image-to-video.
    • (45:58) Adobe’s generative tools: Occasional use, specific purposes.
    • (48:08) Firefly and Gen Phil integrate into Adobe.
    • (57:19) AI models continually improve, likely accelerating progress.
    • (59:09) Focus on storytelling, not just technical details.

    Mentions:

    Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

    Check Out Matt’s Stuff:

    • Future Tools – https://futuretools.beehiiv.com/

    • Blog – https://www.mattwolfe.com/

    • YouTube- https://www.youtube.com/@mreflow

    Check Out Nathan’s Stuff:

    The Next Wave is a HubSpot Original Podcast // Brought to you by The HubSpot Podcast Network // Production by Darren Clarke // Editing by Ezra Bakker Trupiano