AI transcript
0:00:10 You can’t just scrape data from the internet, like the record labels will come and do you.
0:00:17 It’s sort of one of the first really truly social kind of forays into AI, image and video generation.
0:00:25 Now when you post a photo on X, you can like long click and press and immediately turn it into a video.
0:00:34 The fact that someone who’s completely non-technical can build something that a couple thousand people can use overnight is like amazing and so exciting.
0:00:39 None of the existing social platforms have leaned that heavily into AI creative content.
0:00:45 And a lot of the AI creative tools I think can and should and will sort of integrate more social.
0:00:49 Things in consumer AI are moving fast.
0:00:54 In this episode, Olivia Moore and Justine Moore, creators and partners at A16Z,
0:00:57 break down what’s new and what’s next across the consumer AI space.
0:01:03 You’ll hear about the latest updates from Grok’s Imagine and what makes it so different from other creative tools on the market.
0:01:07 They break down the release of Genie 3, Google’s new 3D world model,
0:01:11 and why it might be the start of an entirely new kind of gaming and media format.
0:01:13 And of course, they discuss GPT-5.
0:01:18 Not just what’s new, but what’s missing and why some users want their old chatbot friend back.
0:01:24 Along the way, we’ll hear about AI-generated music and Olivia’s very own vibe-coded selfie app starring Jensen Huang.
0:01:25 Let’s get into it.
0:01:31 Welcome back to This Week in Consumer.
0:01:32 I’m Justine.
0:01:32 I’m Olivia.
0:01:40 And we have a bunch of fun topics we want to cover this week, starting in the creative tools ecosystem with Grok Imagine.
0:01:44 And then we’re also going to talk about Genie 3 and the 11 Labs music model.
0:01:49 And then we’ll cover GBT-5 and the deprecation of GPT-4-0.
0:01:51 And we’ll cover our new vibe-coding thesis.
0:01:58 So this week, we are going to start with Grok, which has had a bunch of big updates over the last month or so, I’d say.
0:02:00 So obviously, Grok 4 came out.
0:02:04 The Grok companions caused a huge stir, particularly Annie and Valentine.
0:02:11 But I think more recently, what’s been really interesting is all of the image and video generation features on Grok Imagine.
0:02:18 Yeah, so Grok released an image and video generation model called Imagine, which is offered standalone through the Grok app.
0:02:20 And they’re also bringing it to the web.
0:02:23 And it’s now embedded into the Core X app as well, which is really exciting.
0:02:27 Yeah, I think that’s one of the things that’s really unique about it.
0:02:33 I would say it’s not the most powerful kind of image or video generation model that exists.
0:02:36 Elon has tweeted a bunch about how they’re training a much bigger model.
0:02:45 But I think what’s really cool about it is it’s sort of one of the first really truly social kind of forays into AI image and video generation.
0:02:58 And what you mean by it being integrated into the X app is like now when you post a photo on X, you can like long click and press and immediately turn it into a video animated in the Grok app.
0:03:05 Or even if you see someone else’s photo posted on X, you can turn it into a video or also edit the image with Grok, which is really exciting.
0:03:05 Totally.
0:03:11 I think one of the coolest things about Grok Imagine, to your point, it’s not the most powerful model.
0:03:12 It’s not VO3.
0:03:12 Yeah.
0:03:16 On video, I would say the audio generation is okay, but not great.
0:03:16 Yeah.
0:03:17 But it’s fast.
0:03:17 So fast.
0:03:24 Which I think for a lot of people has been kind of the real barrier to doing image and video generation more seriously.
0:03:24 Yeah.
0:03:30 Is you put in a prompt, you press go, and then sometimes you’re waiting 30, 60, 90 seconds for a generation.
0:03:32 Yeah, often minutes, honestly, for a generation.
0:03:36 And Grok images are basically instant, and the videos are pretty fast as well.
0:03:39 And so I found myself iterating very frequently.
0:03:44 It’s now become, in less than a week, like my go-to tool for image generation on mobile.
0:03:44 Yep.
0:03:48 And even the video, I would say, is getting there, especially if they’re training a better model.
0:03:49 Totally, yeah.
0:03:52 I think for many people, they’re not professional creators.
0:03:52 Yeah.
0:03:59 And so they don’t want to make an image in one place and then go and download it and then port it into another website.
0:04:04 Because often, very few other tools are on mobile, especially for video generation.
0:04:10 So I think it’s sort of huge that you kind of can click one button and get a video on your phone in less than a minute.
0:04:15 That feels like a massive step forward for consumer applications of AI creative tools.
0:04:18 And I think Elon and a bunch of folks on the XAI team have been tweeting about this.
0:04:26 One of the big use cases is, like, animating memes or animating old photos or, like, things that you already have on your phone.
0:04:26 Yeah.
0:04:30 Because you can access the camera roll so quickly through the Grok mobile app.
0:04:30 Yeah.
0:04:32 It will also generate real people.
0:04:33 Yes.
0:04:37 Elon has been tweeting many imagine-generated photos and videos of himself.
0:04:37 Yes.
0:04:39 Which I think is another big differentiator.
0:04:40 Totally.
0:04:42 And something we’ve really only seen from VO3.
0:04:44 And even then, it’s mostly characters versus celebrities.
0:04:44 Yes.
0:04:50 But it comes from Grok’s kind of uncensored nature, which is pretty, I think, cool and unlocks a whole bunch of new use cases.
0:04:51 Yeah, for sure.
0:04:54 And I think that allows the meme generation.
0:05:00 And even VO3, half the time I try to do an image of myself, it’ll be, like, blocked due to our prominent person thing.
0:05:01 And I’m like, I’m not a—what do you mean?
0:05:01 Yeah.
0:05:02 I’m not a prominent person.
0:05:05 But, like, in that photo, I guess I look too much like some celebrity.
0:05:06 Yeah.
0:05:07 Or prominent person.
0:05:08 And it decided to block it.
0:05:09 Yeah.
0:05:13 And I’ve never had that problem on Grok, which makes it just so fun and easy to play around with.
0:05:13 Yeah.
0:05:15 I’m excited to see where they take it.
0:05:21 It feels like we’ve seen Meta experiment a little bit with kind of AI within their core products.
0:05:28 They’ve done the AI characters you can talk to and then uploading a photo to get an avatar where you can generate photos of yourself.
0:05:29 But none of it felt quite right.
0:05:30 Yeah.
0:05:30 I would say.
0:05:30 Yeah.
0:05:35 In terms of baking into the core existing experience on Instagram or Facebook.
0:05:37 And Grok feels a little bit different.
0:05:39 So I’m excited to see where they go with it.
0:05:39 Yeah.
0:05:44 I’d say none of the existing social platforms have leaned that heavily into AI creative content.
0:05:44 Yeah.
0:05:51 And a lot of the AI creative tools I think can and should and will sort of integrate more social.
0:05:57 But today, most of them have just done relatively basic feeds and liking, not really comments, not really like a following model.
0:05:58 Yeah.
0:06:03 And so I think this is going to be a super interesting proof point about what a more social AI creative tool looks like.
0:06:04 Great.
0:06:11 The other big model news of this week, which is not just big for consumer, but for pretty much all of AI, was the GPT-5 release.
0:06:12 Yes.
0:06:19 And the corresponding deprecation of GPT-4-0, which I think ended up being even bigger news in consumer land specifically.
0:06:20 Yeah.
0:06:27 This one was sort of fascinating because obviously it’s been a while since OpenAI had had a major LLM release since GPT-4.
0:06:30 And so people were like very eagerly awaiting GPT-5.
0:06:39 But yeah, as soon as I got access to GPT-5, I wanted to compare the outputs to GPT-4 and I immediately noticed GPT-4 was gone.
0:06:39 Yeah.
0:06:45 And so how are people – because I’ve seen a lot of posts with people kind of up in arms about 4-0 disappearing.
0:06:51 How would you describe kind of the main differences between the models, at least how they’re manifesting in user experiences?
0:06:51 Yeah.
0:06:53 So I’ve talked to a bunch of folks about this.
0:06:59 I think a widespread conclusion is GPT-5 is really good at – especially like front-end code.
0:07:08 I think a lot of the model companies are really focusing on coding as a major use case, a major driver of economic value, something they can be really good at.
0:07:10 And you can tell in the results from GPT-5.
0:07:12 And they emphasize it in the live stream pretty significantly.
0:07:17 And you can see from the examples people use, it’s much better generating things, it’s much better debugging, et cetera.
0:07:20 But a lot of consumers aren’t using it for code.
0:07:22 A lot of consumers just want to chat with it.
0:07:29 And there’s a bunch of examples of how it’s a lot less expressive, emotional, and fun.
0:07:31 Like it doesn’t really use exclamation points.
0:07:33 It doesn’t really use emojis.
0:07:35 It doesn’t send things in all caps like it used to.
0:07:37 It doesn’t do the classic 4-0.
0:07:37 It’s not just good.
0:07:38 It’s great.
0:07:39 Yes, exactly.
0:07:42 And I think there are kind of two separate issues here.
0:07:46 So one is the sort of like glazing excessive validation.
0:07:47 Like it said like, you’re the best.
0:07:49 You should totally do that.
0:07:56 That’s the best decision for like everything you said, even if it was ridiculous, which is like a problem that I think I’m glad they’re working on and getting rid of.
0:08:03 Because let alone like everyone’s concerns about GBT psychosis or whatever, you just can’t trust something that always tells you you’re right.
0:08:10 The second thing is, does it just have a fun and engaging and more casual human feeling personality?
0:08:16 And I think that actually maybe took a step back from GPT-4-0 to GPT-5.
0:08:20 And that is what people, like if you look at the chat GPT subreddit.
0:08:21 People are freaking out.
0:08:22 People are freaking out.
0:08:25 And I think that’s why Sam sort of rolled it back.
0:08:35 And I think he actually may have announced this on Reddit in a comment in response to all of this backlash where he was like, we hear you guys will bring back 4-0 for the paid users.
0:08:38 I was actually kind of surprised they even got rid of 4-0.
0:08:45 I know there had been a lot of jokes about what a pain it is to have to select the model and kind of the dashboard was always getting bigger.
0:08:50 But they had even started building some UI around 4-0 image generation.
0:08:52 They had some preset templates you could use.
0:08:59 And so the fact that they didn’t just add on 5 as an option but took away your ability to use every other model was a little bit surprising to me.
0:09:02 Yeah, I think there’s image generation on 5, right?
0:09:06 Like I imagine some of the templates and the editing tools, they plan to just move over between the models.
0:09:08 They may not have gotten there yet.
0:09:16 I think it’s so funny because if you imagine yourself in the shoes of one of these researchers, you’re like, we trained what is on the benchmarks clearly a much better model.
0:09:17 Like it’s smarter.
0:09:18 It’s better at math.
0:09:19 It’s better at coding.
0:09:23 It can answer medical questions now, which they really focused on in the live stream.
0:09:26 So, of course, everyone will love and embrace with open arms.
0:09:28 It’s like step forward and model intelligence.
0:09:29 A move towards AGI.
0:09:30 Exactly.
0:09:34 And, of course, classic consumer is, no, we don’t want that.
0:09:35 Give us the old toy back.
0:09:45 Give us our fun friend who kind of mirrored the way we spoke to it and was over the top and sometimes kind of crazy but was like really fun to chat with.
0:10:01 And I think to me, honestly, this exemplifies something I’ve suspected for a long time, which is I don’t necessarily think the, like, smartest model that scores the best on sort of all of these objective benchmarks of intelligence will be the model that people want to chat with.
0:10:13 I think there’s going to be a huge market for more of these companionship, entertainment, just having fun type models that doesn’t need to be, like, the highest IQ person you know.
0:10:14 Yeah, I agree.
0:10:18 I do want to spend 30 seconds on that mental health and health overall use case, though.
0:10:29 It’s interesting timing because also last week the state of Illinois just passed a law banning AI for mental health or therapy without kind of the supervision of a licensed professional.
0:10:41 And it’s pretty interesting because the law is wide-ranging to the extent that some AI mental health companies have already shut down new operations in Illinois or kind of prohibited new users from signing up.
0:10:41 Yeah.
0:10:52 It’s basically anything that’s kind of ongoing support or even personalized advice around specific emotional and mental issues is now counted as therapy and is technically illegal in Illinois.
0:10:56 I am confident ChatGVT is doing this and honestly is doing it well.
0:10:57 Yes.
0:11:03 For a lot of people and so I guess my question is to what extent is this ever going to be enforced because they can’t see people’s individual chats.
0:11:05 I feel like Illinois always does weird stuff.
0:11:08 Like, we’ve been consumer investors for too long.
0:11:23 And remember in, like, 2017, 2018, we would literally talk to social apps, like, consumer social apps that were, like, we’ve launched everywhere except for Illinois because they have all these, like, crazy regulations around, like, people, like, data and sharing and, like, all of these things.
0:11:33 Which, obviously, it’s good to have those, but, like, went way beyond other states to the point where it made it difficult for apps to operate there, which is, in my opinion, then bad for the consumer.
0:11:42 I think there are, like, a lot of people are sort of now grappling with this question of what does it mean for AI to offer medical support or mental health support.
0:11:49 I don’t expect we’ll see the other states go in the direction of Illinois, partially because it’s just so hard to regulate.
0:11:55 Like, how can you control what someone is talking to their chat GPT or clot or whatever about?
0:12:01 And especially because GPT-5 was kind of trained or fine-tuned, at least, with data from real physicians.
0:12:01 Yeah, yeah.
0:12:02 Yeah.
0:12:04 So they talked about this a lot in the live stream.
0:12:06 And I was surprised they leaned in on this.
0:12:09 I’m sure we’ve all seen the viral Reddit posts about, like, chat GPT saved my life.
0:12:11 My doctor said, wait for this imaging scan.
0:12:14 It turns out I had this horrible thing that I was able to get treated immediately.
0:12:23 And Sam and Greg Brockman had been retweeting these posts for a while, which I was like, that’s interesting because you think they’d, from a liability perspective, they’d want to avoid that.
0:12:38 But they had a whole section of the GPT-5 live stream where they brought up someone who had cancer and was using ChatGPT to upload all of her documents, get suggestions about treatment, kind of talk through the diagnosis and what she could do next.
0:12:51 And they talked about how GPT-5 was kind of the highest-scoring model on this thing called HealthBench, which is a benchmark they trained with 250-plus physicians to measure how good an LLM is at answering medical questions.
0:13:01 And so I think it’s kind of a really big statement that OpenAI has leaned into this space so heavily versus being like, hey, there’s a lot of liability around medical stuff.
0:13:04 Our AI chatbot is not a licensed doctor.
0:13:07 We’re going to kind of let people do this off-label, but we’re not going to endorse it.
0:13:10 It seems like now they’re really endorsing it.
0:13:11 Yeah, I’m excited.
0:13:11 Me too.
0:13:14 I upload all sorts of stuff and get all kinds of advice.
0:13:17 And it can be really smart and really helpful in a lot of cases.
0:13:18 I agree.
0:13:22 There were two other big actually creative tool model releases this week.
0:13:22 Yes.
0:13:26 Genie 3 from Google and then a new music model from Eleven Labs.
0:13:28 So maybe let’s start with Genie 3.
0:13:30 I’ve seen the videos, but what is it?
0:13:30 Yes.
0:13:32 Genie 3 took Twitter by storm.
0:13:32 Yeah.
0:13:38 So Google has a bunch of kind of different initiatives around image, video, 3D world.
0:13:45 I think various teams like VO3 and the Genie team working towards this idea of like an interactive world model.
0:13:54 Which is basically you are able to have a scene that you can walk through in real time or interact with that kind of generates on the fly.
0:13:57 And you can imagine it sort of like a personal video game.
0:14:05 Yeah, I saw some of the videos of kind of taking famous paintings and for the first time you’re able to step into them and kind of swivel around and move around in the world.
0:14:10 Almost like you have a VR headset on or something and you’re kind of turning around and seeing the full environment.
0:14:11 Those were really cool.
0:14:12 And it’s not just famous paintings.
0:14:17 They’ve shown a bunch of examples of from a text prompt you can create a world, from an image you can create a world.
0:14:18 Yeah, amazing.
0:14:23 They’ve even shown taking VO3 videos and creating a world around it with Genie 3.
0:14:28 And the cool thing about Genie 3 is there’s controls where you can move the character around.
0:14:35 So you can control like now go to the left and then the scene sort of regenerates to show you what you would see on the left.
0:14:36 It’s incredible.
0:14:38 They haven’t released it publicly yet.
0:14:42 They invited some folks to try it out at their office who were kind of sharing results.
0:14:43 They shared a bunch of clips.
0:14:45 I’m personally really excited to get my hands on it.
0:14:52 I think the natural question we’ve all had with this use case and seeing the demos is like this looks amazing.
0:14:53 Like what are we going to do with it?
0:14:53 Exactly.
0:14:55 And it’s expensive.
0:14:55 Yeah.
0:14:57 And probably takes a long time.
0:14:59 Like they haven’t released the stats around that.
0:14:59 Yeah.
0:15:00 Exactly.
0:15:02 I think there’ll be a couple use cases.
0:15:11 So I think video is an obvious one where if you’re generating the scene in real time and then controlling how you or any character or objects are moving through it,
0:15:19 that enables much more control over a video that you could then kind of screen capture what is happening than you would get from a traditional video model.
0:15:27 So you’re almost recording the video – you’re recording video as you move through the 3D world model, which then becomes like a movie or a film, essentially.
0:15:34 Our portfolio company, World Labs, has a really cool product out that I’m on and a number of folks are on that does this.
0:15:40 And Martine on our team shares a bunch of really cool examples of stuff he makes with exactly that use case.
0:15:40 Yeah.
0:15:43 So much more controllable video generation, which is huge.
0:15:49 I think in gaming, like there’s kind of two paths that this can go and it could go both.
0:15:57 One is it allows real game developers to create games much more quickly and easily where you don’t have to kind of code up and render up an entire world.
0:16:03 It can just generate from the initial image or text prompt you provide and the guidance you give it.
0:16:11 And then you can imagine, like, could a game developer freeze that world and allow other people to play it like a traditional game?
0:16:19 So it’s like the game then would be the same for every person versus in the first example, the game almost regenerates for everyone fresh as they move through it.
0:16:19 Right.
0:16:20 Okay.
0:16:25 But I think the second gaming example is more like kind of what you’re alluding to, which is more personal gaming.
0:16:33 Which is like every person puts in an image or video or text prompt and then is sort of creating their own minigame where they’re wandering through a scene.
0:16:33 Yeah.
0:16:38 Which is sort of a totally new market that I think a lot of people will love.
0:16:38 Yeah.
0:16:49 And then the third example, which is a little out of our wheelhouse, but a lot of folks are talking about how creating sort of these interactive dynamic worlds are really good RL environments for agents to be trained on.
0:16:55 And how to interact with the world, like how things move, going around scenes, interacting with objects.
0:16:57 It’s been a big space of conversation right now.
0:16:57 Yeah.
0:16:59 And sort of there’s a desperate need for more.
0:17:04 There’s so many companies now selling these RL environments for agents that they’re manually creating.
0:17:04 Yeah.
0:17:12 And something like a Genie 3 could make that much easier and sort of allow you to generate unlimited environments for these agents to wander through and learn from.
0:17:16 I could see that for digital agents, but even like physical agents operating within robots or something like that.
0:17:17 Yeah, yeah. Totally.
0:17:17 Yeah.
0:17:21 I think for all sort of agents or like self-learning systems, it’s going to be fascinating.
0:17:22 That’s awesome.
0:17:24 So eagerly awaiting that one to come out.
0:17:27 And then, yes, our portfolio company, Eleven Labs, also released their music model.
0:17:28 Yeah.
0:17:29 Which is super exciting.
0:17:30 I did not know they were working on music.
0:17:31 Yes.
0:17:32 It’s been in the works for a bit.
0:17:42 The really interesting thing about it is it’s trained on fully licensed music, which means – so music is one of those spaces where the rights holders are extremely litigious.
0:17:43 Yeah.
0:17:51 And so compared to things like image or video, it’s been harder for music companies to sort of avoid stepping on the toes.
0:17:52 Yeah, because you can’t just scrape data from the internet.
0:17:55 Like the record labels will come and sue you.
0:18:02 Yes, and the artists and like – it’s often a very complicated ecosystem of who owns the rights to a specific song or to an artist’s voice or something like that.
0:18:16 And so, yeah, I think a lot of folks have thought that you could not get a good quality music model training on licensed data because it’s hard and it’s expensive and it takes a long time and it’s hard to get them to agree to license you the data.
0:18:21 But from what I’ve seen and from my own experiments, folks have been really impressed by Eleven’s outputs.
0:18:26 And so what does the licensed data open up in terms of use cases for the music model, do you think?
0:18:26 Yeah.
0:18:39 So I think a lot of consumers basically don’t care if they’re using a music model that’s trained on licensed data or not because they’re not really monetizing or many of them are not monetizing stuff that they make with this music.
0:18:44 They’re generating like a birthday song for their friend or a meme clip or something like that.
0:18:46 Or like background music for their AI video.
0:18:46 Yep, yep.
0:19:01 Whereas businesses, enterprises, big media companies, gaming companies, like they care and they need to be able to say this music model we used was trained on fully licensed data to not kind of open themselves up to any liability issues.
0:19:07 So they could use this music hypothetically in like advertisements or films or TV shows or things like that.
0:19:08 Exactly.
0:19:12 Which I think is a big step forward for AI music as a whole.
0:19:15 And I think we should expect to see more from Eleven on this front, which is very exciting.
0:19:16 Awesome.
0:19:21 And then our last big topic of this week is around vibe coding, which continues to explode.
0:19:22 Yes.
0:19:23 I think we have two things to talk about here.
0:19:25 One would be our own experiments in the world of vibe coding.
0:19:26 Yes.
0:19:33 Which relates to a piece that you and Anish Acharya put out this past week about how we’re seeing the vibe coding market start to fragment.
0:19:33 Yes.
0:19:36 Your experiment is the more interesting part.
0:19:36 So my—
0:19:37 So let’s start with that.
0:19:37 Yeah.
0:19:43 Maybe to give a real world example, for the first time I vibe coded an app that I fully published and made available to the internet.
0:19:48 Essentially what I did was I thought, hey, I’m seeing on my X feed all the time.
0:19:50 Everyone has a selfie with Jensen and NVIDIA.
0:19:51 How did they get this?
0:19:52 In his classic leather jacket.
0:19:57 He must be spending all of his time taking selfies now because everyone has one and I don’t.
0:20:06 And so I was thinking there’s all these new amazing models out there like Flux Contacts that can kind of take an image, say, of Jensen taking a selfie with someone else and put my—stick me in there instead.
0:20:07 Replace me in the selfie.
0:20:08 Yes, I should have been in the video.
0:20:09 I should have been in the video.
0:20:10 Exactly.
0:20:12 So I did that.
0:20:14 I generated that myself on Kriya.
0:20:18 And then I thought, I bet other people might feel like me and might want this.
0:20:22 And so I’d love to create an app where anyone can upload a photo and get a selfie with Jensen.
0:20:23 Yes.
0:20:25 And so I thought, okay, I can vibe code this.
0:20:31 So I vibe coded on Lovable, an app that connected to Fall to pull in the Flux Contacts API.
0:20:31 Yep.
0:20:35 And then you could upload your own photo, would generate the selfie with Jensen, which you can then download.
0:20:35 I used it. It was great.
0:20:37 It worked super well.
0:20:38 So I published it on Twitter.
0:20:39 Yes.
0:20:40 A lot of people used it.
0:20:50 It got used by like 3,000 people overnight to the point where when I woke up, I had exhausted my self-imposed budget of $100 to spend on API calls here.
0:20:50 Yes.
0:20:52 Because you were funding it.
0:20:53 I was funding it myself.
0:20:54 You weren’t making people put in their API key.
0:20:57 I was not making anyone pay for it or put in their own API key.
0:20:59 So it was to the point where I had exhausted it.
0:21:04 So instead of calling the model, it was just kind of stitching together half of your photo with half of Jensen’s photo.
0:21:04 Love that.
0:21:09 To produce a really kind of 2005 Microsoft Paint looking output, which has its own charm.
0:21:10 Yeah.
0:21:25 Anyway, but the surprise was, so one, the fact that someone who’s completely non-technical can build something that thousands of people, and I did it in like a couple hours in an evening, if that, that a couple thousand people can use overnight is like amazing and so exciting.
0:21:25 Yes.
0:21:33 My second learning was, however, we’re early in vibe coding because these products are definitely built for people who are already technical.
0:21:34 Yeah, there were some issues.
0:21:35 We had some issues.
0:21:35 Which you should talk about.
0:21:38 You should not expose your public API key.
0:21:43 Well, the problem is you didn’t even know you were exposing your public API key until some nice man DM’d you and was like.
0:21:48 So the vibe coding platforms are, I think they assume that you have a certain level of knowledge already.
0:21:53 So if you go to publish a website or an application, they won’t stop you and say, hey, here’s a security issue.
0:21:54 Here’s a compliance issue.
0:21:56 Fix this before you publish.
0:22:03 And so it was a really interesting learning experiment for me, I think, first of like, I think there will be, and this is what you got at in your blog post.
0:22:03 Yes.
0:22:09 There’ll hopefully be a V2, V3, V5 of these vibe coding platforms that are built for people who don’t know these things already.
0:22:15 So two things people flagged to you that the vibe coding platforms did not was, one, your API key was exposed.
0:22:21 And two, you had not created like a protected database for the photos.
0:22:22 For the photos that were updated.
0:22:26 So it was like if you knew how you could go and access the selfies that were uploaded.
0:22:26 Yes.
0:22:26 Yeah.
0:22:28 And I’ve vibe coded many things.
0:22:29 Which I fixed, to be clear.
0:22:39 I’ve had similar problems, vibe coding a lot of apps, where like, I feel like they assume you have a level of technical knowledge to be able to fix things or to even know what a potential problem could be.
0:22:49 And so, yeah, Anish and I published, Anish was actually an engineer, and I collaborated on this post around basically how we think vibe coding will evolve in the future.
0:22:54 I think today you have a bunch of awesome platforms that are trying to be everything to everyone.
0:23:04 They’re saying like an engineer at a company can use this to develop internal tools, or someone can use this to build a SaaS app that scales to hundreds of thousands of users.
0:23:08 And a consumer can also use this to create a fun meme app.
0:23:08 Yep.
0:23:19 But I think the truth, in terms of what we’ve seen at least, is those are very different products, both in terms of the use cases, the integration, and the level of complexity required.
0:23:33 And there probably should be, for example, a platform that is like the training wheels version of vibe coding for like consumer non-developers like us that does not allow you to make mistakes like exposing the API key.
0:23:36 Yes, even if it then means less flexibility in the product.
0:23:37 Exactly.
0:23:40 Like I wasn’t super opinionated about what it looked like, all of the specific features.
0:23:41 I just wanted it to work.
0:23:49 Yeah, you probably weren’t super opinionated on like the coding language or exactly what database was using, like all the back-end stuff of what it was built on top of.
0:23:51 You didn’t really care.
0:23:52 You just wanted it to work.
0:24:00 Whereas there’s many like enterprise or true developer use cases where they very much want to control every element of the stack.
0:24:05 And that level of inflexibility, that product would just not work for them.
0:24:06 Yeah.
0:24:18 And so I think what we are hoping to see basically is like specialized players emerge that offer the best product for a particular demographic of user or for a particular use case.
0:24:27 And that will probably imply very different product experiences, product constraints, and also go-to-market strategies.
0:24:39 Like if you are allowing any consumer to vibe code like a fun app to share with their friends or partner or whatever, you probably want to be going viral on TikTok and Reels.
0:24:52 Versus if you are building a vibe coding platform for designers to prototype new features in an enterprise or for engineers to make internal tools, you might want to even have top-down sales.
0:24:52 Yeah.
0:24:55 Or at least be kind of product-led growth within businesses.
0:24:59 Yeah, and maybe invest in like deep integrations into core business systems or things like that.
0:25:05 Whereas the consumer version, you might actually just want people to vibe code on mobile and get something that works in five minutes.
0:25:14 And that’s a great point too, which is like the consumer users often just want something to look cool and work and not have security issues.
0:25:20 Whereas more business-oriented users, it often needs to integrate with what already exists for the business.
0:25:32 Whether that’s like a design system and aesthetic or whether that’s, you know, their CRM or the emailing platform they use or sort of all of these different products that it needs to connect to that are external to the vibe coding tool.
0:25:37 And so I think the conclusion of the piece was like we’re seeing early winners in vibe coding already.
0:25:41 These are some of the fastest growing companies in the AI application space.
0:25:45 But we probably expect to see even more because it feels like we’re so, so early.
0:25:49 And many of the users of these products are probably still pretty technical.
0:25:50 Yes.
0:25:53 And so there’ll be a version of vibe coding that’s truly consumer grade.
0:25:53 Yes.
0:25:56 That I’m personally very excited to unlock.
0:26:00 And I think we’ve seen this in a lot of AI markets because these markets are large enough.
0:26:02 They can have multiple winners that are specialized.
0:26:03 Like we’ve seen this for LLMs.
0:26:03 Yep.
0:26:07 Open AI, Anthropic, Google, Mistral.
0:26:08 Like there’s XAI.
0:26:12 There’s all of these companies that have models that are really good at particular things.
0:26:13 Different things, yeah.
0:26:20 And we’ve also seen this a ton in image and video, which I think has a lot of corollaries to vibe coding, which is based on,
0:26:27 what type of user you are or what you care about, to what extent you need to reference an existing character, an existing design format, something like that.
0:26:30 Do you want it on your phone and super fast?
0:26:30 Yeah.
0:26:33 Or do you want it in the browser and slower and the highest quality?
0:26:39 Like there are many companies that are doing super well focused on different segments or verticals of this giant market.
0:26:39 Yeah.
0:26:40 Super exciting.
0:26:42 Well, thanks for joining us this week.
0:26:48 If you’ve tried out any of these creative models or had any vibe coding experiments yourself, we’d love to hear from you.
0:26:50 Please comment below and let us know.
0:26:56 And also, please feel free to ping us here or on Twitter if you have ideas of what we should cover in a future episode.
0:27:01 Thanks for listening to the A16Z podcast.
0:27:07 If you enjoyed the episode, let us know by leaving a review at ratethispodcast.com slash A16Z.
0:27:09 We’ve got more great conversations coming your way.
0:27:11 See you next time.
0:27:26 As a reminder, the content here is for informational purposes only, should not be taken as legal business, tax, or investment advice, or be used to evaluate any investment or security, and is not directed at any investors or potential investors in any A16Z fund.
0:27:31 Please note that A16Z and its affiliates may also maintain investments in the companies discussed in this podcast.
0:27:39 For more details, including a link to our investments, please see A16Z.com forward slash disclosures.
0:27:43 you
a16z partners Olivia and Justine Moore unpack the latest in consumer AI including:
– Grok’s “Imagine” and its instant, social-first creative tools
– Google’s Genie 3 and the future of 3D worlds
– GPT-5: what’s new, what’s missing, and why some want their old chatbot back
– AI-generated music from ElevenLabs
– Olivia’s vibecoded Jensen Huang selfie app
Timecodes:
0:00 Introduction & This Week’s Topics
0:24 Grok Imagine: Social AI Image & Video Generation
4:48 GPT-5 Release & GPT-4 Deprecation
5:36 Comparing GPT-5 and GPT-4: Coding vs. Personality
9:13 AI for Mental Health: Illinois Law & Industry Impact
12:29 Genie 3: Interactive World Models from Google
16:53 ElevenLabs Music Model: Licensed AI Music Generation
19:16 Vibecoding: Consumer Experiments & Platform Evolution
24:14 The Future of Vibecoding & AI Tools
27:05 Conclusions
Resources:
Find Olivia on X: https://x.com/omooretweets
Find Justine on X: https://x.com/venturetwins
Read Anish and Justine’s vibecoding post: https://a16z.com/specialized-app-gen-platforms/
Stay Updated:
Let us know what you think: https://ratethispodcast.com/a16z
Find a16z on Twitter: https://twitter.com/a16z
Find a16z on LinkedIn: https://www.linkedin.com/company/a16z
Subscribe on your favorite podcast app: https://a16z.simplecast.com/
Follow our host: https://x.com/eriktorenberg
Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.