I Got Early Access to Runway 4.5 + Kling AI Demo

0
0
AI transcript
0:00:06 Hey, welcome to the Next Wave Podcast. I’m Matt Wolfe. And once again, I’m here with
0:00:11 Maria Garib, and we’re going to talk about some of the coolest things that are happening
0:00:16 in the AI world right now. And at this current moment in time, the things that are happening
0:00:22 in the AI world seem to all be mostly around AI video. We’re getting a lot of cool AI video
0:00:27 models from companies like Runway, from Kling. There was even the little tease about a new
0:00:32 Apple video model that they tease. We don’t actually have any footage of that, so there’s
0:00:36 nothing to show there, but Apple’s been teasing a video model. So in this episode, we’re going
0:00:41 to dive in. We’re going to test some of these AI video models. One of these models I even
0:00:46 got early access to, so we’re testing it. We’re going to be seeing what the model creates for
0:00:51 us with these prompts for the very first time on this recording using Runway. So it’s going
0:00:54 to be a fun episode. Thanks again, Maria, for hanging out with me on this one. It’s going
0:00:58 to be interesting. We’re going to put these AI video models through their motions here.
0:01:07 Being a know-it-all used to be considered a bad thing, but in business, it’s everything.
0:01:14 Because right now, most businesses only use 20% of their data, unless you have HubSpot,
0:01:19 where data that’s buried in emails, call logs, and meeting notes become insights that help
0:01:25 you grow your business. Because when you know more, you grow more. Visit HubSpot.com to learn more.
0:01:32 Well, yeah. Thanks for having me again, honestly. I’m having so much fun on these episodes,
0:01:37 honestly. And I’ve been getting really good feedback so far. So yeah. Let’s go into what
0:01:42 Runway has for us today. You did say you have access to something people don’t have access to,
0:01:43 so I really want to see that, like now.
0:01:50 Yeah, yeah. So here’s the thing. I have access to it right now as of this recording. But by the time
0:01:55 this recording’s out, I’m pretty sure everybody’s going to have access to it. So what we’re showing off
0:02:01 in this video, you should be able to do by the time this video is out. But right now,
0:02:04 at the time of this recording, this is brand new, fresh stuff. We haven’t even seen what
0:02:08 this model can generate yet. Yeah. But we really want to see like, what’s new with it? Because
0:02:13 we did cover it last week, but like, I want to know what’s new. What do they have up their sleeves
0:02:23 right now? All right. So this is the Runway ML backend. And they just released Runway Gen 4.5.
0:02:29 In fact, as of this recording, it’s not even released yet. And I have no idea what to expect.
0:02:34 We’re doing it from scratch right now with you. This is from scratch. This is our first time
0:02:39 seeing it. I have seen some of the like demos and stuff that were circulating on X. However,
0:02:44 those are almost always cherry picked, right? Like if a company is putting up a demo video of like,
0:02:49 here’s what our model can generate, they’re going to take the best of the best generations they’ve made
0:02:55 and share those. So we’re going to test some real stuff and find out if it really is good.
0:02:59 Yeah. You know, one of the things about being a content creator that also shares some of these
0:03:05 video models is there’s a lot of sitting around and waiting, right? Like you enter a prompt and then you
0:03:10 sit there for anywhere from two to five minutes and wait for your video. Yeah. So I want to get a few
0:03:15 different prompts started. So we have some stuff going here. Yeah. I don’t know. Something came up.
0:03:21 I had like a meeting with like a coworker today and she was talking about like her creating like an AI
0:03:26 sort of thing that can help people with meal prepping and stuff. Let’s say someone has like a
0:03:30 company of meal preppers and like, what if they were in the marketing team and they don’t have the
0:03:36 budget to have all these actors and they want to create like an AI video? What would be the product
0:03:42 of that? You know, using runway, like a meal prepping ad, let’s say. So what should I enter
0:03:48 for the prompt then here? You want to try something like very generic, like an ad for a meal prepping
0:03:54 business and see what it does for us? Yes. And then if not, let’s just do the extra work with all these
0:04:01 specific things. All right. Let’s see what it does. If we just give it the generic plan of like create an ad
0:04:07 for a meal prepping business. All right. I’m going to leave it at five seconds just to make sure we’re
0:04:15 getting faster generations here and I’ll kind of keep an eye on how long that takes, but we can get
0:04:20 another prompt queued up and going as well. I always test this prompt, a monkey on roller skates. I don’t
0:04:25 know why. That’s like my benchmark. I tested a monkey on roller skates back when model scope came out
0:04:32 in, uh, I believe it was beginning of 2023 or end of 2022. There was this model called model scope and
0:04:38 it was one of the very first AI video generation models we’ve ever seen. And the very first thing
0:04:42 that popped into my head, like two and a half years ago when that model came out was a monkey on
0:04:47 roller skates. So that was the first thing I ever tested. And it just looked like a little blur. If you
0:04:52 squinted and like got real close to your computer, maybe you could kind of see a monkey out of it,
0:04:56 but it was very, very primitive. But ever since I’ve always kind of used that as like my test,
0:05:01 like how far have we come generating monkeys on roller skates? We’ve come so far, honestly,
0:05:06 I don’t know what we’re going to come up with. Maybe even have not an avatar, but like a hologram.
0:05:10 Oh, a hologram. Yeah. Yeah. Can you imagine hologramming your whole house? So like having
0:05:14 like your favorite celebrity in your house, talking to you, like giving you a pep talk or something.
0:05:18 Yeah. Yeah. Yeah. Because you AI generated it. That would be so funny.
0:05:23 So we’re 61% here. It’s pretty fast. Not bad. All right. So what’s another prompt we can get
0:05:26 going? I don’t know how many it’s going to let me do simultaneously, but let’s keep going.
0:05:32 Let’s do, you know, the dinosaur that has like mittens. I’ve seen this picture before. It makes
0:05:36 me so funny. Oh, like a T-Rex? Yes. That would be so cool. Please do that.
0:05:38 A T-Rex. Making brownies.
0:05:43 Wearing mittens, making brownies on Mars. Why not?
0:05:49 I saw someone literally just take a picture from Majority because people on the Discover,
0:05:54 you know, Majority keeps the results of a lot of people. And like he put it on click and it was
0:05:59 like image to video. And it basically came to, that was so cool.
0:06:03 Yeah. We should test that as well. Cause we can start with an image here. So that would be
0:06:08 something we should definitely test. Yeah. We did get this back, create an ad for a meal prepping
0:06:14 business. Let’s see what it did. Oh, what is going on in that shot? When she’s trying to eat,
0:06:16 this is like flapping around. What is that thing she’s trying to eat?
0:06:20 That’s so funny. Okay. Okay. So I mean, like,
0:06:25 what are your initial impressions? Let’s just freeze frame on something here.
0:06:30 I think the more detailed you are with your prompts, the better the results in my opinion.
0:06:35 I agree. I agree. I think it looks a bit weird, obviously, but I think the more that you give
0:06:38 it, the more that you feel it, because I don’t know what she’s eating. What is she eating though?
0:06:41 Yeah. I don’t know. It actually looks like she’s spitting something out in the beginning.
0:06:44 Yeah. Where did this tray just come from? It just like appeared out of nowhere.
0:06:50 She generated it. So there’s definitely some like physics issues going on, but I mean,
0:06:53 the person looks pretty realistic, especially in this shot.
0:06:56 The kitchen looks pretty good. In the beginning with the food falling out, it was weird, but
0:07:01 yeah. I don’t know. That was funny. That was funny. We also didn’t give the prompt much to
0:07:06 work with. No, it was just like one five word sentence. Yeah. Yeah. All right. Let’s see. So
0:07:12 now we’ve got our monkey on roller skates here. And I mean, that’s cute. That’s pretty good. Actually.
0:07:17 I would play it in the background. Yeah. That’s like a nice screen saver. That is pretty cute. Honestly.
0:07:22 Yeah. This has come a long way. No question. Right. You send that to somebody. That’s a monkey
0:07:26 on roller skates. You’re not going to question what you just saw. No, but it’s pretty cute actually.
0:07:33 And then, all right, here’s our T-Rex wearing mittens, making brownies on Mars. So one thing
0:07:36 to notice, and this is something I always look for when I’m looking at new video or image models
0:07:41 is when you stuff a bunch of things into the prompt, did it get all the things? Right. And we have
0:07:47 a T-Rex. We got some mittens. They’re kind of weird mittens, but we got some mittens and they’re
0:07:53 definitely on Mars. It’s a T-Rex, Matt. It’s going to wear weird mittens. That is true. What kind of
0:07:58 mittens would a T-Rex wear? I mean, it’s decent, right? Because it got all of the elements. And that’s
0:08:02 one of the things that a lot of these models struggled with early on, where if you give it like
0:08:07 three or four elements that you want them all in the same image, it would really struggle to put all
0:08:11 those. You might get one or two, but the other couple things you added to the prompt, it would miss.
0:08:17 These models have gotten so, so, so much better at making sure everything you plugged into the
0:08:23 prompt actually makes it into the video. Yeah, that’s pretty cool. Honestly, runway have come so far.
0:08:29 All right. So for the next test, let’s go to ChatGPT here and have ChatGPT generate a prompt. And the
0:08:35 reason I like doing this is I tend to come up with fairly vague prompts where ChatGPT will sort of
0:08:41 flesh it out for us, right? It’ll add more details. It’ll add scene details and color details and lighting
0:08:47 and maybe even like what camera it was shot on. ChatGPT helps with a lot of that kind of stuff to
0:08:51 give you a more of like fleshed out prompt where you could just start from like a general concept.
0:08:53 What’s another concept?
0:08:57 Let’s give it a role. Like let’s tell it like photographer, like something and like tell it
0:09:02 like you are this and I need to create an ad for this company, for example.
0:09:12 All right. You’re a world-class cinematographer and DP, generate a video prompt for me. Include
0:09:23 scene details, camera, camera angle, and any other details necessary to get a good shot.
0:09:28 Create the prompt about, let’s see, what’s a scene we can try?
0:09:33 Even though I read fantasy, I don’t know why the scenes are not. Let’s do a dragon though.
0:09:41 Okay. Create a prompt about a dragon in a colorful fantasy world. Well, let’s try that one next and
0:09:49 we’ll test to see if it can do like actual IP. All right. So it’s giving us a pretty detailed prompt
0:09:53 because look at this. It’s actually giving us like broken out bullets of camera and movement,
0:09:58 lighting and atmosphere, world aesthetic. Okay. So this was it sort of breaking down the scene and
0:10:02 then this is the actual prompt. And then there’s a prompt. I thought it was going to have me paste
0:10:06 this whole thing in, which it probably would have worked, but I’d prefer to just give it a prompt.
0:10:09 But runway would break down. Runway would cry. Runway would run away.
0:10:13 All right. Should we make it like more than five seconds?
0:10:19 Well, it’ll let me generate up to eight. It gives me five or eight. So I’ll generate eight this time.
0:10:24 And let’s get that one going. Yeah. The other thing I want to test is I want to test starting
0:10:31 from an image. So let’s go to mid journey. Mid journey. Yeah. Oh, is that you? Oh yeah. That’s
0:10:36 supposed to be me holding a fire extinguisher and putting out a fire, but it’s an extinguisher
0:10:41 that’s creating a fire instead of putting out a fire. Setting you on fire. Yeah. Setting me on fire.
0:10:47 All right. So let’s start with an image prompt. Any ideas come to you for an image that we can start
0:10:56 with. Let’s do Hansel and Gretel. Yeah. I should do it in 16.9 actually. Use this prompt again.
0:11:04 Let’s go 16.9 and I’m going to generate it again. Oh, these are all very like cartoony style.
0:11:11 And by the way, you can search in my journey in the images like above of someone that already did
0:11:15 it and it could give you better results. Oh yeah. That’s a good idea. But the tab above it.
0:11:20 Although some of these are pretty good. Yeah. So their video generator doesn’t give you the option
0:11:26 to start with an image if you’re using Gen 4.5. So we only have text as the starting, it looks like on
0:11:32 this model. Yeah. Okay. Well, that’s good to know. So yeah, as of right now, it doesn’t look like we can
0:11:40 start from an image. All right. If you want the full breakdown of all these AI video tools,
0:11:47 we put together a complete guide between runway, cling, VO, and Sora. It includes universal prompts
0:11:53 that work across all these platforms, real test results with the same prompts, pricing breakdowns,
0:11:58 and a framework on when to use which tool. Get it right now. Click the link in the description.
0:12:06 Now, let’s get back to the show. All right. So here’s our cinematic shot of an iridescent dragon.
0:12:13 Let’s see. Oh. I mean, that’s pretty good. I’m not seeing a whole lot wrong with it.
0:12:19 It is pretty good. It’s very consistent. I’m not seeing like the dragon change shape. I’m not seeing
0:12:25 any of this stuff in the background disappear. I think it’s pretty impressive. That was pretty good.
0:12:30 I’m a little bummed. We can’t do an image prompt as a start. So that’s runway. Honestly,
0:12:34 it’s not that bad. I mean, the more that you give it, obviously the details, as we said,
0:12:39 the more it gives you the better results. The dragon thing is like absolutely amazing. I think the more
0:12:45 that you write into it and like you give it like, you know, the context of what it is and stuff,
0:12:52 like imagine doing it like Ghibli style or anime style. And it just comes out of the style of these
0:12:57 animes that you watch and stuff. So, so good. Honestly, so good. They’ve come so far,
0:13:03 10 times better than VO or Sora, honestly. When it comes to Nano Banana at all, Nano Banana is still
0:13:04 up here. But yeah.
0:13:13 Yeah. In my opinion on this 4.5 right now, I think it’s fairly on par with VO 3.1. I’m not seeing
0:13:19 anything about this video model that makes me like super, super blown away. Like it’s a huge improvement.
0:13:22 Like it’s definitely a huge improvement over what we got a year ago, two years ago,
0:13:29 but over the sort of like current crop of available models, I’m not really seeing like a huge leap
0:13:36 between VO 3.1 in this. And you know, Sora too is good as well, but Sora is like pretty good at
0:13:39 generating people and like sort of realistic scenarios.
0:13:39 Yeah.
0:13:43 We should give that to Sora. Honestly, I keep bashing Sora every single time I go on the show
0:13:48 and I don’t apologize by the way. I don’t know if like, I’m not going to say sorry about that
0:13:52 because it’s still sometimes sucks. Sora’s probably the most realistic people still.
0:13:53 People things, yes.
0:13:58 Like if you’re generating videos of people doing something, almost anything else outside of people
0:14:00 doing stuff, I don’t really feel like it’s that good.
0:14:07 And to be fair, you test a lot of these video tools, you know, like I write a newsletter and like,
0:14:11 sometimes I don’t have the time to do everything, but you’re like on the internet 24 seven.
0:14:13 So like you’re going to do that.
0:14:15 Yeah, you can just call me a huge nerd. It’s okay. I get it.
0:14:21 Last episode I did that. But I feel like you’d see the comparison, like for someone that doesn’t
0:14:25 really test them, I think it’s pretty good for me, but like, I think you’ve seen better,
0:14:25 you know, to be fair.
0:14:32 Yeah. I would say it’s like probably fairly on par with VO 3.1. To me, I’m not like blown away as in
0:14:37 like, this is a huge leap over what we already had. And one of the things that I always like to think
0:14:41 about, especially when I’m making videos on my main YouTube channel is like, what can we do now
0:14:47 that we couldn’t do yesterday? And if like the news I’m talking about doesn’t really have any sort of
0:14:53 implication in that world, like if there’s something new that got released and I try it and I go, okay,
0:14:57 but I’ve already been able to do that. This doesn’t make it easier. It doesn’t make it faster. It
0:15:02 doesn’t make it better. So what is the point of this announcement? And if I feel that way about
0:15:06 something, I usually just don’t even talk about it in my videos because it’s solving a problem
0:15:10 that’s been solved by another product already that I’ve probably already talked to someone did it
0:15:15 before. Yeah. And so with runway 4.5, I think my sort of initial gut reaction to it is like,
0:15:22 it’s pretty much doing the same thing we’ve already been getting from something like VO 3.1 or something
0:15:27 like the newest models of cling. It feels pretty on par with those. So I don’t really feel like
0:15:34 this is any sort of huge leap saying that VO 3.1 is very, very expensive to use. I think runway is more
0:15:39 economical. I don’t have the pricing in front of me, but I do think it’s a little bit less expensive to
0:15:46 use to do a lot of generations with VO. How much is VO? I think on like their $20 a month plan, you can
0:15:52 generate something like 10 images a month or something like that on their ultra plan. I don’t know if there’s
0:16:00 a limit of how much you can generate, but it costs $250 a month. Oh my God. It costs a car payment to
0:16:08 actually have an ultra plan from Google. It’s pretty crazy. So most people who use VO don’t actually use it
0:16:13 straight through Google. They’ll go to a tool like Leonardo or Cria. I think Higgs field has it in it.
0:16:19 There’s all these other platforms that use the API and baked it into their product. And you can actually
0:16:24 use it a lot cheaper if you don’t use it directly through Google for whatever reason. But anyway,
0:16:30 I feel like this model is pretty on par with what we’ve gotten already, but it might be a more
0:16:35 economical way to do what we’ve already gotten. Yeah. If something is just as good as what we’ve already
0:16:39 got, but you can do it faster and cheaper, that’s still a plus for people. And I feel like this might
0:16:44 do it slightly faster and slightly cheaper than VO. But would I say it’s a giant leap over VO? Not
0:16:49 really. That’s kind of my take. I’m pretty sure we’re going to see everything flooded on TikTok right
0:16:55 now because TikTok is like the, I think the age range is a bit on the, you know, Gen Z side and
0:17:00 like a bit of millennials and a lot of Gen Alpha. So I don’t know if like all of them have the amount of
0:17:05 money that people can buy their memberships into video. I think they’re going to go with like runway
0:17:12 for now, just to be able to get these high end videos because they blow me away. And this generation
0:17:17 is so creative, extremely creative. I will always joke about this. And a lot of people on the internet
0:17:24 joke about it. We’re definitely going to get scammed eventually by something because you wouldn’t know
0:17:28 the difference, but it’s pretty good. It’s pretty good. These generations, I think we’re going to see
0:17:33 a lot of it flooded on TikTok. Oh yeah. But I want to see on like Kling, do you think like we can do a
0:17:39 comparison of what runway can do? Like what Kling can do? Yeah. So a couple of weeks ago, Kling did like
0:17:43 five days of announcements where every day for five days, they rolled out like a new Kling model.
0:17:50 One of the models was called Kling 01. Kling 01, they’re claiming it’s like the nano banana of video,
0:17:55 right? So like you can give it an initial video and tell it to edit anything about the video and it
0:18:00 will edit that stuff, right? You can re-skin it with a different design, take a real life video,
0:18:07 turn it into a studio Ghibli style. It’ll sort of re-skin the video in studio Ghibli, or you can have,
0:18:10 you know, somebody walking through the shot and then say, Hey, remove the people in the background
0:18:16 and it’ll remove those in the video or add Voldemort in the background and it’ll add a walking up to you
0:18:20 or whatever. Right? So it’s like, they’re claiming it’s like the nano banana video and that you can give it
0:18:26 a video image or text like any of these things. And it will try to combine them and edit them
0:18:33 together and create a video out of it. But there was also Kling image model, which was like identical
0:18:37 to nano banana. It created images, but you can edit images with it. Pretty much anything you can do
0:18:41 with nano banana, you can do with the Kling 01 model, but it was an image model, not a video model.
0:18:51 Then they released Kling video 2.6. Kling video 2.6 is kind of like their version of VO. It will
0:18:55 generate audio. The one we were just looking at runway ML was not generating audio in the video.
0:19:01 This new Kling 2.6 generates video or audio into the video, just like VO and Sora do.
0:19:05 They also released an avatar model, which I’ve never played with. Creating AI avatars isn’t really
0:19:10 something that I have a ton of interest in. So I haven’t touched that one yet, but I have played
0:19:18 around a little bit with video 2.6 and Kling 01, the nano banana of video. So those are the ones I have
0:19:23 played with. What I can do is I can start by showing some of the stuff I already generated while
0:19:26 we’re waiting for some of our new generations to happen.
0:19:30 Yeah. And that’s also run what GPT wrote for like the dragon thing and like put it here.
0:19:33 It’s just so it compares to see what it comes up with.
0:19:42 Yeah. So this is the video 2.6. This is like their version of runway 4.5 VO 3.1 Sora 2. Like this is
0:19:49 their top of the line video generator here. So I’m going to grab our prompt from runway, paste it back
0:19:56 here. So this is the same dragon shot that I’m submitting. This one can do five or 10 seconds.
0:20:04 I’ll do 10 seconds on this one and we’ll go ahead and generate our dragon. We can throw in our T-Rex
0:20:10 wearing mittens making brownies on Mars. I played around today because like I was really curious
0:20:15 about like what it can do and stuff. And I was like creating content for like my page. And I don’t know
0:20:21 if you’ve seen that movie, Death Becomes Her, Meryl Streep. Oh yeah, yeah. It was a long time ago.
0:20:25 That’s like an older movie, right? There’s an amulet that basically gives you immortality and stuff.
0:20:30 And like it has some kind of potion. And so I took that picture and I put it in Kling and I said,
0:20:35 the amulet is sort of like glowing or like kind of like radiating. And the results were pretty cute.
0:20:36 Like I’ll send it to you.
0:20:40 Oh, cool. Actually, let’s throw in our image to video too, because we did generate an image
0:20:45 inside of mid journey. We didn’t get to play with that one. So I’ll use that as our starting frame.
0:20:51 We’ll pull in our Hansel and Gretel image here. All right. Let’s generate that one. Okay. So we’ve got
0:20:56 those generating. While those are generating, I can kind of show you some of the other stuff that I did.
0:21:01 Yes, please. So the first thing I tested was I wanted to test like the lip syncing.
0:21:08 So I made this video of a rapper wrapping into the camera because this one does come with audio.
0:21:19 So, you know, the audio wasn’t great. He was supposed to say lights, flash, systems, crash,
0:21:23 moving in a digital dash. That’s what he was supposed to say. But he kind of goes lights,
0:21:28 lights, flash, flash, right? And he’s just kind of like repeating himself. So the words I asked him to
0:21:33 say didn’t quite come out correctly. But I mean, what are your thoughts on the video?
0:21:38 I mean, the hand stuff are still like a bit iffy. But like, yeah, it feels like he broke
0:21:43 seven phalanges and stuff. But yeah, yeah, his fingers look totally jacked right there.
0:21:48 Or he has like, he’s double jointed and like, you know, like, like me, because I’m double jointed as
0:21:54 I could be that. But yeah, I mean, it looks good. I think if like you gave it more details. Yeah,
0:21:59 the speech thing was not on point, but it’s pretty good. Right? I don’t think others could do the same.
0:22:04 Yeah. I mean, VO, I think does that better. Honestly, VO 3.1, you’re going to get speech.
0:22:09 Yeah. But like, who has $200? Yeah. Okay. I mean, Sora does it pretty well, too. Like,
0:22:12 Sora’s lip syncing. I have a beef with Sora. I don’t want to talk about her today.
0:22:19 So the next thing I wanted to test here was sort of like sound generation for like,
0:22:24 does it generate somebody walking through leaves? So I basically said, have them walk on leaves,
0:22:29 have them walk on wood, have them walk on concrete. The obsession of men. And I wanted to see if the
0:22:33 sound sort of changed based on what they were walking on. It is an obsession. Men in general feel like
0:22:39 this obsession to basically pick up ice or something that’s frozen and I throw it away as well as walk on
0:22:45 dead leaves. I’ve seen it all over the world. It’s universal. There’s videos online of men basically
0:22:50 picking up the crunchiest leave and just stepping. It’s a men thing. That’s funny. I’ve never heard
0:22:55 of that trend. I didn’t know it was a thing. It’s a thing. But here’s what this video came out like.
0:23:05 I mean, it’s okay. To me, it doesn’t really sound like crunching leaves. It just kind of sounds like,
0:23:10 like, I don’t know. Yeah. Someone opening up like a chip bag or something. It’s fine.
0:23:15 Yeah. It’s, it’s okay. Like I’m not like blown away, but it’s, you know, it’s decent. So then I
0:23:20 wanted to do one to see if like the audio would change drastically where like they’re in a loud
0:23:25 nightclub and then they walk outside and then it’s like quiet outside. And this is what I got with that.
0:23:33 Oh, interesting. I mean, yeah, but then it sounds like a horror movie when you come out,
0:23:39 like, you know, yeah, it’s definitely got some issues. Like, first of all, what kind of nightclub
0:23:45 just has like a gate to the outside when it’s snowing outside, right? Like it’s not a door.
0:23:49 It’s like a gate that you can see through. I think, yeah, I haven’t been to Berlin,
0:23:54 but I heard this is how they do their nightclubs. So I’m not going to assume. I think this is how the
0:24:01 nightclubs in Berlin are. But then, you know, when you get outside, the noise stops. But if you were
0:24:05 just behind a gate, would the noise actually stop? Because it’s just a gate, not a door.
0:24:08 Anyway, that’s probably a little too much to ask of the AI.
0:24:13 Also, why are they jumping so much? Like what kind of a party is that?
0:24:16 Yeah. Yeah. It’s just a giant cult gathering. That’s all it is.
0:24:19 I didn’t want to say it.
0:24:22 Oh, are we fighting? What is this?
0:24:27 So this is like, I was going to test dialogue. I wanted to see if the guy starts talking,
0:24:31 if the woman can interrupt. This is let’s just watch it.
0:24:34 I really think we should consider the stop.
0:24:38 I don’t want to stop him like she looks like she’s about to slap him.
0:24:45 The funny thing is, though, like when she yells stop, he also yells stop at the same time.
0:24:47 Watch his face when she yells stop.
0:24:54 So like they both yell stop together for some reason, which is weird imitating her or something.
0:24:59 It’s like, I know you’re about to yell stop, so I’m just going to say it before you can.
0:25:03 And then I also say like the prompt, she’s supposed to raise her hand and go stop.
0:25:07 But she yells stop and then raises her hand like she’s about to slap him.
0:25:11 She’s about to mark his face with like five fingers.
0:25:18 And then this one was meant to be like a little kind of like ASMR test with a marble rolling down
0:25:23 a plastic track. And then it was supposed to hit the dominoes and then smack the balloon and pop
0:25:28 the balloon. That’s what it was supposed to do. Keyword supposed to do contraption is that
0:25:33 I was just wanting to see if it would make the good sound of the ball rolling and the dominoes falling
0:25:36 and stuff didn’t quite come through.
0:25:39 Oh, what the?
0:25:44 What’s that thing coming out of the balloon?
0:25:48 I don’t know. It looks like shaving cream coming out of the balloon or something. I don’t know.
0:25:53 I think because it picks up like the ASMR that happens like on social media where like people
0:25:56 fill up balloons with like shaving cream. I don’t know if you’ve seen them.
0:25:58 Oh, is that a thing people do? I never see those.
0:26:03 It is. That’s probably why it’s picking up that kind of algorithm and like kind of getting,
0:26:04 you know, inspired by it, probably.
0:26:10 Yeah. Yeah. But I mean, the sound isn’t bad, but the visuals didn’t match my prompt at all,
0:26:14 right? The dominoes were supposed to be at the end of the thing. So the ball was supposed to hit
0:26:19 the dominoes, knock the dominoes down. The dominoes were supposed to bump the balloon,
0:26:24 right? It was supposed to be this chain of events, but it put the dominoes next to the like slide that
0:26:29 the ball came down and only knocks it over one domino. Then the ball itself hits the balloon.
0:26:33 And instead of popping, the balloon just kind of gets a little hole in it. And then shaving cream
0:26:37 starts coming out. That’s why. Yeah. So if you’re listening to the audio of this podcast and you’re
0:26:41 like, what’s the big deal? What is happening? Watch it. You need to come watch this.
0:26:45 Yeah. Just for you to understand what we’re seeing right now. You know?
0:26:52 Yeah, exactly. All right. So that was all video 2.6. Let’s see if our generations of video 2.6 are
0:26:58 done. So we’ve got our dragon shot that we’re comparing against runways. The dragon looks better
0:27:02 than runway, honestly. Let’s take a look at the runway just so we can compare them back to back.
0:27:06 So here’s what runway generated. No audio on runway. So if you’re listening on audio,
0:27:11 we’re watching the dragon video that runway generated and it’s pretty decent.
0:27:15 Pretty good. It’s pretty good. I like the colors of it. I think the colors are nice.
0:27:18 They’re vibrant, which is, you don’t see it a lot. Yeah.
0:27:23 All right. So now I’m going to jump back over here to cling 2.6. This is the exact same prompt.
0:27:25 I just copied and pasted it. Wow.
0:27:28 Much better physics. Much better physics. Oh yeah.
0:27:35 There is something wonky with the legs right here in the beginning. I get really nitpicky with these.
0:27:39 most people probably don’t care. But if you notice right here, he’s got three legs on one side and
0:27:45 one leg on the other side. Maybe he has a deformity. Like we shouldn’t judge. Oh yeah, I guess. Yeah.
0:27:47 Sorry. I’m dragon shaming now.
0:27:57 Oh, okay. But I actually think Kling did it better. Look at this face. It’s really good. Look at the scales.
0:28:02 Yeah. I like it. Overall, Kling did it better, I think. I mean, and Kling has audio where Runway
0:28:09 didn’t. But just like the physics and like the speed that the dragon moves and the speed the wings are
0:28:14 moving, it just feels more like what you’d see in a movie than the Runway version. It looks majestic,
0:28:20 honestly. That’s how you should see a dragon. Yeah. Yep. All right. Let’s try our dinosaur. There’s
0:28:25 a dinosaur wearing mittens. There we go. Yeah. Looks good. All right. So a T-Rex wearing mittens,
0:28:28 making brownies on Mars. Here’s what we got. This one’s going to have audio.
0:28:33 Almost done with the brownies. It was not expecting that. Me neither.
0:28:37 I was not expecting the dinosaur to start talking to us.
0:28:41 Different barefoot contesta or something. Like what is happening? BBC food right now.
0:28:47 I mean, it’s pretty good. There’s not a whole lot of motion to the video. It’s just the dragon
0:28:52 sort of stirring. And then he talks to the camera saying almost done with the brownies.
0:28:55 And the mittens look better. The mittens do look better. They look like mitten.
0:28:57 They look like what you’d expect mittens to look like.
0:29:01 Yeah. I think they gave them too long of arms though. I don’t think T-Rex has had that long arms.
0:29:03 Yeah. T-Rex should be like up here with the arms.
0:29:06 All right. Monkey on roller skates.
0:29:07 Having a blast on these skates.
0:29:15 It’s funny because the monkey’s just riding one roller skate. Like it’s a monkey on a roller skate.
0:29:18 Why is he hunched over like this? What are you doing?
0:29:23 Yeah. That one, I think I give it to Runway. Runway’s was a little more cartoony. This one went a little
0:29:27 bit more for realism. You could see the fur. Yeah. You could see the fur on the monkey.
0:29:32 I think it’s funny how like literal Kling makes these prompts where it makes the character turn
0:29:38 to the camera and say what they’re doing. Yeah. Almost done with the brownies. That’s the funniest
0:29:44 thing I’ve ever seen. Whenever you need to test out like other, we should probably run with the T-Rex
0:29:49 Rax making brownies on Mars. Yeah. Yeah. That’s a nice like new benchmark to work off of, I think.
0:29:54 All right. So create an ad for a meal prepping business. I wanted to test again of just like
0:29:58 vague prompt. Let’s see what it gives us. All right. So we have like a, you know,
0:30:02 a fitness looking woman on camera. Let’s see what she does and says.
0:30:05 Get your week’s meals ready in minutes.
0:30:07 I mean, there’s a lot less going on than the other one.
0:30:10 I mean, she’s not freaking out. She’s not spitting stuff from her mouth.
0:30:14 But I also feel like if you give it like more details, like I think she cut the container in
0:30:19 half, which is weird. Yeah. Right there at the very end, her tongs are almost going through the
0:30:25 container. Also, it becomes a knife. Is that right? No, it does. You’re right. It goes from tongs to
0:30:30 knife. That’s weird. But I think visually way better than the freak out. Yeah. She’s not like
0:30:36 vomiting food back out. She looks very realistic, right? Maybe a little like too sort of
0:30:42 polished and realistic. Like you get this thing with AI images and videos where the people almost
0:30:47 start to look like plasticky because they’re like almost too perfect skin looking and stuff, you know?
0:30:51 Yeah. So you have a little bit of that, but it’s, you know, more realistic looking of a person than
0:30:58 runway. Like if we look at runway here, the motion is fast, but the people, there’s just some like
0:31:02 weird uncanniness about like all their movements. I still don’t understand what happened to that
0:31:07 woman. Like someone ask her. So she’s spitting out. So we’re this one. Like it definitely looks
0:31:12 more like a real person. You don’t have some of the weird, like artifacts you get with the movement,
0:31:17 but there’s also a lot less movement going on. Yeah. This is better. All right. Let’s see what
0:31:22 our image to video did. So here’s our Hansel and Gretel. My prompt was Hansel and Gretel skipped
0:31:29 towards the cabin as the camera slowly rotates around them. Cute. Cute. And that actually came out pretty
0:31:34 decent. Yeah. Yeah. I was just scrolling on X because I generally, I’m not on X because I feel
0:31:41 like it’s too much for me. So I saw this thing. I was like really curious about like how people are
0:31:47 using cling. And like, I saw this person kind of take an image from the journey, put it on cling and
0:31:52 it turned out to be cinematic. It was like cinema. It was really good. Yeah. No, I tend to find that
0:31:58 when you’re using video models, you almost always get a better result. If you start from an image and
0:32:05 generate a video off the image saying that I much prefer to test text to video because what it should
0:32:11 be doing is you give it a text prompt. It always seems to generate videos better. If you start from
0:32:16 an image generated into a video, why isn’t it when I give it a text prompt, it generates an image that’s
0:32:20 a decent image and then use that as the starting image to generate the rest of the video. So you’d
0:32:26 think you would be getting fairly equal results, but for whatever reason, starting from an image, almost
0:32:32 always, always creates a better result. And I don’t quite understand it because I do feel like a text
0:32:38 to video should have a very similar result. Just start with an image, just do that part behind the
0:32:43 scene. So it feels like we’re starting with text, you know? Yeah. But also I think the image that comes
0:32:51 out, it is formed in a way that looks pro rather than you generating something from scratch. You know,
0:32:56 like this has the details. It doesn’t need more detail. Right. So when camera angle prompts or
0:33:01 lighting prompts or color prompts, all of that kind of stuff, you don’t necessarily have to put it to the
0:33:06 text prompt because it’s there. They have the, you know, the example already. Yeah. All right. So the
0:33:13 other model that they came out with was this new cling 01 model. And this one, you can actually feed it
0:33:21 images, video and text, any combination of image, video and texts. So some of the prompts that I gave
0:33:26 this one, just to kind of give you an example of what it was capable of and see how the prompt has
0:33:32 like references to the images that I uploaded. So I gave it an image of like this Bugatti supercar here.
0:33:36 And I gave it this image of like a closeup of a pizza slice. No, I’m hungry.
0:33:43 And then I said, animate image one, which was the car driving fast on a racetrack. The car is made
0:33:49 entirely of the material from image two, which is the cheese pizza. As the car turns corners,
0:33:55 the body should stretch and melt slightly due to heat and centrifugal force dripping cheese onto the
0:34:01 asphalt. It’s very technical, by the way. Yeah. This is what it generated. And this one does not have
0:34:07 audio, but you could see it took the cheese from the pizza made that like the paint job of the car.
0:34:13 And I was able to edit those two things together and then create a video from the sort of two blended
0:34:19 things. Now the cheese was supposed to shift around like as it went around corners, but the video sort of
0:34:25 stops right before it goes around the corner. So I don’t know how well it did the physics portion of it,
0:34:29 but it definitely took the texture of the pizza and added it to the car.
0:34:34 Yeah. It looks good though. The second prompt was I gave it this image of like, um,
0:34:40 the internals of a watch. And then I gave it a video of like a drone flying through a canyon.
0:34:45 But I said, bring this image, which was the internals of the clock to life using the camera
0:34:50 movement and speed of, and this was the drone flying through a canyon. We’re flying deep inside
0:34:55 the watch mechanism. The gears must rotate in sync with the speed of the camera movement.
0:35:00 And so the idea was it was supposed to look like a drone flying through the sort of like
0:35:05 inner workings of this watch. Didn’t quite work out super well. In fact, I’m going to unmute this one
0:35:05 because it’s kind of funny.
0:35:12 It took the audio from the drone, but none of the video from the drone.
0:35:18 Yeah. So if you’re just listening to this in audio, what that was, was just a picture of like a watch,
0:35:24 but with drone sounds over it. The other thing that I find super interesting, you probably can’t see this,
0:35:31 but down in the corner, it says Vio. I’m in cling right now. But the reason it did that was because
0:35:36 my video of a drone flying through the canyon was generated with Vio. I generated that with Vio,
0:35:40 brought it over to here, and then tried to make that drone flying through the clock.
0:35:44 And it kept the Vio watermark in the cling video.
0:35:49 Oh my God. So that’s what happened there. This was actually one of my favorite ones. I gave it
0:35:55 a video of me sitting in my office, reading a book here. And then I gave it an image that I generated,
0:36:00 I don’t know, in mid journey or something of like a futuristic transparent tablet. And I said,
0:36:06 replace the book in this video with the tablet from this image, adjust the hand grips to hold the thin
0:36:13 tablet. Naturally, the light from the tablet must illuminate the face and palms, right? And the original
0:36:18 video was me holding up a book. And this is what it generated. You can see, I’m sitting here looking
0:36:23 at this tablet now. Now it’s not lighting up my face like it’s supposed to. And when I’m done,
0:36:28 it kind of folds up like a book. It’s not so bad though. And then at the end of the video,
0:36:33 I folded the book and set it down. And I wanted to see what it would do with the tablet and the tablet
0:36:41 sort of folded. But I mean, the first five seconds of this before the folding kind of looks wonky,
0:36:45 I think it looks really good. I’m telling you, I think you can like sell this,
0:36:49 put it out like on the internet and stuff and like freak people out and like make an ad be like,
0:36:53 hey, buy the new clear tablet. Yeah. Apple came out with this new clear tablet. And like,
0:36:58 I’m giving it away for people for like, you know, that would be so cool. Do it on like April Fool’s.
0:37:02 It’s a good idea. I should try that. See if people buy it. Well, I just said it on the internet.
0:37:08 Now it’s out. Yeah. Yeah. I’ll do something similar. So if I’m promoting some sort of weird
0:37:14 tablet or gadget on April 1st, you’ll know what’s going on. All right. So I gave it a video of a man
0:37:20 on a treadmill. This is just a stock video I found on story blocks. And I said, change the aspect ratio
0:37:27 to horizontal because this original video is 916 vertical view. Replace the gym environment with
0:37:32 the surface of Mars. The person is now walking in a space suit. The treadmill is gone. They’re
0:37:36 walking on a dusty red terrain in the expanded background. Show the futuristic colony in the
0:37:42 distance. So first thing to notice, it did not change my aspect ratio. It kept it at nine by 16.
0:37:46 16 didn’t switch it to 16 by nine. It doesn’t listen. Yeah. When it comes to that, by the way,
0:37:50 I’ve tried it on several occasions, not just, but this one, I just don’t think they’re capable of it
0:37:55 yet for some reason. I don’t care. They should be in Vans. I know. It feels like all the image models
0:38:01 are capable of outpainting, right? They’re all capable of me giving it a nine by 16 vertical image,
0:38:05 and then outpainting that image to a 16 by nine. So making it a 16 by nine and using AI to figure out
0:38:10 what goes in the sides of the image, right? Yeah. If image models are capable of that,
0:38:15 why aren’t video models? Because theoretically, a video is moving pictures. It’s a whole bunch
0:38:19 of pictures moving at a rapid frame rate. I don’t understand why we can’t do it on video
0:38:23 if we can do it on image. And if we can do this stuff starting from an image and we have outpainting
0:38:28 for images, it doesn’t compute why they haven’t figured out how to get this to work on video.
0:38:32 If it can like generate the most complex prompt in the history of everything, because I’m pretty sure
0:38:39 someone out there has neurodivergence like me and like wants everything to be so compacted and like
0:38:45 given the most detailed, you know, prompt ever. So that the end result would be if it can do that
0:38:49 and generate that video, it can definitely change the aspect ratio. Don’t piss me off.
0:38:55 Well, here’s the video that it generated. So the original was somebody running on a treadmill
0:39:01 and this video, it’s the same motion of them running on the treadmill, but the person got changed.
0:39:07 You can’t really tell. I know it was small on my screen, but the original person was
0:39:13 a black guy and this person’s a white guy. So it totally changed the person in the video.
0:39:21 Cling? Should we have a talk? That’s not good. That’s not good, honey. We should fix that.
0:39:27 So it changed the face. It did put him on Mars and in a space suit, but it definitely changed the person
0:39:32 in the space suit. It didn’t change my aspect ratio. You can change a race, but you can change an aspect
0:39:40 ratio really? Fix it. Yeah. So anyway, that’s what it did. But what I was testing here was,
0:39:44 you know, giving it a starting video and seeing if it can sort of re-skin the starting video.
0:39:51 And it literally re-skinned. Bad. Bad cling. Bad cling. So my last one,
0:39:55 this one was actually kind of hilarious. So I gave it an image of myself just sort of standing there with
0:40:01 my arms out. Right. And then I said, “Performing a tap dance in the style of,” and I gave it the old
0:40:06 like rubber hose style, like animation screenshot. The character must completely adopt the rubber hose
0:40:12 body proportions, noodle limbs, black and white aesthetic, but the face must remain recognizable.
0:40:18 The face must remain recognizable. Key ingredient. Who is that? The original image was an image of me.
0:40:23 What is that? And this is the recognizable
0:40:28 face that it generated. Could be you in another life. The funny thing is, it kept my shirt perfectly.
0:40:32 The shirt that this cartoon character is wearing is the exact shirt that I’m wearing in my original
0:40:37 image. It just ignored the part about making it look like me. Why don’t you have pants is my question.
0:40:45 Where are your pants? Because the old doodle cartoons, I guess they didn’t wear pants maybe? I don’t know.
0:40:51 It took off your pants. It did take off the pants. Pretty sus. But like,
0:40:55 my question is like, where did you take that picture? Like, why do you have like photo shoots
0:40:59 and stuff? I want to have photo shoots as well. This was for like a podcast photo shoot I did like
0:41:05 10 years ago. This image is like a 10-year-old image. 10? You look the same. It’s like, drop your
0:41:11 skincare mat right now. So yeah, those were the tests that I did on Kling 01.
0:41:17 Do the public, normies like me, have access to Kling 01? Yes. In fact, I think you can generate
0:41:23 a certain amount for free per day. Oh. Let me actually just check. All right. So here we go.
0:41:30 Basic is free forever. Oh. So I’m on this plan right here, which is a little under $9 a month.
0:41:34 That’s pretty good. That’s not bad. For the stuff that we’ve done, that’s pretty good.
0:41:41 Yeah. So it says you can generate about 33 videos a month. And I mean, I’ve actually never hit my limits
0:41:45 because I’m not sitting here like generating tons and tons and tons of videos. I’m doing like 10 tests,
0:41:51 and then I’m moving on. I’m pretty sure you do more, but you don’t talk about it. But probably go to other
0:41:56 stuff and like generate more. You wouldn’t even know your limits because I know for a fact you test them all.
0:42:01 I do test them all. But some of them I do have like special access, you know, like they hooked me up.
0:42:06 Oh, yes. We don’t. And like you do. Yes. Just flaunt that around. Why don’t you?
0:42:10 So yeah, those are the new video models that have come out. They’re pretty good.
0:42:14 There was also the McDonald’s ad. We wanted to talk about that really quickly because it’s,
0:42:18 you know, somewhat related. I wanted for us to like, you know,
0:42:23 like how we did the runway thing with the meal prepping thing. And like, basically,
0:42:28 they just pulled it down when it comes to the AI ads that McDonald’s did. And everyone,
0:42:34 everyone hated it. Everyone. It’s not as wholesome as the Coca-Cola one. The Coca-Cola one was fine.
0:42:40 But the McDonald’s one was so bad. It was nightmarish. You’re using the McDonald’s cup without a straw.
0:42:46 It’s disgusting on normal days. Are you not going to use a straw? No.
0:42:48 I just think it’s funny that that was the thing that you picked up on.
0:42:53 That’s the only thing. Everything else is like, because I read a lot of like weird things and
0:42:57 everything is like very thrillery and like dark stuff. I didn’t have an issue with that. But this,
0:43:04 I had an issue. That’s so hilarious. It’s like you’ve got AI people getting pulled by like trolleys and
0:43:09 flanging out windows. And the one thing you pick up on is they didn’t use a straw.
0:43:14 You couldn’t try to write that. Anyway, that’s my take on the McDonald’s thing.
0:43:22 Yeah. Yeah. I think if I’m being honest, I don’t think the ad looks that bad. But saying that,
0:43:28 I feel like people have been so inundated with like AI video and images on social media lately.
0:43:33 Like you can’t go on Facebook and scroll like there is so much AI slop just in the Facebook feed,
0:43:38 right? Like there’s AI generated videos. And my wife gets confused all the time. Like I remember
0:43:42 during Halloween, there was all sorts of videos. She’s like, look at this awesome Halloween costume.
0:43:47 And I’m like, that’s AI, honey, right? Your wife gets confused. My 59 year old mother
0:43:53 is always sending me things, Matt, asking me if I could find a way if it’s real or not. What do you mean,
0:44:00 like, yeah, I think people are just like saturated with that stuff. People have gotten so much of that
0:44:05 in all of their feeds. And they’re just so over this, like AI slop video that like everybody is
0:44:11 putting out now. And they’re so over seeing stuff in their feeds and going, is this real or AI? I don’t
0:44:17 even know anymore. And now you have like companies like McDonald’s and Coca-Cola going and doing the same
0:44:21 thing. And people are just like, oh my God, we’re already exhausted from normal people doing this.
0:44:24 Why do we have to see this from the big freaking corporations now too?
0:44:28 Right. And I feel like that’s more of the blowback than anything is like,
0:44:33 we’re already inundated and frustrated with how much slop we’re seeing.
0:44:39 Why are you guys spending money on slop too? You guys are mega corporations with billions and
0:44:45 billions and billions of dollars. Hire VFX artists, hire actors. Like, why are you giving us slop too?
0:44:48 You know, that I think is like what a lot of people have a problem with.
0:44:53 Not even McDonald’s gonna get broke if like they hired someone to make their ads. Honestly,
0:44:57 I didn’t like the idea that they’re using AI ads. And I’m a person that advocates for AI every single
0:44:59 day. Yeah, that’s me.
0:45:05 So I think AI needs to be used as like an assistive tool, not as like the main tool. And I feel like
0:45:11 things like this ad, they just went all in and made everything AI and made it obviously AI,
0:45:17 where I think if they were to do like, let’s say 80% of the video is real people, real actors,
0:45:24 real VFX, you know, CGI, not actual AI generated stuff. But then maybe there’s like one or two scenes
0:45:30 in there that they can’t figure out how to get with real humans, or it would be cost prohibitive to do
0:45:36 it with CGI. So you do AI to finish that little like 10%. I don’t think anybody would have a problem
0:45:42 with that. People are like the big movies already doing that, right? Big movie studios are already
0:45:47 making movies still the old fashioned way. But there’s little shots here and there where AI makes
0:45:54 it a lot faster, but nobody really notices. It’s not like AI slop within a bigger movie. It’s like,
0:45:59 yeah, okay, we needed to de-age this person. So we used AI to make their face look a little younger
0:46:03 for that one scene. That’s matching up. I don’t think people have a problem with that. Yeah. People
0:46:09 have a problem with, all right, you just went and gave them 20 prompts and stitched the video together
0:46:14 and said, look, here’s our Christmas ad this year. People have an issue with that. Yeah. With no straws
0:46:21 also. But anyway, that was a nice look at the world of AI video. It started real fun talking about
0:46:24 all of the latest models that we can play with. And then we spoke about McDonald’s. Yeah.
0:46:30 Yeah. But you know, in some scenarios, AI kind of sucks in these areas. Yes, honestly. But yeah,
0:46:34 I had so much fun on this episode, honestly. Awesome. Yeah. I mean, it’s been a blast making
0:46:38 these last handful of episodes, just sort of recapping and playing around with what’s going
0:46:43 on in the AI world. I’d love feedback from the audience. If you’re listening on Spotify or watching
0:46:48 on YouTube, let us know in the comments. Spotify actually has comments now, so you can comment there as
0:46:52 well. Yes. But let us know what you think of this. Do you want us to like do more podcasts like this,
0:46:58 where we go deeper on individual tools and test about and see what they can do? We always love
0:47:02 hearing your feedback and we try to steer these episodes towards what you want to hear and learn
0:47:07 from us. So always appreciate it. And thank you so much for tuning in. If you like this episode,
0:47:12 give it a thumbs up and consider subscribing wherever you listen to your podcasts. And thanks again
0:47:25 for tuning in this week. Bye.

Get our AI Video Guide with 5+ prompts and real results: https://clickhubspot.com/rhk

Episode 89: How big of a leap is the latest generation of AI video models—and do they really live up to the hype? Matt Wolfe (https://x.com/mreflow) and Maria Gharib (https://uk.linkedin.com/in/maria-gharib-091779b9), an AI writer and newsletter creator, dive into hands-on testing and candid discussion about the brand-new Runway 4.5, Kling AI, and more.

In this episode, Matt and Maria put early-access Runway 4.5 through its paces, experiment with quirky video prompts, and compare outputs from the top AI video tools including Kling’s latest models. Is Runway 4.5 a massive leap forward, or just playing catchup with VEo and Sora? What kinds of content can creative teams actually produce with these new generative video AIs? Plus, Matt and Maria get real about the mixed reactions to AI-driven brand ads—like the recent McDonald’s spot—and discuss where this fast-evolving field is headed.

Check out The Next Wave YouTube Channel if you want to see Matt and Nathan on screen: https://lnk.to/thenextwavepd

—

Show Notes:

  • (00:00) AI Video Innovations Podcast

  • (03:53) Monkey on Roller Skates

  • (07:22) AI Prompt Success Evolution

  • (12:00) Nano Banana: Still Superior

  • (13:56) Incremental Update, Limited Impact

  • (17:32) AI Video & Image Editing

  • (20:27) Lip Sync Test Analysis

  • (25:50) Domino Effect Gone Awry

  • (27:29) Kling’s Dragon Feels Cinematic

  • (31:21) Image-Based Video Generation Preference

  • (34:02) Drone Flight Through Watch

  • (37:38) Why Can’t Video Models Work?

  • (39:20) Rubber Hose Tap Dance Fail

  • (44:30 AI as Assistive, Not Primary

  • (46:03)  Podcast Feedback Wanted

—

Mentions:

Get the guide to build your own Custom GPT: https://clickhubspot.com/tnw

—

Check Out Matt’s Stuff:

• Future Tools – https://futuretools.beehiiv.com/

• Blog – https://www.mattwolfe.com/

• YouTube- https://www.youtube.com/@mreflow

—

Check Out Nathan’s Stuff:

The Next Wave is a HubSpot Original Podcast // Brought to you by Hubspot Media // Production by Darren Clarke // Editing by Ezra Bakker Trupiano

Leave a Reply

The Next Wave - AI and The Future of TechnologyThe Next Wave – AI and The Future of Technology
Let's Evolve Together
Logo