Hi, this is Wayne again with a topic “Project Veo: Google Reveals New AI Video Generator”.
Today, I’m so excited to introduce imagine 3, it’s our most capable image generation model. Yet imagine 3 is more photorealistic. You can literally count the whiskers on its snout with richer details like this incredible sunlight in the shot and fewer visual, artifacts or distorted images. It understands prompts written the way people write the more creative and detailed. You are the better and imagine three remember to incorporate small details like the wild flowers or small blue bird. In this longer prompt plus.
This is our best model. Yet for rendering text, which has been a challenge for image generation models in side-by-side comparisons, independent evaluators preferred imagine three over other popular image generation models in some imagine 3 is our highest quality image generation model. So far you can sign up today to try. Imagine 3 in image FX part of our suite of AI tools at labs., gooogle and it’ll be coming soon to developers and Enterprise customers in vertex AI. Another area full of creative possibility is generative music. I’Ve been working in this space for over 20 years, and this is by far the most exciting year of my career, we’re exploring ways of working with artists to expand their creativity with AI, together with YouTube. We’Ve been building music, AI sandbox, a suite of profession, profal, music, AI tools that can create new instrumental sections from scratch, transfer Styles between tracks and more to help us design and test them.
We’Ve been working closely with Incredible musicians, songwriters and producers. Some of them made even entirely new songs in ways that would have not been possible without these tools. Let’S hear from some of the artists, we’ve been working: [ Applause ] with I’m going to put this right back into the music AI.
To the same, what happens if Haiti meets Brazil dude? I have no clue what’s about to be sprad out. This is what excites me as a hip-hop producer. We dug in the crates we playing these vinyls and the part where there’s no vocal, we pull it.
We sample it and we create an entire song around that so right now we digging in the infinite crate is endless where I found the AI really useful for me this way to like fill in the sparer sort of elements of my Loops. Okay, let’s try Bongos we’re going to put Viola we’re going to put rhythmic clapping and we’re going to see what happens there wo and it makes it sound ironically, at the end of the day, a little more human. So then, this is entirely Google’s Loops right here. These are gloops, it’s like having like this weird friend. That’S just like try this try that and then you’re like. Oh okay, yeah.
No, that’s pretty dope. The tools are capable of speeding up the process of. What’S in my head, getting it out you’re able to move light speed with your creativity. This is amazing that right there, I think this really shows what’s possible.
When we work with the artist community on the future of music, you can find some brand new songs from these acclaimed artists and songwriters on their YouTube channels. Now, there’s one more era, I’m really excited to share with you. Our teams have made some incredible progress in genitive video today, I’m excited to announce our newest, most capable genitive video model called vo. Vo creates high quality 1080p videos from text image and video prompts it can capture the details of your instructions in different, Visual and cinematic Styles. You can prompt for things like aerial shots of a landscape or time lapse and further edit, your videos using additional, prompts you can use vo in our new experimental tool.
Called video effects. We’Re exploring features like storyboarding and generating longer scenes vo gives you unprecedented. Creative control techniques for generating static images have come a long way, but generating video is a different challenge altogether.
Not only is it important to understand where an object or subject should be in space. It needs to maintain this consistency over time. Just like the car in this article VI builds upon years of our pioneer ing, generative video model work, including gqn faki, Walt video poet, Lumiere and much more. We combined the best of these architectures and techniques to improve consistency, quality and output resolution to see what vo can do. We put it in the hands of an amazing filmmaker. Let’S take a look. Well, I’ve been interested in AI for a couple years. Now we got in contact with some of the people at Google and they had been working on something of their own, so we’re all meeting here at gilgal farms to make a short film.
The Core technology is Google deep mind’s, generative video model that has been trained to convert input text into output, video. It looks good. We are able to bring ideas to life that were otherwise not possible. We can visualize things on a time scale, that’s 10 or 100 times faster than before, when you’re shooting you can’t really iterate as much as you wish, and so we we’ve been hearing that feedback is that it allows for like more more optionality, more iteration, more improvisation, But that’s what’s cool about it is like you can make a mistake faster.
That’S all you really want at the end of the day, at least in art. It’S just to make mistakes fast, so using Gemini’s, multimodal capabilities to optimize the model training process. Vo is able to better capture the Nuance from promps, so this includes cinematic techniques and visual effects, giving you total creative control everybody’s going to become a director, and everybody should be a director cuz at the heart of all of this is just storytelling the closer we Are to being able to tell each other our stories the more we’ll understand each other. These models are really enabling us to be more creative and to share that creativity with each other over the coming weeks. Some of these features will be available to select creators through video effects at labs. gooogle and the weit list is open.
Now, of course, these advances and generative video go beyond the beautiful visuals you’ve seen today by teaching future AI models, how to solve problems creatively or in effect, simulate the physics of our world. We can build more useful systems that can help people commun, communicate in new ways and thereby Advance the frontiers of AI. When we first began this journey to build AI more than 15 years ago, we knew that one day it would change everything now that time is here, and we continue to be amazed by the progress we see and inspired by the advances still to come on the Path to AGI .