In a web world dominated by video, it is necessary to help create it, even if we only have a reference image.
This is what Google worked on, enabling the DeepMind neural network to create short videos from a single frame.
This AI model called “Transframer” generates text based on partial prompts and is now capable of generating 30-second videos from a single frame.
The AI uses context images to guess the surroundings of the images and can thus show what the outline of a piece of furniture looks like without actually seeing it. It imagines a real object from every angle.
It’s easy to imagine how he does it. If I give a program thousands of pictures of a chair from all possible angles and then send it a photo of a chair from the front, the program can imagine the rest thanks to previous training.
The artificial depth perception and perspective is noticeable in the demonstration, which helps to envision how video games can improve with something like this, not just making videos for social media.
Transframer is a general generative framework capable of handling many image and video tasks in a probabilistic environment. New work shows it excels in video prediction and view synthesis and can generate 30s videos from a single frame: https://t.co/wX3nrrYEEa1/ pic.twitter.com/gQk6f9nZyg
— DeepMind (@DeepMind) August 15, 2022
If we can create realistic images and videos from them with DALL-E, we are one step away from automatically creating videos from a single text message.
Similar Posts:
- Adjust the image-generating AI to make it compatible with 3D animations
- Meta’s new AI creates images from text and sketches
- Do you think like an artificial intelligence? this game tells you
- Meta’s new AI produces videos from simple sentences
- A Balrog playing chess, an example of what DALL-E mini can create for free