A year ago we watched the rise of miraculous text-to-image generators with wide-open eyes. Fast forward to now: they are not hot news anymore. For the last few months, the tech race has been more about AI video generators, and naturally, Google couldn’t miss out. Last week, company researchers presented their attempt at synthesizing full-frame-rate clips. Their video diffusion model, named after the fathers of cinema – the Lumiere brothers – shows surprisingly consistent results and realistic, coherent motion. Unfortunately, it’s not available for public testing yet. What kind of future in AI video does Google Lumiere depict? Let’s take a closer look! Curiously, Google already played around with a couple of deep-learning models for generating videos: Phenaki and Imagen Video (we wrote about them here). Somehow, neither of them saw the world, remaining in research papers rather than becoming real-life tools. Does the same fate await Google Lumiere? Time will tell, but we hope not, as some of the features presented seem very useful for independent filmmakers. Introducing Google Lumiere So, what can the new diffusion model, fresh out of their AI research center, do? According to the official paper and the announcements on social media, everything that other contemporary AI video generators offer. That includes creating 5-second clips based on text description, video stylization, animating still images, or bringing motion only to the selected parts of the pictures. Introducing Lumiere ????️The new video diffusion model we've been working on @GoogleAI* Text-to-Video* Image-to-Video* Stylized Generation* Inpainting* Cinemagraphsand more ????W/...
Published By: CineD - Friday, 2 February, 2024