In this video I take a look at the cutting-edge realm of AI Text to Video as OpenAI introduces the extraordinary Sora!
Just unveiled a few days ago, this groundbreaking technology has already wowed audiences with its incredible capabilities.
While Sora isn't yet available to the public, OpenAI has granted access to a select group of beta testers, ushering in an exclusive preview of this groundbreaking innovation. Join me on this journey to witness the future of AI in action!
Sora is OpenAI's text-to-video generative AI model. That means you write a text prompt, and it creates a video that matches the description of the prompt. Here's an example from the OpenAI site:
Like text-to-image generative AI models such as DALL·E 3, StableDiffusion, and Midjourney, Sora is a diffusion model. That means that it starts with each frame of the video consisting of static noise, and uses machine learning to gradually transform the images into something resembling the description in the prompt. Sora videos can be up to 60 seconds long.
One area of innovation in Sora is that it considers several video frames at once, which solves the problem of keeping objects consistent when they move in and out of view. In the following video, notice that the kangaroo's hand moves out of the shot several times, and when it returns, the hand looks the same as before.
🚀✨ #OpenAISora #openai #AI
Ещё видео!