Join My Newsletter
[sibwp_form id=2]
Creating consistent AI video can be tricky, but with the right methods, you can achieve stunning results.
In this blog post, I’ll dive into Tokyojab’s advanced techniques for ensuring temporal consistency in AI videos. We’ll explore his latest hacks to help you transform your video with stable, flicker-free results. Whether you’re new to AI filmmaking or an experienced creator, these tips will help you take your video production to the next level.
Tokyojab, a highly innovative AI video creator, has developed several hacks that ensure consistency across frames in AI-generated videos. Temporal consistency ensures that elements, like facial features, stay stable across frames, preventing flickering and jittering.
In Part 1, I introduced two clever hacks from Tokyojab that significantly improve AI video quality. In this follow-up, we’ll dive deeper into these methods, share additional insights, and explore new tools like Tiled VAE and ControlNet for more detailed and longer video projects.
The first step in creating a consistent AI video is installing the Tiled VAE, which helps distribute the GPU workload across frames. This extension is a part of the Multi-Diffusion extension, and here’s how to install it:
These tools are essential for anyone working with smaller GPUs, as they help make the process smoother.
To make longer, 9×16 resolution videos, you’ll need to create a grid of 9 images (or more, depending on your project). Here’s how:
This technique allows for more flexibility when editing longer videos while maintaining temporal consistency.
After creating the grid, use the Depth Map extension to enhance your results:
This process ensures that not only the foreground is consistent, but the background stays stable as well, avoiding any unwanted flicker.
For even more control over your AI video, use multiple ControlNet units. Here’s how to set it up:
This method gives you the flexibility to manage complex movements and facial expressions with greater precision.
When generating video content with Stable Diffusion, creating precise prompts is key. For this tutorial, I used the new Deliberate Model and combined it with a LoRa LoRa model to enhance details. Here are some helpful prompt tips:
By following these prompt techniques, you can ensure more consistent facial features and better overall detail.
One common problem with AI-generated videos is the inconsistency of facial features. I encountered this issue with teeth in my project, and after trying different techniques (like adjusting CFG scales and sampling steps), I found a workaround:
This allows you to keep critical details, like the mouth or teeth, consistent throughout the video.
By following these advanced techniques, you’ll be able to create consistent AI videos with stable backgrounds and flicker-free results. Whether you’re working with smaller GPUs or high-end hardware, these methods will help you produce professional-looking videos.
If you found this tutorial helpful, consider subscribing to my YouTube channel for more AI filmmaking tutorials. Let me know in the comments if you have any questions or thoughts!
