NVIDIA UniRelight: AI Video Relighting for Filmmakers
Share this post:
NVIDIA UniRelight: AI Video Relighting for Filmmakers
NVIDIA's Toronto AI Lab published UniRelight, a system that changes lighting conditions in existing videos while preserving scene geometry, material properties, and temporal consistency. The work was accepted at NeurIPS 2025 and builds on NVIDIA's Cosmos World Foundation Models. The research is led by Kai He and Ruofan Liang at NVIDIA, the University of Toronto, and the Vector Institute.
UniRelight, NVIDIA Toronto AI Lab
Why Relighting Existing Footage Is Difficult
Standard relighting pipelines treat albedo estimation and lighting synthesis as separate steps. The disconnect produces artifacts where shadows, specular highlights, and material responses in one frame disagree with the next, breaking temporal coherence across the clip.
Failure modes of isolated relighting approaches
Complex surfaces compound the problem. Glass, transparent objects, and anisotropic materials produce lighting responses that conventional decomposition methods cannot model accurately, because the physical assumptions built into most approaches break down at those surface types.
Joint Decomposition and Synthesis
UniRelight addresses the separation problem by predicting the relit video output and an albedo map simultaneously in a single forward pass. Each prediction constrains the other, giving the model a structural prior that single task approaches lack.
Joint estimation: simultaneous albedo and relit output prediction
The architecture uses a finetuned DiT video diffusion model with a VAE encoder and decoder for the latent space. Target lighting is supplied as environment maps (HDRI), whose features are concatenated with video latents at inference time. Training combines synthetic multi illumination datasets with automatically labeled real world video footage, covering the range of materials and conditions the model encounters in practice.
What the Model Produces
UniRelight handles specular highlights on metallic and polished surfaces, glass and transparent objects, and anisotropic materials that respond differently based on viewing and lighting angles. All outputs remain consistent from frame to frame throughout the video sequence.
Dramatic lighting changes are within scope: switching from daylight to nighttime conditions, hard directional sources, and colored environments. The authors report that UniRelight outperforms DiffusionRenderer on standard benchmarks for intrinsic decomposition and relighting quality. For commercial video generation workflows without research restrictions, AI FILMS Studio provides a licensed suite of tools.
Gallery
Availability and License
UniRelight's code is on GitHub at nv-tlabs/UniRelight and model weights are published on Hugging Face. The project is released under the NVIDIA OneWay Noncommercial License. Commercial deployment is not permitted. Use is restricted to research and academic purposes.
This release follows the same noncommercial model as other NVIDIA NVLabs projects. NVIDIA LongLive, the lab's real time interactive video generation system, and NVIDIA SANA Video, an efficient long form text-to-video model, both ship under the same terms. NVIDIA Lyra 2.0, from the Spatial Intelligence Lab, takes the opposite approach with an Apache 2.0 license that permits commercial deployment.
Sources
arXiv: UniRelight: Learning Joint Decomposition and Synthesis for Video Relighting
GitHub: nv-tlabs/UniRelight
Project Page: research.nvidia.com/labs/toronto-ai/UniRelight
Continue Reading
Video & LipSync
- Video Generator
- Text to Video
- Image to Video
- Start-End Frame to Video
- Draw to Video
- Motion Control
- Video Enhancer
- Video Upscaler
- Video to Video LipSync
- Audio to Video LipSync
- Image to Video LipSync
- Video FaceSwap
- Seedance 2
- OpenAI Sora 2
- Kling 3.0
- Kling O1
- Google Veo 3.1
- LTX 2.3
- Kling O1
- Hailuo AI
- Luma Ray
- Kling 3.0 Motion
- Topaz Upscaler
- InfiniteTalk Face Swap