
Advancing Colour Fidelity in Virtual Production
Research Team
Spencer Idenouye
Kevin Santos
James Rowan
Partners
Dark Slope Studios
NRC IRAP
CTO
Impact
- Cluster rendering optimization for V.P. in UE5.
- Latency reduction in in LED volume systems
- Color pipeline and calibration workflow for V.P
Developing a Managed Colour Workflow and Engine Distribution System for In-Camera VFX
This research project focused on advancing virtual production (V.P) workflows, specifically addressing the critical need for consistent colour representation between LED wall displays and camera capture.
Recognizing significant discrepancies like colour shifts and desaturation, the team developed a comprehensive, three-step managed colour workflow. This involved establishing a consistent baseline colour space across all hardware and software, implementing a custom camera-specific calibration using OpenVPCal to generate OCIO profiles, and utilizing Unreal Engine’s tools for technical colour correction. Beyond colour, the project also improved cluster rendering configurations and significantly reduced LED volume latency, enhancing real-time interactivity and synchronization.
The outcomes include greater visual consistency, seamless integration of virtual and physical elements, and a repeatable, production-ready calibration workflow suitable for high-end in-camera VFX projects.
The Challenge of Visual Consistency in Virtual Production
In virtual production environments, a significant challenge arises when the image displayed on an LED wall differs noticeably from what the camera captures. This was observed as a clear differential, where reds shifted towards pink, and there was desaturation and diminished contrast in the captured image compared to the LED wall display. To ensure visual consistency and fidelity, especially when shooting with cameras like the RED Komodo X in a Rec. 709 environment and displaying on advanced monitors, a colour managed workflow is essential.
The objective of this research was to establish such a workflow within the Darkslope ICVFX (In-Camera Visual Effects) environment using Unreal Engine 5.5, aiming to make images “look right”. This project recognized the need to address these colour discrepancies and workflow complexities, especially considering that desktop monitors used by operators might not be calibrated, leading to further visual mismatches with the LED wall. For larger projects or those requiring high precision, a dedicated and robust colour management solution becomes imperative, even if it demands significant time and resource.
A Reproducible, Scalable Pipeline for ICVFX Productions
The new refined workflow provides a scalable framework for managing colour in real-time virtual production environments. Key outcomes include:
- A camera-specific OCIO configuration system for accurate LED colour reproduction.
- Real-time tuning using Unreal’s nDisplay Colour Grading Panel to align LED content with physical sets.
- Source-controlled Unreal Engine distribution for consistent engine/plugin states across teams.
This technical infrastructure supports both visual accuracy and production scalability, offering measurable improvements in latency, synchronization, and visual blending between virtual and physical assets. The system is well-suited for high-end productions requiring precise visual fidelity and is adaptable across various LED panel specifications and camera systems.