Dynamic Mixed-Reality Compositing with Unity

Joanna Tarko, Christian Richardt, Peter Hall

Research output: Contribution to conferencePaper


We present a system for dynamic mixed-reality compositing, or how to insert dynamic computer-generated (CG) elements into live-action video footage in real time. The goal of compositing is to combine visual content from different sources, such as live-action footage, still images and animations, in a way that they match each other regarding colour, lighting, scale, perspective, camera movement and timing. Most of these aspects can be matched using geometric calibration of the camera and mixed- reality rendering techniques. To ensure that both sources of visual content are composited seamlessly, our approach combines the accuracy of off-line camera tracking with real-time mixed-reality rendering performed in the Unity game engine.
Original languageEnglish
Number of pages1
Publication statusPublished - 11 Dec 2017
EventCVMP 2017: The European Conference on Visual Media Production -
Duration: 11 Dec 201712 Dec 2017


ConferenceCVMP 2017: The European Conference on Visual Media Production


Dive into the research topics of 'Dynamic Mixed-Reality Compositing with Unity'. Together they form a unique fingerprint.

Cite this