Blind Augmentation: Calibration-free Camera Distortion Model Estimation for Real-time Mixed-reality Consistency

To appear in IEEE Transactions on Visualization and Computer Graphics (2025)

1 2 3

Abstract

Real camera footage is subject to noise, motion blur (MB) and depth of field (DoF). In some applications these might be considered distortions to be removed, but in others it is important to model them because it would be ineffective, or interfere with an aesthetic choice, to simply remove them. In augmented reality applications where virtual content is composed into a live video feed, we can model noise, MB and DoF to make the virtual content visually consistent with the video. Existing methods for this typically suffer two main limitations. First, they require a camera calibration step to relate a known calibration target to the specific cameras response. Second, existing work require methods that can be (differentiably) tuned to the calibration, such as slow and specialized neural networks. We propose a method which estimates parameters for noise, MB and DoF instantly, which allows using off-the-shelf real-time simulation methods from e.g., a game engine in compositing augmented content. Our main idea is to unlock both features by showing how to use modern computer vision methods that can remove noise, MB and DoF from the video stream, essentially providing self-calibration. This allows to auto-tune any black-box real-time noise+MB+DoF method to deliver fast and high-fidelity augmentation consistency.

Summary

Teaser

Our method “blindly” estimates a model of noise, motion blur (MB) and depth of field (DoF) from input frames (left), i.e., without requiring any known calibration markers / objects. The model can then augment other images with virtual objects that appear visually consistent (middle).

Supplementary Video

BibTeX

@Article{prakash2025blind,
  author       = "Prakash, Siddhant and Walton, David R. and dos Anjos, Rafael K. and Steed, Anthony and Ritschel, Tobias",
  title        = "Blind Augmentation: Calibration-free Camera Distortion Model Estimation for Real-time Mixed-reality Consistency",
  journal      = "To appear in IEEE Transactions on Visualization and Computer Graphics (2025)",
  year         = "2025",
	  
 

Acknowledgments and Funding

The authors wish to thank Sebastian Friston and David Swapp for early discussions, and David Mandl for providing the code for Neural Cameras [MandlEtAl 2021]. This work was funded by the EPSRC/UKRI project EP/T01346X/1.