Recent approaches on 3D-informed camera control in video diffusion models (VDMs) often create anchor videos to guide diffusion models as a structured prior by rendering from estimated point clouds following annotated camera trajectories. However, errors inherent in point cloud estimation often lead to inaccurate anchor videos. Moreover, the requirement for extensive camera trajectory annotations further increases resource demands. To address these limitations, we introduce EPiC, an efficient and precise camera control learning framework that automatically constructs high-quality anchor videos without expensive camera trajectory annotations. Concretely, we create highly precise anchor videos for training by masking source videos based on first-frame visibility. This approach ensures high alignment, eliminates the need for camera trajectory annotations, and thus can be readily applied to any in-the-wild video to generate image-to-video (I2V) training pairs. Furthermore, we introduce Anchor-ControlNet, a lightweight conditioning module that integrates anchor video guidance in visible regions to pretrained VDMs, with less than 1% of backbone model parameters. By combining the proposed anchor video data and ControlNet module, EPiC achieves efficient training with substantially fewer parameters, training steps, and less data, without requiring modifications to the diffusion model backbone typically needed to mitigate rendering misalignments. Although being trained on masking-based anchor videos, our method generalizes robustly to anchor videos made with point clouds during inference, enabling precise 3D-informed camera control. EPiC achieves SOTA performance on RealEstate10K and MiraData for I2V camera control task, demonstrating precise and robust camera control ability both quantitatively and qualitatively. Notably, EPiC also exhibits strong zero-shot generalization to video-to-video scenarios.
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
Source
Video
Generated
Video
(a) Due to errors in point cloud estimation, rendered anchor videos from reconstructed point clouds often exhibit misaligned regions compared to the source video, and also require access to the source video’s camera trajectory for rendering. (b) We create anchor videos based on first-frame visibility, which ensures better alignment with the source video and eliminates the need for camera trajectory annotations. (c) Our full construction pipeline ends with the addition of dashed rays to simulate the flying-pixel artifacts commonly seen in point-cloud-rendered anchor videos.
Source
Video
Masked
Source
Video
Anchor
Video
(a) shows an overview of our EPiC framework. EPiC supports multiple inference scenarios. (b) and (c) illustrate our I2V inference scenarios using full and masked point clouds, respectively. The masked variant is designed to enable more dynamic video generation by selectively masking the control signals from potentially moving objects in the point clouds and the resulting anchor videos. (d) depicts the V2V inference scenario employing dynamic point clouds.
Source
Image
Anchor Videos
rendered
from Masked
Point Clouds
Generated
Videos with
mode (c)
Source
Image
Anchor Videos
rendered
from Full
Point Clouds
Generated
Videos with
mode (b)
Source
Image
Anchor Videos
rendered
from Masked
Point Clouds
Generated
Videos with
mode (c)
Source
Image
Anchor Videos
rendered
from Full
Point Clouds
Generated
Videos with
mode (b)
@article{zun2025epic,
author = {Zun Wang and Jaemin Cho and Jialu Li and Han Lin and Jaehong Yoon and Yue Zhang and Mohit Bansal},
title = {
EPiC: Efficient Video Camera Control Learning with Precise Anchor-Video},
journal = {arxiv},
year = {2025},
url = {http://arxiv.org/abs/2505.21876}
}