FreeScale: Scaling 3D scenes via
Certainty-Aware Free-View Generation
Overview
FreeScale leverages reconstructed scene geometry to generate high-fidelity free-view images with accurate poses. Training feed-forward models (e.g., LVSM) with 22% additional generated views significantly boosts generalization to large camera motion (PSNR 18.75 → 21.45).
Abstract
The development of generalizable Novel View Synthesis (NVS) models is critically limited by the scarcity of large-scale training data featuring diverse and precise camera trajectories. While real-world captures are photorealistic, they are typically sparse and discrete. Conversely, synthetic data scales but suffers from a domain gap and often lacks realistic semantics. We introduce FreeScale, a novel framework that leverages the power of scene reconstruction to transform limited real-world image sequences into a scalable source of high-quality training data. Our key insight is that an imperfect reconstructed scene serves as a rich geometric proxy, but naively sampling from it amplifies artifacts. To this end, we propose a certainty-aware free-view sampling strategy identifying novel viewpoints that are both semantically meaningful and minimally affected by reconstruction errors. We demonstrate FreeScale's effectiveness by scaling up the training of feedforward NVS models, achieving a notable gain of 2.6 dB in PSNR on challenging out-of-distribution benchmarks. Furthermore, we show that the generated data can actively enhance per-scene 3D Gaussian Splatting optimization, leading to consistent improvements across multiple datasets. Our work provides a practical and powerful data generation engine to overcome a fundamental bottleneck in 3D vision.
Method: Certainty-Aware View Generation
- Certainty Grid – models geometric uncertainty from sparse inputs.
- View Graph – establishes correspondences between generated and real views.
- Active Exploration – selects high-information, low-artifact viewpoints.
- Scaling Effect – 1.22× training data expansion, +2.6 dB PSNR gain.
Demo: Free-View Rendering Results
We compare FreeScale against 3D Gaussian Splatting (3DGS) and Difix3D+ on scenes from DL3DV-10K and Nerfbusters. Each clip below uses the same scene and playback controls for side-by-side free-view comparison.
Videos keep their native portrait or landscape framing automatically, and both comparison sliders stay synchronized to the same scene.
Quantitative Results
Feed-forward models on viewpoint generalization
| Method | In-Domain (DL3DV) | MipNeRF360 | Tanks & Temples | ||||||
|---|---|---|---|---|---|---|---|---|---|
| PSNR↑ | SSIM↑ | LPIPS↓ | PSNR↑ | SSIM↑ | LPIPS↓ | PSNR↑ | SSIM↑ | LPIPS↓ | |
| Small camera motion | |||||||||
| LVSM | 22.20 | 0.680 | 0.216 | 15.84 | 0.285 | 0.583 | 13.07 | 0.336 | 0.674 |
| LVSM w/ FreeScale | 24.20 | 0.767 | 0.165 | 18.30 | 0.386 | 0.460 | 13.80 | 0.652 | 0.361 |
| Large camera motion | |||||||||
| 3DGS | 16.22 | 0.592 | 0.345 | 13.47 | 0.334 | 0.529 | 12.12 | 0.351 | 0.569 |
| LVSM | 18.75 | 0.522 | 0.352 | 13.88 | 0.293 | 0.622 | 13.89 | 0.352 | 0.650 |
| LVSM w/ FreeScale | 21.45 | 0.661 | 0.247 | 17.27 | 0.432 | 0.398 | 14.67 | 0.391 | 0.609 |
Per-scene reconstruction (3DGS enhancement)
| Method | DL3DV | Nerfbuster | Tanks & Temples | |||||||
|---|---|---|---|---|---|---|---|---|---|---|
| PSNR↑ | SSIM↑ | LPIPS↓ | Time↓ | PSNR↑ | SSIM↑ | LPIPS↓ | PSNR↑ | SSIM↑ | LPIPS↓ | |
| Nerfbusters | 17.45 | 0.606 | 0.370 | - | 17.72 | 0.647 | 0.352 | - | - | - |
| DIFIX3D+ | 17.99 | 0.601 | 0.293 | 81.40 | 18.07 | 0.642 | 0.279 | 18.59 | 0.623 | 0.317 |
| 3DGS | 19.18 | 0.714 | 0.233 | 35.19 | 18.14 | 0.643 | 0.265 | 20.37 | 0.680 | 0.253 |
| 3DGS w/ DIFIX3D | 19.12 | 0.680 | 0.211 | 39.75 | 17.69 | 0.606 | 0.264 | 19.75 | 0.630 | 0.210 |
| 3DGS w/ FreeScale | 19.57 | 0.723 | 0.219 | 37.22 | 18.40 | 0.648 | 0.258 | 20.66 | 0.685 | 0.251 |
Resources & Citation
BibTeX
@inproceedings{jiang2026freescale,
title={FreeScale: Scaling 3D scenes via Certainty-Aware Free-View Generation},
author={Jiang, Chenhan and Chen, Yu and Zhang, Qingwen and Song, Jifei and Xu, Songcen and Yeung, Dit-Yan and Deng, Jiankang},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2026}
}