面向挑战性重建场景的相机参数与3D高斯联合优化
Joint Optimization of Camera Parameters and 3D Gaussians for Challenging Reconstruction Scenarios
DOI: 10.12677/mos.2026.153040, PDF,   
作者: 贾贤志:重庆三峡学院计算机科学与工程学院,重庆
关键词: 3D高斯泼溅相机优化新视角合成3D Gaussian Splatting Camera Optimization Novel View Synthesis
摘要: 3D高斯泼溅(3D Gaussian Splatting, 3DGS)凭借其高效的实时渲染能力与快速收敛特性,已成为新视角合成领域的重要技术路线。然而,其重建质量高度依赖于精确的相机内参与外参标定,这些参数传统上依赖运动恢复结构(SfM)流程(如COLMAP)预先估计。值得注意的是,在低纹理、重复纹理或视角受限等挑战性场景中,SfM估计往往产生误差,限制了重建精度。为此,本文提出一种在训练阶段同步精细化相机参数与3D高斯表示的联合优化框架。具体而言,我们通过偏移参数化策略增强优化稳定性,采用对数尺度焦距表示改善数值条件,并引入余弦衰减自适应正则化项来优化相机参数,提升建模质量。
Abstract: 3D Gaussian Splatting (3DGS) has emerged as a prominent technique in novel view synthesis, distinguished by its efficient real-time rendering capabilities and rapid convergence properties. However, its reconstruction quality heavily relies on accurate estimation of camera intrinsics and extrinsics, which are traditionally obtained through Structure-from-Motion (SfM) pipelines such as COLMAP. Notably, in challenging scenarios characterized by low-texture regions, repetitive patterns, or limited viewpoint distributions, SfM estimations often suffer from inaccuracies, thereby compromising reconstruction fidelity. To address this limitation, we propose a joint optimization framework that simultaneously refines camera parameters and 3D Gaussian representations during the training stage. Specifically, we employ offset parameterization to stabilize optimization, adopt logarithmic focal length representation for improved numerical conditioning, and introduce a cosine annealing adaptive regularization strategy to refine camera parameters and enhance modeling quality.
文章引用:贾贤志. 面向挑战性重建场景的相机参数与3D高斯联合优化 [J]. 建模与仿真, 2026, 15(3): 23-30. https://doi.org/10.12677/mos.2026.153040

参考文献

[1] Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R. and Ng, R. (2020) NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis. In: Vedaldi, A., Bischof, H., Brox, T. and Frahm, JM., Eds., Lecture Notes in Computer Science, Springer International Publishing, 405-421. [Google Scholar] [CrossRef
[2] Kerbl, B., Kopanas, G., Leimkuehler, T. and Drettakis, G. (2023) 3D Gaussian Splatting for Real-Time Radiance Field Rendering. ACM Transactions on Graphics, 42, 1-14. [Google Scholar] [CrossRef
[3] Schonberger, J.L. and Frahm, J. (2016) Structure-from-Motion Revisited. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, 27-30 June 2016, 4104-4113. [Google Scholar] [CrossRef
[4] Yu, Z., Chen, A., Huang, B., Sattler, T. and Geiger, A. (2024) Mip-Splatting: Alias-Free 3D Gaussian Splatting. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, 17-21 June 2024, 19447-19456. [Google Scholar] [CrossRef
[5] Liu, S., Yang, M., Xing, T. and Yang, R. (2025) A Survey of 3D Reconstruction: The Evolution from Multi-View Geometry to NeRF and 3DGS. Sensors, 25, Article 5748. [Google Scholar] [CrossRef
[6] Luiten, J., Kopanas, G., Leibe, B. and Ramanan, D. (2024) Dynamic 3D Gaussians: Tracking by Persistent Dynamic View Synthesis. 2024 International Conference on 3D Vision (3DV), Davos, 18-21 March 2024, 800-809. [Google Scholar] [CrossRef
[7] Lin, C.H., Ma, W.C., Torralba, A. and Lucey, S. (2021) BARF: Bundle-Adjusting Neural Radiance Fields. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, 11-17 October 2021, 5741-5751. [Google Scholar] [CrossRef
[8] Ku, K., Jubery, T.Z., Rodriguez, E., Balu, A., Sarkar, S., Krishnamurthy, A. and Ganapathysubramanian, B. (2025) SC-NeRF: NeRF-Based Point Cloud Reconstruction using a Stationary Camera for Agricultural Applications. 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, 11-15 June 2025, 5472-5481. [Google Scholar] [CrossRef
[9] Triggs, B., McLauchlan, P.F., Hartley, R.I. and Fitzgibbon, A.W. (2000) Bundle Adjustment—A Modern Synthesis. In: Triggs, B., Zisserman, A. and Szeliski, R., Eds., Lecture Notes in Computer Science, Springer Berlin Heidelberg, 298-372. [Google Scholar] [CrossRef
[10] Wei, Y., Liu, S., Rao, Y., Zhao, W., Lu, J. and Zhou, J. (2021) NerfingMVS: Guided Optimization of Neural Radiance Fields for Indoor Multi-View Stereo. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, 11-17 October 2021, 5610-5619. [Google Scholar] [CrossRef
[11] Park, K., Henzler, P., Mildenhall, B., Barron, J.T. and Martin-Brualla, R. (2023) CamP: Camera Preconditioning for Neural Radiance Fields. ACM Transactions on Graphics (TOG), 42, 1-11. [Google Scholar] [CrossRef