We use cookies to improve your experience with our site.
Xu K, Li J, Li ZQ et al. SG-NeRF: Sparse-input generalized neural radiance fields for novel view synthesis. JOURNAL OFCOMPUTER SCIENCE AND TECHNOLOGY 39(4): 785−797 July 2024. DOI: 10.1007/s11390-024-4157-6.
Citation: Xu K, Li J, Li ZQ et al. SG-NeRF: Sparse-input generalized neural radiance fields for novel view synthesis. JOURNAL OFCOMPUTER SCIENCE AND TECHNOLOGY 39(4): 785−797 July 2024. DOI: 10.1007/s11390-024-4157-6.

SG-NeRF: Sparse-Input Generalized Neural Radiance Fields for Novel View Synthesis

  • Traditional neural radiance fields for rendering novel views require intensive input images and pre-scene optimization, which limits their practical applications. We propose a generalization method to infer scenes from input images and perform high-quality rendering without pre-scene optimization named SG-NeRF (Sparse-Input Generalized Neural Radiance Fields). Firstly, we construct an improved multi-view stereo structure based on the convolutional attention and multi-level fusion mechanism to obtain the geometric features and appearance features of the scene from the sparse input images, and then these features are aggregated by multi-head attention as the input of the neural radiance fields. This strategy of utilizing neural radiance fields to decode scene features instead of mapping positions and orientations enables our method to perform cross-scene training as well as inference, thus enabling neural radiance fields to generalize for novel view synthesis on unseen scenes. We tested the generalization ability on DTU dataset, and our PSNR (peak signal-to-noise ratio) improved by 3.14 compared with the baseline method under the same input conditions. In addition, if the scene has dense input views available, the average PSNR can be improved by 1.04 through further refinement training in a short time, and a higher quality rendering effect can be obtained.
  • loading

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return