AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
Article Link
Collect
Submit Manuscript
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Regular Paper

SinGRAV: Learning a Generative Radiance Volume from a Single Natural Scene

School of Computer Science and Technology, Shandong University, Qingdao 266237, China
State Key Laboratory of General Artificial Intelligence, Beijing 100871, China
School of Intelligence Science and Technology, Peking University, Beijing 100871, China
Tencent AI Lab, Tencent Holdings Limited, Shenzhen 518057, China
Show Author Information

Abstract

We present SinGRAV, an attempt to learn a generative radiance volume from multi-view observations of a single natural scene, in stark contrast to existing category-level 3D generative models that learn from images of many object-centric scenes. Inspired by SinGAN, we also learn the internal distribution of the input scene, which necessitates our key designs w.r.t. the scene representation and network architecture. Unlike popular multi-layer perceptrons (MLP)-based architectures, we particularly employ convolutional generators and discriminators, which inherently possess spatial locality bias, to operate over voxelized volumes for learning the internal distribution over a plethora of overlapping regions. On the other hand, localizing the adversarial generators and discriminators over confined areas with limited receptive fields easily leads to highly implausible geometric structures in the spatial. Our remedy is to use spatial inductive bias and joint discrimination on geometric clues in the form of 2D depth maps. This strategy is effective in improving spatial arrangement while incurring negligible additional computational cost. Experimental results demonstrate the ability of SinGRAV in generating plausible and diverse variations from a single scene, the merits of SinGRAV over state-of-the-art generative neural scene models, and the versatility of SinGRAV by its use in a variety of applications. Code and data will be released to facilitate further research.

Electronic Supplementary Material

Download File(s)
JCST-2307-13596-Highlights.pdf (270.9 KB)
Journal of Computer Science and Technology
Pages 305-319
Cite this article:
Wang Y-J, Chen X-L, Chen B-Q. SinGRAV: Learning a Generative Radiance Volume from a Single Natural Scene. Journal of Computer Science and Technology, 2024, 39(2): 305-319. https://doi.org/10.1007/s11390-023-3596-9

48

Views

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Altmetrics

Received: 14 July 2023
Accepted: 12 January 2024
Published: 30 March 2024
© Institute of Computing Technology, Chinese Academy of Sciences 2024
Return