PeRFception: Perception using Radiance Fields
Abstract
The recent progress in implicit 3D representation, i.e., Neural Radiance Fields (NeRFs), has made accurate and photorealistic 3D reconstruction possible in a differentiable manner. This new representation can effectively convey the information of hundreds of high-resolution images in one compact format and allows photorealistic synthesis of novel views. In this work, using the variant of NeRF called Plenoxels, we create the first large-scale implicit representation datasets for perception tasks, called the PeRFception, which consists of two parts that incorporate both object-centric and scene-centric scans for classification and segmentation. It shows a significant memory compression rate (96.4\%) from the original dataset, while containing both 2D and 3D information in a unified form. We construct the classification and segmentation models that directly take as input this implicit format and also propose a novel augmentation technique to avoid overfitting on backgrounds of images. The code and data are publicly available in https://postech-cvlab.github.io/PeRFception
Additional Information
Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)Attached Files
Submitted - 2208.11537.pdf
Files
Name | Size | Download all |
---|---|---|
md5:050da49f8865d7ae9a1cdd7f4ce9e4c7
|
34.3 MB | Preview Download |
Additional details
- Eprint ID
- 118542
- Resolver ID
- CaltechAUTHORS:20221221-004646749
- Created
-
2022-12-22Created from EPrint's datestamp field
- Updated
-
2023-06-02Created from EPrint's last_modified field