MomentsNeRF:
Incorporating Orthogonal Moments in Convolutional Neural Networks for One or Few-Shot Neural Rendering

a sample
a sample
a sample
PixelNeRF
PSNR=25.624 SSIM=0.792 LPIPS=0.403 DISTS=0.235
Ours
PSNR=28.501 SSIM=0.83 LPIPS=0.336 DISTS=0.175
Reference

Qualitative comparison on DTU dataset. We show novel views rendered by PixelNeRF and MomentsNeRF compared to the reference in 3 views settings for scan114 scene.

Abstract

We propose MomentsNeRF, one and a few-shot learning framework that predicts a neural representation of a scene using Orthogonal Moments. Our architecture offers a transfer learning method to pre-train on multi-scenes and incorporate a per-scene optimisation at test time using one or a few images. Our method leverages transfer learning which learns from features extracted from orthogonal Gabor and Zernike moments. Our approach shows a better performance in synthesising the scene details in terms of complex texture and shape capturing, noise reduction, artefact elimination, and completing the missing parts compared to the recent one and a few-shot neural rendering frameworks. We conduct extensive experiments on real scenes from the DTU dataset. MomentsNeRF improves the existing approaches by 3.39 dB PSNR, 11.1% SSIM, 17.9% LPIPS, and 8.3% DISTS metrics. Moreover, MomentsNeRF achieves state of the art performance for both, novel view synthesis and single-image 3D view reconstruction.

proposed method

Our proposed framework diagram, which outlines the entire workflow from a set of videos to any NeRF-like application. The data representation used in the diagram consists of four input videos that were taken from the Nutrition5k dataset

Rendering

PixelNeRF
Ours
Scan21/1v
PixelNeRF
Ours
Scan55/3v
PixelNeRF
Ours
Scan110/6v
PixelNeRF
Ours
Scan114/9v

Rendered images from PixelNeRF and our model.

Citation

Acknowledgements

The website template was borrowed from here.