MaRF: Representing Mars as Neural Radiance Fields

The aim of this work is to introduce MaRF, a novel framework able to synthesize the Martian environment using several collections of images from rover cameras. The idea is to generate a 3D scene of Mars' surface to address key challenges in planetary surface exploration such as: planetary geology, simulated navigation and shape analysis. Although there exist different methods to enable a 3D reconstruction of Mars' surface, they rely on classical computer graphics techniques that incur high amounts of computational resources during the reconstruction process, and have limitations with generalizing reconstructions to unseen scenes and adapting to new images coming from rover cameras. The proposed framework solves the aforementioned limitations by exploiting Neural Radiance Fields (NeRFs), a method that synthesize complex scenes by optimizing a continuous volumetric scene function using a sparse set of images. To speed up the learning process, we replaced the sparse set of rover images with their neural graphics primitives (NGPs), a set of vectors of fixed length that are learned to preserve the information of the original images in a significantly smaller size. In the experimental section, we demonstrate the environments created from actual Mars datasets captured by Curiosity rover, Perseverance rover and Ingenuity helicopter, all of which are available on the Planetary Data System (PDS).

[1]  T. Müller,et al.  Instant neural graphics primitives with a multiresolution hash encoding , 2022, ACM Trans. Graph..

[2]  Benjamin Recht,et al.  Plenoxels: Radiance Fields without Neural Networks , 2021, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  J. Balaram,et al.  The Ingenuity Helicopter on the Perseverance Rover , 2021, Space Science Reviews.

[4]  K. Herkenhoff,et al.  The Mars 2020 Perseverance Rover Mast Camera Zoom (Mastcam-Z) Multispectral, Stereoscopic Imaging Investigation , 2021, Space Science Reviews.

[5]  Angjoo Kanazawa,et al.  pixelNeRF: Neural Radiance Fields from One or Few Images , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Jonathan T. Barron,et al.  Nerfies: Deformable Neural Radiance Fields , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[7]  Dava J. Newman,et al.  Mission enhancing capabilities for science-driven exploration extravehicular activity derived from the NASA BASALT research program , 2020 .

[8]  Kai Zhang,et al.  NeRF++: Analyzing and Improving Neural Radiance Fields , 2020, ArXiv.

[9]  Gordon Wetzstein,et al.  Implicit Neural Representations with Periodic Activation Functions , 2020, NeurIPS.

[10]  Nargess Memarsadeghi,et al.  Virtual and Augmented Reality Applications in Science and Engineering , 2020, Comput. Sci. Eng..

[11]  Pratul P. Srinivasan,et al.  NeRF , 2020, ECCV.

[12]  Joon Chung,et al.  Review of augmented reality in aerospace industry , 2019, Aircraft Engineering and Aerospace Technology.

[13]  Aidong Lu,et al.  Improving Information Sharing and Collaborative Analysis for Remote GeoSpatial Visualization Using Mixed Reality , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[14]  S. P. Abercrombie,et al.  OnSight: Multi-platform Visualization of the Surface of Mars , 2017 .

[15]  Mark T. Lemmon,et al.  The Mars Science Laboratory Curiosity rover Mastcam instruments: Preflight and in‐flight calibration, validation, and data archiving , 2017 .

[16]  Jan-Michael Frahm,et al.  Pixelwise View Selection for Unstructured Multi-View Stereo , 2016, ECCV.

[17]  Jan-Michael Frahm,et al.  Structure-from-Motion Revisited , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Yoshua Bengio,et al.  Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..

[19]  Volker Paelke,et al.  Integrating agile methods for mixed reality design space exploration , 2008, DIS '08.

[20]  Kaichang Di,et al.  CAHVOR camera model and its photogrammetric conversion for planetary applications , 2004 .

[21]  Nelson L. Max,et al.  Optical Models for Direct Volume Rendering , 1995, IEEE Trans. Vis. Comput. Graph..

[22]  A. Hayes,et al.  MASTCAM-Z GEOMETRIC CALIBRATION: AN ALTERNATIVE APPROACH BASED ON PHOTOGRAMMETRIC AND AFFINE SOLUTIONS SPECIFIC TO FILTER, FOCUS, AND ZOOM , 2020 .

[23]  G. Raj,et al.  Blur image detection using Laplacian operator and Open-CV , 2016, 2016 International Conference System Modeling & Advancement in Research Trends (SMART).

[24]  T. Blum,et al.  COMPUTER AIDED MEDICAL DIAGNOSIS AND SURGERY SYSTEM : TOWARDS AUTOMATED MEDICAL DIAGNOSIS FOR LONG TERM SPACE MISSIONS , 2011 .

[25]  Christoph Zauner,et al.  Implementation and Benchmarking of Perceptual Image Hash Functions , 2010 .

[26]  M. Kenward,et al.  An Introduction to the Bootstrap , 2007 .

[27]  Markus H. Gross,et al.  Optimized Spatial Hashing for Collision Detection of Deformable Objects , 2003, VMV.

[28]  R. P. G. Collinson,et al.  Introduction to Avionics Systems , 2003 .

[29]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[30]  Susan K. McMahon,et al.  Overview of the Planetary Data System , 1996 .

[31]  M. Vona,et al.  Multi-platform Immersive Visualization of Planetary , Asteroid , and Terrestrial Analog Terrain , 2022 .