Skip to content

Instantly share code, notes, and snippets.

ALICE_Scitech_extended_abstract_gen_output

In the realm of autonomous space operations, the work presented by James Ecker, Benjamin Kelley, and Danette Allen at AIAA SciTech 2021 addresses a critical component: the creation and utilization of synthetic data for 3D mesh prediction and spatial reasoning in multi-agent robotic missions. The context for this research is the challenging environment of in-space assembly where conventional computer vision systems encounter unique difficulties such as variable illumination, complex angles, orientation, and movement, alongside the constraints of energy usage and mass.

The presentation delves into the intricate balance required in space robotics: the need for extensive sensor data to manage the high variability in the operational environment, which paradoxically increases the energy and mass burden of the spacecraft. To mitigate these challenges, the team emphasizes the need to maximize information extraction while minimizing the additional resource expenditure. A notable innovation presented is the prediction of 3D meshes from a single camera view, reducing the sensor load.

The research builds upon foundational work in computer vision, notably Mask R-CNN and Mesh R-CNN, which provide frameworks for object detection and 3D reconstruction. These models benefit from transfer learning, which allows for the initialization of network weights from pre-trained models, significantly reducing both training time and generalization error. The presentation also discusses the performance of Mask R-CNN, citing impressive instance segmentation accuracy rates.

Advancing further, the team outlines their process for synthesizing data, utilizing tools like Blender and simulation environments such as ROS, Gazebo, and Mujoco. This process accounts for varying observational conditions, such as camera and light source orientation, and the complexity of the scene. The technique of domain randomization is highlighted as a means to bridge the sim-to-reality gap, ensuring that the synthetic data covers a broad spectrum of real-world scenarios.

The dataset creation is a rigorous process, generating a large pool of image and metadata pairs, from which training and validation sets are sampled. This metadata is crucial for training the Mask R-CNN and Mesh R-CNN models, enabling them to understand and predict the complex spatial arrangements found in space operations.

In terms of model architecture, the presentation notes the use of Resnet-50-FPN as the backbone for the Mask R-CNN, which uses a region proposal network and mask prediction to effectively identify and delineate objects within images.

Looking at the results, the team evaluates their models using metrics such as the Chamfer distance and the F1 score, comparing their custom synthetic data approach to existing methods. They acknowledge the necessity for hyperparameter tuning specific to hardware configurations, which included powerful GPUs like the Quadro 6000 RTX and Tesla V100.

The conclusion of the presentation points to the success in generating a synthetic dataset capable of training advanced 2D mask and 3D mesh prediction models. Future work is set to focus on further improving these models through hyperparameter tuning, enhanced domain randomization, and the exploration of Generative Adversarial Networks for point cloud generation, aiming for higher resolution 3D mesh prediction.

This extended abstract encapsulates the efforts of the team in pushing the boundaries of synthetic data generation for the advancement of autonomous robotic missions in space, highlighting both the achievements and the roadmap for future developments in this cutting-edge field.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment