Robotic grasping in scenes with transparent and specular objects presents great challenges for methods relying on accurate depth information. In this paper, we introduce NeuGrasp, a neural surface reconstruction method that leverages background priors for material-agnostic grasp detection. NeuGrasp integrates transformers and global prior volumes to aggregate multi-view features with spatial encoding, enabling robust surface reconstruction in narrow and sparse viewing conditions. By focusing on foreground objects through residual feature enhancement and refining spatial perception with an occupancy-prior volume, NeuGrasp excels in handling objects with transparent and specular surfaces. Extensive experiments in both simulated and real-world scenarios show that NeuGrasp outperforms state-of-the-art methods in grasping while maintaining comparable reconstruction quality.
Fig. 2. Framework of NeuGrasp. NeuGrasp leverages background priors for neural surface reconstruction and material-agnostic grasp detection. A Residual Feature Enhancement module is proposed to enhance the model attention on foreground objects instead of irrelevant background information. We build an occupancy-prior volume from residual features and a shape-prior volume from scene features. These volumes are then combined with multi-view features using Residual and Source View Transformers, which are further refined by a Ray Transformer to capture geometric details. The resulting unified view feature and attention-weighted features are decoded into a signed distance function and converted into a radiance field. Finally, the grasping module maps the reconstructed geometry to 6-DoF grasp poses, enabling end-to-end training.
If you have any questions, please feel free to contact us: