NVIDIA Unveils GraspGen: Revolutionizing 6-DOF Robotic Grasping with Diffusion Models
NVIDIA introduces GraspGen, a diffusion-based framework that significantly improves 6-DOF robotic grasping using large-scale synthetic data and innovative training, achieving superior performance in simulation and real-world tests.
The Challenge of 6-DOF Robotic Grasping
Robotic grasping is fundamental for automation across many fields, from industrial picking to humanoid robots. Achieving reliable 6-degree-of-freedom (6-DOF) grasping, which involves precise position and orientation in 3D space, remains a difficult challenge due to the need to generalize across unknown objects, various gripper types, and complex environments with partial data and clutter.
Limitations of Existing Approaches
Traditional model-based planners rely heavily on accurate object pose estimation or multiple views, limiting their use in real-world, unpredictable settings. Data-driven methods show promise but often struggle with generalization, scalability, and require expensive real-world data collection that is not easily transferable.
GraspGen: Leveraging Large-Scale Simulation and Diffusion Models
NVIDIA’s GraspGen framework sidesteps costly real-world data by utilizing large-scale synthetic data generated in simulation. It uses over 8,000 diverse object meshes from the Objaverse dataset and simulates more than 53 million grasp attempts.
GraspGen frames grasp generation as a denoising diffusion probabilistic model (DDPM) over SE(3) pose space, including rotations and translations. This approach iteratively refines noise into realistic grasp poses conditioned on object point clouds, naturally modeling the multiple valid grasps for complex objects and enabling spatial diversity required for cluttered scenes.
Innovative Architecture and Training
GraspGen uses a Diffusion Transformer Encoder combining a PointTransformerV3 backbone to encode raw 3D point clouds into latent features, followed by diffusion steps predicting noise residuals in grasp poses. This provides better grasp quality and efficiency compared to previous methods.
A novel on-generator training strategy is applied to the discriminator, which learns from grasp poses generated during training rather than static datasets. This exposes the discriminator to typical model errors, improving its ability to filter out false positives during inference.
Efficient weight sharing allows the discriminator to reuse the frozen object encoder from the generator, with only a small multilayer perceptron trained for grasp success classification, reducing memory use by 21 times compared to earlier architectures.
Translation normalization and rotation encoding via Lie algebra or 6D representations ensure stable and accurate pose predictions.
Versatility Across Grippers and Environments
GraspGen supports multiple gripper types, including parallel-jaw, suction, and plans for multi-fingered grippers. It performs robustly with partial or complete point clouds and handles both single objects and cluttered scenes effectively.
Benchmark Success and Real-World Performance
On the FetchBench cluttered grasping benchmark, GraspGen achieved top task and grasp success rates, outperforming state-of-the-art baselines by a significant margin. In simulation, it improved task success by nearly 17% over Contact-GraspNet.
In real robot experiments using a UR10 with RealSense sensors, GraspGen achieved an 81.3% grasp success rate, exceeding other methods by 28%, demonstrating strong zero-shot sim-to-real transfer.
Open Dataset and Code Release
NVIDIA has publicly released the GraspGen dataset, containing approximately 53 million simulated grasps over 8,515 object meshes, along with the codebase and pretrained models under open-source licenses.
Impact on Robotic Manipulation
GraspGen marks a significant advancement in robotic grasping by combining diffusion-based generative modeling with large-scale synthetic data and innovative training techniques. This framework enhances grasping reliability and adaptability across diverse scenarios, pushing the boundaries of real-world robotic manipulation.
For more details, visit the GitHub repository and the project website.
Сменить язык
Читать эту статью на русском