Grasp Anything
"Grasp Anything" research focuses on enabling robots to robustly grasp a wide variety of objects in cluttered environments, mimicking human dexterity. Current efforts concentrate on developing vision-based systems that leverage deep learning models, such as those based on diffusion models, transformer networks, and graph neural networks, often incorporating multimodal data (vision and language) and incorporating techniques like prompt engineering and hierarchical policy learning to improve grasp detection and planning. This field is crucial for advancing robotics in various sectors, including manufacturing, logistics, and assistive technologies, by enabling more versatile and adaptable robotic manipulation.
Papers
Grasp-Anything: Large-scale Grasp Dataset from Foundation Models
An Dinh Vuong, Minh Nhat Vu, Hieu Le, Baoru Huang, Binh Huynh, Thieu Vo, Andreas Kugi, Anh Nguyen
Affordance-Driven Next-Best-View Planning for Robotic Grasping
Xuechao Zhang, Dong Wang, Sun Han, Weichuang Li, Bin Zhao, Zhigang Wang, Xiaoming Duan, Chongrong Fang, Xuelong Li, Jianping He