Home

Awesome

DoUnseen: Segment-&-Classify-Anthing for Robotic Grasping

This library contains a pipeline to detect object without training.

robot grasping

The 3 main features of the library:

1- Unseen object segmentation

2- object identification to find a specific object or classify all objects

3- Suction point calculation

They can be use separately or cascaded.

installing dependencies

First, install other dependencies with pip:

pip install open3d opencv-python argparse torch torchvision

Second,

if you want to use Segment Anything (SAM) for classification then install it from here

If you want use the Mask R-CNN model from this repository then install detectron2 library from here.

if you use ROS, then install

sudo apt install ros-"$ROS_DISTRO"-ros-numpy

Python example

To test the model directly without ROS

git clone https://github.com/AnasIbrahim/image_agnostic_segmentation.git
cd image_agnostic_segmentation
wget 'https://drive.usercontent.google.com/download?id=1WxNVDGhhdces-qpgA5bagdi1geVBqYUX&export=download&authuser=0&confirm=t&uuid=e6d22b99-b6cc-4844-a5d3-412d4d6b0f20&at=APZUnTUsRwb78hJd0-B2dL-BHuPe:1705549231572' -c -O 'models.zip'
unzip models.zip -d models
cd scripts

# to run the example that run. The examples runs:
# 1- unseen object segmentation
# 2- classify all objects
# 3- find a specific object
# 4- calculate suction grasp for all objects
# Note: increas batch size to whatever fits in your GPU
python segment_image.py --batch-size 200 --compute-suction-pts --detect-all-objects --detect-one-object --use-buffered-gallery

# To run the unseen object segmentation only with RGB images
python segment_image.py --batch-size  --rgb-image-path RGB_IMAGE_PATH

# To detect a specific object from RGB images from a gallery 
python segment_image.py --batch-size FIT_IN_GPU --rgb-image-path RGB_IMAGE_PATH --detect-one-object --object-name OBJECT_NAME --gallery_path GALLERY_PATH

# To detect all objects from RGB images with a pre-taken image of the object
python segment_image.py --batch-size FIT_IN_GPU --rgb-image-path RGB_IMAGE_PATH --detect-all-objects --gallery_path GALLERY_PATH

# To segment an image and compute grasps
python segment_image.py --batch-size FIT_IN_GPU --rgb-image-path RGB_IMAGE_PATH --depth-image-path DEPTH_IMAGE_PATH --depth-scale DEPTH_SCALE -c-matrix FX 0.0 CX 0.0 FY CY 0.0 0.0 1.0 --compute-suction-pts

The examples shows the following scene: grasp computation

ROS

To install the ROS driver (the ROS package is currently broken):

mkdir -p catkin_ws/src
cd catkin_ws/
catkin init
cd src/
git clone https://github.com/AnasIbrahim/image_agnostic_segmentation.git
wget 'https://drive.usercontent.google.com/download?id=1WxNVDGhhdces-qpgA5bagdi1geVBqYUX&export=download&authuser=0&confirm=t&uuid=e6d22b99-b6cc-4844-a5d3-412d4d6b0f20&at=APZUnTUsRwb78hJd0-B2dL-BHuPe:1705549231572' -c -O 'models.zip'
unzip models.zip -d models
cd ../../..
catkin build
echo "source $(pwd)/devel/setup.bash" >> ~/.bashrc
source ~/.bashrc

To run ROS example (unseen object segmentation only):

roslaunch image_agnostic_segmentation test_example.launch

Then wait till the segmentation image then grasping image appears (~10 second)

DoPose Dataset

Th unseen object segmentation model was trained with our Dopose data. The dataset can be downloaded here. The dataset is saved in the BOP format. It includes multi-view of storage bin (KLT Euro container) and tabletop scenes. The annotations include RGB and depth images, 6D pose of each object, segmentation masks, COCO json annotations. Also the dataset includes camera transformations between different views of the same scene (this is extra non-standard to BOP format).

Samples from the dataset: DoPose dataset sample

Papers and Citation

For more details about the DoPose dataset please refer to our DoPose-6D paper (Arxiv) and use this citation:

@INPROCEEDINGS{10069586,
  author={Gouda, Anas and Ghanem, Abraham and Reining, Christopher},
  booktitle={2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA)}, 
  title={DoPose-6D dataset for object segmentation and 6D pose estimation}, 
  year={2022},
  volume={},
  number={},
  pages={477-483},
  doi={10.1109/ICMLA55696.2022.00077}}

For more details about the classification please refer to our DoUnseen paper (Arvix). This paper gives information about the classificaiton with the siamese network up till this commit. For any work with the classification after this commit another paper will be added soon.:

@misc{gouda2023dounseen,
      title={DoUnseen: Tuning-Free Class-Adaptive Object Detection of Unseen Objects for Robotic Grasping}, 
      author={Anas Gouda and Moritz Roidl},
      year={2023},
      eprint={2304.02833},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Latest updates

Jan 18 2024:

This research is supported by the LAMARR institute

<img src="images/lamarr_logo.png" width="200">