Feiyu Chen
2019. Mar 22th
The goal of the project is to locate the 6D pose of an object from the RGB-D image. This task is divided into 4 components. I've completed the first 3 components as shown in following paragraphs.
Scan objects 3D point cloud model by Baxter robot and depth camera.
See this Github:
https://github.com/felixchenfy/3D-Scanner-by-Baxter
A figure and a GIF of scanning a bottle is shown below.
Use a depth camera to detect objects on table and get their mask in 2D image. The image and mask are used for training Yolo, which is a deep learning framework for object detection.
See this Github:
https://github.com/felixchenfy/Mask-Objects-from-RGBD
A figure of the image, point cloud, and the detected object and mask is shown below.
The masked objects obtained in the last step were put onto different background images with random locations, scales, rotations, and shears to complete the data augmentation step. Then, these generated images are used for training Yolo.
See this Github:
https://github.com/felixchenfy/Data-Augment-and-Train-Yolo
Four generated images are shown below.
The final result of Yolo is shown below. Video are at here: bottle, meter.
Bottle | Meter |
---|---|
(TODO)
Since I've get the object's 3D model and its 2D pos in RGB image, I can locate the object's 6D pose by fitting the 3D model onto the RGB-D image.