Skip to main content

3D Reconstruction with markers for multi view object alignment

In one of my projects, I needed to reconstruct an object for automatic painting by a robot arm. The target object is a hollow electricity meter box and the robot needs to paint the insides of this object. It is easy to paint the exterior of an object, but to paint the insides of a hollow object we need a decent enough point cloud to make sure our path planning works properly.

An illustration of the object to draw is given below


As indicated by the red arrows, the area just behind the small area is the target of the painter.

Since I don't have electricity boxes lying around, I tried my experiments on a cereal carton.



This is a long winding road of trials and errors and is divided into several parts. The first part is about the point cloud generation. Followed by experiments on camera movement tracking and scene and object reconstruction from point clouds. Finally a method to determine path points for the robot painter will be discussed.

Preliminary research

One of the first things to try was 3D reconstruction of the scene. Hopefully we can get a good representation of the object that is enough for the path finding algorithm.

3D Scene Reconstruction

Using a Realsense D415 depth camera, I attempted to capture the cereal box as point cloud.

Point cloud representation of the cereal box

This approach has a few problems, mainly:
  • Edges have a lot of noise
  • Sides that are not visible from the camera are not detected
To solve these problems, I looked into ways to merge multiple view point clouds so we can get a full representation of the box. I tried first with ICP and performing iterative point registration but the accuracy is not good enough so I tried using a marker to capture various views and merge the point clouds together.

Marker based multi-view point cloud reconstruction

The following is my attempt to get a better point cloud reconstruction from multi view using a marker of known pattern and size. The marker chosen was the chessboard calibration pattern widely used in camera calibrations. The following images depict the scene setup.







Three views of a sofa in the living room along with the chessboard pattern

The usage of a marker improved the reconstruction a lot and I managed to get the reconstructed sofa as shown below.


3D point cloud representation of the sofa



Color image of the sofa and its surroundings from the same view

Implementing marker based multi-view registration on the cereal box

So the next step will be to try the reconstruction process on the cereal box. The marker has also to be visible to the camera so this poses a small challenge in camera placement but overall I found that overcoming this is not so hard.




Two views of the cereal box next to a known marker

With this method, there is no need to incrementally register point clouds and we can just rely on the transformation matrix of the camera relative to the marker.

The result reconstruction is shown below.


We can see that the transformation matrix from the marker really helped in aligning the two views, even though they are far apart and we don't have incremental positions in between. The accuracy needs more work however and I believe this can be overcome by better hardware, and also more thorough calibration of the camera parameters. Using multiple markers may also improve the transformation matrix estimation.

Future work

A better calibration process, along with multiple markers seem to be promising to increase accuracy of alignment.

Comments

Popular posts from this blog

Object detection with Google Colab and Tensorflow

This is just a memo of the challenges I faced when running a model training on Google Colab, while following a great tutorial here . Mind the versions Tensorflow is currently at version 2.2.0 but most tutorials are still using the contrib package, and there is no known easy way to update the code to remove dependency on contrib. So my best bet is to downgrade the tensorflow version to 1.x. Since Google Colab only gives the options of either 1.x or 2.x and we cannot specify the exact version, I ended up with version 1.15.2. Even with the command :  %tensorflow_version  1.15.0 I ended up with : 1.15.2 Another pitfall was the version of numpy. Installing numpy gives us the version 1.18.3 but for some reason this generates the error : TypeError: 'numpy.float64' object cannot be interpreted as an integer Downgrading numpy to version 1.17.4 solved this for me. It seems we don't need ngrok for tensorboard With the command :  %load_ext tensorboard W...

Installing a custom ROM on Android (on the GT-N8013)

It's been a while since my last entry and since it is a new start in 2019, I thought I'd write something about "gone with the old and in with the new". I've had my Samsung Galaxy Note 10.1 (pnotewifi) since 2014, and it's one of the early Galaxy Note tablet series. It has served me well all this years but now it just sits there collecting dust. My old Samsung GT-N8013 I've known a long time about custom Android ROMs like CyanogenMod but has never had the motivation to try them out, until now ! Overview of the process For beginners like me, I didn't have an understanding of the installation process and so it looked complicated and it was one of the reasons I was put off in trying the custom ROM. I just want to say, it's not complicated at all!   Basically you will need to Prepare an SD card and install Android SDK (you need adb ). Install a custom boot loader ( TWRP is the de facto tool at the moment). Use adb to copy custom...

Using FCM with the new HTTP v1 API and NodeJS

When trying to send FCM notifications I found out that Google has changed their API specifications. The legacy API still works but if you want to use the latest v1 API you need to make several changes. The list of changes is listed on their site so I won't be repeating them again but I'll just mention some of the things that caused some trial and error on my project. The official guide from Google is here : Official Migration Guide to v1 . The request must have a Body with the JSON containing the message data. Most importantly it needs to have "message" field which must contain the target of the notification. Usually this is a Topic, or Device IDs. Since my previous project was using GAS, my request had a field called "payload" instead of "body". Using the request from my previous project, my request in Node JS was as follows: request ({ url: 'https://fcm.googleapis.com/v1/projects/safe-door-278108/messages:send' , method: ...