Skip to main content

Image Processing with Intel Realsense for Digital Interactive Installations (1/3)

This is a three part post on using Intel's Realsense depth cameras in a digital theme park, particularly for detecting people from a top-view position. The detected positions are used as input to provide visual effects projected back onto the floor, or in some cases, onto a ball pool.

The following image will probably better explain the setup used.


If you look closely, there are three Realsense cameras attached to the ceiling close to the projectors.

The targets of the project are:
  • We want to project wave ripple effects whenever someone is present.
  • The effect center is where the person is standing and ripples out just like a ripple in the water.
In this first part, I will share how we decided to detect a person using Realsense D415 cameras.

Person Detection Process

There are several ways to detect people, but since we do not need a very accurate position information I found that one of the easiest ways is to utilize Realsense's depth sensor to detect people's heads. This is done by thresholding the depth image at a certain value as shown in the following figure.


In the actual implementation, because I need to detect kids as well as adults, I ended up using two range values as thresholds.

After thresholding depth, we get an image similar to the image below.

In reality, the image is noisy and the blobs are not clean circles. However, the above illustration should convey the general idea of the algorithm. Therefore the output of the sensor is an array of blob positions that correspond to people's positions.

[(x1, y1), (x2, y2), ... , (xn, yn)]

Keep in mind that the above positions are seen from one Realsense device, with a resolution of 1280x720. In the next part, I will share on the data merging between several Realsense cameras.

Appendix : Trial and Experiments

When considering how to easily detect people from top view, and still maintain a certain accuracy, I tried to implement the detection method using paper cups as experiments. It took several tuning and experiments before I could be sure the method would work relatively well.

Paper cups taped to the wall and detected by the depth camera

This setup also served as the basis of multi camera calibration for my system later on.

The paper cups height from the wall

As seen in the image, the depth camera was able to determine the height (depth) of the paper cups. This eventually became the base for the people detection system mentioned in the previous section.

See Part 3 here: http://www.dailytinkerer.com/2020/04/image-processing-with-intel-realsense_18.html


Comments

Popular posts from this blog

Using FCM with the new HTTP v1 API and NodeJS

When trying to send FCM notifications I found out that Google has changed their API specifications. The legacy API still works but if you want to use the latest v1 API you need to make several changes. The list of changes is listed on their site so I won't be repeating them again but I'll just mention some of the things that caused some trial and error on my project. The official guide from Google is here : Official Migration Guide to v1 . The request must have a Body with the JSON containing the message data. Most importantly it needs to have "message" field which must contain the target of the notification. Usually this is a Topic, or Device IDs. Since my previous project was using GAS, my request had a field called "payload" instead of "body". Using the request from my previous project, my request in Node JS was as follows: request ({ url: 'https://fcm.googleapis.com/v1/projects/safe-door-278108/messages:send' , method: ...

Building a native plugin for Intel Realsense D415 for Unity

Based on a previous post , I decided to write a plugin for the Intel Realsense SDK methods so we can use these methods from within Unity. FYI Intel also has their own Unity wrapper in their Github repository , but for our projects, I needed to perform image processing with OpenCV and passing the results to Unity instead of just the raw image/depth data. There is a plugin called OpenCVForUnity to use OpenCV functions from Unity but previous experiments indicate the image processing inside Unity can take a long time. I hope this post can help someone else who wants to use Intel's cameras or any other devices natively in Unity. Test Environment Windows 10 64bit Unity 2017.2.0f3 x64 bit Realsense SDK from Intel CMake 3.0 or higher Steps Checkout the native plugin code here . Don't worry about the other projects in the same repository. The relevant code is in the link above. Checkout the Unity sample project here . However, instead of master you need to go to the br...

Microsoft Azure Face API and Unity

During one of my projects, I came across Microsoft's face recognition API (Azure Face API) and it looked good enough to recognize people's faces and detect if a person is a newcomer or a repeating customer to our store. As our installations mainly use the game engine Unity, I wanted to be able to use the Face API from Unity. Face API does not have an SDK for Unity but their requests are just HTTP requests so the Networking classes in Unity can be wrapped into methods to make it easy to call these APIs. First of all, to those who just want to see the code, here it is . My tests focus on the identification of a face in an input image. The full tutorial I followed can be found here . The Main scene goes through the steps in the tutorial such as creating a PersonGroup and adding Persons to the group if it is not created yet. Just make sure you: Change the API key. I used a free trial key which is no longer valid. Use whatever images you want. I don't mind you us...