Next Article in Journal
Hybrid Learning Models for IMU-Based HAR with Feature Analysis and Data Correction
Previous Article in Journal
An Adaptive Multi-Scale Network Based on Depth Information for Crowd Counting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over

1
Neuroinformatics and Cognitive Robotics Lab, Technische Universität Ilmenau, 98693 Ilmenau, Germany
2
Group for Quality Assurance and Industrial Image Processing, Technische Universität Ilmenau, 98693 Ilmenau, Germany
3
Fraunhofer Institute for Applied Optics and Precision Engineering, IOF Jena, 07745 Jena, Germany
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(18), 7807; https://doi.org/10.3390/s23187807
Submission received: 17 August 2023 / Revised: 1 September 2023 / Accepted: 10 September 2023 / Published: 11 September 2023

Abstract

:
In the context of collaborative robotics, handing over hand-held objects to a robot is a safety-critical task. Therefore, a robust distinction between human hands and presented objects in image data is essential to avoid contact with robotic grippers. To be able to develop machine learning methods for solving this problem, we created the OHO (Object Hand-Over) dataset of tools and other everyday objects being held by human hands. Our dataset consists of color, depth, and thermal images with the addition of pose and shape information about the objects in a real-world scenario. Although the focus of this paper is on instance segmentation, our dataset also enables training for different tasks such as 3D pose estimation or shape estimation of objects. For the instance segmentation task, we present a pipeline for automated label generation in point clouds, as well as image data. Through baseline experiments, we show that these labels are suitable for training an instance segmentation to distinguish hands from objects on a per-pixel basis. Moreover, we present qualitative results for applying our trained model in a real-world application.

1. Introduction

In the course of Industry 4.0, collaborative robots (cobots) are gaining more and more attention. For collaboration, successfully handing over objects between humans and cobots plays a major role. As possible injury to the human needs to be strictly avoided, the robot’s gripper should not touch the human hand. Therefore, the cobot needs to be aware of its surroundings and recognize objects of interest, as well as the human hand. To enable further processing steps, such as grasp planning and the execution of robotic motion trajectories, robust pixel-wise instance segmentation is required. State-of-the-art methods for instance segmentation such as Mask-RCNN [1] or PointRend [2] process RGB image data to achieve this goal. Recently, Transformer-based methods have outperformed CNN-based architectures. These models need a large amount of data or extensive pretraining to achieve comparable or better results [3]. If the necessary amount of training data are available, Transformers can be used as a replacement backbone for methods such as Mask-RCNN.
If additional depth data are available, the resulting pixel-wise segmentation mask can be utilized to reconstruct a labeled 3D scene for further usage, as described in [4]. Alternatively, a point cloud could be created from the RGB-D raw data, which can then be segmented using techniques such as PointNet [5], RandLA-Net [6], or SO-Net [7]. In [8], these and other point cloud segmentation methods were evaluated on the presented dataset.
In other applications, multi-spectral imaging has proven to be helpful when it comes to the segmentation of organic structures [9,10]. Therefore, we are interested in adapting the aforementioned methods designed for RGB images to incorporate multi-modal data, including thermal images. To incorporate depth information, architectures have been proposed that extract information from both depth and color images and fuse them at different stages within their backbone [11,12].
In addition to a method for generating instance segmentation masks, a suitable dataset is of major importance for both training and testing a model. There are several datasets with segmentation labels for hands [13,14,15,16,17,18,19], but few of these contain objects as well as labels. An overview of the datasets most similar to the dataset presented in this paper is presented in Table 1, and the sizes, modalities, and available label types in these datasets are also listed. HandNet [13] and the dataset presented in [14] contain depth and even thermal images but lack labeled objects. In contrast, the WorkingHands dataset [17] contains segmentation labels for objects but no thermal images. However, the available labels are merely semantic segmentation labels. Overlapping instances of the same category can, therefore, not be distinguished, which is a major requirement for grasping individual objects. Moreover, the majority of pixels in WorkingHands are labeled as void, which is an ill-defined class. This often leads to unnecessary object pixels predicted in the background (as seen in their results) and might be a problem for applications where it is necessary to locate individual objects.
Most similar to our dataset is the ContactPose [19] dataset. Even though the intention behind ContactPose is contact modeling between human hands and objects, it also includes instance segmentation labels for hands and objects. However, all objects are 3D-printed in blue, which limits its applicability to real-world instance segmentation. The only dataset that contains thermal images is the one of Kim et al. [14].
When collecting a dataset with the intended properties, one of the most time-consuming tasks is labeling. Ideally, either the recorded data or the setup allows for automated label generation. However, for good generalization, the collected data should still resemble real-world data. For example, the recording of thermal data in [14] allowed for the automated generation of instance segmentation labels for hands. However, the segmentation of the objects is just as important when objects are handed over to the cobot. Therefore, we additionally designed a recording setup so that the automated segmentation of hands and objects is easily possible while keeping the input data close to a real-world setting.
We want to emphasize that our automated labeling allows us to easily expand our dataset with new objects. This distinguishes our approach from those of other fixed datasets, which require manual labeling, as in WorkingHands [17]. In contrast to Kim et al. [14], we found that thermal images were not a useful basis for automated label generation. On the one hand, there are many warm background objects, and objects become warm in the hand during manipulation. On the other hand, thermal images have a much lower resolution than color images, which limits the label quality.
For good generalization of trained models, the diversity of training data is essential, which is realized through diverse augmentation techniques. We addressed this problem by recording the background and the hands with objects separately, which were then combined pairwise.
Due to the presented shortcomings of existing datasets in the context of our intended application, we collected a new multi-modal, multi-purpose dataset—named OHO—for human–robot Object Hand-Over. In this paper, we present the setup and collection process of the OHO dataset, as well as the automatic pipeline for generating instance segmentation labels. Then, we use the data to train state-of-the-art models in instance segmentation. Additionally, we implement the trained models in a real-world application to prove that the collected dataset is suitable for solving the instance segmentation task required for handing over objects. We created our dataset with other tasks in mind, such as 6D pose estimation and object shape reconstruction, which will be addressed in future work.
In summary, our main contributions are:
  • A multi-modal dataset, including RGB, depth, and thermal data.
  • A multi-purpose dataset for instance segmentation, point cloud segmentation, object shape, and pose estimation.
  • A recording procedure and pipeline for automated instance segmentation labeling for hands and hand-held objects in 2D images, as well as 3D point clouds.
  • Quantitative and qualitative baseline results for instance segmentation in 2D images.

2. Multi-Modal Object Hand-Over Dataset (OHO)

One goal of our ongoing research is the development of multi-modal sensors for safe human–robot interactions. Therefore, we want to compare different cameras and modalities, such as RGB, depth, and thermal data, in an application-relevant scenario.
This means that we did not want to restrict our recordings to one type of RGB-D camera. Hence, we equipped our mobile robot, TIAGo, with an Azure Kinect and an i3-System TE-Q1 thermal camera (see Figure 1). Together with its internal Orbbec Astra S camera, we had a spectrum of cameras using different methods for depth perception (time of flight in Azure Kinect, and active stereo in Astra Orbbec). This resulted in three modalities in our dataset: color, depth, and thermal (and, indirectly, stereo images). Each sample in our dataset consisted of the following images, as shown in Figure 2: RGB images from two RGB-D cameras—Azure Kinect, with a resolution of 4096 × 3072 , and Orbbec Astra S, with a resolution of 640 × 480 —as well as the registered depth images from both cameras. Additionally, each sample contained a thermal image from an i3-System TE-Q1, with a resolution of 384 × 288 .
For the selection of objects used in our dataset, we focused on handy items that a robot is able to grasp. Additionally, we captured shape data in the form of a triangle mesh for every recorded object by scanning them with a 3D scanner or modeling them in a CAD program. Moreover, for each sample, the relative poses of the cameras and the object (see Section 2.1), as well as the metadata for the scene, were generated. These included which hand (left or right) was holding the object and whether the object had been masked using masking tape to make it visible to the depth cameras. This was necessary for very dark or glossy metal objects (e.g., a black rubber hammer or the blade of a screwdriver), since both cameras had difficulties capturing depth data for some objects.
For recording, we used a green screen, which enabled automated label generation and foreground–background composition, as discussed in detail in Section 2.3. To replace the green screen background in every modality, we also recorded a set of backgrounds of 234 office scenes with the same camera configuration on the mobile robot. Afterward, the data were split into 171 backgrounds for training, 33 for validation, and 30 for testing.

2.1. Setup for Dataset Recording

The cameras had to be calibrated internally and externally to allow for the registration of thermal, depth, and RGB data. This was performed using a special checkerboard, which is visible in the binarized thermal image, as well as in the visual and near-IR images from the Azure Kinect and Astra Orbbec cameras. For details on the external calibration of the thermal and near-IR cameras, we refer the reader to [8], where along with the mathematics, the active calibration target is described precisely. Note that the Orbbec Astra S camera suffers from a thermal problem that affects the depth data over time. To compensate for this, a parametric transformation of the depth data was performed, which scaled the depth values by a factor that was linearly interpolated between two manually set parameters along the x-axis of the depth image. These scaling factors were manually set by visually aligning the resulting point clouds from both depth cameras.
As a prerequisite for potentially training pose estimation models and for our automatic labeling process, we needed the exact 6D pose of the object in the scene. Therefore, each object was fixed statically on a tripod in front of the green screen, and the point cloud was cropped so that only object points remained in the region of interest. The region of interest was defined relative to the pose of the tripod and was captured through the use of ArUco [20] markers on the tripod (see Figure 3 left). These markers were removed or covered by a green screen before recording the actual data. Note that the green screen was intentionally small to limit the reflections of green ambient light on the object. Afterward, we computed the object’s pose using an ICP (Iterative Closest Point) method [21], registering the given 3D model of the object to the point cloud of the Orbbec Astra S camera. The ArUco marker-based tracking of the tripod pose helped initialize the object pose after the pose had been changed. To minimize the required effort for recording, we used one object pose for multiple samples of a human hand holding the object. For each of the object poses, we recorded a reference image that only showed the object without a hand holding it. This has several advantages, such as providing useful information for automatic label generation (described later), having images of the single object for training a model that needs to be able to recognize the objects regardless of whether they are being held, and recomputing the object pose after the recording has taken place. After the reference image was taken, different people placed one of their hands on the fixed object and pretended to hold it. The hands in the images belong to seven different males and females. During recording, the point clouds were labeled automatically in real time, as described in Section 2.2. This allowed the operator to use the result as feedback to assess whether the setup and the current sample were correct. We made sure to manually check the results during recording and adjusted the ground-truth object poses when necessary.
Following this procedure, which is summarized in Figure 3, we recorded about 10 poses for each of the 43 objects in the 32 categories, where each object pose comprised about 10 different hand positions. This yielded a dataset of 5300 samples. We carefully split all the samples according to the object poses into training, validation, and test (one of the object poses for validation and one for test). Therefore, similar samples were not included in different splits.

2.2. Automated Labeling in Point Clouds

The first step in the automated labeling of the dataset was the annotation of point cloud points belonging to either the object or the human hand. Background points were segmented using a 3D region-of-interest box in relation to the fixed tripod position. Since we knew the 3D shape of the object and its pose, we computed the distances of all remaining point cloud points to the surface of the object model. By thresholding these distances ( t = 6 mm), a coarse segmentation of the object and hand points could be performed. Unfortunately, this resulted in some points on the hand that were close to the object being counted as object points. To compensate for this, the hand segments were dilated by 3–6 mm, depending on the current pose and object properties. This means that object points with a distance to a hand point smaller than the dilation radius were switched to hand points. In the end, we obtained a labeled point cloud, as shown in Figure 2, which can be used for either training point cloud segmentation methods or as a starting point for labeling color images, as described below. Note that the point cloud annotation only worked with the Orbbec Astra S point clouds, which used an active stereo approach. In contrast, the Azure Kinect point cloud did not produce usable segmentation results due to artifacts caused by abrupt changes in depth values. The time-of-flight depth image from the Azure Kinect contained interpolated points at object borders, which resulted in phantom 3D points in free space, making them difficult to filter out automatically. Thus, we used automated labeling of point clouds only for the images from the Orbbec Astra S data.

2.3. Automated Instance Segmentation Label Generation

To train a model to segment hands and objects, instance segmentation labels are required. Since it is time-consuming and, therefore, costly to label the samples by hand, an automated approach was used, leveraging the green screen and the recording of the reference samples of objects without hands. Our approach can be divided into two steps. First, a segmentation of the images into the foreground (hand and object) and the background is performed by removing the green screen. In the next step, the results of this foreground–background segmentation and the labeled point clouds (Section 2.2) projected onto the color image are used to compute the final instance segmentation labels with GrabCut [22]. These two steps are discussed in detail below.
To remove the green screen, we used the images from the Kinect Azure camera due to their better image quality, which made color keying easier. However, the generation pipeline described below can also be adjusted for use with color images recorded by the Orbbec Astra camera.

2.3.1. Green Screen Removal

The first step in generating instance segmentation labels is to remove the green screen from the images to obtain a rough segmentation of the foreground—containing the hand and object—and the background. The free graphics software Blender (v2.82.7, https://www.blender.org/, accessed on 1 January 2020) comes with an implementation of green screen removal and the option to use it via a Python interface. For these reasons, the keying node in Blender was used to segment the foreground from the background. An example of the green screen removal results can be seen in Figure 4. Besides removing the green screen, this node is also capable of removing the green spill (green light reflected from the green screen and visible on the objects and hands). This is important for the images to be more realistic when replacing the background later, as described in Section 3. Because in our setup the green color of the fabric used in the background was slightly different than the green used to hide the mounting stand, we used two keying nodes in succession.

2.3.2. GrabCut

Based on the green screen foreground segmentation and a projection of the segmented point cloud, we used GrabCut [22] to compute a detailed segmentation mask for the hand. In some cases, the skin color segmentation performed in Blender might be sufficient, but for most cases, this refinement is necessary to differentiate the fingers from the object because of shadows. With this hand mask, we subtracted the hand pixels from the foreground segmentation of the reference image to generate the final mask for the object in the sample image.
GrabCut performs the segmentation of an image into the foreground (in our case, the hand) and the background by formulating the segmentation as a graph cut problem, which aims to maximize the difference in the color histograms of foreground and background pixels. Therefore, we needed to first define prior label regions that specify the initial color histograms of the foreground and background. GrabCut differentiates four labels: foreground (FGD), probably foreground (PR_FGD), background (BGD), and probably background (PR_BGD). Figure 5 shows an overview of how the Blender results were used to assign parts of the image to these classes.
First, we preprocessed the foreground segmentations and the projected point cloud by applying opening and closing operations to remove segmentation artifacts. Then, the images were cropped around the center of the object to reduce computational complexity.
For removing static parts like the wall, which was not covered by the green screen, we computed a difference image between the reference recording and the recording containing the hand. After applying opening and closing operations on the difference image, we used a connected component analysis (CCA) to filter out the remaining small components, leaving only the hand and object areas. For the reference image, the walls were eliminated by removing components touching the border of the image, as the object was always located in the middle.
The resulting coarse hand segmentation served as the foreground for GrabCut, whereas the preprocessed projected point cloud of the object served as the background.
As we observed erroneous results when applying GrabCut with only a few object parts visible, the unoccluded reference object was inserted into the image by copying it to a space that was not occupied by the original object or the hand. Defining these pixels as background helped GrabCut segment parts of the object that were similar to the hand in color.
Finally, we ran the GrabCut optimization and obtained a refined segmentation of the hand. By subtracting this hand segmentation from the segmentation of the reference object, the final object segmentation mask was obtained. Figure 6 shows examples of the final instance segmentation labels overlaid on the color image.

3. Dataset Generation for Instance Segmentation

Before any of the generated labels could be used for training, the green screen on the input images needed to be replaced. Otherwise, segmentation could become trivial for a model specializing in the green screen. By using the captured background recordings and the foreground segmentation computed by Blender, the background of the images could easily be replaced.

3.1. Augmentation

For each sample of foreground, we randomly choose 20 background images of the same split. To further augment the appearance, we randomly cropped both the foreground and the background and applied random rotation and color jitter to the foreground. Afterward, the foreground and the background images were combined.

3.2. Combination of Foreground and Background

When combining the foreground and background images, special care needed to be taken at the edges where the object and hand ended and the background began. Simply stitching both color images together may result in artifacts, which could be learned by the model for easier recognition and bypassing the actual task. Therefore, specific methods, such as Gaussian blur, as employed by Dwibedi et al. [23], could be employed to blend both images. Figure 7 shows a comparison of simple overlaying and blending using Gaussian blur at the edge of the mask. Finally, the shorter edge of the combined image was resized to 448 pixels to further reduce any remaining artifacts.

3.3. Incorporating Thermal Data

For training on combined RGB and thermal images, the raw thermal images were registered to the RGB camera’s point of view using the depth data. To this end, reconstructed 3D points from the registered depth image were projected onto the thermal image plane using the respective intrinsic camera parameters. Thus, the thermal layer had the same size as the RGB image, but it inherited missing pixels from the incomplete depth images, which were set to zero. The registered thermal images of the foreground and background underwent the same augmentation and combination procedures as the RGB images, except for the color jittering.

3.4. Dataset Statistics

After generation, our dataset included 33 categories—32 object categories and the hand category. The training split contained 75,480 images with 146,912 annotations, the validation split contained 10,860 images with 21,052 annotations, and the test split contained 11,800 images with 23,067 annotations.

4. Experiments and Results

To demonstrate the effectiveness of our labeling and generation pipeline and establish baselines for our dataset, we trained state-of-the-art instance segmentation methods, such as Mask R-CNN [1] and PointRend [2], as well as the recently proposed, efficient YolactEdge [24]. YolactEdge achieved 61 FPS compared to 14 FPS for Mask-RCNN on an RTX 2080 Ti [24], making it a suitable choice for deployment in robotic applications. Due to their required amount of training data, we did not train Transformer-based models. PointRend is built upon Mask R-CNN and iteratively refines segmentation masks with higher resolutions, similar to rendering in computer graphics. Thus, the resulting segmentation masks were much more fine-grained than those from Mask R-CNN. In contrast, YolactEdge computed category-agnostic segmentation prototypes in parallel with bounding boxes and coefficients for combining the segmentation prototypes for each instance. Due to this parallel computation, YolactEdge was much more efficient and achieved up to 30.7 FPS with a ResNet-50 backbone on an NVIDIA Jetson AGX Xavier, which is common hardware for deployment on a mobile robot platform. We used Detectron2 [25] to train Mask R-CNN and PointRend. Unlike the default configuration, we did not freeze any part of the network. YolactEdge was trained using its officially provided code. For a fair comparison, we used a ResNet-50 [26] backbone for all three architectures. Moreover, we compared the results when transfer learning from an instance segmentation model pretrained on COCO [27] to simply using an ImageNet-pretrained [28] backbone.
Additionally, we present baseline results for incorporating thermal data. To keep it simple, we used only a four-channel input. As the pretrained weights of the first convolution only accounted for three input channels, we added a randomly initialized fourth channel for the new modality. Note that more advanced architectures and fusing methods of different modalities, such as in ESANet [11], will probably perform better.
We evaluated the performance of our trained models using the typical COCO metrics, including average precision (AP) with different intersection-over-union thresholds. The primary challenge metric, AP50:95, is the mean of 10 average precision values with IoU thresholds in the range of [0.5, 0.95]. Since we are especially interested in the segmentation of the hands, we also report the AP50:95 for the hand category. Our results are presented in Table 2. As expected, the models pretrained on COCO outperformed those with pure ImageNet pretraining. The segmentation AP of PointRend was better, especially for higher IoU thresholds, which can be attributed to its iterative mask refinement.
Remarkably, the performance of the much more efficient YolactEdge was on a par with that of Mask R-CNN. Training with additional thermal data further improved the segmentation AP slightly for higher IoU thresholds. Unintuitively, thermal data did not improve the AP Hand . A possible explanation could be that due to cold hands or warm electronic devices in the background—as shown in the second column in Figure 8—distinguishing the hand from the object and background was not trivial. Therefore, we assume that using a simple four-channel input is not sufficient for effectively incorporating thermal data, and more sophisticated multi-modal architectures should be explored.
In addition to the quantitative evaluation and to demonstrate the generalization capabilities, we also present qualitative comparisons in a real-world setting in Figure 9.
Note that our models had never seen full-sized people, and the ground-truth segmentation ended at various positions on the arm. Therefore, the segmentation performance in terms of the right side of the arm should be ignored. It can be seen that the fingers were segmented in much more detail through the iterative refinement in PointRend. As already observed in the quantitative metrics, the far more efficient YolactEdge output segmentation masks that were on a par with those of Mask R-CNN.
When grasping objects, one might only be interested in distinguishing the hand from the held object, but not the exact category of the object. Therefore, we combined all object categories and trained the instance segmentation models on only the two categories: hand and object. Such a category-agnostic instance segmentation of held objects offers the potential to operate the robot with unknown objects. This approach demonstrated improved AP compared to the multi-class problem (shown at the bottom in Table 2). In future research, we will also evaluate the performance on unseen objects.
In the last experiment, we demonstrated that the models trained on our OHO dataset did not simply focus on stitching edges by applying them to cross-domain datasets. In Figure 10, we present the qualitative results for the application of the YolactEdge model working on RGB inputs on WorkingHands [17] and ContactPose [19]. The segmentation results on both the synthetic and real-world images are impressive. The objects of known categories (wrench and scissors) were successfully detected, whereas the unknown object under the left hand in the first image was considered to belong to the background. The object in the ContactPose example was recognized, although it did not belong to the known object categories. As we only trained on samples where a hand was grasping the object, the pliers in the WorkingHands example were not detected, as they were not being grasped by a hand. If an application necessitates the detection of these objects, the reference samples in our dataset could be included in the training data to remove the bias toward grasped objects. These results show that although we recorded our dataset using a green screen, the blending of the images and the bleeding of the green screen onto the objects still resulted in trained models that could be applied to different scenarios.

5. Discussion

The first experiments using CNN models trained on our dataset revealed some limitations of the methods. In particular, the segmentation masks from the Mask R-CNN and YolactEdge models were of lower resolution, sometimes partially missing fingertips (see Figure 9), which was related to the network architecture. Nevertheless, there were situations in which parts of hands and objects were misclassified. This might be related to the limited diversity of the training samples in the OHO dataset. To combat this, more diverse data could be captured and automatically labeled using our label-generation pipeline. For example, by incorporating different skin colors, gloves on the hands, or more diverse backgrounds, including people, the training data could be diversified, leading to better-trained models. The data recording setup, unfortunately, requires a uniform, colored background (green screen), which is associated with many restrictions. The experiments nevertheless showed that with a good stitching method, the artifacts introduced are of minor relevance for a generalization to real-world applications. On the contrary, the background replacement improves the generalization capabilities of the networks due to the increased diversity of the samples.
On the one hand, the manual intervention during data recording is a limiting factor for scaling up the number of objects, but on the other hand, it ensures high-quality segmentation masks in 3D, as well as 2D.

6. Conclusions

In summary, we described how we recorded a comprehensive dataset of hand-held objects. We were able to automatically generate instance segmentation labels for our newly recorded multi-modal dataset of hand-held objects by utilizing green screen background substitution and 3D registration of previously known object models to the captured 3D point cloud data. By training the shelf segmentation networks, we achieved basic real-time capable segmentation results, which can be used in a robotic grasping pipeline. Despite achieving satisfactory segmentation results with the recorded data, for safety-critical applications involving industrial robots with the potential to harm people, the predicted segmentation mask still needs improvement. Moreover, the potential of thermal and depth data, which are already included in the dataset, needs to be further evaluated by utilizing more advanced architectures and fusing mechanisms. We hope that with these additional modalities, the robustness of segmentation methods can be improved to a level that is acceptable for real-world robotic applications.
Besides instance segmentation, the OHO dataset offers the opportunity to investigate further tasks for robotic grasping, such as object pose estimation or object shape reconstruction which, in combination with multiple modalities, make our dataset unique. The presented OHO dataset is publicly available for scientific purposes at https://www.tu-ilmenau.de/neurob/data-sets-code/oho-dataset.

Author Contributions

Conceptualization, B.S., S.M. and Y.Z.; Data curation, S.M. and Y.Z.; Funding acquisition, H.-M.G. and G.N.; Methodology, B.S., M.K., S.M. and Y.Z.; Project administration, H.-M.G. and G.N.; Software, B.S., M.K. and S.M.; Supervision, H.-M.G. and G.N.; Writing—original draft, B.S., M.K. and S.M.; Writing—review and editing, B.S., S.M. and H.-M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work received funding from the Free State of Thuringia of the European Social Fund (ESF) for the research group SONARO and from the Carl Zeiss Foundation as part of the project Engineering for Smart Manufacturing (E4SM).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The OHO dataset is available for scientific use at https://www.tu-ilmenau.de/neurob/data-sets-code/oho-dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  2. Kirillov, A.; Wu, Y.; He, K.; Girshick, R. PointRend: Image Segmentation as Rendering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9799–9808. [Google Scholar]
  3. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations—ICLR 2021, Vienna, Austria, 4 May 2021. [Google Scholar]
  4. Seichter, D.; Langer, P.; Wengefeld, T.; Lewandowski, B.; Hoechemer, D.; Gross, H.M. Efficient and Robust Semantic Mapping for Indoor Environments. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 9221–9227. [Google Scholar]
  5. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  6. Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. Randla-net: Efficient semantic segmentation of large-scale point clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11108–11117. [Google Scholar]
  7. Li, J.; Chen, B.M.; Lee, G.H. So-net: Self-organizing network for point cloud analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 9397–9406. [Google Scholar]
  8. Zhang, Y.; Müller, S.; Stephan, B.; Gross, H.M.; Notni, G. Point cloud hand–object segmentation using multimodal imaging with thermal and color data for safe robotic object handover. Sensors 2021, 21, 5676. [Google Scholar] [CrossRef] [PubMed]
  9. Shivakumar, S.S.; Rodrigues, N.; Zhou, A.; Miller, I.D.; Kumar, V.; Taylor, C.J. Pst900: Rgb-thermal calibration, dataset and segmentation network. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9441–9447. [Google Scholar]
  10. Palmero, C.; Clapés, A.; Bahnsen, C.; Møgelmose, A.; Moeslund, T.B.; Escalera, S. Multi-modal rgb–depth–thermal human body segmentation. Int. J. Comput. Vis. (IJCV) 2016, 118, 217–239. [Google Scholar] [CrossRef]
  11. Seichter, D.; Köhler, M.; Lewandowski, B.; Wengefeld, T.; Gross, H.M. Efficient rgb-d semantic segmentation for indoor scene analysis. In Proceedings of the 2021 IEEE international conference on robotics and automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 13525–13531. [Google Scholar]
  12. Fischedick, S.; Seichter, D.; Schmidt, R.; Rabes, L.; Gross, H.M. Efficient Multi-Task Scene Analysis with RGB-D Transformers. In Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia, 18–23 June 2023; pp. 1–10. [Google Scholar] [CrossRef]
  13. Wetzler, A.; Slossberg, R.; Kimmel, R. Rule Of Thumb: Deep derotation for improved fingertip detection. In Proceedings of the British Machine Vision Conference (BMVC), Swansea, UK, 7–10 September 2015. [Google Scholar]
  14. Kim, S.; Chi, H.G.; Hu, X.; Vegesana, A.; Ramani, K. First-Person View Hand Segmentation of Multi-Modal Hand Activity Video Dataset. In Proceedings of the British Machine Vision Conference (BMVC), Virtual Event, 7–10 September 2020. [Google Scholar]
  15. Urooj, A.; Borji, A. Analysis of hand segmentation in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4710–4719. [Google Scholar]
  16. Narasimhaswamy, S.; Wei, Z.; Wang, Y.; Zhang, J.; Hoai, M. Contextual attention for hand detection in the wild. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9567–9576. [Google Scholar]
  17. Shilkrot, R.; Narasimhaswamy, S.; Vazir, S.; Nguyen, M.H. WorkingHands: A Hand-Tool Assembly Dataset for Image Segmentation and Activity Mining. In Proceedings of the British Machine Vision Conference (BMVC), Cardiff, UK, 9–12 September 2019. [Google Scholar]
  18. Bambach, S.; Lee, S.; Crandall, D.J.; Yu, C. Lending a hand: Detecting hands and recognizing activities in complex egocentric interactions. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1949–1957. [Google Scholar]
  19. Brahmbhatt, S.; Tang, C.; Twigg, C.D.; Kemp, C.C.; Hays, J. ContactPose: A dataset of grasps with object contact and hand pose. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 361–378. [Google Scholar]
  20. Garrido-Jurado, S.; Muñoz-Salinas, R.; Madrid-Cuevas, F.J.; Marín-Jiménez, M.J. Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognit. 2014, 47, 2280–2292. [Google Scholar] [CrossRef]
  21. Arun, K.S.; Huang, T.S.; Blostein, S.D. Least-squares fitting of two 3-D point sets. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 1987, PAMI-9, 698–700. [Google Scholar] [CrossRef] [PubMed]
  22. Rother, C.; Kolmogorov, V.; Blake, A. ‘GrabCut’ interactive foreground extraction using iterated graph cuts. ACM Trans. Graph. (TOG) 2004, 23, 309–314. [Google Scholar] [CrossRef]
  23. Dwibedi, D.; Misra, I.; Hebert, M. Cut, paste and learn: Surprisingly easy synthesis for instance detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1301–1310. [Google Scholar]
  24. Liu, H.; Soto, R.A.R.; Xiao, F.; Lee, Y.J. Yolactedge: Real-time instance segmentation on the edge. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 9579–9585. [Google Scholar]
  25. Wu, Y.; Kirillov, A.; Massa, F.; Lo, W.Y.; Girshick, R. Detectron2. 2019. Available online: https://github.com/facebookresearch/detectron2 (accessed on 1 January 2020).
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13 2014. Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  28. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef]
Figure 1. Cameras mounted on top of the TIAGo robot’s head.
Figure 1. Cameras mounted on top of the TIAGo robot’s head.
Sensors 23 07807 g001
Figure 2. Examples of all modalities for one sample in our dataset, including color and depth images from two RGB-D cameras, a segmented point cloud (hand, object, and background), and object shape and pose, in addition to thermal data.
Figure 2. Examples of all modalities for one sample in our dataset, including color and depth images from two RGB-D cameras, a segmented point cloud (hand, object, and background), and object shape and pose, in addition to thermal data.
Sensors 23 07807 g002
Figure 3. Sequence of recording one object for the OHO dataset.
Figure 3. Sequence of recording one object for the OHO dataset.
Sensors 23 07807 g003
Figure 4. Example result of using the keying node in Blender on a reference recording (left) and the corresponding sample with a hand (right).
Figure 4. Example result of using the keying node in Blender on a reference recording (left) and the corresponding sample with a hand (right).
Sensors 23 07807 g004
Figure 5. Pipeline for generation of segmentation labels. Color codes for GrabCut labels are as follows: BGD = blue, PR_BGD = red, PR_FGD = yellow, and FGD = green.
Figure 5. Pipeline for generation of segmentation labels. Color codes for GrabCut labels are as follows: BGD = blue, PR_BGD = red, PR_FGD = yellow, and FGD = green.
Sensors 23 07807 g005
Figure 6. Examples of automatically generated masks for objects (blue) and hands (red) on top of raw RGB images from the dataset.
Figure 6. Examples of automatically generated masks for objects (blue) and hands (red) on top of raw RGB images from the dataset.
Sensors 23 07807 g006
Figure 7. Result of using Gaussian blur on edges vs. simple overlaying of the foreground and background.
Figure 7. Result of using Gaussian blur on edges vs. simple overlaying of the foreground and background.
Sensors 23 07807 g007
Figure 8. Examples of generated images for instance segmentation.
Figure 8. Examples of generated images for instance segmentation.
Sensors 23 07807 g008
Figure 9. Qualitative comparison of the segmentation performance of different models (all with COCO pretraining) in a real-world setting (no stitched images).
Figure 9. Qualitative comparison of the segmentation performance of different models (all with COCO pretraining) in a real-world setting (no stitched images).
Sensors 23 07807 g009
Figure 10. Qualitative results of cross-domain datasets generated by YolactEdge [24] trained on our OHO dataset (hand vs. object). left: applied on WorkingHands [17] (real), middle: applied on WorkingHands [17] (synthetic), right: applied on ContactPose [19].
Figure 10. Qualitative results of cross-domain datasets generated by YolactEdge [24] trained on our OHO dataset (hand vs. object). left: applied on WorkingHands [17] (real), middle: applied on WorkingHands [17] (synthetic), right: applied on ContactPose [19].
Sensors 23 07807 g010
Table 1. Overview of existing datasets, including ours.
Table 1. Overview of existing datasets, including ours.
Dataset#Frames#Objects (#Categories)DepthThermal
HandNet [13]202.9 K0-
Hand-CNN [16]40.5 K0 *--
WorkingHands [17]7.9 K37 (13)-
EgoHands [18]4.8 K0--
Kim et al. [14]401 K0
ContactPose [19]2.5 M25 (25) **-
OHO (ours)5.3 K43 (32)
* contains samples from COCO, and it is not specified how many labels can be reused; ** objects are 3D-printed in blue.
Table 2. Evaluation of instance segmentation on the validation set of our OHO dataset. All models used a ResNet-50 backbone. The models trained on 33 categories were supposed to distinguish different objects, whereas the models trained on 2 categories were simply trained on hand vs. object. Best results by input modality are highlighted.
Table 2. Evaluation of instance segmentation on the validation set of our OHO dataset. All models used a ResNet-50 backbone. The models trained on 33 categories were supposed to distinguish different objects, whereas the models trained on 2 categories were simply trained on hand vs. object. Best results by input modality are highlighted.
#CatsModalityPretraining Bounding BoxSegmentation
AP50:95AP50AP75AP50:95AP50AP75APHand
33RGBImageNetMask R-CNN [1]67.1287.3977.7158.1188.7162.9278.45
YolactEdge [24]62.9388.4273.6961.6991.1567.1977.78
COCOMask R-CNN [1]73.1591.6082.5364.0093.8970.8878.40
PointRend [2]72.4690.1882.5666.3192.9675.1582.57
YolactEdge [24]66.3389.0976.6763.5091.5869.9380.20
RGB +
Thermal
COCOMask R-CNN [1]73.3590.5282.8463.5192.7070.7178.92
PointRend [2]72.7590.6582.5967.0892.1877.6082.73
YolactEdge [24]66.6888.2876.4265.4191.4173.6480.51
2RGB +
Thermal
COCOMask R-CNN [1]79.3396.5787.5970.0997.1079.4378.42
PointRend [2]79.2196.6987.5074.1597.5383.0382.34
YolactEdge [24]69.3194.2180.0870.5995.9476.5280.13
RGB +
Thermal
Mask R-CNN [1]79.6096.6488.2170.3896.8779.5378.67
PointRend [2]79.3496.9287.5575.1097.5385.5582.87
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Stephan, B.; Köhler, M.; Müller, S.; Zhang, Y.; Gross, H.-M.; Notni, G. OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over. Sensors 2023, 23, 7807. https://doi.org/10.3390/s23187807

AMA Style

Stephan B, Köhler M, Müller S, Zhang Y, Gross H-M, Notni G. OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over. Sensors. 2023; 23(18):7807. https://doi.org/10.3390/s23187807

Chicago/Turabian Style

Stephan, Benedict, Mona Köhler, Steffen Müller, Yan Zhang, Horst-Michael Gross, and Gunther Notni. 2023. "OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over" Sensors 23, no. 18: 7807. https://doi.org/10.3390/s23187807

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop