Published March 9, 2022 | Version v1
Dataset Open

ObChange Dataset

Creators

  • 1. TU Wien, Vienna, Austria

Description

This dataset that can be used to evaluate methods, which are able to detect changed objects when comparing two recordings of the same environment at different time instances. Based on the labeled ground truth objects, it is possible to differentiate between static, moved, removed and novel objects.

Dataset Description

The dataset was recorded with an Asus Xtion PRO Live mounted on the HSR robot. We provide scenes from five different rooms or parts of rooms, namely a big room, a small room, a living area, a kitchen counter and an office desk. Each room is visited by the robot at least five times while between each run a subset of objects from the YCB Object and Model Set (YCB)[1] is re-arranged in the room. In total we generated 26 recordings. For each recording between three and 17 objects are placed (219 in total). Furthermore, furniture and permanent background objects are slightly rearranged. These changes are not labeled because for most service robot tasks, this is not relevant.

Assuming most objects are placed on horizontal surfaces, we extracted planes in each room in a pre-processing step (excluding the floor). For each surface, all frames from the recording where it is visible are extracted and used as the input for ElasticFusion[2]. This results in a total of 34 reconstructed surfaces.

We provide pointwise annotation of the YCB objects for each surface reconstruction from each recording.

Images of exemplary surface reconstructions can be found here: https://www.acin.tuwien.ac.at/vision-for-robotics/software-tools/obchange/

 

Dataset Structure

The file structure of ObChange.zip is the following:

Room

   - scene2  

      - planes    

         - 0      

            - merged_plane_clouds_ds002.pcd      

            - merged_plane_clouds_ds002.anno      

            - merged_plane_clouds_ds002_GT.anno    

         - 1      

            - merged_plane_clouds_ds002.pcd      

            - merged_plane_clouds_ds002.anno      

            - merged_plane_clouds_ds002_GT.anno    

         - ...  

      table.txt

   - scene3

The pcd-file contains the reconstruction of the surface. The merged_plane_clouds_ds002.anno lists the YCB objects visible in the reconstruction and merged_plane_clouds_ds002_GT.anno contains the point indices of the reconstruction corresponding to the YCB objects together with the corresponding object name. The last element for each object is a bool value indicating if the object is on the floor (and was reconstructed by chance). The table.txt lists for each detected plane the centroid, height, convex hull points and plane coefficients.

We provide the original input data for each room. The zip-files contain the rosbag file for each recording. Each rosbag contains the tf-tree and the RGB and depth stream, as well as the camera intrinsic. Additionally, the semantically annotated Voxblox[3] reconstruction created with SparseConvNet[4] is provided for each recording. 

 

You may also be interested in Object Change Detection Dataset of Indoor Environments. It uses the same input data, but the ground truth annotation is based on a full room reconstruction instead of individual planes.

 

Acknowledgements

The research leading to these results has received funding from the Austrian Science Fund (FWF) under grant agreement Nos. I3969-N30 (InDex), I3967-N30 (BURG) and from the Austrian Research Promotion Agency (FFG) under grant agreement 879878 (K4R).

References

[1] B. Calli, A. Singh, J. Bruce, A. Walsman, K. Konolige, S. Srinivasa, P. Abbeel, A. M. Dollar, Yale-CMU-Berkeley dataset for robotic manipulation research, The International Journal of Robotics Research, vol. 36, Issue 3, pp. 261 – 268, April 2017.

[2] T. Whelan, S. Leutenegger, R. Salas-Moreno, B. Glocker, A. Davison, ElasticFusion: Dense SLAM without a pose graph, Proceedings of Robotics: Science and Systems, July 2015.

[3] H. Oleynikova, Z. Taylor, M. Fehr, R. Siegwart, J. Nieto, Juan, Voxblox: Incremental 3D Euclidean Signed Distance Fields for On-Board MAV Planning, in Proceedings of IEEE International Conference on Intelligent Robots and Systems (IROS), pp. 1366-1373, 2017.

[4] B. Graham, M. Engelcke, L. van der Maaten, 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks, Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 9224 – 9232, 2018.

Files

BigRoom.zip

Files (109.0 GiB)

Name Size
md5:48f6b80e135301f45bd4c84a9f17955c
55.1 GiB Preview Download
md5:d9e51e01f1fcfc850f9824bf7fa15aa2
6.1 GiB Preview Download
md5:4765dc17ba21c95b527b15d8d592240f
12.0 GiB Preview Download
md5:76b5f6e899fe30d2d7830507e6d05410
1005.8 MiB Preview Download
md5:27696dd4685a9c42d3ffaee0f8de1101
7.8 GiB Preview Download
md5:a2d7cba9a8c75fff56a62bbfb12273be
27.0 GiB Preview Download

Additional details