Published May 27, 2024 | Version v1
Dataset Open

IAS-Lab Collaborative Draping HAR dataset

  • 1. University of Padova
  • 2. Politecnico di Torino

Description

Description

This dataset contains movement data for several subjects performing actions related to human-robot collaboration in an industrial carbon fiber draping process, such as draping and collaborative transport of carbon fiber plies. The collected dataset has been used to train and evaluate skeleton-based Human Action Recognition (HAR) models developed to provide a simple and intuitive way for the human operator to interact with the robot, such as signaling start and stop of the process or requesting robot’s assistance with specific tasks (e.g., inspection of specific parts).


Actions of interest

The dataset includes gestures designed to provide a simple and intuitive way for the operator to interact with the robot (e.g., “OK/NEXT” and “POINT” actions), short duration actions related to the beginning and end of collaborative transport operations (e.g., “PICK” and “PLACE” actions), and long duration actions related to the draping process (e.g., “TRANSPORT” and “DRAPING” actions) or general movements of the operator (e.g., “REST” and “WALK”). 
The dataset also includes an additional unknown class (“UNKWN”) which represents various random movements that the operator might make during the collaborative process but that do not correspond to any of the actions of interest.

Action ID Action Name Description
A001 OK/NEXT Raise one arm to signal the robot to continue the draping process
A002 POINT Point at a desired location with a straight right arm to trigger inspection
A003 PICK Raise the ply to trigger the collaborative transport
A004 PLACE Place the ply on the mold
A005 TRANSPORT Collaborative transportation
A006 DRAPE Manual draping of a ply
A007 REST Resting position, mainly waiting for the robot to complete its task
A008 WALK Walking across the workcell
A009 UNKWN Operator movements not related to the draping process

 

Dataset

Data has been collected from 7 participants, 2 female and 5 male, average age 27 (SD=3.0). Each participant performed 6 repetitions of each of the 8 actions considered for a collaborative draping process, and 18 repetitions of random movements for the unknow class. This results in a collected dataset containing 462 trimmed samples, where each sample is a sequence of 3D skeletons lasting approximately 3 seconds, containing only one action being performed. For all samples, 3D skeletons were obtained by means of the camera network installed in the laboratory, providing: (i) 3D skeletons from each camera in the network and (ii) 3D skeletons obtained by fusing the detections from each camera with a tracking algorithm; the output of the tracking algorithm provides a 3D skeleton representation robust to occlusions. All the 3D skeletons acquired are composed of 15 joints, with 3D coordinates expressed with respect to the camera network reference frame.

Skeleton data for all sequences are provided in the `skeleton_data` folder. The dataset consists of a folder for each action of interest, with a subfolder for each participant and an individual text file for each action repetition performed by the participant. The naming convention for these files follows a pattern of the type “AxxxPyyyRzzzCwww.skeleton”, where "Axxx" represents the action id, "Pyyy" represents the id assigned to the participant, "Rzzz" represents the repeat number, and "Cwww" represents the id of the camera from which the 3D skeleton is derived; a network of 4 cameras was used to acquire the data, so “C001” denotes the first camera, “C002” the second, and so on, while “C000” represents the 3D skeletons obtained by merging all views.
Each of these files is provided as a “.skeleton” text files, similar to popular human action recognition dataset (e.g., NTU RGB+D and NTU RGB+D 120 action recognition datasets). In particular, each file includes a sequence of 3D skeletons with 25 joints following the OpenPose convention for joint numbering, but only the first 15 joints contain valid values since face keypoints were not considered during the acquisitions.

Example python code to read/write and visualize the skeleton data is also provided in the `scripts` folder:
```python
python3 plot_sequence.py --data_dir ../skeleton_data
```

 

References

  1. Allegro, D., Terreran, M., & Ghidoni, S. (2023). METRIC—Multi-Eye to Robot Indoor Calibration Dataset. Information, 14(6), 314. https://doi.org/10.3390/info14060314
  2. Terreran, M., Barcellona, L., & Ghidoni, S. (2023). A general skeleton-based action and gesture recognition framework for human–robot collaboration. Robotics and Autonomous Systems, 170, 104523. https://doi.org/10.1016/j.robot.2023.104523
  3. Terreran, M., Lazzaretto, M., & Ghidoni, S. (2022, June). Skeleton-based action and gesture recognition for human-robot collaboration. In International Conference on Intelligent Autonomous Systems (pp. 29-45). Cham: Springer Nature Switzerland.
  4. Carraro, M., Munaro, M., Burke, J., & Menegatti, E. (2019). Real-time marker-less multi-person 3D pose estimation in RGB-depth camera networks. In Intelligent Autonomous Systems 15: Proceedings of the 15th International Conference IAS-15 (pp. 534-545). Springer International Publishing.

Files

iaslab_char_dataset.zip

Files (55.5 MB)

Name Size Download all
md5:68b6b466c4641d135c9c107dc5e96a18
55.5 MB Preview Download
md5:ff5eda09e4e9fa82126e87cab4f3adec
5.5 kB Preview Download

Additional details

Funding

European Commission
DrapeBot – Collaborative draping of carbon fiber parts 101006732

References

  • Allegro, D., Terreran, M., & Ghidoni, S. (2023). METRIC—Multi-Eye to Robot Indoor Calibration Dataset. Information, 14(6), 314. https://doi.org/10.3390/info14060314
  • Terreran, M., Barcellona, L., & Ghidoni, S. (2023). A general skeleton-based action and gesture recognition framework for human–robot collaboration. Robotics and Autonomous Systems, 170, 104523. https://doi.org/10.1016/j.robot.2023.104523
  • Terreran, M., Lazzaretto, M., & Ghidoni, S. (2022, June). Skeleton-based action and gesture recognition for human-robot collaboration. In International Conference on Intelligent Autonomous Systems (pp. 29-45). Cham: Springer Nature Switzerland.
  • Carraro, M., Munaro, M., Burke, J., & Menegatti, E. (2019). Real-time marker-less multi-person 3D pose estimation in RGB-depth camera networks. In Intelligent Autonomous Systems 15: Proceedings of the 15th International Conference IAS-15 (pp. 534-545). Springer International Publishing.