Published July 16, 2020 | Version v1
Dataset Open

Object bounding box annotations for the GTEA Gaze+ dataset

  • 1. DeustoTech Institute, University of Deusto
  • 2. IXA NLP Group, University of the Basque Country
  • 3. Department of Computer Science and Artificial Intelligence, University of the Basque Country

Description

Object bounding box annotations for the GTEA Gaze+ dataset of the works Learning to recognize daily actions using gaze (Fathi et al., 2012) and Delving into Egocentric Actions (Li et al., 2015). The dataset contains folders for each of the subjects, within each of them folders for actions, and, within each action folder, a folder for each video. The video folder has a name composed of <name of the original video>_<start frame>_<end frame>. Within this folder, a json file for some frames can be found.

The json contains two keys: filename and objects. filename refers to the path to the image and objects to a dictionary of objects. The keys of the dictionary are the objects present in the image. Each of the objects values is a list containing bounding box coordinates. Each coordinate list is composed of the ymin, xmin, ymax and ymax values.

Files

bb_coordinates.zip

Files (1.0 MB)

Name Size Download all
md5:3b131042847208b0ffa3f39a4e89d466
1.0 MB Preview Download

Additional details

Related works

References
Conference paper: 10.1007/978-3-642-33718-5_23 (DOI)
Conference paper: 10.1109/CVPR.2015.7298625 (DOI)

References

  • Fathi, A., Li, Y., & Rehg, J. M. (2012, October). Learning to recognize daily actions using gaze. In European Conference on Computer Vision (pp. 314-327). Springer, Berlin, Heidelberg.
  • Li, Y., Ye, Z., & Rehg, J. M. (2015). Delving into egocentric actions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 287-295).