Published April 7, 2017 | Version v1
Dataset Open

CAD 120 affordance dataset

  • 1. University of Bonn
  • 2. Carl Zeiss AG

Description

% ==============================================================================
% CAD 120 Affordance Dataset
% Version 1.0
% ------------------------------------------------------------------------------
% If you use the dataset please cite:
%
% Johann Sawatzky, Abhilash Srikantha, Juergen Gall.
% Weakly Supervised Affordance Detection.
% IEEE Conference on Computer Vision and Pattern Recognition (CVPR'17)
%
% and
%
% H. S. Koppula and A. Saxena.
% Physically grounded spatio-temporal object affordances.
% European Conference on Computer Vision (ECCV'14)
%
% Any bugs or questions, please email sawatzky AT iai DOT uni-bonn DOT de.
% ==============================================================================

This is the CAD 120 Affordance Segmentation Dataset based on the Cornell Activity
Dataset CAD 120 (see http://pr.cs.cornell.edu/humanactivities/data.php).

Content

frames/*.png:
RGB frames selected from Cornell Activity Dataset. To find out the location of the frame
in the original videos, see video_info.txt.

object_crop_images/*.png
image crops taken from the selected frames and resized to 321*321. Each crop is a padded
bounding box of an object the human interacts with in the video. Due to the padding,
the crops may contain background and other objects.
In each selected frame, each bounding box was processed. The bounding boxes are already
given in the Cornell Activity Dataset.
The 5-digit number gives the frame number, the second number gives the bounding box number
within the frame.

segmentation_mat/*.mat
321*321*6 segmentation masks for the image crops. Each channel corresponds to an
affordance (openabe, cuttable, pourable, containable, supportable, holdable, in this order).
All pixels belonging to a particular affordance are labeled 1 in the respective channel,
otherwise 0.  

segmentation_png/*.png
321*321 png images, each containing the binary mask for one of the affordances.

lists/*.txt
Lists containing the train and test sets for two splits. The actor split ensures that
train and test images stem from different videos with different actors while the object split ensures
that train and test data have no (central) object classes in common.
The train sets are additionally subdivided into 3 subsets A,B and C. For the actor split,
the subsets stem from different videos. For the object split, each subset contains
every third crop of the train set.

crop_coordinate_info.txt
Maps image crops to their coordinates in the frames.

hpose_info.txt
Maps frames to 2d human pose coordinates. Hand annotated by us.

object_info.txt
Maps image crops to the (central) object it contains.

visible_affordance_info.txt
Maps image crops to affordances visible in this crop

 

%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%55
The crops contain the following object classes:
1.table
2.kettle
3.plate
4.bottle
5.thermal cup
6.knife
7.medicine box
8.can
9.microwave
10.paper box
11.bowl
12.mug

Affordances in our set:
1.openable
2.cuttable
3.pourable
4.containable
5.supportable
6.holdable

Note that our object affordance labeling differs from the Cornell Activity Dataset:
E.g. the cap of a pizza box is considered to be supportable.

 

Notes

Acknowledgments. The work has been financially sup- ported by the DFG projects GA 1927/5-1 (DFG Research Unit FOR 2535 Anticipating Human Behavior) and GA 1927/2-2 (DFG Research Unit FOR 1505 Mapping on De- mand).

Files

Files (2.6 GB)

Name Size Download all
md5:be832f74fa4a3db7644f9e47175d4bc3
2.6 GB Download

Additional details

References

  • Sawatzky, J., Srikantha, A., Gall, J.: Weakly supervised affordance detection. CVPR (2017)