Dataset Open Access

CAD 120 affordance dataset

Sawatzky, Johann; Srikantha, Abhilash; Gall, Juergen


Dublin Core Export

<?xml version='1.0' encoding='utf-8'?>
<oai_dc:dc xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
  <dc:creator>Sawatzky, Johann</dc:creator>
  <dc:creator>Srikantha, Abhilash</dc:creator>
  <dc:creator>Gall, Juergen</dc:creator>
  <dc:date>2017-04-07</dc:date>
  <dc:description>% ==============================================================================
% CAD 120 Affordance Dataset
% Version 1.0
% ------------------------------------------------------------------------------
% If you use the dataset please cite:
%
% Johann Sawatzky, Abhilash Srikantha, Juergen Gall.
% Weakly Supervised Affordance Detection.
% IEEE Conference on Computer Vision and Pattern Recognition (CVPR'17)
%
% and
%
% H. S. Koppula and A. Saxena.
% Physically grounded spatio-temporal object affordances.
% European Conference on Computer Vision (ECCV'14)
%
% Any bugs or questions, please email sawatzky AT iai DOT uni-bonn DOT de.
% ==============================================================================

This is the CAD 120 Affordance Segmentation Dataset based on the Cornell Activity
Dataset CAD 120 (see http://pr.cs.cornell.edu/humanactivities/data.php).

Content

frames/*.png:
RGB frames selected from Cornell Activity Dataset. To find out the location of the frame
in the original videos, see video_info.txt.

object_crop_images/*.png
image crops taken from the selected frames and resized to 321*321. Each crop is a padded
bounding box of an object the human interacts with in the video. Due to the padding,
the crops may contain background and other objects.
In each selected frame, each bounding box was processed. The bounding boxes are already
given in the Cornell Activity Dataset.
The 5-digit number gives the frame number, the second number gives the bounding box number
within the frame.

segmentation_mat/*.mat
321*321*6 segmentation masks for the image crops. Each channel corresponds to an
affordance (openabe, cuttable, pourable, containable, supportable, holdable, in this order).
All pixels belonging to a particular affordance are labeled 1 in the respective channel,
otherwise 0.  

segmentation_png/*.png
321*321 png images, each containing the binary mask for one of the affordances.

lists/*.txt
Lists containing the train and test sets for two splits. The actor split ensures that
train and test images stem from different videos with different actors while the object split ensures
that train and test data have no (central) object classes in common.
The train sets are additionally subdivided into 3 subsets A,B and C. For the actor split,
the subsets stem from different videos. For the object split, each subset contains
every third crop of the train set.

crop_coordinate_info.txt
Maps image crops to their coordinates in the frames.

hpose_info.txt
Maps frames to 2d human pose coordinates. Hand annotated by us.

object_info.txt
Maps image crops to the (central) object it contains.

visible_affordance_info.txt
Maps image crops to affordances visible in this crop

 

%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%55
The crops contain the following object classes:
1.table
2.kettle
3.plate
4.bottle
5.thermal cup
6.knife
7.medicine box
8.can
9.microwave
10.paper box
11.bowl
12.mug

Affordances in our set:
1.openable
2.cuttable
3.pourable
4.containable
5.supportable
6.holdable

Note that our object affordance labeling differs from the Cornell Activity Dataset:
E.g. the cap of a pizza box is considered to be supportable.

 </dc:description>
  <dc:identifier>https://zenodo.org/record/495570</dc:identifier>
  <dc:identifier>10.5281/zenodo.495570</dc:identifier>
  <dc:identifier>oai:zenodo.org:495570</dc:identifier>
  <dc:relation>url:https://pages.iai.uni-bonn.de/gall_juergen/download/jgall_affordancedetection_cvpr17.pdf</dc:relation>
  <dc:relation>url:https://github.com/ykztawas/Weakly-Supervised-Affordance-Detection</dc:relation>
  <dc:rights>info:eu-repo/semantics/openAccess</dc:rights>
  <dc:rights>https://creativecommons.org/licenses/by/4.0/</dc:rights>
  <dc:subject>computer vision</dc:subject>
  <dc:subject>affordances</dc:subject>
  <dc:subject>attributes</dc:subject>
  <dc:subject>semantic image segmentation</dc:subject>
  <dc:subject>robotics</dc:subject>
  <dc:subject>weakly supervised learning</dc:subject>
  <dc:subject>convolutional neural network</dc:subject>
  <dc:subject>anticipating human behavior</dc:subject>
  <dc:subject>mapping on demand</dc:subject>
  <dc:title>CAD 120 affordance dataset</dc:title>
  <dc:type>info:eu-repo/semantics/other</dc:type>
  <dc:type>dataset</dc:type>
</oai_dc:dc>

Share

Cite as