Published December 15, 2023 | Version v2
Dataset Open

FOCAL Dataset: Ford-OLIVES Collaboration on Active Learning

Description

In this dataset, we introduce the FOCAL (Ford-OLIVES Collaboration on Active Learning) dataset which enables the study of the impact of annotation-cost within a video active learning setting. Annotation-cost refers to the time it takes an annotator to label and quality-assure a given video sequence. A practical motivation for active learning research is to minimize annotation-cost by selectively labeling informative samples that will maximize performance within a given budget constraint. However, previous work in video active learning lacks real-time annotation labels for accurately assessing cost minimization and instead operates under the assumption that annotation-cost scales linearly with the amount of data to annotate. This assumption does not take into account a variety of real-world confounding factors that contribute to  a nonlinear cost such as the effect of an assistive labeling tool and the variety of interactions within a scene such as occluded objects, weather, and motion of objects. FOCAL addresses this discrepancy by providing real annotation-cost labels for 126 video sequences across 69 unique city scenes with a variety of weather, lighting, and seasonal conditions. These videos have a wide range of interactions that are at the intersection of infrastructure-assisted autonomy and autonomous vehicle communities. We show through a statistical analysis of the FOCAL dataset that cost is more correlated with a variety of factors beyond just the length of a video sequence. We also introduce a set of conformal active learning algorithms that take advantage of the sequential structure of video data in order to achieve a better trade-off between annotation-cost and performance while also reducing floating point operations (FLOPS) overhead by at least 77.67%. We show how these approaches better reflect how annotations on videos are done in practice through a sequence selection framework. We further demonstrate the advantage of these approaches by introducing two performance-cost metrics and show that the best conformal active learning method is cheaper than the best traditional active learning method by 113 hours.  

This work took place at the OLIVES Lab @ Georgia Tech. 

The codebase associated with this work can be found at this Github.

Please refer to our lab-wide github for more information regarding the code associated with our other papers.

Notes

  • The data presented here is a processed version of the full FOCAL dataset in YOLO object detection format.

 

  • The associated github for this dataset can be located at: https://github.com/olivesgatech/FOCAL_Dataset.

 

  • The images and labels in Yolo format are located in the file focal_yolo.tar.gz.

 

  • Information regarding the annotation cost of every sequence in the dataset is located in the associated excel file.

 

  • Every image has the format:   nameofsequence _ nameofimage.jpg

 

  • The folder called split ids contains files that delineate which sequences belonged to the train, test, and val sets throughout the creation of the paper.

Files

FOCAL.zip

Files (45.2 GB)

Name Size Download all
md5:8009e9a4b29b4214e1b3b4b036420569
45.2 GB Preview Download

Additional details

References

  • K. Kokilepersaud*, Y. Logan*, R. Benkert, C. Zhou, M. Prabhushankar, G. AlRegib, E. Corona, K. Singh, A. Parchami, "FOCAL: A Cost-Aware, Video Dataset for Active Learning," in IEEE Conference on Big Data 2023, Sorento, Italy, Dec. 15-18, 2023.