Multi-view Object Tracking Datasets
For all datasets, videos in each sequence are synchronized. The groundtruth trajectories are fully annotated for all the videos in all the sequences using Vatic.
Video sequences and groundtruth annotations for each camera view are available for download.
Each row in the annotation is organized as follows:
track_id xmin ymin xmax ymax frame_number lost occluded generated label 1 track_id. All rows with the same ID belong to the same path. 2 xmin. The top left x-coordinate of the bounding box. 3 ymin. The top left y-coordinate of the bounding box. 4 xmax. The bottom right x-coordinate of the bounding box. 5 ymax. The bottom right y-coordinate of the bounding box. 6 frame_number. The frame that this annotation represents. 7 lost. If 1, the annotation is outside of the view screen. 8 occluded. If 1, the annotation is occluded. 9 generated. If 1, the annotation was automatically interpolated. 10 label. human, car/vehicle, bicycle.
We assign an unique ID for each object, whether it appears once or several times in the scene. Since the ultimate task of multi-view multi-object tracking is to discover the complete 3D trajectory of any targeted individual under a camera network, we believe uniquely assigned ID should be the groundtruth to fully evaluate the trackers, which poses higher requirements than conventional tracking tasks.
PETS 2009 S2/L1 Dataset
This dataset is originally from PETS 2009. We use 1 surveillance cameras (001) and 4 DV cameras (005, 006, 007, 008), with 10 pedestrians are recorded entering, passing through, staying and exiting the pictured area. The video is down-sampled to 720 x 576 and 7 fps. For convenience, we annotate 2D bounding boxes for each view using cross-view consistent ID assignment.
This dataset is originally collected from EPFL. Note the original groundtruth annotation is only provided for the top view (3D ground plane). We further augment annotations for each view.
There are 5 sequences in this dataset: Laboratory, Campus, Terrace, Passageway, and Basketball. In general, each sequence consists of 3-4 different views and films 4-11 pedestrians walking or running around, lasting 3.5-6 minutes. Each view is shot at 25fps and in a relatively low resolution 360x288.
This dataset is collected and annotated by Yuanlu Xu, Hang Qi, Yang Liu, Yansong Tang and Nawin Waree. We are from Center for Vision, Cognition, Learning, and Autonomy (VCLA), University of California, Los Angeles (UCLA).
This dataset is designed for multi-view object tracking algorithms to test on dense foreground, complex scenarios, various object scales.
- dense foreground: around 15-25 objects, frequent conjunctions and occlusions.
- complex scenarios: objects conducting diverse activities, dynamic background, interactions between objects and background.
- various object scales: tracking targets sometimes either too tiny or huge to be accommodated in certain cameras.
There are 4 sequences in this dataset: Garden 1, Garden 2, Auditorium and Parking Lot. Each sequence is shot by 3-4 high-quality DV cameras mounted around 1.5m-2m above ground and each camera covers both overlapping regions and non-overlapping regions with other cameras.
The videos are recorded with frame rate 30fps and duration about 3-4 minutes. The resolution is preserved in 1920X1080, for better precision and richer information.
Please site our paper if you use this dataset.
"Cross-view People Tracking by Scene-centered Spatio-temporal Parsing", Yuanlu Xu, Xiaobai Liu, Lei Qin, Song-Chun Zhu, AAAI Conference on Artificial Intelligence (AAAI), 2017.
"Multi-view People Tracking via Hierarchical Trajectory Composition", Yuanlu Xu, Xiaobai Liu, Yang Liu, Song-Chun Zhu, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
Please contact Yuanlu Xu (firstname.lastname@example.org) if you have any questions.