Skip to content

Latest commit

 

History

History

mot

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 

Multiple Object Tracking Models of BDD100K

The multiple object tracking (MOT) task involves detecting and tracking objects of interest throughout each video sequence.

box_track1

The BDD100K dataset contains MOT annotations for 2K videos (1.4K/200/400 for train/val/test) with 8 categories. Each video is approximately 40 seconds and annotated at 5 fps, resulting in around 200 frames per video. For details about downloading the data and the annotation format for this task, see the official documentation.

Model Zoo

QDTrack

Quasi-Dense Similarity Learning for Multiple Object Tracking [CVPR 2021 Oral]

Authors: Jiangmiao Pang, Linlu Qiu, Xia Li, Haofeng Chen, Qi Li, Trevor Darrell, Fisher Yu

Abstract Similarity learning has been recognized as a crucial step for object tracking. However, existing multiple object tracking methods only use sparse ground truth matching as the training objective, while ignoring the majority of the informative regions on the images. In this paper, we present Quasi-Dense Similarity Learning, which densely samples hundreds of region proposals on a pair of images for contrastive learning. We can naturally combine this similarity learning with existing detection methods to build Quasi-Dense Tracking (QDTrack) without turning to displacement regression or motion priors. We also find that the resulting distinctive feature space admits a simple nearest neighbor search at the inference time. Despite its simplicity, QDTrack outperforms all existing methods on MOT, BDD100K, Waymo, and TAO tracking benchmarks. It achieves 68.7 MOTA at 20.3 FPS on MOT17 without using external training data. Compared to methods with similar detectors, it boosts almost 10 points of MOTA and significantly decreases the number of ID switches on BDD100K and Waymo datasets.

Results

Detector mMOTA-val mIDF1-val ID Sw.-val Scores-val mMOTA-test mIDF1-test ID Sw.-test Scores-test Config Weights Preds Visuals
ResNet-50 36.6 51.6 6193 scores 35.7 52.3 10822 scores config model | MD5 preds visuals

[Code] [Usage Instructions]

Usage

Model Inference

For model inference, please refer to the usage instructions of the corresponding model.

Output Evaluation

Validation Set

To evaluate the MOT performance on the BDD100K validation set, you can follow the official evaluation scripts provided by BDD100K:

python -m bdd100k.eval.run -t box_track \
    -g ../data/bdd100k/labels/box_track_20/${SET_NAME} \
    -r ${OUTPUT_FILE} \
    [--out-file ${RESULTS_FILE}] [--nproc ${NUM_PROCESS}]

Test Set

You can obtain the performance on the BDD100K test set by submitting your model predictions to our evaluation server hosted on EvalAI.

Output Visualization

For visualization, you can use the visualization tool provided by Scalabel.

Below is an example:

import os
import numpy as np
from PIL import Image
from scalabel.label.io import load
from scalabel.vis.label import LabelViewer

# load prediction frames
frames = load('$OUTPUT_FILE').frames

viewer = LabelViewer()
for frame in frames:
    img = np.array(Image.open(os.path.join('$IMG_DIR', frame.name)))
    viewer.draw(img, frame)
    viewer.save(os.path.join('$VIS_DIR', frame.videoName, frame.name))

Contribution

You can include your models in this repo as well! Please follow the contribution instructions.