DEEP_TAMA: Online Multiple Pedestrian Tracking with Deep Temporal Appearance Matching Association


Video not available.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Rendering of new sequences is currently deactivated due to heavy load.

Benchmark:

MOT17 |

Short name:

DEEP_TAMA

Detector:

Public

Description:

In online multi-target tracking, it is of great importance to model appearance and geometric similarity between pedestrians which have been tracked and appeared in a new frame. The dimension of the inherent feature vector in the appearance model is higher than that in the geometric model, which causes many problems in general. However, the recent success of deep learning-based methods makes it possible to handle high dimensional appearance information successfully. Among many deep networks, the Siamese network with triplet loss is popularly
adopted as an appearance feature extractor. Since the Siamese network can extract features of each input independently, it is possible to update and maintain target-specific features. However, it is not suitable for multi-target settings that require comparison with other inputs. In this paper, to address this issue, we propose a novel track appearance model based on the joint-inference network. The proposed method enables a comparison of two inputs to be used for adaptive appearance modeling, and contributes to disambiguating the process of target-observation matching and consolidating identity consistency. Diverse experimental results support the effectiveness of our method. Our work has been awarded as a 3rd-highest tracker on MOTChallenge19, held in CVPR2019.1 The code is available on https://github.com/yyc9268/Deep-TAMA.

Reference:

Y. Yoon, D. Kim, Y. Song, K. Yoon, M. Jeon. Online Multiple Pedestrians Tracking using Deep Temporal Appearance Matching Association. In Information Sciences, 2020.

Last submitted:

February 07, 2019 (5 years ago)

Published:

February 07, 2019 at 02:56:08 CET

Submissions:

1

Open source:

No

Hardware:

3.7 GHZ, 1 Core(no GPU)

Runtime:

1.5 Hz

Benchmark performance:

Sequence MOTA IDF1 HOTA MT ML FP FN Rcll Prcn AssA DetA AssRe AssPr DetRe DetPr LocA FAF ID Sw. Frag
MOT1750.353.542.0453 (19.2)883 (37.5)25,479252,99655.292.443.341.046.973.143.873.479.71.42,192 (39.7)3,978 (72.1)

Detailed performance:

Sequence MOTA IDF1 HOTA MT ML FP FN Rcll Prcn AssA DetA AssRe AssPr DetRe DetPr LocA FAF ID Sw. Frag
MOT17-01-DPM38.743.932.94111123,82340.795.937.129.339.968.930.371.376.60.22048
MOT17-01-FRCNN28.340.936.5871,5493,05152.768.739.134.542.874.442.955.978.43.42645
MOT17-01-SDP44.349.539.9941,2022,35463.577.339.740.542.372.149.159.878.22.73565
MOT17-03-DPM54.253.540.329292,33145,40756.696.238.742.140.975.644.275.079.21.6248520
MOT17-03-FRCNN59.657.845.040252,31139,82662.096.643.447.146.375.049.577.280.51.5186329
MOT17-03-SDP75.970.655.576141,48623,51577.598.252.558.956.474.762.078.681.21.0183488
MOT17-06-DPM43.255.141.0341097585,85950.388.747.735.754.066.438.467.876.60.682164
MOT17-06-FRCNN47.354.943.252691,3694,73259.883.743.543.353.561.348.668.179.01.1115208
MOT17-06-SDP52.258.845.374691,3144,20664.385.246.244.856.861.450.667.078.41.1108191
MOT17-07-DPM38.543.632.15265389,77742.193.033.630.935.669.432.371.377.01.179166
MOT17-07-FRCNN31.739.431.33152,0119,37344.578.932.331.535.563.235.362.576.74.0150290
MOT17-07-SDP46.449.237.614181,0977,85953.589.236.738.740.367.042.070.178.72.296192
MOT17-08-DPM26.530.426.093947514,97829.192.829.223.231.175.224.076.782.30.87071
MOT17-08-FRCNN23.130.327.373989215,28027.786.834.621.737.077.122.971.781.81.46583
MOT17-08-SDP32.636.632.4113451113,62635.593.637.828.039.976.429.176.981.80.8103136
MOT17-12-DPM41.050.238.819452634,81644.493.645.333.347.776.335.073.879.80.33748
MOT17-12-FRCNN36.550.039.015456354,83844.285.846.333.150.175.936.069.980.40.72837
MOT17-12-SDP40.854.443.421428264,27150.784.251.236.955.077.441.268.481.30.93545
MOT17-14-DPM24.935.825.9610448213,33027.991.432.420.735.167.221.570.778.10.664107
MOT17-14-FRCNN16.934.327.37743,30811,80336.166.930.624.935.756.228.853.374.04.4243410
MOT17-14-SDP32.444.932.310652,00910,27244.480.334.630.439.163.534.061.576.22.7219335

Raw data: