No Arabic abstract
This paper proposes an online visual multi-object tracking algorithm using a top-down Bayesian formulation that seamlessly integrates state estimation, track management, clutter rejection, occlusion and mis-detection handling into a single recursion. This is achieved by modeling the multi-object state as labeled random finite set and using the Bayes recursion to propagate the multi-object filtering density forward in time. The proposed filter updates tracks with detections but switches to image data when mis-detection occurs, thereby exploiting the efficiency of detection data and the accuracy of image data. Furthermore the labeled random finite set framework enables the incorporation of prior knowledge that mis-detections of long tracks which occur in the middle of the scene are likely to be due to occlusions. Such prior knowledge can be exploited to improve occlusion handling, especially long occlusions that can lead to premature track termination in on-line multi-object tracking. Tracking performance are compared to state-of-the-art algorithms on well-known benchmark video datasets.
This paper provides a scalable, multi-sensor measurement adaptive track initiation technique for labeled random finite set filters. A naive construction of the multi-sensor measurement adaptive birth set leads to an exponential number of newborn components in the number of sensors. A truncation criterion is established for a multi-sensor measurement-generated labeled multi-Bernoulli random finite set that provably minimizes the L1-truncation error in the generalized labeled multi-Bernoulli posterior distribution. This criterion is used to construct a Gibbs sampler that produces a truncated measurement-generated labeled multi-Bernoulli birth distribution with quadratic complexity in the number of sensors. A closed form solution of the conditional sampling distribution assuming linear (or linearized) Gaussian likelihoods is provided, alongside an approximate solution using Monte Carlo importance sampling. Multiple simulation results are provided to verify the efficacy of the truncation criterion, as well as the reduction in complexity.
This paper proposes an on-line multiple object tracking algorithm that can operate in unknown background. In a majority of multiple object tracking applications, model parameters for background processes such as clutter and detection are unknown and vary with time, hence the ability of the algorithm to adaptively learn the these parameters is essential in practice. In this work, we detail how the Generalized Labeled Multi Bernouli (GLMB) filter a tractable and provably Bayes optimal multi-object tracker can be tailored to learn clutter and detection parameters on the fly while tracking. Provided that these background model parameters do not fluctuate rapidly compared to the data rate, the proposed algorithm can adapt to the unknown background yielding better tracking performance.
Determining the trajectories of cells and their lineages or ancestries in live-cell experiments are fundamental to the understanding of how cells behave and divide. This paper proposes novel online algorithms for jointly tracking and resolving lineages of an unknown and time-varying number of cells from time-lapse video data. Our approach involves modeling the cell ensemble as a labeled random finite set with labels representing cell identities and lineages. A spawning model is developed to take into account cell lineages and changes in cell appearance prior to division. We then derive analytic filters to propagate multi-object distributions that contain information on the current cell ensemble including their lineages. We also develop numerical implementations of the resulting multi-object filters. Experiments using simulation, synthetic cell migration video, and real time-lapse sequence, are presented to demonstrate the capability of the solutions.
Online updating of the object model via samples from historical frames is of great importance for accurate visual object tracking. Recent works mainly focus on constructing effective and efficient updating methods while neglecting the training samples for learning discriminative object models, which is also a key part of a learning problem. In this paper, we propose the DeepMix that takes historical samples embeddings as input and generates augmented embeddings online, enhancing the state-of-the-art online learning methods for visual object tracking. More specifically, we first propose the online data augmentation for tracking that online augments the historical samples through object-aware filtering. Then, we propose MixNet which is an offline trained network for performing online data augmentation within one-step, enhancing the tracking accuracy while preserving high speeds of the state-of-the-art online learning methods. The extensive experiments on three different tracking frameworks, i.e., DiMP, DSiam, and SiamRPN++, and three large-scale and challenging datasets, ie, OTB-2015, LaSOT, and VOT, demonstrate the effectiveness and advantages of the proposed method.
The main challenge of online multi-object tracking is to reliably associate object trajectories with detections in each video frame based on their tracking history. In this work, we propose the Recurrent Autoregressive Network (RAN), a temporal generative modeling framework to characterize the appearance and motion dynamics of multiple objects over time. The RAN couples an external memory and an internal memory. The external memory explicitly stores previous inputs of each trajectory in a time window, while the internal memory learns to summarize long-term tracking history and associate detections by processing the external memory. We conduct experiments on the MOT 2015 and 2016 datasets to demonstrate the robustness of our tracking method in highly crowded and occluded scenes. Our method achieves top-ranked results on the two benchmarks.