A General Framework for Tracking Multiple People from a Moving Camera
We present a general framework for tracking multiple, possibly interacting, people from a mobile vision platform. In order to determine all of the trajectories robustly and in a 3D coordinate system, we estimate both the camera's ego-motion and the people's paths within a single coherent framework. The tracking problem is framed as .nding the MAP solution of a posterior probability, and is solved using the Reversible Jump Markov Chain Monte Carlo Particle Filtering method. We evaluate our system on challenging datasets taken from moving cameras, including an outdoor street scene video dataset, as well as an indoor RGB-D dataset collected in an of.ce. Experimental evidence shows that the proposed method can robustly estimate a camera's motion from dynamic scenes and stably track people who are moving independently or interacting.
W. Choi, C. Pantofaru, S. Savarese. "A General Framework for Tracking Multiple People from a Moving Camera" in Pattern Analysis and Machine Intelligence (PAMI), 2013. [preprint][bibtex]
W. Choi, C. Pantofaru, S. Savarese. "Detecting and Tracking People using an RGB-D Camera via Multiple Detector Fusion" in CORP (in conjunction with ICCV), 2011. [pdf][bibtex]
W. Choi and S. Savarese. "Multiple Target Tracking in World Coordinate with Single, Minimally Calibrated Camera" in ECCV , 2010. [pdf][bibtex]
The previous project webpage can be found here. This project is sponsored by NSF AEGER (award #1052762) and Toyota. This work is in collaboration with Dr. Caroline Pantofaru (@Willow Garage).
[NEW] Stanford CVG Lab is now hosting the dataset. (07/25/2014).
The source code is available below. (01/21/2013).
The datasets are uploaded (01/18/2013).
This page is created. (01/18/2013).
The source code (monocular camera version) used in the PAMI13 publication is available below. To download the code, launch an SVN client and checkout the code from the following repository (e.g. svn co link):
- Version 0.5 release(August 7th, 2012): C++ code with a demo at http://mtt-umich.googlecode.com/svn/tags/release-0.5
The source code for PR2 platform will be available soon.
The code is released under BSD license.
The recommended method for downloading the data is to use wget
1) PNG + DAT tar balls
There are two RGB+D datasets we used in the PAMI paper. To download all tarballs of PR2 dataset, run
wget -c -i http://cvgl.stanford.edu/data2/pr2dataset/pr2pngdat_all.txt
To download all tarballs of Umich (static) dataset, run
wget -c -i http://cvgl.stanford.edu/data2/pr2dataset/umpngdat_all.txt
2) ROS Bag files
You can also download rosbag
files that are compatible with ROS fuerte. To download all bag files of PR2 dataset, run
wget -c -i http://cvgl.stanford.edu/data2/pr2dataset/pr2bag_all.txt
To download all bag files of Umich (static) dataset, run
wget -c -i http://cvgl.stanford.edu/data2/pr2dataset/umbag_all.txt
The annotations of the PR2 dataset can be downloaded here
and the annotations of the UM dataset can be downloaded here
4) Matlab tools
You can download necessary matlab tools for the dataset (e.g. dat file reader, annotation reader, visualization samples, etc) here
The data is released under Creative Common license.
Last updated on 07/25/2014