This project implements the paper "Cross-View Tracking for Multi-Human 3D Pose Estimation at Over 100 FPS" (CVPR 2020).
This implementation focuses on multi-person 3D pose estimation and tracking from multiple camera views. The key features include:
- Cross-view tracking for robust 3D pose estimation
- Real-time performance optimization
- Support for multiple people tracking
- Handling of occlusions and view changes
- Python 3.7+
- PyTorch
- NumPy
- OpenCV
- Matplotlib
- tqdm
You can download the Campus dataset from:
After downloading:
- Extract the
Campus_Seq1.tar.gz
file - Place the extracted
Campus_Seq1
folder in the project root directory - The dataset should follow the structure shown above
Note: The Campus dataset contains synchronized videos from three calibrated cameras, along with ground truth 3D pose annotations.
The code supports datasets in the following structure:
The main components include:
pose_matcher.py
: Cross-view pose matchingpose_estimator.py
: 3D pose estimation from 2D detectionsdisplay.py
: Visualization utilities
The system utilizes multiple synchronized cameras to capture different viewpoints of the scene. This multi-view setup enables:
- Comprehensive coverage of the tracking area
- Better handling of occlusions
- More accurate 3D pose reconstruction
The visualization demonstrates:
- Accurate reconstruction of human poses in 3D space
- Real-time tracking of multiple subjects
- Robust pose estimation across different views
- Multi-view synchronized capture
- Real-time 3D pose estimation
- Robust person tracking
- Occlusion handling
- High-performance processing
This implementation is based on:
[1] Chen, Long, et al. "Cross-View Tracking for Multi-Human 3D Pose Estimation at Over 100 FPS." The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.