Sports analysis requires processing large amounts of data, which is time-consuming and costly. Advancements in neural networks have significantly alleviated this burden, enabling highly accurate ball tracking in sports broadcasts. However, relying solely on 2D ball tracking is limiting, as it depends on the camera's viewpoint and falls short of supporting comprehensive game analysis. To address this limitation, we propose a novel approach for reconstructing precise 3D ball trajectories from online table tennis match recordings. Our method leverages the underlying physics of the ball's motion to identify the bounce state that minimizes the reprojection error of the ball's flying trajectory, hence ensuring an accurate and reliable 3D reconstruction. A key advantage of our approach is its ability to infer ball spin without relying on human pose estimation or racket tracking, which are often unreliable or unavailable in broadcast footage. We developed an automated camera calibration method capable of reliably tracking camera movements. Additionally, we adapted an existing 3D pose estimation model, which lacks depth motion capture, to accurately track player movements. Together, these contributions enable the full 3D reconstruction of a table tennis rally.
We release an evaluation dataset for benchmarking 3D reconstruction of table tennis rallies from monocular video containing 130 ball trajectories. The dataset includes:
The dataset is designed to evaluate the performance of monocular 3D ball tracking and camera calibration algorithms. It can be used to assess reprojection accuracy, bounce localization, and trajectory smoothness.
We also provide the raw dataset which includes the videos for the oblique and side points of view.
@InProceedings{gossard2025,
author = {Gossard, Thomas and Ziegler, Andreas and Zell, Andreas},
title = {TT3D: Table Tennis 3D Reconstruction},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2025},