This repository will contain the official implementation of the paper: FreeGave: 3D Physics Learning from Dynamic Videos by Gaussian Velocity. FreeGave is a framework that learns 3D geometry, appearance and velocity purely from multi-view videos, achieving SOTA performance in future extrapolation task on four datasets.
Please feel free to contact us via jinxi.li@connect.polyu.hk or open an issue if you have any questions or suggestions.
- 2025-06-14: Our poster is successfully presented in Nashville!
- 2025-02-26: FreeGave is accepted by CVPR2025 🎉!
- Submit the paper onto arXiv.
- Release training codes.
- Release FreeGave-GoPro dataset.
- Release pretrained checkpoints.
git clone https://github.com/vLAR-group/FreeGave.git --recursive
cd FreeGave
### CUDA 11.6
conda env create -f env.yml
conda activate freegave
# install pytorch 1.13.1
conda install pytorch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1 pytorch-cuda=11.6 -c pytorch -c nvidia
# install gaussian requirements
pip install submodules/depth-diff-gaussian-rasterization
pip install submodules/simple-knn
We include in-detail training (both commends and per-scene hyperparameters) and evaluation
instructions in train_eval.sh
file. You just need to download the dataset and modify the path to the
data accordingly.
Notably, we provide training visualization, for which you just need to add
--gui
flag to the training command (but this will increase the GPU usage and the training time
significantly, so we only recommend using it for debugging).
After modifying the path, simply run the following commend:
bash train_eval.sh
All the datasets will be uploaded soon. We organize the dataset following D-NeRF convention. We split the dataset as:
- train: contains the frames within observed time interval, used for training the model.
- val: contains the frames within observed time interval but for novel views, used for evaluating novel-view interpolation.
- test: contains the frames in unobserved future time for both observed and novel views, used for evaluating future extrapolation.
This work is adapted from Deformable-3DGS and NVFi, and we would like to thank the authors for their great work.
If you find our work helpful, please consider citing:
@article{li2025freegave,
title={FreeGave: 3D Physics Learning from Dynamic Videos by Gaussian Velocity},
author={Jinxi Li and Ziyang Song and Siyuan Zhou and Bo Yang},
year={2025},
journal={CVPR}
}