base on Official Implementation of Self-Supervised Street Gaussians for Autonomous Driving # <i>S</i><sup>3</sup>Gaussian: Self-Supervised Street Gaussians for Autonomous Driving ### [Paper](https://arxiv.org/abs/2405.20323) | [Project Page](https://wzzheng.net/S3Gaussian) > <i>S</i><sup>3</sup>Gaussian: Self-Supervised Street Gaussians for Autonomous Driving > [Nan Huang](https://github.com/nnanhuang)\*<sup>§</sup>, [Xiaobao Wei](https://ucwxb.github.io/)\*, [Wenzhao Zheng](https://wzzheng.net/)<sup>†</sup>, Pengju An, [Ming Lu](https://lu-m13.github.io/), [Wei Zhan](https://zhanwei.site/), [Masayoshi Tomizuka](https://me.berkeley.edu/people/masayoshi-tomizuka/), [Kurt Keutzer](https://people.eecs.berkeley.edu/~keutzer/), [Shanghang Zhang](https://www.shanghangzhang.com/)<sup>‡</sup> \* Equal contribution § Work done while interning at UC Berkeley † Project leader ‡ Corresponding author <i>S</i><sup>3</sup>Gaussian employs 3D Gaussians to model dynamic scenes for autonomous driving ***without*** other supervisions (e.g., 3D bounding boxes). ![vis](./assets/vis2.png) ## News - **[2026/1/31]** <i>S</i><sup>3</sup>Gaussian has been accepted by ICRA 2026. - **[2025/7/9]** **🚀 Check out our new work [EMD (ICCV2025)](https://qingpowuwu.github.io/emd/): a plug-and-play module for street reconstruction.** - **[2024/5/31]** Paper released on [arXiv](https://arxiv.org/abs/2405.20323). - **[2023/5/31]** Training & evaluation code release! ## Demo ![demo](./assets/visual.gif) ## Overview ![overview](./assets/pipeline.png) To tackle the challenges in self-supervised street scene decomposition, we propose a multi-resolution hexplane-based encoder to encode 4D grid into feature planes and a multi-head Gaussian decoder to decode them into deformed 4D Gaussians. We optimize the overall model without extra annotations in a self-supervised manner and achieve superior scene decomposition ability and rendering quality. ## Results ![overview](./assets/results.png) ## Getting Started ### Environmental Setups Our code is developed on Ubuntu 22.04 using Python 3.9 and pytorch=1.13.1+cu116. We also tested on pytorch=2.2.1+cu118. We recommend using conda for the installation of dependencies. ```bash git clone https://github.com/nnanhuang/S3Gaussian.git --recursive cd S3Gaussian conda create -n S3Gaussian python=3.9 conda activate S3Gaussian pip install -r requirements.txt pip install -e submodules/depth-diff-gaussian-rasterization pip install -e submodules/simple-knn ``` ### Preparing Dataset Follow detailed instructions in [Prepare Dataset](docs/prepare_data.md). We only use dynamic32 and static32 split. ### Training For training first clip (eg. 0-50 frames), run ``` python train.py -s $data_dir --port 6017 --expname "waymo" --model_path $model_path ``` If you want to try novel view synthesis, use ``` --configs "arguments/nvs.py" ``` For instance, you can try: ``` python train.py -s "./data/processed/dynamic32/training/022" --expname "waymo" --model_path "./work_dirs/phase1/dynamic/recon/022" ``` For training next clip (eg. 51-100 frames), run ``` python train.py -s $data_dir --port 6017 --expname "waymo" --model_path $model_path --prior_checkpoint "$prior_dir/chkpnt_fine_50000.pth" --configs "arguments/stage2.py" ``` For instance, you can try: ``` python train.py -s "./data/processed/dynamic32/training/022" --expname "waymo" --model_path "./work_dirs/phase1/dynamic/recon/p2/022" --prior_checkpoint "./work_dirs/phase1/dynamic/recon/022/chkpnt_fine_50000.pth" --configs "arguments/stage2.py" ``` Also, you can load an existing checkpoint with: ```python python train.py -s $data_dir --port 6017 --expname "waymo" --start_checkpoint "$ckpt_dir/chkpnt_fine_30000.pth" --model_path $model_path ``` For more scripts examples, please check [here](scripts). ### Evaluation and Visualization You can visualize and eval a checkpoints follow: ```python python train.py -s $data_dir --port 6017 --expname "waymo" --start_checkpoint "$ckpt_dir/chkpnt_fine_50000.pth" --model_path $model_path --eval_only ``` If you use different configs, you will need to add them as well: ``` --configs "arguments/nvs.py" ``` Then you can get rendering RGB videos, ground truth RGB videos, depth videos, dynamic rgb videos and static rgb videos. ## Acknowledgments Credits to @[Korace0v0](https://github.com/korace0v0) for building 3D Gaussians for street scenes. Many thanks! Special thanks to [StreetGaussians](https://github.com/zju3dv/street_gaussians) for sharing visualization results! Our code is based on [4D Gaussians](https://github.com/hustvl/4DGaussians/tree/master) and [EmerNeRF](https://github.com/NVlabs/EmerNeRF?tab=readme-ov-file). Thanks to these excellent open-sourced repos! ## Citation If you find this project helpful, please consider citing the following paper: ``` @inproceedings{huang2026s3gaussian, title={S3Gaussian: Self-Supervised Street Gaussians for Autonomous Driving}, author={Huang, Nan and Wei, Xiaobao and Zheng, Wenzhao and An, Pengju and Lu, Ming and Zhan, Wei and Tomizuka, Masayoshi and Keutzer, Kurt and Zhang, Shanghang}, booktitle={Proceedings of the IEEE International Conference on Robotics and Automation (ICRA)}, year={2026}, } ``` ", Assign "at most 3 tags" to the expected json: {"id":"10643","tags":[]} "only from the tags list I provide: []" returns me the "expected json"