Loading...

Overview

This challenge workshop aims to accelerate the transition of volumetric video technology from laboratory prototypes to robust, production- and consumer-ready systems, enabling breakthroughs in immersive interactive experiences.

Compression Track

Focuses on balancing storage efficiency and playback performance for practical deployment in real-world applications.

Sparse-View Track

Focuses on accurately reconstructing dynamic 3D humans from minimal camera viewpoints, reducing hardware costs and simplifying capture setups.

Our challenge fosters cross-disciplinary discussion between computer vision, computer graphics, and AR/VR practitioners, uniting expertise in 3D reconstruction, generative AI, and immersive media systems.

News

 October 15 (AoE)   Add FAQ section to answer some common questions.

 October 14 (AoE)   We have uploaded new timecode related files to the dataset (timecode.json). Now you can use them to get better FreeTimeGS rendering (render_ftgs_new.py). Besides, for all participants please complete our information collection Google Form as soon as possible.

 September 4 (AoE)    ⚠️ The compression and sparse-view datasets have been released. For compression track, the FreeTimeGS models and their rendering script are now available. Access these resources at: Google Drive

 August 31 (AoE)    We are currently completing the final processing of our dataset, and the release is anticipated by September 5. Thank you for your patience.

Important Dates (AoE)

Registration Opens

Get ready to participate in the challenge

August 15, 2025

Dataset Release

Full dataset available for download

September 1, 2025  By September 5, 2025

Submission Deadline

Final submission of your results

November 23, 2025

Results Announcement

Winners will be notified

November 30, 2025

Workshop & Awards

Join us at SIGGRAPH Asia 2025

December 16, 2025

Challenge Tracks

The challenge features two tracks focusing on cutting-edge compression and sparse-view reconstruction techniques for volumetric videos.

Compression Track

Optimize file size while maintaining high reconstruction quality metrics. Perfect for teams working on efficient storage and streaming solutions.

Dataset: 2 validation sequences + 5 test sequences
Download link

Sparse-View Track

Reconstruct dynamic 3D humans from limited camera views. Ideal for teams exploring neural rendering and view synthesis.

Dataset: 2 validation sequences + 5 test sequences
Download link

Dataset

Our high-fidelity dataset features diverse dynamic human subjects with:

Mixed Focal Lengths

Cinematic-Grade Visual Quality

Challenging Motions

Dataset Structure:

.
├── intri.yml               # Camera intrinsics for training views
├── extri.yml               # Camera extrinsics for training views
├── test_intri.yml          # Camera intrinsics for testing views
├── test_extri.yml          # Camera extrinsics for testing views
├── images/                 # Multi-view images for training
│   ├── 00/                 # Camera name
│   │   ├── 000000.jpg      # Image name using format {frame:06d}.jpg
│   │   └── ...
│   ├── 01/
│   └── ...
├── masks/                  # Multi-view masks for training
│   ├── 00/                 # Camera name
│   │   ├── 000000.jpg
│   │   └── ...
│   ├── 01/
│   └── ...
└── pcds/                # Foreground point clouds
    ├── 000000.ply
    └── ...

Note: We provide official Python scripts for data parsing and visualization. All data are provided for research use only.

Evaluation

Hardware: All evaluations conducted on Linux workstation with one NVIDIA RTX 4090 GPU

Evaluatoin: We use PSNR, SSIM, and LPIPS to measure the foreground-only reconstruction quality

Compression Track

Rank = [rank(PSNR) + rank(SSIM) + rank(LPIPS)]/6 + [rank(Size) + rank(Time)]/4

Size: Total on-disk bytes of all content-dependent artifacts required for rendering. Content-agnostic parts (e.g., shared backbones, shared decoders, ...) are excluded. NOTE: Background regions are not involved in computation, so please remove them before your submission.
Time: Average rendering time of a fixed number of test images with batch size = 1, including preprocessing, decoding, and rendering.

Sparse-View Track

Rank = [rank(PSNR) + rank(SSIM) + rank(LPIPS)]/3

Detailed Requirements

Compression Track

Dataset Split

Validation Set (2 sequences)

  • • Full 60-view videos & masks
  • • Full 60-view camera parameters

Test Set (5 sequences)

  • • 48-view videos & masks
  • • Training & testing camera parameters

Baseline

We provide a FreeTimeGS result as baseline. Participants can use it as a starting point (e.g., linear/non-linear quantization and pruning) or develop novel compression methods.

Submission Requirements

1

Technical Report (PDF, max 4 pages)

SIGGRAPH Asia Technical Communications template recommended

2

Rendered Results (ZIP)

A zip file containing testing-view images in a specified directory structure:

output/
├── 00  # Testing camera name defined by test_intri.yml and test_extri.yml
│   ├── 000000.jpg
│   ├── ...
├── 01
│   ├── 000000.jpg
│   ├── ...
...
3

Model & Scripts (ZIP)

A zip file containing:

  • • A .txt file describing all content-dependent files for size computing. One file per line.
  • • Your Conda environment .yml file
  • • Your compressed models and rendering scripts
  • ▪ The rendering code should support the following evaluation commands:
  • conda env create -f [YOUR_ENV_FILE].yml
    conda activate [YOUR_ENV_NAME]
    python3 render.py --model [YOUR_MODEL] --intri test_intri.yml --extri test_extri.yml (default)
    // or
    python3 render.py --model [YOUR_MODEL] --intri test_intri.yml --extri test_extri.yml --no_image (no image saving for time computing)
    
  • ▪ The default command should generate a folder named output, with the same structure as Rendered Results (ZIP).
  • ▪ Please ensure your code runs non-interactively and reproducibly in a clean environment with the above commands.

Sparse-View Track

Dataset Split

Validation Set (2 sequences)

  • • Full 60-view videos & masks
  • • Full 60-view camera parameters

Test Set (5 sequences)

  • • 8-view videos & masks
  • • Training & testing camera parameters

Submission Requirements

1

Technical Report (PDF, max 4 pages)

SIGGRAPH Asia Technical Communications template recommended

2

Rendered Results (ZIP)

A zip file containing testing-view images in a specified directory structure:

output/
├── 00  # Testing camera name defined by test_intri.yml and test_extri.yml
│   ├── 000000.jpg
│   ├── ...
├── 01
│   ├── 000000.jpg
│   ├── ...
...

Submission Guidelines

Submission

  • One registration per team
  • Team name serves as official identifier
  • Maximum 3 submissions per track

File Naming Convention

Report: TeamName.pdf

Results: TeamName.zip

Model: TeamName_Model.zip

Note: All reports are non-archival. Submission link will be provided soon.

Awards

Each track features two prestigious prizes

🥇

First Prize

$2,500

🥈

Second Prize

$1,500

Winners will present their work via a 5–10 minute pre-recorded video at the workshop.

Workshop Schedule

The workshop will take place at the Hong Kong Convention and Exhibition Centre on 16 December 2025, 9:00 am–12:00 pm.

Time (HKT) Event
09:00 - 09:15 Welcome Remarks & Challenge Results
09:15 - 10:15 Winner Teams Presentation
10:15 - 10:30 Coffee Break & Networking
10:30 - 12:00 Invited Speaker Presentations

Keynote Speakers

Distinguished experts in volumetric video and 3D reconstruction

TBD

Speaker 1

Details coming soon

TBD

Speaker 2

Details coming soon

TBD

Speaker 3

Details coming soon

FAQ

Compression

Q: Can we remove background points to reduce the model size? They seem irrelevant to the metric calculation.

A: Yes. Since background regions are not involved in the computation, we recommend removing them before submission.

Q: Will inaccurate foreground masks affect the calculation of rankings?

A: We will exclude samples with incorrect masks in the test set to avoid their impact on metric calculation.

Q: The images rendered by the provided FreeTimeGS model have floating-point artifacts from some perspectives. How should this be resolved?

A: You can try increasing the value of "near" in the rendering script.

Sparse-View

Organizers

Zhiyuan Yu

Zhiyuan Yu

Zhejiang University

Homepage
Jiaming Sun

Jiaming Sun

4DV.ai

Homepage
Siyu Zhang

Siyu Zhang

4DV.ai

Homepage
Sida Peng

Sida Peng

Zhejiang University

Homepage
Ruizhen Hu

Ruizhen Hu

Shenzhen University

Homepage
Xiaowei Zhou

Xiaowei Zhou

Zhejiang University

Homepage

Acknowledgements

We thank the following institutions for their support:

We also thank Yuanhong Yu and Yuxuan Lin for their valuable contributions to the development of the website and the dataset preparation.

Contact

For any questions, please contact us at

[email protected]