Datasets:
image image | label class label |
|---|---|
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images | |
0images |
π§βπ€βπ§ Crowd-3DGS Dataset
A sequential-frame dataset of a real-world crowd protest scene, fully prepared for 3D Gaussian Splatting (3DGS) reconstruction β including high- and low-resolution image sets, COLMAP sparse reconstruction, stereo depth maps, and ready-to-run COLMAP scripts.
Dataset Description
Crowd-3DGS is a single-scene video-derived dataset capturing a crowd protest, intended for 3D scene reconstruction and novel view synthesis using 3D Gaussian Splatting (3DGS). The dataset ships as a complete, pipeline-ready package: sequential JPG frames extracted from video, camera calibration via COLMAP Structure-from-Motion (SfM), stereo dense reconstruction outputs, and helper shell scripts to re-run COLMAP from scratch.
Crowd and protest scenes present unique challenges for 3DGS: heavy human occlusion, transient moving subjects, non-rigid motion, and complex lighting. This dataset provides a realistic real-world benchmark for researchers tackling robust 3DGS under these conditions.
- Curated by: siyah1
- Funded by:
- Language(s): N/A (visual/3D dataset)
- License: CC-BY 4.0
- Total size: ~392 MB
Related Links
| Resource | Link |
|---|---|
| π€ Dataset Repository | siyah1/crowd-3dgs |
| π Paper | |
| π» Code / Training Script | |
| π Project Page |
Background: What is 3D Gaussian Splatting?
3D Gaussian Splatting (3DGS) (Kerbl et al., 2023) is an explicit 3D scene representation that models a scene as a collection of anisotropic 3D Gaussian primitives, each defined by:
- Position (mean): 3D location in space
- Covariance (shape/orientation): size and orientation of the Gaussian
- Opacity: transparency level
- Spherical Harmonics (SH) coefficients: view-dependent colour and appearance
3DGS achieves real-time rendering at high visual fidelity and is widely used in VR, AR, robotics, and digital heritage. However, crowd scenes β with heavy occlusion, dynamic humans, and transient elements β remain an open challenge for the method.
Uses
Direct Use
- 3D reconstruction of a real-world crowd/protest scene using 3DGS or NeRF-based methods
- Novel view synthesis benchmarking in challenging, densely populated environments
- Research into transient object handling, occlusion-robust 3DGS, and dynamic scene reconstruction
- Prototyping end-to-end COLMAP-to-3DGS pipelines using the included scripts and pre-processed data
Out-of-Scope Use
This dataset is not intended for:
- Facial recognition, individual identity tracking, or surveillance
- Training crowd-counting or person-detection models (see CrowdHuman, UCF-QNRF for those)
- Commercial deployment without appropriate licensing clearance
- Any use that violates the privacy or dignity of individuals depicted
Dataset Structure
File Tree
crowd-3dgs/ (~392 MB total)
β
βββ images/ (~202 MB) β High-resolution frames
β βββ frame_0001.jpg (~675 kB)
β βββ frame_0002.jpg (~664 kB)
β βββ frame_0003.jpg (~666 kB)
β βββ ... (sequential video frames)
β
βββ input/ (~95.1 MB) β Low-resolution frames
β βββ frame_0001.jpg (~324 kB)
β βββ frame_0002.jpg (~314 kB)
β βββ frame_0003.jpg (~316 kB)
β βββ ... (same sequence, downscaled)
β
βββ sparse/ β COLMAP SfM output
β βββ 0/
β βββ cameras.bin (camera intrinsics)
β βββ images.bin (per-frame pose / extrinsics)
β βββ points3D.bin (sparse 3D point cloud)
β
βββ stereo/ (~12.9 kB) β COLMAP stereo config files
β βββ fusion.cfg (4.5 kB β stereo fusion settings)
β βββ patch-match.cfg (8.4 kB β patch-match stereo settings)
β
βββ run-colmap-geometric.sh β Script: geometric stereo fusion
βββ run-colmap-photometric.sh β Script: photometric stereo fusion
βββ .gitattributes
βββ README.md
Folder Descriptions
| Folder / File | Size | Description |
|---|---|---|
images/ |
~202 MB | Full-resolution sequential JPG frames from source video. Primary input to 3DGS. |
input/ |
~95.1 MB | Downscaled version of the same frames. Useful for faster COLMAP or low-VRAM training. |
sparse/ |
β | COLMAP SfM output: camera intrinsics, per-frame poses, sparse 3D point cloud for 3DGS initialisation. |
stereo/ |
~12.9 kB | COLMAP stereo configuration files: fusion.cfg (fusion settings) and patch-match.cfg (patch-match stereo settings). Run the included .sh scripts to generate depth/normal maps. |
run-colmap-geometric.sh |
β | Re-run COLMAP dense stereo with geometric consistency filtering. |
run-colmap-photometric.sh |
β | Re-run COLMAP dense stereo with photometric consistency filtering. |
Image Statistics
| Property | images/ (hi-res) |
input/ (lo-res) |
|---|---|---|
| Format | JPEG | JPEG |
| Avg. file size | ~660 kB | ~315 kB |
| Naming convention | frame_XXXX.jpg |
frame_XXXX.jpg |
| Scene | Crowd protest | Crowd protest |
Frame count:
How to Use
Quick Start with 3DGS
The sparse/ folder already contains pre-computed COLMAP output, so you can jump straight to training:
# 1. Clone the dataset
git clone https://huggingface.co/datasets/siyah1/crowd-3dgs
# 2. Train with the official 3DGS repo (https://github.com/graphdeco-inria/gaussian-splatting)
python train.py -s crowd-3dgs/
Re-running COLMAP Dense Stereo (Optional)
# Geometric stereo (recommended for cleaner reconstructions)
bash run-colmap-geometric.sh
# Photometric stereo (alternative)
bash run-colmap-photometric.sh
Using input/ for Fast Iteration
The downscaled input/ frames (~315 kB vs ~660 kB) are useful for:
- Rapid prototyping without loading the full 202 MB image set
- COLMAP runs on machines with limited RAM
- Initial hyperparameter sweeps before full-resolution training
Dataset Creation
Curation Rationale
Standard 3DGS benchmarks (Tanks & Temples, Mip-NeRF 360, Deep Blending) are dominated by static or near-static scenes with sparse human presence. Crowd-3DGS targets the gap for densely-populated, real-world protest/event scenarios β where occlusion, transient people, and motion make reconstruction significantly harder and more practically relevant.
Source Data
Data Collection and Processing
- Scene type: Outdoor crowd protest
- Capture method: [e.g., handheld smartphone / DSLR video β More Information Needed]
- Frame extraction: Sequential frames extracted from video (
frame_0001.jpg,frame_0002.jpg, β¦) - Camera calibration: COLMAP SfM run on
input/frames; camera poses stored insparse/0/ - Dense reconstruction: COLMAP patch-match stereo configured via
stereo/patch-match.cfgandstereo/fusion.cfg; run the included.shscripts to generate depth/normal maps - Multi-resolution: Full-res in
images/, downscaled copies ininput/
Who are the source data producers?
Video captured by [More Information Needed]. All individuals appear in a publicly filmed protest scene.
Personal and Sensitive Information
This dataset contains images of people participating in a public protest:
- Anonymization applied: [Yes / No β More Information Needed]
- EXIF metadata: Verify GPS/timestamp data is stripped before publishing
- This dataset is not suitable for individual identification, political tracking, or surveillance
- Protest participation is a form of political expression β treat subjects' dignity accordingly
β οΈ If face anonymization has not been applied, add appropriate usage restrictions and warnings.
Bias, Risks, and Limitations
Technical Limitations
| Issue | Impact on 3DGS |
|---|---|
| Transient moving people | Inconsistency across frames β floaters, ghosting, blurring in Gaussians |
| Video-derived viewpoints | Camera path is sequential, not a multi-rig; viewpoint diversity is limited |
| Motion blur | Fast crowd movement may degrade per-frame sharpness |
| Single scene | Results may not generalise to other crowd types, densities, or locations |
| Stereo depth noise | Dense depth maps from patch-match are noisy on crowd/human regions |
Sociotechnical Risks
- Individuals are participants in a public protest β a politically sensitive context
- Use for surveillance, identity recognition, or political profiling is explicitly prohibited
- Geographic and demographic biases exist based on the specific protest location and demographics
Recommendations
- Apply transient masking during 3DGS training (semantic segmentation of humans, or RobustNeRF-style robust loss)
- Compute PSNR / SSIM / LPIPS on static background regions, masking dynamic foreground humans
- Do not use this dataset or derived models for crowd monitoring, surveillance, or profiling
Evaluation
Recommended metrics for benchmarking 3DGS on this dataset:
| Metric | Description | Direction |
|---|---|---|
| PSNR | Peak Signal-to-Noise Ratio | β Higher is better |
| SSIM | Structural Similarity Index | β Higher is better |
| LPIPS | Learned Perceptual Image Patch Similarity | β Lower is better |
| FPS | Rendering speed | β Higher is better |
Citation
BibTeX:
@dataset{crowd3dgs2026,
author = {siyah1},
title = {Crowd-3DGS: A Sequential Frame Dataset of a Crowd Protest Scene for 3D Gaussian Splatting},
year = {2026},
publisher = {Hugging Face},
url = {https://huggingface.co/datasets/siyah1/crowd-3dgs}
}
APA:
siyah1. (2026). Crowd-3DGS: A Sequential Frame Dataset of a Crowd Protest Scene for 3D Gaussian Splatting [Dataset]. Hugging Face. https://huggingface.co/datasets/siyah1/crowd-3dgs
Glossary
| Term | Definition |
|---|---|
| 3DGS | 3D Gaussian Splatting β real-time novel view synthesis using anisotropic 3D Gaussian primitives |
| NVS | Novel View Synthesis β rendering a scene from camera positions not seen during training |
| SfM | Structure-from-Motion β recovering 3D camera poses from 2D images |
| COLMAP | Open-source SfM + MVS pipeline; standard preprocessing tool for 3DGS |
| Sparse reconstruction | COLMAP's SfM output: sparse 3D points + per-frame camera poses |
| Dense / stereo reconstruction | Patch-match stereo producing per-pixel depth and normal maps |
| Geometric stereo | COLMAP mode enforcing cross-frame geometric consistency |
| Photometric stereo | COLMAP mode enforcing cross-frame photometric (appearance) consistency |
| Transient objects | Dynamic elements (moving people) that appear inconsistently across frames |
| Spherical Harmonics (SH) | Basis functions encoding view-dependent colour in 3DGS |
Dataset Card Contact
Fields marked "[More Information Needed]" should be completed by the dataset author. Card follows the HuggingFace dataset card spec.
- Downloads last month
- 540