The dataset viewer is not available because its heuristics could not detect any supported data files. You can try uploading some data files, or configuring the data files location manually.
DX.GL Multi-View Datasets for NeRF & 3D Gaussian Splatting
Multi-view training datasets rendered from CC0 3D models via DX.GL. Each dataset includes calibrated camera poses, depth maps, normal maps, binary masks, and point clouds — ready for nerfstudio out of the box.
10 objects × 196 views × 1024×1024 resolution × full sphere coverage.
Quick Start
# Download a dataset (Apple, 196 views, 1024x1024)
wget https://dx.gl/api/v/EJbs8npt2RVM/vCHDLxjWG65d/dataset -O apple.zip
unzip apple.zip -d apple
# Train with nerfstudio
pip install nerfstudio
ns-train splatfacto --data ./apple \
--max-num-iterations 20000 \
--pipeline.model.sh-degree 3 \
--pipeline.model.background-color white
Or use the download script:
pip install requests
python download_all.py
What's in Each Dataset ZIP
dataset/
├── images/ # RGB frames (PNG, transparent background)
│ ├── frame_00000.png
│ └── ...
├── depth/ # 8-bit grayscale depth maps
├── depth_16bit/ # 16-bit grayscale depth maps (higher precision)
├── normals/ # World-space normal maps
├── masks/ # Binary alpha masks
├── transforms.json # Camera poses (nerfstudio / instant-ngp format)
└── points3D.ply # Sparse point cloud for initialization
transforms.json Format
Compatible with both nerfstudio and instant-ngp:
{
"camera_angle_x": 0.857,
"camera_angle_y": 0.857,
"fl_x": 693.5,
"fl_y": 693.5,
"cx": 400,
"cy": 400,
"w": 800,
"h": 800,
"depth_near": 0.85,
"depth_far": 2.35,
"ply_file_path": "points3D.ply",
"frames": [
{
"file_path": "images/frame_00000.png",
"depth_file_path": "depth/frame_00000.png",
"normal_file_path": "normals/frame_00000.png",
"mask_file_path": "masks/frame_00000.png",
"transform_matrix": [[...], [...], [...], [0, 0, 0, 1]]
}
]
}
Specs
| Property | Value |
|---|---|
| Views | 196 per object |
| Resolution | 1024×1024 |
| Coverage | Full sphere (±89° elevation) |
| Point cloud | ~200k points |
| Camera distribution | Fibonacci golden-angle spiral |
| Background | Transparent (RGBA) |
| Lighting | Studio HDRI + directional lights |
Camera Distribution
Views are distributed on a full sphere (±89° elevation) using a golden-angle Fibonacci spiral. The distribution is uniform in solid angle — more views near the equator, fewer near the poles — optimized for NeRF/3DGS training.
Objects
| # | Object | Category | Download | Browse |
|---|---|---|---|---|
| 1 | Apple | organic | ZIP | View |
| 2 | Cash Register | electronics | ZIP | View |
| 3 | Drill | tool | ZIP | View |
| 4 | Fire Extinguisher | metallic | ZIP | View |
| 5 | LED Lightbulb | glass | ZIP | View |
| 6 | Measuring Tape | tool | ZIP | View |
| 7 | Modern Arm Chair | furniture | ZIP | View |
| 8 | Multi Cleaner 5L | product | ZIP | View |
| 9 | Potted Plant | organic | ZIP | View |
| 10 | Wet Floor Sign | plastic | ZIP | View |
All source models from Polyhaven (CC0).
Pre-trained 3DGS Splats
We include pre-trained Gaussian Splat .ply files (nerfstudio splatfacto, 20k iterations, SH degree 3) for each object. Download them with:
python download_all.py --splats
Or view them directly:
- DX.GL Splat Viewer (all 10 models, use ← → to browse)
- SuperSplat Editor (drag-drop the .ply)
- nerfstudio viewer:
ns-viewer --load-config outputs/*/config.yml
Training Parameters
ns-train splatfacto --data ./dataset \
--max-num-iterations 20000 \
--pipeline.model.sh-degree 3 \
--pipeline.model.background-color white \
--pipeline.model.cull-alpha-thresh 0.2 \
--pipeline.model.densify-size-thresh 0.005 \
--pipeline.model.use-scale-regularization True \
--pipeline.model.max-gauss-ratio 5.0
Training time: ~10 minutes on RTX 4000 Pro Ada (70W) at the 196×1024 tier.
Rendering Pipeline
Datasets are rendered using DX.GL's cloud GPU rendering pipeline:
- Lighting: Studio HDRI environment with PBR materials
- Camera: Fibonacci golden-angle sphere distribution
- Depth: Tight near/far planes from model bounding sphere for maximum precision
- Point cloud: Back-projected from depth maps, ~1000 points per view
- Background: Transparent (RGBA)
Modalities
| Modality | Format | Notes |
|---|---|---|
| RGB | PNG, RGBA | Transparent background, PBR-lit |
| Depth (8-bit) | PNG, grayscale | Normalized to near/far range |
| Depth (16-bit) | PNG, grayscale | RG-encoded, higher precision |
| Normals | PNG, RGB | World-space, MeshNormalMaterial |
| Masks | PNG, grayscale | Binary alpha from RGB alpha channel |
| Point Cloud | PLY, binary | XYZ + RGB, ~100k points |
| Camera Poses | JSON | 4×4 camera-to-world matrices |
License
All source 3D models are CC0 (public domain) from Polyhaven. The rendered datasets inherit this license — use them for anything, no attribution required.
Citation
@misc{dxgl_multiview_2026,
title = {DX.GL Multi-View Datasets for NeRF and 3D Gaussian Splatting},
author = {DXGL},
year = {2026},
url = {https://huggingface.co/datasets/dxgl/multiview-datasets},
note = {Multi-view datasets with depth, normals, masks, and point clouds. Rendered via DX.GL.}
}
Links
- This collection: dx.gl/datasets/polyhaven-10
- Browse all datasets: dx.gl/datasets
- Pipeline details: dx.gl/for-research
- API documentation: dx.gl/portal/docs
- Generate your own: dx.gl/signup (2 free renders included)
Feedback
We're actively improving the rendering pipeline. If you find issues with depth accuracy, mask quality, camera calibration, or view distribution — please open a Discussion on this repo. Specific feedback we're looking for:
- Depth map accuracy at object edges
- Mask quality for transparent/reflective materials
- Point cloud alignment with RGB views
- View distribution quality for your training method
- Missing modalities or metadata
- Any other issues or suggestions?
- Downloads last month
- 18
