Nerfstudio
v1.1.0Modular framework for neural radiance field development supporting NeRF, Gaussian Splatting, and many method variants
Development Activity
Sample Renders
Overview
Best for
Researchers and engineers who need a unified platform for experimenting with multiple neural rendering methods, and as the standard framework for academic neural rendering publications
Not ideal for
Production deployment requiring minimal dependencies, projects targeting non-NVIDIA hardware, or use cases demanding maximum single-method training speed
Strengths
- The most comprehensive neural rendering framework — supports NeRF, Gaussian Splatting, and dozens of method variants through a unified, modular architecture with swappable components
- Interactive web-based viewer (viser) provides real-time visualization of training progress, camera path creation, scene exploration, and export — the best developer experience in the neural rendering space
- First-class data pipeline supporting diverse input formats including COLMAP, Polycam, Record3D, and Blender synthetic datasets with automatic preprocessing
- Well-documented with tutorials, API reference, and an active community — the most approachable entry point for researchers new to neural rendering
- Actively maintained by a dedicated team at UC Berkeley with regular releases, new method integrations, and strong ecosystem partnerships with gsplat, nerfacc, and viser
Limitations
- Significant installation complexity — requires PyTorch with CUDA, tiny-cuda-nn, and multiple compiled extensions that can break across CUDA and PyTorch version combinations
- Higher memory overhead than specialized single-method implementations due to the generality and abstraction layers of the framework
- Training speed for individual methods can be slower than optimized standalone implementations (e.g., original 3DGS code may train faster than Nerfstudio's splatfacto)
- Rapid development pace means breaking API changes between versions — code targeting v0.3 may not run on v1.x without significant modification
- Requires NVIDIA GPU for any practical use — CPU and non-CUDA GPU paths are not viable for training or real-time rendering
Background
Nerfstudio is a comprehensive, modular framework for neural radiance field research and development, created at UC Berkeley. It provides a unified platform for training, evaluating, and deploying multiple neural rendering methods — including NeRF variants (nerfacto, instant-ngp integration), Gaussian Splatting methods (splatfacto via gsplat), and experimental approaches — through a consistent API and data pipeline.
The framework's standout feature is its interactive web-based viewer (powered by viser), which provides real-time visualization of training progress, camera path creation, and scene exploration. Its modular architecture allows researchers to swap components — data parsers, field representations, renderers, and loss functions — independently, enabling rapid prototyping of new methods without reimplementing infrastructure. Nerfstudio supports diverse input formats including COLMAP, Polycam, Record3D, and Blender synthetic datasets.
Maintained by a dedicated team with regular releases and an active open-source community, Nerfstudio has become the de facto standard framework for academic neural rendering research. It provides comprehensive documentation, tutorials, and a growing ecosystem of compatible tools (gsplat for Gaussian rasterization, nerfacc for volumetric acceleration, viser for visualization). The framework is designed to lower the barrier to entry for neural rendering research while providing the flexibility needed for state-of-the-art work.
Quick Start
pip install nerfstudioRelated Renderers
Community & Resources
Paper & Citations
Tutorials & Resources
Performance Benchmarks
No benchmark data available for Nerfstudio yet.
Benchmarks will be added as more renderers are tested across our standard scene suite.
Learn about our methodology