3D scene generation research is fragmented. Methods like DiffuScene, Holodeck, HSM, and LayoutGPT use formats, metrics, and evaluation protocols incompatible with each other. This fragmentation creates critical challenges:
- Incomparability: "Method A achieves X% on dataset D1" vs "Method B achieves Y% on dataset D2" - it is difficult to know which is better
- Unreproducibility: Each paper reimplements infrastructure from scratch, making validation difficult
- Slow iteration: Researchers spend time on boilerplate instead of novel algorithms
- Component isolation: Cannot test whether learned placement improves rule-based methods, or vice versa
The field needs infrastructure that enables systematic comparison and accelerates research velocity.
sisglib aims to become the standard infrastructure for 3D scene generation research - analogous to what OpenAI Gym did for reinforcement learning or ImageNet/COCO did for computer vision.
-
Common Interface, Diverse Approaches
- Any method (linear, learned, agentic, hybrid) implements the same interface:
prompt → SceneState - Internal implementation is unconstrained - methods can be fundamentally different while remaining interoperable
- Any method (linear, learned, agentic, hybrid) implements the same interface:
-
Standards Enable Science
- sissf (Spatial Intelligence Scene State Format): Standard representation for scenes, enabling fair comparison
- SceneGenerationStrategy Interface: Standard abstraction for algorithms, enabling component reuse
- Unified Evaluation: Same benchmarks and metrics across all methods
-
Composability Accelerates Innovation
- Mix components from different methods (e.g. Holodeck's object selection + DiffuScene's placement)
- Build hybrid approaches by combining linear, learned, and agentic components
- Ablate individual components without reimplementing entire systems
-
Research to Production
- Prototype locally with minimal dependencies
- Scale to production by swapping backends (storage, vectors, metadata)
- Same research code, different configuration
# Compare fundamentally different approaches on identical benchmarks
methods = [
HolodeckStrategy.build(), # Linear pipeline
DiffuSceneStrategy.build(), # End-to-end learned
AgenticLLMStrategy.build(), # LLM-directed
]
benchmark = SceneBenchmark(
dataset=load_dataset("structured3d_prompts"),
metrics=[CollisionRate(), SemanticCoherence(), SpatialPlausibility()]
)
for method in methods:
results = await benchmark.evaluate(method)
# Now we can actually answer: "Which method is better?"# Question: Does learned placement beat constraint-based?
baseline = LinearStrategy.builder()
.add_stage(RuleBasedArchitecture())
.add_stage(ConstraintBasedPlacement())
.build()
learned = LinearStrategy.builder()
.add_stage(RuleBasedArchitecture()) # Same architecture
.add_stage(LearnedPlacement()) # Only placement differs
.build()
# Isolate the impact of placement method
compare([baseline, learned], metrics=[CollisionRate(), SceneQuality()])# "Does Holodeck's object selection improve DiffuScene?"
hybrid = DAGStrategy.builder()
.add_node("holodeck_selection", HolodeckObjectSelection())
.add_node("diffusion_placement", DiffusionPlacement())
.add_edge("holodeck_selection", "diffusion_placement")
.build()
# Test combinations impossible without a common framework# Share complete experimental configuration
strategy:
type: linear
stages:
- type: holodeck_architecture
version: 1.0.2
- type: learned_placement
checkpoint: weights/best.pt
benchmark:
dataset: structured3d_v2
metrics: [collision, coherence]
seed: 42Anyone can reproduce your exact experiment with identical results.
sisglib succeeds when:
- Methods are comparable: Researchers routinely compare their work against baselines using sisglib benchmarks
- Components are reusable: Papers cite using "X's object selection with our placement method"
- Reproduction is trivial: "We used config Y from the sisglib repository"
- Innovation accelerates: Time from idea to validated result decreases significantly
- Community adopts: Major papers implement their methods in sisglib alongside custom codebases
sisglib is open-source (Apache 2.0) because standardization only works through community adoption. We invite:
- Researchers: Implement your methods, contribute benchmarks
- Labs: Integrate with your existing work
- Industry: Build production systems on research foundations
- Everyone: Provide feedback, report issues, suggest improvements
Together, we can accelerate 3D scene generation research.