Scene Generation
62 papers with code • 5 benchmarks • 8 datasets
Libraries
Use these libraries to find Scene Generation models and implementationsLatest papers
WcDT: World-centric Diffusion Transformer for Traffic Scene Generation
To enhance the scene diversity and stochasticity, the historical trajectory data is first preprocessed and encoded into latent space using Denoising Diffusion Probabilistic Models (DDPM) enhanced with Diffusion with Transformer (DiT) blocks.
Towards Realistic Scene Generation with LiDAR Diffusion Models
In this paper, we propose LiDAR Diffusion Models (LiDMs) to generate LiDAR-realistic scenes from a latent space tailored to capture the realism of LiDAR scenes by incorporating geometric priors into the learning pipeline.
SemCity: Semantic Scene Generation with Triplane Diffusion
In this paper, we concentrate on generating a real-outdoor scene through learning a diffusion model on a real-world outdoor dataset.
WoVoGen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation
Generating multi-camera street-view videos is critical for augmenting autonomous driving datasets, addressing the urgent demand for extensive and varied data.
Pyramid Diffusion for Fine 3D Large Scene Generation
Directly transferring the 2D techniques to 3D scene generation is challenging due to significant resolution reduction and the scarcity of comprehensive real-world 3D scene datasets.
High-fidelity Person-centric Subject-to-Image Synthesis
Specifically, we first develop two specialized pre-trained diffusion models, i. e., Text-driven Diffusion Model (TDM) and Subject-augmented Diffusion Model (SDM), for scene and person generation, respectively.
LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts
Diffusion-based generative models have significantly advanced text-to-image generation but encounter challenges when processing lengthy and intricate text prompts describing complex scenes with multiple objects.
RoomDesigner: Encoding Anchor-latents for Style-consistent and Shape-compatible Indoor Scene Generation
Indoor scene generation aims at creating shape-compatible, style-consistent furniture arrangements within a spatially reasonable layout.
CityDreamer: Compositional Generative Model of Unbounded 3D Cities
3D city generation is a desirable yet challenging task, since humans are more sensitive to structural distortions in urban environments.
On the Generation of a Synthetic Event-Based Vision Dataset for Navigation and Landing
We anticipate that novel event-based vision datasets can be generated using this pipeline to support various spacecraft pose reconstruction problems given events as input, and we hope that the proposed methodology would attract the attention of researchers working at the intersection of neuromorphic vision and guidance navigation and control.