We represent the scene in two parts: the 2D top‑down semantic map with a fixed physical unit per pixel (middle) and the 3D bounding boxes with orientations for object‑level attributes (right).
Note: The left image is the corresponding rendered scene.
From left: (a) is the unified diffusion model that is conditioned on the room mask, and room type croom. During the denoising process, the arch mask or floor mask embedding is added to the noise input embedding. The room type embedding is added to the timestep embedding. (b) is the object attribute prediction model with a semantic layout map as input. si, pi, ri indicate the ith instance's size, position, and orientation. At training time, we use ground-truth instance masks. During inference the semantic layout map is split into instance masks by using connected component analysis. The layout feature and the mask feature are passed to a cross-attention layer to get the final object instance feature, which is used to predict attributes. (c) During inference, objects are retrieved to match the category ci and size si, and arranged using the position pi and orientation ri. Our SemLayoutDiff generates scenes with fewer errors and respects architectural constraints by keeping furniture within room boundaries and maintaining clear spaces around doors (red) and windows (pink), whereas DiffuScene and MiDiffusion do not.
Examples of generated scenes using our SemLayoutDiff mixed-condition model under different condition types from top-down view with Blender rendering. The condition room mask is shown on the top-left of each generated scene (for arch mask, door is red and window is pink).
Sample results showing diverse textured scene layouts generated by our mixed-condition model.
@article{sun2025semlayoutdiff,
title={{SemLayoutDiff}: Semantic Layout Generation with Diffusion Model for Indoor Scene Synthesis},
author={Xiaohao Sun and Divyam Goel and Angle X. Chang},
year={2025},
eprint={2508.18597},
archivePrefix={arXiv},
}
This work was funded in part by a CIFAR AI Chair and NSERC Discovery Grants, and enabled by support from the Digital Research Alliance of Canada and a CFI/BCKDF JELF. We thank Ivan Tam for help with running SceneEval; Yiming Zhang and Jiayi Liu for suggestions on figures; Derek Pun, Dongchen Yang, Xingguang Yan, and Manolis Savva for discussions, proofreading, and paper suggestions. We also thank the anonymous reviewers for their feedback.