LayoutDM: Discrete Diffusion Model
for Controllable Layout Generation
CVPR 2023



Top: LayoutDM is trained to gradually generate a complete layout from a blank state in discrete state space.
Bottom: During sampling, we can steer LayoutDM to perform various conditional generation tasks
without additional training or external models.


Abstract

Controllable layout generation aims at synthesizing plausible arrangement of element bounding boxes with optional constraints, such as type or position of a specific element. In this work, we try to solve a broad range of layout generation tasks in a single model that is based on discrete state-space diffusion models. Our model, named LayoutDM, naturally handles the structured layout data in the discrete representation and learns to progressively infer a noiseless layout from the initial input, where we model the layout corruption process by modality-wise discrete diffusion. For conditional generation, we propose to inject layout constraints in the form of masking or logit adjustment during inference. We show in the experiments that our LayoutDM successfully generates high-quality layouts and outperforms both task-specific and task-agnostic baselines on several layout tasks.

Results

We show some conditional generation results using masking.

We show some conditional generation results using a combination of masking and logit adjustment.

Please refer to the appendix of our paper for more details.

Video

Citation

@inproceedings{inoue2023layout,
    title={{LayoutDM: Discrete Diffusion Model for Controllable Layout Generation}},
    author={Naoto Inoue and Kotaro Kikuchi and Edgar Simo-Serra and Mayu Otani and Kota Yamaguchi},
    booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    year={2023},
    pages={10167-10176},
  }

Acknowledgements

The website template was borrowed from Mip-NeRF 360.