diff --git a/index.html b/index.html index 08e3278..06246f2 100644 --- a/index.html +++ b/index.html @@ -3,10 +3,10 @@
+ content="EditRoom: LLM-parameterized Graph Diffusion for Composable 3D Room Layout Editing"> -Figure 2. MiniGPT-5 pipeline.
+Figure 2. Scene Editor aims to provide accurate, coherent editing results according to the given source scene and language commands. + It consists of two graph transformer-based conditional diffusion models. One diffusion model generates semantic target scene graphs. + Another diffusion model can estimate accurate poses and size information for each object inside the generated target scene graphs. + All diffusion processes are conditioned on the source scene and breakdown command.
@@ -183,12 +202,16 @@- Qualitative examples from MiniGPT-5 and baselines on the CC3M, VIST, and MMDialog datasets. From the comparisons, we can find the MiniGPT-5 and SD 2 have similar results on single-image generation. When we evaluate with multi-step multimodal prompts, MiniGPT-5 can produce more coherent and high-quality images. + Qualitative examples from EditRoom and baselines on single- and multi-operation editing. From the comparisons, we can find the EditRoom can provide more accurate and coherent editing results than other baselines, and it can generalize to multi-operation editing tasks without training on such data.
Figure 3. Comparison with other baselines on single-operation editing.
+ +Figure 3. Comparison with other baselines.
+Figure 4. Comparison with other baselines on multi-operation editing.
@misc{zheng2023minigpt5,
- title={MiniGPT-5: Interleaved Vision-and-Language Generation via Generative Vokens},
- author={Kaizhi Zheng and Xuehai He and Xin Eric Wang},
- year={2023},
- journal={arXiv preprint arXiv:2310.02239}
+ @misc{zheng2024editroomllmparameterizedgraphdiffusion,
+ title={EditRoom: LLM-parameterized Graph Diffusion for Composable 3D Room Layout Editing},
+ author={Kaizhi Zheng and Xiaotong Chen and Xuehai He and Jing Gu and Linjie Li and Zhengyuan Yang and Kevin Lin and Jianfeng Wang and Lijuan Wang and Xin Eric Wang},
+ year={2024},
+ eprint={2410.12836},
+ archivePrefix={arXiv},
+ primaryClass={cs.GR},
+ url={https://arxiv.org/abs/2410.12836},
}