Multi-LoRA Composition for Image Generation

February 26, 2024 ยท Entered Twilight ยท ๐Ÿ› Trans. Mach. Learn. Res.

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

"No code URL or promise found in abstract"
"Derived repo from GitHub Pages (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, README.md, analyze.py, anime_lora_info.json, callbacks.py, compose_anime.sh, compose_lora.py, compose_reality.sh, docs, eval.sh, evaluate.py, example.png, example.py, human_eval, images, lcm_example.py, lcm_lora_example.py, models, pipelines, reality_lora_info.json, requirements.txt, sdxl_example.py, utils.py

Authors Ming Zhong, Yelong Shen, Shuohang Wang, Yadong Lu, Yizhu Jiao, Siru Ouyang, Donghan Yu, Jiawei Han, Weizhu Chen arXiv ID 2402.16843 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.CL, cs.GR, cs.LG Citations 76 Venue Trans. Mach. Learn. Res. Repository https://github.com/maszhongming/Multi-LoRA-Composition. โญ 492 Last Checked 11 days ago
Abstract
Low-Rank Adaptation (LoRA) is extensively utilized in text-to-image models for the accurate rendition of specific elements like distinct characters or unique styles in generated images. Nonetheless, existing methods face challenges in effectively composing multiple LoRAs, especially as the number of LoRAs to be integrated grows, thus hindering the creation of complex imagery. In this paper, we study multi-LoRA composition through a decoding-centric perspective. We present two training-free methods: LoRA Switch, which alternates between different LoRAs at each denoising step, and LoRA Composite, which simultaneously incorporates all LoRAs to guide more cohesive image synthesis. To evaluate the proposed approaches, we establish ComposLoRA, a new comprehensive testbed as part of this research. It features a diverse range of LoRA categories with 480 composition sets. Utilizing an evaluation framework based on GPT-4V, our findings demonstrate a clear improvement in performance with our methods over the prevalent baseline, particularly evident when increasing the number of LoRAs in a composition. The code, benchmarks, LoRA weights, and all evaluation details are available on our project website: https://maszhongming.github.io/Multi-LoRA-Composition.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision