Low-rank Adaptation (LoRA) models have revolutionized the personalization of pre-trained diffusion models by enabling fine-tuning through low-rank, factorized weight matrices specifically optimized for attention layers. These models facilitate the generation of highly customized content across a variety of objects, individuals, and artistic styles without the need for extensive retraining. Despite the availability of over 100K LoRA adapters on platforms like Civit.ai, users often face challenges in navigating, selecting, and effectively utilizing the most suitable adapters due to their sheer volume, diversity, and lack of structured organization. This paper addresses the problem of selecting the most relevant and diverse LoRA models from this vast database by framing the task as a combinatorial optimization problem and proposing a novel submodular framework. Our quantitative and qualitative experiments demonstrate that our method generates diverse outputs across a wide range of domains.
翻译:低秩适配(LoRA)模型通过针对注意力层优化的低秩分解权重矩阵实现微调,彻底改变了预训练扩散模型的个性化定制。这些模型能够生成涵盖各类物体、人物及艺术风格的高度定制化内容,无需进行大规模重新训练。尽管在Civit.ai等平台上已有超过10万个LoRA适配器可用,但由于其数量庞大、种类繁多且缺乏结构化组织,用户在导航、选择和有效利用最适配的模型时仍面临挑战。本文通过将该任务构建为组合优化问题,并提出一种新颖的子模框架,解决了从这一庞大数据库中选取最相关且多样化LoRA模型的问题。我们的定量与定性实验表明,该方法能在广泛领域内生成多样化的输出结果。