Introducing LoRA: A faster way to fine-tune Stable Diffusion
Blog post from Replicate
LoRA, or Low-Rank Adaptation, is a novel technique applied to Stable Diffusion for image generation, offering a faster and more efficient alternative to the previously released DreamBooth. Developed by Microsoft researchers and implemented by Simo Ryu, LoRA allows users to train models using only a few images, producing smaller output files around 5MB in just eight minutes, compared to DreamBooth's larger files and longer training times. LoRA's approach reduces the number of trainable parameters by creating a "diff" of the model rather than saving the entire model, facilitating easy sharing, storage, and reuse. While LoRA excels in generating styles and combining multiple concepts within a single image, it struggles with accurately rendering faces, often landing in the uncanny valley. Users can train their own LoRA concepts by uploading images to a public URL and using pre-set or customizable training models on Replicate, where predictions can be generated instantly without cold boots. The process involves gathering training images, uploading them, training the concept, and using LoRA's prediction model to generate new images. Future updates promise support for Stable Diffusion 2.1 and additional features, with a community space for sharing models and ideas on Discord.