Closed
Description
Apologies if API Design is not the right tag,
I was reading the docs for this page https://huggingface.co/docs/diffusers/training/lora and for a number of reasons it seemed very confusing to me as a end-user:
- At the top of the page "Currently, LoRA is only supported for the attention layers of the UNet2DConditionalModel." I don't understand what the implications of this are or if I should care
- The docs do not mention anything about creating my own dataset, only how to make pokemons...
- It does not specify whether or not we need captions for each image. With DreamBooth, each fine-tune was a single concept and hijacked an existing word. I see mentions of BLIP captions so I believe LoRA has this capability of adjusting of nuanced prompt and if so it should be stated on the page.
- After the training command before moving onto inference, the docs should let me know around how long it will take so I can plan accordingly, and exactly what will happen which files will be created, etc.
Cheers, this library is definitely the best API for using deep learning models of any kind