-
Notifications
You must be signed in to change notification settings - Fork 5.9k
[docs] Adapters #11331
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
[docs] Adapters #11331
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looking quite good. Wondering if merging of info from the different pages in the single page is okay for our users?
- sections: | ||
- local: tutorials/using_peft_for_inference | ||
title: LoRA | ||
- local: using-diffusers/ip_adapter | ||
title: IP-Adapter | ||
- local: using-diffusers/dreambooth | ||
title: DreamBooth | ||
- local: using-diffusers/textual_inversion_inference | ||
title: Textual inversion | ||
title: Adapters |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks crisp. Should we automatically fold this when a user lands on https://huggingface.co/docs/diffusers/main/en/index?
|
||
pipe_id = "stabilityai/stable-diffusion-xl-base-1.0" | ||
pipe = DiffusionPipeline.from_pretrained(pipe_id, torch_dtype=torch.float16).to("cuda") | ||
[torch.compile](../optimization/torch2.0#torchcompile) speeds up inference by compiling the PyTorch model to use optimized kernels. Before compiling, the LoRA weights need to be fused into the base model and unloaded first. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
[torch.compile](../optimization/torch2.0#torchcompile) speeds up inference by compiling the PyTorch model to use optimized kernels. Before compiling, the LoRA weights need to be fused into the base model and unloaded first. | |
[torch.compile](../optimization/torch2.0#torchcompile) speeds up inference by compiling the PyTorch model to benefit from PyTorch's low-level hardware-dependent optimizations. Before compiling, the LoRA weights need to be fused into the base model and unloaded first to take the most advantage of `torch.compile()`. |
Wondering if we should also link to the hotswap
docs?
Refactors the adapter docs.