Skip to content

Commit 31058cd

Browse files
sayakpaulstevhliu
andauthored
[LoRA] allow loras to be loaded with low_cpu_mem_usage. (#9510)
* allow loras to be loaded with low_cpu_mem_usage. * add flux support but note https://github.com/huggingface/diffusers/pull/9510\#issuecomment-2378316687 * low_cpu_mem_usage. * fix-copies * fix-copies again * tests * _LOW_CPU_MEM_USAGE_DEFAULT_LORA * _peft_version default. * version checks. * version check. * version check. * version check. * require peft 0.13.1. * explicitly specify low_cpu_mem_usage=False. * docs. * transformers version 4.45.2. * update * fix * empty * better name initialize_dummy_state_dict. * doc todos. * Apply suggestions from code review Co-authored-by: Steven Liu <[email protected]> * style * fix-copies --------- Co-authored-by: Steven Liu <[email protected]>
1 parent ec9e526 commit 31058cd

File tree

5 files changed

+411
-24
lines changed

5 files changed

+411
-24
lines changed

docs/source/en/tutorials/using_peft_for_inference.md

+6
Original file line numberDiff line numberDiff line change
@@ -75,6 +75,12 @@ image
7575

7676
![pixel-art](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/peft_integration/diffusers_peft_lora_inference_12_1.png)
7777

78+
<Tip>
79+
80+
By default, if the most up-to-date versions of PEFT and Transformers are detected, `low_cpu_mem_usage` is set to `True` to speed up the loading time of LoRA checkpoints.
81+
82+
</Tip>
83+
7884
## Merge adapters
7985

8086
You can also merge different adapter checkpoints for inference to blend their styles together.

0 commit comments

Comments
 (0)