Skip to content

Instantly share code, notes, and snippets.

@oscarnevarezleal
Created March 12, 2023 22:41
Show Gist options
  • Save oscarnevarezleal/a0ce5834b0d52915095f005ca9f73a80 to your computer and use it in GitHub Desktop.
Save oscarnevarezleal/a0ce5834b0d52915095f005ca9f73a80 to your computer and use it in GitHub Desktop.
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=down_blocks.0.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=down_blocks.0.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=down_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=down_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=down_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=down_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.1.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.1.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.1.attentions.2.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.2.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.2.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.2.attentions.2.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.3.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.3.attentions.1.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=up_blocks.3.attentions.2.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_q_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_k_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_v_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn1.processor.to_out_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_q_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_k_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_v_lora.up.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.down.weight
key=mid_block.attentions.0.transformer_blocks.0.attn2.processor.to_out_lora.up.weight
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment