mirror of
https://github.com/ostris/ai-toolkit.git
synced 2026-04-27 09:44:02 +00:00
fixed issues with converting and saving models. Cleaned keys. Improved testing for cycle load saving.
This commit is contained in:
@@ -4,8 +4,6 @@
|
||||
"cond_stage_model.model.ln_final.weight": "te_text_model.final_layer_norm.weight",
|
||||
"cond_stage_model.model.positional_embedding": "te_text_model.embeddings.position_embedding.weight",
|
||||
"cond_stage_model.model.token_embedding.weight": "te_text_model.embeddings.token_embedding.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.0.attn.in_proj_bias": "te_text_model.encoder.layers.0.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.0.attn.in_proj_weight": "te_text_model.encoder.layers.0.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.0.attn.out_proj.bias": "te_text_model.encoder.layers.0.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.0.attn.out_proj.weight": "te_text_model.encoder.layers.0.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.0.ln_1.bias": "te_text_model.encoder.layers.0.layer_norm1.bias",
|
||||
@@ -16,8 +14,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.0.mlp.c_fc.weight": "te_text_model.encoder.layers.0.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.0.mlp.c_proj.bias": "te_text_model.encoder.layers.0.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.0.mlp.c_proj.weight": "te_text_model.encoder.layers.0.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.1.attn.in_proj_bias": "te_text_model.encoder.layers.1.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.1.attn.in_proj_weight": "te_text_model.encoder.layers.1.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.1.attn.out_proj.bias": "te_text_model.encoder.layers.1.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.1.attn.out_proj.weight": "te_text_model.encoder.layers.1.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.1.ln_1.bias": "te_text_model.encoder.layers.1.layer_norm1.bias",
|
||||
@@ -28,8 +24,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.1.mlp.c_fc.weight": "te_text_model.encoder.layers.1.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.1.mlp.c_proj.bias": "te_text_model.encoder.layers.1.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.1.mlp.c_proj.weight": "te_text_model.encoder.layers.1.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.10.attn.in_proj_bias": "te_text_model.encoder.layers.10.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.10.attn.in_proj_weight": "te_text_model.encoder.layers.10.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.10.attn.out_proj.bias": "te_text_model.encoder.layers.10.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.10.attn.out_proj.weight": "te_text_model.encoder.layers.10.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.10.ln_1.bias": "te_text_model.encoder.layers.10.layer_norm1.bias",
|
||||
@@ -40,8 +34,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.10.mlp.c_fc.weight": "te_text_model.encoder.layers.10.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.10.mlp.c_proj.bias": "te_text_model.encoder.layers.10.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.10.mlp.c_proj.weight": "te_text_model.encoder.layers.10.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.11.attn.in_proj_bias": "te_text_model.encoder.layers.11.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.11.attn.in_proj_weight": "te_text_model.encoder.layers.11.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.11.attn.out_proj.bias": "te_text_model.encoder.layers.11.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.11.attn.out_proj.weight": "te_text_model.encoder.layers.11.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.11.ln_1.bias": "te_text_model.encoder.layers.11.layer_norm1.bias",
|
||||
@@ -52,8 +44,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.11.mlp.c_fc.weight": "te_text_model.encoder.layers.11.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.11.mlp.c_proj.bias": "te_text_model.encoder.layers.11.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.11.mlp.c_proj.weight": "te_text_model.encoder.layers.11.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.12.attn.in_proj_bias": "te_text_model.encoder.layers.12.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.12.attn.in_proj_weight": "te_text_model.encoder.layers.12.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.12.attn.out_proj.bias": "te_text_model.encoder.layers.12.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.12.attn.out_proj.weight": "te_text_model.encoder.layers.12.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.12.ln_1.bias": "te_text_model.encoder.layers.12.layer_norm1.bias",
|
||||
@@ -64,8 +54,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.12.mlp.c_fc.weight": "te_text_model.encoder.layers.12.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.12.mlp.c_proj.bias": "te_text_model.encoder.layers.12.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.12.mlp.c_proj.weight": "te_text_model.encoder.layers.12.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.13.attn.in_proj_bias": "te_text_model.encoder.layers.13.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.13.attn.in_proj_weight": "te_text_model.encoder.layers.13.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.13.attn.out_proj.bias": "te_text_model.encoder.layers.13.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.13.attn.out_proj.weight": "te_text_model.encoder.layers.13.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.13.ln_1.bias": "te_text_model.encoder.layers.13.layer_norm1.bias",
|
||||
@@ -76,8 +64,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.13.mlp.c_fc.weight": "te_text_model.encoder.layers.13.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.13.mlp.c_proj.bias": "te_text_model.encoder.layers.13.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.13.mlp.c_proj.weight": "te_text_model.encoder.layers.13.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.14.attn.in_proj_bias": "te_text_model.encoder.layers.14.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.14.attn.in_proj_weight": "te_text_model.encoder.layers.14.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.14.attn.out_proj.bias": "te_text_model.encoder.layers.14.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.14.attn.out_proj.weight": "te_text_model.encoder.layers.14.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.14.ln_1.bias": "te_text_model.encoder.layers.14.layer_norm1.bias",
|
||||
@@ -88,8 +74,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.14.mlp.c_fc.weight": "te_text_model.encoder.layers.14.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.14.mlp.c_proj.bias": "te_text_model.encoder.layers.14.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.14.mlp.c_proj.weight": "te_text_model.encoder.layers.14.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.15.attn.in_proj_bias": "te_text_model.encoder.layers.15.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.15.attn.in_proj_weight": "te_text_model.encoder.layers.15.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.15.attn.out_proj.bias": "te_text_model.encoder.layers.15.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.15.attn.out_proj.weight": "te_text_model.encoder.layers.15.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.15.ln_1.bias": "te_text_model.encoder.layers.15.layer_norm1.bias",
|
||||
@@ -100,8 +84,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.15.mlp.c_fc.weight": "te_text_model.encoder.layers.15.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.15.mlp.c_proj.bias": "te_text_model.encoder.layers.15.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.15.mlp.c_proj.weight": "te_text_model.encoder.layers.15.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.16.attn.in_proj_bias": "te_text_model.encoder.layers.16.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.16.attn.in_proj_weight": "te_text_model.encoder.layers.16.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.16.attn.out_proj.bias": "te_text_model.encoder.layers.16.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.16.attn.out_proj.weight": "te_text_model.encoder.layers.16.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.16.ln_1.bias": "te_text_model.encoder.layers.16.layer_norm1.bias",
|
||||
@@ -112,8 +94,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.16.mlp.c_fc.weight": "te_text_model.encoder.layers.16.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.16.mlp.c_proj.bias": "te_text_model.encoder.layers.16.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.16.mlp.c_proj.weight": "te_text_model.encoder.layers.16.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.17.attn.in_proj_bias": "te_text_model.encoder.layers.17.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.17.attn.in_proj_weight": "te_text_model.encoder.layers.17.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.17.attn.out_proj.bias": "te_text_model.encoder.layers.17.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.17.attn.out_proj.weight": "te_text_model.encoder.layers.17.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.17.ln_1.bias": "te_text_model.encoder.layers.17.layer_norm1.bias",
|
||||
@@ -124,8 +104,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.17.mlp.c_fc.weight": "te_text_model.encoder.layers.17.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.17.mlp.c_proj.bias": "te_text_model.encoder.layers.17.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.17.mlp.c_proj.weight": "te_text_model.encoder.layers.17.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.18.attn.in_proj_bias": "te_text_model.encoder.layers.18.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.18.attn.in_proj_weight": "te_text_model.encoder.layers.18.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.18.attn.out_proj.bias": "te_text_model.encoder.layers.18.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.18.attn.out_proj.weight": "te_text_model.encoder.layers.18.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.18.ln_1.bias": "te_text_model.encoder.layers.18.layer_norm1.bias",
|
||||
@@ -136,8 +114,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.18.mlp.c_fc.weight": "te_text_model.encoder.layers.18.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.18.mlp.c_proj.bias": "te_text_model.encoder.layers.18.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.18.mlp.c_proj.weight": "te_text_model.encoder.layers.18.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.19.attn.in_proj_bias": "te_text_model.encoder.layers.19.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.19.attn.in_proj_weight": "te_text_model.encoder.layers.19.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.19.attn.out_proj.bias": "te_text_model.encoder.layers.19.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.19.attn.out_proj.weight": "te_text_model.encoder.layers.19.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.19.ln_1.bias": "te_text_model.encoder.layers.19.layer_norm1.bias",
|
||||
@@ -148,8 +124,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.19.mlp.c_fc.weight": "te_text_model.encoder.layers.19.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.19.mlp.c_proj.bias": "te_text_model.encoder.layers.19.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.19.mlp.c_proj.weight": "te_text_model.encoder.layers.19.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.2.attn.in_proj_bias": "te_text_model.encoder.layers.2.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.2.attn.in_proj_weight": "te_text_model.encoder.layers.2.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.2.attn.out_proj.bias": "te_text_model.encoder.layers.2.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.2.attn.out_proj.weight": "te_text_model.encoder.layers.2.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.2.ln_1.bias": "te_text_model.encoder.layers.2.layer_norm1.bias",
|
||||
@@ -160,8 +134,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.2.mlp.c_fc.weight": "te_text_model.encoder.layers.2.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.2.mlp.c_proj.bias": "te_text_model.encoder.layers.2.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.2.mlp.c_proj.weight": "te_text_model.encoder.layers.2.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.20.attn.in_proj_bias": "te_text_model.encoder.layers.20.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.20.attn.in_proj_weight": "te_text_model.encoder.layers.20.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.20.attn.out_proj.bias": "te_text_model.encoder.layers.20.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.20.attn.out_proj.weight": "te_text_model.encoder.layers.20.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.20.ln_1.bias": "te_text_model.encoder.layers.20.layer_norm1.bias",
|
||||
@@ -172,8 +144,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.20.mlp.c_fc.weight": "te_text_model.encoder.layers.20.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.20.mlp.c_proj.bias": "te_text_model.encoder.layers.20.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.20.mlp.c_proj.weight": "te_text_model.encoder.layers.20.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.21.attn.in_proj_bias": "te_text_model.encoder.layers.21.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.21.attn.in_proj_weight": "te_text_model.encoder.layers.21.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.21.attn.out_proj.bias": "te_text_model.encoder.layers.21.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.21.attn.out_proj.weight": "te_text_model.encoder.layers.21.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.21.ln_1.bias": "te_text_model.encoder.layers.21.layer_norm1.bias",
|
||||
@@ -184,8 +154,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.21.mlp.c_fc.weight": "te_text_model.encoder.layers.21.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.21.mlp.c_proj.bias": "te_text_model.encoder.layers.21.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.21.mlp.c_proj.weight": "te_text_model.encoder.layers.21.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.22.attn.in_proj_bias": "te_text_model.encoder.layers.22.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.22.attn.in_proj_weight": "te_text_model.encoder.layers.22.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.22.attn.out_proj.bias": "te_text_model.encoder.layers.22.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.22.attn.out_proj.weight": "te_text_model.encoder.layers.22.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.22.ln_1.bias": "te_text_model.encoder.layers.22.layer_norm1.bias",
|
||||
@@ -196,8 +164,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.22.mlp.c_fc.weight": "te_text_model.encoder.layers.22.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.22.mlp.c_proj.bias": "te_text_model.encoder.layers.22.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.22.mlp.c_proj.weight": "te_text_model.encoder.layers.22.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.3.attn.in_proj_bias": "te_text_model.encoder.layers.3.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.3.attn.in_proj_weight": "te_text_model.encoder.layers.3.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.3.attn.out_proj.bias": "te_text_model.encoder.layers.3.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.3.attn.out_proj.weight": "te_text_model.encoder.layers.3.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.3.ln_1.bias": "te_text_model.encoder.layers.3.layer_norm1.bias",
|
||||
@@ -208,8 +174,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.3.mlp.c_fc.weight": "te_text_model.encoder.layers.3.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.3.mlp.c_proj.bias": "te_text_model.encoder.layers.3.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.3.mlp.c_proj.weight": "te_text_model.encoder.layers.3.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.4.attn.in_proj_bias": "te_text_model.encoder.layers.4.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.4.attn.in_proj_weight": "te_text_model.encoder.layers.4.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.4.attn.out_proj.bias": "te_text_model.encoder.layers.4.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.4.attn.out_proj.weight": "te_text_model.encoder.layers.4.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.4.ln_1.bias": "te_text_model.encoder.layers.4.layer_norm1.bias",
|
||||
@@ -220,8 +184,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.4.mlp.c_fc.weight": "te_text_model.encoder.layers.4.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.4.mlp.c_proj.bias": "te_text_model.encoder.layers.4.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.4.mlp.c_proj.weight": "te_text_model.encoder.layers.4.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.5.attn.in_proj_bias": "te_text_model.encoder.layers.5.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.5.attn.in_proj_weight": "te_text_model.encoder.layers.5.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.5.attn.out_proj.bias": "te_text_model.encoder.layers.5.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.5.attn.out_proj.weight": "te_text_model.encoder.layers.5.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.5.ln_1.bias": "te_text_model.encoder.layers.5.layer_norm1.bias",
|
||||
@@ -232,8 +194,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.5.mlp.c_fc.weight": "te_text_model.encoder.layers.5.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.5.mlp.c_proj.bias": "te_text_model.encoder.layers.5.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.5.mlp.c_proj.weight": "te_text_model.encoder.layers.5.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.6.attn.in_proj_bias": "te_text_model.encoder.layers.6.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.6.attn.in_proj_weight": "te_text_model.encoder.layers.6.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.6.attn.out_proj.bias": "te_text_model.encoder.layers.6.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.6.attn.out_proj.weight": "te_text_model.encoder.layers.6.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.6.ln_1.bias": "te_text_model.encoder.layers.6.layer_norm1.bias",
|
||||
@@ -244,8 +204,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.6.mlp.c_fc.weight": "te_text_model.encoder.layers.6.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.6.mlp.c_proj.bias": "te_text_model.encoder.layers.6.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.6.mlp.c_proj.weight": "te_text_model.encoder.layers.6.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.7.attn.in_proj_bias": "te_text_model.encoder.layers.7.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.7.attn.in_proj_weight": "te_text_model.encoder.layers.7.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.7.attn.out_proj.bias": "te_text_model.encoder.layers.7.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.7.attn.out_proj.weight": "te_text_model.encoder.layers.7.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.7.ln_1.bias": "te_text_model.encoder.layers.7.layer_norm1.bias",
|
||||
@@ -256,8 +214,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.7.mlp.c_fc.weight": "te_text_model.encoder.layers.7.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.7.mlp.c_proj.bias": "te_text_model.encoder.layers.7.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.7.mlp.c_proj.weight": "te_text_model.encoder.layers.7.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.8.attn.in_proj_bias": "te_text_model.encoder.layers.8.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.8.attn.in_proj_weight": "te_text_model.encoder.layers.8.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.8.attn.out_proj.bias": "te_text_model.encoder.layers.8.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.8.attn.out_proj.weight": "te_text_model.encoder.layers.8.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.8.ln_1.bias": "te_text_model.encoder.layers.8.layer_norm1.bias",
|
||||
@@ -268,8 +224,6 @@
|
||||
"cond_stage_model.model.transformer.resblocks.8.mlp.c_fc.weight": "te_text_model.encoder.layers.8.mlp.fc1.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.8.mlp.c_proj.bias": "te_text_model.encoder.layers.8.mlp.fc2.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.8.mlp.c_proj.weight": "te_text_model.encoder.layers.8.mlp.fc2.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.9.attn.in_proj_bias": "te_text_model.encoder.layers.9.self_attn.MERGED.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.9.attn.in_proj_weight": "te_text_model.encoder.layers.9.self_attn.MERGED.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.9.attn.out_proj.bias": "te_text_model.encoder.layers.9.self_attn.out_proj.bias",
|
||||
"cond_stage_model.model.transformer.resblocks.9.attn.out_proj.weight": "te_text_model.encoder.layers.9.self_attn.out_proj.weight",
|
||||
"cond_stage_model.model.transformer.resblocks.9.ln_1.bias": "te_text_model.encoder.layers.9.layer_norm1.bias",
|
||||
@@ -1264,62 +1218,6 @@
|
||||
512
|
||||
]
|
||||
],
|
||||
"first_stage_model.decoder.up.0.block.0.nin_shortcut.weight": [
|
||||
[
|
||||
128,
|
||||
256,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
128,
|
||||
256,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"first_stage_model.decoder.up.1.block.0.nin_shortcut.weight": [
|
||||
[
|
||||
256,
|
||||
512,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
256,
|
||||
512,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"first_stage_model.encoder.down.1.block.0.nin_shortcut.weight": [
|
||||
[
|
||||
256,
|
||||
128,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
256,
|
||||
128,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"first_stage_model.encoder.down.2.block.0.nin_shortcut.weight": [
|
||||
[
|
||||
512,
|
||||
256,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
512,
|
||||
256,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"first_stage_model.encoder.mid.attn_1.k.weight": [
|
||||
[
|
||||
512,
|
||||
@@ -1367,230 +1265,6 @@
|
||||
512,
|
||||
512
|
||||
]
|
||||
],
|
||||
"first_stage_model.post_quant_conv.weight": [
|
||||
[
|
||||
4,
|
||||
4,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
4,
|
||||
4,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"first_stage_model.quant_conv.weight": [
|
||||
[
|
||||
8,
|
||||
8,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
8,
|
||||
8,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.input_blocks.4.0.skip_connection.weight": [
|
||||
[
|
||||
640,
|
||||
320,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
640,
|
||||
320,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.input_blocks.7.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
640,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
640,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.0.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.1.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.10.0.skip_connection.weight": [
|
||||
[
|
||||
320,
|
||||
640,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
320,
|
||||
640,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.11.0.skip_connection.weight": [
|
||||
[
|
||||
320,
|
||||
640,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
320,
|
||||
640,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.2.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.3.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.4.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
2560,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.5.0.skip_connection.weight": [
|
||||
[
|
||||
1280,
|
||||
1920,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
1280,
|
||||
1920,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.6.0.skip_connection.weight": [
|
||||
[
|
||||
640,
|
||||
1920,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
640,
|
||||
1920,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.7.0.skip_connection.weight": [
|
||||
[
|
||||
640,
|
||||
1280,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
640,
|
||||
1280,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.8.0.skip_connection.weight": [
|
||||
[
|
||||
640,
|
||||
960,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
640,
|
||||
960,
|
||||
1,
|
||||
1
|
||||
]
|
||||
],
|
||||
"model.diffusion_model.output_blocks.9.0.skip_connection.weight": [
|
||||
[
|
||||
320,
|
||||
960,
|
||||
1,
|
||||
1
|
||||
],
|
||||
[
|
||||
320,
|
||||
960,
|
||||
1,
|
||||
1
|
||||
]
|
||||
]
|
||||
},
|
||||
"ldm_diffusers_operator_map": {
|
||||
@@ -1606,8 +1280,7 @@
|
||||
"te_text_model.encoder.layers.0.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.0.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.0.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.0.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.1.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1621,8 +1294,7 @@
|
||||
"te_text_model.encoder.layers.1.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.1.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.1.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.1.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.10.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1636,8 +1308,7 @@
|
||||
"te_text_model.encoder.layers.10.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.10.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.10.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.10.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.11.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1651,8 +1322,7 @@
|
||||
"te_text_model.encoder.layers.11.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.11.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.11.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.11.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.12.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1666,8 +1336,7 @@
|
||||
"te_text_model.encoder.layers.12.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.12.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.12.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.12.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.13.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1681,8 +1350,7 @@
|
||||
"te_text_model.encoder.layers.13.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.13.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.13.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.13.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.14.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1696,8 +1364,7 @@
|
||||
"te_text_model.encoder.layers.14.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.14.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.14.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.14.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.15.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1711,8 +1378,7 @@
|
||||
"te_text_model.encoder.layers.15.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.15.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.15.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.15.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.16.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1726,8 +1392,7 @@
|
||||
"te_text_model.encoder.layers.16.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.16.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.16.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.16.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.17.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1741,8 +1406,7 @@
|
||||
"te_text_model.encoder.layers.17.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.17.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.17.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.17.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.18.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1756,8 +1420,7 @@
|
||||
"te_text_model.encoder.layers.18.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.18.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.18.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.18.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.19.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1771,8 +1434,7 @@
|
||||
"te_text_model.encoder.layers.19.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.19.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.19.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.19.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.2.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1786,8 +1448,7 @@
|
||||
"te_text_model.encoder.layers.2.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.2.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.2.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.2.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.20.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1801,8 +1462,7 @@
|
||||
"te_text_model.encoder.layers.20.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.20.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.20.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.20.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.21.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1816,8 +1476,7 @@
|
||||
"te_text_model.encoder.layers.21.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.21.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.21.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.21.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.22.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1831,8 +1490,7 @@
|
||||
"te_text_model.encoder.layers.22.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.22.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.22.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.22.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.3.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1846,8 +1504,7 @@
|
||||
"te_text_model.encoder.layers.3.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.3.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.3.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.3.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.4.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1861,8 +1518,7 @@
|
||||
"te_text_model.encoder.layers.4.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.4.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.4.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.4.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.5.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1876,8 +1532,7 @@
|
||||
"te_text_model.encoder.layers.5.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.5.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.5.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.5.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.6.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1891,8 +1546,7 @@
|
||||
"te_text_model.encoder.layers.6.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.6.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.6.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.6.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.7.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1906,8 +1560,7 @@
|
||||
"te_text_model.encoder.layers.7.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.7.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.7.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.7.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.8.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1921,8 +1574,7 @@
|
||||
"te_text_model.encoder.layers.8.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.8.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.8.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.8.self_attn.MERGED.weight"
|
||||
]
|
||||
},
|
||||
"cond_stage_model.model.transformer.resblocks.9.attn.in_proj_bias": {
|
||||
"cat": [
|
||||
@@ -1936,8 +1588,7 @@
|
||||
"te_text_model.encoder.layers.9.self_attn.q_proj.weight",
|
||||
"te_text_model.encoder.layers.9.self_attn.k_proj.weight",
|
||||
"te_text_model.encoder.layers.9.self_attn.v_proj.weight"
|
||||
],
|
||||
"target": "te_text_model.encoder.layers.9.self_attn.MERGED.weight"
|
||||
]
|
||||
}
|
||||
},
|
||||
"diffusers_ldm_operator_map": {
|
||||
|
||||
Reference in New Issue
Block a user