diffsynth 1.0.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- diffsynth/__init__.py +6 -0
- diffsynth/configs/__init__.py +0 -0
- diffsynth/configs/model_config.py +243 -0
- diffsynth/controlnets/__init__.py +2 -0
- diffsynth/controlnets/controlnet_unit.py +53 -0
- diffsynth/controlnets/processors.py +51 -0
- diffsynth/data/__init__.py +1 -0
- diffsynth/data/simple_text_image.py +35 -0
- diffsynth/data/video.py +148 -0
- diffsynth/extensions/ESRGAN/__init__.py +118 -0
- diffsynth/extensions/FastBlend/__init__.py +63 -0
- diffsynth/extensions/FastBlend/api.py +397 -0
- diffsynth/extensions/FastBlend/cupy_kernels.py +119 -0
- diffsynth/extensions/FastBlend/data.py +146 -0
- diffsynth/extensions/FastBlend/patch_match.py +298 -0
- diffsynth/extensions/FastBlend/runners/__init__.py +4 -0
- diffsynth/extensions/FastBlend/runners/accurate.py +35 -0
- diffsynth/extensions/FastBlend/runners/balanced.py +46 -0
- diffsynth/extensions/FastBlend/runners/fast.py +141 -0
- diffsynth/extensions/FastBlend/runners/interpolation.py +121 -0
- diffsynth/extensions/RIFE/__init__.py +242 -0
- diffsynth/extensions/__init__.py +0 -0
- diffsynth/models/__init__.py +1 -0
- diffsynth/models/attention.py +89 -0
- diffsynth/models/downloader.py +66 -0
- diffsynth/models/hunyuan_dit.py +451 -0
- diffsynth/models/hunyuan_dit_text_encoder.py +163 -0
- diffsynth/models/kolors_text_encoder.py +1363 -0
- diffsynth/models/lora.py +195 -0
- diffsynth/models/model_manager.py +536 -0
- diffsynth/models/sd3_dit.py +798 -0
- diffsynth/models/sd3_text_encoder.py +1107 -0
- diffsynth/models/sd3_vae_decoder.py +81 -0
- diffsynth/models/sd3_vae_encoder.py +95 -0
- diffsynth/models/sd_controlnet.py +588 -0
- diffsynth/models/sd_ipadapter.py +57 -0
- diffsynth/models/sd_motion.py +199 -0
- diffsynth/models/sd_text_encoder.py +321 -0
- diffsynth/models/sd_unet.py +1108 -0
- diffsynth/models/sd_vae_decoder.py +336 -0
- diffsynth/models/sd_vae_encoder.py +282 -0
- diffsynth/models/sdxl_ipadapter.py +122 -0
- diffsynth/models/sdxl_motion.py +104 -0
- diffsynth/models/sdxl_text_encoder.py +759 -0
- diffsynth/models/sdxl_unet.py +1899 -0
- diffsynth/models/sdxl_vae_decoder.py +24 -0
- diffsynth/models/sdxl_vae_encoder.py +24 -0
- diffsynth/models/svd_image_encoder.py +505 -0
- diffsynth/models/svd_unet.py +2004 -0
- diffsynth/models/svd_vae_decoder.py +578 -0
- diffsynth/models/svd_vae_encoder.py +139 -0
- diffsynth/models/tiler.py +106 -0
- diffsynth/pipelines/__init__.py +9 -0
- diffsynth/pipelines/base.py +34 -0
- diffsynth/pipelines/dancer.py +178 -0
- diffsynth/pipelines/hunyuan_image.py +274 -0
- diffsynth/pipelines/pipeline_runner.py +105 -0
- diffsynth/pipelines/sd3_image.py +132 -0
- diffsynth/pipelines/sd_image.py +173 -0
- diffsynth/pipelines/sd_video.py +266 -0
- diffsynth/pipelines/sdxl_image.py +191 -0
- diffsynth/pipelines/sdxl_video.py +223 -0
- diffsynth/pipelines/svd_video.py +297 -0
- diffsynth/processors/FastBlend.py +142 -0
- diffsynth/processors/PILEditor.py +28 -0
- diffsynth/processors/RIFE.py +77 -0
- diffsynth/processors/__init__.py +0 -0
- diffsynth/processors/base.py +6 -0
- diffsynth/processors/sequencial_processor.py +41 -0
- diffsynth/prompters/__init__.py +6 -0
- diffsynth/prompters/base_prompter.py +57 -0
- diffsynth/prompters/hunyuan_dit_prompter.py +69 -0
- diffsynth/prompters/kolors_prompter.py +353 -0
- diffsynth/prompters/prompt_refiners.py +77 -0
- diffsynth/prompters/sd3_prompter.py +92 -0
- diffsynth/prompters/sd_prompter.py +73 -0
- diffsynth/prompters/sdxl_prompter.py +61 -0
- diffsynth/schedulers/__init__.py +3 -0
- diffsynth/schedulers/continuous_ode.py +59 -0
- diffsynth/schedulers/ddim.py +79 -0
- diffsynth/schedulers/flow_match.py +51 -0
- diffsynth/tokenizer_configs/__init__.py +0 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer/special_tokens_map.json +7 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer/tokenizer_config.json +16 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer/vocab.txt +47020 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer/vocab_org.txt +21128 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer_t5/config.json +28 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer_t5/special_tokens_map.json +1 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer_t5/spiece.model +0 -0
- diffsynth/tokenizer_configs/hunyuan_dit/tokenizer_t5/tokenizer_config.json +1 -0
- diffsynth/tokenizer_configs/kolors/tokenizer/tokenizer.model +0 -0
- diffsynth/tokenizer_configs/kolors/tokenizer/tokenizer_config.json +12 -0
- diffsynth/tokenizer_configs/kolors/tokenizer/vocab.txt +0 -0
- diffsynth/tokenizer_configs/stable_diffusion/tokenizer/merges.txt +48895 -0
- diffsynth/tokenizer_configs/stable_diffusion/tokenizer/special_tokens_map.json +24 -0
- diffsynth/tokenizer_configs/stable_diffusion/tokenizer/tokenizer_config.json +34 -0
- diffsynth/tokenizer_configs/stable_diffusion/tokenizer/vocab.json +49410 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_1/merges.txt +48895 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_1/special_tokens_map.json +30 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_1/tokenizer_config.json +30 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_1/vocab.json +49410 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_2/merges.txt +48895 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_2/special_tokens_map.json +30 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_2/tokenizer_config.json +38 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_2/vocab.json +49410 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_3/special_tokens_map.json +125 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_3/spiece.model +0 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_3/tokenizer.json +129428 -0
- diffsynth/tokenizer_configs/stable_diffusion_3/tokenizer_3/tokenizer_config.json +940 -0
- diffsynth/tokenizer_configs/stable_diffusion_xl/tokenizer_2/merges.txt +40213 -0
- diffsynth/tokenizer_configs/stable_diffusion_xl/tokenizer_2/special_tokens_map.json +24 -0
- diffsynth/tokenizer_configs/stable_diffusion_xl/tokenizer_2/tokenizer_config.json +38 -0
- diffsynth/tokenizer_configs/stable_diffusion_xl/tokenizer_2/vocab.json +49411 -0
- diffsynth/trainers/__init__.py +0 -0
- diffsynth/trainers/text_to_image.py +253 -0
- diffsynth-1.0.0.dist-info/LICENSE +201 -0
- diffsynth-1.0.0.dist-info/METADATA +23 -0
- diffsynth-1.0.0.dist-info/RECORD +120 -0
- diffsynth-1.0.0.dist-info/WHEEL +5 -0
- diffsynth-1.0.0.dist-info/top_level.txt +1 -0
|
@@ -0,0 +1,57 @@
|
|
|
1
|
+
from .svd_image_encoder import SVDImageEncoder
|
|
2
|
+
from .sdxl_ipadapter import IpAdapterImageProjModel, IpAdapterModule, SDXLIpAdapterStateDictConverter
|
|
3
|
+
from transformers import CLIPImageProcessor
|
|
4
|
+
import torch
|
|
5
|
+
|
|
6
|
+
|
|
7
|
+
class IpAdapterCLIPImageEmbedder(SVDImageEncoder):
|
|
8
|
+
def __init__(self):
|
|
9
|
+
super().__init__()
|
|
10
|
+
self.image_processor = CLIPImageProcessor()
|
|
11
|
+
|
|
12
|
+
def forward(self, image):
|
|
13
|
+
pixel_values = self.image_processor(images=image, return_tensors="pt").pixel_values
|
|
14
|
+
pixel_values = pixel_values.to(device=self.embeddings.class_embedding.device, dtype=self.embeddings.class_embedding.dtype)
|
|
15
|
+
return super().forward(pixel_values)
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
class SDIpAdapter(torch.nn.Module):
|
|
19
|
+
def __init__(self):
|
|
20
|
+
super().__init__()
|
|
21
|
+
shape_list = [(768, 320)] * 2 + [(768, 640)] * 2 + [(768, 1280)] * 5 + [(768, 640)] * 3 + [(768, 320)] * 3 + [(768, 1280)] * 1
|
|
22
|
+
self.ipadapter_modules = torch.nn.ModuleList([IpAdapterModule(*shape) for shape in shape_list])
|
|
23
|
+
self.image_proj = IpAdapterImageProjModel(cross_attention_dim=768, clip_embeddings_dim=1024, clip_extra_context_tokens=4)
|
|
24
|
+
self.set_full_adapter()
|
|
25
|
+
|
|
26
|
+
def set_full_adapter(self):
|
|
27
|
+
block_ids = [1, 4, 9, 12, 17, 20, 40, 43, 46, 50, 53, 56, 60, 63, 66, 29]
|
|
28
|
+
self.call_block_id = {(i, 0): j for j, i in enumerate(block_ids)}
|
|
29
|
+
|
|
30
|
+
def set_less_adapter(self):
|
|
31
|
+
# IP-Adapter for SD v1.5 doesn't support this feature.
|
|
32
|
+
self.set_full_adapter()
|
|
33
|
+
|
|
34
|
+
def forward(self, hidden_states, scale=1.0):
|
|
35
|
+
hidden_states = self.image_proj(hidden_states)
|
|
36
|
+
hidden_states = hidden_states.view(1, -1, hidden_states.shape[-1])
|
|
37
|
+
ip_kv_dict = {}
|
|
38
|
+
for (block_id, transformer_id) in self.call_block_id:
|
|
39
|
+
ipadapter_id = self.call_block_id[(block_id, transformer_id)]
|
|
40
|
+
ip_k, ip_v = self.ipadapter_modules[ipadapter_id](hidden_states)
|
|
41
|
+
if block_id not in ip_kv_dict:
|
|
42
|
+
ip_kv_dict[block_id] = {}
|
|
43
|
+
ip_kv_dict[block_id][transformer_id] = {
|
|
44
|
+
"ip_k": ip_k,
|
|
45
|
+
"ip_v": ip_v,
|
|
46
|
+
"scale": scale
|
|
47
|
+
}
|
|
48
|
+
return ip_kv_dict
|
|
49
|
+
|
|
50
|
+
@staticmethod
|
|
51
|
+
def state_dict_converter():
|
|
52
|
+
return SDIpAdapterStateDictConverter()
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
class SDIpAdapterStateDictConverter(SDXLIpAdapterStateDictConverter):
|
|
56
|
+
def __init__(self):
|
|
57
|
+
pass
|
|
@@ -0,0 +1,199 @@
|
|
|
1
|
+
from .sd_unet import SDUNet, Attention, GEGLU
|
|
2
|
+
import torch
|
|
3
|
+
from einops import rearrange, repeat
|
|
4
|
+
|
|
5
|
+
|
|
6
|
+
class TemporalTransformerBlock(torch.nn.Module):
|
|
7
|
+
|
|
8
|
+
def __init__(self, dim, num_attention_heads, attention_head_dim, max_position_embeddings=32):
|
|
9
|
+
super().__init__()
|
|
10
|
+
|
|
11
|
+
# 1. Self-Attn
|
|
12
|
+
self.pe1 = torch.nn.Parameter(torch.zeros(1, max_position_embeddings, dim))
|
|
13
|
+
self.norm1 = torch.nn.LayerNorm(dim, elementwise_affine=True)
|
|
14
|
+
self.attn1 = Attention(q_dim=dim, num_heads=num_attention_heads, head_dim=attention_head_dim, bias_out=True)
|
|
15
|
+
|
|
16
|
+
# 2. Cross-Attn
|
|
17
|
+
self.pe2 = torch.nn.Parameter(torch.zeros(1, max_position_embeddings, dim))
|
|
18
|
+
self.norm2 = torch.nn.LayerNorm(dim, elementwise_affine=True)
|
|
19
|
+
self.attn2 = Attention(q_dim=dim, num_heads=num_attention_heads, head_dim=attention_head_dim, bias_out=True)
|
|
20
|
+
|
|
21
|
+
# 3. Feed-forward
|
|
22
|
+
self.norm3 = torch.nn.LayerNorm(dim, elementwise_affine=True)
|
|
23
|
+
self.act_fn = GEGLU(dim, dim * 4)
|
|
24
|
+
self.ff = torch.nn.Linear(dim * 4, dim)
|
|
25
|
+
|
|
26
|
+
|
|
27
|
+
def forward(self, hidden_states, batch_size=1):
|
|
28
|
+
|
|
29
|
+
# 1. Self-Attention
|
|
30
|
+
norm_hidden_states = self.norm1(hidden_states)
|
|
31
|
+
norm_hidden_states = rearrange(norm_hidden_states, "(b f) h c -> (b h) f c", b=batch_size)
|
|
32
|
+
attn_output = self.attn1(norm_hidden_states + self.pe1[:, :norm_hidden_states.shape[1]])
|
|
33
|
+
attn_output = rearrange(attn_output, "(b h) f c -> (b f) h c", b=batch_size)
|
|
34
|
+
hidden_states = attn_output + hidden_states
|
|
35
|
+
|
|
36
|
+
# 2. Cross-Attention
|
|
37
|
+
norm_hidden_states = self.norm2(hidden_states)
|
|
38
|
+
norm_hidden_states = rearrange(norm_hidden_states, "(b f) h c -> (b h) f c", b=batch_size)
|
|
39
|
+
attn_output = self.attn2(norm_hidden_states + self.pe2[:, :norm_hidden_states.shape[1]])
|
|
40
|
+
attn_output = rearrange(attn_output, "(b h) f c -> (b f) h c", b=batch_size)
|
|
41
|
+
hidden_states = attn_output + hidden_states
|
|
42
|
+
|
|
43
|
+
# 3. Feed-forward
|
|
44
|
+
norm_hidden_states = self.norm3(hidden_states)
|
|
45
|
+
ff_output = self.act_fn(norm_hidden_states)
|
|
46
|
+
ff_output = self.ff(ff_output)
|
|
47
|
+
hidden_states = ff_output + hidden_states
|
|
48
|
+
|
|
49
|
+
return hidden_states
|
|
50
|
+
|
|
51
|
+
|
|
52
|
+
class TemporalBlock(torch.nn.Module):
|
|
53
|
+
|
|
54
|
+
def __init__(self, num_attention_heads, attention_head_dim, in_channels, num_layers=1, norm_num_groups=32, eps=1e-5):
|
|
55
|
+
super().__init__()
|
|
56
|
+
inner_dim = num_attention_heads * attention_head_dim
|
|
57
|
+
|
|
58
|
+
self.norm = torch.nn.GroupNorm(num_groups=norm_num_groups, num_channels=in_channels, eps=eps, affine=True)
|
|
59
|
+
self.proj_in = torch.nn.Linear(in_channels, inner_dim)
|
|
60
|
+
|
|
61
|
+
self.transformer_blocks = torch.nn.ModuleList([
|
|
62
|
+
TemporalTransformerBlock(
|
|
63
|
+
inner_dim,
|
|
64
|
+
num_attention_heads,
|
|
65
|
+
attention_head_dim
|
|
66
|
+
)
|
|
67
|
+
for d in range(num_layers)
|
|
68
|
+
])
|
|
69
|
+
|
|
70
|
+
self.proj_out = torch.nn.Linear(inner_dim, in_channels)
|
|
71
|
+
|
|
72
|
+
def forward(self, hidden_states, time_emb, text_emb, res_stack, batch_size=1):
|
|
73
|
+
batch, _, height, width = hidden_states.shape
|
|
74
|
+
residual = hidden_states
|
|
75
|
+
|
|
76
|
+
hidden_states = self.norm(hidden_states)
|
|
77
|
+
inner_dim = hidden_states.shape[1]
|
|
78
|
+
hidden_states = hidden_states.permute(0, 2, 3, 1).reshape(batch, height * width, inner_dim)
|
|
79
|
+
hidden_states = self.proj_in(hidden_states)
|
|
80
|
+
|
|
81
|
+
for block in self.transformer_blocks:
|
|
82
|
+
hidden_states = block(
|
|
83
|
+
hidden_states,
|
|
84
|
+
batch_size=batch_size
|
|
85
|
+
)
|
|
86
|
+
|
|
87
|
+
hidden_states = self.proj_out(hidden_states)
|
|
88
|
+
hidden_states = hidden_states.reshape(batch, height, width, inner_dim).permute(0, 3, 1, 2).contiguous()
|
|
89
|
+
hidden_states = hidden_states + residual
|
|
90
|
+
|
|
91
|
+
return hidden_states, time_emb, text_emb, res_stack
|
|
92
|
+
|
|
93
|
+
|
|
94
|
+
class SDMotionModel(torch.nn.Module):
|
|
95
|
+
def __init__(self):
|
|
96
|
+
super().__init__()
|
|
97
|
+
self.motion_modules = torch.nn.ModuleList([
|
|
98
|
+
TemporalBlock(8, 40, 320, eps=1e-6),
|
|
99
|
+
TemporalBlock(8, 40, 320, eps=1e-6),
|
|
100
|
+
TemporalBlock(8, 80, 640, eps=1e-6),
|
|
101
|
+
TemporalBlock(8, 80, 640, eps=1e-6),
|
|
102
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
103
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
104
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
105
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
106
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
107
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
108
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
109
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
110
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
111
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
112
|
+
TemporalBlock(8, 160, 1280, eps=1e-6),
|
|
113
|
+
TemporalBlock(8, 80, 640, eps=1e-6),
|
|
114
|
+
TemporalBlock(8, 80, 640, eps=1e-6),
|
|
115
|
+
TemporalBlock(8, 80, 640, eps=1e-6),
|
|
116
|
+
TemporalBlock(8, 40, 320, eps=1e-6),
|
|
117
|
+
TemporalBlock(8, 40, 320, eps=1e-6),
|
|
118
|
+
TemporalBlock(8, 40, 320, eps=1e-6),
|
|
119
|
+
])
|
|
120
|
+
self.call_block_id = {
|
|
121
|
+
1: 0,
|
|
122
|
+
4: 1,
|
|
123
|
+
9: 2,
|
|
124
|
+
12: 3,
|
|
125
|
+
17: 4,
|
|
126
|
+
20: 5,
|
|
127
|
+
24: 6,
|
|
128
|
+
26: 7,
|
|
129
|
+
29: 8,
|
|
130
|
+
32: 9,
|
|
131
|
+
34: 10,
|
|
132
|
+
36: 11,
|
|
133
|
+
40: 12,
|
|
134
|
+
43: 13,
|
|
135
|
+
46: 14,
|
|
136
|
+
50: 15,
|
|
137
|
+
53: 16,
|
|
138
|
+
56: 17,
|
|
139
|
+
60: 18,
|
|
140
|
+
63: 19,
|
|
141
|
+
66: 20
|
|
142
|
+
}
|
|
143
|
+
|
|
144
|
+
def forward(self):
|
|
145
|
+
pass
|
|
146
|
+
|
|
147
|
+
@staticmethod
|
|
148
|
+
def state_dict_converter():
|
|
149
|
+
return SDMotionModelStateDictConverter()
|
|
150
|
+
|
|
151
|
+
|
|
152
|
+
class SDMotionModelStateDictConverter:
|
|
153
|
+
def __init__(self):
|
|
154
|
+
pass
|
|
155
|
+
|
|
156
|
+
def from_diffusers(self, state_dict):
|
|
157
|
+
rename_dict = {
|
|
158
|
+
"norm": "norm",
|
|
159
|
+
"proj_in": "proj_in",
|
|
160
|
+
"transformer_blocks.0.attention_blocks.0.to_q": "transformer_blocks.0.attn1.to_q",
|
|
161
|
+
"transformer_blocks.0.attention_blocks.0.to_k": "transformer_blocks.0.attn1.to_k",
|
|
162
|
+
"transformer_blocks.0.attention_blocks.0.to_v": "transformer_blocks.0.attn1.to_v",
|
|
163
|
+
"transformer_blocks.0.attention_blocks.0.to_out.0": "transformer_blocks.0.attn1.to_out",
|
|
164
|
+
"transformer_blocks.0.attention_blocks.0.pos_encoder": "transformer_blocks.0.pe1",
|
|
165
|
+
"transformer_blocks.0.attention_blocks.1.to_q": "transformer_blocks.0.attn2.to_q",
|
|
166
|
+
"transformer_blocks.0.attention_blocks.1.to_k": "transformer_blocks.0.attn2.to_k",
|
|
167
|
+
"transformer_blocks.0.attention_blocks.1.to_v": "transformer_blocks.0.attn2.to_v",
|
|
168
|
+
"transformer_blocks.0.attention_blocks.1.to_out.0": "transformer_blocks.0.attn2.to_out",
|
|
169
|
+
"transformer_blocks.0.attention_blocks.1.pos_encoder": "transformer_blocks.0.pe2",
|
|
170
|
+
"transformer_blocks.0.norms.0": "transformer_blocks.0.norm1",
|
|
171
|
+
"transformer_blocks.0.norms.1": "transformer_blocks.0.norm2",
|
|
172
|
+
"transformer_blocks.0.ff.net.0.proj": "transformer_blocks.0.act_fn.proj",
|
|
173
|
+
"transformer_blocks.0.ff.net.2": "transformer_blocks.0.ff",
|
|
174
|
+
"transformer_blocks.0.ff_norm": "transformer_blocks.0.norm3",
|
|
175
|
+
"proj_out": "proj_out",
|
|
176
|
+
}
|
|
177
|
+
name_list = sorted([i for i in state_dict if i.startswith("down_blocks.")])
|
|
178
|
+
name_list += sorted([i for i in state_dict if i.startswith("mid_block.")])
|
|
179
|
+
name_list += sorted([i for i in state_dict if i.startswith("up_blocks.")])
|
|
180
|
+
state_dict_ = {}
|
|
181
|
+
last_prefix, module_id = "", -1
|
|
182
|
+
for name in name_list:
|
|
183
|
+
names = name.split(".")
|
|
184
|
+
prefix_index = names.index("temporal_transformer") + 1
|
|
185
|
+
prefix = ".".join(names[:prefix_index])
|
|
186
|
+
if prefix != last_prefix:
|
|
187
|
+
last_prefix = prefix
|
|
188
|
+
module_id += 1
|
|
189
|
+
middle_name = ".".join(names[prefix_index:-1])
|
|
190
|
+
suffix = names[-1]
|
|
191
|
+
if "pos_encoder" in names:
|
|
192
|
+
rename = ".".join(["motion_modules", str(module_id), rename_dict[middle_name]])
|
|
193
|
+
else:
|
|
194
|
+
rename = ".".join(["motion_modules", str(module_id), rename_dict[middle_name], suffix])
|
|
195
|
+
state_dict_[rename] = state_dict[name]
|
|
196
|
+
return state_dict_
|
|
197
|
+
|
|
198
|
+
def from_civitai(self, state_dict):
|
|
199
|
+
return self.from_diffusers(state_dict)
|
|
@@ -0,0 +1,321 @@
|
|
|
1
|
+
import torch
|
|
2
|
+
from .attention import Attention
|
|
3
|
+
|
|
4
|
+
|
|
5
|
+
class CLIPEncoderLayer(torch.nn.Module):
|
|
6
|
+
def __init__(self, embed_dim, intermediate_size, num_heads=12, head_dim=64, use_quick_gelu=True):
|
|
7
|
+
super().__init__()
|
|
8
|
+
self.attn = Attention(q_dim=embed_dim, num_heads=num_heads, head_dim=head_dim, bias_q=True, bias_kv=True, bias_out=True)
|
|
9
|
+
self.layer_norm1 = torch.nn.LayerNorm(embed_dim)
|
|
10
|
+
self.layer_norm2 = torch.nn.LayerNorm(embed_dim)
|
|
11
|
+
self.fc1 = torch.nn.Linear(embed_dim, intermediate_size)
|
|
12
|
+
self.fc2 = torch.nn.Linear(intermediate_size, embed_dim)
|
|
13
|
+
|
|
14
|
+
self.use_quick_gelu = use_quick_gelu
|
|
15
|
+
|
|
16
|
+
def quickGELU(self, x):
|
|
17
|
+
return x * torch.sigmoid(1.702 * x)
|
|
18
|
+
|
|
19
|
+
def forward(self, hidden_states, attn_mask=None):
|
|
20
|
+
residual = hidden_states
|
|
21
|
+
|
|
22
|
+
hidden_states = self.layer_norm1(hidden_states)
|
|
23
|
+
hidden_states = self.attn(hidden_states, attn_mask=attn_mask)
|
|
24
|
+
hidden_states = residual + hidden_states
|
|
25
|
+
|
|
26
|
+
residual = hidden_states
|
|
27
|
+
hidden_states = self.layer_norm2(hidden_states)
|
|
28
|
+
hidden_states = self.fc1(hidden_states)
|
|
29
|
+
if self.use_quick_gelu:
|
|
30
|
+
hidden_states = self.quickGELU(hidden_states)
|
|
31
|
+
else:
|
|
32
|
+
hidden_states = torch.nn.functional.gelu(hidden_states)
|
|
33
|
+
hidden_states = self.fc2(hidden_states)
|
|
34
|
+
hidden_states = residual + hidden_states
|
|
35
|
+
|
|
36
|
+
return hidden_states
|
|
37
|
+
|
|
38
|
+
|
|
39
|
+
class SDTextEncoder(torch.nn.Module):
|
|
40
|
+
def __init__(self, embed_dim=768, vocab_size=49408, max_position_embeddings=77, num_encoder_layers=12, encoder_intermediate_size=3072):
|
|
41
|
+
super().__init__()
|
|
42
|
+
|
|
43
|
+
# token_embedding
|
|
44
|
+
self.token_embedding = torch.nn.Embedding(vocab_size, embed_dim)
|
|
45
|
+
|
|
46
|
+
# position_embeds (This is a fixed tensor)
|
|
47
|
+
self.position_embeds = torch.nn.Parameter(torch.zeros(1, max_position_embeddings, embed_dim))
|
|
48
|
+
|
|
49
|
+
# encoders
|
|
50
|
+
self.encoders = torch.nn.ModuleList([CLIPEncoderLayer(embed_dim, encoder_intermediate_size) for _ in range(num_encoder_layers)])
|
|
51
|
+
|
|
52
|
+
# attn_mask
|
|
53
|
+
self.attn_mask = self.attention_mask(max_position_embeddings)
|
|
54
|
+
|
|
55
|
+
# final_layer_norm
|
|
56
|
+
self.final_layer_norm = torch.nn.LayerNorm(embed_dim)
|
|
57
|
+
|
|
58
|
+
def attention_mask(self, length):
|
|
59
|
+
mask = torch.empty(length, length)
|
|
60
|
+
mask.fill_(float("-inf"))
|
|
61
|
+
mask.triu_(1)
|
|
62
|
+
return mask
|
|
63
|
+
|
|
64
|
+
def forward(self, input_ids, clip_skip=1):
|
|
65
|
+
embeds = self.token_embedding(input_ids) + self.position_embeds
|
|
66
|
+
attn_mask = self.attn_mask.to(device=embeds.device, dtype=embeds.dtype)
|
|
67
|
+
for encoder_id, encoder in enumerate(self.encoders):
|
|
68
|
+
embeds = encoder(embeds, attn_mask=attn_mask)
|
|
69
|
+
if encoder_id + clip_skip == len(self.encoders):
|
|
70
|
+
break
|
|
71
|
+
embeds = self.final_layer_norm(embeds)
|
|
72
|
+
return embeds
|
|
73
|
+
|
|
74
|
+
@staticmethod
|
|
75
|
+
def state_dict_converter():
|
|
76
|
+
return SDTextEncoderStateDictConverter()
|
|
77
|
+
|
|
78
|
+
|
|
79
|
+
class SDTextEncoderStateDictConverter:
|
|
80
|
+
def __init__(self):
|
|
81
|
+
pass
|
|
82
|
+
|
|
83
|
+
def from_diffusers(self, state_dict):
|
|
84
|
+
rename_dict = {
|
|
85
|
+
"text_model.embeddings.token_embedding.weight": "token_embedding.weight",
|
|
86
|
+
"text_model.embeddings.position_embedding.weight": "position_embeds",
|
|
87
|
+
"text_model.final_layer_norm.weight": "final_layer_norm.weight",
|
|
88
|
+
"text_model.final_layer_norm.bias": "final_layer_norm.bias"
|
|
89
|
+
}
|
|
90
|
+
attn_rename_dict = {
|
|
91
|
+
"self_attn.q_proj": "attn.to_q",
|
|
92
|
+
"self_attn.k_proj": "attn.to_k",
|
|
93
|
+
"self_attn.v_proj": "attn.to_v",
|
|
94
|
+
"self_attn.out_proj": "attn.to_out",
|
|
95
|
+
"layer_norm1": "layer_norm1",
|
|
96
|
+
"layer_norm2": "layer_norm2",
|
|
97
|
+
"mlp.fc1": "fc1",
|
|
98
|
+
"mlp.fc2": "fc2",
|
|
99
|
+
}
|
|
100
|
+
state_dict_ = {}
|
|
101
|
+
for name in state_dict:
|
|
102
|
+
if name in rename_dict:
|
|
103
|
+
param = state_dict[name]
|
|
104
|
+
if name == "text_model.embeddings.position_embedding.weight":
|
|
105
|
+
param = param.reshape((1, param.shape[0], param.shape[1]))
|
|
106
|
+
state_dict_[rename_dict[name]] = param
|
|
107
|
+
elif name.startswith("text_model.encoder.layers."):
|
|
108
|
+
param = state_dict[name]
|
|
109
|
+
names = name.split(".")
|
|
110
|
+
layer_id, layer_type, tail = names[3], ".".join(names[4:-1]), names[-1]
|
|
111
|
+
name_ = ".".join(["encoders", layer_id, attn_rename_dict[layer_type], tail])
|
|
112
|
+
state_dict_[name_] = param
|
|
113
|
+
return state_dict_
|
|
114
|
+
|
|
115
|
+
def from_civitai(self, state_dict):
|
|
116
|
+
rename_dict = {
|
|
117
|
+
"cond_stage_model.transformer.text_model.embeddings.token_embedding.weight": "token_embedding.weight",
|
|
118
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.layer_norm1.bias": "encoders.0.layer_norm1.bias",
|
|
119
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.layer_norm1.weight": "encoders.0.layer_norm1.weight",
|
|
120
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.layer_norm2.bias": "encoders.0.layer_norm2.bias",
|
|
121
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.layer_norm2.weight": "encoders.0.layer_norm2.weight",
|
|
122
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.mlp.fc1.bias": "encoders.0.fc1.bias",
|
|
123
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.mlp.fc1.weight": "encoders.0.fc1.weight",
|
|
124
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.mlp.fc2.bias": "encoders.0.fc2.bias",
|
|
125
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.mlp.fc2.weight": "encoders.0.fc2.weight",
|
|
126
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.k_proj.bias": "encoders.0.attn.to_k.bias",
|
|
127
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.k_proj.weight": "encoders.0.attn.to_k.weight",
|
|
128
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.out_proj.bias": "encoders.0.attn.to_out.bias",
|
|
129
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.out_proj.weight": "encoders.0.attn.to_out.weight",
|
|
130
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.q_proj.bias": "encoders.0.attn.to_q.bias",
|
|
131
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.q_proj.weight": "encoders.0.attn.to_q.weight",
|
|
132
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.v_proj.bias": "encoders.0.attn.to_v.bias",
|
|
133
|
+
"cond_stage_model.transformer.text_model.encoder.layers.0.self_attn.v_proj.weight": "encoders.0.attn.to_v.weight",
|
|
134
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.layer_norm1.bias": "encoders.1.layer_norm1.bias",
|
|
135
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.layer_norm1.weight": "encoders.1.layer_norm1.weight",
|
|
136
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.layer_norm2.bias": "encoders.1.layer_norm2.bias",
|
|
137
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.layer_norm2.weight": "encoders.1.layer_norm2.weight",
|
|
138
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.mlp.fc1.bias": "encoders.1.fc1.bias",
|
|
139
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.mlp.fc1.weight": "encoders.1.fc1.weight",
|
|
140
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.mlp.fc2.bias": "encoders.1.fc2.bias",
|
|
141
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.mlp.fc2.weight": "encoders.1.fc2.weight",
|
|
142
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.k_proj.bias": "encoders.1.attn.to_k.bias",
|
|
143
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.k_proj.weight": "encoders.1.attn.to_k.weight",
|
|
144
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.out_proj.bias": "encoders.1.attn.to_out.bias",
|
|
145
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.out_proj.weight": "encoders.1.attn.to_out.weight",
|
|
146
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.q_proj.bias": "encoders.1.attn.to_q.bias",
|
|
147
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.q_proj.weight": "encoders.1.attn.to_q.weight",
|
|
148
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.v_proj.bias": "encoders.1.attn.to_v.bias",
|
|
149
|
+
"cond_stage_model.transformer.text_model.encoder.layers.1.self_attn.v_proj.weight": "encoders.1.attn.to_v.weight",
|
|
150
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.layer_norm1.bias": "encoders.10.layer_norm1.bias",
|
|
151
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.layer_norm1.weight": "encoders.10.layer_norm1.weight",
|
|
152
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.layer_norm2.bias": "encoders.10.layer_norm2.bias",
|
|
153
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.layer_norm2.weight": "encoders.10.layer_norm2.weight",
|
|
154
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.mlp.fc1.bias": "encoders.10.fc1.bias",
|
|
155
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.mlp.fc1.weight": "encoders.10.fc1.weight",
|
|
156
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.mlp.fc2.bias": "encoders.10.fc2.bias",
|
|
157
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.mlp.fc2.weight": "encoders.10.fc2.weight",
|
|
158
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.k_proj.bias": "encoders.10.attn.to_k.bias",
|
|
159
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.k_proj.weight": "encoders.10.attn.to_k.weight",
|
|
160
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.out_proj.bias": "encoders.10.attn.to_out.bias",
|
|
161
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.out_proj.weight": "encoders.10.attn.to_out.weight",
|
|
162
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.q_proj.bias": "encoders.10.attn.to_q.bias",
|
|
163
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.q_proj.weight": "encoders.10.attn.to_q.weight",
|
|
164
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.v_proj.bias": "encoders.10.attn.to_v.bias",
|
|
165
|
+
"cond_stage_model.transformer.text_model.encoder.layers.10.self_attn.v_proj.weight": "encoders.10.attn.to_v.weight",
|
|
166
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.layer_norm1.bias": "encoders.11.layer_norm1.bias",
|
|
167
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.layer_norm1.weight": "encoders.11.layer_norm1.weight",
|
|
168
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.layer_norm2.bias": "encoders.11.layer_norm2.bias",
|
|
169
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.layer_norm2.weight": "encoders.11.layer_norm2.weight",
|
|
170
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.mlp.fc1.bias": "encoders.11.fc1.bias",
|
|
171
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.mlp.fc1.weight": "encoders.11.fc1.weight",
|
|
172
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.mlp.fc2.bias": "encoders.11.fc2.bias",
|
|
173
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.mlp.fc2.weight": "encoders.11.fc2.weight",
|
|
174
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.k_proj.bias": "encoders.11.attn.to_k.bias",
|
|
175
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.k_proj.weight": "encoders.11.attn.to_k.weight",
|
|
176
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.out_proj.bias": "encoders.11.attn.to_out.bias",
|
|
177
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.out_proj.weight": "encoders.11.attn.to_out.weight",
|
|
178
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.q_proj.bias": "encoders.11.attn.to_q.bias",
|
|
179
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.q_proj.weight": "encoders.11.attn.to_q.weight",
|
|
180
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.v_proj.bias": "encoders.11.attn.to_v.bias",
|
|
181
|
+
"cond_stage_model.transformer.text_model.encoder.layers.11.self_attn.v_proj.weight": "encoders.11.attn.to_v.weight",
|
|
182
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.layer_norm1.bias": "encoders.2.layer_norm1.bias",
|
|
183
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.layer_norm1.weight": "encoders.2.layer_norm1.weight",
|
|
184
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.layer_norm2.bias": "encoders.2.layer_norm2.bias",
|
|
185
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.layer_norm2.weight": "encoders.2.layer_norm2.weight",
|
|
186
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.mlp.fc1.bias": "encoders.2.fc1.bias",
|
|
187
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.mlp.fc1.weight": "encoders.2.fc1.weight",
|
|
188
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.mlp.fc2.bias": "encoders.2.fc2.bias",
|
|
189
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.mlp.fc2.weight": "encoders.2.fc2.weight",
|
|
190
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.k_proj.bias": "encoders.2.attn.to_k.bias",
|
|
191
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.k_proj.weight": "encoders.2.attn.to_k.weight",
|
|
192
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.out_proj.bias": "encoders.2.attn.to_out.bias",
|
|
193
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.out_proj.weight": "encoders.2.attn.to_out.weight",
|
|
194
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.q_proj.bias": "encoders.2.attn.to_q.bias",
|
|
195
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.q_proj.weight": "encoders.2.attn.to_q.weight",
|
|
196
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.v_proj.bias": "encoders.2.attn.to_v.bias",
|
|
197
|
+
"cond_stage_model.transformer.text_model.encoder.layers.2.self_attn.v_proj.weight": "encoders.2.attn.to_v.weight",
|
|
198
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.layer_norm1.bias": "encoders.3.layer_norm1.bias",
|
|
199
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.layer_norm1.weight": "encoders.3.layer_norm1.weight",
|
|
200
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.layer_norm2.bias": "encoders.3.layer_norm2.bias",
|
|
201
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.layer_norm2.weight": "encoders.3.layer_norm2.weight",
|
|
202
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.mlp.fc1.bias": "encoders.3.fc1.bias",
|
|
203
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.mlp.fc1.weight": "encoders.3.fc1.weight",
|
|
204
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.mlp.fc2.bias": "encoders.3.fc2.bias",
|
|
205
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.mlp.fc2.weight": "encoders.3.fc2.weight",
|
|
206
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.k_proj.bias": "encoders.3.attn.to_k.bias",
|
|
207
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.k_proj.weight": "encoders.3.attn.to_k.weight",
|
|
208
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.out_proj.bias": "encoders.3.attn.to_out.bias",
|
|
209
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.out_proj.weight": "encoders.3.attn.to_out.weight",
|
|
210
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.q_proj.bias": "encoders.3.attn.to_q.bias",
|
|
211
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.q_proj.weight": "encoders.3.attn.to_q.weight",
|
|
212
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.v_proj.bias": "encoders.3.attn.to_v.bias",
|
|
213
|
+
"cond_stage_model.transformer.text_model.encoder.layers.3.self_attn.v_proj.weight": "encoders.3.attn.to_v.weight",
|
|
214
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.layer_norm1.bias": "encoders.4.layer_norm1.bias",
|
|
215
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.layer_norm1.weight": "encoders.4.layer_norm1.weight",
|
|
216
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.layer_norm2.bias": "encoders.4.layer_norm2.bias",
|
|
217
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.layer_norm2.weight": "encoders.4.layer_norm2.weight",
|
|
218
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.mlp.fc1.bias": "encoders.4.fc1.bias",
|
|
219
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.mlp.fc1.weight": "encoders.4.fc1.weight",
|
|
220
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.mlp.fc2.bias": "encoders.4.fc2.bias",
|
|
221
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.mlp.fc2.weight": "encoders.4.fc2.weight",
|
|
222
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.k_proj.bias": "encoders.4.attn.to_k.bias",
|
|
223
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.k_proj.weight": "encoders.4.attn.to_k.weight",
|
|
224
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.out_proj.bias": "encoders.4.attn.to_out.bias",
|
|
225
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.out_proj.weight": "encoders.4.attn.to_out.weight",
|
|
226
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.q_proj.bias": "encoders.4.attn.to_q.bias",
|
|
227
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.q_proj.weight": "encoders.4.attn.to_q.weight",
|
|
228
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.v_proj.bias": "encoders.4.attn.to_v.bias",
|
|
229
|
+
"cond_stage_model.transformer.text_model.encoder.layers.4.self_attn.v_proj.weight": "encoders.4.attn.to_v.weight",
|
|
230
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.layer_norm1.bias": "encoders.5.layer_norm1.bias",
|
|
231
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.layer_norm1.weight": "encoders.5.layer_norm1.weight",
|
|
232
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.layer_norm2.bias": "encoders.5.layer_norm2.bias",
|
|
233
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.layer_norm2.weight": "encoders.5.layer_norm2.weight",
|
|
234
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.mlp.fc1.bias": "encoders.5.fc1.bias",
|
|
235
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.mlp.fc1.weight": "encoders.5.fc1.weight",
|
|
236
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.mlp.fc2.bias": "encoders.5.fc2.bias",
|
|
237
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.mlp.fc2.weight": "encoders.5.fc2.weight",
|
|
238
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.k_proj.bias": "encoders.5.attn.to_k.bias",
|
|
239
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.k_proj.weight": "encoders.5.attn.to_k.weight",
|
|
240
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.out_proj.bias": "encoders.5.attn.to_out.bias",
|
|
241
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.out_proj.weight": "encoders.5.attn.to_out.weight",
|
|
242
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.q_proj.bias": "encoders.5.attn.to_q.bias",
|
|
243
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.q_proj.weight": "encoders.5.attn.to_q.weight",
|
|
244
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.v_proj.bias": "encoders.5.attn.to_v.bias",
|
|
245
|
+
"cond_stage_model.transformer.text_model.encoder.layers.5.self_attn.v_proj.weight": "encoders.5.attn.to_v.weight",
|
|
246
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.layer_norm1.bias": "encoders.6.layer_norm1.bias",
|
|
247
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.layer_norm1.weight": "encoders.6.layer_norm1.weight",
|
|
248
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.layer_norm2.bias": "encoders.6.layer_norm2.bias",
|
|
249
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.layer_norm2.weight": "encoders.6.layer_norm2.weight",
|
|
250
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.mlp.fc1.bias": "encoders.6.fc1.bias",
|
|
251
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.mlp.fc1.weight": "encoders.6.fc1.weight",
|
|
252
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.mlp.fc2.bias": "encoders.6.fc2.bias",
|
|
253
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.mlp.fc2.weight": "encoders.6.fc2.weight",
|
|
254
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.k_proj.bias": "encoders.6.attn.to_k.bias",
|
|
255
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.k_proj.weight": "encoders.6.attn.to_k.weight",
|
|
256
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.out_proj.bias": "encoders.6.attn.to_out.bias",
|
|
257
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.out_proj.weight": "encoders.6.attn.to_out.weight",
|
|
258
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.q_proj.bias": "encoders.6.attn.to_q.bias",
|
|
259
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.q_proj.weight": "encoders.6.attn.to_q.weight",
|
|
260
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.v_proj.bias": "encoders.6.attn.to_v.bias",
|
|
261
|
+
"cond_stage_model.transformer.text_model.encoder.layers.6.self_attn.v_proj.weight": "encoders.6.attn.to_v.weight",
|
|
262
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.layer_norm1.bias": "encoders.7.layer_norm1.bias",
|
|
263
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.layer_norm1.weight": "encoders.7.layer_norm1.weight",
|
|
264
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.layer_norm2.bias": "encoders.7.layer_norm2.bias",
|
|
265
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.layer_norm2.weight": "encoders.7.layer_norm2.weight",
|
|
266
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.mlp.fc1.bias": "encoders.7.fc1.bias",
|
|
267
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.mlp.fc1.weight": "encoders.7.fc1.weight",
|
|
268
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.mlp.fc2.bias": "encoders.7.fc2.bias",
|
|
269
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.mlp.fc2.weight": "encoders.7.fc2.weight",
|
|
270
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.k_proj.bias": "encoders.7.attn.to_k.bias",
|
|
271
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.k_proj.weight": "encoders.7.attn.to_k.weight",
|
|
272
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.out_proj.bias": "encoders.7.attn.to_out.bias",
|
|
273
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.out_proj.weight": "encoders.7.attn.to_out.weight",
|
|
274
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.q_proj.bias": "encoders.7.attn.to_q.bias",
|
|
275
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.q_proj.weight": "encoders.7.attn.to_q.weight",
|
|
276
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.v_proj.bias": "encoders.7.attn.to_v.bias",
|
|
277
|
+
"cond_stage_model.transformer.text_model.encoder.layers.7.self_attn.v_proj.weight": "encoders.7.attn.to_v.weight",
|
|
278
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.layer_norm1.bias": "encoders.8.layer_norm1.bias",
|
|
279
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.layer_norm1.weight": "encoders.8.layer_norm1.weight",
|
|
280
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.layer_norm2.bias": "encoders.8.layer_norm2.bias",
|
|
281
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.layer_norm2.weight": "encoders.8.layer_norm2.weight",
|
|
282
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.mlp.fc1.bias": "encoders.8.fc1.bias",
|
|
283
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.mlp.fc1.weight": "encoders.8.fc1.weight",
|
|
284
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.mlp.fc2.bias": "encoders.8.fc2.bias",
|
|
285
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.mlp.fc2.weight": "encoders.8.fc2.weight",
|
|
286
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.k_proj.bias": "encoders.8.attn.to_k.bias",
|
|
287
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.k_proj.weight": "encoders.8.attn.to_k.weight",
|
|
288
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.out_proj.bias": "encoders.8.attn.to_out.bias",
|
|
289
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.out_proj.weight": "encoders.8.attn.to_out.weight",
|
|
290
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.q_proj.bias": "encoders.8.attn.to_q.bias",
|
|
291
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.q_proj.weight": "encoders.8.attn.to_q.weight",
|
|
292
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.v_proj.bias": "encoders.8.attn.to_v.bias",
|
|
293
|
+
"cond_stage_model.transformer.text_model.encoder.layers.8.self_attn.v_proj.weight": "encoders.8.attn.to_v.weight",
|
|
294
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.layer_norm1.bias": "encoders.9.layer_norm1.bias",
|
|
295
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.layer_norm1.weight": "encoders.9.layer_norm1.weight",
|
|
296
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.layer_norm2.bias": "encoders.9.layer_norm2.bias",
|
|
297
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.layer_norm2.weight": "encoders.9.layer_norm2.weight",
|
|
298
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.mlp.fc1.bias": "encoders.9.fc1.bias",
|
|
299
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.mlp.fc1.weight": "encoders.9.fc1.weight",
|
|
300
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.mlp.fc2.bias": "encoders.9.fc2.bias",
|
|
301
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.mlp.fc2.weight": "encoders.9.fc2.weight",
|
|
302
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.k_proj.bias": "encoders.9.attn.to_k.bias",
|
|
303
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.k_proj.weight": "encoders.9.attn.to_k.weight",
|
|
304
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.out_proj.bias": "encoders.9.attn.to_out.bias",
|
|
305
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.out_proj.weight": "encoders.9.attn.to_out.weight",
|
|
306
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.q_proj.bias": "encoders.9.attn.to_q.bias",
|
|
307
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.q_proj.weight": "encoders.9.attn.to_q.weight",
|
|
308
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.v_proj.bias": "encoders.9.attn.to_v.bias",
|
|
309
|
+
"cond_stage_model.transformer.text_model.encoder.layers.9.self_attn.v_proj.weight": "encoders.9.attn.to_v.weight",
|
|
310
|
+
"cond_stage_model.transformer.text_model.final_layer_norm.bias": "final_layer_norm.bias",
|
|
311
|
+
"cond_stage_model.transformer.text_model.final_layer_norm.weight": "final_layer_norm.weight",
|
|
312
|
+
"cond_stage_model.transformer.text_model.embeddings.position_embedding.weight": "position_embeds"
|
|
313
|
+
}
|
|
314
|
+
state_dict_ = {}
|
|
315
|
+
for name in state_dict:
|
|
316
|
+
if name in rename_dict:
|
|
317
|
+
param = state_dict[name]
|
|
318
|
+
if name == "cond_stage_model.transformer.text_model.embeddings.position_embedding.weight":
|
|
319
|
+
param = param.reshape((1, param.shape[0], param.shape[1]))
|
|
320
|
+
state_dict_[rename_dict[name]] = param
|
|
321
|
+
return state_dict_
|