model.safetensors.index.json
102.8 KB · 1175 lines · json Raw
1 {
2 "metadata": {
3 "total_parameters": 11078526194,
4 "total_size": 22157052464
5 },
6 "weight_map": {
7 "action_in_proj.encoder.trunk.0.bias": "model-00005-of-00005.safetensors",
8 "action_in_proj.encoder.trunk.0.weight": "model-00005-of-00005.safetensors",
9 "action_in_proj.encoder.trunk.2.weight": "model-00005-of-00005.safetensors",
10 "action_in_proj.encoder.trunk.3.bias": "model-00005-of-00005.safetensors",
11 "action_in_proj.encoder.trunk.3.weight": "model-00005-of-00005.safetensors",
12 "action_in_proj.encoder.trunk.5.weight": "model-00005-of-00005.safetensors",
13 "action_in_proj.encoder.trunk.6.bias": "model-00005-of-00005.safetensors",
14 "action_in_proj.encoder.trunk.6.weight": "model-00005-of-00005.safetensors",
15 "action_in_proj.norm.bias": "model-00005-of-00005.safetensors",
16 "action_in_proj.norm.weight": "model-00005-of-00005.safetensors",
17 "action_in_proj.sinus.0.freqs": "model-00005-of-00005.safetensors",
18 "action_in_proj.sinus.1.freqs": "model-00005-of-00005.safetensors",
19 "action_in_proj.timestep_fourier_encoder.freqs": "model-00005-of-00005.safetensors",
20 "action_out_proj.bias": "model-00005-of-00005.safetensors",
21 "action_out_proj.weight": "model-00005-of-00005.safetensors",
22 "action_space.accel_mean": "model-00005-of-00005.safetensors",
23 "action_space.accel_std": "model-00005-of-00005.safetensors",
24 "action_space.curvature_mean": "model-00005-of-00005.safetensors",
25 "action_space.curvature_std": "model-00005-of-00005.safetensors",
26 "expert.layers.0.input_layernorm.weight": "model-00004-of-00005.safetensors",
27 "expert.layers.0.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
28 "expert.layers.0.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
29 "expert.layers.0.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
30 "expert.layers.0.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
31 "expert.layers.0.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
32 "expert.layers.0.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
33 "expert.layers.0.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
34 "expert.layers.0.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
35 "expert.layers.0.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
36 "expert.layers.0.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
37 "expert.layers.1.input_layernorm.weight": "model-00004-of-00005.safetensors",
38 "expert.layers.1.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
39 "expert.layers.1.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
40 "expert.layers.1.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
41 "expert.layers.1.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
42 "expert.layers.1.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
43 "expert.layers.1.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
44 "expert.layers.1.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
45 "expert.layers.1.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
46 "expert.layers.1.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
47 "expert.layers.1.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
48 "expert.layers.10.input_layernorm.weight": "model-00004-of-00005.safetensors",
49 "expert.layers.10.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
50 "expert.layers.10.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
51 "expert.layers.10.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
52 "expert.layers.10.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
53 "expert.layers.10.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
54 "expert.layers.10.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
55 "expert.layers.10.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
56 "expert.layers.10.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
57 "expert.layers.10.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
58 "expert.layers.10.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
59 "expert.layers.11.input_layernorm.weight": "model-00004-of-00005.safetensors",
60 "expert.layers.11.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
61 "expert.layers.11.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
62 "expert.layers.11.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
63 "expert.layers.11.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
64 "expert.layers.11.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
65 "expert.layers.11.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
66 "expert.layers.11.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
67 "expert.layers.11.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
68 "expert.layers.11.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
69 "expert.layers.11.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
70 "expert.layers.12.input_layernorm.weight": "model-00004-of-00005.safetensors",
71 "expert.layers.12.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
72 "expert.layers.12.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
73 "expert.layers.12.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
74 "expert.layers.12.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
75 "expert.layers.12.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
76 "expert.layers.12.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
77 "expert.layers.12.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
78 "expert.layers.12.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
79 "expert.layers.12.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
80 "expert.layers.12.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
81 "expert.layers.13.input_layernorm.weight": "model-00004-of-00005.safetensors",
82 "expert.layers.13.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
83 "expert.layers.13.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
84 "expert.layers.13.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
85 "expert.layers.13.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
86 "expert.layers.13.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
87 "expert.layers.13.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
88 "expert.layers.13.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
89 "expert.layers.13.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
90 "expert.layers.13.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
91 "expert.layers.13.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
92 "expert.layers.14.input_layernorm.weight": "model-00004-of-00005.safetensors",
93 "expert.layers.14.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
94 "expert.layers.14.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
95 "expert.layers.14.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
96 "expert.layers.14.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
97 "expert.layers.14.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
98 "expert.layers.14.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
99 "expert.layers.14.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
100 "expert.layers.14.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
101 "expert.layers.14.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
102 "expert.layers.14.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
103 "expert.layers.15.input_layernorm.weight": "model-00004-of-00005.safetensors",
104 "expert.layers.15.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
105 "expert.layers.15.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
106 "expert.layers.15.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
107 "expert.layers.15.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
108 "expert.layers.15.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
109 "expert.layers.15.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
110 "expert.layers.15.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
111 "expert.layers.15.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
112 "expert.layers.15.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
113 "expert.layers.15.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
114 "expert.layers.16.input_layernorm.weight": "model-00004-of-00005.safetensors",
115 "expert.layers.16.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
116 "expert.layers.16.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
117 "expert.layers.16.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
118 "expert.layers.16.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
119 "expert.layers.16.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
120 "expert.layers.16.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
121 "expert.layers.16.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
122 "expert.layers.16.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
123 "expert.layers.16.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
124 "expert.layers.16.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
125 "expert.layers.17.input_layernorm.weight": "model-00005-of-00005.safetensors",
126 "expert.layers.17.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
127 "expert.layers.17.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
128 "expert.layers.17.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
129 "expert.layers.17.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
130 "expert.layers.17.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
131 "expert.layers.17.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
132 "expert.layers.17.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
133 "expert.layers.17.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
134 "expert.layers.17.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
135 "expert.layers.17.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
136 "expert.layers.18.input_layernorm.weight": "model-00005-of-00005.safetensors",
137 "expert.layers.18.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
138 "expert.layers.18.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
139 "expert.layers.18.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
140 "expert.layers.18.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
141 "expert.layers.18.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
142 "expert.layers.18.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
143 "expert.layers.18.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
144 "expert.layers.18.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
145 "expert.layers.18.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
146 "expert.layers.18.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
147 "expert.layers.19.input_layernorm.weight": "model-00005-of-00005.safetensors",
148 "expert.layers.19.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
149 "expert.layers.19.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
150 "expert.layers.19.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
151 "expert.layers.19.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
152 "expert.layers.19.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
153 "expert.layers.19.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
154 "expert.layers.19.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
155 "expert.layers.19.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
156 "expert.layers.19.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
157 "expert.layers.19.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
158 "expert.layers.2.input_layernorm.weight": "model-00004-of-00005.safetensors",
159 "expert.layers.2.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
160 "expert.layers.2.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
161 "expert.layers.2.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
162 "expert.layers.2.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
163 "expert.layers.2.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
164 "expert.layers.2.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
165 "expert.layers.2.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
166 "expert.layers.2.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
167 "expert.layers.2.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
168 "expert.layers.2.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
169 "expert.layers.20.input_layernorm.weight": "model-00005-of-00005.safetensors",
170 "expert.layers.20.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
171 "expert.layers.20.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
172 "expert.layers.20.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
173 "expert.layers.20.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
174 "expert.layers.20.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
175 "expert.layers.20.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
176 "expert.layers.20.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
177 "expert.layers.20.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
178 "expert.layers.20.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
179 "expert.layers.20.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
180 "expert.layers.21.input_layernorm.weight": "model-00005-of-00005.safetensors",
181 "expert.layers.21.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
182 "expert.layers.21.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
183 "expert.layers.21.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
184 "expert.layers.21.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
185 "expert.layers.21.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
186 "expert.layers.21.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
187 "expert.layers.21.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
188 "expert.layers.21.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
189 "expert.layers.21.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
190 "expert.layers.21.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
191 "expert.layers.22.input_layernorm.weight": "model-00005-of-00005.safetensors",
192 "expert.layers.22.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
193 "expert.layers.22.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
194 "expert.layers.22.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
195 "expert.layers.22.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
196 "expert.layers.22.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
197 "expert.layers.22.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
198 "expert.layers.22.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
199 "expert.layers.22.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
200 "expert.layers.22.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
201 "expert.layers.22.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
202 "expert.layers.23.input_layernorm.weight": "model-00005-of-00005.safetensors",
203 "expert.layers.23.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
204 "expert.layers.23.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
205 "expert.layers.23.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
206 "expert.layers.23.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
207 "expert.layers.23.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
208 "expert.layers.23.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
209 "expert.layers.23.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
210 "expert.layers.23.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
211 "expert.layers.23.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
212 "expert.layers.23.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
213 "expert.layers.24.input_layernorm.weight": "model-00005-of-00005.safetensors",
214 "expert.layers.24.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
215 "expert.layers.24.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
216 "expert.layers.24.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
217 "expert.layers.24.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
218 "expert.layers.24.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
219 "expert.layers.24.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
220 "expert.layers.24.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
221 "expert.layers.24.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
222 "expert.layers.24.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
223 "expert.layers.24.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
224 "expert.layers.25.input_layernorm.weight": "model-00005-of-00005.safetensors",
225 "expert.layers.25.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
226 "expert.layers.25.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
227 "expert.layers.25.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
228 "expert.layers.25.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
229 "expert.layers.25.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
230 "expert.layers.25.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
231 "expert.layers.25.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
232 "expert.layers.25.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
233 "expert.layers.25.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
234 "expert.layers.25.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
235 "expert.layers.26.input_layernorm.weight": "model-00005-of-00005.safetensors",
236 "expert.layers.26.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
237 "expert.layers.26.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
238 "expert.layers.26.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
239 "expert.layers.26.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
240 "expert.layers.26.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
241 "expert.layers.26.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
242 "expert.layers.26.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
243 "expert.layers.26.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
244 "expert.layers.26.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
245 "expert.layers.26.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
246 "expert.layers.27.input_layernorm.weight": "model-00005-of-00005.safetensors",
247 "expert.layers.27.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
248 "expert.layers.27.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
249 "expert.layers.27.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
250 "expert.layers.27.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
251 "expert.layers.27.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
252 "expert.layers.27.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
253 "expert.layers.27.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
254 "expert.layers.27.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
255 "expert.layers.27.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
256 "expert.layers.27.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
257 "expert.layers.28.input_layernorm.weight": "model-00005-of-00005.safetensors",
258 "expert.layers.28.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
259 "expert.layers.28.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
260 "expert.layers.28.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
261 "expert.layers.28.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
262 "expert.layers.28.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
263 "expert.layers.28.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
264 "expert.layers.28.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
265 "expert.layers.28.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
266 "expert.layers.28.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
267 "expert.layers.28.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
268 "expert.layers.29.input_layernorm.weight": "model-00005-of-00005.safetensors",
269 "expert.layers.29.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
270 "expert.layers.29.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
271 "expert.layers.29.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
272 "expert.layers.29.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
273 "expert.layers.29.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
274 "expert.layers.29.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
275 "expert.layers.29.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
276 "expert.layers.29.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
277 "expert.layers.29.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
278 "expert.layers.29.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
279 "expert.layers.3.input_layernorm.weight": "model-00004-of-00005.safetensors",
280 "expert.layers.3.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
281 "expert.layers.3.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
282 "expert.layers.3.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
283 "expert.layers.3.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
284 "expert.layers.3.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
285 "expert.layers.3.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
286 "expert.layers.3.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
287 "expert.layers.3.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
288 "expert.layers.3.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
289 "expert.layers.3.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
290 "expert.layers.30.input_layernorm.weight": "model-00005-of-00005.safetensors",
291 "expert.layers.30.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
292 "expert.layers.30.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
293 "expert.layers.30.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
294 "expert.layers.30.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
295 "expert.layers.30.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
296 "expert.layers.30.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
297 "expert.layers.30.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
298 "expert.layers.30.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
299 "expert.layers.30.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
300 "expert.layers.30.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
301 "expert.layers.31.input_layernorm.weight": "model-00005-of-00005.safetensors",
302 "expert.layers.31.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
303 "expert.layers.31.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
304 "expert.layers.31.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
305 "expert.layers.31.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
306 "expert.layers.31.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
307 "expert.layers.31.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
308 "expert.layers.31.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
309 "expert.layers.31.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
310 "expert.layers.31.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
311 "expert.layers.31.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
312 "expert.layers.32.input_layernorm.weight": "model-00005-of-00005.safetensors",
313 "expert.layers.32.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
314 "expert.layers.32.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
315 "expert.layers.32.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
316 "expert.layers.32.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
317 "expert.layers.32.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
318 "expert.layers.32.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
319 "expert.layers.32.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
320 "expert.layers.32.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
321 "expert.layers.32.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
322 "expert.layers.32.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
323 "expert.layers.33.input_layernorm.weight": "model-00005-of-00005.safetensors",
324 "expert.layers.33.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
325 "expert.layers.33.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
326 "expert.layers.33.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
327 "expert.layers.33.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
328 "expert.layers.33.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
329 "expert.layers.33.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
330 "expert.layers.33.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
331 "expert.layers.33.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
332 "expert.layers.33.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
333 "expert.layers.33.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
334 "expert.layers.34.input_layernorm.weight": "model-00005-of-00005.safetensors",
335 "expert.layers.34.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
336 "expert.layers.34.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
337 "expert.layers.34.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
338 "expert.layers.34.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
339 "expert.layers.34.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
340 "expert.layers.34.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
341 "expert.layers.34.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
342 "expert.layers.34.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
343 "expert.layers.34.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
344 "expert.layers.34.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
345 "expert.layers.35.input_layernorm.weight": "model-00005-of-00005.safetensors",
346 "expert.layers.35.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
347 "expert.layers.35.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
348 "expert.layers.35.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
349 "expert.layers.35.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
350 "expert.layers.35.self_attn.k_norm.weight": "model-00005-of-00005.safetensors",
351 "expert.layers.35.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
352 "expert.layers.35.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
353 "expert.layers.35.self_attn.q_norm.weight": "model-00005-of-00005.safetensors",
354 "expert.layers.35.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
355 "expert.layers.35.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
356 "expert.layers.4.input_layernorm.weight": "model-00004-of-00005.safetensors",
357 "expert.layers.4.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
358 "expert.layers.4.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
359 "expert.layers.4.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
360 "expert.layers.4.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
361 "expert.layers.4.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
362 "expert.layers.4.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
363 "expert.layers.4.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
364 "expert.layers.4.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
365 "expert.layers.4.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
366 "expert.layers.4.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
367 "expert.layers.5.input_layernorm.weight": "model-00004-of-00005.safetensors",
368 "expert.layers.5.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
369 "expert.layers.5.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
370 "expert.layers.5.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
371 "expert.layers.5.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
372 "expert.layers.5.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
373 "expert.layers.5.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
374 "expert.layers.5.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
375 "expert.layers.5.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
376 "expert.layers.5.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
377 "expert.layers.5.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
378 "expert.layers.6.input_layernorm.weight": "model-00004-of-00005.safetensors",
379 "expert.layers.6.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
380 "expert.layers.6.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
381 "expert.layers.6.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
382 "expert.layers.6.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
383 "expert.layers.6.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
384 "expert.layers.6.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
385 "expert.layers.6.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
386 "expert.layers.6.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
387 "expert.layers.6.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
388 "expert.layers.6.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
389 "expert.layers.7.input_layernorm.weight": "model-00004-of-00005.safetensors",
390 "expert.layers.7.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
391 "expert.layers.7.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
392 "expert.layers.7.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
393 "expert.layers.7.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
394 "expert.layers.7.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
395 "expert.layers.7.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
396 "expert.layers.7.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
397 "expert.layers.7.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
398 "expert.layers.7.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
399 "expert.layers.7.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
400 "expert.layers.8.input_layernorm.weight": "model-00004-of-00005.safetensors",
401 "expert.layers.8.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
402 "expert.layers.8.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
403 "expert.layers.8.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
404 "expert.layers.8.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
405 "expert.layers.8.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
406 "expert.layers.8.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
407 "expert.layers.8.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
408 "expert.layers.8.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
409 "expert.layers.8.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
410 "expert.layers.8.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
411 "expert.layers.9.input_layernorm.weight": "model-00004-of-00005.safetensors",
412 "expert.layers.9.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
413 "expert.layers.9.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
414 "expert.layers.9.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
415 "expert.layers.9.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
416 "expert.layers.9.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
417 "expert.layers.9.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
418 "expert.layers.9.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
419 "expert.layers.9.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
420 "expert.layers.9.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
421 "expert.layers.9.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
422 "expert.norm.weight": "model-00005-of-00005.safetensors",
423 "vlm.lm_head.weight": "model-00004-of-00005.safetensors",
424 "vlm.model.language_model.embed_tokens.weight": "model-00001-of-00005.safetensors",
425 "vlm.model.language_model.layers.0.input_layernorm.weight": "model-00001-of-00005.safetensors",
426 "vlm.model.language_model.layers.0.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
427 "vlm.model.language_model.layers.0.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
428 "vlm.model.language_model.layers.0.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
429 "vlm.model.language_model.layers.0.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
430 "vlm.model.language_model.layers.0.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
431 "vlm.model.language_model.layers.0.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
432 "vlm.model.language_model.layers.0.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
433 "vlm.model.language_model.layers.0.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
434 "vlm.model.language_model.layers.0.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
435 "vlm.model.language_model.layers.0.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
436 "vlm.model.language_model.layers.1.input_layernorm.weight": "model-00001-of-00005.safetensors",
437 "vlm.model.language_model.layers.1.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
438 "vlm.model.language_model.layers.1.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
439 "vlm.model.language_model.layers.1.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
440 "vlm.model.language_model.layers.1.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
441 "vlm.model.language_model.layers.1.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
442 "vlm.model.language_model.layers.1.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
443 "vlm.model.language_model.layers.1.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
444 "vlm.model.language_model.layers.1.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
445 "vlm.model.language_model.layers.1.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
446 "vlm.model.language_model.layers.1.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
447 "vlm.model.language_model.layers.10.input_layernorm.weight": "model-00002-of-00005.safetensors",
448 "vlm.model.language_model.layers.10.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
449 "vlm.model.language_model.layers.10.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
450 "vlm.model.language_model.layers.10.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
451 "vlm.model.language_model.layers.10.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
452 "vlm.model.language_model.layers.10.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
453 "vlm.model.language_model.layers.10.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
454 "vlm.model.language_model.layers.10.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
455 "vlm.model.language_model.layers.10.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
456 "vlm.model.language_model.layers.10.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
457 "vlm.model.language_model.layers.10.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
458 "vlm.model.language_model.layers.11.input_layernorm.weight": "model-00002-of-00005.safetensors",
459 "vlm.model.language_model.layers.11.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
460 "vlm.model.language_model.layers.11.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
461 "vlm.model.language_model.layers.11.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
462 "vlm.model.language_model.layers.11.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
463 "vlm.model.language_model.layers.11.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
464 "vlm.model.language_model.layers.11.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
465 "vlm.model.language_model.layers.11.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
466 "vlm.model.language_model.layers.11.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
467 "vlm.model.language_model.layers.11.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
468 "vlm.model.language_model.layers.11.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
469 "vlm.model.language_model.layers.12.input_layernorm.weight": "model-00002-of-00005.safetensors",
470 "vlm.model.language_model.layers.12.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
471 "vlm.model.language_model.layers.12.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
472 "vlm.model.language_model.layers.12.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
473 "vlm.model.language_model.layers.12.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
474 "vlm.model.language_model.layers.12.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
475 "vlm.model.language_model.layers.12.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
476 "vlm.model.language_model.layers.12.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
477 "vlm.model.language_model.layers.12.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
478 "vlm.model.language_model.layers.12.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
479 "vlm.model.language_model.layers.12.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
480 "vlm.model.language_model.layers.13.input_layernorm.weight": "model-00002-of-00005.safetensors",
481 "vlm.model.language_model.layers.13.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
482 "vlm.model.language_model.layers.13.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
483 "vlm.model.language_model.layers.13.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
484 "vlm.model.language_model.layers.13.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
485 "vlm.model.language_model.layers.13.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
486 "vlm.model.language_model.layers.13.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
487 "vlm.model.language_model.layers.13.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
488 "vlm.model.language_model.layers.13.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
489 "vlm.model.language_model.layers.13.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
490 "vlm.model.language_model.layers.13.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
491 "vlm.model.language_model.layers.14.input_layernorm.weight": "model-00002-of-00005.safetensors",
492 "vlm.model.language_model.layers.14.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
493 "vlm.model.language_model.layers.14.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
494 "vlm.model.language_model.layers.14.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
495 "vlm.model.language_model.layers.14.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
496 "vlm.model.language_model.layers.14.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
497 "vlm.model.language_model.layers.14.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
498 "vlm.model.language_model.layers.14.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
499 "vlm.model.language_model.layers.14.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
500 "vlm.model.language_model.layers.14.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
501 "vlm.model.language_model.layers.14.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
502 "vlm.model.language_model.layers.15.input_layernorm.weight": "model-00002-of-00005.safetensors",
503 "vlm.model.language_model.layers.15.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
504 "vlm.model.language_model.layers.15.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
505 "vlm.model.language_model.layers.15.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
506 "vlm.model.language_model.layers.15.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
507 "vlm.model.language_model.layers.15.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
508 "vlm.model.language_model.layers.15.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
509 "vlm.model.language_model.layers.15.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
510 "vlm.model.language_model.layers.15.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
511 "vlm.model.language_model.layers.15.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
512 "vlm.model.language_model.layers.15.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
513 "vlm.model.language_model.layers.16.input_layernorm.weight": "model-00002-of-00005.safetensors",
514 "vlm.model.language_model.layers.16.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
515 "vlm.model.language_model.layers.16.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
516 "vlm.model.language_model.layers.16.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
517 "vlm.model.language_model.layers.16.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
518 "vlm.model.language_model.layers.16.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
519 "vlm.model.language_model.layers.16.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
520 "vlm.model.language_model.layers.16.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
521 "vlm.model.language_model.layers.16.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
522 "vlm.model.language_model.layers.16.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
523 "vlm.model.language_model.layers.16.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
524 "vlm.model.language_model.layers.17.input_layernorm.weight": "model-00002-of-00005.safetensors",
525 "vlm.model.language_model.layers.17.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
526 "vlm.model.language_model.layers.17.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
527 "vlm.model.language_model.layers.17.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
528 "vlm.model.language_model.layers.17.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
529 "vlm.model.language_model.layers.17.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
530 "vlm.model.language_model.layers.17.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
531 "vlm.model.language_model.layers.17.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
532 "vlm.model.language_model.layers.17.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
533 "vlm.model.language_model.layers.17.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
534 "vlm.model.language_model.layers.17.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
535 "vlm.model.language_model.layers.18.input_layernorm.weight": "model-00002-of-00005.safetensors",
536 "vlm.model.language_model.layers.18.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
537 "vlm.model.language_model.layers.18.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
538 "vlm.model.language_model.layers.18.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
539 "vlm.model.language_model.layers.18.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
540 "vlm.model.language_model.layers.18.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
541 "vlm.model.language_model.layers.18.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
542 "vlm.model.language_model.layers.18.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
543 "vlm.model.language_model.layers.18.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
544 "vlm.model.language_model.layers.18.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
545 "vlm.model.language_model.layers.18.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
546 "vlm.model.language_model.layers.19.input_layernorm.weight": "model-00003-of-00005.safetensors",
547 "vlm.model.language_model.layers.19.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
548 "vlm.model.language_model.layers.19.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
549 "vlm.model.language_model.layers.19.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
550 "vlm.model.language_model.layers.19.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
551 "vlm.model.language_model.layers.19.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
552 "vlm.model.language_model.layers.19.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
553 "vlm.model.language_model.layers.19.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
554 "vlm.model.language_model.layers.19.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
555 "vlm.model.language_model.layers.19.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
556 "vlm.model.language_model.layers.19.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
557 "vlm.model.language_model.layers.2.input_layernorm.weight": "model-00001-of-00005.safetensors",
558 "vlm.model.language_model.layers.2.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
559 "vlm.model.language_model.layers.2.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
560 "vlm.model.language_model.layers.2.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
561 "vlm.model.language_model.layers.2.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
562 "vlm.model.language_model.layers.2.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
563 "vlm.model.language_model.layers.2.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
564 "vlm.model.language_model.layers.2.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
565 "vlm.model.language_model.layers.2.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
566 "vlm.model.language_model.layers.2.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
567 "vlm.model.language_model.layers.2.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
568 "vlm.model.language_model.layers.20.input_layernorm.weight": "model-00003-of-00005.safetensors",
569 "vlm.model.language_model.layers.20.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
570 "vlm.model.language_model.layers.20.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
571 "vlm.model.language_model.layers.20.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
572 "vlm.model.language_model.layers.20.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
573 "vlm.model.language_model.layers.20.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
574 "vlm.model.language_model.layers.20.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
575 "vlm.model.language_model.layers.20.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
576 "vlm.model.language_model.layers.20.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
577 "vlm.model.language_model.layers.20.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
578 "vlm.model.language_model.layers.20.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
579 "vlm.model.language_model.layers.21.input_layernorm.weight": "model-00003-of-00005.safetensors",
580 "vlm.model.language_model.layers.21.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
581 "vlm.model.language_model.layers.21.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
582 "vlm.model.language_model.layers.21.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
583 "vlm.model.language_model.layers.21.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
584 "vlm.model.language_model.layers.21.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
585 "vlm.model.language_model.layers.21.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
586 "vlm.model.language_model.layers.21.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
587 "vlm.model.language_model.layers.21.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
588 "vlm.model.language_model.layers.21.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
589 "vlm.model.language_model.layers.21.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
590 "vlm.model.language_model.layers.22.input_layernorm.weight": "model-00003-of-00005.safetensors",
591 "vlm.model.language_model.layers.22.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
592 "vlm.model.language_model.layers.22.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
593 "vlm.model.language_model.layers.22.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
594 "vlm.model.language_model.layers.22.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
595 "vlm.model.language_model.layers.22.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
596 "vlm.model.language_model.layers.22.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
597 "vlm.model.language_model.layers.22.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
598 "vlm.model.language_model.layers.22.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
599 "vlm.model.language_model.layers.22.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
600 "vlm.model.language_model.layers.22.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
601 "vlm.model.language_model.layers.23.input_layernorm.weight": "model-00003-of-00005.safetensors",
602 "vlm.model.language_model.layers.23.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
603 "vlm.model.language_model.layers.23.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
604 "vlm.model.language_model.layers.23.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
605 "vlm.model.language_model.layers.23.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
606 "vlm.model.language_model.layers.23.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
607 "vlm.model.language_model.layers.23.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
608 "vlm.model.language_model.layers.23.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
609 "vlm.model.language_model.layers.23.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
610 "vlm.model.language_model.layers.23.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
611 "vlm.model.language_model.layers.23.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
612 "vlm.model.language_model.layers.24.input_layernorm.weight": "model-00003-of-00005.safetensors",
613 "vlm.model.language_model.layers.24.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
614 "vlm.model.language_model.layers.24.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
615 "vlm.model.language_model.layers.24.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
616 "vlm.model.language_model.layers.24.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
617 "vlm.model.language_model.layers.24.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
618 "vlm.model.language_model.layers.24.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
619 "vlm.model.language_model.layers.24.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
620 "vlm.model.language_model.layers.24.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
621 "vlm.model.language_model.layers.24.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
622 "vlm.model.language_model.layers.24.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
623 "vlm.model.language_model.layers.25.input_layernorm.weight": "model-00003-of-00005.safetensors",
624 "vlm.model.language_model.layers.25.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
625 "vlm.model.language_model.layers.25.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
626 "vlm.model.language_model.layers.25.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
627 "vlm.model.language_model.layers.25.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
628 "vlm.model.language_model.layers.25.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
629 "vlm.model.language_model.layers.25.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
630 "vlm.model.language_model.layers.25.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
631 "vlm.model.language_model.layers.25.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
632 "vlm.model.language_model.layers.25.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
633 "vlm.model.language_model.layers.25.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
634 "vlm.model.language_model.layers.26.input_layernorm.weight": "model-00003-of-00005.safetensors",
635 "vlm.model.language_model.layers.26.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
636 "vlm.model.language_model.layers.26.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
637 "vlm.model.language_model.layers.26.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
638 "vlm.model.language_model.layers.26.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
639 "vlm.model.language_model.layers.26.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
640 "vlm.model.language_model.layers.26.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
641 "vlm.model.language_model.layers.26.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
642 "vlm.model.language_model.layers.26.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
643 "vlm.model.language_model.layers.26.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
644 "vlm.model.language_model.layers.26.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
645 "vlm.model.language_model.layers.27.input_layernorm.weight": "model-00003-of-00005.safetensors",
646 "vlm.model.language_model.layers.27.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
647 "vlm.model.language_model.layers.27.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
648 "vlm.model.language_model.layers.27.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
649 "vlm.model.language_model.layers.27.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
650 "vlm.model.language_model.layers.27.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
651 "vlm.model.language_model.layers.27.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
652 "vlm.model.language_model.layers.27.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
653 "vlm.model.language_model.layers.27.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
654 "vlm.model.language_model.layers.27.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
655 "vlm.model.language_model.layers.27.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
656 "vlm.model.language_model.layers.28.input_layernorm.weight": "model-00003-of-00005.safetensors",
657 "vlm.model.language_model.layers.28.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
658 "vlm.model.language_model.layers.28.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
659 "vlm.model.language_model.layers.28.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
660 "vlm.model.language_model.layers.28.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
661 "vlm.model.language_model.layers.28.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
662 "vlm.model.language_model.layers.28.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
663 "vlm.model.language_model.layers.28.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
664 "vlm.model.language_model.layers.28.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
665 "vlm.model.language_model.layers.28.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
666 "vlm.model.language_model.layers.28.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
667 "vlm.model.language_model.layers.29.input_layernorm.weight": "model-00003-of-00005.safetensors",
668 "vlm.model.language_model.layers.29.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
669 "vlm.model.language_model.layers.29.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
670 "vlm.model.language_model.layers.29.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
671 "vlm.model.language_model.layers.29.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
672 "vlm.model.language_model.layers.29.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
673 "vlm.model.language_model.layers.29.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
674 "vlm.model.language_model.layers.29.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
675 "vlm.model.language_model.layers.29.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
676 "vlm.model.language_model.layers.29.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
677 "vlm.model.language_model.layers.29.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
678 "vlm.model.language_model.layers.3.input_layernorm.weight": "model-00001-of-00005.safetensors",
679 "vlm.model.language_model.layers.3.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
680 "vlm.model.language_model.layers.3.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
681 "vlm.model.language_model.layers.3.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
682 "vlm.model.language_model.layers.3.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
683 "vlm.model.language_model.layers.3.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
684 "vlm.model.language_model.layers.3.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
685 "vlm.model.language_model.layers.3.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
686 "vlm.model.language_model.layers.3.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
687 "vlm.model.language_model.layers.3.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
688 "vlm.model.language_model.layers.3.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
689 "vlm.model.language_model.layers.30.input_layernorm.weight": "model-00003-of-00005.safetensors",
690 "vlm.model.language_model.layers.30.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
691 "vlm.model.language_model.layers.30.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
692 "vlm.model.language_model.layers.30.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
693 "vlm.model.language_model.layers.30.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
694 "vlm.model.language_model.layers.30.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
695 "vlm.model.language_model.layers.30.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
696 "vlm.model.language_model.layers.30.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
697 "vlm.model.language_model.layers.30.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
698 "vlm.model.language_model.layers.30.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
699 "vlm.model.language_model.layers.30.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
700 "vlm.model.language_model.layers.31.input_layernorm.weight": "model-00003-of-00005.safetensors",
701 "vlm.model.language_model.layers.31.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
702 "vlm.model.language_model.layers.31.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
703 "vlm.model.language_model.layers.31.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
704 "vlm.model.language_model.layers.31.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
705 "vlm.model.language_model.layers.31.self_attn.k_norm.weight": "model-00003-of-00005.safetensors",
706 "vlm.model.language_model.layers.31.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
707 "vlm.model.language_model.layers.31.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
708 "vlm.model.language_model.layers.31.self_attn.q_norm.weight": "model-00003-of-00005.safetensors",
709 "vlm.model.language_model.layers.31.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
710 "vlm.model.language_model.layers.31.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
711 "vlm.model.language_model.layers.32.input_layernorm.weight": "model-00004-of-00005.safetensors",
712 "vlm.model.language_model.layers.32.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
713 "vlm.model.language_model.layers.32.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
714 "vlm.model.language_model.layers.32.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
715 "vlm.model.language_model.layers.32.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
716 "vlm.model.language_model.layers.32.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
717 "vlm.model.language_model.layers.32.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
718 "vlm.model.language_model.layers.32.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
719 "vlm.model.language_model.layers.32.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
720 "vlm.model.language_model.layers.32.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
721 "vlm.model.language_model.layers.32.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
722 "vlm.model.language_model.layers.33.input_layernorm.weight": "model-00004-of-00005.safetensors",
723 "vlm.model.language_model.layers.33.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
724 "vlm.model.language_model.layers.33.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
725 "vlm.model.language_model.layers.33.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
726 "vlm.model.language_model.layers.33.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
727 "vlm.model.language_model.layers.33.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
728 "vlm.model.language_model.layers.33.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
729 "vlm.model.language_model.layers.33.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
730 "vlm.model.language_model.layers.33.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
731 "vlm.model.language_model.layers.33.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
732 "vlm.model.language_model.layers.33.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
733 "vlm.model.language_model.layers.34.input_layernorm.weight": "model-00004-of-00005.safetensors",
734 "vlm.model.language_model.layers.34.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
735 "vlm.model.language_model.layers.34.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
736 "vlm.model.language_model.layers.34.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
737 "vlm.model.language_model.layers.34.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
738 "vlm.model.language_model.layers.34.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
739 "vlm.model.language_model.layers.34.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
740 "vlm.model.language_model.layers.34.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
741 "vlm.model.language_model.layers.34.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
742 "vlm.model.language_model.layers.34.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
743 "vlm.model.language_model.layers.34.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
744 "vlm.model.language_model.layers.35.input_layernorm.weight": "model-00004-of-00005.safetensors",
745 "vlm.model.language_model.layers.35.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
746 "vlm.model.language_model.layers.35.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
747 "vlm.model.language_model.layers.35.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
748 "vlm.model.language_model.layers.35.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
749 "vlm.model.language_model.layers.35.self_attn.k_norm.weight": "model-00004-of-00005.safetensors",
750 "vlm.model.language_model.layers.35.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
751 "vlm.model.language_model.layers.35.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
752 "vlm.model.language_model.layers.35.self_attn.q_norm.weight": "model-00004-of-00005.safetensors",
753 "vlm.model.language_model.layers.35.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
754 "vlm.model.language_model.layers.35.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
755 "vlm.model.language_model.layers.4.input_layernorm.weight": "model-00001-of-00005.safetensors",
756 "vlm.model.language_model.layers.4.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
757 "vlm.model.language_model.layers.4.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
758 "vlm.model.language_model.layers.4.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
759 "vlm.model.language_model.layers.4.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
760 "vlm.model.language_model.layers.4.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
761 "vlm.model.language_model.layers.4.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
762 "vlm.model.language_model.layers.4.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
763 "vlm.model.language_model.layers.4.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
764 "vlm.model.language_model.layers.4.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
765 "vlm.model.language_model.layers.4.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
766 "vlm.model.language_model.layers.5.input_layernorm.weight": "model-00001-of-00005.safetensors",
767 "vlm.model.language_model.layers.5.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
768 "vlm.model.language_model.layers.5.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
769 "vlm.model.language_model.layers.5.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
770 "vlm.model.language_model.layers.5.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
771 "vlm.model.language_model.layers.5.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
772 "vlm.model.language_model.layers.5.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
773 "vlm.model.language_model.layers.5.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
774 "vlm.model.language_model.layers.5.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
775 "vlm.model.language_model.layers.5.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
776 "vlm.model.language_model.layers.5.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
777 "vlm.model.language_model.layers.6.input_layernorm.weight": "model-00002-of-00005.safetensors",
778 "vlm.model.language_model.layers.6.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
779 "vlm.model.language_model.layers.6.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
780 "vlm.model.language_model.layers.6.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
781 "vlm.model.language_model.layers.6.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
782 "vlm.model.language_model.layers.6.self_attn.k_norm.weight": "model-00001-of-00005.safetensors",
783 "vlm.model.language_model.layers.6.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
784 "vlm.model.language_model.layers.6.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
785 "vlm.model.language_model.layers.6.self_attn.q_norm.weight": "model-00001-of-00005.safetensors",
786 "vlm.model.language_model.layers.6.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
787 "vlm.model.language_model.layers.6.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
788 "vlm.model.language_model.layers.7.input_layernorm.weight": "model-00002-of-00005.safetensors",
789 "vlm.model.language_model.layers.7.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
790 "vlm.model.language_model.layers.7.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
791 "vlm.model.language_model.layers.7.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
792 "vlm.model.language_model.layers.7.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
793 "vlm.model.language_model.layers.7.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
794 "vlm.model.language_model.layers.7.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
795 "vlm.model.language_model.layers.7.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
796 "vlm.model.language_model.layers.7.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
797 "vlm.model.language_model.layers.7.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
798 "vlm.model.language_model.layers.7.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
799 "vlm.model.language_model.layers.8.input_layernorm.weight": "model-00002-of-00005.safetensors",
800 "vlm.model.language_model.layers.8.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
801 "vlm.model.language_model.layers.8.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
802 "vlm.model.language_model.layers.8.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
803 "vlm.model.language_model.layers.8.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
804 "vlm.model.language_model.layers.8.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
805 "vlm.model.language_model.layers.8.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
806 "vlm.model.language_model.layers.8.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
807 "vlm.model.language_model.layers.8.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
808 "vlm.model.language_model.layers.8.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
809 "vlm.model.language_model.layers.8.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
810 "vlm.model.language_model.layers.9.input_layernorm.weight": "model-00002-of-00005.safetensors",
811 "vlm.model.language_model.layers.9.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
812 "vlm.model.language_model.layers.9.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
813 "vlm.model.language_model.layers.9.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
814 "vlm.model.language_model.layers.9.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
815 "vlm.model.language_model.layers.9.self_attn.k_norm.weight": "model-00002-of-00005.safetensors",
816 "vlm.model.language_model.layers.9.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
817 "vlm.model.language_model.layers.9.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
818 "vlm.model.language_model.layers.9.self_attn.q_norm.weight": "model-00002-of-00005.safetensors",
819 "vlm.model.language_model.layers.9.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
820 "vlm.model.language_model.layers.9.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
821 "vlm.model.language_model.norm.weight": "model-00004-of-00005.safetensors",
822 "vlm.model.visual.blocks.0.attn.proj.bias": "model-00001-of-00005.safetensors",
823 "vlm.model.visual.blocks.0.attn.proj.weight": "model-00001-of-00005.safetensors",
824 "vlm.model.visual.blocks.0.attn.qkv.bias": "model-00001-of-00005.safetensors",
825 "vlm.model.visual.blocks.0.attn.qkv.weight": "model-00001-of-00005.safetensors",
826 "vlm.model.visual.blocks.0.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
827 "vlm.model.visual.blocks.0.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
828 "vlm.model.visual.blocks.0.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
829 "vlm.model.visual.blocks.0.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
830 "vlm.model.visual.blocks.0.norm1.bias": "model-00001-of-00005.safetensors",
831 "vlm.model.visual.blocks.0.norm1.weight": "model-00001-of-00005.safetensors",
832 "vlm.model.visual.blocks.0.norm2.bias": "model-00001-of-00005.safetensors",
833 "vlm.model.visual.blocks.0.norm2.weight": "model-00001-of-00005.safetensors",
834 "vlm.model.visual.blocks.1.attn.proj.bias": "model-00001-of-00005.safetensors",
835 "vlm.model.visual.blocks.1.attn.proj.weight": "model-00001-of-00005.safetensors",
836 "vlm.model.visual.blocks.1.attn.qkv.bias": "model-00001-of-00005.safetensors",
837 "vlm.model.visual.blocks.1.attn.qkv.weight": "model-00001-of-00005.safetensors",
838 "vlm.model.visual.blocks.1.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
839 "vlm.model.visual.blocks.1.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
840 "vlm.model.visual.blocks.1.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
841 "vlm.model.visual.blocks.1.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
842 "vlm.model.visual.blocks.1.norm1.bias": "model-00001-of-00005.safetensors",
843 "vlm.model.visual.blocks.1.norm1.weight": "model-00001-of-00005.safetensors",
844 "vlm.model.visual.blocks.1.norm2.bias": "model-00001-of-00005.safetensors",
845 "vlm.model.visual.blocks.1.norm2.weight": "model-00001-of-00005.safetensors",
846 "vlm.model.visual.blocks.10.attn.proj.bias": "model-00001-of-00005.safetensors",
847 "vlm.model.visual.blocks.10.attn.proj.weight": "model-00001-of-00005.safetensors",
848 "vlm.model.visual.blocks.10.attn.qkv.bias": "model-00001-of-00005.safetensors",
849 "vlm.model.visual.blocks.10.attn.qkv.weight": "model-00001-of-00005.safetensors",
850 "vlm.model.visual.blocks.10.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
851 "vlm.model.visual.blocks.10.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
852 "vlm.model.visual.blocks.10.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
853 "vlm.model.visual.blocks.10.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
854 "vlm.model.visual.blocks.10.norm1.bias": "model-00001-of-00005.safetensors",
855 "vlm.model.visual.blocks.10.norm1.weight": "model-00001-of-00005.safetensors",
856 "vlm.model.visual.blocks.10.norm2.bias": "model-00001-of-00005.safetensors",
857 "vlm.model.visual.blocks.10.norm2.weight": "model-00001-of-00005.safetensors",
858 "vlm.model.visual.blocks.11.attn.proj.bias": "model-00001-of-00005.safetensors",
859 "vlm.model.visual.blocks.11.attn.proj.weight": "model-00001-of-00005.safetensors",
860 "vlm.model.visual.blocks.11.attn.qkv.bias": "model-00001-of-00005.safetensors",
861 "vlm.model.visual.blocks.11.attn.qkv.weight": "model-00001-of-00005.safetensors",
862 "vlm.model.visual.blocks.11.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
863 "vlm.model.visual.blocks.11.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
864 "vlm.model.visual.blocks.11.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
865 "vlm.model.visual.blocks.11.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
866 "vlm.model.visual.blocks.11.norm1.bias": "model-00001-of-00005.safetensors",
867 "vlm.model.visual.blocks.11.norm1.weight": "model-00001-of-00005.safetensors",
868 "vlm.model.visual.blocks.11.norm2.bias": "model-00001-of-00005.safetensors",
869 "vlm.model.visual.blocks.11.norm2.weight": "model-00001-of-00005.safetensors",
870 "vlm.model.visual.blocks.12.attn.proj.bias": "model-00001-of-00005.safetensors",
871 "vlm.model.visual.blocks.12.attn.proj.weight": "model-00001-of-00005.safetensors",
872 "vlm.model.visual.blocks.12.attn.qkv.bias": "model-00001-of-00005.safetensors",
873 "vlm.model.visual.blocks.12.attn.qkv.weight": "model-00001-of-00005.safetensors",
874 "vlm.model.visual.blocks.12.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
875 "vlm.model.visual.blocks.12.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
876 "vlm.model.visual.blocks.12.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
877 "vlm.model.visual.blocks.12.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
878 "vlm.model.visual.blocks.12.norm1.bias": "model-00001-of-00005.safetensors",
879 "vlm.model.visual.blocks.12.norm1.weight": "model-00001-of-00005.safetensors",
880 "vlm.model.visual.blocks.12.norm2.bias": "model-00001-of-00005.safetensors",
881 "vlm.model.visual.blocks.12.norm2.weight": "model-00001-of-00005.safetensors",
882 "vlm.model.visual.blocks.13.attn.proj.bias": "model-00001-of-00005.safetensors",
883 "vlm.model.visual.blocks.13.attn.proj.weight": "model-00001-of-00005.safetensors",
884 "vlm.model.visual.blocks.13.attn.qkv.bias": "model-00001-of-00005.safetensors",
885 "vlm.model.visual.blocks.13.attn.qkv.weight": "model-00001-of-00005.safetensors",
886 "vlm.model.visual.blocks.13.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
887 "vlm.model.visual.blocks.13.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
888 "vlm.model.visual.blocks.13.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
889 "vlm.model.visual.blocks.13.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
890 "vlm.model.visual.blocks.13.norm1.bias": "model-00001-of-00005.safetensors",
891 "vlm.model.visual.blocks.13.norm1.weight": "model-00001-of-00005.safetensors",
892 "vlm.model.visual.blocks.13.norm2.bias": "model-00001-of-00005.safetensors",
893 "vlm.model.visual.blocks.13.norm2.weight": "model-00001-of-00005.safetensors",
894 "vlm.model.visual.blocks.14.attn.proj.bias": "model-00001-of-00005.safetensors",
895 "vlm.model.visual.blocks.14.attn.proj.weight": "model-00001-of-00005.safetensors",
896 "vlm.model.visual.blocks.14.attn.qkv.bias": "model-00001-of-00005.safetensors",
897 "vlm.model.visual.blocks.14.attn.qkv.weight": "model-00001-of-00005.safetensors",
898 "vlm.model.visual.blocks.14.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
899 "vlm.model.visual.blocks.14.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
900 "vlm.model.visual.blocks.14.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
901 "vlm.model.visual.blocks.14.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
902 "vlm.model.visual.blocks.14.norm1.bias": "model-00001-of-00005.safetensors",
903 "vlm.model.visual.blocks.14.norm1.weight": "model-00001-of-00005.safetensors",
904 "vlm.model.visual.blocks.14.norm2.bias": "model-00001-of-00005.safetensors",
905 "vlm.model.visual.blocks.14.norm2.weight": "model-00001-of-00005.safetensors",
906 "vlm.model.visual.blocks.15.attn.proj.bias": "model-00001-of-00005.safetensors",
907 "vlm.model.visual.blocks.15.attn.proj.weight": "model-00001-of-00005.safetensors",
908 "vlm.model.visual.blocks.15.attn.qkv.bias": "model-00001-of-00005.safetensors",
909 "vlm.model.visual.blocks.15.attn.qkv.weight": "model-00001-of-00005.safetensors",
910 "vlm.model.visual.blocks.15.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
911 "vlm.model.visual.blocks.15.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
912 "vlm.model.visual.blocks.15.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
913 "vlm.model.visual.blocks.15.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
914 "vlm.model.visual.blocks.15.norm1.bias": "model-00001-of-00005.safetensors",
915 "vlm.model.visual.blocks.15.norm1.weight": "model-00001-of-00005.safetensors",
916 "vlm.model.visual.blocks.15.norm2.bias": "model-00001-of-00005.safetensors",
917 "vlm.model.visual.blocks.15.norm2.weight": "model-00001-of-00005.safetensors",
918 "vlm.model.visual.blocks.16.attn.proj.bias": "model-00001-of-00005.safetensors",
919 "vlm.model.visual.blocks.16.attn.proj.weight": "model-00001-of-00005.safetensors",
920 "vlm.model.visual.blocks.16.attn.qkv.bias": "model-00001-of-00005.safetensors",
921 "vlm.model.visual.blocks.16.attn.qkv.weight": "model-00001-of-00005.safetensors",
922 "vlm.model.visual.blocks.16.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
923 "vlm.model.visual.blocks.16.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
924 "vlm.model.visual.blocks.16.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
925 "vlm.model.visual.blocks.16.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
926 "vlm.model.visual.blocks.16.norm1.bias": "model-00001-of-00005.safetensors",
927 "vlm.model.visual.blocks.16.norm1.weight": "model-00001-of-00005.safetensors",
928 "vlm.model.visual.blocks.16.norm2.bias": "model-00001-of-00005.safetensors",
929 "vlm.model.visual.blocks.16.norm2.weight": "model-00001-of-00005.safetensors",
930 "vlm.model.visual.blocks.17.attn.proj.bias": "model-00001-of-00005.safetensors",
931 "vlm.model.visual.blocks.17.attn.proj.weight": "model-00001-of-00005.safetensors",
932 "vlm.model.visual.blocks.17.attn.qkv.bias": "model-00001-of-00005.safetensors",
933 "vlm.model.visual.blocks.17.attn.qkv.weight": "model-00001-of-00005.safetensors",
934 "vlm.model.visual.blocks.17.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
935 "vlm.model.visual.blocks.17.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
936 "vlm.model.visual.blocks.17.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
937 "vlm.model.visual.blocks.17.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
938 "vlm.model.visual.blocks.17.norm1.bias": "model-00001-of-00005.safetensors",
939 "vlm.model.visual.blocks.17.norm1.weight": "model-00001-of-00005.safetensors",
940 "vlm.model.visual.blocks.17.norm2.bias": "model-00001-of-00005.safetensors",
941 "vlm.model.visual.blocks.17.norm2.weight": "model-00001-of-00005.safetensors",
942 "vlm.model.visual.blocks.18.attn.proj.bias": "model-00001-of-00005.safetensors",
943 "vlm.model.visual.blocks.18.attn.proj.weight": "model-00001-of-00005.safetensors",
944 "vlm.model.visual.blocks.18.attn.qkv.bias": "model-00001-of-00005.safetensors",
945 "vlm.model.visual.blocks.18.attn.qkv.weight": "model-00001-of-00005.safetensors",
946 "vlm.model.visual.blocks.18.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
947 "vlm.model.visual.blocks.18.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
948 "vlm.model.visual.blocks.18.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
949 "vlm.model.visual.blocks.18.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
950 "vlm.model.visual.blocks.18.norm1.bias": "model-00001-of-00005.safetensors",
951 "vlm.model.visual.blocks.18.norm1.weight": "model-00001-of-00005.safetensors",
952 "vlm.model.visual.blocks.18.norm2.bias": "model-00001-of-00005.safetensors",
953 "vlm.model.visual.blocks.18.norm2.weight": "model-00001-of-00005.safetensors",
954 "vlm.model.visual.blocks.19.attn.proj.bias": "model-00001-of-00005.safetensors",
955 "vlm.model.visual.blocks.19.attn.proj.weight": "model-00001-of-00005.safetensors",
956 "vlm.model.visual.blocks.19.attn.qkv.bias": "model-00001-of-00005.safetensors",
957 "vlm.model.visual.blocks.19.attn.qkv.weight": "model-00001-of-00005.safetensors",
958 "vlm.model.visual.blocks.19.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
959 "vlm.model.visual.blocks.19.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
960 "vlm.model.visual.blocks.19.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
961 "vlm.model.visual.blocks.19.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
962 "vlm.model.visual.blocks.19.norm1.bias": "model-00001-of-00005.safetensors",
963 "vlm.model.visual.blocks.19.norm1.weight": "model-00001-of-00005.safetensors",
964 "vlm.model.visual.blocks.19.norm2.bias": "model-00001-of-00005.safetensors",
965 "vlm.model.visual.blocks.19.norm2.weight": "model-00001-of-00005.safetensors",
966 "vlm.model.visual.blocks.2.attn.proj.bias": "model-00001-of-00005.safetensors",
967 "vlm.model.visual.blocks.2.attn.proj.weight": "model-00001-of-00005.safetensors",
968 "vlm.model.visual.blocks.2.attn.qkv.bias": "model-00001-of-00005.safetensors",
969 "vlm.model.visual.blocks.2.attn.qkv.weight": "model-00001-of-00005.safetensors",
970 "vlm.model.visual.blocks.2.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
971 "vlm.model.visual.blocks.2.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
972 "vlm.model.visual.blocks.2.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
973 "vlm.model.visual.blocks.2.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
974 "vlm.model.visual.blocks.2.norm1.bias": "model-00001-of-00005.safetensors",
975 "vlm.model.visual.blocks.2.norm1.weight": "model-00001-of-00005.safetensors",
976 "vlm.model.visual.blocks.2.norm2.bias": "model-00001-of-00005.safetensors",
977 "vlm.model.visual.blocks.2.norm2.weight": "model-00001-of-00005.safetensors",
978 "vlm.model.visual.blocks.20.attn.proj.bias": "model-00001-of-00005.safetensors",
979 "vlm.model.visual.blocks.20.attn.proj.weight": "model-00001-of-00005.safetensors",
980 "vlm.model.visual.blocks.20.attn.qkv.bias": "model-00001-of-00005.safetensors",
981 "vlm.model.visual.blocks.20.attn.qkv.weight": "model-00001-of-00005.safetensors",
982 "vlm.model.visual.blocks.20.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
983 "vlm.model.visual.blocks.20.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
984 "vlm.model.visual.blocks.20.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
985 "vlm.model.visual.blocks.20.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
986 "vlm.model.visual.blocks.20.norm1.bias": "model-00001-of-00005.safetensors",
987 "vlm.model.visual.blocks.20.norm1.weight": "model-00001-of-00005.safetensors",
988 "vlm.model.visual.blocks.20.norm2.bias": "model-00001-of-00005.safetensors",
989 "vlm.model.visual.blocks.20.norm2.weight": "model-00001-of-00005.safetensors",
990 "vlm.model.visual.blocks.21.attn.proj.bias": "model-00001-of-00005.safetensors",
991 "vlm.model.visual.blocks.21.attn.proj.weight": "model-00001-of-00005.safetensors",
992 "vlm.model.visual.blocks.21.attn.qkv.bias": "model-00001-of-00005.safetensors",
993 "vlm.model.visual.blocks.21.attn.qkv.weight": "model-00001-of-00005.safetensors",
994 "vlm.model.visual.blocks.21.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
995 "vlm.model.visual.blocks.21.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
996 "vlm.model.visual.blocks.21.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
997 "vlm.model.visual.blocks.21.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
998 "vlm.model.visual.blocks.21.norm1.bias": "model-00001-of-00005.safetensors",
999 "vlm.model.visual.blocks.21.norm1.weight": "model-00001-of-00005.safetensors",
1000 "vlm.model.visual.blocks.21.norm2.bias": "model-00001-of-00005.safetensors",
1001 "vlm.model.visual.blocks.21.norm2.weight": "model-00001-of-00005.safetensors",
1002 "vlm.model.visual.blocks.22.attn.proj.bias": "model-00001-of-00005.safetensors",
1003 "vlm.model.visual.blocks.22.attn.proj.weight": "model-00001-of-00005.safetensors",
1004 "vlm.model.visual.blocks.22.attn.qkv.bias": "model-00001-of-00005.safetensors",
1005 "vlm.model.visual.blocks.22.attn.qkv.weight": "model-00001-of-00005.safetensors",
1006 "vlm.model.visual.blocks.22.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1007 "vlm.model.visual.blocks.22.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1008 "vlm.model.visual.blocks.22.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1009 "vlm.model.visual.blocks.22.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1010 "vlm.model.visual.blocks.22.norm1.bias": "model-00001-of-00005.safetensors",
1011 "vlm.model.visual.blocks.22.norm1.weight": "model-00001-of-00005.safetensors",
1012 "vlm.model.visual.blocks.22.norm2.bias": "model-00001-of-00005.safetensors",
1013 "vlm.model.visual.blocks.22.norm2.weight": "model-00001-of-00005.safetensors",
1014 "vlm.model.visual.blocks.23.attn.proj.bias": "model-00001-of-00005.safetensors",
1015 "vlm.model.visual.blocks.23.attn.proj.weight": "model-00001-of-00005.safetensors",
1016 "vlm.model.visual.blocks.23.attn.qkv.bias": "model-00001-of-00005.safetensors",
1017 "vlm.model.visual.blocks.23.attn.qkv.weight": "model-00001-of-00005.safetensors",
1018 "vlm.model.visual.blocks.23.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1019 "vlm.model.visual.blocks.23.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1020 "vlm.model.visual.blocks.23.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1021 "vlm.model.visual.blocks.23.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1022 "vlm.model.visual.blocks.23.norm1.bias": "model-00001-of-00005.safetensors",
1023 "vlm.model.visual.blocks.23.norm1.weight": "model-00001-of-00005.safetensors",
1024 "vlm.model.visual.blocks.23.norm2.bias": "model-00001-of-00005.safetensors",
1025 "vlm.model.visual.blocks.23.norm2.weight": "model-00001-of-00005.safetensors",
1026 "vlm.model.visual.blocks.24.attn.proj.bias": "model-00001-of-00005.safetensors",
1027 "vlm.model.visual.blocks.24.attn.proj.weight": "model-00001-of-00005.safetensors",
1028 "vlm.model.visual.blocks.24.attn.qkv.bias": "model-00001-of-00005.safetensors",
1029 "vlm.model.visual.blocks.24.attn.qkv.weight": "model-00001-of-00005.safetensors",
1030 "vlm.model.visual.blocks.24.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1031 "vlm.model.visual.blocks.24.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1032 "vlm.model.visual.blocks.24.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1033 "vlm.model.visual.blocks.24.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1034 "vlm.model.visual.blocks.24.norm1.bias": "model-00001-of-00005.safetensors",
1035 "vlm.model.visual.blocks.24.norm1.weight": "model-00001-of-00005.safetensors",
1036 "vlm.model.visual.blocks.24.norm2.bias": "model-00001-of-00005.safetensors",
1037 "vlm.model.visual.blocks.24.norm2.weight": "model-00001-of-00005.safetensors",
1038 "vlm.model.visual.blocks.25.attn.proj.bias": "model-00001-of-00005.safetensors",
1039 "vlm.model.visual.blocks.25.attn.proj.weight": "model-00001-of-00005.safetensors",
1040 "vlm.model.visual.blocks.25.attn.qkv.bias": "model-00001-of-00005.safetensors",
1041 "vlm.model.visual.blocks.25.attn.qkv.weight": "model-00001-of-00005.safetensors",
1042 "vlm.model.visual.blocks.25.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1043 "vlm.model.visual.blocks.25.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1044 "vlm.model.visual.blocks.25.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1045 "vlm.model.visual.blocks.25.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1046 "vlm.model.visual.blocks.25.norm1.bias": "model-00001-of-00005.safetensors",
1047 "vlm.model.visual.blocks.25.norm1.weight": "model-00001-of-00005.safetensors",
1048 "vlm.model.visual.blocks.25.norm2.bias": "model-00001-of-00005.safetensors",
1049 "vlm.model.visual.blocks.25.norm2.weight": "model-00001-of-00005.safetensors",
1050 "vlm.model.visual.blocks.26.attn.proj.bias": "model-00001-of-00005.safetensors",
1051 "vlm.model.visual.blocks.26.attn.proj.weight": "model-00001-of-00005.safetensors",
1052 "vlm.model.visual.blocks.26.attn.qkv.bias": "model-00001-of-00005.safetensors",
1053 "vlm.model.visual.blocks.26.attn.qkv.weight": "model-00001-of-00005.safetensors",
1054 "vlm.model.visual.blocks.26.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1055 "vlm.model.visual.blocks.26.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1056 "vlm.model.visual.blocks.26.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1057 "vlm.model.visual.blocks.26.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1058 "vlm.model.visual.blocks.26.norm1.bias": "model-00001-of-00005.safetensors",
1059 "vlm.model.visual.blocks.26.norm1.weight": "model-00001-of-00005.safetensors",
1060 "vlm.model.visual.blocks.26.norm2.bias": "model-00001-of-00005.safetensors",
1061 "vlm.model.visual.blocks.26.norm2.weight": "model-00001-of-00005.safetensors",
1062 "vlm.model.visual.blocks.3.attn.proj.bias": "model-00001-of-00005.safetensors",
1063 "vlm.model.visual.blocks.3.attn.proj.weight": "model-00001-of-00005.safetensors",
1064 "vlm.model.visual.blocks.3.attn.qkv.bias": "model-00001-of-00005.safetensors",
1065 "vlm.model.visual.blocks.3.attn.qkv.weight": "model-00001-of-00005.safetensors",
1066 "vlm.model.visual.blocks.3.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1067 "vlm.model.visual.blocks.3.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1068 "vlm.model.visual.blocks.3.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1069 "vlm.model.visual.blocks.3.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1070 "vlm.model.visual.blocks.3.norm1.bias": "model-00001-of-00005.safetensors",
1071 "vlm.model.visual.blocks.3.norm1.weight": "model-00001-of-00005.safetensors",
1072 "vlm.model.visual.blocks.3.norm2.bias": "model-00001-of-00005.safetensors",
1073 "vlm.model.visual.blocks.3.norm2.weight": "model-00001-of-00005.safetensors",
1074 "vlm.model.visual.blocks.4.attn.proj.bias": "model-00001-of-00005.safetensors",
1075 "vlm.model.visual.blocks.4.attn.proj.weight": "model-00001-of-00005.safetensors",
1076 "vlm.model.visual.blocks.4.attn.qkv.bias": "model-00001-of-00005.safetensors",
1077 "vlm.model.visual.blocks.4.attn.qkv.weight": "model-00001-of-00005.safetensors",
1078 "vlm.model.visual.blocks.4.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1079 "vlm.model.visual.blocks.4.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1080 "vlm.model.visual.blocks.4.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1081 "vlm.model.visual.blocks.4.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1082 "vlm.model.visual.blocks.4.norm1.bias": "model-00001-of-00005.safetensors",
1083 "vlm.model.visual.blocks.4.norm1.weight": "model-00001-of-00005.safetensors",
1084 "vlm.model.visual.blocks.4.norm2.bias": "model-00001-of-00005.safetensors",
1085 "vlm.model.visual.blocks.4.norm2.weight": "model-00001-of-00005.safetensors",
1086 "vlm.model.visual.blocks.5.attn.proj.bias": "model-00001-of-00005.safetensors",
1087 "vlm.model.visual.blocks.5.attn.proj.weight": "model-00001-of-00005.safetensors",
1088 "vlm.model.visual.blocks.5.attn.qkv.bias": "model-00001-of-00005.safetensors",
1089 "vlm.model.visual.blocks.5.attn.qkv.weight": "model-00001-of-00005.safetensors",
1090 "vlm.model.visual.blocks.5.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1091 "vlm.model.visual.blocks.5.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1092 "vlm.model.visual.blocks.5.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1093 "vlm.model.visual.blocks.5.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1094 "vlm.model.visual.blocks.5.norm1.bias": "model-00001-of-00005.safetensors",
1095 "vlm.model.visual.blocks.5.norm1.weight": "model-00001-of-00005.safetensors",
1096 "vlm.model.visual.blocks.5.norm2.bias": "model-00001-of-00005.safetensors",
1097 "vlm.model.visual.blocks.5.norm2.weight": "model-00001-of-00005.safetensors",
1098 "vlm.model.visual.blocks.6.attn.proj.bias": "model-00001-of-00005.safetensors",
1099 "vlm.model.visual.blocks.6.attn.proj.weight": "model-00001-of-00005.safetensors",
1100 "vlm.model.visual.blocks.6.attn.qkv.bias": "model-00001-of-00005.safetensors",
1101 "vlm.model.visual.blocks.6.attn.qkv.weight": "model-00001-of-00005.safetensors",
1102 "vlm.model.visual.blocks.6.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1103 "vlm.model.visual.blocks.6.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1104 "vlm.model.visual.blocks.6.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1105 "vlm.model.visual.blocks.6.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1106 "vlm.model.visual.blocks.6.norm1.bias": "model-00001-of-00005.safetensors",
1107 "vlm.model.visual.blocks.6.norm1.weight": "model-00001-of-00005.safetensors",
1108 "vlm.model.visual.blocks.6.norm2.bias": "model-00001-of-00005.safetensors",
1109 "vlm.model.visual.blocks.6.norm2.weight": "model-00001-of-00005.safetensors",
1110 "vlm.model.visual.blocks.7.attn.proj.bias": "model-00001-of-00005.safetensors",
1111 "vlm.model.visual.blocks.7.attn.proj.weight": "model-00001-of-00005.safetensors",
1112 "vlm.model.visual.blocks.7.attn.qkv.bias": "model-00001-of-00005.safetensors",
1113 "vlm.model.visual.blocks.7.attn.qkv.weight": "model-00001-of-00005.safetensors",
1114 "vlm.model.visual.blocks.7.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1115 "vlm.model.visual.blocks.7.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1116 "vlm.model.visual.blocks.7.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1117 "vlm.model.visual.blocks.7.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1118 "vlm.model.visual.blocks.7.norm1.bias": "model-00001-of-00005.safetensors",
1119 "vlm.model.visual.blocks.7.norm1.weight": "model-00001-of-00005.safetensors",
1120 "vlm.model.visual.blocks.7.norm2.bias": "model-00001-of-00005.safetensors",
1121 "vlm.model.visual.blocks.7.norm2.weight": "model-00001-of-00005.safetensors",
1122 "vlm.model.visual.blocks.8.attn.proj.bias": "model-00001-of-00005.safetensors",
1123 "vlm.model.visual.blocks.8.attn.proj.weight": "model-00001-of-00005.safetensors",
1124 "vlm.model.visual.blocks.8.attn.qkv.bias": "model-00001-of-00005.safetensors",
1125 "vlm.model.visual.blocks.8.attn.qkv.weight": "model-00001-of-00005.safetensors",
1126 "vlm.model.visual.blocks.8.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1127 "vlm.model.visual.blocks.8.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1128 "vlm.model.visual.blocks.8.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1129 "vlm.model.visual.blocks.8.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1130 "vlm.model.visual.blocks.8.norm1.bias": "model-00001-of-00005.safetensors",
1131 "vlm.model.visual.blocks.8.norm1.weight": "model-00001-of-00005.safetensors",
1132 "vlm.model.visual.blocks.8.norm2.bias": "model-00001-of-00005.safetensors",
1133 "vlm.model.visual.blocks.8.norm2.weight": "model-00001-of-00005.safetensors",
1134 "vlm.model.visual.blocks.9.attn.proj.bias": "model-00001-of-00005.safetensors",
1135 "vlm.model.visual.blocks.9.attn.proj.weight": "model-00001-of-00005.safetensors",
1136 "vlm.model.visual.blocks.9.attn.qkv.bias": "model-00001-of-00005.safetensors",
1137 "vlm.model.visual.blocks.9.attn.qkv.weight": "model-00001-of-00005.safetensors",
1138 "vlm.model.visual.blocks.9.mlp.linear_fc1.bias": "model-00001-of-00005.safetensors",
1139 "vlm.model.visual.blocks.9.mlp.linear_fc1.weight": "model-00001-of-00005.safetensors",
1140 "vlm.model.visual.blocks.9.mlp.linear_fc2.bias": "model-00001-of-00005.safetensors",
1141 "vlm.model.visual.blocks.9.mlp.linear_fc2.weight": "model-00001-of-00005.safetensors",
1142 "vlm.model.visual.blocks.9.norm1.bias": "model-00001-of-00005.safetensors",
1143 "vlm.model.visual.blocks.9.norm1.weight": "model-00001-of-00005.safetensors",
1144 "vlm.model.visual.blocks.9.norm2.bias": "model-00001-of-00005.safetensors",
1145 "vlm.model.visual.blocks.9.norm2.weight": "model-00001-of-00005.safetensors",
1146 "vlm.model.visual.deepstack_merger_list.0.linear_fc1.bias": "model-00001-of-00005.safetensors",
1147 "vlm.model.visual.deepstack_merger_list.0.linear_fc1.weight": "model-00001-of-00005.safetensors",
1148 "vlm.model.visual.deepstack_merger_list.0.linear_fc2.bias": "model-00001-of-00005.safetensors",
1149 "vlm.model.visual.deepstack_merger_list.0.linear_fc2.weight": "model-00001-of-00005.safetensors",
1150 "vlm.model.visual.deepstack_merger_list.0.norm.bias": "model-00001-of-00005.safetensors",
1151 "vlm.model.visual.deepstack_merger_list.0.norm.weight": "model-00001-of-00005.safetensors",
1152 "vlm.model.visual.deepstack_merger_list.1.linear_fc1.bias": "model-00001-of-00005.safetensors",
1153 "vlm.model.visual.deepstack_merger_list.1.linear_fc1.weight": "model-00001-of-00005.safetensors",
1154 "vlm.model.visual.deepstack_merger_list.1.linear_fc2.bias": "model-00001-of-00005.safetensors",
1155 "vlm.model.visual.deepstack_merger_list.1.linear_fc2.weight": "model-00001-of-00005.safetensors",
1156 "vlm.model.visual.deepstack_merger_list.1.norm.bias": "model-00001-of-00005.safetensors",
1157 "vlm.model.visual.deepstack_merger_list.1.norm.weight": "model-00001-of-00005.safetensors",
1158 "vlm.model.visual.deepstack_merger_list.2.linear_fc1.bias": "model-00001-of-00005.safetensors",
1159 "vlm.model.visual.deepstack_merger_list.2.linear_fc1.weight": "model-00001-of-00005.safetensors",
1160 "vlm.model.visual.deepstack_merger_list.2.linear_fc2.bias": "model-00001-of-00005.safetensors",
1161 "vlm.model.visual.deepstack_merger_list.2.linear_fc2.weight": "model-00001-of-00005.safetensors",
1162 "vlm.model.visual.deepstack_merger_list.2.norm.bias": "model-00001-of-00005.safetensors",
1163 "vlm.model.visual.deepstack_merger_list.2.norm.weight": "model-00001-of-00005.safetensors",
1164 "vlm.model.visual.merger.linear_fc1.bias": "model-00001-of-00005.safetensors",
1165 "vlm.model.visual.merger.linear_fc1.weight": "model-00001-of-00005.safetensors",
1166 "vlm.model.visual.merger.linear_fc2.bias": "model-00001-of-00005.safetensors",
1167 "vlm.model.visual.merger.linear_fc2.weight": "model-00001-of-00005.safetensors",
1168 "vlm.model.visual.merger.norm.bias": "model-00001-of-00005.safetensors",
1169 "vlm.model.visual.merger.norm.weight": "model-00001-of-00005.safetensors",
1170 "vlm.model.visual.patch_embed.proj.bias": "model-00001-of-00005.safetensors",
1171 "vlm.model.visual.patch_embed.proj.weight": "model-00001-of-00005.safetensors",
1172 "vlm.model.visual.pos_embed.weight": "model-00001-of-00005.safetensors"
1173 }
1174 }
1175