text_encoder/model.safetensors.index.json
106.1 KB · 1074 lines · json Raw
1 {
2 "metadata": {
3 "total_parameters": 12187325040,
4 "total_size": 48749300160
5 },
6 "weight_map": {
7 "language_model.model.embed_tokens.weight": "model-00002-of-00011.safetensors",
8 "language_model.model.layers.0.input_layernorm.weight": "model-00002-of-00011.safetensors",
9 "language_model.model.layers.0.mlp.down_proj.weight": "model-00002-of-00011.safetensors",
10 "language_model.model.layers.0.mlp.gate_proj.weight": "model-00002-of-00011.safetensors",
11 "language_model.model.layers.0.mlp.up_proj.weight": "model-00002-of-00011.safetensors",
12 "language_model.model.layers.0.post_attention_layernorm.weight": "model-00002-of-00011.safetensors",
13 "language_model.model.layers.0.post_feedforward_layernorm.weight": "model-00002-of-00011.safetensors",
14 "language_model.model.layers.0.pre_feedforward_layernorm.weight": "model-00002-of-00011.safetensors",
15 "language_model.model.layers.0.self_attn.k_norm.weight": "model-00002-of-00011.safetensors",
16 "language_model.model.layers.0.self_attn.k_proj.weight": "model-00002-of-00011.safetensors",
17 "language_model.model.layers.0.self_attn.o_proj.weight": "model-00002-of-00011.safetensors",
18 "language_model.model.layers.0.self_attn.q_norm.weight": "model-00002-of-00011.safetensors",
19 "language_model.model.layers.0.self_attn.q_proj.weight": "model-00002-of-00011.safetensors",
20 "language_model.model.layers.0.self_attn.v_proj.weight": "model-00002-of-00011.safetensors",
21 "language_model.model.layers.1.input_layernorm.weight": "model-00003-of-00011.safetensors",
22 "language_model.model.layers.1.mlp.down_proj.weight": "model-00003-of-00011.safetensors",
23 "language_model.model.layers.1.mlp.gate_proj.weight": "model-00003-of-00011.safetensors",
24 "language_model.model.layers.1.mlp.up_proj.weight": "model-00003-of-00011.safetensors",
25 "language_model.model.layers.1.post_attention_layernorm.weight": "model-00003-of-00011.safetensors",
26 "language_model.model.layers.1.post_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
27 "language_model.model.layers.1.pre_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
28 "language_model.model.layers.1.self_attn.k_norm.weight": "model-00003-of-00011.safetensors",
29 "language_model.model.layers.1.self_attn.k_proj.weight": "model-00003-of-00011.safetensors",
30 "language_model.model.layers.1.self_attn.o_proj.weight": "model-00003-of-00011.safetensors",
31 "language_model.model.layers.1.self_attn.q_norm.weight": "model-00003-of-00011.safetensors",
32 "language_model.model.layers.1.self_attn.q_proj.weight": "model-00002-of-00011.safetensors",
33 "language_model.model.layers.1.self_attn.v_proj.weight": "model-00003-of-00011.safetensors",
34 "language_model.model.layers.10.input_layernorm.weight": "model-00004-of-00011.safetensors",
35 "language_model.model.layers.10.mlp.down_proj.weight": "model-00004-of-00011.safetensors",
36 "language_model.model.layers.10.mlp.gate_proj.weight": "model-00004-of-00011.safetensors",
37 "language_model.model.layers.10.mlp.up_proj.weight": "model-00004-of-00011.safetensors",
38 "language_model.model.layers.10.post_attention_layernorm.weight": "model-00004-of-00011.safetensors",
39 "language_model.model.layers.10.post_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
40 "language_model.model.layers.10.pre_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
41 "language_model.model.layers.10.self_attn.k_norm.weight": "model-00004-of-00011.safetensors",
42 "language_model.model.layers.10.self_attn.k_proj.weight": "model-00004-of-00011.safetensors",
43 "language_model.model.layers.10.self_attn.o_proj.weight": "model-00004-of-00011.safetensors",
44 "language_model.model.layers.10.self_attn.q_norm.weight": "model-00004-of-00011.safetensors",
45 "language_model.model.layers.10.self_attn.q_proj.weight": "model-00004-of-00011.safetensors",
46 "language_model.model.layers.10.self_attn.v_proj.weight": "model-00004-of-00011.safetensors",
47 "language_model.model.layers.11.input_layernorm.weight": "model-00004-of-00011.safetensors",
48 "language_model.model.layers.11.mlp.down_proj.weight": "model-00004-of-00011.safetensors",
49 "language_model.model.layers.11.mlp.gate_proj.weight": "model-00004-of-00011.safetensors",
50 "language_model.model.layers.11.mlp.up_proj.weight": "model-00004-of-00011.safetensors",
51 "language_model.model.layers.11.post_attention_layernorm.weight": "model-00004-of-00011.safetensors",
52 "language_model.model.layers.11.post_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
53 "language_model.model.layers.11.pre_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
54 "language_model.model.layers.11.self_attn.k_norm.weight": "model-00004-of-00011.safetensors",
55 "language_model.model.layers.11.self_attn.k_proj.weight": "model-00004-of-00011.safetensors",
56 "language_model.model.layers.11.self_attn.o_proj.weight": "model-00004-of-00011.safetensors",
57 "language_model.model.layers.11.self_attn.q_norm.weight": "model-00004-of-00011.safetensors",
58 "language_model.model.layers.11.self_attn.q_proj.weight": "model-00004-of-00011.safetensors",
59 "language_model.model.layers.11.self_attn.v_proj.weight": "model-00004-of-00011.safetensors",
60 "language_model.model.layers.12.input_layernorm.weight": "model-00005-of-00011.safetensors",
61 "language_model.model.layers.12.mlp.down_proj.weight": "model-00005-of-00011.safetensors",
62 "language_model.model.layers.12.mlp.gate_proj.weight": "model-00005-of-00011.safetensors",
63 "language_model.model.layers.12.mlp.up_proj.weight": "model-00005-of-00011.safetensors",
64 "language_model.model.layers.12.post_attention_layernorm.weight": "model-00005-of-00011.safetensors",
65 "language_model.model.layers.12.post_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
66 "language_model.model.layers.12.pre_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
67 "language_model.model.layers.12.self_attn.k_norm.weight": "model-00005-of-00011.safetensors",
68 "language_model.model.layers.12.self_attn.k_proj.weight": "model-00005-of-00011.safetensors",
69 "language_model.model.layers.12.self_attn.o_proj.weight": "model-00005-of-00011.safetensors",
70 "language_model.model.layers.12.self_attn.q_norm.weight": "model-00005-of-00011.safetensors",
71 "language_model.model.layers.12.self_attn.q_proj.weight": "model-00005-of-00011.safetensors",
72 "language_model.model.layers.12.self_attn.v_proj.weight": "model-00005-of-00011.safetensors",
73 "language_model.model.layers.13.input_layernorm.weight": "model-00005-of-00011.safetensors",
74 "language_model.model.layers.13.mlp.down_proj.weight": "model-00005-of-00011.safetensors",
75 "language_model.model.layers.13.mlp.gate_proj.weight": "model-00005-of-00011.safetensors",
76 "language_model.model.layers.13.mlp.up_proj.weight": "model-00005-of-00011.safetensors",
77 "language_model.model.layers.13.post_attention_layernorm.weight": "model-00005-of-00011.safetensors",
78 "language_model.model.layers.13.post_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
79 "language_model.model.layers.13.pre_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
80 "language_model.model.layers.13.self_attn.k_norm.weight": "model-00005-of-00011.safetensors",
81 "language_model.model.layers.13.self_attn.k_proj.weight": "model-00005-of-00011.safetensors",
82 "language_model.model.layers.13.self_attn.o_proj.weight": "model-00005-of-00011.safetensors",
83 "language_model.model.layers.13.self_attn.q_norm.weight": "model-00005-of-00011.safetensors",
84 "language_model.model.layers.13.self_attn.q_proj.weight": "model-00005-of-00011.safetensors",
85 "language_model.model.layers.13.self_attn.v_proj.weight": "model-00005-of-00011.safetensors",
86 "language_model.model.layers.14.input_layernorm.weight": "model-00005-of-00011.safetensors",
87 "language_model.model.layers.14.mlp.down_proj.weight": "model-00005-of-00011.safetensors",
88 "language_model.model.layers.14.mlp.gate_proj.weight": "model-00005-of-00011.safetensors",
89 "language_model.model.layers.14.mlp.up_proj.weight": "model-00005-of-00011.safetensors",
90 "language_model.model.layers.14.post_attention_layernorm.weight": "model-00005-of-00011.safetensors",
91 "language_model.model.layers.14.post_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
92 "language_model.model.layers.14.pre_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
93 "language_model.model.layers.14.self_attn.k_norm.weight": "model-00005-of-00011.safetensors",
94 "language_model.model.layers.14.self_attn.k_proj.weight": "model-00005-of-00011.safetensors",
95 "language_model.model.layers.14.self_attn.o_proj.weight": "model-00005-of-00011.safetensors",
96 "language_model.model.layers.14.self_attn.q_norm.weight": "model-00005-of-00011.safetensors",
97 "language_model.model.layers.14.self_attn.q_proj.weight": "model-00005-of-00011.safetensors",
98 "language_model.model.layers.14.self_attn.v_proj.weight": "model-00005-of-00011.safetensors",
99 "language_model.model.layers.15.input_layernorm.weight": "model-00005-of-00011.safetensors",
100 "language_model.model.layers.15.mlp.down_proj.weight": "model-00005-of-00011.safetensors",
101 "language_model.model.layers.15.mlp.gate_proj.weight": "model-00005-of-00011.safetensors",
102 "language_model.model.layers.15.mlp.up_proj.weight": "model-00005-of-00011.safetensors",
103 "language_model.model.layers.15.post_attention_layernorm.weight": "model-00005-of-00011.safetensors",
104 "language_model.model.layers.15.post_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
105 "language_model.model.layers.15.pre_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
106 "language_model.model.layers.15.self_attn.k_norm.weight": "model-00005-of-00011.safetensors",
107 "language_model.model.layers.15.self_attn.k_proj.weight": "model-00005-of-00011.safetensors",
108 "language_model.model.layers.15.self_attn.o_proj.weight": "model-00005-of-00011.safetensors",
109 "language_model.model.layers.15.self_attn.q_norm.weight": "model-00005-of-00011.safetensors",
110 "language_model.model.layers.15.self_attn.q_proj.weight": "model-00005-of-00011.safetensors",
111 "language_model.model.layers.15.self_attn.v_proj.weight": "model-00005-of-00011.safetensors",
112 "language_model.model.layers.16.input_layernorm.weight": "model-00005-of-00011.safetensors",
113 "language_model.model.layers.16.mlp.down_proj.weight": "model-00005-of-00011.safetensors",
114 "language_model.model.layers.16.mlp.gate_proj.weight": "model-00005-of-00011.safetensors",
115 "language_model.model.layers.16.mlp.up_proj.weight": "model-00005-of-00011.safetensors",
116 "language_model.model.layers.16.post_attention_layernorm.weight": "model-00005-of-00011.safetensors",
117 "language_model.model.layers.16.post_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
118 "language_model.model.layers.16.pre_feedforward_layernorm.weight": "model-00005-of-00011.safetensors",
119 "language_model.model.layers.16.self_attn.k_norm.weight": "model-00005-of-00011.safetensors",
120 "language_model.model.layers.16.self_attn.k_proj.weight": "model-00005-of-00011.safetensors",
121 "language_model.model.layers.16.self_attn.o_proj.weight": "model-00005-of-00011.safetensors",
122 "language_model.model.layers.16.self_attn.q_norm.weight": "model-00005-of-00011.safetensors",
123 "language_model.model.layers.16.self_attn.q_proj.weight": "model-00005-of-00011.safetensors",
124 "language_model.model.layers.16.self_attn.v_proj.weight": "model-00005-of-00011.safetensors",
125 "language_model.model.layers.17.input_layernorm.weight": "model-00006-of-00011.safetensors",
126 "language_model.model.layers.17.mlp.down_proj.weight": "model-00006-of-00011.safetensors",
127 "language_model.model.layers.17.mlp.gate_proj.weight": "model-00005-of-00011.safetensors",
128 "language_model.model.layers.17.mlp.up_proj.weight": "model-00006-of-00011.safetensors",
129 "language_model.model.layers.17.post_attention_layernorm.weight": "model-00006-of-00011.safetensors",
130 "language_model.model.layers.17.post_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
131 "language_model.model.layers.17.pre_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
132 "language_model.model.layers.17.self_attn.k_norm.weight": "model-00005-of-00011.safetensors",
133 "language_model.model.layers.17.self_attn.k_proj.weight": "model-00005-of-00011.safetensors",
134 "language_model.model.layers.17.self_attn.o_proj.weight": "model-00005-of-00011.safetensors",
135 "language_model.model.layers.17.self_attn.q_norm.weight": "model-00005-of-00011.safetensors",
136 "language_model.model.layers.17.self_attn.q_proj.weight": "model-00005-of-00011.safetensors",
137 "language_model.model.layers.17.self_attn.v_proj.weight": "model-00005-of-00011.safetensors",
138 "language_model.model.layers.18.input_layernorm.weight": "model-00006-of-00011.safetensors",
139 "language_model.model.layers.18.mlp.down_proj.weight": "model-00006-of-00011.safetensors",
140 "language_model.model.layers.18.mlp.gate_proj.weight": "model-00006-of-00011.safetensors",
141 "language_model.model.layers.18.mlp.up_proj.weight": "model-00006-of-00011.safetensors",
142 "language_model.model.layers.18.post_attention_layernorm.weight": "model-00006-of-00011.safetensors",
143 "language_model.model.layers.18.post_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
144 "language_model.model.layers.18.pre_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
145 "language_model.model.layers.18.self_attn.k_norm.weight": "model-00006-of-00011.safetensors",
146 "language_model.model.layers.18.self_attn.k_proj.weight": "model-00006-of-00011.safetensors",
147 "language_model.model.layers.18.self_attn.o_proj.weight": "model-00006-of-00011.safetensors",
148 "language_model.model.layers.18.self_attn.q_norm.weight": "model-00006-of-00011.safetensors",
149 "language_model.model.layers.18.self_attn.q_proj.weight": "model-00006-of-00011.safetensors",
150 "language_model.model.layers.18.self_attn.v_proj.weight": "model-00006-of-00011.safetensors",
151 "language_model.model.layers.19.input_layernorm.weight": "model-00006-of-00011.safetensors",
152 "language_model.model.layers.19.mlp.down_proj.weight": "model-00006-of-00011.safetensors",
153 "language_model.model.layers.19.mlp.gate_proj.weight": "model-00006-of-00011.safetensors",
154 "language_model.model.layers.19.mlp.up_proj.weight": "model-00006-of-00011.safetensors",
155 "language_model.model.layers.19.post_attention_layernorm.weight": "model-00006-of-00011.safetensors",
156 "language_model.model.layers.19.post_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
157 "language_model.model.layers.19.pre_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
158 "language_model.model.layers.19.self_attn.k_norm.weight": "model-00006-of-00011.safetensors",
159 "language_model.model.layers.19.self_attn.k_proj.weight": "model-00006-of-00011.safetensors",
160 "language_model.model.layers.19.self_attn.o_proj.weight": "model-00006-of-00011.safetensors",
161 "language_model.model.layers.19.self_attn.q_norm.weight": "model-00006-of-00011.safetensors",
162 "language_model.model.layers.19.self_attn.q_proj.weight": "model-00006-of-00011.safetensors",
163 "language_model.model.layers.19.self_attn.v_proj.weight": "model-00006-of-00011.safetensors",
164 "language_model.model.layers.2.input_layernorm.weight": "model-00003-of-00011.safetensors",
165 "language_model.model.layers.2.mlp.down_proj.weight": "model-00003-of-00011.safetensors",
166 "language_model.model.layers.2.mlp.gate_proj.weight": "model-00003-of-00011.safetensors",
167 "language_model.model.layers.2.mlp.up_proj.weight": "model-00003-of-00011.safetensors",
168 "language_model.model.layers.2.post_attention_layernorm.weight": "model-00003-of-00011.safetensors",
169 "language_model.model.layers.2.post_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
170 "language_model.model.layers.2.pre_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
171 "language_model.model.layers.2.self_attn.k_norm.weight": "model-00003-of-00011.safetensors",
172 "language_model.model.layers.2.self_attn.k_proj.weight": "model-00003-of-00011.safetensors",
173 "language_model.model.layers.2.self_attn.o_proj.weight": "model-00003-of-00011.safetensors",
174 "language_model.model.layers.2.self_attn.q_norm.weight": "model-00003-of-00011.safetensors",
175 "language_model.model.layers.2.self_attn.q_proj.weight": "model-00003-of-00011.safetensors",
176 "language_model.model.layers.2.self_attn.v_proj.weight": "model-00003-of-00011.safetensors",
177 "language_model.model.layers.20.input_layernorm.weight": "model-00006-of-00011.safetensors",
178 "language_model.model.layers.20.mlp.down_proj.weight": "model-00006-of-00011.safetensors",
179 "language_model.model.layers.20.mlp.gate_proj.weight": "model-00006-of-00011.safetensors",
180 "language_model.model.layers.20.mlp.up_proj.weight": "model-00006-of-00011.safetensors",
181 "language_model.model.layers.20.post_attention_layernorm.weight": "model-00006-of-00011.safetensors",
182 "language_model.model.layers.20.post_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
183 "language_model.model.layers.20.pre_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
184 "language_model.model.layers.20.self_attn.k_norm.weight": "model-00006-of-00011.safetensors",
185 "language_model.model.layers.20.self_attn.k_proj.weight": "model-00006-of-00011.safetensors",
186 "language_model.model.layers.20.self_attn.o_proj.weight": "model-00006-of-00011.safetensors",
187 "language_model.model.layers.20.self_attn.q_norm.weight": "model-00006-of-00011.safetensors",
188 "language_model.model.layers.20.self_attn.q_proj.weight": "model-00006-of-00011.safetensors",
189 "language_model.model.layers.20.self_attn.v_proj.weight": "model-00006-of-00011.safetensors",
190 "language_model.model.layers.21.input_layernorm.weight": "model-00006-of-00011.safetensors",
191 "language_model.model.layers.21.mlp.down_proj.weight": "model-00006-of-00011.safetensors",
192 "language_model.model.layers.21.mlp.gate_proj.weight": "model-00006-of-00011.safetensors",
193 "language_model.model.layers.21.mlp.up_proj.weight": "model-00006-of-00011.safetensors",
194 "language_model.model.layers.21.post_attention_layernorm.weight": "model-00006-of-00011.safetensors",
195 "language_model.model.layers.21.post_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
196 "language_model.model.layers.21.pre_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
197 "language_model.model.layers.21.self_attn.k_norm.weight": "model-00006-of-00011.safetensors",
198 "language_model.model.layers.21.self_attn.k_proj.weight": "model-00006-of-00011.safetensors",
199 "language_model.model.layers.21.self_attn.o_proj.weight": "model-00006-of-00011.safetensors",
200 "language_model.model.layers.21.self_attn.q_norm.weight": "model-00006-of-00011.safetensors",
201 "language_model.model.layers.21.self_attn.q_proj.weight": "model-00006-of-00011.safetensors",
202 "language_model.model.layers.21.self_attn.v_proj.weight": "model-00006-of-00011.safetensors",
203 "language_model.model.layers.22.input_layernorm.weight": "model-00006-of-00011.safetensors",
204 "language_model.model.layers.22.mlp.down_proj.weight": "model-00006-of-00011.safetensors",
205 "language_model.model.layers.22.mlp.gate_proj.weight": "model-00006-of-00011.safetensors",
206 "language_model.model.layers.22.mlp.up_proj.weight": "model-00006-of-00011.safetensors",
207 "language_model.model.layers.22.post_attention_layernorm.weight": "model-00006-of-00011.safetensors",
208 "language_model.model.layers.22.post_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
209 "language_model.model.layers.22.pre_feedforward_layernorm.weight": "model-00006-of-00011.safetensors",
210 "language_model.model.layers.22.self_attn.k_norm.weight": "model-00006-of-00011.safetensors",
211 "language_model.model.layers.22.self_attn.k_proj.weight": "model-00006-of-00011.safetensors",
212 "language_model.model.layers.22.self_attn.o_proj.weight": "model-00006-of-00011.safetensors",
213 "language_model.model.layers.22.self_attn.q_norm.weight": "model-00006-of-00011.safetensors",
214 "language_model.model.layers.22.self_attn.q_proj.weight": "model-00006-of-00011.safetensors",
215 "language_model.model.layers.22.self_attn.v_proj.weight": "model-00006-of-00011.safetensors",
216 "language_model.model.layers.23.input_layernorm.weight": "model-00007-of-00011.safetensors",
217 "language_model.model.layers.23.mlp.down_proj.weight": "model-00007-of-00011.safetensors",
218 "language_model.model.layers.23.mlp.gate_proj.weight": "model-00007-of-00011.safetensors",
219 "language_model.model.layers.23.mlp.up_proj.weight": "model-00007-of-00011.safetensors",
220 "language_model.model.layers.23.post_attention_layernorm.weight": "model-00007-of-00011.safetensors",
221 "language_model.model.layers.23.post_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
222 "language_model.model.layers.23.pre_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
223 "language_model.model.layers.23.self_attn.k_norm.weight": "model-00007-of-00011.safetensors",
224 "language_model.model.layers.23.self_attn.k_proj.weight": "model-00007-of-00011.safetensors",
225 "language_model.model.layers.23.self_attn.o_proj.weight": "model-00007-of-00011.safetensors",
226 "language_model.model.layers.23.self_attn.q_norm.weight": "model-00007-of-00011.safetensors",
227 "language_model.model.layers.23.self_attn.q_proj.weight": "model-00007-of-00011.safetensors",
228 "language_model.model.layers.23.self_attn.v_proj.weight": "model-00007-of-00011.safetensors",
229 "language_model.model.layers.24.input_layernorm.weight": "model-00007-of-00011.safetensors",
230 "language_model.model.layers.24.mlp.down_proj.weight": "model-00007-of-00011.safetensors",
231 "language_model.model.layers.24.mlp.gate_proj.weight": "model-00007-of-00011.safetensors",
232 "language_model.model.layers.24.mlp.up_proj.weight": "model-00007-of-00011.safetensors",
233 "language_model.model.layers.24.post_attention_layernorm.weight": "model-00007-of-00011.safetensors",
234 "language_model.model.layers.24.post_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
235 "language_model.model.layers.24.pre_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
236 "language_model.model.layers.24.self_attn.k_norm.weight": "model-00007-of-00011.safetensors",
237 "language_model.model.layers.24.self_attn.k_proj.weight": "model-00007-of-00011.safetensors",
238 "language_model.model.layers.24.self_attn.o_proj.weight": "model-00007-of-00011.safetensors",
239 "language_model.model.layers.24.self_attn.q_norm.weight": "model-00007-of-00011.safetensors",
240 "language_model.model.layers.24.self_attn.q_proj.weight": "model-00007-of-00011.safetensors",
241 "language_model.model.layers.24.self_attn.v_proj.weight": "model-00007-of-00011.safetensors",
242 "language_model.model.layers.25.input_layernorm.weight": "model-00007-of-00011.safetensors",
243 "language_model.model.layers.25.mlp.down_proj.weight": "model-00007-of-00011.safetensors",
244 "language_model.model.layers.25.mlp.gate_proj.weight": "model-00007-of-00011.safetensors",
245 "language_model.model.layers.25.mlp.up_proj.weight": "model-00007-of-00011.safetensors",
246 "language_model.model.layers.25.post_attention_layernorm.weight": "model-00007-of-00011.safetensors",
247 "language_model.model.layers.25.post_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
248 "language_model.model.layers.25.pre_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
249 "language_model.model.layers.25.self_attn.k_norm.weight": "model-00007-of-00011.safetensors",
250 "language_model.model.layers.25.self_attn.k_proj.weight": "model-00007-of-00011.safetensors",
251 "language_model.model.layers.25.self_attn.o_proj.weight": "model-00007-of-00011.safetensors",
252 "language_model.model.layers.25.self_attn.q_norm.weight": "model-00007-of-00011.safetensors",
253 "language_model.model.layers.25.self_attn.q_proj.weight": "model-00007-of-00011.safetensors",
254 "language_model.model.layers.25.self_attn.v_proj.weight": "model-00007-of-00011.safetensors",
255 "language_model.model.layers.26.input_layernorm.weight": "model-00007-of-00011.safetensors",
256 "language_model.model.layers.26.mlp.down_proj.weight": "model-00007-of-00011.safetensors",
257 "language_model.model.layers.26.mlp.gate_proj.weight": "model-00007-of-00011.safetensors",
258 "language_model.model.layers.26.mlp.up_proj.weight": "model-00007-of-00011.safetensors",
259 "language_model.model.layers.26.post_attention_layernorm.weight": "model-00007-of-00011.safetensors",
260 "language_model.model.layers.26.post_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
261 "language_model.model.layers.26.pre_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
262 "language_model.model.layers.26.self_attn.k_norm.weight": "model-00007-of-00011.safetensors",
263 "language_model.model.layers.26.self_attn.k_proj.weight": "model-00007-of-00011.safetensors",
264 "language_model.model.layers.26.self_attn.o_proj.weight": "model-00007-of-00011.safetensors",
265 "language_model.model.layers.26.self_attn.q_norm.weight": "model-00007-of-00011.safetensors",
266 "language_model.model.layers.26.self_attn.q_proj.weight": "model-00007-of-00011.safetensors",
267 "language_model.model.layers.26.self_attn.v_proj.weight": "model-00007-of-00011.safetensors",
268 "language_model.model.layers.27.input_layernorm.weight": "model-00007-of-00011.safetensors",
269 "language_model.model.layers.27.mlp.down_proj.weight": "model-00007-of-00011.safetensors",
270 "language_model.model.layers.27.mlp.gate_proj.weight": "model-00007-of-00011.safetensors",
271 "language_model.model.layers.27.mlp.up_proj.weight": "model-00007-of-00011.safetensors",
272 "language_model.model.layers.27.post_attention_layernorm.weight": "model-00007-of-00011.safetensors",
273 "language_model.model.layers.27.post_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
274 "language_model.model.layers.27.pre_feedforward_layernorm.weight": "model-00007-of-00011.safetensors",
275 "language_model.model.layers.27.self_attn.k_norm.weight": "model-00007-of-00011.safetensors",
276 "language_model.model.layers.27.self_attn.k_proj.weight": "model-00007-of-00011.safetensors",
277 "language_model.model.layers.27.self_attn.o_proj.weight": "model-00007-of-00011.safetensors",
278 "language_model.model.layers.27.self_attn.q_norm.weight": "model-00007-of-00011.safetensors",
279 "language_model.model.layers.27.self_attn.q_proj.weight": "model-00007-of-00011.safetensors",
280 "language_model.model.layers.27.self_attn.v_proj.weight": "model-00007-of-00011.safetensors",
281 "language_model.model.layers.28.input_layernorm.weight": "model-00008-of-00011.safetensors",
282 "language_model.model.layers.28.mlp.down_proj.weight": "model-00008-of-00011.safetensors",
283 "language_model.model.layers.28.mlp.gate_proj.weight": "model-00007-of-00011.safetensors",
284 "language_model.model.layers.28.mlp.up_proj.weight": "model-00008-of-00011.safetensors",
285 "language_model.model.layers.28.post_attention_layernorm.weight": "model-00008-of-00011.safetensors",
286 "language_model.model.layers.28.post_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
287 "language_model.model.layers.28.pre_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
288 "language_model.model.layers.28.self_attn.k_norm.weight": "model-00007-of-00011.safetensors",
289 "language_model.model.layers.28.self_attn.k_proj.weight": "model-00007-of-00011.safetensors",
290 "language_model.model.layers.28.self_attn.o_proj.weight": "model-00007-of-00011.safetensors",
291 "language_model.model.layers.28.self_attn.q_norm.weight": "model-00007-of-00011.safetensors",
292 "language_model.model.layers.28.self_attn.q_proj.weight": "model-00007-of-00011.safetensors",
293 "language_model.model.layers.28.self_attn.v_proj.weight": "model-00007-of-00011.safetensors",
294 "language_model.model.layers.29.input_layernorm.weight": "model-00008-of-00011.safetensors",
295 "language_model.model.layers.29.mlp.down_proj.weight": "model-00008-of-00011.safetensors",
296 "language_model.model.layers.29.mlp.gate_proj.weight": "model-00008-of-00011.safetensors",
297 "language_model.model.layers.29.mlp.up_proj.weight": "model-00008-of-00011.safetensors",
298 "language_model.model.layers.29.post_attention_layernorm.weight": "model-00008-of-00011.safetensors",
299 "language_model.model.layers.29.post_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
300 "language_model.model.layers.29.pre_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
301 "language_model.model.layers.29.self_attn.k_norm.weight": "model-00008-of-00011.safetensors",
302 "language_model.model.layers.29.self_attn.k_proj.weight": "model-00008-of-00011.safetensors",
303 "language_model.model.layers.29.self_attn.o_proj.weight": "model-00008-of-00011.safetensors",
304 "language_model.model.layers.29.self_attn.q_norm.weight": "model-00008-of-00011.safetensors",
305 "language_model.model.layers.29.self_attn.q_proj.weight": "model-00008-of-00011.safetensors",
306 "language_model.model.layers.29.self_attn.v_proj.weight": "model-00008-of-00011.safetensors",
307 "language_model.model.layers.3.input_layernorm.weight": "model-00003-of-00011.safetensors",
308 "language_model.model.layers.3.mlp.down_proj.weight": "model-00003-of-00011.safetensors",
309 "language_model.model.layers.3.mlp.gate_proj.weight": "model-00003-of-00011.safetensors",
310 "language_model.model.layers.3.mlp.up_proj.weight": "model-00003-of-00011.safetensors",
311 "language_model.model.layers.3.post_attention_layernorm.weight": "model-00003-of-00011.safetensors",
312 "language_model.model.layers.3.post_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
313 "language_model.model.layers.3.pre_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
314 "language_model.model.layers.3.self_attn.k_norm.weight": "model-00003-of-00011.safetensors",
315 "language_model.model.layers.3.self_attn.k_proj.weight": "model-00003-of-00011.safetensors",
316 "language_model.model.layers.3.self_attn.o_proj.weight": "model-00003-of-00011.safetensors",
317 "language_model.model.layers.3.self_attn.q_norm.weight": "model-00003-of-00011.safetensors",
318 "language_model.model.layers.3.self_attn.q_proj.weight": "model-00003-of-00011.safetensors",
319 "language_model.model.layers.3.self_attn.v_proj.weight": "model-00003-of-00011.safetensors",
320 "language_model.model.layers.30.input_layernorm.weight": "model-00008-of-00011.safetensors",
321 "language_model.model.layers.30.mlp.down_proj.weight": "model-00008-of-00011.safetensors",
322 "language_model.model.layers.30.mlp.gate_proj.weight": "model-00008-of-00011.safetensors",
323 "language_model.model.layers.30.mlp.up_proj.weight": "model-00008-of-00011.safetensors",
324 "language_model.model.layers.30.post_attention_layernorm.weight": "model-00008-of-00011.safetensors",
325 "language_model.model.layers.30.post_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
326 "language_model.model.layers.30.pre_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
327 "language_model.model.layers.30.self_attn.k_norm.weight": "model-00008-of-00011.safetensors",
328 "language_model.model.layers.30.self_attn.k_proj.weight": "model-00008-of-00011.safetensors",
329 "language_model.model.layers.30.self_attn.o_proj.weight": "model-00008-of-00011.safetensors",
330 "language_model.model.layers.30.self_attn.q_norm.weight": "model-00008-of-00011.safetensors",
331 "language_model.model.layers.30.self_attn.q_proj.weight": "model-00008-of-00011.safetensors",
332 "language_model.model.layers.30.self_attn.v_proj.weight": "model-00008-of-00011.safetensors",
333 "language_model.model.layers.31.input_layernorm.weight": "model-00008-of-00011.safetensors",
334 "language_model.model.layers.31.mlp.down_proj.weight": "model-00008-of-00011.safetensors",
335 "language_model.model.layers.31.mlp.gate_proj.weight": "model-00008-of-00011.safetensors",
336 "language_model.model.layers.31.mlp.up_proj.weight": "model-00008-of-00011.safetensors",
337 "language_model.model.layers.31.post_attention_layernorm.weight": "model-00008-of-00011.safetensors",
338 "language_model.model.layers.31.post_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
339 "language_model.model.layers.31.pre_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
340 "language_model.model.layers.31.self_attn.k_norm.weight": "model-00008-of-00011.safetensors",
341 "language_model.model.layers.31.self_attn.k_proj.weight": "model-00008-of-00011.safetensors",
342 "language_model.model.layers.31.self_attn.o_proj.weight": "model-00008-of-00011.safetensors",
343 "language_model.model.layers.31.self_attn.q_norm.weight": "model-00008-of-00011.safetensors",
344 "language_model.model.layers.31.self_attn.q_proj.weight": "model-00008-of-00011.safetensors",
345 "language_model.model.layers.31.self_attn.v_proj.weight": "model-00008-of-00011.safetensors",
346 "language_model.model.layers.32.input_layernorm.weight": "model-00008-of-00011.safetensors",
347 "language_model.model.layers.32.mlp.down_proj.weight": "model-00008-of-00011.safetensors",
348 "language_model.model.layers.32.mlp.gate_proj.weight": "model-00008-of-00011.safetensors",
349 "language_model.model.layers.32.mlp.up_proj.weight": "model-00008-of-00011.safetensors",
350 "language_model.model.layers.32.post_attention_layernorm.weight": "model-00008-of-00011.safetensors",
351 "language_model.model.layers.32.post_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
352 "language_model.model.layers.32.pre_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
353 "language_model.model.layers.32.self_attn.k_norm.weight": "model-00008-of-00011.safetensors",
354 "language_model.model.layers.32.self_attn.k_proj.weight": "model-00008-of-00011.safetensors",
355 "language_model.model.layers.32.self_attn.o_proj.weight": "model-00008-of-00011.safetensors",
356 "language_model.model.layers.32.self_attn.q_norm.weight": "model-00008-of-00011.safetensors",
357 "language_model.model.layers.32.self_attn.q_proj.weight": "model-00008-of-00011.safetensors",
358 "language_model.model.layers.32.self_attn.v_proj.weight": "model-00008-of-00011.safetensors",
359 "language_model.model.layers.33.input_layernorm.weight": "model-00008-of-00011.safetensors",
360 "language_model.model.layers.33.mlp.down_proj.weight": "model-00008-of-00011.safetensors",
361 "language_model.model.layers.33.mlp.gate_proj.weight": "model-00008-of-00011.safetensors",
362 "language_model.model.layers.33.mlp.up_proj.weight": "model-00008-of-00011.safetensors",
363 "language_model.model.layers.33.post_attention_layernorm.weight": "model-00008-of-00011.safetensors",
364 "language_model.model.layers.33.post_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
365 "language_model.model.layers.33.pre_feedforward_layernorm.weight": "model-00008-of-00011.safetensors",
366 "language_model.model.layers.33.self_attn.k_norm.weight": "model-00008-of-00011.safetensors",
367 "language_model.model.layers.33.self_attn.k_proj.weight": "model-00008-of-00011.safetensors",
368 "language_model.model.layers.33.self_attn.o_proj.weight": "model-00008-of-00011.safetensors",
369 "language_model.model.layers.33.self_attn.q_norm.weight": "model-00008-of-00011.safetensors",
370 "language_model.model.layers.33.self_attn.q_proj.weight": "model-00008-of-00011.safetensors",
371 "language_model.model.layers.33.self_attn.v_proj.weight": "model-00008-of-00011.safetensors",
372 "language_model.model.layers.34.input_layernorm.weight": "model-00009-of-00011.safetensors",
373 "language_model.model.layers.34.mlp.down_proj.weight": "model-00009-of-00011.safetensors",
374 "language_model.model.layers.34.mlp.gate_proj.weight": "model-00009-of-00011.safetensors",
375 "language_model.model.layers.34.mlp.up_proj.weight": "model-00009-of-00011.safetensors",
376 "language_model.model.layers.34.post_attention_layernorm.weight": "model-00009-of-00011.safetensors",
377 "language_model.model.layers.34.post_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
378 "language_model.model.layers.34.pre_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
379 "language_model.model.layers.34.self_attn.k_norm.weight": "model-00009-of-00011.safetensors",
380 "language_model.model.layers.34.self_attn.k_proj.weight": "model-00009-of-00011.safetensors",
381 "language_model.model.layers.34.self_attn.o_proj.weight": "model-00009-of-00011.safetensors",
382 "language_model.model.layers.34.self_attn.q_norm.weight": "model-00009-of-00011.safetensors",
383 "language_model.model.layers.34.self_attn.q_proj.weight": "model-00009-of-00011.safetensors",
384 "language_model.model.layers.34.self_attn.v_proj.weight": "model-00009-of-00011.safetensors",
385 "language_model.model.layers.35.input_layernorm.weight": "model-00009-of-00011.safetensors",
386 "language_model.model.layers.35.mlp.down_proj.weight": "model-00009-of-00011.safetensors",
387 "language_model.model.layers.35.mlp.gate_proj.weight": "model-00009-of-00011.safetensors",
388 "language_model.model.layers.35.mlp.up_proj.weight": "model-00009-of-00011.safetensors",
389 "language_model.model.layers.35.post_attention_layernorm.weight": "model-00009-of-00011.safetensors",
390 "language_model.model.layers.35.post_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
391 "language_model.model.layers.35.pre_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
392 "language_model.model.layers.35.self_attn.k_norm.weight": "model-00009-of-00011.safetensors",
393 "language_model.model.layers.35.self_attn.k_proj.weight": "model-00009-of-00011.safetensors",
394 "language_model.model.layers.35.self_attn.o_proj.weight": "model-00009-of-00011.safetensors",
395 "language_model.model.layers.35.self_attn.q_norm.weight": "model-00009-of-00011.safetensors",
396 "language_model.model.layers.35.self_attn.q_proj.weight": "model-00009-of-00011.safetensors",
397 "language_model.model.layers.35.self_attn.v_proj.weight": "model-00009-of-00011.safetensors",
398 "language_model.model.layers.36.input_layernorm.weight": "model-00009-of-00011.safetensors",
399 "language_model.model.layers.36.mlp.down_proj.weight": "model-00009-of-00011.safetensors",
400 "language_model.model.layers.36.mlp.gate_proj.weight": "model-00009-of-00011.safetensors",
401 "language_model.model.layers.36.mlp.up_proj.weight": "model-00009-of-00011.safetensors",
402 "language_model.model.layers.36.post_attention_layernorm.weight": "model-00009-of-00011.safetensors",
403 "language_model.model.layers.36.post_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
404 "language_model.model.layers.36.pre_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
405 "language_model.model.layers.36.self_attn.k_norm.weight": "model-00009-of-00011.safetensors",
406 "language_model.model.layers.36.self_attn.k_proj.weight": "model-00009-of-00011.safetensors",
407 "language_model.model.layers.36.self_attn.o_proj.weight": "model-00009-of-00011.safetensors",
408 "language_model.model.layers.36.self_attn.q_norm.weight": "model-00009-of-00011.safetensors",
409 "language_model.model.layers.36.self_attn.q_proj.weight": "model-00009-of-00011.safetensors",
410 "language_model.model.layers.36.self_attn.v_proj.weight": "model-00009-of-00011.safetensors",
411 "language_model.model.layers.37.input_layernorm.weight": "model-00009-of-00011.safetensors",
412 "language_model.model.layers.37.mlp.down_proj.weight": "model-00009-of-00011.safetensors",
413 "language_model.model.layers.37.mlp.gate_proj.weight": "model-00009-of-00011.safetensors",
414 "language_model.model.layers.37.mlp.up_proj.weight": "model-00009-of-00011.safetensors",
415 "language_model.model.layers.37.post_attention_layernorm.weight": "model-00009-of-00011.safetensors",
416 "language_model.model.layers.37.post_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
417 "language_model.model.layers.37.pre_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
418 "language_model.model.layers.37.self_attn.k_norm.weight": "model-00009-of-00011.safetensors",
419 "language_model.model.layers.37.self_attn.k_proj.weight": "model-00009-of-00011.safetensors",
420 "language_model.model.layers.37.self_attn.o_proj.weight": "model-00009-of-00011.safetensors",
421 "language_model.model.layers.37.self_attn.q_norm.weight": "model-00009-of-00011.safetensors",
422 "language_model.model.layers.37.self_attn.q_proj.weight": "model-00009-of-00011.safetensors",
423 "language_model.model.layers.37.self_attn.v_proj.weight": "model-00009-of-00011.safetensors",
424 "language_model.model.layers.38.input_layernorm.weight": "model-00009-of-00011.safetensors",
425 "language_model.model.layers.38.mlp.down_proj.weight": "model-00009-of-00011.safetensors",
426 "language_model.model.layers.38.mlp.gate_proj.weight": "model-00009-of-00011.safetensors",
427 "language_model.model.layers.38.mlp.up_proj.weight": "model-00009-of-00011.safetensors",
428 "language_model.model.layers.38.post_attention_layernorm.weight": "model-00009-of-00011.safetensors",
429 "language_model.model.layers.38.post_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
430 "language_model.model.layers.38.pre_feedforward_layernorm.weight": "model-00009-of-00011.safetensors",
431 "language_model.model.layers.38.self_attn.k_norm.weight": "model-00009-of-00011.safetensors",
432 "language_model.model.layers.38.self_attn.k_proj.weight": "model-00009-of-00011.safetensors",
433 "language_model.model.layers.38.self_attn.o_proj.weight": "model-00009-of-00011.safetensors",
434 "language_model.model.layers.38.self_attn.q_norm.weight": "model-00009-of-00011.safetensors",
435 "language_model.model.layers.38.self_attn.q_proj.weight": "model-00009-of-00011.safetensors",
436 "language_model.model.layers.38.self_attn.v_proj.weight": "model-00009-of-00011.safetensors",
437 "language_model.model.layers.39.input_layernorm.weight": "model-00010-of-00011.safetensors",
438 "language_model.model.layers.39.mlp.down_proj.weight": "model-00010-of-00011.safetensors",
439 "language_model.model.layers.39.mlp.gate_proj.weight": "model-00009-of-00011.safetensors",
440 "language_model.model.layers.39.mlp.up_proj.weight": "model-00010-of-00011.safetensors",
441 "language_model.model.layers.39.post_attention_layernorm.weight": "model-00010-of-00011.safetensors",
442 "language_model.model.layers.39.post_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
443 "language_model.model.layers.39.pre_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
444 "language_model.model.layers.39.self_attn.k_norm.weight": "model-00009-of-00011.safetensors",
445 "language_model.model.layers.39.self_attn.k_proj.weight": "model-00009-of-00011.safetensors",
446 "language_model.model.layers.39.self_attn.o_proj.weight": "model-00009-of-00011.safetensors",
447 "language_model.model.layers.39.self_attn.q_norm.weight": "model-00009-of-00011.safetensors",
448 "language_model.model.layers.39.self_attn.q_proj.weight": "model-00009-of-00011.safetensors",
449 "language_model.model.layers.39.self_attn.v_proj.weight": "model-00009-of-00011.safetensors",
450 "language_model.model.layers.4.input_layernorm.weight": "model-00003-of-00011.safetensors",
451 "language_model.model.layers.4.mlp.down_proj.weight": "model-00003-of-00011.safetensors",
452 "language_model.model.layers.4.mlp.gate_proj.weight": "model-00003-of-00011.safetensors",
453 "language_model.model.layers.4.mlp.up_proj.weight": "model-00003-of-00011.safetensors",
454 "language_model.model.layers.4.post_attention_layernorm.weight": "model-00003-of-00011.safetensors",
455 "language_model.model.layers.4.post_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
456 "language_model.model.layers.4.pre_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
457 "language_model.model.layers.4.self_attn.k_norm.weight": "model-00003-of-00011.safetensors",
458 "language_model.model.layers.4.self_attn.k_proj.weight": "model-00003-of-00011.safetensors",
459 "language_model.model.layers.4.self_attn.o_proj.weight": "model-00003-of-00011.safetensors",
460 "language_model.model.layers.4.self_attn.q_norm.weight": "model-00003-of-00011.safetensors",
461 "language_model.model.layers.4.self_attn.q_proj.weight": "model-00003-of-00011.safetensors",
462 "language_model.model.layers.4.self_attn.v_proj.weight": "model-00003-of-00011.safetensors",
463 "language_model.model.layers.40.input_layernorm.weight": "model-00010-of-00011.safetensors",
464 "language_model.model.layers.40.mlp.down_proj.weight": "model-00010-of-00011.safetensors",
465 "language_model.model.layers.40.mlp.gate_proj.weight": "model-00010-of-00011.safetensors",
466 "language_model.model.layers.40.mlp.up_proj.weight": "model-00010-of-00011.safetensors",
467 "language_model.model.layers.40.post_attention_layernorm.weight": "model-00010-of-00011.safetensors",
468 "language_model.model.layers.40.post_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
469 "language_model.model.layers.40.pre_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
470 "language_model.model.layers.40.self_attn.k_norm.weight": "model-00010-of-00011.safetensors",
471 "language_model.model.layers.40.self_attn.k_proj.weight": "model-00010-of-00011.safetensors",
472 "language_model.model.layers.40.self_attn.o_proj.weight": "model-00010-of-00011.safetensors",
473 "language_model.model.layers.40.self_attn.q_norm.weight": "model-00010-of-00011.safetensors",
474 "language_model.model.layers.40.self_attn.q_proj.weight": "model-00010-of-00011.safetensors",
475 "language_model.model.layers.40.self_attn.v_proj.weight": "model-00010-of-00011.safetensors",
476 "language_model.model.layers.41.input_layernorm.weight": "model-00010-of-00011.safetensors",
477 "language_model.model.layers.41.mlp.down_proj.weight": "model-00010-of-00011.safetensors",
478 "language_model.model.layers.41.mlp.gate_proj.weight": "model-00010-of-00011.safetensors",
479 "language_model.model.layers.41.mlp.up_proj.weight": "model-00010-of-00011.safetensors",
480 "language_model.model.layers.41.post_attention_layernorm.weight": "model-00010-of-00011.safetensors",
481 "language_model.model.layers.41.post_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
482 "language_model.model.layers.41.pre_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
483 "language_model.model.layers.41.self_attn.k_norm.weight": "model-00010-of-00011.safetensors",
484 "language_model.model.layers.41.self_attn.k_proj.weight": "model-00010-of-00011.safetensors",
485 "language_model.model.layers.41.self_attn.o_proj.weight": "model-00010-of-00011.safetensors",
486 "language_model.model.layers.41.self_attn.q_norm.weight": "model-00010-of-00011.safetensors",
487 "language_model.model.layers.41.self_attn.q_proj.weight": "model-00010-of-00011.safetensors",
488 "language_model.model.layers.41.self_attn.v_proj.weight": "model-00010-of-00011.safetensors",
489 "language_model.model.layers.42.input_layernorm.weight": "model-00010-of-00011.safetensors",
490 "language_model.model.layers.42.mlp.down_proj.weight": "model-00010-of-00011.safetensors",
491 "language_model.model.layers.42.mlp.gate_proj.weight": "model-00010-of-00011.safetensors",
492 "language_model.model.layers.42.mlp.up_proj.weight": "model-00010-of-00011.safetensors",
493 "language_model.model.layers.42.post_attention_layernorm.weight": "model-00010-of-00011.safetensors",
494 "language_model.model.layers.42.post_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
495 "language_model.model.layers.42.pre_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
496 "language_model.model.layers.42.self_attn.k_norm.weight": "model-00010-of-00011.safetensors",
497 "language_model.model.layers.42.self_attn.k_proj.weight": "model-00010-of-00011.safetensors",
498 "language_model.model.layers.42.self_attn.o_proj.weight": "model-00010-of-00011.safetensors",
499 "language_model.model.layers.42.self_attn.q_norm.weight": "model-00010-of-00011.safetensors",
500 "language_model.model.layers.42.self_attn.q_proj.weight": "model-00010-of-00011.safetensors",
501 "language_model.model.layers.42.self_attn.v_proj.weight": "model-00010-of-00011.safetensors",
502 "language_model.model.layers.43.input_layernorm.weight": "model-00010-of-00011.safetensors",
503 "language_model.model.layers.43.mlp.down_proj.weight": "model-00010-of-00011.safetensors",
504 "language_model.model.layers.43.mlp.gate_proj.weight": "model-00010-of-00011.safetensors",
505 "language_model.model.layers.43.mlp.up_proj.weight": "model-00010-of-00011.safetensors",
506 "language_model.model.layers.43.post_attention_layernorm.weight": "model-00010-of-00011.safetensors",
507 "language_model.model.layers.43.post_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
508 "language_model.model.layers.43.pre_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
509 "language_model.model.layers.43.self_attn.k_norm.weight": "model-00010-of-00011.safetensors",
510 "language_model.model.layers.43.self_attn.k_proj.weight": "model-00010-of-00011.safetensors",
511 "language_model.model.layers.43.self_attn.o_proj.weight": "model-00010-of-00011.safetensors",
512 "language_model.model.layers.43.self_attn.q_norm.weight": "model-00010-of-00011.safetensors",
513 "language_model.model.layers.43.self_attn.q_proj.weight": "model-00010-of-00011.safetensors",
514 "language_model.model.layers.43.self_attn.v_proj.weight": "model-00010-of-00011.safetensors",
515 "language_model.model.layers.44.input_layernorm.weight": "model-00010-of-00011.safetensors",
516 "language_model.model.layers.44.mlp.down_proj.weight": "model-00010-of-00011.safetensors",
517 "language_model.model.layers.44.mlp.gate_proj.weight": "model-00010-of-00011.safetensors",
518 "language_model.model.layers.44.mlp.up_proj.weight": "model-00010-of-00011.safetensors",
519 "language_model.model.layers.44.post_attention_layernorm.weight": "model-00010-of-00011.safetensors",
520 "language_model.model.layers.44.post_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
521 "language_model.model.layers.44.pre_feedforward_layernorm.weight": "model-00010-of-00011.safetensors",
522 "language_model.model.layers.44.self_attn.k_norm.weight": "model-00010-of-00011.safetensors",
523 "language_model.model.layers.44.self_attn.k_proj.weight": "model-00010-of-00011.safetensors",
524 "language_model.model.layers.44.self_attn.o_proj.weight": "model-00010-of-00011.safetensors",
525 "language_model.model.layers.44.self_attn.q_norm.weight": "model-00010-of-00011.safetensors",
526 "language_model.model.layers.44.self_attn.q_proj.weight": "model-00010-of-00011.safetensors",
527 "language_model.model.layers.44.self_attn.v_proj.weight": "model-00010-of-00011.safetensors",
528 "language_model.model.layers.45.input_layernorm.weight": "model-00011-of-00011.safetensors",
529 "language_model.model.layers.45.mlp.down_proj.weight": "model-00011-of-00011.safetensors",
530 "language_model.model.layers.45.mlp.gate_proj.weight": "model-00011-of-00011.safetensors",
531 "language_model.model.layers.45.mlp.up_proj.weight": "model-00011-of-00011.safetensors",
532 "language_model.model.layers.45.post_attention_layernorm.weight": "model-00011-of-00011.safetensors",
533 "language_model.model.layers.45.post_feedforward_layernorm.weight": "model-00011-of-00011.safetensors",
534 "language_model.model.layers.45.pre_feedforward_layernorm.weight": "model-00011-of-00011.safetensors",
535 "language_model.model.layers.45.self_attn.k_norm.weight": "model-00011-of-00011.safetensors",
536 "language_model.model.layers.45.self_attn.k_proj.weight": "model-00011-of-00011.safetensors",
537 "language_model.model.layers.45.self_attn.o_proj.weight": "model-00011-of-00011.safetensors",
538 "language_model.model.layers.45.self_attn.q_norm.weight": "model-00011-of-00011.safetensors",
539 "language_model.model.layers.45.self_attn.q_proj.weight": "model-00011-of-00011.safetensors",
540 "language_model.model.layers.45.self_attn.v_proj.weight": "model-00011-of-00011.safetensors",
541 "language_model.model.layers.46.input_layernorm.weight": "model-00011-of-00011.safetensors",
542 "language_model.model.layers.46.mlp.down_proj.weight": "model-00011-of-00011.safetensors",
543 "language_model.model.layers.46.mlp.gate_proj.weight": "model-00011-of-00011.safetensors",
544 "language_model.model.layers.46.mlp.up_proj.weight": "model-00011-of-00011.safetensors",
545 "language_model.model.layers.46.post_attention_layernorm.weight": "model-00011-of-00011.safetensors",
546 "language_model.model.layers.46.post_feedforward_layernorm.weight": "model-00011-of-00011.safetensors",
547 "language_model.model.layers.46.pre_feedforward_layernorm.weight": "model-00011-of-00011.safetensors",
548 "language_model.model.layers.46.self_attn.k_norm.weight": "model-00011-of-00011.safetensors",
549 "language_model.model.layers.46.self_attn.k_proj.weight": "model-00011-of-00011.safetensors",
550 "language_model.model.layers.46.self_attn.o_proj.weight": "model-00011-of-00011.safetensors",
551 "language_model.model.layers.46.self_attn.q_norm.weight": "model-00011-of-00011.safetensors",
552 "language_model.model.layers.46.self_attn.q_proj.weight": "model-00011-of-00011.safetensors",
553 "language_model.model.layers.46.self_attn.v_proj.weight": "model-00011-of-00011.safetensors",
554 "language_model.model.layers.47.input_layernorm.weight": "model-00011-of-00011.safetensors",
555 "language_model.model.layers.47.mlp.down_proj.weight": "model-00011-of-00011.safetensors",
556 "language_model.model.layers.47.mlp.gate_proj.weight": "model-00011-of-00011.safetensors",
557 "language_model.model.layers.47.mlp.up_proj.weight": "model-00011-of-00011.safetensors",
558 "language_model.model.layers.47.post_attention_layernorm.weight": "model-00011-of-00011.safetensors",
559 "language_model.model.layers.47.post_feedforward_layernorm.weight": "model-00011-of-00011.safetensors",
560 "language_model.model.layers.47.pre_feedforward_layernorm.weight": "model-00011-of-00011.safetensors",
561 "language_model.model.layers.47.self_attn.k_norm.weight": "model-00011-of-00011.safetensors",
562 "language_model.model.layers.47.self_attn.k_proj.weight": "model-00011-of-00011.safetensors",
563 "language_model.model.layers.47.self_attn.o_proj.weight": "model-00011-of-00011.safetensors",
564 "language_model.model.layers.47.self_attn.q_norm.weight": "model-00011-of-00011.safetensors",
565 "language_model.model.layers.47.self_attn.q_proj.weight": "model-00011-of-00011.safetensors",
566 "language_model.model.layers.47.self_attn.v_proj.weight": "model-00011-of-00011.safetensors",
567 "language_model.model.layers.5.input_layernorm.weight": "model-00003-of-00011.safetensors",
568 "language_model.model.layers.5.mlp.down_proj.weight": "model-00003-of-00011.safetensors",
569 "language_model.model.layers.5.mlp.gate_proj.weight": "model-00003-of-00011.safetensors",
570 "language_model.model.layers.5.mlp.up_proj.weight": "model-00003-of-00011.safetensors",
571 "language_model.model.layers.5.post_attention_layernorm.weight": "model-00003-of-00011.safetensors",
572 "language_model.model.layers.5.post_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
573 "language_model.model.layers.5.pre_feedforward_layernorm.weight": "model-00003-of-00011.safetensors",
574 "language_model.model.layers.5.self_attn.k_norm.weight": "model-00003-of-00011.safetensors",
575 "language_model.model.layers.5.self_attn.k_proj.weight": "model-00003-of-00011.safetensors",
576 "language_model.model.layers.5.self_attn.o_proj.weight": "model-00003-of-00011.safetensors",
577 "language_model.model.layers.5.self_attn.q_norm.weight": "model-00003-of-00011.safetensors",
578 "language_model.model.layers.5.self_attn.q_proj.weight": "model-00003-of-00011.safetensors",
579 "language_model.model.layers.5.self_attn.v_proj.weight": "model-00003-of-00011.safetensors",
580 "language_model.model.layers.6.input_layernorm.weight": "model-00004-of-00011.safetensors",
581 "language_model.model.layers.6.mlp.down_proj.weight": "model-00004-of-00011.safetensors",
582 "language_model.model.layers.6.mlp.gate_proj.weight": "model-00003-of-00011.safetensors",
583 "language_model.model.layers.6.mlp.up_proj.weight": "model-00004-of-00011.safetensors",
584 "language_model.model.layers.6.post_attention_layernorm.weight": "model-00004-of-00011.safetensors",
585 "language_model.model.layers.6.post_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
586 "language_model.model.layers.6.pre_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
587 "language_model.model.layers.6.self_attn.k_norm.weight": "model-00003-of-00011.safetensors",
588 "language_model.model.layers.6.self_attn.k_proj.weight": "model-00003-of-00011.safetensors",
589 "language_model.model.layers.6.self_attn.o_proj.weight": "model-00003-of-00011.safetensors",
590 "language_model.model.layers.6.self_attn.q_norm.weight": "model-00003-of-00011.safetensors",
591 "language_model.model.layers.6.self_attn.q_proj.weight": "model-00003-of-00011.safetensors",
592 "language_model.model.layers.6.self_attn.v_proj.weight": "model-00003-of-00011.safetensors",
593 "language_model.model.layers.7.input_layernorm.weight": "model-00004-of-00011.safetensors",
594 "language_model.model.layers.7.mlp.down_proj.weight": "model-00004-of-00011.safetensors",
595 "language_model.model.layers.7.mlp.gate_proj.weight": "model-00004-of-00011.safetensors",
596 "language_model.model.layers.7.mlp.up_proj.weight": "model-00004-of-00011.safetensors",
597 "language_model.model.layers.7.post_attention_layernorm.weight": "model-00004-of-00011.safetensors",
598 "language_model.model.layers.7.post_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
599 "language_model.model.layers.7.pre_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
600 "language_model.model.layers.7.self_attn.k_norm.weight": "model-00004-of-00011.safetensors",
601 "language_model.model.layers.7.self_attn.k_proj.weight": "model-00004-of-00011.safetensors",
602 "language_model.model.layers.7.self_attn.o_proj.weight": "model-00004-of-00011.safetensors",
603 "language_model.model.layers.7.self_attn.q_norm.weight": "model-00004-of-00011.safetensors",
604 "language_model.model.layers.7.self_attn.q_proj.weight": "model-00004-of-00011.safetensors",
605 "language_model.model.layers.7.self_attn.v_proj.weight": "model-00004-of-00011.safetensors",
606 "language_model.model.layers.8.input_layernorm.weight": "model-00004-of-00011.safetensors",
607 "language_model.model.layers.8.mlp.down_proj.weight": "model-00004-of-00011.safetensors",
608 "language_model.model.layers.8.mlp.gate_proj.weight": "model-00004-of-00011.safetensors",
609 "language_model.model.layers.8.mlp.up_proj.weight": "model-00004-of-00011.safetensors",
610 "language_model.model.layers.8.post_attention_layernorm.weight": "model-00004-of-00011.safetensors",
611 "language_model.model.layers.8.post_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
612 "language_model.model.layers.8.pre_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
613 "language_model.model.layers.8.self_attn.k_norm.weight": "model-00004-of-00011.safetensors",
614 "language_model.model.layers.8.self_attn.k_proj.weight": "model-00004-of-00011.safetensors",
615 "language_model.model.layers.8.self_attn.o_proj.weight": "model-00004-of-00011.safetensors",
616 "language_model.model.layers.8.self_attn.q_norm.weight": "model-00004-of-00011.safetensors",
617 "language_model.model.layers.8.self_attn.q_proj.weight": "model-00004-of-00011.safetensors",
618 "language_model.model.layers.8.self_attn.v_proj.weight": "model-00004-of-00011.safetensors",
619 "language_model.model.layers.9.input_layernorm.weight": "model-00004-of-00011.safetensors",
620 "language_model.model.layers.9.mlp.down_proj.weight": "model-00004-of-00011.safetensors",
621 "language_model.model.layers.9.mlp.gate_proj.weight": "model-00004-of-00011.safetensors",
622 "language_model.model.layers.9.mlp.up_proj.weight": "model-00004-of-00011.safetensors",
623 "language_model.model.layers.9.post_attention_layernorm.weight": "model-00004-of-00011.safetensors",
624 "language_model.model.layers.9.post_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
625 "language_model.model.layers.9.pre_feedforward_layernorm.weight": "model-00004-of-00011.safetensors",
626 "language_model.model.layers.9.self_attn.k_norm.weight": "model-00004-of-00011.safetensors",
627 "language_model.model.layers.9.self_attn.k_proj.weight": "model-00004-of-00011.safetensors",
628 "language_model.model.layers.9.self_attn.o_proj.weight": "model-00004-of-00011.safetensors",
629 "language_model.model.layers.9.self_attn.q_norm.weight": "model-00004-of-00011.safetensors",
630 "language_model.model.layers.9.self_attn.q_proj.weight": "model-00004-of-00011.safetensors",
631 "language_model.model.layers.9.self_attn.v_proj.weight": "model-00004-of-00011.safetensors",
632 "language_model.model.norm.weight": "model-00011-of-00011.safetensors",
633 "multi_modal_projector.mm_input_projection_weight": "model-00001-of-00011.safetensors",
634 "multi_modal_projector.mm_soft_emb_norm.weight": "model-00001-of-00011.safetensors",
635 "vision_tower.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00011.safetensors",
636 "vision_tower.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00011.safetensors",
637 "vision_tower.vision_model.embeddings.position_embedding.weight": "model-00001-of-00011.safetensors",
638 "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00011.safetensors",
639 "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00011.safetensors",
640 "vision_tower.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00011.safetensors",
641 "vision_tower.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00011.safetensors",
642 "vision_tower.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00011.safetensors",
643 "vision_tower.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00011.safetensors",
644 "vision_tower.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00011.safetensors",
645 "vision_tower.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00011.safetensors",
646 "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
647 "vision_tower.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
648 "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
649 "vision_tower.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
650 "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
651 "vision_tower.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
652 "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
653 "vision_tower.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
654 "vision_tower.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00011.safetensors",
655 "vision_tower.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00011.safetensors",
656 "vision_tower.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00011.safetensors",
657 "vision_tower.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00011.safetensors",
658 "vision_tower.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00011.safetensors",
659 "vision_tower.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00011.safetensors",
660 "vision_tower.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00011.safetensors",
661 "vision_tower.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00011.safetensors",
662 "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
663 "vision_tower.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
664 "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
665 "vision_tower.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
666 "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
667 "vision_tower.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
668 "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
669 "vision_tower.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
670 "vision_tower.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00011.safetensors",
671 "vision_tower.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00011.safetensors",
672 "vision_tower.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00011.safetensors",
673 "vision_tower.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00011.safetensors",
674 "vision_tower.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00011.safetensors",
675 "vision_tower.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00011.safetensors",
676 "vision_tower.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00011.safetensors",
677 "vision_tower.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00011.safetensors",
678 "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
679 "vision_tower.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
680 "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
681 "vision_tower.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
682 "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
683 "vision_tower.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
684 "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
685 "vision_tower.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
686 "vision_tower.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00011.safetensors",
687 "vision_tower.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00011.safetensors",
688 "vision_tower.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00011.safetensors",
689 "vision_tower.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00011.safetensors",
690 "vision_tower.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00011.safetensors",
691 "vision_tower.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00011.safetensors",
692 "vision_tower.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00011.safetensors",
693 "vision_tower.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00011.safetensors",
694 "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
695 "vision_tower.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
696 "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
697 "vision_tower.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
698 "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
699 "vision_tower.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
700 "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
701 "vision_tower.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
702 "vision_tower.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00011.safetensors",
703 "vision_tower.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00011.safetensors",
704 "vision_tower.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00011.safetensors",
705 "vision_tower.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00011.safetensors",
706 "vision_tower.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00011.safetensors",
707 "vision_tower.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00011.safetensors",
708 "vision_tower.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00011.safetensors",
709 "vision_tower.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00011.safetensors",
710 "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
711 "vision_tower.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
712 "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
713 "vision_tower.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
714 "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
715 "vision_tower.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
716 "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
717 "vision_tower.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
718 "vision_tower.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00011.safetensors",
719 "vision_tower.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00011.safetensors",
720 "vision_tower.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00011.safetensors",
721 "vision_tower.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00011.safetensors",
722 "vision_tower.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00011.safetensors",
723 "vision_tower.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00011.safetensors",
724 "vision_tower.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00011.safetensors",
725 "vision_tower.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00011.safetensors",
726 "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
727 "vision_tower.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
728 "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
729 "vision_tower.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
730 "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
731 "vision_tower.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
732 "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
733 "vision_tower.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
734 "vision_tower.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00011.safetensors",
735 "vision_tower.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00011.safetensors",
736 "vision_tower.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00011.safetensors",
737 "vision_tower.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00011.safetensors",
738 "vision_tower.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00011.safetensors",
739 "vision_tower.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00011.safetensors",
740 "vision_tower.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00011.safetensors",
741 "vision_tower.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00011.safetensors",
742 "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
743 "vision_tower.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
744 "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
745 "vision_tower.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
746 "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
747 "vision_tower.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
748 "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
749 "vision_tower.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
750 "vision_tower.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00011.safetensors",
751 "vision_tower.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00011.safetensors",
752 "vision_tower.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00011.safetensors",
753 "vision_tower.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00011.safetensors",
754 "vision_tower.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00011.safetensors",
755 "vision_tower.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00011.safetensors",
756 "vision_tower.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00011.safetensors",
757 "vision_tower.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00011.safetensors",
758 "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
759 "vision_tower.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
760 "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
761 "vision_tower.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
762 "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
763 "vision_tower.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
764 "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
765 "vision_tower.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
766 "vision_tower.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00011.safetensors",
767 "vision_tower.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00011.safetensors",
768 "vision_tower.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00011.safetensors",
769 "vision_tower.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00011.safetensors",
770 "vision_tower.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00011.safetensors",
771 "vision_tower.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00011.safetensors",
772 "vision_tower.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00011.safetensors",
773 "vision_tower.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00011.safetensors",
774 "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
775 "vision_tower.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
776 "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
777 "vision_tower.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
778 "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
779 "vision_tower.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
780 "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
781 "vision_tower.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
782 "vision_tower.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00011.safetensors",
783 "vision_tower.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00011.safetensors",
784 "vision_tower.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00011.safetensors",
785 "vision_tower.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00011.safetensors",
786 "vision_tower.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00011.safetensors",
787 "vision_tower.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00011.safetensors",
788 "vision_tower.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00011.safetensors",
789 "vision_tower.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00011.safetensors",
790 "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
791 "vision_tower.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
792 "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
793 "vision_tower.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
794 "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
795 "vision_tower.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
796 "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
797 "vision_tower.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
798 "vision_tower.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00011.safetensors",
799 "vision_tower.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00011.safetensors",
800 "vision_tower.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00011.safetensors",
801 "vision_tower.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00011.safetensors",
802 "vision_tower.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00011.safetensors",
803 "vision_tower.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00011.safetensors",
804 "vision_tower.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00011.safetensors",
805 "vision_tower.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00011.safetensors",
806 "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
807 "vision_tower.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
808 "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
809 "vision_tower.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
810 "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
811 "vision_tower.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
812 "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
813 "vision_tower.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
814 "vision_tower.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00011.safetensors",
815 "vision_tower.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00011.safetensors",
816 "vision_tower.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00011.safetensors",
817 "vision_tower.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00011.safetensors",
818 "vision_tower.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00011.safetensors",
819 "vision_tower.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00011.safetensors",
820 "vision_tower.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00011.safetensors",
821 "vision_tower.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00011.safetensors",
822 "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
823 "vision_tower.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
824 "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
825 "vision_tower.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
826 "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
827 "vision_tower.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
828 "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
829 "vision_tower.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
830 "vision_tower.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00011.safetensors",
831 "vision_tower.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00011.safetensors",
832 "vision_tower.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00011.safetensors",
833 "vision_tower.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00011.safetensors",
834 "vision_tower.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00011.safetensors",
835 "vision_tower.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00011.safetensors",
836 "vision_tower.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00011.safetensors",
837 "vision_tower.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00011.safetensors",
838 "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
839 "vision_tower.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
840 "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
841 "vision_tower.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
842 "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
843 "vision_tower.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
844 "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
845 "vision_tower.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
846 "vision_tower.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00011.safetensors",
847 "vision_tower.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00011.safetensors",
848 "vision_tower.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00011.safetensors",
849 "vision_tower.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00011.safetensors",
850 "vision_tower.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00011.safetensors",
851 "vision_tower.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00011.safetensors",
852 "vision_tower.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00011.safetensors",
853 "vision_tower.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00011.safetensors",
854 "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
855 "vision_tower.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
856 "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
857 "vision_tower.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
858 "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
859 "vision_tower.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
860 "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
861 "vision_tower.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
862 "vision_tower.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00011.safetensors",
863 "vision_tower.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00011.safetensors",
864 "vision_tower.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00011.safetensors",
865 "vision_tower.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00011.safetensors",
866 "vision_tower.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00011.safetensors",
867 "vision_tower.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00011.safetensors",
868 "vision_tower.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00011.safetensors",
869 "vision_tower.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00011.safetensors",
870 "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
871 "vision_tower.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
872 "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
873 "vision_tower.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
874 "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
875 "vision_tower.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
876 "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
877 "vision_tower.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
878 "vision_tower.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00011.safetensors",
879 "vision_tower.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00011.safetensors",
880 "vision_tower.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00011.safetensors",
881 "vision_tower.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00011.safetensors",
882 "vision_tower.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00011.safetensors",
883 "vision_tower.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00011.safetensors",
884 "vision_tower.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00011.safetensors",
885 "vision_tower.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00011.safetensors",
886 "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
887 "vision_tower.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
888 "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
889 "vision_tower.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
890 "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
891 "vision_tower.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
892 "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
893 "vision_tower.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
894 "vision_tower.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00011.safetensors",
895 "vision_tower.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00011.safetensors",
896 "vision_tower.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00011.safetensors",
897 "vision_tower.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00011.safetensors",
898 "vision_tower.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00011.safetensors",
899 "vision_tower.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00011.safetensors",
900 "vision_tower.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00011.safetensors",
901 "vision_tower.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00011.safetensors",
902 "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
903 "vision_tower.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
904 "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
905 "vision_tower.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
906 "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
907 "vision_tower.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
908 "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
909 "vision_tower.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
910 "vision_tower.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00011.safetensors",
911 "vision_tower.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00011.safetensors",
912 "vision_tower.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00011.safetensors",
913 "vision_tower.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00011.safetensors",
914 "vision_tower.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00011.safetensors",
915 "vision_tower.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00011.safetensors",
916 "vision_tower.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00011.safetensors",
917 "vision_tower.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00011.safetensors",
918 "vision_tower.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
919 "vision_tower.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
920 "vision_tower.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
921 "vision_tower.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
922 "vision_tower.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
923 "vision_tower.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
924 "vision_tower.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
925 "vision_tower.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
926 "vision_tower.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00011.safetensors",
927 "vision_tower.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00011.safetensors",
928 "vision_tower.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00011.safetensors",
929 "vision_tower.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00011.safetensors",
930 "vision_tower.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00011.safetensors",
931 "vision_tower.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00011.safetensors",
932 "vision_tower.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00011.safetensors",
933 "vision_tower.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00011.safetensors",
934 "vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
935 "vision_tower.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
936 "vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
937 "vision_tower.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
938 "vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
939 "vision_tower.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
940 "vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
941 "vision_tower.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
942 "vision_tower.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00011.safetensors",
943 "vision_tower.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00011.safetensors",
944 "vision_tower.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00011.safetensors",
945 "vision_tower.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00011.safetensors",
946 "vision_tower.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00011.safetensors",
947 "vision_tower.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00011.safetensors",
948 "vision_tower.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00011.safetensors",
949 "vision_tower.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00011.safetensors",
950 "vision_tower.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
951 "vision_tower.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
952 "vision_tower.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
953 "vision_tower.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
954 "vision_tower.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
955 "vision_tower.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
956 "vision_tower.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
957 "vision_tower.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
958 "vision_tower.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00011.safetensors",
959 "vision_tower.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00011.safetensors",
960 "vision_tower.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00011.safetensors",
961 "vision_tower.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00011.safetensors",
962 "vision_tower.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00011.safetensors",
963 "vision_tower.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00011.safetensors",
964 "vision_tower.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00011.safetensors",
965 "vision_tower.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00011.safetensors",
966 "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
967 "vision_tower.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
968 "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
969 "vision_tower.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
970 "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
971 "vision_tower.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
972 "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
973 "vision_tower.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
974 "vision_tower.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00011.safetensors",
975 "vision_tower.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00011.safetensors",
976 "vision_tower.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00011.safetensors",
977 "vision_tower.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00011.safetensors",
978 "vision_tower.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00011.safetensors",
979 "vision_tower.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00011.safetensors",
980 "vision_tower.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00011.safetensors",
981 "vision_tower.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00011.safetensors",
982 "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
983 "vision_tower.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
984 "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
985 "vision_tower.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
986 "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
987 "vision_tower.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
988 "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
989 "vision_tower.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
990 "vision_tower.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00011.safetensors",
991 "vision_tower.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00011.safetensors",
992 "vision_tower.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00011.safetensors",
993 "vision_tower.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00011.safetensors",
994 "vision_tower.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00011.safetensors",
995 "vision_tower.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00011.safetensors",
996 "vision_tower.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00011.safetensors",
997 "vision_tower.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00011.safetensors",
998 "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
999 "vision_tower.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
1000 "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
1001 "vision_tower.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
1002 "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
1003 "vision_tower.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
1004 "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
1005 "vision_tower.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
1006 "vision_tower.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00011.safetensors",
1007 "vision_tower.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00011.safetensors",
1008 "vision_tower.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00011.safetensors",
1009 "vision_tower.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00011.safetensors",
1010 "vision_tower.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00011.safetensors",
1011 "vision_tower.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00011.safetensors",
1012 "vision_tower.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00011.safetensors",
1013 "vision_tower.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00011.safetensors",
1014 "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
1015 "vision_tower.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
1016 "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
1017 "vision_tower.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
1018 "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
1019 "vision_tower.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
1020 "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
1021 "vision_tower.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
1022 "vision_tower.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00011.safetensors",
1023 "vision_tower.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00011.safetensors",
1024 "vision_tower.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00011.safetensors",
1025 "vision_tower.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00011.safetensors",
1026 "vision_tower.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00011.safetensors",
1027 "vision_tower.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00011.safetensors",
1028 "vision_tower.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00011.safetensors",
1029 "vision_tower.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00011.safetensors",
1030 "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
1031 "vision_tower.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
1032 "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
1033 "vision_tower.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
1034 "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
1035 "vision_tower.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
1036 "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
1037 "vision_tower.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
1038 "vision_tower.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00011.safetensors",
1039 "vision_tower.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00011.safetensors",
1040 "vision_tower.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00011.safetensors",
1041 "vision_tower.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00011.safetensors",
1042 "vision_tower.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00011.safetensors",
1043 "vision_tower.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00011.safetensors",
1044 "vision_tower.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00011.safetensors",
1045 "vision_tower.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00011.safetensors",
1046 "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
1047 "vision_tower.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
1048 "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
1049 "vision_tower.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
1050 "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
1051 "vision_tower.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
1052 "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
1053 "vision_tower.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
1054 "vision_tower.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00011.safetensors",
1055 "vision_tower.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00011.safetensors",
1056 "vision_tower.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00011.safetensors",
1057 "vision_tower.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00011.safetensors",
1058 "vision_tower.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00011.safetensors",
1059 "vision_tower.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00011.safetensors",
1060 "vision_tower.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00011.safetensors",
1061 "vision_tower.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00011.safetensors",
1062 "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00011.safetensors",
1063 "vision_tower.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00011.safetensors",
1064 "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00011.safetensors",
1065 "vision_tower.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00011.safetensors",
1066 "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00011.safetensors",
1067 "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00011.safetensors",
1068 "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00011.safetensors",
1069 "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00011.safetensors",
1070 "vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00011.safetensors",
1071 "vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00011.safetensors"
1072 }
1073 }
1074