openvino_model.xml
500.9 KB · 14529 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="torch_jit" version="11">
3 <layers>
4 <layer id="1" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <rt_info>
7 <attribute name="fused_names" version="0" value="input_ids" />
8 </rt_info>
9 <output>
10 <port id="0" precision="I64" names="input_ids">
11 <dim>-1</dim>
12 <dim>-1</dim>
13 </port>
14 </output>
15 </layer>
16 <layer id="0" name="attention_mask" type="Parameter" version="opset1">
17 <data shape="?,?" element_type="i64" />
18 <rt_info>
19 <attribute name="fused_names" version="0" value="attention_mask" />
20 </rt_info>
21 <output>
22 <port id="0" precision="I64" names="attention_mask">
23 <dim>-1</dim>
24 <dim>-1</dim>
25 </port>
26 </output>
27 </layer>
28 <layer id="2" name="qa_outputs.bias" type="Const" version="opset1">
29 <data element_type="f32" shape="2" offset="0" size="8" />
30 <rt_info>
31 <attribute name="fused_names" version="0" value="qa_outputs.bias" />
32 </rt_info>
33 <output>
34 <port id="0" precision="FP32" names="qa_outputs.bias">
35 <dim>2</dim>
36 </port>
37 </output>
38 </layer>
39 <layer id="3" name="distilbert.transformer.layer.5.ffn.lin2.bias" type="Const" version="opset1">
40 <data element_type="f32" shape="768" offset="8" size="3072" />
41 <rt_info>
42 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.ffn.lin2.bias" />
43 </rt_info>
44 <output>
45 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.ffn.lin2.bias">
46 <dim>768</dim>
47 </port>
48 </output>
49 </layer>
50 <layer id="4" name="distilbert.transformer.layer.5.ffn.lin1.bias" type="Const" version="opset1">
51 <data element_type="f32" shape="3072" offset="3080" size="12288" />
52 <rt_info>
53 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.ffn.lin1.bias" />
54 </rt_info>
55 <output>
56 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.ffn.lin1.bias">
57 <dim>3072</dim>
58 </port>
59 </output>
60 </layer>
61 <layer id="5" name="distilbert.transformer.layer.5.attention.out_lin.bias" type="Const" version="opset1">
62 <data element_type="f32" shape="768" offset="15368" size="3072" />
63 <rt_info>
64 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.attention.out_lin.bias" />
65 </rt_info>
66 <output>
67 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.attention.out_lin.bias">
68 <dim>768</dim>
69 </port>
70 </output>
71 </layer>
72 <layer id="6" name="/distilbert/transformer/layer.0/attention/Constant_12" type="Const" version="opset1">
73 <data element_type="i64" shape="" offset="18440" size="8" />
74 <rt_info>
75 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_12" />
76 </rt_info>
77 <output>
78 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_12_output_0" />
79 </output>
80 </layer>
81 <layer id="7" name="/distilbert/transformer/layer.0/attention/Equal" type="Equal" version="opset1">
82 <data auto_broadcast="numpy" />
83 <rt_info>
84 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Equal" />
85 </rt_info>
86 <input>
87 <port id="0" precision="I64">
88 <dim>-1</dim>
89 <dim>-1</dim>
90 </port>
91 <port id="1" precision="I64" />
92 </input>
93 <output>
94 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.0/attention/Equal_output_0">
95 <dim>-1</dim>
96 <dim>-1</dim>
97 </port>
98 </output>
99 </layer>
100 <layer id="8" name="distilbert.transformer.layer.4.ffn.lin2.bias" type="Const" version="opset1">
101 <data element_type="f32" shape="768" offset="18448" size="3072" />
102 <rt_info>
103 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.ffn.lin2.bias" />
104 </rt_info>
105 <output>
106 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.ffn.lin2.bias">
107 <dim>768</dim>
108 </port>
109 </output>
110 </layer>
111 <layer id="9" name="distilbert.transformer.layer.4.ffn.lin1.bias" type="Const" version="opset1">
112 <data element_type="f32" shape="3072" offset="21520" size="12288" />
113 <rt_info>
114 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.ffn.lin1.bias" />
115 </rt_info>
116 <output>
117 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.ffn.lin1.bias">
118 <dim>3072</dim>
119 </port>
120 </output>
121 </layer>
122 <layer id="10" name="distilbert.transformer.layer.4.attention.out_lin.bias" type="Const" version="opset1">
123 <data element_type="f32" shape="768" offset="33808" size="3072" />
124 <rt_info>
125 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.attention.out_lin.bias" />
126 </rt_info>
127 <output>
128 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.attention.out_lin.bias">
129 <dim>768</dim>
130 </port>
131 </output>
132 </layer>
133 <layer id="11" name="distilbert.transformer.layer.3.ffn.lin2.bias" type="Const" version="opset1">
134 <data element_type="f32" shape="768" offset="36880" size="3072" />
135 <rt_info>
136 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.ffn.lin2.bias" />
137 </rt_info>
138 <output>
139 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.ffn.lin2.bias">
140 <dim>768</dim>
141 </port>
142 </output>
143 </layer>
144 <layer id="12" name="distilbert.transformer.layer.3.ffn.lin1.bias" type="Const" version="opset1">
145 <data element_type="f32" shape="3072" offset="39952" size="12288" />
146 <rt_info>
147 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.ffn.lin1.bias" />
148 </rt_info>
149 <output>
150 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.ffn.lin1.bias">
151 <dim>3072</dim>
152 </port>
153 </output>
154 </layer>
155 <layer id="13" name="distilbert.transformer.layer.3.attention.out_lin.bias" type="Const" version="opset1">
156 <data element_type="f32" shape="768" offset="52240" size="3072" />
157 <rt_info>
158 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.attention.out_lin.bias" />
159 </rt_info>
160 <output>
161 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.attention.out_lin.bias">
162 <dim>768</dim>
163 </port>
164 </output>
165 </layer>
166 <layer id="14" name="distilbert.transformer.layer.2.ffn.lin2.bias" type="Const" version="opset1">
167 <data element_type="f32" shape="768" offset="55312" size="3072" />
168 <rt_info>
169 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.ffn.lin2.bias" />
170 </rt_info>
171 <output>
172 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.ffn.lin2.bias">
173 <dim>768</dim>
174 </port>
175 </output>
176 </layer>
177 <layer id="15" name="distilbert.transformer.layer.2.ffn.lin1.bias" type="Const" version="opset1">
178 <data element_type="f32" shape="3072" offset="58384" size="12288" />
179 <rt_info>
180 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.ffn.lin1.bias" />
181 </rt_info>
182 <output>
183 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.ffn.lin1.bias">
184 <dim>3072</dim>
185 </port>
186 </output>
187 </layer>
188 <layer id="16" name="distilbert.transformer.layer.2.attention.out_lin.bias" type="Const" version="opset1">
189 <data element_type="f32" shape="768" offset="70672" size="3072" />
190 <rt_info>
191 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.attention.out_lin.bias" />
192 </rt_info>
193 <output>
194 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.attention.out_lin.bias">
195 <dim>768</dim>
196 </port>
197 </output>
198 </layer>
199 <layer id="17" name="distilbert.transformer.layer.1.ffn.lin2.bias" type="Const" version="opset1">
200 <data element_type="f32" shape="768" offset="73744" size="3072" />
201 <rt_info>
202 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.ffn.lin2.bias" />
203 </rt_info>
204 <output>
205 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.ffn.lin2.bias">
206 <dim>768</dim>
207 </port>
208 </output>
209 </layer>
210 <layer id="18" name="distilbert.transformer.layer.1.ffn.lin1.bias" type="Const" version="opset1">
211 <data element_type="f32" shape="3072" offset="76816" size="12288" />
212 <rt_info>
213 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.ffn.lin1.bias" />
214 </rt_info>
215 <output>
216 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.ffn.lin1.bias">
217 <dim>3072</dim>
218 </port>
219 </output>
220 </layer>
221 <layer id="19" name="distilbert.transformer.layer.1.attention.out_lin.bias" type="Const" version="opset1">
222 <data element_type="f32" shape="768" offset="89104" size="3072" />
223 <rt_info>
224 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.attention.out_lin.bias" />
225 </rt_info>
226 <output>
227 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.attention.out_lin.bias">
228 <dim>768</dim>
229 </port>
230 </output>
231 </layer>
232 <layer id="20" name="distilbert.transformer.layer.0.ffn.lin2.bias" type="Const" version="opset1">
233 <data element_type="f32" shape="768" offset="92176" size="3072" />
234 <rt_info>
235 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.ffn.lin2.bias" />
236 </rt_info>
237 <output>
238 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.ffn.lin2.bias">
239 <dim>768</dim>
240 </port>
241 </output>
242 </layer>
243 <layer id="21" name="distilbert.transformer.layer.0.ffn.lin1.bias" type="Const" version="opset1">
244 <data element_type="f32" shape="3072" offset="95248" size="12288" />
245 <rt_info>
246 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.ffn.lin1.bias" />
247 </rt_info>
248 <output>
249 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.ffn.lin1.bias">
250 <dim>3072</dim>
251 </port>
252 </output>
253 </layer>
254 <layer id="22" name="distilbert.transformer.layer.0.attention.out_lin.bias" type="Const" version="opset1">
255 <data element_type="f32" shape="768" offset="107536" size="3072" />
256 <rt_info>
257 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.attention.out_lin.bias" />
258 </rt_info>
259 <output>
260 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.attention.out_lin.bias">
261 <dim>768</dim>
262 </port>
263 </output>
264 </layer>
265 <layer id="23" name="distilbert.embeddings.word_embeddings.weight" type="Const" version="opset1">
266 <data element_type="f32" shape="28996, 768" offset="110608" size="89075712" />
267 <rt_info>
268 <attribute name="fused_names" version="0" value="distilbert.embeddings.word_embeddings.weight" />
269 </rt_info>
270 <output>
271 <port id="0" precision="FP32" names="distilbert.embeddings.word_embeddings.weight">
272 <dim>28996</dim>
273 <dim>768</dim>
274 </port>
275 </output>
276 </layer>
277 <layer id="24" name="Constant_255188" type="Const" version="opset1">
278 <data element_type="i64" shape="" offset="18440" size="8" />
279 <rt_info>
280 <attribute name="fused_names" version="0" value="Constant_255188" />
281 </rt_info>
282 <output>
283 <port id="0" precision="I64" />
284 </output>
285 </layer>
286 <layer id="25" name="/distilbert/embeddings/word_embeddings/Gather" type="Gather" version="opset8">
287 <data batch_dims="0" />
288 <rt_info>
289 <attribute name="fused_names" version="0" value="/distilbert/embeddings/word_embeddings/Gather" />
290 </rt_info>
291 <input>
292 <port id="0" precision="FP32">
293 <dim>28996</dim>
294 <dim>768</dim>
295 </port>
296 <port id="1" precision="I64">
297 <dim>-1</dim>
298 <dim>-1</dim>
299 </port>
300 <port id="2" precision="I64" />
301 </input>
302 <output>
303 <port id="3" precision="FP32" names="/distilbert/embeddings/word_embeddings/Gather_output_0">
304 <dim>-1</dim>
305 <dim>-1</dim>
306 <dim>768</dim>
307 </port>
308 </output>
309 </layer>
310 <layer id="26" name="distilbert.embeddings.position_embeddings.weight" type="Const" version="opset1">
311 <data element_type="f32" shape="512, 768" offset="89186320" size="1572864" />
312 <rt_info>
313 <attribute name="fused_names" version="0" value="distilbert.embeddings.position_embeddings.weight" />
314 </rt_info>
315 <output>
316 <port id="0" precision="FP32" names="distilbert.embeddings.position_embeddings.weight">
317 <dim>512</dim>
318 <dim>768</dim>
319 </port>
320 </output>
321 </layer>
322 <layer id="27" name="Constant_4" type="Const" version="opset1">
323 <data element_type="i64" shape="1, 512" offset="90759184" size="4096" />
324 <rt_info>
325 <attribute name="fused_names" version="0" value="Constant_4" />
326 </rt_info>
327 <output>
328 <port id="0" precision="I64" names="onnx::Slice_110">
329 <dim>1</dim>
330 <dim>512</dim>
331 </port>
332 </output>
333 </layer>
334 <layer id="28" name="/distilbert/embeddings/Constant_1" type="Const" version="opset1">
335 <data element_type="i64" shape="1" offset="18440" size="8" />
336 <rt_info>
337 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Constant_1" />
338 </rt_info>
339 <output>
340 <port id="0" precision="I64" names="/distilbert/embeddings/Constant_1_output_0">
341 <dim>1</dim>
342 </port>
343 </output>
344 </layer>
345 <layer id="29" name="/distilbert/embeddings/Shape" type="ShapeOf" version="opset3">
346 <data output_type="i64" />
347 <rt_info>
348 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Shape" />
349 </rt_info>
350 <input>
351 <port id="0" precision="FP32">
352 <dim>-1</dim>
353 <dim>-1</dim>
354 <dim>768</dim>
355 </port>
356 </input>
357 <output>
358 <port id="1" precision="I64" names="/distilbert/embeddings/Shape_output_0">
359 <dim>3</dim>
360 </port>
361 </output>
362 </layer>
363 <layer id="30" name="/distilbert/embeddings/Constant" type="Const" version="opset1">
364 <data element_type="i64" shape="" offset="90763280" size="8" />
365 <rt_info>
366 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Constant" />
367 </rt_info>
368 <output>
369 <port id="0" precision="I64" names="/distilbert/embeddings/Constant_output_0" />
370 </output>
371 </layer>
372 <layer id="31" name="Constant_255192" type="Const" version="opset1">
373 <data element_type="i64" shape="" offset="18440" size="8" />
374 <rt_info>
375 <attribute name="fused_names" version="0" value="Constant_255192" />
376 </rt_info>
377 <output>
378 <port id="0" precision="I64" />
379 </output>
380 </layer>
381 <layer id="32" name="/distilbert/embeddings/Gather" type="Gather" version="opset8">
382 <data batch_dims="0" />
383 <rt_info>
384 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Constant, /distilbert/embeddings/Gather, Constant_255192" />
385 </rt_info>
386 <input>
387 <port id="0" precision="I64">
388 <dim>3</dim>
389 </port>
390 <port id="1" precision="I64" />
391 <port id="2" precision="I64" />
392 </input>
393 <output>
394 <port id="3" precision="I64" names="/distilbert/embeddings/Gather_output_0" />
395 </output>
396 </layer>
397 <layer id="33" name="Constant_255196" type="Const" version="opset1">
398 <data element_type="i64" shape="1" offset="18440" size="8" />
399 <rt_info>
400 <attribute name="fused_names" version="0" value="Constant_255196" />
401 </rt_info>
402 <output>
403 <port id="0" precision="I64">
404 <dim>1</dim>
405 </port>
406 </output>
407 </layer>
408 <layer id="34" name="/distilbert/embeddings/Unsqueeze" type="Unsqueeze" version="opset1">
409 <rt_info>
410 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Unsqueeze, Constant_255196" />
411 </rt_info>
412 <input>
413 <port id="0" precision="I64" />
414 <port id="1" precision="I64">
415 <dim>1</dim>
416 </port>
417 </input>
418 <output>
419 <port id="2" precision="I64" names="/distilbert/embeddings/Unsqueeze_output_0">
420 <dim>1</dim>
421 </port>
422 </output>
423 </layer>
424 <layer id="35" name="/distilbert/embeddings/Constant_3" type="Const" version="opset1">
425 <data element_type="i64" shape="1" offset="90763280" size="8" />
426 <rt_info>
427 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Constant_3" />
428 </rt_info>
429 <output>
430 <port id="0" precision="I64" names="/distilbert/embeddings/Constant_3_output_0">
431 <dim>1</dim>
432 </port>
433 </output>
434 </layer>
435 <layer id="36" name="/distilbert/embeddings/Constant_2" type="Const" version="opset1">
436 <data element_type="i64" shape="1" offset="90763280" size="8" />
437 <rt_info>
438 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Constant_2" />
439 </rt_info>
440 <output>
441 <port id="0" precision="I64" names="/distilbert/embeddings/Constant_2_output_0">
442 <dim>1</dim>
443 </port>
444 </output>
445 </layer>
446 <layer id="37" name="/distilbert/embeddings/Slice" type="Slice" version="opset8">
447 <rt_info>
448 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Slice" />
449 </rt_info>
450 <input>
451 <port id="0" precision="I64">
452 <dim>1</dim>
453 <dim>512</dim>
454 </port>
455 <port id="1" precision="I64">
456 <dim>1</dim>
457 </port>
458 <port id="2" precision="I64">
459 <dim>1</dim>
460 </port>
461 <port id="3" precision="I64">
462 <dim>1</dim>
463 </port>
464 <port id="4" precision="I64">
465 <dim>1</dim>
466 </port>
467 </input>
468 <output>
469 <port id="5" precision="I64" names="/distilbert/embeddings/Slice_output_0">
470 <dim>1</dim>
471 <dim>-1</dim>
472 </port>
473 </output>
474 </layer>
475 <layer id="38" name="Constant_255207" type="Const" version="opset1">
476 <data element_type="i64" shape="" offset="18440" size="8" />
477 <rt_info>
478 <attribute name="fused_names" version="0" value="Constant_255207" />
479 </rt_info>
480 <output>
481 <port id="0" precision="I64" />
482 </output>
483 </layer>
484 <layer id="39" name="/distilbert/embeddings/position_embeddings/Gather" type="Gather" version="opset8">
485 <data batch_dims="0" />
486 <rt_info>
487 <attribute name="fused_names" version="0" value="/distilbert/embeddings/position_embeddings/Gather" />
488 </rt_info>
489 <input>
490 <port id="0" precision="FP32">
491 <dim>512</dim>
492 <dim>768</dim>
493 </port>
494 <port id="1" precision="I64">
495 <dim>1</dim>
496 <dim>-1</dim>
497 </port>
498 <port id="2" precision="I64" />
499 </input>
500 <output>
501 <port id="3" precision="FP32" names="/distilbert/embeddings/position_embeddings/Gather_output_0">
502 <dim>1</dim>
503 <dim>-1</dim>
504 <dim>768</dim>
505 </port>
506 </output>
507 </layer>
508 <layer id="40" name="/distilbert/embeddings/Add" type="Add" version="opset1">
509 <data auto_broadcast="numpy" />
510 <rt_info>
511 <attribute name="fused_names" version="0" value="/distilbert/embeddings/Add" />
512 </rt_info>
513 <input>
514 <port id="0" precision="FP32">
515 <dim>-1</dim>
516 <dim>-1</dim>
517 <dim>768</dim>
518 </port>
519 <port id="1" precision="FP32">
520 <dim>1</dim>
521 <dim>-1</dim>
522 <dim>768</dim>
523 </port>
524 </input>
525 <output>
526 <port id="2" precision="FP32" names="/distilbert/embeddings/Add_output_0">
527 <dim>-1</dim>
528 <dim>-1</dim>
529 <dim>768</dim>
530 </port>
531 </output>
532 </layer>
533 <layer id="41" name="Constant_255210" type="Const" version="opset1">
534 <data element_type="i64" shape="1" offset="90763288" size="8" />
535 <rt_info>
536 <attribute name="fused_names" version="0" value="Constant_255210" />
537 </rt_info>
538 <output>
539 <port id="0" precision="I64">
540 <dim>1</dim>
541 </port>
542 </output>
543 </layer>
544 <layer id="42" name="/distilbert/embeddings/LayerNorm/ReduceMean" type="ReduceMean" version="opset1">
545 <data keep_dims="true" />
546 <rt_info>
547 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/ReduceMean" />
548 </rt_info>
549 <input>
550 <port id="0" precision="FP32">
551 <dim>-1</dim>
552 <dim>-1</dim>
553 <dim>768</dim>
554 </port>
555 <port id="1" precision="I64">
556 <dim>1</dim>
557 </port>
558 </input>
559 <output>
560 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/ReduceMean_output_0">
561 <dim>-1</dim>
562 <dim>-1</dim>
563 <dim>1</dim>
564 </port>
565 </output>
566 </layer>
567 <layer id="43" name="/distilbert/embeddings/LayerNorm/Sub" type="Subtract" version="opset1">
568 <data auto_broadcast="numpy" />
569 <rt_info>
570 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Sub" />
571 </rt_info>
572 <input>
573 <port id="0" precision="FP32">
574 <dim>-1</dim>
575 <dim>-1</dim>
576 <dim>768</dim>
577 </port>
578 <port id="1" precision="FP32">
579 <dim>-1</dim>
580 <dim>-1</dim>
581 <dim>1</dim>
582 </port>
583 </input>
584 <output>
585 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/Sub_output_0">
586 <dim>-1</dim>
587 <dim>-1</dim>
588 <dim>768</dim>
589 </port>
590 </output>
591 </layer>
592 <layer id="44" name="/distilbert/embeddings/LayerNorm/Constant" type="Const" version="opset1">
593 <data element_type="f32" shape="" offset="90763296" size="4" />
594 <rt_info>
595 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Constant" />
596 </rt_info>
597 <output>
598 <port id="0" precision="FP32" names="/distilbert/embeddings/LayerNorm/Constant_output_0" />
599 </output>
600 </layer>
601 <layer id="45" name="/distilbert/embeddings/LayerNorm/Pow" type="Power" version="opset1">
602 <data auto_broadcast="numpy" />
603 <rt_info>
604 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Pow" />
605 </rt_info>
606 <input>
607 <port id="0" precision="FP32">
608 <dim>-1</dim>
609 <dim>-1</dim>
610 <dim>768</dim>
611 </port>
612 <port id="1" precision="FP32" />
613 </input>
614 <output>
615 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/Pow_output_0">
616 <dim>-1</dim>
617 <dim>-1</dim>
618 <dim>768</dim>
619 </port>
620 </output>
621 </layer>
622 <layer id="46" name="Constant_255215" type="Const" version="opset1">
623 <data element_type="i64" shape="1" offset="90763288" size="8" />
624 <rt_info>
625 <attribute name="fused_names" version="0" value="Constant_255215" />
626 </rt_info>
627 <output>
628 <port id="0" precision="I64">
629 <dim>1</dim>
630 </port>
631 </output>
632 </layer>
633 <layer id="47" name="/distilbert/embeddings/LayerNorm/ReduceMean_1" type="ReduceMean" version="opset1">
634 <data keep_dims="true" />
635 <rt_info>
636 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/ReduceMean_1" />
637 </rt_info>
638 <input>
639 <port id="0" precision="FP32">
640 <dim>-1</dim>
641 <dim>-1</dim>
642 <dim>768</dim>
643 </port>
644 <port id="1" precision="I64">
645 <dim>1</dim>
646 </port>
647 </input>
648 <output>
649 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/ReduceMean_1_output_0">
650 <dim>-1</dim>
651 <dim>-1</dim>
652 <dim>1</dim>
653 </port>
654 </output>
655 </layer>
656 <layer id="48" name="/distilbert/embeddings/LayerNorm/Constant_1" type="Const" version="opset1">
657 <data element_type="f32" shape="" offset="90763300" size="4" />
658 <rt_info>
659 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Constant_1" />
660 </rt_info>
661 <output>
662 <port id="0" precision="FP32" names="/distilbert/embeddings/LayerNorm/Constant_1_output_0" />
663 </output>
664 </layer>
665 <layer id="49" name="/distilbert/embeddings/LayerNorm/Add" type="Add" version="opset1">
666 <data auto_broadcast="numpy" />
667 <rt_info>
668 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Add" />
669 </rt_info>
670 <input>
671 <port id="0" precision="FP32">
672 <dim>-1</dim>
673 <dim>-1</dim>
674 <dim>1</dim>
675 </port>
676 <port id="1" precision="FP32" />
677 </input>
678 <output>
679 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/Add_output_0">
680 <dim>-1</dim>
681 <dim>-1</dim>
682 <dim>1</dim>
683 </port>
684 </output>
685 </layer>
686 <layer id="50" name="/distilbert/embeddings/LayerNorm/Sqrt" type="Sqrt" version="opset1">
687 <rt_info>
688 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Sqrt" />
689 </rt_info>
690 <input>
691 <port id="0" precision="FP32">
692 <dim>-1</dim>
693 <dim>-1</dim>
694 <dim>1</dim>
695 </port>
696 </input>
697 <output>
698 <port id="1" precision="FP32" names="/distilbert/embeddings/LayerNorm/Sqrt_output_0">
699 <dim>-1</dim>
700 <dim>-1</dim>
701 <dim>1</dim>
702 </port>
703 </output>
704 </layer>
705 <layer id="51" name="/distilbert/embeddings/LayerNorm/Div" type="Divide" version="opset1">
706 <data auto_broadcast="numpy" m_pythondiv="true" />
707 <rt_info>
708 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Div" />
709 </rt_info>
710 <input>
711 <port id="0" precision="FP32">
712 <dim>-1</dim>
713 <dim>-1</dim>
714 <dim>768</dim>
715 </port>
716 <port id="1" precision="FP32">
717 <dim>-1</dim>
718 <dim>-1</dim>
719 <dim>1</dim>
720 </port>
721 </input>
722 <output>
723 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/Div_output_0">
724 <dim>-1</dim>
725 <dim>-1</dim>
726 <dim>768</dim>
727 </port>
728 </output>
729 </layer>
730 <layer id="52" name="distilbert.embeddings.LayerNorm.weight" type="Const" version="opset1">
731 <data element_type="f32" shape="768" offset="90763304" size="3072" />
732 <rt_info>
733 <attribute name="fused_names" version="0" value="distilbert.embeddings.LayerNorm.weight" />
734 </rt_info>
735 <output>
736 <port id="0" precision="FP32" names="distilbert.embeddings.LayerNorm.weight">
737 <dim>768</dim>
738 </port>
739 </output>
740 </layer>
741 <layer id="53" name="/distilbert/embeddings/LayerNorm/Mul" type="Multiply" version="opset1">
742 <data auto_broadcast="numpy" />
743 <rt_info>
744 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Mul" />
745 </rt_info>
746 <input>
747 <port id="0" precision="FP32">
748 <dim>-1</dim>
749 <dim>-1</dim>
750 <dim>768</dim>
751 </port>
752 <port id="1" precision="FP32">
753 <dim>768</dim>
754 </port>
755 </input>
756 <output>
757 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/Mul_output_0">
758 <dim>-1</dim>
759 <dim>-1</dim>
760 <dim>768</dim>
761 </port>
762 </output>
763 </layer>
764 <layer id="54" name="distilbert.embeddings.LayerNorm.bias" type="Const" version="opset1">
765 <data element_type="f32" shape="768" offset="90766376" size="3072" />
766 <rt_info>
767 <attribute name="fused_names" version="0" value="distilbert.embeddings.LayerNorm.bias" />
768 </rt_info>
769 <output>
770 <port id="0" precision="FP32" names="distilbert.embeddings.LayerNorm.bias">
771 <dim>768</dim>
772 </port>
773 </output>
774 </layer>
775 <layer id="55" name="/distilbert/embeddings/LayerNorm/Add_1" type="Add" version="opset1">
776 <data auto_broadcast="numpy" />
777 <rt_info>
778 <attribute name="fused_names" version="0" value="/distilbert/embeddings/LayerNorm/Add_1" />
779 </rt_info>
780 <input>
781 <port id="0" precision="FP32">
782 <dim>-1</dim>
783 <dim>-1</dim>
784 <dim>768</dim>
785 </port>
786 <port id="1" precision="FP32">
787 <dim>768</dim>
788 </port>
789 </input>
790 <output>
791 <port id="2" precision="FP32" names="/distilbert/embeddings/LayerNorm/Add_1_output_0">
792 <dim>-1</dim>
793 <dim>-1</dim>
794 <dim>768</dim>
795 </port>
796 </output>
797 </layer>
798 <layer id="56" name="/distilbert/transformer/layer.0/attention/Shape" type="ShapeOf" version="opset3">
799 <data output_type="i64" />
800 <rt_info>
801 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Shape" />
802 </rt_info>
803 <input>
804 <port id="0" precision="FP32">
805 <dim>-1</dim>
806 <dim>-1</dim>
807 <dim>768</dim>
808 </port>
809 </input>
810 <output>
811 <port id="1" precision="I64" names="/distilbert/transformer/layer.0/attention/Shape_output_0">
812 <dim>3</dim>
813 </port>
814 </output>
815 </layer>
816 <layer id="57" name="/distilbert/transformer/layer.0/attention/Constant" type="Const" version="opset1">
817 <data element_type="i64" shape="" offset="18440" size="8" />
818 <rt_info>
819 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant" />
820 </rt_info>
821 <output>
822 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_output_0" />
823 </output>
824 </layer>
825 <layer id="58" name="Constant_255225" type="Const" version="opset1">
826 <data element_type="i64" shape="" offset="18440" size="8" />
827 <rt_info>
828 <attribute name="fused_names" version="0" value="Constant_255225" />
829 </rt_info>
830 <output>
831 <port id="0" precision="I64" />
832 </output>
833 </layer>
834 <layer id="59" name="/distilbert/transformer/layer.0/attention/Gather" type="Gather" version="opset8">
835 <data batch_dims="0" />
836 <rt_info>
837 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant, /distilbert/transformer/layer.0/attention/Gather, Constant_255225" />
838 </rt_info>
839 <input>
840 <port id="0" precision="I64">
841 <dim>3</dim>
842 </port>
843 <port id="1" precision="I64" />
844 <port id="2" precision="I64" />
845 </input>
846 <output>
847 <port id="3" precision="I64" names="/distilbert/transformer/layer.0/attention/Gather_output_0" />
848 </output>
849 </layer>
850 <layer id="60" name="Constant_255287" type="Const" version="opset1">
851 <data element_type="i64" shape="1" offset="18440" size="8" />
852 <rt_info>
853 <attribute name="fused_names" version="0" value="Constant_255287" />
854 </rt_info>
855 <output>
856 <port id="0" precision="I64">
857 <dim>1</dim>
858 </port>
859 </output>
860 </layer>
861 <layer id="61" name="/distilbert/transformer/layer.0/attention/Unsqueeze_3" type="Unsqueeze" version="opset1">
862 <rt_info>
863 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Unsqueeze_3, Constant_255287" />
864 </rt_info>
865 <input>
866 <port id="0" precision="I64" />
867 <port id="1" precision="I64">
868 <dim>1</dim>
869 </port>
870 </input>
871 <output>
872 <port id="2" precision="I64" names="/distilbert/transformer/layer.0/attention/Unsqueeze_3_output_0">
873 <dim>1</dim>
874 </port>
875 </output>
876 </layer>
877 <layer id="62" name="/distilbert/transformer/layer.0/attention/Constant_13" type="Const" version="opset1">
878 <data element_type="i64" shape="1" offset="90763280" size="8" />
879 <rt_info>
880 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_13" />
881 </rt_info>
882 <output>
883 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_13_output_0">
884 <dim>1</dim>
885 </port>
886 </output>
887 </layer>
888 <layer id="63" name="/distilbert/transformer/layer.0/attention/Constant_14" type="Const" version="opset1">
889 <data element_type="i64" shape="1" offset="90763280" size="8" />
890 <rt_info>
891 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_14" />
892 </rt_info>
893 <output>
894 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_14_output_0">
895 <dim>1</dim>
896 </port>
897 </output>
898 </layer>
899 <layer id="64" name="/distilbert/transformer/layer.0/attention/Shape_1" type="ShapeOf" version="opset3">
900 <data output_type="i64" />
901 <rt_info>
902 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Shape_1" />
903 </rt_info>
904 <input>
905 <port id="0" precision="FP32">
906 <dim>-1</dim>
907 <dim>-1</dim>
908 <dim>768</dim>
909 </port>
910 </input>
911 <output>
912 <port id="1" precision="I64" names="/distilbert/transformer/layer.0/attention/Shape_1_output_0">
913 <dim>3</dim>
914 </port>
915 </output>
916 </layer>
917 <layer id="65" name="/distilbert/transformer/layer.0/attention/Constant_1" type="Const" version="opset1">
918 <data element_type="i64" shape="" offset="90763280" size="8" />
919 <rt_info>
920 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_1" />
921 </rt_info>
922 <output>
923 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_1_output_0" />
924 </output>
925 </layer>
926 <layer id="66" name="Constant_255229" type="Const" version="opset1">
927 <data element_type="i64" shape="" offset="18440" size="8" />
928 <rt_info>
929 <attribute name="fused_names" version="0" value="Constant_255229" />
930 </rt_info>
931 <output>
932 <port id="0" precision="I64" />
933 </output>
934 </layer>
935 <layer id="67" name="/distilbert/transformer/layer.0/attention/Gather_1" type="Gather" version="opset8">
936 <data batch_dims="0" />
937 <rt_info>
938 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_1, /distilbert/transformer/layer.0/attention/Gather_1, Constant_255229" />
939 </rt_info>
940 <input>
941 <port id="0" precision="I64">
942 <dim>3</dim>
943 </port>
944 <port id="1" precision="I64" />
945 <port id="2" precision="I64" />
946 </input>
947 <output>
948 <port id="3" precision="I64" names="/distilbert/transformer/layer.0/attention/Gather_1_output_0" />
949 </output>
950 </layer>
951 <layer id="68" name="Constant_255291" type="Const" version="opset1">
952 <data element_type="i64" shape="1" offset="18440" size="8" />
953 <rt_info>
954 <attribute name="fused_names" version="0" value="Constant_255291" />
955 </rt_info>
956 <output>
957 <port id="0" precision="I64">
958 <dim>1</dim>
959 </port>
960 </output>
961 </layer>
962 <layer id="69" name="/distilbert/transformer/layer.0/attention/Unsqueeze_4" type="Unsqueeze" version="opset1">
963 <rt_info>
964 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Unsqueeze_4, Constant_255291" />
965 </rt_info>
966 <input>
967 <port id="0" precision="I64" />
968 <port id="1" precision="I64">
969 <dim>1</dim>
970 </port>
971 </input>
972 <output>
973 <port id="2" precision="I64" names="/distilbert/transformer/layer.0/attention/Unsqueeze_4_output_0">
974 <dim>1</dim>
975 </port>
976 </output>
977 </layer>
978 <layer id="70" name="/distilbert/transformer/layer.0/attention/Concat_3" type="Concat" version="opset1">
979 <data axis="0" />
980 <rt_info>
981 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Concat_3, /distilbert/transformer/layer.0/attention/Constant_13, /distilbert/transformer/layer.0/attention/Constant_14" />
982 </rt_info>
983 <input>
984 <port id="0" precision="I64">
985 <dim>1</dim>
986 </port>
987 <port id="1" precision="I64">
988 <dim>1</dim>
989 </port>
990 <port id="2" precision="I64">
991 <dim>1</dim>
992 </port>
993 <port id="3" precision="I64">
994 <dim>1</dim>
995 </port>
996 </input>
997 <output>
998 <port id="4" precision="I64" names="/distilbert/transformer/layer.0/attention/Concat_3_output_0">
999 <dim>4</dim>
1000 </port>
1001 </output>
1002 </layer>
1003 <layer id="71" name="/distilbert/transformer/layer.0/attention/Reshape_3" type="Reshape" version="opset1">
1004 <data special_zero="true" />
1005 <rt_info>
1006 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Reshape_3" />
1007 </rt_info>
1008 <input>
1009 <port id="0" precision="BOOL">
1010 <dim>-1</dim>
1011 <dim>-1</dim>
1012 </port>
1013 <port id="1" precision="I64">
1014 <dim>4</dim>
1015 </port>
1016 </input>
1017 <output>
1018 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.0/attention/Reshape_3_output_0">
1019 <dim>-1</dim>
1020 <dim>1</dim>
1021 <dim>1</dim>
1022 <dim>-1</dim>
1023 </port>
1024 </output>
1025 </layer>
1026 <layer id="72" name="distilbert.transformer.layer.0.attention.q_lin.bias" type="Const" version="opset1">
1027 <data element_type="f32" shape="768" offset="90769448" size="3072" />
1028 <rt_info>
1029 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.attention.q_lin.bias" />
1030 </rt_info>
1031 <output>
1032 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.attention.q_lin.bias">
1033 <dim>768</dim>
1034 </port>
1035 </output>
1036 </layer>
1037 <layer id="73" name="onnx::MatMul_738" type="Const" version="opset1">
1038 <data element_type="f32" shape="768, 768" offset="90772520" size="2359296" />
1039 <rt_info>
1040 <attribute name="fused_names" version="0" value="onnx::MatMul_738" />
1041 </rt_info>
1042 <output>
1043 <port id="0" precision="FP32" names="onnx::MatMul_738">
1044 <dim>768</dim>
1045 <dim>768</dim>
1046 </port>
1047 </output>
1048 </layer>
1049 <layer id="74" name="/distilbert/transformer/layer.0/attention/q_lin/MatMul" type="MatMul" version="opset1">
1050 <data transpose_a="false" transpose_b="false" />
1051 <rt_info>
1052 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/q_lin/MatMul" />
1053 </rt_info>
1054 <input>
1055 <port id="0" precision="FP32">
1056 <dim>-1</dim>
1057 <dim>-1</dim>
1058 <dim>768</dim>
1059 </port>
1060 <port id="1" precision="FP32">
1061 <dim>768</dim>
1062 <dim>768</dim>
1063 </port>
1064 </input>
1065 <output>
1066 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/q_lin/MatMul_output_0">
1067 <dim>-1</dim>
1068 <dim>-1</dim>
1069 <dim>768</dim>
1070 </port>
1071 </output>
1072 </layer>
1073 <layer id="75" name="/distilbert/transformer/layer.0/attention/q_lin/Add" type="Add" version="opset1">
1074 <data auto_broadcast="numpy" />
1075 <rt_info>
1076 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/q_lin/Add" />
1077 </rt_info>
1078 <input>
1079 <port id="0" precision="FP32">
1080 <dim>768</dim>
1081 </port>
1082 <port id="1" precision="FP32">
1083 <dim>-1</dim>
1084 <dim>-1</dim>
1085 <dim>768</dim>
1086 </port>
1087 </input>
1088 <output>
1089 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/q_lin/Add_output_0">
1090 <dim>-1</dim>
1091 <dim>-1</dim>
1092 <dim>768</dim>
1093 </port>
1094 </output>
1095 </layer>
1096 <layer id="76" name="Constant_255233" type="Const" version="opset1">
1097 <data element_type="i64" shape="1" offset="18440" size="8" />
1098 <rt_info>
1099 <attribute name="fused_names" version="0" value="Constant_255233" />
1100 </rt_info>
1101 <output>
1102 <port id="0" precision="I64">
1103 <dim>1</dim>
1104 </port>
1105 </output>
1106 </layer>
1107 <layer id="77" name="/distilbert/transformer/layer.0/attention/Unsqueeze" type="Unsqueeze" version="opset1">
1108 <rt_info>
1109 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Unsqueeze, Constant_255233" />
1110 </rt_info>
1111 <input>
1112 <port id="0" precision="I64" />
1113 <port id="1" precision="I64">
1114 <dim>1</dim>
1115 </port>
1116 </input>
1117 <output>
1118 <port id="2" precision="I64" names="/distilbert/transformer/layer.0/attention/Unsqueeze_output_0">
1119 <dim>1</dim>
1120 </port>
1121 </output>
1122 </layer>
1123 <layer id="78" name="/distilbert/transformer/layer.0/attention/Constant_2" type="Const" version="opset1">
1124 <data element_type="i64" shape="1" offset="90763288" size="8" />
1125 <rt_info>
1126 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_2" />
1127 </rt_info>
1128 <output>
1129 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_2_output_0">
1130 <dim>1</dim>
1131 </port>
1132 </output>
1133 </layer>
1134 <layer id="79" name="/distilbert/transformer/layer.0/attention/Constant_3" type="Const" version="opset1">
1135 <data element_type="i64" shape="1" offset="93131816" size="8" />
1136 <rt_info>
1137 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_3" />
1138 </rt_info>
1139 <output>
1140 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_3_output_0">
1141 <dim>1</dim>
1142 </port>
1143 </output>
1144 </layer>
1145 <layer id="80" name="/distilbert/transformer/layer.0/attention/Constant_4" type="Const" version="opset1">
1146 <data element_type="i64" shape="1" offset="93131824" size="8" />
1147 <rt_info>
1148 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_4" />
1149 </rt_info>
1150 <output>
1151 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_4_output_0">
1152 <dim>1</dim>
1153 </port>
1154 </output>
1155 </layer>
1156 <layer id="81" name="/distilbert/transformer/layer.0/attention/Concat" type="Concat" version="opset1">
1157 <data axis="0" />
1158 <rt_info>
1159 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Concat, /distilbert/transformer/layer.0/attention/Constant_2, /distilbert/transformer/layer.0/attention/Constant_3, /distilbert/transformer/layer.0/attention/Constant_4" />
1160 </rt_info>
1161 <input>
1162 <port id="0" precision="I64">
1163 <dim>1</dim>
1164 </port>
1165 <port id="1" precision="I64">
1166 <dim>1</dim>
1167 </port>
1168 <port id="2" precision="I64">
1169 <dim>1</dim>
1170 </port>
1171 <port id="3" precision="I64">
1172 <dim>1</dim>
1173 </port>
1174 </input>
1175 <output>
1176 <port id="4" precision="I64" names="/distilbert/transformer/layer.0/attention/Concat_output_0">
1177 <dim>4</dim>
1178 </port>
1179 </output>
1180 </layer>
1181 <layer id="82" name="/distilbert/transformer/layer.0/attention/Reshape" type="Reshape" version="opset1">
1182 <data special_zero="true" />
1183 <rt_info>
1184 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Reshape" />
1185 </rt_info>
1186 <input>
1187 <port id="0" precision="FP32">
1188 <dim>-1</dim>
1189 <dim>-1</dim>
1190 <dim>768</dim>
1191 </port>
1192 <port id="1" precision="I64">
1193 <dim>4</dim>
1194 </port>
1195 </input>
1196 <output>
1197 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Reshape_output_0">
1198 <dim>-1</dim>
1199 <dim>-1</dim>
1200 <dim>12</dim>
1201 <dim>64</dim>
1202 </port>
1203 </output>
1204 </layer>
1205 <layer id="83" name="Constant_255258" type="Const" version="opset1">
1206 <data element_type="i64" shape="4" offset="93131832" size="32" />
1207 <rt_info>
1208 <attribute name="fused_names" version="0" value="Constant_255258" />
1209 </rt_info>
1210 <output>
1211 <port id="0" precision="I64">
1212 <dim>4</dim>
1213 </port>
1214 </output>
1215 </layer>
1216 <layer id="84" name="/distilbert/transformer/layer.0/attention/Transpose" type="Transpose" version="opset1">
1217 <rt_info>
1218 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Transpose" />
1219 </rt_info>
1220 <input>
1221 <port id="0" precision="FP32">
1222 <dim>-1</dim>
1223 <dim>-1</dim>
1224 <dim>12</dim>
1225 <dim>64</dim>
1226 </port>
1227 <port id="1" precision="I64">
1228 <dim>4</dim>
1229 </port>
1230 </input>
1231 <output>
1232 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Transpose_output_0">
1233 <dim>-1</dim>
1234 <dim>12</dim>
1235 <dim>-1</dim>
1236 <dim>64</dim>
1237 </port>
1238 </output>
1239 </layer>
1240 <layer id="85" name="/distilbert/transformer/layer.0/attention/Constant_11" type="Const" version="opset1">
1241 <data element_type="f32" shape="" offset="93131864" size="4" />
1242 <rt_info>
1243 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_11" />
1244 </rt_info>
1245 <output>
1246 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/attention/Constant_11_output_0" />
1247 </output>
1248 </layer>
1249 <layer id="86" name="/distilbert/transformer/layer.0/attention/Div" type="Divide" version="opset1">
1250 <data auto_broadcast="numpy" m_pythondiv="true" />
1251 <rt_info>
1252 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Div" />
1253 </rt_info>
1254 <input>
1255 <port id="0" precision="FP32">
1256 <dim>-1</dim>
1257 <dim>12</dim>
1258 <dim>-1</dim>
1259 <dim>64</dim>
1260 </port>
1261 <port id="1" precision="FP32" />
1262 </input>
1263 <output>
1264 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Div_output_0">
1265 <dim>-1</dim>
1266 <dim>12</dim>
1267 <dim>-1</dim>
1268 <dim>64</dim>
1269 </port>
1270 </output>
1271 </layer>
1272 <layer id="87" name="distilbert.transformer.layer.0.attention.k_lin.bias" type="Const" version="opset1">
1273 <data element_type="f32" shape="768" offset="93131868" size="3072" />
1274 <rt_info>
1275 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.attention.k_lin.bias" />
1276 </rt_info>
1277 <output>
1278 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.attention.k_lin.bias">
1279 <dim>768</dim>
1280 </port>
1281 </output>
1282 </layer>
1283 <layer id="88" name="onnx::MatMul_748" type="Const" version="opset1">
1284 <data element_type="f32" shape="768, 768" offset="93134940" size="2359296" />
1285 <rt_info>
1286 <attribute name="fused_names" version="0" value="onnx::MatMul_748" />
1287 </rt_info>
1288 <output>
1289 <port id="0" precision="FP32" names="onnx::MatMul_748">
1290 <dim>768</dim>
1291 <dim>768</dim>
1292 </port>
1293 </output>
1294 </layer>
1295 <layer id="89" name="/distilbert/transformer/layer.0/attention/k_lin/MatMul" type="MatMul" version="opset1">
1296 <data transpose_a="false" transpose_b="false" />
1297 <rt_info>
1298 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/k_lin/MatMul" />
1299 </rt_info>
1300 <input>
1301 <port id="0" precision="FP32">
1302 <dim>-1</dim>
1303 <dim>-1</dim>
1304 <dim>768</dim>
1305 </port>
1306 <port id="1" precision="FP32">
1307 <dim>768</dim>
1308 <dim>768</dim>
1309 </port>
1310 </input>
1311 <output>
1312 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/k_lin/MatMul_output_0">
1313 <dim>-1</dim>
1314 <dim>-1</dim>
1315 <dim>768</dim>
1316 </port>
1317 </output>
1318 </layer>
1319 <layer id="90" name="/distilbert/transformer/layer.0/attention/k_lin/Add" type="Add" version="opset1">
1320 <data auto_broadcast="numpy" />
1321 <rt_info>
1322 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/k_lin/Add" />
1323 </rt_info>
1324 <input>
1325 <port id="0" precision="FP32">
1326 <dim>768</dim>
1327 </port>
1328 <port id="1" precision="FP32">
1329 <dim>-1</dim>
1330 <dim>-1</dim>
1331 <dim>768</dim>
1332 </port>
1333 </input>
1334 <output>
1335 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/k_lin/Add_output_0">
1336 <dim>-1</dim>
1337 <dim>-1</dim>
1338 <dim>768</dim>
1339 </port>
1340 </output>
1341 </layer>
1342 <layer id="91" name="Constant_255239" type="Const" version="opset1">
1343 <data element_type="i64" shape="1" offset="18440" size="8" />
1344 <rt_info>
1345 <attribute name="fused_names" version="0" value="Constant_255239" />
1346 </rt_info>
1347 <output>
1348 <port id="0" precision="I64">
1349 <dim>1</dim>
1350 </port>
1351 </output>
1352 </layer>
1353 <layer id="92" name="/distilbert/transformer/layer.0/attention/Unsqueeze_1" type="Unsqueeze" version="opset1">
1354 <rt_info>
1355 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Unsqueeze_1, Constant_255239" />
1356 </rt_info>
1357 <input>
1358 <port id="0" precision="I64" />
1359 <port id="1" precision="I64">
1360 <dim>1</dim>
1361 </port>
1362 </input>
1363 <output>
1364 <port id="2" precision="I64" names="/distilbert/transformer/layer.0/attention/Unsqueeze_1_output_0">
1365 <dim>1</dim>
1366 </port>
1367 </output>
1368 </layer>
1369 <layer id="93" name="/distilbert/transformer/layer.0/attention/Constant_5" type="Const" version="opset1">
1370 <data element_type="i64" shape="1" offset="90763288" size="8" />
1371 <rt_info>
1372 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_5" />
1373 </rt_info>
1374 <output>
1375 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_5_output_0">
1376 <dim>1</dim>
1377 </port>
1378 </output>
1379 </layer>
1380 <layer id="94" name="/distilbert/transformer/layer.0/attention/Constant_6" type="Const" version="opset1">
1381 <data element_type="i64" shape="1" offset="93131816" size="8" />
1382 <rt_info>
1383 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_6" />
1384 </rt_info>
1385 <output>
1386 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_6_output_0">
1387 <dim>1</dim>
1388 </port>
1389 </output>
1390 </layer>
1391 <layer id="95" name="/distilbert/transformer/layer.0/attention/Constant_7" type="Const" version="opset1">
1392 <data element_type="i64" shape="1" offset="93131824" size="8" />
1393 <rt_info>
1394 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_7" />
1395 </rt_info>
1396 <output>
1397 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_7_output_0">
1398 <dim>1</dim>
1399 </port>
1400 </output>
1401 </layer>
1402 <layer id="96" name="/distilbert/transformer/layer.0/attention/Concat_1" type="Concat" version="opset1">
1403 <data axis="0" />
1404 <rt_info>
1405 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Concat_1, /distilbert/transformer/layer.0/attention/Constant_5, /distilbert/transformer/layer.0/attention/Constant_6, /distilbert/transformer/layer.0/attention/Constant_7" />
1406 </rt_info>
1407 <input>
1408 <port id="0" precision="I64">
1409 <dim>1</dim>
1410 </port>
1411 <port id="1" precision="I64">
1412 <dim>1</dim>
1413 </port>
1414 <port id="2" precision="I64">
1415 <dim>1</dim>
1416 </port>
1417 <port id="3" precision="I64">
1418 <dim>1</dim>
1419 </port>
1420 </input>
1421 <output>
1422 <port id="4" precision="I64" names="/distilbert/transformer/layer.0/attention/Concat_1_output_0">
1423 <dim>4</dim>
1424 </port>
1425 </output>
1426 </layer>
1427 <layer id="97" name="/distilbert/transformer/layer.0/attention/Reshape_1" type="Reshape" version="opset1">
1428 <data special_zero="true" />
1429 <rt_info>
1430 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Reshape_1" />
1431 </rt_info>
1432 <input>
1433 <port id="0" precision="FP32">
1434 <dim>-1</dim>
1435 <dim>-1</dim>
1436 <dim>768</dim>
1437 </port>
1438 <port id="1" precision="I64">
1439 <dim>4</dim>
1440 </port>
1441 </input>
1442 <output>
1443 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Reshape_1_output_0">
1444 <dim>-1</dim>
1445 <dim>-1</dim>
1446 <dim>12</dim>
1447 <dim>64</dim>
1448 </port>
1449 </output>
1450 </layer>
1451 <layer id="98" name="Constant_255282" type="Const" version="opset1">
1452 <data element_type="i64" shape="4" offset="95494236" size="32" />
1453 <rt_info>
1454 <attribute name="fused_names" version="0" value="Constant_255282" />
1455 </rt_info>
1456 <output>
1457 <port id="0" precision="I64">
1458 <dim>4</dim>
1459 </port>
1460 </output>
1461 </layer>
1462 <layer id="99" name="/distilbert/transformer/layer.0/attention/Transpose_2" type="Transpose" version="opset1">
1463 <rt_info>
1464 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Transpose_2" />
1465 </rt_info>
1466 <input>
1467 <port id="0" precision="FP32">
1468 <dim>-1</dim>
1469 <dim>-1</dim>
1470 <dim>12</dim>
1471 <dim>64</dim>
1472 </port>
1473 <port id="1" precision="I64">
1474 <dim>4</dim>
1475 </port>
1476 </input>
1477 <output>
1478 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Transpose_2_output_0">
1479 <dim>-1</dim>
1480 <dim>12</dim>
1481 <dim>64</dim>
1482 <dim>-1</dim>
1483 </port>
1484 </output>
1485 </layer>
1486 <layer id="100" name="/distilbert/transformer/layer.0/attention/MatMul" type="MatMul" version="opset1">
1487 <data transpose_a="false" transpose_b="false" />
1488 <rt_info>
1489 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/MatMul" />
1490 </rt_info>
1491 <input>
1492 <port id="0" precision="FP32">
1493 <dim>-1</dim>
1494 <dim>12</dim>
1495 <dim>-1</dim>
1496 <dim>64</dim>
1497 </port>
1498 <port id="1" precision="FP32">
1499 <dim>-1</dim>
1500 <dim>12</dim>
1501 <dim>64</dim>
1502 <dim>-1</dim>
1503 </port>
1504 </input>
1505 <output>
1506 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/MatMul_output_0">
1507 <dim>-1</dim>
1508 <dim>12</dim>
1509 <dim>-1</dim>
1510 <dim>-1</dim>
1511 </port>
1512 </output>
1513 </layer>
1514 <layer id="101" name="/distilbert/transformer/layer.0/attention/Shape_2" type="ShapeOf" version="opset3">
1515 <data output_type="i64" />
1516 <rt_info>
1517 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Shape_2" />
1518 </rt_info>
1519 <input>
1520 <port id="0" precision="FP32">
1521 <dim>-1</dim>
1522 <dim>12</dim>
1523 <dim>-1</dim>
1524 <dim>-1</dim>
1525 </port>
1526 </input>
1527 <output>
1528 <port id="1" precision="I64" names="/distilbert/transformer/layer.0/attention/Shape_2_output_0">
1529 <dim>4</dim>
1530 </port>
1531 </output>
1532 </layer>
1533 <layer id="102" name="/distilbert/transformer/layer.0/attention/Expand" type="Broadcast" version="opset3">
1534 <data mode="bidirectional" />
1535 <rt_info>
1536 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Expand" />
1537 </rt_info>
1538 <input>
1539 <port id="0" precision="BOOL">
1540 <dim>-1</dim>
1541 <dim>1</dim>
1542 <dim>1</dim>
1543 <dim>-1</dim>
1544 </port>
1545 <port id="1" precision="I64">
1546 <dim>4</dim>
1547 </port>
1548 </input>
1549 <output>
1550 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.0/attention/Expand_output_0">
1551 <dim>-1</dim>
1552 <dim>12</dim>
1553 <dim>-1</dim>
1554 <dim>-1</dim>
1555 </port>
1556 </output>
1557 </layer>
1558 <layer id="103" name="/distilbert/transformer/layer.0/attention/Cast" type="Convert" version="opset1">
1559 <data destination_type="boolean" />
1560 <rt_info>
1561 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Cast" />
1562 </rt_info>
1563 <input>
1564 <port id="0" precision="BOOL">
1565 <dim>-1</dim>
1566 <dim>12</dim>
1567 <dim>-1</dim>
1568 <dim>-1</dim>
1569 </port>
1570 </input>
1571 <output>
1572 <port id="1" precision="BOOL" names="/distilbert/transformer/layer.0/attention/Cast_output_0">
1573 <dim>-1</dim>
1574 <dim>12</dim>
1575 <dim>-1</dim>
1576 <dim>-1</dim>
1577 </port>
1578 </output>
1579 </layer>
1580 <layer id="104" name="/distilbert/transformer/layer.0/attention/Constant_15" type="Const" version="opset1">
1581 <data element_type="f32" shape="" offset="95494268" size="4" />
1582 <rt_info>
1583 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_15" />
1584 </rt_info>
1585 <output>
1586 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/attention/Constant_15_output_0" />
1587 </output>
1588 </layer>
1589 <layer id="105" name="/distilbert/transformer/layer.0/attention/Where" type="Select" version="opset1">
1590 <data auto_broadcast="numpy" />
1591 <rt_info>
1592 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Where" />
1593 </rt_info>
1594 <input>
1595 <port id="0" precision="BOOL">
1596 <dim>-1</dim>
1597 <dim>12</dim>
1598 <dim>-1</dim>
1599 <dim>-1</dim>
1600 </port>
1601 <port id="1" precision="FP32" />
1602 <port id="2" precision="FP32">
1603 <dim>-1</dim>
1604 <dim>12</dim>
1605 <dim>-1</dim>
1606 <dim>-1</dim>
1607 </port>
1608 </input>
1609 <output>
1610 <port id="3" precision="FP32" names="/distilbert/transformer/layer.0/attention/Where_output_0">
1611 <dim>-1</dim>
1612 <dim>12</dim>
1613 <dim>-1</dim>
1614 <dim>-1</dim>
1615 </port>
1616 </output>
1617 </layer>
1618 <layer id="106" name="/distilbert/transformer/layer.0/attention/Softmax" type="SoftMax" version="opset8">
1619 <data axis="3" />
1620 <rt_info>
1621 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Softmax" />
1622 </rt_info>
1623 <input>
1624 <port id="0" precision="FP32">
1625 <dim>-1</dim>
1626 <dim>12</dim>
1627 <dim>-1</dim>
1628 <dim>-1</dim>
1629 </port>
1630 </input>
1631 <output>
1632 <port id="1" precision="FP32" names="/distilbert/transformer/layer.0/attention/Softmax_output_0">
1633 <dim>-1</dim>
1634 <dim>12</dim>
1635 <dim>-1</dim>
1636 <dim>-1</dim>
1637 </port>
1638 </output>
1639 </layer>
1640 <layer id="107" name="distilbert.transformer.layer.0.attention.v_lin.bias" type="Const" version="opset1">
1641 <data element_type="f32" shape="768" offset="95494272" size="3072" />
1642 <rt_info>
1643 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.attention.v_lin.bias" />
1644 </rt_info>
1645 <output>
1646 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.attention.v_lin.bias">
1647 <dim>768</dim>
1648 </port>
1649 </output>
1650 </layer>
1651 <layer id="108" name="onnx::MatMul_749" type="Const" version="opset1">
1652 <data element_type="f32" shape="768, 768" offset="95497344" size="2359296" />
1653 <rt_info>
1654 <attribute name="fused_names" version="0" value="onnx::MatMul_749" />
1655 </rt_info>
1656 <output>
1657 <port id="0" precision="FP32" names="onnx::MatMul_749">
1658 <dim>768</dim>
1659 <dim>768</dim>
1660 </port>
1661 </output>
1662 </layer>
1663 <layer id="109" name="/distilbert/transformer/layer.0/attention/v_lin/MatMul" type="MatMul" version="opset1">
1664 <data transpose_a="false" transpose_b="false" />
1665 <rt_info>
1666 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/v_lin/MatMul" />
1667 </rt_info>
1668 <input>
1669 <port id="0" precision="FP32">
1670 <dim>-1</dim>
1671 <dim>-1</dim>
1672 <dim>768</dim>
1673 </port>
1674 <port id="1" precision="FP32">
1675 <dim>768</dim>
1676 <dim>768</dim>
1677 </port>
1678 </input>
1679 <output>
1680 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/v_lin/MatMul_output_0">
1681 <dim>-1</dim>
1682 <dim>-1</dim>
1683 <dim>768</dim>
1684 </port>
1685 </output>
1686 </layer>
1687 <layer id="110" name="/distilbert/transformer/layer.0/attention/v_lin/Add" type="Add" version="opset1">
1688 <data auto_broadcast="numpy" />
1689 <rt_info>
1690 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/v_lin/Add" />
1691 </rt_info>
1692 <input>
1693 <port id="0" precision="FP32">
1694 <dim>768</dim>
1695 </port>
1696 <port id="1" precision="FP32">
1697 <dim>-1</dim>
1698 <dim>-1</dim>
1699 <dim>768</dim>
1700 </port>
1701 </input>
1702 <output>
1703 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/v_lin/Add_output_0">
1704 <dim>-1</dim>
1705 <dim>-1</dim>
1706 <dim>768</dim>
1707 </port>
1708 </output>
1709 </layer>
1710 <layer id="111" name="Constant_255245" type="Const" version="opset1">
1711 <data element_type="i64" shape="1" offset="18440" size="8" />
1712 <rt_info>
1713 <attribute name="fused_names" version="0" value="Constant_255245" />
1714 </rt_info>
1715 <output>
1716 <port id="0" precision="I64">
1717 <dim>1</dim>
1718 </port>
1719 </output>
1720 </layer>
1721 <layer id="112" name="/distilbert/transformer/layer.0/attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
1722 <rt_info>
1723 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Unsqueeze_2, Constant_255245" />
1724 </rt_info>
1725 <input>
1726 <port id="0" precision="I64" />
1727 <port id="1" precision="I64">
1728 <dim>1</dim>
1729 </port>
1730 </input>
1731 <output>
1732 <port id="2" precision="I64" names="/distilbert/transformer/layer.0/attention/Unsqueeze_2_output_0">
1733 <dim>1</dim>
1734 </port>
1735 </output>
1736 </layer>
1737 <layer id="113" name="/distilbert/transformer/layer.0/attention/Constant_8" type="Const" version="opset1">
1738 <data element_type="i64" shape="1" offset="90763288" size="8" />
1739 <rt_info>
1740 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_8" />
1741 </rt_info>
1742 <output>
1743 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_8_output_0">
1744 <dim>1</dim>
1745 </port>
1746 </output>
1747 </layer>
1748 <layer id="114" name="/distilbert/transformer/layer.0/attention/Constant_9" type="Const" version="opset1">
1749 <data element_type="i64" shape="1" offset="93131816" size="8" />
1750 <rt_info>
1751 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_9" />
1752 </rt_info>
1753 <output>
1754 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_9_output_0">
1755 <dim>1</dim>
1756 </port>
1757 </output>
1758 </layer>
1759 <layer id="115" name="/distilbert/transformer/layer.0/attention/Constant_10" type="Const" version="opset1">
1760 <data element_type="i64" shape="1" offset="93131824" size="8" />
1761 <rt_info>
1762 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_10" />
1763 </rt_info>
1764 <output>
1765 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_10_output_0">
1766 <dim>1</dim>
1767 </port>
1768 </output>
1769 </layer>
1770 <layer id="116" name="/distilbert/transformer/layer.0/attention/Concat_2" type="Concat" version="opset1">
1771 <data axis="0" />
1772 <rt_info>
1773 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Concat_2, /distilbert/transformer/layer.0/attention/Constant_10, /distilbert/transformer/layer.0/attention/Constant_8, /distilbert/transformer/layer.0/attention/Constant_9" />
1774 </rt_info>
1775 <input>
1776 <port id="0" precision="I64">
1777 <dim>1</dim>
1778 </port>
1779 <port id="1" precision="I64">
1780 <dim>1</dim>
1781 </port>
1782 <port id="2" precision="I64">
1783 <dim>1</dim>
1784 </port>
1785 <port id="3" precision="I64">
1786 <dim>1</dim>
1787 </port>
1788 </input>
1789 <output>
1790 <port id="4" precision="I64" names="/distilbert/transformer/layer.0/attention/Concat_2_output_0">
1791 <dim>4</dim>
1792 </port>
1793 </output>
1794 </layer>
1795 <layer id="117" name="/distilbert/transformer/layer.0/attention/Reshape_2" type="Reshape" version="opset1">
1796 <data special_zero="true" />
1797 <rt_info>
1798 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Reshape_2" />
1799 </rt_info>
1800 <input>
1801 <port id="0" precision="FP32">
1802 <dim>-1</dim>
1803 <dim>-1</dim>
1804 <dim>768</dim>
1805 </port>
1806 <port id="1" precision="I64">
1807 <dim>4</dim>
1808 </port>
1809 </input>
1810 <output>
1811 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Reshape_2_output_0">
1812 <dim>-1</dim>
1813 <dim>-1</dim>
1814 <dim>12</dim>
1815 <dim>64</dim>
1816 </port>
1817 </output>
1818 </layer>
1819 <layer id="118" name="Constant_255278" type="Const" version="opset1">
1820 <data element_type="i64" shape="4" offset="93131832" size="32" />
1821 <rt_info>
1822 <attribute name="fused_names" version="0" value="Constant_255278" />
1823 </rt_info>
1824 <output>
1825 <port id="0" precision="I64">
1826 <dim>4</dim>
1827 </port>
1828 </output>
1829 </layer>
1830 <layer id="119" name="/distilbert/transformer/layer.0/attention/Transpose_1" type="Transpose" version="opset1">
1831 <rt_info>
1832 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Transpose_1" />
1833 </rt_info>
1834 <input>
1835 <port id="0" precision="FP32">
1836 <dim>-1</dim>
1837 <dim>-1</dim>
1838 <dim>12</dim>
1839 <dim>64</dim>
1840 </port>
1841 <port id="1" precision="I64">
1842 <dim>4</dim>
1843 </port>
1844 </input>
1845 <output>
1846 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Transpose_1_output_0">
1847 <dim>-1</dim>
1848 <dim>12</dim>
1849 <dim>-1</dim>
1850 <dim>64</dim>
1851 </port>
1852 </output>
1853 </layer>
1854 <layer id="120" name="/distilbert/transformer/layer.0/attention/MatMul_1" type="MatMul" version="opset1">
1855 <data transpose_a="false" transpose_b="false" />
1856 <rt_info>
1857 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/MatMul_1" />
1858 </rt_info>
1859 <input>
1860 <port id="0" precision="FP32">
1861 <dim>-1</dim>
1862 <dim>12</dim>
1863 <dim>-1</dim>
1864 <dim>-1</dim>
1865 </port>
1866 <port id="1" precision="FP32">
1867 <dim>-1</dim>
1868 <dim>12</dim>
1869 <dim>-1</dim>
1870 <dim>64</dim>
1871 </port>
1872 </input>
1873 <output>
1874 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/MatMul_1_output_0">
1875 <dim>-1</dim>
1876 <dim>12</dim>
1877 <dim>-1</dim>
1878 <dim>64</dim>
1879 </port>
1880 </output>
1881 </layer>
1882 <layer id="121" name="Constant_255316" type="Const" version="opset1">
1883 <data element_type="i64" shape="4" offset="93131832" size="32" />
1884 <rt_info>
1885 <attribute name="fused_names" version="0" value="Constant_255316" />
1886 </rt_info>
1887 <output>
1888 <port id="0" precision="I64">
1889 <dim>4</dim>
1890 </port>
1891 </output>
1892 </layer>
1893 <layer id="122" name="/distilbert/transformer/layer.0/attention/Transpose_3" type="Transpose" version="opset1">
1894 <rt_info>
1895 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Transpose_3" />
1896 </rt_info>
1897 <input>
1898 <port id="0" precision="FP32">
1899 <dim>-1</dim>
1900 <dim>12</dim>
1901 <dim>-1</dim>
1902 <dim>64</dim>
1903 </port>
1904 <port id="1" precision="I64">
1905 <dim>4</dim>
1906 </port>
1907 </input>
1908 <output>
1909 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Transpose_3_output_0">
1910 <dim>-1</dim>
1911 <dim>-1</dim>
1912 <dim>12</dim>
1913 <dim>64</dim>
1914 </port>
1915 </output>
1916 </layer>
1917 <layer id="123" name="Constant_255318" type="Const" version="opset1">
1918 <data element_type="i64" shape="1" offset="18440" size="8" />
1919 <rt_info>
1920 <attribute name="fused_names" version="0" value="Constant_255318" />
1921 </rt_info>
1922 <output>
1923 <port id="0" precision="I64">
1924 <dim>1</dim>
1925 </port>
1926 </output>
1927 </layer>
1928 <layer id="124" name="/distilbert/transformer/layer.0/attention/Unsqueeze_5" type="Unsqueeze" version="opset1">
1929 <rt_info>
1930 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Unsqueeze_5, Constant_255318" />
1931 </rt_info>
1932 <input>
1933 <port id="0" precision="I64" />
1934 <port id="1" precision="I64">
1935 <dim>1</dim>
1936 </port>
1937 </input>
1938 <output>
1939 <port id="2" precision="I64" names="/distilbert/transformer/layer.0/attention/Unsqueeze_5_output_0">
1940 <dim>1</dim>
1941 </port>
1942 </output>
1943 </layer>
1944 <layer id="125" name="/distilbert/transformer/layer.0/attention/Constant_16" type="Const" version="opset1">
1945 <data element_type="i64" shape="1" offset="90763288" size="8" />
1946 <rt_info>
1947 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_16" />
1948 </rt_info>
1949 <output>
1950 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_16_output_0">
1951 <dim>1</dim>
1952 </port>
1953 </output>
1954 </layer>
1955 <layer id="126" name="/distilbert/transformer/layer.0/attention/Constant_17" type="Const" version="opset1">
1956 <data element_type="i64" shape="1" offset="97856640" size="8" />
1957 <rt_info>
1958 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Constant_17" />
1959 </rt_info>
1960 <output>
1961 <port id="0" precision="I64" names="/distilbert/transformer/layer.0/attention/Constant_17_output_0">
1962 <dim>1</dim>
1963 </port>
1964 </output>
1965 </layer>
1966 <layer id="127" name="/distilbert/transformer/layer.0/attention/Concat_4" type="Concat" version="opset1">
1967 <data axis="0" />
1968 <rt_info>
1969 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Concat_4, /distilbert/transformer/layer.0/attention/Constant_16, /distilbert/transformer/layer.0/attention/Constant_17" />
1970 </rt_info>
1971 <input>
1972 <port id="0" precision="I64">
1973 <dim>1</dim>
1974 </port>
1975 <port id="1" precision="I64">
1976 <dim>1</dim>
1977 </port>
1978 <port id="2" precision="I64">
1979 <dim>1</dim>
1980 </port>
1981 </input>
1982 <output>
1983 <port id="3" precision="I64" names="/distilbert/transformer/layer.0/attention/Concat_4_output_0">
1984 <dim>3</dim>
1985 </port>
1986 </output>
1987 </layer>
1988 <layer id="128" name="/distilbert/transformer/layer.0/attention/Reshape_4" type="Reshape" version="opset1">
1989 <data special_zero="true" />
1990 <rt_info>
1991 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/Reshape_4" />
1992 </rt_info>
1993 <input>
1994 <port id="0" precision="FP32">
1995 <dim>-1</dim>
1996 <dim>-1</dim>
1997 <dim>12</dim>
1998 <dim>64</dim>
1999 </port>
2000 <port id="1" precision="I64">
2001 <dim>3</dim>
2002 </port>
2003 </input>
2004 <output>
2005 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/Reshape_4_output_0">
2006 <dim>-1</dim>
2007 <dim>-1</dim>
2008 <dim>768</dim>
2009 </port>
2010 </output>
2011 </layer>
2012 <layer id="129" name="onnx::MatMul_754" type="Const" version="opset1">
2013 <data element_type="f32" shape="768, 768" offset="97856648" size="2359296" />
2014 <rt_info>
2015 <attribute name="fused_names" version="0" value="onnx::MatMul_754" />
2016 </rt_info>
2017 <output>
2018 <port id="0" precision="FP32" names="onnx::MatMul_754">
2019 <dim>768</dim>
2020 <dim>768</dim>
2021 </port>
2022 </output>
2023 </layer>
2024 <layer id="130" name="/distilbert/transformer/layer.0/attention/out_lin/MatMul" type="MatMul" version="opset1">
2025 <data transpose_a="false" transpose_b="false" />
2026 <rt_info>
2027 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/out_lin/MatMul" />
2028 </rt_info>
2029 <input>
2030 <port id="0" precision="FP32">
2031 <dim>-1</dim>
2032 <dim>-1</dim>
2033 <dim>768</dim>
2034 </port>
2035 <port id="1" precision="FP32">
2036 <dim>768</dim>
2037 <dim>768</dim>
2038 </port>
2039 </input>
2040 <output>
2041 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/out_lin/MatMul_output_0">
2042 <dim>-1</dim>
2043 <dim>-1</dim>
2044 <dim>768</dim>
2045 </port>
2046 </output>
2047 </layer>
2048 <layer id="131" name="/distilbert/transformer/layer.0/attention/out_lin/Add" type="Add" version="opset1">
2049 <data auto_broadcast="numpy" />
2050 <rt_info>
2051 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/attention/out_lin/Add" />
2052 </rt_info>
2053 <input>
2054 <port id="0" precision="FP32">
2055 <dim>768</dim>
2056 </port>
2057 <port id="1" precision="FP32">
2058 <dim>-1</dim>
2059 <dim>-1</dim>
2060 <dim>768</dim>
2061 </port>
2062 </input>
2063 <output>
2064 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/attention/out_lin/Add_output_0">
2065 <dim>-1</dim>
2066 <dim>-1</dim>
2067 <dim>768</dim>
2068 </port>
2069 </output>
2070 </layer>
2071 <layer id="132" name="/distilbert/transformer/layer.0/Add" type="Add" version="opset1">
2072 <data auto_broadcast="numpy" />
2073 <rt_info>
2074 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/Add" />
2075 </rt_info>
2076 <input>
2077 <port id="0" precision="FP32">
2078 <dim>-1</dim>
2079 <dim>-1</dim>
2080 <dim>768</dim>
2081 </port>
2082 <port id="1" precision="FP32">
2083 <dim>-1</dim>
2084 <dim>-1</dim>
2085 <dim>768</dim>
2086 </port>
2087 </input>
2088 <output>
2089 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/Add_output_0">
2090 <dim>-1</dim>
2091 <dim>-1</dim>
2092 <dim>768</dim>
2093 </port>
2094 </output>
2095 </layer>
2096 <layer id="133" name="Constant_255333" type="Const" version="opset1">
2097 <data element_type="i64" shape="1" offset="90763288" size="8" />
2098 <rt_info>
2099 <attribute name="fused_names" version="0" value="Constant_255333" />
2100 </rt_info>
2101 <output>
2102 <port id="0" precision="I64">
2103 <dim>1</dim>
2104 </port>
2105 </output>
2106 </layer>
2107 <layer id="134" name="/distilbert/transformer/layer.0/sa_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
2108 <data keep_dims="true" />
2109 <rt_info>
2110 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/ReduceMean" />
2111 </rt_info>
2112 <input>
2113 <port id="0" precision="FP32">
2114 <dim>-1</dim>
2115 <dim>-1</dim>
2116 <dim>768</dim>
2117 </port>
2118 <port id="1" precision="I64">
2119 <dim>1</dim>
2120 </port>
2121 </input>
2122 <output>
2123 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/ReduceMean_output_0">
2124 <dim>-1</dim>
2125 <dim>-1</dim>
2126 <dim>1</dim>
2127 </port>
2128 </output>
2129 </layer>
2130 <layer id="135" name="/distilbert/transformer/layer.0/sa_layer_norm/Sub" type="Subtract" version="opset1">
2131 <data auto_broadcast="numpy" />
2132 <rt_info>
2133 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Sub" />
2134 </rt_info>
2135 <input>
2136 <port id="0" precision="FP32">
2137 <dim>-1</dim>
2138 <dim>-1</dim>
2139 <dim>768</dim>
2140 </port>
2141 <port id="1" precision="FP32">
2142 <dim>-1</dim>
2143 <dim>-1</dim>
2144 <dim>1</dim>
2145 </port>
2146 </input>
2147 <output>
2148 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Sub_output_0">
2149 <dim>-1</dim>
2150 <dim>-1</dim>
2151 <dim>768</dim>
2152 </port>
2153 </output>
2154 </layer>
2155 <layer id="136" name="/distilbert/transformer/layer.0/sa_layer_norm/Constant" type="Const" version="opset1">
2156 <data element_type="f32" shape="" offset="90763296" size="4" />
2157 <rt_info>
2158 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Constant" />
2159 </rt_info>
2160 <output>
2161 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Constant_output_0" />
2162 </output>
2163 </layer>
2164 <layer id="137" name="/distilbert/transformer/layer.0/sa_layer_norm/Pow" type="Power" version="opset1">
2165 <data auto_broadcast="numpy" />
2166 <rt_info>
2167 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Pow" />
2168 </rt_info>
2169 <input>
2170 <port id="0" precision="FP32">
2171 <dim>-1</dim>
2172 <dim>-1</dim>
2173 <dim>768</dim>
2174 </port>
2175 <port id="1" precision="FP32" />
2176 </input>
2177 <output>
2178 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Pow_output_0">
2179 <dim>-1</dim>
2180 <dim>-1</dim>
2181 <dim>768</dim>
2182 </port>
2183 </output>
2184 </layer>
2185 <layer id="138" name="Constant_255338" type="Const" version="opset1">
2186 <data element_type="i64" shape="1" offset="90763288" size="8" />
2187 <rt_info>
2188 <attribute name="fused_names" version="0" value="Constant_255338" />
2189 </rt_info>
2190 <output>
2191 <port id="0" precision="I64">
2192 <dim>1</dim>
2193 </port>
2194 </output>
2195 </layer>
2196 <layer id="139" name="/distilbert/transformer/layer.0/sa_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
2197 <data keep_dims="true" />
2198 <rt_info>
2199 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/ReduceMean_1" />
2200 </rt_info>
2201 <input>
2202 <port id="0" precision="FP32">
2203 <dim>-1</dim>
2204 <dim>-1</dim>
2205 <dim>768</dim>
2206 </port>
2207 <port id="1" precision="I64">
2208 <dim>1</dim>
2209 </port>
2210 </input>
2211 <output>
2212 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/ReduceMean_1_output_0">
2213 <dim>-1</dim>
2214 <dim>-1</dim>
2215 <dim>1</dim>
2216 </port>
2217 </output>
2218 </layer>
2219 <layer id="140" name="/distilbert/transformer/layer.0/sa_layer_norm/Constant_1" type="Const" version="opset1">
2220 <data element_type="f32" shape="" offset="90763300" size="4" />
2221 <rt_info>
2222 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Constant_1" />
2223 </rt_info>
2224 <output>
2225 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Constant_1_output_0" />
2226 </output>
2227 </layer>
2228 <layer id="141" name="/distilbert/transformer/layer.0/sa_layer_norm/Add" type="Add" version="opset1">
2229 <data auto_broadcast="numpy" />
2230 <rt_info>
2231 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Add" />
2232 </rt_info>
2233 <input>
2234 <port id="0" precision="FP32">
2235 <dim>-1</dim>
2236 <dim>-1</dim>
2237 <dim>1</dim>
2238 </port>
2239 <port id="1" precision="FP32" />
2240 </input>
2241 <output>
2242 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Add_output_0">
2243 <dim>-1</dim>
2244 <dim>-1</dim>
2245 <dim>1</dim>
2246 </port>
2247 </output>
2248 </layer>
2249 <layer id="142" name="/distilbert/transformer/layer.0/sa_layer_norm/Sqrt" type="Sqrt" version="opset1">
2250 <rt_info>
2251 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Sqrt" />
2252 </rt_info>
2253 <input>
2254 <port id="0" precision="FP32">
2255 <dim>-1</dim>
2256 <dim>-1</dim>
2257 <dim>1</dim>
2258 </port>
2259 </input>
2260 <output>
2261 <port id="1" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Sqrt_output_0">
2262 <dim>-1</dim>
2263 <dim>-1</dim>
2264 <dim>1</dim>
2265 </port>
2266 </output>
2267 </layer>
2268 <layer id="143" name="/distilbert/transformer/layer.0/sa_layer_norm/Div" type="Divide" version="opset1">
2269 <data auto_broadcast="numpy" m_pythondiv="true" />
2270 <rt_info>
2271 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Div" />
2272 </rt_info>
2273 <input>
2274 <port id="0" precision="FP32">
2275 <dim>-1</dim>
2276 <dim>-1</dim>
2277 <dim>768</dim>
2278 </port>
2279 <port id="1" precision="FP32">
2280 <dim>-1</dim>
2281 <dim>-1</dim>
2282 <dim>1</dim>
2283 </port>
2284 </input>
2285 <output>
2286 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Div_output_0">
2287 <dim>-1</dim>
2288 <dim>-1</dim>
2289 <dim>768</dim>
2290 </port>
2291 </output>
2292 </layer>
2293 <layer id="144" name="distilbert.transformer.layer.0.sa_layer_norm.weight" type="Const" version="opset1">
2294 <data element_type="f32" shape="768" offset="100215944" size="3072" />
2295 <rt_info>
2296 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.sa_layer_norm.weight" />
2297 </rt_info>
2298 <output>
2299 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.sa_layer_norm.weight">
2300 <dim>768</dim>
2301 </port>
2302 </output>
2303 </layer>
2304 <layer id="145" name="/distilbert/transformer/layer.0/sa_layer_norm/Mul" type="Multiply" version="opset1">
2305 <data auto_broadcast="numpy" />
2306 <rt_info>
2307 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Mul" />
2308 </rt_info>
2309 <input>
2310 <port id="0" precision="FP32">
2311 <dim>-1</dim>
2312 <dim>-1</dim>
2313 <dim>768</dim>
2314 </port>
2315 <port id="1" precision="FP32">
2316 <dim>768</dim>
2317 </port>
2318 </input>
2319 <output>
2320 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Mul_output_0">
2321 <dim>-1</dim>
2322 <dim>-1</dim>
2323 <dim>768</dim>
2324 </port>
2325 </output>
2326 </layer>
2327 <layer id="146" name="distilbert.transformer.layer.0.sa_layer_norm.bias" type="Const" version="opset1">
2328 <data element_type="f32" shape="768" offset="100219016" size="3072" />
2329 <rt_info>
2330 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.sa_layer_norm.bias" />
2331 </rt_info>
2332 <output>
2333 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.sa_layer_norm.bias">
2334 <dim>768</dim>
2335 </port>
2336 </output>
2337 </layer>
2338 <layer id="147" name="/distilbert/transformer/layer.0/sa_layer_norm/Add_1" type="Add" version="opset1">
2339 <data auto_broadcast="numpy" />
2340 <rt_info>
2341 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/sa_layer_norm/Add_1" />
2342 </rt_info>
2343 <input>
2344 <port id="0" precision="FP32">
2345 <dim>-1</dim>
2346 <dim>-1</dim>
2347 <dim>768</dim>
2348 </port>
2349 <port id="1" precision="FP32">
2350 <dim>768</dim>
2351 </port>
2352 </input>
2353 <output>
2354 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/sa_layer_norm/Add_1_output_0">
2355 <dim>-1</dim>
2356 <dim>-1</dim>
2357 <dim>768</dim>
2358 </port>
2359 </output>
2360 </layer>
2361 <layer id="148" name="onnx::MatMul_755" type="Const" version="opset1">
2362 <data element_type="f32" shape="768, 3072" offset="100222088" size="9437184" />
2363 <rt_info>
2364 <attribute name="fused_names" version="0" value="onnx::MatMul_755" />
2365 </rt_info>
2366 <output>
2367 <port id="0" precision="FP32" names="onnx::MatMul_755">
2368 <dim>768</dim>
2369 <dim>3072</dim>
2370 </port>
2371 </output>
2372 </layer>
2373 <layer id="149" name="/distilbert/transformer/layer.0/ffn/lin1/MatMul" type="MatMul" version="opset1">
2374 <data transpose_a="false" transpose_b="false" />
2375 <rt_info>
2376 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/lin1/MatMul" />
2377 </rt_info>
2378 <input>
2379 <port id="0" precision="FP32">
2380 <dim>-1</dim>
2381 <dim>-1</dim>
2382 <dim>768</dim>
2383 </port>
2384 <port id="1" precision="FP32">
2385 <dim>768</dim>
2386 <dim>3072</dim>
2387 </port>
2388 </input>
2389 <output>
2390 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/lin1/MatMul_output_0">
2391 <dim>-1</dim>
2392 <dim>-1</dim>
2393 <dim>3072</dim>
2394 </port>
2395 </output>
2396 </layer>
2397 <layer id="150" name="/distilbert/transformer/layer.0/ffn/lin1/Add" type="Add" version="opset1">
2398 <data auto_broadcast="numpy" />
2399 <rt_info>
2400 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/lin1/Add" />
2401 </rt_info>
2402 <input>
2403 <port id="0" precision="FP32">
2404 <dim>3072</dim>
2405 </port>
2406 <port id="1" precision="FP32">
2407 <dim>-1</dim>
2408 <dim>-1</dim>
2409 <dim>3072</dim>
2410 </port>
2411 </input>
2412 <output>
2413 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/lin1/Add_output_0">
2414 <dim>-1</dim>
2415 <dim>-1</dim>
2416 <dim>3072</dim>
2417 </port>
2418 </output>
2419 </layer>
2420 <layer id="151" name="/distilbert/transformer/layer.0/ffn/activation/Constant" type="Const" version="opset1">
2421 <data element_type="f32" shape="" offset="109659272" size="4" />
2422 <rt_info>
2423 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Constant" />
2424 </rt_info>
2425 <output>
2426 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Constant_output_0" />
2427 </output>
2428 </layer>
2429 <layer id="152" name="/distilbert/transformer/layer.0/ffn/activation/Div" type="Divide" version="opset1">
2430 <data auto_broadcast="numpy" m_pythondiv="true" />
2431 <rt_info>
2432 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Div" />
2433 </rt_info>
2434 <input>
2435 <port id="0" precision="FP32">
2436 <dim>-1</dim>
2437 <dim>-1</dim>
2438 <dim>3072</dim>
2439 </port>
2440 <port id="1" precision="FP32" />
2441 </input>
2442 <output>
2443 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Div_output_0">
2444 <dim>-1</dim>
2445 <dim>-1</dim>
2446 <dim>3072</dim>
2447 </port>
2448 </output>
2449 </layer>
2450 <layer id="153" name="/distilbert/transformer/layer.0/ffn/activation/Erf" type="Erf" version="opset1">
2451 <rt_info>
2452 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Erf" />
2453 </rt_info>
2454 <input>
2455 <port id="0" precision="FP32">
2456 <dim>-1</dim>
2457 <dim>-1</dim>
2458 <dim>3072</dim>
2459 </port>
2460 </input>
2461 <output>
2462 <port id="1" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Erf_output_0">
2463 <dim>-1</dim>
2464 <dim>-1</dim>
2465 <dim>3072</dim>
2466 </port>
2467 </output>
2468 </layer>
2469 <layer id="154" name="/distilbert/transformer/layer.0/ffn/activation/Constant_1" type="Const" version="opset1">
2470 <data element_type="f32" shape="" offset="109659276" size="4" />
2471 <rt_info>
2472 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Constant_1" />
2473 </rt_info>
2474 <output>
2475 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Constant_1_output_0" />
2476 </output>
2477 </layer>
2478 <layer id="155" name="/distilbert/transformer/layer.0/ffn/activation/Add" type="Add" version="opset1">
2479 <data auto_broadcast="numpy" />
2480 <rt_info>
2481 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Add" />
2482 </rt_info>
2483 <input>
2484 <port id="0" precision="FP32">
2485 <dim>-1</dim>
2486 <dim>-1</dim>
2487 <dim>3072</dim>
2488 </port>
2489 <port id="1" precision="FP32" />
2490 </input>
2491 <output>
2492 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Add_output_0">
2493 <dim>-1</dim>
2494 <dim>-1</dim>
2495 <dim>3072</dim>
2496 </port>
2497 </output>
2498 </layer>
2499 <layer id="156" name="/distilbert/transformer/layer.0/ffn/activation/Mul" type="Multiply" version="opset1">
2500 <data auto_broadcast="numpy" />
2501 <rt_info>
2502 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Mul" />
2503 </rt_info>
2504 <input>
2505 <port id="0" precision="FP32">
2506 <dim>-1</dim>
2507 <dim>-1</dim>
2508 <dim>3072</dim>
2509 </port>
2510 <port id="1" precision="FP32">
2511 <dim>-1</dim>
2512 <dim>-1</dim>
2513 <dim>3072</dim>
2514 </port>
2515 </input>
2516 <output>
2517 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Mul_output_0">
2518 <dim>-1</dim>
2519 <dim>-1</dim>
2520 <dim>3072</dim>
2521 </port>
2522 </output>
2523 </layer>
2524 <layer id="157" name="/distilbert/transformer/layer.0/ffn/activation/Constant_2" type="Const" version="opset1">
2525 <data element_type="f32" shape="" offset="109659280" size="4" />
2526 <rt_info>
2527 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Constant_2" />
2528 </rt_info>
2529 <output>
2530 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Constant_2_output_0" />
2531 </output>
2532 </layer>
2533 <layer id="158" name="/distilbert/transformer/layer.0/ffn/activation/Mul_1" type="Multiply" version="opset1">
2534 <data auto_broadcast="numpy" />
2535 <rt_info>
2536 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/activation/Mul_1" />
2537 </rt_info>
2538 <input>
2539 <port id="0" precision="FP32">
2540 <dim>-1</dim>
2541 <dim>-1</dim>
2542 <dim>3072</dim>
2543 </port>
2544 <port id="1" precision="FP32" />
2545 </input>
2546 <output>
2547 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/activation/Mul_1_output_0">
2548 <dim>-1</dim>
2549 <dim>-1</dim>
2550 <dim>3072</dim>
2551 </port>
2552 </output>
2553 </layer>
2554 <layer id="159" name="onnx::MatMul_756" type="Const" version="opset1">
2555 <data element_type="f32" shape="3072, 768" offset="109659284" size="9437184" />
2556 <rt_info>
2557 <attribute name="fused_names" version="0" value="onnx::MatMul_756" />
2558 </rt_info>
2559 <output>
2560 <port id="0" precision="FP32" names="onnx::MatMul_756">
2561 <dim>3072</dim>
2562 <dim>768</dim>
2563 </port>
2564 </output>
2565 </layer>
2566 <layer id="160" name="/distilbert/transformer/layer.0/ffn/lin2/MatMul" type="MatMul" version="opset1">
2567 <data transpose_a="false" transpose_b="false" />
2568 <rt_info>
2569 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/lin2/MatMul" />
2570 </rt_info>
2571 <input>
2572 <port id="0" precision="FP32">
2573 <dim>-1</dim>
2574 <dim>-1</dim>
2575 <dim>3072</dim>
2576 </port>
2577 <port id="1" precision="FP32">
2578 <dim>3072</dim>
2579 <dim>768</dim>
2580 </port>
2581 </input>
2582 <output>
2583 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/lin2/MatMul_output_0">
2584 <dim>-1</dim>
2585 <dim>-1</dim>
2586 <dim>768</dim>
2587 </port>
2588 </output>
2589 </layer>
2590 <layer id="161" name="/distilbert/transformer/layer.0/ffn/lin2/Add" type="Add" version="opset1">
2591 <data auto_broadcast="numpy" />
2592 <rt_info>
2593 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/ffn/lin2/Add" />
2594 </rt_info>
2595 <input>
2596 <port id="0" precision="FP32">
2597 <dim>768</dim>
2598 </port>
2599 <port id="1" precision="FP32">
2600 <dim>-1</dim>
2601 <dim>-1</dim>
2602 <dim>768</dim>
2603 </port>
2604 </input>
2605 <output>
2606 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/ffn/lin2/Add_output_0">
2607 <dim>-1</dim>
2608 <dim>-1</dim>
2609 <dim>768</dim>
2610 </port>
2611 </output>
2612 </layer>
2613 <layer id="162" name="/distilbert/transformer/layer.0/Add_1" type="Add" version="opset1">
2614 <data auto_broadcast="numpy" />
2615 <rt_info>
2616 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/Add_1" />
2617 </rt_info>
2618 <input>
2619 <port id="0" precision="FP32">
2620 <dim>-1</dim>
2621 <dim>-1</dim>
2622 <dim>768</dim>
2623 </port>
2624 <port id="1" precision="FP32">
2625 <dim>-1</dim>
2626 <dim>-1</dim>
2627 <dim>768</dim>
2628 </port>
2629 </input>
2630 <output>
2631 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/Add_1_output_0">
2632 <dim>-1</dim>
2633 <dim>-1</dim>
2634 <dim>768</dim>
2635 </port>
2636 </output>
2637 </layer>
2638 <layer id="163" name="Constant_255359" type="Const" version="opset1">
2639 <data element_type="i64" shape="1" offset="90763288" size="8" />
2640 <rt_info>
2641 <attribute name="fused_names" version="0" value="Constant_255359" />
2642 </rt_info>
2643 <output>
2644 <port id="0" precision="I64">
2645 <dim>1</dim>
2646 </port>
2647 </output>
2648 </layer>
2649 <layer id="164" name="/distilbert/transformer/layer.0/output_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
2650 <data keep_dims="true" />
2651 <rt_info>
2652 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/ReduceMean" />
2653 </rt_info>
2654 <input>
2655 <port id="0" precision="FP32">
2656 <dim>-1</dim>
2657 <dim>-1</dim>
2658 <dim>768</dim>
2659 </port>
2660 <port id="1" precision="I64">
2661 <dim>1</dim>
2662 </port>
2663 </input>
2664 <output>
2665 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/ReduceMean_output_0">
2666 <dim>-1</dim>
2667 <dim>-1</dim>
2668 <dim>1</dim>
2669 </port>
2670 </output>
2671 </layer>
2672 <layer id="165" name="/distilbert/transformer/layer.0/output_layer_norm/Sub" type="Subtract" version="opset1">
2673 <data auto_broadcast="numpy" />
2674 <rt_info>
2675 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Sub" />
2676 </rt_info>
2677 <input>
2678 <port id="0" precision="FP32">
2679 <dim>-1</dim>
2680 <dim>-1</dim>
2681 <dim>768</dim>
2682 </port>
2683 <port id="1" precision="FP32">
2684 <dim>-1</dim>
2685 <dim>-1</dim>
2686 <dim>1</dim>
2687 </port>
2688 </input>
2689 <output>
2690 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Sub_output_0">
2691 <dim>-1</dim>
2692 <dim>-1</dim>
2693 <dim>768</dim>
2694 </port>
2695 </output>
2696 </layer>
2697 <layer id="166" name="/distilbert/transformer/layer.0/output_layer_norm/Constant" type="Const" version="opset1">
2698 <data element_type="f32" shape="" offset="90763296" size="4" />
2699 <rt_info>
2700 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Constant" />
2701 </rt_info>
2702 <output>
2703 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Constant_output_0" />
2704 </output>
2705 </layer>
2706 <layer id="167" name="/distilbert/transformer/layer.0/output_layer_norm/Pow" type="Power" version="opset1">
2707 <data auto_broadcast="numpy" />
2708 <rt_info>
2709 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Pow" />
2710 </rt_info>
2711 <input>
2712 <port id="0" precision="FP32">
2713 <dim>-1</dim>
2714 <dim>-1</dim>
2715 <dim>768</dim>
2716 </port>
2717 <port id="1" precision="FP32" />
2718 </input>
2719 <output>
2720 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Pow_output_0">
2721 <dim>-1</dim>
2722 <dim>-1</dim>
2723 <dim>768</dim>
2724 </port>
2725 </output>
2726 </layer>
2727 <layer id="168" name="Constant_255364" type="Const" version="opset1">
2728 <data element_type="i64" shape="1" offset="90763288" size="8" />
2729 <rt_info>
2730 <attribute name="fused_names" version="0" value="Constant_255364" />
2731 </rt_info>
2732 <output>
2733 <port id="0" precision="I64">
2734 <dim>1</dim>
2735 </port>
2736 </output>
2737 </layer>
2738 <layer id="169" name="/distilbert/transformer/layer.0/output_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
2739 <data keep_dims="true" />
2740 <rt_info>
2741 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/ReduceMean_1" />
2742 </rt_info>
2743 <input>
2744 <port id="0" precision="FP32">
2745 <dim>-1</dim>
2746 <dim>-1</dim>
2747 <dim>768</dim>
2748 </port>
2749 <port id="1" precision="I64">
2750 <dim>1</dim>
2751 </port>
2752 </input>
2753 <output>
2754 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/ReduceMean_1_output_0">
2755 <dim>-1</dim>
2756 <dim>-1</dim>
2757 <dim>1</dim>
2758 </port>
2759 </output>
2760 </layer>
2761 <layer id="170" name="/distilbert/transformer/layer.0/output_layer_norm/Constant_1" type="Const" version="opset1">
2762 <data element_type="f32" shape="" offset="90763300" size="4" />
2763 <rt_info>
2764 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Constant_1" />
2765 </rt_info>
2766 <output>
2767 <port id="0" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Constant_1_output_0" />
2768 </output>
2769 </layer>
2770 <layer id="171" name="/distilbert/transformer/layer.0/output_layer_norm/Add" type="Add" version="opset1">
2771 <data auto_broadcast="numpy" />
2772 <rt_info>
2773 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Add" />
2774 </rt_info>
2775 <input>
2776 <port id="0" precision="FP32">
2777 <dim>-1</dim>
2778 <dim>-1</dim>
2779 <dim>1</dim>
2780 </port>
2781 <port id="1" precision="FP32" />
2782 </input>
2783 <output>
2784 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Add_output_0">
2785 <dim>-1</dim>
2786 <dim>-1</dim>
2787 <dim>1</dim>
2788 </port>
2789 </output>
2790 </layer>
2791 <layer id="172" name="/distilbert/transformer/layer.0/output_layer_norm/Sqrt" type="Sqrt" version="opset1">
2792 <rt_info>
2793 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Sqrt" />
2794 </rt_info>
2795 <input>
2796 <port id="0" precision="FP32">
2797 <dim>-1</dim>
2798 <dim>-1</dim>
2799 <dim>1</dim>
2800 </port>
2801 </input>
2802 <output>
2803 <port id="1" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Sqrt_output_0">
2804 <dim>-1</dim>
2805 <dim>-1</dim>
2806 <dim>1</dim>
2807 </port>
2808 </output>
2809 </layer>
2810 <layer id="173" name="/distilbert/transformer/layer.0/output_layer_norm/Div" type="Divide" version="opset1">
2811 <data auto_broadcast="numpy" m_pythondiv="true" />
2812 <rt_info>
2813 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Div" />
2814 </rt_info>
2815 <input>
2816 <port id="0" precision="FP32">
2817 <dim>-1</dim>
2818 <dim>-1</dim>
2819 <dim>768</dim>
2820 </port>
2821 <port id="1" precision="FP32">
2822 <dim>-1</dim>
2823 <dim>-1</dim>
2824 <dim>1</dim>
2825 </port>
2826 </input>
2827 <output>
2828 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Div_output_0">
2829 <dim>-1</dim>
2830 <dim>-1</dim>
2831 <dim>768</dim>
2832 </port>
2833 </output>
2834 </layer>
2835 <layer id="174" name="distilbert.transformer.layer.0.output_layer_norm.weight" type="Const" version="opset1">
2836 <data element_type="f32" shape="768" offset="119096468" size="3072" />
2837 <rt_info>
2838 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.output_layer_norm.weight" />
2839 </rt_info>
2840 <output>
2841 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.output_layer_norm.weight">
2842 <dim>768</dim>
2843 </port>
2844 </output>
2845 </layer>
2846 <layer id="175" name="/distilbert/transformer/layer.0/output_layer_norm/Mul" type="Multiply" version="opset1">
2847 <data auto_broadcast="numpy" />
2848 <rt_info>
2849 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Mul" />
2850 </rt_info>
2851 <input>
2852 <port id="0" precision="FP32">
2853 <dim>-1</dim>
2854 <dim>-1</dim>
2855 <dim>768</dim>
2856 </port>
2857 <port id="1" precision="FP32">
2858 <dim>768</dim>
2859 </port>
2860 </input>
2861 <output>
2862 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Mul_output_0">
2863 <dim>-1</dim>
2864 <dim>-1</dim>
2865 <dim>768</dim>
2866 </port>
2867 </output>
2868 </layer>
2869 <layer id="176" name="distilbert.transformer.layer.0.output_layer_norm.bias" type="Const" version="opset1">
2870 <data element_type="f32" shape="768" offset="119099540" size="3072" />
2871 <rt_info>
2872 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.0.output_layer_norm.bias" />
2873 </rt_info>
2874 <output>
2875 <port id="0" precision="FP32" names="distilbert.transformer.layer.0.output_layer_norm.bias">
2876 <dim>768</dim>
2877 </port>
2878 </output>
2879 </layer>
2880 <layer id="177" name="/distilbert/transformer/layer.0/output_layer_norm/Add_1" type="Add" version="opset1">
2881 <data auto_broadcast="numpy" />
2882 <rt_info>
2883 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.0/output_layer_norm/Add_1" />
2884 </rt_info>
2885 <input>
2886 <port id="0" precision="FP32">
2887 <dim>-1</dim>
2888 <dim>-1</dim>
2889 <dim>768</dim>
2890 </port>
2891 <port id="1" precision="FP32">
2892 <dim>768</dim>
2893 </port>
2894 </input>
2895 <output>
2896 <port id="2" precision="FP32" names="/distilbert/transformer/layer.0/output_layer_norm/Add_1_output_0">
2897 <dim>-1</dim>
2898 <dim>-1</dim>
2899 <dim>768</dim>
2900 </port>
2901 </output>
2902 </layer>
2903 <layer id="178" name="/distilbert/transformer/layer.1/attention/Shape" type="ShapeOf" version="opset3">
2904 <data output_type="i64" />
2905 <rt_info>
2906 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Shape" />
2907 </rt_info>
2908 <input>
2909 <port id="0" precision="FP32">
2910 <dim>-1</dim>
2911 <dim>-1</dim>
2912 <dim>768</dim>
2913 </port>
2914 </input>
2915 <output>
2916 <port id="1" precision="I64" names="/distilbert/transformer/layer.1/attention/Shape_output_0">
2917 <dim>3</dim>
2918 </port>
2919 </output>
2920 </layer>
2921 <layer id="179" name="/distilbert/transformer/layer.1/attention/Constant" type="Const" version="opset1">
2922 <data element_type="i64" shape="" offset="18440" size="8" />
2923 <rt_info>
2924 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant" />
2925 </rt_info>
2926 <output>
2927 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_output_0" />
2928 </output>
2929 </layer>
2930 <layer id="180" name="Constant_255374" type="Const" version="opset1">
2931 <data element_type="i64" shape="" offset="18440" size="8" />
2932 <rt_info>
2933 <attribute name="fused_names" version="0" value="Constant_255374" />
2934 </rt_info>
2935 <output>
2936 <port id="0" precision="I64" />
2937 </output>
2938 </layer>
2939 <layer id="181" name="/distilbert/transformer/layer.1/attention/Gather" type="Gather" version="opset8">
2940 <data batch_dims="0" />
2941 <rt_info>
2942 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant, /distilbert/transformer/layer.1/attention/Gather, Constant_255374" />
2943 </rt_info>
2944 <input>
2945 <port id="0" precision="I64">
2946 <dim>3</dim>
2947 </port>
2948 <port id="1" precision="I64" />
2949 <port id="2" precision="I64" />
2950 </input>
2951 <output>
2952 <port id="3" precision="I64" names="/distilbert/transformer/layer.1/attention/Gather_output_0" />
2953 </output>
2954 </layer>
2955 <layer id="182" name="Constant_255434" type="Const" version="opset1">
2956 <data element_type="i64" shape="1" offset="18440" size="8" />
2957 <rt_info>
2958 <attribute name="fused_names" version="0" value="Constant_255434" />
2959 </rt_info>
2960 <output>
2961 <port id="0" precision="I64">
2962 <dim>1</dim>
2963 </port>
2964 </output>
2965 </layer>
2966 <layer id="183" name="/distilbert/transformer/layer.1/attention/Unsqueeze_3" type="Unsqueeze" version="opset1">
2967 <rt_info>
2968 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Unsqueeze_3, Constant_255434" />
2969 </rt_info>
2970 <input>
2971 <port id="0" precision="I64" />
2972 <port id="1" precision="I64">
2973 <dim>1</dim>
2974 </port>
2975 </input>
2976 <output>
2977 <port id="2" precision="I64" names="/distilbert/transformer/layer.1/attention/Unsqueeze_3_output_0">
2978 <dim>1</dim>
2979 </port>
2980 </output>
2981 </layer>
2982 <layer id="184" name="/distilbert/transformer/layer.1/attention/Constant_12" type="Const" version="opset1">
2983 <data element_type="i64" shape="1" offset="90763280" size="8" />
2984 <rt_info>
2985 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_12" />
2986 </rt_info>
2987 <output>
2988 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_12_output_0">
2989 <dim>1</dim>
2990 </port>
2991 </output>
2992 </layer>
2993 <layer id="185" name="/distilbert/transformer/layer.1/attention/Constant_13" type="Const" version="opset1">
2994 <data element_type="i64" shape="1" offset="90763280" size="8" />
2995 <rt_info>
2996 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_13" />
2997 </rt_info>
2998 <output>
2999 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_13_output_0">
3000 <dim>1</dim>
3001 </port>
3002 </output>
3003 </layer>
3004 <layer id="186" name="/distilbert/transformer/layer.1/attention/Shape_1" type="ShapeOf" version="opset3">
3005 <data output_type="i64" />
3006 <rt_info>
3007 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Shape_1" />
3008 </rt_info>
3009 <input>
3010 <port id="0" precision="FP32">
3011 <dim>-1</dim>
3012 <dim>-1</dim>
3013 <dim>768</dim>
3014 </port>
3015 </input>
3016 <output>
3017 <port id="1" precision="I64" names="/distilbert/transformer/layer.1/attention/Shape_1_output_0">
3018 <dim>3</dim>
3019 </port>
3020 </output>
3021 </layer>
3022 <layer id="187" name="/distilbert/transformer/layer.1/attention/Constant_1" type="Const" version="opset1">
3023 <data element_type="i64" shape="" offset="90763280" size="8" />
3024 <rt_info>
3025 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_1" />
3026 </rt_info>
3027 <output>
3028 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_1_output_0" />
3029 </output>
3030 </layer>
3031 <layer id="188" name="Constant_255378" type="Const" version="opset1">
3032 <data element_type="i64" shape="" offset="18440" size="8" />
3033 <rt_info>
3034 <attribute name="fused_names" version="0" value="Constant_255378" />
3035 </rt_info>
3036 <output>
3037 <port id="0" precision="I64" />
3038 </output>
3039 </layer>
3040 <layer id="189" name="/distilbert/transformer/layer.1/attention/Gather_1" type="Gather" version="opset8">
3041 <data batch_dims="0" />
3042 <rt_info>
3043 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_1, /distilbert/transformer/layer.1/attention/Gather_1, Constant_255378" />
3044 </rt_info>
3045 <input>
3046 <port id="0" precision="I64">
3047 <dim>3</dim>
3048 </port>
3049 <port id="1" precision="I64" />
3050 <port id="2" precision="I64" />
3051 </input>
3052 <output>
3053 <port id="3" precision="I64" names="/distilbert/transformer/layer.1/attention/Gather_1_output_0" />
3054 </output>
3055 </layer>
3056 <layer id="190" name="Constant_255438" type="Const" version="opset1">
3057 <data element_type="i64" shape="1" offset="18440" size="8" />
3058 <rt_info>
3059 <attribute name="fused_names" version="0" value="Constant_255438" />
3060 </rt_info>
3061 <output>
3062 <port id="0" precision="I64">
3063 <dim>1</dim>
3064 </port>
3065 </output>
3066 </layer>
3067 <layer id="191" name="/distilbert/transformer/layer.1/attention/Unsqueeze_4" type="Unsqueeze" version="opset1">
3068 <rt_info>
3069 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Unsqueeze_4, Constant_255438" />
3070 </rt_info>
3071 <input>
3072 <port id="0" precision="I64" />
3073 <port id="1" precision="I64">
3074 <dim>1</dim>
3075 </port>
3076 </input>
3077 <output>
3078 <port id="2" precision="I64" names="/distilbert/transformer/layer.1/attention/Unsqueeze_4_output_0">
3079 <dim>1</dim>
3080 </port>
3081 </output>
3082 </layer>
3083 <layer id="192" name="/distilbert/transformer/layer.1/attention/Concat_3" type="Concat" version="opset1">
3084 <data axis="0" />
3085 <rt_info>
3086 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Concat_3, /distilbert/transformer/layer.1/attention/Constant_12, /distilbert/transformer/layer.1/attention/Constant_13" />
3087 </rt_info>
3088 <input>
3089 <port id="0" precision="I64">
3090 <dim>1</dim>
3091 </port>
3092 <port id="1" precision="I64">
3093 <dim>1</dim>
3094 </port>
3095 <port id="2" precision="I64">
3096 <dim>1</dim>
3097 </port>
3098 <port id="3" precision="I64">
3099 <dim>1</dim>
3100 </port>
3101 </input>
3102 <output>
3103 <port id="4" precision="I64" names="/distilbert/transformer/layer.1/attention/Concat_3_output_0">
3104 <dim>4</dim>
3105 </port>
3106 </output>
3107 </layer>
3108 <layer id="193" name="/distilbert/transformer/layer.1/attention/Reshape_3" type="Reshape" version="opset1">
3109 <data special_zero="true" />
3110 <rt_info>
3111 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Reshape_3" />
3112 </rt_info>
3113 <input>
3114 <port id="0" precision="BOOL">
3115 <dim>-1</dim>
3116 <dim>-1</dim>
3117 </port>
3118 <port id="1" precision="I64">
3119 <dim>4</dim>
3120 </port>
3121 </input>
3122 <output>
3123 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.1/attention/Reshape_3_output_0">
3124 <dim>-1</dim>
3125 <dim>1</dim>
3126 <dim>1</dim>
3127 <dim>-1</dim>
3128 </port>
3129 </output>
3130 </layer>
3131 <layer id="194" name="distilbert.transformer.layer.1.attention.q_lin.bias" type="Const" version="opset1">
3132 <data element_type="f32" shape="768" offset="119102612" size="3072" />
3133 <rt_info>
3134 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.attention.q_lin.bias" />
3135 </rt_info>
3136 <output>
3137 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.attention.q_lin.bias">
3138 <dim>768</dim>
3139 </port>
3140 </output>
3141 </layer>
3142 <layer id="195" name="onnx::MatMul_757" type="Const" version="opset1">
3143 <data element_type="f32" shape="768, 768" offset="119105684" size="2359296" />
3144 <rt_info>
3145 <attribute name="fused_names" version="0" value="onnx::MatMul_757" />
3146 </rt_info>
3147 <output>
3148 <port id="0" precision="FP32" names="onnx::MatMul_757">
3149 <dim>768</dim>
3150 <dim>768</dim>
3151 </port>
3152 </output>
3153 </layer>
3154 <layer id="196" name="/distilbert/transformer/layer.1/attention/q_lin/MatMul" type="MatMul" version="opset1">
3155 <data transpose_a="false" transpose_b="false" />
3156 <rt_info>
3157 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/q_lin/MatMul" />
3158 </rt_info>
3159 <input>
3160 <port id="0" precision="FP32">
3161 <dim>-1</dim>
3162 <dim>-1</dim>
3163 <dim>768</dim>
3164 </port>
3165 <port id="1" precision="FP32">
3166 <dim>768</dim>
3167 <dim>768</dim>
3168 </port>
3169 </input>
3170 <output>
3171 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/q_lin/MatMul_output_0">
3172 <dim>-1</dim>
3173 <dim>-1</dim>
3174 <dim>768</dim>
3175 </port>
3176 </output>
3177 </layer>
3178 <layer id="197" name="/distilbert/transformer/layer.1/attention/q_lin/Add" type="Add" version="opset1">
3179 <data auto_broadcast="numpy" />
3180 <rt_info>
3181 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/q_lin/Add" />
3182 </rt_info>
3183 <input>
3184 <port id="0" precision="FP32">
3185 <dim>768</dim>
3186 </port>
3187 <port id="1" precision="FP32">
3188 <dim>-1</dim>
3189 <dim>-1</dim>
3190 <dim>768</dim>
3191 </port>
3192 </input>
3193 <output>
3194 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/q_lin/Add_output_0">
3195 <dim>-1</dim>
3196 <dim>-1</dim>
3197 <dim>768</dim>
3198 </port>
3199 </output>
3200 </layer>
3201 <layer id="198" name="Constant_255382" type="Const" version="opset1">
3202 <data element_type="i64" shape="1" offset="18440" size="8" />
3203 <rt_info>
3204 <attribute name="fused_names" version="0" value="Constant_255382" />
3205 </rt_info>
3206 <output>
3207 <port id="0" precision="I64">
3208 <dim>1</dim>
3209 </port>
3210 </output>
3211 </layer>
3212 <layer id="199" name="/distilbert/transformer/layer.1/attention/Unsqueeze" type="Unsqueeze" version="opset1">
3213 <rt_info>
3214 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Unsqueeze, Constant_255382" />
3215 </rt_info>
3216 <input>
3217 <port id="0" precision="I64" />
3218 <port id="1" precision="I64">
3219 <dim>1</dim>
3220 </port>
3221 </input>
3222 <output>
3223 <port id="2" precision="I64" names="/distilbert/transformer/layer.1/attention/Unsqueeze_output_0">
3224 <dim>1</dim>
3225 </port>
3226 </output>
3227 </layer>
3228 <layer id="200" name="/distilbert/transformer/layer.1/attention/Constant_2" type="Const" version="opset1">
3229 <data element_type="i64" shape="1" offset="90763288" size="8" />
3230 <rt_info>
3231 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_2" />
3232 </rt_info>
3233 <output>
3234 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_2_output_0">
3235 <dim>1</dim>
3236 </port>
3237 </output>
3238 </layer>
3239 <layer id="201" name="/distilbert/transformer/layer.1/attention/Constant_3" type="Const" version="opset1">
3240 <data element_type="i64" shape="1" offset="93131816" size="8" />
3241 <rt_info>
3242 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_3" />
3243 </rt_info>
3244 <output>
3245 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_3_output_0">
3246 <dim>1</dim>
3247 </port>
3248 </output>
3249 </layer>
3250 <layer id="202" name="/distilbert/transformer/layer.1/attention/Constant_4" type="Const" version="opset1">
3251 <data element_type="i64" shape="1" offset="93131824" size="8" />
3252 <rt_info>
3253 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_4" />
3254 </rt_info>
3255 <output>
3256 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_4_output_0">
3257 <dim>1</dim>
3258 </port>
3259 </output>
3260 </layer>
3261 <layer id="203" name="/distilbert/transformer/layer.1/attention/Concat" type="Concat" version="opset1">
3262 <data axis="0" />
3263 <rt_info>
3264 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Concat, /distilbert/transformer/layer.1/attention/Constant_2, /distilbert/transformer/layer.1/attention/Constant_3, /distilbert/transformer/layer.1/attention/Constant_4" />
3265 </rt_info>
3266 <input>
3267 <port id="0" precision="I64">
3268 <dim>1</dim>
3269 </port>
3270 <port id="1" precision="I64">
3271 <dim>1</dim>
3272 </port>
3273 <port id="2" precision="I64">
3274 <dim>1</dim>
3275 </port>
3276 <port id="3" precision="I64">
3277 <dim>1</dim>
3278 </port>
3279 </input>
3280 <output>
3281 <port id="4" precision="I64" names="/distilbert/transformer/layer.1/attention/Concat_output_0">
3282 <dim>4</dim>
3283 </port>
3284 </output>
3285 </layer>
3286 <layer id="204" name="/distilbert/transformer/layer.1/attention/Reshape" type="Reshape" version="opset1">
3287 <data special_zero="true" />
3288 <rt_info>
3289 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Reshape" />
3290 </rt_info>
3291 <input>
3292 <port id="0" precision="FP32">
3293 <dim>-1</dim>
3294 <dim>-1</dim>
3295 <dim>768</dim>
3296 </port>
3297 <port id="1" precision="I64">
3298 <dim>4</dim>
3299 </port>
3300 </input>
3301 <output>
3302 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Reshape_output_0">
3303 <dim>-1</dim>
3304 <dim>-1</dim>
3305 <dim>12</dim>
3306 <dim>64</dim>
3307 </port>
3308 </output>
3309 </layer>
3310 <layer id="205" name="Constant_255407" type="Const" version="opset1">
3311 <data element_type="i64" shape="4" offset="93131832" size="32" />
3312 <rt_info>
3313 <attribute name="fused_names" version="0" value="Constant_255407" />
3314 </rt_info>
3315 <output>
3316 <port id="0" precision="I64">
3317 <dim>4</dim>
3318 </port>
3319 </output>
3320 </layer>
3321 <layer id="206" name="/distilbert/transformer/layer.1/attention/Transpose" type="Transpose" version="opset1">
3322 <rt_info>
3323 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Transpose" />
3324 </rt_info>
3325 <input>
3326 <port id="0" precision="FP32">
3327 <dim>-1</dim>
3328 <dim>-1</dim>
3329 <dim>12</dim>
3330 <dim>64</dim>
3331 </port>
3332 <port id="1" precision="I64">
3333 <dim>4</dim>
3334 </port>
3335 </input>
3336 <output>
3337 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Transpose_output_0">
3338 <dim>-1</dim>
3339 <dim>12</dim>
3340 <dim>-1</dim>
3341 <dim>64</dim>
3342 </port>
3343 </output>
3344 </layer>
3345 <layer id="207" name="/distilbert/transformer/layer.1/attention/Constant_11" type="Const" version="opset1">
3346 <data element_type="f32" shape="" offset="93131864" size="4" />
3347 <rt_info>
3348 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_11" />
3349 </rt_info>
3350 <output>
3351 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/attention/Constant_11_output_0" />
3352 </output>
3353 </layer>
3354 <layer id="208" name="/distilbert/transformer/layer.1/attention/Div" type="Divide" version="opset1">
3355 <data auto_broadcast="numpy" m_pythondiv="true" />
3356 <rt_info>
3357 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Div" />
3358 </rt_info>
3359 <input>
3360 <port id="0" precision="FP32">
3361 <dim>-1</dim>
3362 <dim>12</dim>
3363 <dim>-1</dim>
3364 <dim>64</dim>
3365 </port>
3366 <port id="1" precision="FP32" />
3367 </input>
3368 <output>
3369 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Div_output_0">
3370 <dim>-1</dim>
3371 <dim>12</dim>
3372 <dim>-1</dim>
3373 <dim>64</dim>
3374 </port>
3375 </output>
3376 </layer>
3377 <layer id="209" name="distilbert.transformer.layer.1.attention.k_lin.bias" type="Const" version="opset1">
3378 <data element_type="f32" shape="768" offset="121464980" size="3072" />
3379 <rt_info>
3380 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.attention.k_lin.bias" />
3381 </rt_info>
3382 <output>
3383 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.attention.k_lin.bias">
3384 <dim>768</dim>
3385 </port>
3386 </output>
3387 </layer>
3388 <layer id="210" name="onnx::MatMul_767" type="Const" version="opset1">
3389 <data element_type="f32" shape="768, 768" offset="121468052" size="2359296" />
3390 <rt_info>
3391 <attribute name="fused_names" version="0" value="onnx::MatMul_767" />
3392 </rt_info>
3393 <output>
3394 <port id="0" precision="FP32" names="onnx::MatMul_767">
3395 <dim>768</dim>
3396 <dim>768</dim>
3397 </port>
3398 </output>
3399 </layer>
3400 <layer id="211" name="/distilbert/transformer/layer.1/attention/k_lin/MatMul" type="MatMul" version="opset1">
3401 <data transpose_a="false" transpose_b="false" />
3402 <rt_info>
3403 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/k_lin/MatMul" />
3404 </rt_info>
3405 <input>
3406 <port id="0" precision="FP32">
3407 <dim>-1</dim>
3408 <dim>-1</dim>
3409 <dim>768</dim>
3410 </port>
3411 <port id="1" precision="FP32">
3412 <dim>768</dim>
3413 <dim>768</dim>
3414 </port>
3415 </input>
3416 <output>
3417 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/k_lin/MatMul_output_0">
3418 <dim>-1</dim>
3419 <dim>-1</dim>
3420 <dim>768</dim>
3421 </port>
3422 </output>
3423 </layer>
3424 <layer id="212" name="/distilbert/transformer/layer.1/attention/k_lin/Add" type="Add" version="opset1">
3425 <data auto_broadcast="numpy" />
3426 <rt_info>
3427 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/k_lin/Add" />
3428 </rt_info>
3429 <input>
3430 <port id="0" precision="FP32">
3431 <dim>768</dim>
3432 </port>
3433 <port id="1" precision="FP32">
3434 <dim>-1</dim>
3435 <dim>-1</dim>
3436 <dim>768</dim>
3437 </port>
3438 </input>
3439 <output>
3440 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/k_lin/Add_output_0">
3441 <dim>-1</dim>
3442 <dim>-1</dim>
3443 <dim>768</dim>
3444 </port>
3445 </output>
3446 </layer>
3447 <layer id="213" name="Constant_255388" type="Const" version="opset1">
3448 <data element_type="i64" shape="1" offset="18440" size="8" />
3449 <rt_info>
3450 <attribute name="fused_names" version="0" value="Constant_255388" />
3451 </rt_info>
3452 <output>
3453 <port id="0" precision="I64">
3454 <dim>1</dim>
3455 </port>
3456 </output>
3457 </layer>
3458 <layer id="214" name="/distilbert/transformer/layer.1/attention/Unsqueeze_1" type="Unsqueeze" version="opset1">
3459 <rt_info>
3460 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Unsqueeze_1, Constant_255388" />
3461 </rt_info>
3462 <input>
3463 <port id="0" precision="I64" />
3464 <port id="1" precision="I64">
3465 <dim>1</dim>
3466 </port>
3467 </input>
3468 <output>
3469 <port id="2" precision="I64" names="/distilbert/transformer/layer.1/attention/Unsqueeze_1_output_0">
3470 <dim>1</dim>
3471 </port>
3472 </output>
3473 </layer>
3474 <layer id="215" name="/distilbert/transformer/layer.1/attention/Constant_5" type="Const" version="opset1">
3475 <data element_type="i64" shape="1" offset="90763288" size="8" />
3476 <rt_info>
3477 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_5" />
3478 </rt_info>
3479 <output>
3480 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_5_output_0">
3481 <dim>1</dim>
3482 </port>
3483 </output>
3484 </layer>
3485 <layer id="216" name="/distilbert/transformer/layer.1/attention/Constant_6" type="Const" version="opset1">
3486 <data element_type="i64" shape="1" offset="93131816" size="8" />
3487 <rt_info>
3488 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_6" />
3489 </rt_info>
3490 <output>
3491 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_6_output_0">
3492 <dim>1</dim>
3493 </port>
3494 </output>
3495 </layer>
3496 <layer id="217" name="/distilbert/transformer/layer.1/attention/Constant_7" type="Const" version="opset1">
3497 <data element_type="i64" shape="1" offset="93131824" size="8" />
3498 <rt_info>
3499 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_7" />
3500 </rt_info>
3501 <output>
3502 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_7_output_0">
3503 <dim>1</dim>
3504 </port>
3505 </output>
3506 </layer>
3507 <layer id="218" name="/distilbert/transformer/layer.1/attention/Concat_1" type="Concat" version="opset1">
3508 <data axis="0" />
3509 <rt_info>
3510 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Concat_1, /distilbert/transformer/layer.1/attention/Constant_5, /distilbert/transformer/layer.1/attention/Constant_6, /distilbert/transformer/layer.1/attention/Constant_7" />
3511 </rt_info>
3512 <input>
3513 <port id="0" precision="I64">
3514 <dim>1</dim>
3515 </port>
3516 <port id="1" precision="I64">
3517 <dim>1</dim>
3518 </port>
3519 <port id="2" precision="I64">
3520 <dim>1</dim>
3521 </port>
3522 <port id="3" precision="I64">
3523 <dim>1</dim>
3524 </port>
3525 </input>
3526 <output>
3527 <port id="4" precision="I64" names="/distilbert/transformer/layer.1/attention/Concat_1_output_0">
3528 <dim>4</dim>
3529 </port>
3530 </output>
3531 </layer>
3532 <layer id="219" name="/distilbert/transformer/layer.1/attention/Reshape_1" type="Reshape" version="opset1">
3533 <data special_zero="true" />
3534 <rt_info>
3535 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Reshape_1" />
3536 </rt_info>
3537 <input>
3538 <port id="0" precision="FP32">
3539 <dim>-1</dim>
3540 <dim>-1</dim>
3541 <dim>768</dim>
3542 </port>
3543 <port id="1" precision="I64">
3544 <dim>4</dim>
3545 </port>
3546 </input>
3547 <output>
3548 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Reshape_1_output_0">
3549 <dim>-1</dim>
3550 <dim>-1</dim>
3551 <dim>12</dim>
3552 <dim>64</dim>
3553 </port>
3554 </output>
3555 </layer>
3556 <layer id="220" name="Constant_255431" type="Const" version="opset1">
3557 <data element_type="i64" shape="4" offset="95494236" size="32" />
3558 <rt_info>
3559 <attribute name="fused_names" version="0" value="Constant_255431" />
3560 </rt_info>
3561 <output>
3562 <port id="0" precision="I64">
3563 <dim>4</dim>
3564 </port>
3565 </output>
3566 </layer>
3567 <layer id="221" name="/distilbert/transformer/layer.1/attention/Transpose_2" type="Transpose" version="opset1">
3568 <rt_info>
3569 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Transpose_2" />
3570 </rt_info>
3571 <input>
3572 <port id="0" precision="FP32">
3573 <dim>-1</dim>
3574 <dim>-1</dim>
3575 <dim>12</dim>
3576 <dim>64</dim>
3577 </port>
3578 <port id="1" precision="I64">
3579 <dim>4</dim>
3580 </port>
3581 </input>
3582 <output>
3583 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Transpose_2_output_0">
3584 <dim>-1</dim>
3585 <dim>12</dim>
3586 <dim>64</dim>
3587 <dim>-1</dim>
3588 </port>
3589 </output>
3590 </layer>
3591 <layer id="222" name="/distilbert/transformer/layer.1/attention/MatMul" type="MatMul" version="opset1">
3592 <data transpose_a="false" transpose_b="false" />
3593 <rt_info>
3594 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/MatMul" />
3595 </rt_info>
3596 <input>
3597 <port id="0" precision="FP32">
3598 <dim>-1</dim>
3599 <dim>12</dim>
3600 <dim>-1</dim>
3601 <dim>64</dim>
3602 </port>
3603 <port id="1" precision="FP32">
3604 <dim>-1</dim>
3605 <dim>12</dim>
3606 <dim>64</dim>
3607 <dim>-1</dim>
3608 </port>
3609 </input>
3610 <output>
3611 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/MatMul_output_0">
3612 <dim>-1</dim>
3613 <dim>12</dim>
3614 <dim>-1</dim>
3615 <dim>-1</dim>
3616 </port>
3617 </output>
3618 </layer>
3619 <layer id="223" name="/distilbert/transformer/layer.1/attention/Shape_2" type="ShapeOf" version="opset3">
3620 <data output_type="i64" />
3621 <rt_info>
3622 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Shape_2" />
3623 </rt_info>
3624 <input>
3625 <port id="0" precision="FP32">
3626 <dim>-1</dim>
3627 <dim>12</dim>
3628 <dim>-1</dim>
3629 <dim>-1</dim>
3630 </port>
3631 </input>
3632 <output>
3633 <port id="1" precision="I64" names="/distilbert/transformer/layer.1/attention/Shape_2_output_0">
3634 <dim>4</dim>
3635 </port>
3636 </output>
3637 </layer>
3638 <layer id="224" name="/distilbert/transformer/layer.1/attention/Expand" type="Broadcast" version="opset3">
3639 <data mode="bidirectional" />
3640 <rt_info>
3641 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Expand" />
3642 </rt_info>
3643 <input>
3644 <port id="0" precision="BOOL">
3645 <dim>-1</dim>
3646 <dim>1</dim>
3647 <dim>1</dim>
3648 <dim>-1</dim>
3649 </port>
3650 <port id="1" precision="I64">
3651 <dim>4</dim>
3652 </port>
3653 </input>
3654 <output>
3655 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.1/attention/Expand_output_0">
3656 <dim>-1</dim>
3657 <dim>12</dim>
3658 <dim>-1</dim>
3659 <dim>-1</dim>
3660 </port>
3661 </output>
3662 </layer>
3663 <layer id="225" name="/distilbert/transformer/layer.1/attention/Cast" type="Convert" version="opset1">
3664 <data destination_type="boolean" />
3665 <rt_info>
3666 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Cast" />
3667 </rt_info>
3668 <input>
3669 <port id="0" precision="BOOL">
3670 <dim>-1</dim>
3671 <dim>12</dim>
3672 <dim>-1</dim>
3673 <dim>-1</dim>
3674 </port>
3675 </input>
3676 <output>
3677 <port id="1" precision="BOOL" names="/distilbert/transformer/layer.1/attention/Cast_output_0">
3678 <dim>-1</dim>
3679 <dim>12</dim>
3680 <dim>-1</dim>
3681 <dim>-1</dim>
3682 </port>
3683 </output>
3684 </layer>
3685 <layer id="226" name="/distilbert/transformer/layer.1/attention/Constant_14" type="Const" version="opset1">
3686 <data element_type="f32" shape="" offset="95494268" size="4" />
3687 <rt_info>
3688 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_14" />
3689 </rt_info>
3690 <output>
3691 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/attention/Constant_14_output_0" />
3692 </output>
3693 </layer>
3694 <layer id="227" name="/distilbert/transformer/layer.1/attention/Where" type="Select" version="opset1">
3695 <data auto_broadcast="numpy" />
3696 <rt_info>
3697 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Where" />
3698 </rt_info>
3699 <input>
3700 <port id="0" precision="BOOL">
3701 <dim>-1</dim>
3702 <dim>12</dim>
3703 <dim>-1</dim>
3704 <dim>-1</dim>
3705 </port>
3706 <port id="1" precision="FP32" />
3707 <port id="2" precision="FP32">
3708 <dim>-1</dim>
3709 <dim>12</dim>
3710 <dim>-1</dim>
3711 <dim>-1</dim>
3712 </port>
3713 </input>
3714 <output>
3715 <port id="3" precision="FP32" names="/distilbert/transformer/layer.1/attention/Where_output_0">
3716 <dim>-1</dim>
3717 <dim>12</dim>
3718 <dim>-1</dim>
3719 <dim>-1</dim>
3720 </port>
3721 </output>
3722 </layer>
3723 <layer id="228" name="/distilbert/transformer/layer.1/attention/Softmax" type="SoftMax" version="opset8">
3724 <data axis="3" />
3725 <rt_info>
3726 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Softmax" />
3727 </rt_info>
3728 <input>
3729 <port id="0" precision="FP32">
3730 <dim>-1</dim>
3731 <dim>12</dim>
3732 <dim>-1</dim>
3733 <dim>-1</dim>
3734 </port>
3735 </input>
3736 <output>
3737 <port id="1" precision="FP32" names="/distilbert/transformer/layer.1/attention/Softmax_output_0">
3738 <dim>-1</dim>
3739 <dim>12</dim>
3740 <dim>-1</dim>
3741 <dim>-1</dim>
3742 </port>
3743 </output>
3744 </layer>
3745 <layer id="229" name="distilbert.transformer.layer.1.attention.v_lin.bias" type="Const" version="opset1">
3746 <data element_type="f32" shape="768" offset="123827348" size="3072" />
3747 <rt_info>
3748 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.attention.v_lin.bias" />
3749 </rt_info>
3750 <output>
3751 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.attention.v_lin.bias">
3752 <dim>768</dim>
3753 </port>
3754 </output>
3755 </layer>
3756 <layer id="230" name="onnx::MatMul_768" type="Const" version="opset1">
3757 <data element_type="f32" shape="768, 768" offset="123830420" size="2359296" />
3758 <rt_info>
3759 <attribute name="fused_names" version="0" value="onnx::MatMul_768" />
3760 </rt_info>
3761 <output>
3762 <port id="0" precision="FP32" names="onnx::MatMul_768">
3763 <dim>768</dim>
3764 <dim>768</dim>
3765 </port>
3766 </output>
3767 </layer>
3768 <layer id="231" name="/distilbert/transformer/layer.1/attention/v_lin/MatMul" type="MatMul" version="opset1">
3769 <data transpose_a="false" transpose_b="false" />
3770 <rt_info>
3771 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/v_lin/MatMul" />
3772 </rt_info>
3773 <input>
3774 <port id="0" precision="FP32">
3775 <dim>-1</dim>
3776 <dim>-1</dim>
3777 <dim>768</dim>
3778 </port>
3779 <port id="1" precision="FP32">
3780 <dim>768</dim>
3781 <dim>768</dim>
3782 </port>
3783 </input>
3784 <output>
3785 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/v_lin/MatMul_output_0">
3786 <dim>-1</dim>
3787 <dim>-1</dim>
3788 <dim>768</dim>
3789 </port>
3790 </output>
3791 </layer>
3792 <layer id="232" name="/distilbert/transformer/layer.1/attention/v_lin/Add" type="Add" version="opset1">
3793 <data auto_broadcast="numpy" />
3794 <rt_info>
3795 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/v_lin/Add" />
3796 </rt_info>
3797 <input>
3798 <port id="0" precision="FP32">
3799 <dim>768</dim>
3800 </port>
3801 <port id="1" precision="FP32">
3802 <dim>-1</dim>
3803 <dim>-1</dim>
3804 <dim>768</dim>
3805 </port>
3806 </input>
3807 <output>
3808 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/v_lin/Add_output_0">
3809 <dim>-1</dim>
3810 <dim>-1</dim>
3811 <dim>768</dim>
3812 </port>
3813 </output>
3814 </layer>
3815 <layer id="233" name="Constant_255394" type="Const" version="opset1">
3816 <data element_type="i64" shape="1" offset="18440" size="8" />
3817 <rt_info>
3818 <attribute name="fused_names" version="0" value="Constant_255394" />
3819 </rt_info>
3820 <output>
3821 <port id="0" precision="I64">
3822 <dim>1</dim>
3823 </port>
3824 </output>
3825 </layer>
3826 <layer id="234" name="/distilbert/transformer/layer.1/attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
3827 <rt_info>
3828 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Unsqueeze_2, Constant_255394" />
3829 </rt_info>
3830 <input>
3831 <port id="0" precision="I64" />
3832 <port id="1" precision="I64">
3833 <dim>1</dim>
3834 </port>
3835 </input>
3836 <output>
3837 <port id="2" precision="I64" names="/distilbert/transformer/layer.1/attention/Unsqueeze_2_output_0">
3838 <dim>1</dim>
3839 </port>
3840 </output>
3841 </layer>
3842 <layer id="235" name="/distilbert/transformer/layer.1/attention/Constant_8" type="Const" version="opset1">
3843 <data element_type="i64" shape="1" offset="90763288" size="8" />
3844 <rt_info>
3845 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_8" />
3846 </rt_info>
3847 <output>
3848 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_8_output_0">
3849 <dim>1</dim>
3850 </port>
3851 </output>
3852 </layer>
3853 <layer id="236" name="/distilbert/transformer/layer.1/attention/Constant_9" type="Const" version="opset1">
3854 <data element_type="i64" shape="1" offset="93131816" size="8" />
3855 <rt_info>
3856 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_9" />
3857 </rt_info>
3858 <output>
3859 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_9_output_0">
3860 <dim>1</dim>
3861 </port>
3862 </output>
3863 </layer>
3864 <layer id="237" name="/distilbert/transformer/layer.1/attention/Constant_10" type="Const" version="opset1">
3865 <data element_type="i64" shape="1" offset="93131824" size="8" />
3866 <rt_info>
3867 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_10" />
3868 </rt_info>
3869 <output>
3870 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_10_output_0">
3871 <dim>1</dim>
3872 </port>
3873 </output>
3874 </layer>
3875 <layer id="238" name="/distilbert/transformer/layer.1/attention/Concat_2" type="Concat" version="opset1">
3876 <data axis="0" />
3877 <rt_info>
3878 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Concat_2, /distilbert/transformer/layer.1/attention/Constant_10, /distilbert/transformer/layer.1/attention/Constant_8, /distilbert/transformer/layer.1/attention/Constant_9" />
3879 </rt_info>
3880 <input>
3881 <port id="0" precision="I64">
3882 <dim>1</dim>
3883 </port>
3884 <port id="1" precision="I64">
3885 <dim>1</dim>
3886 </port>
3887 <port id="2" precision="I64">
3888 <dim>1</dim>
3889 </port>
3890 <port id="3" precision="I64">
3891 <dim>1</dim>
3892 </port>
3893 </input>
3894 <output>
3895 <port id="4" precision="I64" names="/distilbert/transformer/layer.1/attention/Concat_2_output_0">
3896 <dim>4</dim>
3897 </port>
3898 </output>
3899 </layer>
3900 <layer id="239" name="/distilbert/transformer/layer.1/attention/Reshape_2" type="Reshape" version="opset1">
3901 <data special_zero="true" />
3902 <rt_info>
3903 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Reshape_2" />
3904 </rt_info>
3905 <input>
3906 <port id="0" precision="FP32">
3907 <dim>-1</dim>
3908 <dim>-1</dim>
3909 <dim>768</dim>
3910 </port>
3911 <port id="1" precision="I64">
3912 <dim>4</dim>
3913 </port>
3914 </input>
3915 <output>
3916 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Reshape_2_output_0">
3917 <dim>-1</dim>
3918 <dim>-1</dim>
3919 <dim>12</dim>
3920 <dim>64</dim>
3921 </port>
3922 </output>
3923 </layer>
3924 <layer id="240" name="Constant_255427" type="Const" version="opset1">
3925 <data element_type="i64" shape="4" offset="93131832" size="32" />
3926 <rt_info>
3927 <attribute name="fused_names" version="0" value="Constant_255427" />
3928 </rt_info>
3929 <output>
3930 <port id="0" precision="I64">
3931 <dim>4</dim>
3932 </port>
3933 </output>
3934 </layer>
3935 <layer id="241" name="/distilbert/transformer/layer.1/attention/Transpose_1" type="Transpose" version="opset1">
3936 <rt_info>
3937 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Transpose_1" />
3938 </rt_info>
3939 <input>
3940 <port id="0" precision="FP32">
3941 <dim>-1</dim>
3942 <dim>-1</dim>
3943 <dim>12</dim>
3944 <dim>64</dim>
3945 </port>
3946 <port id="1" precision="I64">
3947 <dim>4</dim>
3948 </port>
3949 </input>
3950 <output>
3951 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Transpose_1_output_0">
3952 <dim>-1</dim>
3953 <dim>12</dim>
3954 <dim>-1</dim>
3955 <dim>64</dim>
3956 </port>
3957 </output>
3958 </layer>
3959 <layer id="242" name="/distilbert/transformer/layer.1/attention/MatMul_1" type="MatMul" version="opset1">
3960 <data transpose_a="false" transpose_b="false" />
3961 <rt_info>
3962 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/MatMul_1" />
3963 </rt_info>
3964 <input>
3965 <port id="0" precision="FP32">
3966 <dim>-1</dim>
3967 <dim>12</dim>
3968 <dim>-1</dim>
3969 <dim>-1</dim>
3970 </port>
3971 <port id="1" precision="FP32">
3972 <dim>-1</dim>
3973 <dim>12</dim>
3974 <dim>-1</dim>
3975 <dim>64</dim>
3976 </port>
3977 </input>
3978 <output>
3979 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/MatMul_1_output_0">
3980 <dim>-1</dim>
3981 <dim>12</dim>
3982 <dim>-1</dim>
3983 <dim>64</dim>
3984 </port>
3985 </output>
3986 </layer>
3987 <layer id="243" name="Constant_255463" type="Const" version="opset1">
3988 <data element_type="i64" shape="4" offset="93131832" size="32" />
3989 <rt_info>
3990 <attribute name="fused_names" version="0" value="Constant_255463" />
3991 </rt_info>
3992 <output>
3993 <port id="0" precision="I64">
3994 <dim>4</dim>
3995 </port>
3996 </output>
3997 </layer>
3998 <layer id="244" name="/distilbert/transformer/layer.1/attention/Transpose_3" type="Transpose" version="opset1">
3999 <rt_info>
4000 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Transpose_3" />
4001 </rt_info>
4002 <input>
4003 <port id="0" precision="FP32">
4004 <dim>-1</dim>
4005 <dim>12</dim>
4006 <dim>-1</dim>
4007 <dim>64</dim>
4008 </port>
4009 <port id="1" precision="I64">
4010 <dim>4</dim>
4011 </port>
4012 </input>
4013 <output>
4014 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Transpose_3_output_0">
4015 <dim>-1</dim>
4016 <dim>-1</dim>
4017 <dim>12</dim>
4018 <dim>64</dim>
4019 </port>
4020 </output>
4021 </layer>
4022 <layer id="245" name="Constant_255465" type="Const" version="opset1">
4023 <data element_type="i64" shape="1" offset="18440" size="8" />
4024 <rt_info>
4025 <attribute name="fused_names" version="0" value="Constant_255465" />
4026 </rt_info>
4027 <output>
4028 <port id="0" precision="I64">
4029 <dim>1</dim>
4030 </port>
4031 </output>
4032 </layer>
4033 <layer id="246" name="/distilbert/transformer/layer.1/attention/Unsqueeze_5" type="Unsqueeze" version="opset1">
4034 <rt_info>
4035 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Unsqueeze_5, Constant_255465" />
4036 </rt_info>
4037 <input>
4038 <port id="0" precision="I64" />
4039 <port id="1" precision="I64">
4040 <dim>1</dim>
4041 </port>
4042 </input>
4043 <output>
4044 <port id="2" precision="I64" names="/distilbert/transformer/layer.1/attention/Unsqueeze_5_output_0">
4045 <dim>1</dim>
4046 </port>
4047 </output>
4048 </layer>
4049 <layer id="247" name="/distilbert/transformer/layer.1/attention/Constant_15" type="Const" version="opset1">
4050 <data element_type="i64" shape="1" offset="90763288" size="8" />
4051 <rt_info>
4052 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_15" />
4053 </rt_info>
4054 <output>
4055 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_15_output_0">
4056 <dim>1</dim>
4057 </port>
4058 </output>
4059 </layer>
4060 <layer id="248" name="/distilbert/transformer/layer.1/attention/Constant_16" type="Const" version="opset1">
4061 <data element_type="i64" shape="1" offset="97856640" size="8" />
4062 <rt_info>
4063 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Constant_16" />
4064 </rt_info>
4065 <output>
4066 <port id="0" precision="I64" names="/distilbert/transformer/layer.1/attention/Constant_16_output_0">
4067 <dim>1</dim>
4068 </port>
4069 </output>
4070 </layer>
4071 <layer id="249" name="/distilbert/transformer/layer.1/attention/Concat_4" type="Concat" version="opset1">
4072 <data axis="0" />
4073 <rt_info>
4074 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Concat_4, /distilbert/transformer/layer.1/attention/Constant_15, /distilbert/transformer/layer.1/attention/Constant_16" />
4075 </rt_info>
4076 <input>
4077 <port id="0" precision="I64">
4078 <dim>1</dim>
4079 </port>
4080 <port id="1" precision="I64">
4081 <dim>1</dim>
4082 </port>
4083 <port id="2" precision="I64">
4084 <dim>1</dim>
4085 </port>
4086 </input>
4087 <output>
4088 <port id="3" precision="I64" names="/distilbert/transformer/layer.1/attention/Concat_4_output_0">
4089 <dim>3</dim>
4090 </port>
4091 </output>
4092 </layer>
4093 <layer id="250" name="/distilbert/transformer/layer.1/attention/Reshape_4" type="Reshape" version="opset1">
4094 <data special_zero="true" />
4095 <rt_info>
4096 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/Reshape_4" />
4097 </rt_info>
4098 <input>
4099 <port id="0" precision="FP32">
4100 <dim>-1</dim>
4101 <dim>-1</dim>
4102 <dim>12</dim>
4103 <dim>64</dim>
4104 </port>
4105 <port id="1" precision="I64">
4106 <dim>3</dim>
4107 </port>
4108 </input>
4109 <output>
4110 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/Reshape_4_output_0">
4111 <dim>-1</dim>
4112 <dim>-1</dim>
4113 <dim>768</dim>
4114 </port>
4115 </output>
4116 </layer>
4117 <layer id="251" name="onnx::MatMul_773" type="Const" version="opset1">
4118 <data element_type="f32" shape="768, 768" offset="126189716" size="2359296" />
4119 <rt_info>
4120 <attribute name="fused_names" version="0" value="onnx::MatMul_773" />
4121 </rt_info>
4122 <output>
4123 <port id="0" precision="FP32" names="onnx::MatMul_773">
4124 <dim>768</dim>
4125 <dim>768</dim>
4126 </port>
4127 </output>
4128 </layer>
4129 <layer id="252" name="/distilbert/transformer/layer.1/attention/out_lin/MatMul" type="MatMul" version="opset1">
4130 <data transpose_a="false" transpose_b="false" />
4131 <rt_info>
4132 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/out_lin/MatMul" />
4133 </rt_info>
4134 <input>
4135 <port id="0" precision="FP32">
4136 <dim>-1</dim>
4137 <dim>-1</dim>
4138 <dim>768</dim>
4139 </port>
4140 <port id="1" precision="FP32">
4141 <dim>768</dim>
4142 <dim>768</dim>
4143 </port>
4144 </input>
4145 <output>
4146 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/out_lin/MatMul_output_0">
4147 <dim>-1</dim>
4148 <dim>-1</dim>
4149 <dim>768</dim>
4150 </port>
4151 </output>
4152 </layer>
4153 <layer id="253" name="/distilbert/transformer/layer.1/attention/out_lin/Add" type="Add" version="opset1">
4154 <data auto_broadcast="numpy" />
4155 <rt_info>
4156 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/attention/out_lin/Add" />
4157 </rt_info>
4158 <input>
4159 <port id="0" precision="FP32">
4160 <dim>768</dim>
4161 </port>
4162 <port id="1" precision="FP32">
4163 <dim>-1</dim>
4164 <dim>-1</dim>
4165 <dim>768</dim>
4166 </port>
4167 </input>
4168 <output>
4169 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/attention/out_lin/Add_output_0">
4170 <dim>-1</dim>
4171 <dim>-1</dim>
4172 <dim>768</dim>
4173 </port>
4174 </output>
4175 </layer>
4176 <layer id="254" name="/distilbert/transformer/layer.1/Add" type="Add" version="opset1">
4177 <data auto_broadcast="numpy" />
4178 <rt_info>
4179 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/Add" />
4180 </rt_info>
4181 <input>
4182 <port id="0" precision="FP32">
4183 <dim>-1</dim>
4184 <dim>-1</dim>
4185 <dim>768</dim>
4186 </port>
4187 <port id="1" precision="FP32">
4188 <dim>-1</dim>
4189 <dim>-1</dim>
4190 <dim>768</dim>
4191 </port>
4192 </input>
4193 <output>
4194 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/Add_output_0">
4195 <dim>-1</dim>
4196 <dim>-1</dim>
4197 <dim>768</dim>
4198 </port>
4199 </output>
4200 </layer>
4201 <layer id="255" name="Constant_255480" type="Const" version="opset1">
4202 <data element_type="i64" shape="1" offset="90763288" size="8" />
4203 <rt_info>
4204 <attribute name="fused_names" version="0" value="Constant_255480" />
4205 </rt_info>
4206 <output>
4207 <port id="0" precision="I64">
4208 <dim>1</dim>
4209 </port>
4210 </output>
4211 </layer>
4212 <layer id="256" name="/distilbert/transformer/layer.1/sa_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
4213 <data keep_dims="true" />
4214 <rt_info>
4215 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/ReduceMean" />
4216 </rt_info>
4217 <input>
4218 <port id="0" precision="FP32">
4219 <dim>-1</dim>
4220 <dim>-1</dim>
4221 <dim>768</dim>
4222 </port>
4223 <port id="1" precision="I64">
4224 <dim>1</dim>
4225 </port>
4226 </input>
4227 <output>
4228 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/ReduceMean_output_0">
4229 <dim>-1</dim>
4230 <dim>-1</dim>
4231 <dim>1</dim>
4232 </port>
4233 </output>
4234 </layer>
4235 <layer id="257" name="/distilbert/transformer/layer.1/sa_layer_norm/Sub" type="Subtract" version="opset1">
4236 <data auto_broadcast="numpy" />
4237 <rt_info>
4238 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Sub" />
4239 </rt_info>
4240 <input>
4241 <port id="0" precision="FP32">
4242 <dim>-1</dim>
4243 <dim>-1</dim>
4244 <dim>768</dim>
4245 </port>
4246 <port id="1" precision="FP32">
4247 <dim>-1</dim>
4248 <dim>-1</dim>
4249 <dim>1</dim>
4250 </port>
4251 </input>
4252 <output>
4253 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Sub_output_0">
4254 <dim>-1</dim>
4255 <dim>-1</dim>
4256 <dim>768</dim>
4257 </port>
4258 </output>
4259 </layer>
4260 <layer id="258" name="/distilbert/transformer/layer.1/sa_layer_norm/Constant" type="Const" version="opset1">
4261 <data element_type="f32" shape="" offset="90763296" size="4" />
4262 <rt_info>
4263 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Constant" />
4264 </rt_info>
4265 <output>
4266 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Constant_output_0" />
4267 </output>
4268 </layer>
4269 <layer id="259" name="/distilbert/transformer/layer.1/sa_layer_norm/Pow" type="Power" version="opset1">
4270 <data auto_broadcast="numpy" />
4271 <rt_info>
4272 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Pow" />
4273 </rt_info>
4274 <input>
4275 <port id="0" precision="FP32">
4276 <dim>-1</dim>
4277 <dim>-1</dim>
4278 <dim>768</dim>
4279 </port>
4280 <port id="1" precision="FP32" />
4281 </input>
4282 <output>
4283 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Pow_output_0">
4284 <dim>-1</dim>
4285 <dim>-1</dim>
4286 <dim>768</dim>
4287 </port>
4288 </output>
4289 </layer>
4290 <layer id="260" name="Constant_255485" type="Const" version="opset1">
4291 <data element_type="i64" shape="1" offset="90763288" size="8" />
4292 <rt_info>
4293 <attribute name="fused_names" version="0" value="Constant_255485" />
4294 </rt_info>
4295 <output>
4296 <port id="0" precision="I64">
4297 <dim>1</dim>
4298 </port>
4299 </output>
4300 </layer>
4301 <layer id="261" name="/distilbert/transformer/layer.1/sa_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
4302 <data keep_dims="true" />
4303 <rt_info>
4304 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/ReduceMean_1" />
4305 </rt_info>
4306 <input>
4307 <port id="0" precision="FP32">
4308 <dim>-1</dim>
4309 <dim>-1</dim>
4310 <dim>768</dim>
4311 </port>
4312 <port id="1" precision="I64">
4313 <dim>1</dim>
4314 </port>
4315 </input>
4316 <output>
4317 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/ReduceMean_1_output_0">
4318 <dim>-1</dim>
4319 <dim>-1</dim>
4320 <dim>1</dim>
4321 </port>
4322 </output>
4323 </layer>
4324 <layer id="262" name="/distilbert/transformer/layer.1/sa_layer_norm/Constant_1" type="Const" version="opset1">
4325 <data element_type="f32" shape="" offset="90763300" size="4" />
4326 <rt_info>
4327 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Constant_1" />
4328 </rt_info>
4329 <output>
4330 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Constant_1_output_0" />
4331 </output>
4332 </layer>
4333 <layer id="263" name="/distilbert/transformer/layer.1/sa_layer_norm/Add" type="Add" version="opset1">
4334 <data auto_broadcast="numpy" />
4335 <rt_info>
4336 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Add" />
4337 </rt_info>
4338 <input>
4339 <port id="0" precision="FP32">
4340 <dim>-1</dim>
4341 <dim>-1</dim>
4342 <dim>1</dim>
4343 </port>
4344 <port id="1" precision="FP32" />
4345 </input>
4346 <output>
4347 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Add_output_0">
4348 <dim>-1</dim>
4349 <dim>-1</dim>
4350 <dim>1</dim>
4351 </port>
4352 </output>
4353 </layer>
4354 <layer id="264" name="/distilbert/transformer/layer.1/sa_layer_norm/Sqrt" type="Sqrt" version="opset1">
4355 <rt_info>
4356 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Sqrt" />
4357 </rt_info>
4358 <input>
4359 <port id="0" precision="FP32">
4360 <dim>-1</dim>
4361 <dim>-1</dim>
4362 <dim>1</dim>
4363 </port>
4364 </input>
4365 <output>
4366 <port id="1" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Sqrt_output_0">
4367 <dim>-1</dim>
4368 <dim>-1</dim>
4369 <dim>1</dim>
4370 </port>
4371 </output>
4372 </layer>
4373 <layer id="265" name="/distilbert/transformer/layer.1/sa_layer_norm/Div" type="Divide" version="opset1">
4374 <data auto_broadcast="numpy" m_pythondiv="true" />
4375 <rt_info>
4376 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Div" />
4377 </rt_info>
4378 <input>
4379 <port id="0" precision="FP32">
4380 <dim>-1</dim>
4381 <dim>-1</dim>
4382 <dim>768</dim>
4383 </port>
4384 <port id="1" precision="FP32">
4385 <dim>-1</dim>
4386 <dim>-1</dim>
4387 <dim>1</dim>
4388 </port>
4389 </input>
4390 <output>
4391 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Div_output_0">
4392 <dim>-1</dim>
4393 <dim>-1</dim>
4394 <dim>768</dim>
4395 </port>
4396 </output>
4397 </layer>
4398 <layer id="266" name="distilbert.transformer.layer.1.sa_layer_norm.weight" type="Const" version="opset1">
4399 <data element_type="f32" shape="768" offset="128549012" size="3072" />
4400 <rt_info>
4401 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.sa_layer_norm.weight" />
4402 </rt_info>
4403 <output>
4404 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.sa_layer_norm.weight">
4405 <dim>768</dim>
4406 </port>
4407 </output>
4408 </layer>
4409 <layer id="267" name="/distilbert/transformer/layer.1/sa_layer_norm/Mul" type="Multiply" version="opset1">
4410 <data auto_broadcast="numpy" />
4411 <rt_info>
4412 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Mul" />
4413 </rt_info>
4414 <input>
4415 <port id="0" precision="FP32">
4416 <dim>-1</dim>
4417 <dim>-1</dim>
4418 <dim>768</dim>
4419 </port>
4420 <port id="1" precision="FP32">
4421 <dim>768</dim>
4422 </port>
4423 </input>
4424 <output>
4425 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Mul_output_0">
4426 <dim>-1</dim>
4427 <dim>-1</dim>
4428 <dim>768</dim>
4429 </port>
4430 </output>
4431 </layer>
4432 <layer id="268" name="distilbert.transformer.layer.1.sa_layer_norm.bias" type="Const" version="opset1">
4433 <data element_type="f32" shape="768" offset="128552084" size="3072" />
4434 <rt_info>
4435 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.sa_layer_norm.bias" />
4436 </rt_info>
4437 <output>
4438 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.sa_layer_norm.bias">
4439 <dim>768</dim>
4440 </port>
4441 </output>
4442 </layer>
4443 <layer id="269" name="/distilbert/transformer/layer.1/sa_layer_norm/Add_1" type="Add" version="opset1">
4444 <data auto_broadcast="numpy" />
4445 <rt_info>
4446 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/sa_layer_norm/Add_1" />
4447 </rt_info>
4448 <input>
4449 <port id="0" precision="FP32">
4450 <dim>-1</dim>
4451 <dim>-1</dim>
4452 <dim>768</dim>
4453 </port>
4454 <port id="1" precision="FP32">
4455 <dim>768</dim>
4456 </port>
4457 </input>
4458 <output>
4459 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/sa_layer_norm/Add_1_output_0">
4460 <dim>-1</dim>
4461 <dim>-1</dim>
4462 <dim>768</dim>
4463 </port>
4464 </output>
4465 </layer>
4466 <layer id="270" name="onnx::MatMul_774" type="Const" version="opset1">
4467 <data element_type="f32" shape="768, 3072" offset="128555156" size="9437184" />
4468 <rt_info>
4469 <attribute name="fused_names" version="0" value="onnx::MatMul_774" />
4470 </rt_info>
4471 <output>
4472 <port id="0" precision="FP32" names="onnx::MatMul_774">
4473 <dim>768</dim>
4474 <dim>3072</dim>
4475 </port>
4476 </output>
4477 </layer>
4478 <layer id="271" name="/distilbert/transformer/layer.1/ffn/lin1/MatMul" type="MatMul" version="opset1">
4479 <data transpose_a="false" transpose_b="false" />
4480 <rt_info>
4481 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/lin1/MatMul" />
4482 </rt_info>
4483 <input>
4484 <port id="0" precision="FP32">
4485 <dim>-1</dim>
4486 <dim>-1</dim>
4487 <dim>768</dim>
4488 </port>
4489 <port id="1" precision="FP32">
4490 <dim>768</dim>
4491 <dim>3072</dim>
4492 </port>
4493 </input>
4494 <output>
4495 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/lin1/MatMul_output_0">
4496 <dim>-1</dim>
4497 <dim>-1</dim>
4498 <dim>3072</dim>
4499 </port>
4500 </output>
4501 </layer>
4502 <layer id="272" name="/distilbert/transformer/layer.1/ffn/lin1/Add" type="Add" version="opset1">
4503 <data auto_broadcast="numpy" />
4504 <rt_info>
4505 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/lin1/Add" />
4506 </rt_info>
4507 <input>
4508 <port id="0" precision="FP32">
4509 <dim>3072</dim>
4510 </port>
4511 <port id="1" precision="FP32">
4512 <dim>-1</dim>
4513 <dim>-1</dim>
4514 <dim>3072</dim>
4515 </port>
4516 </input>
4517 <output>
4518 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/lin1/Add_output_0">
4519 <dim>-1</dim>
4520 <dim>-1</dim>
4521 <dim>3072</dim>
4522 </port>
4523 </output>
4524 </layer>
4525 <layer id="273" name="/distilbert/transformer/layer.1/ffn/activation/Constant" type="Const" version="opset1">
4526 <data element_type="f32" shape="" offset="109659272" size="4" />
4527 <rt_info>
4528 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Constant" />
4529 </rt_info>
4530 <output>
4531 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Constant_output_0" />
4532 </output>
4533 </layer>
4534 <layer id="274" name="/distilbert/transformer/layer.1/ffn/activation/Div" type="Divide" version="opset1">
4535 <data auto_broadcast="numpy" m_pythondiv="true" />
4536 <rt_info>
4537 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Div" />
4538 </rt_info>
4539 <input>
4540 <port id="0" precision="FP32">
4541 <dim>-1</dim>
4542 <dim>-1</dim>
4543 <dim>3072</dim>
4544 </port>
4545 <port id="1" precision="FP32" />
4546 </input>
4547 <output>
4548 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Div_output_0">
4549 <dim>-1</dim>
4550 <dim>-1</dim>
4551 <dim>3072</dim>
4552 </port>
4553 </output>
4554 </layer>
4555 <layer id="275" name="/distilbert/transformer/layer.1/ffn/activation/Erf" type="Erf" version="opset1">
4556 <rt_info>
4557 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Erf" />
4558 </rt_info>
4559 <input>
4560 <port id="0" precision="FP32">
4561 <dim>-1</dim>
4562 <dim>-1</dim>
4563 <dim>3072</dim>
4564 </port>
4565 </input>
4566 <output>
4567 <port id="1" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Erf_output_0">
4568 <dim>-1</dim>
4569 <dim>-1</dim>
4570 <dim>3072</dim>
4571 </port>
4572 </output>
4573 </layer>
4574 <layer id="276" name="/distilbert/transformer/layer.1/ffn/activation/Constant_1" type="Const" version="opset1">
4575 <data element_type="f32" shape="" offset="109659276" size="4" />
4576 <rt_info>
4577 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Constant_1" />
4578 </rt_info>
4579 <output>
4580 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Constant_1_output_0" />
4581 </output>
4582 </layer>
4583 <layer id="277" name="/distilbert/transformer/layer.1/ffn/activation/Add" type="Add" version="opset1">
4584 <data auto_broadcast="numpy" />
4585 <rt_info>
4586 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Add" />
4587 </rt_info>
4588 <input>
4589 <port id="0" precision="FP32">
4590 <dim>-1</dim>
4591 <dim>-1</dim>
4592 <dim>3072</dim>
4593 </port>
4594 <port id="1" precision="FP32" />
4595 </input>
4596 <output>
4597 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Add_output_0">
4598 <dim>-1</dim>
4599 <dim>-1</dim>
4600 <dim>3072</dim>
4601 </port>
4602 </output>
4603 </layer>
4604 <layer id="278" name="/distilbert/transformer/layer.1/ffn/activation/Mul" type="Multiply" version="opset1">
4605 <data auto_broadcast="numpy" />
4606 <rt_info>
4607 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Mul" />
4608 </rt_info>
4609 <input>
4610 <port id="0" precision="FP32">
4611 <dim>-1</dim>
4612 <dim>-1</dim>
4613 <dim>3072</dim>
4614 </port>
4615 <port id="1" precision="FP32">
4616 <dim>-1</dim>
4617 <dim>-1</dim>
4618 <dim>3072</dim>
4619 </port>
4620 </input>
4621 <output>
4622 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Mul_output_0">
4623 <dim>-1</dim>
4624 <dim>-1</dim>
4625 <dim>3072</dim>
4626 </port>
4627 </output>
4628 </layer>
4629 <layer id="279" name="/distilbert/transformer/layer.1/ffn/activation/Constant_2" type="Const" version="opset1">
4630 <data element_type="f32" shape="" offset="109659280" size="4" />
4631 <rt_info>
4632 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Constant_2" />
4633 </rt_info>
4634 <output>
4635 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Constant_2_output_0" />
4636 </output>
4637 </layer>
4638 <layer id="280" name="/distilbert/transformer/layer.1/ffn/activation/Mul_1" type="Multiply" version="opset1">
4639 <data auto_broadcast="numpy" />
4640 <rt_info>
4641 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/activation/Mul_1" />
4642 </rt_info>
4643 <input>
4644 <port id="0" precision="FP32">
4645 <dim>-1</dim>
4646 <dim>-1</dim>
4647 <dim>3072</dim>
4648 </port>
4649 <port id="1" precision="FP32" />
4650 </input>
4651 <output>
4652 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/activation/Mul_1_output_0">
4653 <dim>-1</dim>
4654 <dim>-1</dim>
4655 <dim>3072</dim>
4656 </port>
4657 </output>
4658 </layer>
4659 <layer id="281" name="onnx::MatMul_775" type="Const" version="opset1">
4660 <data element_type="f32" shape="3072, 768" offset="137992340" size="9437184" />
4661 <rt_info>
4662 <attribute name="fused_names" version="0" value="onnx::MatMul_775" />
4663 </rt_info>
4664 <output>
4665 <port id="0" precision="FP32" names="onnx::MatMul_775">
4666 <dim>3072</dim>
4667 <dim>768</dim>
4668 </port>
4669 </output>
4670 </layer>
4671 <layer id="282" name="/distilbert/transformer/layer.1/ffn/lin2/MatMul" type="MatMul" version="opset1">
4672 <data transpose_a="false" transpose_b="false" />
4673 <rt_info>
4674 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/lin2/MatMul" />
4675 </rt_info>
4676 <input>
4677 <port id="0" precision="FP32">
4678 <dim>-1</dim>
4679 <dim>-1</dim>
4680 <dim>3072</dim>
4681 </port>
4682 <port id="1" precision="FP32">
4683 <dim>3072</dim>
4684 <dim>768</dim>
4685 </port>
4686 </input>
4687 <output>
4688 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/lin2/MatMul_output_0">
4689 <dim>-1</dim>
4690 <dim>-1</dim>
4691 <dim>768</dim>
4692 </port>
4693 </output>
4694 </layer>
4695 <layer id="283" name="/distilbert/transformer/layer.1/ffn/lin2/Add" type="Add" version="opset1">
4696 <data auto_broadcast="numpy" />
4697 <rt_info>
4698 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/ffn/lin2/Add" />
4699 </rt_info>
4700 <input>
4701 <port id="0" precision="FP32">
4702 <dim>768</dim>
4703 </port>
4704 <port id="1" precision="FP32">
4705 <dim>-1</dim>
4706 <dim>-1</dim>
4707 <dim>768</dim>
4708 </port>
4709 </input>
4710 <output>
4711 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/ffn/lin2/Add_output_0">
4712 <dim>-1</dim>
4713 <dim>-1</dim>
4714 <dim>768</dim>
4715 </port>
4716 </output>
4717 </layer>
4718 <layer id="284" name="/distilbert/transformer/layer.1/Add_1" type="Add" version="opset1">
4719 <data auto_broadcast="numpy" />
4720 <rt_info>
4721 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/Add_1" />
4722 </rt_info>
4723 <input>
4724 <port id="0" precision="FP32">
4725 <dim>-1</dim>
4726 <dim>-1</dim>
4727 <dim>768</dim>
4728 </port>
4729 <port id="1" precision="FP32">
4730 <dim>-1</dim>
4731 <dim>-1</dim>
4732 <dim>768</dim>
4733 </port>
4734 </input>
4735 <output>
4736 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/Add_1_output_0">
4737 <dim>-1</dim>
4738 <dim>-1</dim>
4739 <dim>768</dim>
4740 </port>
4741 </output>
4742 </layer>
4743 <layer id="285" name="Constant_255506" type="Const" version="opset1">
4744 <data element_type="i64" shape="1" offset="90763288" size="8" />
4745 <rt_info>
4746 <attribute name="fused_names" version="0" value="Constant_255506" />
4747 </rt_info>
4748 <output>
4749 <port id="0" precision="I64">
4750 <dim>1</dim>
4751 </port>
4752 </output>
4753 </layer>
4754 <layer id="286" name="/distilbert/transformer/layer.1/output_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
4755 <data keep_dims="true" />
4756 <rt_info>
4757 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/ReduceMean" />
4758 </rt_info>
4759 <input>
4760 <port id="0" precision="FP32">
4761 <dim>-1</dim>
4762 <dim>-1</dim>
4763 <dim>768</dim>
4764 </port>
4765 <port id="1" precision="I64">
4766 <dim>1</dim>
4767 </port>
4768 </input>
4769 <output>
4770 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/ReduceMean_output_0">
4771 <dim>-1</dim>
4772 <dim>-1</dim>
4773 <dim>1</dim>
4774 </port>
4775 </output>
4776 </layer>
4777 <layer id="287" name="/distilbert/transformer/layer.1/output_layer_norm/Sub" type="Subtract" version="opset1">
4778 <data auto_broadcast="numpy" />
4779 <rt_info>
4780 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Sub" />
4781 </rt_info>
4782 <input>
4783 <port id="0" precision="FP32">
4784 <dim>-1</dim>
4785 <dim>-1</dim>
4786 <dim>768</dim>
4787 </port>
4788 <port id="1" precision="FP32">
4789 <dim>-1</dim>
4790 <dim>-1</dim>
4791 <dim>1</dim>
4792 </port>
4793 </input>
4794 <output>
4795 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Sub_output_0">
4796 <dim>-1</dim>
4797 <dim>-1</dim>
4798 <dim>768</dim>
4799 </port>
4800 </output>
4801 </layer>
4802 <layer id="288" name="/distilbert/transformer/layer.1/output_layer_norm/Constant" type="Const" version="opset1">
4803 <data element_type="f32" shape="" offset="90763296" size="4" />
4804 <rt_info>
4805 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Constant" />
4806 </rt_info>
4807 <output>
4808 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Constant_output_0" />
4809 </output>
4810 </layer>
4811 <layer id="289" name="/distilbert/transformer/layer.1/output_layer_norm/Pow" type="Power" version="opset1">
4812 <data auto_broadcast="numpy" />
4813 <rt_info>
4814 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Pow" />
4815 </rt_info>
4816 <input>
4817 <port id="0" precision="FP32">
4818 <dim>-1</dim>
4819 <dim>-1</dim>
4820 <dim>768</dim>
4821 </port>
4822 <port id="1" precision="FP32" />
4823 </input>
4824 <output>
4825 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Pow_output_0">
4826 <dim>-1</dim>
4827 <dim>-1</dim>
4828 <dim>768</dim>
4829 </port>
4830 </output>
4831 </layer>
4832 <layer id="290" name="Constant_255511" type="Const" version="opset1">
4833 <data element_type="i64" shape="1" offset="90763288" size="8" />
4834 <rt_info>
4835 <attribute name="fused_names" version="0" value="Constant_255511" />
4836 </rt_info>
4837 <output>
4838 <port id="0" precision="I64">
4839 <dim>1</dim>
4840 </port>
4841 </output>
4842 </layer>
4843 <layer id="291" name="/distilbert/transformer/layer.1/output_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
4844 <data keep_dims="true" />
4845 <rt_info>
4846 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/ReduceMean_1" />
4847 </rt_info>
4848 <input>
4849 <port id="0" precision="FP32">
4850 <dim>-1</dim>
4851 <dim>-1</dim>
4852 <dim>768</dim>
4853 </port>
4854 <port id="1" precision="I64">
4855 <dim>1</dim>
4856 </port>
4857 </input>
4858 <output>
4859 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/ReduceMean_1_output_0">
4860 <dim>-1</dim>
4861 <dim>-1</dim>
4862 <dim>1</dim>
4863 </port>
4864 </output>
4865 </layer>
4866 <layer id="292" name="/distilbert/transformer/layer.1/output_layer_norm/Constant_1" type="Const" version="opset1">
4867 <data element_type="f32" shape="" offset="90763300" size="4" />
4868 <rt_info>
4869 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Constant_1" />
4870 </rt_info>
4871 <output>
4872 <port id="0" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Constant_1_output_0" />
4873 </output>
4874 </layer>
4875 <layer id="293" name="/distilbert/transformer/layer.1/output_layer_norm/Add" type="Add" version="opset1">
4876 <data auto_broadcast="numpy" />
4877 <rt_info>
4878 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Add" />
4879 </rt_info>
4880 <input>
4881 <port id="0" precision="FP32">
4882 <dim>-1</dim>
4883 <dim>-1</dim>
4884 <dim>1</dim>
4885 </port>
4886 <port id="1" precision="FP32" />
4887 </input>
4888 <output>
4889 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Add_output_0">
4890 <dim>-1</dim>
4891 <dim>-1</dim>
4892 <dim>1</dim>
4893 </port>
4894 </output>
4895 </layer>
4896 <layer id="294" name="/distilbert/transformer/layer.1/output_layer_norm/Sqrt" type="Sqrt" version="opset1">
4897 <rt_info>
4898 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Sqrt" />
4899 </rt_info>
4900 <input>
4901 <port id="0" precision="FP32">
4902 <dim>-1</dim>
4903 <dim>-1</dim>
4904 <dim>1</dim>
4905 </port>
4906 </input>
4907 <output>
4908 <port id="1" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Sqrt_output_0">
4909 <dim>-1</dim>
4910 <dim>-1</dim>
4911 <dim>1</dim>
4912 </port>
4913 </output>
4914 </layer>
4915 <layer id="295" name="/distilbert/transformer/layer.1/output_layer_norm/Div" type="Divide" version="opset1">
4916 <data auto_broadcast="numpy" m_pythondiv="true" />
4917 <rt_info>
4918 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Div" />
4919 </rt_info>
4920 <input>
4921 <port id="0" precision="FP32">
4922 <dim>-1</dim>
4923 <dim>-1</dim>
4924 <dim>768</dim>
4925 </port>
4926 <port id="1" precision="FP32">
4927 <dim>-1</dim>
4928 <dim>-1</dim>
4929 <dim>1</dim>
4930 </port>
4931 </input>
4932 <output>
4933 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Div_output_0">
4934 <dim>-1</dim>
4935 <dim>-1</dim>
4936 <dim>768</dim>
4937 </port>
4938 </output>
4939 </layer>
4940 <layer id="296" name="distilbert.transformer.layer.1.output_layer_norm.weight" type="Const" version="opset1">
4941 <data element_type="f32" shape="768" offset="147429524" size="3072" />
4942 <rt_info>
4943 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.output_layer_norm.weight" />
4944 </rt_info>
4945 <output>
4946 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.output_layer_norm.weight">
4947 <dim>768</dim>
4948 </port>
4949 </output>
4950 </layer>
4951 <layer id="297" name="/distilbert/transformer/layer.1/output_layer_norm/Mul" type="Multiply" version="opset1">
4952 <data auto_broadcast="numpy" />
4953 <rt_info>
4954 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Mul" />
4955 </rt_info>
4956 <input>
4957 <port id="0" precision="FP32">
4958 <dim>-1</dim>
4959 <dim>-1</dim>
4960 <dim>768</dim>
4961 </port>
4962 <port id="1" precision="FP32">
4963 <dim>768</dim>
4964 </port>
4965 </input>
4966 <output>
4967 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Mul_output_0">
4968 <dim>-1</dim>
4969 <dim>-1</dim>
4970 <dim>768</dim>
4971 </port>
4972 </output>
4973 </layer>
4974 <layer id="298" name="distilbert.transformer.layer.1.output_layer_norm.bias" type="Const" version="opset1">
4975 <data element_type="f32" shape="768" offset="147432596" size="3072" />
4976 <rt_info>
4977 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.1.output_layer_norm.bias" />
4978 </rt_info>
4979 <output>
4980 <port id="0" precision="FP32" names="distilbert.transformer.layer.1.output_layer_norm.bias">
4981 <dim>768</dim>
4982 </port>
4983 </output>
4984 </layer>
4985 <layer id="299" name="/distilbert/transformer/layer.1/output_layer_norm/Add_1" type="Add" version="opset1">
4986 <data auto_broadcast="numpy" />
4987 <rt_info>
4988 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.1/output_layer_norm/Add_1" />
4989 </rt_info>
4990 <input>
4991 <port id="0" precision="FP32">
4992 <dim>-1</dim>
4993 <dim>-1</dim>
4994 <dim>768</dim>
4995 </port>
4996 <port id="1" precision="FP32">
4997 <dim>768</dim>
4998 </port>
4999 </input>
5000 <output>
5001 <port id="2" precision="FP32" names="/distilbert/transformer/layer.1/output_layer_norm/Add_1_output_0">
5002 <dim>-1</dim>
5003 <dim>-1</dim>
5004 <dim>768</dim>
5005 </port>
5006 </output>
5007 </layer>
5008 <layer id="300" name="/distilbert/transformer/layer.2/attention/Shape" type="ShapeOf" version="opset3">
5009 <data output_type="i64" />
5010 <rt_info>
5011 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Shape" />
5012 </rt_info>
5013 <input>
5014 <port id="0" precision="FP32">
5015 <dim>-1</dim>
5016 <dim>-1</dim>
5017 <dim>768</dim>
5018 </port>
5019 </input>
5020 <output>
5021 <port id="1" precision="I64" names="/distilbert/transformer/layer.2/attention/Shape_output_0">
5022 <dim>3</dim>
5023 </port>
5024 </output>
5025 </layer>
5026 <layer id="301" name="/distilbert/transformer/layer.2/attention/Constant" type="Const" version="opset1">
5027 <data element_type="i64" shape="" offset="18440" size="8" />
5028 <rt_info>
5029 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant" />
5030 </rt_info>
5031 <output>
5032 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_output_0" />
5033 </output>
5034 </layer>
5035 <layer id="302" name="Constant_255521" type="Const" version="opset1">
5036 <data element_type="i64" shape="" offset="18440" size="8" />
5037 <rt_info>
5038 <attribute name="fused_names" version="0" value="Constant_255521" />
5039 </rt_info>
5040 <output>
5041 <port id="0" precision="I64" />
5042 </output>
5043 </layer>
5044 <layer id="303" name="/distilbert/transformer/layer.2/attention/Gather" type="Gather" version="opset8">
5045 <data batch_dims="0" />
5046 <rt_info>
5047 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant, /distilbert/transformer/layer.2/attention/Gather, Constant_255521" />
5048 </rt_info>
5049 <input>
5050 <port id="0" precision="I64">
5051 <dim>3</dim>
5052 </port>
5053 <port id="1" precision="I64" />
5054 <port id="2" precision="I64" />
5055 </input>
5056 <output>
5057 <port id="3" precision="I64" names="/distilbert/transformer/layer.2/attention/Gather_output_0" />
5058 </output>
5059 </layer>
5060 <layer id="304" name="Constant_255581" type="Const" version="opset1">
5061 <data element_type="i64" shape="1" offset="18440" size="8" />
5062 <rt_info>
5063 <attribute name="fused_names" version="0" value="Constant_255581" />
5064 </rt_info>
5065 <output>
5066 <port id="0" precision="I64">
5067 <dim>1</dim>
5068 </port>
5069 </output>
5070 </layer>
5071 <layer id="305" name="/distilbert/transformer/layer.2/attention/Unsqueeze_3" type="Unsqueeze" version="opset1">
5072 <rt_info>
5073 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Unsqueeze_3, Constant_255581" />
5074 </rt_info>
5075 <input>
5076 <port id="0" precision="I64" />
5077 <port id="1" precision="I64">
5078 <dim>1</dim>
5079 </port>
5080 </input>
5081 <output>
5082 <port id="2" precision="I64" names="/distilbert/transformer/layer.2/attention/Unsqueeze_3_output_0">
5083 <dim>1</dim>
5084 </port>
5085 </output>
5086 </layer>
5087 <layer id="306" name="/distilbert/transformer/layer.2/attention/Constant_12" type="Const" version="opset1">
5088 <data element_type="i64" shape="1" offset="90763280" size="8" />
5089 <rt_info>
5090 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_12" />
5091 </rt_info>
5092 <output>
5093 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_12_output_0">
5094 <dim>1</dim>
5095 </port>
5096 </output>
5097 </layer>
5098 <layer id="307" name="/distilbert/transformer/layer.2/attention/Constant_13" type="Const" version="opset1">
5099 <data element_type="i64" shape="1" offset="90763280" size="8" />
5100 <rt_info>
5101 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_13" />
5102 </rt_info>
5103 <output>
5104 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_13_output_0">
5105 <dim>1</dim>
5106 </port>
5107 </output>
5108 </layer>
5109 <layer id="308" name="/distilbert/transformer/layer.2/attention/Shape_1" type="ShapeOf" version="opset3">
5110 <data output_type="i64" />
5111 <rt_info>
5112 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Shape_1" />
5113 </rt_info>
5114 <input>
5115 <port id="0" precision="FP32">
5116 <dim>-1</dim>
5117 <dim>-1</dim>
5118 <dim>768</dim>
5119 </port>
5120 </input>
5121 <output>
5122 <port id="1" precision="I64" names="/distilbert/transformer/layer.2/attention/Shape_1_output_0">
5123 <dim>3</dim>
5124 </port>
5125 </output>
5126 </layer>
5127 <layer id="309" name="/distilbert/transformer/layer.2/attention/Constant_1" type="Const" version="opset1">
5128 <data element_type="i64" shape="" offset="90763280" size="8" />
5129 <rt_info>
5130 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_1" />
5131 </rt_info>
5132 <output>
5133 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_1_output_0" />
5134 </output>
5135 </layer>
5136 <layer id="310" name="Constant_255525" type="Const" version="opset1">
5137 <data element_type="i64" shape="" offset="18440" size="8" />
5138 <rt_info>
5139 <attribute name="fused_names" version="0" value="Constant_255525" />
5140 </rt_info>
5141 <output>
5142 <port id="0" precision="I64" />
5143 </output>
5144 </layer>
5145 <layer id="311" name="/distilbert/transformer/layer.2/attention/Gather_1" type="Gather" version="opset8">
5146 <data batch_dims="0" />
5147 <rt_info>
5148 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_1, /distilbert/transformer/layer.2/attention/Gather_1, Constant_255525" />
5149 </rt_info>
5150 <input>
5151 <port id="0" precision="I64">
5152 <dim>3</dim>
5153 </port>
5154 <port id="1" precision="I64" />
5155 <port id="2" precision="I64" />
5156 </input>
5157 <output>
5158 <port id="3" precision="I64" names="/distilbert/transformer/layer.2/attention/Gather_1_output_0" />
5159 </output>
5160 </layer>
5161 <layer id="312" name="Constant_255585" type="Const" version="opset1">
5162 <data element_type="i64" shape="1" offset="18440" size="8" />
5163 <rt_info>
5164 <attribute name="fused_names" version="0" value="Constant_255585" />
5165 </rt_info>
5166 <output>
5167 <port id="0" precision="I64">
5168 <dim>1</dim>
5169 </port>
5170 </output>
5171 </layer>
5172 <layer id="313" name="/distilbert/transformer/layer.2/attention/Unsqueeze_4" type="Unsqueeze" version="opset1">
5173 <rt_info>
5174 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Unsqueeze_4, Constant_255585" />
5175 </rt_info>
5176 <input>
5177 <port id="0" precision="I64" />
5178 <port id="1" precision="I64">
5179 <dim>1</dim>
5180 </port>
5181 </input>
5182 <output>
5183 <port id="2" precision="I64" names="/distilbert/transformer/layer.2/attention/Unsqueeze_4_output_0">
5184 <dim>1</dim>
5185 </port>
5186 </output>
5187 </layer>
5188 <layer id="314" name="/distilbert/transformer/layer.2/attention/Concat_3" type="Concat" version="opset1">
5189 <data axis="0" />
5190 <rt_info>
5191 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Concat_3, /distilbert/transformer/layer.2/attention/Constant_12, /distilbert/transformer/layer.2/attention/Constant_13" />
5192 </rt_info>
5193 <input>
5194 <port id="0" precision="I64">
5195 <dim>1</dim>
5196 </port>
5197 <port id="1" precision="I64">
5198 <dim>1</dim>
5199 </port>
5200 <port id="2" precision="I64">
5201 <dim>1</dim>
5202 </port>
5203 <port id="3" precision="I64">
5204 <dim>1</dim>
5205 </port>
5206 </input>
5207 <output>
5208 <port id="4" precision="I64" names="/distilbert/transformer/layer.2/attention/Concat_3_output_0">
5209 <dim>4</dim>
5210 </port>
5211 </output>
5212 </layer>
5213 <layer id="315" name="/distilbert/transformer/layer.2/attention/Reshape_3" type="Reshape" version="opset1">
5214 <data special_zero="true" />
5215 <rt_info>
5216 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Reshape_3" />
5217 </rt_info>
5218 <input>
5219 <port id="0" precision="BOOL">
5220 <dim>-1</dim>
5221 <dim>-1</dim>
5222 </port>
5223 <port id="1" precision="I64">
5224 <dim>4</dim>
5225 </port>
5226 </input>
5227 <output>
5228 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.2/attention/Reshape_3_output_0">
5229 <dim>-1</dim>
5230 <dim>1</dim>
5231 <dim>1</dim>
5232 <dim>-1</dim>
5233 </port>
5234 </output>
5235 </layer>
5236 <layer id="316" name="distilbert.transformer.layer.2.attention.q_lin.bias" type="Const" version="opset1">
5237 <data element_type="f32" shape="768" offset="147435668" size="3072" />
5238 <rt_info>
5239 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.attention.q_lin.bias" />
5240 </rt_info>
5241 <output>
5242 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.attention.q_lin.bias">
5243 <dim>768</dim>
5244 </port>
5245 </output>
5246 </layer>
5247 <layer id="317" name="onnx::MatMul_776" type="Const" version="opset1">
5248 <data element_type="f32" shape="768, 768" offset="147438740" size="2359296" />
5249 <rt_info>
5250 <attribute name="fused_names" version="0" value="onnx::MatMul_776" />
5251 </rt_info>
5252 <output>
5253 <port id="0" precision="FP32" names="onnx::MatMul_776">
5254 <dim>768</dim>
5255 <dim>768</dim>
5256 </port>
5257 </output>
5258 </layer>
5259 <layer id="318" name="/distilbert/transformer/layer.2/attention/q_lin/MatMul" type="MatMul" version="opset1">
5260 <data transpose_a="false" transpose_b="false" />
5261 <rt_info>
5262 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/q_lin/MatMul" />
5263 </rt_info>
5264 <input>
5265 <port id="0" precision="FP32">
5266 <dim>-1</dim>
5267 <dim>-1</dim>
5268 <dim>768</dim>
5269 </port>
5270 <port id="1" precision="FP32">
5271 <dim>768</dim>
5272 <dim>768</dim>
5273 </port>
5274 </input>
5275 <output>
5276 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/q_lin/MatMul_output_0">
5277 <dim>-1</dim>
5278 <dim>-1</dim>
5279 <dim>768</dim>
5280 </port>
5281 </output>
5282 </layer>
5283 <layer id="319" name="/distilbert/transformer/layer.2/attention/q_lin/Add" type="Add" version="opset1">
5284 <data auto_broadcast="numpy" />
5285 <rt_info>
5286 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/q_lin/Add" />
5287 </rt_info>
5288 <input>
5289 <port id="0" precision="FP32">
5290 <dim>768</dim>
5291 </port>
5292 <port id="1" precision="FP32">
5293 <dim>-1</dim>
5294 <dim>-1</dim>
5295 <dim>768</dim>
5296 </port>
5297 </input>
5298 <output>
5299 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/q_lin/Add_output_0">
5300 <dim>-1</dim>
5301 <dim>-1</dim>
5302 <dim>768</dim>
5303 </port>
5304 </output>
5305 </layer>
5306 <layer id="320" name="Constant_255529" type="Const" version="opset1">
5307 <data element_type="i64" shape="1" offset="18440" size="8" />
5308 <rt_info>
5309 <attribute name="fused_names" version="0" value="Constant_255529" />
5310 </rt_info>
5311 <output>
5312 <port id="0" precision="I64">
5313 <dim>1</dim>
5314 </port>
5315 </output>
5316 </layer>
5317 <layer id="321" name="/distilbert/transformer/layer.2/attention/Unsqueeze" type="Unsqueeze" version="opset1">
5318 <rt_info>
5319 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Unsqueeze, Constant_255529" />
5320 </rt_info>
5321 <input>
5322 <port id="0" precision="I64" />
5323 <port id="1" precision="I64">
5324 <dim>1</dim>
5325 </port>
5326 </input>
5327 <output>
5328 <port id="2" precision="I64" names="/distilbert/transformer/layer.2/attention/Unsqueeze_output_0">
5329 <dim>1</dim>
5330 </port>
5331 </output>
5332 </layer>
5333 <layer id="322" name="/distilbert/transformer/layer.2/attention/Constant_2" type="Const" version="opset1">
5334 <data element_type="i64" shape="1" offset="90763288" size="8" />
5335 <rt_info>
5336 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_2" />
5337 </rt_info>
5338 <output>
5339 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_2_output_0">
5340 <dim>1</dim>
5341 </port>
5342 </output>
5343 </layer>
5344 <layer id="323" name="/distilbert/transformer/layer.2/attention/Constant_3" type="Const" version="opset1">
5345 <data element_type="i64" shape="1" offset="93131816" size="8" />
5346 <rt_info>
5347 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_3" />
5348 </rt_info>
5349 <output>
5350 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_3_output_0">
5351 <dim>1</dim>
5352 </port>
5353 </output>
5354 </layer>
5355 <layer id="324" name="/distilbert/transformer/layer.2/attention/Constant_4" type="Const" version="opset1">
5356 <data element_type="i64" shape="1" offset="93131824" size="8" />
5357 <rt_info>
5358 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_4" />
5359 </rt_info>
5360 <output>
5361 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_4_output_0">
5362 <dim>1</dim>
5363 </port>
5364 </output>
5365 </layer>
5366 <layer id="325" name="/distilbert/transformer/layer.2/attention/Concat" type="Concat" version="opset1">
5367 <data axis="0" />
5368 <rt_info>
5369 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Concat, /distilbert/transformer/layer.2/attention/Constant_2, /distilbert/transformer/layer.2/attention/Constant_3, /distilbert/transformer/layer.2/attention/Constant_4" />
5370 </rt_info>
5371 <input>
5372 <port id="0" precision="I64">
5373 <dim>1</dim>
5374 </port>
5375 <port id="1" precision="I64">
5376 <dim>1</dim>
5377 </port>
5378 <port id="2" precision="I64">
5379 <dim>1</dim>
5380 </port>
5381 <port id="3" precision="I64">
5382 <dim>1</dim>
5383 </port>
5384 </input>
5385 <output>
5386 <port id="4" precision="I64" names="/distilbert/transformer/layer.2/attention/Concat_output_0">
5387 <dim>4</dim>
5388 </port>
5389 </output>
5390 </layer>
5391 <layer id="326" name="/distilbert/transformer/layer.2/attention/Reshape" type="Reshape" version="opset1">
5392 <data special_zero="true" />
5393 <rt_info>
5394 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Reshape" />
5395 </rt_info>
5396 <input>
5397 <port id="0" precision="FP32">
5398 <dim>-1</dim>
5399 <dim>-1</dim>
5400 <dim>768</dim>
5401 </port>
5402 <port id="1" precision="I64">
5403 <dim>4</dim>
5404 </port>
5405 </input>
5406 <output>
5407 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Reshape_output_0">
5408 <dim>-1</dim>
5409 <dim>-1</dim>
5410 <dim>12</dim>
5411 <dim>64</dim>
5412 </port>
5413 </output>
5414 </layer>
5415 <layer id="327" name="Constant_255554" type="Const" version="opset1">
5416 <data element_type="i64" shape="4" offset="93131832" size="32" />
5417 <rt_info>
5418 <attribute name="fused_names" version="0" value="Constant_255554" />
5419 </rt_info>
5420 <output>
5421 <port id="0" precision="I64">
5422 <dim>4</dim>
5423 </port>
5424 </output>
5425 </layer>
5426 <layer id="328" name="/distilbert/transformer/layer.2/attention/Transpose" type="Transpose" version="opset1">
5427 <rt_info>
5428 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Transpose" />
5429 </rt_info>
5430 <input>
5431 <port id="0" precision="FP32">
5432 <dim>-1</dim>
5433 <dim>-1</dim>
5434 <dim>12</dim>
5435 <dim>64</dim>
5436 </port>
5437 <port id="1" precision="I64">
5438 <dim>4</dim>
5439 </port>
5440 </input>
5441 <output>
5442 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Transpose_output_0">
5443 <dim>-1</dim>
5444 <dim>12</dim>
5445 <dim>-1</dim>
5446 <dim>64</dim>
5447 </port>
5448 </output>
5449 </layer>
5450 <layer id="329" name="/distilbert/transformer/layer.2/attention/Constant_11" type="Const" version="opset1">
5451 <data element_type="f32" shape="" offset="93131864" size="4" />
5452 <rt_info>
5453 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_11" />
5454 </rt_info>
5455 <output>
5456 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/attention/Constant_11_output_0" />
5457 </output>
5458 </layer>
5459 <layer id="330" name="/distilbert/transformer/layer.2/attention/Div" type="Divide" version="opset1">
5460 <data auto_broadcast="numpy" m_pythondiv="true" />
5461 <rt_info>
5462 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Div" />
5463 </rt_info>
5464 <input>
5465 <port id="0" precision="FP32">
5466 <dim>-1</dim>
5467 <dim>12</dim>
5468 <dim>-1</dim>
5469 <dim>64</dim>
5470 </port>
5471 <port id="1" precision="FP32" />
5472 </input>
5473 <output>
5474 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Div_output_0">
5475 <dim>-1</dim>
5476 <dim>12</dim>
5477 <dim>-1</dim>
5478 <dim>64</dim>
5479 </port>
5480 </output>
5481 </layer>
5482 <layer id="331" name="distilbert.transformer.layer.2.attention.k_lin.bias" type="Const" version="opset1">
5483 <data element_type="f32" shape="768" offset="149798036" size="3072" />
5484 <rt_info>
5485 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.attention.k_lin.bias" />
5486 </rt_info>
5487 <output>
5488 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.attention.k_lin.bias">
5489 <dim>768</dim>
5490 </port>
5491 </output>
5492 </layer>
5493 <layer id="332" name="onnx::MatMul_786" type="Const" version="opset1">
5494 <data element_type="f32" shape="768, 768" offset="149801108" size="2359296" />
5495 <rt_info>
5496 <attribute name="fused_names" version="0" value="onnx::MatMul_786" />
5497 </rt_info>
5498 <output>
5499 <port id="0" precision="FP32" names="onnx::MatMul_786">
5500 <dim>768</dim>
5501 <dim>768</dim>
5502 </port>
5503 </output>
5504 </layer>
5505 <layer id="333" name="/distilbert/transformer/layer.2/attention/k_lin/MatMul" type="MatMul" version="opset1">
5506 <data transpose_a="false" transpose_b="false" />
5507 <rt_info>
5508 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/k_lin/MatMul" />
5509 </rt_info>
5510 <input>
5511 <port id="0" precision="FP32">
5512 <dim>-1</dim>
5513 <dim>-1</dim>
5514 <dim>768</dim>
5515 </port>
5516 <port id="1" precision="FP32">
5517 <dim>768</dim>
5518 <dim>768</dim>
5519 </port>
5520 </input>
5521 <output>
5522 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/k_lin/MatMul_output_0">
5523 <dim>-1</dim>
5524 <dim>-1</dim>
5525 <dim>768</dim>
5526 </port>
5527 </output>
5528 </layer>
5529 <layer id="334" name="/distilbert/transformer/layer.2/attention/k_lin/Add" type="Add" version="opset1">
5530 <data auto_broadcast="numpy" />
5531 <rt_info>
5532 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/k_lin/Add" />
5533 </rt_info>
5534 <input>
5535 <port id="0" precision="FP32">
5536 <dim>768</dim>
5537 </port>
5538 <port id="1" precision="FP32">
5539 <dim>-1</dim>
5540 <dim>-1</dim>
5541 <dim>768</dim>
5542 </port>
5543 </input>
5544 <output>
5545 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/k_lin/Add_output_0">
5546 <dim>-1</dim>
5547 <dim>-1</dim>
5548 <dim>768</dim>
5549 </port>
5550 </output>
5551 </layer>
5552 <layer id="335" name="Constant_255535" type="Const" version="opset1">
5553 <data element_type="i64" shape="1" offset="18440" size="8" />
5554 <rt_info>
5555 <attribute name="fused_names" version="0" value="Constant_255535" />
5556 </rt_info>
5557 <output>
5558 <port id="0" precision="I64">
5559 <dim>1</dim>
5560 </port>
5561 </output>
5562 </layer>
5563 <layer id="336" name="/distilbert/transformer/layer.2/attention/Unsqueeze_1" type="Unsqueeze" version="opset1">
5564 <rt_info>
5565 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Unsqueeze_1, Constant_255535" />
5566 </rt_info>
5567 <input>
5568 <port id="0" precision="I64" />
5569 <port id="1" precision="I64">
5570 <dim>1</dim>
5571 </port>
5572 </input>
5573 <output>
5574 <port id="2" precision="I64" names="/distilbert/transformer/layer.2/attention/Unsqueeze_1_output_0">
5575 <dim>1</dim>
5576 </port>
5577 </output>
5578 </layer>
5579 <layer id="337" name="/distilbert/transformer/layer.2/attention/Constant_5" type="Const" version="opset1">
5580 <data element_type="i64" shape="1" offset="90763288" size="8" />
5581 <rt_info>
5582 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_5" />
5583 </rt_info>
5584 <output>
5585 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_5_output_0">
5586 <dim>1</dim>
5587 </port>
5588 </output>
5589 </layer>
5590 <layer id="338" name="/distilbert/transformer/layer.2/attention/Constant_6" type="Const" version="opset1">
5591 <data element_type="i64" shape="1" offset="93131816" size="8" />
5592 <rt_info>
5593 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_6" />
5594 </rt_info>
5595 <output>
5596 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_6_output_0">
5597 <dim>1</dim>
5598 </port>
5599 </output>
5600 </layer>
5601 <layer id="339" name="/distilbert/transformer/layer.2/attention/Constant_7" type="Const" version="opset1">
5602 <data element_type="i64" shape="1" offset="93131824" size="8" />
5603 <rt_info>
5604 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_7" />
5605 </rt_info>
5606 <output>
5607 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_7_output_0">
5608 <dim>1</dim>
5609 </port>
5610 </output>
5611 </layer>
5612 <layer id="340" name="/distilbert/transformer/layer.2/attention/Concat_1" type="Concat" version="opset1">
5613 <data axis="0" />
5614 <rt_info>
5615 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Concat_1, /distilbert/transformer/layer.2/attention/Constant_5, /distilbert/transformer/layer.2/attention/Constant_6, /distilbert/transformer/layer.2/attention/Constant_7" />
5616 </rt_info>
5617 <input>
5618 <port id="0" precision="I64">
5619 <dim>1</dim>
5620 </port>
5621 <port id="1" precision="I64">
5622 <dim>1</dim>
5623 </port>
5624 <port id="2" precision="I64">
5625 <dim>1</dim>
5626 </port>
5627 <port id="3" precision="I64">
5628 <dim>1</dim>
5629 </port>
5630 </input>
5631 <output>
5632 <port id="4" precision="I64" names="/distilbert/transformer/layer.2/attention/Concat_1_output_0">
5633 <dim>4</dim>
5634 </port>
5635 </output>
5636 </layer>
5637 <layer id="341" name="/distilbert/transformer/layer.2/attention/Reshape_1" type="Reshape" version="opset1">
5638 <data special_zero="true" />
5639 <rt_info>
5640 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Reshape_1" />
5641 </rt_info>
5642 <input>
5643 <port id="0" precision="FP32">
5644 <dim>-1</dim>
5645 <dim>-1</dim>
5646 <dim>768</dim>
5647 </port>
5648 <port id="1" precision="I64">
5649 <dim>4</dim>
5650 </port>
5651 </input>
5652 <output>
5653 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Reshape_1_output_0">
5654 <dim>-1</dim>
5655 <dim>-1</dim>
5656 <dim>12</dim>
5657 <dim>64</dim>
5658 </port>
5659 </output>
5660 </layer>
5661 <layer id="342" name="Constant_255578" type="Const" version="opset1">
5662 <data element_type="i64" shape="4" offset="95494236" size="32" />
5663 <rt_info>
5664 <attribute name="fused_names" version="0" value="Constant_255578" />
5665 </rt_info>
5666 <output>
5667 <port id="0" precision="I64">
5668 <dim>4</dim>
5669 </port>
5670 </output>
5671 </layer>
5672 <layer id="343" name="/distilbert/transformer/layer.2/attention/Transpose_2" type="Transpose" version="opset1">
5673 <rt_info>
5674 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Transpose_2" />
5675 </rt_info>
5676 <input>
5677 <port id="0" precision="FP32">
5678 <dim>-1</dim>
5679 <dim>-1</dim>
5680 <dim>12</dim>
5681 <dim>64</dim>
5682 </port>
5683 <port id="1" precision="I64">
5684 <dim>4</dim>
5685 </port>
5686 </input>
5687 <output>
5688 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Transpose_2_output_0">
5689 <dim>-1</dim>
5690 <dim>12</dim>
5691 <dim>64</dim>
5692 <dim>-1</dim>
5693 </port>
5694 </output>
5695 </layer>
5696 <layer id="344" name="/distilbert/transformer/layer.2/attention/MatMul" type="MatMul" version="opset1">
5697 <data transpose_a="false" transpose_b="false" />
5698 <rt_info>
5699 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/MatMul" />
5700 </rt_info>
5701 <input>
5702 <port id="0" precision="FP32">
5703 <dim>-1</dim>
5704 <dim>12</dim>
5705 <dim>-1</dim>
5706 <dim>64</dim>
5707 </port>
5708 <port id="1" precision="FP32">
5709 <dim>-1</dim>
5710 <dim>12</dim>
5711 <dim>64</dim>
5712 <dim>-1</dim>
5713 </port>
5714 </input>
5715 <output>
5716 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/MatMul_output_0">
5717 <dim>-1</dim>
5718 <dim>12</dim>
5719 <dim>-1</dim>
5720 <dim>-1</dim>
5721 </port>
5722 </output>
5723 </layer>
5724 <layer id="345" name="/distilbert/transformer/layer.2/attention/Shape_2" type="ShapeOf" version="opset3">
5725 <data output_type="i64" />
5726 <rt_info>
5727 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Shape_2" />
5728 </rt_info>
5729 <input>
5730 <port id="0" precision="FP32">
5731 <dim>-1</dim>
5732 <dim>12</dim>
5733 <dim>-1</dim>
5734 <dim>-1</dim>
5735 </port>
5736 </input>
5737 <output>
5738 <port id="1" precision="I64" names="/distilbert/transformer/layer.2/attention/Shape_2_output_0">
5739 <dim>4</dim>
5740 </port>
5741 </output>
5742 </layer>
5743 <layer id="346" name="/distilbert/transformer/layer.2/attention/Expand" type="Broadcast" version="opset3">
5744 <data mode="bidirectional" />
5745 <rt_info>
5746 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Expand" />
5747 </rt_info>
5748 <input>
5749 <port id="0" precision="BOOL">
5750 <dim>-1</dim>
5751 <dim>1</dim>
5752 <dim>1</dim>
5753 <dim>-1</dim>
5754 </port>
5755 <port id="1" precision="I64">
5756 <dim>4</dim>
5757 </port>
5758 </input>
5759 <output>
5760 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.2/attention/Expand_output_0">
5761 <dim>-1</dim>
5762 <dim>12</dim>
5763 <dim>-1</dim>
5764 <dim>-1</dim>
5765 </port>
5766 </output>
5767 </layer>
5768 <layer id="347" name="/distilbert/transformer/layer.2/attention/Cast" type="Convert" version="opset1">
5769 <data destination_type="boolean" />
5770 <rt_info>
5771 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Cast" />
5772 </rt_info>
5773 <input>
5774 <port id="0" precision="BOOL">
5775 <dim>-1</dim>
5776 <dim>12</dim>
5777 <dim>-1</dim>
5778 <dim>-1</dim>
5779 </port>
5780 </input>
5781 <output>
5782 <port id="1" precision="BOOL" names="/distilbert/transformer/layer.2/attention/Cast_output_0">
5783 <dim>-1</dim>
5784 <dim>12</dim>
5785 <dim>-1</dim>
5786 <dim>-1</dim>
5787 </port>
5788 </output>
5789 </layer>
5790 <layer id="348" name="/distilbert/transformer/layer.2/attention/Constant_14" type="Const" version="opset1">
5791 <data element_type="f32" shape="" offset="95494268" size="4" />
5792 <rt_info>
5793 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_14" />
5794 </rt_info>
5795 <output>
5796 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/attention/Constant_14_output_0" />
5797 </output>
5798 </layer>
5799 <layer id="349" name="/distilbert/transformer/layer.2/attention/Where" type="Select" version="opset1">
5800 <data auto_broadcast="numpy" />
5801 <rt_info>
5802 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Where" />
5803 </rt_info>
5804 <input>
5805 <port id="0" precision="BOOL">
5806 <dim>-1</dim>
5807 <dim>12</dim>
5808 <dim>-1</dim>
5809 <dim>-1</dim>
5810 </port>
5811 <port id="1" precision="FP32" />
5812 <port id="2" precision="FP32">
5813 <dim>-1</dim>
5814 <dim>12</dim>
5815 <dim>-1</dim>
5816 <dim>-1</dim>
5817 </port>
5818 </input>
5819 <output>
5820 <port id="3" precision="FP32" names="/distilbert/transformer/layer.2/attention/Where_output_0">
5821 <dim>-1</dim>
5822 <dim>12</dim>
5823 <dim>-1</dim>
5824 <dim>-1</dim>
5825 </port>
5826 </output>
5827 </layer>
5828 <layer id="350" name="/distilbert/transformer/layer.2/attention/Softmax" type="SoftMax" version="opset8">
5829 <data axis="3" />
5830 <rt_info>
5831 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Softmax" />
5832 </rt_info>
5833 <input>
5834 <port id="0" precision="FP32">
5835 <dim>-1</dim>
5836 <dim>12</dim>
5837 <dim>-1</dim>
5838 <dim>-1</dim>
5839 </port>
5840 </input>
5841 <output>
5842 <port id="1" precision="FP32" names="/distilbert/transformer/layer.2/attention/Softmax_output_0">
5843 <dim>-1</dim>
5844 <dim>12</dim>
5845 <dim>-1</dim>
5846 <dim>-1</dim>
5847 </port>
5848 </output>
5849 </layer>
5850 <layer id="351" name="distilbert.transformer.layer.2.attention.v_lin.bias" type="Const" version="opset1">
5851 <data element_type="f32" shape="768" offset="152160404" size="3072" />
5852 <rt_info>
5853 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.attention.v_lin.bias" />
5854 </rt_info>
5855 <output>
5856 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.attention.v_lin.bias">
5857 <dim>768</dim>
5858 </port>
5859 </output>
5860 </layer>
5861 <layer id="352" name="onnx::MatMul_787" type="Const" version="opset1">
5862 <data element_type="f32" shape="768, 768" offset="152163476" size="2359296" />
5863 <rt_info>
5864 <attribute name="fused_names" version="0" value="onnx::MatMul_787" />
5865 </rt_info>
5866 <output>
5867 <port id="0" precision="FP32" names="onnx::MatMul_787">
5868 <dim>768</dim>
5869 <dim>768</dim>
5870 </port>
5871 </output>
5872 </layer>
5873 <layer id="353" name="/distilbert/transformer/layer.2/attention/v_lin/MatMul" type="MatMul" version="opset1">
5874 <data transpose_a="false" transpose_b="false" />
5875 <rt_info>
5876 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/v_lin/MatMul" />
5877 </rt_info>
5878 <input>
5879 <port id="0" precision="FP32">
5880 <dim>-1</dim>
5881 <dim>-1</dim>
5882 <dim>768</dim>
5883 </port>
5884 <port id="1" precision="FP32">
5885 <dim>768</dim>
5886 <dim>768</dim>
5887 </port>
5888 </input>
5889 <output>
5890 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/v_lin/MatMul_output_0">
5891 <dim>-1</dim>
5892 <dim>-1</dim>
5893 <dim>768</dim>
5894 </port>
5895 </output>
5896 </layer>
5897 <layer id="354" name="/distilbert/transformer/layer.2/attention/v_lin/Add" type="Add" version="opset1">
5898 <data auto_broadcast="numpy" />
5899 <rt_info>
5900 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/v_lin/Add" />
5901 </rt_info>
5902 <input>
5903 <port id="0" precision="FP32">
5904 <dim>768</dim>
5905 </port>
5906 <port id="1" precision="FP32">
5907 <dim>-1</dim>
5908 <dim>-1</dim>
5909 <dim>768</dim>
5910 </port>
5911 </input>
5912 <output>
5913 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/v_lin/Add_output_0">
5914 <dim>-1</dim>
5915 <dim>-1</dim>
5916 <dim>768</dim>
5917 </port>
5918 </output>
5919 </layer>
5920 <layer id="355" name="Constant_255541" type="Const" version="opset1">
5921 <data element_type="i64" shape="1" offset="18440" size="8" />
5922 <rt_info>
5923 <attribute name="fused_names" version="0" value="Constant_255541" />
5924 </rt_info>
5925 <output>
5926 <port id="0" precision="I64">
5927 <dim>1</dim>
5928 </port>
5929 </output>
5930 </layer>
5931 <layer id="356" name="/distilbert/transformer/layer.2/attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
5932 <rt_info>
5933 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Unsqueeze_2, Constant_255541" />
5934 </rt_info>
5935 <input>
5936 <port id="0" precision="I64" />
5937 <port id="1" precision="I64">
5938 <dim>1</dim>
5939 </port>
5940 </input>
5941 <output>
5942 <port id="2" precision="I64" names="/distilbert/transformer/layer.2/attention/Unsqueeze_2_output_0">
5943 <dim>1</dim>
5944 </port>
5945 </output>
5946 </layer>
5947 <layer id="357" name="/distilbert/transformer/layer.2/attention/Constant_8" type="Const" version="opset1">
5948 <data element_type="i64" shape="1" offset="90763288" size="8" />
5949 <rt_info>
5950 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_8" />
5951 </rt_info>
5952 <output>
5953 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_8_output_0">
5954 <dim>1</dim>
5955 </port>
5956 </output>
5957 </layer>
5958 <layer id="358" name="/distilbert/transformer/layer.2/attention/Constant_9" type="Const" version="opset1">
5959 <data element_type="i64" shape="1" offset="93131816" size="8" />
5960 <rt_info>
5961 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_9" />
5962 </rt_info>
5963 <output>
5964 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_9_output_0">
5965 <dim>1</dim>
5966 </port>
5967 </output>
5968 </layer>
5969 <layer id="359" name="/distilbert/transformer/layer.2/attention/Constant_10" type="Const" version="opset1">
5970 <data element_type="i64" shape="1" offset="93131824" size="8" />
5971 <rt_info>
5972 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_10" />
5973 </rt_info>
5974 <output>
5975 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_10_output_0">
5976 <dim>1</dim>
5977 </port>
5978 </output>
5979 </layer>
5980 <layer id="360" name="/distilbert/transformer/layer.2/attention/Concat_2" type="Concat" version="opset1">
5981 <data axis="0" />
5982 <rt_info>
5983 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Concat_2, /distilbert/transformer/layer.2/attention/Constant_10, /distilbert/transformer/layer.2/attention/Constant_8, /distilbert/transformer/layer.2/attention/Constant_9" />
5984 </rt_info>
5985 <input>
5986 <port id="0" precision="I64">
5987 <dim>1</dim>
5988 </port>
5989 <port id="1" precision="I64">
5990 <dim>1</dim>
5991 </port>
5992 <port id="2" precision="I64">
5993 <dim>1</dim>
5994 </port>
5995 <port id="3" precision="I64">
5996 <dim>1</dim>
5997 </port>
5998 </input>
5999 <output>
6000 <port id="4" precision="I64" names="/distilbert/transformer/layer.2/attention/Concat_2_output_0">
6001 <dim>4</dim>
6002 </port>
6003 </output>
6004 </layer>
6005 <layer id="361" name="/distilbert/transformer/layer.2/attention/Reshape_2" type="Reshape" version="opset1">
6006 <data special_zero="true" />
6007 <rt_info>
6008 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Reshape_2" />
6009 </rt_info>
6010 <input>
6011 <port id="0" precision="FP32">
6012 <dim>-1</dim>
6013 <dim>-1</dim>
6014 <dim>768</dim>
6015 </port>
6016 <port id="1" precision="I64">
6017 <dim>4</dim>
6018 </port>
6019 </input>
6020 <output>
6021 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Reshape_2_output_0">
6022 <dim>-1</dim>
6023 <dim>-1</dim>
6024 <dim>12</dim>
6025 <dim>64</dim>
6026 </port>
6027 </output>
6028 </layer>
6029 <layer id="362" name="Constant_255574" type="Const" version="opset1">
6030 <data element_type="i64" shape="4" offset="93131832" size="32" />
6031 <rt_info>
6032 <attribute name="fused_names" version="0" value="Constant_255574" />
6033 </rt_info>
6034 <output>
6035 <port id="0" precision="I64">
6036 <dim>4</dim>
6037 </port>
6038 </output>
6039 </layer>
6040 <layer id="363" name="/distilbert/transformer/layer.2/attention/Transpose_1" type="Transpose" version="opset1">
6041 <rt_info>
6042 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Transpose_1" />
6043 </rt_info>
6044 <input>
6045 <port id="0" precision="FP32">
6046 <dim>-1</dim>
6047 <dim>-1</dim>
6048 <dim>12</dim>
6049 <dim>64</dim>
6050 </port>
6051 <port id="1" precision="I64">
6052 <dim>4</dim>
6053 </port>
6054 </input>
6055 <output>
6056 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Transpose_1_output_0">
6057 <dim>-1</dim>
6058 <dim>12</dim>
6059 <dim>-1</dim>
6060 <dim>64</dim>
6061 </port>
6062 </output>
6063 </layer>
6064 <layer id="364" name="/distilbert/transformer/layer.2/attention/MatMul_1" type="MatMul" version="opset1">
6065 <data transpose_a="false" transpose_b="false" />
6066 <rt_info>
6067 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/MatMul_1" />
6068 </rt_info>
6069 <input>
6070 <port id="0" precision="FP32">
6071 <dim>-1</dim>
6072 <dim>12</dim>
6073 <dim>-1</dim>
6074 <dim>-1</dim>
6075 </port>
6076 <port id="1" precision="FP32">
6077 <dim>-1</dim>
6078 <dim>12</dim>
6079 <dim>-1</dim>
6080 <dim>64</dim>
6081 </port>
6082 </input>
6083 <output>
6084 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/MatMul_1_output_0">
6085 <dim>-1</dim>
6086 <dim>12</dim>
6087 <dim>-1</dim>
6088 <dim>64</dim>
6089 </port>
6090 </output>
6091 </layer>
6092 <layer id="365" name="Constant_255610" type="Const" version="opset1">
6093 <data element_type="i64" shape="4" offset="93131832" size="32" />
6094 <rt_info>
6095 <attribute name="fused_names" version="0" value="Constant_255610" />
6096 </rt_info>
6097 <output>
6098 <port id="0" precision="I64">
6099 <dim>4</dim>
6100 </port>
6101 </output>
6102 </layer>
6103 <layer id="366" name="/distilbert/transformer/layer.2/attention/Transpose_3" type="Transpose" version="opset1">
6104 <rt_info>
6105 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Transpose_3" />
6106 </rt_info>
6107 <input>
6108 <port id="0" precision="FP32">
6109 <dim>-1</dim>
6110 <dim>12</dim>
6111 <dim>-1</dim>
6112 <dim>64</dim>
6113 </port>
6114 <port id="1" precision="I64">
6115 <dim>4</dim>
6116 </port>
6117 </input>
6118 <output>
6119 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Transpose_3_output_0">
6120 <dim>-1</dim>
6121 <dim>-1</dim>
6122 <dim>12</dim>
6123 <dim>64</dim>
6124 </port>
6125 </output>
6126 </layer>
6127 <layer id="367" name="Constant_255612" type="Const" version="opset1">
6128 <data element_type="i64" shape="1" offset="18440" size="8" />
6129 <rt_info>
6130 <attribute name="fused_names" version="0" value="Constant_255612" />
6131 </rt_info>
6132 <output>
6133 <port id="0" precision="I64">
6134 <dim>1</dim>
6135 </port>
6136 </output>
6137 </layer>
6138 <layer id="368" name="/distilbert/transformer/layer.2/attention/Unsqueeze_5" type="Unsqueeze" version="opset1">
6139 <rt_info>
6140 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Unsqueeze_5, Constant_255612" />
6141 </rt_info>
6142 <input>
6143 <port id="0" precision="I64" />
6144 <port id="1" precision="I64">
6145 <dim>1</dim>
6146 </port>
6147 </input>
6148 <output>
6149 <port id="2" precision="I64" names="/distilbert/transformer/layer.2/attention/Unsqueeze_5_output_0">
6150 <dim>1</dim>
6151 </port>
6152 </output>
6153 </layer>
6154 <layer id="369" name="/distilbert/transformer/layer.2/attention/Constant_15" type="Const" version="opset1">
6155 <data element_type="i64" shape="1" offset="90763288" size="8" />
6156 <rt_info>
6157 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_15" />
6158 </rt_info>
6159 <output>
6160 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_15_output_0">
6161 <dim>1</dim>
6162 </port>
6163 </output>
6164 </layer>
6165 <layer id="370" name="/distilbert/transformer/layer.2/attention/Constant_16" type="Const" version="opset1">
6166 <data element_type="i64" shape="1" offset="97856640" size="8" />
6167 <rt_info>
6168 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Constant_16" />
6169 </rt_info>
6170 <output>
6171 <port id="0" precision="I64" names="/distilbert/transformer/layer.2/attention/Constant_16_output_0">
6172 <dim>1</dim>
6173 </port>
6174 </output>
6175 </layer>
6176 <layer id="371" name="/distilbert/transformer/layer.2/attention/Concat_4" type="Concat" version="opset1">
6177 <data axis="0" />
6178 <rt_info>
6179 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Concat_4, /distilbert/transformer/layer.2/attention/Constant_15, /distilbert/transformer/layer.2/attention/Constant_16" />
6180 </rt_info>
6181 <input>
6182 <port id="0" precision="I64">
6183 <dim>1</dim>
6184 </port>
6185 <port id="1" precision="I64">
6186 <dim>1</dim>
6187 </port>
6188 <port id="2" precision="I64">
6189 <dim>1</dim>
6190 </port>
6191 </input>
6192 <output>
6193 <port id="3" precision="I64" names="/distilbert/transformer/layer.2/attention/Concat_4_output_0">
6194 <dim>3</dim>
6195 </port>
6196 </output>
6197 </layer>
6198 <layer id="372" name="/distilbert/transformer/layer.2/attention/Reshape_4" type="Reshape" version="opset1">
6199 <data special_zero="true" />
6200 <rt_info>
6201 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/Reshape_4" />
6202 </rt_info>
6203 <input>
6204 <port id="0" precision="FP32">
6205 <dim>-1</dim>
6206 <dim>-1</dim>
6207 <dim>12</dim>
6208 <dim>64</dim>
6209 </port>
6210 <port id="1" precision="I64">
6211 <dim>3</dim>
6212 </port>
6213 </input>
6214 <output>
6215 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/Reshape_4_output_0">
6216 <dim>-1</dim>
6217 <dim>-1</dim>
6218 <dim>768</dim>
6219 </port>
6220 </output>
6221 </layer>
6222 <layer id="373" name="onnx::MatMul_792" type="Const" version="opset1">
6223 <data element_type="f32" shape="768, 768" offset="154522772" size="2359296" />
6224 <rt_info>
6225 <attribute name="fused_names" version="0" value="onnx::MatMul_792" />
6226 </rt_info>
6227 <output>
6228 <port id="0" precision="FP32" names="onnx::MatMul_792">
6229 <dim>768</dim>
6230 <dim>768</dim>
6231 </port>
6232 </output>
6233 </layer>
6234 <layer id="374" name="/distilbert/transformer/layer.2/attention/out_lin/MatMul" type="MatMul" version="opset1">
6235 <data transpose_a="false" transpose_b="false" />
6236 <rt_info>
6237 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/out_lin/MatMul" />
6238 </rt_info>
6239 <input>
6240 <port id="0" precision="FP32">
6241 <dim>-1</dim>
6242 <dim>-1</dim>
6243 <dim>768</dim>
6244 </port>
6245 <port id="1" precision="FP32">
6246 <dim>768</dim>
6247 <dim>768</dim>
6248 </port>
6249 </input>
6250 <output>
6251 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/out_lin/MatMul_output_0">
6252 <dim>-1</dim>
6253 <dim>-1</dim>
6254 <dim>768</dim>
6255 </port>
6256 </output>
6257 </layer>
6258 <layer id="375" name="/distilbert/transformer/layer.2/attention/out_lin/Add" type="Add" version="opset1">
6259 <data auto_broadcast="numpy" />
6260 <rt_info>
6261 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/attention/out_lin/Add" />
6262 </rt_info>
6263 <input>
6264 <port id="0" precision="FP32">
6265 <dim>768</dim>
6266 </port>
6267 <port id="1" precision="FP32">
6268 <dim>-1</dim>
6269 <dim>-1</dim>
6270 <dim>768</dim>
6271 </port>
6272 </input>
6273 <output>
6274 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/attention/out_lin/Add_output_0">
6275 <dim>-1</dim>
6276 <dim>-1</dim>
6277 <dim>768</dim>
6278 </port>
6279 </output>
6280 </layer>
6281 <layer id="376" name="/distilbert/transformer/layer.2/Add" type="Add" version="opset1">
6282 <data auto_broadcast="numpy" />
6283 <rt_info>
6284 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/Add" />
6285 </rt_info>
6286 <input>
6287 <port id="0" precision="FP32">
6288 <dim>-1</dim>
6289 <dim>-1</dim>
6290 <dim>768</dim>
6291 </port>
6292 <port id="1" precision="FP32">
6293 <dim>-1</dim>
6294 <dim>-1</dim>
6295 <dim>768</dim>
6296 </port>
6297 </input>
6298 <output>
6299 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/Add_output_0">
6300 <dim>-1</dim>
6301 <dim>-1</dim>
6302 <dim>768</dim>
6303 </port>
6304 </output>
6305 </layer>
6306 <layer id="377" name="Constant_255627" type="Const" version="opset1">
6307 <data element_type="i64" shape="1" offset="90763288" size="8" />
6308 <rt_info>
6309 <attribute name="fused_names" version="0" value="Constant_255627" />
6310 </rt_info>
6311 <output>
6312 <port id="0" precision="I64">
6313 <dim>1</dim>
6314 </port>
6315 </output>
6316 </layer>
6317 <layer id="378" name="/distilbert/transformer/layer.2/sa_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
6318 <data keep_dims="true" />
6319 <rt_info>
6320 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/ReduceMean" />
6321 </rt_info>
6322 <input>
6323 <port id="0" precision="FP32">
6324 <dim>-1</dim>
6325 <dim>-1</dim>
6326 <dim>768</dim>
6327 </port>
6328 <port id="1" precision="I64">
6329 <dim>1</dim>
6330 </port>
6331 </input>
6332 <output>
6333 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/ReduceMean_output_0">
6334 <dim>-1</dim>
6335 <dim>-1</dim>
6336 <dim>1</dim>
6337 </port>
6338 </output>
6339 </layer>
6340 <layer id="379" name="/distilbert/transformer/layer.2/sa_layer_norm/Sub" type="Subtract" version="opset1">
6341 <data auto_broadcast="numpy" />
6342 <rt_info>
6343 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Sub" />
6344 </rt_info>
6345 <input>
6346 <port id="0" precision="FP32">
6347 <dim>-1</dim>
6348 <dim>-1</dim>
6349 <dim>768</dim>
6350 </port>
6351 <port id="1" precision="FP32">
6352 <dim>-1</dim>
6353 <dim>-1</dim>
6354 <dim>1</dim>
6355 </port>
6356 </input>
6357 <output>
6358 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Sub_output_0">
6359 <dim>-1</dim>
6360 <dim>-1</dim>
6361 <dim>768</dim>
6362 </port>
6363 </output>
6364 </layer>
6365 <layer id="380" name="/distilbert/transformer/layer.2/sa_layer_norm/Constant" type="Const" version="opset1">
6366 <data element_type="f32" shape="" offset="90763296" size="4" />
6367 <rt_info>
6368 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Constant" />
6369 </rt_info>
6370 <output>
6371 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Constant_output_0" />
6372 </output>
6373 </layer>
6374 <layer id="381" name="/distilbert/transformer/layer.2/sa_layer_norm/Pow" type="Power" version="opset1">
6375 <data auto_broadcast="numpy" />
6376 <rt_info>
6377 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Pow" />
6378 </rt_info>
6379 <input>
6380 <port id="0" precision="FP32">
6381 <dim>-1</dim>
6382 <dim>-1</dim>
6383 <dim>768</dim>
6384 </port>
6385 <port id="1" precision="FP32" />
6386 </input>
6387 <output>
6388 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Pow_output_0">
6389 <dim>-1</dim>
6390 <dim>-1</dim>
6391 <dim>768</dim>
6392 </port>
6393 </output>
6394 </layer>
6395 <layer id="382" name="Constant_255632" type="Const" version="opset1">
6396 <data element_type="i64" shape="1" offset="90763288" size="8" />
6397 <rt_info>
6398 <attribute name="fused_names" version="0" value="Constant_255632" />
6399 </rt_info>
6400 <output>
6401 <port id="0" precision="I64">
6402 <dim>1</dim>
6403 </port>
6404 </output>
6405 </layer>
6406 <layer id="383" name="/distilbert/transformer/layer.2/sa_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
6407 <data keep_dims="true" />
6408 <rt_info>
6409 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/ReduceMean_1" />
6410 </rt_info>
6411 <input>
6412 <port id="0" precision="FP32">
6413 <dim>-1</dim>
6414 <dim>-1</dim>
6415 <dim>768</dim>
6416 </port>
6417 <port id="1" precision="I64">
6418 <dim>1</dim>
6419 </port>
6420 </input>
6421 <output>
6422 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/ReduceMean_1_output_0">
6423 <dim>-1</dim>
6424 <dim>-1</dim>
6425 <dim>1</dim>
6426 </port>
6427 </output>
6428 </layer>
6429 <layer id="384" name="/distilbert/transformer/layer.2/sa_layer_norm/Constant_1" type="Const" version="opset1">
6430 <data element_type="f32" shape="" offset="90763300" size="4" />
6431 <rt_info>
6432 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Constant_1" />
6433 </rt_info>
6434 <output>
6435 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Constant_1_output_0" />
6436 </output>
6437 </layer>
6438 <layer id="385" name="/distilbert/transformer/layer.2/sa_layer_norm/Add" type="Add" version="opset1">
6439 <data auto_broadcast="numpy" />
6440 <rt_info>
6441 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Add" />
6442 </rt_info>
6443 <input>
6444 <port id="0" precision="FP32">
6445 <dim>-1</dim>
6446 <dim>-1</dim>
6447 <dim>1</dim>
6448 </port>
6449 <port id="1" precision="FP32" />
6450 </input>
6451 <output>
6452 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Add_output_0">
6453 <dim>-1</dim>
6454 <dim>-1</dim>
6455 <dim>1</dim>
6456 </port>
6457 </output>
6458 </layer>
6459 <layer id="386" name="/distilbert/transformer/layer.2/sa_layer_norm/Sqrt" type="Sqrt" version="opset1">
6460 <rt_info>
6461 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Sqrt" />
6462 </rt_info>
6463 <input>
6464 <port id="0" precision="FP32">
6465 <dim>-1</dim>
6466 <dim>-1</dim>
6467 <dim>1</dim>
6468 </port>
6469 </input>
6470 <output>
6471 <port id="1" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Sqrt_output_0">
6472 <dim>-1</dim>
6473 <dim>-1</dim>
6474 <dim>1</dim>
6475 </port>
6476 </output>
6477 </layer>
6478 <layer id="387" name="/distilbert/transformer/layer.2/sa_layer_norm/Div" type="Divide" version="opset1">
6479 <data auto_broadcast="numpy" m_pythondiv="true" />
6480 <rt_info>
6481 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Div" />
6482 </rt_info>
6483 <input>
6484 <port id="0" precision="FP32">
6485 <dim>-1</dim>
6486 <dim>-1</dim>
6487 <dim>768</dim>
6488 </port>
6489 <port id="1" precision="FP32">
6490 <dim>-1</dim>
6491 <dim>-1</dim>
6492 <dim>1</dim>
6493 </port>
6494 </input>
6495 <output>
6496 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Div_output_0">
6497 <dim>-1</dim>
6498 <dim>-1</dim>
6499 <dim>768</dim>
6500 </port>
6501 </output>
6502 </layer>
6503 <layer id="388" name="distilbert.transformer.layer.2.sa_layer_norm.weight" type="Const" version="opset1">
6504 <data element_type="f32" shape="768" offset="156882068" size="3072" />
6505 <rt_info>
6506 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.sa_layer_norm.weight" />
6507 </rt_info>
6508 <output>
6509 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.sa_layer_norm.weight">
6510 <dim>768</dim>
6511 </port>
6512 </output>
6513 </layer>
6514 <layer id="389" name="/distilbert/transformer/layer.2/sa_layer_norm/Mul" type="Multiply" version="opset1">
6515 <data auto_broadcast="numpy" />
6516 <rt_info>
6517 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Mul" />
6518 </rt_info>
6519 <input>
6520 <port id="0" precision="FP32">
6521 <dim>-1</dim>
6522 <dim>-1</dim>
6523 <dim>768</dim>
6524 </port>
6525 <port id="1" precision="FP32">
6526 <dim>768</dim>
6527 </port>
6528 </input>
6529 <output>
6530 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Mul_output_0">
6531 <dim>-1</dim>
6532 <dim>-1</dim>
6533 <dim>768</dim>
6534 </port>
6535 </output>
6536 </layer>
6537 <layer id="390" name="distilbert.transformer.layer.2.sa_layer_norm.bias" type="Const" version="opset1">
6538 <data element_type="f32" shape="768" offset="156885140" size="3072" />
6539 <rt_info>
6540 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.sa_layer_norm.bias" />
6541 </rt_info>
6542 <output>
6543 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.sa_layer_norm.bias">
6544 <dim>768</dim>
6545 </port>
6546 </output>
6547 </layer>
6548 <layer id="391" name="/distilbert/transformer/layer.2/sa_layer_norm/Add_1" type="Add" version="opset1">
6549 <data auto_broadcast="numpy" />
6550 <rt_info>
6551 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/sa_layer_norm/Add_1" />
6552 </rt_info>
6553 <input>
6554 <port id="0" precision="FP32">
6555 <dim>-1</dim>
6556 <dim>-1</dim>
6557 <dim>768</dim>
6558 </port>
6559 <port id="1" precision="FP32">
6560 <dim>768</dim>
6561 </port>
6562 </input>
6563 <output>
6564 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/sa_layer_norm/Add_1_output_0">
6565 <dim>-1</dim>
6566 <dim>-1</dim>
6567 <dim>768</dim>
6568 </port>
6569 </output>
6570 </layer>
6571 <layer id="392" name="onnx::MatMul_793" type="Const" version="opset1">
6572 <data element_type="f32" shape="768, 3072" offset="156888212" size="9437184" />
6573 <rt_info>
6574 <attribute name="fused_names" version="0" value="onnx::MatMul_793" />
6575 </rt_info>
6576 <output>
6577 <port id="0" precision="FP32" names="onnx::MatMul_793">
6578 <dim>768</dim>
6579 <dim>3072</dim>
6580 </port>
6581 </output>
6582 </layer>
6583 <layer id="393" name="/distilbert/transformer/layer.2/ffn/lin1/MatMul" type="MatMul" version="opset1">
6584 <data transpose_a="false" transpose_b="false" />
6585 <rt_info>
6586 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/lin1/MatMul" />
6587 </rt_info>
6588 <input>
6589 <port id="0" precision="FP32">
6590 <dim>-1</dim>
6591 <dim>-1</dim>
6592 <dim>768</dim>
6593 </port>
6594 <port id="1" precision="FP32">
6595 <dim>768</dim>
6596 <dim>3072</dim>
6597 </port>
6598 </input>
6599 <output>
6600 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/lin1/MatMul_output_0">
6601 <dim>-1</dim>
6602 <dim>-1</dim>
6603 <dim>3072</dim>
6604 </port>
6605 </output>
6606 </layer>
6607 <layer id="394" name="/distilbert/transformer/layer.2/ffn/lin1/Add" type="Add" version="opset1">
6608 <data auto_broadcast="numpy" />
6609 <rt_info>
6610 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/lin1/Add" />
6611 </rt_info>
6612 <input>
6613 <port id="0" precision="FP32">
6614 <dim>3072</dim>
6615 </port>
6616 <port id="1" precision="FP32">
6617 <dim>-1</dim>
6618 <dim>-1</dim>
6619 <dim>3072</dim>
6620 </port>
6621 </input>
6622 <output>
6623 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/lin1/Add_output_0">
6624 <dim>-1</dim>
6625 <dim>-1</dim>
6626 <dim>3072</dim>
6627 </port>
6628 </output>
6629 </layer>
6630 <layer id="395" name="/distilbert/transformer/layer.2/ffn/activation/Constant" type="Const" version="opset1">
6631 <data element_type="f32" shape="" offset="109659272" size="4" />
6632 <rt_info>
6633 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Constant" />
6634 </rt_info>
6635 <output>
6636 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Constant_output_0" />
6637 </output>
6638 </layer>
6639 <layer id="396" name="/distilbert/transformer/layer.2/ffn/activation/Div" type="Divide" version="opset1">
6640 <data auto_broadcast="numpy" m_pythondiv="true" />
6641 <rt_info>
6642 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Div" />
6643 </rt_info>
6644 <input>
6645 <port id="0" precision="FP32">
6646 <dim>-1</dim>
6647 <dim>-1</dim>
6648 <dim>3072</dim>
6649 </port>
6650 <port id="1" precision="FP32" />
6651 </input>
6652 <output>
6653 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Div_output_0">
6654 <dim>-1</dim>
6655 <dim>-1</dim>
6656 <dim>3072</dim>
6657 </port>
6658 </output>
6659 </layer>
6660 <layer id="397" name="/distilbert/transformer/layer.2/ffn/activation/Erf" type="Erf" version="opset1">
6661 <rt_info>
6662 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Erf" />
6663 </rt_info>
6664 <input>
6665 <port id="0" precision="FP32">
6666 <dim>-1</dim>
6667 <dim>-1</dim>
6668 <dim>3072</dim>
6669 </port>
6670 </input>
6671 <output>
6672 <port id="1" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Erf_output_0">
6673 <dim>-1</dim>
6674 <dim>-1</dim>
6675 <dim>3072</dim>
6676 </port>
6677 </output>
6678 </layer>
6679 <layer id="398" name="/distilbert/transformer/layer.2/ffn/activation/Constant_1" type="Const" version="opset1">
6680 <data element_type="f32" shape="" offset="109659276" size="4" />
6681 <rt_info>
6682 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Constant_1" />
6683 </rt_info>
6684 <output>
6685 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Constant_1_output_0" />
6686 </output>
6687 </layer>
6688 <layer id="399" name="/distilbert/transformer/layer.2/ffn/activation/Add" type="Add" version="opset1">
6689 <data auto_broadcast="numpy" />
6690 <rt_info>
6691 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Add" />
6692 </rt_info>
6693 <input>
6694 <port id="0" precision="FP32">
6695 <dim>-1</dim>
6696 <dim>-1</dim>
6697 <dim>3072</dim>
6698 </port>
6699 <port id="1" precision="FP32" />
6700 </input>
6701 <output>
6702 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Add_output_0">
6703 <dim>-1</dim>
6704 <dim>-1</dim>
6705 <dim>3072</dim>
6706 </port>
6707 </output>
6708 </layer>
6709 <layer id="400" name="/distilbert/transformer/layer.2/ffn/activation/Mul" type="Multiply" version="opset1">
6710 <data auto_broadcast="numpy" />
6711 <rt_info>
6712 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Mul" />
6713 </rt_info>
6714 <input>
6715 <port id="0" precision="FP32">
6716 <dim>-1</dim>
6717 <dim>-1</dim>
6718 <dim>3072</dim>
6719 </port>
6720 <port id="1" precision="FP32">
6721 <dim>-1</dim>
6722 <dim>-1</dim>
6723 <dim>3072</dim>
6724 </port>
6725 </input>
6726 <output>
6727 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Mul_output_0">
6728 <dim>-1</dim>
6729 <dim>-1</dim>
6730 <dim>3072</dim>
6731 </port>
6732 </output>
6733 </layer>
6734 <layer id="401" name="/distilbert/transformer/layer.2/ffn/activation/Constant_2" type="Const" version="opset1">
6735 <data element_type="f32" shape="" offset="109659280" size="4" />
6736 <rt_info>
6737 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Constant_2" />
6738 </rt_info>
6739 <output>
6740 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Constant_2_output_0" />
6741 </output>
6742 </layer>
6743 <layer id="402" name="/distilbert/transformer/layer.2/ffn/activation/Mul_1" type="Multiply" version="opset1">
6744 <data auto_broadcast="numpy" />
6745 <rt_info>
6746 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/activation/Mul_1" />
6747 </rt_info>
6748 <input>
6749 <port id="0" precision="FP32">
6750 <dim>-1</dim>
6751 <dim>-1</dim>
6752 <dim>3072</dim>
6753 </port>
6754 <port id="1" precision="FP32" />
6755 </input>
6756 <output>
6757 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/activation/Mul_1_output_0">
6758 <dim>-1</dim>
6759 <dim>-1</dim>
6760 <dim>3072</dim>
6761 </port>
6762 </output>
6763 </layer>
6764 <layer id="403" name="onnx::MatMul_794" type="Const" version="opset1">
6765 <data element_type="f32" shape="3072, 768" offset="166325396" size="9437184" />
6766 <rt_info>
6767 <attribute name="fused_names" version="0" value="onnx::MatMul_794" />
6768 </rt_info>
6769 <output>
6770 <port id="0" precision="FP32" names="onnx::MatMul_794">
6771 <dim>3072</dim>
6772 <dim>768</dim>
6773 </port>
6774 </output>
6775 </layer>
6776 <layer id="404" name="/distilbert/transformer/layer.2/ffn/lin2/MatMul" type="MatMul" version="opset1">
6777 <data transpose_a="false" transpose_b="false" />
6778 <rt_info>
6779 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/lin2/MatMul" />
6780 </rt_info>
6781 <input>
6782 <port id="0" precision="FP32">
6783 <dim>-1</dim>
6784 <dim>-1</dim>
6785 <dim>3072</dim>
6786 </port>
6787 <port id="1" precision="FP32">
6788 <dim>3072</dim>
6789 <dim>768</dim>
6790 </port>
6791 </input>
6792 <output>
6793 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/lin2/MatMul_output_0">
6794 <dim>-1</dim>
6795 <dim>-1</dim>
6796 <dim>768</dim>
6797 </port>
6798 </output>
6799 </layer>
6800 <layer id="405" name="/distilbert/transformer/layer.2/ffn/lin2/Add" type="Add" version="opset1">
6801 <data auto_broadcast="numpy" />
6802 <rt_info>
6803 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/ffn/lin2/Add" />
6804 </rt_info>
6805 <input>
6806 <port id="0" precision="FP32">
6807 <dim>768</dim>
6808 </port>
6809 <port id="1" precision="FP32">
6810 <dim>-1</dim>
6811 <dim>-1</dim>
6812 <dim>768</dim>
6813 </port>
6814 </input>
6815 <output>
6816 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/ffn/lin2/Add_output_0">
6817 <dim>-1</dim>
6818 <dim>-1</dim>
6819 <dim>768</dim>
6820 </port>
6821 </output>
6822 </layer>
6823 <layer id="406" name="/distilbert/transformer/layer.2/Add_1" type="Add" version="opset1">
6824 <data auto_broadcast="numpy" />
6825 <rt_info>
6826 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/Add_1" />
6827 </rt_info>
6828 <input>
6829 <port id="0" precision="FP32">
6830 <dim>-1</dim>
6831 <dim>-1</dim>
6832 <dim>768</dim>
6833 </port>
6834 <port id="1" precision="FP32">
6835 <dim>-1</dim>
6836 <dim>-1</dim>
6837 <dim>768</dim>
6838 </port>
6839 </input>
6840 <output>
6841 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/Add_1_output_0">
6842 <dim>-1</dim>
6843 <dim>-1</dim>
6844 <dim>768</dim>
6845 </port>
6846 </output>
6847 </layer>
6848 <layer id="407" name="Constant_255653" type="Const" version="opset1">
6849 <data element_type="i64" shape="1" offset="90763288" size="8" />
6850 <rt_info>
6851 <attribute name="fused_names" version="0" value="Constant_255653" />
6852 </rt_info>
6853 <output>
6854 <port id="0" precision="I64">
6855 <dim>1</dim>
6856 </port>
6857 </output>
6858 </layer>
6859 <layer id="408" name="/distilbert/transformer/layer.2/output_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
6860 <data keep_dims="true" />
6861 <rt_info>
6862 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/ReduceMean" />
6863 </rt_info>
6864 <input>
6865 <port id="0" precision="FP32">
6866 <dim>-1</dim>
6867 <dim>-1</dim>
6868 <dim>768</dim>
6869 </port>
6870 <port id="1" precision="I64">
6871 <dim>1</dim>
6872 </port>
6873 </input>
6874 <output>
6875 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/ReduceMean_output_0">
6876 <dim>-1</dim>
6877 <dim>-1</dim>
6878 <dim>1</dim>
6879 </port>
6880 </output>
6881 </layer>
6882 <layer id="409" name="/distilbert/transformer/layer.2/output_layer_norm/Sub" type="Subtract" version="opset1">
6883 <data auto_broadcast="numpy" />
6884 <rt_info>
6885 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Sub" />
6886 </rt_info>
6887 <input>
6888 <port id="0" precision="FP32">
6889 <dim>-1</dim>
6890 <dim>-1</dim>
6891 <dim>768</dim>
6892 </port>
6893 <port id="1" precision="FP32">
6894 <dim>-1</dim>
6895 <dim>-1</dim>
6896 <dim>1</dim>
6897 </port>
6898 </input>
6899 <output>
6900 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Sub_output_0">
6901 <dim>-1</dim>
6902 <dim>-1</dim>
6903 <dim>768</dim>
6904 </port>
6905 </output>
6906 </layer>
6907 <layer id="410" name="/distilbert/transformer/layer.2/output_layer_norm/Constant" type="Const" version="opset1">
6908 <data element_type="f32" shape="" offset="90763296" size="4" />
6909 <rt_info>
6910 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Constant" />
6911 </rt_info>
6912 <output>
6913 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Constant_output_0" />
6914 </output>
6915 </layer>
6916 <layer id="411" name="/distilbert/transformer/layer.2/output_layer_norm/Pow" type="Power" version="opset1">
6917 <data auto_broadcast="numpy" />
6918 <rt_info>
6919 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Pow" />
6920 </rt_info>
6921 <input>
6922 <port id="0" precision="FP32">
6923 <dim>-1</dim>
6924 <dim>-1</dim>
6925 <dim>768</dim>
6926 </port>
6927 <port id="1" precision="FP32" />
6928 </input>
6929 <output>
6930 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Pow_output_0">
6931 <dim>-1</dim>
6932 <dim>-1</dim>
6933 <dim>768</dim>
6934 </port>
6935 </output>
6936 </layer>
6937 <layer id="412" name="Constant_255658" type="Const" version="opset1">
6938 <data element_type="i64" shape="1" offset="90763288" size="8" />
6939 <rt_info>
6940 <attribute name="fused_names" version="0" value="Constant_255658" />
6941 </rt_info>
6942 <output>
6943 <port id="0" precision="I64">
6944 <dim>1</dim>
6945 </port>
6946 </output>
6947 </layer>
6948 <layer id="413" name="/distilbert/transformer/layer.2/output_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
6949 <data keep_dims="true" />
6950 <rt_info>
6951 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/ReduceMean_1" />
6952 </rt_info>
6953 <input>
6954 <port id="0" precision="FP32">
6955 <dim>-1</dim>
6956 <dim>-1</dim>
6957 <dim>768</dim>
6958 </port>
6959 <port id="1" precision="I64">
6960 <dim>1</dim>
6961 </port>
6962 </input>
6963 <output>
6964 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/ReduceMean_1_output_0">
6965 <dim>-1</dim>
6966 <dim>-1</dim>
6967 <dim>1</dim>
6968 </port>
6969 </output>
6970 </layer>
6971 <layer id="414" name="/distilbert/transformer/layer.2/output_layer_norm/Constant_1" type="Const" version="opset1">
6972 <data element_type="f32" shape="" offset="90763300" size="4" />
6973 <rt_info>
6974 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Constant_1" />
6975 </rt_info>
6976 <output>
6977 <port id="0" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Constant_1_output_0" />
6978 </output>
6979 </layer>
6980 <layer id="415" name="/distilbert/transformer/layer.2/output_layer_norm/Add" type="Add" version="opset1">
6981 <data auto_broadcast="numpy" />
6982 <rt_info>
6983 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Add" />
6984 </rt_info>
6985 <input>
6986 <port id="0" precision="FP32">
6987 <dim>-1</dim>
6988 <dim>-1</dim>
6989 <dim>1</dim>
6990 </port>
6991 <port id="1" precision="FP32" />
6992 </input>
6993 <output>
6994 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Add_output_0">
6995 <dim>-1</dim>
6996 <dim>-1</dim>
6997 <dim>1</dim>
6998 </port>
6999 </output>
7000 </layer>
7001 <layer id="416" name="/distilbert/transformer/layer.2/output_layer_norm/Sqrt" type="Sqrt" version="opset1">
7002 <rt_info>
7003 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Sqrt" />
7004 </rt_info>
7005 <input>
7006 <port id="0" precision="FP32">
7007 <dim>-1</dim>
7008 <dim>-1</dim>
7009 <dim>1</dim>
7010 </port>
7011 </input>
7012 <output>
7013 <port id="1" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Sqrt_output_0">
7014 <dim>-1</dim>
7015 <dim>-1</dim>
7016 <dim>1</dim>
7017 </port>
7018 </output>
7019 </layer>
7020 <layer id="417" name="/distilbert/transformer/layer.2/output_layer_norm/Div" type="Divide" version="opset1">
7021 <data auto_broadcast="numpy" m_pythondiv="true" />
7022 <rt_info>
7023 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Div" />
7024 </rt_info>
7025 <input>
7026 <port id="0" precision="FP32">
7027 <dim>-1</dim>
7028 <dim>-1</dim>
7029 <dim>768</dim>
7030 </port>
7031 <port id="1" precision="FP32">
7032 <dim>-1</dim>
7033 <dim>-1</dim>
7034 <dim>1</dim>
7035 </port>
7036 </input>
7037 <output>
7038 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Div_output_0">
7039 <dim>-1</dim>
7040 <dim>-1</dim>
7041 <dim>768</dim>
7042 </port>
7043 </output>
7044 </layer>
7045 <layer id="418" name="distilbert.transformer.layer.2.output_layer_norm.weight" type="Const" version="opset1">
7046 <data element_type="f32" shape="768" offset="175762580" size="3072" />
7047 <rt_info>
7048 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.output_layer_norm.weight" />
7049 </rt_info>
7050 <output>
7051 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.output_layer_norm.weight">
7052 <dim>768</dim>
7053 </port>
7054 </output>
7055 </layer>
7056 <layer id="419" name="/distilbert/transformer/layer.2/output_layer_norm/Mul" type="Multiply" version="opset1">
7057 <data auto_broadcast="numpy" />
7058 <rt_info>
7059 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Mul" />
7060 </rt_info>
7061 <input>
7062 <port id="0" precision="FP32">
7063 <dim>-1</dim>
7064 <dim>-1</dim>
7065 <dim>768</dim>
7066 </port>
7067 <port id="1" precision="FP32">
7068 <dim>768</dim>
7069 </port>
7070 </input>
7071 <output>
7072 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Mul_output_0">
7073 <dim>-1</dim>
7074 <dim>-1</dim>
7075 <dim>768</dim>
7076 </port>
7077 </output>
7078 </layer>
7079 <layer id="420" name="distilbert.transformer.layer.2.output_layer_norm.bias" type="Const" version="opset1">
7080 <data element_type="f32" shape="768" offset="175765652" size="3072" />
7081 <rt_info>
7082 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.2.output_layer_norm.bias" />
7083 </rt_info>
7084 <output>
7085 <port id="0" precision="FP32" names="distilbert.transformer.layer.2.output_layer_norm.bias">
7086 <dim>768</dim>
7087 </port>
7088 </output>
7089 </layer>
7090 <layer id="421" name="/distilbert/transformer/layer.2/output_layer_norm/Add_1" type="Add" version="opset1">
7091 <data auto_broadcast="numpy" />
7092 <rt_info>
7093 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.2/output_layer_norm/Add_1" />
7094 </rt_info>
7095 <input>
7096 <port id="0" precision="FP32">
7097 <dim>-1</dim>
7098 <dim>-1</dim>
7099 <dim>768</dim>
7100 </port>
7101 <port id="1" precision="FP32">
7102 <dim>768</dim>
7103 </port>
7104 </input>
7105 <output>
7106 <port id="2" precision="FP32" names="/distilbert/transformer/layer.2/output_layer_norm/Add_1_output_0">
7107 <dim>-1</dim>
7108 <dim>-1</dim>
7109 <dim>768</dim>
7110 </port>
7111 </output>
7112 </layer>
7113 <layer id="422" name="/distilbert/transformer/layer.3/attention/Shape" type="ShapeOf" version="opset3">
7114 <data output_type="i64" />
7115 <rt_info>
7116 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Shape" />
7117 </rt_info>
7118 <input>
7119 <port id="0" precision="FP32">
7120 <dim>-1</dim>
7121 <dim>-1</dim>
7122 <dim>768</dim>
7123 </port>
7124 </input>
7125 <output>
7126 <port id="1" precision="I64" names="/distilbert/transformer/layer.3/attention/Shape_output_0">
7127 <dim>3</dim>
7128 </port>
7129 </output>
7130 </layer>
7131 <layer id="423" name="/distilbert/transformer/layer.3/attention/Constant" type="Const" version="opset1">
7132 <data element_type="i64" shape="" offset="18440" size="8" />
7133 <rt_info>
7134 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant" />
7135 </rt_info>
7136 <output>
7137 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_output_0" />
7138 </output>
7139 </layer>
7140 <layer id="424" name="Constant_255668" type="Const" version="opset1">
7141 <data element_type="i64" shape="" offset="18440" size="8" />
7142 <rt_info>
7143 <attribute name="fused_names" version="0" value="Constant_255668" />
7144 </rt_info>
7145 <output>
7146 <port id="0" precision="I64" />
7147 </output>
7148 </layer>
7149 <layer id="425" name="/distilbert/transformer/layer.3/attention/Gather" type="Gather" version="opset8">
7150 <data batch_dims="0" />
7151 <rt_info>
7152 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant, /distilbert/transformer/layer.3/attention/Gather, Constant_255668" />
7153 </rt_info>
7154 <input>
7155 <port id="0" precision="I64">
7156 <dim>3</dim>
7157 </port>
7158 <port id="1" precision="I64" />
7159 <port id="2" precision="I64" />
7160 </input>
7161 <output>
7162 <port id="3" precision="I64" names="/distilbert/transformer/layer.3/attention/Gather_output_0" />
7163 </output>
7164 </layer>
7165 <layer id="426" name="Constant_255728" type="Const" version="opset1">
7166 <data element_type="i64" shape="1" offset="18440" size="8" />
7167 <rt_info>
7168 <attribute name="fused_names" version="0" value="Constant_255728" />
7169 </rt_info>
7170 <output>
7171 <port id="0" precision="I64">
7172 <dim>1</dim>
7173 </port>
7174 </output>
7175 </layer>
7176 <layer id="427" name="/distilbert/transformer/layer.3/attention/Unsqueeze_3" type="Unsqueeze" version="opset1">
7177 <rt_info>
7178 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Unsqueeze_3, Constant_255728" />
7179 </rt_info>
7180 <input>
7181 <port id="0" precision="I64" />
7182 <port id="1" precision="I64">
7183 <dim>1</dim>
7184 </port>
7185 </input>
7186 <output>
7187 <port id="2" precision="I64" names="/distilbert/transformer/layer.3/attention/Unsqueeze_3_output_0">
7188 <dim>1</dim>
7189 </port>
7190 </output>
7191 </layer>
7192 <layer id="428" name="/distilbert/transformer/layer.3/attention/Constant_12" type="Const" version="opset1">
7193 <data element_type="i64" shape="1" offset="90763280" size="8" />
7194 <rt_info>
7195 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_12" />
7196 </rt_info>
7197 <output>
7198 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_12_output_0">
7199 <dim>1</dim>
7200 </port>
7201 </output>
7202 </layer>
7203 <layer id="429" name="/distilbert/transformer/layer.3/attention/Constant_13" type="Const" version="opset1">
7204 <data element_type="i64" shape="1" offset="90763280" size="8" />
7205 <rt_info>
7206 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_13" />
7207 </rt_info>
7208 <output>
7209 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_13_output_0">
7210 <dim>1</dim>
7211 </port>
7212 </output>
7213 </layer>
7214 <layer id="430" name="/distilbert/transformer/layer.3/attention/Shape_1" type="ShapeOf" version="opset3">
7215 <data output_type="i64" />
7216 <rt_info>
7217 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Shape_1" />
7218 </rt_info>
7219 <input>
7220 <port id="0" precision="FP32">
7221 <dim>-1</dim>
7222 <dim>-1</dim>
7223 <dim>768</dim>
7224 </port>
7225 </input>
7226 <output>
7227 <port id="1" precision="I64" names="/distilbert/transformer/layer.3/attention/Shape_1_output_0">
7228 <dim>3</dim>
7229 </port>
7230 </output>
7231 </layer>
7232 <layer id="431" name="/distilbert/transformer/layer.3/attention/Constant_1" type="Const" version="opset1">
7233 <data element_type="i64" shape="" offset="90763280" size="8" />
7234 <rt_info>
7235 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_1" />
7236 </rt_info>
7237 <output>
7238 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_1_output_0" />
7239 </output>
7240 </layer>
7241 <layer id="432" name="Constant_255672" type="Const" version="opset1">
7242 <data element_type="i64" shape="" offset="18440" size="8" />
7243 <rt_info>
7244 <attribute name="fused_names" version="0" value="Constant_255672" />
7245 </rt_info>
7246 <output>
7247 <port id="0" precision="I64" />
7248 </output>
7249 </layer>
7250 <layer id="433" name="/distilbert/transformer/layer.3/attention/Gather_1" type="Gather" version="opset8">
7251 <data batch_dims="0" />
7252 <rt_info>
7253 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_1, /distilbert/transformer/layer.3/attention/Gather_1, Constant_255672" />
7254 </rt_info>
7255 <input>
7256 <port id="0" precision="I64">
7257 <dim>3</dim>
7258 </port>
7259 <port id="1" precision="I64" />
7260 <port id="2" precision="I64" />
7261 </input>
7262 <output>
7263 <port id="3" precision="I64" names="/distilbert/transformer/layer.3/attention/Gather_1_output_0" />
7264 </output>
7265 </layer>
7266 <layer id="434" name="Constant_255732" type="Const" version="opset1">
7267 <data element_type="i64" shape="1" offset="18440" size="8" />
7268 <rt_info>
7269 <attribute name="fused_names" version="0" value="Constant_255732" />
7270 </rt_info>
7271 <output>
7272 <port id="0" precision="I64">
7273 <dim>1</dim>
7274 </port>
7275 </output>
7276 </layer>
7277 <layer id="435" name="/distilbert/transformer/layer.3/attention/Unsqueeze_4" type="Unsqueeze" version="opset1">
7278 <rt_info>
7279 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Unsqueeze_4, Constant_255732" />
7280 </rt_info>
7281 <input>
7282 <port id="0" precision="I64" />
7283 <port id="1" precision="I64">
7284 <dim>1</dim>
7285 </port>
7286 </input>
7287 <output>
7288 <port id="2" precision="I64" names="/distilbert/transformer/layer.3/attention/Unsqueeze_4_output_0">
7289 <dim>1</dim>
7290 </port>
7291 </output>
7292 </layer>
7293 <layer id="436" name="/distilbert/transformer/layer.3/attention/Concat_3" type="Concat" version="opset1">
7294 <data axis="0" />
7295 <rt_info>
7296 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Concat_3, /distilbert/transformer/layer.3/attention/Constant_12, /distilbert/transformer/layer.3/attention/Constant_13" />
7297 </rt_info>
7298 <input>
7299 <port id="0" precision="I64">
7300 <dim>1</dim>
7301 </port>
7302 <port id="1" precision="I64">
7303 <dim>1</dim>
7304 </port>
7305 <port id="2" precision="I64">
7306 <dim>1</dim>
7307 </port>
7308 <port id="3" precision="I64">
7309 <dim>1</dim>
7310 </port>
7311 </input>
7312 <output>
7313 <port id="4" precision="I64" names="/distilbert/transformer/layer.3/attention/Concat_3_output_0">
7314 <dim>4</dim>
7315 </port>
7316 </output>
7317 </layer>
7318 <layer id="437" name="/distilbert/transformer/layer.3/attention/Reshape_3" type="Reshape" version="opset1">
7319 <data special_zero="true" />
7320 <rt_info>
7321 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Reshape_3" />
7322 </rt_info>
7323 <input>
7324 <port id="0" precision="BOOL">
7325 <dim>-1</dim>
7326 <dim>-1</dim>
7327 </port>
7328 <port id="1" precision="I64">
7329 <dim>4</dim>
7330 </port>
7331 </input>
7332 <output>
7333 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.3/attention/Reshape_3_output_0">
7334 <dim>-1</dim>
7335 <dim>1</dim>
7336 <dim>1</dim>
7337 <dim>-1</dim>
7338 </port>
7339 </output>
7340 </layer>
7341 <layer id="438" name="distilbert.transformer.layer.3.attention.q_lin.bias" type="Const" version="opset1">
7342 <data element_type="f32" shape="768" offset="175768724" size="3072" />
7343 <rt_info>
7344 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.attention.q_lin.bias" />
7345 </rt_info>
7346 <output>
7347 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.attention.q_lin.bias">
7348 <dim>768</dim>
7349 </port>
7350 </output>
7351 </layer>
7352 <layer id="439" name="onnx::MatMul_795" type="Const" version="opset1">
7353 <data element_type="f32" shape="768, 768" offset="175771796" size="2359296" />
7354 <rt_info>
7355 <attribute name="fused_names" version="0" value="onnx::MatMul_795" />
7356 </rt_info>
7357 <output>
7358 <port id="0" precision="FP32" names="onnx::MatMul_795">
7359 <dim>768</dim>
7360 <dim>768</dim>
7361 </port>
7362 </output>
7363 </layer>
7364 <layer id="440" name="/distilbert/transformer/layer.3/attention/q_lin/MatMul" type="MatMul" version="opset1">
7365 <data transpose_a="false" transpose_b="false" />
7366 <rt_info>
7367 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/q_lin/MatMul" />
7368 </rt_info>
7369 <input>
7370 <port id="0" precision="FP32">
7371 <dim>-1</dim>
7372 <dim>-1</dim>
7373 <dim>768</dim>
7374 </port>
7375 <port id="1" precision="FP32">
7376 <dim>768</dim>
7377 <dim>768</dim>
7378 </port>
7379 </input>
7380 <output>
7381 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/q_lin/MatMul_output_0">
7382 <dim>-1</dim>
7383 <dim>-1</dim>
7384 <dim>768</dim>
7385 </port>
7386 </output>
7387 </layer>
7388 <layer id="441" name="/distilbert/transformer/layer.3/attention/q_lin/Add" type="Add" version="opset1">
7389 <data auto_broadcast="numpy" />
7390 <rt_info>
7391 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/q_lin/Add" />
7392 </rt_info>
7393 <input>
7394 <port id="0" precision="FP32">
7395 <dim>768</dim>
7396 </port>
7397 <port id="1" precision="FP32">
7398 <dim>-1</dim>
7399 <dim>-1</dim>
7400 <dim>768</dim>
7401 </port>
7402 </input>
7403 <output>
7404 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/q_lin/Add_output_0">
7405 <dim>-1</dim>
7406 <dim>-1</dim>
7407 <dim>768</dim>
7408 </port>
7409 </output>
7410 </layer>
7411 <layer id="442" name="Constant_255676" type="Const" version="opset1">
7412 <data element_type="i64" shape="1" offset="18440" size="8" />
7413 <rt_info>
7414 <attribute name="fused_names" version="0" value="Constant_255676" />
7415 </rt_info>
7416 <output>
7417 <port id="0" precision="I64">
7418 <dim>1</dim>
7419 </port>
7420 </output>
7421 </layer>
7422 <layer id="443" name="/distilbert/transformer/layer.3/attention/Unsqueeze" type="Unsqueeze" version="opset1">
7423 <rt_info>
7424 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Unsqueeze, Constant_255676" />
7425 </rt_info>
7426 <input>
7427 <port id="0" precision="I64" />
7428 <port id="1" precision="I64">
7429 <dim>1</dim>
7430 </port>
7431 </input>
7432 <output>
7433 <port id="2" precision="I64" names="/distilbert/transformer/layer.3/attention/Unsqueeze_output_0">
7434 <dim>1</dim>
7435 </port>
7436 </output>
7437 </layer>
7438 <layer id="444" name="/distilbert/transformer/layer.3/attention/Constant_2" type="Const" version="opset1">
7439 <data element_type="i64" shape="1" offset="90763288" size="8" />
7440 <rt_info>
7441 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_2" />
7442 </rt_info>
7443 <output>
7444 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_2_output_0">
7445 <dim>1</dim>
7446 </port>
7447 </output>
7448 </layer>
7449 <layer id="445" name="/distilbert/transformer/layer.3/attention/Constant_3" type="Const" version="opset1">
7450 <data element_type="i64" shape="1" offset="93131816" size="8" />
7451 <rt_info>
7452 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_3" />
7453 </rt_info>
7454 <output>
7455 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_3_output_0">
7456 <dim>1</dim>
7457 </port>
7458 </output>
7459 </layer>
7460 <layer id="446" name="/distilbert/transformer/layer.3/attention/Constant_4" type="Const" version="opset1">
7461 <data element_type="i64" shape="1" offset="93131824" size="8" />
7462 <rt_info>
7463 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_4" />
7464 </rt_info>
7465 <output>
7466 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_4_output_0">
7467 <dim>1</dim>
7468 </port>
7469 </output>
7470 </layer>
7471 <layer id="447" name="/distilbert/transformer/layer.3/attention/Concat" type="Concat" version="opset1">
7472 <data axis="0" />
7473 <rt_info>
7474 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Concat, /distilbert/transformer/layer.3/attention/Constant_2, /distilbert/transformer/layer.3/attention/Constant_3, /distilbert/transformer/layer.3/attention/Constant_4" />
7475 </rt_info>
7476 <input>
7477 <port id="0" precision="I64">
7478 <dim>1</dim>
7479 </port>
7480 <port id="1" precision="I64">
7481 <dim>1</dim>
7482 </port>
7483 <port id="2" precision="I64">
7484 <dim>1</dim>
7485 </port>
7486 <port id="3" precision="I64">
7487 <dim>1</dim>
7488 </port>
7489 </input>
7490 <output>
7491 <port id="4" precision="I64" names="/distilbert/transformer/layer.3/attention/Concat_output_0">
7492 <dim>4</dim>
7493 </port>
7494 </output>
7495 </layer>
7496 <layer id="448" name="/distilbert/transformer/layer.3/attention/Reshape" type="Reshape" version="opset1">
7497 <data special_zero="true" />
7498 <rt_info>
7499 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Reshape" />
7500 </rt_info>
7501 <input>
7502 <port id="0" precision="FP32">
7503 <dim>-1</dim>
7504 <dim>-1</dim>
7505 <dim>768</dim>
7506 </port>
7507 <port id="1" precision="I64">
7508 <dim>4</dim>
7509 </port>
7510 </input>
7511 <output>
7512 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Reshape_output_0">
7513 <dim>-1</dim>
7514 <dim>-1</dim>
7515 <dim>12</dim>
7516 <dim>64</dim>
7517 </port>
7518 </output>
7519 </layer>
7520 <layer id="449" name="Constant_255701" type="Const" version="opset1">
7521 <data element_type="i64" shape="4" offset="93131832" size="32" />
7522 <rt_info>
7523 <attribute name="fused_names" version="0" value="Constant_255701" />
7524 </rt_info>
7525 <output>
7526 <port id="0" precision="I64">
7527 <dim>4</dim>
7528 </port>
7529 </output>
7530 </layer>
7531 <layer id="450" name="/distilbert/transformer/layer.3/attention/Transpose" type="Transpose" version="opset1">
7532 <rt_info>
7533 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Transpose" />
7534 </rt_info>
7535 <input>
7536 <port id="0" precision="FP32">
7537 <dim>-1</dim>
7538 <dim>-1</dim>
7539 <dim>12</dim>
7540 <dim>64</dim>
7541 </port>
7542 <port id="1" precision="I64">
7543 <dim>4</dim>
7544 </port>
7545 </input>
7546 <output>
7547 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Transpose_output_0">
7548 <dim>-1</dim>
7549 <dim>12</dim>
7550 <dim>-1</dim>
7551 <dim>64</dim>
7552 </port>
7553 </output>
7554 </layer>
7555 <layer id="451" name="/distilbert/transformer/layer.3/attention/Constant_11" type="Const" version="opset1">
7556 <data element_type="f32" shape="" offset="93131864" size="4" />
7557 <rt_info>
7558 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_11" />
7559 </rt_info>
7560 <output>
7561 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/attention/Constant_11_output_0" />
7562 </output>
7563 </layer>
7564 <layer id="452" name="/distilbert/transformer/layer.3/attention/Div" type="Divide" version="opset1">
7565 <data auto_broadcast="numpy" m_pythondiv="true" />
7566 <rt_info>
7567 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Div" />
7568 </rt_info>
7569 <input>
7570 <port id="0" precision="FP32">
7571 <dim>-1</dim>
7572 <dim>12</dim>
7573 <dim>-1</dim>
7574 <dim>64</dim>
7575 </port>
7576 <port id="1" precision="FP32" />
7577 </input>
7578 <output>
7579 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Div_output_0">
7580 <dim>-1</dim>
7581 <dim>12</dim>
7582 <dim>-1</dim>
7583 <dim>64</dim>
7584 </port>
7585 </output>
7586 </layer>
7587 <layer id="453" name="distilbert.transformer.layer.3.attention.k_lin.bias" type="Const" version="opset1">
7588 <data element_type="f32" shape="768" offset="178131092" size="3072" />
7589 <rt_info>
7590 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.attention.k_lin.bias" />
7591 </rt_info>
7592 <output>
7593 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.attention.k_lin.bias">
7594 <dim>768</dim>
7595 </port>
7596 </output>
7597 </layer>
7598 <layer id="454" name="onnx::MatMul_805" type="Const" version="opset1">
7599 <data element_type="f32" shape="768, 768" offset="178134164" size="2359296" />
7600 <rt_info>
7601 <attribute name="fused_names" version="0" value="onnx::MatMul_805" />
7602 </rt_info>
7603 <output>
7604 <port id="0" precision="FP32" names="onnx::MatMul_805">
7605 <dim>768</dim>
7606 <dim>768</dim>
7607 </port>
7608 </output>
7609 </layer>
7610 <layer id="455" name="/distilbert/transformer/layer.3/attention/k_lin/MatMul" type="MatMul" version="opset1">
7611 <data transpose_a="false" transpose_b="false" />
7612 <rt_info>
7613 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/k_lin/MatMul" />
7614 </rt_info>
7615 <input>
7616 <port id="0" precision="FP32">
7617 <dim>-1</dim>
7618 <dim>-1</dim>
7619 <dim>768</dim>
7620 </port>
7621 <port id="1" precision="FP32">
7622 <dim>768</dim>
7623 <dim>768</dim>
7624 </port>
7625 </input>
7626 <output>
7627 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/k_lin/MatMul_output_0">
7628 <dim>-1</dim>
7629 <dim>-1</dim>
7630 <dim>768</dim>
7631 </port>
7632 </output>
7633 </layer>
7634 <layer id="456" name="/distilbert/transformer/layer.3/attention/k_lin/Add" type="Add" version="opset1">
7635 <data auto_broadcast="numpy" />
7636 <rt_info>
7637 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/k_lin/Add" />
7638 </rt_info>
7639 <input>
7640 <port id="0" precision="FP32">
7641 <dim>768</dim>
7642 </port>
7643 <port id="1" precision="FP32">
7644 <dim>-1</dim>
7645 <dim>-1</dim>
7646 <dim>768</dim>
7647 </port>
7648 </input>
7649 <output>
7650 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/k_lin/Add_output_0">
7651 <dim>-1</dim>
7652 <dim>-1</dim>
7653 <dim>768</dim>
7654 </port>
7655 </output>
7656 </layer>
7657 <layer id="457" name="Constant_255682" type="Const" version="opset1">
7658 <data element_type="i64" shape="1" offset="18440" size="8" />
7659 <rt_info>
7660 <attribute name="fused_names" version="0" value="Constant_255682" />
7661 </rt_info>
7662 <output>
7663 <port id="0" precision="I64">
7664 <dim>1</dim>
7665 </port>
7666 </output>
7667 </layer>
7668 <layer id="458" name="/distilbert/transformer/layer.3/attention/Unsqueeze_1" type="Unsqueeze" version="opset1">
7669 <rt_info>
7670 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Unsqueeze_1, Constant_255682" />
7671 </rt_info>
7672 <input>
7673 <port id="0" precision="I64" />
7674 <port id="1" precision="I64">
7675 <dim>1</dim>
7676 </port>
7677 </input>
7678 <output>
7679 <port id="2" precision="I64" names="/distilbert/transformer/layer.3/attention/Unsqueeze_1_output_0">
7680 <dim>1</dim>
7681 </port>
7682 </output>
7683 </layer>
7684 <layer id="459" name="/distilbert/transformer/layer.3/attention/Constant_5" type="Const" version="opset1">
7685 <data element_type="i64" shape="1" offset="90763288" size="8" />
7686 <rt_info>
7687 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_5" />
7688 </rt_info>
7689 <output>
7690 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_5_output_0">
7691 <dim>1</dim>
7692 </port>
7693 </output>
7694 </layer>
7695 <layer id="460" name="/distilbert/transformer/layer.3/attention/Constant_6" type="Const" version="opset1">
7696 <data element_type="i64" shape="1" offset="93131816" size="8" />
7697 <rt_info>
7698 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_6" />
7699 </rt_info>
7700 <output>
7701 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_6_output_0">
7702 <dim>1</dim>
7703 </port>
7704 </output>
7705 </layer>
7706 <layer id="461" name="/distilbert/transformer/layer.3/attention/Constant_7" type="Const" version="opset1">
7707 <data element_type="i64" shape="1" offset="93131824" size="8" />
7708 <rt_info>
7709 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_7" />
7710 </rt_info>
7711 <output>
7712 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_7_output_0">
7713 <dim>1</dim>
7714 </port>
7715 </output>
7716 </layer>
7717 <layer id="462" name="/distilbert/transformer/layer.3/attention/Concat_1" type="Concat" version="opset1">
7718 <data axis="0" />
7719 <rt_info>
7720 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Concat_1, /distilbert/transformer/layer.3/attention/Constant_5, /distilbert/transformer/layer.3/attention/Constant_6, /distilbert/transformer/layer.3/attention/Constant_7" />
7721 </rt_info>
7722 <input>
7723 <port id="0" precision="I64">
7724 <dim>1</dim>
7725 </port>
7726 <port id="1" precision="I64">
7727 <dim>1</dim>
7728 </port>
7729 <port id="2" precision="I64">
7730 <dim>1</dim>
7731 </port>
7732 <port id="3" precision="I64">
7733 <dim>1</dim>
7734 </port>
7735 </input>
7736 <output>
7737 <port id="4" precision="I64" names="/distilbert/transformer/layer.3/attention/Concat_1_output_0">
7738 <dim>4</dim>
7739 </port>
7740 </output>
7741 </layer>
7742 <layer id="463" name="/distilbert/transformer/layer.3/attention/Reshape_1" type="Reshape" version="opset1">
7743 <data special_zero="true" />
7744 <rt_info>
7745 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Reshape_1" />
7746 </rt_info>
7747 <input>
7748 <port id="0" precision="FP32">
7749 <dim>-1</dim>
7750 <dim>-1</dim>
7751 <dim>768</dim>
7752 </port>
7753 <port id="1" precision="I64">
7754 <dim>4</dim>
7755 </port>
7756 </input>
7757 <output>
7758 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Reshape_1_output_0">
7759 <dim>-1</dim>
7760 <dim>-1</dim>
7761 <dim>12</dim>
7762 <dim>64</dim>
7763 </port>
7764 </output>
7765 </layer>
7766 <layer id="464" name="Constant_255725" type="Const" version="opset1">
7767 <data element_type="i64" shape="4" offset="95494236" size="32" />
7768 <rt_info>
7769 <attribute name="fused_names" version="0" value="Constant_255725" />
7770 </rt_info>
7771 <output>
7772 <port id="0" precision="I64">
7773 <dim>4</dim>
7774 </port>
7775 </output>
7776 </layer>
7777 <layer id="465" name="/distilbert/transformer/layer.3/attention/Transpose_2" type="Transpose" version="opset1">
7778 <rt_info>
7779 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Transpose_2" />
7780 </rt_info>
7781 <input>
7782 <port id="0" precision="FP32">
7783 <dim>-1</dim>
7784 <dim>-1</dim>
7785 <dim>12</dim>
7786 <dim>64</dim>
7787 </port>
7788 <port id="1" precision="I64">
7789 <dim>4</dim>
7790 </port>
7791 </input>
7792 <output>
7793 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Transpose_2_output_0">
7794 <dim>-1</dim>
7795 <dim>12</dim>
7796 <dim>64</dim>
7797 <dim>-1</dim>
7798 </port>
7799 </output>
7800 </layer>
7801 <layer id="466" name="/distilbert/transformer/layer.3/attention/MatMul" type="MatMul" version="opset1">
7802 <data transpose_a="false" transpose_b="false" />
7803 <rt_info>
7804 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/MatMul" />
7805 </rt_info>
7806 <input>
7807 <port id="0" precision="FP32">
7808 <dim>-1</dim>
7809 <dim>12</dim>
7810 <dim>-1</dim>
7811 <dim>64</dim>
7812 </port>
7813 <port id="1" precision="FP32">
7814 <dim>-1</dim>
7815 <dim>12</dim>
7816 <dim>64</dim>
7817 <dim>-1</dim>
7818 </port>
7819 </input>
7820 <output>
7821 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/MatMul_output_0">
7822 <dim>-1</dim>
7823 <dim>12</dim>
7824 <dim>-1</dim>
7825 <dim>-1</dim>
7826 </port>
7827 </output>
7828 </layer>
7829 <layer id="467" name="/distilbert/transformer/layer.3/attention/Shape_2" type="ShapeOf" version="opset3">
7830 <data output_type="i64" />
7831 <rt_info>
7832 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Shape_2" />
7833 </rt_info>
7834 <input>
7835 <port id="0" precision="FP32">
7836 <dim>-1</dim>
7837 <dim>12</dim>
7838 <dim>-1</dim>
7839 <dim>-1</dim>
7840 </port>
7841 </input>
7842 <output>
7843 <port id="1" precision="I64" names="/distilbert/transformer/layer.3/attention/Shape_2_output_0">
7844 <dim>4</dim>
7845 </port>
7846 </output>
7847 </layer>
7848 <layer id="468" name="/distilbert/transformer/layer.3/attention/Expand" type="Broadcast" version="opset3">
7849 <data mode="bidirectional" />
7850 <rt_info>
7851 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Expand" />
7852 </rt_info>
7853 <input>
7854 <port id="0" precision="BOOL">
7855 <dim>-1</dim>
7856 <dim>1</dim>
7857 <dim>1</dim>
7858 <dim>-1</dim>
7859 </port>
7860 <port id="1" precision="I64">
7861 <dim>4</dim>
7862 </port>
7863 </input>
7864 <output>
7865 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.3/attention/Expand_output_0">
7866 <dim>-1</dim>
7867 <dim>12</dim>
7868 <dim>-1</dim>
7869 <dim>-1</dim>
7870 </port>
7871 </output>
7872 </layer>
7873 <layer id="469" name="/distilbert/transformer/layer.3/attention/Cast" type="Convert" version="opset1">
7874 <data destination_type="boolean" />
7875 <rt_info>
7876 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Cast" />
7877 </rt_info>
7878 <input>
7879 <port id="0" precision="BOOL">
7880 <dim>-1</dim>
7881 <dim>12</dim>
7882 <dim>-1</dim>
7883 <dim>-1</dim>
7884 </port>
7885 </input>
7886 <output>
7887 <port id="1" precision="BOOL" names="/distilbert/transformer/layer.3/attention/Cast_output_0">
7888 <dim>-1</dim>
7889 <dim>12</dim>
7890 <dim>-1</dim>
7891 <dim>-1</dim>
7892 </port>
7893 </output>
7894 </layer>
7895 <layer id="470" name="/distilbert/transformer/layer.3/attention/Constant_14" type="Const" version="opset1">
7896 <data element_type="f32" shape="" offset="95494268" size="4" />
7897 <rt_info>
7898 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_14" />
7899 </rt_info>
7900 <output>
7901 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/attention/Constant_14_output_0" />
7902 </output>
7903 </layer>
7904 <layer id="471" name="/distilbert/transformer/layer.3/attention/Where" type="Select" version="opset1">
7905 <data auto_broadcast="numpy" />
7906 <rt_info>
7907 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Where" />
7908 </rt_info>
7909 <input>
7910 <port id="0" precision="BOOL">
7911 <dim>-1</dim>
7912 <dim>12</dim>
7913 <dim>-1</dim>
7914 <dim>-1</dim>
7915 </port>
7916 <port id="1" precision="FP32" />
7917 <port id="2" precision="FP32">
7918 <dim>-1</dim>
7919 <dim>12</dim>
7920 <dim>-1</dim>
7921 <dim>-1</dim>
7922 </port>
7923 </input>
7924 <output>
7925 <port id="3" precision="FP32" names="/distilbert/transformer/layer.3/attention/Where_output_0">
7926 <dim>-1</dim>
7927 <dim>12</dim>
7928 <dim>-1</dim>
7929 <dim>-1</dim>
7930 </port>
7931 </output>
7932 </layer>
7933 <layer id="472" name="/distilbert/transformer/layer.3/attention/Softmax" type="SoftMax" version="opset8">
7934 <data axis="3" />
7935 <rt_info>
7936 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Softmax" />
7937 </rt_info>
7938 <input>
7939 <port id="0" precision="FP32">
7940 <dim>-1</dim>
7941 <dim>12</dim>
7942 <dim>-1</dim>
7943 <dim>-1</dim>
7944 </port>
7945 </input>
7946 <output>
7947 <port id="1" precision="FP32" names="/distilbert/transformer/layer.3/attention/Softmax_output_0">
7948 <dim>-1</dim>
7949 <dim>12</dim>
7950 <dim>-1</dim>
7951 <dim>-1</dim>
7952 </port>
7953 </output>
7954 </layer>
7955 <layer id="473" name="distilbert.transformer.layer.3.attention.v_lin.bias" type="Const" version="opset1">
7956 <data element_type="f32" shape="768" offset="180493460" size="3072" />
7957 <rt_info>
7958 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.attention.v_lin.bias" />
7959 </rt_info>
7960 <output>
7961 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.attention.v_lin.bias">
7962 <dim>768</dim>
7963 </port>
7964 </output>
7965 </layer>
7966 <layer id="474" name="onnx::MatMul_806" type="Const" version="opset1">
7967 <data element_type="f32" shape="768, 768" offset="180496532" size="2359296" />
7968 <rt_info>
7969 <attribute name="fused_names" version="0" value="onnx::MatMul_806" />
7970 </rt_info>
7971 <output>
7972 <port id="0" precision="FP32" names="onnx::MatMul_806">
7973 <dim>768</dim>
7974 <dim>768</dim>
7975 </port>
7976 </output>
7977 </layer>
7978 <layer id="475" name="/distilbert/transformer/layer.3/attention/v_lin/MatMul" type="MatMul" version="opset1">
7979 <data transpose_a="false" transpose_b="false" />
7980 <rt_info>
7981 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/v_lin/MatMul" />
7982 </rt_info>
7983 <input>
7984 <port id="0" precision="FP32">
7985 <dim>-1</dim>
7986 <dim>-1</dim>
7987 <dim>768</dim>
7988 </port>
7989 <port id="1" precision="FP32">
7990 <dim>768</dim>
7991 <dim>768</dim>
7992 </port>
7993 </input>
7994 <output>
7995 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/v_lin/MatMul_output_0">
7996 <dim>-1</dim>
7997 <dim>-1</dim>
7998 <dim>768</dim>
7999 </port>
8000 </output>
8001 </layer>
8002 <layer id="476" name="/distilbert/transformer/layer.3/attention/v_lin/Add" type="Add" version="opset1">
8003 <data auto_broadcast="numpy" />
8004 <rt_info>
8005 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/v_lin/Add" />
8006 </rt_info>
8007 <input>
8008 <port id="0" precision="FP32">
8009 <dim>768</dim>
8010 </port>
8011 <port id="1" precision="FP32">
8012 <dim>-1</dim>
8013 <dim>-1</dim>
8014 <dim>768</dim>
8015 </port>
8016 </input>
8017 <output>
8018 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/v_lin/Add_output_0">
8019 <dim>-1</dim>
8020 <dim>-1</dim>
8021 <dim>768</dim>
8022 </port>
8023 </output>
8024 </layer>
8025 <layer id="477" name="Constant_255688" type="Const" version="opset1">
8026 <data element_type="i64" shape="1" offset="18440" size="8" />
8027 <rt_info>
8028 <attribute name="fused_names" version="0" value="Constant_255688" />
8029 </rt_info>
8030 <output>
8031 <port id="0" precision="I64">
8032 <dim>1</dim>
8033 </port>
8034 </output>
8035 </layer>
8036 <layer id="478" name="/distilbert/transformer/layer.3/attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
8037 <rt_info>
8038 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Unsqueeze_2, Constant_255688" />
8039 </rt_info>
8040 <input>
8041 <port id="0" precision="I64" />
8042 <port id="1" precision="I64">
8043 <dim>1</dim>
8044 </port>
8045 </input>
8046 <output>
8047 <port id="2" precision="I64" names="/distilbert/transformer/layer.3/attention/Unsqueeze_2_output_0">
8048 <dim>1</dim>
8049 </port>
8050 </output>
8051 </layer>
8052 <layer id="479" name="/distilbert/transformer/layer.3/attention/Constant_8" type="Const" version="opset1">
8053 <data element_type="i64" shape="1" offset="90763288" size="8" />
8054 <rt_info>
8055 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_8" />
8056 </rt_info>
8057 <output>
8058 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_8_output_0">
8059 <dim>1</dim>
8060 </port>
8061 </output>
8062 </layer>
8063 <layer id="480" name="/distilbert/transformer/layer.3/attention/Constant_9" type="Const" version="opset1">
8064 <data element_type="i64" shape="1" offset="93131816" size="8" />
8065 <rt_info>
8066 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_9" />
8067 </rt_info>
8068 <output>
8069 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_9_output_0">
8070 <dim>1</dim>
8071 </port>
8072 </output>
8073 </layer>
8074 <layer id="481" name="/distilbert/transformer/layer.3/attention/Constant_10" type="Const" version="opset1">
8075 <data element_type="i64" shape="1" offset="93131824" size="8" />
8076 <rt_info>
8077 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_10" />
8078 </rt_info>
8079 <output>
8080 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_10_output_0">
8081 <dim>1</dim>
8082 </port>
8083 </output>
8084 </layer>
8085 <layer id="482" name="/distilbert/transformer/layer.3/attention/Concat_2" type="Concat" version="opset1">
8086 <data axis="0" />
8087 <rt_info>
8088 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Concat_2, /distilbert/transformer/layer.3/attention/Constant_10, /distilbert/transformer/layer.3/attention/Constant_8, /distilbert/transformer/layer.3/attention/Constant_9" />
8089 </rt_info>
8090 <input>
8091 <port id="0" precision="I64">
8092 <dim>1</dim>
8093 </port>
8094 <port id="1" precision="I64">
8095 <dim>1</dim>
8096 </port>
8097 <port id="2" precision="I64">
8098 <dim>1</dim>
8099 </port>
8100 <port id="3" precision="I64">
8101 <dim>1</dim>
8102 </port>
8103 </input>
8104 <output>
8105 <port id="4" precision="I64" names="/distilbert/transformer/layer.3/attention/Concat_2_output_0">
8106 <dim>4</dim>
8107 </port>
8108 </output>
8109 </layer>
8110 <layer id="483" name="/distilbert/transformer/layer.3/attention/Reshape_2" type="Reshape" version="opset1">
8111 <data special_zero="true" />
8112 <rt_info>
8113 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Reshape_2" />
8114 </rt_info>
8115 <input>
8116 <port id="0" precision="FP32">
8117 <dim>-1</dim>
8118 <dim>-1</dim>
8119 <dim>768</dim>
8120 </port>
8121 <port id="1" precision="I64">
8122 <dim>4</dim>
8123 </port>
8124 </input>
8125 <output>
8126 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Reshape_2_output_0">
8127 <dim>-1</dim>
8128 <dim>-1</dim>
8129 <dim>12</dim>
8130 <dim>64</dim>
8131 </port>
8132 </output>
8133 </layer>
8134 <layer id="484" name="Constant_255721" type="Const" version="opset1">
8135 <data element_type="i64" shape="4" offset="93131832" size="32" />
8136 <rt_info>
8137 <attribute name="fused_names" version="0" value="Constant_255721" />
8138 </rt_info>
8139 <output>
8140 <port id="0" precision="I64">
8141 <dim>4</dim>
8142 </port>
8143 </output>
8144 </layer>
8145 <layer id="485" name="/distilbert/transformer/layer.3/attention/Transpose_1" type="Transpose" version="opset1">
8146 <rt_info>
8147 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Transpose_1" />
8148 </rt_info>
8149 <input>
8150 <port id="0" precision="FP32">
8151 <dim>-1</dim>
8152 <dim>-1</dim>
8153 <dim>12</dim>
8154 <dim>64</dim>
8155 </port>
8156 <port id="1" precision="I64">
8157 <dim>4</dim>
8158 </port>
8159 </input>
8160 <output>
8161 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Transpose_1_output_0">
8162 <dim>-1</dim>
8163 <dim>12</dim>
8164 <dim>-1</dim>
8165 <dim>64</dim>
8166 </port>
8167 </output>
8168 </layer>
8169 <layer id="486" name="/distilbert/transformer/layer.3/attention/MatMul_1" type="MatMul" version="opset1">
8170 <data transpose_a="false" transpose_b="false" />
8171 <rt_info>
8172 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/MatMul_1" />
8173 </rt_info>
8174 <input>
8175 <port id="0" precision="FP32">
8176 <dim>-1</dim>
8177 <dim>12</dim>
8178 <dim>-1</dim>
8179 <dim>-1</dim>
8180 </port>
8181 <port id="1" precision="FP32">
8182 <dim>-1</dim>
8183 <dim>12</dim>
8184 <dim>-1</dim>
8185 <dim>64</dim>
8186 </port>
8187 </input>
8188 <output>
8189 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/MatMul_1_output_0">
8190 <dim>-1</dim>
8191 <dim>12</dim>
8192 <dim>-1</dim>
8193 <dim>64</dim>
8194 </port>
8195 </output>
8196 </layer>
8197 <layer id="487" name="Constant_255757" type="Const" version="opset1">
8198 <data element_type="i64" shape="4" offset="93131832" size="32" />
8199 <rt_info>
8200 <attribute name="fused_names" version="0" value="Constant_255757" />
8201 </rt_info>
8202 <output>
8203 <port id="0" precision="I64">
8204 <dim>4</dim>
8205 </port>
8206 </output>
8207 </layer>
8208 <layer id="488" name="/distilbert/transformer/layer.3/attention/Transpose_3" type="Transpose" version="opset1">
8209 <rt_info>
8210 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Transpose_3" />
8211 </rt_info>
8212 <input>
8213 <port id="0" precision="FP32">
8214 <dim>-1</dim>
8215 <dim>12</dim>
8216 <dim>-1</dim>
8217 <dim>64</dim>
8218 </port>
8219 <port id="1" precision="I64">
8220 <dim>4</dim>
8221 </port>
8222 </input>
8223 <output>
8224 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Transpose_3_output_0">
8225 <dim>-1</dim>
8226 <dim>-1</dim>
8227 <dim>12</dim>
8228 <dim>64</dim>
8229 </port>
8230 </output>
8231 </layer>
8232 <layer id="489" name="Constant_255759" type="Const" version="opset1">
8233 <data element_type="i64" shape="1" offset="18440" size="8" />
8234 <rt_info>
8235 <attribute name="fused_names" version="0" value="Constant_255759" />
8236 </rt_info>
8237 <output>
8238 <port id="0" precision="I64">
8239 <dim>1</dim>
8240 </port>
8241 </output>
8242 </layer>
8243 <layer id="490" name="/distilbert/transformer/layer.3/attention/Unsqueeze_5" type="Unsqueeze" version="opset1">
8244 <rt_info>
8245 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Unsqueeze_5, Constant_255759" />
8246 </rt_info>
8247 <input>
8248 <port id="0" precision="I64" />
8249 <port id="1" precision="I64">
8250 <dim>1</dim>
8251 </port>
8252 </input>
8253 <output>
8254 <port id="2" precision="I64" names="/distilbert/transformer/layer.3/attention/Unsqueeze_5_output_0">
8255 <dim>1</dim>
8256 </port>
8257 </output>
8258 </layer>
8259 <layer id="491" name="/distilbert/transformer/layer.3/attention/Constant_15" type="Const" version="opset1">
8260 <data element_type="i64" shape="1" offset="90763288" size="8" />
8261 <rt_info>
8262 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_15" />
8263 </rt_info>
8264 <output>
8265 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_15_output_0">
8266 <dim>1</dim>
8267 </port>
8268 </output>
8269 </layer>
8270 <layer id="492" name="/distilbert/transformer/layer.3/attention/Constant_16" type="Const" version="opset1">
8271 <data element_type="i64" shape="1" offset="97856640" size="8" />
8272 <rt_info>
8273 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Constant_16" />
8274 </rt_info>
8275 <output>
8276 <port id="0" precision="I64" names="/distilbert/transformer/layer.3/attention/Constant_16_output_0">
8277 <dim>1</dim>
8278 </port>
8279 </output>
8280 </layer>
8281 <layer id="493" name="/distilbert/transformer/layer.3/attention/Concat_4" type="Concat" version="opset1">
8282 <data axis="0" />
8283 <rt_info>
8284 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Concat_4, /distilbert/transformer/layer.3/attention/Constant_15, /distilbert/transformer/layer.3/attention/Constant_16" />
8285 </rt_info>
8286 <input>
8287 <port id="0" precision="I64">
8288 <dim>1</dim>
8289 </port>
8290 <port id="1" precision="I64">
8291 <dim>1</dim>
8292 </port>
8293 <port id="2" precision="I64">
8294 <dim>1</dim>
8295 </port>
8296 </input>
8297 <output>
8298 <port id="3" precision="I64" names="/distilbert/transformer/layer.3/attention/Concat_4_output_0">
8299 <dim>3</dim>
8300 </port>
8301 </output>
8302 </layer>
8303 <layer id="494" name="/distilbert/transformer/layer.3/attention/Reshape_4" type="Reshape" version="opset1">
8304 <data special_zero="true" />
8305 <rt_info>
8306 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/Reshape_4" />
8307 </rt_info>
8308 <input>
8309 <port id="0" precision="FP32">
8310 <dim>-1</dim>
8311 <dim>-1</dim>
8312 <dim>12</dim>
8313 <dim>64</dim>
8314 </port>
8315 <port id="1" precision="I64">
8316 <dim>3</dim>
8317 </port>
8318 </input>
8319 <output>
8320 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/Reshape_4_output_0">
8321 <dim>-1</dim>
8322 <dim>-1</dim>
8323 <dim>768</dim>
8324 </port>
8325 </output>
8326 </layer>
8327 <layer id="495" name="onnx::MatMul_811" type="Const" version="opset1">
8328 <data element_type="f32" shape="768, 768" offset="182855828" size="2359296" />
8329 <rt_info>
8330 <attribute name="fused_names" version="0" value="onnx::MatMul_811" />
8331 </rt_info>
8332 <output>
8333 <port id="0" precision="FP32" names="onnx::MatMul_811">
8334 <dim>768</dim>
8335 <dim>768</dim>
8336 </port>
8337 </output>
8338 </layer>
8339 <layer id="496" name="/distilbert/transformer/layer.3/attention/out_lin/MatMul" type="MatMul" version="opset1">
8340 <data transpose_a="false" transpose_b="false" />
8341 <rt_info>
8342 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/out_lin/MatMul" />
8343 </rt_info>
8344 <input>
8345 <port id="0" precision="FP32">
8346 <dim>-1</dim>
8347 <dim>-1</dim>
8348 <dim>768</dim>
8349 </port>
8350 <port id="1" precision="FP32">
8351 <dim>768</dim>
8352 <dim>768</dim>
8353 </port>
8354 </input>
8355 <output>
8356 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/out_lin/MatMul_output_0">
8357 <dim>-1</dim>
8358 <dim>-1</dim>
8359 <dim>768</dim>
8360 </port>
8361 </output>
8362 </layer>
8363 <layer id="497" name="/distilbert/transformer/layer.3/attention/out_lin/Add" type="Add" version="opset1">
8364 <data auto_broadcast="numpy" />
8365 <rt_info>
8366 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/attention/out_lin/Add" />
8367 </rt_info>
8368 <input>
8369 <port id="0" precision="FP32">
8370 <dim>768</dim>
8371 </port>
8372 <port id="1" precision="FP32">
8373 <dim>-1</dim>
8374 <dim>-1</dim>
8375 <dim>768</dim>
8376 </port>
8377 </input>
8378 <output>
8379 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/attention/out_lin/Add_output_0">
8380 <dim>-1</dim>
8381 <dim>-1</dim>
8382 <dim>768</dim>
8383 </port>
8384 </output>
8385 </layer>
8386 <layer id="498" name="/distilbert/transformer/layer.3/Add" type="Add" version="opset1">
8387 <data auto_broadcast="numpy" />
8388 <rt_info>
8389 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/Add" />
8390 </rt_info>
8391 <input>
8392 <port id="0" precision="FP32">
8393 <dim>-1</dim>
8394 <dim>-1</dim>
8395 <dim>768</dim>
8396 </port>
8397 <port id="1" precision="FP32">
8398 <dim>-1</dim>
8399 <dim>-1</dim>
8400 <dim>768</dim>
8401 </port>
8402 </input>
8403 <output>
8404 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/Add_output_0">
8405 <dim>-1</dim>
8406 <dim>-1</dim>
8407 <dim>768</dim>
8408 </port>
8409 </output>
8410 </layer>
8411 <layer id="499" name="Constant_255774" type="Const" version="opset1">
8412 <data element_type="i64" shape="1" offset="90763288" size="8" />
8413 <rt_info>
8414 <attribute name="fused_names" version="0" value="Constant_255774" />
8415 </rt_info>
8416 <output>
8417 <port id="0" precision="I64">
8418 <dim>1</dim>
8419 </port>
8420 </output>
8421 </layer>
8422 <layer id="500" name="/distilbert/transformer/layer.3/sa_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
8423 <data keep_dims="true" />
8424 <rt_info>
8425 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/ReduceMean" />
8426 </rt_info>
8427 <input>
8428 <port id="0" precision="FP32">
8429 <dim>-1</dim>
8430 <dim>-1</dim>
8431 <dim>768</dim>
8432 </port>
8433 <port id="1" precision="I64">
8434 <dim>1</dim>
8435 </port>
8436 </input>
8437 <output>
8438 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/ReduceMean_output_0">
8439 <dim>-1</dim>
8440 <dim>-1</dim>
8441 <dim>1</dim>
8442 </port>
8443 </output>
8444 </layer>
8445 <layer id="501" name="/distilbert/transformer/layer.3/sa_layer_norm/Sub" type="Subtract" version="opset1">
8446 <data auto_broadcast="numpy" />
8447 <rt_info>
8448 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Sub" />
8449 </rt_info>
8450 <input>
8451 <port id="0" precision="FP32">
8452 <dim>-1</dim>
8453 <dim>-1</dim>
8454 <dim>768</dim>
8455 </port>
8456 <port id="1" precision="FP32">
8457 <dim>-1</dim>
8458 <dim>-1</dim>
8459 <dim>1</dim>
8460 </port>
8461 </input>
8462 <output>
8463 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Sub_output_0">
8464 <dim>-1</dim>
8465 <dim>-1</dim>
8466 <dim>768</dim>
8467 </port>
8468 </output>
8469 </layer>
8470 <layer id="502" name="/distilbert/transformer/layer.3/sa_layer_norm/Constant" type="Const" version="opset1">
8471 <data element_type="f32" shape="" offset="90763296" size="4" />
8472 <rt_info>
8473 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Constant" />
8474 </rt_info>
8475 <output>
8476 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Constant_output_0" />
8477 </output>
8478 </layer>
8479 <layer id="503" name="/distilbert/transformer/layer.3/sa_layer_norm/Pow" type="Power" version="opset1">
8480 <data auto_broadcast="numpy" />
8481 <rt_info>
8482 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Pow" />
8483 </rt_info>
8484 <input>
8485 <port id="0" precision="FP32">
8486 <dim>-1</dim>
8487 <dim>-1</dim>
8488 <dim>768</dim>
8489 </port>
8490 <port id="1" precision="FP32" />
8491 </input>
8492 <output>
8493 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Pow_output_0">
8494 <dim>-1</dim>
8495 <dim>-1</dim>
8496 <dim>768</dim>
8497 </port>
8498 </output>
8499 </layer>
8500 <layer id="504" name="Constant_255779" type="Const" version="opset1">
8501 <data element_type="i64" shape="1" offset="90763288" size="8" />
8502 <rt_info>
8503 <attribute name="fused_names" version="0" value="Constant_255779" />
8504 </rt_info>
8505 <output>
8506 <port id="0" precision="I64">
8507 <dim>1</dim>
8508 </port>
8509 </output>
8510 </layer>
8511 <layer id="505" name="/distilbert/transformer/layer.3/sa_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
8512 <data keep_dims="true" />
8513 <rt_info>
8514 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/ReduceMean_1" />
8515 </rt_info>
8516 <input>
8517 <port id="0" precision="FP32">
8518 <dim>-1</dim>
8519 <dim>-1</dim>
8520 <dim>768</dim>
8521 </port>
8522 <port id="1" precision="I64">
8523 <dim>1</dim>
8524 </port>
8525 </input>
8526 <output>
8527 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/ReduceMean_1_output_0">
8528 <dim>-1</dim>
8529 <dim>-1</dim>
8530 <dim>1</dim>
8531 </port>
8532 </output>
8533 </layer>
8534 <layer id="506" name="/distilbert/transformer/layer.3/sa_layer_norm/Constant_1" type="Const" version="opset1">
8535 <data element_type="f32" shape="" offset="90763300" size="4" />
8536 <rt_info>
8537 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Constant_1" />
8538 </rt_info>
8539 <output>
8540 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Constant_1_output_0" />
8541 </output>
8542 </layer>
8543 <layer id="507" name="/distilbert/transformer/layer.3/sa_layer_norm/Add" type="Add" version="opset1">
8544 <data auto_broadcast="numpy" />
8545 <rt_info>
8546 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Add" />
8547 </rt_info>
8548 <input>
8549 <port id="0" precision="FP32">
8550 <dim>-1</dim>
8551 <dim>-1</dim>
8552 <dim>1</dim>
8553 </port>
8554 <port id="1" precision="FP32" />
8555 </input>
8556 <output>
8557 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Add_output_0">
8558 <dim>-1</dim>
8559 <dim>-1</dim>
8560 <dim>1</dim>
8561 </port>
8562 </output>
8563 </layer>
8564 <layer id="508" name="/distilbert/transformer/layer.3/sa_layer_norm/Sqrt" type="Sqrt" version="opset1">
8565 <rt_info>
8566 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Sqrt" />
8567 </rt_info>
8568 <input>
8569 <port id="0" precision="FP32">
8570 <dim>-1</dim>
8571 <dim>-1</dim>
8572 <dim>1</dim>
8573 </port>
8574 </input>
8575 <output>
8576 <port id="1" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Sqrt_output_0">
8577 <dim>-1</dim>
8578 <dim>-1</dim>
8579 <dim>1</dim>
8580 </port>
8581 </output>
8582 </layer>
8583 <layer id="509" name="/distilbert/transformer/layer.3/sa_layer_norm/Div" type="Divide" version="opset1">
8584 <data auto_broadcast="numpy" m_pythondiv="true" />
8585 <rt_info>
8586 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Div" />
8587 </rt_info>
8588 <input>
8589 <port id="0" precision="FP32">
8590 <dim>-1</dim>
8591 <dim>-1</dim>
8592 <dim>768</dim>
8593 </port>
8594 <port id="1" precision="FP32">
8595 <dim>-1</dim>
8596 <dim>-1</dim>
8597 <dim>1</dim>
8598 </port>
8599 </input>
8600 <output>
8601 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Div_output_0">
8602 <dim>-1</dim>
8603 <dim>-1</dim>
8604 <dim>768</dim>
8605 </port>
8606 </output>
8607 </layer>
8608 <layer id="510" name="distilbert.transformer.layer.3.sa_layer_norm.weight" type="Const" version="opset1">
8609 <data element_type="f32" shape="768" offset="185215124" size="3072" />
8610 <rt_info>
8611 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.sa_layer_norm.weight" />
8612 </rt_info>
8613 <output>
8614 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.sa_layer_norm.weight">
8615 <dim>768</dim>
8616 </port>
8617 </output>
8618 </layer>
8619 <layer id="511" name="/distilbert/transformer/layer.3/sa_layer_norm/Mul" type="Multiply" version="opset1">
8620 <data auto_broadcast="numpy" />
8621 <rt_info>
8622 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Mul" />
8623 </rt_info>
8624 <input>
8625 <port id="0" precision="FP32">
8626 <dim>-1</dim>
8627 <dim>-1</dim>
8628 <dim>768</dim>
8629 </port>
8630 <port id="1" precision="FP32">
8631 <dim>768</dim>
8632 </port>
8633 </input>
8634 <output>
8635 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Mul_output_0">
8636 <dim>-1</dim>
8637 <dim>-1</dim>
8638 <dim>768</dim>
8639 </port>
8640 </output>
8641 </layer>
8642 <layer id="512" name="distilbert.transformer.layer.3.sa_layer_norm.bias" type="Const" version="opset1">
8643 <data element_type="f32" shape="768" offset="185218196" size="3072" />
8644 <rt_info>
8645 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.sa_layer_norm.bias" />
8646 </rt_info>
8647 <output>
8648 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.sa_layer_norm.bias">
8649 <dim>768</dim>
8650 </port>
8651 </output>
8652 </layer>
8653 <layer id="513" name="/distilbert/transformer/layer.3/sa_layer_norm/Add_1" type="Add" version="opset1">
8654 <data auto_broadcast="numpy" />
8655 <rt_info>
8656 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/sa_layer_norm/Add_1" />
8657 </rt_info>
8658 <input>
8659 <port id="0" precision="FP32">
8660 <dim>-1</dim>
8661 <dim>-1</dim>
8662 <dim>768</dim>
8663 </port>
8664 <port id="1" precision="FP32">
8665 <dim>768</dim>
8666 </port>
8667 </input>
8668 <output>
8669 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/sa_layer_norm/Add_1_output_0">
8670 <dim>-1</dim>
8671 <dim>-1</dim>
8672 <dim>768</dim>
8673 </port>
8674 </output>
8675 </layer>
8676 <layer id="514" name="onnx::MatMul_812" type="Const" version="opset1">
8677 <data element_type="f32" shape="768, 3072" offset="185221268" size="9437184" />
8678 <rt_info>
8679 <attribute name="fused_names" version="0" value="onnx::MatMul_812" />
8680 </rt_info>
8681 <output>
8682 <port id="0" precision="FP32" names="onnx::MatMul_812">
8683 <dim>768</dim>
8684 <dim>3072</dim>
8685 </port>
8686 </output>
8687 </layer>
8688 <layer id="515" name="/distilbert/transformer/layer.3/ffn/lin1/MatMul" type="MatMul" version="opset1">
8689 <data transpose_a="false" transpose_b="false" />
8690 <rt_info>
8691 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/lin1/MatMul" />
8692 </rt_info>
8693 <input>
8694 <port id="0" precision="FP32">
8695 <dim>-1</dim>
8696 <dim>-1</dim>
8697 <dim>768</dim>
8698 </port>
8699 <port id="1" precision="FP32">
8700 <dim>768</dim>
8701 <dim>3072</dim>
8702 </port>
8703 </input>
8704 <output>
8705 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/lin1/MatMul_output_0">
8706 <dim>-1</dim>
8707 <dim>-1</dim>
8708 <dim>3072</dim>
8709 </port>
8710 </output>
8711 </layer>
8712 <layer id="516" name="/distilbert/transformer/layer.3/ffn/lin1/Add" type="Add" version="opset1">
8713 <data auto_broadcast="numpy" />
8714 <rt_info>
8715 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/lin1/Add" />
8716 </rt_info>
8717 <input>
8718 <port id="0" precision="FP32">
8719 <dim>3072</dim>
8720 </port>
8721 <port id="1" precision="FP32">
8722 <dim>-1</dim>
8723 <dim>-1</dim>
8724 <dim>3072</dim>
8725 </port>
8726 </input>
8727 <output>
8728 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/lin1/Add_output_0">
8729 <dim>-1</dim>
8730 <dim>-1</dim>
8731 <dim>3072</dim>
8732 </port>
8733 </output>
8734 </layer>
8735 <layer id="517" name="/distilbert/transformer/layer.3/ffn/activation/Constant" type="Const" version="opset1">
8736 <data element_type="f32" shape="" offset="109659272" size="4" />
8737 <rt_info>
8738 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Constant" />
8739 </rt_info>
8740 <output>
8741 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Constant_output_0" />
8742 </output>
8743 </layer>
8744 <layer id="518" name="/distilbert/transformer/layer.3/ffn/activation/Div" type="Divide" version="opset1">
8745 <data auto_broadcast="numpy" m_pythondiv="true" />
8746 <rt_info>
8747 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Div" />
8748 </rt_info>
8749 <input>
8750 <port id="0" precision="FP32">
8751 <dim>-1</dim>
8752 <dim>-1</dim>
8753 <dim>3072</dim>
8754 </port>
8755 <port id="1" precision="FP32" />
8756 </input>
8757 <output>
8758 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Div_output_0">
8759 <dim>-1</dim>
8760 <dim>-1</dim>
8761 <dim>3072</dim>
8762 </port>
8763 </output>
8764 </layer>
8765 <layer id="519" name="/distilbert/transformer/layer.3/ffn/activation/Erf" type="Erf" version="opset1">
8766 <rt_info>
8767 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Erf" />
8768 </rt_info>
8769 <input>
8770 <port id="0" precision="FP32">
8771 <dim>-1</dim>
8772 <dim>-1</dim>
8773 <dim>3072</dim>
8774 </port>
8775 </input>
8776 <output>
8777 <port id="1" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Erf_output_0">
8778 <dim>-1</dim>
8779 <dim>-1</dim>
8780 <dim>3072</dim>
8781 </port>
8782 </output>
8783 </layer>
8784 <layer id="520" name="/distilbert/transformer/layer.3/ffn/activation/Constant_1" type="Const" version="opset1">
8785 <data element_type="f32" shape="" offset="109659276" size="4" />
8786 <rt_info>
8787 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Constant_1" />
8788 </rt_info>
8789 <output>
8790 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Constant_1_output_0" />
8791 </output>
8792 </layer>
8793 <layer id="521" name="/distilbert/transformer/layer.3/ffn/activation/Add" type="Add" version="opset1">
8794 <data auto_broadcast="numpy" />
8795 <rt_info>
8796 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Add" />
8797 </rt_info>
8798 <input>
8799 <port id="0" precision="FP32">
8800 <dim>-1</dim>
8801 <dim>-1</dim>
8802 <dim>3072</dim>
8803 </port>
8804 <port id="1" precision="FP32" />
8805 </input>
8806 <output>
8807 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Add_output_0">
8808 <dim>-1</dim>
8809 <dim>-1</dim>
8810 <dim>3072</dim>
8811 </port>
8812 </output>
8813 </layer>
8814 <layer id="522" name="/distilbert/transformer/layer.3/ffn/activation/Mul" type="Multiply" version="opset1">
8815 <data auto_broadcast="numpy" />
8816 <rt_info>
8817 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Mul" />
8818 </rt_info>
8819 <input>
8820 <port id="0" precision="FP32">
8821 <dim>-1</dim>
8822 <dim>-1</dim>
8823 <dim>3072</dim>
8824 </port>
8825 <port id="1" precision="FP32">
8826 <dim>-1</dim>
8827 <dim>-1</dim>
8828 <dim>3072</dim>
8829 </port>
8830 </input>
8831 <output>
8832 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Mul_output_0">
8833 <dim>-1</dim>
8834 <dim>-1</dim>
8835 <dim>3072</dim>
8836 </port>
8837 </output>
8838 </layer>
8839 <layer id="523" name="/distilbert/transformer/layer.3/ffn/activation/Constant_2" type="Const" version="opset1">
8840 <data element_type="f32" shape="" offset="109659280" size="4" />
8841 <rt_info>
8842 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Constant_2" />
8843 </rt_info>
8844 <output>
8845 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Constant_2_output_0" />
8846 </output>
8847 </layer>
8848 <layer id="524" name="/distilbert/transformer/layer.3/ffn/activation/Mul_1" type="Multiply" version="opset1">
8849 <data auto_broadcast="numpy" />
8850 <rt_info>
8851 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/activation/Mul_1" />
8852 </rt_info>
8853 <input>
8854 <port id="0" precision="FP32">
8855 <dim>-1</dim>
8856 <dim>-1</dim>
8857 <dim>3072</dim>
8858 </port>
8859 <port id="1" precision="FP32" />
8860 </input>
8861 <output>
8862 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/activation/Mul_1_output_0">
8863 <dim>-1</dim>
8864 <dim>-1</dim>
8865 <dim>3072</dim>
8866 </port>
8867 </output>
8868 </layer>
8869 <layer id="525" name="onnx::MatMul_813" type="Const" version="opset1">
8870 <data element_type="f32" shape="3072, 768" offset="194658452" size="9437184" />
8871 <rt_info>
8872 <attribute name="fused_names" version="0" value="onnx::MatMul_813" />
8873 </rt_info>
8874 <output>
8875 <port id="0" precision="FP32" names="onnx::MatMul_813">
8876 <dim>3072</dim>
8877 <dim>768</dim>
8878 </port>
8879 </output>
8880 </layer>
8881 <layer id="526" name="/distilbert/transformer/layer.3/ffn/lin2/MatMul" type="MatMul" version="opset1">
8882 <data transpose_a="false" transpose_b="false" />
8883 <rt_info>
8884 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/lin2/MatMul" />
8885 </rt_info>
8886 <input>
8887 <port id="0" precision="FP32">
8888 <dim>-1</dim>
8889 <dim>-1</dim>
8890 <dim>3072</dim>
8891 </port>
8892 <port id="1" precision="FP32">
8893 <dim>3072</dim>
8894 <dim>768</dim>
8895 </port>
8896 </input>
8897 <output>
8898 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/lin2/MatMul_output_0">
8899 <dim>-1</dim>
8900 <dim>-1</dim>
8901 <dim>768</dim>
8902 </port>
8903 </output>
8904 </layer>
8905 <layer id="527" name="/distilbert/transformer/layer.3/ffn/lin2/Add" type="Add" version="opset1">
8906 <data auto_broadcast="numpy" />
8907 <rt_info>
8908 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/ffn/lin2/Add" />
8909 </rt_info>
8910 <input>
8911 <port id="0" precision="FP32">
8912 <dim>768</dim>
8913 </port>
8914 <port id="1" precision="FP32">
8915 <dim>-1</dim>
8916 <dim>-1</dim>
8917 <dim>768</dim>
8918 </port>
8919 </input>
8920 <output>
8921 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/ffn/lin2/Add_output_0">
8922 <dim>-1</dim>
8923 <dim>-1</dim>
8924 <dim>768</dim>
8925 </port>
8926 </output>
8927 </layer>
8928 <layer id="528" name="/distilbert/transformer/layer.3/Add_1" type="Add" version="opset1">
8929 <data auto_broadcast="numpy" />
8930 <rt_info>
8931 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/Add_1" />
8932 </rt_info>
8933 <input>
8934 <port id="0" precision="FP32">
8935 <dim>-1</dim>
8936 <dim>-1</dim>
8937 <dim>768</dim>
8938 </port>
8939 <port id="1" precision="FP32">
8940 <dim>-1</dim>
8941 <dim>-1</dim>
8942 <dim>768</dim>
8943 </port>
8944 </input>
8945 <output>
8946 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/Add_1_output_0">
8947 <dim>-1</dim>
8948 <dim>-1</dim>
8949 <dim>768</dim>
8950 </port>
8951 </output>
8952 </layer>
8953 <layer id="529" name="Constant_255800" type="Const" version="opset1">
8954 <data element_type="i64" shape="1" offset="90763288" size="8" />
8955 <rt_info>
8956 <attribute name="fused_names" version="0" value="Constant_255800" />
8957 </rt_info>
8958 <output>
8959 <port id="0" precision="I64">
8960 <dim>1</dim>
8961 </port>
8962 </output>
8963 </layer>
8964 <layer id="530" name="/distilbert/transformer/layer.3/output_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
8965 <data keep_dims="true" />
8966 <rt_info>
8967 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/ReduceMean" />
8968 </rt_info>
8969 <input>
8970 <port id="0" precision="FP32">
8971 <dim>-1</dim>
8972 <dim>-1</dim>
8973 <dim>768</dim>
8974 </port>
8975 <port id="1" precision="I64">
8976 <dim>1</dim>
8977 </port>
8978 </input>
8979 <output>
8980 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/ReduceMean_output_0">
8981 <dim>-1</dim>
8982 <dim>-1</dim>
8983 <dim>1</dim>
8984 </port>
8985 </output>
8986 </layer>
8987 <layer id="531" name="/distilbert/transformer/layer.3/output_layer_norm/Sub" type="Subtract" version="opset1">
8988 <data auto_broadcast="numpy" />
8989 <rt_info>
8990 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Sub" />
8991 </rt_info>
8992 <input>
8993 <port id="0" precision="FP32">
8994 <dim>-1</dim>
8995 <dim>-1</dim>
8996 <dim>768</dim>
8997 </port>
8998 <port id="1" precision="FP32">
8999 <dim>-1</dim>
9000 <dim>-1</dim>
9001 <dim>1</dim>
9002 </port>
9003 </input>
9004 <output>
9005 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Sub_output_0">
9006 <dim>-1</dim>
9007 <dim>-1</dim>
9008 <dim>768</dim>
9009 </port>
9010 </output>
9011 </layer>
9012 <layer id="532" name="/distilbert/transformer/layer.3/output_layer_norm/Constant" type="Const" version="opset1">
9013 <data element_type="f32" shape="" offset="90763296" size="4" />
9014 <rt_info>
9015 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Constant" />
9016 </rt_info>
9017 <output>
9018 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Constant_output_0" />
9019 </output>
9020 </layer>
9021 <layer id="533" name="/distilbert/transformer/layer.3/output_layer_norm/Pow" type="Power" version="opset1">
9022 <data auto_broadcast="numpy" />
9023 <rt_info>
9024 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Pow" />
9025 </rt_info>
9026 <input>
9027 <port id="0" precision="FP32">
9028 <dim>-1</dim>
9029 <dim>-1</dim>
9030 <dim>768</dim>
9031 </port>
9032 <port id="1" precision="FP32" />
9033 </input>
9034 <output>
9035 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Pow_output_0">
9036 <dim>-1</dim>
9037 <dim>-1</dim>
9038 <dim>768</dim>
9039 </port>
9040 </output>
9041 </layer>
9042 <layer id="534" name="Constant_255805" type="Const" version="opset1">
9043 <data element_type="i64" shape="1" offset="90763288" size="8" />
9044 <rt_info>
9045 <attribute name="fused_names" version="0" value="Constant_255805" />
9046 </rt_info>
9047 <output>
9048 <port id="0" precision="I64">
9049 <dim>1</dim>
9050 </port>
9051 </output>
9052 </layer>
9053 <layer id="535" name="/distilbert/transformer/layer.3/output_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
9054 <data keep_dims="true" />
9055 <rt_info>
9056 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/ReduceMean_1" />
9057 </rt_info>
9058 <input>
9059 <port id="0" precision="FP32">
9060 <dim>-1</dim>
9061 <dim>-1</dim>
9062 <dim>768</dim>
9063 </port>
9064 <port id="1" precision="I64">
9065 <dim>1</dim>
9066 </port>
9067 </input>
9068 <output>
9069 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/ReduceMean_1_output_0">
9070 <dim>-1</dim>
9071 <dim>-1</dim>
9072 <dim>1</dim>
9073 </port>
9074 </output>
9075 </layer>
9076 <layer id="536" name="/distilbert/transformer/layer.3/output_layer_norm/Constant_1" type="Const" version="opset1">
9077 <data element_type="f32" shape="" offset="90763300" size="4" />
9078 <rt_info>
9079 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Constant_1" />
9080 </rt_info>
9081 <output>
9082 <port id="0" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Constant_1_output_0" />
9083 </output>
9084 </layer>
9085 <layer id="537" name="/distilbert/transformer/layer.3/output_layer_norm/Add" type="Add" version="opset1">
9086 <data auto_broadcast="numpy" />
9087 <rt_info>
9088 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Add" />
9089 </rt_info>
9090 <input>
9091 <port id="0" precision="FP32">
9092 <dim>-1</dim>
9093 <dim>-1</dim>
9094 <dim>1</dim>
9095 </port>
9096 <port id="1" precision="FP32" />
9097 </input>
9098 <output>
9099 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Add_output_0">
9100 <dim>-1</dim>
9101 <dim>-1</dim>
9102 <dim>1</dim>
9103 </port>
9104 </output>
9105 </layer>
9106 <layer id="538" name="/distilbert/transformer/layer.3/output_layer_norm/Sqrt" type="Sqrt" version="opset1">
9107 <rt_info>
9108 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Sqrt" />
9109 </rt_info>
9110 <input>
9111 <port id="0" precision="FP32">
9112 <dim>-1</dim>
9113 <dim>-1</dim>
9114 <dim>1</dim>
9115 </port>
9116 </input>
9117 <output>
9118 <port id="1" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Sqrt_output_0">
9119 <dim>-1</dim>
9120 <dim>-1</dim>
9121 <dim>1</dim>
9122 </port>
9123 </output>
9124 </layer>
9125 <layer id="539" name="/distilbert/transformer/layer.3/output_layer_norm/Div" type="Divide" version="opset1">
9126 <data auto_broadcast="numpy" m_pythondiv="true" />
9127 <rt_info>
9128 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Div" />
9129 </rt_info>
9130 <input>
9131 <port id="0" precision="FP32">
9132 <dim>-1</dim>
9133 <dim>-1</dim>
9134 <dim>768</dim>
9135 </port>
9136 <port id="1" precision="FP32">
9137 <dim>-1</dim>
9138 <dim>-1</dim>
9139 <dim>1</dim>
9140 </port>
9141 </input>
9142 <output>
9143 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Div_output_0">
9144 <dim>-1</dim>
9145 <dim>-1</dim>
9146 <dim>768</dim>
9147 </port>
9148 </output>
9149 </layer>
9150 <layer id="540" name="distilbert.transformer.layer.3.output_layer_norm.weight" type="Const" version="opset1">
9151 <data element_type="f32" shape="768" offset="204095636" size="3072" />
9152 <rt_info>
9153 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.output_layer_norm.weight" />
9154 </rt_info>
9155 <output>
9156 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.output_layer_norm.weight">
9157 <dim>768</dim>
9158 </port>
9159 </output>
9160 </layer>
9161 <layer id="541" name="/distilbert/transformer/layer.3/output_layer_norm/Mul" type="Multiply" version="opset1">
9162 <data auto_broadcast="numpy" />
9163 <rt_info>
9164 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Mul" />
9165 </rt_info>
9166 <input>
9167 <port id="0" precision="FP32">
9168 <dim>-1</dim>
9169 <dim>-1</dim>
9170 <dim>768</dim>
9171 </port>
9172 <port id="1" precision="FP32">
9173 <dim>768</dim>
9174 </port>
9175 </input>
9176 <output>
9177 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Mul_output_0">
9178 <dim>-1</dim>
9179 <dim>-1</dim>
9180 <dim>768</dim>
9181 </port>
9182 </output>
9183 </layer>
9184 <layer id="542" name="distilbert.transformer.layer.3.output_layer_norm.bias" type="Const" version="opset1">
9185 <data element_type="f32" shape="768" offset="204098708" size="3072" />
9186 <rt_info>
9187 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.3.output_layer_norm.bias" />
9188 </rt_info>
9189 <output>
9190 <port id="0" precision="FP32" names="distilbert.transformer.layer.3.output_layer_norm.bias">
9191 <dim>768</dim>
9192 </port>
9193 </output>
9194 </layer>
9195 <layer id="543" name="/distilbert/transformer/layer.3/output_layer_norm/Add_1" type="Add" version="opset1">
9196 <data auto_broadcast="numpy" />
9197 <rt_info>
9198 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.3/output_layer_norm/Add_1" />
9199 </rt_info>
9200 <input>
9201 <port id="0" precision="FP32">
9202 <dim>-1</dim>
9203 <dim>-1</dim>
9204 <dim>768</dim>
9205 </port>
9206 <port id="1" precision="FP32">
9207 <dim>768</dim>
9208 </port>
9209 </input>
9210 <output>
9211 <port id="2" precision="FP32" names="/distilbert/transformer/layer.3/output_layer_norm/Add_1_output_0">
9212 <dim>-1</dim>
9213 <dim>-1</dim>
9214 <dim>768</dim>
9215 </port>
9216 </output>
9217 </layer>
9218 <layer id="544" name="/distilbert/transformer/layer.4/attention/Shape" type="ShapeOf" version="opset3">
9219 <data output_type="i64" />
9220 <rt_info>
9221 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Shape" />
9222 </rt_info>
9223 <input>
9224 <port id="0" precision="FP32">
9225 <dim>-1</dim>
9226 <dim>-1</dim>
9227 <dim>768</dim>
9228 </port>
9229 </input>
9230 <output>
9231 <port id="1" precision="I64" names="/distilbert/transformer/layer.4/attention/Shape_output_0">
9232 <dim>3</dim>
9233 </port>
9234 </output>
9235 </layer>
9236 <layer id="545" name="/distilbert/transformer/layer.4/attention/Constant" type="Const" version="opset1">
9237 <data element_type="i64" shape="" offset="18440" size="8" />
9238 <rt_info>
9239 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant" />
9240 </rt_info>
9241 <output>
9242 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_output_0" />
9243 </output>
9244 </layer>
9245 <layer id="546" name="Constant_255815" type="Const" version="opset1">
9246 <data element_type="i64" shape="" offset="18440" size="8" />
9247 <rt_info>
9248 <attribute name="fused_names" version="0" value="Constant_255815" />
9249 </rt_info>
9250 <output>
9251 <port id="0" precision="I64" />
9252 </output>
9253 </layer>
9254 <layer id="547" name="/distilbert/transformer/layer.4/attention/Gather" type="Gather" version="opset8">
9255 <data batch_dims="0" />
9256 <rt_info>
9257 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant, /distilbert/transformer/layer.4/attention/Gather, Constant_255815" />
9258 </rt_info>
9259 <input>
9260 <port id="0" precision="I64">
9261 <dim>3</dim>
9262 </port>
9263 <port id="1" precision="I64" />
9264 <port id="2" precision="I64" />
9265 </input>
9266 <output>
9267 <port id="3" precision="I64" names="/distilbert/transformer/layer.4/attention/Gather_output_0" />
9268 </output>
9269 </layer>
9270 <layer id="548" name="Constant_255875" type="Const" version="opset1">
9271 <data element_type="i64" shape="1" offset="18440" size="8" />
9272 <rt_info>
9273 <attribute name="fused_names" version="0" value="Constant_255875" />
9274 </rt_info>
9275 <output>
9276 <port id="0" precision="I64">
9277 <dim>1</dim>
9278 </port>
9279 </output>
9280 </layer>
9281 <layer id="549" name="/distilbert/transformer/layer.4/attention/Unsqueeze_3" type="Unsqueeze" version="opset1">
9282 <rt_info>
9283 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Unsqueeze_3, Constant_255875" />
9284 </rt_info>
9285 <input>
9286 <port id="0" precision="I64" />
9287 <port id="1" precision="I64">
9288 <dim>1</dim>
9289 </port>
9290 </input>
9291 <output>
9292 <port id="2" precision="I64" names="/distilbert/transformer/layer.4/attention/Unsqueeze_3_output_0">
9293 <dim>1</dim>
9294 </port>
9295 </output>
9296 </layer>
9297 <layer id="550" name="/distilbert/transformer/layer.4/attention/Constant_12" type="Const" version="opset1">
9298 <data element_type="i64" shape="1" offset="90763280" size="8" />
9299 <rt_info>
9300 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_12" />
9301 </rt_info>
9302 <output>
9303 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_12_output_0">
9304 <dim>1</dim>
9305 </port>
9306 </output>
9307 </layer>
9308 <layer id="551" name="/distilbert/transformer/layer.4/attention/Constant_13" type="Const" version="opset1">
9309 <data element_type="i64" shape="1" offset="90763280" size="8" />
9310 <rt_info>
9311 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_13" />
9312 </rt_info>
9313 <output>
9314 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_13_output_0">
9315 <dim>1</dim>
9316 </port>
9317 </output>
9318 </layer>
9319 <layer id="552" name="/distilbert/transformer/layer.4/attention/Shape_1" type="ShapeOf" version="opset3">
9320 <data output_type="i64" />
9321 <rt_info>
9322 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Shape_1" />
9323 </rt_info>
9324 <input>
9325 <port id="0" precision="FP32">
9326 <dim>-1</dim>
9327 <dim>-1</dim>
9328 <dim>768</dim>
9329 </port>
9330 </input>
9331 <output>
9332 <port id="1" precision="I64" names="/distilbert/transformer/layer.4/attention/Shape_1_output_0">
9333 <dim>3</dim>
9334 </port>
9335 </output>
9336 </layer>
9337 <layer id="553" name="/distilbert/transformer/layer.4/attention/Constant_1" type="Const" version="opset1">
9338 <data element_type="i64" shape="" offset="90763280" size="8" />
9339 <rt_info>
9340 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_1" />
9341 </rt_info>
9342 <output>
9343 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_1_output_0" />
9344 </output>
9345 </layer>
9346 <layer id="554" name="Constant_255819" type="Const" version="opset1">
9347 <data element_type="i64" shape="" offset="18440" size="8" />
9348 <rt_info>
9349 <attribute name="fused_names" version="0" value="Constant_255819" />
9350 </rt_info>
9351 <output>
9352 <port id="0" precision="I64" />
9353 </output>
9354 </layer>
9355 <layer id="555" name="/distilbert/transformer/layer.4/attention/Gather_1" type="Gather" version="opset8">
9356 <data batch_dims="0" />
9357 <rt_info>
9358 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_1, /distilbert/transformer/layer.4/attention/Gather_1, Constant_255819" />
9359 </rt_info>
9360 <input>
9361 <port id="0" precision="I64">
9362 <dim>3</dim>
9363 </port>
9364 <port id="1" precision="I64" />
9365 <port id="2" precision="I64" />
9366 </input>
9367 <output>
9368 <port id="3" precision="I64" names="/distilbert/transformer/layer.4/attention/Gather_1_output_0" />
9369 </output>
9370 </layer>
9371 <layer id="556" name="Constant_255879" type="Const" version="opset1">
9372 <data element_type="i64" shape="1" offset="18440" size="8" />
9373 <rt_info>
9374 <attribute name="fused_names" version="0" value="Constant_255879" />
9375 </rt_info>
9376 <output>
9377 <port id="0" precision="I64">
9378 <dim>1</dim>
9379 </port>
9380 </output>
9381 </layer>
9382 <layer id="557" name="/distilbert/transformer/layer.4/attention/Unsqueeze_4" type="Unsqueeze" version="opset1">
9383 <rt_info>
9384 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Unsqueeze_4, Constant_255879" />
9385 </rt_info>
9386 <input>
9387 <port id="0" precision="I64" />
9388 <port id="1" precision="I64">
9389 <dim>1</dim>
9390 </port>
9391 </input>
9392 <output>
9393 <port id="2" precision="I64" names="/distilbert/transformer/layer.4/attention/Unsqueeze_4_output_0">
9394 <dim>1</dim>
9395 </port>
9396 </output>
9397 </layer>
9398 <layer id="558" name="/distilbert/transformer/layer.4/attention/Concat_3" type="Concat" version="opset1">
9399 <data axis="0" />
9400 <rt_info>
9401 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Concat_3, /distilbert/transformer/layer.4/attention/Constant_12, /distilbert/transformer/layer.4/attention/Constant_13" />
9402 </rt_info>
9403 <input>
9404 <port id="0" precision="I64">
9405 <dim>1</dim>
9406 </port>
9407 <port id="1" precision="I64">
9408 <dim>1</dim>
9409 </port>
9410 <port id="2" precision="I64">
9411 <dim>1</dim>
9412 </port>
9413 <port id="3" precision="I64">
9414 <dim>1</dim>
9415 </port>
9416 </input>
9417 <output>
9418 <port id="4" precision="I64" names="/distilbert/transformer/layer.4/attention/Concat_3_output_0">
9419 <dim>4</dim>
9420 </port>
9421 </output>
9422 </layer>
9423 <layer id="559" name="/distilbert/transformer/layer.4/attention/Reshape_3" type="Reshape" version="opset1">
9424 <data special_zero="true" />
9425 <rt_info>
9426 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Reshape_3" />
9427 </rt_info>
9428 <input>
9429 <port id="0" precision="BOOL">
9430 <dim>-1</dim>
9431 <dim>-1</dim>
9432 </port>
9433 <port id="1" precision="I64">
9434 <dim>4</dim>
9435 </port>
9436 </input>
9437 <output>
9438 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.4/attention/Reshape_3_output_0">
9439 <dim>-1</dim>
9440 <dim>1</dim>
9441 <dim>1</dim>
9442 <dim>-1</dim>
9443 </port>
9444 </output>
9445 </layer>
9446 <layer id="560" name="distilbert.transformer.layer.4.attention.q_lin.bias" type="Const" version="opset1">
9447 <data element_type="f32" shape="768" offset="204101780" size="3072" />
9448 <rt_info>
9449 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.attention.q_lin.bias" />
9450 </rt_info>
9451 <output>
9452 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.attention.q_lin.bias">
9453 <dim>768</dim>
9454 </port>
9455 </output>
9456 </layer>
9457 <layer id="561" name="onnx::MatMul_814" type="Const" version="opset1">
9458 <data element_type="f32" shape="768, 768" offset="204104852" size="2359296" />
9459 <rt_info>
9460 <attribute name="fused_names" version="0" value="onnx::MatMul_814" />
9461 </rt_info>
9462 <output>
9463 <port id="0" precision="FP32" names="onnx::MatMul_814">
9464 <dim>768</dim>
9465 <dim>768</dim>
9466 </port>
9467 </output>
9468 </layer>
9469 <layer id="562" name="/distilbert/transformer/layer.4/attention/q_lin/MatMul" type="MatMul" version="opset1">
9470 <data transpose_a="false" transpose_b="false" />
9471 <rt_info>
9472 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/q_lin/MatMul" />
9473 </rt_info>
9474 <input>
9475 <port id="0" precision="FP32">
9476 <dim>-1</dim>
9477 <dim>-1</dim>
9478 <dim>768</dim>
9479 </port>
9480 <port id="1" precision="FP32">
9481 <dim>768</dim>
9482 <dim>768</dim>
9483 </port>
9484 </input>
9485 <output>
9486 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/q_lin/MatMul_output_0">
9487 <dim>-1</dim>
9488 <dim>-1</dim>
9489 <dim>768</dim>
9490 </port>
9491 </output>
9492 </layer>
9493 <layer id="563" name="/distilbert/transformer/layer.4/attention/q_lin/Add" type="Add" version="opset1">
9494 <data auto_broadcast="numpy" />
9495 <rt_info>
9496 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/q_lin/Add" />
9497 </rt_info>
9498 <input>
9499 <port id="0" precision="FP32">
9500 <dim>768</dim>
9501 </port>
9502 <port id="1" precision="FP32">
9503 <dim>-1</dim>
9504 <dim>-1</dim>
9505 <dim>768</dim>
9506 </port>
9507 </input>
9508 <output>
9509 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/q_lin/Add_output_0">
9510 <dim>-1</dim>
9511 <dim>-1</dim>
9512 <dim>768</dim>
9513 </port>
9514 </output>
9515 </layer>
9516 <layer id="564" name="Constant_255823" type="Const" version="opset1">
9517 <data element_type="i64" shape="1" offset="18440" size="8" />
9518 <rt_info>
9519 <attribute name="fused_names" version="0" value="Constant_255823" />
9520 </rt_info>
9521 <output>
9522 <port id="0" precision="I64">
9523 <dim>1</dim>
9524 </port>
9525 </output>
9526 </layer>
9527 <layer id="565" name="/distilbert/transformer/layer.4/attention/Unsqueeze" type="Unsqueeze" version="opset1">
9528 <rt_info>
9529 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Unsqueeze, Constant_255823" />
9530 </rt_info>
9531 <input>
9532 <port id="0" precision="I64" />
9533 <port id="1" precision="I64">
9534 <dim>1</dim>
9535 </port>
9536 </input>
9537 <output>
9538 <port id="2" precision="I64" names="/distilbert/transformer/layer.4/attention/Unsqueeze_output_0">
9539 <dim>1</dim>
9540 </port>
9541 </output>
9542 </layer>
9543 <layer id="566" name="/distilbert/transformer/layer.4/attention/Constant_2" type="Const" version="opset1">
9544 <data element_type="i64" shape="1" offset="90763288" size="8" />
9545 <rt_info>
9546 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_2" />
9547 </rt_info>
9548 <output>
9549 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_2_output_0">
9550 <dim>1</dim>
9551 </port>
9552 </output>
9553 </layer>
9554 <layer id="567" name="/distilbert/transformer/layer.4/attention/Constant_3" type="Const" version="opset1">
9555 <data element_type="i64" shape="1" offset="93131816" size="8" />
9556 <rt_info>
9557 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_3" />
9558 </rt_info>
9559 <output>
9560 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_3_output_0">
9561 <dim>1</dim>
9562 </port>
9563 </output>
9564 </layer>
9565 <layer id="568" name="/distilbert/transformer/layer.4/attention/Constant_4" type="Const" version="opset1">
9566 <data element_type="i64" shape="1" offset="93131824" size="8" />
9567 <rt_info>
9568 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_4" />
9569 </rt_info>
9570 <output>
9571 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_4_output_0">
9572 <dim>1</dim>
9573 </port>
9574 </output>
9575 </layer>
9576 <layer id="569" name="/distilbert/transformer/layer.4/attention/Concat" type="Concat" version="opset1">
9577 <data axis="0" />
9578 <rt_info>
9579 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Concat, /distilbert/transformer/layer.4/attention/Constant_2, /distilbert/transformer/layer.4/attention/Constant_3, /distilbert/transformer/layer.4/attention/Constant_4" />
9580 </rt_info>
9581 <input>
9582 <port id="0" precision="I64">
9583 <dim>1</dim>
9584 </port>
9585 <port id="1" precision="I64">
9586 <dim>1</dim>
9587 </port>
9588 <port id="2" precision="I64">
9589 <dim>1</dim>
9590 </port>
9591 <port id="3" precision="I64">
9592 <dim>1</dim>
9593 </port>
9594 </input>
9595 <output>
9596 <port id="4" precision="I64" names="/distilbert/transformer/layer.4/attention/Concat_output_0">
9597 <dim>4</dim>
9598 </port>
9599 </output>
9600 </layer>
9601 <layer id="570" name="/distilbert/transformer/layer.4/attention/Reshape" type="Reshape" version="opset1">
9602 <data special_zero="true" />
9603 <rt_info>
9604 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Reshape" />
9605 </rt_info>
9606 <input>
9607 <port id="0" precision="FP32">
9608 <dim>-1</dim>
9609 <dim>-1</dim>
9610 <dim>768</dim>
9611 </port>
9612 <port id="1" precision="I64">
9613 <dim>4</dim>
9614 </port>
9615 </input>
9616 <output>
9617 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Reshape_output_0">
9618 <dim>-1</dim>
9619 <dim>-1</dim>
9620 <dim>12</dim>
9621 <dim>64</dim>
9622 </port>
9623 </output>
9624 </layer>
9625 <layer id="571" name="Constant_255848" type="Const" version="opset1">
9626 <data element_type="i64" shape="4" offset="93131832" size="32" />
9627 <rt_info>
9628 <attribute name="fused_names" version="0" value="Constant_255848" />
9629 </rt_info>
9630 <output>
9631 <port id="0" precision="I64">
9632 <dim>4</dim>
9633 </port>
9634 </output>
9635 </layer>
9636 <layer id="572" name="/distilbert/transformer/layer.4/attention/Transpose" type="Transpose" version="opset1">
9637 <rt_info>
9638 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Transpose" />
9639 </rt_info>
9640 <input>
9641 <port id="0" precision="FP32">
9642 <dim>-1</dim>
9643 <dim>-1</dim>
9644 <dim>12</dim>
9645 <dim>64</dim>
9646 </port>
9647 <port id="1" precision="I64">
9648 <dim>4</dim>
9649 </port>
9650 </input>
9651 <output>
9652 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Transpose_output_0">
9653 <dim>-1</dim>
9654 <dim>12</dim>
9655 <dim>-1</dim>
9656 <dim>64</dim>
9657 </port>
9658 </output>
9659 </layer>
9660 <layer id="573" name="/distilbert/transformer/layer.4/attention/Constant_11" type="Const" version="opset1">
9661 <data element_type="f32" shape="" offset="93131864" size="4" />
9662 <rt_info>
9663 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_11" />
9664 </rt_info>
9665 <output>
9666 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/attention/Constant_11_output_0" />
9667 </output>
9668 </layer>
9669 <layer id="574" name="/distilbert/transformer/layer.4/attention/Div" type="Divide" version="opset1">
9670 <data auto_broadcast="numpy" m_pythondiv="true" />
9671 <rt_info>
9672 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Div" />
9673 </rt_info>
9674 <input>
9675 <port id="0" precision="FP32">
9676 <dim>-1</dim>
9677 <dim>12</dim>
9678 <dim>-1</dim>
9679 <dim>64</dim>
9680 </port>
9681 <port id="1" precision="FP32" />
9682 </input>
9683 <output>
9684 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Div_output_0">
9685 <dim>-1</dim>
9686 <dim>12</dim>
9687 <dim>-1</dim>
9688 <dim>64</dim>
9689 </port>
9690 </output>
9691 </layer>
9692 <layer id="575" name="distilbert.transformer.layer.4.attention.k_lin.bias" type="Const" version="opset1">
9693 <data element_type="f32" shape="768" offset="206464148" size="3072" />
9694 <rt_info>
9695 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.attention.k_lin.bias" />
9696 </rt_info>
9697 <output>
9698 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.attention.k_lin.bias">
9699 <dim>768</dim>
9700 </port>
9701 </output>
9702 </layer>
9703 <layer id="576" name="onnx::MatMul_824" type="Const" version="opset1">
9704 <data element_type="f32" shape="768, 768" offset="206467220" size="2359296" />
9705 <rt_info>
9706 <attribute name="fused_names" version="0" value="onnx::MatMul_824" />
9707 </rt_info>
9708 <output>
9709 <port id="0" precision="FP32" names="onnx::MatMul_824">
9710 <dim>768</dim>
9711 <dim>768</dim>
9712 </port>
9713 </output>
9714 </layer>
9715 <layer id="577" name="/distilbert/transformer/layer.4/attention/k_lin/MatMul" type="MatMul" version="opset1">
9716 <data transpose_a="false" transpose_b="false" />
9717 <rt_info>
9718 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/k_lin/MatMul" />
9719 </rt_info>
9720 <input>
9721 <port id="0" precision="FP32">
9722 <dim>-1</dim>
9723 <dim>-1</dim>
9724 <dim>768</dim>
9725 </port>
9726 <port id="1" precision="FP32">
9727 <dim>768</dim>
9728 <dim>768</dim>
9729 </port>
9730 </input>
9731 <output>
9732 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/k_lin/MatMul_output_0">
9733 <dim>-1</dim>
9734 <dim>-1</dim>
9735 <dim>768</dim>
9736 </port>
9737 </output>
9738 </layer>
9739 <layer id="578" name="/distilbert/transformer/layer.4/attention/k_lin/Add" type="Add" version="opset1">
9740 <data auto_broadcast="numpy" />
9741 <rt_info>
9742 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/k_lin/Add" />
9743 </rt_info>
9744 <input>
9745 <port id="0" precision="FP32">
9746 <dim>768</dim>
9747 </port>
9748 <port id="1" precision="FP32">
9749 <dim>-1</dim>
9750 <dim>-1</dim>
9751 <dim>768</dim>
9752 </port>
9753 </input>
9754 <output>
9755 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/k_lin/Add_output_0">
9756 <dim>-1</dim>
9757 <dim>-1</dim>
9758 <dim>768</dim>
9759 </port>
9760 </output>
9761 </layer>
9762 <layer id="579" name="Constant_255829" type="Const" version="opset1">
9763 <data element_type="i64" shape="1" offset="18440" size="8" />
9764 <rt_info>
9765 <attribute name="fused_names" version="0" value="Constant_255829" />
9766 </rt_info>
9767 <output>
9768 <port id="0" precision="I64">
9769 <dim>1</dim>
9770 </port>
9771 </output>
9772 </layer>
9773 <layer id="580" name="/distilbert/transformer/layer.4/attention/Unsqueeze_1" type="Unsqueeze" version="opset1">
9774 <rt_info>
9775 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Unsqueeze_1, Constant_255829" />
9776 </rt_info>
9777 <input>
9778 <port id="0" precision="I64" />
9779 <port id="1" precision="I64">
9780 <dim>1</dim>
9781 </port>
9782 </input>
9783 <output>
9784 <port id="2" precision="I64" names="/distilbert/transformer/layer.4/attention/Unsqueeze_1_output_0">
9785 <dim>1</dim>
9786 </port>
9787 </output>
9788 </layer>
9789 <layer id="581" name="/distilbert/transformer/layer.4/attention/Constant_5" type="Const" version="opset1">
9790 <data element_type="i64" shape="1" offset="90763288" size="8" />
9791 <rt_info>
9792 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_5" />
9793 </rt_info>
9794 <output>
9795 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_5_output_0">
9796 <dim>1</dim>
9797 </port>
9798 </output>
9799 </layer>
9800 <layer id="582" name="/distilbert/transformer/layer.4/attention/Constant_6" type="Const" version="opset1">
9801 <data element_type="i64" shape="1" offset="93131816" size="8" />
9802 <rt_info>
9803 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_6" />
9804 </rt_info>
9805 <output>
9806 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_6_output_0">
9807 <dim>1</dim>
9808 </port>
9809 </output>
9810 </layer>
9811 <layer id="583" name="/distilbert/transformer/layer.4/attention/Constant_7" type="Const" version="opset1">
9812 <data element_type="i64" shape="1" offset="93131824" size="8" />
9813 <rt_info>
9814 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_7" />
9815 </rt_info>
9816 <output>
9817 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_7_output_0">
9818 <dim>1</dim>
9819 </port>
9820 </output>
9821 </layer>
9822 <layer id="584" name="/distilbert/transformer/layer.4/attention/Concat_1" type="Concat" version="opset1">
9823 <data axis="0" />
9824 <rt_info>
9825 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Concat_1, /distilbert/transformer/layer.4/attention/Constant_5, /distilbert/transformer/layer.4/attention/Constant_6, /distilbert/transformer/layer.4/attention/Constant_7" />
9826 </rt_info>
9827 <input>
9828 <port id="0" precision="I64">
9829 <dim>1</dim>
9830 </port>
9831 <port id="1" precision="I64">
9832 <dim>1</dim>
9833 </port>
9834 <port id="2" precision="I64">
9835 <dim>1</dim>
9836 </port>
9837 <port id="3" precision="I64">
9838 <dim>1</dim>
9839 </port>
9840 </input>
9841 <output>
9842 <port id="4" precision="I64" names="/distilbert/transformer/layer.4/attention/Concat_1_output_0">
9843 <dim>4</dim>
9844 </port>
9845 </output>
9846 </layer>
9847 <layer id="585" name="/distilbert/transformer/layer.4/attention/Reshape_1" type="Reshape" version="opset1">
9848 <data special_zero="true" />
9849 <rt_info>
9850 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Reshape_1" />
9851 </rt_info>
9852 <input>
9853 <port id="0" precision="FP32">
9854 <dim>-1</dim>
9855 <dim>-1</dim>
9856 <dim>768</dim>
9857 </port>
9858 <port id="1" precision="I64">
9859 <dim>4</dim>
9860 </port>
9861 </input>
9862 <output>
9863 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Reshape_1_output_0">
9864 <dim>-1</dim>
9865 <dim>-1</dim>
9866 <dim>12</dim>
9867 <dim>64</dim>
9868 </port>
9869 </output>
9870 </layer>
9871 <layer id="586" name="Constant_255872" type="Const" version="opset1">
9872 <data element_type="i64" shape="4" offset="95494236" size="32" />
9873 <rt_info>
9874 <attribute name="fused_names" version="0" value="Constant_255872" />
9875 </rt_info>
9876 <output>
9877 <port id="0" precision="I64">
9878 <dim>4</dim>
9879 </port>
9880 </output>
9881 </layer>
9882 <layer id="587" name="/distilbert/transformer/layer.4/attention/Transpose_2" type="Transpose" version="opset1">
9883 <rt_info>
9884 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Transpose_2" />
9885 </rt_info>
9886 <input>
9887 <port id="0" precision="FP32">
9888 <dim>-1</dim>
9889 <dim>-1</dim>
9890 <dim>12</dim>
9891 <dim>64</dim>
9892 </port>
9893 <port id="1" precision="I64">
9894 <dim>4</dim>
9895 </port>
9896 </input>
9897 <output>
9898 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Transpose_2_output_0">
9899 <dim>-1</dim>
9900 <dim>12</dim>
9901 <dim>64</dim>
9902 <dim>-1</dim>
9903 </port>
9904 </output>
9905 </layer>
9906 <layer id="588" name="/distilbert/transformer/layer.4/attention/MatMul" type="MatMul" version="opset1">
9907 <data transpose_a="false" transpose_b="false" />
9908 <rt_info>
9909 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/MatMul" />
9910 </rt_info>
9911 <input>
9912 <port id="0" precision="FP32">
9913 <dim>-1</dim>
9914 <dim>12</dim>
9915 <dim>-1</dim>
9916 <dim>64</dim>
9917 </port>
9918 <port id="1" precision="FP32">
9919 <dim>-1</dim>
9920 <dim>12</dim>
9921 <dim>64</dim>
9922 <dim>-1</dim>
9923 </port>
9924 </input>
9925 <output>
9926 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/MatMul_output_0">
9927 <dim>-1</dim>
9928 <dim>12</dim>
9929 <dim>-1</dim>
9930 <dim>-1</dim>
9931 </port>
9932 </output>
9933 </layer>
9934 <layer id="589" name="/distilbert/transformer/layer.4/attention/Shape_2" type="ShapeOf" version="opset3">
9935 <data output_type="i64" />
9936 <rt_info>
9937 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Shape_2" />
9938 </rt_info>
9939 <input>
9940 <port id="0" precision="FP32">
9941 <dim>-1</dim>
9942 <dim>12</dim>
9943 <dim>-1</dim>
9944 <dim>-1</dim>
9945 </port>
9946 </input>
9947 <output>
9948 <port id="1" precision="I64" names="/distilbert/transformer/layer.4/attention/Shape_2_output_0">
9949 <dim>4</dim>
9950 </port>
9951 </output>
9952 </layer>
9953 <layer id="590" name="/distilbert/transformer/layer.4/attention/Expand" type="Broadcast" version="opset3">
9954 <data mode="bidirectional" />
9955 <rt_info>
9956 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Expand" />
9957 </rt_info>
9958 <input>
9959 <port id="0" precision="BOOL">
9960 <dim>-1</dim>
9961 <dim>1</dim>
9962 <dim>1</dim>
9963 <dim>-1</dim>
9964 </port>
9965 <port id="1" precision="I64">
9966 <dim>4</dim>
9967 </port>
9968 </input>
9969 <output>
9970 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.4/attention/Expand_output_0">
9971 <dim>-1</dim>
9972 <dim>12</dim>
9973 <dim>-1</dim>
9974 <dim>-1</dim>
9975 </port>
9976 </output>
9977 </layer>
9978 <layer id="591" name="/distilbert/transformer/layer.4/attention/Cast" type="Convert" version="opset1">
9979 <data destination_type="boolean" />
9980 <rt_info>
9981 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Cast" />
9982 </rt_info>
9983 <input>
9984 <port id="0" precision="BOOL">
9985 <dim>-1</dim>
9986 <dim>12</dim>
9987 <dim>-1</dim>
9988 <dim>-1</dim>
9989 </port>
9990 </input>
9991 <output>
9992 <port id="1" precision="BOOL" names="/distilbert/transformer/layer.4/attention/Cast_output_0">
9993 <dim>-1</dim>
9994 <dim>12</dim>
9995 <dim>-1</dim>
9996 <dim>-1</dim>
9997 </port>
9998 </output>
9999 </layer>
10000 <layer id="592" name="/distilbert/transformer/layer.4/attention/Constant_14" type="Const" version="opset1">
10001 <data element_type="f32" shape="" offset="95494268" size="4" />
10002 <rt_info>
10003 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_14" />
10004 </rt_info>
10005 <output>
10006 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/attention/Constant_14_output_0" />
10007 </output>
10008 </layer>
10009 <layer id="593" name="/distilbert/transformer/layer.4/attention/Where" type="Select" version="opset1">
10010 <data auto_broadcast="numpy" />
10011 <rt_info>
10012 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Where" />
10013 </rt_info>
10014 <input>
10015 <port id="0" precision="BOOL">
10016 <dim>-1</dim>
10017 <dim>12</dim>
10018 <dim>-1</dim>
10019 <dim>-1</dim>
10020 </port>
10021 <port id="1" precision="FP32" />
10022 <port id="2" precision="FP32">
10023 <dim>-1</dim>
10024 <dim>12</dim>
10025 <dim>-1</dim>
10026 <dim>-1</dim>
10027 </port>
10028 </input>
10029 <output>
10030 <port id="3" precision="FP32" names="/distilbert/transformer/layer.4/attention/Where_output_0">
10031 <dim>-1</dim>
10032 <dim>12</dim>
10033 <dim>-1</dim>
10034 <dim>-1</dim>
10035 </port>
10036 </output>
10037 </layer>
10038 <layer id="594" name="/distilbert/transformer/layer.4/attention/Softmax" type="SoftMax" version="opset8">
10039 <data axis="3" />
10040 <rt_info>
10041 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Softmax" />
10042 </rt_info>
10043 <input>
10044 <port id="0" precision="FP32">
10045 <dim>-1</dim>
10046 <dim>12</dim>
10047 <dim>-1</dim>
10048 <dim>-1</dim>
10049 </port>
10050 </input>
10051 <output>
10052 <port id="1" precision="FP32" names="/distilbert/transformer/layer.4/attention/Softmax_output_0">
10053 <dim>-1</dim>
10054 <dim>12</dim>
10055 <dim>-1</dim>
10056 <dim>-1</dim>
10057 </port>
10058 </output>
10059 </layer>
10060 <layer id="595" name="distilbert.transformer.layer.4.attention.v_lin.bias" type="Const" version="opset1">
10061 <data element_type="f32" shape="768" offset="208826516" size="3072" />
10062 <rt_info>
10063 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.attention.v_lin.bias" />
10064 </rt_info>
10065 <output>
10066 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.attention.v_lin.bias">
10067 <dim>768</dim>
10068 </port>
10069 </output>
10070 </layer>
10071 <layer id="596" name="onnx::MatMul_825" type="Const" version="opset1">
10072 <data element_type="f32" shape="768, 768" offset="208829588" size="2359296" />
10073 <rt_info>
10074 <attribute name="fused_names" version="0" value="onnx::MatMul_825" />
10075 </rt_info>
10076 <output>
10077 <port id="0" precision="FP32" names="onnx::MatMul_825">
10078 <dim>768</dim>
10079 <dim>768</dim>
10080 </port>
10081 </output>
10082 </layer>
10083 <layer id="597" name="/distilbert/transformer/layer.4/attention/v_lin/MatMul" type="MatMul" version="opset1">
10084 <data transpose_a="false" transpose_b="false" />
10085 <rt_info>
10086 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/v_lin/MatMul" />
10087 </rt_info>
10088 <input>
10089 <port id="0" precision="FP32">
10090 <dim>-1</dim>
10091 <dim>-1</dim>
10092 <dim>768</dim>
10093 </port>
10094 <port id="1" precision="FP32">
10095 <dim>768</dim>
10096 <dim>768</dim>
10097 </port>
10098 </input>
10099 <output>
10100 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/v_lin/MatMul_output_0">
10101 <dim>-1</dim>
10102 <dim>-1</dim>
10103 <dim>768</dim>
10104 </port>
10105 </output>
10106 </layer>
10107 <layer id="598" name="/distilbert/transformer/layer.4/attention/v_lin/Add" type="Add" version="opset1">
10108 <data auto_broadcast="numpy" />
10109 <rt_info>
10110 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/v_lin/Add" />
10111 </rt_info>
10112 <input>
10113 <port id="0" precision="FP32">
10114 <dim>768</dim>
10115 </port>
10116 <port id="1" precision="FP32">
10117 <dim>-1</dim>
10118 <dim>-1</dim>
10119 <dim>768</dim>
10120 </port>
10121 </input>
10122 <output>
10123 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/v_lin/Add_output_0">
10124 <dim>-1</dim>
10125 <dim>-1</dim>
10126 <dim>768</dim>
10127 </port>
10128 </output>
10129 </layer>
10130 <layer id="599" name="Constant_255835" type="Const" version="opset1">
10131 <data element_type="i64" shape="1" offset="18440" size="8" />
10132 <rt_info>
10133 <attribute name="fused_names" version="0" value="Constant_255835" />
10134 </rt_info>
10135 <output>
10136 <port id="0" precision="I64">
10137 <dim>1</dim>
10138 </port>
10139 </output>
10140 </layer>
10141 <layer id="600" name="/distilbert/transformer/layer.4/attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
10142 <rt_info>
10143 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Unsqueeze_2, Constant_255835" />
10144 </rt_info>
10145 <input>
10146 <port id="0" precision="I64" />
10147 <port id="1" precision="I64">
10148 <dim>1</dim>
10149 </port>
10150 </input>
10151 <output>
10152 <port id="2" precision="I64" names="/distilbert/transformer/layer.4/attention/Unsqueeze_2_output_0">
10153 <dim>1</dim>
10154 </port>
10155 </output>
10156 </layer>
10157 <layer id="601" name="/distilbert/transformer/layer.4/attention/Constant_8" type="Const" version="opset1">
10158 <data element_type="i64" shape="1" offset="90763288" size="8" />
10159 <rt_info>
10160 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_8" />
10161 </rt_info>
10162 <output>
10163 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_8_output_0">
10164 <dim>1</dim>
10165 </port>
10166 </output>
10167 </layer>
10168 <layer id="602" name="/distilbert/transformer/layer.4/attention/Constant_9" type="Const" version="opset1">
10169 <data element_type="i64" shape="1" offset="93131816" size="8" />
10170 <rt_info>
10171 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_9" />
10172 </rt_info>
10173 <output>
10174 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_9_output_0">
10175 <dim>1</dim>
10176 </port>
10177 </output>
10178 </layer>
10179 <layer id="603" name="/distilbert/transformer/layer.4/attention/Constant_10" type="Const" version="opset1">
10180 <data element_type="i64" shape="1" offset="93131824" size="8" />
10181 <rt_info>
10182 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_10" />
10183 </rt_info>
10184 <output>
10185 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_10_output_0">
10186 <dim>1</dim>
10187 </port>
10188 </output>
10189 </layer>
10190 <layer id="604" name="/distilbert/transformer/layer.4/attention/Concat_2" type="Concat" version="opset1">
10191 <data axis="0" />
10192 <rt_info>
10193 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Concat_2, /distilbert/transformer/layer.4/attention/Constant_10, /distilbert/transformer/layer.4/attention/Constant_8, /distilbert/transformer/layer.4/attention/Constant_9" />
10194 </rt_info>
10195 <input>
10196 <port id="0" precision="I64">
10197 <dim>1</dim>
10198 </port>
10199 <port id="1" precision="I64">
10200 <dim>1</dim>
10201 </port>
10202 <port id="2" precision="I64">
10203 <dim>1</dim>
10204 </port>
10205 <port id="3" precision="I64">
10206 <dim>1</dim>
10207 </port>
10208 </input>
10209 <output>
10210 <port id="4" precision="I64" names="/distilbert/transformer/layer.4/attention/Concat_2_output_0">
10211 <dim>4</dim>
10212 </port>
10213 </output>
10214 </layer>
10215 <layer id="605" name="/distilbert/transformer/layer.4/attention/Reshape_2" type="Reshape" version="opset1">
10216 <data special_zero="true" />
10217 <rt_info>
10218 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Reshape_2" />
10219 </rt_info>
10220 <input>
10221 <port id="0" precision="FP32">
10222 <dim>-1</dim>
10223 <dim>-1</dim>
10224 <dim>768</dim>
10225 </port>
10226 <port id="1" precision="I64">
10227 <dim>4</dim>
10228 </port>
10229 </input>
10230 <output>
10231 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Reshape_2_output_0">
10232 <dim>-1</dim>
10233 <dim>-1</dim>
10234 <dim>12</dim>
10235 <dim>64</dim>
10236 </port>
10237 </output>
10238 </layer>
10239 <layer id="606" name="Constant_255868" type="Const" version="opset1">
10240 <data element_type="i64" shape="4" offset="93131832" size="32" />
10241 <rt_info>
10242 <attribute name="fused_names" version="0" value="Constant_255868" />
10243 </rt_info>
10244 <output>
10245 <port id="0" precision="I64">
10246 <dim>4</dim>
10247 </port>
10248 </output>
10249 </layer>
10250 <layer id="607" name="/distilbert/transformer/layer.4/attention/Transpose_1" type="Transpose" version="opset1">
10251 <rt_info>
10252 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Transpose_1" />
10253 </rt_info>
10254 <input>
10255 <port id="0" precision="FP32">
10256 <dim>-1</dim>
10257 <dim>-1</dim>
10258 <dim>12</dim>
10259 <dim>64</dim>
10260 </port>
10261 <port id="1" precision="I64">
10262 <dim>4</dim>
10263 </port>
10264 </input>
10265 <output>
10266 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Transpose_1_output_0">
10267 <dim>-1</dim>
10268 <dim>12</dim>
10269 <dim>-1</dim>
10270 <dim>64</dim>
10271 </port>
10272 </output>
10273 </layer>
10274 <layer id="608" name="/distilbert/transformer/layer.4/attention/MatMul_1" type="MatMul" version="opset1">
10275 <data transpose_a="false" transpose_b="false" />
10276 <rt_info>
10277 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/MatMul_1" />
10278 </rt_info>
10279 <input>
10280 <port id="0" precision="FP32">
10281 <dim>-1</dim>
10282 <dim>12</dim>
10283 <dim>-1</dim>
10284 <dim>-1</dim>
10285 </port>
10286 <port id="1" precision="FP32">
10287 <dim>-1</dim>
10288 <dim>12</dim>
10289 <dim>-1</dim>
10290 <dim>64</dim>
10291 </port>
10292 </input>
10293 <output>
10294 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/MatMul_1_output_0">
10295 <dim>-1</dim>
10296 <dim>12</dim>
10297 <dim>-1</dim>
10298 <dim>64</dim>
10299 </port>
10300 </output>
10301 </layer>
10302 <layer id="609" name="Constant_255904" type="Const" version="opset1">
10303 <data element_type="i64" shape="4" offset="93131832" size="32" />
10304 <rt_info>
10305 <attribute name="fused_names" version="0" value="Constant_255904" />
10306 </rt_info>
10307 <output>
10308 <port id="0" precision="I64">
10309 <dim>4</dim>
10310 </port>
10311 </output>
10312 </layer>
10313 <layer id="610" name="/distilbert/transformer/layer.4/attention/Transpose_3" type="Transpose" version="opset1">
10314 <rt_info>
10315 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Transpose_3" />
10316 </rt_info>
10317 <input>
10318 <port id="0" precision="FP32">
10319 <dim>-1</dim>
10320 <dim>12</dim>
10321 <dim>-1</dim>
10322 <dim>64</dim>
10323 </port>
10324 <port id="1" precision="I64">
10325 <dim>4</dim>
10326 </port>
10327 </input>
10328 <output>
10329 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Transpose_3_output_0">
10330 <dim>-1</dim>
10331 <dim>-1</dim>
10332 <dim>12</dim>
10333 <dim>64</dim>
10334 </port>
10335 </output>
10336 </layer>
10337 <layer id="611" name="Constant_255906" type="Const" version="opset1">
10338 <data element_type="i64" shape="1" offset="18440" size="8" />
10339 <rt_info>
10340 <attribute name="fused_names" version="0" value="Constant_255906" />
10341 </rt_info>
10342 <output>
10343 <port id="0" precision="I64">
10344 <dim>1</dim>
10345 </port>
10346 </output>
10347 </layer>
10348 <layer id="612" name="/distilbert/transformer/layer.4/attention/Unsqueeze_5" type="Unsqueeze" version="opset1">
10349 <rt_info>
10350 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Unsqueeze_5, Constant_255906" />
10351 </rt_info>
10352 <input>
10353 <port id="0" precision="I64" />
10354 <port id="1" precision="I64">
10355 <dim>1</dim>
10356 </port>
10357 </input>
10358 <output>
10359 <port id="2" precision="I64" names="/distilbert/transformer/layer.4/attention/Unsqueeze_5_output_0">
10360 <dim>1</dim>
10361 </port>
10362 </output>
10363 </layer>
10364 <layer id="613" name="/distilbert/transformer/layer.4/attention/Constant_15" type="Const" version="opset1">
10365 <data element_type="i64" shape="1" offset="90763288" size="8" />
10366 <rt_info>
10367 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_15" />
10368 </rt_info>
10369 <output>
10370 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_15_output_0">
10371 <dim>1</dim>
10372 </port>
10373 </output>
10374 </layer>
10375 <layer id="614" name="/distilbert/transformer/layer.4/attention/Constant_16" type="Const" version="opset1">
10376 <data element_type="i64" shape="1" offset="97856640" size="8" />
10377 <rt_info>
10378 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Constant_16" />
10379 </rt_info>
10380 <output>
10381 <port id="0" precision="I64" names="/distilbert/transformer/layer.4/attention/Constant_16_output_0">
10382 <dim>1</dim>
10383 </port>
10384 </output>
10385 </layer>
10386 <layer id="615" name="/distilbert/transformer/layer.4/attention/Concat_4" type="Concat" version="opset1">
10387 <data axis="0" />
10388 <rt_info>
10389 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Concat_4, /distilbert/transformer/layer.4/attention/Constant_15, /distilbert/transformer/layer.4/attention/Constant_16" />
10390 </rt_info>
10391 <input>
10392 <port id="0" precision="I64">
10393 <dim>1</dim>
10394 </port>
10395 <port id="1" precision="I64">
10396 <dim>1</dim>
10397 </port>
10398 <port id="2" precision="I64">
10399 <dim>1</dim>
10400 </port>
10401 </input>
10402 <output>
10403 <port id="3" precision="I64" names="/distilbert/transformer/layer.4/attention/Concat_4_output_0">
10404 <dim>3</dim>
10405 </port>
10406 </output>
10407 </layer>
10408 <layer id="616" name="/distilbert/transformer/layer.4/attention/Reshape_4" type="Reshape" version="opset1">
10409 <data special_zero="true" />
10410 <rt_info>
10411 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/Reshape_4" />
10412 </rt_info>
10413 <input>
10414 <port id="0" precision="FP32">
10415 <dim>-1</dim>
10416 <dim>-1</dim>
10417 <dim>12</dim>
10418 <dim>64</dim>
10419 </port>
10420 <port id="1" precision="I64">
10421 <dim>3</dim>
10422 </port>
10423 </input>
10424 <output>
10425 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/Reshape_4_output_0">
10426 <dim>-1</dim>
10427 <dim>-1</dim>
10428 <dim>768</dim>
10429 </port>
10430 </output>
10431 </layer>
10432 <layer id="617" name="onnx::MatMul_830" type="Const" version="opset1">
10433 <data element_type="f32" shape="768, 768" offset="211188884" size="2359296" />
10434 <rt_info>
10435 <attribute name="fused_names" version="0" value="onnx::MatMul_830" />
10436 </rt_info>
10437 <output>
10438 <port id="0" precision="FP32" names="onnx::MatMul_830">
10439 <dim>768</dim>
10440 <dim>768</dim>
10441 </port>
10442 </output>
10443 </layer>
10444 <layer id="618" name="/distilbert/transformer/layer.4/attention/out_lin/MatMul" type="MatMul" version="opset1">
10445 <data transpose_a="false" transpose_b="false" />
10446 <rt_info>
10447 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/out_lin/MatMul" />
10448 </rt_info>
10449 <input>
10450 <port id="0" precision="FP32">
10451 <dim>-1</dim>
10452 <dim>-1</dim>
10453 <dim>768</dim>
10454 </port>
10455 <port id="1" precision="FP32">
10456 <dim>768</dim>
10457 <dim>768</dim>
10458 </port>
10459 </input>
10460 <output>
10461 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/out_lin/MatMul_output_0">
10462 <dim>-1</dim>
10463 <dim>-1</dim>
10464 <dim>768</dim>
10465 </port>
10466 </output>
10467 </layer>
10468 <layer id="619" name="/distilbert/transformer/layer.4/attention/out_lin/Add" type="Add" version="opset1">
10469 <data auto_broadcast="numpy" />
10470 <rt_info>
10471 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/attention/out_lin/Add" />
10472 </rt_info>
10473 <input>
10474 <port id="0" precision="FP32">
10475 <dim>768</dim>
10476 </port>
10477 <port id="1" precision="FP32">
10478 <dim>-1</dim>
10479 <dim>-1</dim>
10480 <dim>768</dim>
10481 </port>
10482 </input>
10483 <output>
10484 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/attention/out_lin/Add_output_0">
10485 <dim>-1</dim>
10486 <dim>-1</dim>
10487 <dim>768</dim>
10488 </port>
10489 </output>
10490 </layer>
10491 <layer id="620" name="/distilbert/transformer/layer.4/Add" type="Add" version="opset1">
10492 <data auto_broadcast="numpy" />
10493 <rt_info>
10494 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/Add" />
10495 </rt_info>
10496 <input>
10497 <port id="0" precision="FP32">
10498 <dim>-1</dim>
10499 <dim>-1</dim>
10500 <dim>768</dim>
10501 </port>
10502 <port id="1" precision="FP32">
10503 <dim>-1</dim>
10504 <dim>-1</dim>
10505 <dim>768</dim>
10506 </port>
10507 </input>
10508 <output>
10509 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/Add_output_0">
10510 <dim>-1</dim>
10511 <dim>-1</dim>
10512 <dim>768</dim>
10513 </port>
10514 </output>
10515 </layer>
10516 <layer id="621" name="Constant_255921" type="Const" version="opset1">
10517 <data element_type="i64" shape="1" offset="90763288" size="8" />
10518 <rt_info>
10519 <attribute name="fused_names" version="0" value="Constant_255921" />
10520 </rt_info>
10521 <output>
10522 <port id="0" precision="I64">
10523 <dim>1</dim>
10524 </port>
10525 </output>
10526 </layer>
10527 <layer id="622" name="/distilbert/transformer/layer.4/sa_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
10528 <data keep_dims="true" />
10529 <rt_info>
10530 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/ReduceMean" />
10531 </rt_info>
10532 <input>
10533 <port id="0" precision="FP32">
10534 <dim>-1</dim>
10535 <dim>-1</dim>
10536 <dim>768</dim>
10537 </port>
10538 <port id="1" precision="I64">
10539 <dim>1</dim>
10540 </port>
10541 </input>
10542 <output>
10543 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/ReduceMean_output_0">
10544 <dim>-1</dim>
10545 <dim>-1</dim>
10546 <dim>1</dim>
10547 </port>
10548 </output>
10549 </layer>
10550 <layer id="623" name="/distilbert/transformer/layer.4/sa_layer_norm/Sub" type="Subtract" version="opset1">
10551 <data auto_broadcast="numpy" />
10552 <rt_info>
10553 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Sub" />
10554 </rt_info>
10555 <input>
10556 <port id="0" precision="FP32">
10557 <dim>-1</dim>
10558 <dim>-1</dim>
10559 <dim>768</dim>
10560 </port>
10561 <port id="1" precision="FP32">
10562 <dim>-1</dim>
10563 <dim>-1</dim>
10564 <dim>1</dim>
10565 </port>
10566 </input>
10567 <output>
10568 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Sub_output_0">
10569 <dim>-1</dim>
10570 <dim>-1</dim>
10571 <dim>768</dim>
10572 </port>
10573 </output>
10574 </layer>
10575 <layer id="624" name="/distilbert/transformer/layer.4/sa_layer_norm/Constant" type="Const" version="opset1">
10576 <data element_type="f32" shape="" offset="90763296" size="4" />
10577 <rt_info>
10578 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Constant" />
10579 </rt_info>
10580 <output>
10581 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Constant_output_0" />
10582 </output>
10583 </layer>
10584 <layer id="625" name="/distilbert/transformer/layer.4/sa_layer_norm/Pow" type="Power" version="opset1">
10585 <data auto_broadcast="numpy" />
10586 <rt_info>
10587 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Pow" />
10588 </rt_info>
10589 <input>
10590 <port id="0" precision="FP32">
10591 <dim>-1</dim>
10592 <dim>-1</dim>
10593 <dim>768</dim>
10594 </port>
10595 <port id="1" precision="FP32" />
10596 </input>
10597 <output>
10598 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Pow_output_0">
10599 <dim>-1</dim>
10600 <dim>-1</dim>
10601 <dim>768</dim>
10602 </port>
10603 </output>
10604 </layer>
10605 <layer id="626" name="Constant_255926" type="Const" version="opset1">
10606 <data element_type="i64" shape="1" offset="90763288" size="8" />
10607 <rt_info>
10608 <attribute name="fused_names" version="0" value="Constant_255926" />
10609 </rt_info>
10610 <output>
10611 <port id="0" precision="I64">
10612 <dim>1</dim>
10613 </port>
10614 </output>
10615 </layer>
10616 <layer id="627" name="/distilbert/transformer/layer.4/sa_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
10617 <data keep_dims="true" />
10618 <rt_info>
10619 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/ReduceMean_1" />
10620 </rt_info>
10621 <input>
10622 <port id="0" precision="FP32">
10623 <dim>-1</dim>
10624 <dim>-1</dim>
10625 <dim>768</dim>
10626 </port>
10627 <port id="1" precision="I64">
10628 <dim>1</dim>
10629 </port>
10630 </input>
10631 <output>
10632 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/ReduceMean_1_output_0">
10633 <dim>-1</dim>
10634 <dim>-1</dim>
10635 <dim>1</dim>
10636 </port>
10637 </output>
10638 </layer>
10639 <layer id="628" name="/distilbert/transformer/layer.4/sa_layer_norm/Constant_1" type="Const" version="opset1">
10640 <data element_type="f32" shape="" offset="90763300" size="4" />
10641 <rt_info>
10642 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Constant_1" />
10643 </rt_info>
10644 <output>
10645 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Constant_1_output_0" />
10646 </output>
10647 </layer>
10648 <layer id="629" name="/distilbert/transformer/layer.4/sa_layer_norm/Add" type="Add" version="opset1">
10649 <data auto_broadcast="numpy" />
10650 <rt_info>
10651 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Add" />
10652 </rt_info>
10653 <input>
10654 <port id="0" precision="FP32">
10655 <dim>-1</dim>
10656 <dim>-1</dim>
10657 <dim>1</dim>
10658 </port>
10659 <port id="1" precision="FP32" />
10660 </input>
10661 <output>
10662 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Add_output_0">
10663 <dim>-1</dim>
10664 <dim>-1</dim>
10665 <dim>1</dim>
10666 </port>
10667 </output>
10668 </layer>
10669 <layer id="630" name="/distilbert/transformer/layer.4/sa_layer_norm/Sqrt" type="Sqrt" version="opset1">
10670 <rt_info>
10671 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Sqrt" />
10672 </rt_info>
10673 <input>
10674 <port id="0" precision="FP32">
10675 <dim>-1</dim>
10676 <dim>-1</dim>
10677 <dim>1</dim>
10678 </port>
10679 </input>
10680 <output>
10681 <port id="1" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Sqrt_output_0">
10682 <dim>-1</dim>
10683 <dim>-1</dim>
10684 <dim>1</dim>
10685 </port>
10686 </output>
10687 </layer>
10688 <layer id="631" name="/distilbert/transformer/layer.4/sa_layer_norm/Div" type="Divide" version="opset1">
10689 <data auto_broadcast="numpy" m_pythondiv="true" />
10690 <rt_info>
10691 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Div" />
10692 </rt_info>
10693 <input>
10694 <port id="0" precision="FP32">
10695 <dim>-1</dim>
10696 <dim>-1</dim>
10697 <dim>768</dim>
10698 </port>
10699 <port id="1" precision="FP32">
10700 <dim>-1</dim>
10701 <dim>-1</dim>
10702 <dim>1</dim>
10703 </port>
10704 </input>
10705 <output>
10706 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Div_output_0">
10707 <dim>-1</dim>
10708 <dim>-1</dim>
10709 <dim>768</dim>
10710 </port>
10711 </output>
10712 </layer>
10713 <layer id="632" name="distilbert.transformer.layer.4.sa_layer_norm.weight" type="Const" version="opset1">
10714 <data element_type="f32" shape="768" offset="213548180" size="3072" />
10715 <rt_info>
10716 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.sa_layer_norm.weight" />
10717 </rt_info>
10718 <output>
10719 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.sa_layer_norm.weight">
10720 <dim>768</dim>
10721 </port>
10722 </output>
10723 </layer>
10724 <layer id="633" name="/distilbert/transformer/layer.4/sa_layer_norm/Mul" type="Multiply" version="opset1">
10725 <data auto_broadcast="numpy" />
10726 <rt_info>
10727 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Mul" />
10728 </rt_info>
10729 <input>
10730 <port id="0" precision="FP32">
10731 <dim>-1</dim>
10732 <dim>-1</dim>
10733 <dim>768</dim>
10734 </port>
10735 <port id="1" precision="FP32">
10736 <dim>768</dim>
10737 </port>
10738 </input>
10739 <output>
10740 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Mul_output_0">
10741 <dim>-1</dim>
10742 <dim>-1</dim>
10743 <dim>768</dim>
10744 </port>
10745 </output>
10746 </layer>
10747 <layer id="634" name="distilbert.transformer.layer.4.sa_layer_norm.bias" type="Const" version="opset1">
10748 <data element_type="f32" shape="768" offset="213551252" size="3072" />
10749 <rt_info>
10750 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.sa_layer_norm.bias" />
10751 </rt_info>
10752 <output>
10753 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.sa_layer_norm.bias">
10754 <dim>768</dim>
10755 </port>
10756 </output>
10757 </layer>
10758 <layer id="635" name="/distilbert/transformer/layer.4/sa_layer_norm/Add_1" type="Add" version="opset1">
10759 <data auto_broadcast="numpy" />
10760 <rt_info>
10761 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/sa_layer_norm/Add_1" />
10762 </rt_info>
10763 <input>
10764 <port id="0" precision="FP32">
10765 <dim>-1</dim>
10766 <dim>-1</dim>
10767 <dim>768</dim>
10768 </port>
10769 <port id="1" precision="FP32">
10770 <dim>768</dim>
10771 </port>
10772 </input>
10773 <output>
10774 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/sa_layer_norm/Add_1_output_0">
10775 <dim>-1</dim>
10776 <dim>-1</dim>
10777 <dim>768</dim>
10778 </port>
10779 </output>
10780 </layer>
10781 <layer id="636" name="onnx::MatMul_831" type="Const" version="opset1">
10782 <data element_type="f32" shape="768, 3072" offset="213554324" size="9437184" />
10783 <rt_info>
10784 <attribute name="fused_names" version="0" value="onnx::MatMul_831" />
10785 </rt_info>
10786 <output>
10787 <port id="0" precision="FP32" names="onnx::MatMul_831">
10788 <dim>768</dim>
10789 <dim>3072</dim>
10790 </port>
10791 </output>
10792 </layer>
10793 <layer id="637" name="/distilbert/transformer/layer.4/ffn/lin1/MatMul" type="MatMul" version="opset1">
10794 <data transpose_a="false" transpose_b="false" />
10795 <rt_info>
10796 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/lin1/MatMul" />
10797 </rt_info>
10798 <input>
10799 <port id="0" precision="FP32">
10800 <dim>-1</dim>
10801 <dim>-1</dim>
10802 <dim>768</dim>
10803 </port>
10804 <port id="1" precision="FP32">
10805 <dim>768</dim>
10806 <dim>3072</dim>
10807 </port>
10808 </input>
10809 <output>
10810 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/lin1/MatMul_output_0">
10811 <dim>-1</dim>
10812 <dim>-1</dim>
10813 <dim>3072</dim>
10814 </port>
10815 </output>
10816 </layer>
10817 <layer id="638" name="/distilbert/transformer/layer.4/ffn/lin1/Add" type="Add" version="opset1">
10818 <data auto_broadcast="numpy" />
10819 <rt_info>
10820 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/lin1/Add" />
10821 </rt_info>
10822 <input>
10823 <port id="0" precision="FP32">
10824 <dim>3072</dim>
10825 </port>
10826 <port id="1" precision="FP32">
10827 <dim>-1</dim>
10828 <dim>-1</dim>
10829 <dim>3072</dim>
10830 </port>
10831 </input>
10832 <output>
10833 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/lin1/Add_output_0">
10834 <dim>-1</dim>
10835 <dim>-1</dim>
10836 <dim>3072</dim>
10837 </port>
10838 </output>
10839 </layer>
10840 <layer id="639" name="/distilbert/transformer/layer.4/ffn/activation/Constant" type="Const" version="opset1">
10841 <data element_type="f32" shape="" offset="109659272" size="4" />
10842 <rt_info>
10843 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Constant" />
10844 </rt_info>
10845 <output>
10846 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Constant_output_0" />
10847 </output>
10848 </layer>
10849 <layer id="640" name="/distilbert/transformer/layer.4/ffn/activation/Div" type="Divide" version="opset1">
10850 <data auto_broadcast="numpy" m_pythondiv="true" />
10851 <rt_info>
10852 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Div" />
10853 </rt_info>
10854 <input>
10855 <port id="0" precision="FP32">
10856 <dim>-1</dim>
10857 <dim>-1</dim>
10858 <dim>3072</dim>
10859 </port>
10860 <port id="1" precision="FP32" />
10861 </input>
10862 <output>
10863 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Div_output_0">
10864 <dim>-1</dim>
10865 <dim>-1</dim>
10866 <dim>3072</dim>
10867 </port>
10868 </output>
10869 </layer>
10870 <layer id="641" name="/distilbert/transformer/layer.4/ffn/activation/Erf" type="Erf" version="opset1">
10871 <rt_info>
10872 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Erf" />
10873 </rt_info>
10874 <input>
10875 <port id="0" precision="FP32">
10876 <dim>-1</dim>
10877 <dim>-1</dim>
10878 <dim>3072</dim>
10879 </port>
10880 </input>
10881 <output>
10882 <port id="1" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Erf_output_0">
10883 <dim>-1</dim>
10884 <dim>-1</dim>
10885 <dim>3072</dim>
10886 </port>
10887 </output>
10888 </layer>
10889 <layer id="642" name="/distilbert/transformer/layer.4/ffn/activation/Constant_1" type="Const" version="opset1">
10890 <data element_type="f32" shape="" offset="109659276" size="4" />
10891 <rt_info>
10892 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Constant_1" />
10893 </rt_info>
10894 <output>
10895 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Constant_1_output_0" />
10896 </output>
10897 </layer>
10898 <layer id="643" name="/distilbert/transformer/layer.4/ffn/activation/Add" type="Add" version="opset1">
10899 <data auto_broadcast="numpy" />
10900 <rt_info>
10901 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Add" />
10902 </rt_info>
10903 <input>
10904 <port id="0" precision="FP32">
10905 <dim>-1</dim>
10906 <dim>-1</dim>
10907 <dim>3072</dim>
10908 </port>
10909 <port id="1" precision="FP32" />
10910 </input>
10911 <output>
10912 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Add_output_0">
10913 <dim>-1</dim>
10914 <dim>-1</dim>
10915 <dim>3072</dim>
10916 </port>
10917 </output>
10918 </layer>
10919 <layer id="644" name="/distilbert/transformer/layer.4/ffn/activation/Mul" type="Multiply" version="opset1">
10920 <data auto_broadcast="numpy" />
10921 <rt_info>
10922 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Mul" />
10923 </rt_info>
10924 <input>
10925 <port id="0" precision="FP32">
10926 <dim>-1</dim>
10927 <dim>-1</dim>
10928 <dim>3072</dim>
10929 </port>
10930 <port id="1" precision="FP32">
10931 <dim>-1</dim>
10932 <dim>-1</dim>
10933 <dim>3072</dim>
10934 </port>
10935 </input>
10936 <output>
10937 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Mul_output_0">
10938 <dim>-1</dim>
10939 <dim>-1</dim>
10940 <dim>3072</dim>
10941 </port>
10942 </output>
10943 </layer>
10944 <layer id="645" name="/distilbert/transformer/layer.4/ffn/activation/Constant_2" type="Const" version="opset1">
10945 <data element_type="f32" shape="" offset="109659280" size="4" />
10946 <rt_info>
10947 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Constant_2" />
10948 </rt_info>
10949 <output>
10950 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Constant_2_output_0" />
10951 </output>
10952 </layer>
10953 <layer id="646" name="/distilbert/transformer/layer.4/ffn/activation/Mul_1" type="Multiply" version="opset1">
10954 <data auto_broadcast="numpy" />
10955 <rt_info>
10956 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/activation/Mul_1" />
10957 </rt_info>
10958 <input>
10959 <port id="0" precision="FP32">
10960 <dim>-1</dim>
10961 <dim>-1</dim>
10962 <dim>3072</dim>
10963 </port>
10964 <port id="1" precision="FP32" />
10965 </input>
10966 <output>
10967 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/activation/Mul_1_output_0">
10968 <dim>-1</dim>
10969 <dim>-1</dim>
10970 <dim>3072</dim>
10971 </port>
10972 </output>
10973 </layer>
10974 <layer id="647" name="onnx::MatMul_832" type="Const" version="opset1">
10975 <data element_type="f32" shape="3072, 768" offset="222991508" size="9437184" />
10976 <rt_info>
10977 <attribute name="fused_names" version="0" value="onnx::MatMul_832" />
10978 </rt_info>
10979 <output>
10980 <port id="0" precision="FP32" names="onnx::MatMul_832">
10981 <dim>3072</dim>
10982 <dim>768</dim>
10983 </port>
10984 </output>
10985 </layer>
10986 <layer id="648" name="/distilbert/transformer/layer.4/ffn/lin2/MatMul" type="MatMul" version="opset1">
10987 <data transpose_a="false" transpose_b="false" />
10988 <rt_info>
10989 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/lin2/MatMul" />
10990 </rt_info>
10991 <input>
10992 <port id="0" precision="FP32">
10993 <dim>-1</dim>
10994 <dim>-1</dim>
10995 <dim>3072</dim>
10996 </port>
10997 <port id="1" precision="FP32">
10998 <dim>3072</dim>
10999 <dim>768</dim>
11000 </port>
11001 </input>
11002 <output>
11003 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/lin2/MatMul_output_0">
11004 <dim>-1</dim>
11005 <dim>-1</dim>
11006 <dim>768</dim>
11007 </port>
11008 </output>
11009 </layer>
11010 <layer id="649" name="/distilbert/transformer/layer.4/ffn/lin2/Add" type="Add" version="opset1">
11011 <data auto_broadcast="numpy" />
11012 <rt_info>
11013 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/ffn/lin2/Add" />
11014 </rt_info>
11015 <input>
11016 <port id="0" precision="FP32">
11017 <dim>768</dim>
11018 </port>
11019 <port id="1" precision="FP32">
11020 <dim>-1</dim>
11021 <dim>-1</dim>
11022 <dim>768</dim>
11023 </port>
11024 </input>
11025 <output>
11026 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/ffn/lin2/Add_output_0">
11027 <dim>-1</dim>
11028 <dim>-1</dim>
11029 <dim>768</dim>
11030 </port>
11031 </output>
11032 </layer>
11033 <layer id="650" name="/distilbert/transformer/layer.4/Add_1" type="Add" version="opset1">
11034 <data auto_broadcast="numpy" />
11035 <rt_info>
11036 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/Add_1" />
11037 </rt_info>
11038 <input>
11039 <port id="0" precision="FP32">
11040 <dim>-1</dim>
11041 <dim>-1</dim>
11042 <dim>768</dim>
11043 </port>
11044 <port id="1" precision="FP32">
11045 <dim>-1</dim>
11046 <dim>-1</dim>
11047 <dim>768</dim>
11048 </port>
11049 </input>
11050 <output>
11051 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/Add_1_output_0">
11052 <dim>-1</dim>
11053 <dim>-1</dim>
11054 <dim>768</dim>
11055 </port>
11056 </output>
11057 </layer>
11058 <layer id="651" name="Constant_255947" type="Const" version="opset1">
11059 <data element_type="i64" shape="1" offset="90763288" size="8" />
11060 <rt_info>
11061 <attribute name="fused_names" version="0" value="Constant_255947" />
11062 </rt_info>
11063 <output>
11064 <port id="0" precision="I64">
11065 <dim>1</dim>
11066 </port>
11067 </output>
11068 </layer>
11069 <layer id="652" name="/distilbert/transformer/layer.4/output_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
11070 <data keep_dims="true" />
11071 <rt_info>
11072 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/ReduceMean" />
11073 </rt_info>
11074 <input>
11075 <port id="0" precision="FP32">
11076 <dim>-1</dim>
11077 <dim>-1</dim>
11078 <dim>768</dim>
11079 </port>
11080 <port id="1" precision="I64">
11081 <dim>1</dim>
11082 </port>
11083 </input>
11084 <output>
11085 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/ReduceMean_output_0">
11086 <dim>-1</dim>
11087 <dim>-1</dim>
11088 <dim>1</dim>
11089 </port>
11090 </output>
11091 </layer>
11092 <layer id="653" name="/distilbert/transformer/layer.4/output_layer_norm/Sub" type="Subtract" version="opset1">
11093 <data auto_broadcast="numpy" />
11094 <rt_info>
11095 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Sub" />
11096 </rt_info>
11097 <input>
11098 <port id="0" precision="FP32">
11099 <dim>-1</dim>
11100 <dim>-1</dim>
11101 <dim>768</dim>
11102 </port>
11103 <port id="1" precision="FP32">
11104 <dim>-1</dim>
11105 <dim>-1</dim>
11106 <dim>1</dim>
11107 </port>
11108 </input>
11109 <output>
11110 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Sub_output_0">
11111 <dim>-1</dim>
11112 <dim>-1</dim>
11113 <dim>768</dim>
11114 </port>
11115 </output>
11116 </layer>
11117 <layer id="654" name="/distilbert/transformer/layer.4/output_layer_norm/Constant" type="Const" version="opset1">
11118 <data element_type="f32" shape="" offset="90763296" size="4" />
11119 <rt_info>
11120 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Constant" />
11121 </rt_info>
11122 <output>
11123 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Constant_output_0" />
11124 </output>
11125 </layer>
11126 <layer id="655" name="/distilbert/transformer/layer.4/output_layer_norm/Pow" type="Power" version="opset1">
11127 <data auto_broadcast="numpy" />
11128 <rt_info>
11129 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Pow" />
11130 </rt_info>
11131 <input>
11132 <port id="0" precision="FP32">
11133 <dim>-1</dim>
11134 <dim>-1</dim>
11135 <dim>768</dim>
11136 </port>
11137 <port id="1" precision="FP32" />
11138 </input>
11139 <output>
11140 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Pow_output_0">
11141 <dim>-1</dim>
11142 <dim>-1</dim>
11143 <dim>768</dim>
11144 </port>
11145 </output>
11146 </layer>
11147 <layer id="656" name="Constant_255952" type="Const" version="opset1">
11148 <data element_type="i64" shape="1" offset="90763288" size="8" />
11149 <rt_info>
11150 <attribute name="fused_names" version="0" value="Constant_255952" />
11151 </rt_info>
11152 <output>
11153 <port id="0" precision="I64">
11154 <dim>1</dim>
11155 </port>
11156 </output>
11157 </layer>
11158 <layer id="657" name="/distilbert/transformer/layer.4/output_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
11159 <data keep_dims="true" />
11160 <rt_info>
11161 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/ReduceMean_1" />
11162 </rt_info>
11163 <input>
11164 <port id="0" precision="FP32">
11165 <dim>-1</dim>
11166 <dim>-1</dim>
11167 <dim>768</dim>
11168 </port>
11169 <port id="1" precision="I64">
11170 <dim>1</dim>
11171 </port>
11172 </input>
11173 <output>
11174 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/ReduceMean_1_output_0">
11175 <dim>-1</dim>
11176 <dim>-1</dim>
11177 <dim>1</dim>
11178 </port>
11179 </output>
11180 </layer>
11181 <layer id="658" name="/distilbert/transformer/layer.4/output_layer_norm/Constant_1" type="Const" version="opset1">
11182 <data element_type="f32" shape="" offset="90763300" size="4" />
11183 <rt_info>
11184 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Constant_1" />
11185 </rt_info>
11186 <output>
11187 <port id="0" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Constant_1_output_0" />
11188 </output>
11189 </layer>
11190 <layer id="659" name="/distilbert/transformer/layer.4/output_layer_norm/Add" type="Add" version="opset1">
11191 <data auto_broadcast="numpy" />
11192 <rt_info>
11193 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Add" />
11194 </rt_info>
11195 <input>
11196 <port id="0" precision="FP32">
11197 <dim>-1</dim>
11198 <dim>-1</dim>
11199 <dim>1</dim>
11200 </port>
11201 <port id="1" precision="FP32" />
11202 </input>
11203 <output>
11204 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Add_output_0">
11205 <dim>-1</dim>
11206 <dim>-1</dim>
11207 <dim>1</dim>
11208 </port>
11209 </output>
11210 </layer>
11211 <layer id="660" name="/distilbert/transformer/layer.4/output_layer_norm/Sqrt" type="Sqrt" version="opset1">
11212 <rt_info>
11213 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Sqrt" />
11214 </rt_info>
11215 <input>
11216 <port id="0" precision="FP32">
11217 <dim>-1</dim>
11218 <dim>-1</dim>
11219 <dim>1</dim>
11220 </port>
11221 </input>
11222 <output>
11223 <port id="1" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Sqrt_output_0">
11224 <dim>-1</dim>
11225 <dim>-1</dim>
11226 <dim>1</dim>
11227 </port>
11228 </output>
11229 </layer>
11230 <layer id="661" name="/distilbert/transformer/layer.4/output_layer_norm/Div" type="Divide" version="opset1">
11231 <data auto_broadcast="numpy" m_pythondiv="true" />
11232 <rt_info>
11233 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Div" />
11234 </rt_info>
11235 <input>
11236 <port id="0" precision="FP32">
11237 <dim>-1</dim>
11238 <dim>-1</dim>
11239 <dim>768</dim>
11240 </port>
11241 <port id="1" precision="FP32">
11242 <dim>-1</dim>
11243 <dim>-1</dim>
11244 <dim>1</dim>
11245 </port>
11246 </input>
11247 <output>
11248 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Div_output_0">
11249 <dim>-1</dim>
11250 <dim>-1</dim>
11251 <dim>768</dim>
11252 </port>
11253 </output>
11254 </layer>
11255 <layer id="662" name="distilbert.transformer.layer.4.output_layer_norm.weight" type="Const" version="opset1">
11256 <data element_type="f32" shape="768" offset="232428692" size="3072" />
11257 <rt_info>
11258 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.output_layer_norm.weight" />
11259 </rt_info>
11260 <output>
11261 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.output_layer_norm.weight">
11262 <dim>768</dim>
11263 </port>
11264 </output>
11265 </layer>
11266 <layer id="663" name="/distilbert/transformer/layer.4/output_layer_norm/Mul" type="Multiply" version="opset1">
11267 <data auto_broadcast="numpy" />
11268 <rt_info>
11269 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Mul" />
11270 </rt_info>
11271 <input>
11272 <port id="0" precision="FP32">
11273 <dim>-1</dim>
11274 <dim>-1</dim>
11275 <dim>768</dim>
11276 </port>
11277 <port id="1" precision="FP32">
11278 <dim>768</dim>
11279 </port>
11280 </input>
11281 <output>
11282 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Mul_output_0">
11283 <dim>-1</dim>
11284 <dim>-1</dim>
11285 <dim>768</dim>
11286 </port>
11287 </output>
11288 </layer>
11289 <layer id="664" name="distilbert.transformer.layer.4.output_layer_norm.bias" type="Const" version="opset1">
11290 <data element_type="f32" shape="768" offset="232431764" size="3072" />
11291 <rt_info>
11292 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.4.output_layer_norm.bias" />
11293 </rt_info>
11294 <output>
11295 <port id="0" precision="FP32" names="distilbert.transformer.layer.4.output_layer_norm.bias">
11296 <dim>768</dim>
11297 </port>
11298 </output>
11299 </layer>
11300 <layer id="665" name="/distilbert/transformer/layer.4/output_layer_norm/Add_1" type="Add" version="opset1">
11301 <data auto_broadcast="numpy" />
11302 <rt_info>
11303 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.4/output_layer_norm/Add_1" />
11304 </rt_info>
11305 <input>
11306 <port id="0" precision="FP32">
11307 <dim>-1</dim>
11308 <dim>-1</dim>
11309 <dim>768</dim>
11310 </port>
11311 <port id="1" precision="FP32">
11312 <dim>768</dim>
11313 </port>
11314 </input>
11315 <output>
11316 <port id="2" precision="FP32" names="/distilbert/transformer/layer.4/output_layer_norm/Add_1_output_0">
11317 <dim>-1</dim>
11318 <dim>-1</dim>
11319 <dim>768</dim>
11320 </port>
11321 </output>
11322 </layer>
11323 <layer id="666" name="/distilbert/transformer/layer.5/attention/Shape" type="ShapeOf" version="opset3">
11324 <data output_type="i64" />
11325 <rt_info>
11326 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Shape" />
11327 </rt_info>
11328 <input>
11329 <port id="0" precision="FP32">
11330 <dim>-1</dim>
11331 <dim>-1</dim>
11332 <dim>768</dim>
11333 </port>
11334 </input>
11335 <output>
11336 <port id="1" precision="I64" names="/distilbert/transformer/layer.5/attention/Shape_output_0">
11337 <dim>3</dim>
11338 </port>
11339 </output>
11340 </layer>
11341 <layer id="667" name="/distilbert/transformer/layer.5/attention/Constant" type="Const" version="opset1">
11342 <data element_type="i64" shape="" offset="18440" size="8" />
11343 <rt_info>
11344 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant" />
11345 </rt_info>
11346 <output>
11347 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_output_0" />
11348 </output>
11349 </layer>
11350 <layer id="668" name="Constant_255962" type="Const" version="opset1">
11351 <data element_type="i64" shape="" offset="18440" size="8" />
11352 <rt_info>
11353 <attribute name="fused_names" version="0" value="Constant_255962" />
11354 </rt_info>
11355 <output>
11356 <port id="0" precision="I64" />
11357 </output>
11358 </layer>
11359 <layer id="669" name="/distilbert/transformer/layer.5/attention/Gather" type="Gather" version="opset8">
11360 <data batch_dims="0" />
11361 <rt_info>
11362 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant, /distilbert/transformer/layer.5/attention/Gather, Constant_255962" />
11363 </rt_info>
11364 <input>
11365 <port id="0" precision="I64">
11366 <dim>3</dim>
11367 </port>
11368 <port id="1" precision="I64" />
11369 <port id="2" precision="I64" />
11370 </input>
11371 <output>
11372 <port id="3" precision="I64" names="/distilbert/transformer/layer.5/attention/Gather_output_0" />
11373 </output>
11374 </layer>
11375 <layer id="670" name="Constant_256022" type="Const" version="opset1">
11376 <data element_type="i64" shape="1" offset="18440" size="8" />
11377 <rt_info>
11378 <attribute name="fused_names" version="0" value="Constant_256022" />
11379 </rt_info>
11380 <output>
11381 <port id="0" precision="I64">
11382 <dim>1</dim>
11383 </port>
11384 </output>
11385 </layer>
11386 <layer id="671" name="/distilbert/transformer/layer.5/attention/Unsqueeze_3" type="Unsqueeze" version="opset1">
11387 <rt_info>
11388 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Unsqueeze_3, Constant_256022" />
11389 </rt_info>
11390 <input>
11391 <port id="0" precision="I64" />
11392 <port id="1" precision="I64">
11393 <dim>1</dim>
11394 </port>
11395 </input>
11396 <output>
11397 <port id="2" precision="I64" names="/distilbert/transformer/layer.5/attention/Unsqueeze_3_output_0">
11398 <dim>1</dim>
11399 </port>
11400 </output>
11401 </layer>
11402 <layer id="672" name="/distilbert/transformer/layer.5/attention/Constant_12" type="Const" version="opset1">
11403 <data element_type="i64" shape="1" offset="90763280" size="8" />
11404 <rt_info>
11405 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_12" />
11406 </rt_info>
11407 <output>
11408 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_12_output_0">
11409 <dim>1</dim>
11410 </port>
11411 </output>
11412 </layer>
11413 <layer id="673" name="/distilbert/transformer/layer.5/attention/Constant_13" type="Const" version="opset1">
11414 <data element_type="i64" shape="1" offset="90763280" size="8" />
11415 <rt_info>
11416 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_13" />
11417 </rt_info>
11418 <output>
11419 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_13_output_0">
11420 <dim>1</dim>
11421 </port>
11422 </output>
11423 </layer>
11424 <layer id="674" name="/distilbert/transformer/layer.5/attention/Shape_1" type="ShapeOf" version="opset3">
11425 <data output_type="i64" />
11426 <rt_info>
11427 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Shape_1" />
11428 </rt_info>
11429 <input>
11430 <port id="0" precision="FP32">
11431 <dim>-1</dim>
11432 <dim>-1</dim>
11433 <dim>768</dim>
11434 </port>
11435 </input>
11436 <output>
11437 <port id="1" precision="I64" names="/distilbert/transformer/layer.5/attention/Shape_1_output_0">
11438 <dim>3</dim>
11439 </port>
11440 </output>
11441 </layer>
11442 <layer id="675" name="/distilbert/transformer/layer.5/attention/Constant_1" type="Const" version="opset1">
11443 <data element_type="i64" shape="" offset="90763280" size="8" />
11444 <rt_info>
11445 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_1" />
11446 </rt_info>
11447 <output>
11448 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_1_output_0" />
11449 </output>
11450 </layer>
11451 <layer id="676" name="Constant_255966" type="Const" version="opset1">
11452 <data element_type="i64" shape="" offset="18440" size="8" />
11453 <rt_info>
11454 <attribute name="fused_names" version="0" value="Constant_255966" />
11455 </rt_info>
11456 <output>
11457 <port id="0" precision="I64" />
11458 </output>
11459 </layer>
11460 <layer id="677" name="/distilbert/transformer/layer.5/attention/Gather_1" type="Gather" version="opset8">
11461 <data batch_dims="0" />
11462 <rt_info>
11463 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_1, /distilbert/transformer/layer.5/attention/Gather_1, Constant_255966" />
11464 </rt_info>
11465 <input>
11466 <port id="0" precision="I64">
11467 <dim>3</dim>
11468 </port>
11469 <port id="1" precision="I64" />
11470 <port id="2" precision="I64" />
11471 </input>
11472 <output>
11473 <port id="3" precision="I64" names="/distilbert/transformer/layer.5/attention/Gather_1_output_0" />
11474 </output>
11475 </layer>
11476 <layer id="678" name="Constant_256026" type="Const" version="opset1">
11477 <data element_type="i64" shape="1" offset="18440" size="8" />
11478 <rt_info>
11479 <attribute name="fused_names" version="0" value="Constant_256026" />
11480 </rt_info>
11481 <output>
11482 <port id="0" precision="I64">
11483 <dim>1</dim>
11484 </port>
11485 </output>
11486 </layer>
11487 <layer id="679" name="/distilbert/transformer/layer.5/attention/Unsqueeze_4" type="Unsqueeze" version="opset1">
11488 <rt_info>
11489 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Unsqueeze_4, Constant_256026" />
11490 </rt_info>
11491 <input>
11492 <port id="0" precision="I64" />
11493 <port id="1" precision="I64">
11494 <dim>1</dim>
11495 </port>
11496 </input>
11497 <output>
11498 <port id="2" precision="I64" names="/distilbert/transformer/layer.5/attention/Unsqueeze_4_output_0">
11499 <dim>1</dim>
11500 </port>
11501 </output>
11502 </layer>
11503 <layer id="680" name="/distilbert/transformer/layer.5/attention/Concat_3" type="Concat" version="opset1">
11504 <data axis="0" />
11505 <rt_info>
11506 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Concat_3, /distilbert/transformer/layer.5/attention/Constant_12, /distilbert/transformer/layer.5/attention/Constant_13" />
11507 </rt_info>
11508 <input>
11509 <port id="0" precision="I64">
11510 <dim>1</dim>
11511 </port>
11512 <port id="1" precision="I64">
11513 <dim>1</dim>
11514 </port>
11515 <port id="2" precision="I64">
11516 <dim>1</dim>
11517 </port>
11518 <port id="3" precision="I64">
11519 <dim>1</dim>
11520 </port>
11521 </input>
11522 <output>
11523 <port id="4" precision="I64" names="/distilbert/transformer/layer.5/attention/Concat_3_output_0">
11524 <dim>4</dim>
11525 </port>
11526 </output>
11527 </layer>
11528 <layer id="681" name="/distilbert/transformer/layer.5/attention/Reshape_3" type="Reshape" version="opset1">
11529 <data special_zero="true" />
11530 <rt_info>
11531 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Reshape_3" />
11532 </rt_info>
11533 <input>
11534 <port id="0" precision="BOOL">
11535 <dim>-1</dim>
11536 <dim>-1</dim>
11537 </port>
11538 <port id="1" precision="I64">
11539 <dim>4</dim>
11540 </port>
11541 </input>
11542 <output>
11543 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.5/attention/Reshape_3_output_0">
11544 <dim>-1</dim>
11545 <dim>1</dim>
11546 <dim>1</dim>
11547 <dim>-1</dim>
11548 </port>
11549 </output>
11550 </layer>
11551 <layer id="682" name="distilbert.transformer.layer.5.attention.q_lin.bias" type="Const" version="opset1">
11552 <data element_type="f32" shape="768" offset="232434836" size="3072" />
11553 <rt_info>
11554 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.attention.q_lin.bias" />
11555 </rt_info>
11556 <output>
11557 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.attention.q_lin.bias">
11558 <dim>768</dim>
11559 </port>
11560 </output>
11561 </layer>
11562 <layer id="683" name="onnx::MatMul_833" type="Const" version="opset1">
11563 <data element_type="f32" shape="768, 768" offset="232437908" size="2359296" />
11564 <rt_info>
11565 <attribute name="fused_names" version="0" value="onnx::MatMul_833" />
11566 </rt_info>
11567 <output>
11568 <port id="0" precision="FP32" names="onnx::MatMul_833">
11569 <dim>768</dim>
11570 <dim>768</dim>
11571 </port>
11572 </output>
11573 </layer>
11574 <layer id="684" name="/distilbert/transformer/layer.5/attention/q_lin/MatMul" type="MatMul" version="opset1">
11575 <data transpose_a="false" transpose_b="false" />
11576 <rt_info>
11577 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/q_lin/MatMul" />
11578 </rt_info>
11579 <input>
11580 <port id="0" precision="FP32">
11581 <dim>-1</dim>
11582 <dim>-1</dim>
11583 <dim>768</dim>
11584 </port>
11585 <port id="1" precision="FP32">
11586 <dim>768</dim>
11587 <dim>768</dim>
11588 </port>
11589 </input>
11590 <output>
11591 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/q_lin/MatMul_output_0">
11592 <dim>-1</dim>
11593 <dim>-1</dim>
11594 <dim>768</dim>
11595 </port>
11596 </output>
11597 </layer>
11598 <layer id="685" name="/distilbert/transformer/layer.5/attention/q_lin/Add" type="Add" version="opset1">
11599 <data auto_broadcast="numpy" />
11600 <rt_info>
11601 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/q_lin/Add" />
11602 </rt_info>
11603 <input>
11604 <port id="0" precision="FP32">
11605 <dim>768</dim>
11606 </port>
11607 <port id="1" precision="FP32">
11608 <dim>-1</dim>
11609 <dim>-1</dim>
11610 <dim>768</dim>
11611 </port>
11612 </input>
11613 <output>
11614 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/q_lin/Add_output_0">
11615 <dim>-1</dim>
11616 <dim>-1</dim>
11617 <dim>768</dim>
11618 </port>
11619 </output>
11620 </layer>
11621 <layer id="686" name="Constant_255970" type="Const" version="opset1">
11622 <data element_type="i64" shape="1" offset="18440" size="8" />
11623 <rt_info>
11624 <attribute name="fused_names" version="0" value="Constant_255970" />
11625 </rt_info>
11626 <output>
11627 <port id="0" precision="I64">
11628 <dim>1</dim>
11629 </port>
11630 </output>
11631 </layer>
11632 <layer id="687" name="/distilbert/transformer/layer.5/attention/Unsqueeze" type="Unsqueeze" version="opset1">
11633 <rt_info>
11634 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Unsqueeze, Constant_255970" />
11635 </rt_info>
11636 <input>
11637 <port id="0" precision="I64" />
11638 <port id="1" precision="I64">
11639 <dim>1</dim>
11640 </port>
11641 </input>
11642 <output>
11643 <port id="2" precision="I64" names="/distilbert/transformer/layer.5/attention/Unsqueeze_output_0">
11644 <dim>1</dim>
11645 </port>
11646 </output>
11647 </layer>
11648 <layer id="688" name="/distilbert/transformer/layer.5/attention/Constant_2" type="Const" version="opset1">
11649 <data element_type="i64" shape="1" offset="90763288" size="8" />
11650 <rt_info>
11651 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_2" />
11652 </rt_info>
11653 <output>
11654 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_2_output_0">
11655 <dim>1</dim>
11656 </port>
11657 </output>
11658 </layer>
11659 <layer id="689" name="/distilbert/transformer/layer.5/attention/Constant_3" type="Const" version="opset1">
11660 <data element_type="i64" shape="1" offset="93131816" size="8" />
11661 <rt_info>
11662 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_3" />
11663 </rt_info>
11664 <output>
11665 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_3_output_0">
11666 <dim>1</dim>
11667 </port>
11668 </output>
11669 </layer>
11670 <layer id="690" name="/distilbert/transformer/layer.5/attention/Constant_4" type="Const" version="opset1">
11671 <data element_type="i64" shape="1" offset="93131824" size="8" />
11672 <rt_info>
11673 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_4" />
11674 </rt_info>
11675 <output>
11676 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_4_output_0">
11677 <dim>1</dim>
11678 </port>
11679 </output>
11680 </layer>
11681 <layer id="691" name="/distilbert/transformer/layer.5/attention/Concat" type="Concat" version="opset1">
11682 <data axis="0" />
11683 <rt_info>
11684 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Concat, /distilbert/transformer/layer.5/attention/Constant_2, /distilbert/transformer/layer.5/attention/Constant_3, /distilbert/transformer/layer.5/attention/Constant_4" />
11685 </rt_info>
11686 <input>
11687 <port id="0" precision="I64">
11688 <dim>1</dim>
11689 </port>
11690 <port id="1" precision="I64">
11691 <dim>1</dim>
11692 </port>
11693 <port id="2" precision="I64">
11694 <dim>1</dim>
11695 </port>
11696 <port id="3" precision="I64">
11697 <dim>1</dim>
11698 </port>
11699 </input>
11700 <output>
11701 <port id="4" precision="I64" names="/distilbert/transformer/layer.5/attention/Concat_output_0">
11702 <dim>4</dim>
11703 </port>
11704 </output>
11705 </layer>
11706 <layer id="692" name="/distilbert/transformer/layer.5/attention/Reshape" type="Reshape" version="opset1">
11707 <data special_zero="true" />
11708 <rt_info>
11709 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Reshape" />
11710 </rt_info>
11711 <input>
11712 <port id="0" precision="FP32">
11713 <dim>-1</dim>
11714 <dim>-1</dim>
11715 <dim>768</dim>
11716 </port>
11717 <port id="1" precision="I64">
11718 <dim>4</dim>
11719 </port>
11720 </input>
11721 <output>
11722 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Reshape_output_0">
11723 <dim>-1</dim>
11724 <dim>-1</dim>
11725 <dim>12</dim>
11726 <dim>64</dim>
11727 </port>
11728 </output>
11729 </layer>
11730 <layer id="693" name="Constant_255995" type="Const" version="opset1">
11731 <data element_type="i64" shape="4" offset="93131832" size="32" />
11732 <rt_info>
11733 <attribute name="fused_names" version="0" value="Constant_255995" />
11734 </rt_info>
11735 <output>
11736 <port id="0" precision="I64">
11737 <dim>4</dim>
11738 </port>
11739 </output>
11740 </layer>
11741 <layer id="694" name="/distilbert/transformer/layer.5/attention/Transpose" type="Transpose" version="opset1">
11742 <rt_info>
11743 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Transpose" />
11744 </rt_info>
11745 <input>
11746 <port id="0" precision="FP32">
11747 <dim>-1</dim>
11748 <dim>-1</dim>
11749 <dim>12</dim>
11750 <dim>64</dim>
11751 </port>
11752 <port id="1" precision="I64">
11753 <dim>4</dim>
11754 </port>
11755 </input>
11756 <output>
11757 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Transpose_output_0">
11758 <dim>-1</dim>
11759 <dim>12</dim>
11760 <dim>-1</dim>
11761 <dim>64</dim>
11762 </port>
11763 </output>
11764 </layer>
11765 <layer id="695" name="/distilbert/transformer/layer.5/attention/Constant_11" type="Const" version="opset1">
11766 <data element_type="f32" shape="" offset="93131864" size="4" />
11767 <rt_info>
11768 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_11" />
11769 </rt_info>
11770 <output>
11771 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/attention/Constant_11_output_0" />
11772 </output>
11773 </layer>
11774 <layer id="696" name="/distilbert/transformer/layer.5/attention/Div" type="Divide" version="opset1">
11775 <data auto_broadcast="numpy" m_pythondiv="true" />
11776 <rt_info>
11777 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Div" />
11778 </rt_info>
11779 <input>
11780 <port id="0" precision="FP32">
11781 <dim>-1</dim>
11782 <dim>12</dim>
11783 <dim>-1</dim>
11784 <dim>64</dim>
11785 </port>
11786 <port id="1" precision="FP32" />
11787 </input>
11788 <output>
11789 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Div_output_0">
11790 <dim>-1</dim>
11791 <dim>12</dim>
11792 <dim>-1</dim>
11793 <dim>64</dim>
11794 </port>
11795 </output>
11796 </layer>
11797 <layer id="697" name="distilbert.transformer.layer.5.attention.k_lin.bias" type="Const" version="opset1">
11798 <data element_type="f32" shape="768" offset="234797204" size="3072" />
11799 <rt_info>
11800 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.attention.k_lin.bias" />
11801 </rt_info>
11802 <output>
11803 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.attention.k_lin.bias">
11804 <dim>768</dim>
11805 </port>
11806 </output>
11807 </layer>
11808 <layer id="698" name="onnx::MatMul_843" type="Const" version="opset1">
11809 <data element_type="f32" shape="768, 768" offset="234800276" size="2359296" />
11810 <rt_info>
11811 <attribute name="fused_names" version="0" value="onnx::MatMul_843" />
11812 </rt_info>
11813 <output>
11814 <port id="0" precision="FP32" names="onnx::MatMul_843">
11815 <dim>768</dim>
11816 <dim>768</dim>
11817 </port>
11818 </output>
11819 </layer>
11820 <layer id="699" name="/distilbert/transformer/layer.5/attention/k_lin/MatMul" type="MatMul" version="opset1">
11821 <data transpose_a="false" transpose_b="false" />
11822 <rt_info>
11823 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/k_lin/MatMul" />
11824 </rt_info>
11825 <input>
11826 <port id="0" precision="FP32">
11827 <dim>-1</dim>
11828 <dim>-1</dim>
11829 <dim>768</dim>
11830 </port>
11831 <port id="1" precision="FP32">
11832 <dim>768</dim>
11833 <dim>768</dim>
11834 </port>
11835 </input>
11836 <output>
11837 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/k_lin/MatMul_output_0">
11838 <dim>-1</dim>
11839 <dim>-1</dim>
11840 <dim>768</dim>
11841 </port>
11842 </output>
11843 </layer>
11844 <layer id="700" name="/distilbert/transformer/layer.5/attention/k_lin/Add" type="Add" version="opset1">
11845 <data auto_broadcast="numpy" />
11846 <rt_info>
11847 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/k_lin/Add" />
11848 </rt_info>
11849 <input>
11850 <port id="0" precision="FP32">
11851 <dim>768</dim>
11852 </port>
11853 <port id="1" precision="FP32">
11854 <dim>-1</dim>
11855 <dim>-1</dim>
11856 <dim>768</dim>
11857 </port>
11858 </input>
11859 <output>
11860 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/k_lin/Add_output_0">
11861 <dim>-1</dim>
11862 <dim>-1</dim>
11863 <dim>768</dim>
11864 </port>
11865 </output>
11866 </layer>
11867 <layer id="701" name="Constant_255976" type="Const" version="opset1">
11868 <data element_type="i64" shape="1" offset="18440" size="8" />
11869 <rt_info>
11870 <attribute name="fused_names" version="0" value="Constant_255976" />
11871 </rt_info>
11872 <output>
11873 <port id="0" precision="I64">
11874 <dim>1</dim>
11875 </port>
11876 </output>
11877 </layer>
11878 <layer id="702" name="/distilbert/transformer/layer.5/attention/Unsqueeze_1" type="Unsqueeze" version="opset1">
11879 <rt_info>
11880 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Unsqueeze_1, Constant_255976" />
11881 </rt_info>
11882 <input>
11883 <port id="0" precision="I64" />
11884 <port id="1" precision="I64">
11885 <dim>1</dim>
11886 </port>
11887 </input>
11888 <output>
11889 <port id="2" precision="I64" names="/distilbert/transformer/layer.5/attention/Unsqueeze_1_output_0">
11890 <dim>1</dim>
11891 </port>
11892 </output>
11893 </layer>
11894 <layer id="703" name="/distilbert/transformer/layer.5/attention/Constant_5" type="Const" version="opset1">
11895 <data element_type="i64" shape="1" offset="90763288" size="8" />
11896 <rt_info>
11897 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_5" />
11898 </rt_info>
11899 <output>
11900 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_5_output_0">
11901 <dim>1</dim>
11902 </port>
11903 </output>
11904 </layer>
11905 <layer id="704" name="/distilbert/transformer/layer.5/attention/Constant_6" type="Const" version="opset1">
11906 <data element_type="i64" shape="1" offset="93131816" size="8" />
11907 <rt_info>
11908 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_6" />
11909 </rt_info>
11910 <output>
11911 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_6_output_0">
11912 <dim>1</dim>
11913 </port>
11914 </output>
11915 </layer>
11916 <layer id="705" name="/distilbert/transformer/layer.5/attention/Constant_7" type="Const" version="opset1">
11917 <data element_type="i64" shape="1" offset="93131824" size="8" />
11918 <rt_info>
11919 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_7" />
11920 </rt_info>
11921 <output>
11922 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_7_output_0">
11923 <dim>1</dim>
11924 </port>
11925 </output>
11926 </layer>
11927 <layer id="706" name="/distilbert/transformer/layer.5/attention/Concat_1" type="Concat" version="opset1">
11928 <data axis="0" />
11929 <rt_info>
11930 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Concat_1, /distilbert/transformer/layer.5/attention/Constant_5, /distilbert/transformer/layer.5/attention/Constant_6, /distilbert/transformer/layer.5/attention/Constant_7" />
11931 </rt_info>
11932 <input>
11933 <port id="0" precision="I64">
11934 <dim>1</dim>
11935 </port>
11936 <port id="1" precision="I64">
11937 <dim>1</dim>
11938 </port>
11939 <port id="2" precision="I64">
11940 <dim>1</dim>
11941 </port>
11942 <port id="3" precision="I64">
11943 <dim>1</dim>
11944 </port>
11945 </input>
11946 <output>
11947 <port id="4" precision="I64" names="/distilbert/transformer/layer.5/attention/Concat_1_output_0">
11948 <dim>4</dim>
11949 </port>
11950 </output>
11951 </layer>
11952 <layer id="707" name="/distilbert/transformer/layer.5/attention/Reshape_1" type="Reshape" version="opset1">
11953 <data special_zero="true" />
11954 <rt_info>
11955 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Reshape_1" />
11956 </rt_info>
11957 <input>
11958 <port id="0" precision="FP32">
11959 <dim>-1</dim>
11960 <dim>-1</dim>
11961 <dim>768</dim>
11962 </port>
11963 <port id="1" precision="I64">
11964 <dim>4</dim>
11965 </port>
11966 </input>
11967 <output>
11968 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Reshape_1_output_0">
11969 <dim>-1</dim>
11970 <dim>-1</dim>
11971 <dim>12</dim>
11972 <dim>64</dim>
11973 </port>
11974 </output>
11975 </layer>
11976 <layer id="708" name="Constant_256019" type="Const" version="opset1">
11977 <data element_type="i64" shape="4" offset="95494236" size="32" />
11978 <rt_info>
11979 <attribute name="fused_names" version="0" value="Constant_256019" />
11980 </rt_info>
11981 <output>
11982 <port id="0" precision="I64">
11983 <dim>4</dim>
11984 </port>
11985 </output>
11986 </layer>
11987 <layer id="709" name="/distilbert/transformer/layer.5/attention/Transpose_2" type="Transpose" version="opset1">
11988 <rt_info>
11989 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Transpose_2" />
11990 </rt_info>
11991 <input>
11992 <port id="0" precision="FP32">
11993 <dim>-1</dim>
11994 <dim>-1</dim>
11995 <dim>12</dim>
11996 <dim>64</dim>
11997 </port>
11998 <port id="1" precision="I64">
11999 <dim>4</dim>
12000 </port>
12001 </input>
12002 <output>
12003 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Transpose_2_output_0">
12004 <dim>-1</dim>
12005 <dim>12</dim>
12006 <dim>64</dim>
12007 <dim>-1</dim>
12008 </port>
12009 </output>
12010 </layer>
12011 <layer id="710" name="/distilbert/transformer/layer.5/attention/MatMul" type="MatMul" version="opset1">
12012 <data transpose_a="false" transpose_b="false" />
12013 <rt_info>
12014 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/MatMul" />
12015 </rt_info>
12016 <input>
12017 <port id="0" precision="FP32">
12018 <dim>-1</dim>
12019 <dim>12</dim>
12020 <dim>-1</dim>
12021 <dim>64</dim>
12022 </port>
12023 <port id="1" precision="FP32">
12024 <dim>-1</dim>
12025 <dim>12</dim>
12026 <dim>64</dim>
12027 <dim>-1</dim>
12028 </port>
12029 </input>
12030 <output>
12031 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/MatMul_output_0">
12032 <dim>-1</dim>
12033 <dim>12</dim>
12034 <dim>-1</dim>
12035 <dim>-1</dim>
12036 </port>
12037 </output>
12038 </layer>
12039 <layer id="711" name="/distilbert/transformer/layer.5/attention/Shape_2" type="ShapeOf" version="opset3">
12040 <data output_type="i64" />
12041 <rt_info>
12042 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Shape_2" />
12043 </rt_info>
12044 <input>
12045 <port id="0" precision="FP32">
12046 <dim>-1</dim>
12047 <dim>12</dim>
12048 <dim>-1</dim>
12049 <dim>-1</dim>
12050 </port>
12051 </input>
12052 <output>
12053 <port id="1" precision="I64" names="/distilbert/transformer/layer.5/attention/Shape_2_output_0">
12054 <dim>4</dim>
12055 </port>
12056 </output>
12057 </layer>
12058 <layer id="712" name="/distilbert/transformer/layer.5/attention/Expand" type="Broadcast" version="opset3">
12059 <data mode="bidirectional" />
12060 <rt_info>
12061 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Expand" />
12062 </rt_info>
12063 <input>
12064 <port id="0" precision="BOOL">
12065 <dim>-1</dim>
12066 <dim>1</dim>
12067 <dim>1</dim>
12068 <dim>-1</dim>
12069 </port>
12070 <port id="1" precision="I64">
12071 <dim>4</dim>
12072 </port>
12073 </input>
12074 <output>
12075 <port id="2" precision="BOOL" names="/distilbert/transformer/layer.5/attention/Expand_output_0">
12076 <dim>-1</dim>
12077 <dim>12</dim>
12078 <dim>-1</dim>
12079 <dim>-1</dim>
12080 </port>
12081 </output>
12082 </layer>
12083 <layer id="713" name="/distilbert/transformer/layer.5/attention/Cast" type="Convert" version="opset1">
12084 <data destination_type="boolean" />
12085 <rt_info>
12086 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Cast" />
12087 </rt_info>
12088 <input>
12089 <port id="0" precision="BOOL">
12090 <dim>-1</dim>
12091 <dim>12</dim>
12092 <dim>-1</dim>
12093 <dim>-1</dim>
12094 </port>
12095 </input>
12096 <output>
12097 <port id="1" precision="BOOL" names="/distilbert/transformer/layer.5/attention/Cast_output_0">
12098 <dim>-1</dim>
12099 <dim>12</dim>
12100 <dim>-1</dim>
12101 <dim>-1</dim>
12102 </port>
12103 </output>
12104 </layer>
12105 <layer id="714" name="/distilbert/transformer/layer.5/attention/Constant_14" type="Const" version="opset1">
12106 <data element_type="f32" shape="" offset="95494268" size="4" />
12107 <rt_info>
12108 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_14" />
12109 </rt_info>
12110 <output>
12111 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/attention/Constant_14_output_0" />
12112 </output>
12113 </layer>
12114 <layer id="715" name="/distilbert/transformer/layer.5/attention/Where" type="Select" version="opset1">
12115 <data auto_broadcast="numpy" />
12116 <rt_info>
12117 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Where" />
12118 </rt_info>
12119 <input>
12120 <port id="0" precision="BOOL">
12121 <dim>-1</dim>
12122 <dim>12</dim>
12123 <dim>-1</dim>
12124 <dim>-1</dim>
12125 </port>
12126 <port id="1" precision="FP32" />
12127 <port id="2" precision="FP32">
12128 <dim>-1</dim>
12129 <dim>12</dim>
12130 <dim>-1</dim>
12131 <dim>-1</dim>
12132 </port>
12133 </input>
12134 <output>
12135 <port id="3" precision="FP32" names="/distilbert/transformer/layer.5/attention/Where_output_0">
12136 <dim>-1</dim>
12137 <dim>12</dim>
12138 <dim>-1</dim>
12139 <dim>-1</dim>
12140 </port>
12141 </output>
12142 </layer>
12143 <layer id="716" name="/distilbert/transformer/layer.5/attention/Softmax" type="SoftMax" version="opset8">
12144 <data axis="3" />
12145 <rt_info>
12146 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Softmax" />
12147 </rt_info>
12148 <input>
12149 <port id="0" precision="FP32">
12150 <dim>-1</dim>
12151 <dim>12</dim>
12152 <dim>-1</dim>
12153 <dim>-1</dim>
12154 </port>
12155 </input>
12156 <output>
12157 <port id="1" precision="FP32" names="/distilbert/transformer/layer.5/attention/Softmax_output_0">
12158 <dim>-1</dim>
12159 <dim>12</dim>
12160 <dim>-1</dim>
12161 <dim>-1</dim>
12162 </port>
12163 </output>
12164 </layer>
12165 <layer id="717" name="distilbert.transformer.layer.5.attention.v_lin.bias" type="Const" version="opset1">
12166 <data element_type="f32" shape="768" offset="237159572" size="3072" />
12167 <rt_info>
12168 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.attention.v_lin.bias" />
12169 </rt_info>
12170 <output>
12171 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.attention.v_lin.bias">
12172 <dim>768</dim>
12173 </port>
12174 </output>
12175 </layer>
12176 <layer id="718" name="onnx::MatMul_844" type="Const" version="opset1">
12177 <data element_type="f32" shape="768, 768" offset="237162644" size="2359296" />
12178 <rt_info>
12179 <attribute name="fused_names" version="0" value="onnx::MatMul_844" />
12180 </rt_info>
12181 <output>
12182 <port id="0" precision="FP32" names="onnx::MatMul_844">
12183 <dim>768</dim>
12184 <dim>768</dim>
12185 </port>
12186 </output>
12187 </layer>
12188 <layer id="719" name="/distilbert/transformer/layer.5/attention/v_lin/MatMul" type="MatMul" version="opset1">
12189 <data transpose_a="false" transpose_b="false" />
12190 <rt_info>
12191 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/v_lin/MatMul" />
12192 </rt_info>
12193 <input>
12194 <port id="0" precision="FP32">
12195 <dim>-1</dim>
12196 <dim>-1</dim>
12197 <dim>768</dim>
12198 </port>
12199 <port id="1" precision="FP32">
12200 <dim>768</dim>
12201 <dim>768</dim>
12202 </port>
12203 </input>
12204 <output>
12205 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/v_lin/MatMul_output_0">
12206 <dim>-1</dim>
12207 <dim>-1</dim>
12208 <dim>768</dim>
12209 </port>
12210 </output>
12211 </layer>
12212 <layer id="720" name="/distilbert/transformer/layer.5/attention/v_lin/Add" type="Add" version="opset1">
12213 <data auto_broadcast="numpy" />
12214 <rt_info>
12215 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/v_lin/Add" />
12216 </rt_info>
12217 <input>
12218 <port id="0" precision="FP32">
12219 <dim>768</dim>
12220 </port>
12221 <port id="1" precision="FP32">
12222 <dim>-1</dim>
12223 <dim>-1</dim>
12224 <dim>768</dim>
12225 </port>
12226 </input>
12227 <output>
12228 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/v_lin/Add_output_0">
12229 <dim>-1</dim>
12230 <dim>-1</dim>
12231 <dim>768</dim>
12232 </port>
12233 </output>
12234 </layer>
12235 <layer id="721" name="Constant_255982" type="Const" version="opset1">
12236 <data element_type="i64" shape="1" offset="18440" size="8" />
12237 <rt_info>
12238 <attribute name="fused_names" version="0" value="Constant_255982" />
12239 </rt_info>
12240 <output>
12241 <port id="0" precision="I64">
12242 <dim>1</dim>
12243 </port>
12244 </output>
12245 </layer>
12246 <layer id="722" name="/distilbert/transformer/layer.5/attention/Unsqueeze_2" type="Unsqueeze" version="opset1">
12247 <rt_info>
12248 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Unsqueeze_2, Constant_255982" />
12249 </rt_info>
12250 <input>
12251 <port id="0" precision="I64" />
12252 <port id="1" precision="I64">
12253 <dim>1</dim>
12254 </port>
12255 </input>
12256 <output>
12257 <port id="2" precision="I64" names="/distilbert/transformer/layer.5/attention/Unsqueeze_2_output_0">
12258 <dim>1</dim>
12259 </port>
12260 </output>
12261 </layer>
12262 <layer id="723" name="/distilbert/transformer/layer.5/attention/Constant_8" type="Const" version="opset1">
12263 <data element_type="i64" shape="1" offset="90763288" size="8" />
12264 <rt_info>
12265 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_8" />
12266 </rt_info>
12267 <output>
12268 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_8_output_0">
12269 <dim>1</dim>
12270 </port>
12271 </output>
12272 </layer>
12273 <layer id="724" name="/distilbert/transformer/layer.5/attention/Constant_9" type="Const" version="opset1">
12274 <data element_type="i64" shape="1" offset="93131816" size="8" />
12275 <rt_info>
12276 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_9" />
12277 </rt_info>
12278 <output>
12279 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_9_output_0">
12280 <dim>1</dim>
12281 </port>
12282 </output>
12283 </layer>
12284 <layer id="725" name="/distilbert/transformer/layer.5/attention/Constant_10" type="Const" version="opset1">
12285 <data element_type="i64" shape="1" offset="93131824" size="8" />
12286 <rt_info>
12287 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_10" />
12288 </rt_info>
12289 <output>
12290 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_10_output_0">
12291 <dim>1</dim>
12292 </port>
12293 </output>
12294 </layer>
12295 <layer id="726" name="/distilbert/transformer/layer.5/attention/Concat_2" type="Concat" version="opset1">
12296 <data axis="0" />
12297 <rt_info>
12298 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Concat_2, /distilbert/transformer/layer.5/attention/Constant_10, /distilbert/transformer/layer.5/attention/Constant_8, /distilbert/transformer/layer.5/attention/Constant_9" />
12299 </rt_info>
12300 <input>
12301 <port id="0" precision="I64">
12302 <dim>1</dim>
12303 </port>
12304 <port id="1" precision="I64">
12305 <dim>1</dim>
12306 </port>
12307 <port id="2" precision="I64">
12308 <dim>1</dim>
12309 </port>
12310 <port id="3" precision="I64">
12311 <dim>1</dim>
12312 </port>
12313 </input>
12314 <output>
12315 <port id="4" precision="I64" names="/distilbert/transformer/layer.5/attention/Concat_2_output_0">
12316 <dim>4</dim>
12317 </port>
12318 </output>
12319 </layer>
12320 <layer id="727" name="/distilbert/transformer/layer.5/attention/Reshape_2" type="Reshape" version="opset1">
12321 <data special_zero="true" />
12322 <rt_info>
12323 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Reshape_2" />
12324 </rt_info>
12325 <input>
12326 <port id="0" precision="FP32">
12327 <dim>-1</dim>
12328 <dim>-1</dim>
12329 <dim>768</dim>
12330 </port>
12331 <port id="1" precision="I64">
12332 <dim>4</dim>
12333 </port>
12334 </input>
12335 <output>
12336 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Reshape_2_output_0">
12337 <dim>-1</dim>
12338 <dim>-1</dim>
12339 <dim>12</dim>
12340 <dim>64</dim>
12341 </port>
12342 </output>
12343 </layer>
12344 <layer id="728" name="Constant_256015" type="Const" version="opset1">
12345 <data element_type="i64" shape="4" offset="93131832" size="32" />
12346 <rt_info>
12347 <attribute name="fused_names" version="0" value="Constant_256015" />
12348 </rt_info>
12349 <output>
12350 <port id="0" precision="I64">
12351 <dim>4</dim>
12352 </port>
12353 </output>
12354 </layer>
12355 <layer id="729" name="/distilbert/transformer/layer.5/attention/Transpose_1" type="Transpose" version="opset1">
12356 <rt_info>
12357 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Transpose_1" />
12358 </rt_info>
12359 <input>
12360 <port id="0" precision="FP32">
12361 <dim>-1</dim>
12362 <dim>-1</dim>
12363 <dim>12</dim>
12364 <dim>64</dim>
12365 </port>
12366 <port id="1" precision="I64">
12367 <dim>4</dim>
12368 </port>
12369 </input>
12370 <output>
12371 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Transpose_1_output_0">
12372 <dim>-1</dim>
12373 <dim>12</dim>
12374 <dim>-1</dim>
12375 <dim>64</dim>
12376 </port>
12377 </output>
12378 </layer>
12379 <layer id="730" name="/distilbert/transformer/layer.5/attention/MatMul_1" type="MatMul" version="opset1">
12380 <data transpose_a="false" transpose_b="false" />
12381 <rt_info>
12382 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/MatMul_1" />
12383 </rt_info>
12384 <input>
12385 <port id="0" precision="FP32">
12386 <dim>-1</dim>
12387 <dim>12</dim>
12388 <dim>-1</dim>
12389 <dim>-1</dim>
12390 </port>
12391 <port id="1" precision="FP32">
12392 <dim>-1</dim>
12393 <dim>12</dim>
12394 <dim>-1</dim>
12395 <dim>64</dim>
12396 </port>
12397 </input>
12398 <output>
12399 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/MatMul_1_output_0">
12400 <dim>-1</dim>
12401 <dim>12</dim>
12402 <dim>-1</dim>
12403 <dim>64</dim>
12404 </port>
12405 </output>
12406 </layer>
12407 <layer id="731" name="Constant_256051" type="Const" version="opset1">
12408 <data element_type="i64" shape="4" offset="93131832" size="32" />
12409 <rt_info>
12410 <attribute name="fused_names" version="0" value="Constant_256051" />
12411 </rt_info>
12412 <output>
12413 <port id="0" precision="I64">
12414 <dim>4</dim>
12415 </port>
12416 </output>
12417 </layer>
12418 <layer id="732" name="/distilbert/transformer/layer.5/attention/Transpose_3" type="Transpose" version="opset1">
12419 <rt_info>
12420 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Transpose_3" />
12421 </rt_info>
12422 <input>
12423 <port id="0" precision="FP32">
12424 <dim>-1</dim>
12425 <dim>12</dim>
12426 <dim>-1</dim>
12427 <dim>64</dim>
12428 </port>
12429 <port id="1" precision="I64">
12430 <dim>4</dim>
12431 </port>
12432 </input>
12433 <output>
12434 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Transpose_3_output_0">
12435 <dim>-1</dim>
12436 <dim>-1</dim>
12437 <dim>12</dim>
12438 <dim>64</dim>
12439 </port>
12440 </output>
12441 </layer>
12442 <layer id="733" name="Constant_256053" type="Const" version="opset1">
12443 <data element_type="i64" shape="1" offset="18440" size="8" />
12444 <rt_info>
12445 <attribute name="fused_names" version="0" value="Constant_256053" />
12446 </rt_info>
12447 <output>
12448 <port id="0" precision="I64">
12449 <dim>1</dim>
12450 </port>
12451 </output>
12452 </layer>
12453 <layer id="734" name="/distilbert/transformer/layer.5/attention/Unsqueeze_5" type="Unsqueeze" version="opset1">
12454 <rt_info>
12455 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Unsqueeze_5, Constant_256053" />
12456 </rt_info>
12457 <input>
12458 <port id="0" precision="I64" />
12459 <port id="1" precision="I64">
12460 <dim>1</dim>
12461 </port>
12462 </input>
12463 <output>
12464 <port id="2" precision="I64" names="/distilbert/transformer/layer.5/attention/Unsqueeze_5_output_0">
12465 <dim>1</dim>
12466 </port>
12467 </output>
12468 </layer>
12469 <layer id="735" name="/distilbert/transformer/layer.5/attention/Constant_15" type="Const" version="opset1">
12470 <data element_type="i64" shape="1" offset="90763288" size="8" />
12471 <rt_info>
12472 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_15" />
12473 </rt_info>
12474 <output>
12475 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_15_output_0">
12476 <dim>1</dim>
12477 </port>
12478 </output>
12479 </layer>
12480 <layer id="736" name="/distilbert/transformer/layer.5/attention/Constant_16" type="Const" version="opset1">
12481 <data element_type="i64" shape="1" offset="97856640" size="8" />
12482 <rt_info>
12483 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Constant_16" />
12484 </rt_info>
12485 <output>
12486 <port id="0" precision="I64" names="/distilbert/transformer/layer.5/attention/Constant_16_output_0">
12487 <dim>1</dim>
12488 </port>
12489 </output>
12490 </layer>
12491 <layer id="737" name="/distilbert/transformer/layer.5/attention/Concat_4" type="Concat" version="opset1">
12492 <data axis="0" />
12493 <rt_info>
12494 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Concat_4, /distilbert/transformer/layer.5/attention/Constant_15, /distilbert/transformer/layer.5/attention/Constant_16" />
12495 </rt_info>
12496 <input>
12497 <port id="0" precision="I64">
12498 <dim>1</dim>
12499 </port>
12500 <port id="1" precision="I64">
12501 <dim>1</dim>
12502 </port>
12503 <port id="2" precision="I64">
12504 <dim>1</dim>
12505 </port>
12506 </input>
12507 <output>
12508 <port id="3" precision="I64" names="/distilbert/transformer/layer.5/attention/Concat_4_output_0">
12509 <dim>3</dim>
12510 </port>
12511 </output>
12512 </layer>
12513 <layer id="738" name="/distilbert/transformer/layer.5/attention/Reshape_4" type="Reshape" version="opset1">
12514 <data special_zero="true" />
12515 <rt_info>
12516 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/Reshape_4" />
12517 </rt_info>
12518 <input>
12519 <port id="0" precision="FP32">
12520 <dim>-1</dim>
12521 <dim>-1</dim>
12522 <dim>12</dim>
12523 <dim>64</dim>
12524 </port>
12525 <port id="1" precision="I64">
12526 <dim>3</dim>
12527 </port>
12528 </input>
12529 <output>
12530 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/Reshape_4_output_0">
12531 <dim>-1</dim>
12532 <dim>-1</dim>
12533 <dim>768</dim>
12534 </port>
12535 </output>
12536 </layer>
12537 <layer id="739" name="onnx::MatMul_849" type="Const" version="opset1">
12538 <data element_type="f32" shape="768, 768" offset="239521940" size="2359296" />
12539 <rt_info>
12540 <attribute name="fused_names" version="0" value="onnx::MatMul_849" />
12541 </rt_info>
12542 <output>
12543 <port id="0" precision="FP32" names="onnx::MatMul_849">
12544 <dim>768</dim>
12545 <dim>768</dim>
12546 </port>
12547 </output>
12548 </layer>
12549 <layer id="740" name="/distilbert/transformer/layer.5/attention/out_lin/MatMul" type="MatMul" version="opset1">
12550 <data transpose_a="false" transpose_b="false" />
12551 <rt_info>
12552 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/out_lin/MatMul" />
12553 </rt_info>
12554 <input>
12555 <port id="0" precision="FP32">
12556 <dim>-1</dim>
12557 <dim>-1</dim>
12558 <dim>768</dim>
12559 </port>
12560 <port id="1" precision="FP32">
12561 <dim>768</dim>
12562 <dim>768</dim>
12563 </port>
12564 </input>
12565 <output>
12566 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/out_lin/MatMul_output_0">
12567 <dim>-1</dim>
12568 <dim>-1</dim>
12569 <dim>768</dim>
12570 </port>
12571 </output>
12572 </layer>
12573 <layer id="741" name="/distilbert/transformer/layer.5/attention/out_lin/Add" type="Add" version="opset1">
12574 <data auto_broadcast="numpy" />
12575 <rt_info>
12576 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/attention/out_lin/Add" />
12577 </rt_info>
12578 <input>
12579 <port id="0" precision="FP32">
12580 <dim>768</dim>
12581 </port>
12582 <port id="1" precision="FP32">
12583 <dim>-1</dim>
12584 <dim>-1</dim>
12585 <dim>768</dim>
12586 </port>
12587 </input>
12588 <output>
12589 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/attention/out_lin/Add_output_0">
12590 <dim>-1</dim>
12591 <dim>-1</dim>
12592 <dim>768</dim>
12593 </port>
12594 </output>
12595 </layer>
12596 <layer id="742" name="/distilbert/transformer/layer.5/Add" type="Add" version="opset1">
12597 <data auto_broadcast="numpy" />
12598 <rt_info>
12599 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/Add" />
12600 </rt_info>
12601 <input>
12602 <port id="0" precision="FP32">
12603 <dim>-1</dim>
12604 <dim>-1</dim>
12605 <dim>768</dim>
12606 </port>
12607 <port id="1" precision="FP32">
12608 <dim>-1</dim>
12609 <dim>-1</dim>
12610 <dim>768</dim>
12611 </port>
12612 </input>
12613 <output>
12614 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/Add_output_0">
12615 <dim>-1</dim>
12616 <dim>-1</dim>
12617 <dim>768</dim>
12618 </port>
12619 </output>
12620 </layer>
12621 <layer id="743" name="Constant_256068" type="Const" version="opset1">
12622 <data element_type="i64" shape="1" offset="90763288" size="8" />
12623 <rt_info>
12624 <attribute name="fused_names" version="0" value="Constant_256068" />
12625 </rt_info>
12626 <output>
12627 <port id="0" precision="I64">
12628 <dim>1</dim>
12629 </port>
12630 </output>
12631 </layer>
12632 <layer id="744" name="/distilbert/transformer/layer.5/sa_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
12633 <data keep_dims="true" />
12634 <rt_info>
12635 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/ReduceMean" />
12636 </rt_info>
12637 <input>
12638 <port id="0" precision="FP32">
12639 <dim>-1</dim>
12640 <dim>-1</dim>
12641 <dim>768</dim>
12642 </port>
12643 <port id="1" precision="I64">
12644 <dim>1</dim>
12645 </port>
12646 </input>
12647 <output>
12648 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/ReduceMean_output_0">
12649 <dim>-1</dim>
12650 <dim>-1</dim>
12651 <dim>1</dim>
12652 </port>
12653 </output>
12654 </layer>
12655 <layer id="745" name="/distilbert/transformer/layer.5/sa_layer_norm/Sub" type="Subtract" version="opset1">
12656 <data auto_broadcast="numpy" />
12657 <rt_info>
12658 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Sub" />
12659 </rt_info>
12660 <input>
12661 <port id="0" precision="FP32">
12662 <dim>-1</dim>
12663 <dim>-1</dim>
12664 <dim>768</dim>
12665 </port>
12666 <port id="1" precision="FP32">
12667 <dim>-1</dim>
12668 <dim>-1</dim>
12669 <dim>1</dim>
12670 </port>
12671 </input>
12672 <output>
12673 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Sub_output_0">
12674 <dim>-1</dim>
12675 <dim>-1</dim>
12676 <dim>768</dim>
12677 </port>
12678 </output>
12679 </layer>
12680 <layer id="746" name="/distilbert/transformer/layer.5/sa_layer_norm/Constant" type="Const" version="opset1">
12681 <data element_type="f32" shape="" offset="90763296" size="4" />
12682 <rt_info>
12683 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Constant" />
12684 </rt_info>
12685 <output>
12686 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Constant_output_0" />
12687 </output>
12688 </layer>
12689 <layer id="747" name="/distilbert/transformer/layer.5/sa_layer_norm/Pow" type="Power" version="opset1">
12690 <data auto_broadcast="numpy" />
12691 <rt_info>
12692 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Pow" />
12693 </rt_info>
12694 <input>
12695 <port id="0" precision="FP32">
12696 <dim>-1</dim>
12697 <dim>-1</dim>
12698 <dim>768</dim>
12699 </port>
12700 <port id="1" precision="FP32" />
12701 </input>
12702 <output>
12703 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Pow_output_0">
12704 <dim>-1</dim>
12705 <dim>-1</dim>
12706 <dim>768</dim>
12707 </port>
12708 </output>
12709 </layer>
12710 <layer id="748" name="Constant_256073" type="Const" version="opset1">
12711 <data element_type="i64" shape="1" offset="90763288" size="8" />
12712 <rt_info>
12713 <attribute name="fused_names" version="0" value="Constant_256073" />
12714 </rt_info>
12715 <output>
12716 <port id="0" precision="I64">
12717 <dim>1</dim>
12718 </port>
12719 </output>
12720 </layer>
12721 <layer id="749" name="/distilbert/transformer/layer.5/sa_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
12722 <data keep_dims="true" />
12723 <rt_info>
12724 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/ReduceMean_1" />
12725 </rt_info>
12726 <input>
12727 <port id="0" precision="FP32">
12728 <dim>-1</dim>
12729 <dim>-1</dim>
12730 <dim>768</dim>
12731 </port>
12732 <port id="1" precision="I64">
12733 <dim>1</dim>
12734 </port>
12735 </input>
12736 <output>
12737 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/ReduceMean_1_output_0">
12738 <dim>-1</dim>
12739 <dim>-1</dim>
12740 <dim>1</dim>
12741 </port>
12742 </output>
12743 </layer>
12744 <layer id="750" name="/distilbert/transformer/layer.5/sa_layer_norm/Constant_1" type="Const" version="opset1">
12745 <data element_type="f32" shape="" offset="90763300" size="4" />
12746 <rt_info>
12747 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Constant_1" />
12748 </rt_info>
12749 <output>
12750 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Constant_1_output_0" />
12751 </output>
12752 </layer>
12753 <layer id="751" name="/distilbert/transformer/layer.5/sa_layer_norm/Add" type="Add" version="opset1">
12754 <data auto_broadcast="numpy" />
12755 <rt_info>
12756 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Add" />
12757 </rt_info>
12758 <input>
12759 <port id="0" precision="FP32">
12760 <dim>-1</dim>
12761 <dim>-1</dim>
12762 <dim>1</dim>
12763 </port>
12764 <port id="1" precision="FP32" />
12765 </input>
12766 <output>
12767 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Add_output_0">
12768 <dim>-1</dim>
12769 <dim>-1</dim>
12770 <dim>1</dim>
12771 </port>
12772 </output>
12773 </layer>
12774 <layer id="752" name="/distilbert/transformer/layer.5/sa_layer_norm/Sqrt" type="Sqrt" version="opset1">
12775 <rt_info>
12776 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Sqrt" />
12777 </rt_info>
12778 <input>
12779 <port id="0" precision="FP32">
12780 <dim>-1</dim>
12781 <dim>-1</dim>
12782 <dim>1</dim>
12783 </port>
12784 </input>
12785 <output>
12786 <port id="1" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Sqrt_output_0">
12787 <dim>-1</dim>
12788 <dim>-1</dim>
12789 <dim>1</dim>
12790 </port>
12791 </output>
12792 </layer>
12793 <layer id="753" name="/distilbert/transformer/layer.5/sa_layer_norm/Div" type="Divide" version="opset1">
12794 <data auto_broadcast="numpy" m_pythondiv="true" />
12795 <rt_info>
12796 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Div" />
12797 </rt_info>
12798 <input>
12799 <port id="0" precision="FP32">
12800 <dim>-1</dim>
12801 <dim>-1</dim>
12802 <dim>768</dim>
12803 </port>
12804 <port id="1" precision="FP32">
12805 <dim>-1</dim>
12806 <dim>-1</dim>
12807 <dim>1</dim>
12808 </port>
12809 </input>
12810 <output>
12811 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Div_output_0">
12812 <dim>-1</dim>
12813 <dim>-1</dim>
12814 <dim>768</dim>
12815 </port>
12816 </output>
12817 </layer>
12818 <layer id="754" name="distilbert.transformer.layer.5.sa_layer_norm.weight" type="Const" version="opset1">
12819 <data element_type="f32" shape="768" offset="241881236" size="3072" />
12820 <rt_info>
12821 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.sa_layer_norm.weight" />
12822 </rt_info>
12823 <output>
12824 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.sa_layer_norm.weight">
12825 <dim>768</dim>
12826 </port>
12827 </output>
12828 </layer>
12829 <layer id="755" name="/distilbert/transformer/layer.5/sa_layer_norm/Mul" type="Multiply" version="opset1">
12830 <data auto_broadcast="numpy" />
12831 <rt_info>
12832 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Mul" />
12833 </rt_info>
12834 <input>
12835 <port id="0" precision="FP32">
12836 <dim>-1</dim>
12837 <dim>-1</dim>
12838 <dim>768</dim>
12839 </port>
12840 <port id="1" precision="FP32">
12841 <dim>768</dim>
12842 </port>
12843 </input>
12844 <output>
12845 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Mul_output_0">
12846 <dim>-1</dim>
12847 <dim>-1</dim>
12848 <dim>768</dim>
12849 </port>
12850 </output>
12851 </layer>
12852 <layer id="756" name="distilbert.transformer.layer.5.sa_layer_norm.bias" type="Const" version="opset1">
12853 <data element_type="f32" shape="768" offset="241884308" size="3072" />
12854 <rt_info>
12855 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.sa_layer_norm.bias" />
12856 </rt_info>
12857 <output>
12858 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.sa_layer_norm.bias">
12859 <dim>768</dim>
12860 </port>
12861 </output>
12862 </layer>
12863 <layer id="757" name="/distilbert/transformer/layer.5/sa_layer_norm/Add_1" type="Add" version="opset1">
12864 <data auto_broadcast="numpy" />
12865 <rt_info>
12866 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/sa_layer_norm/Add_1" />
12867 </rt_info>
12868 <input>
12869 <port id="0" precision="FP32">
12870 <dim>-1</dim>
12871 <dim>-1</dim>
12872 <dim>768</dim>
12873 </port>
12874 <port id="1" precision="FP32">
12875 <dim>768</dim>
12876 </port>
12877 </input>
12878 <output>
12879 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/sa_layer_norm/Add_1_output_0">
12880 <dim>-1</dim>
12881 <dim>-1</dim>
12882 <dim>768</dim>
12883 </port>
12884 </output>
12885 </layer>
12886 <layer id="758" name="onnx::MatMul_850" type="Const" version="opset1">
12887 <data element_type="f32" shape="768, 3072" offset="241887380" size="9437184" />
12888 <rt_info>
12889 <attribute name="fused_names" version="0" value="onnx::MatMul_850" />
12890 </rt_info>
12891 <output>
12892 <port id="0" precision="FP32" names="onnx::MatMul_850">
12893 <dim>768</dim>
12894 <dim>3072</dim>
12895 </port>
12896 </output>
12897 </layer>
12898 <layer id="759" name="/distilbert/transformer/layer.5/ffn/lin1/MatMul" type="MatMul" version="opset1">
12899 <data transpose_a="false" transpose_b="false" />
12900 <rt_info>
12901 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/lin1/MatMul" />
12902 </rt_info>
12903 <input>
12904 <port id="0" precision="FP32">
12905 <dim>-1</dim>
12906 <dim>-1</dim>
12907 <dim>768</dim>
12908 </port>
12909 <port id="1" precision="FP32">
12910 <dim>768</dim>
12911 <dim>3072</dim>
12912 </port>
12913 </input>
12914 <output>
12915 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/lin1/MatMul_output_0">
12916 <dim>-1</dim>
12917 <dim>-1</dim>
12918 <dim>3072</dim>
12919 </port>
12920 </output>
12921 </layer>
12922 <layer id="760" name="/distilbert/transformer/layer.5/ffn/lin1/Add" type="Add" version="opset1">
12923 <data auto_broadcast="numpy" />
12924 <rt_info>
12925 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/lin1/Add" />
12926 </rt_info>
12927 <input>
12928 <port id="0" precision="FP32">
12929 <dim>3072</dim>
12930 </port>
12931 <port id="1" precision="FP32">
12932 <dim>-1</dim>
12933 <dim>-1</dim>
12934 <dim>3072</dim>
12935 </port>
12936 </input>
12937 <output>
12938 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/lin1/Add_output_0">
12939 <dim>-1</dim>
12940 <dim>-1</dim>
12941 <dim>3072</dim>
12942 </port>
12943 </output>
12944 </layer>
12945 <layer id="761" name="/distilbert/transformer/layer.5/ffn/activation/Constant" type="Const" version="opset1">
12946 <data element_type="f32" shape="" offset="109659272" size="4" />
12947 <rt_info>
12948 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Constant" />
12949 </rt_info>
12950 <output>
12951 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Constant_output_0" />
12952 </output>
12953 </layer>
12954 <layer id="762" name="/distilbert/transformer/layer.5/ffn/activation/Div" type="Divide" version="opset1">
12955 <data auto_broadcast="numpy" m_pythondiv="true" />
12956 <rt_info>
12957 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Div" />
12958 </rt_info>
12959 <input>
12960 <port id="0" precision="FP32">
12961 <dim>-1</dim>
12962 <dim>-1</dim>
12963 <dim>3072</dim>
12964 </port>
12965 <port id="1" precision="FP32" />
12966 </input>
12967 <output>
12968 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Div_output_0">
12969 <dim>-1</dim>
12970 <dim>-1</dim>
12971 <dim>3072</dim>
12972 </port>
12973 </output>
12974 </layer>
12975 <layer id="763" name="/distilbert/transformer/layer.5/ffn/activation/Erf" type="Erf" version="opset1">
12976 <rt_info>
12977 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Erf" />
12978 </rt_info>
12979 <input>
12980 <port id="0" precision="FP32">
12981 <dim>-1</dim>
12982 <dim>-1</dim>
12983 <dim>3072</dim>
12984 </port>
12985 </input>
12986 <output>
12987 <port id="1" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Erf_output_0">
12988 <dim>-1</dim>
12989 <dim>-1</dim>
12990 <dim>3072</dim>
12991 </port>
12992 </output>
12993 </layer>
12994 <layer id="764" name="/distilbert/transformer/layer.5/ffn/activation/Constant_1" type="Const" version="opset1">
12995 <data element_type="f32" shape="" offset="109659276" size="4" />
12996 <rt_info>
12997 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Constant_1" />
12998 </rt_info>
12999 <output>
13000 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Constant_1_output_0" />
13001 </output>
13002 </layer>
13003 <layer id="765" name="/distilbert/transformer/layer.5/ffn/activation/Add" type="Add" version="opset1">
13004 <data auto_broadcast="numpy" />
13005 <rt_info>
13006 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Add" />
13007 </rt_info>
13008 <input>
13009 <port id="0" precision="FP32">
13010 <dim>-1</dim>
13011 <dim>-1</dim>
13012 <dim>3072</dim>
13013 </port>
13014 <port id="1" precision="FP32" />
13015 </input>
13016 <output>
13017 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Add_output_0">
13018 <dim>-1</dim>
13019 <dim>-1</dim>
13020 <dim>3072</dim>
13021 </port>
13022 </output>
13023 </layer>
13024 <layer id="766" name="/distilbert/transformer/layer.5/ffn/activation/Mul" type="Multiply" version="opset1">
13025 <data auto_broadcast="numpy" />
13026 <rt_info>
13027 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Mul" />
13028 </rt_info>
13029 <input>
13030 <port id="0" precision="FP32">
13031 <dim>-1</dim>
13032 <dim>-1</dim>
13033 <dim>3072</dim>
13034 </port>
13035 <port id="1" precision="FP32">
13036 <dim>-1</dim>
13037 <dim>-1</dim>
13038 <dim>3072</dim>
13039 </port>
13040 </input>
13041 <output>
13042 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Mul_output_0">
13043 <dim>-1</dim>
13044 <dim>-1</dim>
13045 <dim>3072</dim>
13046 </port>
13047 </output>
13048 </layer>
13049 <layer id="767" name="/distilbert/transformer/layer.5/ffn/activation/Constant_2" type="Const" version="opset1">
13050 <data element_type="f32" shape="" offset="109659280" size="4" />
13051 <rt_info>
13052 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Constant_2" />
13053 </rt_info>
13054 <output>
13055 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Constant_2_output_0" />
13056 </output>
13057 </layer>
13058 <layer id="768" name="/distilbert/transformer/layer.5/ffn/activation/Mul_1" type="Multiply" version="opset1">
13059 <data auto_broadcast="numpy" />
13060 <rt_info>
13061 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/activation/Mul_1" />
13062 </rt_info>
13063 <input>
13064 <port id="0" precision="FP32">
13065 <dim>-1</dim>
13066 <dim>-1</dim>
13067 <dim>3072</dim>
13068 </port>
13069 <port id="1" precision="FP32" />
13070 </input>
13071 <output>
13072 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/activation/Mul_1_output_0">
13073 <dim>-1</dim>
13074 <dim>-1</dim>
13075 <dim>3072</dim>
13076 </port>
13077 </output>
13078 </layer>
13079 <layer id="769" name="onnx::MatMul_851" type="Const" version="opset1">
13080 <data element_type="f32" shape="3072, 768" offset="251324564" size="9437184" />
13081 <rt_info>
13082 <attribute name="fused_names" version="0" value="onnx::MatMul_851" />
13083 </rt_info>
13084 <output>
13085 <port id="0" precision="FP32" names="onnx::MatMul_851">
13086 <dim>3072</dim>
13087 <dim>768</dim>
13088 </port>
13089 </output>
13090 </layer>
13091 <layer id="770" name="/distilbert/transformer/layer.5/ffn/lin2/MatMul" type="MatMul" version="opset1">
13092 <data transpose_a="false" transpose_b="false" />
13093 <rt_info>
13094 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/lin2/MatMul" />
13095 </rt_info>
13096 <input>
13097 <port id="0" precision="FP32">
13098 <dim>-1</dim>
13099 <dim>-1</dim>
13100 <dim>3072</dim>
13101 </port>
13102 <port id="1" precision="FP32">
13103 <dim>3072</dim>
13104 <dim>768</dim>
13105 </port>
13106 </input>
13107 <output>
13108 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/lin2/MatMul_output_0">
13109 <dim>-1</dim>
13110 <dim>-1</dim>
13111 <dim>768</dim>
13112 </port>
13113 </output>
13114 </layer>
13115 <layer id="771" name="/distilbert/transformer/layer.5/ffn/lin2/Add" type="Add" version="opset1">
13116 <data auto_broadcast="numpy" />
13117 <rt_info>
13118 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/ffn/lin2/Add" />
13119 </rt_info>
13120 <input>
13121 <port id="0" precision="FP32">
13122 <dim>768</dim>
13123 </port>
13124 <port id="1" precision="FP32">
13125 <dim>-1</dim>
13126 <dim>-1</dim>
13127 <dim>768</dim>
13128 </port>
13129 </input>
13130 <output>
13131 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/ffn/lin2/Add_output_0">
13132 <dim>-1</dim>
13133 <dim>-1</dim>
13134 <dim>768</dim>
13135 </port>
13136 </output>
13137 </layer>
13138 <layer id="772" name="/distilbert/transformer/layer.5/Add_1" type="Add" version="opset1">
13139 <data auto_broadcast="numpy" />
13140 <rt_info>
13141 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/Add_1" />
13142 </rt_info>
13143 <input>
13144 <port id="0" precision="FP32">
13145 <dim>-1</dim>
13146 <dim>-1</dim>
13147 <dim>768</dim>
13148 </port>
13149 <port id="1" precision="FP32">
13150 <dim>-1</dim>
13151 <dim>-1</dim>
13152 <dim>768</dim>
13153 </port>
13154 </input>
13155 <output>
13156 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/Add_1_output_0">
13157 <dim>-1</dim>
13158 <dim>-1</dim>
13159 <dim>768</dim>
13160 </port>
13161 </output>
13162 </layer>
13163 <layer id="773" name="Constant_256094" type="Const" version="opset1">
13164 <data element_type="i64" shape="1" offset="90763288" size="8" />
13165 <rt_info>
13166 <attribute name="fused_names" version="0" value="Constant_256094" />
13167 </rt_info>
13168 <output>
13169 <port id="0" precision="I64">
13170 <dim>1</dim>
13171 </port>
13172 </output>
13173 </layer>
13174 <layer id="774" name="/distilbert/transformer/layer.5/output_layer_norm/ReduceMean" type="ReduceMean" version="opset1">
13175 <data keep_dims="true" />
13176 <rt_info>
13177 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/ReduceMean" />
13178 </rt_info>
13179 <input>
13180 <port id="0" precision="FP32">
13181 <dim>-1</dim>
13182 <dim>-1</dim>
13183 <dim>768</dim>
13184 </port>
13185 <port id="1" precision="I64">
13186 <dim>1</dim>
13187 </port>
13188 </input>
13189 <output>
13190 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/ReduceMean_output_0">
13191 <dim>-1</dim>
13192 <dim>-1</dim>
13193 <dim>1</dim>
13194 </port>
13195 </output>
13196 </layer>
13197 <layer id="775" name="/distilbert/transformer/layer.5/output_layer_norm/Sub" type="Subtract" version="opset1">
13198 <data auto_broadcast="numpy" />
13199 <rt_info>
13200 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Sub" />
13201 </rt_info>
13202 <input>
13203 <port id="0" precision="FP32">
13204 <dim>-1</dim>
13205 <dim>-1</dim>
13206 <dim>768</dim>
13207 </port>
13208 <port id="1" precision="FP32">
13209 <dim>-1</dim>
13210 <dim>-1</dim>
13211 <dim>1</dim>
13212 </port>
13213 </input>
13214 <output>
13215 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Sub_output_0">
13216 <dim>-1</dim>
13217 <dim>-1</dim>
13218 <dim>768</dim>
13219 </port>
13220 </output>
13221 </layer>
13222 <layer id="776" name="/distilbert/transformer/layer.5/output_layer_norm/Constant" type="Const" version="opset1">
13223 <data element_type="f32" shape="" offset="90763296" size="4" />
13224 <rt_info>
13225 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Constant" />
13226 </rt_info>
13227 <output>
13228 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Constant_output_0" />
13229 </output>
13230 </layer>
13231 <layer id="777" name="/distilbert/transformer/layer.5/output_layer_norm/Pow" type="Power" version="opset1">
13232 <data auto_broadcast="numpy" />
13233 <rt_info>
13234 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Pow" />
13235 </rt_info>
13236 <input>
13237 <port id="0" precision="FP32">
13238 <dim>-1</dim>
13239 <dim>-1</dim>
13240 <dim>768</dim>
13241 </port>
13242 <port id="1" precision="FP32" />
13243 </input>
13244 <output>
13245 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Pow_output_0">
13246 <dim>-1</dim>
13247 <dim>-1</dim>
13248 <dim>768</dim>
13249 </port>
13250 </output>
13251 </layer>
13252 <layer id="778" name="Constant_256099" type="Const" version="opset1">
13253 <data element_type="i64" shape="1" offset="90763288" size="8" />
13254 <rt_info>
13255 <attribute name="fused_names" version="0" value="Constant_256099" />
13256 </rt_info>
13257 <output>
13258 <port id="0" precision="I64">
13259 <dim>1</dim>
13260 </port>
13261 </output>
13262 </layer>
13263 <layer id="779" name="/distilbert/transformer/layer.5/output_layer_norm/ReduceMean_1" type="ReduceMean" version="opset1">
13264 <data keep_dims="true" />
13265 <rt_info>
13266 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/ReduceMean_1" />
13267 </rt_info>
13268 <input>
13269 <port id="0" precision="FP32">
13270 <dim>-1</dim>
13271 <dim>-1</dim>
13272 <dim>768</dim>
13273 </port>
13274 <port id="1" precision="I64">
13275 <dim>1</dim>
13276 </port>
13277 </input>
13278 <output>
13279 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/ReduceMean_1_output_0">
13280 <dim>-1</dim>
13281 <dim>-1</dim>
13282 <dim>1</dim>
13283 </port>
13284 </output>
13285 </layer>
13286 <layer id="780" name="/distilbert/transformer/layer.5/output_layer_norm/Constant_1" type="Const" version="opset1">
13287 <data element_type="f32" shape="" offset="90763300" size="4" />
13288 <rt_info>
13289 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Constant_1" />
13290 </rt_info>
13291 <output>
13292 <port id="0" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Constant_1_output_0" />
13293 </output>
13294 </layer>
13295 <layer id="781" name="/distilbert/transformer/layer.5/output_layer_norm/Add" type="Add" version="opset1">
13296 <data auto_broadcast="numpy" />
13297 <rt_info>
13298 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Add" />
13299 </rt_info>
13300 <input>
13301 <port id="0" precision="FP32">
13302 <dim>-1</dim>
13303 <dim>-1</dim>
13304 <dim>1</dim>
13305 </port>
13306 <port id="1" precision="FP32" />
13307 </input>
13308 <output>
13309 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Add_output_0">
13310 <dim>-1</dim>
13311 <dim>-1</dim>
13312 <dim>1</dim>
13313 </port>
13314 </output>
13315 </layer>
13316 <layer id="782" name="/distilbert/transformer/layer.5/output_layer_norm/Sqrt" type="Sqrt" version="opset1">
13317 <rt_info>
13318 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Sqrt" />
13319 </rt_info>
13320 <input>
13321 <port id="0" precision="FP32">
13322 <dim>-1</dim>
13323 <dim>-1</dim>
13324 <dim>1</dim>
13325 </port>
13326 </input>
13327 <output>
13328 <port id="1" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Sqrt_output_0">
13329 <dim>-1</dim>
13330 <dim>-1</dim>
13331 <dim>1</dim>
13332 </port>
13333 </output>
13334 </layer>
13335 <layer id="783" name="/distilbert/transformer/layer.5/output_layer_norm/Div" type="Divide" version="opset1">
13336 <data auto_broadcast="numpy" m_pythondiv="true" />
13337 <rt_info>
13338 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Div" />
13339 </rt_info>
13340 <input>
13341 <port id="0" precision="FP32">
13342 <dim>-1</dim>
13343 <dim>-1</dim>
13344 <dim>768</dim>
13345 </port>
13346 <port id="1" precision="FP32">
13347 <dim>-1</dim>
13348 <dim>-1</dim>
13349 <dim>1</dim>
13350 </port>
13351 </input>
13352 <output>
13353 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Div_output_0">
13354 <dim>-1</dim>
13355 <dim>-1</dim>
13356 <dim>768</dim>
13357 </port>
13358 </output>
13359 </layer>
13360 <layer id="784" name="distilbert.transformer.layer.5.output_layer_norm.weight" type="Const" version="opset1">
13361 <data element_type="f32" shape="768" offset="260761748" size="3072" />
13362 <rt_info>
13363 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.output_layer_norm.weight" />
13364 </rt_info>
13365 <output>
13366 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.output_layer_norm.weight">
13367 <dim>768</dim>
13368 </port>
13369 </output>
13370 </layer>
13371 <layer id="785" name="/distilbert/transformer/layer.5/output_layer_norm/Mul" type="Multiply" version="opset1">
13372 <data auto_broadcast="numpy" />
13373 <rt_info>
13374 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Mul" />
13375 </rt_info>
13376 <input>
13377 <port id="0" precision="FP32">
13378 <dim>-1</dim>
13379 <dim>-1</dim>
13380 <dim>768</dim>
13381 </port>
13382 <port id="1" precision="FP32">
13383 <dim>768</dim>
13384 </port>
13385 </input>
13386 <output>
13387 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Mul_output_0">
13388 <dim>-1</dim>
13389 <dim>-1</dim>
13390 <dim>768</dim>
13391 </port>
13392 </output>
13393 </layer>
13394 <layer id="786" name="distilbert.transformer.layer.5.output_layer_norm.bias" type="Const" version="opset1">
13395 <data element_type="f32" shape="768" offset="260764820" size="3072" />
13396 <rt_info>
13397 <attribute name="fused_names" version="0" value="distilbert.transformer.layer.5.output_layer_norm.bias" />
13398 </rt_info>
13399 <output>
13400 <port id="0" precision="FP32" names="distilbert.transformer.layer.5.output_layer_norm.bias">
13401 <dim>768</dim>
13402 </port>
13403 </output>
13404 </layer>
13405 <layer id="787" name="/distilbert/transformer/layer.5/output_layer_norm/Add_1" type="Add" version="opset1">
13406 <data auto_broadcast="numpy" />
13407 <rt_info>
13408 <attribute name="fused_names" version="0" value="/distilbert/transformer/layer.5/output_layer_norm/Add_1" />
13409 </rt_info>
13410 <input>
13411 <port id="0" precision="FP32">
13412 <dim>-1</dim>
13413 <dim>-1</dim>
13414 <dim>768</dim>
13415 </port>
13416 <port id="1" precision="FP32">
13417 <dim>768</dim>
13418 </port>
13419 </input>
13420 <output>
13421 <port id="2" precision="FP32" names="/distilbert/transformer/layer.5/output_layer_norm/Add_1_output_0">
13422 <dim>-1</dim>
13423 <dim>-1</dim>
13424 <dim>768</dim>
13425 </port>
13426 </output>
13427 </layer>
13428 <layer id="788" name="onnx::MatMul_852" type="Const" version="opset1">
13429 <data element_type="f32" shape="768, 2" offset="260767892" size="6144" />
13430 <rt_info>
13431 <attribute name="fused_names" version="0" value="onnx::MatMul_852" />
13432 </rt_info>
13433 <output>
13434 <port id="0" precision="FP32" names="onnx::MatMul_852">
13435 <dim>768</dim>
13436 <dim>2</dim>
13437 </port>
13438 </output>
13439 </layer>
13440 <layer id="789" name="/qa_outputs/MatMul" type="MatMul" version="opset1">
13441 <data transpose_a="false" transpose_b="false" />
13442 <rt_info>
13443 <attribute name="fused_names" version="0" value="/qa_outputs/MatMul" />
13444 </rt_info>
13445 <input>
13446 <port id="0" precision="FP32">
13447 <dim>-1</dim>
13448 <dim>-1</dim>
13449 <dim>768</dim>
13450 </port>
13451 <port id="1" precision="FP32">
13452 <dim>768</dim>
13453 <dim>2</dim>
13454 </port>
13455 </input>
13456 <output>
13457 <port id="2" precision="FP32" names="/qa_outputs/MatMul_output_0">
13458 <dim>-1</dim>
13459 <dim>-1</dim>
13460 <dim>2</dim>
13461 </port>
13462 </output>
13463 </layer>
13464 <layer id="790" name="/qa_outputs/Add" type="Add" version="opset1">
13465 <data auto_broadcast="numpy" />
13466 <rt_info>
13467 <attribute name="fused_names" version="0" value="/qa_outputs/Add" />
13468 </rt_info>
13469 <input>
13470 <port id="0" precision="FP32">
13471 <dim>2</dim>
13472 </port>
13473 <port id="1" precision="FP32">
13474 <dim>-1</dim>
13475 <dim>-1</dim>
13476 <dim>2</dim>
13477 </port>
13478 </input>
13479 <output>
13480 <port id="2" precision="FP32" names="/qa_outputs/Add_output_0">
13481 <dim>-1</dim>
13482 <dim>-1</dim>
13483 <dim>2</dim>
13484 </port>
13485 </output>
13486 </layer>
13487 <layer id="791" name="Constant_256109" type="Const" version="opset1">
13488 <data element_type="i64" shape="" offset="90763288" size="8" />
13489 <rt_info>
13490 <attribute name="fused_names" version="0" value="Constant_256109" />
13491 </rt_info>
13492 <output>
13493 <port id="0" precision="I64" />
13494 </output>
13495 </layer>
13496 <layer id="792" name="Constant_256110" type="Const" version="opset1">
13497 <data element_type="i64" shape="2" offset="260774036" size="16" />
13498 <rt_info>
13499 <attribute name="fused_names" version="0" value="Constant_256110" />
13500 </rt_info>
13501 <output>
13502 <port id="0" precision="I64">
13503 <dim>2</dim>
13504 </port>
13505 </output>
13506 </layer>
13507 <layer id="793" name="/Split" type="VariadicSplit" version="opset1">
13508 <rt_info>
13509 <attribute name="fused_names" version="0" value="/Split" />
13510 </rt_info>
13511 <input>
13512 <port id="0" precision="FP32">
13513 <dim>-1</dim>
13514 <dim>-1</dim>
13515 <dim>2</dim>
13516 </port>
13517 <port id="1" precision="I64" />
13518 <port id="2" precision="I64">
13519 <dim>2</dim>
13520 </port>
13521 </input>
13522 <output>
13523 <port id="3" precision="FP32" names="/Split_output_0">
13524 <dim>-1</dim>
13525 <dim>-1</dim>
13526 <dim>1</dim>
13527 </port>
13528 <port id="4" precision="FP32" names="/Split_output_1">
13529 <dim>-1</dim>
13530 <dim>-1</dim>
13531 <dim>1</dim>
13532 </port>
13533 </output>
13534 </layer>
13535 <layer id="794" name="Constant_256114" type="Const" version="opset1">
13536 <data element_type="u64" shape="1" offset="260774052" size="8" />
13537 <rt_info>
13538 <attribute name="fused_names" version="0" value="Constant_256114" />
13539 </rt_info>
13540 <output>
13541 <port id="0" precision="U64">
13542 <dim>1</dim>
13543 </port>
13544 </output>
13545 </layer>
13546 <layer id="795" name="end_logits" type="Squeeze" version="opset1">
13547 <rt_info>
13548 <attribute name="fused_names" version="0" value="end_logits" />
13549 </rt_info>
13550 <input>
13551 <port id="0" precision="FP32">
13552 <dim>-1</dim>
13553 <dim>-1</dim>
13554 <dim>1</dim>
13555 </port>
13556 <port id="1" precision="U64">
13557 <dim>1</dim>
13558 </port>
13559 </input>
13560 <output>
13561 <port id="2" precision="FP32" names="end_logits">
13562 <dim>-1</dim>
13563 <dim>-1</dim>
13564 </port>
13565 </output>
13566 </layer>
13567 <layer id="797" name="Constant_256112" type="Const" version="opset1">
13568 <data element_type="u64" shape="1" offset="260774052" size="8" />
13569 <rt_info>
13570 <attribute name="fused_names" version="0" value="Constant_256112" />
13571 </rt_info>
13572 <output>
13573 <port id="0" precision="U64">
13574 <dim>1</dim>
13575 </port>
13576 </output>
13577 </layer>
13578 <layer id="798" name="start_logits" type="Squeeze" version="opset1">
13579 <rt_info>
13580 <attribute name="fused_names" version="0" value="start_logits" />
13581 </rt_info>
13582 <input>
13583 <port id="0" precision="FP32">
13584 <dim>-1</dim>
13585 <dim>-1</dim>
13586 <dim>1</dim>
13587 </port>
13588 <port id="1" precision="U64">
13589 <dim>1</dim>
13590 </port>
13591 </input>
13592 <output>
13593 <port id="2" precision="FP32" names="start_logits">
13594 <dim>-1</dim>
13595 <dim>-1</dim>
13596 </port>
13597 </output>
13598 </layer>
13599 <layer id="799" name="start_logits/sink_port_0" type="Result" version="opset1">
13600 <rt_info>
13601 <attribute name="fused_names" version="0" value="start_logits/sink_port_0" />
13602 </rt_info>
13603 <input>
13604 <port id="0" precision="FP32">
13605 <dim>-1</dim>
13606 <dim>-1</dim>
13607 </port>
13608 </input>
13609 </layer>
13610 <layer id="796" name="end_logits/sink_port_0" type="Result" version="opset1">
13611 <rt_info>
13612 <attribute name="fused_names" version="0" value="end_logits/sink_port_0" />
13613 </rt_info>
13614 <input>
13615 <port id="0" precision="FP32">
13616 <dim>-1</dim>
13617 <dim>-1</dim>
13618 </port>
13619 </input>
13620 </layer>
13621 </layers>
13622 <edges>
13623 <edge from-layer="0" from-port="0" to-layer="7" to-port="0" />
13624 <edge from-layer="1" from-port="0" to-layer="25" to-port="1" />
13625 <edge from-layer="2" from-port="0" to-layer="790" to-port="0" />
13626 <edge from-layer="3" from-port="0" to-layer="771" to-port="0" />
13627 <edge from-layer="4" from-port="0" to-layer="760" to-port="0" />
13628 <edge from-layer="5" from-port="0" to-layer="741" to-port="0" />
13629 <edge from-layer="6" from-port="0" to-layer="7" to-port="1" />
13630 <edge from-layer="7" from-port="2" to-layer="437" to-port="0" />
13631 <edge from-layer="7" from-port="2" to-layer="559" to-port="0" />
13632 <edge from-layer="7" from-port="2" to-layer="681" to-port="0" />
13633 <edge from-layer="7" from-port="2" to-layer="315" to-port="0" />
13634 <edge from-layer="7" from-port="2" to-layer="193" to-port="0" />
13635 <edge from-layer="7" from-port="2" to-layer="71" to-port="0" />
13636 <edge from-layer="8" from-port="0" to-layer="649" to-port="0" />
13637 <edge from-layer="9" from-port="0" to-layer="638" to-port="0" />
13638 <edge from-layer="10" from-port="0" to-layer="619" to-port="0" />
13639 <edge from-layer="11" from-port="0" to-layer="527" to-port="0" />
13640 <edge from-layer="12" from-port="0" to-layer="516" to-port="0" />
13641 <edge from-layer="13" from-port="0" to-layer="497" to-port="0" />
13642 <edge from-layer="14" from-port="0" to-layer="405" to-port="0" />
13643 <edge from-layer="15" from-port="0" to-layer="394" to-port="0" />
13644 <edge from-layer="16" from-port="0" to-layer="375" to-port="0" />
13645 <edge from-layer="17" from-port="0" to-layer="283" to-port="0" />
13646 <edge from-layer="18" from-port="0" to-layer="272" to-port="0" />
13647 <edge from-layer="19" from-port="0" to-layer="253" to-port="0" />
13648 <edge from-layer="20" from-port="0" to-layer="161" to-port="0" />
13649 <edge from-layer="21" from-port="0" to-layer="150" to-port="0" />
13650 <edge from-layer="22" from-port="0" to-layer="131" to-port="0" />
13651 <edge from-layer="23" from-port="0" to-layer="25" to-port="0" />
13652 <edge from-layer="24" from-port="0" to-layer="25" to-port="2" />
13653 <edge from-layer="25" from-port="3" to-layer="29" to-port="0" />
13654 <edge from-layer="25" from-port="3" to-layer="40" to-port="0" />
13655 <edge from-layer="26" from-port="0" to-layer="39" to-port="0" />
13656 <edge from-layer="27" from-port="0" to-layer="37" to-port="0" />
13657 <edge from-layer="28" from-port="0" to-layer="37" to-port="1" />
13658 <edge from-layer="29" from-port="1" to-layer="32" to-port="0" />
13659 <edge from-layer="30" from-port="0" to-layer="32" to-port="1" />
13660 <edge from-layer="31" from-port="0" to-layer="32" to-port="2" />
13661 <edge from-layer="32" from-port="3" to-layer="34" to-port="0" />
13662 <edge from-layer="33" from-port="0" to-layer="34" to-port="1" />
13663 <edge from-layer="34" from-port="2" to-layer="37" to-port="2" />
13664 <edge from-layer="35" from-port="0" to-layer="37" to-port="3" />
13665 <edge from-layer="36" from-port="0" to-layer="37" to-port="4" />
13666 <edge from-layer="37" from-port="5" to-layer="39" to-port="1" />
13667 <edge from-layer="38" from-port="0" to-layer="39" to-port="2" />
13668 <edge from-layer="39" from-port="3" to-layer="40" to-port="1" />
13669 <edge from-layer="40" from-port="2" to-layer="42" to-port="0" />
13670 <edge from-layer="40" from-port="2" to-layer="43" to-port="0" />
13671 <edge from-layer="41" from-port="0" to-layer="42" to-port="1" />
13672 <edge from-layer="42" from-port="2" to-layer="43" to-port="1" />
13673 <edge from-layer="43" from-port="2" to-layer="51" to-port="0" />
13674 <edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
13675 <edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
13676 <edge from-layer="45" from-port="2" to-layer="47" to-port="0" />
13677 <edge from-layer="46" from-port="0" to-layer="47" to-port="1" />
13678 <edge from-layer="47" from-port="2" to-layer="49" to-port="0" />
13679 <edge from-layer="48" from-port="0" to-layer="49" to-port="1" />
13680 <edge from-layer="49" from-port="2" to-layer="50" to-port="0" />
13681 <edge from-layer="50" from-port="1" to-layer="51" to-port="1" />
13682 <edge from-layer="51" from-port="2" to-layer="53" to-port="0" />
13683 <edge from-layer="52" from-port="0" to-layer="53" to-port="1" />
13684 <edge from-layer="53" from-port="2" to-layer="55" to-port="0" />
13685 <edge from-layer="54" from-port="0" to-layer="55" to-port="1" />
13686 <edge from-layer="55" from-port="2" to-layer="132" to-port="1" />
13687 <edge from-layer="55" from-port="2" to-layer="89" to-port="0" />
13688 <edge from-layer="55" from-port="2" to-layer="109" to-port="0" />
13689 <edge from-layer="55" from-port="2" to-layer="56" to-port="0" />
13690 <edge from-layer="55" from-port="2" to-layer="74" to-port="0" />
13691 <edge from-layer="55" from-port="2" to-layer="64" to-port="0" />
13692 <edge from-layer="56" from-port="1" to-layer="59" to-port="0" />
13693 <edge from-layer="57" from-port="0" to-layer="59" to-port="1" />
13694 <edge from-layer="58" from-port="0" to-layer="59" to-port="2" />
13695 <edge from-layer="59" from-port="3" to-layer="61" to-port="0" />
13696 <edge from-layer="59" from-port="3" to-layer="77" to-port="0" />
13697 <edge from-layer="59" from-port="3" to-layer="124" to-port="0" />
13698 <edge from-layer="59" from-port="3" to-layer="112" to-port="0" />
13699 <edge from-layer="59" from-port="3" to-layer="92" to-port="0" />
13700 <edge from-layer="60" from-port="0" to-layer="61" to-port="1" />
13701 <edge from-layer="61" from-port="2" to-layer="70" to-port="0" />
13702 <edge from-layer="62" from-port="0" to-layer="70" to-port="1" />
13703 <edge from-layer="63" from-port="0" to-layer="70" to-port="2" />
13704 <edge from-layer="64" from-port="1" to-layer="67" to-port="0" />
13705 <edge from-layer="65" from-port="0" to-layer="67" to-port="1" />
13706 <edge from-layer="66" from-port="0" to-layer="67" to-port="2" />
13707 <edge from-layer="67" from-port="3" to-layer="69" to-port="0" />
13708 <edge from-layer="68" from-port="0" to-layer="69" to-port="1" />
13709 <edge from-layer="69" from-port="2" to-layer="70" to-port="3" />
13710 <edge from-layer="70" from-port="4" to-layer="71" to-port="1" />
13711 <edge from-layer="71" from-port="2" to-layer="102" to-port="0" />
13712 <edge from-layer="72" from-port="0" to-layer="75" to-port="0" />
13713 <edge from-layer="73" from-port="0" to-layer="74" to-port="1" />
13714 <edge from-layer="74" from-port="2" to-layer="75" to-port="1" />
13715 <edge from-layer="75" from-port="2" to-layer="82" to-port="0" />
13716 <edge from-layer="76" from-port="0" to-layer="77" to-port="1" />
13717 <edge from-layer="77" from-port="2" to-layer="81" to-port="0" />
13718 <edge from-layer="78" from-port="0" to-layer="81" to-port="1" />
13719 <edge from-layer="79" from-port="0" to-layer="81" to-port="2" />
13720 <edge from-layer="80" from-port="0" to-layer="81" to-port="3" />
13721 <edge from-layer="81" from-port="4" to-layer="82" to-port="1" />
13722 <edge from-layer="82" from-port="2" to-layer="84" to-port="0" />
13723 <edge from-layer="83" from-port="0" to-layer="84" to-port="1" />
13724 <edge from-layer="84" from-port="2" to-layer="86" to-port="0" />
13725 <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
13726 <edge from-layer="86" from-port="2" to-layer="100" to-port="0" />
13727 <edge from-layer="87" from-port="0" to-layer="90" to-port="0" />
13728 <edge from-layer="88" from-port="0" to-layer="89" to-port="1" />
13729 <edge from-layer="89" from-port="2" to-layer="90" to-port="1" />
13730 <edge from-layer="90" from-port="2" to-layer="97" to-port="0" />
13731 <edge from-layer="91" from-port="0" to-layer="92" to-port="1" />
13732 <edge from-layer="92" from-port="2" to-layer="96" to-port="0" />
13733 <edge from-layer="93" from-port="0" to-layer="96" to-port="1" />
13734 <edge from-layer="94" from-port="0" to-layer="96" to-port="2" />
13735 <edge from-layer="95" from-port="0" to-layer="96" to-port="3" />
13736 <edge from-layer="96" from-port="4" to-layer="97" to-port="1" />
13737 <edge from-layer="97" from-port="2" to-layer="99" to-port="0" />
13738 <edge from-layer="98" from-port="0" to-layer="99" to-port="1" />
13739 <edge from-layer="99" from-port="2" to-layer="100" to-port="1" />
13740 <edge from-layer="100" from-port="2" to-layer="105" to-port="2" />
13741 <edge from-layer="100" from-port="2" to-layer="101" to-port="0" />
13742 <edge from-layer="101" from-port="1" to-layer="102" to-port="1" />
13743 <edge from-layer="102" from-port="2" to-layer="103" to-port="0" />
13744 <edge from-layer="103" from-port="1" to-layer="105" to-port="0" />
13745 <edge from-layer="104" from-port="0" to-layer="105" to-port="1" />
13746 <edge from-layer="105" from-port="3" to-layer="106" to-port="0" />
13747 <edge from-layer="106" from-port="1" to-layer="120" to-port="0" />
13748 <edge from-layer="107" from-port="0" to-layer="110" to-port="0" />
13749 <edge from-layer="108" from-port="0" to-layer="109" to-port="1" />
13750 <edge from-layer="109" from-port="2" to-layer="110" to-port="1" />
13751 <edge from-layer="110" from-port="2" to-layer="117" to-port="0" />
13752 <edge from-layer="111" from-port="0" to-layer="112" to-port="1" />
13753 <edge from-layer="112" from-port="2" to-layer="116" to-port="0" />
13754 <edge from-layer="113" from-port="0" to-layer="116" to-port="1" />
13755 <edge from-layer="114" from-port="0" to-layer="116" to-port="2" />
13756 <edge from-layer="115" from-port="0" to-layer="116" to-port="3" />
13757 <edge from-layer="116" from-port="4" to-layer="117" to-port="1" />
13758 <edge from-layer="117" from-port="2" to-layer="119" to-port="0" />
13759 <edge from-layer="118" from-port="0" to-layer="119" to-port="1" />
13760 <edge from-layer="119" from-port="2" to-layer="120" to-port="1" />
13761 <edge from-layer="120" from-port="2" to-layer="122" to-port="0" />
13762 <edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
13763 <edge from-layer="122" from-port="2" to-layer="128" to-port="0" />
13764 <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
13765 <edge from-layer="124" from-port="2" to-layer="127" to-port="0" />
13766 <edge from-layer="125" from-port="0" to-layer="127" to-port="1" />
13767 <edge from-layer="126" from-port="0" to-layer="127" to-port="2" />
13768 <edge from-layer="127" from-port="3" to-layer="128" to-port="1" />
13769 <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
13770 <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
13771 <edge from-layer="130" from-port="2" to-layer="131" to-port="1" />
13772 <edge from-layer="131" from-port="2" to-layer="132" to-port="0" />
13773 <edge from-layer="132" from-port="2" to-layer="134" to-port="0" />
13774 <edge from-layer="132" from-port="2" to-layer="135" to-port="0" />
13775 <edge from-layer="133" from-port="0" to-layer="134" to-port="1" />
13776 <edge from-layer="134" from-port="2" to-layer="135" to-port="1" />
13777 <edge from-layer="135" from-port="2" to-layer="143" to-port="0" />
13778 <edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
13779 <edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
13780 <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
13781 <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
13782 <edge from-layer="139" from-port="2" to-layer="141" to-port="0" />
13783 <edge from-layer="140" from-port="0" to-layer="141" to-port="1" />
13784 <edge from-layer="141" from-port="2" to-layer="142" to-port="0" />
13785 <edge from-layer="142" from-port="1" to-layer="143" to-port="1" />
13786 <edge from-layer="143" from-port="2" to-layer="145" to-port="0" />
13787 <edge from-layer="144" from-port="0" to-layer="145" to-port="1" />
13788 <edge from-layer="145" from-port="2" to-layer="147" to-port="0" />
13789 <edge from-layer="146" from-port="0" to-layer="147" to-port="1" />
13790 <edge from-layer="147" from-port="2" to-layer="162" to-port="1" />
13791 <edge from-layer="147" from-port="2" to-layer="149" to-port="0" />
13792 <edge from-layer="148" from-port="0" to-layer="149" to-port="1" />
13793 <edge from-layer="149" from-port="2" to-layer="150" to-port="1" />
13794 <edge from-layer="150" from-port="2" to-layer="156" to-port="0" />
13795 <edge from-layer="150" from-port="2" to-layer="152" to-port="0" />
13796 <edge from-layer="151" from-port="0" to-layer="152" to-port="1" />
13797 <edge from-layer="152" from-port="2" to-layer="153" to-port="0" />
13798 <edge from-layer="153" from-port="1" to-layer="155" to-port="0" />
13799 <edge from-layer="154" from-port="0" to-layer="155" to-port="1" />
13800 <edge from-layer="155" from-port="2" to-layer="156" to-port="1" />
13801 <edge from-layer="156" from-port="2" to-layer="158" to-port="0" />
13802 <edge from-layer="157" from-port="0" to-layer="158" to-port="1" />
13803 <edge from-layer="158" from-port="2" to-layer="160" to-port="0" />
13804 <edge from-layer="159" from-port="0" to-layer="160" to-port="1" />
13805 <edge from-layer="160" from-port="2" to-layer="161" to-port="1" />
13806 <edge from-layer="161" from-port="2" to-layer="162" to-port="0" />
13807 <edge from-layer="162" from-port="2" to-layer="164" to-port="0" />
13808 <edge from-layer="162" from-port="2" to-layer="165" to-port="0" />
13809 <edge from-layer="163" from-port="0" to-layer="164" to-port="1" />
13810 <edge from-layer="164" from-port="2" to-layer="165" to-port="1" />
13811 <edge from-layer="165" from-port="2" to-layer="167" to-port="0" />
13812 <edge from-layer="165" from-port="2" to-layer="173" to-port="0" />
13813 <edge from-layer="166" from-port="0" to-layer="167" to-port="1" />
13814 <edge from-layer="167" from-port="2" to-layer="169" to-port="0" />
13815 <edge from-layer="168" from-port="0" to-layer="169" to-port="1" />
13816 <edge from-layer="169" from-port="2" to-layer="171" to-port="0" />
13817 <edge from-layer="170" from-port="0" to-layer="171" to-port="1" />
13818 <edge from-layer="171" from-port="2" to-layer="172" to-port="0" />
13819 <edge from-layer="172" from-port="1" to-layer="173" to-port="1" />
13820 <edge from-layer="173" from-port="2" to-layer="175" to-port="0" />
13821 <edge from-layer="174" from-port="0" to-layer="175" to-port="1" />
13822 <edge from-layer="175" from-port="2" to-layer="177" to-port="0" />
13823 <edge from-layer="176" from-port="0" to-layer="177" to-port="1" />
13824 <edge from-layer="177" from-port="2" to-layer="231" to-port="0" />
13825 <edge from-layer="177" from-port="2" to-layer="211" to-port="0" />
13826 <edge from-layer="177" from-port="2" to-layer="196" to-port="0" />
13827 <edge from-layer="177" from-port="2" to-layer="186" to-port="0" />
13828 <edge from-layer="177" from-port="2" to-layer="254" to-port="1" />
13829 <edge from-layer="177" from-port="2" to-layer="178" to-port="0" />
13830 <edge from-layer="178" from-port="1" to-layer="181" to-port="0" />
13831 <edge from-layer="179" from-port="0" to-layer="181" to-port="1" />
13832 <edge from-layer="180" from-port="0" to-layer="181" to-port="2" />
13833 <edge from-layer="181" from-port="3" to-layer="246" to-port="0" />
13834 <edge from-layer="181" from-port="3" to-layer="234" to-port="0" />
13835 <edge from-layer="181" from-port="3" to-layer="214" to-port="0" />
13836 <edge from-layer="181" from-port="3" to-layer="183" to-port="0" />
13837 <edge from-layer="181" from-port="3" to-layer="199" to-port="0" />
13838 <edge from-layer="182" from-port="0" to-layer="183" to-port="1" />
13839 <edge from-layer="183" from-port="2" to-layer="192" to-port="0" />
13840 <edge from-layer="184" from-port="0" to-layer="192" to-port="1" />
13841 <edge from-layer="185" from-port="0" to-layer="192" to-port="2" />
13842 <edge from-layer="186" from-port="1" to-layer="189" to-port="0" />
13843 <edge from-layer="187" from-port="0" to-layer="189" to-port="1" />
13844 <edge from-layer="188" from-port="0" to-layer="189" to-port="2" />
13845 <edge from-layer="189" from-port="3" to-layer="191" to-port="0" />
13846 <edge from-layer="190" from-port="0" to-layer="191" to-port="1" />
13847 <edge from-layer="191" from-port="2" to-layer="192" to-port="3" />
13848 <edge from-layer="192" from-port="4" to-layer="193" to-port="1" />
13849 <edge from-layer="193" from-port="2" to-layer="224" to-port="0" />
13850 <edge from-layer="194" from-port="0" to-layer="197" to-port="0" />
13851 <edge from-layer="195" from-port="0" to-layer="196" to-port="1" />
13852 <edge from-layer="196" from-port="2" to-layer="197" to-port="1" />
13853 <edge from-layer="197" from-port="2" to-layer="204" to-port="0" />
13854 <edge from-layer="198" from-port="0" to-layer="199" to-port="1" />
13855 <edge from-layer="199" from-port="2" to-layer="203" to-port="0" />
13856 <edge from-layer="200" from-port="0" to-layer="203" to-port="1" />
13857 <edge from-layer="201" from-port="0" to-layer="203" to-port="2" />
13858 <edge from-layer="202" from-port="0" to-layer="203" to-port="3" />
13859 <edge from-layer="203" from-port="4" to-layer="204" to-port="1" />
13860 <edge from-layer="204" from-port="2" to-layer="206" to-port="0" />
13861 <edge from-layer="205" from-port="0" to-layer="206" to-port="1" />
13862 <edge from-layer="206" from-port="2" to-layer="208" to-port="0" />
13863 <edge from-layer="207" from-port="0" to-layer="208" to-port="1" />
13864 <edge from-layer="208" from-port="2" to-layer="222" to-port="0" />
13865 <edge from-layer="209" from-port="0" to-layer="212" to-port="0" />
13866 <edge from-layer="210" from-port="0" to-layer="211" to-port="1" />
13867 <edge from-layer="211" from-port="2" to-layer="212" to-port="1" />
13868 <edge from-layer="212" from-port="2" to-layer="219" to-port="0" />
13869 <edge from-layer="213" from-port="0" to-layer="214" to-port="1" />
13870 <edge from-layer="214" from-port="2" to-layer="218" to-port="0" />
13871 <edge from-layer="215" from-port="0" to-layer="218" to-port="1" />
13872 <edge from-layer="216" from-port="0" to-layer="218" to-port="2" />
13873 <edge from-layer="217" from-port="0" to-layer="218" to-port="3" />
13874 <edge from-layer="218" from-port="4" to-layer="219" to-port="1" />
13875 <edge from-layer="219" from-port="2" to-layer="221" to-port="0" />
13876 <edge from-layer="220" from-port="0" to-layer="221" to-port="1" />
13877 <edge from-layer="221" from-port="2" to-layer="222" to-port="1" />
13878 <edge from-layer="222" from-port="2" to-layer="223" to-port="0" />
13879 <edge from-layer="222" from-port="2" to-layer="227" to-port="2" />
13880 <edge from-layer="223" from-port="1" to-layer="224" to-port="1" />
13881 <edge from-layer="224" from-port="2" to-layer="225" to-port="0" />
13882 <edge from-layer="225" from-port="1" to-layer="227" to-port="0" />
13883 <edge from-layer="226" from-port="0" to-layer="227" to-port="1" />
13884 <edge from-layer="227" from-port="3" to-layer="228" to-port="0" />
13885 <edge from-layer="228" from-port="1" to-layer="242" to-port="0" />
13886 <edge from-layer="229" from-port="0" to-layer="232" to-port="0" />
13887 <edge from-layer="230" from-port="0" to-layer="231" to-port="1" />
13888 <edge from-layer="231" from-port="2" to-layer="232" to-port="1" />
13889 <edge from-layer="232" from-port="2" to-layer="239" to-port="0" />
13890 <edge from-layer="233" from-port="0" to-layer="234" to-port="1" />
13891 <edge from-layer="234" from-port="2" to-layer="238" to-port="0" />
13892 <edge from-layer="235" from-port="0" to-layer="238" to-port="1" />
13893 <edge from-layer="236" from-port="0" to-layer="238" to-port="2" />
13894 <edge from-layer="237" from-port="0" to-layer="238" to-port="3" />
13895 <edge from-layer="238" from-port="4" to-layer="239" to-port="1" />
13896 <edge from-layer="239" from-port="2" to-layer="241" to-port="0" />
13897 <edge from-layer="240" from-port="0" to-layer="241" to-port="1" />
13898 <edge from-layer="241" from-port="2" to-layer="242" to-port="1" />
13899 <edge from-layer="242" from-port="2" to-layer="244" to-port="0" />
13900 <edge from-layer="243" from-port="0" to-layer="244" to-port="1" />
13901 <edge from-layer="244" from-port="2" to-layer="250" to-port="0" />
13902 <edge from-layer="245" from-port="0" to-layer="246" to-port="1" />
13903 <edge from-layer="246" from-port="2" to-layer="249" to-port="0" />
13904 <edge from-layer="247" from-port="0" to-layer="249" to-port="1" />
13905 <edge from-layer="248" from-port="0" to-layer="249" to-port="2" />
13906 <edge from-layer="249" from-port="3" to-layer="250" to-port="1" />
13907 <edge from-layer="250" from-port="2" to-layer="252" to-port="0" />
13908 <edge from-layer="251" from-port="0" to-layer="252" to-port="1" />
13909 <edge from-layer="252" from-port="2" to-layer="253" to-port="1" />
13910 <edge from-layer="253" from-port="2" to-layer="254" to-port="0" />
13911 <edge from-layer="254" from-port="2" to-layer="256" to-port="0" />
13912 <edge from-layer="254" from-port="2" to-layer="257" to-port="0" />
13913 <edge from-layer="255" from-port="0" to-layer="256" to-port="1" />
13914 <edge from-layer="256" from-port="2" to-layer="257" to-port="1" />
13915 <edge from-layer="257" from-port="2" to-layer="259" to-port="0" />
13916 <edge from-layer="257" from-port="2" to-layer="265" to-port="0" />
13917 <edge from-layer="258" from-port="0" to-layer="259" to-port="1" />
13918 <edge from-layer="259" from-port="2" to-layer="261" to-port="0" />
13919 <edge from-layer="260" from-port="0" to-layer="261" to-port="1" />
13920 <edge from-layer="261" from-port="2" to-layer="263" to-port="0" />
13921 <edge from-layer="262" from-port="0" to-layer="263" to-port="1" />
13922 <edge from-layer="263" from-port="2" to-layer="264" to-port="0" />
13923 <edge from-layer="264" from-port="1" to-layer="265" to-port="1" />
13924 <edge from-layer="265" from-port="2" to-layer="267" to-port="0" />
13925 <edge from-layer="266" from-port="0" to-layer="267" to-port="1" />
13926 <edge from-layer="267" from-port="2" to-layer="269" to-port="0" />
13927 <edge from-layer="268" from-port="0" to-layer="269" to-port="1" />
13928 <edge from-layer="269" from-port="2" to-layer="271" to-port="0" />
13929 <edge from-layer="269" from-port="2" to-layer="284" to-port="1" />
13930 <edge from-layer="270" from-port="0" to-layer="271" to-port="1" />
13931 <edge from-layer="271" from-port="2" to-layer="272" to-port="1" />
13932 <edge from-layer="272" from-port="2" to-layer="274" to-port="0" />
13933 <edge from-layer="272" from-port="2" to-layer="278" to-port="0" />
13934 <edge from-layer="273" from-port="0" to-layer="274" to-port="1" />
13935 <edge from-layer="274" from-port="2" to-layer="275" to-port="0" />
13936 <edge from-layer="275" from-port="1" to-layer="277" to-port="0" />
13937 <edge from-layer="276" from-port="0" to-layer="277" to-port="1" />
13938 <edge from-layer="277" from-port="2" to-layer="278" to-port="1" />
13939 <edge from-layer="278" from-port="2" to-layer="280" to-port="0" />
13940 <edge from-layer="279" from-port="0" to-layer="280" to-port="1" />
13941 <edge from-layer="280" from-port="2" to-layer="282" to-port="0" />
13942 <edge from-layer="281" from-port="0" to-layer="282" to-port="1" />
13943 <edge from-layer="282" from-port="2" to-layer="283" to-port="1" />
13944 <edge from-layer="283" from-port="2" to-layer="284" to-port="0" />
13945 <edge from-layer="284" from-port="2" to-layer="287" to-port="0" />
13946 <edge from-layer="284" from-port="2" to-layer="286" to-port="0" />
13947 <edge from-layer="285" from-port="0" to-layer="286" to-port="1" />
13948 <edge from-layer="286" from-port="2" to-layer="287" to-port="1" />
13949 <edge from-layer="287" from-port="2" to-layer="289" to-port="0" />
13950 <edge from-layer="287" from-port="2" to-layer="295" to-port="0" />
13951 <edge from-layer="288" from-port="0" to-layer="289" to-port="1" />
13952 <edge from-layer="289" from-port="2" to-layer="291" to-port="0" />
13953 <edge from-layer="290" from-port="0" to-layer="291" to-port="1" />
13954 <edge from-layer="291" from-port="2" to-layer="293" to-port="0" />
13955 <edge from-layer="292" from-port="0" to-layer="293" to-port="1" />
13956 <edge from-layer="293" from-port="2" to-layer="294" to-port="0" />
13957 <edge from-layer="294" from-port="1" to-layer="295" to-port="1" />
13958 <edge from-layer="295" from-port="2" to-layer="297" to-port="0" />
13959 <edge from-layer="296" from-port="0" to-layer="297" to-port="1" />
13960 <edge from-layer="297" from-port="2" to-layer="299" to-port="0" />
13961 <edge from-layer="298" from-port="0" to-layer="299" to-port="1" />
13962 <edge from-layer="299" from-port="2" to-layer="333" to-port="0" />
13963 <edge from-layer="299" from-port="2" to-layer="318" to-port="0" />
13964 <edge from-layer="299" from-port="2" to-layer="353" to-port="0" />
13965 <edge from-layer="299" from-port="2" to-layer="300" to-port="0" />
13966 <edge from-layer="299" from-port="2" to-layer="308" to-port="0" />
13967 <edge from-layer="299" from-port="2" to-layer="376" to-port="1" />
13968 <edge from-layer="300" from-port="1" to-layer="303" to-port="0" />
13969 <edge from-layer="301" from-port="0" to-layer="303" to-port="1" />
13970 <edge from-layer="302" from-port="0" to-layer="303" to-port="2" />
13971 <edge from-layer="303" from-port="3" to-layer="356" to-port="0" />
13972 <edge from-layer="303" from-port="3" to-layer="368" to-port="0" />
13973 <edge from-layer="303" from-port="3" to-layer="336" to-port="0" />
13974 <edge from-layer="303" from-port="3" to-layer="321" to-port="0" />
13975 <edge from-layer="303" from-port="3" to-layer="305" to-port="0" />
13976 <edge from-layer="304" from-port="0" to-layer="305" to-port="1" />
13977 <edge from-layer="305" from-port="2" to-layer="314" to-port="0" />
13978 <edge from-layer="306" from-port="0" to-layer="314" to-port="1" />
13979 <edge from-layer="307" from-port="0" to-layer="314" to-port="2" />
13980 <edge from-layer="308" from-port="1" to-layer="311" to-port="0" />
13981 <edge from-layer="309" from-port="0" to-layer="311" to-port="1" />
13982 <edge from-layer="310" from-port="0" to-layer="311" to-port="2" />
13983 <edge from-layer="311" from-port="3" to-layer="313" to-port="0" />
13984 <edge from-layer="312" from-port="0" to-layer="313" to-port="1" />
13985 <edge from-layer="313" from-port="2" to-layer="314" to-port="3" />
13986 <edge from-layer="314" from-port="4" to-layer="315" to-port="1" />
13987 <edge from-layer="315" from-port="2" to-layer="346" to-port="0" />
13988 <edge from-layer="316" from-port="0" to-layer="319" to-port="0" />
13989 <edge from-layer="317" from-port="0" to-layer="318" to-port="1" />
13990 <edge from-layer="318" from-port="2" to-layer="319" to-port="1" />
13991 <edge from-layer="319" from-port="2" to-layer="326" to-port="0" />
13992 <edge from-layer="320" from-port="0" to-layer="321" to-port="1" />
13993 <edge from-layer="321" from-port="2" to-layer="325" to-port="0" />
13994 <edge from-layer="322" from-port="0" to-layer="325" to-port="1" />
13995 <edge from-layer="323" from-port="0" to-layer="325" to-port="2" />
13996 <edge from-layer="324" from-port="0" to-layer="325" to-port="3" />
13997 <edge from-layer="325" from-port="4" to-layer="326" to-port="1" />
13998 <edge from-layer="326" from-port="2" to-layer="328" to-port="0" />
13999 <edge from-layer="327" from-port="0" to-layer="328" to-port="1" />
14000 <edge from-layer="328" from-port="2" to-layer="330" to-port="0" />
14001 <edge from-layer="329" from-port="0" to-layer="330" to-port="1" />
14002 <edge from-layer="330" from-port="2" to-layer="344" to-port="0" />
14003 <edge from-layer="331" from-port="0" to-layer="334" to-port="0" />
14004 <edge from-layer="332" from-port="0" to-layer="333" to-port="1" />
14005 <edge from-layer="333" from-port="2" to-layer="334" to-port="1" />
14006 <edge from-layer="334" from-port="2" to-layer="341" to-port="0" />
14007 <edge from-layer="335" from-port="0" to-layer="336" to-port="1" />
14008 <edge from-layer="336" from-port="2" to-layer="340" to-port="0" />
14009 <edge from-layer="337" from-port="0" to-layer="340" to-port="1" />
14010 <edge from-layer="338" from-port="0" to-layer="340" to-port="2" />
14011 <edge from-layer="339" from-port="0" to-layer="340" to-port="3" />
14012 <edge from-layer="340" from-port="4" to-layer="341" to-port="1" />
14013 <edge from-layer="341" from-port="2" to-layer="343" to-port="0" />
14014 <edge from-layer="342" from-port="0" to-layer="343" to-port="1" />
14015 <edge from-layer="343" from-port="2" to-layer="344" to-port="1" />
14016 <edge from-layer="344" from-port="2" to-layer="349" to-port="2" />
14017 <edge from-layer="344" from-port="2" to-layer="345" to-port="0" />
14018 <edge from-layer="345" from-port="1" to-layer="346" to-port="1" />
14019 <edge from-layer="346" from-port="2" to-layer="347" to-port="0" />
14020 <edge from-layer="347" from-port="1" to-layer="349" to-port="0" />
14021 <edge from-layer="348" from-port="0" to-layer="349" to-port="1" />
14022 <edge from-layer="349" from-port="3" to-layer="350" to-port="0" />
14023 <edge from-layer="350" from-port="1" to-layer="364" to-port="0" />
14024 <edge from-layer="351" from-port="0" to-layer="354" to-port="0" />
14025 <edge from-layer="352" from-port="0" to-layer="353" to-port="1" />
14026 <edge from-layer="353" from-port="2" to-layer="354" to-port="1" />
14027 <edge from-layer="354" from-port="2" to-layer="361" to-port="0" />
14028 <edge from-layer="355" from-port="0" to-layer="356" to-port="1" />
14029 <edge from-layer="356" from-port="2" to-layer="360" to-port="0" />
14030 <edge from-layer="357" from-port="0" to-layer="360" to-port="1" />
14031 <edge from-layer="358" from-port="0" to-layer="360" to-port="2" />
14032 <edge from-layer="359" from-port="0" to-layer="360" to-port="3" />
14033 <edge from-layer="360" from-port="4" to-layer="361" to-port="1" />
14034 <edge from-layer="361" from-port="2" to-layer="363" to-port="0" />
14035 <edge from-layer="362" from-port="0" to-layer="363" to-port="1" />
14036 <edge from-layer="363" from-port="2" to-layer="364" to-port="1" />
14037 <edge from-layer="364" from-port="2" to-layer="366" to-port="0" />
14038 <edge from-layer="365" from-port="0" to-layer="366" to-port="1" />
14039 <edge from-layer="366" from-port="2" to-layer="372" to-port="0" />
14040 <edge from-layer="367" from-port="0" to-layer="368" to-port="1" />
14041 <edge from-layer="368" from-port="2" to-layer="371" to-port="0" />
14042 <edge from-layer="369" from-port="0" to-layer="371" to-port="1" />
14043 <edge from-layer="370" from-port="0" to-layer="371" to-port="2" />
14044 <edge from-layer="371" from-port="3" to-layer="372" to-port="1" />
14045 <edge from-layer="372" from-port="2" to-layer="374" to-port="0" />
14046 <edge from-layer="373" from-port="0" to-layer="374" to-port="1" />
14047 <edge from-layer="374" from-port="2" to-layer="375" to-port="1" />
14048 <edge from-layer="375" from-port="2" to-layer="376" to-port="0" />
14049 <edge from-layer="376" from-port="2" to-layer="379" to-port="0" />
14050 <edge from-layer="376" from-port="2" to-layer="378" to-port="0" />
14051 <edge from-layer="377" from-port="0" to-layer="378" to-port="1" />
14052 <edge from-layer="378" from-port="2" to-layer="379" to-port="1" />
14053 <edge from-layer="379" from-port="2" to-layer="387" to-port="0" />
14054 <edge from-layer="379" from-port="2" to-layer="381" to-port="0" />
14055 <edge from-layer="380" from-port="0" to-layer="381" to-port="1" />
14056 <edge from-layer="381" from-port="2" to-layer="383" to-port="0" />
14057 <edge from-layer="382" from-port="0" to-layer="383" to-port="1" />
14058 <edge from-layer="383" from-port="2" to-layer="385" to-port="0" />
14059 <edge from-layer="384" from-port="0" to-layer="385" to-port="1" />
14060 <edge from-layer="385" from-port="2" to-layer="386" to-port="0" />
14061 <edge from-layer="386" from-port="1" to-layer="387" to-port="1" />
14062 <edge from-layer="387" from-port="2" to-layer="389" to-port="0" />
14063 <edge from-layer="388" from-port="0" to-layer="389" to-port="1" />
14064 <edge from-layer="389" from-port="2" to-layer="391" to-port="0" />
14065 <edge from-layer="390" from-port="0" to-layer="391" to-port="1" />
14066 <edge from-layer="391" from-port="2" to-layer="393" to-port="0" />
14067 <edge from-layer="391" from-port="2" to-layer="406" to-port="1" />
14068 <edge from-layer="392" from-port="0" to-layer="393" to-port="1" />
14069 <edge from-layer="393" from-port="2" to-layer="394" to-port="1" />
14070 <edge from-layer="394" from-port="2" to-layer="396" to-port="0" />
14071 <edge from-layer="394" from-port="2" to-layer="400" to-port="0" />
14072 <edge from-layer="395" from-port="0" to-layer="396" to-port="1" />
14073 <edge from-layer="396" from-port="2" to-layer="397" to-port="0" />
14074 <edge from-layer="397" from-port="1" to-layer="399" to-port="0" />
14075 <edge from-layer="398" from-port="0" to-layer="399" to-port="1" />
14076 <edge from-layer="399" from-port="2" to-layer="400" to-port="1" />
14077 <edge from-layer="400" from-port="2" to-layer="402" to-port="0" />
14078 <edge from-layer="401" from-port="0" to-layer="402" to-port="1" />
14079 <edge from-layer="402" from-port="2" to-layer="404" to-port="0" />
14080 <edge from-layer="403" from-port="0" to-layer="404" to-port="1" />
14081 <edge from-layer="404" from-port="2" to-layer="405" to-port="1" />
14082 <edge from-layer="405" from-port="2" to-layer="406" to-port="0" />
14083 <edge from-layer="406" from-port="2" to-layer="409" to-port="0" />
14084 <edge from-layer="406" from-port="2" to-layer="408" to-port="0" />
14085 <edge from-layer="407" from-port="0" to-layer="408" to-port="1" />
14086 <edge from-layer="408" from-port="2" to-layer="409" to-port="1" />
14087 <edge from-layer="409" from-port="2" to-layer="411" to-port="0" />
14088 <edge from-layer="409" from-port="2" to-layer="417" to-port="0" />
14089 <edge from-layer="410" from-port="0" to-layer="411" to-port="1" />
14090 <edge from-layer="411" from-port="2" to-layer="413" to-port="0" />
14091 <edge from-layer="412" from-port="0" to-layer="413" to-port="1" />
14092 <edge from-layer="413" from-port="2" to-layer="415" to-port="0" />
14093 <edge from-layer="414" from-port="0" to-layer="415" to-port="1" />
14094 <edge from-layer="415" from-port="2" to-layer="416" to-port="0" />
14095 <edge from-layer="416" from-port="1" to-layer="417" to-port="1" />
14096 <edge from-layer="417" from-port="2" to-layer="419" to-port="0" />
14097 <edge from-layer="418" from-port="0" to-layer="419" to-port="1" />
14098 <edge from-layer="419" from-port="2" to-layer="421" to-port="0" />
14099 <edge from-layer="420" from-port="0" to-layer="421" to-port="1" />
14100 <edge from-layer="421" from-port="2" to-layer="430" to-port="0" />
14101 <edge from-layer="421" from-port="2" to-layer="498" to-port="1" />
14102 <edge from-layer="421" from-port="2" to-layer="422" to-port="0" />
14103 <edge from-layer="421" from-port="2" to-layer="440" to-port="0" />
14104 <edge from-layer="421" from-port="2" to-layer="475" to-port="0" />
14105 <edge from-layer="421" from-port="2" to-layer="455" to-port="0" />
14106 <edge from-layer="422" from-port="1" to-layer="425" to-port="0" />
14107 <edge from-layer="423" from-port="0" to-layer="425" to-port="1" />
14108 <edge from-layer="424" from-port="0" to-layer="425" to-port="2" />
14109 <edge from-layer="425" from-port="3" to-layer="443" to-port="0" />
14110 <edge from-layer="425" from-port="3" to-layer="427" to-port="0" />
14111 <edge from-layer="425" from-port="3" to-layer="490" to-port="0" />
14112 <edge from-layer="425" from-port="3" to-layer="478" to-port="0" />
14113 <edge from-layer="425" from-port="3" to-layer="458" to-port="0" />
14114 <edge from-layer="426" from-port="0" to-layer="427" to-port="1" />
14115 <edge from-layer="427" from-port="2" to-layer="436" to-port="0" />
14116 <edge from-layer="428" from-port="0" to-layer="436" to-port="1" />
14117 <edge from-layer="429" from-port="0" to-layer="436" to-port="2" />
14118 <edge from-layer="430" from-port="1" to-layer="433" to-port="0" />
14119 <edge from-layer="431" from-port="0" to-layer="433" to-port="1" />
14120 <edge from-layer="432" from-port="0" to-layer="433" to-port="2" />
14121 <edge from-layer="433" from-port="3" to-layer="435" to-port="0" />
14122 <edge from-layer="434" from-port="0" to-layer="435" to-port="1" />
14123 <edge from-layer="435" from-port="2" to-layer="436" to-port="3" />
14124 <edge from-layer="436" from-port="4" to-layer="437" to-port="1" />
14125 <edge from-layer="437" from-port="2" to-layer="468" to-port="0" />
14126 <edge from-layer="438" from-port="0" to-layer="441" to-port="0" />
14127 <edge from-layer="439" from-port="0" to-layer="440" to-port="1" />
14128 <edge from-layer="440" from-port="2" to-layer="441" to-port="1" />
14129 <edge from-layer="441" from-port="2" to-layer="448" to-port="0" />
14130 <edge from-layer="442" from-port="0" to-layer="443" to-port="1" />
14131 <edge from-layer="443" from-port="2" to-layer="447" to-port="0" />
14132 <edge from-layer="444" from-port="0" to-layer="447" to-port="1" />
14133 <edge from-layer="445" from-port="0" to-layer="447" to-port="2" />
14134 <edge from-layer="446" from-port="0" to-layer="447" to-port="3" />
14135 <edge from-layer="447" from-port="4" to-layer="448" to-port="1" />
14136 <edge from-layer="448" from-port="2" to-layer="450" to-port="0" />
14137 <edge from-layer="449" from-port="0" to-layer="450" to-port="1" />
14138 <edge from-layer="450" from-port="2" to-layer="452" to-port="0" />
14139 <edge from-layer="451" from-port="0" to-layer="452" to-port="1" />
14140 <edge from-layer="452" from-port="2" to-layer="466" to-port="0" />
14141 <edge from-layer="453" from-port="0" to-layer="456" to-port="0" />
14142 <edge from-layer="454" from-port="0" to-layer="455" to-port="1" />
14143 <edge from-layer="455" from-port="2" to-layer="456" to-port="1" />
14144 <edge from-layer="456" from-port="2" to-layer="463" to-port="0" />
14145 <edge from-layer="457" from-port="0" to-layer="458" to-port="1" />
14146 <edge from-layer="458" from-port="2" to-layer="462" to-port="0" />
14147 <edge from-layer="459" from-port="0" to-layer="462" to-port="1" />
14148 <edge from-layer="460" from-port="0" to-layer="462" to-port="2" />
14149 <edge from-layer="461" from-port="0" to-layer="462" to-port="3" />
14150 <edge from-layer="462" from-port="4" to-layer="463" to-port="1" />
14151 <edge from-layer="463" from-port="2" to-layer="465" to-port="0" />
14152 <edge from-layer="464" from-port="0" to-layer="465" to-port="1" />
14153 <edge from-layer="465" from-port="2" to-layer="466" to-port="1" />
14154 <edge from-layer="466" from-port="2" to-layer="471" to-port="2" />
14155 <edge from-layer="466" from-port="2" to-layer="467" to-port="0" />
14156 <edge from-layer="467" from-port="1" to-layer="468" to-port="1" />
14157 <edge from-layer="468" from-port="2" to-layer="469" to-port="0" />
14158 <edge from-layer="469" from-port="1" to-layer="471" to-port="0" />
14159 <edge from-layer="470" from-port="0" to-layer="471" to-port="1" />
14160 <edge from-layer="471" from-port="3" to-layer="472" to-port="0" />
14161 <edge from-layer="472" from-port="1" to-layer="486" to-port="0" />
14162 <edge from-layer="473" from-port="0" to-layer="476" to-port="0" />
14163 <edge from-layer="474" from-port="0" to-layer="475" to-port="1" />
14164 <edge from-layer="475" from-port="2" to-layer="476" to-port="1" />
14165 <edge from-layer="476" from-port="2" to-layer="483" to-port="0" />
14166 <edge from-layer="477" from-port="0" to-layer="478" to-port="1" />
14167 <edge from-layer="478" from-port="2" to-layer="482" to-port="0" />
14168 <edge from-layer="479" from-port="0" to-layer="482" to-port="1" />
14169 <edge from-layer="480" from-port="0" to-layer="482" to-port="2" />
14170 <edge from-layer="481" from-port="0" to-layer="482" to-port="3" />
14171 <edge from-layer="482" from-port="4" to-layer="483" to-port="1" />
14172 <edge from-layer="483" from-port="2" to-layer="485" to-port="0" />
14173 <edge from-layer="484" from-port="0" to-layer="485" to-port="1" />
14174 <edge from-layer="485" from-port="2" to-layer="486" to-port="1" />
14175 <edge from-layer="486" from-port="2" to-layer="488" to-port="0" />
14176 <edge from-layer="487" from-port="0" to-layer="488" to-port="1" />
14177 <edge from-layer="488" from-port="2" to-layer="494" to-port="0" />
14178 <edge from-layer="489" from-port="0" to-layer="490" to-port="1" />
14179 <edge from-layer="490" from-port="2" to-layer="493" to-port="0" />
14180 <edge from-layer="491" from-port="0" to-layer="493" to-port="1" />
14181 <edge from-layer="492" from-port="0" to-layer="493" to-port="2" />
14182 <edge from-layer="493" from-port="3" to-layer="494" to-port="1" />
14183 <edge from-layer="494" from-port="2" to-layer="496" to-port="0" />
14184 <edge from-layer="495" from-port="0" to-layer="496" to-port="1" />
14185 <edge from-layer="496" from-port="2" to-layer="497" to-port="1" />
14186 <edge from-layer="497" from-port="2" to-layer="498" to-port="0" />
14187 <edge from-layer="498" from-port="2" to-layer="500" to-port="0" />
14188 <edge from-layer="498" from-port="2" to-layer="501" to-port="0" />
14189 <edge from-layer="499" from-port="0" to-layer="500" to-port="1" />
14190 <edge from-layer="500" from-port="2" to-layer="501" to-port="1" />
14191 <edge from-layer="501" from-port="2" to-layer="509" to-port="0" />
14192 <edge from-layer="501" from-port="2" to-layer="503" to-port="0" />
14193 <edge from-layer="502" from-port="0" to-layer="503" to-port="1" />
14194 <edge from-layer="503" from-port="2" to-layer="505" to-port="0" />
14195 <edge from-layer="504" from-port="0" to-layer="505" to-port="1" />
14196 <edge from-layer="505" from-port="2" to-layer="507" to-port="0" />
14197 <edge from-layer="506" from-port="0" to-layer="507" to-port="1" />
14198 <edge from-layer="507" from-port="2" to-layer="508" to-port="0" />
14199 <edge from-layer="508" from-port="1" to-layer="509" to-port="1" />
14200 <edge from-layer="509" from-port="2" to-layer="511" to-port="0" />
14201 <edge from-layer="510" from-port="0" to-layer="511" to-port="1" />
14202 <edge from-layer="511" from-port="2" to-layer="513" to-port="0" />
14203 <edge from-layer="512" from-port="0" to-layer="513" to-port="1" />
14204 <edge from-layer="513" from-port="2" to-layer="528" to-port="1" />
14205 <edge from-layer="513" from-port="2" to-layer="515" to-port="0" />
14206 <edge from-layer="514" from-port="0" to-layer="515" to-port="1" />
14207 <edge from-layer="515" from-port="2" to-layer="516" to-port="1" />
14208 <edge from-layer="516" from-port="2" to-layer="522" to-port="0" />
14209 <edge from-layer="516" from-port="2" to-layer="518" to-port="0" />
14210 <edge from-layer="517" from-port="0" to-layer="518" to-port="1" />
14211 <edge from-layer="518" from-port="2" to-layer="519" to-port="0" />
14212 <edge from-layer="519" from-port="1" to-layer="521" to-port="0" />
14213 <edge from-layer="520" from-port="0" to-layer="521" to-port="1" />
14214 <edge from-layer="521" from-port="2" to-layer="522" to-port="1" />
14215 <edge from-layer="522" from-port="2" to-layer="524" to-port="0" />
14216 <edge from-layer="523" from-port="0" to-layer="524" to-port="1" />
14217 <edge from-layer="524" from-port="2" to-layer="526" to-port="0" />
14218 <edge from-layer="525" from-port="0" to-layer="526" to-port="1" />
14219 <edge from-layer="526" from-port="2" to-layer="527" to-port="1" />
14220 <edge from-layer="527" from-port="2" to-layer="528" to-port="0" />
14221 <edge from-layer="528" from-port="2" to-layer="531" to-port="0" />
14222 <edge from-layer="528" from-port="2" to-layer="530" to-port="0" />
14223 <edge from-layer="529" from-port="0" to-layer="530" to-port="1" />
14224 <edge from-layer="530" from-port="2" to-layer="531" to-port="1" />
14225 <edge from-layer="531" from-port="2" to-layer="533" to-port="0" />
14226 <edge from-layer="531" from-port="2" to-layer="539" to-port="0" />
14227 <edge from-layer="532" from-port="0" to-layer="533" to-port="1" />
14228 <edge from-layer="533" from-port="2" to-layer="535" to-port="0" />
14229 <edge from-layer="534" from-port="0" to-layer="535" to-port="1" />
14230 <edge from-layer="535" from-port="2" to-layer="537" to-port="0" />
14231 <edge from-layer="536" from-port="0" to-layer="537" to-port="1" />
14232 <edge from-layer="537" from-port="2" to-layer="538" to-port="0" />
14233 <edge from-layer="538" from-port="1" to-layer="539" to-port="1" />
14234 <edge from-layer="539" from-port="2" to-layer="541" to-port="0" />
14235 <edge from-layer="540" from-port="0" to-layer="541" to-port="1" />
14236 <edge from-layer="541" from-port="2" to-layer="543" to-port="0" />
14237 <edge from-layer="542" from-port="0" to-layer="543" to-port="1" />
14238 <edge from-layer="543" from-port="2" to-layer="597" to-port="0" />
14239 <edge from-layer="543" from-port="2" to-layer="577" to-port="0" />
14240 <edge from-layer="543" from-port="2" to-layer="562" to-port="0" />
14241 <edge from-layer="543" from-port="2" to-layer="620" to-port="1" />
14242 <edge from-layer="543" from-port="2" to-layer="544" to-port="0" />
14243 <edge from-layer="543" from-port="2" to-layer="552" to-port="0" />
14244 <edge from-layer="544" from-port="1" to-layer="547" to-port="0" />
14245 <edge from-layer="545" from-port="0" to-layer="547" to-port="1" />
14246 <edge from-layer="546" from-port="0" to-layer="547" to-port="2" />
14247 <edge from-layer="547" from-port="3" to-layer="549" to-port="0" />
14248 <edge from-layer="547" from-port="3" to-layer="565" to-port="0" />
14249 <edge from-layer="547" from-port="3" to-layer="612" to-port="0" />
14250 <edge from-layer="547" from-port="3" to-layer="600" to-port="0" />
14251 <edge from-layer="547" from-port="3" to-layer="580" to-port="0" />
14252 <edge from-layer="548" from-port="0" to-layer="549" to-port="1" />
14253 <edge from-layer="549" from-port="2" to-layer="558" to-port="0" />
14254 <edge from-layer="550" from-port="0" to-layer="558" to-port="1" />
14255 <edge from-layer="551" from-port="0" to-layer="558" to-port="2" />
14256 <edge from-layer="552" from-port="1" to-layer="555" to-port="0" />
14257 <edge from-layer="553" from-port="0" to-layer="555" to-port="1" />
14258 <edge from-layer="554" from-port="0" to-layer="555" to-port="2" />
14259 <edge from-layer="555" from-port="3" to-layer="557" to-port="0" />
14260 <edge from-layer="556" from-port="0" to-layer="557" to-port="1" />
14261 <edge from-layer="557" from-port="2" to-layer="558" to-port="3" />
14262 <edge from-layer="558" from-port="4" to-layer="559" to-port="1" />
14263 <edge from-layer="559" from-port="2" to-layer="590" to-port="0" />
14264 <edge from-layer="560" from-port="0" to-layer="563" to-port="0" />
14265 <edge from-layer="561" from-port="0" to-layer="562" to-port="1" />
14266 <edge from-layer="562" from-port="2" to-layer="563" to-port="1" />
14267 <edge from-layer="563" from-port="2" to-layer="570" to-port="0" />
14268 <edge from-layer="564" from-port="0" to-layer="565" to-port="1" />
14269 <edge from-layer="565" from-port="2" to-layer="569" to-port="0" />
14270 <edge from-layer="566" from-port="0" to-layer="569" to-port="1" />
14271 <edge from-layer="567" from-port="0" to-layer="569" to-port="2" />
14272 <edge from-layer="568" from-port="0" to-layer="569" to-port="3" />
14273 <edge from-layer="569" from-port="4" to-layer="570" to-port="1" />
14274 <edge from-layer="570" from-port="2" to-layer="572" to-port="0" />
14275 <edge from-layer="571" from-port="0" to-layer="572" to-port="1" />
14276 <edge from-layer="572" from-port="2" to-layer="574" to-port="0" />
14277 <edge from-layer="573" from-port="0" to-layer="574" to-port="1" />
14278 <edge from-layer="574" from-port="2" to-layer="588" to-port="0" />
14279 <edge from-layer="575" from-port="0" to-layer="578" to-port="0" />
14280 <edge from-layer="576" from-port="0" to-layer="577" to-port="1" />
14281 <edge from-layer="577" from-port="2" to-layer="578" to-port="1" />
14282 <edge from-layer="578" from-port="2" to-layer="585" to-port="0" />
14283 <edge from-layer="579" from-port="0" to-layer="580" to-port="1" />
14284 <edge from-layer="580" from-port="2" to-layer="584" to-port="0" />
14285 <edge from-layer="581" from-port="0" to-layer="584" to-port="1" />
14286 <edge from-layer="582" from-port="0" to-layer="584" to-port="2" />
14287 <edge from-layer="583" from-port="0" to-layer="584" to-port="3" />
14288 <edge from-layer="584" from-port="4" to-layer="585" to-port="1" />
14289 <edge from-layer="585" from-port="2" to-layer="587" to-port="0" />
14290 <edge from-layer="586" from-port="0" to-layer="587" to-port="1" />
14291 <edge from-layer="587" from-port="2" to-layer="588" to-port="1" />
14292 <edge from-layer="588" from-port="2" to-layer="589" to-port="0" />
14293 <edge from-layer="588" from-port="2" to-layer="593" to-port="2" />
14294 <edge from-layer="589" from-port="1" to-layer="590" to-port="1" />
14295 <edge from-layer="590" from-port="2" to-layer="591" to-port="0" />
14296 <edge from-layer="591" from-port="1" to-layer="593" to-port="0" />
14297 <edge from-layer="592" from-port="0" to-layer="593" to-port="1" />
14298 <edge from-layer="593" from-port="3" to-layer="594" to-port="0" />
14299 <edge from-layer="594" from-port="1" to-layer="608" to-port="0" />
14300 <edge from-layer="595" from-port="0" to-layer="598" to-port="0" />
14301 <edge from-layer="596" from-port="0" to-layer="597" to-port="1" />
14302 <edge from-layer="597" from-port="2" to-layer="598" to-port="1" />
14303 <edge from-layer="598" from-port="2" to-layer="605" to-port="0" />
14304 <edge from-layer="599" from-port="0" to-layer="600" to-port="1" />
14305 <edge from-layer="600" from-port="2" to-layer="604" to-port="0" />
14306 <edge from-layer="601" from-port="0" to-layer="604" to-port="1" />
14307 <edge from-layer="602" from-port="0" to-layer="604" to-port="2" />
14308 <edge from-layer="603" from-port="0" to-layer="604" to-port="3" />
14309 <edge from-layer="604" from-port="4" to-layer="605" to-port="1" />
14310 <edge from-layer="605" from-port="2" to-layer="607" to-port="0" />
14311 <edge from-layer="606" from-port="0" to-layer="607" to-port="1" />
14312 <edge from-layer="607" from-port="2" to-layer="608" to-port="1" />
14313 <edge from-layer="608" from-port="2" to-layer="610" to-port="0" />
14314 <edge from-layer="609" from-port="0" to-layer="610" to-port="1" />
14315 <edge from-layer="610" from-port="2" to-layer="616" to-port="0" />
14316 <edge from-layer="611" from-port="0" to-layer="612" to-port="1" />
14317 <edge from-layer="612" from-port="2" to-layer="615" to-port="0" />
14318 <edge from-layer="613" from-port="0" to-layer="615" to-port="1" />
14319 <edge from-layer="614" from-port="0" to-layer="615" to-port="2" />
14320 <edge from-layer="615" from-port="3" to-layer="616" to-port="1" />
14321 <edge from-layer="616" from-port="2" to-layer="618" to-port="0" />
14322 <edge from-layer="617" from-port="0" to-layer="618" to-port="1" />
14323 <edge from-layer="618" from-port="2" to-layer="619" to-port="1" />
14324 <edge from-layer="619" from-port="2" to-layer="620" to-port="0" />
14325 <edge from-layer="620" from-port="2" to-layer="623" to-port="0" />
14326 <edge from-layer="620" from-port="2" to-layer="622" to-port="0" />
14327 <edge from-layer="621" from-port="0" to-layer="622" to-port="1" />
14328 <edge from-layer="622" from-port="2" to-layer="623" to-port="1" />
14329 <edge from-layer="623" from-port="2" to-layer="631" to-port="0" />
14330 <edge from-layer="623" from-port="2" to-layer="625" to-port="0" />
14331 <edge from-layer="624" from-port="0" to-layer="625" to-port="1" />
14332 <edge from-layer="625" from-port="2" to-layer="627" to-port="0" />
14333 <edge from-layer="626" from-port="0" to-layer="627" to-port="1" />
14334 <edge from-layer="627" from-port="2" to-layer="629" to-port="0" />
14335 <edge from-layer="628" from-port="0" to-layer="629" to-port="1" />
14336 <edge from-layer="629" from-port="2" to-layer="630" to-port="0" />
14337 <edge from-layer="630" from-port="1" to-layer="631" to-port="1" />
14338 <edge from-layer="631" from-port="2" to-layer="633" to-port="0" />
14339 <edge from-layer="632" from-port="0" to-layer="633" to-port="1" />
14340 <edge from-layer="633" from-port="2" to-layer="635" to-port="0" />
14341 <edge from-layer="634" from-port="0" to-layer="635" to-port="1" />
14342 <edge from-layer="635" from-port="2" to-layer="637" to-port="0" />
14343 <edge from-layer="635" from-port="2" to-layer="650" to-port="1" />
14344 <edge from-layer="636" from-port="0" to-layer="637" to-port="1" />
14345 <edge from-layer="637" from-port="2" to-layer="638" to-port="1" />
14346 <edge from-layer="638" from-port="2" to-layer="644" to-port="0" />
14347 <edge from-layer="638" from-port="2" to-layer="640" to-port="0" />
14348 <edge from-layer="639" from-port="0" to-layer="640" to-port="1" />
14349 <edge from-layer="640" from-port="2" to-layer="641" to-port="0" />
14350 <edge from-layer="641" from-port="1" to-layer="643" to-port="0" />
14351 <edge from-layer="642" from-port="0" to-layer="643" to-port="1" />
14352 <edge from-layer="643" from-port="2" to-layer="644" to-port="1" />
14353 <edge from-layer="644" from-port="2" to-layer="646" to-port="0" />
14354 <edge from-layer="645" from-port="0" to-layer="646" to-port="1" />
14355 <edge from-layer="646" from-port="2" to-layer="648" to-port="0" />
14356 <edge from-layer="647" from-port="0" to-layer="648" to-port="1" />
14357 <edge from-layer="648" from-port="2" to-layer="649" to-port="1" />
14358 <edge from-layer="649" from-port="2" to-layer="650" to-port="0" />
14359 <edge from-layer="650" from-port="2" to-layer="653" to-port="0" />
14360 <edge from-layer="650" from-port="2" to-layer="652" to-port="0" />
14361 <edge from-layer="651" from-port="0" to-layer="652" to-port="1" />
14362 <edge from-layer="652" from-port="2" to-layer="653" to-port="1" />
14363 <edge from-layer="653" from-port="2" to-layer="655" to-port="0" />
14364 <edge from-layer="653" from-port="2" to-layer="661" to-port="0" />
14365 <edge from-layer="654" from-port="0" to-layer="655" to-port="1" />
14366 <edge from-layer="655" from-port="2" to-layer="657" to-port="0" />
14367 <edge from-layer="656" from-port="0" to-layer="657" to-port="1" />
14368 <edge from-layer="657" from-port="2" to-layer="659" to-port="0" />
14369 <edge from-layer="658" from-port="0" to-layer="659" to-port="1" />
14370 <edge from-layer="659" from-port="2" to-layer="660" to-port="0" />
14371 <edge from-layer="660" from-port="1" to-layer="661" to-port="1" />
14372 <edge from-layer="661" from-port="2" to-layer="663" to-port="0" />
14373 <edge from-layer="662" from-port="0" to-layer="663" to-port="1" />
14374 <edge from-layer="663" from-port="2" to-layer="665" to-port="0" />
14375 <edge from-layer="664" from-port="0" to-layer="665" to-port="1" />
14376 <edge from-layer="665" from-port="2" to-layer="699" to-port="0" />
14377 <edge from-layer="665" from-port="2" to-layer="719" to-port="0" />
14378 <edge from-layer="665" from-port="2" to-layer="742" to-port="1" />
14379 <edge from-layer="665" from-port="2" to-layer="684" to-port="0" />
14380 <edge from-layer="665" from-port="2" to-layer="666" to-port="0" />
14381 <edge from-layer="665" from-port="2" to-layer="674" to-port="0" />
14382 <edge from-layer="666" from-port="1" to-layer="669" to-port="0" />
14383 <edge from-layer="667" from-port="0" to-layer="669" to-port="1" />
14384 <edge from-layer="668" from-port="0" to-layer="669" to-port="2" />
14385 <edge from-layer="669" from-port="3" to-layer="702" to-port="0" />
14386 <edge from-layer="669" from-port="3" to-layer="722" to-port="0" />
14387 <edge from-layer="669" from-port="3" to-layer="734" to-port="0" />
14388 <edge from-layer="669" from-port="3" to-layer="687" to-port="0" />
14389 <edge from-layer="669" from-port="3" to-layer="671" to-port="0" />
14390 <edge from-layer="670" from-port="0" to-layer="671" to-port="1" />
14391 <edge from-layer="671" from-port="2" to-layer="680" to-port="0" />
14392 <edge from-layer="672" from-port="0" to-layer="680" to-port="1" />
14393 <edge from-layer="673" from-port="0" to-layer="680" to-port="2" />
14394 <edge from-layer="674" from-port="1" to-layer="677" to-port="0" />
14395 <edge from-layer="675" from-port="0" to-layer="677" to-port="1" />
14396 <edge from-layer="676" from-port="0" to-layer="677" to-port="2" />
14397 <edge from-layer="677" from-port="3" to-layer="679" to-port="0" />
14398 <edge from-layer="678" from-port="0" to-layer="679" to-port="1" />
14399 <edge from-layer="679" from-port="2" to-layer="680" to-port="3" />
14400 <edge from-layer="680" from-port="4" to-layer="681" to-port="1" />
14401 <edge from-layer="681" from-port="2" to-layer="712" to-port="0" />
14402 <edge from-layer="682" from-port="0" to-layer="685" to-port="0" />
14403 <edge from-layer="683" from-port="0" to-layer="684" to-port="1" />
14404 <edge from-layer="684" from-port="2" to-layer="685" to-port="1" />
14405 <edge from-layer="685" from-port="2" to-layer="692" to-port="0" />
14406 <edge from-layer="686" from-port="0" to-layer="687" to-port="1" />
14407 <edge from-layer="687" from-port="2" to-layer="691" to-port="0" />
14408 <edge from-layer="688" from-port="0" to-layer="691" to-port="1" />
14409 <edge from-layer="689" from-port="0" to-layer="691" to-port="2" />
14410 <edge from-layer="690" from-port="0" to-layer="691" to-port="3" />
14411 <edge from-layer="691" from-port="4" to-layer="692" to-port="1" />
14412 <edge from-layer="692" from-port="2" to-layer="694" to-port="0" />
14413 <edge from-layer="693" from-port="0" to-layer="694" to-port="1" />
14414 <edge from-layer="694" from-port="2" to-layer="696" to-port="0" />
14415 <edge from-layer="695" from-port="0" to-layer="696" to-port="1" />
14416 <edge from-layer="696" from-port="2" to-layer="710" to-port="0" />
14417 <edge from-layer="697" from-port="0" to-layer="700" to-port="0" />
14418 <edge from-layer="698" from-port="0" to-layer="699" to-port="1" />
14419 <edge from-layer="699" from-port="2" to-layer="700" to-port="1" />
14420 <edge from-layer="700" from-port="2" to-layer="707" to-port="0" />
14421 <edge from-layer="701" from-port="0" to-layer="702" to-port="1" />
14422 <edge from-layer="702" from-port="2" to-layer="706" to-port="0" />
14423 <edge from-layer="703" from-port="0" to-layer="706" to-port="1" />
14424 <edge from-layer="704" from-port="0" to-layer="706" to-port="2" />
14425 <edge from-layer="705" from-port="0" to-layer="706" to-port="3" />
14426 <edge from-layer="706" from-port="4" to-layer="707" to-port="1" />
14427 <edge from-layer="707" from-port="2" to-layer="709" to-port="0" />
14428 <edge from-layer="708" from-port="0" to-layer="709" to-port="1" />
14429 <edge from-layer="709" from-port="2" to-layer="710" to-port="1" />
14430 <edge from-layer="710" from-port="2" to-layer="715" to-port="2" />
14431 <edge from-layer="710" from-port="2" to-layer="711" to-port="0" />
14432 <edge from-layer="711" from-port="1" to-layer="712" to-port="1" />
14433 <edge from-layer="712" from-port="2" to-layer="713" to-port="0" />
14434 <edge from-layer="713" from-port="1" to-layer="715" to-port="0" />
14435 <edge from-layer="714" from-port="0" to-layer="715" to-port="1" />
14436 <edge from-layer="715" from-port="3" to-layer="716" to-port="0" />
14437 <edge from-layer="716" from-port="1" to-layer="730" to-port="0" />
14438 <edge from-layer="717" from-port="0" to-layer="720" to-port="0" />
14439 <edge from-layer="718" from-port="0" to-layer="719" to-port="1" />
14440 <edge from-layer="719" from-port="2" to-layer="720" to-port="1" />
14441 <edge from-layer="720" from-port="2" to-layer="727" to-port="0" />
14442 <edge from-layer="721" from-port="0" to-layer="722" to-port="1" />
14443 <edge from-layer="722" from-port="2" to-layer="726" to-port="0" />
14444 <edge from-layer="723" from-port="0" to-layer="726" to-port="1" />
14445 <edge from-layer="724" from-port="0" to-layer="726" to-port="2" />
14446 <edge from-layer="725" from-port="0" to-layer="726" to-port="3" />
14447 <edge from-layer="726" from-port="4" to-layer="727" to-port="1" />
14448 <edge from-layer="727" from-port="2" to-layer="729" to-port="0" />
14449 <edge from-layer="728" from-port="0" to-layer="729" to-port="1" />
14450 <edge from-layer="729" from-port="2" to-layer="730" to-port="1" />
14451 <edge from-layer="730" from-port="2" to-layer="732" to-port="0" />
14452 <edge from-layer="731" from-port="0" to-layer="732" to-port="1" />
14453 <edge from-layer="732" from-port="2" to-layer="738" to-port="0" />
14454 <edge from-layer="733" from-port="0" to-layer="734" to-port="1" />
14455 <edge from-layer="734" from-port="2" to-layer="737" to-port="0" />
14456 <edge from-layer="735" from-port="0" to-layer="737" to-port="1" />
14457 <edge from-layer="736" from-port="0" to-layer="737" to-port="2" />
14458 <edge from-layer="737" from-port="3" to-layer="738" to-port="1" />
14459 <edge from-layer="738" from-port="2" to-layer="740" to-port="0" />
14460 <edge from-layer="739" from-port="0" to-layer="740" to-port="1" />
14461 <edge from-layer="740" from-port="2" to-layer="741" to-port="1" />
14462 <edge from-layer="741" from-port="2" to-layer="742" to-port="0" />
14463 <edge from-layer="742" from-port="2" to-layer="744" to-port="0" />
14464 <edge from-layer="742" from-port="2" to-layer="745" to-port="0" />
14465 <edge from-layer="743" from-port="0" to-layer="744" to-port="1" />
14466 <edge from-layer="744" from-port="2" to-layer="745" to-port="1" />
14467 <edge from-layer="745" from-port="2" to-layer="747" to-port="0" />
14468 <edge from-layer="745" from-port="2" to-layer="753" to-port="0" />
14469 <edge from-layer="746" from-port="0" to-layer="747" to-port="1" />
14470 <edge from-layer="747" from-port="2" to-layer="749" to-port="0" />
14471 <edge from-layer="748" from-port="0" to-layer="749" to-port="1" />
14472 <edge from-layer="749" from-port="2" to-layer="751" to-port="0" />
14473 <edge from-layer="750" from-port="0" to-layer="751" to-port="1" />
14474 <edge from-layer="751" from-port="2" to-layer="752" to-port="0" />
14475 <edge from-layer="752" from-port="1" to-layer="753" to-port="1" />
14476 <edge from-layer="753" from-port="2" to-layer="755" to-port="0" />
14477 <edge from-layer="754" from-port="0" to-layer="755" to-port="1" />
14478 <edge from-layer="755" from-port="2" to-layer="757" to-port="0" />
14479 <edge from-layer="756" from-port="0" to-layer="757" to-port="1" />
14480 <edge from-layer="757" from-port="2" to-layer="759" to-port="0" />
14481 <edge from-layer="757" from-port="2" to-layer="772" to-port="1" />
14482 <edge from-layer="758" from-port="0" to-layer="759" to-port="1" />
14483 <edge from-layer="759" from-port="2" to-layer="760" to-port="1" />
14484 <edge from-layer="760" from-port="2" to-layer="762" to-port="0" />
14485 <edge from-layer="760" from-port="2" to-layer="766" to-port="0" />
14486 <edge from-layer="761" from-port="0" to-layer="762" to-port="1" />
14487 <edge from-layer="762" from-port="2" to-layer="763" to-port="0" />
14488 <edge from-layer="763" from-port="1" to-layer="765" to-port="0" />
14489 <edge from-layer="764" from-port="0" to-layer="765" to-port="1" />
14490 <edge from-layer="765" from-port="2" to-layer="766" to-port="1" />
14491 <edge from-layer="766" from-port="2" to-layer="768" to-port="0" />
14492 <edge from-layer="767" from-port="0" to-layer="768" to-port="1" />
14493 <edge from-layer="768" from-port="2" to-layer="770" to-port="0" />
14494 <edge from-layer="769" from-port="0" to-layer="770" to-port="1" />
14495 <edge from-layer="770" from-port="2" to-layer="771" to-port="1" />
14496 <edge from-layer="771" from-port="2" to-layer="772" to-port="0" />
14497 <edge from-layer="772" from-port="2" to-layer="774" to-port="0" />
14498 <edge from-layer="772" from-port="2" to-layer="775" to-port="0" />
14499 <edge from-layer="773" from-port="0" to-layer="774" to-port="1" />
14500 <edge from-layer="774" from-port="2" to-layer="775" to-port="1" />
14501 <edge from-layer="775" from-port="2" to-layer="783" to-port="0" />
14502 <edge from-layer="775" from-port="2" to-layer="777" to-port="0" />
14503 <edge from-layer="776" from-port="0" to-layer="777" to-port="1" />
14504 <edge from-layer="777" from-port="2" to-layer="779" to-port="0" />
14505 <edge from-layer="778" from-port="0" to-layer="779" to-port="1" />
14506 <edge from-layer="779" from-port="2" to-layer="781" to-port="0" />
14507 <edge from-layer="780" from-port="0" to-layer="781" to-port="1" />
14508 <edge from-layer="781" from-port="2" to-layer="782" to-port="0" />
14509 <edge from-layer="782" from-port="1" to-layer="783" to-port="1" />
14510 <edge from-layer="783" from-port="2" to-layer="785" to-port="0" />
14511 <edge from-layer="784" from-port="0" to-layer="785" to-port="1" />
14512 <edge from-layer="785" from-port="2" to-layer="787" to-port="0" />
14513 <edge from-layer="786" from-port="0" to-layer="787" to-port="1" />
14514 <edge from-layer="787" from-port="2" to-layer="789" to-port="0" />
14515 <edge from-layer="788" from-port="0" to-layer="789" to-port="1" />
14516 <edge from-layer="789" from-port="2" to-layer="790" to-port="1" />
14517 <edge from-layer="790" from-port="2" to-layer="793" to-port="0" />
14518 <edge from-layer="791" from-port="0" to-layer="793" to-port="1" />
14519 <edge from-layer="792" from-port="0" to-layer="793" to-port="2" />
14520 <edge from-layer="793" from-port="4" to-layer="795" to-port="0" />
14521 <edge from-layer="793" from-port="3" to-layer="798" to-port="0" />
14522 <edge from-layer="794" from-port="0" to-layer="795" to-port="1" />
14523 <edge from-layer="795" from-port="2" to-layer="796" to-port="0" />
14524 <edge from-layer="797" from-port="0" to-layer="798" to-port="1" />
14525 <edge from-layer="798" from-port="2" to-layer="799" to-port="0" />
14526 </edges>
14527 <rt_info />
14528 </net>
14529