openvino/openvino_model.xml
200.5 KB · 6618 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model110" version="11">
3 <layers>
4 <layer id="2" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="1" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="0" name="token_type_ids" type="Parameter" version="opset1">
23 <data shape="?,?" element_type="i64" />
24 <output>
25 <port id="0" precision="I64" names="token_type_ids">
26 <dim>-1</dim>
27 <dim>-1</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="self.bert.embeddings.word_embeddings.weight" type="Const" version="opset1">
32 <data element_type="f32" shape="30522, 384" offset="0" size="46881792" />
33 <output>
34 <port id="0" precision="FP32" names="self.bert.embeddings.word_embeddings.weight">
35 <dim>30522</dim>
36 <dim>384</dim>
37 </port>
38 </output>
39 </layer>
40 <layer id="4" name="__module.bert.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
41 <data destination_type="i32" />
42 <input>
43 <port id="0" precision="I64">
44 <dim>-1</dim>
45 <dim>-1</dim>
46 </port>
47 </input>
48 <output>
49 <port id="1" precision="I32">
50 <dim>-1</dim>
51 <dim>-1</dim>
52 </port>
53 </output>
54 </layer>
55 <layer id="5" name="__module.bert.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
56 <data element_type="i32" shape="" offset="46881792" size="4" />
57 <output>
58 <port id="0" precision="I32" />
59 </output>
60 </layer>
61 <layer id="6" name="__module.bert.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
62 <data batch_dims="0" />
63 <input>
64 <port id="0" precision="FP32">
65 <dim>30522</dim>
66 <dim>384</dim>
67 </port>
68 <port id="1" precision="I32">
69 <dim>-1</dim>
70 <dim>-1</dim>
71 </port>
72 <port id="2" precision="I32" />
73 </input>
74 <output>
75 <port id="3" precision="FP32" names="48,inputs_embeds">
76 <dim>-1</dim>
77 <dim>-1</dim>
78 <dim>384</dim>
79 </port>
80 </output>
81 </layer>
82 <layer id="7" name="self.bert.embeddings.token_type_embeddings.weight" type="Const" version="opset1">
83 <data element_type="f32" shape="2, 384" offset="46881796" size="3072" />
84 <output>
85 <port id="0" precision="FP32" names="self.bert.embeddings.token_type_embeddings.weight">
86 <dim>2</dim>
87 <dim>384</dim>
88 </port>
89 </output>
90 </layer>
91 <layer id="8" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
92 <data destination_type="i32" />
93 <input>
94 <port id="0" precision="I64">
95 <dim>-1</dim>
96 <dim>-1</dim>
97 </port>
98 </input>
99 <output>
100 <port id="1" precision="I32">
101 <dim>-1</dim>
102 <dim>-1</dim>
103 </port>
104 </output>
105 </layer>
106 <layer id="9" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
107 <data element_type="i32" shape="" offset="46881792" size="4" />
108 <output>
109 <port id="0" precision="I32" />
110 </output>
111 </layer>
112 <layer id="10" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
113 <data batch_dims="0" />
114 <input>
115 <port id="0" precision="FP32">
116 <dim>2</dim>
117 <dim>384</dim>
118 </port>
119 <port id="1" precision="I32">
120 <dim>-1</dim>
121 <dim>-1</dim>
122 </port>
123 <port id="2" precision="I32" />
124 </input>
125 <output>
126 <port id="3" precision="FP32" names="50,token_type_embeddings.1">
127 <dim>-1</dim>
128 <dim>-1</dim>
129 <dim>384</dim>
130 </port>
131 </output>
132 </layer>
133 <layer id="11" name="__module.bert.embeddings/aten::add/Add" type="Add" version="opset1">
134 <data auto_broadcast="numpy" />
135 <input>
136 <port id="0" precision="FP32">
137 <dim>-1</dim>
138 <dim>-1</dim>
139 <dim>384</dim>
140 </port>
141 <port id="1" precision="FP32">
142 <dim>-1</dim>
143 <dim>-1</dim>
144 <dim>384</dim>
145 </port>
146 </input>
147 <output>
148 <port id="2" precision="FP32" names="51_1">
149 <dim>-1</dim>
150 <dim>-1</dim>
151 <dim>384</dim>
152 </port>
153 </output>
154 </layer>
155 <layer id="12" name="self.bert.embeddings.position_embeddings.weight" type="Const" version="opset1">
156 <data element_type="f32" shape="512, 384" offset="46884868" size="786432" />
157 <output>
158 <port id="0" precision="FP32" names="self.bert.embeddings.position_embeddings.weight">
159 <dim>512</dim>
160 <dim>384</dim>
161 </port>
162 </output>
163 </layer>
164 <layer id="13" name="__module.bert.embeddings/aten::slice/Slice" type="Const" version="opset1">
165 <data element_type="i64" shape="1, 512" offset="47671300" size="4096" />
166 <output>
167 <port id="0" precision="I64" names="45">
168 <dim>1</dim>
169 <dim>512</dim>
170 </port>
171 </output>
172 </layer>
173 <layer id="14" name="__module.bert.embeddings/aten::slice/Reshape" type="Const" version="opset1">
174 <data element_type="i64" shape="1" offset="47675396" size="8" />
175 <output>
176 <port id="0" precision="I64">
177 <dim>1</dim>
178 </port>
179 </output>
180 </layer>
181 <layer id="15" name="ShapeOf_90427" type="ShapeOf" version="opset3">
182 <data output_type="i64" />
183 <input>
184 <port id="0" precision="I64">
185 <dim>-1</dim>
186 <dim>-1</dim>
187 </port>
188 </input>
189 <output>
190 <port id="1" precision="I64">
191 <dim>2</dim>
192 </port>
193 </output>
194 </layer>
195 <layer id="16" name="Constant_90546" type="Const" version="opset1">
196 <data element_type="i64" shape="1" offset="47675404" size="8" />
197 <output>
198 <port id="0" precision="I64">
199 <dim>1</dim>
200 </port>
201 </output>
202 </layer>
203 <layer id="17" name="Constant_90429" type="Const" version="opset1">
204 <data element_type="i64" shape="" offset="47675396" size="8" />
205 <output>
206 <port id="0" precision="I64" />
207 </output>
208 </layer>
209 <layer id="18" name="Gather_90430" type="Gather" version="opset8">
210 <data batch_dims="0" />
211 <input>
212 <port id="0" precision="I64">
213 <dim>2</dim>
214 </port>
215 <port id="1" precision="I64">
216 <dim>1</dim>
217 </port>
218 <port id="2" precision="I64" />
219 </input>
220 <output>
221 <port id="3" precision="I64" names="35,41,43,44,61">
222 <dim>1</dim>
223 </port>
224 </output>
225 </layer>
226 <layer id="19" name="__module.bert.embeddings/aten::slice/Reshape_2" type="Const" version="opset1">
227 <data element_type="i64" shape="1" offset="47675404" size="8" />
228 <output>
229 <port id="0" precision="I64">
230 <dim>1</dim>
231 </port>
232 </output>
233 </layer>
234 <layer id="20" name="__module.bert.embeddings/aten::slice/Reshape_3" type="Const" version="opset1">
235 <data element_type="i64" shape="1" offset="47675404" size="8" />
236 <output>
237 <port id="0" precision="I64">
238 <dim>1</dim>
239 </port>
240 </output>
241 </layer>
242 <layer id="21" name="__module.bert.embeddings/aten::slice/Slice_1" type="Slice" version="opset8">
243 <input>
244 <port id="0" precision="I64">
245 <dim>1</dim>
246 <dim>512</dim>
247 </port>
248 <port id="1" precision="I64">
249 <dim>1</dim>
250 </port>
251 <port id="2" precision="I64">
252 <dim>1</dim>
253 </port>
254 <port id="3" precision="I64">
255 <dim>1</dim>
256 </port>
257 <port id="4" precision="I64">
258 <dim>1</dim>
259 </port>
260 </input>
261 <output>
262 <port id="5" precision="I64" names="46">
263 <dim>1</dim>
264 <dim>-1</dim>
265 </port>
266 </output>
267 </layer>
268 <layer id="22" name="__module.bert.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
269 <data destination_type="i32" />
270 <input>
271 <port id="0" precision="I64">
272 <dim>1</dim>
273 <dim>-1</dim>
274 </port>
275 </input>
276 <output>
277 <port id="1" precision="I32">
278 <dim>1</dim>
279 <dim>-1</dim>
280 </port>
281 </output>
282 </layer>
283 <layer id="23" name="__module.bert.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
284 <data element_type="i32" shape="" offset="46881792" size="4" />
285 <output>
286 <port id="0" precision="I32" />
287 </output>
288 </layer>
289 <layer id="24" name="__module.bert.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
290 <data batch_dims="0" />
291 <input>
292 <port id="0" precision="FP32">
293 <dim>512</dim>
294 <dim>384</dim>
295 </port>
296 <port id="1" precision="I32">
297 <dim>1</dim>
298 <dim>-1</dim>
299 </port>
300 <port id="2" precision="I32" />
301 </input>
302 <output>
303 <port id="3" precision="FP32" names="53,position_embeddings.1">
304 <dim>1</dim>
305 <dim>-1</dim>
306 <dim>384</dim>
307 </port>
308 </output>
309 </layer>
310 <layer id="25" name="__module.bert.embeddings/aten::add_/Add" type="Add" version="opset1">
311 <data auto_broadcast="numpy" />
312 <input>
313 <port id="0" precision="FP32">
314 <dim>-1</dim>
315 <dim>-1</dim>
316 <dim>384</dim>
317 </port>
318 <port id="1" precision="FP32">
319 <dim>1</dim>
320 <dim>-1</dim>
321 <dim>384</dim>
322 </port>
323 </input>
324 <output>
325 <port id="2" precision="FP32" names="51,embeddings.1">
326 <dim>-1</dim>
327 <dim>-1</dim>
328 <dim>384</dim>
329 </port>
330 </output>
331 </layer>
332 <layer id="26" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
333 <data element_type="i32" shape="1" offset="47675412" size="4" />
334 <output>
335 <port id="0" precision="I32">
336 <dim>1</dim>
337 </port>
338 </output>
339 </layer>
340 <layer id="27" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
341 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
342 <input>
343 <port id="0" precision="FP32">
344 <dim>-1</dim>
345 <dim>-1</dim>
346 <dim>384</dim>
347 </port>
348 <port id="1" precision="I32">
349 <dim>1</dim>
350 </port>
351 </input>
352 <output>
353 <port id="2" precision="FP32">
354 <dim>-1</dim>
355 <dim>-1</dim>
356 <dim>384</dim>
357 </port>
358 </output>
359 </layer>
360 <layer id="28" name="Constant_90301" type="Const" version="opset1">
361 <data element_type="f32" shape="1, 1, 384" offset="47675416" size="1536" />
362 <output>
363 <port id="0" precision="FP32">
364 <dim>1</dim>
365 <dim>1</dim>
366 <dim>384</dim>
367 </port>
368 </output>
369 </layer>
370 <layer id="29" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
371 <data auto_broadcast="numpy" />
372 <input>
373 <port id="0" precision="FP32">
374 <dim>-1</dim>
375 <dim>-1</dim>
376 <dim>384</dim>
377 </port>
378 <port id="1" precision="FP32">
379 <dim>1</dim>
380 <dim>1</dim>
381 <dim>384</dim>
382 </port>
383 </input>
384 <output>
385 <port id="2" precision="FP32">
386 <dim>-1</dim>
387 <dim>-1</dim>
388 <dim>384</dim>
389 </port>
390 </output>
391 </layer>
392 <layer id="30" name="Constant_90302" type="Const" version="opset1">
393 <data element_type="f32" shape="1, 1, 384" offset="47676952" size="1536" />
394 <output>
395 <port id="0" precision="FP32">
396 <dim>1</dim>
397 <dim>1</dim>
398 <dim>384</dim>
399 </port>
400 </output>
401 </layer>
402 <layer id="31" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
403 <data auto_broadcast="numpy" />
404 <input>
405 <port id="0" precision="FP32">
406 <dim>-1</dim>
407 <dim>-1</dim>
408 <dim>384</dim>
409 </port>
410 <port id="1" precision="FP32">
411 <dim>1</dim>
412 <dim>1</dim>
413 <dim>384</dim>
414 </port>
415 </input>
416 <output>
417 <port id="2" precision="FP32" names="58,input.1">
418 <dim>-1</dim>
419 <dim>-1</dim>
420 <dim>384</dim>
421 </port>
422 </output>
423 </layer>
424 <layer id="32" name="self.bert.encoder.layer.0.attention.self.query.weight" type="Const" version="opset1">
425 <data element_type="f32" shape="384, 384" offset="47678488" size="589824" />
426 <output>
427 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.self.query.weight">
428 <dim>384</dim>
429 <dim>384</dim>
430 </port>
431 </output>
432 </layer>
433 <layer id="33" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
434 <data transpose_a="false" transpose_b="true" />
435 <input>
436 <port id="0" precision="FP32">
437 <dim>-1</dim>
438 <dim>-1</dim>
439 <dim>384</dim>
440 </port>
441 <port id="1" precision="FP32">
442 <dim>384</dim>
443 <dim>384</dim>
444 </port>
445 </input>
446 <output>
447 <port id="2" precision="FP32">
448 <dim>-1</dim>
449 <dim>-1</dim>
450 <dim>384</dim>
451 </port>
452 </output>
453 </layer>
454 <layer id="34" name="Constant_90303" type="Const" version="opset1">
455 <data element_type="f32" shape="1, 1, 384" offset="48268312" size="1536" />
456 <output>
457 <port id="0" precision="FP32">
458 <dim>1</dim>
459 <dim>1</dim>
460 <dim>384</dim>
461 </port>
462 </output>
463 </layer>
464 <layer id="35" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
465 <data auto_broadcast="numpy" />
466 <input>
467 <port id="0" precision="FP32">
468 <dim>-1</dim>
469 <dim>-1</dim>
470 <dim>384</dim>
471 </port>
472 <port id="1" precision="FP32">
473 <dim>1</dim>
474 <dim>1</dim>
475 <dim>384</dim>
476 </port>
477 </input>
478 <output>
479 <port id="2" precision="FP32" names="96,x.1">
480 <dim>-1</dim>
481 <dim>-1</dim>
482 <dim>384</dim>
483 </port>
484 </output>
485 </layer>
486 <layer id="36" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
487 <data element_type="i64" shape="4" offset="48269848" size="32" />
488 <output>
489 <port id="0" precision="I64">
490 <dim>4</dim>
491 </port>
492 </output>
493 </layer>
494 <layer id="37" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
495 <data special_zero="true" />
496 <input>
497 <port id="0" precision="FP32">
498 <dim>-1</dim>
499 <dim>-1</dim>
500 <dim>384</dim>
501 </port>
502 <port id="1" precision="I64">
503 <dim>4</dim>
504 </port>
505 </input>
506 <output>
507 <port id="2" precision="FP32" names="100,x.3">
508 <dim>-1</dim>
509 <dim>-1</dim>
510 <dim>12</dim>
511 <dim>32</dim>
512 </port>
513 </output>
514 </layer>
515 <layer id="38" name="Constant_84452" type="Const" version="opset1">
516 <data element_type="i64" shape="4" offset="48269880" size="32" />
517 <output>
518 <port id="0" precision="I64" names="101">
519 <dim>4</dim>
520 </port>
521 </output>
522 </layer>
523 <layer id="39" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
524 <input>
525 <port id="0" precision="FP32">
526 <dim>-1</dim>
527 <dim>-1</dim>
528 <dim>12</dim>
529 <dim>32</dim>
530 </port>
531 <port id="1" precision="I64">
532 <dim>4</dim>
533 </port>
534 </input>
535 <output>
536 <port id="2" precision="FP32" names="102">
537 <dim>-1</dim>
538 <dim>12</dim>
539 <dim>-1</dim>
540 <dim>32</dim>
541 </port>
542 </output>
543 </layer>
544 <layer id="40" name="self.bert.encoder.layer.0.attention.self.key.weight" type="Const" version="opset1">
545 <data element_type="f32" shape="384, 384" offset="48269912" size="589824" />
546 <output>
547 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.self.key.weight">
548 <dim>384</dim>
549 <dim>384</dim>
550 </port>
551 </output>
552 </layer>
553 <layer id="41" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
554 <data transpose_a="false" transpose_b="true" />
555 <input>
556 <port id="0" precision="FP32">
557 <dim>-1</dim>
558 <dim>-1</dim>
559 <dim>384</dim>
560 </port>
561 <port id="1" precision="FP32">
562 <dim>384</dim>
563 <dim>384</dim>
564 </port>
565 </input>
566 <output>
567 <port id="2" precision="FP32">
568 <dim>-1</dim>
569 <dim>-1</dim>
570 <dim>384</dim>
571 </port>
572 </output>
573 </layer>
574 <layer id="42" name="Constant_90304" type="Const" version="opset1">
575 <data element_type="f32" shape="1, 1, 384" offset="48859736" size="1536" />
576 <output>
577 <port id="0" precision="FP32">
578 <dim>1</dim>
579 <dim>1</dim>
580 <dim>384</dim>
581 </port>
582 </output>
583 </layer>
584 <layer id="43" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
585 <data auto_broadcast="numpy" />
586 <input>
587 <port id="0" precision="FP32">
588 <dim>-1</dim>
589 <dim>-1</dim>
590 <dim>384</dim>
591 </port>
592 <port id="1" precision="FP32">
593 <dim>1</dim>
594 <dim>1</dim>
595 <dim>384</dim>
596 </port>
597 </input>
598 <output>
599 <port id="2" precision="FP32" names="105,x.5">
600 <dim>-1</dim>
601 <dim>-1</dim>
602 <dim>384</dim>
603 </port>
604 </output>
605 </layer>
606 <layer id="44" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
607 <data element_type="i64" shape="4" offset="48269848" size="32" />
608 <output>
609 <port id="0" precision="I64">
610 <dim>4</dim>
611 </port>
612 </output>
613 </layer>
614 <layer id="45" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
615 <data special_zero="true" />
616 <input>
617 <port id="0" precision="FP32">
618 <dim>-1</dim>
619 <dim>-1</dim>
620 <dim>384</dim>
621 </port>
622 <port id="1" precision="I64">
623 <dim>4</dim>
624 </port>
625 </input>
626 <output>
627 <port id="2" precision="FP32" names="109,x.7">
628 <dim>-1</dim>
629 <dim>-1</dim>
630 <dim>12</dim>
631 <dim>32</dim>
632 </port>
633 </output>
634 </layer>
635 <layer id="46" name="Constant_84477" type="Const" version="opset1">
636 <data element_type="i64" shape="4" offset="48269880" size="32" />
637 <output>
638 <port id="0" precision="I64" names="110">
639 <dim>4</dim>
640 </port>
641 </output>
642 </layer>
643 <layer id="47" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
644 <input>
645 <port id="0" precision="FP32">
646 <dim>-1</dim>
647 <dim>-1</dim>
648 <dim>12</dim>
649 <dim>32</dim>
650 </port>
651 <port id="1" precision="I64">
652 <dim>4</dim>
653 </port>
654 </input>
655 <output>
656 <port id="2" precision="FP32" names="111">
657 <dim>-1</dim>
658 <dim>12</dim>
659 <dim>-1</dim>
660 <dim>32</dim>
661 </port>
662 </output>
663 </layer>
664 <layer id="48" name="self.bert.encoder.layer.0.attention.self.value.weight" type="Const" version="opset1">
665 <data element_type="f32" shape="384, 384" offset="48861272" size="589824" />
666 <output>
667 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.self.value.weight">
668 <dim>384</dim>
669 <dim>384</dim>
670 </port>
671 </output>
672 </layer>
673 <layer id="49" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
674 <data transpose_a="false" transpose_b="true" />
675 <input>
676 <port id="0" precision="FP32">
677 <dim>-1</dim>
678 <dim>-1</dim>
679 <dim>384</dim>
680 </port>
681 <port id="1" precision="FP32">
682 <dim>384</dim>
683 <dim>384</dim>
684 </port>
685 </input>
686 <output>
687 <port id="2" precision="FP32">
688 <dim>-1</dim>
689 <dim>-1</dim>
690 <dim>384</dim>
691 </port>
692 </output>
693 </layer>
694 <layer id="50" name="Constant_90305" type="Const" version="opset1">
695 <data element_type="f32" shape="1, 1, 384" offset="49451096" size="1536" />
696 <output>
697 <port id="0" precision="FP32">
698 <dim>1</dim>
699 <dim>1</dim>
700 <dim>384</dim>
701 </port>
702 </output>
703 </layer>
704 <layer id="51" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
705 <data auto_broadcast="numpy" />
706 <input>
707 <port id="0" precision="FP32">
708 <dim>-1</dim>
709 <dim>-1</dim>
710 <dim>384</dim>
711 </port>
712 <port id="1" precision="FP32">
713 <dim>1</dim>
714 <dim>1</dim>
715 <dim>384</dim>
716 </port>
717 </input>
718 <output>
719 <port id="2" precision="FP32" names="114,x.9">
720 <dim>-1</dim>
721 <dim>-1</dim>
722 <dim>384</dim>
723 </port>
724 </output>
725 </layer>
726 <layer id="52" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
727 <data element_type="i64" shape="4" offset="48269848" size="32" />
728 <output>
729 <port id="0" precision="I64">
730 <dim>4</dim>
731 </port>
732 </output>
733 </layer>
734 <layer id="53" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
735 <data special_zero="true" />
736 <input>
737 <port id="0" precision="FP32">
738 <dim>-1</dim>
739 <dim>-1</dim>
740 <dim>384</dim>
741 </port>
742 <port id="1" precision="I64">
743 <dim>4</dim>
744 </port>
745 </input>
746 <output>
747 <port id="2" precision="FP32" names="118,x.11">
748 <dim>-1</dim>
749 <dim>-1</dim>
750 <dim>12</dim>
751 <dim>32</dim>
752 </port>
753 </output>
754 </layer>
755 <layer id="54" name="Constant_84502" type="Const" version="opset1">
756 <data element_type="i64" shape="4" offset="48269880" size="32" />
757 <output>
758 <port id="0" precision="I64" names="119">
759 <dim>4</dim>
760 </port>
761 </output>
762 </layer>
763 <layer id="55" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
764 <input>
765 <port id="0" precision="FP32">
766 <dim>-1</dim>
767 <dim>-1</dim>
768 <dim>12</dim>
769 <dim>32</dim>
770 </port>
771 <port id="1" precision="I64">
772 <dim>4</dim>
773 </port>
774 </input>
775 <output>
776 <port id="2" precision="FP32" names="120">
777 <dim>-1</dim>
778 <dim>12</dim>
779 <dim>-1</dim>
780 <dim>32</dim>
781 </port>
782 </output>
783 </layer>
784 <layer id="56" name="Constant_90307" type="Const" version="opset1">
785 <data element_type="f32" shape="1, 1, 1, 1" offset="49452632" size="4" />
786 <output>
787 <port id="0" precision="FP32">
788 <dim>1</dim>
789 <dim>1</dim>
790 <dim>1</dim>
791 <dim>1</dim>
792 </port>
793 </output>
794 </layer>
795 <layer id="57" name="31" type="Const" version="opset1">
796 <data element_type="i64" shape="" offset="47675404" size="8" />
797 <output>
798 <port id="0" precision="I64" names="31" />
799 </output>
800 </layer>
801 <layer id="58" name="__module.bert/aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
802 <input>
803 <port id="0" precision="I64">
804 <dim>-1</dim>
805 <dim>-1</dim>
806 </port>
807 <port id="1" precision="I64" />
808 </input>
809 <output>
810 <port id="2" precision="I64" names="63">
811 <dim>-1</dim>
812 <dim>1</dim>
813 <dim>-1</dim>
814 </port>
815 </output>
816 </layer>
817 <layer id="59" name="21" type="Const" version="opset1">
818 <data element_type="i64" shape="" offset="49452636" size="8" />
819 <output>
820 <port id="0" precision="I64" names="21" />
821 </output>
822 </layer>
823 <layer id="60" name="__module.bert/aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
824 <input>
825 <port id="0" precision="I64">
826 <dim>-1</dim>
827 <dim>1</dim>
828 <dim>-1</dim>
829 </port>
830 <port id="1" precision="I64" />
831 </input>
832 <output>
833 <port id="2" precision="I64" names="64,65">
834 <dim>-1</dim>
835 <dim>1</dim>
836 <dim>1</dim>
837 <dim>-1</dim>
838 </port>
839 </output>
840 </layer>
841 <layer id="61" name="Constant_90436" type="Const" version="opset1">
842 <data element_type="i64" shape="1" offset="47675396" size="8" />
843 <output>
844 <port id="0" precision="I64">
845 <dim>1</dim>
846 </port>
847 </output>
848 </layer>
849 <layer id="62" name="Constant_90437" type="Const" version="opset1">
850 <data element_type="i64" shape="" offset="47675396" size="8" />
851 <output>
852 <port id="0" precision="I64" />
853 </output>
854 </layer>
855 <layer id="63" name="Gather_90438" type="Gather" version="opset8">
856 <data batch_dims="0" />
857 <input>
858 <port id="0" precision="I64">
859 <dim>2</dim>
860 </port>
861 <port id="1" precision="I64">
862 <dim>1</dim>
863 </port>
864 <port id="2" precision="I64" />
865 </input>
866 <output>
867 <port id="3" precision="I64" names="60">
868 <dim>1</dim>
869 </port>
870 </output>
871 </layer>
872 <layer id="64" name="Constant_89535" type="Const" version="opset1">
873 <data element_type="i64" shape="1" offset="47675404" size="8" />
874 <output>
875 <port id="0" precision="I64">
876 <dim>1</dim>
877 </port>
878 </output>
879 </layer>
880 <layer id="65" name="Constant_90548" type="Const" version="opset1">
881 <data element_type="i64" shape="2" offset="49452644" size="16" />
882 <output>
883 <port id="0" precision="I64">
884 <dim>2</dim>
885 </port>
886 </output>
887 </layer>
888 <layer id="66" name="Constant_90549" type="Const" version="opset1">
889 <data element_type="i64" shape="" offset="47675396" size="8" />
890 <output>
891 <port id="0" precision="I64" />
892 </output>
893 </layer>
894 <layer id="67" name="Gather_90550" type="Gather" version="opset8">
895 <data batch_dims="0" />
896 <input>
897 <port id="0" precision="I64">
898 <dim>2</dim>
899 </port>
900 <port id="1" precision="I64">
901 <dim>2</dim>
902 </port>
903 <port id="2" precision="I64" />
904 </input>
905 <output>
906 <port id="3" precision="I64">
907 <dim>2</dim>
908 </port>
909 </output>
910 </layer>
911 <layer id="68" name="__module.bert/prim::ListConstruct/Concat" type="Concat" version="opset1">
912 <data axis="0" />
913 <input>
914 <port id="0" precision="I64">
915 <dim>1</dim>
916 </port>
917 <port id="1" precision="I64">
918 <dim>1</dim>
919 </port>
920 <port id="2" precision="I64">
921 <dim>2</dim>
922 </port>
923 </input>
924 <output>
925 <port id="3" precision="I64" names="66">
926 <dim>4</dim>
927 </port>
928 </output>
929 </layer>
930 <layer id="69" name="__module.bert/aten::expand/Broadcast" type="Broadcast" version="opset3">
931 <data mode="bidirectional" />
932 <input>
933 <port id="0" precision="I64">
934 <dim>-1</dim>
935 <dim>1</dim>
936 <dim>1</dim>
937 <dim>-1</dim>
938 </port>
939 <port id="1" precision="I64">
940 <dim>4</dim>
941 </port>
942 </input>
943 <output>
944 <port id="2" precision="I64" names="67">
945 <dim>-1</dim>
946 <dim>1</dim>
947 <dim>-1</dim>
948 <dim>-1</dim>
949 </port>
950 </output>
951 </layer>
952 <layer id="70" name="__module.bert/aten::to/Convert" type="Convert" version="opset1">
953 <data destination_type="f32" />
954 <input>
955 <port id="0" precision="I64">
956 <dim>-1</dim>
957 <dim>1</dim>
958 <dim>-1</dim>
959 <dim>-1</dim>
960 </port>
961 </input>
962 <output>
963 <port id="1" precision="FP32" names="68">
964 <dim>-1</dim>
965 <dim>1</dim>
966 <dim>-1</dim>
967 <dim>-1</dim>
968 </port>
969 </output>
970 </layer>
971 <layer id="71" name="Constant_90306" type="Const" version="opset1">
972 <data element_type="f32" shape="1, 1, 1, 1" offset="49452632" size="4" />
973 <output>
974 <port id="0" precision="FP32">
975 <dim>1</dim>
976 <dim>1</dim>
977 <dim>1</dim>
978 <dim>1</dim>
979 </port>
980 </output>
981 </layer>
982 <layer id="72" name="__module.bert/aten::rsub/Multiply" type="Multiply" version="opset1">
983 <data auto_broadcast="numpy" />
984 <input>
985 <port id="0" precision="FP32">
986 <dim>-1</dim>
987 <dim>1</dim>
988 <dim>-1</dim>
989 <dim>-1</dim>
990 </port>
991 <port id="1" precision="FP32">
992 <dim>1</dim>
993 <dim>1</dim>
994 <dim>1</dim>
995 <dim>1</dim>
996 </port>
997 </input>
998 <output>
999 <port id="2" precision="FP32">
1000 <dim>-1</dim>
1001 <dim>1</dim>
1002 <dim>-1</dim>
1003 <dim>-1</dim>
1004 </port>
1005 </output>
1006 </layer>
1007 <layer id="73" name="__module.bert/aten::rsub/Subtract" type="Subtract" version="opset1">
1008 <data auto_broadcast="numpy" />
1009 <input>
1010 <port id="0" precision="FP32">
1011 <dim>1</dim>
1012 <dim>1</dim>
1013 <dim>1</dim>
1014 <dim>1</dim>
1015 </port>
1016 <port id="1" precision="FP32">
1017 <dim>-1</dim>
1018 <dim>1</dim>
1019 <dim>-1</dim>
1020 <dim>-1</dim>
1021 </port>
1022 </input>
1023 <output>
1024 <port id="2" precision="FP32" names="69,inverted_mask">
1025 <dim>-1</dim>
1026 <dim>1</dim>
1027 <dim>-1</dim>
1028 <dim>-1</dim>
1029 </port>
1030 </output>
1031 </layer>
1032 <layer id="74" name="__module.bert/aten::to/Convert_1" type="Convert" version="opset1">
1033 <data destination_type="boolean" />
1034 <input>
1035 <port id="0" precision="FP32">
1036 <dim>-1</dim>
1037 <dim>1</dim>
1038 <dim>-1</dim>
1039 <dim>-1</dim>
1040 </port>
1041 </input>
1042 <output>
1043 <port id="1" precision="BOOL" names="70">
1044 <dim>-1</dim>
1045 <dim>1</dim>
1046 <dim>-1</dim>
1047 <dim>-1</dim>
1048 </port>
1049 </output>
1050 </layer>
1051 <layer id="75" name="__module.bert/aten::masked_fill/ConvertLike" type="Const" version="opset1">
1052 <data element_type="f32" shape="" offset="49452660" size="4" />
1053 <output>
1054 <port id="0" precision="FP32" />
1055 </output>
1056 </layer>
1057 <layer id="76" name="__module.bert/aten::masked_fill/Select" type="Select" version="opset1">
1058 <data auto_broadcast="numpy" />
1059 <input>
1060 <port id="0" precision="BOOL">
1061 <dim>-1</dim>
1062 <dim>1</dim>
1063 <dim>-1</dim>
1064 <dim>-1</dim>
1065 </port>
1066 <port id="1" precision="FP32" />
1067 <port id="2" precision="FP32">
1068 <dim>-1</dim>
1069 <dim>1</dim>
1070 <dim>-1</dim>
1071 <dim>-1</dim>
1072 </port>
1073 </input>
1074 <output>
1075 <port id="3" precision="FP32" names="71">
1076 <dim>-1</dim>
1077 <dim>1</dim>
1078 <dim>-1</dim>
1079 <dim>-1</dim>
1080 </port>
1081 </output>
1082 </layer>
1083 <layer id="77" name="__module.bert.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1084 <data causal="false" />
1085 <input>
1086 <port id="0" precision="FP32">
1087 <dim>-1</dim>
1088 <dim>12</dim>
1089 <dim>-1</dim>
1090 <dim>32</dim>
1091 </port>
1092 <port id="1" precision="FP32">
1093 <dim>-1</dim>
1094 <dim>12</dim>
1095 <dim>-1</dim>
1096 <dim>32</dim>
1097 </port>
1098 <port id="2" precision="FP32">
1099 <dim>-1</dim>
1100 <dim>12</dim>
1101 <dim>-1</dim>
1102 <dim>32</dim>
1103 </port>
1104 <port id="3" precision="FP32">
1105 <dim>-1</dim>
1106 <dim>1</dim>
1107 <dim>-1</dim>
1108 <dim>-1</dim>
1109 </port>
1110 </input>
1111 <output>
1112 <port id="4" precision="FP32" names="121,attn_output.1">
1113 <dim>-1</dim>
1114 <dim>12</dim>
1115 <dim>-1</dim>
1116 <dim>32</dim>
1117 </port>
1118 </output>
1119 </layer>
1120 <layer id="78" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1121 <data element_type="i32" shape="4" offset="49452664" size="16" />
1122 <output>
1123 <port id="0" precision="I32">
1124 <dim>4</dim>
1125 </port>
1126 </output>
1127 </layer>
1128 <layer id="79" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1129 <input>
1130 <port id="0" precision="FP32">
1131 <dim>-1</dim>
1132 <dim>12</dim>
1133 <dim>-1</dim>
1134 <dim>32</dim>
1135 </port>
1136 <port id="1" precision="I32">
1137 <dim>4</dim>
1138 </port>
1139 </input>
1140 <output>
1141 <port id="2" precision="FP32" names="122,attn_output.3">
1142 <dim>-1</dim>
1143 <dim>-1</dim>
1144 <dim>12</dim>
1145 <dim>32</dim>
1146 </port>
1147 </output>
1148 </layer>
1149 <layer id="80" name="Constant_90443" type="Const" version="opset1">
1150 <data element_type="i64" shape="3" offset="49452680" size="24" />
1151 <output>
1152 <port id="0" precision="I64">
1153 <dim>3</dim>
1154 </port>
1155 </output>
1156 </layer>
1157 <layer id="81" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1158 <data special_zero="true" />
1159 <input>
1160 <port id="0" precision="FP32">
1161 <dim>-1</dim>
1162 <dim>-1</dim>
1163 <dim>12</dim>
1164 <dim>32</dim>
1165 </port>
1166 <port id="1" precision="I64">
1167 <dim>3</dim>
1168 </port>
1169 </input>
1170 <output>
1171 <port id="2" precision="FP32" names="124">
1172 <dim>-1</dim>
1173 <dim>-1</dim>
1174 <dim>384</dim>
1175 </port>
1176 </output>
1177 </layer>
1178 <layer id="82" name="self.bert.encoder.layer.0.attention.output.dense.weight" type="Const" version="opset1">
1179 <data element_type="f32" shape="384, 384" offset="49452704" size="589824" />
1180 <output>
1181 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.output.dense.weight">
1182 <dim>384</dim>
1183 <dim>384</dim>
1184 </port>
1185 </output>
1186 </layer>
1187 <layer id="83" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1188 <data transpose_a="false" transpose_b="true" />
1189 <input>
1190 <port id="0" precision="FP32">
1191 <dim>-1</dim>
1192 <dim>-1</dim>
1193 <dim>384</dim>
1194 </port>
1195 <port id="1" precision="FP32">
1196 <dim>384</dim>
1197 <dim>384</dim>
1198 </port>
1199 </input>
1200 <output>
1201 <port id="2" precision="FP32">
1202 <dim>-1</dim>
1203 <dim>-1</dim>
1204 <dim>384</dim>
1205 </port>
1206 </output>
1207 </layer>
1208 <layer id="84" name="Constant_90308" type="Const" version="opset1">
1209 <data element_type="f32" shape="1, 1, 384" offset="50042528" size="1536" />
1210 <output>
1211 <port id="0" precision="FP32">
1212 <dim>1</dim>
1213 <dim>1</dim>
1214 <dim>384</dim>
1215 </port>
1216 </output>
1217 </layer>
1218 <layer id="85" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1219 <data auto_broadcast="numpy" />
1220 <input>
1221 <port id="0" precision="FP32">
1222 <dim>-1</dim>
1223 <dim>-1</dim>
1224 <dim>384</dim>
1225 </port>
1226 <port id="1" precision="FP32">
1227 <dim>1</dim>
1228 <dim>1</dim>
1229 <dim>384</dim>
1230 </port>
1231 </input>
1232 <output>
1233 <port id="2" precision="FP32" names="129,input.3">
1234 <dim>-1</dim>
1235 <dim>-1</dim>
1236 <dim>384</dim>
1237 </port>
1238 </output>
1239 </layer>
1240 <layer id="86" name="__module.bert.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1241 <data auto_broadcast="numpy" />
1242 <input>
1243 <port id="0" precision="FP32">
1244 <dim>-1</dim>
1245 <dim>-1</dim>
1246 <dim>384</dim>
1247 </port>
1248 <port id="1" precision="FP32">
1249 <dim>-1</dim>
1250 <dim>-1</dim>
1251 <dim>384</dim>
1252 </port>
1253 </input>
1254 <output>
1255 <port id="2" precision="FP32" names="131">
1256 <dim>-1</dim>
1257 <dim>-1</dim>
1258 <dim>384</dim>
1259 </port>
1260 </output>
1261 </layer>
1262 <layer id="87" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1263 <data element_type="i32" shape="1" offset="47675412" size="4" />
1264 <output>
1265 <port id="0" precision="I32">
1266 <dim>1</dim>
1267 </port>
1268 </output>
1269 </layer>
1270 <layer id="88" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1271 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1272 <input>
1273 <port id="0" precision="FP32">
1274 <dim>-1</dim>
1275 <dim>-1</dim>
1276 <dim>384</dim>
1277 </port>
1278 <port id="1" precision="I32">
1279 <dim>1</dim>
1280 </port>
1281 </input>
1282 <output>
1283 <port id="2" precision="FP32">
1284 <dim>-1</dim>
1285 <dim>-1</dim>
1286 <dim>384</dim>
1287 </port>
1288 </output>
1289 </layer>
1290 <layer id="89" name="Constant_90309" type="Const" version="opset1">
1291 <data element_type="f32" shape="1, 1, 384" offset="50044064" size="1536" />
1292 <output>
1293 <port id="0" precision="FP32">
1294 <dim>1</dim>
1295 <dim>1</dim>
1296 <dim>384</dim>
1297 </port>
1298 </output>
1299 </layer>
1300 <layer id="90" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1301 <data auto_broadcast="numpy" />
1302 <input>
1303 <port id="0" precision="FP32">
1304 <dim>-1</dim>
1305 <dim>-1</dim>
1306 <dim>384</dim>
1307 </port>
1308 <port id="1" precision="FP32">
1309 <dim>1</dim>
1310 <dim>1</dim>
1311 <dim>384</dim>
1312 </port>
1313 </input>
1314 <output>
1315 <port id="2" precision="FP32">
1316 <dim>-1</dim>
1317 <dim>-1</dim>
1318 <dim>384</dim>
1319 </port>
1320 </output>
1321 </layer>
1322 <layer id="91" name="Constant_90310" type="Const" version="opset1">
1323 <data element_type="f32" shape="1, 1, 384" offset="50045600" size="1536" />
1324 <output>
1325 <port id="0" precision="FP32">
1326 <dim>1</dim>
1327 <dim>1</dim>
1328 <dim>384</dim>
1329 </port>
1330 </output>
1331 </layer>
1332 <layer id="92" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1333 <data auto_broadcast="numpy" />
1334 <input>
1335 <port id="0" precision="FP32">
1336 <dim>-1</dim>
1337 <dim>-1</dim>
1338 <dim>384</dim>
1339 </port>
1340 <port id="1" precision="FP32">
1341 <dim>1</dim>
1342 <dim>1</dim>
1343 <dim>384</dim>
1344 </port>
1345 </input>
1346 <output>
1347 <port id="2" precision="FP32" names="135,input_tensor.1">
1348 <dim>-1</dim>
1349 <dim>-1</dim>
1350 <dim>384</dim>
1351 </port>
1352 </output>
1353 </layer>
1354 <layer id="93" name="self.bert.encoder.layer.0.intermediate.dense.weight" type="Const" version="opset1">
1355 <data element_type="f32" shape="1536, 384" offset="50047136" size="2359296" />
1356 <output>
1357 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.intermediate.dense.weight">
1358 <dim>1536</dim>
1359 <dim>384</dim>
1360 </port>
1361 </output>
1362 </layer>
1363 <layer id="94" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1364 <data transpose_a="false" transpose_b="true" />
1365 <input>
1366 <port id="0" precision="FP32">
1367 <dim>-1</dim>
1368 <dim>-1</dim>
1369 <dim>384</dim>
1370 </port>
1371 <port id="1" precision="FP32">
1372 <dim>1536</dim>
1373 <dim>384</dim>
1374 </port>
1375 </input>
1376 <output>
1377 <port id="2" precision="FP32">
1378 <dim>-1</dim>
1379 <dim>-1</dim>
1380 <dim>1536</dim>
1381 </port>
1382 </output>
1383 </layer>
1384 <layer id="95" name="Constant_90311" type="Const" version="opset1">
1385 <data element_type="f32" shape="1, 1, 1536" offset="52406432" size="6144" />
1386 <output>
1387 <port id="0" precision="FP32">
1388 <dim>1</dim>
1389 <dim>1</dim>
1390 <dim>1536</dim>
1391 </port>
1392 </output>
1393 </layer>
1394 <layer id="96" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
1395 <data auto_broadcast="numpy" />
1396 <input>
1397 <port id="0" precision="FP32">
1398 <dim>-1</dim>
1399 <dim>-1</dim>
1400 <dim>1536</dim>
1401 </port>
1402 <port id="1" precision="FP32">
1403 <dim>1</dim>
1404 <dim>1</dim>
1405 <dim>1536</dim>
1406 </port>
1407 </input>
1408 <output>
1409 <port id="2" precision="FP32" names="139">
1410 <dim>-1</dim>
1411 <dim>-1</dim>
1412 <dim>1536</dim>
1413 </port>
1414 </output>
1415 </layer>
1416 <layer id="97" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
1417 <data approximation_mode="ERF" />
1418 <input>
1419 <port id="0" precision="FP32">
1420 <dim>-1</dim>
1421 <dim>-1</dim>
1422 <dim>1536</dim>
1423 </port>
1424 </input>
1425 <output>
1426 <port id="1" precision="FP32" names="140">
1427 <dim>-1</dim>
1428 <dim>-1</dim>
1429 <dim>1536</dim>
1430 </port>
1431 </output>
1432 </layer>
1433 <layer id="98" name="self.bert.encoder.layer.0.output.dense.weight" type="Const" version="opset1">
1434 <data element_type="f32" shape="384, 1536" offset="52412576" size="2359296" />
1435 <output>
1436 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.output.dense.weight">
1437 <dim>384</dim>
1438 <dim>1536</dim>
1439 </port>
1440 </output>
1441 </layer>
1442 <layer id="99" name="__module.bert.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1443 <data transpose_a="false" transpose_b="true" />
1444 <input>
1445 <port id="0" precision="FP32">
1446 <dim>-1</dim>
1447 <dim>-1</dim>
1448 <dim>1536</dim>
1449 </port>
1450 <port id="1" precision="FP32">
1451 <dim>384</dim>
1452 <dim>1536</dim>
1453 </port>
1454 </input>
1455 <output>
1456 <port id="2" precision="FP32">
1457 <dim>-1</dim>
1458 <dim>-1</dim>
1459 <dim>384</dim>
1460 </port>
1461 </output>
1462 </layer>
1463 <layer id="100" name="Constant_90312" type="Const" version="opset1">
1464 <data element_type="f32" shape="1, 1, 384" offset="54771872" size="1536" />
1465 <output>
1466 <port id="0" precision="FP32">
1467 <dim>1</dim>
1468 <dim>1</dim>
1469 <dim>384</dim>
1470 </port>
1471 </output>
1472 </layer>
1473 <layer id="101" name="__module.bert.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
1474 <data auto_broadcast="numpy" />
1475 <input>
1476 <port id="0" precision="FP32">
1477 <dim>-1</dim>
1478 <dim>-1</dim>
1479 <dim>384</dim>
1480 </port>
1481 <port id="1" precision="FP32">
1482 <dim>1</dim>
1483 <dim>1</dim>
1484 <dim>384</dim>
1485 </port>
1486 </input>
1487 <output>
1488 <port id="2" precision="FP32" names="145,input.5">
1489 <dim>-1</dim>
1490 <dim>-1</dim>
1491 <dim>384</dim>
1492 </port>
1493 </output>
1494 </layer>
1495 <layer id="102" name="__module.bert.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
1496 <data auto_broadcast="numpy" />
1497 <input>
1498 <port id="0" precision="FP32">
1499 <dim>-1</dim>
1500 <dim>-1</dim>
1501 <dim>384</dim>
1502 </port>
1503 <port id="1" precision="FP32">
1504 <dim>-1</dim>
1505 <dim>-1</dim>
1506 <dim>384</dim>
1507 </port>
1508 </input>
1509 <output>
1510 <port id="2" precision="FP32" names="147">
1511 <dim>-1</dim>
1512 <dim>-1</dim>
1513 <dim>384</dim>
1514 </port>
1515 </output>
1516 </layer>
1517 <layer id="103" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1518 <data element_type="i32" shape="1" offset="47675412" size="4" />
1519 <output>
1520 <port id="0" precision="I32">
1521 <dim>1</dim>
1522 </port>
1523 </output>
1524 </layer>
1525 <layer id="104" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1526 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1527 <input>
1528 <port id="0" precision="FP32">
1529 <dim>-1</dim>
1530 <dim>-1</dim>
1531 <dim>384</dim>
1532 </port>
1533 <port id="1" precision="I32">
1534 <dim>1</dim>
1535 </port>
1536 </input>
1537 <output>
1538 <port id="2" precision="FP32">
1539 <dim>-1</dim>
1540 <dim>-1</dim>
1541 <dim>384</dim>
1542 </port>
1543 </output>
1544 </layer>
1545 <layer id="105" name="Constant_90313" type="Const" version="opset1">
1546 <data element_type="f32" shape="1, 1, 384" offset="54773408" size="1536" />
1547 <output>
1548 <port id="0" precision="FP32">
1549 <dim>1</dim>
1550 <dim>1</dim>
1551 <dim>384</dim>
1552 </port>
1553 </output>
1554 </layer>
1555 <layer id="106" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1556 <data auto_broadcast="numpy" />
1557 <input>
1558 <port id="0" precision="FP32">
1559 <dim>-1</dim>
1560 <dim>-1</dim>
1561 <dim>384</dim>
1562 </port>
1563 <port id="1" precision="FP32">
1564 <dim>1</dim>
1565 <dim>1</dim>
1566 <dim>384</dim>
1567 </port>
1568 </input>
1569 <output>
1570 <port id="2" precision="FP32">
1571 <dim>-1</dim>
1572 <dim>-1</dim>
1573 <dim>384</dim>
1574 </port>
1575 </output>
1576 </layer>
1577 <layer id="107" name="Constant_90314" type="Const" version="opset1">
1578 <data element_type="f32" shape="1, 1, 384" offset="54774944" size="1536" />
1579 <output>
1580 <port id="0" precision="FP32">
1581 <dim>1</dim>
1582 <dim>1</dim>
1583 <dim>384</dim>
1584 </port>
1585 </output>
1586 </layer>
1587 <layer id="108" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1588 <data auto_broadcast="numpy" />
1589 <input>
1590 <port id="0" precision="FP32">
1591 <dim>-1</dim>
1592 <dim>-1</dim>
1593 <dim>384</dim>
1594 </port>
1595 <port id="1" precision="FP32">
1596 <dim>1</dim>
1597 <dim>1</dim>
1598 <dim>384</dim>
1599 </port>
1600 </input>
1601 <output>
1602 <port id="2" precision="FP32" names="151,hidden_states.7">
1603 <dim>-1</dim>
1604 <dim>-1</dim>
1605 <dim>384</dim>
1606 </port>
1607 </output>
1608 </layer>
1609 <layer id="109" name="self.bert.encoder.layer.1.attention.self.query.weight" type="Const" version="opset1">
1610 <data element_type="f32" shape="384, 384" offset="54776480" size="589824" />
1611 <output>
1612 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.self.query.weight">
1613 <dim>384</dim>
1614 <dim>384</dim>
1615 </port>
1616 </output>
1617 </layer>
1618 <layer id="110" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
1619 <data transpose_a="false" transpose_b="true" />
1620 <input>
1621 <port id="0" precision="FP32">
1622 <dim>-1</dim>
1623 <dim>-1</dim>
1624 <dim>384</dim>
1625 </port>
1626 <port id="1" precision="FP32">
1627 <dim>384</dim>
1628 <dim>384</dim>
1629 </port>
1630 </input>
1631 <output>
1632 <port id="2" precision="FP32">
1633 <dim>-1</dim>
1634 <dim>-1</dim>
1635 <dim>384</dim>
1636 </port>
1637 </output>
1638 </layer>
1639 <layer id="111" name="Constant_90315" type="Const" version="opset1">
1640 <data element_type="f32" shape="1, 1, 384" offset="55366304" size="1536" />
1641 <output>
1642 <port id="0" precision="FP32">
1643 <dim>1</dim>
1644 <dim>1</dim>
1645 <dim>384</dim>
1646 </port>
1647 </output>
1648 </layer>
1649 <layer id="112" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
1650 <data auto_broadcast="numpy" />
1651 <input>
1652 <port id="0" precision="FP32">
1653 <dim>-1</dim>
1654 <dim>-1</dim>
1655 <dim>384</dim>
1656 </port>
1657 <port id="1" precision="FP32">
1658 <dim>1</dim>
1659 <dim>1</dim>
1660 <dim>384</dim>
1661 </port>
1662 </input>
1663 <output>
1664 <port id="2" precision="FP32" names="164,x.13">
1665 <dim>-1</dim>
1666 <dim>-1</dim>
1667 <dim>384</dim>
1668 </port>
1669 </output>
1670 </layer>
1671 <layer id="113" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
1672 <data element_type="i64" shape="4" offset="48269848" size="32" />
1673 <output>
1674 <port id="0" precision="I64">
1675 <dim>4</dim>
1676 </port>
1677 </output>
1678 </layer>
1679 <layer id="114" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
1680 <data special_zero="true" />
1681 <input>
1682 <port id="0" precision="FP32">
1683 <dim>-1</dim>
1684 <dim>-1</dim>
1685 <dim>384</dim>
1686 </port>
1687 <port id="1" precision="I64">
1688 <dim>4</dim>
1689 </port>
1690 </input>
1691 <output>
1692 <port id="2" precision="FP32" names="168,x.15">
1693 <dim>-1</dim>
1694 <dim>-1</dim>
1695 <dim>12</dim>
1696 <dim>32</dim>
1697 </port>
1698 </output>
1699 </layer>
1700 <layer id="115" name="Constant_84681" type="Const" version="opset1">
1701 <data element_type="i64" shape="4" offset="48269880" size="32" />
1702 <output>
1703 <port id="0" precision="I64" names="169">
1704 <dim>4</dim>
1705 </port>
1706 </output>
1707 </layer>
1708 <layer id="116" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
1709 <input>
1710 <port id="0" precision="FP32">
1711 <dim>-1</dim>
1712 <dim>-1</dim>
1713 <dim>12</dim>
1714 <dim>32</dim>
1715 </port>
1716 <port id="1" precision="I64">
1717 <dim>4</dim>
1718 </port>
1719 </input>
1720 <output>
1721 <port id="2" precision="FP32" names="170">
1722 <dim>-1</dim>
1723 <dim>12</dim>
1724 <dim>-1</dim>
1725 <dim>32</dim>
1726 </port>
1727 </output>
1728 </layer>
1729 <layer id="117" name="self.bert.encoder.layer.1.attention.self.key.weight" type="Const" version="opset1">
1730 <data element_type="f32" shape="384, 384" offset="55367840" size="589824" />
1731 <output>
1732 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.self.key.weight">
1733 <dim>384</dim>
1734 <dim>384</dim>
1735 </port>
1736 </output>
1737 </layer>
1738 <layer id="118" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
1739 <data transpose_a="false" transpose_b="true" />
1740 <input>
1741 <port id="0" precision="FP32">
1742 <dim>-1</dim>
1743 <dim>-1</dim>
1744 <dim>384</dim>
1745 </port>
1746 <port id="1" precision="FP32">
1747 <dim>384</dim>
1748 <dim>384</dim>
1749 </port>
1750 </input>
1751 <output>
1752 <port id="2" precision="FP32">
1753 <dim>-1</dim>
1754 <dim>-1</dim>
1755 <dim>384</dim>
1756 </port>
1757 </output>
1758 </layer>
1759 <layer id="119" name="Constant_90316" type="Const" version="opset1">
1760 <data element_type="f32" shape="1, 1, 384" offset="55957664" size="1536" />
1761 <output>
1762 <port id="0" precision="FP32">
1763 <dim>1</dim>
1764 <dim>1</dim>
1765 <dim>384</dim>
1766 </port>
1767 </output>
1768 </layer>
1769 <layer id="120" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
1770 <data auto_broadcast="numpy" />
1771 <input>
1772 <port id="0" precision="FP32">
1773 <dim>-1</dim>
1774 <dim>-1</dim>
1775 <dim>384</dim>
1776 </port>
1777 <port id="1" precision="FP32">
1778 <dim>1</dim>
1779 <dim>1</dim>
1780 <dim>384</dim>
1781 </port>
1782 </input>
1783 <output>
1784 <port id="2" precision="FP32" names="173,x.17">
1785 <dim>-1</dim>
1786 <dim>-1</dim>
1787 <dim>384</dim>
1788 </port>
1789 </output>
1790 </layer>
1791 <layer id="121" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
1792 <data element_type="i64" shape="4" offset="48269848" size="32" />
1793 <output>
1794 <port id="0" precision="I64">
1795 <dim>4</dim>
1796 </port>
1797 </output>
1798 </layer>
1799 <layer id="122" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
1800 <data special_zero="true" />
1801 <input>
1802 <port id="0" precision="FP32">
1803 <dim>-1</dim>
1804 <dim>-1</dim>
1805 <dim>384</dim>
1806 </port>
1807 <port id="1" precision="I64">
1808 <dim>4</dim>
1809 </port>
1810 </input>
1811 <output>
1812 <port id="2" precision="FP32" names="177,x.19">
1813 <dim>-1</dim>
1814 <dim>-1</dim>
1815 <dim>12</dim>
1816 <dim>32</dim>
1817 </port>
1818 </output>
1819 </layer>
1820 <layer id="123" name="Constant_84704" type="Const" version="opset1">
1821 <data element_type="i64" shape="4" offset="48269880" size="32" />
1822 <output>
1823 <port id="0" precision="I64" names="178">
1824 <dim>4</dim>
1825 </port>
1826 </output>
1827 </layer>
1828 <layer id="124" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1829 <input>
1830 <port id="0" precision="FP32">
1831 <dim>-1</dim>
1832 <dim>-1</dim>
1833 <dim>12</dim>
1834 <dim>32</dim>
1835 </port>
1836 <port id="1" precision="I64">
1837 <dim>4</dim>
1838 </port>
1839 </input>
1840 <output>
1841 <port id="2" precision="FP32" names="179">
1842 <dim>-1</dim>
1843 <dim>12</dim>
1844 <dim>-1</dim>
1845 <dim>32</dim>
1846 </port>
1847 </output>
1848 </layer>
1849 <layer id="125" name="self.bert.encoder.layer.1.attention.self.value.weight" type="Const" version="opset1">
1850 <data element_type="f32" shape="384, 384" offset="55959200" size="589824" />
1851 <output>
1852 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.self.value.weight">
1853 <dim>384</dim>
1854 <dim>384</dim>
1855 </port>
1856 </output>
1857 </layer>
1858 <layer id="126" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1859 <data transpose_a="false" transpose_b="true" />
1860 <input>
1861 <port id="0" precision="FP32">
1862 <dim>-1</dim>
1863 <dim>-1</dim>
1864 <dim>384</dim>
1865 </port>
1866 <port id="1" precision="FP32">
1867 <dim>384</dim>
1868 <dim>384</dim>
1869 </port>
1870 </input>
1871 <output>
1872 <port id="2" precision="FP32">
1873 <dim>-1</dim>
1874 <dim>-1</dim>
1875 <dim>384</dim>
1876 </port>
1877 </output>
1878 </layer>
1879 <layer id="127" name="Constant_90317" type="Const" version="opset1">
1880 <data element_type="f32" shape="1, 1, 384" offset="56549024" size="1536" />
1881 <output>
1882 <port id="0" precision="FP32">
1883 <dim>1</dim>
1884 <dim>1</dim>
1885 <dim>384</dim>
1886 </port>
1887 </output>
1888 </layer>
1889 <layer id="128" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
1890 <data auto_broadcast="numpy" />
1891 <input>
1892 <port id="0" precision="FP32">
1893 <dim>-1</dim>
1894 <dim>-1</dim>
1895 <dim>384</dim>
1896 </port>
1897 <port id="1" precision="FP32">
1898 <dim>1</dim>
1899 <dim>1</dim>
1900 <dim>384</dim>
1901 </port>
1902 </input>
1903 <output>
1904 <port id="2" precision="FP32" names="182,x.21">
1905 <dim>-1</dim>
1906 <dim>-1</dim>
1907 <dim>384</dim>
1908 </port>
1909 </output>
1910 </layer>
1911 <layer id="129" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
1912 <data element_type="i64" shape="4" offset="48269848" size="32" />
1913 <output>
1914 <port id="0" precision="I64">
1915 <dim>4</dim>
1916 </port>
1917 </output>
1918 </layer>
1919 <layer id="130" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
1920 <data special_zero="true" />
1921 <input>
1922 <port id="0" precision="FP32">
1923 <dim>-1</dim>
1924 <dim>-1</dim>
1925 <dim>384</dim>
1926 </port>
1927 <port id="1" precision="I64">
1928 <dim>4</dim>
1929 </port>
1930 </input>
1931 <output>
1932 <port id="2" precision="FP32" names="186,x.23">
1933 <dim>-1</dim>
1934 <dim>-1</dim>
1935 <dim>12</dim>
1936 <dim>32</dim>
1937 </port>
1938 </output>
1939 </layer>
1940 <layer id="131" name="Constant_84727" type="Const" version="opset1">
1941 <data element_type="i64" shape="4" offset="48269880" size="32" />
1942 <output>
1943 <port id="0" precision="I64" names="187">
1944 <dim>4</dim>
1945 </port>
1946 </output>
1947 </layer>
1948 <layer id="132" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
1949 <input>
1950 <port id="0" precision="FP32">
1951 <dim>-1</dim>
1952 <dim>-1</dim>
1953 <dim>12</dim>
1954 <dim>32</dim>
1955 </port>
1956 <port id="1" precision="I64">
1957 <dim>4</dim>
1958 </port>
1959 </input>
1960 <output>
1961 <port id="2" precision="FP32" names="188">
1962 <dim>-1</dim>
1963 <dim>12</dim>
1964 <dim>-1</dim>
1965 <dim>32</dim>
1966 </port>
1967 </output>
1968 </layer>
1969 <layer id="133" name="__module.bert.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1970 <data causal="false" />
1971 <input>
1972 <port id="0" precision="FP32">
1973 <dim>-1</dim>
1974 <dim>12</dim>
1975 <dim>-1</dim>
1976 <dim>32</dim>
1977 </port>
1978 <port id="1" precision="FP32">
1979 <dim>-1</dim>
1980 <dim>12</dim>
1981 <dim>-1</dim>
1982 <dim>32</dim>
1983 </port>
1984 <port id="2" precision="FP32">
1985 <dim>-1</dim>
1986 <dim>12</dim>
1987 <dim>-1</dim>
1988 <dim>32</dim>
1989 </port>
1990 <port id="3" precision="FP32">
1991 <dim>-1</dim>
1992 <dim>1</dim>
1993 <dim>-1</dim>
1994 <dim>-1</dim>
1995 </port>
1996 </input>
1997 <output>
1998 <port id="4" precision="FP32" names="189,attn_output.5">
1999 <dim>-1</dim>
2000 <dim>12</dim>
2001 <dim>-1</dim>
2002 <dim>32</dim>
2003 </port>
2004 </output>
2005 </layer>
2006 <layer id="134" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
2007 <data element_type="i32" shape="4" offset="49452664" size="16" />
2008 <output>
2009 <port id="0" precision="I32">
2010 <dim>4</dim>
2011 </port>
2012 </output>
2013 </layer>
2014 <layer id="135" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
2015 <input>
2016 <port id="0" precision="FP32">
2017 <dim>-1</dim>
2018 <dim>12</dim>
2019 <dim>-1</dim>
2020 <dim>32</dim>
2021 </port>
2022 <port id="1" precision="I32">
2023 <dim>4</dim>
2024 </port>
2025 </input>
2026 <output>
2027 <port id="2" precision="FP32" names="190,attn_output.7">
2028 <dim>-1</dim>
2029 <dim>-1</dim>
2030 <dim>12</dim>
2031 <dim>32</dim>
2032 </port>
2033 </output>
2034 </layer>
2035 <layer id="136" name="Constant_90444" type="Const" version="opset1">
2036 <data element_type="i64" shape="3" offset="49452680" size="24" />
2037 <output>
2038 <port id="0" precision="I64">
2039 <dim>3</dim>
2040 </port>
2041 </output>
2042 </layer>
2043 <layer id="137" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
2044 <data special_zero="true" />
2045 <input>
2046 <port id="0" precision="FP32">
2047 <dim>-1</dim>
2048 <dim>-1</dim>
2049 <dim>12</dim>
2050 <dim>32</dim>
2051 </port>
2052 <port id="1" precision="I64">
2053 <dim>3</dim>
2054 </port>
2055 </input>
2056 <output>
2057 <port id="2" precision="FP32" names="192">
2058 <dim>-1</dim>
2059 <dim>-1</dim>
2060 <dim>384</dim>
2061 </port>
2062 </output>
2063 </layer>
2064 <layer id="138" name="self.bert.encoder.layer.1.attention.output.dense.weight" type="Const" version="opset1">
2065 <data element_type="f32" shape="384, 384" offset="56550560" size="589824" />
2066 <output>
2067 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.output.dense.weight">
2068 <dim>384</dim>
2069 <dim>384</dim>
2070 </port>
2071 </output>
2072 </layer>
2073 <layer id="139" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2074 <data transpose_a="false" transpose_b="true" />
2075 <input>
2076 <port id="0" precision="FP32">
2077 <dim>-1</dim>
2078 <dim>-1</dim>
2079 <dim>384</dim>
2080 </port>
2081 <port id="1" precision="FP32">
2082 <dim>384</dim>
2083 <dim>384</dim>
2084 </port>
2085 </input>
2086 <output>
2087 <port id="2" precision="FP32">
2088 <dim>-1</dim>
2089 <dim>-1</dim>
2090 <dim>384</dim>
2091 </port>
2092 </output>
2093 </layer>
2094 <layer id="140" name="Constant_90318" type="Const" version="opset1">
2095 <data element_type="f32" shape="1, 1, 384" offset="57140384" size="1536" />
2096 <output>
2097 <port id="0" precision="FP32">
2098 <dim>1</dim>
2099 <dim>1</dim>
2100 <dim>384</dim>
2101 </port>
2102 </output>
2103 </layer>
2104 <layer id="141" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
2105 <data auto_broadcast="numpy" />
2106 <input>
2107 <port id="0" precision="FP32">
2108 <dim>-1</dim>
2109 <dim>-1</dim>
2110 <dim>384</dim>
2111 </port>
2112 <port id="1" precision="FP32">
2113 <dim>1</dim>
2114 <dim>1</dim>
2115 <dim>384</dim>
2116 </port>
2117 </input>
2118 <output>
2119 <port id="2" precision="FP32" names="197,input.7">
2120 <dim>-1</dim>
2121 <dim>-1</dim>
2122 <dim>384</dim>
2123 </port>
2124 </output>
2125 </layer>
2126 <layer id="142" name="__module.bert.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
2127 <data auto_broadcast="numpy" />
2128 <input>
2129 <port id="0" precision="FP32">
2130 <dim>-1</dim>
2131 <dim>-1</dim>
2132 <dim>384</dim>
2133 </port>
2134 <port id="1" precision="FP32">
2135 <dim>-1</dim>
2136 <dim>-1</dim>
2137 <dim>384</dim>
2138 </port>
2139 </input>
2140 <output>
2141 <port id="2" precision="FP32" names="199">
2142 <dim>-1</dim>
2143 <dim>-1</dim>
2144 <dim>384</dim>
2145 </port>
2146 </output>
2147 </layer>
2148 <layer id="143" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2149 <data element_type="i32" shape="1" offset="47675412" size="4" />
2150 <output>
2151 <port id="0" precision="I32">
2152 <dim>1</dim>
2153 </port>
2154 </output>
2155 </layer>
2156 <layer id="144" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2157 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2158 <input>
2159 <port id="0" precision="FP32">
2160 <dim>-1</dim>
2161 <dim>-1</dim>
2162 <dim>384</dim>
2163 </port>
2164 <port id="1" precision="I32">
2165 <dim>1</dim>
2166 </port>
2167 </input>
2168 <output>
2169 <port id="2" precision="FP32">
2170 <dim>-1</dim>
2171 <dim>-1</dim>
2172 <dim>384</dim>
2173 </port>
2174 </output>
2175 </layer>
2176 <layer id="145" name="Constant_90319" type="Const" version="opset1">
2177 <data element_type="f32" shape="1, 1, 384" offset="57141920" size="1536" />
2178 <output>
2179 <port id="0" precision="FP32">
2180 <dim>1</dim>
2181 <dim>1</dim>
2182 <dim>384</dim>
2183 </port>
2184 </output>
2185 </layer>
2186 <layer id="146" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2187 <data auto_broadcast="numpy" />
2188 <input>
2189 <port id="0" precision="FP32">
2190 <dim>-1</dim>
2191 <dim>-1</dim>
2192 <dim>384</dim>
2193 </port>
2194 <port id="1" precision="FP32">
2195 <dim>1</dim>
2196 <dim>1</dim>
2197 <dim>384</dim>
2198 </port>
2199 </input>
2200 <output>
2201 <port id="2" precision="FP32">
2202 <dim>-1</dim>
2203 <dim>-1</dim>
2204 <dim>384</dim>
2205 </port>
2206 </output>
2207 </layer>
2208 <layer id="147" name="Constant_90320" type="Const" version="opset1">
2209 <data element_type="f32" shape="1, 1, 384" offset="57143456" size="1536" />
2210 <output>
2211 <port id="0" precision="FP32">
2212 <dim>1</dim>
2213 <dim>1</dim>
2214 <dim>384</dim>
2215 </port>
2216 </output>
2217 </layer>
2218 <layer id="148" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2219 <data auto_broadcast="numpy" />
2220 <input>
2221 <port id="0" precision="FP32">
2222 <dim>-1</dim>
2223 <dim>-1</dim>
2224 <dim>384</dim>
2225 </port>
2226 <port id="1" precision="FP32">
2227 <dim>1</dim>
2228 <dim>1</dim>
2229 <dim>384</dim>
2230 </port>
2231 </input>
2232 <output>
2233 <port id="2" precision="FP32" names="203,input_tensor.3">
2234 <dim>-1</dim>
2235 <dim>-1</dim>
2236 <dim>384</dim>
2237 </port>
2238 </output>
2239 </layer>
2240 <layer id="149" name="self.bert.encoder.layer.1.intermediate.dense.weight" type="Const" version="opset1">
2241 <data element_type="f32" shape="1536, 384" offset="57144992" size="2359296" />
2242 <output>
2243 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.intermediate.dense.weight">
2244 <dim>1536</dim>
2245 <dim>384</dim>
2246 </port>
2247 </output>
2248 </layer>
2249 <layer id="150" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2250 <data transpose_a="false" transpose_b="true" />
2251 <input>
2252 <port id="0" precision="FP32">
2253 <dim>-1</dim>
2254 <dim>-1</dim>
2255 <dim>384</dim>
2256 </port>
2257 <port id="1" precision="FP32">
2258 <dim>1536</dim>
2259 <dim>384</dim>
2260 </port>
2261 </input>
2262 <output>
2263 <port id="2" precision="FP32">
2264 <dim>-1</dim>
2265 <dim>-1</dim>
2266 <dim>1536</dim>
2267 </port>
2268 </output>
2269 </layer>
2270 <layer id="151" name="Constant_90321" type="Const" version="opset1">
2271 <data element_type="f32" shape="1, 1, 1536" offset="59504288" size="6144" />
2272 <output>
2273 <port id="0" precision="FP32">
2274 <dim>1</dim>
2275 <dim>1</dim>
2276 <dim>1536</dim>
2277 </port>
2278 </output>
2279 </layer>
2280 <layer id="152" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2281 <data auto_broadcast="numpy" />
2282 <input>
2283 <port id="0" precision="FP32">
2284 <dim>-1</dim>
2285 <dim>-1</dim>
2286 <dim>1536</dim>
2287 </port>
2288 <port id="1" precision="FP32">
2289 <dim>1</dim>
2290 <dim>1</dim>
2291 <dim>1536</dim>
2292 </port>
2293 </input>
2294 <output>
2295 <port id="2" precision="FP32" names="207">
2296 <dim>-1</dim>
2297 <dim>-1</dim>
2298 <dim>1536</dim>
2299 </port>
2300 </output>
2301 </layer>
2302 <layer id="153" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2303 <data approximation_mode="ERF" />
2304 <input>
2305 <port id="0" precision="FP32">
2306 <dim>-1</dim>
2307 <dim>-1</dim>
2308 <dim>1536</dim>
2309 </port>
2310 </input>
2311 <output>
2312 <port id="1" precision="FP32" names="208">
2313 <dim>-1</dim>
2314 <dim>-1</dim>
2315 <dim>1536</dim>
2316 </port>
2317 </output>
2318 </layer>
2319 <layer id="154" name="self.bert.encoder.layer.1.output.dense.weight" type="Const" version="opset1">
2320 <data element_type="f32" shape="384, 1536" offset="59510432" size="2359296" />
2321 <output>
2322 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.output.dense.weight">
2323 <dim>384</dim>
2324 <dim>1536</dim>
2325 </port>
2326 </output>
2327 </layer>
2328 <layer id="155" name="__module.bert.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2329 <data transpose_a="false" transpose_b="true" />
2330 <input>
2331 <port id="0" precision="FP32">
2332 <dim>-1</dim>
2333 <dim>-1</dim>
2334 <dim>1536</dim>
2335 </port>
2336 <port id="1" precision="FP32">
2337 <dim>384</dim>
2338 <dim>1536</dim>
2339 </port>
2340 </input>
2341 <output>
2342 <port id="2" precision="FP32">
2343 <dim>-1</dim>
2344 <dim>-1</dim>
2345 <dim>384</dim>
2346 </port>
2347 </output>
2348 </layer>
2349 <layer id="156" name="Constant_90322" type="Const" version="opset1">
2350 <data element_type="f32" shape="1, 1, 384" offset="61869728" size="1536" />
2351 <output>
2352 <port id="0" precision="FP32">
2353 <dim>1</dim>
2354 <dim>1</dim>
2355 <dim>384</dim>
2356 </port>
2357 </output>
2358 </layer>
2359 <layer id="157" name="__module.bert.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
2360 <data auto_broadcast="numpy" />
2361 <input>
2362 <port id="0" precision="FP32">
2363 <dim>-1</dim>
2364 <dim>-1</dim>
2365 <dim>384</dim>
2366 </port>
2367 <port id="1" precision="FP32">
2368 <dim>1</dim>
2369 <dim>1</dim>
2370 <dim>384</dim>
2371 </port>
2372 </input>
2373 <output>
2374 <port id="2" precision="FP32" names="213,input.9">
2375 <dim>-1</dim>
2376 <dim>-1</dim>
2377 <dim>384</dim>
2378 </port>
2379 </output>
2380 </layer>
2381 <layer id="158" name="__module.bert.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
2382 <data auto_broadcast="numpy" />
2383 <input>
2384 <port id="0" precision="FP32">
2385 <dim>-1</dim>
2386 <dim>-1</dim>
2387 <dim>384</dim>
2388 </port>
2389 <port id="1" precision="FP32">
2390 <dim>-1</dim>
2391 <dim>-1</dim>
2392 <dim>384</dim>
2393 </port>
2394 </input>
2395 <output>
2396 <port id="2" precision="FP32" names="215">
2397 <dim>-1</dim>
2398 <dim>-1</dim>
2399 <dim>384</dim>
2400 </port>
2401 </output>
2402 </layer>
2403 <layer id="159" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2404 <data element_type="i32" shape="1" offset="47675412" size="4" />
2405 <output>
2406 <port id="0" precision="I32">
2407 <dim>1</dim>
2408 </port>
2409 </output>
2410 </layer>
2411 <layer id="160" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2412 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2413 <input>
2414 <port id="0" precision="FP32">
2415 <dim>-1</dim>
2416 <dim>-1</dim>
2417 <dim>384</dim>
2418 </port>
2419 <port id="1" precision="I32">
2420 <dim>1</dim>
2421 </port>
2422 </input>
2423 <output>
2424 <port id="2" precision="FP32">
2425 <dim>-1</dim>
2426 <dim>-1</dim>
2427 <dim>384</dim>
2428 </port>
2429 </output>
2430 </layer>
2431 <layer id="161" name="Constant_90323" type="Const" version="opset1">
2432 <data element_type="f32" shape="1, 1, 384" offset="61871264" size="1536" />
2433 <output>
2434 <port id="0" precision="FP32">
2435 <dim>1</dim>
2436 <dim>1</dim>
2437 <dim>384</dim>
2438 </port>
2439 </output>
2440 </layer>
2441 <layer id="162" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2442 <data auto_broadcast="numpy" />
2443 <input>
2444 <port id="0" precision="FP32">
2445 <dim>-1</dim>
2446 <dim>-1</dim>
2447 <dim>384</dim>
2448 </port>
2449 <port id="1" precision="FP32">
2450 <dim>1</dim>
2451 <dim>1</dim>
2452 <dim>384</dim>
2453 </port>
2454 </input>
2455 <output>
2456 <port id="2" precision="FP32">
2457 <dim>-1</dim>
2458 <dim>-1</dim>
2459 <dim>384</dim>
2460 </port>
2461 </output>
2462 </layer>
2463 <layer id="163" name="Constant_90324" type="Const" version="opset1">
2464 <data element_type="f32" shape="1, 1, 384" offset="61872800" size="1536" />
2465 <output>
2466 <port id="0" precision="FP32">
2467 <dim>1</dim>
2468 <dim>1</dim>
2469 <dim>384</dim>
2470 </port>
2471 </output>
2472 </layer>
2473 <layer id="164" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2474 <data auto_broadcast="numpy" />
2475 <input>
2476 <port id="0" precision="FP32">
2477 <dim>-1</dim>
2478 <dim>-1</dim>
2479 <dim>384</dim>
2480 </port>
2481 <port id="1" precision="FP32">
2482 <dim>1</dim>
2483 <dim>1</dim>
2484 <dim>384</dim>
2485 </port>
2486 </input>
2487 <output>
2488 <port id="2" precision="FP32" names="219,hidden_states.13">
2489 <dim>-1</dim>
2490 <dim>-1</dim>
2491 <dim>384</dim>
2492 </port>
2493 </output>
2494 </layer>
2495 <layer id="165" name="self.bert.encoder.layer.2.attention.self.query.weight" type="Const" version="opset1">
2496 <data element_type="f32" shape="384, 384" offset="61874336" size="589824" />
2497 <output>
2498 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.self.query.weight">
2499 <dim>384</dim>
2500 <dim>384</dim>
2501 </port>
2502 </output>
2503 </layer>
2504 <layer id="166" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2505 <data transpose_a="false" transpose_b="true" />
2506 <input>
2507 <port id="0" precision="FP32">
2508 <dim>-1</dim>
2509 <dim>-1</dim>
2510 <dim>384</dim>
2511 </port>
2512 <port id="1" precision="FP32">
2513 <dim>384</dim>
2514 <dim>384</dim>
2515 </port>
2516 </input>
2517 <output>
2518 <port id="2" precision="FP32">
2519 <dim>-1</dim>
2520 <dim>-1</dim>
2521 <dim>384</dim>
2522 </port>
2523 </output>
2524 </layer>
2525 <layer id="167" name="Constant_90325" type="Const" version="opset1">
2526 <data element_type="f32" shape="1, 1, 384" offset="62464160" size="1536" />
2527 <output>
2528 <port id="0" precision="FP32">
2529 <dim>1</dim>
2530 <dim>1</dim>
2531 <dim>384</dim>
2532 </port>
2533 </output>
2534 </layer>
2535 <layer id="168" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2536 <data auto_broadcast="numpy" />
2537 <input>
2538 <port id="0" precision="FP32">
2539 <dim>-1</dim>
2540 <dim>-1</dim>
2541 <dim>384</dim>
2542 </port>
2543 <port id="1" precision="FP32">
2544 <dim>1</dim>
2545 <dim>1</dim>
2546 <dim>384</dim>
2547 </port>
2548 </input>
2549 <output>
2550 <port id="2" precision="FP32" names="232,x.25">
2551 <dim>-1</dim>
2552 <dim>-1</dim>
2553 <dim>384</dim>
2554 </port>
2555 </output>
2556 </layer>
2557 <layer id="169" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2558 <data element_type="i64" shape="4" offset="48269848" size="32" />
2559 <output>
2560 <port id="0" precision="I64">
2561 <dim>4</dim>
2562 </port>
2563 </output>
2564 </layer>
2565 <layer id="170" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2566 <data special_zero="true" />
2567 <input>
2568 <port id="0" precision="FP32">
2569 <dim>-1</dim>
2570 <dim>-1</dim>
2571 <dim>384</dim>
2572 </port>
2573 <port id="1" precision="I64">
2574 <dim>4</dim>
2575 </port>
2576 </input>
2577 <output>
2578 <port id="2" precision="FP32" names="236,x.27">
2579 <dim>-1</dim>
2580 <dim>-1</dim>
2581 <dim>12</dim>
2582 <dim>32</dim>
2583 </port>
2584 </output>
2585 </layer>
2586 <layer id="171" name="Constant_84904" type="Const" version="opset1">
2587 <data element_type="i64" shape="4" offset="48269880" size="32" />
2588 <output>
2589 <port id="0" precision="I64" names="237">
2590 <dim>4</dim>
2591 </port>
2592 </output>
2593 </layer>
2594 <layer id="172" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2595 <input>
2596 <port id="0" precision="FP32">
2597 <dim>-1</dim>
2598 <dim>-1</dim>
2599 <dim>12</dim>
2600 <dim>32</dim>
2601 </port>
2602 <port id="1" precision="I64">
2603 <dim>4</dim>
2604 </port>
2605 </input>
2606 <output>
2607 <port id="2" precision="FP32" names="238">
2608 <dim>-1</dim>
2609 <dim>12</dim>
2610 <dim>-1</dim>
2611 <dim>32</dim>
2612 </port>
2613 </output>
2614 </layer>
2615 <layer id="173" name="self.bert.encoder.layer.2.attention.self.key.weight" type="Const" version="opset1">
2616 <data element_type="f32" shape="384, 384" offset="62465696" size="589824" />
2617 <output>
2618 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.self.key.weight">
2619 <dim>384</dim>
2620 <dim>384</dim>
2621 </port>
2622 </output>
2623 </layer>
2624 <layer id="174" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2625 <data transpose_a="false" transpose_b="true" />
2626 <input>
2627 <port id="0" precision="FP32">
2628 <dim>-1</dim>
2629 <dim>-1</dim>
2630 <dim>384</dim>
2631 </port>
2632 <port id="1" precision="FP32">
2633 <dim>384</dim>
2634 <dim>384</dim>
2635 </port>
2636 </input>
2637 <output>
2638 <port id="2" precision="FP32">
2639 <dim>-1</dim>
2640 <dim>-1</dim>
2641 <dim>384</dim>
2642 </port>
2643 </output>
2644 </layer>
2645 <layer id="175" name="Constant_90326" type="Const" version="opset1">
2646 <data element_type="f32" shape="1, 1, 384" offset="63055520" size="1536" />
2647 <output>
2648 <port id="0" precision="FP32">
2649 <dim>1</dim>
2650 <dim>1</dim>
2651 <dim>384</dim>
2652 </port>
2653 </output>
2654 </layer>
2655 <layer id="176" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2656 <data auto_broadcast="numpy" />
2657 <input>
2658 <port id="0" precision="FP32">
2659 <dim>-1</dim>
2660 <dim>-1</dim>
2661 <dim>384</dim>
2662 </port>
2663 <port id="1" precision="FP32">
2664 <dim>1</dim>
2665 <dim>1</dim>
2666 <dim>384</dim>
2667 </port>
2668 </input>
2669 <output>
2670 <port id="2" precision="FP32" names="241,x.29">
2671 <dim>-1</dim>
2672 <dim>-1</dim>
2673 <dim>384</dim>
2674 </port>
2675 </output>
2676 </layer>
2677 <layer id="177" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2678 <data element_type="i64" shape="4" offset="48269848" size="32" />
2679 <output>
2680 <port id="0" precision="I64">
2681 <dim>4</dim>
2682 </port>
2683 </output>
2684 </layer>
2685 <layer id="178" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2686 <data special_zero="true" />
2687 <input>
2688 <port id="0" precision="FP32">
2689 <dim>-1</dim>
2690 <dim>-1</dim>
2691 <dim>384</dim>
2692 </port>
2693 <port id="1" precision="I64">
2694 <dim>4</dim>
2695 </port>
2696 </input>
2697 <output>
2698 <port id="2" precision="FP32" names="245,x.31">
2699 <dim>-1</dim>
2700 <dim>-1</dim>
2701 <dim>12</dim>
2702 <dim>32</dim>
2703 </port>
2704 </output>
2705 </layer>
2706 <layer id="179" name="Constant_84927" type="Const" version="opset1">
2707 <data element_type="i64" shape="4" offset="48269880" size="32" />
2708 <output>
2709 <port id="0" precision="I64" names="246">
2710 <dim>4</dim>
2711 </port>
2712 </output>
2713 </layer>
2714 <layer id="180" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2715 <input>
2716 <port id="0" precision="FP32">
2717 <dim>-1</dim>
2718 <dim>-1</dim>
2719 <dim>12</dim>
2720 <dim>32</dim>
2721 </port>
2722 <port id="1" precision="I64">
2723 <dim>4</dim>
2724 </port>
2725 </input>
2726 <output>
2727 <port id="2" precision="FP32" names="247">
2728 <dim>-1</dim>
2729 <dim>12</dim>
2730 <dim>-1</dim>
2731 <dim>32</dim>
2732 </port>
2733 </output>
2734 </layer>
2735 <layer id="181" name="self.bert.encoder.layer.2.attention.self.value.weight" type="Const" version="opset1">
2736 <data element_type="f32" shape="384, 384" offset="63057056" size="589824" />
2737 <output>
2738 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.self.value.weight">
2739 <dim>384</dim>
2740 <dim>384</dim>
2741 </port>
2742 </output>
2743 </layer>
2744 <layer id="182" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
2745 <data transpose_a="false" transpose_b="true" />
2746 <input>
2747 <port id="0" precision="FP32">
2748 <dim>-1</dim>
2749 <dim>-1</dim>
2750 <dim>384</dim>
2751 </port>
2752 <port id="1" precision="FP32">
2753 <dim>384</dim>
2754 <dim>384</dim>
2755 </port>
2756 </input>
2757 <output>
2758 <port id="2" precision="FP32">
2759 <dim>-1</dim>
2760 <dim>-1</dim>
2761 <dim>384</dim>
2762 </port>
2763 </output>
2764 </layer>
2765 <layer id="183" name="Constant_90327" type="Const" version="opset1">
2766 <data element_type="f32" shape="1, 1, 384" offset="63646880" size="1536" />
2767 <output>
2768 <port id="0" precision="FP32">
2769 <dim>1</dim>
2770 <dim>1</dim>
2771 <dim>384</dim>
2772 </port>
2773 </output>
2774 </layer>
2775 <layer id="184" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
2776 <data auto_broadcast="numpy" />
2777 <input>
2778 <port id="0" precision="FP32">
2779 <dim>-1</dim>
2780 <dim>-1</dim>
2781 <dim>384</dim>
2782 </port>
2783 <port id="1" precision="FP32">
2784 <dim>1</dim>
2785 <dim>1</dim>
2786 <dim>384</dim>
2787 </port>
2788 </input>
2789 <output>
2790 <port id="2" precision="FP32" names="250,x.33">
2791 <dim>-1</dim>
2792 <dim>-1</dim>
2793 <dim>384</dim>
2794 </port>
2795 </output>
2796 </layer>
2797 <layer id="185" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
2798 <data element_type="i64" shape="4" offset="48269848" size="32" />
2799 <output>
2800 <port id="0" precision="I64">
2801 <dim>4</dim>
2802 </port>
2803 </output>
2804 </layer>
2805 <layer id="186" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
2806 <data special_zero="true" />
2807 <input>
2808 <port id="0" precision="FP32">
2809 <dim>-1</dim>
2810 <dim>-1</dim>
2811 <dim>384</dim>
2812 </port>
2813 <port id="1" precision="I64">
2814 <dim>4</dim>
2815 </port>
2816 </input>
2817 <output>
2818 <port id="2" precision="FP32" names="254,x.35">
2819 <dim>-1</dim>
2820 <dim>-1</dim>
2821 <dim>12</dim>
2822 <dim>32</dim>
2823 </port>
2824 </output>
2825 </layer>
2826 <layer id="187" name="Constant_84950" type="Const" version="opset1">
2827 <data element_type="i64" shape="4" offset="48269880" size="32" />
2828 <output>
2829 <port id="0" precision="I64" names="255">
2830 <dim>4</dim>
2831 </port>
2832 </output>
2833 </layer>
2834 <layer id="188" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
2835 <input>
2836 <port id="0" precision="FP32">
2837 <dim>-1</dim>
2838 <dim>-1</dim>
2839 <dim>12</dim>
2840 <dim>32</dim>
2841 </port>
2842 <port id="1" precision="I64">
2843 <dim>4</dim>
2844 </port>
2845 </input>
2846 <output>
2847 <port id="2" precision="FP32" names="256">
2848 <dim>-1</dim>
2849 <dim>12</dim>
2850 <dim>-1</dim>
2851 <dim>32</dim>
2852 </port>
2853 </output>
2854 </layer>
2855 <layer id="189" name="__module.bert.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
2856 <data causal="false" />
2857 <input>
2858 <port id="0" precision="FP32">
2859 <dim>-1</dim>
2860 <dim>12</dim>
2861 <dim>-1</dim>
2862 <dim>32</dim>
2863 </port>
2864 <port id="1" precision="FP32">
2865 <dim>-1</dim>
2866 <dim>12</dim>
2867 <dim>-1</dim>
2868 <dim>32</dim>
2869 </port>
2870 <port id="2" precision="FP32">
2871 <dim>-1</dim>
2872 <dim>12</dim>
2873 <dim>-1</dim>
2874 <dim>32</dim>
2875 </port>
2876 <port id="3" precision="FP32">
2877 <dim>-1</dim>
2878 <dim>1</dim>
2879 <dim>-1</dim>
2880 <dim>-1</dim>
2881 </port>
2882 </input>
2883 <output>
2884 <port id="4" precision="FP32" names="257,attn_output.9">
2885 <dim>-1</dim>
2886 <dim>12</dim>
2887 <dim>-1</dim>
2888 <dim>32</dim>
2889 </port>
2890 </output>
2891 </layer>
2892 <layer id="190" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
2893 <data element_type="i32" shape="4" offset="49452664" size="16" />
2894 <output>
2895 <port id="0" precision="I32">
2896 <dim>4</dim>
2897 </port>
2898 </output>
2899 </layer>
2900 <layer id="191" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
2901 <input>
2902 <port id="0" precision="FP32">
2903 <dim>-1</dim>
2904 <dim>12</dim>
2905 <dim>-1</dim>
2906 <dim>32</dim>
2907 </port>
2908 <port id="1" precision="I32">
2909 <dim>4</dim>
2910 </port>
2911 </input>
2912 <output>
2913 <port id="2" precision="FP32" names="258,attn_output.11">
2914 <dim>-1</dim>
2915 <dim>-1</dim>
2916 <dim>12</dim>
2917 <dim>32</dim>
2918 </port>
2919 </output>
2920 </layer>
2921 <layer id="192" name="Constant_90445" type="Const" version="opset1">
2922 <data element_type="i64" shape="3" offset="49452680" size="24" />
2923 <output>
2924 <port id="0" precision="I64">
2925 <dim>3</dim>
2926 </port>
2927 </output>
2928 </layer>
2929 <layer id="193" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
2930 <data special_zero="true" />
2931 <input>
2932 <port id="0" precision="FP32">
2933 <dim>-1</dim>
2934 <dim>-1</dim>
2935 <dim>12</dim>
2936 <dim>32</dim>
2937 </port>
2938 <port id="1" precision="I64">
2939 <dim>3</dim>
2940 </port>
2941 </input>
2942 <output>
2943 <port id="2" precision="FP32" names="260">
2944 <dim>-1</dim>
2945 <dim>-1</dim>
2946 <dim>384</dim>
2947 </port>
2948 </output>
2949 </layer>
2950 <layer id="194" name="self.bert.encoder.layer.2.attention.output.dense.weight" type="Const" version="opset1">
2951 <data element_type="f32" shape="384, 384" offset="63648416" size="589824" />
2952 <output>
2953 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.output.dense.weight">
2954 <dim>384</dim>
2955 <dim>384</dim>
2956 </port>
2957 </output>
2958 </layer>
2959 <layer id="195" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2960 <data transpose_a="false" transpose_b="true" />
2961 <input>
2962 <port id="0" precision="FP32">
2963 <dim>-1</dim>
2964 <dim>-1</dim>
2965 <dim>384</dim>
2966 </port>
2967 <port id="1" precision="FP32">
2968 <dim>384</dim>
2969 <dim>384</dim>
2970 </port>
2971 </input>
2972 <output>
2973 <port id="2" precision="FP32">
2974 <dim>-1</dim>
2975 <dim>-1</dim>
2976 <dim>384</dim>
2977 </port>
2978 </output>
2979 </layer>
2980 <layer id="196" name="Constant_90328" type="Const" version="opset1">
2981 <data element_type="f32" shape="1, 1, 384" offset="64238240" size="1536" />
2982 <output>
2983 <port id="0" precision="FP32">
2984 <dim>1</dim>
2985 <dim>1</dim>
2986 <dim>384</dim>
2987 </port>
2988 </output>
2989 </layer>
2990 <layer id="197" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
2991 <data auto_broadcast="numpy" />
2992 <input>
2993 <port id="0" precision="FP32">
2994 <dim>-1</dim>
2995 <dim>-1</dim>
2996 <dim>384</dim>
2997 </port>
2998 <port id="1" precision="FP32">
2999 <dim>1</dim>
3000 <dim>1</dim>
3001 <dim>384</dim>
3002 </port>
3003 </input>
3004 <output>
3005 <port id="2" precision="FP32" names="265,input.11">
3006 <dim>-1</dim>
3007 <dim>-1</dim>
3008 <dim>384</dim>
3009 </port>
3010 </output>
3011 </layer>
3012 <layer id="198" name="__module.bert.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
3013 <data auto_broadcast="numpy" />
3014 <input>
3015 <port id="0" precision="FP32">
3016 <dim>-1</dim>
3017 <dim>-1</dim>
3018 <dim>384</dim>
3019 </port>
3020 <port id="1" precision="FP32">
3021 <dim>-1</dim>
3022 <dim>-1</dim>
3023 <dim>384</dim>
3024 </port>
3025 </input>
3026 <output>
3027 <port id="2" precision="FP32" names="267">
3028 <dim>-1</dim>
3029 <dim>-1</dim>
3030 <dim>384</dim>
3031 </port>
3032 </output>
3033 </layer>
3034 <layer id="199" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3035 <data element_type="i32" shape="1" offset="47675412" size="4" />
3036 <output>
3037 <port id="0" precision="I32">
3038 <dim>1</dim>
3039 </port>
3040 </output>
3041 </layer>
3042 <layer id="200" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3043 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3044 <input>
3045 <port id="0" precision="FP32">
3046 <dim>-1</dim>
3047 <dim>-1</dim>
3048 <dim>384</dim>
3049 </port>
3050 <port id="1" precision="I32">
3051 <dim>1</dim>
3052 </port>
3053 </input>
3054 <output>
3055 <port id="2" precision="FP32">
3056 <dim>-1</dim>
3057 <dim>-1</dim>
3058 <dim>384</dim>
3059 </port>
3060 </output>
3061 </layer>
3062 <layer id="201" name="Constant_90329" type="Const" version="opset1">
3063 <data element_type="f32" shape="1, 1, 384" offset="64239776" size="1536" />
3064 <output>
3065 <port id="0" precision="FP32">
3066 <dim>1</dim>
3067 <dim>1</dim>
3068 <dim>384</dim>
3069 </port>
3070 </output>
3071 </layer>
3072 <layer id="202" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3073 <data auto_broadcast="numpy" />
3074 <input>
3075 <port id="0" precision="FP32">
3076 <dim>-1</dim>
3077 <dim>-1</dim>
3078 <dim>384</dim>
3079 </port>
3080 <port id="1" precision="FP32">
3081 <dim>1</dim>
3082 <dim>1</dim>
3083 <dim>384</dim>
3084 </port>
3085 </input>
3086 <output>
3087 <port id="2" precision="FP32">
3088 <dim>-1</dim>
3089 <dim>-1</dim>
3090 <dim>384</dim>
3091 </port>
3092 </output>
3093 </layer>
3094 <layer id="203" name="Constant_90330" type="Const" version="opset1">
3095 <data element_type="f32" shape="1, 1, 384" offset="64241312" size="1536" />
3096 <output>
3097 <port id="0" precision="FP32">
3098 <dim>1</dim>
3099 <dim>1</dim>
3100 <dim>384</dim>
3101 </port>
3102 </output>
3103 </layer>
3104 <layer id="204" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3105 <data auto_broadcast="numpy" />
3106 <input>
3107 <port id="0" precision="FP32">
3108 <dim>-1</dim>
3109 <dim>-1</dim>
3110 <dim>384</dim>
3111 </port>
3112 <port id="1" precision="FP32">
3113 <dim>1</dim>
3114 <dim>1</dim>
3115 <dim>384</dim>
3116 </port>
3117 </input>
3118 <output>
3119 <port id="2" precision="FP32" names="271,input_tensor.5">
3120 <dim>-1</dim>
3121 <dim>-1</dim>
3122 <dim>384</dim>
3123 </port>
3124 </output>
3125 </layer>
3126 <layer id="205" name="self.bert.encoder.layer.2.intermediate.dense.weight" type="Const" version="opset1">
3127 <data element_type="f32" shape="1536, 384" offset="64242848" size="2359296" />
3128 <output>
3129 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.intermediate.dense.weight">
3130 <dim>1536</dim>
3131 <dim>384</dim>
3132 </port>
3133 </output>
3134 </layer>
3135 <layer id="206" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3136 <data transpose_a="false" transpose_b="true" />
3137 <input>
3138 <port id="0" precision="FP32">
3139 <dim>-1</dim>
3140 <dim>-1</dim>
3141 <dim>384</dim>
3142 </port>
3143 <port id="1" precision="FP32">
3144 <dim>1536</dim>
3145 <dim>384</dim>
3146 </port>
3147 </input>
3148 <output>
3149 <port id="2" precision="FP32">
3150 <dim>-1</dim>
3151 <dim>-1</dim>
3152 <dim>1536</dim>
3153 </port>
3154 </output>
3155 </layer>
3156 <layer id="207" name="Constant_90331" type="Const" version="opset1">
3157 <data element_type="f32" shape="1, 1, 1536" offset="66602144" size="6144" />
3158 <output>
3159 <port id="0" precision="FP32">
3160 <dim>1</dim>
3161 <dim>1</dim>
3162 <dim>1536</dim>
3163 </port>
3164 </output>
3165 </layer>
3166 <layer id="208" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3167 <data auto_broadcast="numpy" />
3168 <input>
3169 <port id="0" precision="FP32">
3170 <dim>-1</dim>
3171 <dim>-1</dim>
3172 <dim>1536</dim>
3173 </port>
3174 <port id="1" precision="FP32">
3175 <dim>1</dim>
3176 <dim>1</dim>
3177 <dim>1536</dim>
3178 </port>
3179 </input>
3180 <output>
3181 <port id="2" precision="FP32" names="275">
3182 <dim>-1</dim>
3183 <dim>-1</dim>
3184 <dim>1536</dim>
3185 </port>
3186 </output>
3187 </layer>
3188 <layer id="209" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3189 <data approximation_mode="ERF" />
3190 <input>
3191 <port id="0" precision="FP32">
3192 <dim>-1</dim>
3193 <dim>-1</dim>
3194 <dim>1536</dim>
3195 </port>
3196 </input>
3197 <output>
3198 <port id="1" precision="FP32" names="276">
3199 <dim>-1</dim>
3200 <dim>-1</dim>
3201 <dim>1536</dim>
3202 </port>
3203 </output>
3204 </layer>
3205 <layer id="210" name="self.bert.encoder.layer.2.output.dense.weight" type="Const" version="opset1">
3206 <data element_type="f32" shape="384, 1536" offset="66608288" size="2359296" />
3207 <output>
3208 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.output.dense.weight">
3209 <dim>384</dim>
3210 <dim>1536</dim>
3211 </port>
3212 </output>
3213 </layer>
3214 <layer id="211" name="__module.bert.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3215 <data transpose_a="false" transpose_b="true" />
3216 <input>
3217 <port id="0" precision="FP32">
3218 <dim>-1</dim>
3219 <dim>-1</dim>
3220 <dim>1536</dim>
3221 </port>
3222 <port id="1" precision="FP32">
3223 <dim>384</dim>
3224 <dim>1536</dim>
3225 </port>
3226 </input>
3227 <output>
3228 <port id="2" precision="FP32">
3229 <dim>-1</dim>
3230 <dim>-1</dim>
3231 <dim>384</dim>
3232 </port>
3233 </output>
3234 </layer>
3235 <layer id="212" name="Constant_90332" type="Const" version="opset1">
3236 <data element_type="f32" shape="1, 1, 384" offset="68967584" size="1536" />
3237 <output>
3238 <port id="0" precision="FP32">
3239 <dim>1</dim>
3240 <dim>1</dim>
3241 <dim>384</dim>
3242 </port>
3243 </output>
3244 </layer>
3245 <layer id="213" name="__module.bert.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
3246 <data auto_broadcast="numpy" />
3247 <input>
3248 <port id="0" precision="FP32">
3249 <dim>-1</dim>
3250 <dim>-1</dim>
3251 <dim>384</dim>
3252 </port>
3253 <port id="1" precision="FP32">
3254 <dim>1</dim>
3255 <dim>1</dim>
3256 <dim>384</dim>
3257 </port>
3258 </input>
3259 <output>
3260 <port id="2" precision="FP32" names="281,input.13">
3261 <dim>-1</dim>
3262 <dim>-1</dim>
3263 <dim>384</dim>
3264 </port>
3265 </output>
3266 </layer>
3267 <layer id="214" name="__module.bert.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
3268 <data auto_broadcast="numpy" />
3269 <input>
3270 <port id="0" precision="FP32">
3271 <dim>-1</dim>
3272 <dim>-1</dim>
3273 <dim>384</dim>
3274 </port>
3275 <port id="1" precision="FP32">
3276 <dim>-1</dim>
3277 <dim>-1</dim>
3278 <dim>384</dim>
3279 </port>
3280 </input>
3281 <output>
3282 <port id="2" precision="FP32" names="283">
3283 <dim>-1</dim>
3284 <dim>-1</dim>
3285 <dim>384</dim>
3286 </port>
3287 </output>
3288 </layer>
3289 <layer id="215" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3290 <data element_type="i32" shape="1" offset="47675412" size="4" />
3291 <output>
3292 <port id="0" precision="I32">
3293 <dim>1</dim>
3294 </port>
3295 </output>
3296 </layer>
3297 <layer id="216" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3298 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3299 <input>
3300 <port id="0" precision="FP32">
3301 <dim>-1</dim>
3302 <dim>-1</dim>
3303 <dim>384</dim>
3304 </port>
3305 <port id="1" precision="I32">
3306 <dim>1</dim>
3307 </port>
3308 </input>
3309 <output>
3310 <port id="2" precision="FP32">
3311 <dim>-1</dim>
3312 <dim>-1</dim>
3313 <dim>384</dim>
3314 </port>
3315 </output>
3316 </layer>
3317 <layer id="217" name="Constant_90333" type="Const" version="opset1">
3318 <data element_type="f32" shape="1, 1, 384" offset="68969120" size="1536" />
3319 <output>
3320 <port id="0" precision="FP32">
3321 <dim>1</dim>
3322 <dim>1</dim>
3323 <dim>384</dim>
3324 </port>
3325 </output>
3326 </layer>
3327 <layer id="218" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3328 <data auto_broadcast="numpy" />
3329 <input>
3330 <port id="0" precision="FP32">
3331 <dim>-1</dim>
3332 <dim>-1</dim>
3333 <dim>384</dim>
3334 </port>
3335 <port id="1" precision="FP32">
3336 <dim>1</dim>
3337 <dim>1</dim>
3338 <dim>384</dim>
3339 </port>
3340 </input>
3341 <output>
3342 <port id="2" precision="FP32">
3343 <dim>-1</dim>
3344 <dim>-1</dim>
3345 <dim>384</dim>
3346 </port>
3347 </output>
3348 </layer>
3349 <layer id="219" name="Constant_90334" type="Const" version="opset1">
3350 <data element_type="f32" shape="1, 1, 384" offset="68970656" size="1536" />
3351 <output>
3352 <port id="0" precision="FP32">
3353 <dim>1</dim>
3354 <dim>1</dim>
3355 <dim>384</dim>
3356 </port>
3357 </output>
3358 </layer>
3359 <layer id="220" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3360 <data auto_broadcast="numpy" />
3361 <input>
3362 <port id="0" precision="FP32">
3363 <dim>-1</dim>
3364 <dim>-1</dim>
3365 <dim>384</dim>
3366 </port>
3367 <port id="1" precision="FP32">
3368 <dim>1</dim>
3369 <dim>1</dim>
3370 <dim>384</dim>
3371 </port>
3372 </input>
3373 <output>
3374 <port id="2" precision="FP32" names="287,hidden_states.19">
3375 <dim>-1</dim>
3376 <dim>-1</dim>
3377 <dim>384</dim>
3378 </port>
3379 </output>
3380 </layer>
3381 <layer id="221" name="self.bert.encoder.layer.3.attention.self.query.weight" type="Const" version="opset1">
3382 <data element_type="f32" shape="384, 384" offset="68972192" size="589824" />
3383 <output>
3384 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.self.query.weight">
3385 <dim>384</dim>
3386 <dim>384</dim>
3387 </port>
3388 </output>
3389 </layer>
3390 <layer id="222" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
3391 <data transpose_a="false" transpose_b="true" />
3392 <input>
3393 <port id="0" precision="FP32">
3394 <dim>-1</dim>
3395 <dim>-1</dim>
3396 <dim>384</dim>
3397 </port>
3398 <port id="1" precision="FP32">
3399 <dim>384</dim>
3400 <dim>384</dim>
3401 </port>
3402 </input>
3403 <output>
3404 <port id="2" precision="FP32">
3405 <dim>-1</dim>
3406 <dim>-1</dim>
3407 <dim>384</dim>
3408 </port>
3409 </output>
3410 </layer>
3411 <layer id="223" name="Constant_90335" type="Const" version="opset1">
3412 <data element_type="f32" shape="1, 1, 384" offset="69562016" size="1536" />
3413 <output>
3414 <port id="0" precision="FP32">
3415 <dim>1</dim>
3416 <dim>1</dim>
3417 <dim>384</dim>
3418 </port>
3419 </output>
3420 </layer>
3421 <layer id="224" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
3422 <data auto_broadcast="numpy" />
3423 <input>
3424 <port id="0" precision="FP32">
3425 <dim>-1</dim>
3426 <dim>-1</dim>
3427 <dim>384</dim>
3428 </port>
3429 <port id="1" precision="FP32">
3430 <dim>1</dim>
3431 <dim>1</dim>
3432 <dim>384</dim>
3433 </port>
3434 </input>
3435 <output>
3436 <port id="2" precision="FP32" names="300,x.37">
3437 <dim>-1</dim>
3438 <dim>-1</dim>
3439 <dim>384</dim>
3440 </port>
3441 </output>
3442 </layer>
3443 <layer id="225" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
3444 <data element_type="i64" shape="4" offset="48269848" size="32" />
3445 <output>
3446 <port id="0" precision="I64">
3447 <dim>4</dim>
3448 </port>
3449 </output>
3450 </layer>
3451 <layer id="226" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
3452 <data special_zero="true" />
3453 <input>
3454 <port id="0" precision="FP32">
3455 <dim>-1</dim>
3456 <dim>-1</dim>
3457 <dim>384</dim>
3458 </port>
3459 <port id="1" precision="I64">
3460 <dim>4</dim>
3461 </port>
3462 </input>
3463 <output>
3464 <port id="2" precision="FP32" names="304,x.39">
3465 <dim>-1</dim>
3466 <dim>-1</dim>
3467 <dim>12</dim>
3468 <dim>32</dim>
3469 </port>
3470 </output>
3471 </layer>
3472 <layer id="227" name="Constant_85127" type="Const" version="opset1">
3473 <data element_type="i64" shape="4" offset="48269880" size="32" />
3474 <output>
3475 <port id="0" precision="I64" names="305">
3476 <dim>4</dim>
3477 </port>
3478 </output>
3479 </layer>
3480 <layer id="228" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
3481 <input>
3482 <port id="0" precision="FP32">
3483 <dim>-1</dim>
3484 <dim>-1</dim>
3485 <dim>12</dim>
3486 <dim>32</dim>
3487 </port>
3488 <port id="1" precision="I64">
3489 <dim>4</dim>
3490 </port>
3491 </input>
3492 <output>
3493 <port id="2" precision="FP32" names="306">
3494 <dim>-1</dim>
3495 <dim>12</dim>
3496 <dim>-1</dim>
3497 <dim>32</dim>
3498 </port>
3499 </output>
3500 </layer>
3501 <layer id="229" name="self.bert.encoder.layer.3.attention.self.key.weight" type="Const" version="opset1">
3502 <data element_type="f32" shape="384, 384" offset="69563552" size="589824" />
3503 <output>
3504 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.self.key.weight">
3505 <dim>384</dim>
3506 <dim>384</dim>
3507 </port>
3508 </output>
3509 </layer>
3510 <layer id="230" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
3511 <data transpose_a="false" transpose_b="true" />
3512 <input>
3513 <port id="0" precision="FP32">
3514 <dim>-1</dim>
3515 <dim>-1</dim>
3516 <dim>384</dim>
3517 </port>
3518 <port id="1" precision="FP32">
3519 <dim>384</dim>
3520 <dim>384</dim>
3521 </port>
3522 </input>
3523 <output>
3524 <port id="2" precision="FP32">
3525 <dim>-1</dim>
3526 <dim>-1</dim>
3527 <dim>384</dim>
3528 </port>
3529 </output>
3530 </layer>
3531 <layer id="231" name="Constant_90336" type="Const" version="opset1">
3532 <data element_type="f32" shape="1, 1, 384" offset="70153376" size="1536" />
3533 <output>
3534 <port id="0" precision="FP32">
3535 <dim>1</dim>
3536 <dim>1</dim>
3537 <dim>384</dim>
3538 </port>
3539 </output>
3540 </layer>
3541 <layer id="232" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
3542 <data auto_broadcast="numpy" />
3543 <input>
3544 <port id="0" precision="FP32">
3545 <dim>-1</dim>
3546 <dim>-1</dim>
3547 <dim>384</dim>
3548 </port>
3549 <port id="1" precision="FP32">
3550 <dim>1</dim>
3551 <dim>1</dim>
3552 <dim>384</dim>
3553 </port>
3554 </input>
3555 <output>
3556 <port id="2" precision="FP32" names="309,x.41">
3557 <dim>-1</dim>
3558 <dim>-1</dim>
3559 <dim>384</dim>
3560 </port>
3561 </output>
3562 </layer>
3563 <layer id="233" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
3564 <data element_type="i64" shape="4" offset="48269848" size="32" />
3565 <output>
3566 <port id="0" precision="I64">
3567 <dim>4</dim>
3568 </port>
3569 </output>
3570 </layer>
3571 <layer id="234" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
3572 <data special_zero="true" />
3573 <input>
3574 <port id="0" precision="FP32">
3575 <dim>-1</dim>
3576 <dim>-1</dim>
3577 <dim>384</dim>
3578 </port>
3579 <port id="1" precision="I64">
3580 <dim>4</dim>
3581 </port>
3582 </input>
3583 <output>
3584 <port id="2" precision="FP32" names="313,x.43">
3585 <dim>-1</dim>
3586 <dim>-1</dim>
3587 <dim>12</dim>
3588 <dim>32</dim>
3589 </port>
3590 </output>
3591 </layer>
3592 <layer id="235" name="Constant_85150" type="Const" version="opset1">
3593 <data element_type="i64" shape="4" offset="48269880" size="32" />
3594 <output>
3595 <port id="0" precision="I64" names="314">
3596 <dim>4</dim>
3597 </port>
3598 </output>
3599 </layer>
3600 <layer id="236" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
3601 <input>
3602 <port id="0" precision="FP32">
3603 <dim>-1</dim>
3604 <dim>-1</dim>
3605 <dim>12</dim>
3606 <dim>32</dim>
3607 </port>
3608 <port id="1" precision="I64">
3609 <dim>4</dim>
3610 </port>
3611 </input>
3612 <output>
3613 <port id="2" precision="FP32" names="315">
3614 <dim>-1</dim>
3615 <dim>12</dim>
3616 <dim>-1</dim>
3617 <dim>32</dim>
3618 </port>
3619 </output>
3620 </layer>
3621 <layer id="237" name="self.bert.encoder.layer.3.attention.self.value.weight" type="Const" version="opset1">
3622 <data element_type="f32" shape="384, 384" offset="70154912" size="589824" />
3623 <output>
3624 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.self.value.weight">
3625 <dim>384</dim>
3626 <dim>384</dim>
3627 </port>
3628 </output>
3629 </layer>
3630 <layer id="238" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
3631 <data transpose_a="false" transpose_b="true" />
3632 <input>
3633 <port id="0" precision="FP32">
3634 <dim>-1</dim>
3635 <dim>-1</dim>
3636 <dim>384</dim>
3637 </port>
3638 <port id="1" precision="FP32">
3639 <dim>384</dim>
3640 <dim>384</dim>
3641 </port>
3642 </input>
3643 <output>
3644 <port id="2" precision="FP32">
3645 <dim>-1</dim>
3646 <dim>-1</dim>
3647 <dim>384</dim>
3648 </port>
3649 </output>
3650 </layer>
3651 <layer id="239" name="Constant_90337" type="Const" version="opset1">
3652 <data element_type="f32" shape="1, 1, 384" offset="70744736" size="1536" />
3653 <output>
3654 <port id="0" precision="FP32">
3655 <dim>1</dim>
3656 <dim>1</dim>
3657 <dim>384</dim>
3658 </port>
3659 </output>
3660 </layer>
3661 <layer id="240" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
3662 <data auto_broadcast="numpy" />
3663 <input>
3664 <port id="0" precision="FP32">
3665 <dim>-1</dim>
3666 <dim>-1</dim>
3667 <dim>384</dim>
3668 </port>
3669 <port id="1" precision="FP32">
3670 <dim>1</dim>
3671 <dim>1</dim>
3672 <dim>384</dim>
3673 </port>
3674 </input>
3675 <output>
3676 <port id="2" precision="FP32" names="318,x.45">
3677 <dim>-1</dim>
3678 <dim>-1</dim>
3679 <dim>384</dim>
3680 </port>
3681 </output>
3682 </layer>
3683 <layer id="241" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
3684 <data element_type="i64" shape="4" offset="48269848" size="32" />
3685 <output>
3686 <port id="0" precision="I64">
3687 <dim>4</dim>
3688 </port>
3689 </output>
3690 </layer>
3691 <layer id="242" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3692 <data special_zero="true" />
3693 <input>
3694 <port id="0" precision="FP32">
3695 <dim>-1</dim>
3696 <dim>-1</dim>
3697 <dim>384</dim>
3698 </port>
3699 <port id="1" precision="I64">
3700 <dim>4</dim>
3701 </port>
3702 </input>
3703 <output>
3704 <port id="2" precision="FP32" names="322,x.47">
3705 <dim>-1</dim>
3706 <dim>-1</dim>
3707 <dim>12</dim>
3708 <dim>32</dim>
3709 </port>
3710 </output>
3711 </layer>
3712 <layer id="243" name="Constant_85173" type="Const" version="opset1">
3713 <data element_type="i64" shape="4" offset="48269880" size="32" />
3714 <output>
3715 <port id="0" precision="I64" names="323">
3716 <dim>4</dim>
3717 </port>
3718 </output>
3719 </layer>
3720 <layer id="244" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3721 <input>
3722 <port id="0" precision="FP32">
3723 <dim>-1</dim>
3724 <dim>-1</dim>
3725 <dim>12</dim>
3726 <dim>32</dim>
3727 </port>
3728 <port id="1" precision="I64">
3729 <dim>4</dim>
3730 </port>
3731 </input>
3732 <output>
3733 <port id="2" precision="FP32" names="324">
3734 <dim>-1</dim>
3735 <dim>12</dim>
3736 <dim>-1</dim>
3737 <dim>32</dim>
3738 </port>
3739 </output>
3740 </layer>
3741 <layer id="245" name="__module.bert.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3742 <data causal="false" />
3743 <input>
3744 <port id="0" precision="FP32">
3745 <dim>-1</dim>
3746 <dim>12</dim>
3747 <dim>-1</dim>
3748 <dim>32</dim>
3749 </port>
3750 <port id="1" precision="FP32">
3751 <dim>-1</dim>
3752 <dim>12</dim>
3753 <dim>-1</dim>
3754 <dim>32</dim>
3755 </port>
3756 <port id="2" precision="FP32">
3757 <dim>-1</dim>
3758 <dim>12</dim>
3759 <dim>-1</dim>
3760 <dim>32</dim>
3761 </port>
3762 <port id="3" precision="FP32">
3763 <dim>-1</dim>
3764 <dim>1</dim>
3765 <dim>-1</dim>
3766 <dim>-1</dim>
3767 </port>
3768 </input>
3769 <output>
3770 <port id="4" precision="FP32" names="325,attn_output.13">
3771 <dim>-1</dim>
3772 <dim>12</dim>
3773 <dim>-1</dim>
3774 <dim>32</dim>
3775 </port>
3776 </output>
3777 </layer>
3778 <layer id="246" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3779 <data element_type="i32" shape="4" offset="49452664" size="16" />
3780 <output>
3781 <port id="0" precision="I32">
3782 <dim>4</dim>
3783 </port>
3784 </output>
3785 </layer>
3786 <layer id="247" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3787 <input>
3788 <port id="0" precision="FP32">
3789 <dim>-1</dim>
3790 <dim>12</dim>
3791 <dim>-1</dim>
3792 <dim>32</dim>
3793 </port>
3794 <port id="1" precision="I32">
3795 <dim>4</dim>
3796 </port>
3797 </input>
3798 <output>
3799 <port id="2" precision="FP32" names="326,attn_output.15">
3800 <dim>-1</dim>
3801 <dim>-1</dim>
3802 <dim>12</dim>
3803 <dim>32</dim>
3804 </port>
3805 </output>
3806 </layer>
3807 <layer id="248" name="Constant_90446" type="Const" version="opset1">
3808 <data element_type="i64" shape="3" offset="49452680" size="24" />
3809 <output>
3810 <port id="0" precision="I64">
3811 <dim>3</dim>
3812 </port>
3813 </output>
3814 </layer>
3815 <layer id="249" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3816 <data special_zero="true" />
3817 <input>
3818 <port id="0" precision="FP32">
3819 <dim>-1</dim>
3820 <dim>-1</dim>
3821 <dim>12</dim>
3822 <dim>32</dim>
3823 </port>
3824 <port id="1" precision="I64">
3825 <dim>3</dim>
3826 </port>
3827 </input>
3828 <output>
3829 <port id="2" precision="FP32" names="328">
3830 <dim>-1</dim>
3831 <dim>-1</dim>
3832 <dim>384</dim>
3833 </port>
3834 </output>
3835 </layer>
3836 <layer id="250" name="self.bert.encoder.layer.3.attention.output.dense.weight" type="Const" version="opset1">
3837 <data element_type="f32" shape="384, 384" offset="70746272" size="589824" />
3838 <output>
3839 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.output.dense.weight">
3840 <dim>384</dim>
3841 <dim>384</dim>
3842 </port>
3843 </output>
3844 </layer>
3845 <layer id="251" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3846 <data transpose_a="false" transpose_b="true" />
3847 <input>
3848 <port id="0" precision="FP32">
3849 <dim>-1</dim>
3850 <dim>-1</dim>
3851 <dim>384</dim>
3852 </port>
3853 <port id="1" precision="FP32">
3854 <dim>384</dim>
3855 <dim>384</dim>
3856 </port>
3857 </input>
3858 <output>
3859 <port id="2" precision="FP32">
3860 <dim>-1</dim>
3861 <dim>-1</dim>
3862 <dim>384</dim>
3863 </port>
3864 </output>
3865 </layer>
3866 <layer id="252" name="Constant_90338" type="Const" version="opset1">
3867 <data element_type="f32" shape="1, 1, 384" offset="71336096" size="1536" />
3868 <output>
3869 <port id="0" precision="FP32">
3870 <dim>1</dim>
3871 <dim>1</dim>
3872 <dim>384</dim>
3873 </port>
3874 </output>
3875 </layer>
3876 <layer id="253" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3877 <data auto_broadcast="numpy" />
3878 <input>
3879 <port id="0" precision="FP32">
3880 <dim>-1</dim>
3881 <dim>-1</dim>
3882 <dim>384</dim>
3883 </port>
3884 <port id="1" precision="FP32">
3885 <dim>1</dim>
3886 <dim>1</dim>
3887 <dim>384</dim>
3888 </port>
3889 </input>
3890 <output>
3891 <port id="2" precision="FP32" names="333,input.15">
3892 <dim>-1</dim>
3893 <dim>-1</dim>
3894 <dim>384</dim>
3895 </port>
3896 </output>
3897 </layer>
3898 <layer id="254" name="__module.bert.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
3899 <data auto_broadcast="numpy" />
3900 <input>
3901 <port id="0" precision="FP32">
3902 <dim>-1</dim>
3903 <dim>-1</dim>
3904 <dim>384</dim>
3905 </port>
3906 <port id="1" precision="FP32">
3907 <dim>-1</dim>
3908 <dim>-1</dim>
3909 <dim>384</dim>
3910 </port>
3911 </input>
3912 <output>
3913 <port id="2" precision="FP32" names="335">
3914 <dim>-1</dim>
3915 <dim>-1</dim>
3916 <dim>384</dim>
3917 </port>
3918 </output>
3919 </layer>
3920 <layer id="255" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3921 <data element_type="i32" shape="1" offset="47675412" size="4" />
3922 <output>
3923 <port id="0" precision="I32">
3924 <dim>1</dim>
3925 </port>
3926 </output>
3927 </layer>
3928 <layer id="256" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3929 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3930 <input>
3931 <port id="0" precision="FP32">
3932 <dim>-1</dim>
3933 <dim>-1</dim>
3934 <dim>384</dim>
3935 </port>
3936 <port id="1" precision="I32">
3937 <dim>1</dim>
3938 </port>
3939 </input>
3940 <output>
3941 <port id="2" precision="FP32">
3942 <dim>-1</dim>
3943 <dim>-1</dim>
3944 <dim>384</dim>
3945 </port>
3946 </output>
3947 </layer>
3948 <layer id="257" name="Constant_90339" type="Const" version="opset1">
3949 <data element_type="f32" shape="1, 1, 384" offset="71337632" size="1536" />
3950 <output>
3951 <port id="0" precision="FP32">
3952 <dim>1</dim>
3953 <dim>1</dim>
3954 <dim>384</dim>
3955 </port>
3956 </output>
3957 </layer>
3958 <layer id="258" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3959 <data auto_broadcast="numpy" />
3960 <input>
3961 <port id="0" precision="FP32">
3962 <dim>-1</dim>
3963 <dim>-1</dim>
3964 <dim>384</dim>
3965 </port>
3966 <port id="1" precision="FP32">
3967 <dim>1</dim>
3968 <dim>1</dim>
3969 <dim>384</dim>
3970 </port>
3971 </input>
3972 <output>
3973 <port id="2" precision="FP32">
3974 <dim>-1</dim>
3975 <dim>-1</dim>
3976 <dim>384</dim>
3977 </port>
3978 </output>
3979 </layer>
3980 <layer id="259" name="Constant_90340" type="Const" version="opset1">
3981 <data element_type="f32" shape="1, 1, 384" offset="71339168" size="1536" />
3982 <output>
3983 <port id="0" precision="FP32">
3984 <dim>1</dim>
3985 <dim>1</dim>
3986 <dim>384</dim>
3987 </port>
3988 </output>
3989 </layer>
3990 <layer id="260" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3991 <data auto_broadcast="numpy" />
3992 <input>
3993 <port id="0" precision="FP32">
3994 <dim>-1</dim>
3995 <dim>-1</dim>
3996 <dim>384</dim>
3997 </port>
3998 <port id="1" precision="FP32">
3999 <dim>1</dim>
4000 <dim>1</dim>
4001 <dim>384</dim>
4002 </port>
4003 </input>
4004 <output>
4005 <port id="2" precision="FP32" names="339,input_tensor.7">
4006 <dim>-1</dim>
4007 <dim>-1</dim>
4008 <dim>384</dim>
4009 </port>
4010 </output>
4011 </layer>
4012 <layer id="261" name="self.bert.encoder.layer.3.intermediate.dense.weight" type="Const" version="opset1">
4013 <data element_type="f32" shape="1536, 384" offset="71340704" size="2359296" />
4014 <output>
4015 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.intermediate.dense.weight">
4016 <dim>1536</dim>
4017 <dim>384</dim>
4018 </port>
4019 </output>
4020 </layer>
4021 <layer id="262" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4022 <data transpose_a="false" transpose_b="true" />
4023 <input>
4024 <port id="0" precision="FP32">
4025 <dim>-1</dim>
4026 <dim>-1</dim>
4027 <dim>384</dim>
4028 </port>
4029 <port id="1" precision="FP32">
4030 <dim>1536</dim>
4031 <dim>384</dim>
4032 </port>
4033 </input>
4034 <output>
4035 <port id="2" precision="FP32">
4036 <dim>-1</dim>
4037 <dim>-1</dim>
4038 <dim>1536</dim>
4039 </port>
4040 </output>
4041 </layer>
4042 <layer id="263" name="Constant_90341" type="Const" version="opset1">
4043 <data element_type="f32" shape="1, 1, 1536" offset="73700000" size="6144" />
4044 <output>
4045 <port id="0" precision="FP32">
4046 <dim>1</dim>
4047 <dim>1</dim>
4048 <dim>1536</dim>
4049 </port>
4050 </output>
4051 </layer>
4052 <layer id="264" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
4053 <data auto_broadcast="numpy" />
4054 <input>
4055 <port id="0" precision="FP32">
4056 <dim>-1</dim>
4057 <dim>-1</dim>
4058 <dim>1536</dim>
4059 </port>
4060 <port id="1" precision="FP32">
4061 <dim>1</dim>
4062 <dim>1</dim>
4063 <dim>1536</dim>
4064 </port>
4065 </input>
4066 <output>
4067 <port id="2" precision="FP32" names="343">
4068 <dim>-1</dim>
4069 <dim>-1</dim>
4070 <dim>1536</dim>
4071 </port>
4072 </output>
4073 </layer>
4074 <layer id="265" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
4075 <data approximation_mode="ERF" />
4076 <input>
4077 <port id="0" precision="FP32">
4078 <dim>-1</dim>
4079 <dim>-1</dim>
4080 <dim>1536</dim>
4081 </port>
4082 </input>
4083 <output>
4084 <port id="1" precision="FP32" names="344">
4085 <dim>-1</dim>
4086 <dim>-1</dim>
4087 <dim>1536</dim>
4088 </port>
4089 </output>
4090 </layer>
4091 <layer id="266" name="self.bert.encoder.layer.3.output.dense.weight" type="Const" version="opset1">
4092 <data element_type="f32" shape="384, 1536" offset="73706144" size="2359296" />
4093 <output>
4094 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.output.dense.weight">
4095 <dim>384</dim>
4096 <dim>1536</dim>
4097 </port>
4098 </output>
4099 </layer>
4100 <layer id="267" name="__module.bert.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4101 <data transpose_a="false" transpose_b="true" />
4102 <input>
4103 <port id="0" precision="FP32">
4104 <dim>-1</dim>
4105 <dim>-1</dim>
4106 <dim>1536</dim>
4107 </port>
4108 <port id="1" precision="FP32">
4109 <dim>384</dim>
4110 <dim>1536</dim>
4111 </port>
4112 </input>
4113 <output>
4114 <port id="2" precision="FP32">
4115 <dim>-1</dim>
4116 <dim>-1</dim>
4117 <dim>384</dim>
4118 </port>
4119 </output>
4120 </layer>
4121 <layer id="268" name="Constant_90342" type="Const" version="opset1">
4122 <data element_type="f32" shape="1, 1, 384" offset="76065440" size="1536" />
4123 <output>
4124 <port id="0" precision="FP32">
4125 <dim>1</dim>
4126 <dim>1</dim>
4127 <dim>384</dim>
4128 </port>
4129 </output>
4130 </layer>
4131 <layer id="269" name="__module.bert.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
4132 <data auto_broadcast="numpy" />
4133 <input>
4134 <port id="0" precision="FP32">
4135 <dim>-1</dim>
4136 <dim>-1</dim>
4137 <dim>384</dim>
4138 </port>
4139 <port id="1" precision="FP32">
4140 <dim>1</dim>
4141 <dim>1</dim>
4142 <dim>384</dim>
4143 </port>
4144 </input>
4145 <output>
4146 <port id="2" precision="FP32" names="349,input.17">
4147 <dim>-1</dim>
4148 <dim>-1</dim>
4149 <dim>384</dim>
4150 </port>
4151 </output>
4152 </layer>
4153 <layer id="270" name="__module.bert.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
4154 <data auto_broadcast="numpy" />
4155 <input>
4156 <port id="0" precision="FP32">
4157 <dim>-1</dim>
4158 <dim>-1</dim>
4159 <dim>384</dim>
4160 </port>
4161 <port id="1" precision="FP32">
4162 <dim>-1</dim>
4163 <dim>-1</dim>
4164 <dim>384</dim>
4165 </port>
4166 </input>
4167 <output>
4168 <port id="2" precision="FP32" names="351">
4169 <dim>-1</dim>
4170 <dim>-1</dim>
4171 <dim>384</dim>
4172 </port>
4173 </output>
4174 </layer>
4175 <layer id="271" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4176 <data element_type="i32" shape="1" offset="47675412" size="4" />
4177 <output>
4178 <port id="0" precision="I32">
4179 <dim>1</dim>
4180 </port>
4181 </output>
4182 </layer>
4183 <layer id="272" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4184 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4185 <input>
4186 <port id="0" precision="FP32">
4187 <dim>-1</dim>
4188 <dim>-1</dim>
4189 <dim>384</dim>
4190 </port>
4191 <port id="1" precision="I32">
4192 <dim>1</dim>
4193 </port>
4194 </input>
4195 <output>
4196 <port id="2" precision="FP32">
4197 <dim>-1</dim>
4198 <dim>-1</dim>
4199 <dim>384</dim>
4200 </port>
4201 </output>
4202 </layer>
4203 <layer id="273" name="Constant_90343" type="Const" version="opset1">
4204 <data element_type="f32" shape="1, 1, 384" offset="76066976" size="1536" />
4205 <output>
4206 <port id="0" precision="FP32">
4207 <dim>1</dim>
4208 <dim>1</dim>
4209 <dim>384</dim>
4210 </port>
4211 </output>
4212 </layer>
4213 <layer id="274" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4214 <data auto_broadcast="numpy" />
4215 <input>
4216 <port id="0" precision="FP32">
4217 <dim>-1</dim>
4218 <dim>-1</dim>
4219 <dim>384</dim>
4220 </port>
4221 <port id="1" precision="FP32">
4222 <dim>1</dim>
4223 <dim>1</dim>
4224 <dim>384</dim>
4225 </port>
4226 </input>
4227 <output>
4228 <port id="2" precision="FP32">
4229 <dim>-1</dim>
4230 <dim>-1</dim>
4231 <dim>384</dim>
4232 </port>
4233 </output>
4234 </layer>
4235 <layer id="275" name="Constant_90344" type="Const" version="opset1">
4236 <data element_type="f32" shape="1, 1, 384" offset="76068512" size="1536" />
4237 <output>
4238 <port id="0" precision="FP32">
4239 <dim>1</dim>
4240 <dim>1</dim>
4241 <dim>384</dim>
4242 </port>
4243 </output>
4244 </layer>
4245 <layer id="276" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4246 <data auto_broadcast="numpy" />
4247 <input>
4248 <port id="0" precision="FP32">
4249 <dim>-1</dim>
4250 <dim>-1</dim>
4251 <dim>384</dim>
4252 </port>
4253 <port id="1" precision="FP32">
4254 <dim>1</dim>
4255 <dim>1</dim>
4256 <dim>384</dim>
4257 </port>
4258 </input>
4259 <output>
4260 <port id="2" precision="FP32" names="355,hidden_states.25">
4261 <dim>-1</dim>
4262 <dim>-1</dim>
4263 <dim>384</dim>
4264 </port>
4265 </output>
4266 </layer>
4267 <layer id="277" name="self.bert.encoder.layer.4.attention.self.query.weight" type="Const" version="opset1">
4268 <data element_type="f32" shape="384, 384" offset="76070048" size="589824" />
4269 <output>
4270 <port id="0" precision="FP32" names="self.bert.encoder.layer.4.attention.self.query.weight">
4271 <dim>384</dim>
4272 <dim>384</dim>
4273 </port>
4274 </output>
4275 </layer>
4276 <layer id="278" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
4277 <data transpose_a="false" transpose_b="true" />
4278 <input>
4279 <port id="0" precision="FP32">
4280 <dim>-1</dim>
4281 <dim>-1</dim>
4282 <dim>384</dim>
4283 </port>
4284 <port id="1" precision="FP32">
4285 <dim>384</dim>
4286 <dim>384</dim>
4287 </port>
4288 </input>
4289 <output>
4290 <port id="2" precision="FP32">
4291 <dim>-1</dim>
4292 <dim>-1</dim>
4293 <dim>384</dim>
4294 </port>
4295 </output>
4296 </layer>
4297 <layer id="279" name="Constant_90345" type="Const" version="opset1">
4298 <data element_type="f32" shape="1, 1, 384" offset="76659872" size="1536" />
4299 <output>
4300 <port id="0" precision="FP32">
4301 <dim>1</dim>
4302 <dim>1</dim>
4303 <dim>384</dim>
4304 </port>
4305 </output>
4306 </layer>
4307 <layer id="280" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add" type="Add" version="opset1">
4308 <data auto_broadcast="numpy" />
4309 <input>
4310 <port id="0" precision="FP32">
4311 <dim>-1</dim>
4312 <dim>-1</dim>
4313 <dim>384</dim>
4314 </port>
4315 <port id="1" precision="FP32">
4316 <dim>1</dim>
4317 <dim>1</dim>
4318 <dim>384</dim>
4319 </port>
4320 </input>
4321 <output>
4322 <port id="2" precision="FP32" names="368,x.49">
4323 <dim>-1</dim>
4324 <dim>-1</dim>
4325 <dim>384</dim>
4326 </port>
4327 </output>
4328 </layer>
4329 <layer id="281" name="__module.bert.encoder.layer.4.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
4330 <data element_type="i64" shape="4" offset="48269848" size="32" />
4331 <output>
4332 <port id="0" precision="I64">
4333 <dim>4</dim>
4334 </port>
4335 </output>
4336 </layer>
4337 <layer id="282" name="__module.bert.encoder.layer.4.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
4338 <data special_zero="true" />
4339 <input>
4340 <port id="0" precision="FP32">
4341 <dim>-1</dim>
4342 <dim>-1</dim>
4343 <dim>384</dim>
4344 </port>
4345 <port id="1" precision="I64">
4346 <dim>4</dim>
4347 </port>
4348 </input>
4349 <output>
4350 <port id="2" precision="FP32" names="372,x.51">
4351 <dim>-1</dim>
4352 <dim>-1</dim>
4353 <dim>12</dim>
4354 <dim>32</dim>
4355 </port>
4356 </output>
4357 </layer>
4358 <layer id="283" name="Constant_85350" type="Const" version="opset1">
4359 <data element_type="i64" shape="4" offset="48269880" size="32" />
4360 <output>
4361 <port id="0" precision="I64" names="373">
4362 <dim>4</dim>
4363 </port>
4364 </output>
4365 </layer>
4366 <layer id="284" name="__module.bert.encoder.layer.4.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
4367 <input>
4368 <port id="0" precision="FP32">
4369 <dim>-1</dim>
4370 <dim>-1</dim>
4371 <dim>12</dim>
4372 <dim>32</dim>
4373 </port>
4374 <port id="1" precision="I64">
4375 <dim>4</dim>
4376 </port>
4377 </input>
4378 <output>
4379 <port id="2" precision="FP32" names="374">
4380 <dim>-1</dim>
4381 <dim>12</dim>
4382 <dim>-1</dim>
4383 <dim>32</dim>
4384 </port>
4385 </output>
4386 </layer>
4387 <layer id="285" name="self.bert.encoder.layer.4.attention.self.key.weight" type="Const" version="opset1">
4388 <data element_type="f32" shape="384, 384" offset="76661408" size="589824" />
4389 <output>
4390 <port id="0" precision="FP32" names="self.bert.encoder.layer.4.attention.self.key.weight">
4391 <dim>384</dim>
4392 <dim>384</dim>
4393 </port>
4394 </output>
4395 </layer>
4396 <layer id="286" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
4397 <data transpose_a="false" transpose_b="true" />
4398 <input>
4399 <port id="0" precision="FP32">
4400 <dim>-1</dim>
4401 <dim>-1</dim>
4402 <dim>384</dim>
4403 </port>
4404 <port id="1" precision="FP32">
4405 <dim>384</dim>
4406 <dim>384</dim>
4407 </port>
4408 </input>
4409 <output>
4410 <port id="2" precision="FP32">
4411 <dim>-1</dim>
4412 <dim>-1</dim>
4413 <dim>384</dim>
4414 </port>
4415 </output>
4416 </layer>
4417 <layer id="287" name="Constant_90346" type="Const" version="opset1">
4418 <data element_type="f32" shape="1, 1, 384" offset="77251232" size="1536" />
4419 <output>
4420 <port id="0" precision="FP32">
4421 <dim>1</dim>
4422 <dim>1</dim>
4423 <dim>384</dim>
4424 </port>
4425 </output>
4426 </layer>
4427 <layer id="288" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add" type="Add" version="opset1">
4428 <data auto_broadcast="numpy" />
4429 <input>
4430 <port id="0" precision="FP32">
4431 <dim>-1</dim>
4432 <dim>-1</dim>
4433 <dim>384</dim>
4434 </port>
4435 <port id="1" precision="FP32">
4436 <dim>1</dim>
4437 <dim>1</dim>
4438 <dim>384</dim>
4439 </port>
4440 </input>
4441 <output>
4442 <port id="2" precision="FP32" names="377,x.53">
4443 <dim>-1</dim>
4444 <dim>-1</dim>
4445 <dim>384</dim>
4446 </port>
4447 </output>
4448 </layer>
4449 <layer id="289" name="__module.bert.encoder.layer.4.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
4450 <data element_type="i64" shape="4" offset="48269848" size="32" />
4451 <output>
4452 <port id="0" precision="I64">
4453 <dim>4</dim>
4454 </port>
4455 </output>
4456 </layer>
4457 <layer id="290" name="__module.bert.encoder.layer.4.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
4458 <data special_zero="true" />
4459 <input>
4460 <port id="0" precision="FP32">
4461 <dim>-1</dim>
4462 <dim>-1</dim>
4463 <dim>384</dim>
4464 </port>
4465 <port id="1" precision="I64">
4466 <dim>4</dim>
4467 </port>
4468 </input>
4469 <output>
4470 <port id="2" precision="FP32" names="381,x.55">
4471 <dim>-1</dim>
4472 <dim>-1</dim>
4473 <dim>12</dim>
4474 <dim>32</dim>
4475 </port>
4476 </output>
4477 </layer>
4478 <layer id="291" name="Constant_85373" type="Const" version="opset1">
4479 <data element_type="i64" shape="4" offset="48269880" size="32" />
4480 <output>
4481 <port id="0" precision="I64" names="382">
4482 <dim>4</dim>
4483 </port>
4484 </output>
4485 </layer>
4486 <layer id="292" name="__module.bert.encoder.layer.4.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
4487 <input>
4488 <port id="0" precision="FP32">
4489 <dim>-1</dim>
4490 <dim>-1</dim>
4491 <dim>12</dim>
4492 <dim>32</dim>
4493 </port>
4494 <port id="1" precision="I64">
4495 <dim>4</dim>
4496 </port>
4497 </input>
4498 <output>
4499 <port id="2" precision="FP32" names="383">
4500 <dim>-1</dim>
4501 <dim>12</dim>
4502 <dim>-1</dim>
4503 <dim>32</dim>
4504 </port>
4505 </output>
4506 </layer>
4507 <layer id="293" name="self.bert.encoder.layer.4.attention.self.value.weight" type="Const" version="opset1">
4508 <data element_type="f32" shape="384, 384" offset="77252768" size="589824" />
4509 <output>
4510 <port id="0" precision="FP32" names="self.bert.encoder.layer.4.attention.self.value.weight">
4511 <dim>384</dim>
4512 <dim>384</dim>
4513 </port>
4514 </output>
4515 </layer>
4516 <layer id="294" name="__module.bert.encoder.layer.4.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
4517 <data transpose_a="false" transpose_b="true" />
4518 <input>
4519 <port id="0" precision="FP32">
4520 <dim>-1</dim>
4521 <dim>-1</dim>
4522 <dim>384</dim>
4523 </port>
4524 <port id="1" precision="FP32">
4525 <dim>384</dim>
4526 <dim>384</dim>
4527 </port>
4528 </input>
4529 <output>
4530 <port id="2" precision="FP32">
4531 <dim>-1</dim>
4532 <dim>-1</dim>
4533 <dim>384</dim>
4534 </port>
4535 </output>
4536 </layer>
4537 <layer id="295" name="Constant_90347" type="Const" version="opset1">
4538 <data element_type="f32" shape="1, 1, 384" offset="77842592" size="1536" />
4539 <output>
4540 <port id="0" precision="FP32">
4541 <dim>1</dim>
4542 <dim>1</dim>
4543 <dim>384</dim>
4544 </port>
4545 </output>
4546 </layer>
4547 <layer id="296" name="__module.bert.encoder.layer.4.attention.self.value/aten::linear/Add" type="Add" version="opset1">
4548 <data auto_broadcast="numpy" />
4549 <input>
4550 <port id="0" precision="FP32">
4551 <dim>-1</dim>
4552 <dim>-1</dim>
4553 <dim>384</dim>
4554 </port>
4555 <port id="1" precision="FP32">
4556 <dim>1</dim>
4557 <dim>1</dim>
4558 <dim>384</dim>
4559 </port>
4560 </input>
4561 <output>
4562 <port id="2" precision="FP32" names="386,x.57">
4563 <dim>-1</dim>
4564 <dim>-1</dim>
4565 <dim>384</dim>
4566 </port>
4567 </output>
4568 </layer>
4569 <layer id="297" name="__module.bert.encoder.layer.4.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
4570 <data element_type="i64" shape="4" offset="48269848" size="32" />
4571 <output>
4572 <port id="0" precision="I64">
4573 <dim>4</dim>
4574 </port>
4575 </output>
4576 </layer>
4577 <layer id="298" name="__module.bert.encoder.layer.4.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
4578 <data special_zero="true" />
4579 <input>
4580 <port id="0" precision="FP32">
4581 <dim>-1</dim>
4582 <dim>-1</dim>
4583 <dim>384</dim>
4584 </port>
4585 <port id="1" precision="I64">
4586 <dim>4</dim>
4587 </port>
4588 </input>
4589 <output>
4590 <port id="2" precision="FP32" names="390,x.59">
4591 <dim>-1</dim>
4592 <dim>-1</dim>
4593 <dim>12</dim>
4594 <dim>32</dim>
4595 </port>
4596 </output>
4597 </layer>
4598 <layer id="299" name="Constant_85396" type="Const" version="opset1">
4599 <data element_type="i64" shape="4" offset="48269880" size="32" />
4600 <output>
4601 <port id="0" precision="I64" names="391">
4602 <dim>4</dim>
4603 </port>
4604 </output>
4605 </layer>
4606 <layer id="300" name="__module.bert.encoder.layer.4.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
4607 <input>
4608 <port id="0" precision="FP32">
4609 <dim>-1</dim>
4610 <dim>-1</dim>
4611 <dim>12</dim>
4612 <dim>32</dim>
4613 </port>
4614 <port id="1" precision="I64">
4615 <dim>4</dim>
4616 </port>
4617 </input>
4618 <output>
4619 <port id="2" precision="FP32" names="392">
4620 <dim>-1</dim>
4621 <dim>12</dim>
4622 <dim>-1</dim>
4623 <dim>32</dim>
4624 </port>
4625 </output>
4626 </layer>
4627 <layer id="301" name="__module.bert.encoder.layer.4.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
4628 <data causal="false" />
4629 <input>
4630 <port id="0" precision="FP32">
4631 <dim>-1</dim>
4632 <dim>12</dim>
4633 <dim>-1</dim>
4634 <dim>32</dim>
4635 </port>
4636 <port id="1" precision="FP32">
4637 <dim>-1</dim>
4638 <dim>12</dim>
4639 <dim>-1</dim>
4640 <dim>32</dim>
4641 </port>
4642 <port id="2" precision="FP32">
4643 <dim>-1</dim>
4644 <dim>12</dim>
4645 <dim>-1</dim>
4646 <dim>32</dim>
4647 </port>
4648 <port id="3" precision="FP32">
4649 <dim>-1</dim>
4650 <dim>1</dim>
4651 <dim>-1</dim>
4652 <dim>-1</dim>
4653 </port>
4654 </input>
4655 <output>
4656 <port id="4" precision="FP32" names="393,attn_output.17">
4657 <dim>-1</dim>
4658 <dim>12</dim>
4659 <dim>-1</dim>
4660 <dim>32</dim>
4661 </port>
4662 </output>
4663 </layer>
4664 <layer id="302" name="__module.bert.encoder.layer.4.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
4665 <data element_type="i32" shape="4" offset="49452664" size="16" />
4666 <output>
4667 <port id="0" precision="I32">
4668 <dim>4</dim>
4669 </port>
4670 </output>
4671 </layer>
4672 <layer id="303" name="__module.bert.encoder.layer.4.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
4673 <input>
4674 <port id="0" precision="FP32">
4675 <dim>-1</dim>
4676 <dim>12</dim>
4677 <dim>-1</dim>
4678 <dim>32</dim>
4679 </port>
4680 <port id="1" precision="I32">
4681 <dim>4</dim>
4682 </port>
4683 </input>
4684 <output>
4685 <port id="2" precision="FP32" names="394,attn_output.19">
4686 <dim>-1</dim>
4687 <dim>-1</dim>
4688 <dim>12</dim>
4689 <dim>32</dim>
4690 </port>
4691 </output>
4692 </layer>
4693 <layer id="304" name="Constant_90447" type="Const" version="opset1">
4694 <data element_type="i64" shape="3" offset="49452680" size="24" />
4695 <output>
4696 <port id="0" precision="I64">
4697 <dim>3</dim>
4698 </port>
4699 </output>
4700 </layer>
4701 <layer id="305" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
4702 <data special_zero="true" />
4703 <input>
4704 <port id="0" precision="FP32">
4705 <dim>-1</dim>
4706 <dim>-1</dim>
4707 <dim>12</dim>
4708 <dim>32</dim>
4709 </port>
4710 <port id="1" precision="I64">
4711 <dim>3</dim>
4712 </port>
4713 </input>
4714 <output>
4715 <port id="2" precision="FP32" names="396">
4716 <dim>-1</dim>
4717 <dim>-1</dim>
4718 <dim>384</dim>
4719 </port>
4720 </output>
4721 </layer>
4722 <layer id="306" name="self.bert.encoder.layer.4.attention.output.dense.weight" type="Const" version="opset1">
4723 <data element_type="f32" shape="384, 384" offset="77844128" size="589824" />
4724 <output>
4725 <port id="0" precision="FP32" names="self.bert.encoder.layer.4.attention.output.dense.weight">
4726 <dim>384</dim>
4727 <dim>384</dim>
4728 </port>
4729 </output>
4730 </layer>
4731 <layer id="307" name="__module.bert.encoder.layer.4.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4732 <data transpose_a="false" transpose_b="true" />
4733 <input>
4734 <port id="0" precision="FP32">
4735 <dim>-1</dim>
4736 <dim>-1</dim>
4737 <dim>384</dim>
4738 </port>
4739 <port id="1" precision="FP32">
4740 <dim>384</dim>
4741 <dim>384</dim>
4742 </port>
4743 </input>
4744 <output>
4745 <port id="2" precision="FP32">
4746 <dim>-1</dim>
4747 <dim>-1</dim>
4748 <dim>384</dim>
4749 </port>
4750 </output>
4751 </layer>
4752 <layer id="308" name="Constant_90348" type="Const" version="opset1">
4753 <data element_type="f32" shape="1, 1, 384" offset="78433952" size="1536" />
4754 <output>
4755 <port id="0" precision="FP32">
4756 <dim>1</dim>
4757 <dim>1</dim>
4758 <dim>384</dim>
4759 </port>
4760 </output>
4761 </layer>
4762 <layer id="309" name="__module.bert.encoder.layer.4.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
4763 <data auto_broadcast="numpy" />
4764 <input>
4765 <port id="0" precision="FP32">
4766 <dim>-1</dim>
4767 <dim>-1</dim>
4768 <dim>384</dim>
4769 </port>
4770 <port id="1" precision="FP32">
4771 <dim>1</dim>
4772 <dim>1</dim>
4773 <dim>384</dim>
4774 </port>
4775 </input>
4776 <output>
4777 <port id="2" precision="FP32" names="401,input.19">
4778 <dim>-1</dim>
4779 <dim>-1</dim>
4780 <dim>384</dim>
4781 </port>
4782 </output>
4783 </layer>
4784 <layer id="310" name="__module.bert.encoder.layer.4.attention.output/aten::add/Add" type="Add" version="opset1">
4785 <data auto_broadcast="numpy" />
4786 <input>
4787 <port id="0" precision="FP32">
4788 <dim>-1</dim>
4789 <dim>-1</dim>
4790 <dim>384</dim>
4791 </port>
4792 <port id="1" precision="FP32">
4793 <dim>-1</dim>
4794 <dim>-1</dim>
4795 <dim>384</dim>
4796 </port>
4797 </input>
4798 <output>
4799 <port id="2" precision="FP32" names="403">
4800 <dim>-1</dim>
4801 <dim>-1</dim>
4802 <dim>384</dim>
4803 </port>
4804 </output>
4805 </layer>
4806 <layer id="311" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4807 <data element_type="i32" shape="1" offset="47675412" size="4" />
4808 <output>
4809 <port id="0" precision="I32">
4810 <dim>1</dim>
4811 </port>
4812 </output>
4813 </layer>
4814 <layer id="312" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4815 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4816 <input>
4817 <port id="0" precision="FP32">
4818 <dim>-1</dim>
4819 <dim>-1</dim>
4820 <dim>384</dim>
4821 </port>
4822 <port id="1" precision="I32">
4823 <dim>1</dim>
4824 </port>
4825 </input>
4826 <output>
4827 <port id="2" precision="FP32">
4828 <dim>-1</dim>
4829 <dim>-1</dim>
4830 <dim>384</dim>
4831 </port>
4832 </output>
4833 </layer>
4834 <layer id="313" name="Constant_90349" type="Const" version="opset1">
4835 <data element_type="f32" shape="1, 1, 384" offset="78435488" size="1536" />
4836 <output>
4837 <port id="0" precision="FP32">
4838 <dim>1</dim>
4839 <dim>1</dim>
4840 <dim>384</dim>
4841 </port>
4842 </output>
4843 </layer>
4844 <layer id="314" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4845 <data auto_broadcast="numpy" />
4846 <input>
4847 <port id="0" precision="FP32">
4848 <dim>-1</dim>
4849 <dim>-1</dim>
4850 <dim>384</dim>
4851 </port>
4852 <port id="1" precision="FP32">
4853 <dim>1</dim>
4854 <dim>1</dim>
4855 <dim>384</dim>
4856 </port>
4857 </input>
4858 <output>
4859 <port id="2" precision="FP32">
4860 <dim>-1</dim>
4861 <dim>-1</dim>
4862 <dim>384</dim>
4863 </port>
4864 </output>
4865 </layer>
4866 <layer id="315" name="Constant_90350" type="Const" version="opset1">
4867 <data element_type="f32" shape="1, 1, 384" offset="78437024" size="1536" />
4868 <output>
4869 <port id="0" precision="FP32">
4870 <dim>1</dim>
4871 <dim>1</dim>
4872 <dim>384</dim>
4873 </port>
4874 </output>
4875 </layer>
4876 <layer id="316" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4877 <data auto_broadcast="numpy" />
4878 <input>
4879 <port id="0" precision="FP32">
4880 <dim>-1</dim>
4881 <dim>-1</dim>
4882 <dim>384</dim>
4883 </port>
4884 <port id="1" precision="FP32">
4885 <dim>1</dim>
4886 <dim>1</dim>
4887 <dim>384</dim>
4888 </port>
4889 </input>
4890 <output>
4891 <port id="2" precision="FP32" names="407,input_tensor.9">
4892 <dim>-1</dim>
4893 <dim>-1</dim>
4894 <dim>384</dim>
4895 </port>
4896 </output>
4897 </layer>
4898 <layer id="317" name="self.bert.encoder.layer.4.intermediate.dense.weight" type="Const" version="opset1">
4899 <data element_type="f32" shape="1536, 384" offset="78438560" size="2359296" />
4900 <output>
4901 <port id="0" precision="FP32" names="self.bert.encoder.layer.4.intermediate.dense.weight">
4902 <dim>1536</dim>
4903 <dim>384</dim>
4904 </port>
4905 </output>
4906 </layer>
4907 <layer id="318" name="__module.bert.encoder.layer.4.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4908 <data transpose_a="false" transpose_b="true" />
4909 <input>
4910 <port id="0" precision="FP32">
4911 <dim>-1</dim>
4912 <dim>-1</dim>
4913 <dim>384</dim>
4914 </port>
4915 <port id="1" precision="FP32">
4916 <dim>1536</dim>
4917 <dim>384</dim>
4918 </port>
4919 </input>
4920 <output>
4921 <port id="2" precision="FP32">
4922 <dim>-1</dim>
4923 <dim>-1</dim>
4924 <dim>1536</dim>
4925 </port>
4926 </output>
4927 </layer>
4928 <layer id="319" name="Constant_90351" type="Const" version="opset1">
4929 <data element_type="f32" shape="1, 1, 1536" offset="80797856" size="6144" />
4930 <output>
4931 <port id="0" precision="FP32">
4932 <dim>1</dim>
4933 <dim>1</dim>
4934 <dim>1536</dim>
4935 </port>
4936 </output>
4937 </layer>
4938 <layer id="320" name="__module.bert.encoder.layer.4.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
4939 <data auto_broadcast="numpy" />
4940 <input>
4941 <port id="0" precision="FP32">
4942 <dim>-1</dim>
4943 <dim>-1</dim>
4944 <dim>1536</dim>
4945 </port>
4946 <port id="1" precision="FP32">
4947 <dim>1</dim>
4948 <dim>1</dim>
4949 <dim>1536</dim>
4950 </port>
4951 </input>
4952 <output>
4953 <port id="2" precision="FP32" names="411">
4954 <dim>-1</dim>
4955 <dim>-1</dim>
4956 <dim>1536</dim>
4957 </port>
4958 </output>
4959 </layer>
4960 <layer id="321" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
4961 <data approximation_mode="ERF" />
4962 <input>
4963 <port id="0" precision="FP32">
4964 <dim>-1</dim>
4965 <dim>-1</dim>
4966 <dim>1536</dim>
4967 </port>
4968 </input>
4969 <output>
4970 <port id="1" precision="FP32" names="412">
4971 <dim>-1</dim>
4972 <dim>-1</dim>
4973 <dim>1536</dim>
4974 </port>
4975 </output>
4976 </layer>
4977 <layer id="322" name="self.bert.encoder.layer.4.output.dense.weight" type="Const" version="opset1">
4978 <data element_type="f32" shape="384, 1536" offset="80804000" size="2359296" />
4979 <output>
4980 <port id="0" precision="FP32" names="self.bert.encoder.layer.4.output.dense.weight">
4981 <dim>384</dim>
4982 <dim>1536</dim>
4983 </port>
4984 </output>
4985 </layer>
4986 <layer id="323" name="__module.bert.encoder.layer.4.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4987 <data transpose_a="false" transpose_b="true" />
4988 <input>
4989 <port id="0" precision="FP32">
4990 <dim>-1</dim>
4991 <dim>-1</dim>
4992 <dim>1536</dim>
4993 </port>
4994 <port id="1" precision="FP32">
4995 <dim>384</dim>
4996 <dim>1536</dim>
4997 </port>
4998 </input>
4999 <output>
5000 <port id="2" precision="FP32">
5001 <dim>-1</dim>
5002 <dim>-1</dim>
5003 <dim>384</dim>
5004 </port>
5005 </output>
5006 </layer>
5007 <layer id="324" name="Constant_90352" type="Const" version="opset1">
5008 <data element_type="f32" shape="1, 1, 384" offset="83163296" size="1536" />
5009 <output>
5010 <port id="0" precision="FP32">
5011 <dim>1</dim>
5012 <dim>1</dim>
5013 <dim>384</dim>
5014 </port>
5015 </output>
5016 </layer>
5017 <layer id="325" name="__module.bert.encoder.layer.4.output.dense/aten::linear/Add" type="Add" version="opset1">
5018 <data auto_broadcast="numpy" />
5019 <input>
5020 <port id="0" precision="FP32">
5021 <dim>-1</dim>
5022 <dim>-1</dim>
5023 <dim>384</dim>
5024 </port>
5025 <port id="1" precision="FP32">
5026 <dim>1</dim>
5027 <dim>1</dim>
5028 <dim>384</dim>
5029 </port>
5030 </input>
5031 <output>
5032 <port id="2" precision="FP32" names="417,input.21">
5033 <dim>-1</dim>
5034 <dim>-1</dim>
5035 <dim>384</dim>
5036 </port>
5037 </output>
5038 </layer>
5039 <layer id="326" name="__module.bert.encoder.layer.4.output/aten::add/Add" type="Add" version="opset1">
5040 <data auto_broadcast="numpy" />
5041 <input>
5042 <port id="0" precision="FP32">
5043 <dim>-1</dim>
5044 <dim>-1</dim>
5045 <dim>384</dim>
5046 </port>
5047 <port id="1" precision="FP32">
5048 <dim>-1</dim>
5049 <dim>-1</dim>
5050 <dim>384</dim>
5051 </port>
5052 </input>
5053 <output>
5054 <port id="2" precision="FP32" names="419">
5055 <dim>-1</dim>
5056 <dim>-1</dim>
5057 <dim>384</dim>
5058 </port>
5059 </output>
5060 </layer>
5061 <layer id="327" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5062 <data element_type="i32" shape="1" offset="47675412" size="4" />
5063 <output>
5064 <port id="0" precision="I32">
5065 <dim>1</dim>
5066 </port>
5067 </output>
5068 </layer>
5069 <layer id="328" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5070 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5071 <input>
5072 <port id="0" precision="FP32">
5073 <dim>-1</dim>
5074 <dim>-1</dim>
5075 <dim>384</dim>
5076 </port>
5077 <port id="1" precision="I32">
5078 <dim>1</dim>
5079 </port>
5080 </input>
5081 <output>
5082 <port id="2" precision="FP32">
5083 <dim>-1</dim>
5084 <dim>-1</dim>
5085 <dim>384</dim>
5086 </port>
5087 </output>
5088 </layer>
5089 <layer id="329" name="Constant_90353" type="Const" version="opset1">
5090 <data element_type="f32" shape="1, 1, 384" offset="83164832" size="1536" />
5091 <output>
5092 <port id="0" precision="FP32">
5093 <dim>1</dim>
5094 <dim>1</dim>
5095 <dim>384</dim>
5096 </port>
5097 </output>
5098 </layer>
5099 <layer id="330" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5100 <data auto_broadcast="numpy" />
5101 <input>
5102 <port id="0" precision="FP32">
5103 <dim>-1</dim>
5104 <dim>-1</dim>
5105 <dim>384</dim>
5106 </port>
5107 <port id="1" precision="FP32">
5108 <dim>1</dim>
5109 <dim>1</dim>
5110 <dim>384</dim>
5111 </port>
5112 </input>
5113 <output>
5114 <port id="2" precision="FP32">
5115 <dim>-1</dim>
5116 <dim>-1</dim>
5117 <dim>384</dim>
5118 </port>
5119 </output>
5120 </layer>
5121 <layer id="331" name="Constant_90354" type="Const" version="opset1">
5122 <data element_type="f32" shape="1, 1, 384" offset="83166368" size="1536" />
5123 <output>
5124 <port id="0" precision="FP32">
5125 <dim>1</dim>
5126 <dim>1</dim>
5127 <dim>384</dim>
5128 </port>
5129 </output>
5130 </layer>
5131 <layer id="332" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5132 <data auto_broadcast="numpy" />
5133 <input>
5134 <port id="0" precision="FP32">
5135 <dim>-1</dim>
5136 <dim>-1</dim>
5137 <dim>384</dim>
5138 </port>
5139 <port id="1" precision="FP32">
5140 <dim>1</dim>
5141 <dim>1</dim>
5142 <dim>384</dim>
5143 </port>
5144 </input>
5145 <output>
5146 <port id="2" precision="FP32" names="423,hidden_states.31">
5147 <dim>-1</dim>
5148 <dim>-1</dim>
5149 <dim>384</dim>
5150 </port>
5151 </output>
5152 </layer>
5153 <layer id="333" name="self.bert.encoder.layer.5.attention.self.query.weight" type="Const" version="opset1">
5154 <data element_type="f32" shape="384, 384" offset="83167904" size="589824" />
5155 <output>
5156 <port id="0" precision="FP32" names="self.bert.encoder.layer.5.attention.self.query.weight">
5157 <dim>384</dim>
5158 <dim>384</dim>
5159 </port>
5160 </output>
5161 </layer>
5162 <layer id="334" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
5163 <data transpose_a="false" transpose_b="true" />
5164 <input>
5165 <port id="0" precision="FP32">
5166 <dim>-1</dim>
5167 <dim>-1</dim>
5168 <dim>384</dim>
5169 </port>
5170 <port id="1" precision="FP32">
5171 <dim>384</dim>
5172 <dim>384</dim>
5173 </port>
5174 </input>
5175 <output>
5176 <port id="2" precision="FP32">
5177 <dim>-1</dim>
5178 <dim>-1</dim>
5179 <dim>384</dim>
5180 </port>
5181 </output>
5182 </layer>
5183 <layer id="335" name="Constant_90355" type="Const" version="opset1">
5184 <data element_type="f32" shape="1, 1, 384" offset="83757728" size="1536" />
5185 <output>
5186 <port id="0" precision="FP32">
5187 <dim>1</dim>
5188 <dim>1</dim>
5189 <dim>384</dim>
5190 </port>
5191 </output>
5192 </layer>
5193 <layer id="336" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add" type="Add" version="opset1">
5194 <data auto_broadcast="numpy" />
5195 <input>
5196 <port id="0" precision="FP32">
5197 <dim>-1</dim>
5198 <dim>-1</dim>
5199 <dim>384</dim>
5200 </port>
5201 <port id="1" precision="FP32">
5202 <dim>1</dim>
5203 <dim>1</dim>
5204 <dim>384</dim>
5205 </port>
5206 </input>
5207 <output>
5208 <port id="2" precision="FP32" names="436,x.61">
5209 <dim>-1</dim>
5210 <dim>-1</dim>
5211 <dim>384</dim>
5212 </port>
5213 </output>
5214 </layer>
5215 <layer id="337" name="__module.bert.encoder.layer.5.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
5216 <data element_type="i64" shape="4" offset="48269848" size="32" />
5217 <output>
5218 <port id="0" precision="I64">
5219 <dim>4</dim>
5220 </port>
5221 </output>
5222 </layer>
5223 <layer id="338" name="__module.bert.encoder.layer.5.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
5224 <data special_zero="true" />
5225 <input>
5226 <port id="0" precision="FP32">
5227 <dim>-1</dim>
5228 <dim>-1</dim>
5229 <dim>384</dim>
5230 </port>
5231 <port id="1" precision="I64">
5232 <dim>4</dim>
5233 </port>
5234 </input>
5235 <output>
5236 <port id="2" precision="FP32" names="440,x.63">
5237 <dim>-1</dim>
5238 <dim>-1</dim>
5239 <dim>12</dim>
5240 <dim>32</dim>
5241 </port>
5242 </output>
5243 </layer>
5244 <layer id="339" name="Constant_85573" type="Const" version="opset1">
5245 <data element_type="i64" shape="4" offset="48269880" size="32" />
5246 <output>
5247 <port id="0" precision="I64" names="441">
5248 <dim>4</dim>
5249 </port>
5250 </output>
5251 </layer>
5252 <layer id="340" name="__module.bert.encoder.layer.5.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
5253 <input>
5254 <port id="0" precision="FP32">
5255 <dim>-1</dim>
5256 <dim>-1</dim>
5257 <dim>12</dim>
5258 <dim>32</dim>
5259 </port>
5260 <port id="1" precision="I64">
5261 <dim>4</dim>
5262 </port>
5263 </input>
5264 <output>
5265 <port id="2" precision="FP32" names="442">
5266 <dim>-1</dim>
5267 <dim>12</dim>
5268 <dim>-1</dim>
5269 <dim>32</dim>
5270 </port>
5271 </output>
5272 </layer>
5273 <layer id="341" name="self.bert.encoder.layer.5.attention.self.key.weight" type="Const" version="opset1">
5274 <data element_type="f32" shape="384, 384" offset="83759264" size="589824" />
5275 <output>
5276 <port id="0" precision="FP32" names="self.bert.encoder.layer.5.attention.self.key.weight">
5277 <dim>384</dim>
5278 <dim>384</dim>
5279 </port>
5280 </output>
5281 </layer>
5282 <layer id="342" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
5283 <data transpose_a="false" transpose_b="true" />
5284 <input>
5285 <port id="0" precision="FP32">
5286 <dim>-1</dim>
5287 <dim>-1</dim>
5288 <dim>384</dim>
5289 </port>
5290 <port id="1" precision="FP32">
5291 <dim>384</dim>
5292 <dim>384</dim>
5293 </port>
5294 </input>
5295 <output>
5296 <port id="2" precision="FP32">
5297 <dim>-1</dim>
5298 <dim>-1</dim>
5299 <dim>384</dim>
5300 </port>
5301 </output>
5302 </layer>
5303 <layer id="343" name="Constant_90356" type="Const" version="opset1">
5304 <data element_type="f32" shape="1, 1, 384" offset="84349088" size="1536" />
5305 <output>
5306 <port id="0" precision="FP32">
5307 <dim>1</dim>
5308 <dim>1</dim>
5309 <dim>384</dim>
5310 </port>
5311 </output>
5312 </layer>
5313 <layer id="344" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add" type="Add" version="opset1">
5314 <data auto_broadcast="numpy" />
5315 <input>
5316 <port id="0" precision="FP32">
5317 <dim>-1</dim>
5318 <dim>-1</dim>
5319 <dim>384</dim>
5320 </port>
5321 <port id="1" precision="FP32">
5322 <dim>1</dim>
5323 <dim>1</dim>
5324 <dim>384</dim>
5325 </port>
5326 </input>
5327 <output>
5328 <port id="2" precision="FP32" names="445,x.65">
5329 <dim>-1</dim>
5330 <dim>-1</dim>
5331 <dim>384</dim>
5332 </port>
5333 </output>
5334 </layer>
5335 <layer id="345" name="__module.bert.encoder.layer.5.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
5336 <data element_type="i64" shape="4" offset="48269848" size="32" />
5337 <output>
5338 <port id="0" precision="I64">
5339 <dim>4</dim>
5340 </port>
5341 </output>
5342 </layer>
5343 <layer id="346" name="__module.bert.encoder.layer.5.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
5344 <data special_zero="true" />
5345 <input>
5346 <port id="0" precision="FP32">
5347 <dim>-1</dim>
5348 <dim>-1</dim>
5349 <dim>384</dim>
5350 </port>
5351 <port id="1" precision="I64">
5352 <dim>4</dim>
5353 </port>
5354 </input>
5355 <output>
5356 <port id="2" precision="FP32" names="449,x.67">
5357 <dim>-1</dim>
5358 <dim>-1</dim>
5359 <dim>12</dim>
5360 <dim>32</dim>
5361 </port>
5362 </output>
5363 </layer>
5364 <layer id="347" name="Constant_85596" type="Const" version="opset1">
5365 <data element_type="i64" shape="4" offset="48269880" size="32" />
5366 <output>
5367 <port id="0" precision="I64" names="450">
5368 <dim>4</dim>
5369 </port>
5370 </output>
5371 </layer>
5372 <layer id="348" name="__module.bert.encoder.layer.5.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
5373 <input>
5374 <port id="0" precision="FP32">
5375 <dim>-1</dim>
5376 <dim>-1</dim>
5377 <dim>12</dim>
5378 <dim>32</dim>
5379 </port>
5380 <port id="1" precision="I64">
5381 <dim>4</dim>
5382 </port>
5383 </input>
5384 <output>
5385 <port id="2" precision="FP32" names="451">
5386 <dim>-1</dim>
5387 <dim>12</dim>
5388 <dim>-1</dim>
5389 <dim>32</dim>
5390 </port>
5391 </output>
5392 </layer>
5393 <layer id="349" name="self.bert.encoder.layer.5.attention.self.value.weight" type="Const" version="opset1">
5394 <data element_type="f32" shape="384, 384" offset="84350624" size="589824" />
5395 <output>
5396 <port id="0" precision="FP32" names="self.bert.encoder.layer.5.attention.self.value.weight">
5397 <dim>384</dim>
5398 <dim>384</dim>
5399 </port>
5400 </output>
5401 </layer>
5402 <layer id="350" name="__module.bert.encoder.layer.5.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
5403 <data transpose_a="false" transpose_b="true" />
5404 <input>
5405 <port id="0" precision="FP32">
5406 <dim>-1</dim>
5407 <dim>-1</dim>
5408 <dim>384</dim>
5409 </port>
5410 <port id="1" precision="FP32">
5411 <dim>384</dim>
5412 <dim>384</dim>
5413 </port>
5414 </input>
5415 <output>
5416 <port id="2" precision="FP32">
5417 <dim>-1</dim>
5418 <dim>-1</dim>
5419 <dim>384</dim>
5420 </port>
5421 </output>
5422 </layer>
5423 <layer id="351" name="Constant_90357" type="Const" version="opset1">
5424 <data element_type="f32" shape="1, 1, 384" offset="84940448" size="1536" />
5425 <output>
5426 <port id="0" precision="FP32">
5427 <dim>1</dim>
5428 <dim>1</dim>
5429 <dim>384</dim>
5430 </port>
5431 </output>
5432 </layer>
5433 <layer id="352" name="__module.bert.encoder.layer.5.attention.self.value/aten::linear/Add" type="Add" version="opset1">
5434 <data auto_broadcast="numpy" />
5435 <input>
5436 <port id="0" precision="FP32">
5437 <dim>-1</dim>
5438 <dim>-1</dim>
5439 <dim>384</dim>
5440 </port>
5441 <port id="1" precision="FP32">
5442 <dim>1</dim>
5443 <dim>1</dim>
5444 <dim>384</dim>
5445 </port>
5446 </input>
5447 <output>
5448 <port id="2" precision="FP32" names="454,x.69">
5449 <dim>-1</dim>
5450 <dim>-1</dim>
5451 <dim>384</dim>
5452 </port>
5453 </output>
5454 </layer>
5455 <layer id="353" name="__module.bert.encoder.layer.5.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
5456 <data element_type="i64" shape="4" offset="48269848" size="32" />
5457 <output>
5458 <port id="0" precision="I64">
5459 <dim>4</dim>
5460 </port>
5461 </output>
5462 </layer>
5463 <layer id="354" name="__module.bert.encoder.layer.5.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
5464 <data special_zero="true" />
5465 <input>
5466 <port id="0" precision="FP32">
5467 <dim>-1</dim>
5468 <dim>-1</dim>
5469 <dim>384</dim>
5470 </port>
5471 <port id="1" precision="I64">
5472 <dim>4</dim>
5473 </port>
5474 </input>
5475 <output>
5476 <port id="2" precision="FP32" names="458,x">
5477 <dim>-1</dim>
5478 <dim>-1</dim>
5479 <dim>12</dim>
5480 <dim>32</dim>
5481 </port>
5482 </output>
5483 </layer>
5484 <layer id="355" name="Constant_85619" type="Const" version="opset1">
5485 <data element_type="i64" shape="4" offset="48269880" size="32" />
5486 <output>
5487 <port id="0" precision="I64" names="459">
5488 <dim>4</dim>
5489 </port>
5490 </output>
5491 </layer>
5492 <layer id="356" name="__module.bert.encoder.layer.5.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
5493 <input>
5494 <port id="0" precision="FP32">
5495 <dim>-1</dim>
5496 <dim>-1</dim>
5497 <dim>12</dim>
5498 <dim>32</dim>
5499 </port>
5500 <port id="1" precision="I64">
5501 <dim>4</dim>
5502 </port>
5503 </input>
5504 <output>
5505 <port id="2" precision="FP32" names="460">
5506 <dim>-1</dim>
5507 <dim>12</dim>
5508 <dim>-1</dim>
5509 <dim>32</dim>
5510 </port>
5511 </output>
5512 </layer>
5513 <layer id="357" name="__module.bert.encoder.layer.5.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
5514 <data causal="false" />
5515 <input>
5516 <port id="0" precision="FP32">
5517 <dim>-1</dim>
5518 <dim>12</dim>
5519 <dim>-1</dim>
5520 <dim>32</dim>
5521 </port>
5522 <port id="1" precision="FP32">
5523 <dim>-1</dim>
5524 <dim>12</dim>
5525 <dim>-1</dim>
5526 <dim>32</dim>
5527 </port>
5528 <port id="2" precision="FP32">
5529 <dim>-1</dim>
5530 <dim>12</dim>
5531 <dim>-1</dim>
5532 <dim>32</dim>
5533 </port>
5534 <port id="3" precision="FP32">
5535 <dim>-1</dim>
5536 <dim>1</dim>
5537 <dim>-1</dim>
5538 <dim>-1</dim>
5539 </port>
5540 </input>
5541 <output>
5542 <port id="4" precision="FP32" names="461,attn_output.21">
5543 <dim>-1</dim>
5544 <dim>12</dim>
5545 <dim>-1</dim>
5546 <dim>32</dim>
5547 </port>
5548 </output>
5549 </layer>
5550 <layer id="358" name="__module.bert.encoder.layer.5.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
5551 <data element_type="i32" shape="4" offset="49452664" size="16" />
5552 <output>
5553 <port id="0" precision="I32">
5554 <dim>4</dim>
5555 </port>
5556 </output>
5557 </layer>
5558 <layer id="359" name="__module.bert.encoder.layer.5.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
5559 <input>
5560 <port id="0" precision="FP32">
5561 <dim>-1</dim>
5562 <dim>12</dim>
5563 <dim>-1</dim>
5564 <dim>32</dim>
5565 </port>
5566 <port id="1" precision="I32">
5567 <dim>4</dim>
5568 </port>
5569 </input>
5570 <output>
5571 <port id="2" precision="FP32" names="462,attn_output">
5572 <dim>-1</dim>
5573 <dim>-1</dim>
5574 <dim>12</dim>
5575 <dim>32</dim>
5576 </port>
5577 </output>
5578 </layer>
5579 <layer id="360" name="Constant_90448" type="Const" version="opset1">
5580 <data element_type="i64" shape="3" offset="49452680" size="24" />
5581 <output>
5582 <port id="0" precision="I64">
5583 <dim>3</dim>
5584 </port>
5585 </output>
5586 </layer>
5587 <layer id="361" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
5588 <data special_zero="true" />
5589 <input>
5590 <port id="0" precision="FP32">
5591 <dim>-1</dim>
5592 <dim>-1</dim>
5593 <dim>12</dim>
5594 <dim>32</dim>
5595 </port>
5596 <port id="1" precision="I64">
5597 <dim>3</dim>
5598 </port>
5599 </input>
5600 <output>
5601 <port id="2" precision="FP32" names="464">
5602 <dim>-1</dim>
5603 <dim>-1</dim>
5604 <dim>384</dim>
5605 </port>
5606 </output>
5607 </layer>
5608 <layer id="362" name="self.bert.encoder.layer.5.attention.output.dense.weight" type="Const" version="opset1">
5609 <data element_type="f32" shape="384, 384" offset="84941984" size="589824" />
5610 <output>
5611 <port id="0" precision="FP32" names="self.bert.encoder.layer.5.attention.output.dense.weight">
5612 <dim>384</dim>
5613 <dim>384</dim>
5614 </port>
5615 </output>
5616 </layer>
5617 <layer id="363" name="__module.bert.encoder.layer.5.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5618 <data transpose_a="false" transpose_b="true" />
5619 <input>
5620 <port id="0" precision="FP32">
5621 <dim>-1</dim>
5622 <dim>-1</dim>
5623 <dim>384</dim>
5624 </port>
5625 <port id="1" precision="FP32">
5626 <dim>384</dim>
5627 <dim>384</dim>
5628 </port>
5629 </input>
5630 <output>
5631 <port id="2" precision="FP32">
5632 <dim>-1</dim>
5633 <dim>-1</dim>
5634 <dim>384</dim>
5635 </port>
5636 </output>
5637 </layer>
5638 <layer id="364" name="Constant_90358" type="Const" version="opset1">
5639 <data element_type="f32" shape="1, 1, 384" offset="85531808" size="1536" />
5640 <output>
5641 <port id="0" precision="FP32">
5642 <dim>1</dim>
5643 <dim>1</dim>
5644 <dim>384</dim>
5645 </port>
5646 </output>
5647 </layer>
5648 <layer id="365" name="__module.bert.encoder.layer.5.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
5649 <data auto_broadcast="numpy" />
5650 <input>
5651 <port id="0" precision="FP32">
5652 <dim>-1</dim>
5653 <dim>-1</dim>
5654 <dim>384</dim>
5655 </port>
5656 <port id="1" precision="FP32">
5657 <dim>1</dim>
5658 <dim>1</dim>
5659 <dim>384</dim>
5660 </port>
5661 </input>
5662 <output>
5663 <port id="2" precision="FP32" names="469,input.23">
5664 <dim>-1</dim>
5665 <dim>-1</dim>
5666 <dim>384</dim>
5667 </port>
5668 </output>
5669 </layer>
5670 <layer id="366" name="__module.bert.encoder.layer.5.attention.output/aten::add/Add" type="Add" version="opset1">
5671 <data auto_broadcast="numpy" />
5672 <input>
5673 <port id="0" precision="FP32">
5674 <dim>-1</dim>
5675 <dim>-1</dim>
5676 <dim>384</dim>
5677 </port>
5678 <port id="1" precision="FP32">
5679 <dim>-1</dim>
5680 <dim>-1</dim>
5681 <dim>384</dim>
5682 </port>
5683 </input>
5684 <output>
5685 <port id="2" precision="FP32" names="471">
5686 <dim>-1</dim>
5687 <dim>-1</dim>
5688 <dim>384</dim>
5689 </port>
5690 </output>
5691 </layer>
5692 <layer id="367" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5693 <data element_type="i32" shape="1" offset="47675412" size="4" />
5694 <output>
5695 <port id="0" precision="I32">
5696 <dim>1</dim>
5697 </port>
5698 </output>
5699 </layer>
5700 <layer id="368" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5701 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5702 <input>
5703 <port id="0" precision="FP32">
5704 <dim>-1</dim>
5705 <dim>-1</dim>
5706 <dim>384</dim>
5707 </port>
5708 <port id="1" precision="I32">
5709 <dim>1</dim>
5710 </port>
5711 </input>
5712 <output>
5713 <port id="2" precision="FP32">
5714 <dim>-1</dim>
5715 <dim>-1</dim>
5716 <dim>384</dim>
5717 </port>
5718 </output>
5719 </layer>
5720 <layer id="369" name="Constant_90359" type="Const" version="opset1">
5721 <data element_type="f32" shape="1, 1, 384" offset="85533344" size="1536" />
5722 <output>
5723 <port id="0" precision="FP32">
5724 <dim>1</dim>
5725 <dim>1</dim>
5726 <dim>384</dim>
5727 </port>
5728 </output>
5729 </layer>
5730 <layer id="370" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5731 <data auto_broadcast="numpy" />
5732 <input>
5733 <port id="0" precision="FP32">
5734 <dim>-1</dim>
5735 <dim>-1</dim>
5736 <dim>384</dim>
5737 </port>
5738 <port id="1" precision="FP32">
5739 <dim>1</dim>
5740 <dim>1</dim>
5741 <dim>384</dim>
5742 </port>
5743 </input>
5744 <output>
5745 <port id="2" precision="FP32">
5746 <dim>-1</dim>
5747 <dim>-1</dim>
5748 <dim>384</dim>
5749 </port>
5750 </output>
5751 </layer>
5752 <layer id="371" name="Constant_90360" type="Const" version="opset1">
5753 <data element_type="f32" shape="1, 1, 384" offset="85534880" size="1536" />
5754 <output>
5755 <port id="0" precision="FP32">
5756 <dim>1</dim>
5757 <dim>1</dim>
5758 <dim>384</dim>
5759 </port>
5760 </output>
5761 </layer>
5762 <layer id="372" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5763 <data auto_broadcast="numpy" />
5764 <input>
5765 <port id="0" precision="FP32">
5766 <dim>-1</dim>
5767 <dim>-1</dim>
5768 <dim>384</dim>
5769 </port>
5770 <port id="1" precision="FP32">
5771 <dim>1</dim>
5772 <dim>1</dim>
5773 <dim>384</dim>
5774 </port>
5775 </input>
5776 <output>
5777 <port id="2" precision="FP32" names="475,input_tensor">
5778 <dim>-1</dim>
5779 <dim>-1</dim>
5780 <dim>384</dim>
5781 </port>
5782 </output>
5783 </layer>
5784 <layer id="373" name="self.bert.encoder.layer.5.intermediate.dense.weight" type="Const" version="opset1">
5785 <data element_type="f32" shape="1536, 384" offset="85536416" size="2359296" />
5786 <output>
5787 <port id="0" precision="FP32" names="self.bert.encoder.layer.5.intermediate.dense.weight">
5788 <dim>1536</dim>
5789 <dim>384</dim>
5790 </port>
5791 </output>
5792 </layer>
5793 <layer id="374" name="__module.bert.encoder.layer.5.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5794 <data transpose_a="false" transpose_b="true" />
5795 <input>
5796 <port id="0" precision="FP32">
5797 <dim>-1</dim>
5798 <dim>-1</dim>
5799 <dim>384</dim>
5800 </port>
5801 <port id="1" precision="FP32">
5802 <dim>1536</dim>
5803 <dim>384</dim>
5804 </port>
5805 </input>
5806 <output>
5807 <port id="2" precision="FP32">
5808 <dim>-1</dim>
5809 <dim>-1</dim>
5810 <dim>1536</dim>
5811 </port>
5812 </output>
5813 </layer>
5814 <layer id="375" name="Constant_90361" type="Const" version="opset1">
5815 <data element_type="f32" shape="1, 1, 1536" offset="87895712" size="6144" />
5816 <output>
5817 <port id="0" precision="FP32">
5818 <dim>1</dim>
5819 <dim>1</dim>
5820 <dim>1536</dim>
5821 </port>
5822 </output>
5823 </layer>
5824 <layer id="376" name="__module.bert.encoder.layer.5.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5825 <data auto_broadcast="numpy" />
5826 <input>
5827 <port id="0" precision="FP32">
5828 <dim>-1</dim>
5829 <dim>-1</dim>
5830 <dim>1536</dim>
5831 </port>
5832 <port id="1" precision="FP32">
5833 <dim>1</dim>
5834 <dim>1</dim>
5835 <dim>1536</dim>
5836 </port>
5837 </input>
5838 <output>
5839 <port id="2" precision="FP32" names="479">
5840 <dim>-1</dim>
5841 <dim>-1</dim>
5842 <dim>1536</dim>
5843 </port>
5844 </output>
5845 </layer>
5846 <layer id="377" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5847 <data approximation_mode="ERF" />
5848 <input>
5849 <port id="0" precision="FP32">
5850 <dim>-1</dim>
5851 <dim>-1</dim>
5852 <dim>1536</dim>
5853 </port>
5854 </input>
5855 <output>
5856 <port id="1" precision="FP32" names="480">
5857 <dim>-1</dim>
5858 <dim>-1</dim>
5859 <dim>1536</dim>
5860 </port>
5861 </output>
5862 </layer>
5863 <layer id="378" name="self.bert.encoder.layer.5.output.dense.weight" type="Const" version="opset1">
5864 <data element_type="f32" shape="384, 1536" offset="87901856" size="2359296" />
5865 <output>
5866 <port id="0" precision="FP32" names="self.bert.encoder.layer.5.output.dense.weight">
5867 <dim>384</dim>
5868 <dim>1536</dim>
5869 </port>
5870 </output>
5871 </layer>
5872 <layer id="379" name="__module.bert.encoder.layer.5.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5873 <data transpose_a="false" transpose_b="true" />
5874 <input>
5875 <port id="0" precision="FP32">
5876 <dim>-1</dim>
5877 <dim>-1</dim>
5878 <dim>1536</dim>
5879 </port>
5880 <port id="1" precision="FP32">
5881 <dim>384</dim>
5882 <dim>1536</dim>
5883 </port>
5884 </input>
5885 <output>
5886 <port id="2" precision="FP32">
5887 <dim>-1</dim>
5888 <dim>-1</dim>
5889 <dim>384</dim>
5890 </port>
5891 </output>
5892 </layer>
5893 <layer id="380" name="Constant_90362" type="Const" version="opset1">
5894 <data element_type="f32" shape="1, 1, 384" offset="90261152" size="1536" />
5895 <output>
5896 <port id="0" precision="FP32">
5897 <dim>1</dim>
5898 <dim>1</dim>
5899 <dim>384</dim>
5900 </port>
5901 </output>
5902 </layer>
5903 <layer id="381" name="__module.bert.encoder.layer.5.output.dense/aten::linear/Add" type="Add" version="opset1">
5904 <data auto_broadcast="numpy" />
5905 <input>
5906 <port id="0" precision="FP32">
5907 <dim>-1</dim>
5908 <dim>-1</dim>
5909 <dim>384</dim>
5910 </port>
5911 <port id="1" precision="FP32">
5912 <dim>1</dim>
5913 <dim>1</dim>
5914 <dim>384</dim>
5915 </port>
5916 </input>
5917 <output>
5918 <port id="2" precision="FP32" names="485,input.25">
5919 <dim>-1</dim>
5920 <dim>-1</dim>
5921 <dim>384</dim>
5922 </port>
5923 </output>
5924 </layer>
5925 <layer id="382" name="__module.bert.encoder.layer.5.output/aten::add/Add" type="Add" version="opset1">
5926 <data auto_broadcast="numpy" />
5927 <input>
5928 <port id="0" precision="FP32">
5929 <dim>-1</dim>
5930 <dim>-1</dim>
5931 <dim>384</dim>
5932 </port>
5933 <port id="1" precision="FP32">
5934 <dim>-1</dim>
5935 <dim>-1</dim>
5936 <dim>384</dim>
5937 </port>
5938 </input>
5939 <output>
5940 <port id="2" precision="FP32" names="487">
5941 <dim>-1</dim>
5942 <dim>-1</dim>
5943 <dim>384</dim>
5944 </port>
5945 </output>
5946 </layer>
5947 <layer id="383" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5948 <data element_type="i32" shape="1" offset="47675412" size="4" />
5949 <output>
5950 <port id="0" precision="I32">
5951 <dim>1</dim>
5952 </port>
5953 </output>
5954 </layer>
5955 <layer id="384" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5956 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5957 <input>
5958 <port id="0" precision="FP32">
5959 <dim>-1</dim>
5960 <dim>-1</dim>
5961 <dim>384</dim>
5962 </port>
5963 <port id="1" precision="I32">
5964 <dim>1</dim>
5965 </port>
5966 </input>
5967 <output>
5968 <port id="2" precision="FP32">
5969 <dim>-1</dim>
5970 <dim>-1</dim>
5971 <dim>384</dim>
5972 </port>
5973 </output>
5974 </layer>
5975 <layer id="385" name="Constant_90363" type="Const" version="opset1">
5976 <data element_type="f32" shape="1, 1, 384" offset="90262688" size="1536" />
5977 <output>
5978 <port id="0" precision="FP32">
5979 <dim>1</dim>
5980 <dim>1</dim>
5981 <dim>384</dim>
5982 </port>
5983 </output>
5984 </layer>
5985 <layer id="386" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5986 <data auto_broadcast="numpy" />
5987 <input>
5988 <port id="0" precision="FP32">
5989 <dim>-1</dim>
5990 <dim>-1</dim>
5991 <dim>384</dim>
5992 </port>
5993 <port id="1" precision="FP32">
5994 <dim>1</dim>
5995 <dim>1</dim>
5996 <dim>384</dim>
5997 </port>
5998 </input>
5999 <output>
6000 <port id="2" precision="FP32">
6001 <dim>-1</dim>
6002 <dim>-1</dim>
6003 <dim>384</dim>
6004 </port>
6005 </output>
6006 </layer>
6007 <layer id="387" name="Constant_90364" type="Const" version="opset1">
6008 <data element_type="f32" shape="1, 1, 384" offset="90264224" size="1536" />
6009 <output>
6010 <port id="0" precision="FP32">
6011 <dim>1</dim>
6012 <dim>1</dim>
6013 <dim>384</dim>
6014 </port>
6015 </output>
6016 </layer>
6017 <layer id="388" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6018 <data auto_broadcast="numpy" />
6019 <input>
6020 <port id="0" precision="FP32">
6021 <dim>-1</dim>
6022 <dim>-1</dim>
6023 <dim>384</dim>
6024 </port>
6025 <port id="1" precision="FP32">
6026 <dim>1</dim>
6027 <dim>1</dim>
6028 <dim>384</dim>
6029 </port>
6030 </input>
6031 <output>
6032 <port id="2" precision="FP32" names="491,493,hidden_states">
6033 <dim>-1</dim>
6034 <dim>-1</dim>
6035 <dim>384</dim>
6036 </port>
6037 </output>
6038 </layer>
6039 <layer id="389" name="23" type="Const" version="opset1">
6040 <data element_type="i64" shape="" offset="47675396" size="8" />
6041 <output>
6042 <port id="0" precision="I64" names="23" />
6043 </output>
6044 </layer>
6045 <layer id="390" name="__module.bert.pooler/aten::select/Gather" type="Gather" version="opset8">
6046 <data batch_dims="0" />
6047 <input>
6048 <port id="0" precision="FP32">
6049 <dim>-1</dim>
6050 <dim>-1</dim>
6051 <dim>384</dim>
6052 </port>
6053 <port id="1" precision="I64" />
6054 <port id="2" precision="I64" />
6055 </input>
6056 <output>
6057 <port id="3" precision="FP32" names="494">
6058 <dim>-1</dim>
6059 <dim>384</dim>
6060 </port>
6061 </output>
6062 </layer>
6063 <layer id="391" name="self.bert.pooler.dense.weight" type="Const" version="opset1">
6064 <data element_type="f32" shape="384, 384" offset="90265760" size="589824" />
6065 <output>
6066 <port id="0" precision="FP32" names="self.bert.pooler.dense.weight">
6067 <dim>384</dim>
6068 <dim>384</dim>
6069 </port>
6070 </output>
6071 </layer>
6072 <layer id="392" name="__module.bert.pooler.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6073 <data transpose_a="false" transpose_b="true" />
6074 <input>
6075 <port id="0" precision="FP32">
6076 <dim>-1</dim>
6077 <dim>384</dim>
6078 </port>
6079 <port id="1" precision="FP32">
6080 <dim>384</dim>
6081 <dim>384</dim>
6082 </port>
6083 </input>
6084 <output>
6085 <port id="2" precision="FP32">
6086 <dim>-1</dim>
6087 <dim>384</dim>
6088 </port>
6089 </output>
6090 </layer>
6091 <layer id="393" name="Constant_90365" type="Const" version="opset1">
6092 <data element_type="f32" shape="1, 384" offset="90855584" size="1536" />
6093 <output>
6094 <port id="0" precision="FP32">
6095 <dim>1</dim>
6096 <dim>384</dim>
6097 </port>
6098 </output>
6099 </layer>
6100 <layer id="394" name="__module.bert.pooler.dense/aten::linear/Add" type="Add" version="opset1">
6101 <data auto_broadcast="numpy" />
6102 <input>
6103 <port id="0" precision="FP32">
6104 <dim>-1</dim>
6105 <dim>384</dim>
6106 </port>
6107 <port id="1" precision="FP32">
6108 <dim>1</dim>
6109 <dim>384</dim>
6110 </port>
6111 </input>
6112 <output>
6113 <port id="2" precision="FP32" names="497">
6114 <dim>-1</dim>
6115 <dim>384</dim>
6116 </port>
6117 </output>
6118 </layer>
6119 <layer id="395" name="__module.bert.pooler.activation/aten::tanh/Tanh" type="Tanh" version="opset1">
6120 <input>
6121 <port id="0" precision="FP32">
6122 <dim>-1</dim>
6123 <dim>384</dim>
6124 </port>
6125 </input>
6126 <output>
6127 <port id="1" precision="FP32" names="498,input">
6128 <dim>-1</dim>
6129 <dim>384</dim>
6130 </port>
6131 </output>
6132 </layer>
6133 <layer id="396" name="self.classifier.weight" type="Const" version="opset1">
6134 <data element_type="f32" shape="1, 384" offset="90857120" size="1536" />
6135 <output>
6136 <port id="0" precision="FP32" names="self.classifier.weight">
6137 <dim>1</dim>
6138 <dim>384</dim>
6139 </port>
6140 </output>
6141 </layer>
6142 <layer id="397" name="__module.classifier/aten::linear/Add" type="MatMul" version="opset1">
6143 <data transpose_a="false" transpose_b="true" />
6144 <input>
6145 <port id="0" precision="FP32">
6146 <dim>-1</dim>
6147 <dim>384</dim>
6148 </port>
6149 <port id="1" precision="FP32">
6150 <dim>1</dim>
6151 <dim>384</dim>
6152 </port>
6153 </input>
6154 <output>
6155 <port id="2" precision="FP32" names="logits">
6156 <dim>-1</dim>
6157 <dim>1</dim>
6158 </port>
6159 </output>
6160 </layer>
6161 <layer id="398" name="Result_86857" type="Result" version="opset1">
6162 <input>
6163 <port id="0" precision="FP32">
6164 <dim>-1</dim>
6165 <dim>1</dim>
6166 </port>
6167 </input>
6168 </layer>
6169 </layers>
6170 <edges>
6171 <edge from-layer="0" from-port="0" to-layer="8" to-port="0" />
6172 <edge from-layer="1" from-port="0" to-layer="58" to-port="0" />
6173 <edge from-layer="2" from-port="0" to-layer="4" to-port="0" />
6174 <edge from-layer="2" from-port="0" to-layer="15" to-port="0" />
6175 <edge from-layer="3" from-port="0" to-layer="6" to-port="0" />
6176 <edge from-layer="4" from-port="1" to-layer="6" to-port="1" />
6177 <edge from-layer="5" from-port="0" to-layer="6" to-port="2" />
6178 <edge from-layer="6" from-port="3" to-layer="11" to-port="0" />
6179 <edge from-layer="7" from-port="0" to-layer="10" to-port="0" />
6180 <edge from-layer="8" from-port="1" to-layer="10" to-port="1" />
6181 <edge from-layer="9" from-port="0" to-layer="10" to-port="2" />
6182 <edge from-layer="10" from-port="3" to-layer="11" to-port="1" />
6183 <edge from-layer="11" from-port="2" to-layer="25" to-port="0" />
6184 <edge from-layer="12" from-port="0" to-layer="24" to-port="0" />
6185 <edge from-layer="13" from-port="0" to-layer="21" to-port="0" />
6186 <edge from-layer="14" from-port="0" to-layer="21" to-port="1" />
6187 <edge from-layer="15" from-port="1" to-layer="18" to-port="0" />
6188 <edge from-layer="15" from-port="1" to-layer="63" to-port="0" />
6189 <edge from-layer="15" from-port="1" to-layer="67" to-port="0" />
6190 <edge from-layer="16" from-port="0" to-layer="18" to-port="1" />
6191 <edge from-layer="17" from-port="0" to-layer="18" to-port="2" />
6192 <edge from-layer="18" from-port="3" to-layer="21" to-port="2" />
6193 <edge from-layer="19" from-port="0" to-layer="21" to-port="3" />
6194 <edge from-layer="20" from-port="0" to-layer="21" to-port="4" />
6195 <edge from-layer="21" from-port="5" to-layer="22" to-port="0" />
6196 <edge from-layer="22" from-port="1" to-layer="24" to-port="1" />
6197 <edge from-layer="23" from-port="0" to-layer="24" to-port="2" />
6198 <edge from-layer="24" from-port="3" to-layer="25" to-port="1" />
6199 <edge from-layer="25" from-port="2" to-layer="27" to-port="0" />
6200 <edge from-layer="26" from-port="0" to-layer="27" to-port="1" />
6201 <edge from-layer="27" from-port="2" to-layer="29" to-port="0" />
6202 <edge from-layer="28" from-port="0" to-layer="29" to-port="1" />
6203 <edge from-layer="29" from-port="2" to-layer="31" to-port="0" />
6204 <edge from-layer="30" from-port="0" to-layer="31" to-port="1" />
6205 <edge from-layer="31" from-port="2" to-layer="33" to-port="0" />
6206 <edge from-layer="31" from-port="2" to-layer="41" to-port="0" />
6207 <edge from-layer="31" from-port="2" to-layer="49" to-port="0" />
6208 <edge from-layer="31" from-port="2" to-layer="86" to-port="1" />
6209 <edge from-layer="32" from-port="0" to-layer="33" to-port="1" />
6210 <edge from-layer="33" from-port="2" to-layer="35" to-port="0" />
6211 <edge from-layer="34" from-port="0" to-layer="35" to-port="1" />
6212 <edge from-layer="35" from-port="2" to-layer="37" to-port="0" />
6213 <edge from-layer="36" from-port="0" to-layer="37" to-port="1" />
6214 <edge from-layer="37" from-port="2" to-layer="39" to-port="0" />
6215 <edge from-layer="38" from-port="0" to-layer="39" to-port="1" />
6216 <edge from-layer="39" from-port="2" to-layer="77" to-port="0" />
6217 <edge from-layer="40" from-port="0" to-layer="41" to-port="1" />
6218 <edge from-layer="41" from-port="2" to-layer="43" to-port="0" />
6219 <edge from-layer="42" from-port="0" to-layer="43" to-port="1" />
6220 <edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
6221 <edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
6222 <edge from-layer="45" from-port="2" to-layer="47" to-port="0" />
6223 <edge from-layer="46" from-port="0" to-layer="47" to-port="1" />
6224 <edge from-layer="47" from-port="2" to-layer="77" to-port="1" />
6225 <edge from-layer="48" from-port="0" to-layer="49" to-port="1" />
6226 <edge from-layer="49" from-port="2" to-layer="51" to-port="0" />
6227 <edge from-layer="50" from-port="0" to-layer="51" to-port="1" />
6228 <edge from-layer="51" from-port="2" to-layer="53" to-port="0" />
6229 <edge from-layer="52" from-port="0" to-layer="53" to-port="1" />
6230 <edge from-layer="53" from-port="2" to-layer="55" to-port="0" />
6231 <edge from-layer="54" from-port="0" to-layer="55" to-port="1" />
6232 <edge from-layer="55" from-port="2" to-layer="77" to-port="2" />
6233 <edge from-layer="56" from-port="0" to-layer="73" to-port="0" />
6234 <edge from-layer="57" from-port="0" to-layer="58" to-port="1" />
6235 <edge from-layer="57" from-port="0" to-layer="390" to-port="2" />
6236 <edge from-layer="58" from-port="2" to-layer="60" to-port="0" />
6237 <edge from-layer="59" from-port="0" to-layer="60" to-port="1" />
6238 <edge from-layer="60" from-port="2" to-layer="69" to-port="0" />
6239 <edge from-layer="61" from-port="0" to-layer="63" to-port="1" />
6240 <edge from-layer="62" from-port="0" to-layer="63" to-port="2" />
6241 <edge from-layer="63" from-port="3" to-layer="68" to-port="0" />
6242 <edge from-layer="64" from-port="0" to-layer="68" to-port="1" />
6243 <edge from-layer="65" from-port="0" to-layer="67" to-port="1" />
6244 <edge from-layer="66" from-port="0" to-layer="67" to-port="2" />
6245 <edge from-layer="67" from-port="3" to-layer="68" to-port="2" />
6246 <edge from-layer="68" from-port="3" to-layer="69" to-port="1" />
6247 <edge from-layer="69" from-port="2" to-layer="70" to-port="0" />
6248 <edge from-layer="70" from-port="1" to-layer="72" to-port="0" />
6249 <edge from-layer="71" from-port="0" to-layer="72" to-port="1" />
6250 <edge from-layer="72" from-port="2" to-layer="73" to-port="1" />
6251 <edge from-layer="73" from-port="2" to-layer="74" to-port="0" />
6252 <edge from-layer="73" from-port="2" to-layer="76" to-port="2" />
6253 <edge from-layer="74" from-port="1" to-layer="76" to-port="0" />
6254 <edge from-layer="75" from-port="0" to-layer="76" to-port="1" />
6255 <edge from-layer="76" from-port="3" to-layer="77" to-port="3" />
6256 <edge from-layer="76" from-port="3" to-layer="357" to-port="3" />
6257 <edge from-layer="76" from-port="3" to-layer="133" to-port="3" />
6258 <edge from-layer="76" from-port="3" to-layer="189" to-port="3" />
6259 <edge from-layer="76" from-port="3" to-layer="245" to-port="3" />
6260 <edge from-layer="76" from-port="3" to-layer="301" to-port="3" />
6261 <edge from-layer="77" from-port="4" to-layer="79" to-port="0" />
6262 <edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
6263 <edge from-layer="79" from-port="2" to-layer="81" to-port="0" />
6264 <edge from-layer="80" from-port="0" to-layer="81" to-port="1" />
6265 <edge from-layer="81" from-port="2" to-layer="83" to-port="0" />
6266 <edge from-layer="82" from-port="0" to-layer="83" to-port="1" />
6267 <edge from-layer="83" from-port="2" to-layer="85" to-port="0" />
6268 <edge from-layer="84" from-port="0" to-layer="85" to-port="1" />
6269 <edge from-layer="85" from-port="2" to-layer="86" to-port="0" />
6270 <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
6271 <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
6272 <edge from-layer="88" from-port="2" to-layer="90" to-port="0" />
6273 <edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
6274 <edge from-layer="90" from-port="2" to-layer="92" to-port="0" />
6275 <edge from-layer="91" from-port="0" to-layer="92" to-port="1" />
6276 <edge from-layer="92" from-port="2" to-layer="94" to-port="0" />
6277 <edge from-layer="92" from-port="2" to-layer="102" to-port="1" />
6278 <edge from-layer="93" from-port="0" to-layer="94" to-port="1" />
6279 <edge from-layer="94" from-port="2" to-layer="96" to-port="0" />
6280 <edge from-layer="95" from-port="0" to-layer="96" to-port="1" />
6281 <edge from-layer="96" from-port="2" to-layer="97" to-port="0" />
6282 <edge from-layer="97" from-port="1" to-layer="99" to-port="0" />
6283 <edge from-layer="98" from-port="0" to-layer="99" to-port="1" />
6284 <edge from-layer="99" from-port="2" to-layer="101" to-port="0" />
6285 <edge from-layer="100" from-port="0" to-layer="101" to-port="1" />
6286 <edge from-layer="101" from-port="2" to-layer="102" to-port="0" />
6287 <edge from-layer="102" from-port="2" to-layer="104" to-port="0" />
6288 <edge from-layer="103" from-port="0" to-layer="104" to-port="1" />
6289 <edge from-layer="104" from-port="2" to-layer="106" to-port="0" />
6290 <edge from-layer="105" from-port="0" to-layer="106" to-port="1" />
6291 <edge from-layer="106" from-port="2" to-layer="108" to-port="0" />
6292 <edge from-layer="107" from-port="0" to-layer="108" to-port="1" />
6293 <edge from-layer="108" from-port="2" to-layer="118" to-port="0" />
6294 <edge from-layer="108" from-port="2" to-layer="142" to-port="1" />
6295 <edge from-layer="108" from-port="2" to-layer="110" to-port="0" />
6296 <edge from-layer="108" from-port="2" to-layer="126" to-port="0" />
6297 <edge from-layer="109" from-port="0" to-layer="110" to-port="1" />
6298 <edge from-layer="110" from-port="2" to-layer="112" to-port="0" />
6299 <edge from-layer="111" from-port="0" to-layer="112" to-port="1" />
6300 <edge from-layer="112" from-port="2" to-layer="114" to-port="0" />
6301 <edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
6302 <edge from-layer="114" from-port="2" to-layer="116" to-port="0" />
6303 <edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
6304 <edge from-layer="116" from-port="2" to-layer="133" to-port="0" />
6305 <edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
6306 <edge from-layer="118" from-port="2" to-layer="120" to-port="0" />
6307 <edge from-layer="119" from-port="0" to-layer="120" to-port="1" />
6308 <edge from-layer="120" from-port="2" to-layer="122" to-port="0" />
6309 <edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
6310 <edge from-layer="122" from-port="2" to-layer="124" to-port="0" />
6311 <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
6312 <edge from-layer="124" from-port="2" to-layer="133" to-port="1" />
6313 <edge from-layer="125" from-port="0" to-layer="126" to-port="1" />
6314 <edge from-layer="126" from-port="2" to-layer="128" to-port="0" />
6315 <edge from-layer="127" from-port="0" to-layer="128" to-port="1" />
6316 <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
6317 <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
6318 <edge from-layer="130" from-port="2" to-layer="132" to-port="0" />
6319 <edge from-layer="131" from-port="0" to-layer="132" to-port="1" />
6320 <edge from-layer="132" from-port="2" to-layer="133" to-port="2" />
6321 <edge from-layer="133" from-port="4" to-layer="135" to-port="0" />
6322 <edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
6323 <edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
6324 <edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
6325 <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
6326 <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
6327 <edge from-layer="139" from-port="2" to-layer="141" to-port="0" />
6328 <edge from-layer="140" from-port="0" to-layer="141" to-port="1" />
6329 <edge from-layer="141" from-port="2" to-layer="142" to-port="0" />
6330 <edge from-layer="142" from-port="2" to-layer="144" to-port="0" />
6331 <edge from-layer="143" from-port="0" to-layer="144" to-port="1" />
6332 <edge from-layer="144" from-port="2" to-layer="146" to-port="0" />
6333 <edge from-layer="145" from-port="0" to-layer="146" to-port="1" />
6334 <edge from-layer="146" from-port="2" to-layer="148" to-port="0" />
6335 <edge from-layer="147" from-port="0" to-layer="148" to-port="1" />
6336 <edge from-layer="148" from-port="2" to-layer="150" to-port="0" />
6337 <edge from-layer="148" from-port="2" to-layer="158" to-port="1" />
6338 <edge from-layer="149" from-port="0" to-layer="150" to-port="1" />
6339 <edge from-layer="150" from-port="2" to-layer="152" to-port="0" />
6340 <edge from-layer="151" from-port="0" to-layer="152" to-port="1" />
6341 <edge from-layer="152" from-port="2" to-layer="153" to-port="0" />
6342 <edge from-layer="153" from-port="1" to-layer="155" to-port="0" />
6343 <edge from-layer="154" from-port="0" to-layer="155" to-port="1" />
6344 <edge from-layer="155" from-port="2" to-layer="157" to-port="0" />
6345 <edge from-layer="156" from-port="0" to-layer="157" to-port="1" />
6346 <edge from-layer="157" from-port="2" to-layer="158" to-port="0" />
6347 <edge from-layer="158" from-port="2" to-layer="160" to-port="0" />
6348 <edge from-layer="159" from-port="0" to-layer="160" to-port="1" />
6349 <edge from-layer="160" from-port="2" to-layer="162" to-port="0" />
6350 <edge from-layer="161" from-port="0" to-layer="162" to-port="1" />
6351 <edge from-layer="162" from-port="2" to-layer="164" to-port="0" />
6352 <edge from-layer="163" from-port="0" to-layer="164" to-port="1" />
6353 <edge from-layer="164" from-port="2" to-layer="166" to-port="0" />
6354 <edge from-layer="164" from-port="2" to-layer="174" to-port="0" />
6355 <edge from-layer="164" from-port="2" to-layer="182" to-port="0" />
6356 <edge from-layer="164" from-port="2" to-layer="198" to-port="1" />
6357 <edge from-layer="165" from-port="0" to-layer="166" to-port="1" />
6358 <edge from-layer="166" from-port="2" to-layer="168" to-port="0" />
6359 <edge from-layer="167" from-port="0" to-layer="168" to-port="1" />
6360 <edge from-layer="168" from-port="2" to-layer="170" to-port="0" />
6361 <edge from-layer="169" from-port="0" to-layer="170" to-port="1" />
6362 <edge from-layer="170" from-port="2" to-layer="172" to-port="0" />
6363 <edge from-layer="171" from-port="0" to-layer="172" to-port="1" />
6364 <edge from-layer="172" from-port="2" to-layer="189" to-port="0" />
6365 <edge from-layer="173" from-port="0" to-layer="174" to-port="1" />
6366 <edge from-layer="174" from-port="2" to-layer="176" to-port="0" />
6367 <edge from-layer="175" from-port="0" to-layer="176" to-port="1" />
6368 <edge from-layer="176" from-port="2" to-layer="178" to-port="0" />
6369 <edge from-layer="177" from-port="0" to-layer="178" to-port="1" />
6370 <edge from-layer="178" from-port="2" to-layer="180" to-port="0" />
6371 <edge from-layer="179" from-port="0" to-layer="180" to-port="1" />
6372 <edge from-layer="180" from-port="2" to-layer="189" to-port="1" />
6373 <edge from-layer="181" from-port="0" to-layer="182" to-port="1" />
6374 <edge from-layer="182" from-port="2" to-layer="184" to-port="0" />
6375 <edge from-layer="183" from-port="0" to-layer="184" to-port="1" />
6376 <edge from-layer="184" from-port="2" to-layer="186" to-port="0" />
6377 <edge from-layer="185" from-port="0" to-layer="186" to-port="1" />
6378 <edge from-layer="186" from-port="2" to-layer="188" to-port="0" />
6379 <edge from-layer="187" from-port="0" to-layer="188" to-port="1" />
6380 <edge from-layer="188" from-port="2" to-layer="189" to-port="2" />
6381 <edge from-layer="189" from-port="4" to-layer="191" to-port="0" />
6382 <edge from-layer="190" from-port="0" to-layer="191" to-port="1" />
6383 <edge from-layer="191" from-port="2" to-layer="193" to-port="0" />
6384 <edge from-layer="192" from-port="0" to-layer="193" to-port="1" />
6385 <edge from-layer="193" from-port="2" to-layer="195" to-port="0" />
6386 <edge from-layer="194" from-port="0" to-layer="195" to-port="1" />
6387 <edge from-layer="195" from-port="2" to-layer="197" to-port="0" />
6388 <edge from-layer="196" from-port="0" to-layer="197" to-port="1" />
6389 <edge from-layer="197" from-port="2" to-layer="198" to-port="0" />
6390 <edge from-layer="198" from-port="2" to-layer="200" to-port="0" />
6391 <edge from-layer="199" from-port="0" to-layer="200" to-port="1" />
6392 <edge from-layer="200" from-port="2" to-layer="202" to-port="0" />
6393 <edge from-layer="201" from-port="0" to-layer="202" to-port="1" />
6394 <edge from-layer="202" from-port="2" to-layer="204" to-port="0" />
6395 <edge from-layer="203" from-port="0" to-layer="204" to-port="1" />
6396 <edge from-layer="204" from-port="2" to-layer="206" to-port="0" />
6397 <edge from-layer="204" from-port="2" to-layer="214" to-port="1" />
6398 <edge from-layer="205" from-port="0" to-layer="206" to-port="1" />
6399 <edge from-layer="206" from-port="2" to-layer="208" to-port="0" />
6400 <edge from-layer="207" from-port="0" to-layer="208" to-port="1" />
6401 <edge from-layer="208" from-port="2" to-layer="209" to-port="0" />
6402 <edge from-layer="209" from-port="1" to-layer="211" to-port="0" />
6403 <edge from-layer="210" from-port="0" to-layer="211" to-port="1" />
6404 <edge from-layer="211" from-port="2" to-layer="213" to-port="0" />
6405 <edge from-layer="212" from-port="0" to-layer="213" to-port="1" />
6406 <edge from-layer="213" from-port="2" to-layer="214" to-port="0" />
6407 <edge from-layer="214" from-port="2" to-layer="216" to-port="0" />
6408 <edge from-layer="215" from-port="0" to-layer="216" to-port="1" />
6409 <edge from-layer="216" from-port="2" to-layer="218" to-port="0" />
6410 <edge from-layer="217" from-port="0" to-layer="218" to-port="1" />
6411 <edge from-layer="218" from-port="2" to-layer="220" to-port="0" />
6412 <edge from-layer="219" from-port="0" to-layer="220" to-port="1" />
6413 <edge from-layer="220" from-port="2" to-layer="230" to-port="0" />
6414 <edge from-layer="220" from-port="2" to-layer="238" to-port="0" />
6415 <edge from-layer="220" from-port="2" to-layer="222" to-port="0" />
6416 <edge from-layer="220" from-port="2" to-layer="254" to-port="1" />
6417 <edge from-layer="221" from-port="0" to-layer="222" to-port="1" />
6418 <edge from-layer="222" from-port="2" to-layer="224" to-port="0" />
6419 <edge from-layer="223" from-port="0" to-layer="224" to-port="1" />
6420 <edge from-layer="224" from-port="2" to-layer="226" to-port="0" />
6421 <edge from-layer="225" from-port="0" to-layer="226" to-port="1" />
6422 <edge from-layer="226" from-port="2" to-layer="228" to-port="0" />
6423 <edge from-layer="227" from-port="0" to-layer="228" to-port="1" />
6424 <edge from-layer="228" from-port="2" to-layer="245" to-port="0" />
6425 <edge from-layer="229" from-port="0" to-layer="230" to-port="1" />
6426 <edge from-layer="230" from-port="2" to-layer="232" to-port="0" />
6427 <edge from-layer="231" from-port="0" to-layer="232" to-port="1" />
6428 <edge from-layer="232" from-port="2" to-layer="234" to-port="0" />
6429 <edge from-layer="233" from-port="0" to-layer="234" to-port="1" />
6430 <edge from-layer="234" from-port="2" to-layer="236" to-port="0" />
6431 <edge from-layer="235" from-port="0" to-layer="236" to-port="1" />
6432 <edge from-layer="236" from-port="2" to-layer="245" to-port="1" />
6433 <edge from-layer="237" from-port="0" to-layer="238" to-port="1" />
6434 <edge from-layer="238" from-port="2" to-layer="240" to-port="0" />
6435 <edge from-layer="239" from-port="0" to-layer="240" to-port="1" />
6436 <edge from-layer="240" from-port="2" to-layer="242" to-port="0" />
6437 <edge from-layer="241" from-port="0" to-layer="242" to-port="1" />
6438 <edge from-layer="242" from-port="2" to-layer="244" to-port="0" />
6439 <edge from-layer="243" from-port="0" to-layer="244" to-port="1" />
6440 <edge from-layer="244" from-port="2" to-layer="245" to-port="2" />
6441 <edge from-layer="245" from-port="4" to-layer="247" to-port="0" />
6442 <edge from-layer="246" from-port="0" to-layer="247" to-port="1" />
6443 <edge from-layer="247" from-port="2" to-layer="249" to-port="0" />
6444 <edge from-layer="248" from-port="0" to-layer="249" to-port="1" />
6445 <edge from-layer="249" from-port="2" to-layer="251" to-port="0" />
6446 <edge from-layer="250" from-port="0" to-layer="251" to-port="1" />
6447 <edge from-layer="251" from-port="2" to-layer="253" to-port="0" />
6448 <edge from-layer="252" from-port="0" to-layer="253" to-port="1" />
6449 <edge from-layer="253" from-port="2" to-layer="254" to-port="0" />
6450 <edge from-layer="254" from-port="2" to-layer="256" to-port="0" />
6451 <edge from-layer="255" from-port="0" to-layer="256" to-port="1" />
6452 <edge from-layer="256" from-port="2" to-layer="258" to-port="0" />
6453 <edge from-layer="257" from-port="0" to-layer="258" to-port="1" />
6454 <edge from-layer="258" from-port="2" to-layer="260" to-port="0" />
6455 <edge from-layer="259" from-port="0" to-layer="260" to-port="1" />
6456 <edge from-layer="260" from-port="2" to-layer="262" to-port="0" />
6457 <edge from-layer="260" from-port="2" to-layer="270" to-port="1" />
6458 <edge from-layer="261" from-port="0" to-layer="262" to-port="1" />
6459 <edge from-layer="262" from-port="2" to-layer="264" to-port="0" />
6460 <edge from-layer="263" from-port="0" to-layer="264" to-port="1" />
6461 <edge from-layer="264" from-port="2" to-layer="265" to-port="0" />
6462 <edge from-layer="265" from-port="1" to-layer="267" to-port="0" />
6463 <edge from-layer="266" from-port="0" to-layer="267" to-port="1" />
6464 <edge from-layer="267" from-port="2" to-layer="269" to-port="0" />
6465 <edge from-layer="268" from-port="0" to-layer="269" to-port="1" />
6466 <edge from-layer="269" from-port="2" to-layer="270" to-port="0" />
6467 <edge from-layer="270" from-port="2" to-layer="272" to-port="0" />
6468 <edge from-layer="271" from-port="0" to-layer="272" to-port="1" />
6469 <edge from-layer="272" from-port="2" to-layer="274" to-port="0" />
6470 <edge from-layer="273" from-port="0" to-layer="274" to-port="1" />
6471 <edge from-layer="274" from-port="2" to-layer="276" to-port="0" />
6472 <edge from-layer="275" from-port="0" to-layer="276" to-port="1" />
6473 <edge from-layer="276" from-port="2" to-layer="278" to-port="0" />
6474 <edge from-layer="276" from-port="2" to-layer="294" to-port="0" />
6475 <edge from-layer="276" from-port="2" to-layer="286" to-port="0" />
6476 <edge from-layer="276" from-port="2" to-layer="310" to-port="1" />
6477 <edge from-layer="277" from-port="0" to-layer="278" to-port="1" />
6478 <edge from-layer="278" from-port="2" to-layer="280" to-port="0" />
6479 <edge from-layer="279" from-port="0" to-layer="280" to-port="1" />
6480 <edge from-layer="280" from-port="2" to-layer="282" to-port="0" />
6481 <edge from-layer="281" from-port="0" to-layer="282" to-port="1" />
6482 <edge from-layer="282" from-port="2" to-layer="284" to-port="0" />
6483 <edge from-layer="283" from-port="0" to-layer="284" to-port="1" />
6484 <edge from-layer="284" from-port="2" to-layer="301" to-port="0" />
6485 <edge from-layer="285" from-port="0" to-layer="286" to-port="1" />
6486 <edge from-layer="286" from-port="2" to-layer="288" to-port="0" />
6487 <edge from-layer="287" from-port="0" to-layer="288" to-port="1" />
6488 <edge from-layer="288" from-port="2" to-layer="290" to-port="0" />
6489 <edge from-layer="289" from-port="0" to-layer="290" to-port="1" />
6490 <edge from-layer="290" from-port="2" to-layer="292" to-port="0" />
6491 <edge from-layer="291" from-port="0" to-layer="292" to-port="1" />
6492 <edge from-layer="292" from-port="2" to-layer="301" to-port="1" />
6493 <edge from-layer="293" from-port="0" to-layer="294" to-port="1" />
6494 <edge from-layer="294" from-port="2" to-layer="296" to-port="0" />
6495 <edge from-layer="295" from-port="0" to-layer="296" to-port="1" />
6496 <edge from-layer="296" from-port="2" to-layer="298" to-port="0" />
6497 <edge from-layer="297" from-port="0" to-layer="298" to-port="1" />
6498 <edge from-layer="298" from-port="2" to-layer="300" to-port="0" />
6499 <edge from-layer="299" from-port="0" to-layer="300" to-port="1" />
6500 <edge from-layer="300" from-port="2" to-layer="301" to-port="2" />
6501 <edge from-layer="301" from-port="4" to-layer="303" to-port="0" />
6502 <edge from-layer="302" from-port="0" to-layer="303" to-port="1" />
6503 <edge from-layer="303" from-port="2" to-layer="305" to-port="0" />
6504 <edge from-layer="304" from-port="0" to-layer="305" to-port="1" />
6505 <edge from-layer="305" from-port="2" to-layer="307" to-port="0" />
6506 <edge from-layer="306" from-port="0" to-layer="307" to-port="1" />
6507 <edge from-layer="307" from-port="2" to-layer="309" to-port="0" />
6508 <edge from-layer="308" from-port="0" to-layer="309" to-port="1" />
6509 <edge from-layer="309" from-port="2" to-layer="310" to-port="0" />
6510 <edge from-layer="310" from-port="2" to-layer="312" to-port="0" />
6511 <edge from-layer="311" from-port="0" to-layer="312" to-port="1" />
6512 <edge from-layer="312" from-port="2" to-layer="314" to-port="0" />
6513 <edge from-layer="313" from-port="0" to-layer="314" to-port="1" />
6514 <edge from-layer="314" from-port="2" to-layer="316" to-port="0" />
6515 <edge from-layer="315" from-port="0" to-layer="316" to-port="1" />
6516 <edge from-layer="316" from-port="2" to-layer="318" to-port="0" />
6517 <edge from-layer="316" from-port="2" to-layer="326" to-port="1" />
6518 <edge from-layer="317" from-port="0" to-layer="318" to-port="1" />
6519 <edge from-layer="318" from-port="2" to-layer="320" to-port="0" />
6520 <edge from-layer="319" from-port="0" to-layer="320" to-port="1" />
6521 <edge from-layer="320" from-port="2" to-layer="321" to-port="0" />
6522 <edge from-layer="321" from-port="1" to-layer="323" to-port="0" />
6523 <edge from-layer="322" from-port="0" to-layer="323" to-port="1" />
6524 <edge from-layer="323" from-port="2" to-layer="325" to-port="0" />
6525 <edge from-layer="324" from-port="0" to-layer="325" to-port="1" />
6526 <edge from-layer="325" from-port="2" to-layer="326" to-port="0" />
6527 <edge from-layer="326" from-port="2" to-layer="328" to-port="0" />
6528 <edge from-layer="327" from-port="0" to-layer="328" to-port="1" />
6529 <edge from-layer="328" from-port="2" to-layer="330" to-port="0" />
6530 <edge from-layer="329" from-port="0" to-layer="330" to-port="1" />
6531 <edge from-layer="330" from-port="2" to-layer="332" to-port="0" />
6532 <edge from-layer="331" from-port="0" to-layer="332" to-port="1" />
6533 <edge from-layer="332" from-port="2" to-layer="334" to-port="0" />
6534 <edge from-layer="332" from-port="2" to-layer="342" to-port="0" />
6535 <edge from-layer="332" from-port="2" to-layer="350" to-port="0" />
6536 <edge from-layer="332" from-port="2" to-layer="366" to-port="1" />
6537 <edge from-layer="333" from-port="0" to-layer="334" to-port="1" />
6538 <edge from-layer="334" from-port="2" to-layer="336" to-port="0" />
6539 <edge from-layer="335" from-port="0" to-layer="336" to-port="1" />
6540 <edge from-layer="336" from-port="2" to-layer="338" to-port="0" />
6541 <edge from-layer="337" from-port="0" to-layer="338" to-port="1" />
6542 <edge from-layer="338" from-port="2" to-layer="340" to-port="0" />
6543 <edge from-layer="339" from-port="0" to-layer="340" to-port="1" />
6544 <edge from-layer="340" from-port="2" to-layer="357" to-port="0" />
6545 <edge from-layer="341" from-port="0" to-layer="342" to-port="1" />
6546 <edge from-layer="342" from-port="2" to-layer="344" to-port="0" />
6547 <edge from-layer="343" from-port="0" to-layer="344" to-port="1" />
6548 <edge from-layer="344" from-port="2" to-layer="346" to-port="0" />
6549 <edge from-layer="345" from-port="0" to-layer="346" to-port="1" />
6550 <edge from-layer="346" from-port="2" to-layer="348" to-port="0" />
6551 <edge from-layer="347" from-port="0" to-layer="348" to-port="1" />
6552 <edge from-layer="348" from-port="2" to-layer="357" to-port="1" />
6553 <edge from-layer="349" from-port="0" to-layer="350" to-port="1" />
6554 <edge from-layer="350" from-port="2" to-layer="352" to-port="0" />
6555 <edge from-layer="351" from-port="0" to-layer="352" to-port="1" />
6556 <edge from-layer="352" from-port="2" to-layer="354" to-port="0" />
6557 <edge from-layer="353" from-port="0" to-layer="354" to-port="1" />
6558 <edge from-layer="354" from-port="2" to-layer="356" to-port="0" />
6559 <edge from-layer="355" from-port="0" to-layer="356" to-port="1" />
6560 <edge from-layer="356" from-port="2" to-layer="357" to-port="2" />
6561 <edge from-layer="357" from-port="4" to-layer="359" to-port="0" />
6562 <edge from-layer="358" from-port="0" to-layer="359" to-port="1" />
6563 <edge from-layer="359" from-port="2" to-layer="361" to-port="0" />
6564 <edge from-layer="360" from-port="0" to-layer="361" to-port="1" />
6565 <edge from-layer="361" from-port="2" to-layer="363" to-port="0" />
6566 <edge from-layer="362" from-port="0" to-layer="363" to-port="1" />
6567 <edge from-layer="363" from-port="2" to-layer="365" to-port="0" />
6568 <edge from-layer="364" from-port="0" to-layer="365" to-port="1" />
6569 <edge from-layer="365" from-port="2" to-layer="366" to-port="0" />
6570 <edge from-layer="366" from-port="2" to-layer="368" to-port="0" />
6571 <edge from-layer="367" from-port="0" to-layer="368" to-port="1" />
6572 <edge from-layer="368" from-port="2" to-layer="370" to-port="0" />
6573 <edge from-layer="369" from-port="0" to-layer="370" to-port="1" />
6574 <edge from-layer="370" from-port="2" to-layer="372" to-port="0" />
6575 <edge from-layer="371" from-port="0" to-layer="372" to-port="1" />
6576 <edge from-layer="372" from-port="2" to-layer="382" to-port="1" />
6577 <edge from-layer="372" from-port="2" to-layer="374" to-port="0" />
6578 <edge from-layer="373" from-port="0" to-layer="374" to-port="1" />
6579 <edge from-layer="374" from-port="2" to-layer="376" to-port="0" />
6580 <edge from-layer="375" from-port="0" to-layer="376" to-port="1" />
6581 <edge from-layer="376" from-port="2" to-layer="377" to-port="0" />
6582 <edge from-layer="377" from-port="1" to-layer="379" to-port="0" />
6583 <edge from-layer="378" from-port="0" to-layer="379" to-port="1" />
6584 <edge from-layer="379" from-port="2" to-layer="381" to-port="0" />
6585 <edge from-layer="380" from-port="0" to-layer="381" to-port="1" />
6586 <edge from-layer="381" from-port="2" to-layer="382" to-port="0" />
6587 <edge from-layer="382" from-port="2" to-layer="384" to-port="0" />
6588 <edge from-layer="383" from-port="0" to-layer="384" to-port="1" />
6589 <edge from-layer="384" from-port="2" to-layer="386" to-port="0" />
6590 <edge from-layer="385" from-port="0" to-layer="386" to-port="1" />
6591 <edge from-layer="386" from-port="2" to-layer="388" to-port="0" />
6592 <edge from-layer="387" from-port="0" to-layer="388" to-port="1" />
6593 <edge from-layer="388" from-port="2" to-layer="390" to-port="0" />
6594 <edge from-layer="389" from-port="0" to-layer="390" to-port="1" />
6595 <edge from-layer="390" from-port="3" to-layer="392" to-port="0" />
6596 <edge from-layer="391" from-port="0" to-layer="392" to-port="1" />
6597 <edge from-layer="392" from-port="2" to-layer="394" to-port="0" />
6598 <edge from-layer="393" from-port="0" to-layer="394" to-port="1" />
6599 <edge from-layer="394" from-port="2" to-layer="395" to-port="0" />
6600 <edge from-layer="395" from-port="1" to-layer="397" to-port="0" />
6601 <edge from-layer="396" from-port="0" to-layer="397" to-port="1" />
6602 <edge from-layer="397" from-port="2" to-layer="398" to-port="0" />
6603 </edges>
6604 <rt_info>
6605 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
6606 <conversion_parameters>
6607 <framework value="pytorch" />
6608 <is_python_object value="True" />
6609 </conversion_parameters>
6610 <optimum>
6611 <optimum_intel_version value="1.20.1" />
6612 <optimum_version value="1.24.0" />
6613 <pytorch_version value="2.6.0+cu124" />
6614 <transformers_version value="4.52.0.dev0" />
6615 </optimum>
6616 </rt_info>
6617 </net>
6618