openvino/openvino_model_qint8_quantized.xml
359.6 KB · 11298 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model0" version="11">
3 <layers>
4 <layer id="2" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="1" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="0" name="token_type_ids" type="Parameter" version="opset1">
23 <data shape="?,?" element_type="i64" />
24 <output>
25 <port id="0" precision="I64" names="token_type_ids">
26 <dim>-1</dim>
27 <dim>-1</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="Constant_27536122" type="Const" version="opset1">
32 <data element_type="i8" shape="30522, 384" offset="0" size="11720448" />
33 <output>
34 <port id="0" precision="I8">
35 <dim>30522</dim>
36 <dim>384</dim>
37 </port>
38 </output>
39 </layer>
40 <layer id="4" name="Convert_27536123" type="Convert" version="opset1">
41 <data destination_type="f32" />
42 <input>
43 <port id="0" precision="I8">
44 <dim>30522</dim>
45 <dim>384</dim>
46 </port>
47 </input>
48 <output>
49 <port id="1" precision="FP32">
50 <dim>30522</dim>
51 <dim>384</dim>
52 </port>
53 </output>
54 </layer>
55 <layer id="5" name="Constant_27536124" type="Const" version="opset1">
56 <data element_type="f32" shape="30522, 1" offset="11720448" size="122088" />
57 <output>
58 <port id="0" precision="FP32">
59 <dim>30522</dim>
60 <dim>1</dim>
61 </port>
62 </output>
63 </layer>
64 <layer id="6" name="__module.embeddings.word_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
65 <data auto_broadcast="numpy" />
66 <input>
67 <port id="0" precision="FP32">
68 <dim>30522</dim>
69 <dim>384</dim>
70 </port>
71 <port id="1" precision="FP32">
72 <dim>30522</dim>
73 <dim>1</dim>
74 </port>
75 </input>
76 <output>
77 <port id="2" precision="FP32">
78 <dim>30522</dim>
79 <dim>384</dim>
80 </port>
81 </output>
82 </layer>
83 <layer id="7" name="__module.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
84 <data destination_type="i32" />
85 <input>
86 <port id="0" precision="I64">
87 <dim>-1</dim>
88 <dim>-1</dim>
89 </port>
90 </input>
91 <output>
92 <port id="1" precision="I32">
93 <dim>-1</dim>
94 <dim>-1</dim>
95 </port>
96 </output>
97 </layer>
98 <layer id="8" name="__module.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
99 <data element_type="i32" shape="" offset="11842536" size="4" />
100 <output>
101 <port id="0" precision="I32" />
102 </output>
103 </layer>
104 <layer id="9" name="__module.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
105 <data batch_dims="0" />
106 <input>
107 <port id="0" precision="FP32">
108 <dim>30522</dim>
109 <dim>384</dim>
110 </port>
111 <port id="1" precision="I32">
112 <dim>-1</dim>
113 <dim>-1</dim>
114 </port>
115 <port id="2" precision="I32" />
116 </input>
117 <output>
118 <port id="3" precision="FP32" names="79,inputs_embeds">
119 <dim>-1</dim>
120 <dim>-1</dim>
121 <dim>384</dim>
122 </port>
123 </output>
124 </layer>
125 <layer id="10" name="Constant_27536126" type="Const" version="opset1">
126 <data element_type="i8" shape="2, 384" offset="11842540" size="768" />
127 <output>
128 <port id="0" precision="I8">
129 <dim>2</dim>
130 <dim>384</dim>
131 </port>
132 </output>
133 </layer>
134 <layer id="11" name="Convert_27536127" type="Convert" version="opset1">
135 <data destination_type="f32" />
136 <input>
137 <port id="0" precision="I8">
138 <dim>2</dim>
139 <dim>384</dim>
140 </port>
141 </input>
142 <output>
143 <port id="1" precision="FP32">
144 <dim>2</dim>
145 <dim>384</dim>
146 </port>
147 </output>
148 </layer>
149 <layer id="12" name="Constant_27536128" type="Const" version="opset1">
150 <data element_type="f32" shape="2, 1" offset="11843308" size="8" />
151 <output>
152 <port id="0" precision="FP32">
153 <dim>2</dim>
154 <dim>1</dim>
155 </port>
156 </output>
157 </layer>
158 <layer id="13" name="__module.embeddings.token_type_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
159 <data auto_broadcast="numpy" />
160 <input>
161 <port id="0" precision="FP32">
162 <dim>2</dim>
163 <dim>384</dim>
164 </port>
165 <port id="1" precision="FP32">
166 <dim>2</dim>
167 <dim>1</dim>
168 </port>
169 </input>
170 <output>
171 <port id="2" precision="FP32">
172 <dim>2</dim>
173 <dim>384</dim>
174 </port>
175 </output>
176 </layer>
177 <layer id="14" name="__module.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
178 <data destination_type="i32" />
179 <input>
180 <port id="0" precision="I64">
181 <dim>-1</dim>
182 <dim>-1</dim>
183 </port>
184 </input>
185 <output>
186 <port id="1" precision="I32">
187 <dim>-1</dim>
188 <dim>-1</dim>
189 </port>
190 </output>
191 </layer>
192 <layer id="15" name="__module.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
193 <data element_type="i32" shape="" offset="11842536" size="4" />
194 <output>
195 <port id="0" precision="I32" />
196 </output>
197 </layer>
198 <layer id="16" name="__module.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
199 <data batch_dims="0" />
200 <input>
201 <port id="0" precision="FP32">
202 <dim>2</dim>
203 <dim>384</dim>
204 </port>
205 <port id="1" precision="I32">
206 <dim>-1</dim>
207 <dim>-1</dim>
208 </port>
209 <port id="2" precision="I32" />
210 </input>
211 <output>
212 <port id="3" precision="FP32" names="81,token_type_embeddings.1">
213 <dim>-1</dim>
214 <dim>-1</dim>
215 <dim>384</dim>
216 </port>
217 </output>
218 </layer>
219 <layer id="17" name="__module.embeddings/aten::add/Add" type="Add" version="opset1">
220 <data auto_broadcast="numpy" />
221 <input>
222 <port id="0" precision="FP32">
223 <dim>-1</dim>
224 <dim>-1</dim>
225 <dim>384</dim>
226 </port>
227 <port id="1" precision="FP32">
228 <dim>-1</dim>
229 <dim>-1</dim>
230 <dim>384</dim>
231 </port>
232 </input>
233 <output>
234 <port id="2" precision="FP32" names="82_1">
235 <dim>-1</dim>
236 <dim>-1</dim>
237 <dim>384</dim>
238 </port>
239 </output>
240 </layer>
241 <layer id="18" name="Constant_27536130" type="Const" version="opset1">
242 <data element_type="i8" shape="512, 384" offset="11843316" size="196608" />
243 <output>
244 <port id="0" precision="I8">
245 <dim>512</dim>
246 <dim>384</dim>
247 </port>
248 </output>
249 </layer>
250 <layer id="19" name="Convert_27536131" type="Convert" version="opset1">
251 <data destination_type="f32" />
252 <input>
253 <port id="0" precision="I8">
254 <dim>512</dim>
255 <dim>384</dim>
256 </port>
257 </input>
258 <output>
259 <port id="1" precision="FP32">
260 <dim>512</dim>
261 <dim>384</dim>
262 </port>
263 </output>
264 </layer>
265 <layer id="20" name="Constant_27536132" type="Const" version="opset1">
266 <data element_type="f32" shape="512, 1" offset="12039924" size="2048" />
267 <output>
268 <port id="0" precision="FP32">
269 <dim>512</dim>
270 <dim>1</dim>
271 </port>
272 </output>
273 </layer>
274 <layer id="21" name="__module.embeddings.position_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
275 <data auto_broadcast="numpy" />
276 <input>
277 <port id="0" precision="FP32">
278 <dim>512</dim>
279 <dim>384</dim>
280 </port>
281 <port id="1" precision="FP32">
282 <dim>512</dim>
283 <dim>1</dim>
284 </port>
285 </input>
286 <output>
287 <port id="2" precision="FP32">
288 <dim>512</dim>
289 <dim>384</dim>
290 </port>
291 </output>
292 </layer>
293 <layer id="22" name="__module.embeddings/aten::slice/Slice" type="Const" version="opset1">
294 <data element_type="i64" shape="1, 512" offset="12041972" size="4096" />
295 <output>
296 <port id="0" precision="I64" names="76">
297 <dim>1</dim>
298 <dim>512</dim>
299 </port>
300 </output>
301 </layer>
302 <layer id="23" name="__module.embeddings/aten::slice/Reshape" type="Const" version="opset1">
303 <data element_type="i64" shape="1" offset="12046068" size="8" />
304 <output>
305 <port id="0" precision="I64">
306 <dim>1</dim>
307 </port>
308 </output>
309 </layer>
310 <layer id="24" name="ShapeOf_6355" type="ShapeOf" version="opset3">
311 <data output_type="i64" />
312 <input>
313 <port id="0" precision="I64">
314 <dim>-1</dim>
315 <dim>-1</dim>
316 </port>
317 </input>
318 <output>
319 <port id="1" precision="I64">
320 <dim>2</dim>
321 </port>
322 </output>
323 </layer>
324 <layer id="25" name="Constant_6476" type="Const" version="opset1">
325 <data element_type="i64" shape="1" offset="12046076" size="8" />
326 <output>
327 <port id="0" precision="I64">
328 <dim>1</dim>
329 </port>
330 </output>
331 </layer>
332 <layer id="26" name="Constant_6357" type="Const" version="opset1">
333 <data element_type="i64" shape="" offset="12046068" size="8" />
334 <output>
335 <port id="0" precision="I64" />
336 </output>
337 </layer>
338 <layer id="27" name="Gather_6358" type="Gather" version="opset8">
339 <data batch_dims="0" />
340 <input>
341 <port id="0" precision="I64">
342 <dim>2</dim>
343 </port>
344 <port id="1" precision="I64">
345 <dim>1</dim>
346 </port>
347 <port id="2" precision="I64" />
348 </input>
349 <output>
350 <port id="3" precision="I64" names="10,72,74,75,8">
351 <dim>1</dim>
352 </port>
353 </output>
354 </layer>
355 <layer id="28" name="__module.embeddings/aten::slice/Reshape_2" type="Const" version="opset1">
356 <data element_type="i64" shape="1" offset="12046076" size="8" />
357 <output>
358 <port id="0" precision="I64">
359 <dim>1</dim>
360 </port>
361 </output>
362 </layer>
363 <layer id="29" name="__module.embeddings/aten::slice/Reshape_3" type="Const" version="opset1">
364 <data element_type="i64" shape="1" offset="12046076" size="8" />
365 <output>
366 <port id="0" precision="I64">
367 <dim>1</dim>
368 </port>
369 </output>
370 </layer>
371 <layer id="30" name="__module.embeddings/aten::slice/Slice_1" type="Slice" version="opset8">
372 <input>
373 <port id="0" precision="I64">
374 <dim>1</dim>
375 <dim>512</dim>
376 </port>
377 <port id="1" precision="I64">
378 <dim>1</dim>
379 </port>
380 <port id="2" precision="I64">
381 <dim>1</dim>
382 </port>
383 <port id="3" precision="I64">
384 <dim>1</dim>
385 </port>
386 <port id="4" precision="I64">
387 <dim>1</dim>
388 </port>
389 </input>
390 <output>
391 <port id="5" precision="I64" names="77">
392 <dim>1</dim>
393 <dim>-1</dim>
394 </port>
395 </output>
396 </layer>
397 <layer id="31" name="__module.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
398 <data destination_type="i32" />
399 <input>
400 <port id="0" precision="I64">
401 <dim>1</dim>
402 <dim>-1</dim>
403 </port>
404 </input>
405 <output>
406 <port id="1" precision="I32">
407 <dim>1</dim>
408 <dim>-1</dim>
409 </port>
410 </output>
411 </layer>
412 <layer id="32" name="__module.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
413 <data element_type="i32" shape="" offset="11842536" size="4" />
414 <output>
415 <port id="0" precision="I32" />
416 </output>
417 </layer>
418 <layer id="33" name="__module.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
419 <data batch_dims="0" />
420 <input>
421 <port id="0" precision="FP32">
422 <dim>512</dim>
423 <dim>384</dim>
424 </port>
425 <port id="1" precision="I32">
426 <dim>1</dim>
427 <dim>-1</dim>
428 </port>
429 <port id="2" precision="I32" />
430 </input>
431 <output>
432 <port id="3" precision="FP32" names="84,position_embeddings.1">
433 <dim>1</dim>
434 <dim>-1</dim>
435 <dim>384</dim>
436 </port>
437 </output>
438 </layer>
439 <layer id="34" name="__module.embeddings/aten::add_/Add" type="Add" version="opset1">
440 <data auto_broadcast="numpy" />
441 <input>
442 <port id="0" precision="FP32">
443 <dim>-1</dim>
444 <dim>-1</dim>
445 <dim>384</dim>
446 </port>
447 <port id="1" precision="FP32">
448 <dim>1</dim>
449 <dim>-1</dim>
450 <dim>384</dim>
451 </port>
452 </input>
453 <output>
454 <port id="2" precision="FP32" names="82,embeddings.1">
455 <dim>-1</dim>
456 <dim>-1</dim>
457 <dim>384</dim>
458 </port>
459 </output>
460 </layer>
461 <layer id="35" name="__module.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
462 <data element_type="i32" shape="1" offset="12046084" size="4" />
463 <output>
464 <port id="0" precision="I32">
465 <dim>1</dim>
466 </port>
467 </output>
468 </layer>
469 <layer id="36" name="__module.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
470 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
471 <input>
472 <port id="0" precision="FP32">
473 <dim>-1</dim>
474 <dim>-1</dim>
475 <dim>384</dim>
476 </port>
477 <port id="1" precision="I32">
478 <dim>1</dim>
479 </port>
480 </input>
481 <output>
482 <port id="2" precision="FP32">
483 <dim>-1</dim>
484 <dim>-1</dim>
485 <dim>384</dim>
486 </port>
487 </output>
488 </layer>
489 <layer id="37" name="Constant_6230" type="Const" version="opset1">
490 <data element_type="f32" shape="1, 1, 384" offset="12046088" size="1536" />
491 <output>
492 <port id="0" precision="FP32">
493 <dim>1</dim>
494 <dim>1</dim>
495 <dim>384</dim>
496 </port>
497 </output>
498 </layer>
499 <layer id="38" name="__module.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
500 <data auto_broadcast="numpy" />
501 <input>
502 <port id="0" precision="FP32">
503 <dim>-1</dim>
504 <dim>-1</dim>
505 <dim>384</dim>
506 </port>
507 <port id="1" precision="FP32">
508 <dim>1</dim>
509 <dim>1</dim>
510 <dim>384</dim>
511 </port>
512 </input>
513 <output>
514 <port id="2" precision="FP32">
515 <dim>-1</dim>
516 <dim>-1</dim>
517 <dim>384</dim>
518 </port>
519 </output>
520 </layer>
521 <layer id="39" name="Constant_6231" type="Const" version="opset1">
522 <data element_type="f32" shape="1, 1, 384" offset="12047624" size="1536" />
523 <output>
524 <port id="0" precision="FP32">
525 <dim>1</dim>
526 <dim>1</dim>
527 <dim>384</dim>
528 </port>
529 </output>
530 </layer>
531 <layer id="40" name="__module.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
532 <data auto_broadcast="numpy" />
533 <input>
534 <port id="0" precision="FP32">
535 <dim>-1</dim>
536 <dim>-1</dim>
537 <dim>384</dim>
538 </port>
539 <port id="1" precision="FP32">
540 <dim>1</dim>
541 <dim>1</dim>
542 <dim>384</dim>
543 </port>
544 </input>
545 <output>
546 <port id="2" precision="FP32" names="89,input.1">
547 <dim>-1</dim>
548 <dim>-1</dim>
549 <dim>384</dim>
550 </port>
551 </output>
552 </layer>
553 <layer id="41" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
554 <data element_type="f32" shape="1, 1, 384" offset="12049160" size="1536" />
555 <output>
556 <port id="0" precision="FP32">
557 <dim>1</dim>
558 <dim>1</dim>
559 <dim>384</dim>
560 </port>
561 </output>
562 </layer>
563 <layer id="42" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
564 <data auto_broadcast="numpy" />
565 <input>
566 <port id="0" precision="FP32">
567 <dim>-1</dim>
568 <dim>-1</dim>
569 <dim>384</dim>
570 </port>
571 <port id="1" precision="FP32">
572 <dim>1</dim>
573 <dim>1</dim>
574 <dim>384</dim>
575 </port>
576 </input>
577 <output>
578 <port id="2" precision="FP32">
579 <dim>-1</dim>
580 <dim>-1</dim>
581 <dim>384</dim>
582 </port>
583 </output>
584 </layer>
585 <layer id="43" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
586 <data element_type="f32" shape="" offset="12050696" size="4" />
587 <output>
588 <port id="0" precision="FP32" />
589 </output>
590 </layer>
591 <layer id="44" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
592 <data element_type="f32" shape="" offset="12050700" size="4" />
593 <output>
594 <port id="0" precision="FP32" />
595 </output>
596 </layer>
597 <layer id="45" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
598 <data element_type="f32" shape="" offset="12050696" size="4" />
599 <output>
600 <port id="0" precision="FP32" />
601 </output>
602 </layer>
603 <layer id="46" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
604 <data element_type="f32" shape="" offset="12050700" size="4" />
605 <output>
606 <port id="0" precision="FP32" />
607 </output>
608 </layer>
609 <layer id="47" name="__module.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
610 <data levels="256" auto_broadcast="numpy" />
611 <input>
612 <port id="0" precision="FP32">
613 <dim>-1</dim>
614 <dim>-1</dim>
615 <dim>384</dim>
616 </port>
617 <port id="1" precision="FP32" />
618 <port id="2" precision="FP32" />
619 <port id="3" precision="FP32" />
620 <port id="4" precision="FP32" />
621 </input>
622 <output>
623 <port id="5" precision="FP32">
624 <dim>-1</dim>
625 <dim>-1</dim>
626 <dim>384</dim>
627 </port>
628 </output>
629 </layer>
630 <layer id="48" name="Constant_27536134" type="Const" version="opset1">
631 <data element_type="i8" shape="384, 384" offset="12050704" size="147456" />
632 <output>
633 <port id="0" precision="I8">
634 <dim>384</dim>
635 <dim>384</dim>
636 </port>
637 </output>
638 </layer>
639 <layer id="49" name="Convert_27536135" type="Convert" version="opset1">
640 <data destination_type="f32" />
641 <input>
642 <port id="0" precision="I8">
643 <dim>384</dim>
644 <dim>384</dim>
645 </port>
646 </input>
647 <output>
648 <port id="1" precision="FP32">
649 <dim>384</dim>
650 <dim>384</dim>
651 </port>
652 </output>
653 </layer>
654 <layer id="50" name="Constant_27536136" type="Const" version="opset1">
655 <data element_type="f32" shape="384, 1" offset="12198160" size="1536" />
656 <output>
657 <port id="0" precision="FP32">
658 <dim>384</dim>
659 <dim>1</dim>
660 </port>
661 </output>
662 </layer>
663 <layer id="51" name="__module.encoder.layer.0.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
664 <data auto_broadcast="numpy" />
665 <input>
666 <port id="0" precision="FP32">
667 <dim>384</dim>
668 <dim>384</dim>
669 </port>
670 <port id="1" precision="FP32">
671 <dim>384</dim>
672 <dim>1</dim>
673 </port>
674 </input>
675 <output>
676 <port id="2" precision="FP32">
677 <dim>384</dim>
678 <dim>384</dim>
679 </port>
680 </output>
681 </layer>
682 <layer id="52" name="__module.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
683 <data transpose_a="false" transpose_b="true" />
684 <input>
685 <port id="0" precision="FP32">
686 <dim>-1</dim>
687 <dim>-1</dim>
688 <dim>384</dim>
689 </port>
690 <port id="1" precision="FP32">
691 <dim>384</dim>
692 <dim>384</dim>
693 </port>
694 </input>
695 <output>
696 <port id="2" precision="FP32">
697 <dim>-1</dim>
698 <dim>-1</dim>
699 <dim>384</dim>
700 </port>
701 </output>
702 </layer>
703 <layer id="53" name="Constant_6232" type="Const" version="opset1">
704 <data element_type="f32" shape="1, 1, 384" offset="12199696" size="1536" />
705 <output>
706 <port id="0" precision="FP32">
707 <dim>1</dim>
708 <dim>1</dim>
709 <dim>384</dim>
710 </port>
711 </output>
712 </layer>
713 <layer id="54" name="__module.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
714 <data auto_broadcast="numpy" />
715 <input>
716 <port id="0" precision="FP32">
717 <dim>-1</dim>
718 <dim>-1</dim>
719 <dim>384</dim>
720 </port>
721 <port id="1" precision="FP32">
722 <dim>1</dim>
723 <dim>1</dim>
724 <dim>384</dim>
725 </port>
726 </input>
727 <output>
728 <port id="2" precision="FP32" names="129,x.1">
729 <dim>-1</dim>
730 <dim>-1</dim>
731 <dim>384</dim>
732 </port>
733 </output>
734 </layer>
735 <layer id="55" name="__module.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
736 <data element_type="f32" shape="" offset="12201232" size="4" />
737 <output>
738 <port id="0" precision="FP32" />
739 </output>
740 </layer>
741 <layer id="56" name="__module.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
742 <data element_type="f32" shape="" offset="12201236" size="4" />
743 <output>
744 <port id="0" precision="FP32" />
745 </output>
746 </layer>
747 <layer id="57" name="__module.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
748 <data element_type="f32" shape="" offset="12201232" size="4" />
749 <output>
750 <port id="0" precision="FP32" />
751 </output>
752 </layer>
753 <layer id="58" name="__module.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
754 <data element_type="f32" shape="" offset="12201236" size="4" />
755 <output>
756 <port id="0" precision="FP32" />
757 </output>
758 </layer>
759 <layer id="59" name="__module.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
760 <data levels="256" auto_broadcast="numpy" />
761 <input>
762 <port id="0" precision="FP32">
763 <dim>-1</dim>
764 <dim>-1</dim>
765 <dim>384</dim>
766 </port>
767 <port id="1" precision="FP32" />
768 <port id="2" precision="FP32" />
769 <port id="3" precision="FP32" />
770 <port id="4" precision="FP32" />
771 </input>
772 <output>
773 <port id="5" precision="FP32">
774 <dim>-1</dim>
775 <dim>-1</dim>
776 <dim>384</dim>
777 </port>
778 </output>
779 </layer>
780 <layer id="60" name="__module.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
781 <data element_type="i64" shape="4" offset="12201240" size="32" />
782 <output>
783 <port id="0" precision="I64">
784 <dim>4</dim>
785 </port>
786 </output>
787 </layer>
788 <layer id="61" name="__module.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
789 <data special_zero="true" />
790 <input>
791 <port id="0" precision="FP32">
792 <dim>-1</dim>
793 <dim>-1</dim>
794 <dim>384</dim>
795 </port>
796 <port id="1" precision="I64">
797 <dim>4</dim>
798 </port>
799 </input>
800 <output>
801 <port id="2" precision="FP32" names="133,x.3">
802 <dim>-1</dim>
803 <dim>-1</dim>
804 <dim>12</dim>
805 <dim>32</dim>
806 </port>
807 </output>
808 </layer>
809 <layer id="62" name="Constant_247" type="Const" version="opset1">
810 <data element_type="i64" shape="4" offset="12201272" size="32" />
811 <output>
812 <port id="0" precision="I64" names="134">
813 <dim>4</dim>
814 </port>
815 </output>
816 </layer>
817 <layer id="63" name="__module.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
818 <input>
819 <port id="0" precision="FP32">
820 <dim>-1</dim>
821 <dim>-1</dim>
822 <dim>12</dim>
823 <dim>32</dim>
824 </port>
825 <port id="1" precision="I64">
826 <dim>4</dim>
827 </port>
828 </input>
829 <output>
830 <port id="2" precision="FP32" names="135">
831 <dim>-1</dim>
832 <dim>12</dim>
833 <dim>-1</dim>
834 <dim>32</dim>
835 </port>
836 </output>
837 </layer>
838 <layer id="64" name="Constant_27536138" type="Const" version="opset1">
839 <data element_type="i8" shape="384, 384" offset="12201304" size="147456" />
840 <output>
841 <port id="0" precision="I8">
842 <dim>384</dim>
843 <dim>384</dim>
844 </port>
845 </output>
846 </layer>
847 <layer id="65" name="Convert_27536139" type="Convert" version="opset1">
848 <data destination_type="f32" />
849 <input>
850 <port id="0" precision="I8">
851 <dim>384</dim>
852 <dim>384</dim>
853 </port>
854 </input>
855 <output>
856 <port id="1" precision="FP32">
857 <dim>384</dim>
858 <dim>384</dim>
859 </port>
860 </output>
861 </layer>
862 <layer id="66" name="Constant_27536140" type="Const" version="opset1">
863 <data element_type="f32" shape="384, 1" offset="12348760" size="1536" />
864 <output>
865 <port id="0" precision="FP32">
866 <dim>384</dim>
867 <dim>1</dim>
868 </port>
869 </output>
870 </layer>
871 <layer id="67" name="__module.encoder.layer.0.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
872 <data auto_broadcast="numpy" />
873 <input>
874 <port id="0" precision="FP32">
875 <dim>384</dim>
876 <dim>384</dim>
877 </port>
878 <port id="1" precision="FP32">
879 <dim>384</dim>
880 <dim>1</dim>
881 </port>
882 </input>
883 <output>
884 <port id="2" precision="FP32">
885 <dim>384</dim>
886 <dim>384</dim>
887 </port>
888 </output>
889 </layer>
890 <layer id="68" name="__module.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
891 <data transpose_a="false" transpose_b="true" />
892 <input>
893 <port id="0" precision="FP32">
894 <dim>-1</dim>
895 <dim>-1</dim>
896 <dim>384</dim>
897 </port>
898 <port id="1" precision="FP32">
899 <dim>384</dim>
900 <dim>384</dim>
901 </port>
902 </input>
903 <output>
904 <port id="2" precision="FP32">
905 <dim>-1</dim>
906 <dim>-1</dim>
907 <dim>384</dim>
908 </port>
909 </output>
910 </layer>
911 <layer id="69" name="Constant_6233" type="Const" version="opset1">
912 <data element_type="f32" shape="1, 1, 384" offset="12350296" size="1536" />
913 <output>
914 <port id="0" precision="FP32">
915 <dim>1</dim>
916 <dim>1</dim>
917 <dim>384</dim>
918 </port>
919 </output>
920 </layer>
921 <layer id="70" name="__module.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
922 <data auto_broadcast="numpy" />
923 <input>
924 <port id="0" precision="FP32">
925 <dim>-1</dim>
926 <dim>-1</dim>
927 <dim>384</dim>
928 </port>
929 <port id="1" precision="FP32">
930 <dim>1</dim>
931 <dim>1</dim>
932 <dim>384</dim>
933 </port>
934 </input>
935 <output>
936 <port id="2" precision="FP32" names="138,x.5">
937 <dim>-1</dim>
938 <dim>-1</dim>
939 <dim>384</dim>
940 </port>
941 </output>
942 </layer>
943 <layer id="71" name="__module.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
944 <data element_type="f32" shape="" offset="12351832" size="4" />
945 <output>
946 <port id="0" precision="FP32" />
947 </output>
948 </layer>
949 <layer id="72" name="__module.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
950 <data element_type="f32" shape="" offset="12351836" size="4" />
951 <output>
952 <port id="0" precision="FP32" />
953 </output>
954 </layer>
955 <layer id="73" name="__module.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
956 <data element_type="f32" shape="" offset="12351832" size="4" />
957 <output>
958 <port id="0" precision="FP32" />
959 </output>
960 </layer>
961 <layer id="74" name="__module.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
962 <data element_type="f32" shape="" offset="12351836" size="4" />
963 <output>
964 <port id="0" precision="FP32" />
965 </output>
966 </layer>
967 <layer id="75" name="__module.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
968 <data levels="256" auto_broadcast="numpy" />
969 <input>
970 <port id="0" precision="FP32">
971 <dim>-1</dim>
972 <dim>-1</dim>
973 <dim>384</dim>
974 </port>
975 <port id="1" precision="FP32" />
976 <port id="2" precision="FP32" />
977 <port id="3" precision="FP32" />
978 <port id="4" precision="FP32" />
979 </input>
980 <output>
981 <port id="5" precision="FP32">
982 <dim>-1</dim>
983 <dim>-1</dim>
984 <dim>384</dim>
985 </port>
986 </output>
987 </layer>
988 <layer id="76" name="__module.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
989 <data element_type="i64" shape="4" offset="12201240" size="32" />
990 <output>
991 <port id="0" precision="I64">
992 <dim>4</dim>
993 </port>
994 </output>
995 </layer>
996 <layer id="77" name="__module.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
997 <data special_zero="true" />
998 <input>
999 <port id="0" precision="FP32">
1000 <dim>-1</dim>
1001 <dim>-1</dim>
1002 <dim>384</dim>
1003 </port>
1004 <port id="1" precision="I64">
1005 <dim>4</dim>
1006 </port>
1007 </input>
1008 <output>
1009 <port id="2" precision="FP32" names="142,x.7">
1010 <dim>-1</dim>
1011 <dim>-1</dim>
1012 <dim>12</dim>
1013 <dim>32</dim>
1014 </port>
1015 </output>
1016 </layer>
1017 <layer id="78" name="Constant_272" type="Const" version="opset1">
1018 <data element_type="i64" shape="4" offset="12201272" size="32" />
1019 <output>
1020 <port id="0" precision="I64" names="143">
1021 <dim>4</dim>
1022 </port>
1023 </output>
1024 </layer>
1025 <layer id="79" name="__module.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1026 <input>
1027 <port id="0" precision="FP32">
1028 <dim>-1</dim>
1029 <dim>-1</dim>
1030 <dim>12</dim>
1031 <dim>32</dim>
1032 </port>
1033 <port id="1" precision="I64">
1034 <dim>4</dim>
1035 </port>
1036 </input>
1037 <output>
1038 <port id="2" precision="FP32" names="144">
1039 <dim>-1</dim>
1040 <dim>12</dim>
1041 <dim>-1</dim>
1042 <dim>32</dim>
1043 </port>
1044 </output>
1045 </layer>
1046 <layer id="80" name="Constant_27536142" type="Const" version="opset1">
1047 <data element_type="i8" shape="384, 384" offset="12351840" size="147456" />
1048 <output>
1049 <port id="0" precision="I8">
1050 <dim>384</dim>
1051 <dim>384</dim>
1052 </port>
1053 </output>
1054 </layer>
1055 <layer id="81" name="Convert_27536143" type="Convert" version="opset1">
1056 <data destination_type="f32" />
1057 <input>
1058 <port id="0" precision="I8">
1059 <dim>384</dim>
1060 <dim>384</dim>
1061 </port>
1062 </input>
1063 <output>
1064 <port id="1" precision="FP32">
1065 <dim>384</dim>
1066 <dim>384</dim>
1067 </port>
1068 </output>
1069 </layer>
1070 <layer id="82" name="Constant_27536144" type="Const" version="opset1">
1071 <data element_type="f32" shape="384, 1" offset="12499296" size="1536" />
1072 <output>
1073 <port id="0" precision="FP32">
1074 <dim>384</dim>
1075 <dim>1</dim>
1076 </port>
1077 </output>
1078 </layer>
1079 <layer id="83" name="__module.encoder.layer.0.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1080 <data auto_broadcast="numpy" />
1081 <input>
1082 <port id="0" precision="FP32">
1083 <dim>384</dim>
1084 <dim>384</dim>
1085 </port>
1086 <port id="1" precision="FP32">
1087 <dim>384</dim>
1088 <dim>1</dim>
1089 </port>
1090 </input>
1091 <output>
1092 <port id="2" precision="FP32">
1093 <dim>384</dim>
1094 <dim>384</dim>
1095 </port>
1096 </output>
1097 </layer>
1098 <layer id="84" name="__module.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1099 <data transpose_a="false" transpose_b="true" />
1100 <input>
1101 <port id="0" precision="FP32">
1102 <dim>-1</dim>
1103 <dim>-1</dim>
1104 <dim>384</dim>
1105 </port>
1106 <port id="1" precision="FP32">
1107 <dim>384</dim>
1108 <dim>384</dim>
1109 </port>
1110 </input>
1111 <output>
1112 <port id="2" precision="FP32">
1113 <dim>-1</dim>
1114 <dim>-1</dim>
1115 <dim>384</dim>
1116 </port>
1117 </output>
1118 </layer>
1119 <layer id="85" name="Constant_6234" type="Const" version="opset1">
1120 <data element_type="f32" shape="1, 1, 384" offset="12500832" size="1536" />
1121 <output>
1122 <port id="0" precision="FP32">
1123 <dim>1</dim>
1124 <dim>1</dim>
1125 <dim>384</dim>
1126 </port>
1127 </output>
1128 </layer>
1129 <layer id="86" name="__module.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
1130 <data auto_broadcast="numpy" />
1131 <input>
1132 <port id="0" precision="FP32">
1133 <dim>-1</dim>
1134 <dim>-1</dim>
1135 <dim>384</dim>
1136 </port>
1137 <port id="1" precision="FP32">
1138 <dim>1</dim>
1139 <dim>1</dim>
1140 <dim>384</dim>
1141 </port>
1142 </input>
1143 <output>
1144 <port id="2" precision="FP32" names="147,x.9">
1145 <dim>-1</dim>
1146 <dim>-1</dim>
1147 <dim>384</dim>
1148 </port>
1149 </output>
1150 </layer>
1151 <layer id="87" name="__module.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
1152 <data element_type="i64" shape="4" offset="12201240" size="32" />
1153 <output>
1154 <port id="0" precision="I64">
1155 <dim>4</dim>
1156 </port>
1157 </output>
1158 </layer>
1159 <layer id="88" name="__module.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
1160 <data special_zero="true" />
1161 <input>
1162 <port id="0" precision="FP32">
1163 <dim>-1</dim>
1164 <dim>-1</dim>
1165 <dim>384</dim>
1166 </port>
1167 <port id="1" precision="I64">
1168 <dim>4</dim>
1169 </port>
1170 </input>
1171 <output>
1172 <port id="2" precision="FP32" names="151,x.11">
1173 <dim>-1</dim>
1174 <dim>-1</dim>
1175 <dim>12</dim>
1176 <dim>32</dim>
1177 </port>
1178 </output>
1179 </layer>
1180 <layer id="89" name="Constant_297" type="Const" version="opset1">
1181 <data element_type="i64" shape="4" offset="12201272" size="32" />
1182 <output>
1183 <port id="0" precision="I64" names="152">
1184 <dim>4</dim>
1185 </port>
1186 </output>
1187 </layer>
1188 <layer id="90" name="__module.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
1189 <input>
1190 <port id="0" precision="FP32">
1191 <dim>-1</dim>
1192 <dim>-1</dim>
1193 <dim>12</dim>
1194 <dim>32</dim>
1195 </port>
1196 <port id="1" precision="I64">
1197 <dim>4</dim>
1198 </port>
1199 </input>
1200 <output>
1201 <port id="2" precision="FP32" names="153">
1202 <dim>-1</dim>
1203 <dim>12</dim>
1204 <dim>-1</dim>
1205 <dim>32</dim>
1206 </port>
1207 </output>
1208 </layer>
1209 <layer id="91" name="Constant_6236" type="Const" version="opset1">
1210 <data element_type="f32" shape="1, 1, 1, 1" offset="12502368" size="4" />
1211 <output>
1212 <port id="0" precision="FP32">
1213 <dim>1</dim>
1214 <dim>1</dim>
1215 <dim>1</dim>
1216 <dim>1</dim>
1217 </port>
1218 </output>
1219 </layer>
1220 <layer id="92" name="25" type="Const" version="opset1">
1221 <data element_type="i64" shape="" offset="12046076" size="8" />
1222 <output>
1223 <port id="0" precision="I64" names="25" />
1224 </output>
1225 </layer>
1226 <layer id="93" name="aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
1227 <input>
1228 <port id="0" precision="I64">
1229 <dim>-1</dim>
1230 <dim>-1</dim>
1231 </port>
1232 <port id="1" precision="I64" />
1233 </input>
1234 <output>
1235 <port id="2" precision="I64" names="26">
1236 <dim>-1</dim>
1237 <dim>1</dim>
1238 <dim>-1</dim>
1239 </port>
1240 </output>
1241 </layer>
1242 <layer id="94" name="27" type="Const" version="opset1">
1243 <data element_type="i64" shape="" offset="12502372" size="8" />
1244 <output>
1245 <port id="0" precision="I64" names="27" />
1246 </output>
1247 </layer>
1248 <layer id="95" name="aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
1249 <input>
1250 <port id="0" precision="I64">
1251 <dim>-1</dim>
1252 <dim>1</dim>
1253 <dim>-1</dim>
1254 </port>
1255 <port id="1" precision="I64" />
1256 </input>
1257 <output>
1258 <port id="2" precision="I64" names="28,33">
1259 <dim>-1</dim>
1260 <dim>1</dim>
1261 <dim>1</dim>
1262 <dim>-1</dim>
1263 </port>
1264 </output>
1265 </layer>
1266 <layer id="96" name="ShapeOf_6363" type="ShapeOf" version="opset3">
1267 <data output_type="i64" />
1268 <input>
1269 <port id="0" precision="I64">
1270 <dim>-1</dim>
1271 <dim>-1</dim>
1272 </port>
1273 </input>
1274 <output>
1275 <port id="1" precision="I64">
1276 <dim>2</dim>
1277 </port>
1278 </output>
1279 </layer>
1280 <layer id="97" name="Constant_6479" type="Const" version="opset1">
1281 <data element_type="i64" shape="1" offset="12046068" size="8" />
1282 <output>
1283 <port id="0" precision="I64">
1284 <dim>1</dim>
1285 </port>
1286 </output>
1287 </layer>
1288 <layer id="98" name="Constant_6365" type="Const" version="opset1">
1289 <data element_type="i64" shape="" offset="12046068" size="8" />
1290 <output>
1291 <port id="0" precision="I64" />
1292 </output>
1293 </layer>
1294 <layer id="99" name="Gather_6366" type="Gather" version="opset8">
1295 <data batch_dims="0" />
1296 <input>
1297 <port id="0" precision="I64">
1298 <dim>2</dim>
1299 </port>
1300 <port id="1" precision="I64">
1301 <dim>1</dim>
1302 </port>
1303 <port id="2" precision="I64" />
1304 </input>
1305 <output>
1306 <port id="3" precision="I64" names="13,15">
1307 <dim>1</dim>
1308 </port>
1309 </output>
1310 </layer>
1311 <layer id="100" name="Constant_5460" type="Const" version="opset1">
1312 <data element_type="i64" shape="1" offset="12046076" size="8" />
1313 <output>
1314 <port id="0" precision="I64">
1315 <dim>1</dim>
1316 </port>
1317 </output>
1318 </layer>
1319 <layer id="101" name="Constant_6482" type="Const" version="opset1">
1320 <data element_type="i64" shape="1" offset="12046076" size="8" />
1321 <output>
1322 <port id="0" precision="I64">
1323 <dim>1</dim>
1324 </port>
1325 </output>
1326 </layer>
1327 <layer id="102" name="Constant_6373" type="Const" version="opset1">
1328 <data element_type="i64" shape="" offset="12046068" size="8" />
1329 <output>
1330 <port id="0" precision="I64" />
1331 </output>
1332 </layer>
1333 <layer id="103" name="Gather_6374" type="Gather" version="opset8">
1334 <data batch_dims="0" />
1335 <input>
1336 <port id="0" precision="I64">
1337 <dim>2</dim>
1338 </port>
1339 <port id="1" precision="I64">
1340 <dim>1</dim>
1341 </port>
1342 <port id="2" precision="I64" />
1343 </input>
1344 <output>
1345 <port id="3" precision="I64" names="17,19">
1346 <dim>1</dim>
1347 </port>
1348 </output>
1349 </layer>
1350 <layer id="104" name="prim::ListConstruct/Concat" type="Concat" version="opset1">
1351 <data axis="0" />
1352 <input>
1353 <port id="0" precision="I64">
1354 <dim>1</dim>
1355 </port>
1356 <port id="1" precision="I64">
1357 <dim>1</dim>
1358 </port>
1359 <port id="2" precision="I64">
1360 <dim>1</dim>
1361 </port>
1362 <port id="3" precision="I64">
1363 <dim>1</dim>
1364 </port>
1365 </input>
1366 <output>
1367 <port id="4" precision="I64" names="35">
1368 <dim>4</dim>
1369 </port>
1370 </output>
1371 </layer>
1372 <layer id="105" name="aten::expand/Broadcast" type="Broadcast" version="opset3">
1373 <data mode="bidirectional" />
1374 <input>
1375 <port id="0" precision="I64">
1376 <dim>-1</dim>
1377 <dim>1</dim>
1378 <dim>1</dim>
1379 <dim>-1</dim>
1380 </port>
1381 <port id="1" precision="I64">
1382 <dim>4</dim>
1383 </port>
1384 </input>
1385 <output>
1386 <port id="2" precision="I64" names="37">
1387 <dim>-1</dim>
1388 <dim>1</dim>
1389 <dim>-1</dim>
1390 <dim>-1</dim>
1391 </port>
1392 </output>
1393 </layer>
1394 <layer id="106" name="aten::to/Convert" type="Convert" version="opset1">
1395 <data destination_type="f32" />
1396 <input>
1397 <port id="0" precision="I64">
1398 <dim>-1</dim>
1399 <dim>1</dim>
1400 <dim>-1</dim>
1401 <dim>-1</dim>
1402 </port>
1403 </input>
1404 <output>
1405 <port id="1" precision="FP32" names="42">
1406 <dim>-1</dim>
1407 <dim>1</dim>
1408 <dim>-1</dim>
1409 <dim>-1</dim>
1410 </port>
1411 </output>
1412 </layer>
1413 <layer id="107" name="Constant_6235" type="Const" version="opset1">
1414 <data element_type="f32" shape="1, 1, 1, 1" offset="12502368" size="4" />
1415 <output>
1416 <port id="0" precision="FP32">
1417 <dim>1</dim>
1418 <dim>1</dim>
1419 <dim>1</dim>
1420 <dim>1</dim>
1421 </port>
1422 </output>
1423 </layer>
1424 <layer id="108" name="aten::rsub/Multiply" type="Multiply" version="opset1">
1425 <data auto_broadcast="numpy" />
1426 <input>
1427 <port id="0" precision="FP32">
1428 <dim>-1</dim>
1429 <dim>1</dim>
1430 <dim>-1</dim>
1431 <dim>-1</dim>
1432 </port>
1433 <port id="1" precision="FP32">
1434 <dim>1</dim>
1435 <dim>1</dim>
1436 <dim>1</dim>
1437 <dim>1</dim>
1438 </port>
1439 </input>
1440 <output>
1441 <port id="2" precision="FP32">
1442 <dim>-1</dim>
1443 <dim>1</dim>
1444 <dim>-1</dim>
1445 <dim>-1</dim>
1446 </port>
1447 </output>
1448 </layer>
1449 <layer id="109" name="aten::rsub/Subtract" type="Subtract" version="opset1">
1450 <data auto_broadcast="numpy" />
1451 <input>
1452 <port id="0" precision="FP32">
1453 <dim>1</dim>
1454 <dim>1</dim>
1455 <dim>1</dim>
1456 <dim>1</dim>
1457 </port>
1458 <port id="1" precision="FP32">
1459 <dim>-1</dim>
1460 <dim>1</dim>
1461 <dim>-1</dim>
1462 <dim>-1</dim>
1463 </port>
1464 </input>
1465 <output>
1466 <port id="2" precision="FP32" names="45,inverted_mask">
1467 <dim>-1</dim>
1468 <dim>1</dim>
1469 <dim>-1</dim>
1470 <dim>-1</dim>
1471 </port>
1472 </output>
1473 </layer>
1474 <layer id="110" name="aten::to/Convert_1" type="Convert" version="opset1">
1475 <data destination_type="boolean" />
1476 <input>
1477 <port id="0" precision="FP32">
1478 <dim>-1</dim>
1479 <dim>1</dim>
1480 <dim>-1</dim>
1481 <dim>-1</dim>
1482 </port>
1483 </input>
1484 <output>
1485 <port id="1" precision="BOOL" names="50">
1486 <dim>-1</dim>
1487 <dim>1</dim>
1488 <dim>-1</dim>
1489 <dim>-1</dim>
1490 </port>
1491 </output>
1492 </layer>
1493 <layer id="111" name="aten::masked_fill/ConvertLike" type="Const" version="opset1">
1494 <data element_type="f32" shape="" offset="12502380" size="4" />
1495 <output>
1496 <port id="0" precision="FP32" />
1497 </output>
1498 </layer>
1499 <layer id="112" name="aten::masked_fill/Select" type="Select" version="opset1">
1500 <data auto_broadcast="numpy" />
1501 <input>
1502 <port id="0" precision="BOOL">
1503 <dim>-1</dim>
1504 <dim>1</dim>
1505 <dim>-1</dim>
1506 <dim>-1</dim>
1507 </port>
1508 <port id="1" precision="FP32" />
1509 <port id="2" precision="FP32">
1510 <dim>-1</dim>
1511 <dim>1</dim>
1512 <dim>-1</dim>
1513 <dim>-1</dim>
1514 </port>
1515 </input>
1516 <output>
1517 <port id="3" precision="FP32" names="52">
1518 <dim>-1</dim>
1519 <dim>1</dim>
1520 <dim>-1</dim>
1521 <dim>-1</dim>
1522 </port>
1523 </output>
1524 </layer>
1525 <layer id="113" name="__module.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1526 <data causal="false" />
1527 <input>
1528 <port id="0" precision="FP32">
1529 <dim>-1</dim>
1530 <dim>12</dim>
1531 <dim>-1</dim>
1532 <dim>32</dim>
1533 </port>
1534 <port id="1" precision="FP32">
1535 <dim>-1</dim>
1536 <dim>12</dim>
1537 <dim>-1</dim>
1538 <dim>32</dim>
1539 </port>
1540 <port id="2" precision="FP32">
1541 <dim>-1</dim>
1542 <dim>12</dim>
1543 <dim>-1</dim>
1544 <dim>32</dim>
1545 </port>
1546 <port id="3" precision="FP32">
1547 <dim>-1</dim>
1548 <dim>1</dim>
1549 <dim>-1</dim>
1550 <dim>-1</dim>
1551 </port>
1552 </input>
1553 <output>
1554 <port id="4" precision="FP32" names="154,attn_output.1">
1555 <dim>-1</dim>
1556 <dim>12</dim>
1557 <dim>-1</dim>
1558 <dim>32</dim>
1559 </port>
1560 </output>
1561 </layer>
1562 <layer id="114" name="__module.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1563 <data element_type="i32" shape="4" offset="12502384" size="16" />
1564 <output>
1565 <port id="0" precision="I32">
1566 <dim>4</dim>
1567 </port>
1568 </output>
1569 </layer>
1570 <layer id="115" name="__module.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1571 <input>
1572 <port id="0" precision="FP32">
1573 <dim>-1</dim>
1574 <dim>12</dim>
1575 <dim>-1</dim>
1576 <dim>32</dim>
1577 </port>
1578 <port id="1" precision="I32">
1579 <dim>4</dim>
1580 </port>
1581 </input>
1582 <output>
1583 <port id="2" precision="FP32" names="155,attn_output.3">
1584 <dim>-1</dim>
1585 <dim>-1</dim>
1586 <dim>12</dim>
1587 <dim>32</dim>
1588 </port>
1589 </output>
1590 </layer>
1591 <layer id="116" name="__module.encoder.layer.0.attention.self/aten::size/ShapeOf_6" type="ShapeOf" version="opset3">
1592 <data output_type="i64" />
1593 <input>
1594 <port id="0" precision="FP32">
1595 <dim>-1</dim>
1596 <dim>-1</dim>
1597 <dim>384</dim>
1598 </port>
1599 </input>
1600 <output>
1601 <port id="1" precision="I64">
1602 <dim>3</dim>
1603 </port>
1604 </output>
1605 </layer>
1606 <layer id="117" name="Constant_5737" type="Const" version="opset1">
1607 <data element_type="i64" shape="2" offset="12502400" size="16" />
1608 <output>
1609 <port id="0" precision="I64">
1610 <dim>2</dim>
1611 </port>
1612 </output>
1613 </layer>
1614 <layer id="118" name="Constant_5738" type="Const" version="opset1">
1615 <data element_type="i64" shape="" offset="12046068" size="8" />
1616 <output>
1617 <port id="0" precision="I64" />
1618 </output>
1619 </layer>
1620 <layer id="119" name="Gather_5739" type="Gather" version="opset8">
1621 <data batch_dims="0" />
1622 <input>
1623 <port id="0" precision="I64">
1624 <dim>3</dim>
1625 </port>
1626 <port id="1" precision="I64">
1627 <dim>2</dim>
1628 </port>
1629 <port id="2" precision="I64" />
1630 </input>
1631 <output>
1632 <port id="3" precision="I64">
1633 <dim>2</dim>
1634 </port>
1635 </output>
1636 </layer>
1637 <layer id="120" name="__module.encoder.layer.0.attention.self/prim::ListConstruct/Reshape_1_3" type="Const" version="opset1">
1638 <data element_type="i64" shape="1" offset="12502416" size="8" />
1639 <output>
1640 <port id="0" precision="I64">
1641 <dim>1</dim>
1642 </port>
1643 </output>
1644 </layer>
1645 <layer id="121" name="__module.encoder.layer.0.attention.self/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
1646 <data axis="0" />
1647 <input>
1648 <port id="0" precision="I64">
1649 <dim>2</dim>
1650 </port>
1651 <port id="1" precision="I64">
1652 <dim>1</dim>
1653 </port>
1654 </input>
1655 <output>
1656 <port id="2" precision="I64" names="156">
1657 <dim>3</dim>
1658 </port>
1659 </output>
1660 </layer>
1661 <layer id="122" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1662 <data special_zero="false" />
1663 <input>
1664 <port id="0" precision="FP32">
1665 <dim>-1</dim>
1666 <dim>-1</dim>
1667 <dim>12</dim>
1668 <dim>32</dim>
1669 </port>
1670 <port id="1" precision="I64">
1671 <dim>3</dim>
1672 </port>
1673 </input>
1674 <output>
1675 <port id="2" precision="FP32" names="157">
1676 <dim>-1</dim>
1677 <dim>-1</dim>
1678 <dim>384</dim>
1679 </port>
1680 </output>
1681 </layer>
1682 <layer id="123" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1683 <data element_type="f32" shape="1, 1, 384" offset="12502424" size="1536" />
1684 <output>
1685 <port id="0" precision="FP32">
1686 <dim>1</dim>
1687 <dim>1</dim>
1688 <dim>384</dim>
1689 </port>
1690 </output>
1691 </layer>
1692 <layer id="124" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1693 <data auto_broadcast="numpy" />
1694 <input>
1695 <port id="0" precision="FP32">
1696 <dim>-1</dim>
1697 <dim>-1</dim>
1698 <dim>384</dim>
1699 </port>
1700 <port id="1" precision="FP32">
1701 <dim>1</dim>
1702 <dim>1</dim>
1703 <dim>384</dim>
1704 </port>
1705 </input>
1706 <output>
1707 <port id="2" precision="FP32">
1708 <dim>-1</dim>
1709 <dim>-1</dim>
1710 <dim>384</dim>
1711 </port>
1712 </output>
1713 </layer>
1714 <layer id="125" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
1715 <data element_type="f32" shape="" offset="12503960" size="4" />
1716 <output>
1717 <port id="0" precision="FP32" />
1718 </output>
1719 </layer>
1720 <layer id="126" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
1721 <data element_type="f32" shape="" offset="12503964" size="4" />
1722 <output>
1723 <port id="0" precision="FP32" />
1724 </output>
1725 </layer>
1726 <layer id="127" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
1727 <data element_type="f32" shape="" offset="12503960" size="4" />
1728 <output>
1729 <port id="0" precision="FP32" />
1730 </output>
1731 </layer>
1732 <layer id="128" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
1733 <data element_type="f32" shape="" offset="12503964" size="4" />
1734 <output>
1735 <port id="0" precision="FP32" />
1736 </output>
1737 </layer>
1738 <layer id="129" name="__module.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
1739 <data levels="256" auto_broadcast="numpy" />
1740 <input>
1741 <port id="0" precision="FP32">
1742 <dim>-1</dim>
1743 <dim>-1</dim>
1744 <dim>384</dim>
1745 </port>
1746 <port id="1" precision="FP32" />
1747 <port id="2" precision="FP32" />
1748 <port id="3" precision="FP32" />
1749 <port id="4" precision="FP32" />
1750 </input>
1751 <output>
1752 <port id="5" precision="FP32">
1753 <dim>-1</dim>
1754 <dim>-1</dim>
1755 <dim>384</dim>
1756 </port>
1757 </output>
1758 </layer>
1759 <layer id="130" name="Constant_27536146" type="Const" version="opset1">
1760 <data element_type="i8" shape="384, 384" offset="12503968" size="147456" />
1761 <output>
1762 <port id="0" precision="I8">
1763 <dim>384</dim>
1764 <dim>384</dim>
1765 </port>
1766 </output>
1767 </layer>
1768 <layer id="131" name="Convert_27536147" type="Convert" version="opset1">
1769 <data destination_type="f32" />
1770 <input>
1771 <port id="0" precision="I8">
1772 <dim>384</dim>
1773 <dim>384</dim>
1774 </port>
1775 </input>
1776 <output>
1777 <port id="1" precision="FP32">
1778 <dim>384</dim>
1779 <dim>384</dim>
1780 </port>
1781 </output>
1782 </layer>
1783 <layer id="132" name="Constant_27536148" type="Const" version="opset1">
1784 <data element_type="f32" shape="384, 1" offset="12651424" size="1536" />
1785 <output>
1786 <port id="0" precision="FP32">
1787 <dim>384</dim>
1788 <dim>1</dim>
1789 </port>
1790 </output>
1791 </layer>
1792 <layer id="133" name="__module.encoder.layer.0.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1793 <data auto_broadcast="numpy" />
1794 <input>
1795 <port id="0" precision="FP32">
1796 <dim>384</dim>
1797 <dim>384</dim>
1798 </port>
1799 <port id="1" precision="FP32">
1800 <dim>384</dim>
1801 <dim>1</dim>
1802 </port>
1803 </input>
1804 <output>
1805 <port id="2" precision="FP32">
1806 <dim>384</dim>
1807 <dim>384</dim>
1808 </port>
1809 </output>
1810 </layer>
1811 <layer id="134" name="__module.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1812 <data transpose_a="false" transpose_b="true" />
1813 <input>
1814 <port id="0" precision="FP32">
1815 <dim>-1</dim>
1816 <dim>-1</dim>
1817 <dim>384</dim>
1818 </port>
1819 <port id="1" precision="FP32">
1820 <dim>384</dim>
1821 <dim>384</dim>
1822 </port>
1823 </input>
1824 <output>
1825 <port id="2" precision="FP32">
1826 <dim>-1</dim>
1827 <dim>-1</dim>
1828 <dim>384</dim>
1829 </port>
1830 </output>
1831 </layer>
1832 <layer id="135" name="Constant_6237" type="Const" version="opset1">
1833 <data element_type="f32" shape="1, 1, 384" offset="12652960" size="1536" />
1834 <output>
1835 <port id="0" precision="FP32">
1836 <dim>1</dim>
1837 <dim>1</dim>
1838 <dim>384</dim>
1839 </port>
1840 </output>
1841 </layer>
1842 <layer id="136" name="__module.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1843 <data auto_broadcast="numpy" />
1844 <input>
1845 <port id="0" precision="FP32">
1846 <dim>-1</dim>
1847 <dim>-1</dim>
1848 <dim>384</dim>
1849 </port>
1850 <port id="1" precision="FP32">
1851 <dim>1</dim>
1852 <dim>1</dim>
1853 <dim>384</dim>
1854 </port>
1855 </input>
1856 <output>
1857 <port id="2" precision="FP32" names="163,input.3">
1858 <dim>-1</dim>
1859 <dim>-1</dim>
1860 <dim>384</dim>
1861 </port>
1862 </output>
1863 </layer>
1864 <layer id="137" name="__module.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1865 <data auto_broadcast="numpy" />
1866 <input>
1867 <port id="0" precision="FP32">
1868 <dim>-1</dim>
1869 <dim>-1</dim>
1870 <dim>384</dim>
1871 </port>
1872 <port id="1" precision="FP32">
1873 <dim>-1</dim>
1874 <dim>-1</dim>
1875 <dim>384</dim>
1876 </port>
1877 </input>
1878 <output>
1879 <port id="2" precision="FP32" names="165">
1880 <dim>-1</dim>
1881 <dim>-1</dim>
1882 <dim>384</dim>
1883 </port>
1884 </output>
1885 </layer>
1886 <layer id="138" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1887 <data element_type="i32" shape="1" offset="12046084" size="4" />
1888 <output>
1889 <port id="0" precision="I32">
1890 <dim>1</dim>
1891 </port>
1892 </output>
1893 </layer>
1894 <layer id="139" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1895 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1896 <input>
1897 <port id="0" precision="FP32">
1898 <dim>-1</dim>
1899 <dim>-1</dim>
1900 <dim>384</dim>
1901 </port>
1902 <port id="1" precision="I32">
1903 <dim>1</dim>
1904 </port>
1905 </input>
1906 <output>
1907 <port id="2" precision="FP32">
1908 <dim>-1</dim>
1909 <dim>-1</dim>
1910 <dim>384</dim>
1911 </port>
1912 </output>
1913 </layer>
1914 <layer id="140" name="Constant_6238" type="Const" version="opset1">
1915 <data element_type="f32" shape="1, 1, 384" offset="12654496" size="1536" />
1916 <output>
1917 <port id="0" precision="FP32">
1918 <dim>1</dim>
1919 <dim>1</dim>
1920 <dim>384</dim>
1921 </port>
1922 </output>
1923 </layer>
1924 <layer id="141" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1925 <data auto_broadcast="numpy" />
1926 <input>
1927 <port id="0" precision="FP32">
1928 <dim>-1</dim>
1929 <dim>-1</dim>
1930 <dim>384</dim>
1931 </port>
1932 <port id="1" precision="FP32">
1933 <dim>1</dim>
1934 <dim>1</dim>
1935 <dim>384</dim>
1936 </port>
1937 </input>
1938 <output>
1939 <port id="2" precision="FP32">
1940 <dim>-1</dim>
1941 <dim>-1</dim>
1942 <dim>384</dim>
1943 </port>
1944 </output>
1945 </layer>
1946 <layer id="142" name="Constant_6239" type="Const" version="opset1">
1947 <data element_type="f32" shape="1, 1, 384" offset="12656032" size="1536" />
1948 <output>
1949 <port id="0" precision="FP32">
1950 <dim>1</dim>
1951 <dim>1</dim>
1952 <dim>384</dim>
1953 </port>
1954 </output>
1955 </layer>
1956 <layer id="143" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1957 <data auto_broadcast="numpy" />
1958 <input>
1959 <port id="0" precision="FP32">
1960 <dim>-1</dim>
1961 <dim>-1</dim>
1962 <dim>384</dim>
1963 </port>
1964 <port id="1" precision="FP32">
1965 <dim>1</dim>
1966 <dim>1</dim>
1967 <dim>384</dim>
1968 </port>
1969 </input>
1970 <output>
1971 <port id="2" precision="FP32" names="169,input_tensor.1">
1972 <dim>-1</dim>
1973 <dim>-1</dim>
1974 <dim>384</dim>
1975 </port>
1976 </output>
1977 </layer>
1978 <layer id="144" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1979 <data element_type="f32" shape="1, 1, 384" offset="12657568" size="1536" />
1980 <output>
1981 <port id="0" precision="FP32">
1982 <dim>1</dim>
1983 <dim>1</dim>
1984 <dim>384</dim>
1985 </port>
1986 </output>
1987 </layer>
1988 <layer id="145" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1989 <data auto_broadcast="numpy" />
1990 <input>
1991 <port id="0" precision="FP32">
1992 <dim>-1</dim>
1993 <dim>-1</dim>
1994 <dim>384</dim>
1995 </port>
1996 <port id="1" precision="FP32">
1997 <dim>1</dim>
1998 <dim>1</dim>
1999 <dim>384</dim>
2000 </port>
2001 </input>
2002 <output>
2003 <port id="2" precision="FP32">
2004 <dim>-1</dim>
2005 <dim>-1</dim>
2006 <dim>384</dim>
2007 </port>
2008 </output>
2009 </layer>
2010 <layer id="146" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2011 <data element_type="f32" shape="" offset="12659104" size="4" />
2012 <output>
2013 <port id="0" precision="FP32" />
2014 </output>
2015 </layer>
2016 <layer id="147" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2017 <data element_type="f32" shape="" offset="12659108" size="4" />
2018 <output>
2019 <port id="0" precision="FP32" />
2020 </output>
2021 </layer>
2022 <layer id="148" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2023 <data element_type="f32" shape="" offset="12659104" size="4" />
2024 <output>
2025 <port id="0" precision="FP32" />
2026 </output>
2027 </layer>
2028 <layer id="149" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2029 <data element_type="f32" shape="" offset="12659108" size="4" />
2030 <output>
2031 <port id="0" precision="FP32" />
2032 </output>
2033 </layer>
2034 <layer id="150" name="__module.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2035 <data levels="256" auto_broadcast="numpy" />
2036 <input>
2037 <port id="0" precision="FP32">
2038 <dim>-1</dim>
2039 <dim>-1</dim>
2040 <dim>384</dim>
2041 </port>
2042 <port id="1" precision="FP32" />
2043 <port id="2" precision="FP32" />
2044 <port id="3" precision="FP32" />
2045 <port id="4" precision="FP32" />
2046 </input>
2047 <output>
2048 <port id="5" precision="FP32">
2049 <dim>-1</dim>
2050 <dim>-1</dim>
2051 <dim>384</dim>
2052 </port>
2053 </output>
2054 </layer>
2055 <layer id="151" name="Constant_27536150" type="Const" version="opset1">
2056 <data element_type="i8" shape="1536, 384" offset="12659112" size="589824" />
2057 <output>
2058 <port id="0" precision="I8">
2059 <dim>1536</dim>
2060 <dim>384</dim>
2061 </port>
2062 </output>
2063 </layer>
2064 <layer id="152" name="Convert_27536151" type="Convert" version="opset1">
2065 <data destination_type="f32" />
2066 <input>
2067 <port id="0" precision="I8">
2068 <dim>1536</dim>
2069 <dim>384</dim>
2070 </port>
2071 </input>
2072 <output>
2073 <port id="1" precision="FP32">
2074 <dim>1536</dim>
2075 <dim>384</dim>
2076 </port>
2077 </output>
2078 </layer>
2079 <layer id="153" name="Constant_27536152" type="Const" version="opset1">
2080 <data element_type="f32" shape="1536, 1" offset="13248936" size="6144" />
2081 <output>
2082 <port id="0" precision="FP32">
2083 <dim>1536</dim>
2084 <dim>1</dim>
2085 </port>
2086 </output>
2087 </layer>
2088 <layer id="154" name="__module.encoder.layer.0.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2089 <data auto_broadcast="numpy" />
2090 <input>
2091 <port id="0" precision="FP32">
2092 <dim>1536</dim>
2093 <dim>384</dim>
2094 </port>
2095 <port id="1" precision="FP32">
2096 <dim>1536</dim>
2097 <dim>1</dim>
2098 </port>
2099 </input>
2100 <output>
2101 <port id="2" precision="FP32">
2102 <dim>1536</dim>
2103 <dim>384</dim>
2104 </port>
2105 </output>
2106 </layer>
2107 <layer id="155" name="__module.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2108 <data transpose_a="false" transpose_b="true" />
2109 <input>
2110 <port id="0" precision="FP32">
2111 <dim>-1</dim>
2112 <dim>-1</dim>
2113 <dim>384</dim>
2114 </port>
2115 <port id="1" precision="FP32">
2116 <dim>1536</dim>
2117 <dim>384</dim>
2118 </port>
2119 </input>
2120 <output>
2121 <port id="2" precision="FP32">
2122 <dim>-1</dim>
2123 <dim>-1</dim>
2124 <dim>1536</dim>
2125 </port>
2126 </output>
2127 </layer>
2128 <layer id="156" name="Constant_6240" type="Const" version="opset1">
2129 <data element_type="f32" shape="1, 1, 1536" offset="13255080" size="6144" />
2130 <output>
2131 <port id="0" precision="FP32">
2132 <dim>1</dim>
2133 <dim>1</dim>
2134 <dim>1536</dim>
2135 </port>
2136 </output>
2137 </layer>
2138 <layer id="157" name="__module.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2139 <data auto_broadcast="numpy" />
2140 <input>
2141 <port id="0" precision="FP32">
2142 <dim>-1</dim>
2143 <dim>-1</dim>
2144 <dim>1536</dim>
2145 </port>
2146 <port id="1" precision="FP32">
2147 <dim>1</dim>
2148 <dim>1</dim>
2149 <dim>1536</dim>
2150 </port>
2151 </input>
2152 <output>
2153 <port id="2" precision="FP32" names="174">
2154 <dim>-1</dim>
2155 <dim>-1</dim>
2156 <dim>1536</dim>
2157 </port>
2158 </output>
2159 </layer>
2160 <layer id="158" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2161 <data approximation_mode="ERF" />
2162 <input>
2163 <port id="0" precision="FP32">
2164 <dim>-1</dim>
2165 <dim>-1</dim>
2166 <dim>1536</dim>
2167 </port>
2168 </input>
2169 <output>
2170 <port id="1" precision="FP32" names="175">
2171 <dim>-1</dim>
2172 <dim>-1</dim>
2173 <dim>1536</dim>
2174 </port>
2175 </output>
2176 </layer>
2177 <layer id="159" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2178 <data element_type="f32" shape="1, 1, 1536" offset="13261224" size="6144" />
2179 <output>
2180 <port id="0" precision="FP32">
2181 <dim>1</dim>
2182 <dim>1</dim>
2183 <dim>1536</dim>
2184 </port>
2185 </output>
2186 </layer>
2187 <layer id="160" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2188 <data auto_broadcast="numpy" />
2189 <input>
2190 <port id="0" precision="FP32">
2191 <dim>-1</dim>
2192 <dim>-1</dim>
2193 <dim>1536</dim>
2194 </port>
2195 <port id="1" precision="FP32">
2196 <dim>1</dim>
2197 <dim>1</dim>
2198 <dim>1536</dim>
2199 </port>
2200 </input>
2201 <output>
2202 <port id="2" precision="FP32">
2203 <dim>-1</dim>
2204 <dim>-1</dim>
2205 <dim>1536</dim>
2206 </port>
2207 </output>
2208 </layer>
2209 <layer id="161" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2210 <data element_type="f32" shape="" offset="13267368" size="4" />
2211 <output>
2212 <port id="0" precision="FP32" />
2213 </output>
2214 </layer>
2215 <layer id="162" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2216 <data element_type="f32" shape="" offset="13267372" size="4" />
2217 <output>
2218 <port id="0" precision="FP32" />
2219 </output>
2220 </layer>
2221 <layer id="163" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2222 <data element_type="f32" shape="" offset="13267368" size="4" />
2223 <output>
2224 <port id="0" precision="FP32" />
2225 </output>
2226 </layer>
2227 <layer id="164" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2228 <data element_type="f32" shape="" offset="13267372" size="4" />
2229 <output>
2230 <port id="0" precision="FP32" />
2231 </output>
2232 </layer>
2233 <layer id="165" name="__module.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2234 <data levels="256" auto_broadcast="numpy" />
2235 <input>
2236 <port id="0" precision="FP32">
2237 <dim>-1</dim>
2238 <dim>-1</dim>
2239 <dim>1536</dim>
2240 </port>
2241 <port id="1" precision="FP32" />
2242 <port id="2" precision="FP32" />
2243 <port id="3" precision="FP32" />
2244 <port id="4" precision="FP32" />
2245 </input>
2246 <output>
2247 <port id="5" precision="FP32">
2248 <dim>-1</dim>
2249 <dim>-1</dim>
2250 <dim>1536</dim>
2251 </port>
2252 </output>
2253 </layer>
2254 <layer id="166" name="Constant_27536154" type="Const" version="opset1">
2255 <data element_type="i8" shape="384, 1536" offset="13267376" size="589824" />
2256 <output>
2257 <port id="0" precision="I8">
2258 <dim>384</dim>
2259 <dim>1536</dim>
2260 </port>
2261 </output>
2262 </layer>
2263 <layer id="167" name="Convert_27536155" type="Convert" version="opset1">
2264 <data destination_type="f32" />
2265 <input>
2266 <port id="0" precision="I8">
2267 <dim>384</dim>
2268 <dim>1536</dim>
2269 </port>
2270 </input>
2271 <output>
2272 <port id="1" precision="FP32">
2273 <dim>384</dim>
2274 <dim>1536</dim>
2275 </port>
2276 </output>
2277 </layer>
2278 <layer id="168" name="Constant_27536156" type="Const" version="opset1">
2279 <data element_type="f32" shape="384, 1" offset="13857200" size="1536" />
2280 <output>
2281 <port id="0" precision="FP32">
2282 <dim>384</dim>
2283 <dim>1</dim>
2284 </port>
2285 </output>
2286 </layer>
2287 <layer id="169" name="__module.encoder.layer.0.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2288 <data auto_broadcast="numpy" />
2289 <input>
2290 <port id="0" precision="FP32">
2291 <dim>384</dim>
2292 <dim>1536</dim>
2293 </port>
2294 <port id="1" precision="FP32">
2295 <dim>384</dim>
2296 <dim>1</dim>
2297 </port>
2298 </input>
2299 <output>
2300 <port id="2" precision="FP32">
2301 <dim>384</dim>
2302 <dim>1536</dim>
2303 </port>
2304 </output>
2305 </layer>
2306 <layer id="170" name="__module.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2307 <data transpose_a="false" transpose_b="true" />
2308 <input>
2309 <port id="0" precision="FP32">
2310 <dim>-1</dim>
2311 <dim>-1</dim>
2312 <dim>1536</dim>
2313 </port>
2314 <port id="1" precision="FP32">
2315 <dim>384</dim>
2316 <dim>1536</dim>
2317 </port>
2318 </input>
2319 <output>
2320 <port id="2" precision="FP32">
2321 <dim>-1</dim>
2322 <dim>-1</dim>
2323 <dim>384</dim>
2324 </port>
2325 </output>
2326 </layer>
2327 <layer id="171" name="Constant_6241" type="Const" version="opset1">
2328 <data element_type="f32" shape="1, 1, 384" offset="13858736" size="1536" />
2329 <output>
2330 <port id="0" precision="FP32">
2331 <dim>1</dim>
2332 <dim>1</dim>
2333 <dim>384</dim>
2334 </port>
2335 </output>
2336 </layer>
2337 <layer id="172" name="__module.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
2338 <data auto_broadcast="numpy" />
2339 <input>
2340 <port id="0" precision="FP32">
2341 <dim>-1</dim>
2342 <dim>-1</dim>
2343 <dim>384</dim>
2344 </port>
2345 <port id="1" precision="FP32">
2346 <dim>1</dim>
2347 <dim>1</dim>
2348 <dim>384</dim>
2349 </port>
2350 </input>
2351 <output>
2352 <port id="2" precision="FP32" names="181,input.5">
2353 <dim>-1</dim>
2354 <dim>-1</dim>
2355 <dim>384</dim>
2356 </port>
2357 </output>
2358 </layer>
2359 <layer id="173" name="__module.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
2360 <data auto_broadcast="numpy" />
2361 <input>
2362 <port id="0" precision="FP32">
2363 <dim>-1</dim>
2364 <dim>-1</dim>
2365 <dim>384</dim>
2366 </port>
2367 <port id="1" precision="FP32">
2368 <dim>-1</dim>
2369 <dim>-1</dim>
2370 <dim>384</dim>
2371 </port>
2372 </input>
2373 <output>
2374 <port id="2" precision="FP32" names="183">
2375 <dim>-1</dim>
2376 <dim>-1</dim>
2377 <dim>384</dim>
2378 </port>
2379 </output>
2380 </layer>
2381 <layer id="174" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2382 <data element_type="i32" shape="1" offset="12046084" size="4" />
2383 <output>
2384 <port id="0" precision="I32">
2385 <dim>1</dim>
2386 </port>
2387 </output>
2388 </layer>
2389 <layer id="175" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2390 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2391 <input>
2392 <port id="0" precision="FP32">
2393 <dim>-1</dim>
2394 <dim>-1</dim>
2395 <dim>384</dim>
2396 </port>
2397 <port id="1" precision="I32">
2398 <dim>1</dim>
2399 </port>
2400 </input>
2401 <output>
2402 <port id="2" precision="FP32">
2403 <dim>-1</dim>
2404 <dim>-1</dim>
2405 <dim>384</dim>
2406 </port>
2407 </output>
2408 </layer>
2409 <layer id="176" name="Constant_6242" type="Const" version="opset1">
2410 <data element_type="f32" shape="1, 1, 384" offset="13860272" size="1536" />
2411 <output>
2412 <port id="0" precision="FP32">
2413 <dim>1</dim>
2414 <dim>1</dim>
2415 <dim>384</dim>
2416 </port>
2417 </output>
2418 </layer>
2419 <layer id="177" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2420 <data auto_broadcast="numpy" />
2421 <input>
2422 <port id="0" precision="FP32">
2423 <dim>-1</dim>
2424 <dim>-1</dim>
2425 <dim>384</dim>
2426 </port>
2427 <port id="1" precision="FP32">
2428 <dim>1</dim>
2429 <dim>1</dim>
2430 <dim>384</dim>
2431 </port>
2432 </input>
2433 <output>
2434 <port id="2" precision="FP32">
2435 <dim>-1</dim>
2436 <dim>-1</dim>
2437 <dim>384</dim>
2438 </port>
2439 </output>
2440 </layer>
2441 <layer id="178" name="Constant_6243" type="Const" version="opset1">
2442 <data element_type="f32" shape="1, 1, 384" offset="13861808" size="1536" />
2443 <output>
2444 <port id="0" precision="FP32">
2445 <dim>1</dim>
2446 <dim>1</dim>
2447 <dim>384</dim>
2448 </port>
2449 </output>
2450 </layer>
2451 <layer id="179" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2452 <data auto_broadcast="numpy" />
2453 <input>
2454 <port id="0" precision="FP32">
2455 <dim>-1</dim>
2456 <dim>-1</dim>
2457 <dim>384</dim>
2458 </port>
2459 <port id="1" precision="FP32">
2460 <dim>1</dim>
2461 <dim>1</dim>
2462 <dim>384</dim>
2463 </port>
2464 </input>
2465 <output>
2466 <port id="2" precision="FP32" names="187,hidden_states.7">
2467 <dim>-1</dim>
2468 <dim>-1</dim>
2469 <dim>384</dim>
2470 </port>
2471 </output>
2472 </layer>
2473 <layer id="180" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2474 <data element_type="f32" shape="1, 1, 384" offset="13863344" size="1536" />
2475 <output>
2476 <port id="0" precision="FP32">
2477 <dim>1</dim>
2478 <dim>1</dim>
2479 <dim>384</dim>
2480 </port>
2481 </output>
2482 </layer>
2483 <layer id="181" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2484 <data auto_broadcast="numpy" />
2485 <input>
2486 <port id="0" precision="FP32">
2487 <dim>-1</dim>
2488 <dim>-1</dim>
2489 <dim>384</dim>
2490 </port>
2491 <port id="1" precision="FP32">
2492 <dim>1</dim>
2493 <dim>1</dim>
2494 <dim>384</dim>
2495 </port>
2496 </input>
2497 <output>
2498 <port id="2" precision="FP32">
2499 <dim>-1</dim>
2500 <dim>-1</dim>
2501 <dim>384</dim>
2502 </port>
2503 </output>
2504 </layer>
2505 <layer id="182" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2506 <data element_type="f32" shape="" offset="13864880" size="4" />
2507 <output>
2508 <port id="0" precision="FP32" />
2509 </output>
2510 </layer>
2511 <layer id="183" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2512 <data element_type="f32" shape="" offset="13864884" size="4" />
2513 <output>
2514 <port id="0" precision="FP32" />
2515 </output>
2516 </layer>
2517 <layer id="184" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2518 <data element_type="f32" shape="" offset="13864880" size="4" />
2519 <output>
2520 <port id="0" precision="FP32" />
2521 </output>
2522 </layer>
2523 <layer id="185" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2524 <data element_type="f32" shape="" offset="13864884" size="4" />
2525 <output>
2526 <port id="0" precision="FP32" />
2527 </output>
2528 </layer>
2529 <layer id="186" name="__module.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2530 <data levels="256" auto_broadcast="numpy" />
2531 <input>
2532 <port id="0" precision="FP32">
2533 <dim>-1</dim>
2534 <dim>-1</dim>
2535 <dim>384</dim>
2536 </port>
2537 <port id="1" precision="FP32" />
2538 <port id="2" precision="FP32" />
2539 <port id="3" precision="FP32" />
2540 <port id="4" precision="FP32" />
2541 </input>
2542 <output>
2543 <port id="5" precision="FP32">
2544 <dim>-1</dim>
2545 <dim>-1</dim>
2546 <dim>384</dim>
2547 </port>
2548 </output>
2549 </layer>
2550 <layer id="187" name="Constant_27536158" type="Const" version="opset1">
2551 <data element_type="i8" shape="384, 384" offset="13864888" size="147456" />
2552 <output>
2553 <port id="0" precision="I8">
2554 <dim>384</dim>
2555 <dim>384</dim>
2556 </port>
2557 </output>
2558 </layer>
2559 <layer id="188" name="Convert_27536159" type="Convert" version="opset1">
2560 <data destination_type="f32" />
2561 <input>
2562 <port id="0" precision="I8">
2563 <dim>384</dim>
2564 <dim>384</dim>
2565 </port>
2566 </input>
2567 <output>
2568 <port id="1" precision="FP32">
2569 <dim>384</dim>
2570 <dim>384</dim>
2571 </port>
2572 </output>
2573 </layer>
2574 <layer id="189" name="Constant_27536160" type="Const" version="opset1">
2575 <data element_type="f32" shape="384, 1" offset="14012344" size="1536" />
2576 <output>
2577 <port id="0" precision="FP32">
2578 <dim>384</dim>
2579 <dim>1</dim>
2580 </port>
2581 </output>
2582 </layer>
2583 <layer id="190" name="__module.encoder.layer.1.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2584 <data auto_broadcast="numpy" />
2585 <input>
2586 <port id="0" precision="FP32">
2587 <dim>384</dim>
2588 <dim>384</dim>
2589 </port>
2590 <port id="1" precision="FP32">
2591 <dim>384</dim>
2592 <dim>1</dim>
2593 </port>
2594 </input>
2595 <output>
2596 <port id="2" precision="FP32">
2597 <dim>384</dim>
2598 <dim>384</dim>
2599 </port>
2600 </output>
2601 </layer>
2602 <layer id="191" name="__module.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2603 <data transpose_a="false" transpose_b="true" />
2604 <input>
2605 <port id="0" precision="FP32">
2606 <dim>-1</dim>
2607 <dim>-1</dim>
2608 <dim>384</dim>
2609 </port>
2610 <port id="1" precision="FP32">
2611 <dim>384</dim>
2612 <dim>384</dim>
2613 </port>
2614 </input>
2615 <output>
2616 <port id="2" precision="FP32">
2617 <dim>-1</dim>
2618 <dim>-1</dim>
2619 <dim>384</dim>
2620 </port>
2621 </output>
2622 </layer>
2623 <layer id="192" name="Constant_6244" type="Const" version="opset1">
2624 <data element_type="f32" shape="1, 1, 384" offset="14013880" size="1536" />
2625 <output>
2626 <port id="0" precision="FP32">
2627 <dim>1</dim>
2628 <dim>1</dim>
2629 <dim>384</dim>
2630 </port>
2631 </output>
2632 </layer>
2633 <layer id="193" name="__module.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2634 <data auto_broadcast="numpy" />
2635 <input>
2636 <port id="0" precision="FP32">
2637 <dim>-1</dim>
2638 <dim>-1</dim>
2639 <dim>384</dim>
2640 </port>
2641 <port id="1" precision="FP32">
2642 <dim>1</dim>
2643 <dim>1</dim>
2644 <dim>384</dim>
2645 </port>
2646 </input>
2647 <output>
2648 <port id="2" precision="FP32" names="200,x.13">
2649 <dim>-1</dim>
2650 <dim>-1</dim>
2651 <dim>384</dim>
2652 </port>
2653 </output>
2654 </layer>
2655 <layer id="194" name="__module.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2656 <data element_type="f32" shape="" offset="14015416" size="4" />
2657 <output>
2658 <port id="0" precision="FP32" />
2659 </output>
2660 </layer>
2661 <layer id="195" name="__module.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2662 <data element_type="f32" shape="" offset="14015420" size="4" />
2663 <output>
2664 <port id="0" precision="FP32" />
2665 </output>
2666 </layer>
2667 <layer id="196" name="__module.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2668 <data element_type="f32" shape="" offset="14015416" size="4" />
2669 <output>
2670 <port id="0" precision="FP32" />
2671 </output>
2672 </layer>
2673 <layer id="197" name="__module.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2674 <data element_type="f32" shape="" offset="14015420" size="4" />
2675 <output>
2676 <port id="0" precision="FP32" />
2677 </output>
2678 </layer>
2679 <layer id="198" name="__module.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2680 <data levels="256" auto_broadcast="numpy" />
2681 <input>
2682 <port id="0" precision="FP32">
2683 <dim>-1</dim>
2684 <dim>-1</dim>
2685 <dim>384</dim>
2686 </port>
2687 <port id="1" precision="FP32" />
2688 <port id="2" precision="FP32" />
2689 <port id="3" precision="FP32" />
2690 <port id="4" precision="FP32" />
2691 </input>
2692 <output>
2693 <port id="5" precision="FP32">
2694 <dim>-1</dim>
2695 <dim>-1</dim>
2696 <dim>384</dim>
2697 </port>
2698 </output>
2699 </layer>
2700 <layer id="199" name="__module.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2701 <data element_type="i64" shape="4" offset="12201240" size="32" />
2702 <output>
2703 <port id="0" precision="I64">
2704 <dim>4</dim>
2705 </port>
2706 </output>
2707 </layer>
2708 <layer id="200" name="__module.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2709 <data special_zero="true" />
2710 <input>
2711 <port id="0" precision="FP32">
2712 <dim>-1</dim>
2713 <dim>-1</dim>
2714 <dim>384</dim>
2715 </port>
2716 <port id="1" precision="I64">
2717 <dim>4</dim>
2718 </port>
2719 </input>
2720 <output>
2721 <port id="2" precision="FP32" names="204,x.15">
2722 <dim>-1</dim>
2723 <dim>-1</dim>
2724 <dim>12</dim>
2725 <dim>32</dim>
2726 </port>
2727 </output>
2728 </layer>
2729 <layer id="201" name="Constant_479" type="Const" version="opset1">
2730 <data element_type="i64" shape="4" offset="12201272" size="32" />
2731 <output>
2732 <port id="0" precision="I64" names="205">
2733 <dim>4</dim>
2734 </port>
2735 </output>
2736 </layer>
2737 <layer id="202" name="__module.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2738 <input>
2739 <port id="0" precision="FP32">
2740 <dim>-1</dim>
2741 <dim>-1</dim>
2742 <dim>12</dim>
2743 <dim>32</dim>
2744 </port>
2745 <port id="1" precision="I64">
2746 <dim>4</dim>
2747 </port>
2748 </input>
2749 <output>
2750 <port id="2" precision="FP32" names="206">
2751 <dim>-1</dim>
2752 <dim>12</dim>
2753 <dim>-1</dim>
2754 <dim>32</dim>
2755 </port>
2756 </output>
2757 </layer>
2758 <layer id="203" name="Constant_27536162" type="Const" version="opset1">
2759 <data element_type="i8" shape="384, 384" offset="14015424" size="147456" />
2760 <output>
2761 <port id="0" precision="I8">
2762 <dim>384</dim>
2763 <dim>384</dim>
2764 </port>
2765 </output>
2766 </layer>
2767 <layer id="204" name="Convert_27536163" type="Convert" version="opset1">
2768 <data destination_type="f32" />
2769 <input>
2770 <port id="0" precision="I8">
2771 <dim>384</dim>
2772 <dim>384</dim>
2773 </port>
2774 </input>
2775 <output>
2776 <port id="1" precision="FP32">
2777 <dim>384</dim>
2778 <dim>384</dim>
2779 </port>
2780 </output>
2781 </layer>
2782 <layer id="205" name="Constant_27536164" type="Const" version="opset1">
2783 <data element_type="f32" shape="384, 1" offset="14162880" size="1536" />
2784 <output>
2785 <port id="0" precision="FP32">
2786 <dim>384</dim>
2787 <dim>1</dim>
2788 </port>
2789 </output>
2790 </layer>
2791 <layer id="206" name="__module.encoder.layer.1.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2792 <data auto_broadcast="numpy" />
2793 <input>
2794 <port id="0" precision="FP32">
2795 <dim>384</dim>
2796 <dim>384</dim>
2797 </port>
2798 <port id="1" precision="FP32">
2799 <dim>384</dim>
2800 <dim>1</dim>
2801 </port>
2802 </input>
2803 <output>
2804 <port id="2" precision="FP32">
2805 <dim>384</dim>
2806 <dim>384</dim>
2807 </port>
2808 </output>
2809 </layer>
2810 <layer id="207" name="__module.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2811 <data transpose_a="false" transpose_b="true" />
2812 <input>
2813 <port id="0" precision="FP32">
2814 <dim>-1</dim>
2815 <dim>-1</dim>
2816 <dim>384</dim>
2817 </port>
2818 <port id="1" precision="FP32">
2819 <dim>384</dim>
2820 <dim>384</dim>
2821 </port>
2822 </input>
2823 <output>
2824 <port id="2" precision="FP32">
2825 <dim>-1</dim>
2826 <dim>-1</dim>
2827 <dim>384</dim>
2828 </port>
2829 </output>
2830 </layer>
2831 <layer id="208" name="Constant_6245" type="Const" version="opset1">
2832 <data element_type="f32" shape="1, 1, 384" offset="14164416" size="1536" />
2833 <output>
2834 <port id="0" precision="FP32">
2835 <dim>1</dim>
2836 <dim>1</dim>
2837 <dim>384</dim>
2838 </port>
2839 </output>
2840 </layer>
2841 <layer id="209" name="__module.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2842 <data auto_broadcast="numpy" />
2843 <input>
2844 <port id="0" precision="FP32">
2845 <dim>-1</dim>
2846 <dim>-1</dim>
2847 <dim>384</dim>
2848 </port>
2849 <port id="1" precision="FP32">
2850 <dim>1</dim>
2851 <dim>1</dim>
2852 <dim>384</dim>
2853 </port>
2854 </input>
2855 <output>
2856 <port id="2" precision="FP32" names="209,x.17">
2857 <dim>-1</dim>
2858 <dim>-1</dim>
2859 <dim>384</dim>
2860 </port>
2861 </output>
2862 </layer>
2863 <layer id="210" name="__module.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2864 <data element_type="f32" shape="" offset="14165952" size="4" />
2865 <output>
2866 <port id="0" precision="FP32" />
2867 </output>
2868 </layer>
2869 <layer id="211" name="__module.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2870 <data element_type="f32" shape="" offset="14165956" size="4" />
2871 <output>
2872 <port id="0" precision="FP32" />
2873 </output>
2874 </layer>
2875 <layer id="212" name="__module.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2876 <data element_type="f32" shape="" offset="14165952" size="4" />
2877 <output>
2878 <port id="0" precision="FP32" />
2879 </output>
2880 </layer>
2881 <layer id="213" name="__module.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2882 <data element_type="f32" shape="" offset="14165956" size="4" />
2883 <output>
2884 <port id="0" precision="FP32" />
2885 </output>
2886 </layer>
2887 <layer id="214" name="__module.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2888 <data levels="256" auto_broadcast="numpy" />
2889 <input>
2890 <port id="0" precision="FP32">
2891 <dim>-1</dim>
2892 <dim>-1</dim>
2893 <dim>384</dim>
2894 </port>
2895 <port id="1" precision="FP32" />
2896 <port id="2" precision="FP32" />
2897 <port id="3" precision="FP32" />
2898 <port id="4" precision="FP32" />
2899 </input>
2900 <output>
2901 <port id="5" precision="FP32">
2902 <dim>-1</dim>
2903 <dim>-1</dim>
2904 <dim>384</dim>
2905 </port>
2906 </output>
2907 </layer>
2908 <layer id="215" name="__module.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2909 <data element_type="i64" shape="4" offset="12201240" size="32" />
2910 <output>
2911 <port id="0" precision="I64">
2912 <dim>4</dim>
2913 </port>
2914 </output>
2915 </layer>
2916 <layer id="216" name="__module.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2917 <data special_zero="true" />
2918 <input>
2919 <port id="0" precision="FP32">
2920 <dim>-1</dim>
2921 <dim>-1</dim>
2922 <dim>384</dim>
2923 </port>
2924 <port id="1" precision="I64">
2925 <dim>4</dim>
2926 </port>
2927 </input>
2928 <output>
2929 <port id="2" precision="FP32" names="213,x.19">
2930 <dim>-1</dim>
2931 <dim>-1</dim>
2932 <dim>12</dim>
2933 <dim>32</dim>
2934 </port>
2935 </output>
2936 </layer>
2937 <layer id="217" name="Constant_502" type="Const" version="opset1">
2938 <data element_type="i64" shape="4" offset="12201272" size="32" />
2939 <output>
2940 <port id="0" precision="I64" names="214">
2941 <dim>4</dim>
2942 </port>
2943 </output>
2944 </layer>
2945 <layer id="218" name="__module.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2946 <input>
2947 <port id="0" precision="FP32">
2948 <dim>-1</dim>
2949 <dim>-1</dim>
2950 <dim>12</dim>
2951 <dim>32</dim>
2952 </port>
2953 <port id="1" precision="I64">
2954 <dim>4</dim>
2955 </port>
2956 </input>
2957 <output>
2958 <port id="2" precision="FP32" names="215">
2959 <dim>-1</dim>
2960 <dim>12</dim>
2961 <dim>-1</dim>
2962 <dim>32</dim>
2963 </port>
2964 </output>
2965 </layer>
2966 <layer id="219" name="Constant_27536166" type="Const" version="opset1">
2967 <data element_type="i8" shape="384, 384" offset="14165960" size="147456" />
2968 <output>
2969 <port id="0" precision="I8">
2970 <dim>384</dim>
2971 <dim>384</dim>
2972 </port>
2973 </output>
2974 </layer>
2975 <layer id="220" name="Convert_27536167" type="Convert" version="opset1">
2976 <data destination_type="f32" />
2977 <input>
2978 <port id="0" precision="I8">
2979 <dim>384</dim>
2980 <dim>384</dim>
2981 </port>
2982 </input>
2983 <output>
2984 <port id="1" precision="FP32">
2985 <dim>384</dim>
2986 <dim>384</dim>
2987 </port>
2988 </output>
2989 </layer>
2990 <layer id="221" name="Constant_27536168" type="Const" version="opset1">
2991 <data element_type="f32" shape="384, 1" offset="14313416" size="1536" />
2992 <output>
2993 <port id="0" precision="FP32">
2994 <dim>384</dim>
2995 <dim>1</dim>
2996 </port>
2997 </output>
2998 </layer>
2999 <layer id="222" name="__module.encoder.layer.1.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3000 <data auto_broadcast="numpy" />
3001 <input>
3002 <port id="0" precision="FP32">
3003 <dim>384</dim>
3004 <dim>384</dim>
3005 </port>
3006 <port id="1" precision="FP32">
3007 <dim>384</dim>
3008 <dim>1</dim>
3009 </port>
3010 </input>
3011 <output>
3012 <port id="2" precision="FP32">
3013 <dim>384</dim>
3014 <dim>384</dim>
3015 </port>
3016 </output>
3017 </layer>
3018 <layer id="223" name="__module.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
3019 <data transpose_a="false" transpose_b="true" />
3020 <input>
3021 <port id="0" precision="FP32">
3022 <dim>-1</dim>
3023 <dim>-1</dim>
3024 <dim>384</dim>
3025 </port>
3026 <port id="1" precision="FP32">
3027 <dim>384</dim>
3028 <dim>384</dim>
3029 </port>
3030 </input>
3031 <output>
3032 <port id="2" precision="FP32">
3033 <dim>-1</dim>
3034 <dim>-1</dim>
3035 <dim>384</dim>
3036 </port>
3037 </output>
3038 </layer>
3039 <layer id="224" name="Constant_6246" type="Const" version="opset1">
3040 <data element_type="f32" shape="1, 1, 384" offset="14314952" size="1536" />
3041 <output>
3042 <port id="0" precision="FP32">
3043 <dim>1</dim>
3044 <dim>1</dim>
3045 <dim>384</dim>
3046 </port>
3047 </output>
3048 </layer>
3049 <layer id="225" name="__module.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
3050 <data auto_broadcast="numpy" />
3051 <input>
3052 <port id="0" precision="FP32">
3053 <dim>-1</dim>
3054 <dim>-1</dim>
3055 <dim>384</dim>
3056 </port>
3057 <port id="1" precision="FP32">
3058 <dim>1</dim>
3059 <dim>1</dim>
3060 <dim>384</dim>
3061 </port>
3062 </input>
3063 <output>
3064 <port id="2" precision="FP32" names="218,x.21">
3065 <dim>-1</dim>
3066 <dim>-1</dim>
3067 <dim>384</dim>
3068 </port>
3069 </output>
3070 </layer>
3071 <layer id="226" name="__module.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
3072 <data element_type="i64" shape="4" offset="12201240" size="32" />
3073 <output>
3074 <port id="0" precision="I64">
3075 <dim>4</dim>
3076 </port>
3077 </output>
3078 </layer>
3079 <layer id="227" name="__module.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3080 <data special_zero="true" />
3081 <input>
3082 <port id="0" precision="FP32">
3083 <dim>-1</dim>
3084 <dim>-1</dim>
3085 <dim>384</dim>
3086 </port>
3087 <port id="1" precision="I64">
3088 <dim>4</dim>
3089 </port>
3090 </input>
3091 <output>
3092 <port id="2" precision="FP32" names="222,x.23">
3093 <dim>-1</dim>
3094 <dim>-1</dim>
3095 <dim>12</dim>
3096 <dim>32</dim>
3097 </port>
3098 </output>
3099 </layer>
3100 <layer id="228" name="Constant_525" type="Const" version="opset1">
3101 <data element_type="i64" shape="4" offset="12201272" size="32" />
3102 <output>
3103 <port id="0" precision="I64" names="223">
3104 <dim>4</dim>
3105 </port>
3106 </output>
3107 </layer>
3108 <layer id="229" name="__module.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3109 <input>
3110 <port id="0" precision="FP32">
3111 <dim>-1</dim>
3112 <dim>-1</dim>
3113 <dim>12</dim>
3114 <dim>32</dim>
3115 </port>
3116 <port id="1" precision="I64">
3117 <dim>4</dim>
3118 </port>
3119 </input>
3120 <output>
3121 <port id="2" precision="FP32" names="224">
3122 <dim>-1</dim>
3123 <dim>12</dim>
3124 <dim>-1</dim>
3125 <dim>32</dim>
3126 </port>
3127 </output>
3128 </layer>
3129 <layer id="230" name="__module.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3130 <data causal="false" />
3131 <input>
3132 <port id="0" precision="FP32">
3133 <dim>-1</dim>
3134 <dim>12</dim>
3135 <dim>-1</dim>
3136 <dim>32</dim>
3137 </port>
3138 <port id="1" precision="FP32">
3139 <dim>-1</dim>
3140 <dim>12</dim>
3141 <dim>-1</dim>
3142 <dim>32</dim>
3143 </port>
3144 <port id="2" precision="FP32">
3145 <dim>-1</dim>
3146 <dim>12</dim>
3147 <dim>-1</dim>
3148 <dim>32</dim>
3149 </port>
3150 <port id="3" precision="FP32">
3151 <dim>-1</dim>
3152 <dim>1</dim>
3153 <dim>-1</dim>
3154 <dim>-1</dim>
3155 </port>
3156 </input>
3157 <output>
3158 <port id="4" precision="FP32" names="225,attn_output.5">
3159 <dim>-1</dim>
3160 <dim>12</dim>
3161 <dim>-1</dim>
3162 <dim>32</dim>
3163 </port>
3164 </output>
3165 </layer>
3166 <layer id="231" name="__module.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3167 <data element_type="i32" shape="4" offset="12502384" size="16" />
3168 <output>
3169 <port id="0" precision="I32">
3170 <dim>4</dim>
3171 </port>
3172 </output>
3173 </layer>
3174 <layer id="232" name="__module.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3175 <input>
3176 <port id="0" precision="FP32">
3177 <dim>-1</dim>
3178 <dim>12</dim>
3179 <dim>-1</dim>
3180 <dim>32</dim>
3181 </port>
3182 <port id="1" precision="I32">
3183 <dim>4</dim>
3184 </port>
3185 </input>
3186 <output>
3187 <port id="2" precision="FP32" names="226,attn_output.7">
3188 <dim>-1</dim>
3189 <dim>-1</dim>
3190 <dim>12</dim>
3191 <dim>32</dim>
3192 </port>
3193 </output>
3194 </layer>
3195 <layer id="233" name="__module.encoder.layer.1.attention.self/aten::size/ShapeOf_6" type="ShapeOf" version="opset3">
3196 <data output_type="i64" />
3197 <input>
3198 <port id="0" precision="FP32">
3199 <dim>-1</dim>
3200 <dim>-1</dim>
3201 <dim>384</dim>
3202 </port>
3203 </input>
3204 <output>
3205 <port id="1" precision="I64">
3206 <dim>3</dim>
3207 </port>
3208 </output>
3209 </layer>
3210 <layer id="234" name="Constant_5757" type="Const" version="opset1">
3211 <data element_type="i64" shape="2" offset="12502400" size="16" />
3212 <output>
3213 <port id="0" precision="I64">
3214 <dim>2</dim>
3215 </port>
3216 </output>
3217 </layer>
3218 <layer id="235" name="Constant_5758" type="Const" version="opset1">
3219 <data element_type="i64" shape="" offset="12046068" size="8" />
3220 <output>
3221 <port id="0" precision="I64" />
3222 </output>
3223 </layer>
3224 <layer id="236" name="Gather_5759" type="Gather" version="opset8">
3225 <data batch_dims="0" />
3226 <input>
3227 <port id="0" precision="I64">
3228 <dim>3</dim>
3229 </port>
3230 <port id="1" precision="I64">
3231 <dim>2</dim>
3232 </port>
3233 <port id="2" precision="I64" />
3234 </input>
3235 <output>
3236 <port id="3" precision="I64">
3237 <dim>2</dim>
3238 </port>
3239 </output>
3240 </layer>
3241 <layer id="237" name="__module.encoder.layer.1.attention.self/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
3242 <data axis="0" />
3243 <input>
3244 <port id="0" precision="I64">
3245 <dim>2</dim>
3246 </port>
3247 <port id="1" precision="I64">
3248 <dim>1</dim>
3249 </port>
3250 </input>
3251 <output>
3252 <port id="2" precision="I64" names="227">
3253 <dim>3</dim>
3254 </port>
3255 </output>
3256 </layer>
3257 <layer id="238" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3258 <data special_zero="false" />
3259 <input>
3260 <port id="0" precision="FP32">
3261 <dim>-1</dim>
3262 <dim>-1</dim>
3263 <dim>12</dim>
3264 <dim>32</dim>
3265 </port>
3266 <port id="1" precision="I64">
3267 <dim>3</dim>
3268 </port>
3269 </input>
3270 <output>
3271 <port id="2" precision="FP32" names="228">
3272 <dim>-1</dim>
3273 <dim>-1</dim>
3274 <dim>384</dim>
3275 </port>
3276 </output>
3277 </layer>
3278 <layer id="239" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3279 <data element_type="f32" shape="1, 1, 384" offset="14316488" size="1536" />
3280 <output>
3281 <port id="0" precision="FP32">
3282 <dim>1</dim>
3283 <dim>1</dim>
3284 <dim>384</dim>
3285 </port>
3286 </output>
3287 </layer>
3288 <layer id="240" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3289 <data auto_broadcast="numpy" />
3290 <input>
3291 <port id="0" precision="FP32">
3292 <dim>-1</dim>
3293 <dim>-1</dim>
3294 <dim>384</dim>
3295 </port>
3296 <port id="1" precision="FP32">
3297 <dim>1</dim>
3298 <dim>1</dim>
3299 <dim>384</dim>
3300 </port>
3301 </input>
3302 <output>
3303 <port id="2" precision="FP32">
3304 <dim>-1</dim>
3305 <dim>-1</dim>
3306 <dim>384</dim>
3307 </port>
3308 </output>
3309 </layer>
3310 <layer id="241" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3311 <data element_type="f32" shape="" offset="14318024" size="4" />
3312 <output>
3313 <port id="0" precision="FP32" />
3314 </output>
3315 </layer>
3316 <layer id="242" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3317 <data element_type="f32" shape="" offset="14318028" size="4" />
3318 <output>
3319 <port id="0" precision="FP32" />
3320 </output>
3321 </layer>
3322 <layer id="243" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3323 <data element_type="f32" shape="" offset="14318024" size="4" />
3324 <output>
3325 <port id="0" precision="FP32" />
3326 </output>
3327 </layer>
3328 <layer id="244" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3329 <data element_type="f32" shape="" offset="14318028" size="4" />
3330 <output>
3331 <port id="0" precision="FP32" />
3332 </output>
3333 </layer>
3334 <layer id="245" name="__module.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3335 <data levels="256" auto_broadcast="numpy" />
3336 <input>
3337 <port id="0" precision="FP32">
3338 <dim>-1</dim>
3339 <dim>-1</dim>
3340 <dim>384</dim>
3341 </port>
3342 <port id="1" precision="FP32" />
3343 <port id="2" precision="FP32" />
3344 <port id="3" precision="FP32" />
3345 <port id="4" precision="FP32" />
3346 </input>
3347 <output>
3348 <port id="5" precision="FP32">
3349 <dim>-1</dim>
3350 <dim>-1</dim>
3351 <dim>384</dim>
3352 </port>
3353 </output>
3354 </layer>
3355 <layer id="246" name="Constant_27536170" type="Const" version="opset1">
3356 <data element_type="i8" shape="384, 384" offset="14318032" size="147456" />
3357 <output>
3358 <port id="0" precision="I8">
3359 <dim>384</dim>
3360 <dim>384</dim>
3361 </port>
3362 </output>
3363 </layer>
3364 <layer id="247" name="Convert_27536171" type="Convert" version="opset1">
3365 <data destination_type="f32" />
3366 <input>
3367 <port id="0" precision="I8">
3368 <dim>384</dim>
3369 <dim>384</dim>
3370 </port>
3371 </input>
3372 <output>
3373 <port id="1" precision="FP32">
3374 <dim>384</dim>
3375 <dim>384</dim>
3376 </port>
3377 </output>
3378 </layer>
3379 <layer id="248" name="Constant_27536172" type="Const" version="opset1">
3380 <data element_type="f32" shape="384, 1" offset="14465488" size="1536" />
3381 <output>
3382 <port id="0" precision="FP32">
3383 <dim>384</dim>
3384 <dim>1</dim>
3385 </port>
3386 </output>
3387 </layer>
3388 <layer id="249" name="__module.encoder.layer.1.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3389 <data auto_broadcast="numpy" />
3390 <input>
3391 <port id="0" precision="FP32">
3392 <dim>384</dim>
3393 <dim>384</dim>
3394 </port>
3395 <port id="1" precision="FP32">
3396 <dim>384</dim>
3397 <dim>1</dim>
3398 </port>
3399 </input>
3400 <output>
3401 <port id="2" precision="FP32">
3402 <dim>384</dim>
3403 <dim>384</dim>
3404 </port>
3405 </output>
3406 </layer>
3407 <layer id="250" name="__module.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3408 <data transpose_a="false" transpose_b="true" />
3409 <input>
3410 <port id="0" precision="FP32">
3411 <dim>-1</dim>
3412 <dim>-1</dim>
3413 <dim>384</dim>
3414 </port>
3415 <port id="1" precision="FP32">
3416 <dim>384</dim>
3417 <dim>384</dim>
3418 </port>
3419 </input>
3420 <output>
3421 <port id="2" precision="FP32">
3422 <dim>-1</dim>
3423 <dim>-1</dim>
3424 <dim>384</dim>
3425 </port>
3426 </output>
3427 </layer>
3428 <layer id="251" name="Constant_6247" type="Const" version="opset1">
3429 <data element_type="f32" shape="1, 1, 384" offset="14467024" size="1536" />
3430 <output>
3431 <port id="0" precision="FP32">
3432 <dim>1</dim>
3433 <dim>1</dim>
3434 <dim>384</dim>
3435 </port>
3436 </output>
3437 </layer>
3438 <layer id="252" name="__module.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3439 <data auto_broadcast="numpy" />
3440 <input>
3441 <port id="0" precision="FP32">
3442 <dim>-1</dim>
3443 <dim>-1</dim>
3444 <dim>384</dim>
3445 </port>
3446 <port id="1" precision="FP32">
3447 <dim>1</dim>
3448 <dim>1</dim>
3449 <dim>384</dim>
3450 </port>
3451 </input>
3452 <output>
3453 <port id="2" precision="FP32" names="234,input.7">
3454 <dim>-1</dim>
3455 <dim>-1</dim>
3456 <dim>384</dim>
3457 </port>
3458 </output>
3459 </layer>
3460 <layer id="253" name="__module.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
3461 <data auto_broadcast="numpy" />
3462 <input>
3463 <port id="0" precision="FP32">
3464 <dim>-1</dim>
3465 <dim>-1</dim>
3466 <dim>384</dim>
3467 </port>
3468 <port id="1" precision="FP32">
3469 <dim>-1</dim>
3470 <dim>-1</dim>
3471 <dim>384</dim>
3472 </port>
3473 </input>
3474 <output>
3475 <port id="2" precision="FP32" names="236">
3476 <dim>-1</dim>
3477 <dim>-1</dim>
3478 <dim>384</dim>
3479 </port>
3480 </output>
3481 </layer>
3482 <layer id="254" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3483 <data element_type="i32" shape="1" offset="12046084" size="4" />
3484 <output>
3485 <port id="0" precision="I32">
3486 <dim>1</dim>
3487 </port>
3488 </output>
3489 </layer>
3490 <layer id="255" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3491 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3492 <input>
3493 <port id="0" precision="FP32">
3494 <dim>-1</dim>
3495 <dim>-1</dim>
3496 <dim>384</dim>
3497 </port>
3498 <port id="1" precision="I32">
3499 <dim>1</dim>
3500 </port>
3501 </input>
3502 <output>
3503 <port id="2" precision="FP32">
3504 <dim>-1</dim>
3505 <dim>-1</dim>
3506 <dim>384</dim>
3507 </port>
3508 </output>
3509 </layer>
3510 <layer id="256" name="Constant_6248" type="Const" version="opset1">
3511 <data element_type="f32" shape="1, 1, 384" offset="14468560" size="1536" />
3512 <output>
3513 <port id="0" precision="FP32">
3514 <dim>1</dim>
3515 <dim>1</dim>
3516 <dim>384</dim>
3517 </port>
3518 </output>
3519 </layer>
3520 <layer id="257" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3521 <data auto_broadcast="numpy" />
3522 <input>
3523 <port id="0" precision="FP32">
3524 <dim>-1</dim>
3525 <dim>-1</dim>
3526 <dim>384</dim>
3527 </port>
3528 <port id="1" precision="FP32">
3529 <dim>1</dim>
3530 <dim>1</dim>
3531 <dim>384</dim>
3532 </port>
3533 </input>
3534 <output>
3535 <port id="2" precision="FP32">
3536 <dim>-1</dim>
3537 <dim>-1</dim>
3538 <dim>384</dim>
3539 </port>
3540 </output>
3541 </layer>
3542 <layer id="258" name="Constant_6249" type="Const" version="opset1">
3543 <data element_type="f32" shape="1, 1, 384" offset="14470096" size="1536" />
3544 <output>
3545 <port id="0" precision="FP32">
3546 <dim>1</dim>
3547 <dim>1</dim>
3548 <dim>384</dim>
3549 </port>
3550 </output>
3551 </layer>
3552 <layer id="259" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3553 <data auto_broadcast="numpy" />
3554 <input>
3555 <port id="0" precision="FP32">
3556 <dim>-1</dim>
3557 <dim>-1</dim>
3558 <dim>384</dim>
3559 </port>
3560 <port id="1" precision="FP32">
3561 <dim>1</dim>
3562 <dim>1</dim>
3563 <dim>384</dim>
3564 </port>
3565 </input>
3566 <output>
3567 <port id="2" precision="FP32" names="240,input_tensor.3">
3568 <dim>-1</dim>
3569 <dim>-1</dim>
3570 <dim>384</dim>
3571 </port>
3572 </output>
3573 </layer>
3574 <layer id="260" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3575 <data element_type="f32" shape="1, 1, 384" offset="14471632" size="1536" />
3576 <output>
3577 <port id="0" precision="FP32">
3578 <dim>1</dim>
3579 <dim>1</dim>
3580 <dim>384</dim>
3581 </port>
3582 </output>
3583 </layer>
3584 <layer id="261" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3585 <data auto_broadcast="numpy" />
3586 <input>
3587 <port id="0" precision="FP32">
3588 <dim>-1</dim>
3589 <dim>-1</dim>
3590 <dim>384</dim>
3591 </port>
3592 <port id="1" precision="FP32">
3593 <dim>1</dim>
3594 <dim>1</dim>
3595 <dim>384</dim>
3596 </port>
3597 </input>
3598 <output>
3599 <port id="2" precision="FP32">
3600 <dim>-1</dim>
3601 <dim>-1</dim>
3602 <dim>384</dim>
3603 </port>
3604 </output>
3605 </layer>
3606 <layer id="262" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3607 <data element_type="f32" shape="" offset="14473168" size="4" />
3608 <output>
3609 <port id="0" precision="FP32" />
3610 </output>
3611 </layer>
3612 <layer id="263" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3613 <data element_type="f32" shape="" offset="14473172" size="4" />
3614 <output>
3615 <port id="0" precision="FP32" />
3616 </output>
3617 </layer>
3618 <layer id="264" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3619 <data element_type="f32" shape="" offset="14473168" size="4" />
3620 <output>
3621 <port id="0" precision="FP32" />
3622 </output>
3623 </layer>
3624 <layer id="265" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3625 <data element_type="f32" shape="" offset="14473172" size="4" />
3626 <output>
3627 <port id="0" precision="FP32" />
3628 </output>
3629 </layer>
3630 <layer id="266" name="__module.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3631 <data levels="256" auto_broadcast="numpy" />
3632 <input>
3633 <port id="0" precision="FP32">
3634 <dim>-1</dim>
3635 <dim>-1</dim>
3636 <dim>384</dim>
3637 </port>
3638 <port id="1" precision="FP32" />
3639 <port id="2" precision="FP32" />
3640 <port id="3" precision="FP32" />
3641 <port id="4" precision="FP32" />
3642 </input>
3643 <output>
3644 <port id="5" precision="FP32">
3645 <dim>-1</dim>
3646 <dim>-1</dim>
3647 <dim>384</dim>
3648 </port>
3649 </output>
3650 </layer>
3651 <layer id="267" name="Constant_27536174" type="Const" version="opset1">
3652 <data element_type="i8" shape="1536, 384" offset="14473176" size="589824" />
3653 <output>
3654 <port id="0" precision="I8">
3655 <dim>1536</dim>
3656 <dim>384</dim>
3657 </port>
3658 </output>
3659 </layer>
3660 <layer id="268" name="Convert_27536175" type="Convert" version="opset1">
3661 <data destination_type="f32" />
3662 <input>
3663 <port id="0" precision="I8">
3664 <dim>1536</dim>
3665 <dim>384</dim>
3666 </port>
3667 </input>
3668 <output>
3669 <port id="1" precision="FP32">
3670 <dim>1536</dim>
3671 <dim>384</dim>
3672 </port>
3673 </output>
3674 </layer>
3675 <layer id="269" name="Constant_27536176" type="Const" version="opset1">
3676 <data element_type="f32" shape="1536, 1" offset="15063000" size="6144" />
3677 <output>
3678 <port id="0" precision="FP32">
3679 <dim>1536</dim>
3680 <dim>1</dim>
3681 </port>
3682 </output>
3683 </layer>
3684 <layer id="270" name="__module.encoder.layer.1.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3685 <data auto_broadcast="numpy" />
3686 <input>
3687 <port id="0" precision="FP32">
3688 <dim>1536</dim>
3689 <dim>384</dim>
3690 </port>
3691 <port id="1" precision="FP32">
3692 <dim>1536</dim>
3693 <dim>1</dim>
3694 </port>
3695 </input>
3696 <output>
3697 <port id="2" precision="FP32">
3698 <dim>1536</dim>
3699 <dim>384</dim>
3700 </port>
3701 </output>
3702 </layer>
3703 <layer id="271" name="__module.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3704 <data transpose_a="false" transpose_b="true" />
3705 <input>
3706 <port id="0" precision="FP32">
3707 <dim>-1</dim>
3708 <dim>-1</dim>
3709 <dim>384</dim>
3710 </port>
3711 <port id="1" precision="FP32">
3712 <dim>1536</dim>
3713 <dim>384</dim>
3714 </port>
3715 </input>
3716 <output>
3717 <port id="2" precision="FP32">
3718 <dim>-1</dim>
3719 <dim>-1</dim>
3720 <dim>1536</dim>
3721 </port>
3722 </output>
3723 </layer>
3724 <layer id="272" name="Constant_6250" type="Const" version="opset1">
3725 <data element_type="f32" shape="1, 1, 1536" offset="15069144" size="6144" />
3726 <output>
3727 <port id="0" precision="FP32">
3728 <dim>1</dim>
3729 <dim>1</dim>
3730 <dim>1536</dim>
3731 </port>
3732 </output>
3733 </layer>
3734 <layer id="273" name="__module.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3735 <data auto_broadcast="numpy" />
3736 <input>
3737 <port id="0" precision="FP32">
3738 <dim>-1</dim>
3739 <dim>-1</dim>
3740 <dim>1536</dim>
3741 </port>
3742 <port id="1" precision="FP32">
3743 <dim>1</dim>
3744 <dim>1</dim>
3745 <dim>1536</dim>
3746 </port>
3747 </input>
3748 <output>
3749 <port id="2" precision="FP32" names="245">
3750 <dim>-1</dim>
3751 <dim>-1</dim>
3752 <dim>1536</dim>
3753 </port>
3754 </output>
3755 </layer>
3756 <layer id="274" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3757 <data approximation_mode="ERF" />
3758 <input>
3759 <port id="0" precision="FP32">
3760 <dim>-1</dim>
3761 <dim>-1</dim>
3762 <dim>1536</dim>
3763 </port>
3764 </input>
3765 <output>
3766 <port id="1" precision="FP32" names="246">
3767 <dim>-1</dim>
3768 <dim>-1</dim>
3769 <dim>1536</dim>
3770 </port>
3771 </output>
3772 </layer>
3773 <layer id="275" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3774 <data element_type="f32" shape="1, 1, 1536" offset="15075288" size="6144" />
3775 <output>
3776 <port id="0" precision="FP32">
3777 <dim>1</dim>
3778 <dim>1</dim>
3779 <dim>1536</dim>
3780 </port>
3781 </output>
3782 </layer>
3783 <layer id="276" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3784 <data auto_broadcast="numpy" />
3785 <input>
3786 <port id="0" precision="FP32">
3787 <dim>-1</dim>
3788 <dim>-1</dim>
3789 <dim>1536</dim>
3790 </port>
3791 <port id="1" precision="FP32">
3792 <dim>1</dim>
3793 <dim>1</dim>
3794 <dim>1536</dim>
3795 </port>
3796 </input>
3797 <output>
3798 <port id="2" precision="FP32">
3799 <dim>-1</dim>
3800 <dim>-1</dim>
3801 <dim>1536</dim>
3802 </port>
3803 </output>
3804 </layer>
3805 <layer id="277" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3806 <data element_type="f32" shape="" offset="15081432" size="4" />
3807 <output>
3808 <port id="0" precision="FP32" />
3809 </output>
3810 </layer>
3811 <layer id="278" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3812 <data element_type="f32" shape="" offset="15081436" size="4" />
3813 <output>
3814 <port id="0" precision="FP32" />
3815 </output>
3816 </layer>
3817 <layer id="279" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3818 <data element_type="f32" shape="" offset="15081432" size="4" />
3819 <output>
3820 <port id="0" precision="FP32" />
3821 </output>
3822 </layer>
3823 <layer id="280" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3824 <data element_type="f32" shape="" offset="15081436" size="4" />
3825 <output>
3826 <port id="0" precision="FP32" />
3827 </output>
3828 </layer>
3829 <layer id="281" name="__module.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3830 <data levels="256" auto_broadcast="numpy" />
3831 <input>
3832 <port id="0" precision="FP32">
3833 <dim>-1</dim>
3834 <dim>-1</dim>
3835 <dim>1536</dim>
3836 </port>
3837 <port id="1" precision="FP32" />
3838 <port id="2" precision="FP32" />
3839 <port id="3" precision="FP32" />
3840 <port id="4" precision="FP32" />
3841 </input>
3842 <output>
3843 <port id="5" precision="FP32">
3844 <dim>-1</dim>
3845 <dim>-1</dim>
3846 <dim>1536</dim>
3847 </port>
3848 </output>
3849 </layer>
3850 <layer id="282" name="Constant_27536178" type="Const" version="opset1">
3851 <data element_type="i8" shape="384, 1536" offset="15081440" size="589824" />
3852 <output>
3853 <port id="0" precision="I8">
3854 <dim>384</dim>
3855 <dim>1536</dim>
3856 </port>
3857 </output>
3858 </layer>
3859 <layer id="283" name="Convert_27536179" type="Convert" version="opset1">
3860 <data destination_type="f32" />
3861 <input>
3862 <port id="0" precision="I8">
3863 <dim>384</dim>
3864 <dim>1536</dim>
3865 </port>
3866 </input>
3867 <output>
3868 <port id="1" precision="FP32">
3869 <dim>384</dim>
3870 <dim>1536</dim>
3871 </port>
3872 </output>
3873 </layer>
3874 <layer id="284" name="Constant_27536180" type="Const" version="opset1">
3875 <data element_type="f32" shape="384, 1" offset="15671264" size="1536" />
3876 <output>
3877 <port id="0" precision="FP32">
3878 <dim>384</dim>
3879 <dim>1</dim>
3880 </port>
3881 </output>
3882 </layer>
3883 <layer id="285" name="__module.encoder.layer.1.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3884 <data auto_broadcast="numpy" />
3885 <input>
3886 <port id="0" precision="FP32">
3887 <dim>384</dim>
3888 <dim>1536</dim>
3889 </port>
3890 <port id="1" precision="FP32">
3891 <dim>384</dim>
3892 <dim>1</dim>
3893 </port>
3894 </input>
3895 <output>
3896 <port id="2" precision="FP32">
3897 <dim>384</dim>
3898 <dim>1536</dim>
3899 </port>
3900 </output>
3901 </layer>
3902 <layer id="286" name="__module.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3903 <data transpose_a="false" transpose_b="true" />
3904 <input>
3905 <port id="0" precision="FP32">
3906 <dim>-1</dim>
3907 <dim>-1</dim>
3908 <dim>1536</dim>
3909 </port>
3910 <port id="1" precision="FP32">
3911 <dim>384</dim>
3912 <dim>1536</dim>
3913 </port>
3914 </input>
3915 <output>
3916 <port id="2" precision="FP32">
3917 <dim>-1</dim>
3918 <dim>-1</dim>
3919 <dim>384</dim>
3920 </port>
3921 </output>
3922 </layer>
3923 <layer id="287" name="Constant_6251" type="Const" version="opset1">
3924 <data element_type="f32" shape="1, 1, 384" offset="15672800" size="1536" />
3925 <output>
3926 <port id="0" precision="FP32">
3927 <dim>1</dim>
3928 <dim>1</dim>
3929 <dim>384</dim>
3930 </port>
3931 </output>
3932 </layer>
3933 <layer id="288" name="__module.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
3934 <data auto_broadcast="numpy" />
3935 <input>
3936 <port id="0" precision="FP32">
3937 <dim>-1</dim>
3938 <dim>-1</dim>
3939 <dim>384</dim>
3940 </port>
3941 <port id="1" precision="FP32">
3942 <dim>1</dim>
3943 <dim>1</dim>
3944 <dim>384</dim>
3945 </port>
3946 </input>
3947 <output>
3948 <port id="2" precision="FP32" names="252,input.9">
3949 <dim>-1</dim>
3950 <dim>-1</dim>
3951 <dim>384</dim>
3952 </port>
3953 </output>
3954 </layer>
3955 <layer id="289" name="__module.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
3956 <data auto_broadcast="numpy" />
3957 <input>
3958 <port id="0" precision="FP32">
3959 <dim>-1</dim>
3960 <dim>-1</dim>
3961 <dim>384</dim>
3962 </port>
3963 <port id="1" precision="FP32">
3964 <dim>-1</dim>
3965 <dim>-1</dim>
3966 <dim>384</dim>
3967 </port>
3968 </input>
3969 <output>
3970 <port id="2" precision="FP32" names="254">
3971 <dim>-1</dim>
3972 <dim>-1</dim>
3973 <dim>384</dim>
3974 </port>
3975 </output>
3976 </layer>
3977 <layer id="290" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3978 <data element_type="i32" shape="1" offset="12046084" size="4" />
3979 <output>
3980 <port id="0" precision="I32">
3981 <dim>1</dim>
3982 </port>
3983 </output>
3984 </layer>
3985 <layer id="291" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3986 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3987 <input>
3988 <port id="0" precision="FP32">
3989 <dim>-1</dim>
3990 <dim>-1</dim>
3991 <dim>384</dim>
3992 </port>
3993 <port id="1" precision="I32">
3994 <dim>1</dim>
3995 </port>
3996 </input>
3997 <output>
3998 <port id="2" precision="FP32">
3999 <dim>-1</dim>
4000 <dim>-1</dim>
4001 <dim>384</dim>
4002 </port>
4003 </output>
4004 </layer>
4005 <layer id="292" name="Constant_6252" type="Const" version="opset1">
4006 <data element_type="f32" shape="1, 1, 384" offset="15674336" size="1536" />
4007 <output>
4008 <port id="0" precision="FP32">
4009 <dim>1</dim>
4010 <dim>1</dim>
4011 <dim>384</dim>
4012 </port>
4013 </output>
4014 </layer>
4015 <layer id="293" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4016 <data auto_broadcast="numpy" />
4017 <input>
4018 <port id="0" precision="FP32">
4019 <dim>-1</dim>
4020 <dim>-1</dim>
4021 <dim>384</dim>
4022 </port>
4023 <port id="1" precision="FP32">
4024 <dim>1</dim>
4025 <dim>1</dim>
4026 <dim>384</dim>
4027 </port>
4028 </input>
4029 <output>
4030 <port id="2" precision="FP32">
4031 <dim>-1</dim>
4032 <dim>-1</dim>
4033 <dim>384</dim>
4034 </port>
4035 </output>
4036 </layer>
4037 <layer id="294" name="Constant_6253" type="Const" version="opset1">
4038 <data element_type="f32" shape="1, 1, 384" offset="15675872" size="1536" />
4039 <output>
4040 <port id="0" precision="FP32">
4041 <dim>1</dim>
4042 <dim>1</dim>
4043 <dim>384</dim>
4044 </port>
4045 </output>
4046 </layer>
4047 <layer id="295" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4048 <data auto_broadcast="numpy" />
4049 <input>
4050 <port id="0" precision="FP32">
4051 <dim>-1</dim>
4052 <dim>-1</dim>
4053 <dim>384</dim>
4054 </port>
4055 <port id="1" precision="FP32">
4056 <dim>1</dim>
4057 <dim>1</dim>
4058 <dim>384</dim>
4059 </port>
4060 </input>
4061 <output>
4062 <port id="2" precision="FP32" names="258,hidden_states.13">
4063 <dim>-1</dim>
4064 <dim>-1</dim>
4065 <dim>384</dim>
4066 </port>
4067 </output>
4068 </layer>
4069 <layer id="296" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4070 <data element_type="f32" shape="1, 1, 384" offset="15677408" size="1536" />
4071 <output>
4072 <port id="0" precision="FP32">
4073 <dim>1</dim>
4074 <dim>1</dim>
4075 <dim>384</dim>
4076 </port>
4077 </output>
4078 </layer>
4079 <layer id="297" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4080 <data auto_broadcast="numpy" />
4081 <input>
4082 <port id="0" precision="FP32">
4083 <dim>-1</dim>
4084 <dim>-1</dim>
4085 <dim>384</dim>
4086 </port>
4087 <port id="1" precision="FP32">
4088 <dim>1</dim>
4089 <dim>1</dim>
4090 <dim>384</dim>
4091 </port>
4092 </input>
4093 <output>
4094 <port id="2" precision="FP32">
4095 <dim>-1</dim>
4096 <dim>-1</dim>
4097 <dim>384</dim>
4098 </port>
4099 </output>
4100 </layer>
4101 <layer id="298" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
4102 <data element_type="f32" shape="" offset="15678944" size="4" />
4103 <output>
4104 <port id="0" precision="FP32" />
4105 </output>
4106 </layer>
4107 <layer id="299" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
4108 <data element_type="f32" shape="" offset="15678948" size="4" />
4109 <output>
4110 <port id="0" precision="FP32" />
4111 </output>
4112 </layer>
4113 <layer id="300" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
4114 <data element_type="f32" shape="" offset="15678944" size="4" />
4115 <output>
4116 <port id="0" precision="FP32" />
4117 </output>
4118 </layer>
4119 <layer id="301" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
4120 <data element_type="f32" shape="" offset="15678948" size="4" />
4121 <output>
4122 <port id="0" precision="FP32" />
4123 </output>
4124 </layer>
4125 <layer id="302" name="__module.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
4126 <data levels="256" auto_broadcast="numpy" />
4127 <input>
4128 <port id="0" precision="FP32">
4129 <dim>-1</dim>
4130 <dim>-1</dim>
4131 <dim>384</dim>
4132 </port>
4133 <port id="1" precision="FP32" />
4134 <port id="2" precision="FP32" />
4135 <port id="3" precision="FP32" />
4136 <port id="4" precision="FP32" />
4137 </input>
4138 <output>
4139 <port id="5" precision="FP32">
4140 <dim>-1</dim>
4141 <dim>-1</dim>
4142 <dim>384</dim>
4143 </port>
4144 </output>
4145 </layer>
4146 <layer id="303" name="Constant_27536182" type="Const" version="opset1">
4147 <data element_type="i8" shape="384, 384" offset="15678952" size="147456" />
4148 <output>
4149 <port id="0" precision="I8">
4150 <dim>384</dim>
4151 <dim>384</dim>
4152 </port>
4153 </output>
4154 </layer>
4155 <layer id="304" name="Convert_27536183" type="Convert" version="opset1">
4156 <data destination_type="f32" />
4157 <input>
4158 <port id="0" precision="I8">
4159 <dim>384</dim>
4160 <dim>384</dim>
4161 </port>
4162 </input>
4163 <output>
4164 <port id="1" precision="FP32">
4165 <dim>384</dim>
4166 <dim>384</dim>
4167 </port>
4168 </output>
4169 </layer>
4170 <layer id="305" name="Constant_27536184" type="Const" version="opset1">
4171 <data element_type="f32" shape="384, 1" offset="15826408" size="1536" />
4172 <output>
4173 <port id="0" precision="FP32">
4174 <dim>384</dim>
4175 <dim>1</dim>
4176 </port>
4177 </output>
4178 </layer>
4179 <layer id="306" name="__module.encoder.layer.2.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4180 <data auto_broadcast="numpy" />
4181 <input>
4182 <port id="0" precision="FP32">
4183 <dim>384</dim>
4184 <dim>384</dim>
4185 </port>
4186 <port id="1" precision="FP32">
4187 <dim>384</dim>
4188 <dim>1</dim>
4189 </port>
4190 </input>
4191 <output>
4192 <port id="2" precision="FP32">
4193 <dim>384</dim>
4194 <dim>384</dim>
4195 </port>
4196 </output>
4197 </layer>
4198 <layer id="307" name="__module.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
4199 <data transpose_a="false" transpose_b="true" />
4200 <input>
4201 <port id="0" precision="FP32">
4202 <dim>-1</dim>
4203 <dim>-1</dim>
4204 <dim>384</dim>
4205 </port>
4206 <port id="1" precision="FP32">
4207 <dim>384</dim>
4208 <dim>384</dim>
4209 </port>
4210 </input>
4211 <output>
4212 <port id="2" precision="FP32">
4213 <dim>-1</dim>
4214 <dim>-1</dim>
4215 <dim>384</dim>
4216 </port>
4217 </output>
4218 </layer>
4219 <layer id="308" name="Constant_6254" type="Const" version="opset1">
4220 <data element_type="f32" shape="1, 1, 384" offset="15827944" size="1536" />
4221 <output>
4222 <port id="0" precision="FP32">
4223 <dim>1</dim>
4224 <dim>1</dim>
4225 <dim>384</dim>
4226 </port>
4227 </output>
4228 </layer>
4229 <layer id="309" name="__module.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
4230 <data auto_broadcast="numpy" />
4231 <input>
4232 <port id="0" precision="FP32">
4233 <dim>-1</dim>
4234 <dim>-1</dim>
4235 <dim>384</dim>
4236 </port>
4237 <port id="1" precision="FP32">
4238 <dim>1</dim>
4239 <dim>1</dim>
4240 <dim>384</dim>
4241 </port>
4242 </input>
4243 <output>
4244 <port id="2" precision="FP32" names="271,x.25">
4245 <dim>-1</dim>
4246 <dim>-1</dim>
4247 <dim>384</dim>
4248 </port>
4249 </output>
4250 </layer>
4251 <layer id="310" name="__module.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4252 <data element_type="f32" shape="" offset="15829480" size="4" />
4253 <output>
4254 <port id="0" precision="FP32" />
4255 </output>
4256 </layer>
4257 <layer id="311" name="__module.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4258 <data element_type="f32" shape="" offset="15829484" size="4" />
4259 <output>
4260 <port id="0" precision="FP32" />
4261 </output>
4262 </layer>
4263 <layer id="312" name="__module.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4264 <data element_type="f32" shape="" offset="15829480" size="4" />
4265 <output>
4266 <port id="0" precision="FP32" />
4267 </output>
4268 </layer>
4269 <layer id="313" name="__module.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4270 <data element_type="f32" shape="" offset="15829484" size="4" />
4271 <output>
4272 <port id="0" precision="FP32" />
4273 </output>
4274 </layer>
4275 <layer id="314" name="__module.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4276 <data levels="256" auto_broadcast="numpy" />
4277 <input>
4278 <port id="0" precision="FP32">
4279 <dim>-1</dim>
4280 <dim>-1</dim>
4281 <dim>384</dim>
4282 </port>
4283 <port id="1" precision="FP32" />
4284 <port id="2" precision="FP32" />
4285 <port id="3" precision="FP32" />
4286 <port id="4" precision="FP32" />
4287 </input>
4288 <output>
4289 <port id="5" precision="FP32">
4290 <dim>-1</dim>
4291 <dim>-1</dim>
4292 <dim>384</dim>
4293 </port>
4294 </output>
4295 </layer>
4296 <layer id="315" name="__module.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
4297 <data element_type="i64" shape="4" offset="12201240" size="32" />
4298 <output>
4299 <port id="0" precision="I64">
4300 <dim>4</dim>
4301 </port>
4302 </output>
4303 </layer>
4304 <layer id="316" name="__module.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
4305 <data special_zero="true" />
4306 <input>
4307 <port id="0" precision="FP32">
4308 <dim>-1</dim>
4309 <dim>-1</dim>
4310 <dim>384</dim>
4311 </port>
4312 <port id="1" precision="I64">
4313 <dim>4</dim>
4314 </port>
4315 </input>
4316 <output>
4317 <port id="2" precision="FP32" names="275,x.27">
4318 <dim>-1</dim>
4319 <dim>-1</dim>
4320 <dim>12</dim>
4321 <dim>32</dim>
4322 </port>
4323 </output>
4324 </layer>
4325 <layer id="317" name="Constant_705" type="Const" version="opset1">
4326 <data element_type="i64" shape="4" offset="12201272" size="32" />
4327 <output>
4328 <port id="0" precision="I64" names="276">
4329 <dim>4</dim>
4330 </port>
4331 </output>
4332 </layer>
4333 <layer id="318" name="__module.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
4334 <input>
4335 <port id="0" precision="FP32">
4336 <dim>-1</dim>
4337 <dim>-1</dim>
4338 <dim>12</dim>
4339 <dim>32</dim>
4340 </port>
4341 <port id="1" precision="I64">
4342 <dim>4</dim>
4343 </port>
4344 </input>
4345 <output>
4346 <port id="2" precision="FP32" names="277">
4347 <dim>-1</dim>
4348 <dim>12</dim>
4349 <dim>-1</dim>
4350 <dim>32</dim>
4351 </port>
4352 </output>
4353 </layer>
4354 <layer id="319" name="Constant_27536186" type="Const" version="opset1">
4355 <data element_type="i8" shape="384, 384" offset="15829488" size="147456" />
4356 <output>
4357 <port id="0" precision="I8">
4358 <dim>384</dim>
4359 <dim>384</dim>
4360 </port>
4361 </output>
4362 </layer>
4363 <layer id="320" name="Convert_27536187" type="Convert" version="opset1">
4364 <data destination_type="f32" />
4365 <input>
4366 <port id="0" precision="I8">
4367 <dim>384</dim>
4368 <dim>384</dim>
4369 </port>
4370 </input>
4371 <output>
4372 <port id="1" precision="FP32">
4373 <dim>384</dim>
4374 <dim>384</dim>
4375 </port>
4376 </output>
4377 </layer>
4378 <layer id="321" name="Constant_27536188" type="Const" version="opset1">
4379 <data element_type="f32" shape="384, 1" offset="15976944" size="1536" />
4380 <output>
4381 <port id="0" precision="FP32">
4382 <dim>384</dim>
4383 <dim>1</dim>
4384 </port>
4385 </output>
4386 </layer>
4387 <layer id="322" name="__module.encoder.layer.2.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4388 <data auto_broadcast="numpy" />
4389 <input>
4390 <port id="0" precision="FP32">
4391 <dim>384</dim>
4392 <dim>384</dim>
4393 </port>
4394 <port id="1" precision="FP32">
4395 <dim>384</dim>
4396 <dim>1</dim>
4397 </port>
4398 </input>
4399 <output>
4400 <port id="2" precision="FP32">
4401 <dim>384</dim>
4402 <dim>384</dim>
4403 </port>
4404 </output>
4405 </layer>
4406 <layer id="323" name="__module.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
4407 <data transpose_a="false" transpose_b="true" />
4408 <input>
4409 <port id="0" precision="FP32">
4410 <dim>-1</dim>
4411 <dim>-1</dim>
4412 <dim>384</dim>
4413 </port>
4414 <port id="1" precision="FP32">
4415 <dim>384</dim>
4416 <dim>384</dim>
4417 </port>
4418 </input>
4419 <output>
4420 <port id="2" precision="FP32">
4421 <dim>-1</dim>
4422 <dim>-1</dim>
4423 <dim>384</dim>
4424 </port>
4425 </output>
4426 </layer>
4427 <layer id="324" name="Constant_6255" type="Const" version="opset1">
4428 <data element_type="f32" shape="1, 1, 384" offset="15978480" size="1536" />
4429 <output>
4430 <port id="0" precision="FP32">
4431 <dim>1</dim>
4432 <dim>1</dim>
4433 <dim>384</dim>
4434 </port>
4435 </output>
4436 </layer>
4437 <layer id="325" name="__module.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
4438 <data auto_broadcast="numpy" />
4439 <input>
4440 <port id="0" precision="FP32">
4441 <dim>-1</dim>
4442 <dim>-1</dim>
4443 <dim>384</dim>
4444 </port>
4445 <port id="1" precision="FP32">
4446 <dim>1</dim>
4447 <dim>1</dim>
4448 <dim>384</dim>
4449 </port>
4450 </input>
4451 <output>
4452 <port id="2" precision="FP32" names="280,x.29">
4453 <dim>-1</dim>
4454 <dim>-1</dim>
4455 <dim>384</dim>
4456 </port>
4457 </output>
4458 </layer>
4459 <layer id="326" name="__module.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4460 <data element_type="f32" shape="" offset="15980016" size="4" />
4461 <output>
4462 <port id="0" precision="FP32" />
4463 </output>
4464 </layer>
4465 <layer id="327" name="__module.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4466 <data element_type="f32" shape="" offset="15980020" size="4" />
4467 <output>
4468 <port id="0" precision="FP32" />
4469 </output>
4470 </layer>
4471 <layer id="328" name="__module.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4472 <data element_type="f32" shape="" offset="15980016" size="4" />
4473 <output>
4474 <port id="0" precision="FP32" />
4475 </output>
4476 </layer>
4477 <layer id="329" name="__module.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4478 <data element_type="f32" shape="" offset="15980020" size="4" />
4479 <output>
4480 <port id="0" precision="FP32" />
4481 </output>
4482 </layer>
4483 <layer id="330" name="__module.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4484 <data levels="256" auto_broadcast="numpy" />
4485 <input>
4486 <port id="0" precision="FP32">
4487 <dim>-1</dim>
4488 <dim>-1</dim>
4489 <dim>384</dim>
4490 </port>
4491 <port id="1" precision="FP32" />
4492 <port id="2" precision="FP32" />
4493 <port id="3" precision="FP32" />
4494 <port id="4" precision="FP32" />
4495 </input>
4496 <output>
4497 <port id="5" precision="FP32">
4498 <dim>-1</dim>
4499 <dim>-1</dim>
4500 <dim>384</dim>
4501 </port>
4502 </output>
4503 </layer>
4504 <layer id="331" name="__module.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
4505 <data element_type="i64" shape="4" offset="12201240" size="32" />
4506 <output>
4507 <port id="0" precision="I64">
4508 <dim>4</dim>
4509 </port>
4510 </output>
4511 </layer>
4512 <layer id="332" name="__module.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
4513 <data special_zero="true" />
4514 <input>
4515 <port id="0" precision="FP32">
4516 <dim>-1</dim>
4517 <dim>-1</dim>
4518 <dim>384</dim>
4519 </port>
4520 <port id="1" precision="I64">
4521 <dim>4</dim>
4522 </port>
4523 </input>
4524 <output>
4525 <port id="2" precision="FP32" names="284,x.31">
4526 <dim>-1</dim>
4527 <dim>-1</dim>
4528 <dim>12</dim>
4529 <dim>32</dim>
4530 </port>
4531 </output>
4532 </layer>
4533 <layer id="333" name="Constant_728" type="Const" version="opset1">
4534 <data element_type="i64" shape="4" offset="12201272" size="32" />
4535 <output>
4536 <port id="0" precision="I64" names="285">
4537 <dim>4</dim>
4538 </port>
4539 </output>
4540 </layer>
4541 <layer id="334" name="__module.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
4542 <input>
4543 <port id="0" precision="FP32">
4544 <dim>-1</dim>
4545 <dim>-1</dim>
4546 <dim>12</dim>
4547 <dim>32</dim>
4548 </port>
4549 <port id="1" precision="I64">
4550 <dim>4</dim>
4551 </port>
4552 </input>
4553 <output>
4554 <port id="2" precision="FP32" names="286">
4555 <dim>-1</dim>
4556 <dim>12</dim>
4557 <dim>-1</dim>
4558 <dim>32</dim>
4559 </port>
4560 </output>
4561 </layer>
4562 <layer id="335" name="Constant_27536190" type="Const" version="opset1">
4563 <data element_type="i8" shape="384, 384" offset="15980024" size="147456" />
4564 <output>
4565 <port id="0" precision="I8">
4566 <dim>384</dim>
4567 <dim>384</dim>
4568 </port>
4569 </output>
4570 </layer>
4571 <layer id="336" name="Convert_27536191" type="Convert" version="opset1">
4572 <data destination_type="f32" />
4573 <input>
4574 <port id="0" precision="I8">
4575 <dim>384</dim>
4576 <dim>384</dim>
4577 </port>
4578 </input>
4579 <output>
4580 <port id="1" precision="FP32">
4581 <dim>384</dim>
4582 <dim>384</dim>
4583 </port>
4584 </output>
4585 </layer>
4586 <layer id="337" name="Constant_27536192" type="Const" version="opset1">
4587 <data element_type="f32" shape="384, 1" offset="16127480" size="1536" />
4588 <output>
4589 <port id="0" precision="FP32">
4590 <dim>384</dim>
4591 <dim>1</dim>
4592 </port>
4593 </output>
4594 </layer>
4595 <layer id="338" name="__module.encoder.layer.2.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4596 <data auto_broadcast="numpy" />
4597 <input>
4598 <port id="0" precision="FP32">
4599 <dim>384</dim>
4600 <dim>384</dim>
4601 </port>
4602 <port id="1" precision="FP32">
4603 <dim>384</dim>
4604 <dim>1</dim>
4605 </port>
4606 </input>
4607 <output>
4608 <port id="2" precision="FP32">
4609 <dim>384</dim>
4610 <dim>384</dim>
4611 </port>
4612 </output>
4613 </layer>
4614 <layer id="339" name="__module.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
4615 <data transpose_a="false" transpose_b="true" />
4616 <input>
4617 <port id="0" precision="FP32">
4618 <dim>-1</dim>
4619 <dim>-1</dim>
4620 <dim>384</dim>
4621 </port>
4622 <port id="1" precision="FP32">
4623 <dim>384</dim>
4624 <dim>384</dim>
4625 </port>
4626 </input>
4627 <output>
4628 <port id="2" precision="FP32">
4629 <dim>-1</dim>
4630 <dim>-1</dim>
4631 <dim>384</dim>
4632 </port>
4633 </output>
4634 </layer>
4635 <layer id="340" name="Constant_6256" type="Const" version="opset1">
4636 <data element_type="f32" shape="1, 1, 384" offset="16129016" size="1536" />
4637 <output>
4638 <port id="0" precision="FP32">
4639 <dim>1</dim>
4640 <dim>1</dim>
4641 <dim>384</dim>
4642 </port>
4643 </output>
4644 </layer>
4645 <layer id="341" name="__module.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
4646 <data auto_broadcast="numpy" />
4647 <input>
4648 <port id="0" precision="FP32">
4649 <dim>-1</dim>
4650 <dim>-1</dim>
4651 <dim>384</dim>
4652 </port>
4653 <port id="1" precision="FP32">
4654 <dim>1</dim>
4655 <dim>1</dim>
4656 <dim>384</dim>
4657 </port>
4658 </input>
4659 <output>
4660 <port id="2" precision="FP32" names="289,x.33">
4661 <dim>-1</dim>
4662 <dim>-1</dim>
4663 <dim>384</dim>
4664 </port>
4665 </output>
4666 </layer>
4667 <layer id="342" name="__module.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
4668 <data element_type="i64" shape="4" offset="12201240" size="32" />
4669 <output>
4670 <port id="0" precision="I64">
4671 <dim>4</dim>
4672 </port>
4673 </output>
4674 </layer>
4675 <layer id="343" name="__module.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
4676 <data special_zero="true" />
4677 <input>
4678 <port id="0" precision="FP32">
4679 <dim>-1</dim>
4680 <dim>-1</dim>
4681 <dim>384</dim>
4682 </port>
4683 <port id="1" precision="I64">
4684 <dim>4</dim>
4685 </port>
4686 </input>
4687 <output>
4688 <port id="2" precision="FP32" names="293,x.35">
4689 <dim>-1</dim>
4690 <dim>-1</dim>
4691 <dim>12</dim>
4692 <dim>32</dim>
4693 </port>
4694 </output>
4695 </layer>
4696 <layer id="344" name="Constant_751" type="Const" version="opset1">
4697 <data element_type="i64" shape="4" offset="12201272" size="32" />
4698 <output>
4699 <port id="0" precision="I64" names="294">
4700 <dim>4</dim>
4701 </port>
4702 </output>
4703 </layer>
4704 <layer id="345" name="__module.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
4705 <input>
4706 <port id="0" precision="FP32">
4707 <dim>-1</dim>
4708 <dim>-1</dim>
4709 <dim>12</dim>
4710 <dim>32</dim>
4711 </port>
4712 <port id="1" precision="I64">
4713 <dim>4</dim>
4714 </port>
4715 </input>
4716 <output>
4717 <port id="2" precision="FP32" names="295">
4718 <dim>-1</dim>
4719 <dim>12</dim>
4720 <dim>-1</dim>
4721 <dim>32</dim>
4722 </port>
4723 </output>
4724 </layer>
4725 <layer id="346" name="__module.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
4726 <data causal="false" />
4727 <input>
4728 <port id="0" precision="FP32">
4729 <dim>-1</dim>
4730 <dim>12</dim>
4731 <dim>-1</dim>
4732 <dim>32</dim>
4733 </port>
4734 <port id="1" precision="FP32">
4735 <dim>-1</dim>
4736 <dim>12</dim>
4737 <dim>-1</dim>
4738 <dim>32</dim>
4739 </port>
4740 <port id="2" precision="FP32">
4741 <dim>-1</dim>
4742 <dim>12</dim>
4743 <dim>-1</dim>
4744 <dim>32</dim>
4745 </port>
4746 <port id="3" precision="FP32">
4747 <dim>-1</dim>
4748 <dim>1</dim>
4749 <dim>-1</dim>
4750 <dim>-1</dim>
4751 </port>
4752 </input>
4753 <output>
4754 <port id="4" precision="FP32" names="296,attn_output.9">
4755 <dim>-1</dim>
4756 <dim>12</dim>
4757 <dim>-1</dim>
4758 <dim>32</dim>
4759 </port>
4760 </output>
4761 </layer>
4762 <layer id="347" name="__module.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
4763 <data element_type="i32" shape="4" offset="12502384" size="16" />
4764 <output>
4765 <port id="0" precision="I32">
4766 <dim>4</dim>
4767 </port>
4768 </output>
4769 </layer>
4770 <layer id="348" name="__module.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
4771 <input>
4772 <port id="0" precision="FP32">
4773 <dim>-1</dim>
4774 <dim>12</dim>
4775 <dim>-1</dim>
4776 <dim>32</dim>
4777 </port>
4778 <port id="1" precision="I32">
4779 <dim>4</dim>
4780 </port>
4781 </input>
4782 <output>
4783 <port id="2" precision="FP32" names="297,attn_output.11">
4784 <dim>-1</dim>
4785 <dim>-1</dim>
4786 <dim>12</dim>
4787 <dim>32</dim>
4788 </port>
4789 </output>
4790 </layer>
4791 <layer id="349" name="__module.encoder.layer.2.attention.self/aten::size/ShapeOf_6" type="ShapeOf" version="opset3">
4792 <data output_type="i64" />
4793 <input>
4794 <port id="0" precision="FP32">
4795 <dim>-1</dim>
4796 <dim>-1</dim>
4797 <dim>384</dim>
4798 </port>
4799 </input>
4800 <output>
4801 <port id="1" precision="I64">
4802 <dim>3</dim>
4803 </port>
4804 </output>
4805 </layer>
4806 <layer id="350" name="Constant_5777" type="Const" version="opset1">
4807 <data element_type="i64" shape="2" offset="12502400" size="16" />
4808 <output>
4809 <port id="0" precision="I64">
4810 <dim>2</dim>
4811 </port>
4812 </output>
4813 </layer>
4814 <layer id="351" name="Constant_5778" type="Const" version="opset1">
4815 <data element_type="i64" shape="" offset="12046068" size="8" />
4816 <output>
4817 <port id="0" precision="I64" />
4818 </output>
4819 </layer>
4820 <layer id="352" name="Gather_5779" type="Gather" version="opset8">
4821 <data batch_dims="0" />
4822 <input>
4823 <port id="0" precision="I64">
4824 <dim>3</dim>
4825 </port>
4826 <port id="1" precision="I64">
4827 <dim>2</dim>
4828 </port>
4829 <port id="2" precision="I64" />
4830 </input>
4831 <output>
4832 <port id="3" precision="I64">
4833 <dim>2</dim>
4834 </port>
4835 </output>
4836 </layer>
4837 <layer id="353" name="__module.encoder.layer.2.attention.self/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
4838 <data axis="0" />
4839 <input>
4840 <port id="0" precision="I64">
4841 <dim>2</dim>
4842 </port>
4843 <port id="1" precision="I64">
4844 <dim>1</dim>
4845 </port>
4846 </input>
4847 <output>
4848 <port id="2" precision="I64" names="298">
4849 <dim>3</dim>
4850 </port>
4851 </output>
4852 </layer>
4853 <layer id="354" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
4854 <data special_zero="false" />
4855 <input>
4856 <port id="0" precision="FP32">
4857 <dim>-1</dim>
4858 <dim>-1</dim>
4859 <dim>12</dim>
4860 <dim>32</dim>
4861 </port>
4862 <port id="1" precision="I64">
4863 <dim>3</dim>
4864 </port>
4865 </input>
4866 <output>
4867 <port id="2" precision="FP32" names="299">
4868 <dim>-1</dim>
4869 <dim>-1</dim>
4870 <dim>384</dim>
4871 </port>
4872 </output>
4873 </layer>
4874 <layer id="355" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4875 <data element_type="f32" shape="1, 1, 384" offset="16130552" size="1536" />
4876 <output>
4877 <port id="0" precision="FP32">
4878 <dim>1</dim>
4879 <dim>1</dim>
4880 <dim>384</dim>
4881 </port>
4882 </output>
4883 </layer>
4884 <layer id="356" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4885 <data auto_broadcast="numpy" />
4886 <input>
4887 <port id="0" precision="FP32">
4888 <dim>-1</dim>
4889 <dim>-1</dim>
4890 <dim>384</dim>
4891 </port>
4892 <port id="1" precision="FP32">
4893 <dim>1</dim>
4894 <dim>1</dim>
4895 <dim>384</dim>
4896 </port>
4897 </input>
4898 <output>
4899 <port id="2" precision="FP32">
4900 <dim>-1</dim>
4901 <dim>-1</dim>
4902 <dim>384</dim>
4903 </port>
4904 </output>
4905 </layer>
4906 <layer id="357" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
4907 <data element_type="f32" shape="" offset="16132088" size="4" />
4908 <output>
4909 <port id="0" precision="FP32" />
4910 </output>
4911 </layer>
4912 <layer id="358" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
4913 <data element_type="f32" shape="" offset="16132092" size="4" />
4914 <output>
4915 <port id="0" precision="FP32" />
4916 </output>
4917 </layer>
4918 <layer id="359" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
4919 <data element_type="f32" shape="" offset="16132088" size="4" />
4920 <output>
4921 <port id="0" precision="FP32" />
4922 </output>
4923 </layer>
4924 <layer id="360" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
4925 <data element_type="f32" shape="" offset="16132092" size="4" />
4926 <output>
4927 <port id="0" precision="FP32" />
4928 </output>
4929 </layer>
4930 <layer id="361" name="__module.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
4931 <data levels="256" auto_broadcast="numpy" />
4932 <input>
4933 <port id="0" precision="FP32">
4934 <dim>-1</dim>
4935 <dim>-1</dim>
4936 <dim>384</dim>
4937 </port>
4938 <port id="1" precision="FP32" />
4939 <port id="2" precision="FP32" />
4940 <port id="3" precision="FP32" />
4941 <port id="4" precision="FP32" />
4942 </input>
4943 <output>
4944 <port id="5" precision="FP32">
4945 <dim>-1</dim>
4946 <dim>-1</dim>
4947 <dim>384</dim>
4948 </port>
4949 </output>
4950 </layer>
4951 <layer id="362" name="Constant_27536194" type="Const" version="opset1">
4952 <data element_type="i8" shape="384, 384" offset="16132096" size="147456" />
4953 <output>
4954 <port id="0" precision="I8">
4955 <dim>384</dim>
4956 <dim>384</dim>
4957 </port>
4958 </output>
4959 </layer>
4960 <layer id="363" name="Convert_27536195" type="Convert" version="opset1">
4961 <data destination_type="f32" />
4962 <input>
4963 <port id="0" precision="I8">
4964 <dim>384</dim>
4965 <dim>384</dim>
4966 </port>
4967 </input>
4968 <output>
4969 <port id="1" precision="FP32">
4970 <dim>384</dim>
4971 <dim>384</dim>
4972 </port>
4973 </output>
4974 </layer>
4975 <layer id="364" name="Constant_27536196" type="Const" version="opset1">
4976 <data element_type="f32" shape="384, 1" offset="16279552" size="1536" />
4977 <output>
4978 <port id="0" precision="FP32">
4979 <dim>384</dim>
4980 <dim>1</dim>
4981 </port>
4982 </output>
4983 </layer>
4984 <layer id="365" name="__module.encoder.layer.2.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4985 <data auto_broadcast="numpy" />
4986 <input>
4987 <port id="0" precision="FP32">
4988 <dim>384</dim>
4989 <dim>384</dim>
4990 </port>
4991 <port id="1" precision="FP32">
4992 <dim>384</dim>
4993 <dim>1</dim>
4994 </port>
4995 </input>
4996 <output>
4997 <port id="2" precision="FP32">
4998 <dim>384</dim>
4999 <dim>384</dim>
5000 </port>
5001 </output>
5002 </layer>
5003 <layer id="366" name="__module.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5004 <data transpose_a="false" transpose_b="true" />
5005 <input>
5006 <port id="0" precision="FP32">
5007 <dim>-1</dim>
5008 <dim>-1</dim>
5009 <dim>384</dim>
5010 </port>
5011 <port id="1" precision="FP32">
5012 <dim>384</dim>
5013 <dim>384</dim>
5014 </port>
5015 </input>
5016 <output>
5017 <port id="2" precision="FP32">
5018 <dim>-1</dim>
5019 <dim>-1</dim>
5020 <dim>384</dim>
5021 </port>
5022 </output>
5023 </layer>
5024 <layer id="367" name="Constant_6257" type="Const" version="opset1">
5025 <data element_type="f32" shape="1, 1, 384" offset="16281088" size="1536" />
5026 <output>
5027 <port id="0" precision="FP32">
5028 <dim>1</dim>
5029 <dim>1</dim>
5030 <dim>384</dim>
5031 </port>
5032 </output>
5033 </layer>
5034 <layer id="368" name="__module.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
5035 <data auto_broadcast="numpy" />
5036 <input>
5037 <port id="0" precision="FP32">
5038 <dim>-1</dim>
5039 <dim>-1</dim>
5040 <dim>384</dim>
5041 </port>
5042 <port id="1" precision="FP32">
5043 <dim>1</dim>
5044 <dim>1</dim>
5045 <dim>384</dim>
5046 </port>
5047 </input>
5048 <output>
5049 <port id="2" precision="FP32" names="305,input.11">
5050 <dim>-1</dim>
5051 <dim>-1</dim>
5052 <dim>384</dim>
5053 </port>
5054 </output>
5055 </layer>
5056 <layer id="369" name="__module.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
5057 <data auto_broadcast="numpy" />
5058 <input>
5059 <port id="0" precision="FP32">
5060 <dim>-1</dim>
5061 <dim>-1</dim>
5062 <dim>384</dim>
5063 </port>
5064 <port id="1" precision="FP32">
5065 <dim>-1</dim>
5066 <dim>-1</dim>
5067 <dim>384</dim>
5068 </port>
5069 </input>
5070 <output>
5071 <port id="2" precision="FP32" names="307">
5072 <dim>-1</dim>
5073 <dim>-1</dim>
5074 <dim>384</dim>
5075 </port>
5076 </output>
5077 </layer>
5078 <layer id="370" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5079 <data element_type="i32" shape="1" offset="12046084" size="4" />
5080 <output>
5081 <port id="0" precision="I32">
5082 <dim>1</dim>
5083 </port>
5084 </output>
5085 </layer>
5086 <layer id="371" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5087 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5088 <input>
5089 <port id="0" precision="FP32">
5090 <dim>-1</dim>
5091 <dim>-1</dim>
5092 <dim>384</dim>
5093 </port>
5094 <port id="1" precision="I32">
5095 <dim>1</dim>
5096 </port>
5097 </input>
5098 <output>
5099 <port id="2" precision="FP32">
5100 <dim>-1</dim>
5101 <dim>-1</dim>
5102 <dim>384</dim>
5103 </port>
5104 </output>
5105 </layer>
5106 <layer id="372" name="Constant_6258" type="Const" version="opset1">
5107 <data element_type="f32" shape="1, 1, 384" offset="16282624" size="1536" />
5108 <output>
5109 <port id="0" precision="FP32">
5110 <dim>1</dim>
5111 <dim>1</dim>
5112 <dim>384</dim>
5113 </port>
5114 </output>
5115 </layer>
5116 <layer id="373" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5117 <data auto_broadcast="numpy" />
5118 <input>
5119 <port id="0" precision="FP32">
5120 <dim>-1</dim>
5121 <dim>-1</dim>
5122 <dim>384</dim>
5123 </port>
5124 <port id="1" precision="FP32">
5125 <dim>1</dim>
5126 <dim>1</dim>
5127 <dim>384</dim>
5128 </port>
5129 </input>
5130 <output>
5131 <port id="2" precision="FP32">
5132 <dim>-1</dim>
5133 <dim>-1</dim>
5134 <dim>384</dim>
5135 </port>
5136 </output>
5137 </layer>
5138 <layer id="374" name="Constant_6259" type="Const" version="opset1">
5139 <data element_type="f32" shape="1, 1, 384" offset="16284160" size="1536" />
5140 <output>
5141 <port id="0" precision="FP32">
5142 <dim>1</dim>
5143 <dim>1</dim>
5144 <dim>384</dim>
5145 </port>
5146 </output>
5147 </layer>
5148 <layer id="375" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5149 <data auto_broadcast="numpy" />
5150 <input>
5151 <port id="0" precision="FP32">
5152 <dim>-1</dim>
5153 <dim>-1</dim>
5154 <dim>384</dim>
5155 </port>
5156 <port id="1" precision="FP32">
5157 <dim>1</dim>
5158 <dim>1</dim>
5159 <dim>384</dim>
5160 </port>
5161 </input>
5162 <output>
5163 <port id="2" precision="FP32" names="311,input_tensor.5">
5164 <dim>-1</dim>
5165 <dim>-1</dim>
5166 <dim>384</dim>
5167 </port>
5168 </output>
5169 </layer>
5170 <layer id="376" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5171 <data element_type="f32" shape="1, 1, 384" offset="16285696" size="1536" />
5172 <output>
5173 <port id="0" precision="FP32">
5174 <dim>1</dim>
5175 <dim>1</dim>
5176 <dim>384</dim>
5177 </port>
5178 </output>
5179 </layer>
5180 <layer id="377" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5181 <data auto_broadcast="numpy" />
5182 <input>
5183 <port id="0" precision="FP32">
5184 <dim>-1</dim>
5185 <dim>-1</dim>
5186 <dim>384</dim>
5187 </port>
5188 <port id="1" precision="FP32">
5189 <dim>1</dim>
5190 <dim>1</dim>
5191 <dim>384</dim>
5192 </port>
5193 </input>
5194 <output>
5195 <port id="2" precision="FP32">
5196 <dim>-1</dim>
5197 <dim>-1</dim>
5198 <dim>384</dim>
5199 </port>
5200 </output>
5201 </layer>
5202 <layer id="378" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5203 <data element_type="f32" shape="" offset="16287232" size="4" />
5204 <output>
5205 <port id="0" precision="FP32" />
5206 </output>
5207 </layer>
5208 <layer id="379" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5209 <data element_type="f32" shape="" offset="16287236" size="4" />
5210 <output>
5211 <port id="0" precision="FP32" />
5212 </output>
5213 </layer>
5214 <layer id="380" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5215 <data element_type="f32" shape="" offset="16287232" size="4" />
5216 <output>
5217 <port id="0" precision="FP32" />
5218 </output>
5219 </layer>
5220 <layer id="381" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5221 <data element_type="f32" shape="" offset="16287236" size="4" />
5222 <output>
5223 <port id="0" precision="FP32" />
5224 </output>
5225 </layer>
5226 <layer id="382" name="__module.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5227 <data levels="256" auto_broadcast="numpy" />
5228 <input>
5229 <port id="0" precision="FP32">
5230 <dim>-1</dim>
5231 <dim>-1</dim>
5232 <dim>384</dim>
5233 </port>
5234 <port id="1" precision="FP32" />
5235 <port id="2" precision="FP32" />
5236 <port id="3" precision="FP32" />
5237 <port id="4" precision="FP32" />
5238 </input>
5239 <output>
5240 <port id="5" precision="FP32">
5241 <dim>-1</dim>
5242 <dim>-1</dim>
5243 <dim>384</dim>
5244 </port>
5245 </output>
5246 </layer>
5247 <layer id="383" name="Constant_27536198" type="Const" version="opset1">
5248 <data element_type="i8" shape="1536, 384" offset="16287240" size="589824" />
5249 <output>
5250 <port id="0" precision="I8">
5251 <dim>1536</dim>
5252 <dim>384</dim>
5253 </port>
5254 </output>
5255 </layer>
5256 <layer id="384" name="Convert_27536199" type="Convert" version="opset1">
5257 <data destination_type="f32" />
5258 <input>
5259 <port id="0" precision="I8">
5260 <dim>1536</dim>
5261 <dim>384</dim>
5262 </port>
5263 </input>
5264 <output>
5265 <port id="1" precision="FP32">
5266 <dim>1536</dim>
5267 <dim>384</dim>
5268 </port>
5269 </output>
5270 </layer>
5271 <layer id="385" name="Constant_27536200" type="Const" version="opset1">
5272 <data element_type="f32" shape="1536, 1" offset="16877064" size="6144" />
5273 <output>
5274 <port id="0" precision="FP32">
5275 <dim>1536</dim>
5276 <dim>1</dim>
5277 </port>
5278 </output>
5279 </layer>
5280 <layer id="386" name="__module.encoder.layer.2.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5281 <data auto_broadcast="numpy" />
5282 <input>
5283 <port id="0" precision="FP32">
5284 <dim>1536</dim>
5285 <dim>384</dim>
5286 </port>
5287 <port id="1" precision="FP32">
5288 <dim>1536</dim>
5289 <dim>1</dim>
5290 </port>
5291 </input>
5292 <output>
5293 <port id="2" precision="FP32">
5294 <dim>1536</dim>
5295 <dim>384</dim>
5296 </port>
5297 </output>
5298 </layer>
5299 <layer id="387" name="__module.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5300 <data transpose_a="false" transpose_b="true" />
5301 <input>
5302 <port id="0" precision="FP32">
5303 <dim>-1</dim>
5304 <dim>-1</dim>
5305 <dim>384</dim>
5306 </port>
5307 <port id="1" precision="FP32">
5308 <dim>1536</dim>
5309 <dim>384</dim>
5310 </port>
5311 </input>
5312 <output>
5313 <port id="2" precision="FP32">
5314 <dim>-1</dim>
5315 <dim>-1</dim>
5316 <dim>1536</dim>
5317 </port>
5318 </output>
5319 </layer>
5320 <layer id="388" name="Constant_6260" type="Const" version="opset1">
5321 <data element_type="f32" shape="1, 1, 1536" offset="16883208" size="6144" />
5322 <output>
5323 <port id="0" precision="FP32">
5324 <dim>1</dim>
5325 <dim>1</dim>
5326 <dim>1536</dim>
5327 </port>
5328 </output>
5329 </layer>
5330 <layer id="389" name="__module.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5331 <data auto_broadcast="numpy" />
5332 <input>
5333 <port id="0" precision="FP32">
5334 <dim>-1</dim>
5335 <dim>-1</dim>
5336 <dim>1536</dim>
5337 </port>
5338 <port id="1" precision="FP32">
5339 <dim>1</dim>
5340 <dim>1</dim>
5341 <dim>1536</dim>
5342 </port>
5343 </input>
5344 <output>
5345 <port id="2" precision="FP32" names="316">
5346 <dim>-1</dim>
5347 <dim>-1</dim>
5348 <dim>1536</dim>
5349 </port>
5350 </output>
5351 </layer>
5352 <layer id="390" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5353 <data approximation_mode="ERF" />
5354 <input>
5355 <port id="0" precision="FP32">
5356 <dim>-1</dim>
5357 <dim>-1</dim>
5358 <dim>1536</dim>
5359 </port>
5360 </input>
5361 <output>
5362 <port id="1" precision="FP32" names="317">
5363 <dim>-1</dim>
5364 <dim>-1</dim>
5365 <dim>1536</dim>
5366 </port>
5367 </output>
5368 </layer>
5369 <layer id="391" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5370 <data element_type="f32" shape="1, 1, 1536" offset="16889352" size="6144" />
5371 <output>
5372 <port id="0" precision="FP32">
5373 <dim>1</dim>
5374 <dim>1</dim>
5375 <dim>1536</dim>
5376 </port>
5377 </output>
5378 </layer>
5379 <layer id="392" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5380 <data auto_broadcast="numpy" />
5381 <input>
5382 <port id="0" precision="FP32">
5383 <dim>-1</dim>
5384 <dim>-1</dim>
5385 <dim>1536</dim>
5386 </port>
5387 <port id="1" precision="FP32">
5388 <dim>1</dim>
5389 <dim>1</dim>
5390 <dim>1536</dim>
5391 </port>
5392 </input>
5393 <output>
5394 <port id="2" precision="FP32">
5395 <dim>-1</dim>
5396 <dim>-1</dim>
5397 <dim>1536</dim>
5398 </port>
5399 </output>
5400 </layer>
5401 <layer id="393" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5402 <data element_type="f32" shape="" offset="16895496" size="4" />
5403 <output>
5404 <port id="0" precision="FP32" />
5405 </output>
5406 </layer>
5407 <layer id="394" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5408 <data element_type="f32" shape="" offset="16895500" size="4" />
5409 <output>
5410 <port id="0" precision="FP32" />
5411 </output>
5412 </layer>
5413 <layer id="395" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5414 <data element_type="f32" shape="" offset="16895496" size="4" />
5415 <output>
5416 <port id="0" precision="FP32" />
5417 </output>
5418 </layer>
5419 <layer id="396" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5420 <data element_type="f32" shape="" offset="16895500" size="4" />
5421 <output>
5422 <port id="0" precision="FP32" />
5423 </output>
5424 </layer>
5425 <layer id="397" name="__module.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5426 <data levels="256" auto_broadcast="numpy" />
5427 <input>
5428 <port id="0" precision="FP32">
5429 <dim>-1</dim>
5430 <dim>-1</dim>
5431 <dim>1536</dim>
5432 </port>
5433 <port id="1" precision="FP32" />
5434 <port id="2" precision="FP32" />
5435 <port id="3" precision="FP32" />
5436 <port id="4" precision="FP32" />
5437 </input>
5438 <output>
5439 <port id="5" precision="FP32">
5440 <dim>-1</dim>
5441 <dim>-1</dim>
5442 <dim>1536</dim>
5443 </port>
5444 </output>
5445 </layer>
5446 <layer id="398" name="Constant_27536202" type="Const" version="opset1">
5447 <data element_type="i8" shape="384, 1536" offset="16895504" size="589824" />
5448 <output>
5449 <port id="0" precision="I8">
5450 <dim>384</dim>
5451 <dim>1536</dim>
5452 </port>
5453 </output>
5454 </layer>
5455 <layer id="399" name="Convert_27536203" type="Convert" version="opset1">
5456 <data destination_type="f32" />
5457 <input>
5458 <port id="0" precision="I8">
5459 <dim>384</dim>
5460 <dim>1536</dim>
5461 </port>
5462 </input>
5463 <output>
5464 <port id="1" precision="FP32">
5465 <dim>384</dim>
5466 <dim>1536</dim>
5467 </port>
5468 </output>
5469 </layer>
5470 <layer id="400" name="Constant_27536204" type="Const" version="opset1">
5471 <data element_type="f32" shape="384, 1" offset="17485328" size="1536" />
5472 <output>
5473 <port id="0" precision="FP32">
5474 <dim>384</dim>
5475 <dim>1</dim>
5476 </port>
5477 </output>
5478 </layer>
5479 <layer id="401" name="__module.encoder.layer.2.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5480 <data auto_broadcast="numpy" />
5481 <input>
5482 <port id="0" precision="FP32">
5483 <dim>384</dim>
5484 <dim>1536</dim>
5485 </port>
5486 <port id="1" precision="FP32">
5487 <dim>384</dim>
5488 <dim>1</dim>
5489 </port>
5490 </input>
5491 <output>
5492 <port id="2" precision="FP32">
5493 <dim>384</dim>
5494 <dim>1536</dim>
5495 </port>
5496 </output>
5497 </layer>
5498 <layer id="402" name="__module.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5499 <data transpose_a="false" transpose_b="true" />
5500 <input>
5501 <port id="0" precision="FP32">
5502 <dim>-1</dim>
5503 <dim>-1</dim>
5504 <dim>1536</dim>
5505 </port>
5506 <port id="1" precision="FP32">
5507 <dim>384</dim>
5508 <dim>1536</dim>
5509 </port>
5510 </input>
5511 <output>
5512 <port id="2" precision="FP32">
5513 <dim>-1</dim>
5514 <dim>-1</dim>
5515 <dim>384</dim>
5516 </port>
5517 </output>
5518 </layer>
5519 <layer id="403" name="Constant_6261" type="Const" version="opset1">
5520 <data element_type="f32" shape="1, 1, 384" offset="17486864" size="1536" />
5521 <output>
5522 <port id="0" precision="FP32">
5523 <dim>1</dim>
5524 <dim>1</dim>
5525 <dim>384</dim>
5526 </port>
5527 </output>
5528 </layer>
5529 <layer id="404" name="__module.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
5530 <data auto_broadcast="numpy" />
5531 <input>
5532 <port id="0" precision="FP32">
5533 <dim>-1</dim>
5534 <dim>-1</dim>
5535 <dim>384</dim>
5536 </port>
5537 <port id="1" precision="FP32">
5538 <dim>1</dim>
5539 <dim>1</dim>
5540 <dim>384</dim>
5541 </port>
5542 </input>
5543 <output>
5544 <port id="2" precision="FP32" names="323,input.13">
5545 <dim>-1</dim>
5546 <dim>-1</dim>
5547 <dim>384</dim>
5548 </port>
5549 </output>
5550 </layer>
5551 <layer id="405" name="__module.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
5552 <data auto_broadcast="numpy" />
5553 <input>
5554 <port id="0" precision="FP32">
5555 <dim>-1</dim>
5556 <dim>-1</dim>
5557 <dim>384</dim>
5558 </port>
5559 <port id="1" precision="FP32">
5560 <dim>-1</dim>
5561 <dim>-1</dim>
5562 <dim>384</dim>
5563 </port>
5564 </input>
5565 <output>
5566 <port id="2" precision="FP32" names="325">
5567 <dim>-1</dim>
5568 <dim>-1</dim>
5569 <dim>384</dim>
5570 </port>
5571 </output>
5572 </layer>
5573 <layer id="406" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5574 <data element_type="i32" shape="1" offset="12046084" size="4" />
5575 <output>
5576 <port id="0" precision="I32">
5577 <dim>1</dim>
5578 </port>
5579 </output>
5580 </layer>
5581 <layer id="407" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5582 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5583 <input>
5584 <port id="0" precision="FP32">
5585 <dim>-1</dim>
5586 <dim>-1</dim>
5587 <dim>384</dim>
5588 </port>
5589 <port id="1" precision="I32">
5590 <dim>1</dim>
5591 </port>
5592 </input>
5593 <output>
5594 <port id="2" precision="FP32">
5595 <dim>-1</dim>
5596 <dim>-1</dim>
5597 <dim>384</dim>
5598 </port>
5599 </output>
5600 </layer>
5601 <layer id="408" name="Constant_6262" type="Const" version="opset1">
5602 <data element_type="f32" shape="1, 1, 384" offset="17488400" size="1536" />
5603 <output>
5604 <port id="0" precision="FP32">
5605 <dim>1</dim>
5606 <dim>1</dim>
5607 <dim>384</dim>
5608 </port>
5609 </output>
5610 </layer>
5611 <layer id="409" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5612 <data auto_broadcast="numpy" />
5613 <input>
5614 <port id="0" precision="FP32">
5615 <dim>-1</dim>
5616 <dim>-1</dim>
5617 <dim>384</dim>
5618 </port>
5619 <port id="1" precision="FP32">
5620 <dim>1</dim>
5621 <dim>1</dim>
5622 <dim>384</dim>
5623 </port>
5624 </input>
5625 <output>
5626 <port id="2" precision="FP32">
5627 <dim>-1</dim>
5628 <dim>-1</dim>
5629 <dim>384</dim>
5630 </port>
5631 </output>
5632 </layer>
5633 <layer id="410" name="Constant_6263" type="Const" version="opset1">
5634 <data element_type="f32" shape="1, 1, 384" offset="17489936" size="1536" />
5635 <output>
5636 <port id="0" precision="FP32">
5637 <dim>1</dim>
5638 <dim>1</dim>
5639 <dim>384</dim>
5640 </port>
5641 </output>
5642 </layer>
5643 <layer id="411" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5644 <data auto_broadcast="numpy" />
5645 <input>
5646 <port id="0" precision="FP32">
5647 <dim>-1</dim>
5648 <dim>-1</dim>
5649 <dim>384</dim>
5650 </port>
5651 <port id="1" precision="FP32">
5652 <dim>1</dim>
5653 <dim>1</dim>
5654 <dim>384</dim>
5655 </port>
5656 </input>
5657 <output>
5658 <port id="2" precision="FP32" names="329,hidden_states.19">
5659 <dim>-1</dim>
5660 <dim>-1</dim>
5661 <dim>384</dim>
5662 </port>
5663 </output>
5664 </layer>
5665 <layer id="412" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5666 <data element_type="f32" shape="1, 1, 384" offset="17491472" size="1536" />
5667 <output>
5668 <port id="0" precision="FP32">
5669 <dim>1</dim>
5670 <dim>1</dim>
5671 <dim>384</dim>
5672 </port>
5673 </output>
5674 </layer>
5675 <layer id="413" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5676 <data auto_broadcast="numpy" />
5677 <input>
5678 <port id="0" precision="FP32">
5679 <dim>-1</dim>
5680 <dim>-1</dim>
5681 <dim>384</dim>
5682 </port>
5683 <port id="1" precision="FP32">
5684 <dim>1</dim>
5685 <dim>1</dim>
5686 <dim>384</dim>
5687 </port>
5688 </input>
5689 <output>
5690 <port id="2" precision="FP32">
5691 <dim>-1</dim>
5692 <dim>-1</dim>
5693 <dim>384</dim>
5694 </port>
5695 </output>
5696 </layer>
5697 <layer id="414" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5698 <data element_type="f32" shape="" offset="17493008" size="4" />
5699 <output>
5700 <port id="0" precision="FP32" />
5701 </output>
5702 </layer>
5703 <layer id="415" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5704 <data element_type="f32" shape="" offset="17493012" size="4" />
5705 <output>
5706 <port id="0" precision="FP32" />
5707 </output>
5708 </layer>
5709 <layer id="416" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5710 <data element_type="f32" shape="" offset="17493008" size="4" />
5711 <output>
5712 <port id="0" precision="FP32" />
5713 </output>
5714 </layer>
5715 <layer id="417" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5716 <data element_type="f32" shape="" offset="17493012" size="4" />
5717 <output>
5718 <port id="0" precision="FP32" />
5719 </output>
5720 </layer>
5721 <layer id="418" name="__module.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5722 <data levels="256" auto_broadcast="numpy" />
5723 <input>
5724 <port id="0" precision="FP32">
5725 <dim>-1</dim>
5726 <dim>-1</dim>
5727 <dim>384</dim>
5728 </port>
5729 <port id="1" precision="FP32" />
5730 <port id="2" precision="FP32" />
5731 <port id="3" precision="FP32" />
5732 <port id="4" precision="FP32" />
5733 </input>
5734 <output>
5735 <port id="5" precision="FP32">
5736 <dim>-1</dim>
5737 <dim>-1</dim>
5738 <dim>384</dim>
5739 </port>
5740 </output>
5741 </layer>
5742 <layer id="419" name="Constant_27536206" type="Const" version="opset1">
5743 <data element_type="i8" shape="384, 384" offset="17493016" size="147456" />
5744 <output>
5745 <port id="0" precision="I8">
5746 <dim>384</dim>
5747 <dim>384</dim>
5748 </port>
5749 </output>
5750 </layer>
5751 <layer id="420" name="Convert_27536207" type="Convert" version="opset1">
5752 <data destination_type="f32" />
5753 <input>
5754 <port id="0" precision="I8">
5755 <dim>384</dim>
5756 <dim>384</dim>
5757 </port>
5758 </input>
5759 <output>
5760 <port id="1" precision="FP32">
5761 <dim>384</dim>
5762 <dim>384</dim>
5763 </port>
5764 </output>
5765 </layer>
5766 <layer id="421" name="Constant_27536208" type="Const" version="opset1">
5767 <data element_type="f32" shape="384, 1" offset="17640472" size="1536" />
5768 <output>
5769 <port id="0" precision="FP32">
5770 <dim>384</dim>
5771 <dim>1</dim>
5772 </port>
5773 </output>
5774 </layer>
5775 <layer id="422" name="__module.encoder.layer.3.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5776 <data auto_broadcast="numpy" />
5777 <input>
5778 <port id="0" precision="FP32">
5779 <dim>384</dim>
5780 <dim>384</dim>
5781 </port>
5782 <port id="1" precision="FP32">
5783 <dim>384</dim>
5784 <dim>1</dim>
5785 </port>
5786 </input>
5787 <output>
5788 <port id="2" precision="FP32">
5789 <dim>384</dim>
5790 <dim>384</dim>
5791 </port>
5792 </output>
5793 </layer>
5794 <layer id="423" name="__module.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
5795 <data transpose_a="false" transpose_b="true" />
5796 <input>
5797 <port id="0" precision="FP32">
5798 <dim>-1</dim>
5799 <dim>-1</dim>
5800 <dim>384</dim>
5801 </port>
5802 <port id="1" precision="FP32">
5803 <dim>384</dim>
5804 <dim>384</dim>
5805 </port>
5806 </input>
5807 <output>
5808 <port id="2" precision="FP32">
5809 <dim>-1</dim>
5810 <dim>-1</dim>
5811 <dim>384</dim>
5812 </port>
5813 </output>
5814 </layer>
5815 <layer id="424" name="Constant_6264" type="Const" version="opset1">
5816 <data element_type="f32" shape="1, 1, 384" offset="17642008" size="1536" />
5817 <output>
5818 <port id="0" precision="FP32">
5819 <dim>1</dim>
5820 <dim>1</dim>
5821 <dim>384</dim>
5822 </port>
5823 </output>
5824 </layer>
5825 <layer id="425" name="__module.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
5826 <data auto_broadcast="numpy" />
5827 <input>
5828 <port id="0" precision="FP32">
5829 <dim>-1</dim>
5830 <dim>-1</dim>
5831 <dim>384</dim>
5832 </port>
5833 <port id="1" precision="FP32">
5834 <dim>1</dim>
5835 <dim>1</dim>
5836 <dim>384</dim>
5837 </port>
5838 </input>
5839 <output>
5840 <port id="2" precision="FP32" names="342,x.37">
5841 <dim>-1</dim>
5842 <dim>-1</dim>
5843 <dim>384</dim>
5844 </port>
5845 </output>
5846 </layer>
5847 <layer id="426" name="__module.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5848 <data element_type="f32" shape="" offset="17643544" size="4" />
5849 <output>
5850 <port id="0" precision="FP32" />
5851 </output>
5852 </layer>
5853 <layer id="427" name="__module.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
5854 <data element_type="f32" shape="" offset="17643548" size="4" />
5855 <output>
5856 <port id="0" precision="FP32" />
5857 </output>
5858 </layer>
5859 <layer id="428" name="__module.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
5860 <data element_type="f32" shape="" offset="17643544" size="4" />
5861 <output>
5862 <port id="0" precision="FP32" />
5863 </output>
5864 </layer>
5865 <layer id="429" name="__module.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
5866 <data element_type="f32" shape="" offset="17643548" size="4" />
5867 <output>
5868 <port id="0" precision="FP32" />
5869 </output>
5870 </layer>
5871 <layer id="430" name="__module.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
5872 <data levels="256" auto_broadcast="numpy" />
5873 <input>
5874 <port id="0" precision="FP32">
5875 <dim>-1</dim>
5876 <dim>-1</dim>
5877 <dim>384</dim>
5878 </port>
5879 <port id="1" precision="FP32" />
5880 <port id="2" precision="FP32" />
5881 <port id="3" precision="FP32" />
5882 <port id="4" precision="FP32" />
5883 </input>
5884 <output>
5885 <port id="5" precision="FP32">
5886 <dim>-1</dim>
5887 <dim>-1</dim>
5888 <dim>384</dim>
5889 </port>
5890 </output>
5891 </layer>
5892 <layer id="431" name="__module.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
5893 <data element_type="i64" shape="4" offset="12201240" size="32" />
5894 <output>
5895 <port id="0" precision="I64">
5896 <dim>4</dim>
5897 </port>
5898 </output>
5899 </layer>
5900 <layer id="432" name="__module.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
5901 <data special_zero="true" />
5902 <input>
5903 <port id="0" precision="FP32">
5904 <dim>-1</dim>
5905 <dim>-1</dim>
5906 <dim>384</dim>
5907 </port>
5908 <port id="1" precision="I64">
5909 <dim>4</dim>
5910 </port>
5911 </input>
5912 <output>
5913 <port id="2" precision="FP32" names="346,x.39">
5914 <dim>-1</dim>
5915 <dim>-1</dim>
5916 <dim>12</dim>
5917 <dim>32</dim>
5918 </port>
5919 </output>
5920 </layer>
5921 <layer id="433" name="Constant_931" type="Const" version="opset1">
5922 <data element_type="i64" shape="4" offset="12201272" size="32" />
5923 <output>
5924 <port id="0" precision="I64" names="347">
5925 <dim>4</dim>
5926 </port>
5927 </output>
5928 </layer>
5929 <layer id="434" name="__module.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
5930 <input>
5931 <port id="0" precision="FP32">
5932 <dim>-1</dim>
5933 <dim>-1</dim>
5934 <dim>12</dim>
5935 <dim>32</dim>
5936 </port>
5937 <port id="1" precision="I64">
5938 <dim>4</dim>
5939 </port>
5940 </input>
5941 <output>
5942 <port id="2" precision="FP32" names="348">
5943 <dim>-1</dim>
5944 <dim>12</dim>
5945 <dim>-1</dim>
5946 <dim>32</dim>
5947 </port>
5948 </output>
5949 </layer>
5950 <layer id="435" name="Constant_27536210" type="Const" version="opset1">
5951 <data element_type="i8" shape="384, 384" offset="17643552" size="147456" />
5952 <output>
5953 <port id="0" precision="I8">
5954 <dim>384</dim>
5955 <dim>384</dim>
5956 </port>
5957 </output>
5958 </layer>
5959 <layer id="436" name="Convert_27536211" type="Convert" version="opset1">
5960 <data destination_type="f32" />
5961 <input>
5962 <port id="0" precision="I8">
5963 <dim>384</dim>
5964 <dim>384</dim>
5965 </port>
5966 </input>
5967 <output>
5968 <port id="1" precision="FP32">
5969 <dim>384</dim>
5970 <dim>384</dim>
5971 </port>
5972 </output>
5973 </layer>
5974 <layer id="437" name="Constant_27536212" type="Const" version="opset1">
5975 <data element_type="f32" shape="384, 1" offset="17791008" size="1536" />
5976 <output>
5977 <port id="0" precision="FP32">
5978 <dim>384</dim>
5979 <dim>1</dim>
5980 </port>
5981 </output>
5982 </layer>
5983 <layer id="438" name="__module.encoder.layer.3.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5984 <data auto_broadcast="numpy" />
5985 <input>
5986 <port id="0" precision="FP32">
5987 <dim>384</dim>
5988 <dim>384</dim>
5989 </port>
5990 <port id="1" precision="FP32">
5991 <dim>384</dim>
5992 <dim>1</dim>
5993 </port>
5994 </input>
5995 <output>
5996 <port id="2" precision="FP32">
5997 <dim>384</dim>
5998 <dim>384</dim>
5999 </port>
6000 </output>
6001 </layer>
6002 <layer id="439" name="__module.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
6003 <data transpose_a="false" transpose_b="true" />
6004 <input>
6005 <port id="0" precision="FP32">
6006 <dim>-1</dim>
6007 <dim>-1</dim>
6008 <dim>384</dim>
6009 </port>
6010 <port id="1" precision="FP32">
6011 <dim>384</dim>
6012 <dim>384</dim>
6013 </port>
6014 </input>
6015 <output>
6016 <port id="2" precision="FP32">
6017 <dim>-1</dim>
6018 <dim>-1</dim>
6019 <dim>384</dim>
6020 </port>
6021 </output>
6022 </layer>
6023 <layer id="440" name="Constant_6265" type="Const" version="opset1">
6024 <data element_type="f32" shape="1, 1, 384" offset="17792544" size="1536" />
6025 <output>
6026 <port id="0" precision="FP32">
6027 <dim>1</dim>
6028 <dim>1</dim>
6029 <dim>384</dim>
6030 </port>
6031 </output>
6032 </layer>
6033 <layer id="441" name="__module.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
6034 <data auto_broadcast="numpy" />
6035 <input>
6036 <port id="0" precision="FP32">
6037 <dim>-1</dim>
6038 <dim>-1</dim>
6039 <dim>384</dim>
6040 </port>
6041 <port id="1" precision="FP32">
6042 <dim>1</dim>
6043 <dim>1</dim>
6044 <dim>384</dim>
6045 </port>
6046 </input>
6047 <output>
6048 <port id="2" precision="FP32" names="351,x.41">
6049 <dim>-1</dim>
6050 <dim>-1</dim>
6051 <dim>384</dim>
6052 </port>
6053 </output>
6054 </layer>
6055 <layer id="442" name="__module.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
6056 <data element_type="f32" shape="" offset="17794080" size="4" />
6057 <output>
6058 <port id="0" precision="FP32" />
6059 </output>
6060 </layer>
6061 <layer id="443" name="__module.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
6062 <data element_type="f32" shape="" offset="17794084" size="4" />
6063 <output>
6064 <port id="0" precision="FP32" />
6065 </output>
6066 </layer>
6067 <layer id="444" name="__module.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
6068 <data element_type="f32" shape="" offset="17794080" size="4" />
6069 <output>
6070 <port id="0" precision="FP32" />
6071 </output>
6072 </layer>
6073 <layer id="445" name="__module.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
6074 <data element_type="f32" shape="" offset="17794084" size="4" />
6075 <output>
6076 <port id="0" precision="FP32" />
6077 </output>
6078 </layer>
6079 <layer id="446" name="__module.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
6080 <data levels="256" auto_broadcast="numpy" />
6081 <input>
6082 <port id="0" precision="FP32">
6083 <dim>-1</dim>
6084 <dim>-1</dim>
6085 <dim>384</dim>
6086 </port>
6087 <port id="1" precision="FP32" />
6088 <port id="2" precision="FP32" />
6089 <port id="3" precision="FP32" />
6090 <port id="4" precision="FP32" />
6091 </input>
6092 <output>
6093 <port id="5" precision="FP32">
6094 <dim>-1</dim>
6095 <dim>-1</dim>
6096 <dim>384</dim>
6097 </port>
6098 </output>
6099 </layer>
6100 <layer id="447" name="__module.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
6101 <data element_type="i64" shape="4" offset="12201240" size="32" />
6102 <output>
6103 <port id="0" precision="I64">
6104 <dim>4</dim>
6105 </port>
6106 </output>
6107 </layer>
6108 <layer id="448" name="__module.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
6109 <data special_zero="true" />
6110 <input>
6111 <port id="0" precision="FP32">
6112 <dim>-1</dim>
6113 <dim>-1</dim>
6114 <dim>384</dim>
6115 </port>
6116 <port id="1" precision="I64">
6117 <dim>4</dim>
6118 </port>
6119 </input>
6120 <output>
6121 <port id="2" precision="FP32" names="355,x.43">
6122 <dim>-1</dim>
6123 <dim>-1</dim>
6124 <dim>12</dim>
6125 <dim>32</dim>
6126 </port>
6127 </output>
6128 </layer>
6129 <layer id="449" name="Constant_954" type="Const" version="opset1">
6130 <data element_type="i64" shape="4" offset="12201272" size="32" />
6131 <output>
6132 <port id="0" precision="I64" names="356">
6133 <dim>4</dim>
6134 </port>
6135 </output>
6136 </layer>
6137 <layer id="450" name="__module.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
6138 <input>
6139 <port id="0" precision="FP32">
6140 <dim>-1</dim>
6141 <dim>-1</dim>
6142 <dim>12</dim>
6143 <dim>32</dim>
6144 </port>
6145 <port id="1" precision="I64">
6146 <dim>4</dim>
6147 </port>
6148 </input>
6149 <output>
6150 <port id="2" precision="FP32" names="357">
6151 <dim>-1</dim>
6152 <dim>12</dim>
6153 <dim>-1</dim>
6154 <dim>32</dim>
6155 </port>
6156 </output>
6157 </layer>
6158 <layer id="451" name="Constant_27536214" type="Const" version="opset1">
6159 <data element_type="i8" shape="384, 384" offset="17794088" size="147456" />
6160 <output>
6161 <port id="0" precision="I8">
6162 <dim>384</dim>
6163 <dim>384</dim>
6164 </port>
6165 </output>
6166 </layer>
6167 <layer id="452" name="Convert_27536215" type="Convert" version="opset1">
6168 <data destination_type="f32" />
6169 <input>
6170 <port id="0" precision="I8">
6171 <dim>384</dim>
6172 <dim>384</dim>
6173 </port>
6174 </input>
6175 <output>
6176 <port id="1" precision="FP32">
6177 <dim>384</dim>
6178 <dim>384</dim>
6179 </port>
6180 </output>
6181 </layer>
6182 <layer id="453" name="Constant_27536216" type="Const" version="opset1">
6183 <data element_type="f32" shape="384, 1" offset="17941544" size="1536" />
6184 <output>
6185 <port id="0" precision="FP32">
6186 <dim>384</dim>
6187 <dim>1</dim>
6188 </port>
6189 </output>
6190 </layer>
6191 <layer id="454" name="__module.encoder.layer.3.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6192 <data auto_broadcast="numpy" />
6193 <input>
6194 <port id="0" precision="FP32">
6195 <dim>384</dim>
6196 <dim>384</dim>
6197 </port>
6198 <port id="1" precision="FP32">
6199 <dim>384</dim>
6200 <dim>1</dim>
6201 </port>
6202 </input>
6203 <output>
6204 <port id="2" precision="FP32">
6205 <dim>384</dim>
6206 <dim>384</dim>
6207 </port>
6208 </output>
6209 </layer>
6210 <layer id="455" name="__module.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
6211 <data transpose_a="false" transpose_b="true" />
6212 <input>
6213 <port id="0" precision="FP32">
6214 <dim>-1</dim>
6215 <dim>-1</dim>
6216 <dim>384</dim>
6217 </port>
6218 <port id="1" precision="FP32">
6219 <dim>384</dim>
6220 <dim>384</dim>
6221 </port>
6222 </input>
6223 <output>
6224 <port id="2" precision="FP32">
6225 <dim>-1</dim>
6226 <dim>-1</dim>
6227 <dim>384</dim>
6228 </port>
6229 </output>
6230 </layer>
6231 <layer id="456" name="Constant_6266" type="Const" version="opset1">
6232 <data element_type="f32" shape="1, 1, 384" offset="17943080" size="1536" />
6233 <output>
6234 <port id="0" precision="FP32">
6235 <dim>1</dim>
6236 <dim>1</dim>
6237 <dim>384</dim>
6238 </port>
6239 </output>
6240 </layer>
6241 <layer id="457" name="__module.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
6242 <data auto_broadcast="numpy" />
6243 <input>
6244 <port id="0" precision="FP32">
6245 <dim>-1</dim>
6246 <dim>-1</dim>
6247 <dim>384</dim>
6248 </port>
6249 <port id="1" precision="FP32">
6250 <dim>1</dim>
6251 <dim>1</dim>
6252 <dim>384</dim>
6253 </port>
6254 </input>
6255 <output>
6256 <port id="2" precision="FP32" names="360,x.45">
6257 <dim>-1</dim>
6258 <dim>-1</dim>
6259 <dim>384</dim>
6260 </port>
6261 </output>
6262 </layer>
6263 <layer id="458" name="__module.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
6264 <data element_type="i64" shape="4" offset="12201240" size="32" />
6265 <output>
6266 <port id="0" precision="I64">
6267 <dim>4</dim>
6268 </port>
6269 </output>
6270 </layer>
6271 <layer id="459" name="__module.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
6272 <data special_zero="true" />
6273 <input>
6274 <port id="0" precision="FP32">
6275 <dim>-1</dim>
6276 <dim>-1</dim>
6277 <dim>384</dim>
6278 </port>
6279 <port id="1" precision="I64">
6280 <dim>4</dim>
6281 </port>
6282 </input>
6283 <output>
6284 <port id="2" precision="FP32" names="364,x.47">
6285 <dim>-1</dim>
6286 <dim>-1</dim>
6287 <dim>12</dim>
6288 <dim>32</dim>
6289 </port>
6290 </output>
6291 </layer>
6292 <layer id="460" name="Constant_977" type="Const" version="opset1">
6293 <data element_type="i64" shape="4" offset="12201272" size="32" />
6294 <output>
6295 <port id="0" precision="I64" names="365">
6296 <dim>4</dim>
6297 </port>
6298 </output>
6299 </layer>
6300 <layer id="461" name="__module.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
6301 <input>
6302 <port id="0" precision="FP32">
6303 <dim>-1</dim>
6304 <dim>-1</dim>
6305 <dim>12</dim>
6306 <dim>32</dim>
6307 </port>
6308 <port id="1" precision="I64">
6309 <dim>4</dim>
6310 </port>
6311 </input>
6312 <output>
6313 <port id="2" precision="FP32" names="366">
6314 <dim>-1</dim>
6315 <dim>12</dim>
6316 <dim>-1</dim>
6317 <dim>32</dim>
6318 </port>
6319 </output>
6320 </layer>
6321 <layer id="462" name="__module.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
6322 <data causal="false" />
6323 <input>
6324 <port id="0" precision="FP32">
6325 <dim>-1</dim>
6326 <dim>12</dim>
6327 <dim>-1</dim>
6328 <dim>32</dim>
6329 </port>
6330 <port id="1" precision="FP32">
6331 <dim>-1</dim>
6332 <dim>12</dim>
6333 <dim>-1</dim>
6334 <dim>32</dim>
6335 </port>
6336 <port id="2" precision="FP32">
6337 <dim>-1</dim>
6338 <dim>12</dim>
6339 <dim>-1</dim>
6340 <dim>32</dim>
6341 </port>
6342 <port id="3" precision="FP32">
6343 <dim>-1</dim>
6344 <dim>1</dim>
6345 <dim>-1</dim>
6346 <dim>-1</dim>
6347 </port>
6348 </input>
6349 <output>
6350 <port id="4" precision="FP32" names="367,attn_output.13">
6351 <dim>-1</dim>
6352 <dim>12</dim>
6353 <dim>-1</dim>
6354 <dim>32</dim>
6355 </port>
6356 </output>
6357 </layer>
6358 <layer id="463" name="__module.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
6359 <data element_type="i32" shape="4" offset="12502384" size="16" />
6360 <output>
6361 <port id="0" precision="I32">
6362 <dim>4</dim>
6363 </port>
6364 </output>
6365 </layer>
6366 <layer id="464" name="__module.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
6367 <input>
6368 <port id="0" precision="FP32">
6369 <dim>-1</dim>
6370 <dim>12</dim>
6371 <dim>-1</dim>
6372 <dim>32</dim>
6373 </port>
6374 <port id="1" precision="I32">
6375 <dim>4</dim>
6376 </port>
6377 </input>
6378 <output>
6379 <port id="2" precision="FP32" names="368,attn_output.15">
6380 <dim>-1</dim>
6381 <dim>-1</dim>
6382 <dim>12</dim>
6383 <dim>32</dim>
6384 </port>
6385 </output>
6386 </layer>
6387 <layer id="465" name="__module.encoder.layer.3.attention.self/aten::size/ShapeOf_6" type="ShapeOf" version="opset3">
6388 <data output_type="i64" />
6389 <input>
6390 <port id="0" precision="FP32">
6391 <dim>-1</dim>
6392 <dim>-1</dim>
6393 <dim>384</dim>
6394 </port>
6395 </input>
6396 <output>
6397 <port id="1" precision="I64">
6398 <dim>3</dim>
6399 </port>
6400 </output>
6401 </layer>
6402 <layer id="466" name="Constant_5797" type="Const" version="opset1">
6403 <data element_type="i64" shape="2" offset="12502400" size="16" />
6404 <output>
6405 <port id="0" precision="I64">
6406 <dim>2</dim>
6407 </port>
6408 </output>
6409 </layer>
6410 <layer id="467" name="Constant_5798" type="Const" version="opset1">
6411 <data element_type="i64" shape="" offset="12046068" size="8" />
6412 <output>
6413 <port id="0" precision="I64" />
6414 </output>
6415 </layer>
6416 <layer id="468" name="Gather_5799" type="Gather" version="opset8">
6417 <data batch_dims="0" />
6418 <input>
6419 <port id="0" precision="I64">
6420 <dim>3</dim>
6421 </port>
6422 <port id="1" precision="I64">
6423 <dim>2</dim>
6424 </port>
6425 <port id="2" precision="I64" />
6426 </input>
6427 <output>
6428 <port id="3" precision="I64">
6429 <dim>2</dim>
6430 </port>
6431 </output>
6432 </layer>
6433 <layer id="469" name="__module.encoder.layer.3.attention.self/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
6434 <data axis="0" />
6435 <input>
6436 <port id="0" precision="I64">
6437 <dim>2</dim>
6438 </port>
6439 <port id="1" precision="I64">
6440 <dim>1</dim>
6441 </port>
6442 </input>
6443 <output>
6444 <port id="2" precision="I64" names="369">
6445 <dim>3</dim>
6446 </port>
6447 </output>
6448 </layer>
6449 <layer id="470" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
6450 <data special_zero="false" />
6451 <input>
6452 <port id="0" precision="FP32">
6453 <dim>-1</dim>
6454 <dim>-1</dim>
6455 <dim>12</dim>
6456 <dim>32</dim>
6457 </port>
6458 <port id="1" precision="I64">
6459 <dim>3</dim>
6460 </port>
6461 </input>
6462 <output>
6463 <port id="2" precision="FP32" names="370">
6464 <dim>-1</dim>
6465 <dim>-1</dim>
6466 <dim>384</dim>
6467 </port>
6468 </output>
6469 </layer>
6470 <layer id="471" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6471 <data element_type="f32" shape="1, 1, 384" offset="17944616" size="1536" />
6472 <output>
6473 <port id="0" precision="FP32">
6474 <dim>1</dim>
6475 <dim>1</dim>
6476 <dim>384</dim>
6477 </port>
6478 </output>
6479 </layer>
6480 <layer id="472" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6481 <data auto_broadcast="numpy" />
6482 <input>
6483 <port id="0" precision="FP32">
6484 <dim>-1</dim>
6485 <dim>-1</dim>
6486 <dim>384</dim>
6487 </port>
6488 <port id="1" precision="FP32">
6489 <dim>1</dim>
6490 <dim>1</dim>
6491 <dim>384</dim>
6492 </port>
6493 </input>
6494 <output>
6495 <port id="2" precision="FP32">
6496 <dim>-1</dim>
6497 <dim>-1</dim>
6498 <dim>384</dim>
6499 </port>
6500 </output>
6501 </layer>
6502 <layer id="473" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6503 <data element_type="f32" shape="" offset="17946152" size="4" />
6504 <output>
6505 <port id="0" precision="FP32" />
6506 </output>
6507 </layer>
6508 <layer id="474" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6509 <data element_type="f32" shape="" offset="17946156" size="4" />
6510 <output>
6511 <port id="0" precision="FP32" />
6512 </output>
6513 </layer>
6514 <layer id="475" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6515 <data element_type="f32" shape="" offset="17946152" size="4" />
6516 <output>
6517 <port id="0" precision="FP32" />
6518 </output>
6519 </layer>
6520 <layer id="476" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6521 <data element_type="f32" shape="" offset="17946156" size="4" />
6522 <output>
6523 <port id="0" precision="FP32" />
6524 </output>
6525 </layer>
6526 <layer id="477" name="__module.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6527 <data levels="256" auto_broadcast="numpy" />
6528 <input>
6529 <port id="0" precision="FP32">
6530 <dim>-1</dim>
6531 <dim>-1</dim>
6532 <dim>384</dim>
6533 </port>
6534 <port id="1" precision="FP32" />
6535 <port id="2" precision="FP32" />
6536 <port id="3" precision="FP32" />
6537 <port id="4" precision="FP32" />
6538 </input>
6539 <output>
6540 <port id="5" precision="FP32">
6541 <dim>-1</dim>
6542 <dim>-1</dim>
6543 <dim>384</dim>
6544 </port>
6545 </output>
6546 </layer>
6547 <layer id="478" name="Constant_27536218" type="Const" version="opset1">
6548 <data element_type="i8" shape="384, 384" offset="17946160" size="147456" />
6549 <output>
6550 <port id="0" precision="I8">
6551 <dim>384</dim>
6552 <dim>384</dim>
6553 </port>
6554 </output>
6555 </layer>
6556 <layer id="479" name="Convert_27536219" type="Convert" version="opset1">
6557 <data destination_type="f32" />
6558 <input>
6559 <port id="0" precision="I8">
6560 <dim>384</dim>
6561 <dim>384</dim>
6562 </port>
6563 </input>
6564 <output>
6565 <port id="1" precision="FP32">
6566 <dim>384</dim>
6567 <dim>384</dim>
6568 </port>
6569 </output>
6570 </layer>
6571 <layer id="480" name="Constant_27536220" type="Const" version="opset1">
6572 <data element_type="f32" shape="384, 1" offset="18093616" size="1536" />
6573 <output>
6574 <port id="0" precision="FP32">
6575 <dim>384</dim>
6576 <dim>1</dim>
6577 </port>
6578 </output>
6579 </layer>
6580 <layer id="481" name="__module.encoder.layer.3.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6581 <data auto_broadcast="numpy" />
6582 <input>
6583 <port id="0" precision="FP32">
6584 <dim>384</dim>
6585 <dim>384</dim>
6586 </port>
6587 <port id="1" precision="FP32">
6588 <dim>384</dim>
6589 <dim>1</dim>
6590 </port>
6591 </input>
6592 <output>
6593 <port id="2" precision="FP32">
6594 <dim>384</dim>
6595 <dim>384</dim>
6596 </port>
6597 </output>
6598 </layer>
6599 <layer id="482" name="__module.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6600 <data transpose_a="false" transpose_b="true" />
6601 <input>
6602 <port id="0" precision="FP32">
6603 <dim>-1</dim>
6604 <dim>-1</dim>
6605 <dim>384</dim>
6606 </port>
6607 <port id="1" precision="FP32">
6608 <dim>384</dim>
6609 <dim>384</dim>
6610 </port>
6611 </input>
6612 <output>
6613 <port id="2" precision="FP32">
6614 <dim>-1</dim>
6615 <dim>-1</dim>
6616 <dim>384</dim>
6617 </port>
6618 </output>
6619 </layer>
6620 <layer id="483" name="Constant_6267" type="Const" version="opset1">
6621 <data element_type="f32" shape="1, 1, 384" offset="18095152" size="1536" />
6622 <output>
6623 <port id="0" precision="FP32">
6624 <dim>1</dim>
6625 <dim>1</dim>
6626 <dim>384</dim>
6627 </port>
6628 </output>
6629 </layer>
6630 <layer id="484" name="__module.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
6631 <data auto_broadcast="numpy" />
6632 <input>
6633 <port id="0" precision="FP32">
6634 <dim>-1</dim>
6635 <dim>-1</dim>
6636 <dim>384</dim>
6637 </port>
6638 <port id="1" precision="FP32">
6639 <dim>1</dim>
6640 <dim>1</dim>
6641 <dim>384</dim>
6642 </port>
6643 </input>
6644 <output>
6645 <port id="2" precision="FP32" names="376,input.15">
6646 <dim>-1</dim>
6647 <dim>-1</dim>
6648 <dim>384</dim>
6649 </port>
6650 </output>
6651 </layer>
6652 <layer id="485" name="__module.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
6653 <data auto_broadcast="numpy" />
6654 <input>
6655 <port id="0" precision="FP32">
6656 <dim>-1</dim>
6657 <dim>-1</dim>
6658 <dim>384</dim>
6659 </port>
6660 <port id="1" precision="FP32">
6661 <dim>-1</dim>
6662 <dim>-1</dim>
6663 <dim>384</dim>
6664 </port>
6665 </input>
6666 <output>
6667 <port id="2" precision="FP32" names="378">
6668 <dim>-1</dim>
6669 <dim>-1</dim>
6670 <dim>384</dim>
6671 </port>
6672 </output>
6673 </layer>
6674 <layer id="486" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6675 <data element_type="i32" shape="1" offset="12046084" size="4" />
6676 <output>
6677 <port id="0" precision="I32">
6678 <dim>1</dim>
6679 </port>
6680 </output>
6681 </layer>
6682 <layer id="487" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6683 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6684 <input>
6685 <port id="0" precision="FP32">
6686 <dim>-1</dim>
6687 <dim>-1</dim>
6688 <dim>384</dim>
6689 </port>
6690 <port id="1" precision="I32">
6691 <dim>1</dim>
6692 </port>
6693 </input>
6694 <output>
6695 <port id="2" precision="FP32">
6696 <dim>-1</dim>
6697 <dim>-1</dim>
6698 <dim>384</dim>
6699 </port>
6700 </output>
6701 </layer>
6702 <layer id="488" name="Constant_6268" type="Const" version="opset1">
6703 <data element_type="f32" shape="1, 1, 384" offset="18096688" size="1536" />
6704 <output>
6705 <port id="0" precision="FP32">
6706 <dim>1</dim>
6707 <dim>1</dim>
6708 <dim>384</dim>
6709 </port>
6710 </output>
6711 </layer>
6712 <layer id="489" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6713 <data auto_broadcast="numpy" />
6714 <input>
6715 <port id="0" precision="FP32">
6716 <dim>-1</dim>
6717 <dim>-1</dim>
6718 <dim>384</dim>
6719 </port>
6720 <port id="1" precision="FP32">
6721 <dim>1</dim>
6722 <dim>1</dim>
6723 <dim>384</dim>
6724 </port>
6725 </input>
6726 <output>
6727 <port id="2" precision="FP32">
6728 <dim>-1</dim>
6729 <dim>-1</dim>
6730 <dim>384</dim>
6731 </port>
6732 </output>
6733 </layer>
6734 <layer id="490" name="Constant_6269" type="Const" version="opset1">
6735 <data element_type="f32" shape="1, 1, 384" offset="18098224" size="1536" />
6736 <output>
6737 <port id="0" precision="FP32">
6738 <dim>1</dim>
6739 <dim>1</dim>
6740 <dim>384</dim>
6741 </port>
6742 </output>
6743 </layer>
6744 <layer id="491" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6745 <data auto_broadcast="numpy" />
6746 <input>
6747 <port id="0" precision="FP32">
6748 <dim>-1</dim>
6749 <dim>-1</dim>
6750 <dim>384</dim>
6751 </port>
6752 <port id="1" precision="FP32">
6753 <dim>1</dim>
6754 <dim>1</dim>
6755 <dim>384</dim>
6756 </port>
6757 </input>
6758 <output>
6759 <port id="2" precision="FP32" names="382,input_tensor.7">
6760 <dim>-1</dim>
6761 <dim>-1</dim>
6762 <dim>384</dim>
6763 </port>
6764 </output>
6765 </layer>
6766 <layer id="492" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6767 <data element_type="f32" shape="1, 1, 384" offset="18099760" size="1536" />
6768 <output>
6769 <port id="0" precision="FP32">
6770 <dim>1</dim>
6771 <dim>1</dim>
6772 <dim>384</dim>
6773 </port>
6774 </output>
6775 </layer>
6776 <layer id="493" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6777 <data auto_broadcast="numpy" />
6778 <input>
6779 <port id="0" precision="FP32">
6780 <dim>-1</dim>
6781 <dim>-1</dim>
6782 <dim>384</dim>
6783 </port>
6784 <port id="1" precision="FP32">
6785 <dim>1</dim>
6786 <dim>1</dim>
6787 <dim>384</dim>
6788 </port>
6789 </input>
6790 <output>
6791 <port id="2" precision="FP32">
6792 <dim>-1</dim>
6793 <dim>-1</dim>
6794 <dim>384</dim>
6795 </port>
6796 </output>
6797 </layer>
6798 <layer id="494" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6799 <data element_type="f32" shape="" offset="18101296" size="4" />
6800 <output>
6801 <port id="0" precision="FP32" />
6802 </output>
6803 </layer>
6804 <layer id="495" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6805 <data element_type="f32" shape="" offset="18101300" size="4" />
6806 <output>
6807 <port id="0" precision="FP32" />
6808 </output>
6809 </layer>
6810 <layer id="496" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6811 <data element_type="f32" shape="" offset="18101296" size="4" />
6812 <output>
6813 <port id="0" precision="FP32" />
6814 </output>
6815 </layer>
6816 <layer id="497" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6817 <data element_type="f32" shape="" offset="18101300" size="4" />
6818 <output>
6819 <port id="0" precision="FP32" />
6820 </output>
6821 </layer>
6822 <layer id="498" name="__module.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6823 <data levels="256" auto_broadcast="numpy" />
6824 <input>
6825 <port id="0" precision="FP32">
6826 <dim>-1</dim>
6827 <dim>-1</dim>
6828 <dim>384</dim>
6829 </port>
6830 <port id="1" precision="FP32" />
6831 <port id="2" precision="FP32" />
6832 <port id="3" precision="FP32" />
6833 <port id="4" precision="FP32" />
6834 </input>
6835 <output>
6836 <port id="5" precision="FP32">
6837 <dim>-1</dim>
6838 <dim>-1</dim>
6839 <dim>384</dim>
6840 </port>
6841 </output>
6842 </layer>
6843 <layer id="499" name="Constant_27536222" type="Const" version="opset1">
6844 <data element_type="i8" shape="1536, 384" offset="18101304" size="589824" />
6845 <output>
6846 <port id="0" precision="I8">
6847 <dim>1536</dim>
6848 <dim>384</dim>
6849 </port>
6850 </output>
6851 </layer>
6852 <layer id="500" name="Convert_27536223" type="Convert" version="opset1">
6853 <data destination_type="f32" />
6854 <input>
6855 <port id="0" precision="I8">
6856 <dim>1536</dim>
6857 <dim>384</dim>
6858 </port>
6859 </input>
6860 <output>
6861 <port id="1" precision="FP32">
6862 <dim>1536</dim>
6863 <dim>384</dim>
6864 </port>
6865 </output>
6866 </layer>
6867 <layer id="501" name="Constant_27536224" type="Const" version="opset1">
6868 <data element_type="f32" shape="1536, 1" offset="18691128" size="6144" />
6869 <output>
6870 <port id="0" precision="FP32">
6871 <dim>1536</dim>
6872 <dim>1</dim>
6873 </port>
6874 </output>
6875 </layer>
6876 <layer id="502" name="__module.encoder.layer.3.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6877 <data auto_broadcast="numpy" />
6878 <input>
6879 <port id="0" precision="FP32">
6880 <dim>1536</dim>
6881 <dim>384</dim>
6882 </port>
6883 <port id="1" precision="FP32">
6884 <dim>1536</dim>
6885 <dim>1</dim>
6886 </port>
6887 </input>
6888 <output>
6889 <port id="2" precision="FP32">
6890 <dim>1536</dim>
6891 <dim>384</dim>
6892 </port>
6893 </output>
6894 </layer>
6895 <layer id="503" name="__module.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6896 <data transpose_a="false" transpose_b="true" />
6897 <input>
6898 <port id="0" precision="FP32">
6899 <dim>-1</dim>
6900 <dim>-1</dim>
6901 <dim>384</dim>
6902 </port>
6903 <port id="1" precision="FP32">
6904 <dim>1536</dim>
6905 <dim>384</dim>
6906 </port>
6907 </input>
6908 <output>
6909 <port id="2" precision="FP32">
6910 <dim>-1</dim>
6911 <dim>-1</dim>
6912 <dim>1536</dim>
6913 </port>
6914 </output>
6915 </layer>
6916 <layer id="504" name="Constant_6270" type="Const" version="opset1">
6917 <data element_type="f32" shape="1, 1, 1536" offset="18697272" size="6144" />
6918 <output>
6919 <port id="0" precision="FP32">
6920 <dim>1</dim>
6921 <dim>1</dim>
6922 <dim>1536</dim>
6923 </port>
6924 </output>
6925 </layer>
6926 <layer id="505" name="__module.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
6927 <data auto_broadcast="numpy" />
6928 <input>
6929 <port id="0" precision="FP32">
6930 <dim>-1</dim>
6931 <dim>-1</dim>
6932 <dim>1536</dim>
6933 </port>
6934 <port id="1" precision="FP32">
6935 <dim>1</dim>
6936 <dim>1</dim>
6937 <dim>1536</dim>
6938 </port>
6939 </input>
6940 <output>
6941 <port id="2" precision="FP32" names="387">
6942 <dim>-1</dim>
6943 <dim>-1</dim>
6944 <dim>1536</dim>
6945 </port>
6946 </output>
6947 </layer>
6948 <layer id="506" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
6949 <data approximation_mode="ERF" />
6950 <input>
6951 <port id="0" precision="FP32">
6952 <dim>-1</dim>
6953 <dim>-1</dim>
6954 <dim>1536</dim>
6955 </port>
6956 </input>
6957 <output>
6958 <port id="1" precision="FP32" names="388">
6959 <dim>-1</dim>
6960 <dim>-1</dim>
6961 <dim>1536</dim>
6962 </port>
6963 </output>
6964 </layer>
6965 <layer id="507" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6966 <data element_type="f32" shape="1, 1, 1536" offset="18703416" size="6144" />
6967 <output>
6968 <port id="0" precision="FP32">
6969 <dim>1</dim>
6970 <dim>1</dim>
6971 <dim>1536</dim>
6972 </port>
6973 </output>
6974 </layer>
6975 <layer id="508" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6976 <data auto_broadcast="numpy" />
6977 <input>
6978 <port id="0" precision="FP32">
6979 <dim>-1</dim>
6980 <dim>-1</dim>
6981 <dim>1536</dim>
6982 </port>
6983 <port id="1" precision="FP32">
6984 <dim>1</dim>
6985 <dim>1</dim>
6986 <dim>1536</dim>
6987 </port>
6988 </input>
6989 <output>
6990 <port id="2" precision="FP32">
6991 <dim>-1</dim>
6992 <dim>-1</dim>
6993 <dim>1536</dim>
6994 </port>
6995 </output>
6996 </layer>
6997 <layer id="509" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6998 <data element_type="f32" shape="" offset="18709560" size="4" />
6999 <output>
7000 <port id="0" precision="FP32" />
7001 </output>
7002 </layer>
7003 <layer id="510" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7004 <data element_type="f32" shape="" offset="18709564" size="4" />
7005 <output>
7006 <port id="0" precision="FP32" />
7007 </output>
7008 </layer>
7009 <layer id="511" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7010 <data element_type="f32" shape="" offset="18709560" size="4" />
7011 <output>
7012 <port id="0" precision="FP32" />
7013 </output>
7014 </layer>
7015 <layer id="512" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7016 <data element_type="f32" shape="" offset="18709564" size="4" />
7017 <output>
7018 <port id="0" precision="FP32" />
7019 </output>
7020 </layer>
7021 <layer id="513" name="__module.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7022 <data levels="256" auto_broadcast="numpy" />
7023 <input>
7024 <port id="0" precision="FP32">
7025 <dim>-1</dim>
7026 <dim>-1</dim>
7027 <dim>1536</dim>
7028 </port>
7029 <port id="1" precision="FP32" />
7030 <port id="2" precision="FP32" />
7031 <port id="3" precision="FP32" />
7032 <port id="4" precision="FP32" />
7033 </input>
7034 <output>
7035 <port id="5" precision="FP32">
7036 <dim>-1</dim>
7037 <dim>-1</dim>
7038 <dim>1536</dim>
7039 </port>
7040 </output>
7041 </layer>
7042 <layer id="514" name="Constant_27536226" type="Const" version="opset1">
7043 <data element_type="i8" shape="384, 1536" offset="18709568" size="589824" />
7044 <output>
7045 <port id="0" precision="I8">
7046 <dim>384</dim>
7047 <dim>1536</dim>
7048 </port>
7049 </output>
7050 </layer>
7051 <layer id="515" name="Convert_27536227" type="Convert" version="opset1">
7052 <data destination_type="f32" />
7053 <input>
7054 <port id="0" precision="I8">
7055 <dim>384</dim>
7056 <dim>1536</dim>
7057 </port>
7058 </input>
7059 <output>
7060 <port id="1" precision="FP32">
7061 <dim>384</dim>
7062 <dim>1536</dim>
7063 </port>
7064 </output>
7065 </layer>
7066 <layer id="516" name="Constant_27536228" type="Const" version="opset1">
7067 <data element_type="f32" shape="384, 1" offset="19299392" size="1536" />
7068 <output>
7069 <port id="0" precision="FP32">
7070 <dim>384</dim>
7071 <dim>1</dim>
7072 </port>
7073 </output>
7074 </layer>
7075 <layer id="517" name="__module.encoder.layer.3.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7076 <data auto_broadcast="numpy" />
7077 <input>
7078 <port id="0" precision="FP32">
7079 <dim>384</dim>
7080 <dim>1536</dim>
7081 </port>
7082 <port id="1" precision="FP32">
7083 <dim>384</dim>
7084 <dim>1</dim>
7085 </port>
7086 </input>
7087 <output>
7088 <port id="2" precision="FP32">
7089 <dim>384</dim>
7090 <dim>1536</dim>
7091 </port>
7092 </output>
7093 </layer>
7094 <layer id="518" name="__module.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
7095 <data transpose_a="false" transpose_b="true" />
7096 <input>
7097 <port id="0" precision="FP32">
7098 <dim>-1</dim>
7099 <dim>-1</dim>
7100 <dim>1536</dim>
7101 </port>
7102 <port id="1" precision="FP32">
7103 <dim>384</dim>
7104 <dim>1536</dim>
7105 </port>
7106 </input>
7107 <output>
7108 <port id="2" precision="FP32">
7109 <dim>-1</dim>
7110 <dim>-1</dim>
7111 <dim>384</dim>
7112 </port>
7113 </output>
7114 </layer>
7115 <layer id="519" name="Constant_6271" type="Const" version="opset1">
7116 <data element_type="f32" shape="1, 1, 384" offset="19300928" size="1536" />
7117 <output>
7118 <port id="0" precision="FP32">
7119 <dim>1</dim>
7120 <dim>1</dim>
7121 <dim>384</dim>
7122 </port>
7123 </output>
7124 </layer>
7125 <layer id="520" name="__module.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
7126 <data auto_broadcast="numpy" />
7127 <input>
7128 <port id="0" precision="FP32">
7129 <dim>-1</dim>
7130 <dim>-1</dim>
7131 <dim>384</dim>
7132 </port>
7133 <port id="1" precision="FP32">
7134 <dim>1</dim>
7135 <dim>1</dim>
7136 <dim>384</dim>
7137 </port>
7138 </input>
7139 <output>
7140 <port id="2" precision="FP32" names="394,input.17">
7141 <dim>-1</dim>
7142 <dim>-1</dim>
7143 <dim>384</dim>
7144 </port>
7145 </output>
7146 </layer>
7147 <layer id="521" name="__module.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
7148 <data auto_broadcast="numpy" />
7149 <input>
7150 <port id="0" precision="FP32">
7151 <dim>-1</dim>
7152 <dim>-1</dim>
7153 <dim>384</dim>
7154 </port>
7155 <port id="1" precision="FP32">
7156 <dim>-1</dim>
7157 <dim>-1</dim>
7158 <dim>384</dim>
7159 </port>
7160 </input>
7161 <output>
7162 <port id="2" precision="FP32" names="396">
7163 <dim>-1</dim>
7164 <dim>-1</dim>
7165 <dim>384</dim>
7166 </port>
7167 </output>
7168 </layer>
7169 <layer id="522" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
7170 <data element_type="i32" shape="1" offset="12046084" size="4" />
7171 <output>
7172 <port id="0" precision="I32">
7173 <dim>1</dim>
7174 </port>
7175 </output>
7176 </layer>
7177 <layer id="523" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
7178 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
7179 <input>
7180 <port id="0" precision="FP32">
7181 <dim>-1</dim>
7182 <dim>-1</dim>
7183 <dim>384</dim>
7184 </port>
7185 <port id="1" precision="I32">
7186 <dim>1</dim>
7187 </port>
7188 </input>
7189 <output>
7190 <port id="2" precision="FP32">
7191 <dim>-1</dim>
7192 <dim>-1</dim>
7193 <dim>384</dim>
7194 </port>
7195 </output>
7196 </layer>
7197 <layer id="524" name="Constant_6272" type="Const" version="opset1">
7198 <data element_type="f32" shape="1, 1, 384" offset="19302464" size="1536" />
7199 <output>
7200 <port id="0" precision="FP32">
7201 <dim>1</dim>
7202 <dim>1</dim>
7203 <dim>384</dim>
7204 </port>
7205 </output>
7206 </layer>
7207 <layer id="525" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
7208 <data auto_broadcast="numpy" />
7209 <input>
7210 <port id="0" precision="FP32">
7211 <dim>-1</dim>
7212 <dim>-1</dim>
7213 <dim>384</dim>
7214 </port>
7215 <port id="1" precision="FP32">
7216 <dim>1</dim>
7217 <dim>1</dim>
7218 <dim>384</dim>
7219 </port>
7220 </input>
7221 <output>
7222 <port id="2" precision="FP32">
7223 <dim>-1</dim>
7224 <dim>-1</dim>
7225 <dim>384</dim>
7226 </port>
7227 </output>
7228 </layer>
7229 <layer id="526" name="Constant_6273" type="Const" version="opset1">
7230 <data element_type="f32" shape="1, 1, 384" offset="19304000" size="1536" />
7231 <output>
7232 <port id="0" precision="FP32">
7233 <dim>1</dim>
7234 <dim>1</dim>
7235 <dim>384</dim>
7236 </port>
7237 </output>
7238 </layer>
7239 <layer id="527" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
7240 <data auto_broadcast="numpy" />
7241 <input>
7242 <port id="0" precision="FP32">
7243 <dim>-1</dim>
7244 <dim>-1</dim>
7245 <dim>384</dim>
7246 </port>
7247 <port id="1" precision="FP32">
7248 <dim>1</dim>
7249 <dim>1</dim>
7250 <dim>384</dim>
7251 </port>
7252 </input>
7253 <output>
7254 <port id="2" precision="FP32" names="400,hidden_states.25">
7255 <dim>-1</dim>
7256 <dim>-1</dim>
7257 <dim>384</dim>
7258 </port>
7259 </output>
7260 </layer>
7261 <layer id="528" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7262 <data element_type="f32" shape="1, 1, 384" offset="19305536" size="1536" />
7263 <output>
7264 <port id="0" precision="FP32">
7265 <dim>1</dim>
7266 <dim>1</dim>
7267 <dim>384</dim>
7268 </port>
7269 </output>
7270 </layer>
7271 <layer id="529" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7272 <data auto_broadcast="numpy" />
7273 <input>
7274 <port id="0" precision="FP32">
7275 <dim>-1</dim>
7276 <dim>-1</dim>
7277 <dim>384</dim>
7278 </port>
7279 <port id="1" precision="FP32">
7280 <dim>1</dim>
7281 <dim>1</dim>
7282 <dim>384</dim>
7283 </port>
7284 </input>
7285 <output>
7286 <port id="2" precision="FP32">
7287 <dim>-1</dim>
7288 <dim>-1</dim>
7289 <dim>384</dim>
7290 </port>
7291 </output>
7292 </layer>
7293 <layer id="530" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7294 <data element_type="f32" shape="" offset="19307072" size="4" />
7295 <output>
7296 <port id="0" precision="FP32" />
7297 </output>
7298 </layer>
7299 <layer id="531" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7300 <data element_type="f32" shape="" offset="19307076" size="4" />
7301 <output>
7302 <port id="0" precision="FP32" />
7303 </output>
7304 </layer>
7305 <layer id="532" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7306 <data element_type="f32" shape="" offset="19307072" size="4" />
7307 <output>
7308 <port id="0" precision="FP32" />
7309 </output>
7310 </layer>
7311 <layer id="533" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7312 <data element_type="f32" shape="" offset="19307076" size="4" />
7313 <output>
7314 <port id="0" precision="FP32" />
7315 </output>
7316 </layer>
7317 <layer id="534" name="__module.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7318 <data levels="256" auto_broadcast="numpy" />
7319 <input>
7320 <port id="0" precision="FP32">
7321 <dim>-1</dim>
7322 <dim>-1</dim>
7323 <dim>384</dim>
7324 </port>
7325 <port id="1" precision="FP32" />
7326 <port id="2" precision="FP32" />
7327 <port id="3" precision="FP32" />
7328 <port id="4" precision="FP32" />
7329 </input>
7330 <output>
7331 <port id="5" precision="FP32">
7332 <dim>-1</dim>
7333 <dim>-1</dim>
7334 <dim>384</dim>
7335 </port>
7336 </output>
7337 </layer>
7338 <layer id="535" name="Constant_27536230" type="Const" version="opset1">
7339 <data element_type="i8" shape="384, 384" offset="19307080" size="147456" />
7340 <output>
7341 <port id="0" precision="I8">
7342 <dim>384</dim>
7343 <dim>384</dim>
7344 </port>
7345 </output>
7346 </layer>
7347 <layer id="536" name="Convert_27536231" type="Convert" version="opset1">
7348 <data destination_type="f32" />
7349 <input>
7350 <port id="0" precision="I8">
7351 <dim>384</dim>
7352 <dim>384</dim>
7353 </port>
7354 </input>
7355 <output>
7356 <port id="1" precision="FP32">
7357 <dim>384</dim>
7358 <dim>384</dim>
7359 </port>
7360 </output>
7361 </layer>
7362 <layer id="537" name="Constant_27536232" type="Const" version="opset1">
7363 <data element_type="f32" shape="384, 1" offset="19454536" size="1536" />
7364 <output>
7365 <port id="0" precision="FP32">
7366 <dim>384</dim>
7367 <dim>1</dim>
7368 </port>
7369 </output>
7370 </layer>
7371 <layer id="538" name="__module.encoder.layer.4.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7372 <data auto_broadcast="numpy" />
7373 <input>
7374 <port id="0" precision="FP32">
7375 <dim>384</dim>
7376 <dim>384</dim>
7377 </port>
7378 <port id="1" precision="FP32">
7379 <dim>384</dim>
7380 <dim>1</dim>
7381 </port>
7382 </input>
7383 <output>
7384 <port id="2" precision="FP32">
7385 <dim>384</dim>
7386 <dim>384</dim>
7387 </port>
7388 </output>
7389 </layer>
7390 <layer id="539" name="__module.encoder.layer.4.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
7391 <data transpose_a="false" transpose_b="true" />
7392 <input>
7393 <port id="0" precision="FP32">
7394 <dim>-1</dim>
7395 <dim>-1</dim>
7396 <dim>384</dim>
7397 </port>
7398 <port id="1" precision="FP32">
7399 <dim>384</dim>
7400 <dim>384</dim>
7401 </port>
7402 </input>
7403 <output>
7404 <port id="2" precision="FP32">
7405 <dim>-1</dim>
7406 <dim>-1</dim>
7407 <dim>384</dim>
7408 </port>
7409 </output>
7410 </layer>
7411 <layer id="540" name="Constant_6274" type="Const" version="opset1">
7412 <data element_type="f32" shape="1, 1, 384" offset="19456072" size="1536" />
7413 <output>
7414 <port id="0" precision="FP32">
7415 <dim>1</dim>
7416 <dim>1</dim>
7417 <dim>384</dim>
7418 </port>
7419 </output>
7420 </layer>
7421 <layer id="541" name="__module.encoder.layer.4.attention.self.query/aten::linear/Add" type="Add" version="opset1">
7422 <data auto_broadcast="numpy" />
7423 <input>
7424 <port id="0" precision="FP32">
7425 <dim>-1</dim>
7426 <dim>-1</dim>
7427 <dim>384</dim>
7428 </port>
7429 <port id="1" precision="FP32">
7430 <dim>1</dim>
7431 <dim>1</dim>
7432 <dim>384</dim>
7433 </port>
7434 </input>
7435 <output>
7436 <port id="2" precision="FP32" names="413,x.49">
7437 <dim>-1</dim>
7438 <dim>-1</dim>
7439 <dim>384</dim>
7440 </port>
7441 </output>
7442 </layer>
7443 <layer id="542" name="__module.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
7444 <data element_type="f32" shape="" offset="19457608" size="4" />
7445 <output>
7446 <port id="0" precision="FP32" />
7447 </output>
7448 </layer>
7449 <layer id="543" name="__module.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
7450 <data element_type="f32" shape="" offset="19457612" size="4" />
7451 <output>
7452 <port id="0" precision="FP32" />
7453 </output>
7454 </layer>
7455 <layer id="544" name="__module.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
7456 <data element_type="f32" shape="" offset="19457608" size="4" />
7457 <output>
7458 <port id="0" precision="FP32" />
7459 </output>
7460 </layer>
7461 <layer id="545" name="__module.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
7462 <data element_type="f32" shape="" offset="19457612" size="4" />
7463 <output>
7464 <port id="0" precision="FP32" />
7465 </output>
7466 </layer>
7467 <layer id="546" name="__module.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
7468 <data levels="256" auto_broadcast="numpy" />
7469 <input>
7470 <port id="0" precision="FP32">
7471 <dim>-1</dim>
7472 <dim>-1</dim>
7473 <dim>384</dim>
7474 </port>
7475 <port id="1" precision="FP32" />
7476 <port id="2" precision="FP32" />
7477 <port id="3" precision="FP32" />
7478 <port id="4" precision="FP32" />
7479 </input>
7480 <output>
7481 <port id="5" precision="FP32">
7482 <dim>-1</dim>
7483 <dim>-1</dim>
7484 <dim>384</dim>
7485 </port>
7486 </output>
7487 </layer>
7488 <layer id="547" name="__module.encoder.layer.4.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
7489 <data element_type="i64" shape="4" offset="12201240" size="32" />
7490 <output>
7491 <port id="0" precision="I64">
7492 <dim>4</dim>
7493 </port>
7494 </output>
7495 </layer>
7496 <layer id="548" name="__module.encoder.layer.4.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
7497 <data special_zero="true" />
7498 <input>
7499 <port id="0" precision="FP32">
7500 <dim>-1</dim>
7501 <dim>-1</dim>
7502 <dim>384</dim>
7503 </port>
7504 <port id="1" precision="I64">
7505 <dim>4</dim>
7506 </port>
7507 </input>
7508 <output>
7509 <port id="2" precision="FP32" names="417,x.51">
7510 <dim>-1</dim>
7511 <dim>-1</dim>
7512 <dim>12</dim>
7513 <dim>32</dim>
7514 </port>
7515 </output>
7516 </layer>
7517 <layer id="549" name="Constant_1157" type="Const" version="opset1">
7518 <data element_type="i64" shape="4" offset="12201272" size="32" />
7519 <output>
7520 <port id="0" precision="I64" names="418">
7521 <dim>4</dim>
7522 </port>
7523 </output>
7524 </layer>
7525 <layer id="550" name="__module.encoder.layer.4.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
7526 <input>
7527 <port id="0" precision="FP32">
7528 <dim>-1</dim>
7529 <dim>-1</dim>
7530 <dim>12</dim>
7531 <dim>32</dim>
7532 </port>
7533 <port id="1" precision="I64">
7534 <dim>4</dim>
7535 </port>
7536 </input>
7537 <output>
7538 <port id="2" precision="FP32" names="419">
7539 <dim>-1</dim>
7540 <dim>12</dim>
7541 <dim>-1</dim>
7542 <dim>32</dim>
7543 </port>
7544 </output>
7545 </layer>
7546 <layer id="551" name="Constant_27536234" type="Const" version="opset1">
7547 <data element_type="i8" shape="384, 384" offset="19457616" size="147456" />
7548 <output>
7549 <port id="0" precision="I8">
7550 <dim>384</dim>
7551 <dim>384</dim>
7552 </port>
7553 </output>
7554 </layer>
7555 <layer id="552" name="Convert_27536235" type="Convert" version="opset1">
7556 <data destination_type="f32" />
7557 <input>
7558 <port id="0" precision="I8">
7559 <dim>384</dim>
7560 <dim>384</dim>
7561 </port>
7562 </input>
7563 <output>
7564 <port id="1" precision="FP32">
7565 <dim>384</dim>
7566 <dim>384</dim>
7567 </port>
7568 </output>
7569 </layer>
7570 <layer id="553" name="Constant_27536236" type="Const" version="opset1">
7571 <data element_type="f32" shape="384, 1" offset="19605072" size="1536" />
7572 <output>
7573 <port id="0" precision="FP32">
7574 <dim>384</dim>
7575 <dim>1</dim>
7576 </port>
7577 </output>
7578 </layer>
7579 <layer id="554" name="__module.encoder.layer.4.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7580 <data auto_broadcast="numpy" />
7581 <input>
7582 <port id="0" precision="FP32">
7583 <dim>384</dim>
7584 <dim>384</dim>
7585 </port>
7586 <port id="1" precision="FP32">
7587 <dim>384</dim>
7588 <dim>1</dim>
7589 </port>
7590 </input>
7591 <output>
7592 <port id="2" precision="FP32">
7593 <dim>384</dim>
7594 <dim>384</dim>
7595 </port>
7596 </output>
7597 </layer>
7598 <layer id="555" name="__module.encoder.layer.4.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
7599 <data transpose_a="false" transpose_b="true" />
7600 <input>
7601 <port id="0" precision="FP32">
7602 <dim>-1</dim>
7603 <dim>-1</dim>
7604 <dim>384</dim>
7605 </port>
7606 <port id="1" precision="FP32">
7607 <dim>384</dim>
7608 <dim>384</dim>
7609 </port>
7610 </input>
7611 <output>
7612 <port id="2" precision="FP32">
7613 <dim>-1</dim>
7614 <dim>-1</dim>
7615 <dim>384</dim>
7616 </port>
7617 </output>
7618 </layer>
7619 <layer id="556" name="Constant_6275" type="Const" version="opset1">
7620 <data element_type="f32" shape="1, 1, 384" offset="19606608" size="1536" />
7621 <output>
7622 <port id="0" precision="FP32">
7623 <dim>1</dim>
7624 <dim>1</dim>
7625 <dim>384</dim>
7626 </port>
7627 </output>
7628 </layer>
7629 <layer id="557" name="__module.encoder.layer.4.attention.self.key/aten::linear/Add" type="Add" version="opset1">
7630 <data auto_broadcast="numpy" />
7631 <input>
7632 <port id="0" precision="FP32">
7633 <dim>-1</dim>
7634 <dim>-1</dim>
7635 <dim>384</dim>
7636 </port>
7637 <port id="1" precision="FP32">
7638 <dim>1</dim>
7639 <dim>1</dim>
7640 <dim>384</dim>
7641 </port>
7642 </input>
7643 <output>
7644 <port id="2" precision="FP32" names="422,x.53">
7645 <dim>-1</dim>
7646 <dim>-1</dim>
7647 <dim>384</dim>
7648 </port>
7649 </output>
7650 </layer>
7651 <layer id="558" name="__module.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
7652 <data element_type="f32" shape="" offset="19608144" size="4" />
7653 <output>
7654 <port id="0" precision="FP32" />
7655 </output>
7656 </layer>
7657 <layer id="559" name="__module.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
7658 <data element_type="f32" shape="" offset="19608148" size="4" />
7659 <output>
7660 <port id="0" precision="FP32" />
7661 </output>
7662 </layer>
7663 <layer id="560" name="__module.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
7664 <data element_type="f32" shape="" offset="19608144" size="4" />
7665 <output>
7666 <port id="0" precision="FP32" />
7667 </output>
7668 </layer>
7669 <layer id="561" name="__module.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
7670 <data element_type="f32" shape="" offset="19608148" size="4" />
7671 <output>
7672 <port id="0" precision="FP32" />
7673 </output>
7674 </layer>
7675 <layer id="562" name="__module.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
7676 <data levels="256" auto_broadcast="numpy" />
7677 <input>
7678 <port id="0" precision="FP32">
7679 <dim>-1</dim>
7680 <dim>-1</dim>
7681 <dim>384</dim>
7682 </port>
7683 <port id="1" precision="FP32" />
7684 <port id="2" precision="FP32" />
7685 <port id="3" precision="FP32" />
7686 <port id="4" precision="FP32" />
7687 </input>
7688 <output>
7689 <port id="5" precision="FP32">
7690 <dim>-1</dim>
7691 <dim>-1</dim>
7692 <dim>384</dim>
7693 </port>
7694 </output>
7695 </layer>
7696 <layer id="563" name="__module.encoder.layer.4.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
7697 <data element_type="i64" shape="4" offset="12201240" size="32" />
7698 <output>
7699 <port id="0" precision="I64">
7700 <dim>4</dim>
7701 </port>
7702 </output>
7703 </layer>
7704 <layer id="564" name="__module.encoder.layer.4.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
7705 <data special_zero="true" />
7706 <input>
7707 <port id="0" precision="FP32">
7708 <dim>-1</dim>
7709 <dim>-1</dim>
7710 <dim>384</dim>
7711 </port>
7712 <port id="1" precision="I64">
7713 <dim>4</dim>
7714 </port>
7715 </input>
7716 <output>
7717 <port id="2" precision="FP32" names="426,x.55">
7718 <dim>-1</dim>
7719 <dim>-1</dim>
7720 <dim>12</dim>
7721 <dim>32</dim>
7722 </port>
7723 </output>
7724 </layer>
7725 <layer id="565" name="Constant_1180" type="Const" version="opset1">
7726 <data element_type="i64" shape="4" offset="12201272" size="32" />
7727 <output>
7728 <port id="0" precision="I64" names="427">
7729 <dim>4</dim>
7730 </port>
7731 </output>
7732 </layer>
7733 <layer id="566" name="__module.encoder.layer.4.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
7734 <input>
7735 <port id="0" precision="FP32">
7736 <dim>-1</dim>
7737 <dim>-1</dim>
7738 <dim>12</dim>
7739 <dim>32</dim>
7740 </port>
7741 <port id="1" precision="I64">
7742 <dim>4</dim>
7743 </port>
7744 </input>
7745 <output>
7746 <port id="2" precision="FP32" names="428">
7747 <dim>-1</dim>
7748 <dim>12</dim>
7749 <dim>-1</dim>
7750 <dim>32</dim>
7751 </port>
7752 </output>
7753 </layer>
7754 <layer id="567" name="Constant_27536238" type="Const" version="opset1">
7755 <data element_type="i8" shape="384, 384" offset="19608152" size="147456" />
7756 <output>
7757 <port id="0" precision="I8">
7758 <dim>384</dim>
7759 <dim>384</dim>
7760 </port>
7761 </output>
7762 </layer>
7763 <layer id="568" name="Convert_27536239" type="Convert" version="opset1">
7764 <data destination_type="f32" />
7765 <input>
7766 <port id="0" precision="I8">
7767 <dim>384</dim>
7768 <dim>384</dim>
7769 </port>
7770 </input>
7771 <output>
7772 <port id="1" precision="FP32">
7773 <dim>384</dim>
7774 <dim>384</dim>
7775 </port>
7776 </output>
7777 </layer>
7778 <layer id="569" name="Constant_27536240" type="Const" version="opset1">
7779 <data element_type="f32" shape="384, 1" offset="19755608" size="1536" />
7780 <output>
7781 <port id="0" precision="FP32">
7782 <dim>384</dim>
7783 <dim>1</dim>
7784 </port>
7785 </output>
7786 </layer>
7787 <layer id="570" name="__module.encoder.layer.4.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7788 <data auto_broadcast="numpy" />
7789 <input>
7790 <port id="0" precision="FP32">
7791 <dim>384</dim>
7792 <dim>384</dim>
7793 </port>
7794 <port id="1" precision="FP32">
7795 <dim>384</dim>
7796 <dim>1</dim>
7797 </port>
7798 </input>
7799 <output>
7800 <port id="2" precision="FP32">
7801 <dim>384</dim>
7802 <dim>384</dim>
7803 </port>
7804 </output>
7805 </layer>
7806 <layer id="571" name="__module.encoder.layer.4.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
7807 <data transpose_a="false" transpose_b="true" />
7808 <input>
7809 <port id="0" precision="FP32">
7810 <dim>-1</dim>
7811 <dim>-1</dim>
7812 <dim>384</dim>
7813 </port>
7814 <port id="1" precision="FP32">
7815 <dim>384</dim>
7816 <dim>384</dim>
7817 </port>
7818 </input>
7819 <output>
7820 <port id="2" precision="FP32">
7821 <dim>-1</dim>
7822 <dim>-1</dim>
7823 <dim>384</dim>
7824 </port>
7825 </output>
7826 </layer>
7827 <layer id="572" name="Constant_6276" type="Const" version="opset1">
7828 <data element_type="f32" shape="1, 1, 384" offset="19757144" size="1536" />
7829 <output>
7830 <port id="0" precision="FP32">
7831 <dim>1</dim>
7832 <dim>1</dim>
7833 <dim>384</dim>
7834 </port>
7835 </output>
7836 </layer>
7837 <layer id="573" name="__module.encoder.layer.4.attention.self.value/aten::linear/Add" type="Add" version="opset1">
7838 <data auto_broadcast="numpy" />
7839 <input>
7840 <port id="0" precision="FP32">
7841 <dim>-1</dim>
7842 <dim>-1</dim>
7843 <dim>384</dim>
7844 </port>
7845 <port id="1" precision="FP32">
7846 <dim>1</dim>
7847 <dim>1</dim>
7848 <dim>384</dim>
7849 </port>
7850 </input>
7851 <output>
7852 <port id="2" precision="FP32" names="431,x.57">
7853 <dim>-1</dim>
7854 <dim>-1</dim>
7855 <dim>384</dim>
7856 </port>
7857 </output>
7858 </layer>
7859 <layer id="574" name="__module.encoder.layer.4.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
7860 <data element_type="i64" shape="4" offset="12201240" size="32" />
7861 <output>
7862 <port id="0" precision="I64">
7863 <dim>4</dim>
7864 </port>
7865 </output>
7866 </layer>
7867 <layer id="575" name="__module.encoder.layer.4.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
7868 <data special_zero="true" />
7869 <input>
7870 <port id="0" precision="FP32">
7871 <dim>-1</dim>
7872 <dim>-1</dim>
7873 <dim>384</dim>
7874 </port>
7875 <port id="1" precision="I64">
7876 <dim>4</dim>
7877 </port>
7878 </input>
7879 <output>
7880 <port id="2" precision="FP32" names="435,x.59">
7881 <dim>-1</dim>
7882 <dim>-1</dim>
7883 <dim>12</dim>
7884 <dim>32</dim>
7885 </port>
7886 </output>
7887 </layer>
7888 <layer id="576" name="Constant_1203" type="Const" version="opset1">
7889 <data element_type="i64" shape="4" offset="12201272" size="32" />
7890 <output>
7891 <port id="0" precision="I64" names="436">
7892 <dim>4</dim>
7893 </port>
7894 </output>
7895 </layer>
7896 <layer id="577" name="__module.encoder.layer.4.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
7897 <input>
7898 <port id="0" precision="FP32">
7899 <dim>-1</dim>
7900 <dim>-1</dim>
7901 <dim>12</dim>
7902 <dim>32</dim>
7903 </port>
7904 <port id="1" precision="I64">
7905 <dim>4</dim>
7906 </port>
7907 </input>
7908 <output>
7909 <port id="2" precision="FP32" names="437">
7910 <dim>-1</dim>
7911 <dim>12</dim>
7912 <dim>-1</dim>
7913 <dim>32</dim>
7914 </port>
7915 </output>
7916 </layer>
7917 <layer id="578" name="__module.encoder.layer.4.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
7918 <data causal="false" />
7919 <input>
7920 <port id="0" precision="FP32">
7921 <dim>-1</dim>
7922 <dim>12</dim>
7923 <dim>-1</dim>
7924 <dim>32</dim>
7925 </port>
7926 <port id="1" precision="FP32">
7927 <dim>-1</dim>
7928 <dim>12</dim>
7929 <dim>-1</dim>
7930 <dim>32</dim>
7931 </port>
7932 <port id="2" precision="FP32">
7933 <dim>-1</dim>
7934 <dim>12</dim>
7935 <dim>-1</dim>
7936 <dim>32</dim>
7937 </port>
7938 <port id="3" precision="FP32">
7939 <dim>-1</dim>
7940 <dim>1</dim>
7941 <dim>-1</dim>
7942 <dim>-1</dim>
7943 </port>
7944 </input>
7945 <output>
7946 <port id="4" precision="FP32" names="438,attn_output.17">
7947 <dim>-1</dim>
7948 <dim>12</dim>
7949 <dim>-1</dim>
7950 <dim>32</dim>
7951 </port>
7952 </output>
7953 </layer>
7954 <layer id="579" name="__module.encoder.layer.4.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
7955 <data element_type="i32" shape="4" offset="12502384" size="16" />
7956 <output>
7957 <port id="0" precision="I32">
7958 <dim>4</dim>
7959 </port>
7960 </output>
7961 </layer>
7962 <layer id="580" name="__module.encoder.layer.4.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
7963 <input>
7964 <port id="0" precision="FP32">
7965 <dim>-1</dim>
7966 <dim>12</dim>
7967 <dim>-1</dim>
7968 <dim>32</dim>
7969 </port>
7970 <port id="1" precision="I32">
7971 <dim>4</dim>
7972 </port>
7973 </input>
7974 <output>
7975 <port id="2" precision="FP32" names="439,attn_output.19">
7976 <dim>-1</dim>
7977 <dim>-1</dim>
7978 <dim>12</dim>
7979 <dim>32</dim>
7980 </port>
7981 </output>
7982 </layer>
7983 <layer id="581" name="__module.encoder.layer.4.attention.self/aten::size/ShapeOf_6" type="ShapeOf" version="opset3">
7984 <data output_type="i64" />
7985 <input>
7986 <port id="0" precision="FP32">
7987 <dim>-1</dim>
7988 <dim>-1</dim>
7989 <dim>384</dim>
7990 </port>
7991 </input>
7992 <output>
7993 <port id="1" precision="I64">
7994 <dim>3</dim>
7995 </port>
7996 </output>
7997 </layer>
7998 <layer id="582" name="Constant_5817" type="Const" version="opset1">
7999 <data element_type="i64" shape="2" offset="12502400" size="16" />
8000 <output>
8001 <port id="0" precision="I64">
8002 <dim>2</dim>
8003 </port>
8004 </output>
8005 </layer>
8006 <layer id="583" name="Constant_5818" type="Const" version="opset1">
8007 <data element_type="i64" shape="" offset="12046068" size="8" />
8008 <output>
8009 <port id="0" precision="I64" />
8010 </output>
8011 </layer>
8012 <layer id="584" name="Gather_5819" type="Gather" version="opset8">
8013 <data batch_dims="0" />
8014 <input>
8015 <port id="0" precision="I64">
8016 <dim>3</dim>
8017 </port>
8018 <port id="1" precision="I64">
8019 <dim>2</dim>
8020 </port>
8021 <port id="2" precision="I64" />
8022 </input>
8023 <output>
8024 <port id="3" precision="I64">
8025 <dim>2</dim>
8026 </port>
8027 </output>
8028 </layer>
8029 <layer id="585" name="__module.encoder.layer.4.attention.self/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
8030 <data axis="0" />
8031 <input>
8032 <port id="0" precision="I64">
8033 <dim>2</dim>
8034 </port>
8035 <port id="1" precision="I64">
8036 <dim>1</dim>
8037 </port>
8038 </input>
8039 <output>
8040 <port id="2" precision="I64" names="440">
8041 <dim>3</dim>
8042 </port>
8043 </output>
8044 </layer>
8045 <layer id="586" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
8046 <data special_zero="false" />
8047 <input>
8048 <port id="0" precision="FP32">
8049 <dim>-1</dim>
8050 <dim>-1</dim>
8051 <dim>12</dim>
8052 <dim>32</dim>
8053 </port>
8054 <port id="1" precision="I64">
8055 <dim>3</dim>
8056 </port>
8057 </input>
8058 <output>
8059 <port id="2" precision="FP32" names="441">
8060 <dim>-1</dim>
8061 <dim>-1</dim>
8062 <dim>384</dim>
8063 </port>
8064 </output>
8065 </layer>
8066 <layer id="587" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8067 <data element_type="f32" shape="1, 1, 384" offset="19758680" size="1536" />
8068 <output>
8069 <port id="0" precision="FP32">
8070 <dim>1</dim>
8071 <dim>1</dim>
8072 <dim>384</dim>
8073 </port>
8074 </output>
8075 </layer>
8076 <layer id="588" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8077 <data auto_broadcast="numpy" />
8078 <input>
8079 <port id="0" precision="FP32">
8080 <dim>-1</dim>
8081 <dim>-1</dim>
8082 <dim>384</dim>
8083 </port>
8084 <port id="1" precision="FP32">
8085 <dim>1</dim>
8086 <dim>1</dim>
8087 <dim>384</dim>
8088 </port>
8089 </input>
8090 <output>
8091 <port id="2" precision="FP32">
8092 <dim>-1</dim>
8093 <dim>-1</dim>
8094 <dim>384</dim>
8095 </port>
8096 </output>
8097 </layer>
8098 <layer id="589" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8099 <data element_type="f32" shape="" offset="19760216" size="4" />
8100 <output>
8101 <port id="0" precision="FP32" />
8102 </output>
8103 </layer>
8104 <layer id="590" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8105 <data element_type="f32" shape="" offset="19760220" size="4" />
8106 <output>
8107 <port id="0" precision="FP32" />
8108 </output>
8109 </layer>
8110 <layer id="591" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8111 <data element_type="f32" shape="" offset="19760216" size="4" />
8112 <output>
8113 <port id="0" precision="FP32" />
8114 </output>
8115 </layer>
8116 <layer id="592" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8117 <data element_type="f32" shape="" offset="19760220" size="4" />
8118 <output>
8119 <port id="0" precision="FP32" />
8120 </output>
8121 </layer>
8122 <layer id="593" name="__module.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8123 <data levels="256" auto_broadcast="numpy" />
8124 <input>
8125 <port id="0" precision="FP32">
8126 <dim>-1</dim>
8127 <dim>-1</dim>
8128 <dim>384</dim>
8129 </port>
8130 <port id="1" precision="FP32" />
8131 <port id="2" precision="FP32" />
8132 <port id="3" precision="FP32" />
8133 <port id="4" precision="FP32" />
8134 </input>
8135 <output>
8136 <port id="5" precision="FP32">
8137 <dim>-1</dim>
8138 <dim>-1</dim>
8139 <dim>384</dim>
8140 </port>
8141 </output>
8142 </layer>
8143 <layer id="594" name="Constant_27536242" type="Const" version="opset1">
8144 <data element_type="i8" shape="384, 384" offset="19760224" size="147456" />
8145 <output>
8146 <port id="0" precision="I8">
8147 <dim>384</dim>
8148 <dim>384</dim>
8149 </port>
8150 </output>
8151 </layer>
8152 <layer id="595" name="Convert_27536243" type="Convert" version="opset1">
8153 <data destination_type="f32" />
8154 <input>
8155 <port id="0" precision="I8">
8156 <dim>384</dim>
8157 <dim>384</dim>
8158 </port>
8159 </input>
8160 <output>
8161 <port id="1" precision="FP32">
8162 <dim>384</dim>
8163 <dim>384</dim>
8164 </port>
8165 </output>
8166 </layer>
8167 <layer id="596" name="Constant_27536244" type="Const" version="opset1">
8168 <data element_type="f32" shape="384, 1" offset="19907680" size="1536" />
8169 <output>
8170 <port id="0" precision="FP32">
8171 <dim>384</dim>
8172 <dim>1</dim>
8173 </port>
8174 </output>
8175 </layer>
8176 <layer id="597" name="__module.encoder.layer.4.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8177 <data auto_broadcast="numpy" />
8178 <input>
8179 <port id="0" precision="FP32">
8180 <dim>384</dim>
8181 <dim>384</dim>
8182 </port>
8183 <port id="1" precision="FP32">
8184 <dim>384</dim>
8185 <dim>1</dim>
8186 </port>
8187 </input>
8188 <output>
8189 <port id="2" precision="FP32">
8190 <dim>384</dim>
8191 <dim>384</dim>
8192 </port>
8193 </output>
8194 </layer>
8195 <layer id="598" name="__module.encoder.layer.4.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8196 <data transpose_a="false" transpose_b="true" />
8197 <input>
8198 <port id="0" precision="FP32">
8199 <dim>-1</dim>
8200 <dim>-1</dim>
8201 <dim>384</dim>
8202 </port>
8203 <port id="1" precision="FP32">
8204 <dim>384</dim>
8205 <dim>384</dim>
8206 </port>
8207 </input>
8208 <output>
8209 <port id="2" precision="FP32">
8210 <dim>-1</dim>
8211 <dim>-1</dim>
8212 <dim>384</dim>
8213 </port>
8214 </output>
8215 </layer>
8216 <layer id="599" name="Constant_6277" type="Const" version="opset1">
8217 <data element_type="f32" shape="1, 1, 384" offset="19909216" size="1536" />
8218 <output>
8219 <port id="0" precision="FP32">
8220 <dim>1</dim>
8221 <dim>1</dim>
8222 <dim>384</dim>
8223 </port>
8224 </output>
8225 </layer>
8226 <layer id="600" name="__module.encoder.layer.4.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
8227 <data auto_broadcast="numpy" />
8228 <input>
8229 <port id="0" precision="FP32">
8230 <dim>-1</dim>
8231 <dim>-1</dim>
8232 <dim>384</dim>
8233 </port>
8234 <port id="1" precision="FP32">
8235 <dim>1</dim>
8236 <dim>1</dim>
8237 <dim>384</dim>
8238 </port>
8239 </input>
8240 <output>
8241 <port id="2" precision="FP32" names="447,input.19">
8242 <dim>-1</dim>
8243 <dim>-1</dim>
8244 <dim>384</dim>
8245 </port>
8246 </output>
8247 </layer>
8248 <layer id="601" name="__module.encoder.layer.4.attention.output/aten::add/Add" type="Add" version="opset1">
8249 <data auto_broadcast="numpy" />
8250 <input>
8251 <port id="0" precision="FP32">
8252 <dim>-1</dim>
8253 <dim>-1</dim>
8254 <dim>384</dim>
8255 </port>
8256 <port id="1" precision="FP32">
8257 <dim>-1</dim>
8258 <dim>-1</dim>
8259 <dim>384</dim>
8260 </port>
8261 </input>
8262 <output>
8263 <port id="2" precision="FP32" names="449">
8264 <dim>-1</dim>
8265 <dim>-1</dim>
8266 <dim>384</dim>
8267 </port>
8268 </output>
8269 </layer>
8270 <layer id="602" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
8271 <data element_type="i32" shape="1" offset="12046084" size="4" />
8272 <output>
8273 <port id="0" precision="I32">
8274 <dim>1</dim>
8275 </port>
8276 </output>
8277 </layer>
8278 <layer id="603" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
8279 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
8280 <input>
8281 <port id="0" precision="FP32">
8282 <dim>-1</dim>
8283 <dim>-1</dim>
8284 <dim>384</dim>
8285 </port>
8286 <port id="1" precision="I32">
8287 <dim>1</dim>
8288 </port>
8289 </input>
8290 <output>
8291 <port id="2" precision="FP32">
8292 <dim>-1</dim>
8293 <dim>-1</dim>
8294 <dim>384</dim>
8295 </port>
8296 </output>
8297 </layer>
8298 <layer id="604" name="Constant_6278" type="Const" version="opset1">
8299 <data element_type="f32" shape="1, 1, 384" offset="19910752" size="1536" />
8300 <output>
8301 <port id="0" precision="FP32">
8302 <dim>1</dim>
8303 <dim>1</dim>
8304 <dim>384</dim>
8305 </port>
8306 </output>
8307 </layer>
8308 <layer id="605" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
8309 <data auto_broadcast="numpy" />
8310 <input>
8311 <port id="0" precision="FP32">
8312 <dim>-1</dim>
8313 <dim>-1</dim>
8314 <dim>384</dim>
8315 </port>
8316 <port id="1" precision="FP32">
8317 <dim>1</dim>
8318 <dim>1</dim>
8319 <dim>384</dim>
8320 </port>
8321 </input>
8322 <output>
8323 <port id="2" precision="FP32">
8324 <dim>-1</dim>
8325 <dim>-1</dim>
8326 <dim>384</dim>
8327 </port>
8328 </output>
8329 </layer>
8330 <layer id="606" name="Constant_6279" type="Const" version="opset1">
8331 <data element_type="f32" shape="1, 1, 384" offset="19912288" size="1536" />
8332 <output>
8333 <port id="0" precision="FP32">
8334 <dim>1</dim>
8335 <dim>1</dim>
8336 <dim>384</dim>
8337 </port>
8338 </output>
8339 </layer>
8340 <layer id="607" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
8341 <data auto_broadcast="numpy" />
8342 <input>
8343 <port id="0" precision="FP32">
8344 <dim>-1</dim>
8345 <dim>-1</dim>
8346 <dim>384</dim>
8347 </port>
8348 <port id="1" precision="FP32">
8349 <dim>1</dim>
8350 <dim>1</dim>
8351 <dim>384</dim>
8352 </port>
8353 </input>
8354 <output>
8355 <port id="2" precision="FP32" names="453,input_tensor.9">
8356 <dim>-1</dim>
8357 <dim>-1</dim>
8358 <dim>384</dim>
8359 </port>
8360 </output>
8361 </layer>
8362 <layer id="608" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8363 <data element_type="f32" shape="1, 1, 384" offset="19913824" size="1536" />
8364 <output>
8365 <port id="0" precision="FP32">
8366 <dim>1</dim>
8367 <dim>1</dim>
8368 <dim>384</dim>
8369 </port>
8370 </output>
8371 </layer>
8372 <layer id="609" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8373 <data auto_broadcast="numpy" />
8374 <input>
8375 <port id="0" precision="FP32">
8376 <dim>-1</dim>
8377 <dim>-1</dim>
8378 <dim>384</dim>
8379 </port>
8380 <port id="1" precision="FP32">
8381 <dim>1</dim>
8382 <dim>1</dim>
8383 <dim>384</dim>
8384 </port>
8385 </input>
8386 <output>
8387 <port id="2" precision="FP32">
8388 <dim>-1</dim>
8389 <dim>-1</dim>
8390 <dim>384</dim>
8391 </port>
8392 </output>
8393 </layer>
8394 <layer id="610" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8395 <data element_type="f32" shape="" offset="19915360" size="4" />
8396 <output>
8397 <port id="0" precision="FP32" />
8398 </output>
8399 </layer>
8400 <layer id="611" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8401 <data element_type="f32" shape="" offset="19915364" size="4" />
8402 <output>
8403 <port id="0" precision="FP32" />
8404 </output>
8405 </layer>
8406 <layer id="612" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8407 <data element_type="f32" shape="" offset="19915360" size="4" />
8408 <output>
8409 <port id="0" precision="FP32" />
8410 </output>
8411 </layer>
8412 <layer id="613" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8413 <data element_type="f32" shape="" offset="19915364" size="4" />
8414 <output>
8415 <port id="0" precision="FP32" />
8416 </output>
8417 </layer>
8418 <layer id="614" name="__module.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8419 <data levels="256" auto_broadcast="numpy" />
8420 <input>
8421 <port id="0" precision="FP32">
8422 <dim>-1</dim>
8423 <dim>-1</dim>
8424 <dim>384</dim>
8425 </port>
8426 <port id="1" precision="FP32" />
8427 <port id="2" precision="FP32" />
8428 <port id="3" precision="FP32" />
8429 <port id="4" precision="FP32" />
8430 </input>
8431 <output>
8432 <port id="5" precision="FP32">
8433 <dim>-1</dim>
8434 <dim>-1</dim>
8435 <dim>384</dim>
8436 </port>
8437 </output>
8438 </layer>
8439 <layer id="615" name="Constant_27536246" type="Const" version="opset1">
8440 <data element_type="i8" shape="1536, 384" offset="19915368" size="589824" />
8441 <output>
8442 <port id="0" precision="I8">
8443 <dim>1536</dim>
8444 <dim>384</dim>
8445 </port>
8446 </output>
8447 </layer>
8448 <layer id="616" name="Convert_27536247" type="Convert" version="opset1">
8449 <data destination_type="f32" />
8450 <input>
8451 <port id="0" precision="I8">
8452 <dim>1536</dim>
8453 <dim>384</dim>
8454 </port>
8455 </input>
8456 <output>
8457 <port id="1" precision="FP32">
8458 <dim>1536</dim>
8459 <dim>384</dim>
8460 </port>
8461 </output>
8462 </layer>
8463 <layer id="617" name="Constant_27536248" type="Const" version="opset1">
8464 <data element_type="f32" shape="1536, 1" offset="20505192" size="6144" />
8465 <output>
8466 <port id="0" precision="FP32">
8467 <dim>1536</dim>
8468 <dim>1</dim>
8469 </port>
8470 </output>
8471 </layer>
8472 <layer id="618" name="__module.encoder.layer.4.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8473 <data auto_broadcast="numpy" />
8474 <input>
8475 <port id="0" precision="FP32">
8476 <dim>1536</dim>
8477 <dim>384</dim>
8478 </port>
8479 <port id="1" precision="FP32">
8480 <dim>1536</dim>
8481 <dim>1</dim>
8482 </port>
8483 </input>
8484 <output>
8485 <port id="2" precision="FP32">
8486 <dim>1536</dim>
8487 <dim>384</dim>
8488 </port>
8489 </output>
8490 </layer>
8491 <layer id="619" name="__module.encoder.layer.4.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8492 <data transpose_a="false" transpose_b="true" />
8493 <input>
8494 <port id="0" precision="FP32">
8495 <dim>-1</dim>
8496 <dim>-1</dim>
8497 <dim>384</dim>
8498 </port>
8499 <port id="1" precision="FP32">
8500 <dim>1536</dim>
8501 <dim>384</dim>
8502 </port>
8503 </input>
8504 <output>
8505 <port id="2" precision="FP32">
8506 <dim>-1</dim>
8507 <dim>-1</dim>
8508 <dim>1536</dim>
8509 </port>
8510 </output>
8511 </layer>
8512 <layer id="620" name="Constant_6280" type="Const" version="opset1">
8513 <data element_type="f32" shape="1, 1, 1536" offset="20511336" size="6144" />
8514 <output>
8515 <port id="0" precision="FP32">
8516 <dim>1</dim>
8517 <dim>1</dim>
8518 <dim>1536</dim>
8519 </port>
8520 </output>
8521 </layer>
8522 <layer id="621" name="__module.encoder.layer.4.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
8523 <data auto_broadcast="numpy" />
8524 <input>
8525 <port id="0" precision="FP32">
8526 <dim>-1</dim>
8527 <dim>-1</dim>
8528 <dim>1536</dim>
8529 </port>
8530 <port id="1" precision="FP32">
8531 <dim>1</dim>
8532 <dim>1</dim>
8533 <dim>1536</dim>
8534 </port>
8535 </input>
8536 <output>
8537 <port id="2" precision="FP32" names="458">
8538 <dim>-1</dim>
8539 <dim>-1</dim>
8540 <dim>1536</dim>
8541 </port>
8542 </output>
8543 </layer>
8544 <layer id="622" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
8545 <data approximation_mode="ERF" />
8546 <input>
8547 <port id="0" precision="FP32">
8548 <dim>-1</dim>
8549 <dim>-1</dim>
8550 <dim>1536</dim>
8551 </port>
8552 </input>
8553 <output>
8554 <port id="1" precision="FP32" names="459">
8555 <dim>-1</dim>
8556 <dim>-1</dim>
8557 <dim>1536</dim>
8558 </port>
8559 </output>
8560 </layer>
8561 <layer id="623" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8562 <data element_type="f32" shape="1, 1, 1536" offset="20517480" size="6144" />
8563 <output>
8564 <port id="0" precision="FP32">
8565 <dim>1</dim>
8566 <dim>1</dim>
8567 <dim>1536</dim>
8568 </port>
8569 </output>
8570 </layer>
8571 <layer id="624" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8572 <data auto_broadcast="numpy" />
8573 <input>
8574 <port id="0" precision="FP32">
8575 <dim>-1</dim>
8576 <dim>-1</dim>
8577 <dim>1536</dim>
8578 </port>
8579 <port id="1" precision="FP32">
8580 <dim>1</dim>
8581 <dim>1</dim>
8582 <dim>1536</dim>
8583 </port>
8584 </input>
8585 <output>
8586 <port id="2" precision="FP32">
8587 <dim>-1</dim>
8588 <dim>-1</dim>
8589 <dim>1536</dim>
8590 </port>
8591 </output>
8592 </layer>
8593 <layer id="625" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8594 <data element_type="f32" shape="" offset="20523624" size="4" />
8595 <output>
8596 <port id="0" precision="FP32" />
8597 </output>
8598 </layer>
8599 <layer id="626" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8600 <data element_type="f32" shape="" offset="20523628" size="4" />
8601 <output>
8602 <port id="0" precision="FP32" />
8603 </output>
8604 </layer>
8605 <layer id="627" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8606 <data element_type="f32" shape="" offset="20523624" size="4" />
8607 <output>
8608 <port id="0" precision="FP32" />
8609 </output>
8610 </layer>
8611 <layer id="628" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8612 <data element_type="f32" shape="" offset="20523628" size="4" />
8613 <output>
8614 <port id="0" precision="FP32" />
8615 </output>
8616 </layer>
8617 <layer id="629" name="__module.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8618 <data levels="256" auto_broadcast="numpy" />
8619 <input>
8620 <port id="0" precision="FP32">
8621 <dim>-1</dim>
8622 <dim>-1</dim>
8623 <dim>1536</dim>
8624 </port>
8625 <port id="1" precision="FP32" />
8626 <port id="2" precision="FP32" />
8627 <port id="3" precision="FP32" />
8628 <port id="4" precision="FP32" />
8629 </input>
8630 <output>
8631 <port id="5" precision="FP32">
8632 <dim>-1</dim>
8633 <dim>-1</dim>
8634 <dim>1536</dim>
8635 </port>
8636 </output>
8637 </layer>
8638 <layer id="630" name="Constant_27536250" type="Const" version="opset1">
8639 <data element_type="i8" shape="384, 1536" offset="20523632" size="589824" />
8640 <output>
8641 <port id="0" precision="I8">
8642 <dim>384</dim>
8643 <dim>1536</dim>
8644 </port>
8645 </output>
8646 </layer>
8647 <layer id="631" name="Convert_27536251" type="Convert" version="opset1">
8648 <data destination_type="f32" />
8649 <input>
8650 <port id="0" precision="I8">
8651 <dim>384</dim>
8652 <dim>1536</dim>
8653 </port>
8654 </input>
8655 <output>
8656 <port id="1" precision="FP32">
8657 <dim>384</dim>
8658 <dim>1536</dim>
8659 </port>
8660 </output>
8661 </layer>
8662 <layer id="632" name="Constant_27536252" type="Const" version="opset1">
8663 <data element_type="f32" shape="384, 1" offset="21113456" size="1536" />
8664 <output>
8665 <port id="0" precision="FP32">
8666 <dim>384</dim>
8667 <dim>1</dim>
8668 </port>
8669 </output>
8670 </layer>
8671 <layer id="633" name="__module.encoder.layer.4.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8672 <data auto_broadcast="numpy" />
8673 <input>
8674 <port id="0" precision="FP32">
8675 <dim>384</dim>
8676 <dim>1536</dim>
8677 </port>
8678 <port id="1" precision="FP32">
8679 <dim>384</dim>
8680 <dim>1</dim>
8681 </port>
8682 </input>
8683 <output>
8684 <port id="2" precision="FP32">
8685 <dim>384</dim>
8686 <dim>1536</dim>
8687 </port>
8688 </output>
8689 </layer>
8690 <layer id="634" name="__module.encoder.layer.4.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8691 <data transpose_a="false" transpose_b="true" />
8692 <input>
8693 <port id="0" precision="FP32">
8694 <dim>-1</dim>
8695 <dim>-1</dim>
8696 <dim>1536</dim>
8697 </port>
8698 <port id="1" precision="FP32">
8699 <dim>384</dim>
8700 <dim>1536</dim>
8701 </port>
8702 </input>
8703 <output>
8704 <port id="2" precision="FP32">
8705 <dim>-1</dim>
8706 <dim>-1</dim>
8707 <dim>384</dim>
8708 </port>
8709 </output>
8710 </layer>
8711 <layer id="635" name="Constant_6281" type="Const" version="opset1">
8712 <data element_type="f32" shape="1, 1, 384" offset="21114992" size="1536" />
8713 <output>
8714 <port id="0" precision="FP32">
8715 <dim>1</dim>
8716 <dim>1</dim>
8717 <dim>384</dim>
8718 </port>
8719 </output>
8720 </layer>
8721 <layer id="636" name="__module.encoder.layer.4.output.dense/aten::linear/Add" type="Add" version="opset1">
8722 <data auto_broadcast="numpy" />
8723 <input>
8724 <port id="0" precision="FP32">
8725 <dim>-1</dim>
8726 <dim>-1</dim>
8727 <dim>384</dim>
8728 </port>
8729 <port id="1" precision="FP32">
8730 <dim>1</dim>
8731 <dim>1</dim>
8732 <dim>384</dim>
8733 </port>
8734 </input>
8735 <output>
8736 <port id="2" precision="FP32" names="465,input.21">
8737 <dim>-1</dim>
8738 <dim>-1</dim>
8739 <dim>384</dim>
8740 </port>
8741 </output>
8742 </layer>
8743 <layer id="637" name="__module.encoder.layer.4.output/aten::add/Add" type="Add" version="opset1">
8744 <data auto_broadcast="numpy" />
8745 <input>
8746 <port id="0" precision="FP32">
8747 <dim>-1</dim>
8748 <dim>-1</dim>
8749 <dim>384</dim>
8750 </port>
8751 <port id="1" precision="FP32">
8752 <dim>-1</dim>
8753 <dim>-1</dim>
8754 <dim>384</dim>
8755 </port>
8756 </input>
8757 <output>
8758 <port id="2" precision="FP32" names="467">
8759 <dim>-1</dim>
8760 <dim>-1</dim>
8761 <dim>384</dim>
8762 </port>
8763 </output>
8764 </layer>
8765 <layer id="638" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
8766 <data element_type="i32" shape="1" offset="12046084" size="4" />
8767 <output>
8768 <port id="0" precision="I32">
8769 <dim>1</dim>
8770 </port>
8771 </output>
8772 </layer>
8773 <layer id="639" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
8774 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
8775 <input>
8776 <port id="0" precision="FP32">
8777 <dim>-1</dim>
8778 <dim>-1</dim>
8779 <dim>384</dim>
8780 </port>
8781 <port id="1" precision="I32">
8782 <dim>1</dim>
8783 </port>
8784 </input>
8785 <output>
8786 <port id="2" precision="FP32">
8787 <dim>-1</dim>
8788 <dim>-1</dim>
8789 <dim>384</dim>
8790 </port>
8791 </output>
8792 </layer>
8793 <layer id="640" name="Constant_6282" type="Const" version="opset1">
8794 <data element_type="f32" shape="1, 1, 384" offset="21116528" size="1536" />
8795 <output>
8796 <port id="0" precision="FP32">
8797 <dim>1</dim>
8798 <dim>1</dim>
8799 <dim>384</dim>
8800 </port>
8801 </output>
8802 </layer>
8803 <layer id="641" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
8804 <data auto_broadcast="numpy" />
8805 <input>
8806 <port id="0" precision="FP32">
8807 <dim>-1</dim>
8808 <dim>-1</dim>
8809 <dim>384</dim>
8810 </port>
8811 <port id="1" precision="FP32">
8812 <dim>1</dim>
8813 <dim>1</dim>
8814 <dim>384</dim>
8815 </port>
8816 </input>
8817 <output>
8818 <port id="2" precision="FP32">
8819 <dim>-1</dim>
8820 <dim>-1</dim>
8821 <dim>384</dim>
8822 </port>
8823 </output>
8824 </layer>
8825 <layer id="642" name="Constant_6283" type="Const" version="opset1">
8826 <data element_type="f32" shape="1, 1, 384" offset="21118064" size="1536" />
8827 <output>
8828 <port id="0" precision="FP32">
8829 <dim>1</dim>
8830 <dim>1</dim>
8831 <dim>384</dim>
8832 </port>
8833 </output>
8834 </layer>
8835 <layer id="643" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
8836 <data auto_broadcast="numpy" />
8837 <input>
8838 <port id="0" precision="FP32">
8839 <dim>-1</dim>
8840 <dim>-1</dim>
8841 <dim>384</dim>
8842 </port>
8843 <port id="1" precision="FP32">
8844 <dim>1</dim>
8845 <dim>1</dim>
8846 <dim>384</dim>
8847 </port>
8848 </input>
8849 <output>
8850 <port id="2" precision="FP32" names="471,hidden_states.31">
8851 <dim>-1</dim>
8852 <dim>-1</dim>
8853 <dim>384</dim>
8854 </port>
8855 </output>
8856 </layer>
8857 <layer id="644" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8858 <data element_type="f32" shape="1, 1, 384" offset="21119600" size="1536" />
8859 <output>
8860 <port id="0" precision="FP32">
8861 <dim>1</dim>
8862 <dim>1</dim>
8863 <dim>384</dim>
8864 </port>
8865 </output>
8866 </layer>
8867 <layer id="645" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8868 <data auto_broadcast="numpy" />
8869 <input>
8870 <port id="0" precision="FP32">
8871 <dim>-1</dim>
8872 <dim>-1</dim>
8873 <dim>384</dim>
8874 </port>
8875 <port id="1" precision="FP32">
8876 <dim>1</dim>
8877 <dim>1</dim>
8878 <dim>384</dim>
8879 </port>
8880 </input>
8881 <output>
8882 <port id="2" precision="FP32">
8883 <dim>-1</dim>
8884 <dim>-1</dim>
8885 <dim>384</dim>
8886 </port>
8887 </output>
8888 </layer>
8889 <layer id="646" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8890 <data element_type="f32" shape="" offset="21121136" size="4" />
8891 <output>
8892 <port id="0" precision="FP32" />
8893 </output>
8894 </layer>
8895 <layer id="647" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8896 <data element_type="f32" shape="" offset="21121140" size="4" />
8897 <output>
8898 <port id="0" precision="FP32" />
8899 </output>
8900 </layer>
8901 <layer id="648" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8902 <data element_type="f32" shape="" offset="21121136" size="4" />
8903 <output>
8904 <port id="0" precision="FP32" />
8905 </output>
8906 </layer>
8907 <layer id="649" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8908 <data element_type="f32" shape="" offset="21121140" size="4" />
8909 <output>
8910 <port id="0" precision="FP32" />
8911 </output>
8912 </layer>
8913 <layer id="650" name="__module.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8914 <data levels="256" auto_broadcast="numpy" />
8915 <input>
8916 <port id="0" precision="FP32">
8917 <dim>-1</dim>
8918 <dim>-1</dim>
8919 <dim>384</dim>
8920 </port>
8921 <port id="1" precision="FP32" />
8922 <port id="2" precision="FP32" />
8923 <port id="3" precision="FP32" />
8924 <port id="4" precision="FP32" />
8925 </input>
8926 <output>
8927 <port id="5" precision="FP32">
8928 <dim>-1</dim>
8929 <dim>-1</dim>
8930 <dim>384</dim>
8931 </port>
8932 </output>
8933 </layer>
8934 <layer id="651" name="Constant_27536254" type="Const" version="opset1">
8935 <data element_type="i8" shape="384, 384" offset="21121144" size="147456" />
8936 <output>
8937 <port id="0" precision="I8">
8938 <dim>384</dim>
8939 <dim>384</dim>
8940 </port>
8941 </output>
8942 </layer>
8943 <layer id="652" name="Convert_27536255" type="Convert" version="opset1">
8944 <data destination_type="f32" />
8945 <input>
8946 <port id="0" precision="I8">
8947 <dim>384</dim>
8948 <dim>384</dim>
8949 </port>
8950 </input>
8951 <output>
8952 <port id="1" precision="FP32">
8953 <dim>384</dim>
8954 <dim>384</dim>
8955 </port>
8956 </output>
8957 </layer>
8958 <layer id="653" name="Constant_27536256" type="Const" version="opset1">
8959 <data element_type="f32" shape="384, 1" offset="21268600" size="1536" />
8960 <output>
8961 <port id="0" precision="FP32">
8962 <dim>384</dim>
8963 <dim>1</dim>
8964 </port>
8965 </output>
8966 </layer>
8967 <layer id="654" name="__module.encoder.layer.5.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8968 <data auto_broadcast="numpy" />
8969 <input>
8970 <port id="0" precision="FP32">
8971 <dim>384</dim>
8972 <dim>384</dim>
8973 </port>
8974 <port id="1" precision="FP32">
8975 <dim>384</dim>
8976 <dim>1</dim>
8977 </port>
8978 </input>
8979 <output>
8980 <port id="2" precision="FP32">
8981 <dim>384</dim>
8982 <dim>384</dim>
8983 </port>
8984 </output>
8985 </layer>
8986 <layer id="655" name="__module.encoder.layer.5.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
8987 <data transpose_a="false" transpose_b="true" />
8988 <input>
8989 <port id="0" precision="FP32">
8990 <dim>-1</dim>
8991 <dim>-1</dim>
8992 <dim>384</dim>
8993 </port>
8994 <port id="1" precision="FP32">
8995 <dim>384</dim>
8996 <dim>384</dim>
8997 </port>
8998 </input>
8999 <output>
9000 <port id="2" precision="FP32">
9001 <dim>-1</dim>
9002 <dim>-1</dim>
9003 <dim>384</dim>
9004 </port>
9005 </output>
9006 </layer>
9007 <layer id="656" name="Constant_6284" type="Const" version="opset1">
9008 <data element_type="f32" shape="1, 1, 384" offset="21270136" size="1536" />
9009 <output>
9010 <port id="0" precision="FP32">
9011 <dim>1</dim>
9012 <dim>1</dim>
9013 <dim>384</dim>
9014 </port>
9015 </output>
9016 </layer>
9017 <layer id="657" name="__module.encoder.layer.5.attention.self.query/aten::linear/Add" type="Add" version="opset1">
9018 <data auto_broadcast="numpy" />
9019 <input>
9020 <port id="0" precision="FP32">
9021 <dim>-1</dim>
9022 <dim>-1</dim>
9023 <dim>384</dim>
9024 </port>
9025 <port id="1" precision="FP32">
9026 <dim>1</dim>
9027 <dim>1</dim>
9028 <dim>384</dim>
9029 </port>
9030 </input>
9031 <output>
9032 <port id="2" precision="FP32" names="484,x.61">
9033 <dim>-1</dim>
9034 <dim>-1</dim>
9035 <dim>384</dim>
9036 </port>
9037 </output>
9038 </layer>
9039 <layer id="658" name="__module.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
9040 <data element_type="f32" shape="" offset="21271672" size="4" />
9041 <output>
9042 <port id="0" precision="FP32" />
9043 </output>
9044 </layer>
9045 <layer id="659" name="__module.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
9046 <data element_type="f32" shape="" offset="21271676" size="4" />
9047 <output>
9048 <port id="0" precision="FP32" />
9049 </output>
9050 </layer>
9051 <layer id="660" name="__module.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
9052 <data element_type="f32" shape="" offset="21271672" size="4" />
9053 <output>
9054 <port id="0" precision="FP32" />
9055 </output>
9056 </layer>
9057 <layer id="661" name="__module.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
9058 <data element_type="f32" shape="" offset="21271676" size="4" />
9059 <output>
9060 <port id="0" precision="FP32" />
9061 </output>
9062 </layer>
9063 <layer id="662" name="__module.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
9064 <data levels="256" auto_broadcast="numpy" />
9065 <input>
9066 <port id="0" precision="FP32">
9067 <dim>-1</dim>
9068 <dim>-1</dim>
9069 <dim>384</dim>
9070 </port>
9071 <port id="1" precision="FP32" />
9072 <port id="2" precision="FP32" />
9073 <port id="3" precision="FP32" />
9074 <port id="4" precision="FP32" />
9075 </input>
9076 <output>
9077 <port id="5" precision="FP32">
9078 <dim>-1</dim>
9079 <dim>-1</dim>
9080 <dim>384</dim>
9081 </port>
9082 </output>
9083 </layer>
9084 <layer id="663" name="__module.encoder.layer.5.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
9085 <data element_type="i64" shape="4" offset="12201240" size="32" />
9086 <output>
9087 <port id="0" precision="I64">
9088 <dim>4</dim>
9089 </port>
9090 </output>
9091 </layer>
9092 <layer id="664" name="__module.encoder.layer.5.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
9093 <data special_zero="true" />
9094 <input>
9095 <port id="0" precision="FP32">
9096 <dim>-1</dim>
9097 <dim>-1</dim>
9098 <dim>384</dim>
9099 </port>
9100 <port id="1" precision="I64">
9101 <dim>4</dim>
9102 </port>
9103 </input>
9104 <output>
9105 <port id="2" precision="FP32" names="488,x.63">
9106 <dim>-1</dim>
9107 <dim>-1</dim>
9108 <dim>12</dim>
9109 <dim>32</dim>
9110 </port>
9111 </output>
9112 </layer>
9113 <layer id="665" name="Constant_1383" type="Const" version="opset1">
9114 <data element_type="i64" shape="4" offset="12201272" size="32" />
9115 <output>
9116 <port id="0" precision="I64" names="489">
9117 <dim>4</dim>
9118 </port>
9119 </output>
9120 </layer>
9121 <layer id="666" name="__module.encoder.layer.5.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
9122 <input>
9123 <port id="0" precision="FP32">
9124 <dim>-1</dim>
9125 <dim>-1</dim>
9126 <dim>12</dim>
9127 <dim>32</dim>
9128 </port>
9129 <port id="1" precision="I64">
9130 <dim>4</dim>
9131 </port>
9132 </input>
9133 <output>
9134 <port id="2" precision="FP32" names="490">
9135 <dim>-1</dim>
9136 <dim>12</dim>
9137 <dim>-1</dim>
9138 <dim>32</dim>
9139 </port>
9140 </output>
9141 </layer>
9142 <layer id="667" name="Constant_27536258" type="Const" version="opset1">
9143 <data element_type="i8" shape="384, 384" offset="21271680" size="147456" />
9144 <output>
9145 <port id="0" precision="I8">
9146 <dim>384</dim>
9147 <dim>384</dim>
9148 </port>
9149 </output>
9150 </layer>
9151 <layer id="668" name="Convert_27536259" type="Convert" version="opset1">
9152 <data destination_type="f32" />
9153 <input>
9154 <port id="0" precision="I8">
9155 <dim>384</dim>
9156 <dim>384</dim>
9157 </port>
9158 </input>
9159 <output>
9160 <port id="1" precision="FP32">
9161 <dim>384</dim>
9162 <dim>384</dim>
9163 </port>
9164 </output>
9165 </layer>
9166 <layer id="669" name="Constant_27536260" type="Const" version="opset1">
9167 <data element_type="f32" shape="384, 1" offset="21419136" size="1536" />
9168 <output>
9169 <port id="0" precision="FP32">
9170 <dim>384</dim>
9171 <dim>1</dim>
9172 </port>
9173 </output>
9174 </layer>
9175 <layer id="670" name="__module.encoder.layer.5.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9176 <data auto_broadcast="numpy" />
9177 <input>
9178 <port id="0" precision="FP32">
9179 <dim>384</dim>
9180 <dim>384</dim>
9181 </port>
9182 <port id="1" precision="FP32">
9183 <dim>384</dim>
9184 <dim>1</dim>
9185 </port>
9186 </input>
9187 <output>
9188 <port id="2" precision="FP32">
9189 <dim>384</dim>
9190 <dim>384</dim>
9191 </port>
9192 </output>
9193 </layer>
9194 <layer id="671" name="__module.encoder.layer.5.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
9195 <data transpose_a="false" transpose_b="true" />
9196 <input>
9197 <port id="0" precision="FP32">
9198 <dim>-1</dim>
9199 <dim>-1</dim>
9200 <dim>384</dim>
9201 </port>
9202 <port id="1" precision="FP32">
9203 <dim>384</dim>
9204 <dim>384</dim>
9205 </port>
9206 </input>
9207 <output>
9208 <port id="2" precision="FP32">
9209 <dim>-1</dim>
9210 <dim>-1</dim>
9211 <dim>384</dim>
9212 </port>
9213 </output>
9214 </layer>
9215 <layer id="672" name="Constant_6285" type="Const" version="opset1">
9216 <data element_type="f32" shape="1, 1, 384" offset="21420672" size="1536" />
9217 <output>
9218 <port id="0" precision="FP32">
9219 <dim>1</dim>
9220 <dim>1</dim>
9221 <dim>384</dim>
9222 </port>
9223 </output>
9224 </layer>
9225 <layer id="673" name="__module.encoder.layer.5.attention.self.key/aten::linear/Add" type="Add" version="opset1">
9226 <data auto_broadcast="numpy" />
9227 <input>
9228 <port id="0" precision="FP32">
9229 <dim>-1</dim>
9230 <dim>-1</dim>
9231 <dim>384</dim>
9232 </port>
9233 <port id="1" precision="FP32">
9234 <dim>1</dim>
9235 <dim>1</dim>
9236 <dim>384</dim>
9237 </port>
9238 </input>
9239 <output>
9240 <port id="2" precision="FP32" names="493,x.65">
9241 <dim>-1</dim>
9242 <dim>-1</dim>
9243 <dim>384</dim>
9244 </port>
9245 </output>
9246 </layer>
9247 <layer id="674" name="__module.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
9248 <data element_type="f32" shape="" offset="21422208" size="4" />
9249 <output>
9250 <port id="0" precision="FP32" />
9251 </output>
9252 </layer>
9253 <layer id="675" name="__module.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
9254 <data element_type="f32" shape="" offset="21422212" size="4" />
9255 <output>
9256 <port id="0" precision="FP32" />
9257 </output>
9258 </layer>
9259 <layer id="676" name="__module.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
9260 <data element_type="f32" shape="" offset="21422208" size="4" />
9261 <output>
9262 <port id="0" precision="FP32" />
9263 </output>
9264 </layer>
9265 <layer id="677" name="__module.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
9266 <data element_type="f32" shape="" offset="21422212" size="4" />
9267 <output>
9268 <port id="0" precision="FP32" />
9269 </output>
9270 </layer>
9271 <layer id="678" name="__module.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
9272 <data levels="256" auto_broadcast="numpy" />
9273 <input>
9274 <port id="0" precision="FP32">
9275 <dim>-1</dim>
9276 <dim>-1</dim>
9277 <dim>384</dim>
9278 </port>
9279 <port id="1" precision="FP32" />
9280 <port id="2" precision="FP32" />
9281 <port id="3" precision="FP32" />
9282 <port id="4" precision="FP32" />
9283 </input>
9284 <output>
9285 <port id="5" precision="FP32">
9286 <dim>-1</dim>
9287 <dim>-1</dim>
9288 <dim>384</dim>
9289 </port>
9290 </output>
9291 </layer>
9292 <layer id="679" name="__module.encoder.layer.5.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
9293 <data element_type="i64" shape="4" offset="12201240" size="32" />
9294 <output>
9295 <port id="0" precision="I64">
9296 <dim>4</dim>
9297 </port>
9298 </output>
9299 </layer>
9300 <layer id="680" name="__module.encoder.layer.5.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
9301 <data special_zero="true" />
9302 <input>
9303 <port id="0" precision="FP32">
9304 <dim>-1</dim>
9305 <dim>-1</dim>
9306 <dim>384</dim>
9307 </port>
9308 <port id="1" precision="I64">
9309 <dim>4</dim>
9310 </port>
9311 </input>
9312 <output>
9313 <port id="2" precision="FP32" names="497,x.67">
9314 <dim>-1</dim>
9315 <dim>-1</dim>
9316 <dim>12</dim>
9317 <dim>32</dim>
9318 </port>
9319 </output>
9320 </layer>
9321 <layer id="681" name="Constant_1406" type="Const" version="opset1">
9322 <data element_type="i64" shape="4" offset="12201272" size="32" />
9323 <output>
9324 <port id="0" precision="I64" names="498">
9325 <dim>4</dim>
9326 </port>
9327 </output>
9328 </layer>
9329 <layer id="682" name="__module.encoder.layer.5.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
9330 <input>
9331 <port id="0" precision="FP32">
9332 <dim>-1</dim>
9333 <dim>-1</dim>
9334 <dim>12</dim>
9335 <dim>32</dim>
9336 </port>
9337 <port id="1" precision="I64">
9338 <dim>4</dim>
9339 </port>
9340 </input>
9341 <output>
9342 <port id="2" precision="FP32" names="499">
9343 <dim>-1</dim>
9344 <dim>12</dim>
9345 <dim>-1</dim>
9346 <dim>32</dim>
9347 </port>
9348 </output>
9349 </layer>
9350 <layer id="683" name="Constant_27536262" type="Const" version="opset1">
9351 <data element_type="i8" shape="384, 384" offset="21422216" size="147456" />
9352 <output>
9353 <port id="0" precision="I8">
9354 <dim>384</dim>
9355 <dim>384</dim>
9356 </port>
9357 </output>
9358 </layer>
9359 <layer id="684" name="Convert_27536263" type="Convert" version="opset1">
9360 <data destination_type="f32" />
9361 <input>
9362 <port id="0" precision="I8">
9363 <dim>384</dim>
9364 <dim>384</dim>
9365 </port>
9366 </input>
9367 <output>
9368 <port id="1" precision="FP32">
9369 <dim>384</dim>
9370 <dim>384</dim>
9371 </port>
9372 </output>
9373 </layer>
9374 <layer id="685" name="Constant_27536264" type="Const" version="opset1">
9375 <data element_type="f32" shape="384, 1" offset="21569672" size="1536" />
9376 <output>
9377 <port id="0" precision="FP32">
9378 <dim>384</dim>
9379 <dim>1</dim>
9380 </port>
9381 </output>
9382 </layer>
9383 <layer id="686" name="__module.encoder.layer.5.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9384 <data auto_broadcast="numpy" />
9385 <input>
9386 <port id="0" precision="FP32">
9387 <dim>384</dim>
9388 <dim>384</dim>
9389 </port>
9390 <port id="1" precision="FP32">
9391 <dim>384</dim>
9392 <dim>1</dim>
9393 </port>
9394 </input>
9395 <output>
9396 <port id="2" precision="FP32">
9397 <dim>384</dim>
9398 <dim>384</dim>
9399 </port>
9400 </output>
9401 </layer>
9402 <layer id="687" name="__module.encoder.layer.5.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
9403 <data transpose_a="false" transpose_b="true" />
9404 <input>
9405 <port id="0" precision="FP32">
9406 <dim>-1</dim>
9407 <dim>-1</dim>
9408 <dim>384</dim>
9409 </port>
9410 <port id="1" precision="FP32">
9411 <dim>384</dim>
9412 <dim>384</dim>
9413 </port>
9414 </input>
9415 <output>
9416 <port id="2" precision="FP32">
9417 <dim>-1</dim>
9418 <dim>-1</dim>
9419 <dim>384</dim>
9420 </port>
9421 </output>
9422 </layer>
9423 <layer id="688" name="Constant_6286" type="Const" version="opset1">
9424 <data element_type="f32" shape="1, 1, 384" offset="21571208" size="1536" />
9425 <output>
9426 <port id="0" precision="FP32">
9427 <dim>1</dim>
9428 <dim>1</dim>
9429 <dim>384</dim>
9430 </port>
9431 </output>
9432 </layer>
9433 <layer id="689" name="__module.encoder.layer.5.attention.self.value/aten::linear/Add" type="Add" version="opset1">
9434 <data auto_broadcast="numpy" />
9435 <input>
9436 <port id="0" precision="FP32">
9437 <dim>-1</dim>
9438 <dim>-1</dim>
9439 <dim>384</dim>
9440 </port>
9441 <port id="1" precision="FP32">
9442 <dim>1</dim>
9443 <dim>1</dim>
9444 <dim>384</dim>
9445 </port>
9446 </input>
9447 <output>
9448 <port id="2" precision="FP32" names="502,x.69">
9449 <dim>-1</dim>
9450 <dim>-1</dim>
9451 <dim>384</dim>
9452 </port>
9453 </output>
9454 </layer>
9455 <layer id="690" name="__module.encoder.layer.5.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
9456 <data element_type="i64" shape="4" offset="12201240" size="32" />
9457 <output>
9458 <port id="0" precision="I64">
9459 <dim>4</dim>
9460 </port>
9461 </output>
9462 </layer>
9463 <layer id="691" name="__module.encoder.layer.5.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
9464 <data special_zero="true" />
9465 <input>
9466 <port id="0" precision="FP32">
9467 <dim>-1</dim>
9468 <dim>-1</dim>
9469 <dim>384</dim>
9470 </port>
9471 <port id="1" precision="I64">
9472 <dim>4</dim>
9473 </port>
9474 </input>
9475 <output>
9476 <port id="2" precision="FP32" names="506,x">
9477 <dim>-1</dim>
9478 <dim>-1</dim>
9479 <dim>12</dim>
9480 <dim>32</dim>
9481 </port>
9482 </output>
9483 </layer>
9484 <layer id="692" name="Constant_1429" type="Const" version="opset1">
9485 <data element_type="i64" shape="4" offset="12201272" size="32" />
9486 <output>
9487 <port id="0" precision="I64" names="507">
9488 <dim>4</dim>
9489 </port>
9490 </output>
9491 </layer>
9492 <layer id="693" name="__module.encoder.layer.5.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
9493 <input>
9494 <port id="0" precision="FP32">
9495 <dim>-1</dim>
9496 <dim>-1</dim>
9497 <dim>12</dim>
9498 <dim>32</dim>
9499 </port>
9500 <port id="1" precision="I64">
9501 <dim>4</dim>
9502 </port>
9503 </input>
9504 <output>
9505 <port id="2" precision="FP32" names="508">
9506 <dim>-1</dim>
9507 <dim>12</dim>
9508 <dim>-1</dim>
9509 <dim>32</dim>
9510 </port>
9511 </output>
9512 </layer>
9513 <layer id="694" name="__module.encoder.layer.5.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
9514 <data causal="false" />
9515 <input>
9516 <port id="0" precision="FP32">
9517 <dim>-1</dim>
9518 <dim>12</dim>
9519 <dim>-1</dim>
9520 <dim>32</dim>
9521 </port>
9522 <port id="1" precision="FP32">
9523 <dim>-1</dim>
9524 <dim>12</dim>
9525 <dim>-1</dim>
9526 <dim>32</dim>
9527 </port>
9528 <port id="2" precision="FP32">
9529 <dim>-1</dim>
9530 <dim>12</dim>
9531 <dim>-1</dim>
9532 <dim>32</dim>
9533 </port>
9534 <port id="3" precision="FP32">
9535 <dim>-1</dim>
9536 <dim>1</dim>
9537 <dim>-1</dim>
9538 <dim>-1</dim>
9539 </port>
9540 </input>
9541 <output>
9542 <port id="4" precision="FP32" names="509,attn_output.21">
9543 <dim>-1</dim>
9544 <dim>12</dim>
9545 <dim>-1</dim>
9546 <dim>32</dim>
9547 </port>
9548 </output>
9549 </layer>
9550 <layer id="695" name="__module.encoder.layer.5.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
9551 <data element_type="i32" shape="4" offset="12502384" size="16" />
9552 <output>
9553 <port id="0" precision="I32">
9554 <dim>4</dim>
9555 </port>
9556 </output>
9557 </layer>
9558 <layer id="696" name="__module.encoder.layer.5.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
9559 <input>
9560 <port id="0" precision="FP32">
9561 <dim>-1</dim>
9562 <dim>12</dim>
9563 <dim>-1</dim>
9564 <dim>32</dim>
9565 </port>
9566 <port id="1" precision="I32">
9567 <dim>4</dim>
9568 </port>
9569 </input>
9570 <output>
9571 <port id="2" precision="FP32" names="510,attn_output">
9572 <dim>-1</dim>
9573 <dim>-1</dim>
9574 <dim>12</dim>
9575 <dim>32</dim>
9576 </port>
9577 </output>
9578 </layer>
9579 <layer id="697" name="__module.encoder.layer.5.attention.self/aten::size/ShapeOf_6" type="ShapeOf" version="opset3">
9580 <data output_type="i64" />
9581 <input>
9582 <port id="0" precision="FP32">
9583 <dim>-1</dim>
9584 <dim>-1</dim>
9585 <dim>384</dim>
9586 </port>
9587 </input>
9588 <output>
9589 <port id="1" precision="I64">
9590 <dim>3</dim>
9591 </port>
9592 </output>
9593 </layer>
9594 <layer id="698" name="Constant_5837" type="Const" version="opset1">
9595 <data element_type="i64" shape="2" offset="12502400" size="16" />
9596 <output>
9597 <port id="0" precision="I64">
9598 <dim>2</dim>
9599 </port>
9600 </output>
9601 </layer>
9602 <layer id="699" name="Constant_5838" type="Const" version="opset1">
9603 <data element_type="i64" shape="" offset="12046068" size="8" />
9604 <output>
9605 <port id="0" precision="I64" />
9606 </output>
9607 </layer>
9608 <layer id="700" name="Gather_5839" type="Gather" version="opset8">
9609 <data batch_dims="0" />
9610 <input>
9611 <port id="0" precision="I64">
9612 <dim>3</dim>
9613 </port>
9614 <port id="1" precision="I64">
9615 <dim>2</dim>
9616 </port>
9617 <port id="2" precision="I64" />
9618 </input>
9619 <output>
9620 <port id="3" precision="I64">
9621 <dim>2</dim>
9622 </port>
9623 </output>
9624 </layer>
9625 <layer id="701" name="__module.encoder.layer.5.attention.self/prim::ListConstruct/Concat_3" type="Concat" version="opset1">
9626 <data axis="0" />
9627 <input>
9628 <port id="0" precision="I64">
9629 <dim>2</dim>
9630 </port>
9631 <port id="1" precision="I64">
9632 <dim>1</dim>
9633 </port>
9634 </input>
9635 <output>
9636 <port id="2" precision="I64" names="511">
9637 <dim>3</dim>
9638 </port>
9639 </output>
9640 </layer>
9641 <layer id="702" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
9642 <data special_zero="false" />
9643 <input>
9644 <port id="0" precision="FP32">
9645 <dim>-1</dim>
9646 <dim>-1</dim>
9647 <dim>12</dim>
9648 <dim>32</dim>
9649 </port>
9650 <port id="1" precision="I64">
9651 <dim>3</dim>
9652 </port>
9653 </input>
9654 <output>
9655 <port id="2" precision="FP32" names="512">
9656 <dim>-1</dim>
9657 <dim>-1</dim>
9658 <dim>384</dim>
9659 </port>
9660 </output>
9661 </layer>
9662 <layer id="703" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9663 <data element_type="f32" shape="1, 1, 384" offset="21572744" size="1536" />
9664 <output>
9665 <port id="0" precision="FP32">
9666 <dim>1</dim>
9667 <dim>1</dim>
9668 <dim>384</dim>
9669 </port>
9670 </output>
9671 </layer>
9672 <layer id="704" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9673 <data auto_broadcast="numpy" />
9674 <input>
9675 <port id="0" precision="FP32">
9676 <dim>-1</dim>
9677 <dim>-1</dim>
9678 <dim>384</dim>
9679 </port>
9680 <port id="1" precision="FP32">
9681 <dim>1</dim>
9682 <dim>1</dim>
9683 <dim>384</dim>
9684 </port>
9685 </input>
9686 <output>
9687 <port id="2" precision="FP32">
9688 <dim>-1</dim>
9689 <dim>-1</dim>
9690 <dim>384</dim>
9691 </port>
9692 </output>
9693 </layer>
9694 <layer id="705" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9695 <data element_type="f32" shape="" offset="21574280" size="4" />
9696 <output>
9697 <port id="0" precision="FP32" />
9698 </output>
9699 </layer>
9700 <layer id="706" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9701 <data element_type="f32" shape="" offset="21574284" size="4" />
9702 <output>
9703 <port id="0" precision="FP32" />
9704 </output>
9705 </layer>
9706 <layer id="707" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9707 <data element_type="f32" shape="" offset="21574280" size="4" />
9708 <output>
9709 <port id="0" precision="FP32" />
9710 </output>
9711 </layer>
9712 <layer id="708" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9713 <data element_type="f32" shape="" offset="21574284" size="4" />
9714 <output>
9715 <port id="0" precision="FP32" />
9716 </output>
9717 </layer>
9718 <layer id="709" name="__module.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9719 <data levels="256" auto_broadcast="numpy" />
9720 <input>
9721 <port id="0" precision="FP32">
9722 <dim>-1</dim>
9723 <dim>-1</dim>
9724 <dim>384</dim>
9725 </port>
9726 <port id="1" precision="FP32" />
9727 <port id="2" precision="FP32" />
9728 <port id="3" precision="FP32" />
9729 <port id="4" precision="FP32" />
9730 </input>
9731 <output>
9732 <port id="5" precision="FP32">
9733 <dim>-1</dim>
9734 <dim>-1</dim>
9735 <dim>384</dim>
9736 </port>
9737 </output>
9738 </layer>
9739 <layer id="710" name="Constant_27536266" type="Const" version="opset1">
9740 <data element_type="i8" shape="384, 384" offset="21574288" size="147456" />
9741 <output>
9742 <port id="0" precision="I8">
9743 <dim>384</dim>
9744 <dim>384</dim>
9745 </port>
9746 </output>
9747 </layer>
9748 <layer id="711" name="Convert_27536267" type="Convert" version="opset1">
9749 <data destination_type="f32" />
9750 <input>
9751 <port id="0" precision="I8">
9752 <dim>384</dim>
9753 <dim>384</dim>
9754 </port>
9755 </input>
9756 <output>
9757 <port id="1" precision="FP32">
9758 <dim>384</dim>
9759 <dim>384</dim>
9760 </port>
9761 </output>
9762 </layer>
9763 <layer id="712" name="Constant_27536268" type="Const" version="opset1">
9764 <data element_type="f32" shape="384, 1" offset="21721744" size="1536" />
9765 <output>
9766 <port id="0" precision="FP32">
9767 <dim>384</dim>
9768 <dim>1</dim>
9769 </port>
9770 </output>
9771 </layer>
9772 <layer id="713" name="__module.encoder.layer.5.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9773 <data auto_broadcast="numpy" />
9774 <input>
9775 <port id="0" precision="FP32">
9776 <dim>384</dim>
9777 <dim>384</dim>
9778 </port>
9779 <port id="1" precision="FP32">
9780 <dim>384</dim>
9781 <dim>1</dim>
9782 </port>
9783 </input>
9784 <output>
9785 <port id="2" precision="FP32">
9786 <dim>384</dim>
9787 <dim>384</dim>
9788 </port>
9789 </output>
9790 </layer>
9791 <layer id="714" name="__module.encoder.layer.5.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
9792 <data transpose_a="false" transpose_b="true" />
9793 <input>
9794 <port id="0" precision="FP32">
9795 <dim>-1</dim>
9796 <dim>-1</dim>
9797 <dim>384</dim>
9798 </port>
9799 <port id="1" precision="FP32">
9800 <dim>384</dim>
9801 <dim>384</dim>
9802 </port>
9803 </input>
9804 <output>
9805 <port id="2" precision="FP32">
9806 <dim>-1</dim>
9807 <dim>-1</dim>
9808 <dim>384</dim>
9809 </port>
9810 </output>
9811 </layer>
9812 <layer id="715" name="Constant_6287" type="Const" version="opset1">
9813 <data element_type="f32" shape="1, 1, 384" offset="21723280" size="1536" />
9814 <output>
9815 <port id="0" precision="FP32">
9816 <dim>1</dim>
9817 <dim>1</dim>
9818 <dim>384</dim>
9819 </port>
9820 </output>
9821 </layer>
9822 <layer id="716" name="__module.encoder.layer.5.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
9823 <data auto_broadcast="numpy" />
9824 <input>
9825 <port id="0" precision="FP32">
9826 <dim>-1</dim>
9827 <dim>-1</dim>
9828 <dim>384</dim>
9829 </port>
9830 <port id="1" precision="FP32">
9831 <dim>1</dim>
9832 <dim>1</dim>
9833 <dim>384</dim>
9834 </port>
9835 </input>
9836 <output>
9837 <port id="2" precision="FP32" names="518,input.23">
9838 <dim>-1</dim>
9839 <dim>-1</dim>
9840 <dim>384</dim>
9841 </port>
9842 </output>
9843 </layer>
9844 <layer id="717" name="__module.encoder.layer.5.attention.output/aten::add/Add" type="Add" version="opset1">
9845 <data auto_broadcast="numpy" />
9846 <input>
9847 <port id="0" precision="FP32">
9848 <dim>-1</dim>
9849 <dim>-1</dim>
9850 <dim>384</dim>
9851 </port>
9852 <port id="1" precision="FP32">
9853 <dim>-1</dim>
9854 <dim>-1</dim>
9855 <dim>384</dim>
9856 </port>
9857 </input>
9858 <output>
9859 <port id="2" precision="FP32" names="520">
9860 <dim>-1</dim>
9861 <dim>-1</dim>
9862 <dim>384</dim>
9863 </port>
9864 </output>
9865 </layer>
9866 <layer id="718" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
9867 <data element_type="i32" shape="1" offset="12046084" size="4" />
9868 <output>
9869 <port id="0" precision="I32">
9870 <dim>1</dim>
9871 </port>
9872 </output>
9873 </layer>
9874 <layer id="719" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
9875 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
9876 <input>
9877 <port id="0" precision="FP32">
9878 <dim>-1</dim>
9879 <dim>-1</dim>
9880 <dim>384</dim>
9881 </port>
9882 <port id="1" precision="I32">
9883 <dim>1</dim>
9884 </port>
9885 </input>
9886 <output>
9887 <port id="2" precision="FP32">
9888 <dim>-1</dim>
9889 <dim>-1</dim>
9890 <dim>384</dim>
9891 </port>
9892 </output>
9893 </layer>
9894 <layer id="720" name="Constant_6288" type="Const" version="opset1">
9895 <data element_type="f32" shape="1, 1, 384" offset="21724816" size="1536" />
9896 <output>
9897 <port id="0" precision="FP32">
9898 <dim>1</dim>
9899 <dim>1</dim>
9900 <dim>384</dim>
9901 </port>
9902 </output>
9903 </layer>
9904 <layer id="721" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
9905 <data auto_broadcast="numpy" />
9906 <input>
9907 <port id="0" precision="FP32">
9908 <dim>-1</dim>
9909 <dim>-1</dim>
9910 <dim>384</dim>
9911 </port>
9912 <port id="1" precision="FP32">
9913 <dim>1</dim>
9914 <dim>1</dim>
9915 <dim>384</dim>
9916 </port>
9917 </input>
9918 <output>
9919 <port id="2" precision="FP32">
9920 <dim>-1</dim>
9921 <dim>-1</dim>
9922 <dim>384</dim>
9923 </port>
9924 </output>
9925 </layer>
9926 <layer id="722" name="Constant_6289" type="Const" version="opset1">
9927 <data element_type="f32" shape="1, 1, 384" offset="21726352" size="1536" />
9928 <output>
9929 <port id="0" precision="FP32">
9930 <dim>1</dim>
9931 <dim>1</dim>
9932 <dim>384</dim>
9933 </port>
9934 </output>
9935 </layer>
9936 <layer id="723" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
9937 <data auto_broadcast="numpy" />
9938 <input>
9939 <port id="0" precision="FP32">
9940 <dim>-1</dim>
9941 <dim>-1</dim>
9942 <dim>384</dim>
9943 </port>
9944 <port id="1" precision="FP32">
9945 <dim>1</dim>
9946 <dim>1</dim>
9947 <dim>384</dim>
9948 </port>
9949 </input>
9950 <output>
9951 <port id="2" precision="FP32" names="524,input_tensor">
9952 <dim>-1</dim>
9953 <dim>-1</dim>
9954 <dim>384</dim>
9955 </port>
9956 </output>
9957 </layer>
9958 <layer id="724" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9959 <data element_type="f32" shape="1, 1, 384" offset="21727888" size="1536" />
9960 <output>
9961 <port id="0" precision="FP32">
9962 <dim>1</dim>
9963 <dim>1</dim>
9964 <dim>384</dim>
9965 </port>
9966 </output>
9967 </layer>
9968 <layer id="725" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9969 <data auto_broadcast="numpy" />
9970 <input>
9971 <port id="0" precision="FP32">
9972 <dim>-1</dim>
9973 <dim>-1</dim>
9974 <dim>384</dim>
9975 </port>
9976 <port id="1" precision="FP32">
9977 <dim>1</dim>
9978 <dim>1</dim>
9979 <dim>384</dim>
9980 </port>
9981 </input>
9982 <output>
9983 <port id="2" precision="FP32">
9984 <dim>-1</dim>
9985 <dim>-1</dim>
9986 <dim>384</dim>
9987 </port>
9988 </output>
9989 </layer>
9990 <layer id="726" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9991 <data element_type="f32" shape="" offset="21729424" size="4" />
9992 <output>
9993 <port id="0" precision="FP32" />
9994 </output>
9995 </layer>
9996 <layer id="727" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9997 <data element_type="f32" shape="" offset="21729428" size="4" />
9998 <output>
9999 <port id="0" precision="FP32" />
10000 </output>
10001 </layer>
10002 <layer id="728" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
10003 <data element_type="f32" shape="" offset="21729424" size="4" />
10004 <output>
10005 <port id="0" precision="FP32" />
10006 </output>
10007 </layer>
10008 <layer id="729" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
10009 <data element_type="f32" shape="" offset="21729428" size="4" />
10010 <output>
10011 <port id="0" precision="FP32" />
10012 </output>
10013 </layer>
10014 <layer id="730" name="__module.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
10015 <data levels="256" auto_broadcast="numpy" />
10016 <input>
10017 <port id="0" precision="FP32">
10018 <dim>-1</dim>
10019 <dim>-1</dim>
10020 <dim>384</dim>
10021 </port>
10022 <port id="1" precision="FP32" />
10023 <port id="2" precision="FP32" />
10024 <port id="3" precision="FP32" />
10025 <port id="4" precision="FP32" />
10026 </input>
10027 <output>
10028 <port id="5" precision="FP32">
10029 <dim>-1</dim>
10030 <dim>-1</dim>
10031 <dim>384</dim>
10032 </port>
10033 </output>
10034 </layer>
10035 <layer id="731" name="Constant_27536270" type="Const" version="opset1">
10036 <data element_type="i8" shape="1536, 384" offset="21729432" size="589824" />
10037 <output>
10038 <port id="0" precision="I8">
10039 <dim>1536</dim>
10040 <dim>384</dim>
10041 </port>
10042 </output>
10043 </layer>
10044 <layer id="732" name="Convert_27536271" type="Convert" version="opset1">
10045 <data destination_type="f32" />
10046 <input>
10047 <port id="0" precision="I8">
10048 <dim>1536</dim>
10049 <dim>384</dim>
10050 </port>
10051 </input>
10052 <output>
10053 <port id="1" precision="FP32">
10054 <dim>1536</dim>
10055 <dim>384</dim>
10056 </port>
10057 </output>
10058 </layer>
10059 <layer id="733" name="Constant_27536272" type="Const" version="opset1">
10060 <data element_type="f32" shape="1536, 1" offset="22319256" size="6144" />
10061 <output>
10062 <port id="0" precision="FP32">
10063 <dim>1536</dim>
10064 <dim>1</dim>
10065 </port>
10066 </output>
10067 </layer>
10068 <layer id="734" name="__module.encoder.layer.5.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
10069 <data auto_broadcast="numpy" />
10070 <input>
10071 <port id="0" precision="FP32">
10072 <dim>1536</dim>
10073 <dim>384</dim>
10074 </port>
10075 <port id="1" precision="FP32">
10076 <dim>1536</dim>
10077 <dim>1</dim>
10078 </port>
10079 </input>
10080 <output>
10081 <port id="2" precision="FP32">
10082 <dim>1536</dim>
10083 <dim>384</dim>
10084 </port>
10085 </output>
10086 </layer>
10087 <layer id="735" name="__module.encoder.layer.5.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
10088 <data transpose_a="false" transpose_b="true" />
10089 <input>
10090 <port id="0" precision="FP32">
10091 <dim>-1</dim>
10092 <dim>-1</dim>
10093 <dim>384</dim>
10094 </port>
10095 <port id="1" precision="FP32">
10096 <dim>1536</dim>
10097 <dim>384</dim>
10098 </port>
10099 </input>
10100 <output>
10101 <port id="2" precision="FP32">
10102 <dim>-1</dim>
10103 <dim>-1</dim>
10104 <dim>1536</dim>
10105 </port>
10106 </output>
10107 </layer>
10108 <layer id="736" name="Constant_6290" type="Const" version="opset1">
10109 <data element_type="f32" shape="1, 1, 1536" offset="22325400" size="6144" />
10110 <output>
10111 <port id="0" precision="FP32">
10112 <dim>1</dim>
10113 <dim>1</dim>
10114 <dim>1536</dim>
10115 </port>
10116 </output>
10117 </layer>
10118 <layer id="737" name="__module.encoder.layer.5.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
10119 <data auto_broadcast="numpy" />
10120 <input>
10121 <port id="0" precision="FP32">
10122 <dim>-1</dim>
10123 <dim>-1</dim>
10124 <dim>1536</dim>
10125 </port>
10126 <port id="1" precision="FP32">
10127 <dim>1</dim>
10128 <dim>1</dim>
10129 <dim>1536</dim>
10130 </port>
10131 </input>
10132 <output>
10133 <port id="2" precision="FP32" names="529">
10134 <dim>-1</dim>
10135 <dim>-1</dim>
10136 <dim>1536</dim>
10137 </port>
10138 </output>
10139 </layer>
10140 <layer id="738" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
10141 <data approximation_mode="ERF" />
10142 <input>
10143 <port id="0" precision="FP32">
10144 <dim>-1</dim>
10145 <dim>-1</dim>
10146 <dim>1536</dim>
10147 </port>
10148 </input>
10149 <output>
10150 <port id="1" precision="FP32" names="530">
10151 <dim>-1</dim>
10152 <dim>-1</dim>
10153 <dim>1536</dim>
10154 </port>
10155 </output>
10156 </layer>
10157 <layer id="739" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
10158 <data element_type="f32" shape="1, 1, 1536" offset="22331544" size="6144" />
10159 <output>
10160 <port id="0" precision="FP32">
10161 <dim>1</dim>
10162 <dim>1</dim>
10163 <dim>1536</dim>
10164 </port>
10165 </output>
10166 </layer>
10167 <layer id="740" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
10168 <data auto_broadcast="numpy" />
10169 <input>
10170 <port id="0" precision="FP32">
10171 <dim>-1</dim>
10172 <dim>-1</dim>
10173 <dim>1536</dim>
10174 </port>
10175 <port id="1" precision="FP32">
10176 <dim>1</dim>
10177 <dim>1</dim>
10178 <dim>1536</dim>
10179 </port>
10180 </input>
10181 <output>
10182 <port id="2" precision="FP32">
10183 <dim>-1</dim>
10184 <dim>-1</dim>
10185 <dim>1536</dim>
10186 </port>
10187 </output>
10188 </layer>
10189 <layer id="741" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
10190 <data element_type="f32" shape="" offset="22337688" size="4" />
10191 <output>
10192 <port id="0" precision="FP32" />
10193 </output>
10194 </layer>
10195 <layer id="742" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
10196 <data element_type="f32" shape="" offset="22337692" size="4" />
10197 <output>
10198 <port id="0" precision="FP32" />
10199 </output>
10200 </layer>
10201 <layer id="743" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
10202 <data element_type="f32" shape="" offset="22337688" size="4" />
10203 <output>
10204 <port id="0" precision="FP32" />
10205 </output>
10206 </layer>
10207 <layer id="744" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
10208 <data element_type="f32" shape="" offset="22337692" size="4" />
10209 <output>
10210 <port id="0" precision="FP32" />
10211 </output>
10212 </layer>
10213 <layer id="745" name="__module.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
10214 <data levels="256" auto_broadcast="numpy" />
10215 <input>
10216 <port id="0" precision="FP32">
10217 <dim>-1</dim>
10218 <dim>-1</dim>
10219 <dim>1536</dim>
10220 </port>
10221 <port id="1" precision="FP32" />
10222 <port id="2" precision="FP32" />
10223 <port id="3" precision="FP32" />
10224 <port id="4" precision="FP32" />
10225 </input>
10226 <output>
10227 <port id="5" precision="FP32">
10228 <dim>-1</dim>
10229 <dim>-1</dim>
10230 <dim>1536</dim>
10231 </port>
10232 </output>
10233 </layer>
10234 <layer id="746" name="Constant_27536274" type="Const" version="opset1">
10235 <data element_type="i8" shape="384, 1536" offset="22337696" size="589824" />
10236 <output>
10237 <port id="0" precision="I8">
10238 <dim>384</dim>
10239 <dim>1536</dim>
10240 </port>
10241 </output>
10242 </layer>
10243 <layer id="747" name="Convert_27536275" type="Convert" version="opset1">
10244 <data destination_type="f32" />
10245 <input>
10246 <port id="0" precision="I8">
10247 <dim>384</dim>
10248 <dim>1536</dim>
10249 </port>
10250 </input>
10251 <output>
10252 <port id="1" precision="FP32">
10253 <dim>384</dim>
10254 <dim>1536</dim>
10255 </port>
10256 </output>
10257 </layer>
10258 <layer id="748" name="Constant_27536276" type="Const" version="opset1">
10259 <data element_type="f32" shape="384, 1" offset="22927520" size="1536" />
10260 <output>
10261 <port id="0" precision="FP32">
10262 <dim>384</dim>
10263 <dim>1</dim>
10264 </port>
10265 </output>
10266 </layer>
10267 <layer id="749" name="__module.encoder.layer.5.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
10268 <data auto_broadcast="numpy" />
10269 <input>
10270 <port id="0" precision="FP32">
10271 <dim>384</dim>
10272 <dim>1536</dim>
10273 </port>
10274 <port id="1" precision="FP32">
10275 <dim>384</dim>
10276 <dim>1</dim>
10277 </port>
10278 </input>
10279 <output>
10280 <port id="2" precision="FP32">
10281 <dim>384</dim>
10282 <dim>1536</dim>
10283 </port>
10284 </output>
10285 </layer>
10286 <layer id="750" name="__module.encoder.layer.5.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
10287 <data transpose_a="false" transpose_b="true" />
10288 <input>
10289 <port id="0" precision="FP32">
10290 <dim>-1</dim>
10291 <dim>-1</dim>
10292 <dim>1536</dim>
10293 </port>
10294 <port id="1" precision="FP32">
10295 <dim>384</dim>
10296 <dim>1536</dim>
10297 </port>
10298 </input>
10299 <output>
10300 <port id="2" precision="FP32">
10301 <dim>-1</dim>
10302 <dim>-1</dim>
10303 <dim>384</dim>
10304 </port>
10305 </output>
10306 </layer>
10307 <layer id="751" name="Constant_6291" type="Const" version="opset1">
10308 <data element_type="f32" shape="1, 1, 384" offset="22929056" size="1536" />
10309 <output>
10310 <port id="0" precision="FP32">
10311 <dim>1</dim>
10312 <dim>1</dim>
10313 <dim>384</dim>
10314 </port>
10315 </output>
10316 </layer>
10317 <layer id="752" name="__module.encoder.layer.5.output.dense/aten::linear/Add" type="Add" version="opset1">
10318 <data auto_broadcast="numpy" />
10319 <input>
10320 <port id="0" precision="FP32">
10321 <dim>-1</dim>
10322 <dim>-1</dim>
10323 <dim>384</dim>
10324 </port>
10325 <port id="1" precision="FP32">
10326 <dim>1</dim>
10327 <dim>1</dim>
10328 <dim>384</dim>
10329 </port>
10330 </input>
10331 <output>
10332 <port id="2" precision="FP32" names="536,input">
10333 <dim>-1</dim>
10334 <dim>-1</dim>
10335 <dim>384</dim>
10336 </port>
10337 </output>
10338 </layer>
10339 <layer id="753" name="__module.encoder.layer.5.output/aten::add/Add" type="Add" version="opset1">
10340 <data auto_broadcast="numpy" />
10341 <input>
10342 <port id="0" precision="FP32">
10343 <dim>-1</dim>
10344 <dim>-1</dim>
10345 <dim>384</dim>
10346 </port>
10347 <port id="1" precision="FP32">
10348 <dim>-1</dim>
10349 <dim>-1</dim>
10350 <dim>384</dim>
10351 </port>
10352 </input>
10353 <output>
10354 <port id="2" precision="FP32" names="538">
10355 <dim>-1</dim>
10356 <dim>-1</dim>
10357 <dim>384</dim>
10358 </port>
10359 </output>
10360 </layer>
10361 <layer id="754" name="__module.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
10362 <data element_type="i32" shape="1" offset="12046084" size="4" />
10363 <output>
10364 <port id="0" precision="I32">
10365 <dim>1</dim>
10366 </port>
10367 </output>
10368 </layer>
10369 <layer id="755" name="__module.encoder.layer.5.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
10370 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
10371 <input>
10372 <port id="0" precision="FP32">
10373 <dim>-1</dim>
10374 <dim>-1</dim>
10375 <dim>384</dim>
10376 </port>
10377 <port id="1" precision="I32">
10378 <dim>1</dim>
10379 </port>
10380 </input>
10381 <output>
10382 <port id="2" precision="FP32">
10383 <dim>-1</dim>
10384 <dim>-1</dim>
10385 <dim>384</dim>
10386 </port>
10387 </output>
10388 </layer>
10389 <layer id="756" name="Constant_6292" type="Const" version="opset1">
10390 <data element_type="f32" shape="1, 1, 384" offset="22930592" size="1536" />
10391 <output>
10392 <port id="0" precision="FP32">
10393 <dim>1</dim>
10394 <dim>1</dim>
10395 <dim>384</dim>
10396 </port>
10397 </output>
10398 </layer>
10399 <layer id="757" name="__module.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
10400 <data auto_broadcast="numpy" />
10401 <input>
10402 <port id="0" precision="FP32">
10403 <dim>-1</dim>
10404 <dim>-1</dim>
10405 <dim>384</dim>
10406 </port>
10407 <port id="1" precision="FP32">
10408 <dim>1</dim>
10409 <dim>1</dim>
10410 <dim>384</dim>
10411 </port>
10412 </input>
10413 <output>
10414 <port id="2" precision="FP32">
10415 <dim>-1</dim>
10416 <dim>-1</dim>
10417 <dim>384</dim>
10418 </port>
10419 </output>
10420 </layer>
10421 <layer id="758" name="Constant_6293" type="Const" version="opset1">
10422 <data element_type="f32" shape="1, 1, 384" offset="22932128" size="1536" />
10423 <output>
10424 <port id="0" precision="FP32">
10425 <dim>1</dim>
10426 <dim>1</dim>
10427 <dim>384</dim>
10428 </port>
10429 </output>
10430 </layer>
10431 <layer id="759" name="__module.encoder.layer.5.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
10432 <data auto_broadcast="numpy" />
10433 <input>
10434 <port id="0" precision="FP32">
10435 <dim>-1</dim>
10436 <dim>-1</dim>
10437 <dim>384</dim>
10438 </port>
10439 <port id="1" precision="FP32">
10440 <dim>1</dim>
10441 <dim>1</dim>
10442 <dim>384</dim>
10443 </port>
10444 </input>
10445 <output>
10446 <port id="2" precision="FP32" names="last_hidden_state">
10447 <dim>-1</dim>
10448 <dim>-1</dim>
10449 <dim>384</dim>
10450 </port>
10451 </output>
10452 </layer>
10453 <layer id="760" name="Result_2691" type="Result" version="opset1">
10454 <input>
10455 <port id="0" precision="FP32">
10456 <dim>-1</dim>
10457 <dim>-1</dim>
10458 <dim>384</dim>
10459 </port>
10460 </input>
10461 </layer>
10462 </layers>
10463 <edges>
10464 <edge from-layer="0" from-port="0" to-layer="14" to-port="0" />
10465 <edge from-layer="1" from-port="0" to-layer="93" to-port="0" />
10466 <edge from-layer="1" from-port="0" to-layer="96" to-port="0" />
10467 <edge from-layer="2" from-port="0" to-layer="7" to-port="0" />
10468 <edge from-layer="2" from-port="0" to-layer="24" to-port="0" />
10469 <edge from-layer="3" from-port="0" to-layer="4" to-port="0" />
10470 <edge from-layer="4" from-port="1" to-layer="6" to-port="0" />
10471 <edge from-layer="5" from-port="0" to-layer="6" to-port="1" />
10472 <edge from-layer="6" from-port="2" to-layer="9" to-port="0" />
10473 <edge from-layer="7" from-port="1" to-layer="9" to-port="1" />
10474 <edge from-layer="8" from-port="0" to-layer="9" to-port="2" />
10475 <edge from-layer="9" from-port="3" to-layer="17" to-port="0" />
10476 <edge from-layer="10" from-port="0" to-layer="11" to-port="0" />
10477 <edge from-layer="11" from-port="1" to-layer="13" to-port="0" />
10478 <edge from-layer="12" from-port="0" to-layer="13" to-port="1" />
10479 <edge from-layer="13" from-port="2" to-layer="16" to-port="0" />
10480 <edge from-layer="14" from-port="1" to-layer="16" to-port="1" />
10481 <edge from-layer="15" from-port="0" to-layer="16" to-port="2" />
10482 <edge from-layer="16" from-port="3" to-layer="17" to-port="1" />
10483 <edge from-layer="17" from-port="2" to-layer="34" to-port="0" />
10484 <edge from-layer="18" from-port="0" to-layer="19" to-port="0" />
10485 <edge from-layer="19" from-port="1" to-layer="21" to-port="0" />
10486 <edge from-layer="20" from-port="0" to-layer="21" to-port="1" />
10487 <edge from-layer="21" from-port="2" to-layer="33" to-port="0" />
10488 <edge from-layer="22" from-port="0" to-layer="30" to-port="0" />
10489 <edge from-layer="23" from-port="0" to-layer="30" to-port="1" />
10490 <edge from-layer="24" from-port="1" to-layer="27" to-port="0" />
10491 <edge from-layer="25" from-port="0" to-layer="27" to-port="1" />
10492 <edge from-layer="26" from-port="0" to-layer="27" to-port="2" />
10493 <edge from-layer="27" from-port="3" to-layer="30" to-port="2" />
10494 <edge from-layer="27" from-port="3" to-layer="104" to-port="2" />
10495 <edge from-layer="28" from-port="0" to-layer="30" to-port="3" />
10496 <edge from-layer="29" from-port="0" to-layer="30" to-port="4" />
10497 <edge from-layer="30" from-port="5" to-layer="31" to-port="0" />
10498 <edge from-layer="31" from-port="1" to-layer="33" to-port="1" />
10499 <edge from-layer="32" from-port="0" to-layer="33" to-port="2" />
10500 <edge from-layer="33" from-port="3" to-layer="34" to-port="1" />
10501 <edge from-layer="34" from-port="2" to-layer="36" to-port="0" />
10502 <edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
10503 <edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
10504 <edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
10505 <edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
10506 <edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
10507 <edge from-layer="40" from-port="2" to-layer="42" to-port="0" />
10508 <edge from-layer="40" from-port="2" to-layer="116" to-port="0" />
10509 <edge from-layer="40" from-port="2" to-layer="137" to-port="1" />
10510 <edge from-layer="41" from-port="0" to-layer="42" to-port="1" />
10511 <edge from-layer="42" from-port="2" to-layer="47" to-port="0" />
10512 <edge from-layer="43" from-port="0" to-layer="47" to-port="1" />
10513 <edge from-layer="44" from-port="0" to-layer="47" to-port="2" />
10514 <edge from-layer="45" from-port="0" to-layer="47" to-port="3" />
10515 <edge from-layer="46" from-port="0" to-layer="47" to-port="4" />
10516 <edge from-layer="47" from-port="5" to-layer="68" to-port="0" />
10517 <edge from-layer="47" from-port="5" to-layer="84" to-port="0" />
10518 <edge from-layer="47" from-port="5" to-layer="52" to-port="0" />
10519 <edge from-layer="48" from-port="0" to-layer="49" to-port="0" />
10520 <edge from-layer="49" from-port="1" to-layer="51" to-port="0" />
10521 <edge from-layer="50" from-port="0" to-layer="51" to-port="1" />
10522 <edge from-layer="51" from-port="2" to-layer="52" to-port="1" />
10523 <edge from-layer="52" from-port="2" to-layer="54" to-port="0" />
10524 <edge from-layer="53" from-port="0" to-layer="54" to-port="1" />
10525 <edge from-layer="54" from-port="2" to-layer="59" to-port="0" />
10526 <edge from-layer="55" from-port="0" to-layer="59" to-port="1" />
10527 <edge from-layer="56" from-port="0" to-layer="59" to-port="2" />
10528 <edge from-layer="57" from-port="0" to-layer="59" to-port="3" />
10529 <edge from-layer="58" from-port="0" to-layer="59" to-port="4" />
10530 <edge from-layer="59" from-port="5" to-layer="61" to-port="0" />
10531 <edge from-layer="60" from-port="0" to-layer="61" to-port="1" />
10532 <edge from-layer="61" from-port="2" to-layer="63" to-port="0" />
10533 <edge from-layer="62" from-port="0" to-layer="63" to-port="1" />
10534 <edge from-layer="63" from-port="2" to-layer="113" to-port="0" />
10535 <edge from-layer="64" from-port="0" to-layer="65" to-port="0" />
10536 <edge from-layer="65" from-port="1" to-layer="67" to-port="0" />
10537 <edge from-layer="66" from-port="0" to-layer="67" to-port="1" />
10538 <edge from-layer="67" from-port="2" to-layer="68" to-port="1" />
10539 <edge from-layer="68" from-port="2" to-layer="70" to-port="0" />
10540 <edge from-layer="69" from-port="0" to-layer="70" to-port="1" />
10541 <edge from-layer="70" from-port="2" to-layer="75" to-port="0" />
10542 <edge from-layer="71" from-port="0" to-layer="75" to-port="1" />
10543 <edge from-layer="72" from-port="0" to-layer="75" to-port="2" />
10544 <edge from-layer="73" from-port="0" to-layer="75" to-port="3" />
10545 <edge from-layer="74" from-port="0" to-layer="75" to-port="4" />
10546 <edge from-layer="75" from-port="5" to-layer="77" to-port="0" />
10547 <edge from-layer="76" from-port="0" to-layer="77" to-port="1" />
10548 <edge from-layer="77" from-port="2" to-layer="79" to-port="0" />
10549 <edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
10550 <edge from-layer="79" from-port="2" to-layer="113" to-port="1" />
10551 <edge from-layer="80" from-port="0" to-layer="81" to-port="0" />
10552 <edge from-layer="81" from-port="1" to-layer="83" to-port="0" />
10553 <edge from-layer="82" from-port="0" to-layer="83" to-port="1" />
10554 <edge from-layer="83" from-port="2" to-layer="84" to-port="1" />
10555 <edge from-layer="84" from-port="2" to-layer="86" to-port="0" />
10556 <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
10557 <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
10558 <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
10559 <edge from-layer="88" from-port="2" to-layer="90" to-port="0" />
10560 <edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
10561 <edge from-layer="90" from-port="2" to-layer="113" to-port="2" />
10562 <edge from-layer="91" from-port="0" to-layer="109" to-port="0" />
10563 <edge from-layer="92" from-port="0" to-layer="93" to-port="1" />
10564 <edge from-layer="93" from-port="2" to-layer="95" to-port="0" />
10565 <edge from-layer="94" from-port="0" to-layer="95" to-port="1" />
10566 <edge from-layer="95" from-port="2" to-layer="105" to-port="0" />
10567 <edge from-layer="96" from-port="1" to-layer="99" to-port="0" />
10568 <edge from-layer="96" from-port="1" to-layer="103" to-port="0" />
10569 <edge from-layer="97" from-port="0" to-layer="99" to-port="1" />
10570 <edge from-layer="98" from-port="0" to-layer="99" to-port="2" />
10571 <edge from-layer="99" from-port="3" to-layer="104" to-port="0" />
10572 <edge from-layer="100" from-port="0" to-layer="104" to-port="1" />
10573 <edge from-layer="101" from-port="0" to-layer="103" to-port="1" />
10574 <edge from-layer="102" from-port="0" to-layer="103" to-port="2" />
10575 <edge from-layer="103" from-port="3" to-layer="104" to-port="3" />
10576 <edge from-layer="104" from-port="4" to-layer="105" to-port="1" />
10577 <edge from-layer="105" from-port="2" to-layer="106" to-port="0" />
10578 <edge from-layer="106" from-port="1" to-layer="108" to-port="0" />
10579 <edge from-layer="107" from-port="0" to-layer="108" to-port="1" />
10580 <edge from-layer="108" from-port="2" to-layer="109" to-port="1" />
10581 <edge from-layer="109" from-port="2" to-layer="110" to-port="0" />
10582 <edge from-layer="109" from-port="2" to-layer="112" to-port="2" />
10583 <edge from-layer="110" from-port="1" to-layer="112" to-port="0" />
10584 <edge from-layer="111" from-port="0" to-layer="112" to-port="1" />
10585 <edge from-layer="112" from-port="3" to-layer="113" to-port="3" />
10586 <edge from-layer="112" from-port="3" to-layer="694" to-port="3" />
10587 <edge from-layer="112" from-port="3" to-layer="578" to-port="3" />
10588 <edge from-layer="112" from-port="3" to-layer="462" to-port="3" />
10589 <edge from-layer="112" from-port="3" to-layer="346" to-port="3" />
10590 <edge from-layer="112" from-port="3" to-layer="230" to-port="3" />
10591 <edge from-layer="113" from-port="4" to-layer="115" to-port="0" />
10592 <edge from-layer="114" from-port="0" to-layer="115" to-port="1" />
10593 <edge from-layer="115" from-port="2" to-layer="122" to-port="0" />
10594 <edge from-layer="116" from-port="1" to-layer="119" to-port="0" />
10595 <edge from-layer="117" from-port="0" to-layer="119" to-port="1" />
10596 <edge from-layer="118" from-port="0" to-layer="119" to-port="2" />
10597 <edge from-layer="119" from-port="3" to-layer="121" to-port="0" />
10598 <edge from-layer="120" from-port="0" to-layer="121" to-port="1" />
10599 <edge from-layer="120" from-port="0" to-layer="469" to-port="1" />
10600 <edge from-layer="120" from-port="0" to-layer="585" to-port="1" />
10601 <edge from-layer="120" from-port="0" to-layer="237" to-port="1" />
10602 <edge from-layer="120" from-port="0" to-layer="701" to-port="1" />
10603 <edge from-layer="120" from-port="0" to-layer="353" to-port="1" />
10604 <edge from-layer="121" from-port="2" to-layer="122" to-port="1" />
10605 <edge from-layer="122" from-port="2" to-layer="124" to-port="0" />
10606 <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
10607 <edge from-layer="124" from-port="2" to-layer="129" to-port="0" />
10608 <edge from-layer="125" from-port="0" to-layer="129" to-port="1" />
10609 <edge from-layer="126" from-port="0" to-layer="129" to-port="2" />
10610 <edge from-layer="127" from-port="0" to-layer="129" to-port="3" />
10611 <edge from-layer="128" from-port="0" to-layer="129" to-port="4" />
10612 <edge from-layer="129" from-port="5" to-layer="134" to-port="0" />
10613 <edge from-layer="130" from-port="0" to-layer="131" to-port="0" />
10614 <edge from-layer="131" from-port="1" to-layer="133" to-port="0" />
10615 <edge from-layer="132" from-port="0" to-layer="133" to-port="1" />
10616 <edge from-layer="133" from-port="2" to-layer="134" to-port="1" />
10617 <edge from-layer="134" from-port="2" to-layer="136" to-port="0" />
10618 <edge from-layer="135" from-port="0" to-layer="136" to-port="1" />
10619 <edge from-layer="136" from-port="2" to-layer="137" to-port="0" />
10620 <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
10621 <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
10622 <edge from-layer="139" from-port="2" to-layer="141" to-port="0" />
10623 <edge from-layer="140" from-port="0" to-layer="141" to-port="1" />
10624 <edge from-layer="141" from-port="2" to-layer="143" to-port="0" />
10625 <edge from-layer="142" from-port="0" to-layer="143" to-port="1" />
10626 <edge from-layer="143" from-port="2" to-layer="173" to-port="1" />
10627 <edge from-layer="143" from-port="2" to-layer="145" to-port="0" />
10628 <edge from-layer="144" from-port="0" to-layer="145" to-port="1" />
10629 <edge from-layer="145" from-port="2" to-layer="150" to-port="0" />
10630 <edge from-layer="146" from-port="0" to-layer="150" to-port="1" />
10631 <edge from-layer="147" from-port="0" to-layer="150" to-port="2" />
10632 <edge from-layer="148" from-port="0" to-layer="150" to-port="3" />
10633 <edge from-layer="149" from-port="0" to-layer="150" to-port="4" />
10634 <edge from-layer="150" from-port="5" to-layer="155" to-port="0" />
10635 <edge from-layer="151" from-port="0" to-layer="152" to-port="0" />
10636 <edge from-layer="152" from-port="1" to-layer="154" to-port="0" />
10637 <edge from-layer="153" from-port="0" to-layer="154" to-port="1" />
10638 <edge from-layer="154" from-port="2" to-layer="155" to-port="1" />
10639 <edge from-layer="155" from-port="2" to-layer="157" to-port="0" />
10640 <edge from-layer="156" from-port="0" to-layer="157" to-port="1" />
10641 <edge from-layer="157" from-port="2" to-layer="158" to-port="0" />
10642 <edge from-layer="158" from-port="1" to-layer="160" to-port="0" />
10643 <edge from-layer="159" from-port="0" to-layer="160" to-port="1" />
10644 <edge from-layer="160" from-port="2" to-layer="165" to-port="0" />
10645 <edge from-layer="161" from-port="0" to-layer="165" to-port="1" />
10646 <edge from-layer="162" from-port="0" to-layer="165" to-port="2" />
10647 <edge from-layer="163" from-port="0" to-layer="165" to-port="3" />
10648 <edge from-layer="164" from-port="0" to-layer="165" to-port="4" />
10649 <edge from-layer="165" from-port="5" to-layer="170" to-port="0" />
10650 <edge from-layer="166" from-port="0" to-layer="167" to-port="0" />
10651 <edge from-layer="167" from-port="1" to-layer="169" to-port="0" />
10652 <edge from-layer="168" from-port="0" to-layer="169" to-port="1" />
10653 <edge from-layer="169" from-port="2" to-layer="170" to-port="1" />
10654 <edge from-layer="170" from-port="2" to-layer="172" to-port="0" />
10655 <edge from-layer="171" from-port="0" to-layer="172" to-port="1" />
10656 <edge from-layer="172" from-port="2" to-layer="173" to-port="0" />
10657 <edge from-layer="173" from-port="2" to-layer="175" to-port="0" />
10658 <edge from-layer="174" from-port="0" to-layer="175" to-port="1" />
10659 <edge from-layer="175" from-port="2" to-layer="177" to-port="0" />
10660 <edge from-layer="176" from-port="0" to-layer="177" to-port="1" />
10661 <edge from-layer="177" from-port="2" to-layer="179" to-port="0" />
10662 <edge from-layer="178" from-port="0" to-layer="179" to-port="1" />
10663 <edge from-layer="179" from-port="2" to-layer="181" to-port="0" />
10664 <edge from-layer="179" from-port="2" to-layer="233" to-port="0" />
10665 <edge from-layer="179" from-port="2" to-layer="253" to-port="1" />
10666 <edge from-layer="180" from-port="0" to-layer="181" to-port="1" />
10667 <edge from-layer="181" from-port="2" to-layer="186" to-port="0" />
10668 <edge from-layer="182" from-port="0" to-layer="186" to-port="1" />
10669 <edge from-layer="183" from-port="0" to-layer="186" to-port="2" />
10670 <edge from-layer="184" from-port="0" to-layer="186" to-port="3" />
10671 <edge from-layer="185" from-port="0" to-layer="186" to-port="4" />
10672 <edge from-layer="186" from-port="5" to-layer="191" to-port="0" />
10673 <edge from-layer="186" from-port="5" to-layer="207" to-port="0" />
10674 <edge from-layer="186" from-port="5" to-layer="223" to-port="0" />
10675 <edge from-layer="187" from-port="0" to-layer="188" to-port="0" />
10676 <edge from-layer="188" from-port="1" to-layer="190" to-port="0" />
10677 <edge from-layer="189" from-port="0" to-layer="190" to-port="1" />
10678 <edge from-layer="190" from-port="2" to-layer="191" to-port="1" />
10679 <edge from-layer="191" from-port="2" to-layer="193" to-port="0" />
10680 <edge from-layer="192" from-port="0" to-layer="193" to-port="1" />
10681 <edge from-layer="193" from-port="2" to-layer="198" to-port="0" />
10682 <edge from-layer="194" from-port="0" to-layer="198" to-port="1" />
10683 <edge from-layer="195" from-port="0" to-layer="198" to-port="2" />
10684 <edge from-layer="196" from-port="0" to-layer="198" to-port="3" />
10685 <edge from-layer="197" from-port="0" to-layer="198" to-port="4" />
10686 <edge from-layer="198" from-port="5" to-layer="200" to-port="0" />
10687 <edge from-layer="199" from-port="0" to-layer="200" to-port="1" />
10688 <edge from-layer="200" from-port="2" to-layer="202" to-port="0" />
10689 <edge from-layer="201" from-port="0" to-layer="202" to-port="1" />
10690 <edge from-layer="202" from-port="2" to-layer="230" to-port="0" />
10691 <edge from-layer="203" from-port="0" to-layer="204" to-port="0" />
10692 <edge from-layer="204" from-port="1" to-layer="206" to-port="0" />
10693 <edge from-layer="205" from-port="0" to-layer="206" to-port="1" />
10694 <edge from-layer="206" from-port="2" to-layer="207" to-port="1" />
10695 <edge from-layer="207" from-port="2" to-layer="209" to-port="0" />
10696 <edge from-layer="208" from-port="0" to-layer="209" to-port="1" />
10697 <edge from-layer="209" from-port="2" to-layer="214" to-port="0" />
10698 <edge from-layer="210" from-port="0" to-layer="214" to-port="1" />
10699 <edge from-layer="211" from-port="0" to-layer="214" to-port="2" />
10700 <edge from-layer="212" from-port="0" to-layer="214" to-port="3" />
10701 <edge from-layer="213" from-port="0" to-layer="214" to-port="4" />
10702 <edge from-layer="214" from-port="5" to-layer="216" to-port="0" />
10703 <edge from-layer="215" from-port="0" to-layer="216" to-port="1" />
10704 <edge from-layer="216" from-port="2" to-layer="218" to-port="0" />
10705 <edge from-layer="217" from-port="0" to-layer="218" to-port="1" />
10706 <edge from-layer="218" from-port="2" to-layer="230" to-port="1" />
10707 <edge from-layer="219" from-port="0" to-layer="220" to-port="0" />
10708 <edge from-layer="220" from-port="1" to-layer="222" to-port="0" />
10709 <edge from-layer="221" from-port="0" to-layer="222" to-port="1" />
10710 <edge from-layer="222" from-port="2" to-layer="223" to-port="1" />
10711 <edge from-layer="223" from-port="2" to-layer="225" to-port="0" />
10712 <edge from-layer="224" from-port="0" to-layer="225" to-port="1" />
10713 <edge from-layer="225" from-port="2" to-layer="227" to-port="0" />
10714 <edge from-layer="226" from-port="0" to-layer="227" to-port="1" />
10715 <edge from-layer="227" from-port="2" to-layer="229" to-port="0" />
10716 <edge from-layer="228" from-port="0" to-layer="229" to-port="1" />
10717 <edge from-layer="229" from-port="2" to-layer="230" to-port="2" />
10718 <edge from-layer="230" from-port="4" to-layer="232" to-port="0" />
10719 <edge from-layer="231" from-port="0" to-layer="232" to-port="1" />
10720 <edge from-layer="232" from-port="2" to-layer="238" to-port="0" />
10721 <edge from-layer="233" from-port="1" to-layer="236" to-port="0" />
10722 <edge from-layer="234" from-port="0" to-layer="236" to-port="1" />
10723 <edge from-layer="235" from-port="0" to-layer="236" to-port="2" />
10724 <edge from-layer="236" from-port="3" to-layer="237" to-port="0" />
10725 <edge from-layer="237" from-port="2" to-layer="238" to-port="1" />
10726 <edge from-layer="238" from-port="2" to-layer="240" to-port="0" />
10727 <edge from-layer="239" from-port="0" to-layer="240" to-port="1" />
10728 <edge from-layer="240" from-port="2" to-layer="245" to-port="0" />
10729 <edge from-layer="241" from-port="0" to-layer="245" to-port="1" />
10730 <edge from-layer="242" from-port="0" to-layer="245" to-port="2" />
10731 <edge from-layer="243" from-port="0" to-layer="245" to-port="3" />
10732 <edge from-layer="244" from-port="0" to-layer="245" to-port="4" />
10733 <edge from-layer="245" from-port="5" to-layer="250" to-port="0" />
10734 <edge from-layer="246" from-port="0" to-layer="247" to-port="0" />
10735 <edge from-layer="247" from-port="1" to-layer="249" to-port="0" />
10736 <edge from-layer="248" from-port="0" to-layer="249" to-port="1" />
10737 <edge from-layer="249" from-port="2" to-layer="250" to-port="1" />
10738 <edge from-layer="250" from-port="2" to-layer="252" to-port="0" />
10739 <edge from-layer="251" from-port="0" to-layer="252" to-port="1" />
10740 <edge from-layer="252" from-port="2" to-layer="253" to-port="0" />
10741 <edge from-layer="253" from-port="2" to-layer="255" to-port="0" />
10742 <edge from-layer="254" from-port="0" to-layer="255" to-port="1" />
10743 <edge from-layer="255" from-port="2" to-layer="257" to-port="0" />
10744 <edge from-layer="256" from-port="0" to-layer="257" to-port="1" />
10745 <edge from-layer="257" from-port="2" to-layer="259" to-port="0" />
10746 <edge from-layer="258" from-port="0" to-layer="259" to-port="1" />
10747 <edge from-layer="259" from-port="2" to-layer="289" to-port="1" />
10748 <edge from-layer="259" from-port="2" to-layer="261" to-port="0" />
10749 <edge from-layer="260" from-port="0" to-layer="261" to-port="1" />
10750 <edge from-layer="261" from-port="2" to-layer="266" to-port="0" />
10751 <edge from-layer="262" from-port="0" to-layer="266" to-port="1" />
10752 <edge from-layer="263" from-port="0" to-layer="266" to-port="2" />
10753 <edge from-layer="264" from-port="0" to-layer="266" to-port="3" />
10754 <edge from-layer="265" from-port="0" to-layer="266" to-port="4" />
10755 <edge from-layer="266" from-port="5" to-layer="271" to-port="0" />
10756 <edge from-layer="267" from-port="0" to-layer="268" to-port="0" />
10757 <edge from-layer="268" from-port="1" to-layer="270" to-port="0" />
10758 <edge from-layer="269" from-port="0" to-layer="270" to-port="1" />
10759 <edge from-layer="270" from-port="2" to-layer="271" to-port="1" />
10760 <edge from-layer="271" from-port="2" to-layer="273" to-port="0" />
10761 <edge from-layer="272" from-port="0" to-layer="273" to-port="1" />
10762 <edge from-layer="273" from-port="2" to-layer="274" to-port="0" />
10763 <edge from-layer="274" from-port="1" to-layer="276" to-port="0" />
10764 <edge from-layer="275" from-port="0" to-layer="276" to-port="1" />
10765 <edge from-layer="276" from-port="2" to-layer="281" to-port="0" />
10766 <edge from-layer="277" from-port="0" to-layer="281" to-port="1" />
10767 <edge from-layer="278" from-port="0" to-layer="281" to-port="2" />
10768 <edge from-layer="279" from-port="0" to-layer="281" to-port="3" />
10769 <edge from-layer="280" from-port="0" to-layer="281" to-port="4" />
10770 <edge from-layer="281" from-port="5" to-layer="286" to-port="0" />
10771 <edge from-layer="282" from-port="0" to-layer="283" to-port="0" />
10772 <edge from-layer="283" from-port="1" to-layer="285" to-port="0" />
10773 <edge from-layer="284" from-port="0" to-layer="285" to-port="1" />
10774 <edge from-layer="285" from-port="2" to-layer="286" to-port="1" />
10775 <edge from-layer="286" from-port="2" to-layer="288" to-port="0" />
10776 <edge from-layer="287" from-port="0" to-layer="288" to-port="1" />
10777 <edge from-layer="288" from-port="2" to-layer="289" to-port="0" />
10778 <edge from-layer="289" from-port="2" to-layer="291" to-port="0" />
10779 <edge from-layer="290" from-port="0" to-layer="291" to-port="1" />
10780 <edge from-layer="291" from-port="2" to-layer="293" to-port="0" />
10781 <edge from-layer="292" from-port="0" to-layer="293" to-port="1" />
10782 <edge from-layer="293" from-port="2" to-layer="295" to-port="0" />
10783 <edge from-layer="294" from-port="0" to-layer="295" to-port="1" />
10784 <edge from-layer="295" from-port="2" to-layer="297" to-port="0" />
10785 <edge from-layer="295" from-port="2" to-layer="349" to-port="0" />
10786 <edge from-layer="295" from-port="2" to-layer="369" to-port="1" />
10787 <edge from-layer="296" from-port="0" to-layer="297" to-port="1" />
10788 <edge from-layer="297" from-port="2" to-layer="302" to-port="0" />
10789 <edge from-layer="298" from-port="0" to-layer="302" to-port="1" />
10790 <edge from-layer="299" from-port="0" to-layer="302" to-port="2" />
10791 <edge from-layer="300" from-port="0" to-layer="302" to-port="3" />
10792 <edge from-layer="301" from-port="0" to-layer="302" to-port="4" />
10793 <edge from-layer="302" from-port="5" to-layer="323" to-port="0" />
10794 <edge from-layer="302" from-port="5" to-layer="339" to-port="0" />
10795 <edge from-layer="302" from-port="5" to-layer="307" to-port="0" />
10796 <edge from-layer="303" from-port="0" to-layer="304" to-port="0" />
10797 <edge from-layer="304" from-port="1" to-layer="306" to-port="0" />
10798 <edge from-layer="305" from-port="0" to-layer="306" to-port="1" />
10799 <edge from-layer="306" from-port="2" to-layer="307" to-port="1" />
10800 <edge from-layer="307" from-port="2" to-layer="309" to-port="0" />
10801 <edge from-layer="308" from-port="0" to-layer="309" to-port="1" />
10802 <edge from-layer="309" from-port="2" to-layer="314" to-port="0" />
10803 <edge from-layer="310" from-port="0" to-layer="314" to-port="1" />
10804 <edge from-layer="311" from-port="0" to-layer="314" to-port="2" />
10805 <edge from-layer="312" from-port="0" to-layer="314" to-port="3" />
10806 <edge from-layer="313" from-port="0" to-layer="314" to-port="4" />
10807 <edge from-layer="314" from-port="5" to-layer="316" to-port="0" />
10808 <edge from-layer="315" from-port="0" to-layer="316" to-port="1" />
10809 <edge from-layer="316" from-port="2" to-layer="318" to-port="0" />
10810 <edge from-layer="317" from-port="0" to-layer="318" to-port="1" />
10811 <edge from-layer="318" from-port="2" to-layer="346" to-port="0" />
10812 <edge from-layer="319" from-port="0" to-layer="320" to-port="0" />
10813 <edge from-layer="320" from-port="1" to-layer="322" to-port="0" />
10814 <edge from-layer="321" from-port="0" to-layer="322" to-port="1" />
10815 <edge from-layer="322" from-port="2" to-layer="323" to-port="1" />
10816 <edge from-layer="323" from-port="2" to-layer="325" to-port="0" />
10817 <edge from-layer="324" from-port="0" to-layer="325" to-port="1" />
10818 <edge from-layer="325" from-port="2" to-layer="330" to-port="0" />
10819 <edge from-layer="326" from-port="0" to-layer="330" to-port="1" />
10820 <edge from-layer="327" from-port="0" to-layer="330" to-port="2" />
10821 <edge from-layer="328" from-port="0" to-layer="330" to-port="3" />
10822 <edge from-layer="329" from-port="0" to-layer="330" to-port="4" />
10823 <edge from-layer="330" from-port="5" to-layer="332" to-port="0" />
10824 <edge from-layer="331" from-port="0" to-layer="332" to-port="1" />
10825 <edge from-layer="332" from-port="2" to-layer="334" to-port="0" />
10826 <edge from-layer="333" from-port="0" to-layer="334" to-port="1" />
10827 <edge from-layer="334" from-port="2" to-layer="346" to-port="1" />
10828 <edge from-layer="335" from-port="0" to-layer="336" to-port="0" />
10829 <edge from-layer="336" from-port="1" to-layer="338" to-port="0" />
10830 <edge from-layer="337" from-port="0" to-layer="338" to-port="1" />
10831 <edge from-layer="338" from-port="2" to-layer="339" to-port="1" />
10832 <edge from-layer="339" from-port="2" to-layer="341" to-port="0" />
10833 <edge from-layer="340" from-port="0" to-layer="341" to-port="1" />
10834 <edge from-layer="341" from-port="2" to-layer="343" to-port="0" />
10835 <edge from-layer="342" from-port="0" to-layer="343" to-port="1" />
10836 <edge from-layer="343" from-port="2" to-layer="345" to-port="0" />
10837 <edge from-layer="344" from-port="0" to-layer="345" to-port="1" />
10838 <edge from-layer="345" from-port="2" to-layer="346" to-port="2" />
10839 <edge from-layer="346" from-port="4" to-layer="348" to-port="0" />
10840 <edge from-layer="347" from-port="0" to-layer="348" to-port="1" />
10841 <edge from-layer="348" from-port="2" to-layer="354" to-port="0" />
10842 <edge from-layer="349" from-port="1" to-layer="352" to-port="0" />
10843 <edge from-layer="350" from-port="0" to-layer="352" to-port="1" />
10844 <edge from-layer="351" from-port="0" to-layer="352" to-port="2" />
10845 <edge from-layer="352" from-port="3" to-layer="353" to-port="0" />
10846 <edge from-layer="353" from-port="2" to-layer="354" to-port="1" />
10847 <edge from-layer="354" from-port="2" to-layer="356" to-port="0" />
10848 <edge from-layer="355" from-port="0" to-layer="356" to-port="1" />
10849 <edge from-layer="356" from-port="2" to-layer="361" to-port="0" />
10850 <edge from-layer="357" from-port="0" to-layer="361" to-port="1" />
10851 <edge from-layer="358" from-port="0" to-layer="361" to-port="2" />
10852 <edge from-layer="359" from-port="0" to-layer="361" to-port="3" />
10853 <edge from-layer="360" from-port="0" to-layer="361" to-port="4" />
10854 <edge from-layer="361" from-port="5" to-layer="366" to-port="0" />
10855 <edge from-layer="362" from-port="0" to-layer="363" to-port="0" />
10856 <edge from-layer="363" from-port="1" to-layer="365" to-port="0" />
10857 <edge from-layer="364" from-port="0" to-layer="365" to-port="1" />
10858 <edge from-layer="365" from-port="2" to-layer="366" to-port="1" />
10859 <edge from-layer="366" from-port="2" to-layer="368" to-port="0" />
10860 <edge from-layer="367" from-port="0" to-layer="368" to-port="1" />
10861 <edge from-layer="368" from-port="2" to-layer="369" to-port="0" />
10862 <edge from-layer="369" from-port="2" to-layer="371" to-port="0" />
10863 <edge from-layer="370" from-port="0" to-layer="371" to-port="1" />
10864 <edge from-layer="371" from-port="2" to-layer="373" to-port="0" />
10865 <edge from-layer="372" from-port="0" to-layer="373" to-port="1" />
10866 <edge from-layer="373" from-port="2" to-layer="375" to-port="0" />
10867 <edge from-layer="374" from-port="0" to-layer="375" to-port="1" />
10868 <edge from-layer="375" from-port="2" to-layer="405" to-port="1" />
10869 <edge from-layer="375" from-port="2" to-layer="377" to-port="0" />
10870 <edge from-layer="376" from-port="0" to-layer="377" to-port="1" />
10871 <edge from-layer="377" from-port="2" to-layer="382" to-port="0" />
10872 <edge from-layer="378" from-port="0" to-layer="382" to-port="1" />
10873 <edge from-layer="379" from-port="0" to-layer="382" to-port="2" />
10874 <edge from-layer="380" from-port="0" to-layer="382" to-port="3" />
10875 <edge from-layer="381" from-port="0" to-layer="382" to-port="4" />
10876 <edge from-layer="382" from-port="5" to-layer="387" to-port="0" />
10877 <edge from-layer="383" from-port="0" to-layer="384" to-port="0" />
10878 <edge from-layer="384" from-port="1" to-layer="386" to-port="0" />
10879 <edge from-layer="385" from-port="0" to-layer="386" to-port="1" />
10880 <edge from-layer="386" from-port="2" to-layer="387" to-port="1" />
10881 <edge from-layer="387" from-port="2" to-layer="389" to-port="0" />
10882 <edge from-layer="388" from-port="0" to-layer="389" to-port="1" />
10883 <edge from-layer="389" from-port="2" to-layer="390" to-port="0" />
10884 <edge from-layer="390" from-port="1" to-layer="392" to-port="0" />
10885 <edge from-layer="391" from-port="0" to-layer="392" to-port="1" />
10886 <edge from-layer="392" from-port="2" to-layer="397" to-port="0" />
10887 <edge from-layer="393" from-port="0" to-layer="397" to-port="1" />
10888 <edge from-layer="394" from-port="0" to-layer="397" to-port="2" />
10889 <edge from-layer="395" from-port="0" to-layer="397" to-port="3" />
10890 <edge from-layer="396" from-port="0" to-layer="397" to-port="4" />
10891 <edge from-layer="397" from-port="5" to-layer="402" to-port="0" />
10892 <edge from-layer="398" from-port="0" to-layer="399" to-port="0" />
10893 <edge from-layer="399" from-port="1" to-layer="401" to-port="0" />
10894 <edge from-layer="400" from-port="0" to-layer="401" to-port="1" />
10895 <edge from-layer="401" from-port="2" to-layer="402" to-port="1" />
10896 <edge from-layer="402" from-port="2" to-layer="404" to-port="0" />
10897 <edge from-layer="403" from-port="0" to-layer="404" to-port="1" />
10898 <edge from-layer="404" from-port="2" to-layer="405" to-port="0" />
10899 <edge from-layer="405" from-port="2" to-layer="407" to-port="0" />
10900 <edge from-layer="406" from-port="0" to-layer="407" to-port="1" />
10901 <edge from-layer="407" from-port="2" to-layer="409" to-port="0" />
10902 <edge from-layer="408" from-port="0" to-layer="409" to-port="1" />
10903 <edge from-layer="409" from-port="2" to-layer="411" to-port="0" />
10904 <edge from-layer="410" from-port="0" to-layer="411" to-port="1" />
10905 <edge from-layer="411" from-port="2" to-layer="413" to-port="0" />
10906 <edge from-layer="411" from-port="2" to-layer="465" to-port="0" />
10907 <edge from-layer="411" from-port="2" to-layer="485" to-port="1" />
10908 <edge from-layer="412" from-port="0" to-layer="413" to-port="1" />
10909 <edge from-layer="413" from-port="2" to-layer="418" to-port="0" />
10910 <edge from-layer="414" from-port="0" to-layer="418" to-port="1" />
10911 <edge from-layer="415" from-port="0" to-layer="418" to-port="2" />
10912 <edge from-layer="416" from-port="0" to-layer="418" to-port="3" />
10913 <edge from-layer="417" from-port="0" to-layer="418" to-port="4" />
10914 <edge from-layer="418" from-port="5" to-layer="423" to-port="0" />
10915 <edge from-layer="418" from-port="5" to-layer="439" to-port="0" />
10916 <edge from-layer="418" from-port="5" to-layer="455" to-port="0" />
10917 <edge from-layer="419" from-port="0" to-layer="420" to-port="0" />
10918 <edge from-layer="420" from-port="1" to-layer="422" to-port="0" />
10919 <edge from-layer="421" from-port="0" to-layer="422" to-port="1" />
10920 <edge from-layer="422" from-port="2" to-layer="423" to-port="1" />
10921 <edge from-layer="423" from-port="2" to-layer="425" to-port="0" />
10922 <edge from-layer="424" from-port="0" to-layer="425" to-port="1" />
10923 <edge from-layer="425" from-port="2" to-layer="430" to-port="0" />
10924 <edge from-layer="426" from-port="0" to-layer="430" to-port="1" />
10925 <edge from-layer="427" from-port="0" to-layer="430" to-port="2" />
10926 <edge from-layer="428" from-port="0" to-layer="430" to-port="3" />
10927 <edge from-layer="429" from-port="0" to-layer="430" to-port="4" />
10928 <edge from-layer="430" from-port="5" to-layer="432" to-port="0" />
10929 <edge from-layer="431" from-port="0" to-layer="432" to-port="1" />
10930 <edge from-layer="432" from-port="2" to-layer="434" to-port="0" />
10931 <edge from-layer="433" from-port="0" to-layer="434" to-port="1" />
10932 <edge from-layer="434" from-port="2" to-layer="462" to-port="0" />
10933 <edge from-layer="435" from-port="0" to-layer="436" to-port="0" />
10934 <edge from-layer="436" from-port="1" to-layer="438" to-port="0" />
10935 <edge from-layer="437" from-port="0" to-layer="438" to-port="1" />
10936 <edge from-layer="438" from-port="2" to-layer="439" to-port="1" />
10937 <edge from-layer="439" from-port="2" to-layer="441" to-port="0" />
10938 <edge from-layer="440" from-port="0" to-layer="441" to-port="1" />
10939 <edge from-layer="441" from-port="2" to-layer="446" to-port="0" />
10940 <edge from-layer="442" from-port="0" to-layer="446" to-port="1" />
10941 <edge from-layer="443" from-port="0" to-layer="446" to-port="2" />
10942 <edge from-layer="444" from-port="0" to-layer="446" to-port="3" />
10943 <edge from-layer="445" from-port="0" to-layer="446" to-port="4" />
10944 <edge from-layer="446" from-port="5" to-layer="448" to-port="0" />
10945 <edge from-layer="447" from-port="0" to-layer="448" to-port="1" />
10946 <edge from-layer="448" from-port="2" to-layer="450" to-port="0" />
10947 <edge from-layer="449" from-port="0" to-layer="450" to-port="1" />
10948 <edge from-layer="450" from-port="2" to-layer="462" to-port="1" />
10949 <edge from-layer="451" from-port="0" to-layer="452" to-port="0" />
10950 <edge from-layer="452" from-port="1" to-layer="454" to-port="0" />
10951 <edge from-layer="453" from-port="0" to-layer="454" to-port="1" />
10952 <edge from-layer="454" from-port="2" to-layer="455" to-port="1" />
10953 <edge from-layer="455" from-port="2" to-layer="457" to-port="0" />
10954 <edge from-layer="456" from-port="0" to-layer="457" to-port="1" />
10955 <edge from-layer="457" from-port="2" to-layer="459" to-port="0" />
10956 <edge from-layer="458" from-port="0" to-layer="459" to-port="1" />
10957 <edge from-layer="459" from-port="2" to-layer="461" to-port="0" />
10958 <edge from-layer="460" from-port="0" to-layer="461" to-port="1" />
10959 <edge from-layer="461" from-port="2" to-layer="462" to-port="2" />
10960 <edge from-layer="462" from-port="4" to-layer="464" to-port="0" />
10961 <edge from-layer="463" from-port="0" to-layer="464" to-port="1" />
10962 <edge from-layer="464" from-port="2" to-layer="470" to-port="0" />
10963 <edge from-layer="465" from-port="1" to-layer="468" to-port="0" />
10964 <edge from-layer="466" from-port="0" to-layer="468" to-port="1" />
10965 <edge from-layer="467" from-port="0" to-layer="468" to-port="2" />
10966 <edge from-layer="468" from-port="3" to-layer="469" to-port="0" />
10967 <edge from-layer="469" from-port="2" to-layer="470" to-port="1" />
10968 <edge from-layer="470" from-port="2" to-layer="472" to-port="0" />
10969 <edge from-layer="471" from-port="0" to-layer="472" to-port="1" />
10970 <edge from-layer="472" from-port="2" to-layer="477" to-port="0" />
10971 <edge from-layer="473" from-port="0" to-layer="477" to-port="1" />
10972 <edge from-layer="474" from-port="0" to-layer="477" to-port="2" />
10973 <edge from-layer="475" from-port="0" to-layer="477" to-port="3" />
10974 <edge from-layer="476" from-port="0" to-layer="477" to-port="4" />
10975 <edge from-layer="477" from-port="5" to-layer="482" to-port="0" />
10976 <edge from-layer="478" from-port="0" to-layer="479" to-port="0" />
10977 <edge from-layer="479" from-port="1" to-layer="481" to-port="0" />
10978 <edge from-layer="480" from-port="0" to-layer="481" to-port="1" />
10979 <edge from-layer="481" from-port="2" to-layer="482" to-port="1" />
10980 <edge from-layer="482" from-port="2" to-layer="484" to-port="0" />
10981 <edge from-layer="483" from-port="0" to-layer="484" to-port="1" />
10982 <edge from-layer="484" from-port="2" to-layer="485" to-port="0" />
10983 <edge from-layer="485" from-port="2" to-layer="487" to-port="0" />
10984 <edge from-layer="486" from-port="0" to-layer="487" to-port="1" />
10985 <edge from-layer="487" from-port="2" to-layer="489" to-port="0" />
10986 <edge from-layer="488" from-port="0" to-layer="489" to-port="1" />
10987 <edge from-layer="489" from-port="2" to-layer="491" to-port="0" />
10988 <edge from-layer="490" from-port="0" to-layer="491" to-port="1" />
10989 <edge from-layer="491" from-port="2" to-layer="493" to-port="0" />
10990 <edge from-layer="491" from-port="2" to-layer="521" to-port="1" />
10991 <edge from-layer="492" from-port="0" to-layer="493" to-port="1" />
10992 <edge from-layer="493" from-port="2" to-layer="498" to-port="0" />
10993 <edge from-layer="494" from-port="0" to-layer="498" to-port="1" />
10994 <edge from-layer="495" from-port="0" to-layer="498" to-port="2" />
10995 <edge from-layer="496" from-port="0" to-layer="498" to-port="3" />
10996 <edge from-layer="497" from-port="0" to-layer="498" to-port="4" />
10997 <edge from-layer="498" from-port="5" to-layer="503" to-port="0" />
10998 <edge from-layer="499" from-port="0" to-layer="500" to-port="0" />
10999 <edge from-layer="500" from-port="1" to-layer="502" to-port="0" />
11000 <edge from-layer="501" from-port="0" to-layer="502" to-port="1" />
11001 <edge from-layer="502" from-port="2" to-layer="503" to-port="1" />
11002 <edge from-layer="503" from-port="2" to-layer="505" to-port="0" />
11003 <edge from-layer="504" from-port="0" to-layer="505" to-port="1" />
11004 <edge from-layer="505" from-port="2" to-layer="506" to-port="0" />
11005 <edge from-layer="506" from-port="1" to-layer="508" to-port="0" />
11006 <edge from-layer="507" from-port="0" to-layer="508" to-port="1" />
11007 <edge from-layer="508" from-port="2" to-layer="513" to-port="0" />
11008 <edge from-layer="509" from-port="0" to-layer="513" to-port="1" />
11009 <edge from-layer="510" from-port="0" to-layer="513" to-port="2" />
11010 <edge from-layer="511" from-port="0" to-layer="513" to-port="3" />
11011 <edge from-layer="512" from-port="0" to-layer="513" to-port="4" />
11012 <edge from-layer="513" from-port="5" to-layer="518" to-port="0" />
11013 <edge from-layer="514" from-port="0" to-layer="515" to-port="0" />
11014 <edge from-layer="515" from-port="1" to-layer="517" to-port="0" />
11015 <edge from-layer="516" from-port="0" to-layer="517" to-port="1" />
11016 <edge from-layer="517" from-port="2" to-layer="518" to-port="1" />
11017 <edge from-layer="518" from-port="2" to-layer="520" to-port="0" />
11018 <edge from-layer="519" from-port="0" to-layer="520" to-port="1" />
11019 <edge from-layer="520" from-port="2" to-layer="521" to-port="0" />
11020 <edge from-layer="521" from-port="2" to-layer="523" to-port="0" />
11021 <edge from-layer="522" from-port="0" to-layer="523" to-port="1" />
11022 <edge from-layer="523" from-port="2" to-layer="525" to-port="0" />
11023 <edge from-layer="524" from-port="0" to-layer="525" to-port="1" />
11024 <edge from-layer="525" from-port="2" to-layer="527" to-port="0" />
11025 <edge from-layer="526" from-port="0" to-layer="527" to-port="1" />
11026 <edge from-layer="527" from-port="2" to-layer="529" to-port="0" />
11027 <edge from-layer="527" from-port="2" to-layer="581" to-port="0" />
11028 <edge from-layer="527" from-port="2" to-layer="601" to-port="1" />
11029 <edge from-layer="528" from-port="0" to-layer="529" to-port="1" />
11030 <edge from-layer="529" from-port="2" to-layer="534" to-port="0" />
11031 <edge from-layer="530" from-port="0" to-layer="534" to-port="1" />
11032 <edge from-layer="531" from-port="0" to-layer="534" to-port="2" />
11033 <edge from-layer="532" from-port="0" to-layer="534" to-port="3" />
11034 <edge from-layer="533" from-port="0" to-layer="534" to-port="4" />
11035 <edge from-layer="534" from-port="5" to-layer="539" to-port="0" />
11036 <edge from-layer="534" from-port="5" to-layer="571" to-port="0" />
11037 <edge from-layer="534" from-port="5" to-layer="555" to-port="0" />
11038 <edge from-layer="535" from-port="0" to-layer="536" to-port="0" />
11039 <edge from-layer="536" from-port="1" to-layer="538" to-port="0" />
11040 <edge from-layer="537" from-port="0" to-layer="538" to-port="1" />
11041 <edge from-layer="538" from-port="2" to-layer="539" to-port="1" />
11042 <edge from-layer="539" from-port="2" to-layer="541" to-port="0" />
11043 <edge from-layer="540" from-port="0" to-layer="541" to-port="1" />
11044 <edge from-layer="541" from-port="2" to-layer="546" to-port="0" />
11045 <edge from-layer="542" from-port="0" to-layer="546" to-port="1" />
11046 <edge from-layer="543" from-port="0" to-layer="546" to-port="2" />
11047 <edge from-layer="544" from-port="0" to-layer="546" to-port="3" />
11048 <edge from-layer="545" from-port="0" to-layer="546" to-port="4" />
11049 <edge from-layer="546" from-port="5" to-layer="548" to-port="0" />
11050 <edge from-layer="547" from-port="0" to-layer="548" to-port="1" />
11051 <edge from-layer="548" from-port="2" to-layer="550" to-port="0" />
11052 <edge from-layer="549" from-port="0" to-layer="550" to-port="1" />
11053 <edge from-layer="550" from-port="2" to-layer="578" to-port="0" />
11054 <edge from-layer="551" from-port="0" to-layer="552" to-port="0" />
11055 <edge from-layer="552" from-port="1" to-layer="554" to-port="0" />
11056 <edge from-layer="553" from-port="0" to-layer="554" to-port="1" />
11057 <edge from-layer="554" from-port="2" to-layer="555" to-port="1" />
11058 <edge from-layer="555" from-port="2" to-layer="557" to-port="0" />
11059 <edge from-layer="556" from-port="0" to-layer="557" to-port="1" />
11060 <edge from-layer="557" from-port="2" to-layer="562" to-port="0" />
11061 <edge from-layer="558" from-port="0" to-layer="562" to-port="1" />
11062 <edge from-layer="559" from-port="0" to-layer="562" to-port="2" />
11063 <edge from-layer="560" from-port="0" to-layer="562" to-port="3" />
11064 <edge from-layer="561" from-port="0" to-layer="562" to-port="4" />
11065 <edge from-layer="562" from-port="5" to-layer="564" to-port="0" />
11066 <edge from-layer="563" from-port="0" to-layer="564" to-port="1" />
11067 <edge from-layer="564" from-port="2" to-layer="566" to-port="0" />
11068 <edge from-layer="565" from-port="0" to-layer="566" to-port="1" />
11069 <edge from-layer="566" from-port="2" to-layer="578" to-port="1" />
11070 <edge from-layer="567" from-port="0" to-layer="568" to-port="0" />
11071 <edge from-layer="568" from-port="1" to-layer="570" to-port="0" />
11072 <edge from-layer="569" from-port="0" to-layer="570" to-port="1" />
11073 <edge from-layer="570" from-port="2" to-layer="571" to-port="1" />
11074 <edge from-layer="571" from-port="2" to-layer="573" to-port="0" />
11075 <edge from-layer="572" from-port="0" to-layer="573" to-port="1" />
11076 <edge from-layer="573" from-port="2" to-layer="575" to-port="0" />
11077 <edge from-layer="574" from-port="0" to-layer="575" to-port="1" />
11078 <edge from-layer="575" from-port="2" to-layer="577" to-port="0" />
11079 <edge from-layer="576" from-port="0" to-layer="577" to-port="1" />
11080 <edge from-layer="577" from-port="2" to-layer="578" to-port="2" />
11081 <edge from-layer="578" from-port="4" to-layer="580" to-port="0" />
11082 <edge from-layer="579" from-port="0" to-layer="580" to-port="1" />
11083 <edge from-layer="580" from-port="2" to-layer="586" to-port="0" />
11084 <edge from-layer="581" from-port="1" to-layer="584" to-port="0" />
11085 <edge from-layer="582" from-port="0" to-layer="584" to-port="1" />
11086 <edge from-layer="583" from-port="0" to-layer="584" to-port="2" />
11087 <edge from-layer="584" from-port="3" to-layer="585" to-port="0" />
11088 <edge from-layer="585" from-port="2" to-layer="586" to-port="1" />
11089 <edge from-layer="586" from-port="2" to-layer="588" to-port="0" />
11090 <edge from-layer="587" from-port="0" to-layer="588" to-port="1" />
11091 <edge from-layer="588" from-port="2" to-layer="593" to-port="0" />
11092 <edge from-layer="589" from-port="0" to-layer="593" to-port="1" />
11093 <edge from-layer="590" from-port="0" to-layer="593" to-port="2" />
11094 <edge from-layer="591" from-port="0" to-layer="593" to-port="3" />
11095 <edge from-layer="592" from-port="0" to-layer="593" to-port="4" />
11096 <edge from-layer="593" from-port="5" to-layer="598" to-port="0" />
11097 <edge from-layer="594" from-port="0" to-layer="595" to-port="0" />
11098 <edge from-layer="595" from-port="1" to-layer="597" to-port="0" />
11099 <edge from-layer="596" from-port="0" to-layer="597" to-port="1" />
11100 <edge from-layer="597" from-port="2" to-layer="598" to-port="1" />
11101 <edge from-layer="598" from-port="2" to-layer="600" to-port="0" />
11102 <edge from-layer="599" from-port="0" to-layer="600" to-port="1" />
11103 <edge from-layer="600" from-port="2" to-layer="601" to-port="0" />
11104 <edge from-layer="601" from-port="2" to-layer="603" to-port="0" />
11105 <edge from-layer="602" from-port="0" to-layer="603" to-port="1" />
11106 <edge from-layer="603" from-port="2" to-layer="605" to-port="0" />
11107 <edge from-layer="604" from-port="0" to-layer="605" to-port="1" />
11108 <edge from-layer="605" from-port="2" to-layer="607" to-port="0" />
11109 <edge from-layer="606" from-port="0" to-layer="607" to-port="1" />
11110 <edge from-layer="607" from-port="2" to-layer="609" to-port="0" />
11111 <edge from-layer="607" from-port="2" to-layer="637" to-port="1" />
11112 <edge from-layer="608" from-port="0" to-layer="609" to-port="1" />
11113 <edge from-layer="609" from-port="2" to-layer="614" to-port="0" />
11114 <edge from-layer="610" from-port="0" to-layer="614" to-port="1" />
11115 <edge from-layer="611" from-port="0" to-layer="614" to-port="2" />
11116 <edge from-layer="612" from-port="0" to-layer="614" to-port="3" />
11117 <edge from-layer="613" from-port="0" to-layer="614" to-port="4" />
11118 <edge from-layer="614" from-port="5" to-layer="619" to-port="0" />
11119 <edge from-layer="615" from-port="0" to-layer="616" to-port="0" />
11120 <edge from-layer="616" from-port="1" to-layer="618" to-port="0" />
11121 <edge from-layer="617" from-port="0" to-layer="618" to-port="1" />
11122 <edge from-layer="618" from-port="2" to-layer="619" to-port="1" />
11123 <edge from-layer="619" from-port="2" to-layer="621" to-port="0" />
11124 <edge from-layer="620" from-port="0" to-layer="621" to-port="1" />
11125 <edge from-layer="621" from-port="2" to-layer="622" to-port="0" />
11126 <edge from-layer="622" from-port="1" to-layer="624" to-port="0" />
11127 <edge from-layer="623" from-port="0" to-layer="624" to-port="1" />
11128 <edge from-layer="624" from-port="2" to-layer="629" to-port="0" />
11129 <edge from-layer="625" from-port="0" to-layer="629" to-port="1" />
11130 <edge from-layer="626" from-port="0" to-layer="629" to-port="2" />
11131 <edge from-layer="627" from-port="0" to-layer="629" to-port="3" />
11132 <edge from-layer="628" from-port="0" to-layer="629" to-port="4" />
11133 <edge from-layer="629" from-port="5" to-layer="634" to-port="0" />
11134 <edge from-layer="630" from-port="0" to-layer="631" to-port="0" />
11135 <edge from-layer="631" from-port="1" to-layer="633" to-port="0" />
11136 <edge from-layer="632" from-port="0" to-layer="633" to-port="1" />
11137 <edge from-layer="633" from-port="2" to-layer="634" to-port="1" />
11138 <edge from-layer="634" from-port="2" to-layer="636" to-port="0" />
11139 <edge from-layer="635" from-port="0" to-layer="636" to-port="1" />
11140 <edge from-layer="636" from-port="2" to-layer="637" to-port="0" />
11141 <edge from-layer="637" from-port="2" to-layer="639" to-port="0" />
11142 <edge from-layer="638" from-port="0" to-layer="639" to-port="1" />
11143 <edge from-layer="639" from-port="2" to-layer="641" to-port="0" />
11144 <edge from-layer="640" from-port="0" to-layer="641" to-port="1" />
11145 <edge from-layer="641" from-port="2" to-layer="643" to-port="0" />
11146 <edge from-layer="642" from-port="0" to-layer="643" to-port="1" />
11147 <edge from-layer="643" from-port="2" to-layer="645" to-port="0" />
11148 <edge from-layer="643" from-port="2" to-layer="717" to-port="1" />
11149 <edge from-layer="643" from-port="2" to-layer="697" to-port="0" />
11150 <edge from-layer="644" from-port="0" to-layer="645" to-port="1" />
11151 <edge from-layer="645" from-port="2" to-layer="650" to-port="0" />
11152 <edge from-layer="646" from-port="0" to-layer="650" to-port="1" />
11153 <edge from-layer="647" from-port="0" to-layer="650" to-port="2" />
11154 <edge from-layer="648" from-port="0" to-layer="650" to-port="3" />
11155 <edge from-layer="649" from-port="0" to-layer="650" to-port="4" />
11156 <edge from-layer="650" from-port="5" to-layer="655" to-port="0" />
11157 <edge from-layer="650" from-port="5" to-layer="687" to-port="0" />
11158 <edge from-layer="650" from-port="5" to-layer="671" to-port="0" />
11159 <edge from-layer="651" from-port="0" to-layer="652" to-port="0" />
11160 <edge from-layer="652" from-port="1" to-layer="654" to-port="0" />
11161 <edge from-layer="653" from-port="0" to-layer="654" to-port="1" />
11162 <edge from-layer="654" from-port="2" to-layer="655" to-port="1" />
11163 <edge from-layer="655" from-port="2" to-layer="657" to-port="0" />
11164 <edge from-layer="656" from-port="0" to-layer="657" to-port="1" />
11165 <edge from-layer="657" from-port="2" to-layer="662" to-port="0" />
11166 <edge from-layer="658" from-port="0" to-layer="662" to-port="1" />
11167 <edge from-layer="659" from-port="0" to-layer="662" to-port="2" />
11168 <edge from-layer="660" from-port="0" to-layer="662" to-port="3" />
11169 <edge from-layer="661" from-port="0" to-layer="662" to-port="4" />
11170 <edge from-layer="662" from-port="5" to-layer="664" to-port="0" />
11171 <edge from-layer="663" from-port="0" to-layer="664" to-port="1" />
11172 <edge from-layer="664" from-port="2" to-layer="666" to-port="0" />
11173 <edge from-layer="665" from-port="0" to-layer="666" to-port="1" />
11174 <edge from-layer="666" from-port="2" to-layer="694" to-port="0" />
11175 <edge from-layer="667" from-port="0" to-layer="668" to-port="0" />
11176 <edge from-layer="668" from-port="1" to-layer="670" to-port="0" />
11177 <edge from-layer="669" from-port="0" to-layer="670" to-port="1" />
11178 <edge from-layer="670" from-port="2" to-layer="671" to-port="1" />
11179 <edge from-layer="671" from-port="2" to-layer="673" to-port="0" />
11180 <edge from-layer="672" from-port="0" to-layer="673" to-port="1" />
11181 <edge from-layer="673" from-port="2" to-layer="678" to-port="0" />
11182 <edge from-layer="674" from-port="0" to-layer="678" to-port="1" />
11183 <edge from-layer="675" from-port="0" to-layer="678" to-port="2" />
11184 <edge from-layer="676" from-port="0" to-layer="678" to-port="3" />
11185 <edge from-layer="677" from-port="0" to-layer="678" to-port="4" />
11186 <edge from-layer="678" from-port="5" to-layer="680" to-port="0" />
11187 <edge from-layer="679" from-port="0" to-layer="680" to-port="1" />
11188 <edge from-layer="680" from-port="2" to-layer="682" to-port="0" />
11189 <edge from-layer="681" from-port="0" to-layer="682" to-port="1" />
11190 <edge from-layer="682" from-port="2" to-layer="694" to-port="1" />
11191 <edge from-layer="683" from-port="0" to-layer="684" to-port="0" />
11192 <edge from-layer="684" from-port="1" to-layer="686" to-port="0" />
11193 <edge from-layer="685" from-port="0" to-layer="686" to-port="1" />
11194 <edge from-layer="686" from-port="2" to-layer="687" to-port="1" />
11195 <edge from-layer="687" from-port="2" to-layer="689" to-port="0" />
11196 <edge from-layer="688" from-port="0" to-layer="689" to-port="1" />
11197 <edge from-layer="689" from-port="2" to-layer="691" to-port="0" />
11198 <edge from-layer="690" from-port="0" to-layer="691" to-port="1" />
11199 <edge from-layer="691" from-port="2" to-layer="693" to-port="0" />
11200 <edge from-layer="692" from-port="0" to-layer="693" to-port="1" />
11201 <edge from-layer="693" from-port="2" to-layer="694" to-port="2" />
11202 <edge from-layer="694" from-port="4" to-layer="696" to-port="0" />
11203 <edge from-layer="695" from-port="0" to-layer="696" to-port="1" />
11204 <edge from-layer="696" from-port="2" to-layer="702" to-port="0" />
11205 <edge from-layer="697" from-port="1" to-layer="700" to-port="0" />
11206 <edge from-layer="698" from-port="0" to-layer="700" to-port="1" />
11207 <edge from-layer="699" from-port="0" to-layer="700" to-port="2" />
11208 <edge from-layer="700" from-port="3" to-layer="701" to-port="0" />
11209 <edge from-layer="701" from-port="2" to-layer="702" to-port="1" />
11210 <edge from-layer="702" from-port="2" to-layer="704" to-port="0" />
11211 <edge from-layer="703" from-port="0" to-layer="704" to-port="1" />
11212 <edge from-layer="704" from-port="2" to-layer="709" to-port="0" />
11213 <edge from-layer="705" from-port="0" to-layer="709" to-port="1" />
11214 <edge from-layer="706" from-port="0" to-layer="709" to-port="2" />
11215 <edge from-layer="707" from-port="0" to-layer="709" to-port="3" />
11216 <edge from-layer="708" from-port="0" to-layer="709" to-port="4" />
11217 <edge from-layer="709" from-port="5" to-layer="714" to-port="0" />
11218 <edge from-layer="710" from-port="0" to-layer="711" to-port="0" />
11219 <edge from-layer="711" from-port="1" to-layer="713" to-port="0" />
11220 <edge from-layer="712" from-port="0" to-layer="713" to-port="1" />
11221 <edge from-layer="713" from-port="2" to-layer="714" to-port="1" />
11222 <edge from-layer="714" from-port="2" to-layer="716" to-port="0" />
11223 <edge from-layer="715" from-port="0" to-layer="716" to-port="1" />
11224 <edge from-layer="716" from-port="2" to-layer="717" to-port="0" />
11225 <edge from-layer="717" from-port="2" to-layer="719" to-port="0" />
11226 <edge from-layer="718" from-port="0" to-layer="719" to-port="1" />
11227 <edge from-layer="719" from-port="2" to-layer="721" to-port="0" />
11228 <edge from-layer="720" from-port="0" to-layer="721" to-port="1" />
11229 <edge from-layer="721" from-port="2" to-layer="723" to-port="0" />
11230 <edge from-layer="722" from-port="0" to-layer="723" to-port="1" />
11231 <edge from-layer="723" from-port="2" to-layer="725" to-port="0" />
11232 <edge from-layer="723" from-port="2" to-layer="753" to-port="1" />
11233 <edge from-layer="724" from-port="0" to-layer="725" to-port="1" />
11234 <edge from-layer="725" from-port="2" to-layer="730" to-port="0" />
11235 <edge from-layer="726" from-port="0" to-layer="730" to-port="1" />
11236 <edge from-layer="727" from-port="0" to-layer="730" to-port="2" />
11237 <edge from-layer="728" from-port="0" to-layer="730" to-port="3" />
11238 <edge from-layer="729" from-port="0" to-layer="730" to-port="4" />
11239 <edge from-layer="730" from-port="5" to-layer="735" to-port="0" />
11240 <edge from-layer="731" from-port="0" to-layer="732" to-port="0" />
11241 <edge from-layer="732" from-port="1" to-layer="734" to-port="0" />
11242 <edge from-layer="733" from-port="0" to-layer="734" to-port="1" />
11243 <edge from-layer="734" from-port="2" to-layer="735" to-port="1" />
11244 <edge from-layer="735" from-port="2" to-layer="737" to-port="0" />
11245 <edge from-layer="736" from-port="0" to-layer="737" to-port="1" />
11246 <edge from-layer="737" from-port="2" to-layer="738" to-port="0" />
11247 <edge from-layer="738" from-port="1" to-layer="740" to-port="0" />
11248 <edge from-layer="739" from-port="0" to-layer="740" to-port="1" />
11249 <edge from-layer="740" from-port="2" to-layer="745" to-port="0" />
11250 <edge from-layer="741" from-port="0" to-layer="745" to-port="1" />
11251 <edge from-layer="742" from-port="0" to-layer="745" to-port="2" />
11252 <edge from-layer="743" from-port="0" to-layer="745" to-port="3" />
11253 <edge from-layer="744" from-port="0" to-layer="745" to-port="4" />
11254 <edge from-layer="745" from-port="5" to-layer="750" to-port="0" />
11255 <edge from-layer="746" from-port="0" to-layer="747" to-port="0" />
11256 <edge from-layer="747" from-port="1" to-layer="749" to-port="0" />
11257 <edge from-layer="748" from-port="0" to-layer="749" to-port="1" />
11258 <edge from-layer="749" from-port="2" to-layer="750" to-port="1" />
11259 <edge from-layer="750" from-port="2" to-layer="752" to-port="0" />
11260 <edge from-layer="751" from-port="0" to-layer="752" to-port="1" />
11261 <edge from-layer="752" from-port="2" to-layer="753" to-port="0" />
11262 <edge from-layer="753" from-port="2" to-layer="755" to-port="0" />
11263 <edge from-layer="754" from-port="0" to-layer="755" to-port="1" />
11264 <edge from-layer="755" from-port="2" to-layer="757" to-port="0" />
11265 <edge from-layer="756" from-port="0" to-layer="757" to-port="1" />
11266 <edge from-layer="757" from-port="2" to-layer="759" to-port="0" />
11267 <edge from-layer="758" from-port="0" to-layer="759" to-port="1" />
11268 <edge from-layer="759" from-port="2" to-layer="760" to-port="0" />
11269 </edges>
11270 <rt_info>
11271 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
11272 <conversion_parameters>
11273 <framework value="pytorch" />
11274 <is_python_object value="True" />
11275 </conversion_parameters>
11276 <nncf>
11277 <friendly_names_were_updated value="True" />
11278 <quantization>
11279 <advanced_parameters value="{'overflow_fix': 'disable', 'quantize_outputs': False, 'inplace_statistics': True, 'disable_channel_alignment': True, 'disable_bias_correction': False, 'batchwise_statistics': None, 'activations_quantization_params': None, 'weights_quantization_params': None, 'activations_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'weights_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'bias_correction_params': {'apply_for_all_nodes': False, 'threshold': None}, 'smooth_quant_alphas': {'convolution': -1, 'matmul': 0.95}, 'smooth_quant_alpha': None, 'backend_params': {}}" />
11280 <fast_bias_correction value="True" />
11281 <ignored_scope>
11282 <types value="['GroupNormalization']" />
11283 </ignored_scope>
11284 <model_type value="transformer" />
11285 <preset value="mixed" />
11286 <subset_size value="300" />
11287 <target_device value="ANY" />
11288 </quantization>
11289 </nncf>
11290 <optimum>
11291 <optimum_intel_version value="1.20.0.dev0+b31524c" />
11292 <optimum_version value="1.23.0" />
11293 <pytorch_version value="2.5.0.dev20240807+cu121" />
11294 <transformers_version value="4.43.4" />
11295 </optimum>
11296 </rt_info>
11297 </net>
11298