openvino/openvino_model_qint8_quantized.xml
362.9 KB · 11290 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model110" version="11">
3 <layers>
4 <layer id="2" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="1" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="0" name="token_type_ids" type="Parameter" version="opset1">
23 <data shape="?,?" element_type="i64" />
24 <output>
25 <port id="0" precision="I64" names="token_type_ids">
26 <dim>-1</dim>
27 <dim>-1</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="Constant_299926" type="Const" version="opset1">
32 <data element_type="i8" shape="30522, 384" offset="0" size="11720448" />
33 <output>
34 <port id="0" precision="I8">
35 <dim>30522</dim>
36 <dim>384</dim>
37 </port>
38 </output>
39 </layer>
40 <layer id="4" name="Convert_299927" type="Convert" version="opset1">
41 <data destination_type="f32" />
42 <input>
43 <port id="0" precision="I8">
44 <dim>30522</dim>
45 <dim>384</dim>
46 </port>
47 </input>
48 <output>
49 <port id="1" precision="FP32">
50 <dim>30522</dim>
51 <dim>384</dim>
52 </port>
53 </output>
54 </layer>
55 <layer id="5" name="Constant_299928" type="Const" version="opset1">
56 <data element_type="f32" shape="30522, 1" offset="11720448" size="122088" />
57 <output>
58 <port id="0" precision="FP32">
59 <dim>30522</dim>
60 <dim>1</dim>
61 </port>
62 </output>
63 </layer>
64 <layer id="6" name="__module.bert.embeddings.word_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
65 <data auto_broadcast="numpy" />
66 <input>
67 <port id="0" precision="FP32">
68 <dim>30522</dim>
69 <dim>384</dim>
70 </port>
71 <port id="1" precision="FP32">
72 <dim>30522</dim>
73 <dim>1</dim>
74 </port>
75 </input>
76 <output>
77 <port id="2" precision="FP32">
78 <dim>30522</dim>
79 <dim>384</dim>
80 </port>
81 </output>
82 </layer>
83 <layer id="7" name="__module.bert.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
84 <data destination_type="i32" />
85 <input>
86 <port id="0" precision="I64">
87 <dim>-1</dim>
88 <dim>-1</dim>
89 </port>
90 </input>
91 <output>
92 <port id="1" precision="I32">
93 <dim>-1</dim>
94 <dim>-1</dim>
95 </port>
96 </output>
97 </layer>
98 <layer id="8" name="__module.bert.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
99 <data element_type="i32" shape="" offset="11842536" size="4" />
100 <output>
101 <port id="0" precision="I32" />
102 </output>
103 </layer>
104 <layer id="9" name="__module.bert.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
105 <data batch_dims="0" />
106 <input>
107 <port id="0" precision="FP32">
108 <dim>30522</dim>
109 <dim>384</dim>
110 </port>
111 <port id="1" precision="I32">
112 <dim>-1</dim>
113 <dim>-1</dim>
114 </port>
115 <port id="2" precision="I32" />
116 </input>
117 <output>
118 <port id="3" precision="FP32" names="48,inputs_embeds">
119 <dim>-1</dim>
120 <dim>-1</dim>
121 <dim>384</dim>
122 </port>
123 </output>
124 </layer>
125 <layer id="10" name="Constant_299930" type="Const" version="opset1">
126 <data element_type="i8" shape="2, 384" offset="11842540" size="768" />
127 <output>
128 <port id="0" precision="I8">
129 <dim>2</dim>
130 <dim>384</dim>
131 </port>
132 </output>
133 </layer>
134 <layer id="11" name="Convert_299931" type="Convert" version="opset1">
135 <data destination_type="f32" />
136 <input>
137 <port id="0" precision="I8">
138 <dim>2</dim>
139 <dim>384</dim>
140 </port>
141 </input>
142 <output>
143 <port id="1" precision="FP32">
144 <dim>2</dim>
145 <dim>384</dim>
146 </port>
147 </output>
148 </layer>
149 <layer id="12" name="Constant_299932" type="Const" version="opset1">
150 <data element_type="f32" shape="2, 1" offset="11843308" size="8" />
151 <output>
152 <port id="0" precision="FP32">
153 <dim>2</dim>
154 <dim>1</dim>
155 </port>
156 </output>
157 </layer>
158 <layer id="13" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
159 <data auto_broadcast="numpy" />
160 <input>
161 <port id="0" precision="FP32">
162 <dim>2</dim>
163 <dim>384</dim>
164 </port>
165 <port id="1" precision="FP32">
166 <dim>2</dim>
167 <dim>1</dim>
168 </port>
169 </input>
170 <output>
171 <port id="2" precision="FP32">
172 <dim>2</dim>
173 <dim>384</dim>
174 </port>
175 </output>
176 </layer>
177 <layer id="14" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
178 <data destination_type="i32" />
179 <input>
180 <port id="0" precision="I64">
181 <dim>-1</dim>
182 <dim>-1</dim>
183 </port>
184 </input>
185 <output>
186 <port id="1" precision="I32">
187 <dim>-1</dim>
188 <dim>-1</dim>
189 </port>
190 </output>
191 </layer>
192 <layer id="15" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
193 <data element_type="i32" shape="" offset="11842536" size="4" />
194 <output>
195 <port id="0" precision="I32" />
196 </output>
197 </layer>
198 <layer id="16" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
199 <data batch_dims="0" />
200 <input>
201 <port id="0" precision="FP32">
202 <dim>2</dim>
203 <dim>384</dim>
204 </port>
205 <port id="1" precision="I32">
206 <dim>-1</dim>
207 <dim>-1</dim>
208 </port>
209 <port id="2" precision="I32" />
210 </input>
211 <output>
212 <port id="3" precision="FP32" names="50,token_type_embeddings.1">
213 <dim>-1</dim>
214 <dim>-1</dim>
215 <dim>384</dim>
216 </port>
217 </output>
218 </layer>
219 <layer id="17" name="__module.bert.embeddings/aten::add/Add" type="Add" version="opset1">
220 <data auto_broadcast="numpy" />
221 <input>
222 <port id="0" precision="FP32">
223 <dim>-1</dim>
224 <dim>-1</dim>
225 <dim>384</dim>
226 </port>
227 <port id="1" precision="FP32">
228 <dim>-1</dim>
229 <dim>-1</dim>
230 <dim>384</dim>
231 </port>
232 </input>
233 <output>
234 <port id="2" precision="FP32" names="51_1">
235 <dim>-1</dim>
236 <dim>-1</dim>
237 <dim>384</dim>
238 </port>
239 </output>
240 </layer>
241 <layer id="18" name="Constant_299934" type="Const" version="opset1">
242 <data element_type="i8" shape="512, 384" offset="11843316" size="196608" />
243 <output>
244 <port id="0" precision="I8">
245 <dim>512</dim>
246 <dim>384</dim>
247 </port>
248 </output>
249 </layer>
250 <layer id="19" name="Convert_299935" type="Convert" version="opset1">
251 <data destination_type="f32" />
252 <input>
253 <port id="0" precision="I8">
254 <dim>512</dim>
255 <dim>384</dim>
256 </port>
257 </input>
258 <output>
259 <port id="1" precision="FP32">
260 <dim>512</dim>
261 <dim>384</dim>
262 </port>
263 </output>
264 </layer>
265 <layer id="20" name="Constant_299936" type="Const" version="opset1">
266 <data element_type="f32" shape="512, 1" offset="12039924" size="2048" />
267 <output>
268 <port id="0" precision="FP32">
269 <dim>512</dim>
270 <dim>1</dim>
271 </port>
272 </output>
273 </layer>
274 <layer id="21" name="__module.bert.embeddings.position_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
275 <data auto_broadcast="numpy" />
276 <input>
277 <port id="0" precision="FP32">
278 <dim>512</dim>
279 <dim>384</dim>
280 </port>
281 <port id="1" precision="FP32">
282 <dim>512</dim>
283 <dim>1</dim>
284 </port>
285 </input>
286 <output>
287 <port id="2" precision="FP32">
288 <dim>512</dim>
289 <dim>384</dim>
290 </port>
291 </output>
292 </layer>
293 <layer id="22" name="__module.bert.embeddings/aten::slice/Slice" type="Const" version="opset1">
294 <data element_type="i64" shape="1, 512" offset="12041972" size="4096" />
295 <output>
296 <port id="0" precision="I64" names="45">
297 <dim>1</dim>
298 <dim>512</dim>
299 </port>
300 </output>
301 </layer>
302 <layer id="23" name="__module.bert.embeddings/aten::slice/Reshape" type="Const" version="opset1">
303 <data element_type="i64" shape="1" offset="12046068" size="8" />
304 <output>
305 <port id="0" precision="I64">
306 <dim>1</dim>
307 </port>
308 </output>
309 </layer>
310 <layer id="24" name="ShapeOf_90427" type="ShapeOf" version="opset3">
311 <data output_type="i64" />
312 <input>
313 <port id="0" precision="I64">
314 <dim>-1</dim>
315 <dim>-1</dim>
316 </port>
317 </input>
318 <output>
319 <port id="1" precision="I64">
320 <dim>2</dim>
321 </port>
322 </output>
323 </layer>
324 <layer id="25" name="Constant_90546" type="Const" version="opset1">
325 <data element_type="i64" shape="1" offset="12046076" size="8" />
326 <output>
327 <port id="0" precision="I64">
328 <dim>1</dim>
329 </port>
330 </output>
331 </layer>
332 <layer id="26" name="Constant_90429" type="Const" version="opset1">
333 <data element_type="i64" shape="" offset="12046068" size="8" />
334 <output>
335 <port id="0" precision="I64" />
336 </output>
337 </layer>
338 <layer id="27" name="Gather_90430" type="Gather" version="opset8">
339 <data batch_dims="0" />
340 <input>
341 <port id="0" precision="I64">
342 <dim>2</dim>
343 </port>
344 <port id="1" precision="I64">
345 <dim>1</dim>
346 </port>
347 <port id="2" precision="I64" />
348 </input>
349 <output>
350 <port id="3" precision="I64" names="35,41,43,44,61">
351 <dim>1</dim>
352 </port>
353 </output>
354 </layer>
355 <layer id="28" name="__module.bert.embeddings/aten::slice/Reshape_2" type="Const" version="opset1">
356 <data element_type="i64" shape="1" offset="12046076" size="8" />
357 <output>
358 <port id="0" precision="I64">
359 <dim>1</dim>
360 </port>
361 </output>
362 </layer>
363 <layer id="29" name="__module.bert.embeddings/aten::slice/Reshape_3" type="Const" version="opset1">
364 <data element_type="i64" shape="1" offset="12046076" size="8" />
365 <output>
366 <port id="0" precision="I64">
367 <dim>1</dim>
368 </port>
369 </output>
370 </layer>
371 <layer id="30" name="__module.bert.embeddings/aten::slice/Slice_1" type="Slice" version="opset8">
372 <input>
373 <port id="0" precision="I64">
374 <dim>1</dim>
375 <dim>512</dim>
376 </port>
377 <port id="1" precision="I64">
378 <dim>1</dim>
379 </port>
380 <port id="2" precision="I64">
381 <dim>1</dim>
382 </port>
383 <port id="3" precision="I64">
384 <dim>1</dim>
385 </port>
386 <port id="4" precision="I64">
387 <dim>1</dim>
388 </port>
389 </input>
390 <output>
391 <port id="5" precision="I64" names="46">
392 <dim>1</dim>
393 <dim>-1</dim>
394 </port>
395 </output>
396 </layer>
397 <layer id="31" name="__module.bert.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
398 <data destination_type="i32" />
399 <input>
400 <port id="0" precision="I64">
401 <dim>1</dim>
402 <dim>-1</dim>
403 </port>
404 </input>
405 <output>
406 <port id="1" precision="I32">
407 <dim>1</dim>
408 <dim>-1</dim>
409 </port>
410 </output>
411 </layer>
412 <layer id="32" name="__module.bert.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
413 <data element_type="i32" shape="" offset="11842536" size="4" />
414 <output>
415 <port id="0" precision="I32" />
416 </output>
417 </layer>
418 <layer id="33" name="__module.bert.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
419 <data batch_dims="0" />
420 <input>
421 <port id="0" precision="FP32">
422 <dim>512</dim>
423 <dim>384</dim>
424 </port>
425 <port id="1" precision="I32">
426 <dim>1</dim>
427 <dim>-1</dim>
428 </port>
429 <port id="2" precision="I32" />
430 </input>
431 <output>
432 <port id="3" precision="FP32" names="53,position_embeddings.1">
433 <dim>1</dim>
434 <dim>-1</dim>
435 <dim>384</dim>
436 </port>
437 </output>
438 </layer>
439 <layer id="34" name="__module.bert.embeddings/aten::add_/Add" type="Add" version="opset1">
440 <data auto_broadcast="numpy" />
441 <input>
442 <port id="0" precision="FP32">
443 <dim>-1</dim>
444 <dim>-1</dim>
445 <dim>384</dim>
446 </port>
447 <port id="1" precision="FP32">
448 <dim>1</dim>
449 <dim>-1</dim>
450 <dim>384</dim>
451 </port>
452 </input>
453 <output>
454 <port id="2" precision="FP32" names="51,embeddings.1">
455 <dim>-1</dim>
456 <dim>-1</dim>
457 <dim>384</dim>
458 </port>
459 </output>
460 </layer>
461 <layer id="35" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
462 <data element_type="i32" shape="1" offset="12046084" size="4" />
463 <output>
464 <port id="0" precision="I32">
465 <dim>1</dim>
466 </port>
467 </output>
468 </layer>
469 <layer id="36" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
470 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
471 <input>
472 <port id="0" precision="FP32">
473 <dim>-1</dim>
474 <dim>-1</dim>
475 <dim>384</dim>
476 </port>
477 <port id="1" precision="I32">
478 <dim>1</dim>
479 </port>
480 </input>
481 <output>
482 <port id="2" precision="FP32">
483 <dim>-1</dim>
484 <dim>-1</dim>
485 <dim>384</dim>
486 </port>
487 </output>
488 </layer>
489 <layer id="37" name="Constant_90301" type="Const" version="opset1">
490 <data element_type="f32" shape="1, 1, 384" offset="12046088" size="1536" />
491 <output>
492 <port id="0" precision="FP32">
493 <dim>1</dim>
494 <dim>1</dim>
495 <dim>384</dim>
496 </port>
497 </output>
498 </layer>
499 <layer id="38" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
500 <data auto_broadcast="numpy" />
501 <input>
502 <port id="0" precision="FP32">
503 <dim>-1</dim>
504 <dim>-1</dim>
505 <dim>384</dim>
506 </port>
507 <port id="1" precision="FP32">
508 <dim>1</dim>
509 <dim>1</dim>
510 <dim>384</dim>
511 </port>
512 </input>
513 <output>
514 <port id="2" precision="FP32">
515 <dim>-1</dim>
516 <dim>-1</dim>
517 <dim>384</dim>
518 </port>
519 </output>
520 </layer>
521 <layer id="39" name="Constant_90302" type="Const" version="opset1">
522 <data element_type="f32" shape="1, 1, 384" offset="12047624" size="1536" />
523 <output>
524 <port id="0" precision="FP32">
525 <dim>1</dim>
526 <dim>1</dim>
527 <dim>384</dim>
528 </port>
529 </output>
530 </layer>
531 <layer id="40" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
532 <data auto_broadcast="numpy" />
533 <input>
534 <port id="0" precision="FP32">
535 <dim>-1</dim>
536 <dim>-1</dim>
537 <dim>384</dim>
538 </port>
539 <port id="1" precision="FP32">
540 <dim>1</dim>
541 <dim>1</dim>
542 <dim>384</dim>
543 </port>
544 </input>
545 <output>
546 <port id="2" precision="FP32" names="58,input.1">
547 <dim>-1</dim>
548 <dim>-1</dim>
549 <dim>384</dim>
550 </port>
551 </output>
552 </layer>
553 <layer id="41" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
554 <data element_type="f32" shape="1, 1, 384" offset="12049160" size="1536" />
555 <output>
556 <port id="0" precision="FP32">
557 <dim>1</dim>
558 <dim>1</dim>
559 <dim>384</dim>
560 </port>
561 </output>
562 </layer>
563 <layer id="42" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
564 <data auto_broadcast="numpy" />
565 <input>
566 <port id="0" precision="FP32">
567 <dim>-1</dim>
568 <dim>-1</dim>
569 <dim>384</dim>
570 </port>
571 <port id="1" precision="FP32">
572 <dim>1</dim>
573 <dim>1</dim>
574 <dim>384</dim>
575 </port>
576 </input>
577 <output>
578 <port id="2" precision="FP32">
579 <dim>-1</dim>
580 <dim>-1</dim>
581 <dim>384</dim>
582 </port>
583 </output>
584 </layer>
585 <layer id="43" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
586 <data element_type="f32" shape="" offset="12050696" size="4" />
587 <output>
588 <port id="0" precision="FP32" />
589 </output>
590 </layer>
591 <layer id="44" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
592 <data element_type="f32" shape="" offset="12050700" size="4" />
593 <output>
594 <port id="0" precision="FP32" />
595 </output>
596 </layer>
597 <layer id="45" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
598 <data element_type="f32" shape="" offset="12050696" size="4" />
599 <output>
600 <port id="0" precision="FP32" />
601 </output>
602 </layer>
603 <layer id="46" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
604 <data element_type="f32" shape="" offset="12050700" size="4" />
605 <output>
606 <port id="0" precision="FP32" />
607 </output>
608 </layer>
609 <layer id="47" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
610 <data levels="256" auto_broadcast="numpy" />
611 <input>
612 <port id="0" precision="FP32">
613 <dim>-1</dim>
614 <dim>-1</dim>
615 <dim>384</dim>
616 </port>
617 <port id="1" precision="FP32" />
618 <port id="2" precision="FP32" />
619 <port id="3" precision="FP32" />
620 <port id="4" precision="FP32" />
621 </input>
622 <output>
623 <port id="5" precision="FP32">
624 <dim>-1</dim>
625 <dim>-1</dim>
626 <dim>384</dim>
627 </port>
628 </output>
629 </layer>
630 <layer id="48" name="Constant_299938" type="Const" version="opset1">
631 <data element_type="i8" shape="384, 384" offset="12050704" size="147456" />
632 <output>
633 <port id="0" precision="I8">
634 <dim>384</dim>
635 <dim>384</dim>
636 </port>
637 </output>
638 </layer>
639 <layer id="49" name="Convert_299939" type="Convert" version="opset1">
640 <data destination_type="f32" />
641 <input>
642 <port id="0" precision="I8">
643 <dim>384</dim>
644 <dim>384</dim>
645 </port>
646 </input>
647 <output>
648 <port id="1" precision="FP32">
649 <dim>384</dim>
650 <dim>384</dim>
651 </port>
652 </output>
653 </layer>
654 <layer id="50" name="Constant_299940" type="Const" version="opset1">
655 <data element_type="f32" shape="384, 1" offset="12198160" size="1536" />
656 <output>
657 <port id="0" precision="FP32">
658 <dim>384</dim>
659 <dim>1</dim>
660 </port>
661 </output>
662 </layer>
663 <layer id="51" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
664 <data auto_broadcast="numpy" />
665 <input>
666 <port id="0" precision="FP32">
667 <dim>384</dim>
668 <dim>384</dim>
669 </port>
670 <port id="1" precision="FP32">
671 <dim>384</dim>
672 <dim>1</dim>
673 </port>
674 </input>
675 <output>
676 <port id="2" precision="FP32">
677 <dim>384</dim>
678 <dim>384</dim>
679 </port>
680 </output>
681 </layer>
682 <layer id="52" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
683 <data transpose_a="false" transpose_b="true" />
684 <input>
685 <port id="0" precision="FP32">
686 <dim>-1</dim>
687 <dim>-1</dim>
688 <dim>384</dim>
689 </port>
690 <port id="1" precision="FP32">
691 <dim>384</dim>
692 <dim>384</dim>
693 </port>
694 </input>
695 <output>
696 <port id="2" precision="FP32">
697 <dim>-1</dim>
698 <dim>-1</dim>
699 <dim>384</dim>
700 </port>
701 </output>
702 </layer>
703 <layer id="53" name="Constant_90303" type="Const" version="opset1">
704 <data element_type="f32" shape="1, 1, 384" offset="12199696" size="1536" />
705 <output>
706 <port id="0" precision="FP32">
707 <dim>1</dim>
708 <dim>1</dim>
709 <dim>384</dim>
710 </port>
711 </output>
712 </layer>
713 <layer id="54" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
714 <data auto_broadcast="numpy" />
715 <input>
716 <port id="0" precision="FP32">
717 <dim>-1</dim>
718 <dim>-1</dim>
719 <dim>384</dim>
720 </port>
721 <port id="1" precision="FP32">
722 <dim>1</dim>
723 <dim>1</dim>
724 <dim>384</dim>
725 </port>
726 </input>
727 <output>
728 <port id="2" precision="FP32" names="96,x.1">
729 <dim>-1</dim>
730 <dim>-1</dim>
731 <dim>384</dim>
732 </port>
733 </output>
734 </layer>
735 <layer id="55" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
736 <data element_type="f32" shape="" offset="12201232" size="4" />
737 <output>
738 <port id="0" precision="FP32" />
739 </output>
740 </layer>
741 <layer id="56" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
742 <data element_type="f32" shape="" offset="12201236" size="4" />
743 <output>
744 <port id="0" precision="FP32" />
745 </output>
746 </layer>
747 <layer id="57" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
748 <data element_type="f32" shape="" offset="12201232" size="4" />
749 <output>
750 <port id="0" precision="FP32" />
751 </output>
752 </layer>
753 <layer id="58" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
754 <data element_type="f32" shape="" offset="12201236" size="4" />
755 <output>
756 <port id="0" precision="FP32" />
757 </output>
758 </layer>
759 <layer id="59" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
760 <data levels="256" auto_broadcast="numpy" />
761 <input>
762 <port id="0" precision="FP32">
763 <dim>-1</dim>
764 <dim>-1</dim>
765 <dim>384</dim>
766 </port>
767 <port id="1" precision="FP32" />
768 <port id="2" precision="FP32" />
769 <port id="3" precision="FP32" />
770 <port id="4" precision="FP32" />
771 </input>
772 <output>
773 <port id="5" precision="FP32">
774 <dim>-1</dim>
775 <dim>-1</dim>
776 <dim>384</dim>
777 </port>
778 </output>
779 </layer>
780 <layer id="60" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
781 <data element_type="i64" shape="4" offset="12201240" size="32" />
782 <output>
783 <port id="0" precision="I64">
784 <dim>4</dim>
785 </port>
786 </output>
787 </layer>
788 <layer id="61" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
789 <data special_zero="true" />
790 <input>
791 <port id="0" precision="FP32">
792 <dim>-1</dim>
793 <dim>-1</dim>
794 <dim>384</dim>
795 </port>
796 <port id="1" precision="I64">
797 <dim>4</dim>
798 </port>
799 </input>
800 <output>
801 <port id="2" precision="FP32" names="100,x.3">
802 <dim>-1</dim>
803 <dim>-1</dim>
804 <dim>12</dim>
805 <dim>32</dim>
806 </port>
807 </output>
808 </layer>
809 <layer id="62" name="Constant_84452" type="Const" version="opset1">
810 <data element_type="i64" shape="4" offset="12201272" size="32" />
811 <output>
812 <port id="0" precision="I64" names="101">
813 <dim>4</dim>
814 </port>
815 </output>
816 </layer>
817 <layer id="63" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
818 <input>
819 <port id="0" precision="FP32">
820 <dim>-1</dim>
821 <dim>-1</dim>
822 <dim>12</dim>
823 <dim>32</dim>
824 </port>
825 <port id="1" precision="I64">
826 <dim>4</dim>
827 </port>
828 </input>
829 <output>
830 <port id="2" precision="FP32" names="102">
831 <dim>-1</dim>
832 <dim>12</dim>
833 <dim>-1</dim>
834 <dim>32</dim>
835 </port>
836 </output>
837 </layer>
838 <layer id="64" name="Constant_299942" type="Const" version="opset1">
839 <data element_type="i8" shape="384, 384" offset="12201304" size="147456" />
840 <output>
841 <port id="0" precision="I8">
842 <dim>384</dim>
843 <dim>384</dim>
844 </port>
845 </output>
846 </layer>
847 <layer id="65" name="Convert_299943" type="Convert" version="opset1">
848 <data destination_type="f32" />
849 <input>
850 <port id="0" precision="I8">
851 <dim>384</dim>
852 <dim>384</dim>
853 </port>
854 </input>
855 <output>
856 <port id="1" precision="FP32">
857 <dim>384</dim>
858 <dim>384</dim>
859 </port>
860 </output>
861 </layer>
862 <layer id="66" name="Constant_299944" type="Const" version="opset1">
863 <data element_type="f32" shape="384, 1" offset="12348760" size="1536" />
864 <output>
865 <port id="0" precision="FP32">
866 <dim>384</dim>
867 <dim>1</dim>
868 </port>
869 </output>
870 </layer>
871 <layer id="67" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
872 <data auto_broadcast="numpy" />
873 <input>
874 <port id="0" precision="FP32">
875 <dim>384</dim>
876 <dim>384</dim>
877 </port>
878 <port id="1" precision="FP32">
879 <dim>384</dim>
880 <dim>1</dim>
881 </port>
882 </input>
883 <output>
884 <port id="2" precision="FP32">
885 <dim>384</dim>
886 <dim>384</dim>
887 </port>
888 </output>
889 </layer>
890 <layer id="68" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
891 <data transpose_a="false" transpose_b="true" />
892 <input>
893 <port id="0" precision="FP32">
894 <dim>-1</dim>
895 <dim>-1</dim>
896 <dim>384</dim>
897 </port>
898 <port id="1" precision="FP32">
899 <dim>384</dim>
900 <dim>384</dim>
901 </port>
902 </input>
903 <output>
904 <port id="2" precision="FP32">
905 <dim>-1</dim>
906 <dim>-1</dim>
907 <dim>384</dim>
908 </port>
909 </output>
910 </layer>
911 <layer id="69" name="Constant_90304" type="Const" version="opset1">
912 <data element_type="f32" shape="1, 1, 384" offset="12350296" size="1536" />
913 <output>
914 <port id="0" precision="FP32">
915 <dim>1</dim>
916 <dim>1</dim>
917 <dim>384</dim>
918 </port>
919 </output>
920 </layer>
921 <layer id="70" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
922 <data auto_broadcast="numpy" />
923 <input>
924 <port id="0" precision="FP32">
925 <dim>-1</dim>
926 <dim>-1</dim>
927 <dim>384</dim>
928 </port>
929 <port id="1" precision="FP32">
930 <dim>1</dim>
931 <dim>1</dim>
932 <dim>384</dim>
933 </port>
934 </input>
935 <output>
936 <port id="2" precision="FP32" names="105,x.5">
937 <dim>-1</dim>
938 <dim>-1</dim>
939 <dim>384</dim>
940 </port>
941 </output>
942 </layer>
943 <layer id="71" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
944 <data element_type="f32" shape="" offset="12351832" size="4" />
945 <output>
946 <port id="0" precision="FP32" />
947 </output>
948 </layer>
949 <layer id="72" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
950 <data element_type="f32" shape="" offset="12351836" size="4" />
951 <output>
952 <port id="0" precision="FP32" />
953 </output>
954 </layer>
955 <layer id="73" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
956 <data element_type="f32" shape="" offset="12351832" size="4" />
957 <output>
958 <port id="0" precision="FP32" />
959 </output>
960 </layer>
961 <layer id="74" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
962 <data element_type="f32" shape="" offset="12351836" size="4" />
963 <output>
964 <port id="0" precision="FP32" />
965 </output>
966 </layer>
967 <layer id="75" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
968 <data levels="256" auto_broadcast="numpy" />
969 <input>
970 <port id="0" precision="FP32">
971 <dim>-1</dim>
972 <dim>-1</dim>
973 <dim>384</dim>
974 </port>
975 <port id="1" precision="FP32" />
976 <port id="2" precision="FP32" />
977 <port id="3" precision="FP32" />
978 <port id="4" precision="FP32" />
979 </input>
980 <output>
981 <port id="5" precision="FP32">
982 <dim>-1</dim>
983 <dim>-1</dim>
984 <dim>384</dim>
985 </port>
986 </output>
987 </layer>
988 <layer id="76" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
989 <data element_type="i64" shape="4" offset="12201240" size="32" />
990 <output>
991 <port id="0" precision="I64">
992 <dim>4</dim>
993 </port>
994 </output>
995 </layer>
996 <layer id="77" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
997 <data special_zero="true" />
998 <input>
999 <port id="0" precision="FP32">
1000 <dim>-1</dim>
1001 <dim>-1</dim>
1002 <dim>384</dim>
1003 </port>
1004 <port id="1" precision="I64">
1005 <dim>4</dim>
1006 </port>
1007 </input>
1008 <output>
1009 <port id="2" precision="FP32" names="109,x.7">
1010 <dim>-1</dim>
1011 <dim>-1</dim>
1012 <dim>12</dim>
1013 <dim>32</dim>
1014 </port>
1015 </output>
1016 </layer>
1017 <layer id="78" name="Constant_84477" type="Const" version="opset1">
1018 <data element_type="i64" shape="4" offset="12201272" size="32" />
1019 <output>
1020 <port id="0" precision="I64" names="110">
1021 <dim>4</dim>
1022 </port>
1023 </output>
1024 </layer>
1025 <layer id="79" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1026 <input>
1027 <port id="0" precision="FP32">
1028 <dim>-1</dim>
1029 <dim>-1</dim>
1030 <dim>12</dim>
1031 <dim>32</dim>
1032 </port>
1033 <port id="1" precision="I64">
1034 <dim>4</dim>
1035 </port>
1036 </input>
1037 <output>
1038 <port id="2" precision="FP32" names="111">
1039 <dim>-1</dim>
1040 <dim>12</dim>
1041 <dim>-1</dim>
1042 <dim>32</dim>
1043 </port>
1044 </output>
1045 </layer>
1046 <layer id="80" name="Constant_299946" type="Const" version="opset1">
1047 <data element_type="i8" shape="384, 384" offset="12351840" size="147456" />
1048 <output>
1049 <port id="0" precision="I8">
1050 <dim>384</dim>
1051 <dim>384</dim>
1052 </port>
1053 </output>
1054 </layer>
1055 <layer id="81" name="Convert_299947" type="Convert" version="opset1">
1056 <data destination_type="f32" />
1057 <input>
1058 <port id="0" precision="I8">
1059 <dim>384</dim>
1060 <dim>384</dim>
1061 </port>
1062 </input>
1063 <output>
1064 <port id="1" precision="FP32">
1065 <dim>384</dim>
1066 <dim>384</dim>
1067 </port>
1068 </output>
1069 </layer>
1070 <layer id="82" name="Constant_299948" type="Const" version="opset1">
1071 <data element_type="f32" shape="384, 1" offset="12499296" size="1536" />
1072 <output>
1073 <port id="0" precision="FP32">
1074 <dim>384</dim>
1075 <dim>1</dim>
1076 </port>
1077 </output>
1078 </layer>
1079 <layer id="83" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1080 <data auto_broadcast="numpy" />
1081 <input>
1082 <port id="0" precision="FP32">
1083 <dim>384</dim>
1084 <dim>384</dim>
1085 </port>
1086 <port id="1" precision="FP32">
1087 <dim>384</dim>
1088 <dim>1</dim>
1089 </port>
1090 </input>
1091 <output>
1092 <port id="2" precision="FP32">
1093 <dim>384</dim>
1094 <dim>384</dim>
1095 </port>
1096 </output>
1097 </layer>
1098 <layer id="84" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1099 <data transpose_a="false" transpose_b="true" />
1100 <input>
1101 <port id="0" precision="FP32">
1102 <dim>-1</dim>
1103 <dim>-1</dim>
1104 <dim>384</dim>
1105 </port>
1106 <port id="1" precision="FP32">
1107 <dim>384</dim>
1108 <dim>384</dim>
1109 </port>
1110 </input>
1111 <output>
1112 <port id="2" precision="FP32">
1113 <dim>-1</dim>
1114 <dim>-1</dim>
1115 <dim>384</dim>
1116 </port>
1117 </output>
1118 </layer>
1119 <layer id="85" name="Constant_90305" type="Const" version="opset1">
1120 <data element_type="f32" shape="1, 1, 384" offset="12500832" size="1536" />
1121 <output>
1122 <port id="0" precision="FP32">
1123 <dim>1</dim>
1124 <dim>1</dim>
1125 <dim>384</dim>
1126 </port>
1127 </output>
1128 </layer>
1129 <layer id="86" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
1130 <data auto_broadcast="numpy" />
1131 <input>
1132 <port id="0" precision="FP32">
1133 <dim>-1</dim>
1134 <dim>-1</dim>
1135 <dim>384</dim>
1136 </port>
1137 <port id="1" precision="FP32">
1138 <dim>1</dim>
1139 <dim>1</dim>
1140 <dim>384</dim>
1141 </port>
1142 </input>
1143 <output>
1144 <port id="2" precision="FP32" names="114,x.9">
1145 <dim>-1</dim>
1146 <dim>-1</dim>
1147 <dim>384</dim>
1148 </port>
1149 </output>
1150 </layer>
1151 <layer id="87" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
1152 <data element_type="i64" shape="4" offset="12201240" size="32" />
1153 <output>
1154 <port id="0" precision="I64">
1155 <dim>4</dim>
1156 </port>
1157 </output>
1158 </layer>
1159 <layer id="88" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
1160 <data special_zero="true" />
1161 <input>
1162 <port id="0" precision="FP32">
1163 <dim>-1</dim>
1164 <dim>-1</dim>
1165 <dim>384</dim>
1166 </port>
1167 <port id="1" precision="I64">
1168 <dim>4</dim>
1169 </port>
1170 </input>
1171 <output>
1172 <port id="2" precision="FP32" names="118,x.11">
1173 <dim>-1</dim>
1174 <dim>-1</dim>
1175 <dim>12</dim>
1176 <dim>32</dim>
1177 </port>
1178 </output>
1179 </layer>
1180 <layer id="89" name="Constant_84502" type="Const" version="opset1">
1181 <data element_type="i64" shape="4" offset="12201272" size="32" />
1182 <output>
1183 <port id="0" precision="I64" names="119">
1184 <dim>4</dim>
1185 </port>
1186 </output>
1187 </layer>
1188 <layer id="90" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
1189 <input>
1190 <port id="0" precision="FP32">
1191 <dim>-1</dim>
1192 <dim>-1</dim>
1193 <dim>12</dim>
1194 <dim>32</dim>
1195 </port>
1196 <port id="1" precision="I64">
1197 <dim>4</dim>
1198 </port>
1199 </input>
1200 <output>
1201 <port id="2" precision="FP32" names="120">
1202 <dim>-1</dim>
1203 <dim>12</dim>
1204 <dim>-1</dim>
1205 <dim>32</dim>
1206 </port>
1207 </output>
1208 </layer>
1209 <layer id="91" name="Constant_90307" type="Const" version="opset1">
1210 <data element_type="f32" shape="1, 1, 1, 1" offset="12502368" size="4" />
1211 <output>
1212 <port id="0" precision="FP32">
1213 <dim>1</dim>
1214 <dim>1</dim>
1215 <dim>1</dim>
1216 <dim>1</dim>
1217 </port>
1218 </output>
1219 </layer>
1220 <layer id="92" name="31" type="Const" version="opset1">
1221 <data element_type="i64" shape="" offset="12046076" size="8" />
1222 <output>
1223 <port id="0" precision="I64" names="31" />
1224 </output>
1225 </layer>
1226 <layer id="93" name="__module.bert/aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
1227 <input>
1228 <port id="0" precision="I64">
1229 <dim>-1</dim>
1230 <dim>-1</dim>
1231 </port>
1232 <port id="1" precision="I64" />
1233 </input>
1234 <output>
1235 <port id="2" precision="I64" names="63">
1236 <dim>-1</dim>
1237 <dim>1</dim>
1238 <dim>-1</dim>
1239 </port>
1240 </output>
1241 </layer>
1242 <layer id="94" name="21" type="Const" version="opset1">
1243 <data element_type="i64" shape="" offset="12502372" size="8" />
1244 <output>
1245 <port id="0" precision="I64" names="21" />
1246 </output>
1247 </layer>
1248 <layer id="95" name="__module.bert/aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
1249 <input>
1250 <port id="0" precision="I64">
1251 <dim>-1</dim>
1252 <dim>1</dim>
1253 <dim>-1</dim>
1254 </port>
1255 <port id="1" precision="I64" />
1256 </input>
1257 <output>
1258 <port id="2" precision="I64" names="64,65">
1259 <dim>-1</dim>
1260 <dim>1</dim>
1261 <dim>1</dim>
1262 <dim>-1</dim>
1263 </port>
1264 </output>
1265 </layer>
1266 <layer id="96" name="Constant_90436" type="Const" version="opset1">
1267 <data element_type="i64" shape="1" offset="12046068" size="8" />
1268 <output>
1269 <port id="0" precision="I64">
1270 <dim>1</dim>
1271 </port>
1272 </output>
1273 </layer>
1274 <layer id="97" name="Constant_90437" type="Const" version="opset1">
1275 <data element_type="i64" shape="" offset="12046068" size="8" />
1276 <output>
1277 <port id="0" precision="I64" />
1278 </output>
1279 </layer>
1280 <layer id="98" name="Gather_90438" type="Gather" version="opset8">
1281 <data batch_dims="0" />
1282 <input>
1283 <port id="0" precision="I64">
1284 <dim>2</dim>
1285 </port>
1286 <port id="1" precision="I64">
1287 <dim>1</dim>
1288 </port>
1289 <port id="2" precision="I64" />
1290 </input>
1291 <output>
1292 <port id="3" precision="I64" names="60">
1293 <dim>1</dim>
1294 </port>
1295 </output>
1296 </layer>
1297 <layer id="99" name="Constant_89535" type="Const" version="opset1">
1298 <data element_type="i64" shape="1" offset="12046076" size="8" />
1299 <output>
1300 <port id="0" precision="I64">
1301 <dim>1</dim>
1302 </port>
1303 </output>
1304 </layer>
1305 <layer id="100" name="Constant_90548" type="Const" version="opset1">
1306 <data element_type="i64" shape="2" offset="12502380" size="16" />
1307 <output>
1308 <port id="0" precision="I64">
1309 <dim>2</dim>
1310 </port>
1311 </output>
1312 </layer>
1313 <layer id="101" name="Constant_90549" type="Const" version="opset1">
1314 <data element_type="i64" shape="" offset="12046068" size="8" />
1315 <output>
1316 <port id="0" precision="I64" />
1317 </output>
1318 </layer>
1319 <layer id="102" name="Gather_90550" type="Gather" version="opset8">
1320 <data batch_dims="0" />
1321 <input>
1322 <port id="0" precision="I64">
1323 <dim>2</dim>
1324 </port>
1325 <port id="1" precision="I64">
1326 <dim>2</dim>
1327 </port>
1328 <port id="2" precision="I64" />
1329 </input>
1330 <output>
1331 <port id="3" precision="I64">
1332 <dim>2</dim>
1333 </port>
1334 </output>
1335 </layer>
1336 <layer id="103" name="__module.bert/prim::ListConstruct/Concat" type="Concat" version="opset1">
1337 <data axis="0" />
1338 <input>
1339 <port id="0" precision="I64">
1340 <dim>1</dim>
1341 </port>
1342 <port id="1" precision="I64">
1343 <dim>1</dim>
1344 </port>
1345 <port id="2" precision="I64">
1346 <dim>2</dim>
1347 </port>
1348 </input>
1349 <output>
1350 <port id="3" precision="I64" names="66">
1351 <dim>4</dim>
1352 </port>
1353 </output>
1354 </layer>
1355 <layer id="104" name="__module.bert/aten::expand/Broadcast" type="Broadcast" version="opset3">
1356 <data mode="bidirectional" />
1357 <input>
1358 <port id="0" precision="I64">
1359 <dim>-1</dim>
1360 <dim>1</dim>
1361 <dim>1</dim>
1362 <dim>-1</dim>
1363 </port>
1364 <port id="1" precision="I64">
1365 <dim>4</dim>
1366 </port>
1367 </input>
1368 <output>
1369 <port id="2" precision="I64" names="67">
1370 <dim>-1</dim>
1371 <dim>1</dim>
1372 <dim>-1</dim>
1373 <dim>-1</dim>
1374 </port>
1375 </output>
1376 </layer>
1377 <layer id="105" name="__module.bert/aten::to/Convert" type="Convert" version="opset1">
1378 <data destination_type="f32" />
1379 <input>
1380 <port id="0" precision="I64">
1381 <dim>-1</dim>
1382 <dim>1</dim>
1383 <dim>-1</dim>
1384 <dim>-1</dim>
1385 </port>
1386 </input>
1387 <output>
1388 <port id="1" precision="FP32" names="68">
1389 <dim>-1</dim>
1390 <dim>1</dim>
1391 <dim>-1</dim>
1392 <dim>-1</dim>
1393 </port>
1394 </output>
1395 </layer>
1396 <layer id="106" name="Constant_90306" type="Const" version="opset1">
1397 <data element_type="f32" shape="1, 1, 1, 1" offset="12502368" size="4" />
1398 <output>
1399 <port id="0" precision="FP32">
1400 <dim>1</dim>
1401 <dim>1</dim>
1402 <dim>1</dim>
1403 <dim>1</dim>
1404 </port>
1405 </output>
1406 </layer>
1407 <layer id="107" name="__module.bert/aten::rsub/Multiply" type="Multiply" version="opset1">
1408 <data auto_broadcast="numpy" />
1409 <input>
1410 <port id="0" precision="FP32">
1411 <dim>-1</dim>
1412 <dim>1</dim>
1413 <dim>-1</dim>
1414 <dim>-1</dim>
1415 </port>
1416 <port id="1" precision="FP32">
1417 <dim>1</dim>
1418 <dim>1</dim>
1419 <dim>1</dim>
1420 <dim>1</dim>
1421 </port>
1422 </input>
1423 <output>
1424 <port id="2" precision="FP32">
1425 <dim>-1</dim>
1426 <dim>1</dim>
1427 <dim>-1</dim>
1428 <dim>-1</dim>
1429 </port>
1430 </output>
1431 </layer>
1432 <layer id="108" name="__module.bert/aten::rsub/Subtract" type="Subtract" version="opset1">
1433 <data auto_broadcast="numpy" />
1434 <input>
1435 <port id="0" precision="FP32">
1436 <dim>1</dim>
1437 <dim>1</dim>
1438 <dim>1</dim>
1439 <dim>1</dim>
1440 </port>
1441 <port id="1" precision="FP32">
1442 <dim>-1</dim>
1443 <dim>1</dim>
1444 <dim>-1</dim>
1445 <dim>-1</dim>
1446 </port>
1447 </input>
1448 <output>
1449 <port id="2" precision="FP32" names="69,inverted_mask">
1450 <dim>-1</dim>
1451 <dim>1</dim>
1452 <dim>-1</dim>
1453 <dim>-1</dim>
1454 </port>
1455 </output>
1456 </layer>
1457 <layer id="109" name="__module.bert/aten::to/Convert_1" type="Convert" version="opset1">
1458 <data destination_type="boolean" />
1459 <input>
1460 <port id="0" precision="FP32">
1461 <dim>-1</dim>
1462 <dim>1</dim>
1463 <dim>-1</dim>
1464 <dim>-1</dim>
1465 </port>
1466 </input>
1467 <output>
1468 <port id="1" precision="BOOL" names="70">
1469 <dim>-1</dim>
1470 <dim>1</dim>
1471 <dim>-1</dim>
1472 <dim>-1</dim>
1473 </port>
1474 </output>
1475 </layer>
1476 <layer id="110" name="__module.bert/aten::masked_fill/ConvertLike" type="Const" version="opset1">
1477 <data element_type="f32" shape="" offset="12502396" size="4" />
1478 <output>
1479 <port id="0" precision="FP32" />
1480 </output>
1481 </layer>
1482 <layer id="111" name="__module.bert/aten::masked_fill/Select" type="Select" version="opset1">
1483 <data auto_broadcast="numpy" />
1484 <input>
1485 <port id="0" precision="BOOL">
1486 <dim>-1</dim>
1487 <dim>1</dim>
1488 <dim>-1</dim>
1489 <dim>-1</dim>
1490 </port>
1491 <port id="1" precision="FP32" />
1492 <port id="2" precision="FP32">
1493 <dim>-1</dim>
1494 <dim>1</dim>
1495 <dim>-1</dim>
1496 <dim>-1</dim>
1497 </port>
1498 </input>
1499 <output>
1500 <port id="3" precision="FP32" names="71">
1501 <dim>-1</dim>
1502 <dim>1</dim>
1503 <dim>-1</dim>
1504 <dim>-1</dim>
1505 </port>
1506 </output>
1507 </layer>
1508 <layer id="112" name="__module.bert.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1509 <data causal="false" />
1510 <input>
1511 <port id="0" precision="FP32">
1512 <dim>-1</dim>
1513 <dim>12</dim>
1514 <dim>-1</dim>
1515 <dim>32</dim>
1516 </port>
1517 <port id="1" precision="FP32">
1518 <dim>-1</dim>
1519 <dim>12</dim>
1520 <dim>-1</dim>
1521 <dim>32</dim>
1522 </port>
1523 <port id="2" precision="FP32">
1524 <dim>-1</dim>
1525 <dim>12</dim>
1526 <dim>-1</dim>
1527 <dim>32</dim>
1528 </port>
1529 <port id="3" precision="FP32">
1530 <dim>-1</dim>
1531 <dim>1</dim>
1532 <dim>-1</dim>
1533 <dim>-1</dim>
1534 </port>
1535 </input>
1536 <output>
1537 <port id="4" precision="FP32" names="121,attn_output.1">
1538 <dim>-1</dim>
1539 <dim>12</dim>
1540 <dim>-1</dim>
1541 <dim>32</dim>
1542 </port>
1543 </output>
1544 </layer>
1545 <layer id="113" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1546 <data element_type="i32" shape="4" offset="12502400" size="16" />
1547 <output>
1548 <port id="0" precision="I32">
1549 <dim>4</dim>
1550 </port>
1551 </output>
1552 </layer>
1553 <layer id="114" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1554 <input>
1555 <port id="0" precision="FP32">
1556 <dim>-1</dim>
1557 <dim>12</dim>
1558 <dim>-1</dim>
1559 <dim>32</dim>
1560 </port>
1561 <port id="1" precision="I32">
1562 <dim>4</dim>
1563 </port>
1564 </input>
1565 <output>
1566 <port id="2" precision="FP32" names="122,attn_output.3">
1567 <dim>-1</dim>
1568 <dim>-1</dim>
1569 <dim>12</dim>
1570 <dim>32</dim>
1571 </port>
1572 </output>
1573 </layer>
1574 <layer id="115" name="Constant_90443" type="Const" version="opset1">
1575 <data element_type="i64" shape="3" offset="12502416" size="24" />
1576 <output>
1577 <port id="0" precision="I64">
1578 <dim>3</dim>
1579 </port>
1580 </output>
1581 </layer>
1582 <layer id="116" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1583 <data special_zero="true" />
1584 <input>
1585 <port id="0" precision="FP32">
1586 <dim>-1</dim>
1587 <dim>-1</dim>
1588 <dim>12</dim>
1589 <dim>32</dim>
1590 </port>
1591 <port id="1" precision="I64">
1592 <dim>3</dim>
1593 </port>
1594 </input>
1595 <output>
1596 <port id="2" precision="FP32" names="124">
1597 <dim>-1</dim>
1598 <dim>-1</dim>
1599 <dim>384</dim>
1600 </port>
1601 </output>
1602 </layer>
1603 <layer id="117" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1604 <data element_type="f32" shape="1, 1, 384" offset="12502440" size="1536" />
1605 <output>
1606 <port id="0" precision="FP32">
1607 <dim>1</dim>
1608 <dim>1</dim>
1609 <dim>384</dim>
1610 </port>
1611 </output>
1612 </layer>
1613 <layer id="118" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1614 <data auto_broadcast="numpy" />
1615 <input>
1616 <port id="0" precision="FP32">
1617 <dim>-1</dim>
1618 <dim>-1</dim>
1619 <dim>384</dim>
1620 </port>
1621 <port id="1" precision="FP32">
1622 <dim>1</dim>
1623 <dim>1</dim>
1624 <dim>384</dim>
1625 </port>
1626 </input>
1627 <output>
1628 <port id="2" precision="FP32">
1629 <dim>-1</dim>
1630 <dim>-1</dim>
1631 <dim>384</dim>
1632 </port>
1633 </output>
1634 </layer>
1635 <layer id="119" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
1636 <data element_type="f32" shape="" offset="12503976" size="4" />
1637 <output>
1638 <port id="0" precision="FP32" />
1639 </output>
1640 </layer>
1641 <layer id="120" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
1642 <data element_type="f32" shape="" offset="12503980" size="4" />
1643 <output>
1644 <port id="0" precision="FP32" />
1645 </output>
1646 </layer>
1647 <layer id="121" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
1648 <data element_type="f32" shape="" offset="12503976" size="4" />
1649 <output>
1650 <port id="0" precision="FP32" />
1651 </output>
1652 </layer>
1653 <layer id="122" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
1654 <data element_type="f32" shape="" offset="12503980" size="4" />
1655 <output>
1656 <port id="0" precision="FP32" />
1657 </output>
1658 </layer>
1659 <layer id="123" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
1660 <data levels="256" auto_broadcast="numpy" />
1661 <input>
1662 <port id="0" precision="FP32">
1663 <dim>-1</dim>
1664 <dim>-1</dim>
1665 <dim>384</dim>
1666 </port>
1667 <port id="1" precision="FP32" />
1668 <port id="2" precision="FP32" />
1669 <port id="3" precision="FP32" />
1670 <port id="4" precision="FP32" />
1671 </input>
1672 <output>
1673 <port id="5" precision="FP32">
1674 <dim>-1</dim>
1675 <dim>-1</dim>
1676 <dim>384</dim>
1677 </port>
1678 </output>
1679 </layer>
1680 <layer id="124" name="Constant_299950" type="Const" version="opset1">
1681 <data element_type="i8" shape="384, 384" offset="12503984" size="147456" />
1682 <output>
1683 <port id="0" precision="I8">
1684 <dim>384</dim>
1685 <dim>384</dim>
1686 </port>
1687 </output>
1688 </layer>
1689 <layer id="125" name="Convert_299951" type="Convert" version="opset1">
1690 <data destination_type="f32" />
1691 <input>
1692 <port id="0" precision="I8">
1693 <dim>384</dim>
1694 <dim>384</dim>
1695 </port>
1696 </input>
1697 <output>
1698 <port id="1" precision="FP32">
1699 <dim>384</dim>
1700 <dim>384</dim>
1701 </port>
1702 </output>
1703 </layer>
1704 <layer id="126" name="Constant_299952" type="Const" version="opset1">
1705 <data element_type="f32" shape="384, 1" offset="12651440" size="1536" />
1706 <output>
1707 <port id="0" precision="FP32">
1708 <dim>384</dim>
1709 <dim>1</dim>
1710 </port>
1711 </output>
1712 </layer>
1713 <layer id="127" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1714 <data auto_broadcast="numpy" />
1715 <input>
1716 <port id="0" precision="FP32">
1717 <dim>384</dim>
1718 <dim>384</dim>
1719 </port>
1720 <port id="1" precision="FP32">
1721 <dim>384</dim>
1722 <dim>1</dim>
1723 </port>
1724 </input>
1725 <output>
1726 <port id="2" precision="FP32">
1727 <dim>384</dim>
1728 <dim>384</dim>
1729 </port>
1730 </output>
1731 </layer>
1732 <layer id="128" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1733 <data transpose_a="false" transpose_b="true" />
1734 <input>
1735 <port id="0" precision="FP32">
1736 <dim>-1</dim>
1737 <dim>-1</dim>
1738 <dim>384</dim>
1739 </port>
1740 <port id="1" precision="FP32">
1741 <dim>384</dim>
1742 <dim>384</dim>
1743 </port>
1744 </input>
1745 <output>
1746 <port id="2" precision="FP32">
1747 <dim>-1</dim>
1748 <dim>-1</dim>
1749 <dim>384</dim>
1750 </port>
1751 </output>
1752 </layer>
1753 <layer id="129" name="Constant_90308" type="Const" version="opset1">
1754 <data element_type="f32" shape="1, 1, 384" offset="12652976" size="1536" />
1755 <output>
1756 <port id="0" precision="FP32">
1757 <dim>1</dim>
1758 <dim>1</dim>
1759 <dim>384</dim>
1760 </port>
1761 </output>
1762 </layer>
1763 <layer id="130" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1764 <data auto_broadcast="numpy" />
1765 <input>
1766 <port id="0" precision="FP32">
1767 <dim>-1</dim>
1768 <dim>-1</dim>
1769 <dim>384</dim>
1770 </port>
1771 <port id="1" precision="FP32">
1772 <dim>1</dim>
1773 <dim>1</dim>
1774 <dim>384</dim>
1775 </port>
1776 </input>
1777 <output>
1778 <port id="2" precision="FP32" names="129,input.3">
1779 <dim>-1</dim>
1780 <dim>-1</dim>
1781 <dim>384</dim>
1782 </port>
1783 </output>
1784 </layer>
1785 <layer id="131" name="__module.bert.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1786 <data auto_broadcast="numpy" />
1787 <input>
1788 <port id="0" precision="FP32">
1789 <dim>-1</dim>
1790 <dim>-1</dim>
1791 <dim>384</dim>
1792 </port>
1793 <port id="1" precision="FP32">
1794 <dim>-1</dim>
1795 <dim>-1</dim>
1796 <dim>384</dim>
1797 </port>
1798 </input>
1799 <output>
1800 <port id="2" precision="FP32" names="131">
1801 <dim>-1</dim>
1802 <dim>-1</dim>
1803 <dim>384</dim>
1804 </port>
1805 </output>
1806 </layer>
1807 <layer id="132" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1808 <data element_type="i32" shape="1" offset="12046084" size="4" />
1809 <output>
1810 <port id="0" precision="I32">
1811 <dim>1</dim>
1812 </port>
1813 </output>
1814 </layer>
1815 <layer id="133" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1816 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1817 <input>
1818 <port id="0" precision="FP32">
1819 <dim>-1</dim>
1820 <dim>-1</dim>
1821 <dim>384</dim>
1822 </port>
1823 <port id="1" precision="I32">
1824 <dim>1</dim>
1825 </port>
1826 </input>
1827 <output>
1828 <port id="2" precision="FP32">
1829 <dim>-1</dim>
1830 <dim>-1</dim>
1831 <dim>384</dim>
1832 </port>
1833 </output>
1834 </layer>
1835 <layer id="134" name="Constant_90309" type="Const" version="opset1">
1836 <data element_type="f32" shape="1, 1, 384" offset="12654512" size="1536" />
1837 <output>
1838 <port id="0" precision="FP32">
1839 <dim>1</dim>
1840 <dim>1</dim>
1841 <dim>384</dim>
1842 </port>
1843 </output>
1844 </layer>
1845 <layer id="135" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1846 <data auto_broadcast="numpy" />
1847 <input>
1848 <port id="0" precision="FP32">
1849 <dim>-1</dim>
1850 <dim>-1</dim>
1851 <dim>384</dim>
1852 </port>
1853 <port id="1" precision="FP32">
1854 <dim>1</dim>
1855 <dim>1</dim>
1856 <dim>384</dim>
1857 </port>
1858 </input>
1859 <output>
1860 <port id="2" precision="FP32">
1861 <dim>-1</dim>
1862 <dim>-1</dim>
1863 <dim>384</dim>
1864 </port>
1865 </output>
1866 </layer>
1867 <layer id="136" name="Constant_90310" type="Const" version="opset1">
1868 <data element_type="f32" shape="1, 1, 384" offset="12656048" size="1536" />
1869 <output>
1870 <port id="0" precision="FP32">
1871 <dim>1</dim>
1872 <dim>1</dim>
1873 <dim>384</dim>
1874 </port>
1875 </output>
1876 </layer>
1877 <layer id="137" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1878 <data auto_broadcast="numpy" />
1879 <input>
1880 <port id="0" precision="FP32">
1881 <dim>-1</dim>
1882 <dim>-1</dim>
1883 <dim>384</dim>
1884 </port>
1885 <port id="1" precision="FP32">
1886 <dim>1</dim>
1887 <dim>1</dim>
1888 <dim>384</dim>
1889 </port>
1890 </input>
1891 <output>
1892 <port id="2" precision="FP32" names="135,input_tensor.1">
1893 <dim>-1</dim>
1894 <dim>-1</dim>
1895 <dim>384</dim>
1896 </port>
1897 </output>
1898 </layer>
1899 <layer id="138" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1900 <data element_type="f32" shape="1, 1, 384" offset="12657584" size="1536" />
1901 <output>
1902 <port id="0" precision="FP32">
1903 <dim>1</dim>
1904 <dim>1</dim>
1905 <dim>384</dim>
1906 </port>
1907 </output>
1908 </layer>
1909 <layer id="139" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1910 <data auto_broadcast="numpy" />
1911 <input>
1912 <port id="0" precision="FP32">
1913 <dim>-1</dim>
1914 <dim>-1</dim>
1915 <dim>384</dim>
1916 </port>
1917 <port id="1" precision="FP32">
1918 <dim>1</dim>
1919 <dim>1</dim>
1920 <dim>384</dim>
1921 </port>
1922 </input>
1923 <output>
1924 <port id="2" precision="FP32">
1925 <dim>-1</dim>
1926 <dim>-1</dim>
1927 <dim>384</dim>
1928 </port>
1929 </output>
1930 </layer>
1931 <layer id="140" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
1932 <data element_type="f32" shape="" offset="12659120" size="4" />
1933 <output>
1934 <port id="0" precision="FP32" />
1935 </output>
1936 </layer>
1937 <layer id="141" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
1938 <data element_type="f32" shape="" offset="12659124" size="4" />
1939 <output>
1940 <port id="0" precision="FP32" />
1941 </output>
1942 </layer>
1943 <layer id="142" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
1944 <data element_type="f32" shape="" offset="12659120" size="4" />
1945 <output>
1946 <port id="0" precision="FP32" />
1947 </output>
1948 </layer>
1949 <layer id="143" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
1950 <data element_type="f32" shape="" offset="12659124" size="4" />
1951 <output>
1952 <port id="0" precision="FP32" />
1953 </output>
1954 </layer>
1955 <layer id="144" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
1956 <data levels="256" auto_broadcast="numpy" />
1957 <input>
1958 <port id="0" precision="FP32">
1959 <dim>-1</dim>
1960 <dim>-1</dim>
1961 <dim>384</dim>
1962 </port>
1963 <port id="1" precision="FP32" />
1964 <port id="2" precision="FP32" />
1965 <port id="3" precision="FP32" />
1966 <port id="4" precision="FP32" />
1967 </input>
1968 <output>
1969 <port id="5" precision="FP32">
1970 <dim>-1</dim>
1971 <dim>-1</dim>
1972 <dim>384</dim>
1973 </port>
1974 </output>
1975 </layer>
1976 <layer id="145" name="Constant_299954" type="Const" version="opset1">
1977 <data element_type="i8" shape="1536, 384" offset="12659128" size="589824" />
1978 <output>
1979 <port id="0" precision="I8">
1980 <dim>1536</dim>
1981 <dim>384</dim>
1982 </port>
1983 </output>
1984 </layer>
1985 <layer id="146" name="Convert_299955" type="Convert" version="opset1">
1986 <data destination_type="f32" />
1987 <input>
1988 <port id="0" precision="I8">
1989 <dim>1536</dim>
1990 <dim>384</dim>
1991 </port>
1992 </input>
1993 <output>
1994 <port id="1" precision="FP32">
1995 <dim>1536</dim>
1996 <dim>384</dim>
1997 </port>
1998 </output>
1999 </layer>
2000 <layer id="147" name="Constant_299956" type="Const" version="opset1">
2001 <data element_type="f32" shape="1536, 1" offset="13248952" size="6144" />
2002 <output>
2003 <port id="0" precision="FP32">
2004 <dim>1536</dim>
2005 <dim>1</dim>
2006 </port>
2007 </output>
2008 </layer>
2009 <layer id="148" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2010 <data auto_broadcast="numpy" />
2011 <input>
2012 <port id="0" precision="FP32">
2013 <dim>1536</dim>
2014 <dim>384</dim>
2015 </port>
2016 <port id="1" precision="FP32">
2017 <dim>1536</dim>
2018 <dim>1</dim>
2019 </port>
2020 </input>
2021 <output>
2022 <port id="2" precision="FP32">
2023 <dim>1536</dim>
2024 <dim>384</dim>
2025 </port>
2026 </output>
2027 </layer>
2028 <layer id="149" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2029 <data transpose_a="false" transpose_b="true" />
2030 <input>
2031 <port id="0" precision="FP32">
2032 <dim>-1</dim>
2033 <dim>-1</dim>
2034 <dim>384</dim>
2035 </port>
2036 <port id="1" precision="FP32">
2037 <dim>1536</dim>
2038 <dim>384</dim>
2039 </port>
2040 </input>
2041 <output>
2042 <port id="2" precision="FP32">
2043 <dim>-1</dim>
2044 <dim>-1</dim>
2045 <dim>1536</dim>
2046 </port>
2047 </output>
2048 </layer>
2049 <layer id="150" name="Constant_90311" type="Const" version="opset1">
2050 <data element_type="f32" shape="1, 1, 1536" offset="13255096" size="6144" />
2051 <output>
2052 <port id="0" precision="FP32">
2053 <dim>1</dim>
2054 <dim>1</dim>
2055 <dim>1536</dim>
2056 </port>
2057 </output>
2058 </layer>
2059 <layer id="151" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2060 <data auto_broadcast="numpy" />
2061 <input>
2062 <port id="0" precision="FP32">
2063 <dim>-1</dim>
2064 <dim>-1</dim>
2065 <dim>1536</dim>
2066 </port>
2067 <port id="1" precision="FP32">
2068 <dim>1</dim>
2069 <dim>1</dim>
2070 <dim>1536</dim>
2071 </port>
2072 </input>
2073 <output>
2074 <port id="2" precision="FP32" names="139">
2075 <dim>-1</dim>
2076 <dim>-1</dim>
2077 <dim>1536</dim>
2078 </port>
2079 </output>
2080 </layer>
2081 <layer id="152" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2082 <data approximation_mode="ERF" />
2083 <input>
2084 <port id="0" precision="FP32">
2085 <dim>-1</dim>
2086 <dim>-1</dim>
2087 <dim>1536</dim>
2088 </port>
2089 </input>
2090 <output>
2091 <port id="1" precision="FP32" names="140">
2092 <dim>-1</dim>
2093 <dim>-1</dim>
2094 <dim>1536</dim>
2095 </port>
2096 </output>
2097 </layer>
2098 <layer id="153" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2099 <data element_type="f32" shape="1, 1, 1536" offset="13261240" size="6144" />
2100 <output>
2101 <port id="0" precision="FP32">
2102 <dim>1</dim>
2103 <dim>1</dim>
2104 <dim>1536</dim>
2105 </port>
2106 </output>
2107 </layer>
2108 <layer id="154" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2109 <data auto_broadcast="numpy" />
2110 <input>
2111 <port id="0" precision="FP32">
2112 <dim>-1</dim>
2113 <dim>-1</dim>
2114 <dim>1536</dim>
2115 </port>
2116 <port id="1" precision="FP32">
2117 <dim>1</dim>
2118 <dim>1</dim>
2119 <dim>1536</dim>
2120 </port>
2121 </input>
2122 <output>
2123 <port id="2" precision="FP32">
2124 <dim>-1</dim>
2125 <dim>-1</dim>
2126 <dim>1536</dim>
2127 </port>
2128 </output>
2129 </layer>
2130 <layer id="155" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2131 <data element_type="f32" shape="" offset="13267384" size="4" />
2132 <output>
2133 <port id="0" precision="FP32" />
2134 </output>
2135 </layer>
2136 <layer id="156" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2137 <data element_type="f32" shape="" offset="13267388" size="4" />
2138 <output>
2139 <port id="0" precision="FP32" />
2140 </output>
2141 </layer>
2142 <layer id="157" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2143 <data element_type="f32" shape="" offset="13267384" size="4" />
2144 <output>
2145 <port id="0" precision="FP32" />
2146 </output>
2147 </layer>
2148 <layer id="158" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2149 <data element_type="f32" shape="" offset="13267388" size="4" />
2150 <output>
2151 <port id="0" precision="FP32" />
2152 </output>
2153 </layer>
2154 <layer id="159" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2155 <data levels="256" auto_broadcast="numpy" />
2156 <input>
2157 <port id="0" precision="FP32">
2158 <dim>-1</dim>
2159 <dim>-1</dim>
2160 <dim>1536</dim>
2161 </port>
2162 <port id="1" precision="FP32" />
2163 <port id="2" precision="FP32" />
2164 <port id="3" precision="FP32" />
2165 <port id="4" precision="FP32" />
2166 </input>
2167 <output>
2168 <port id="5" precision="FP32">
2169 <dim>-1</dim>
2170 <dim>-1</dim>
2171 <dim>1536</dim>
2172 </port>
2173 </output>
2174 </layer>
2175 <layer id="160" name="Constant_299958" type="Const" version="opset1">
2176 <data element_type="i8" shape="384, 1536" offset="13267392" size="589824" />
2177 <output>
2178 <port id="0" precision="I8">
2179 <dim>384</dim>
2180 <dim>1536</dim>
2181 </port>
2182 </output>
2183 </layer>
2184 <layer id="161" name="Convert_299959" type="Convert" version="opset1">
2185 <data destination_type="f32" />
2186 <input>
2187 <port id="0" precision="I8">
2188 <dim>384</dim>
2189 <dim>1536</dim>
2190 </port>
2191 </input>
2192 <output>
2193 <port id="1" precision="FP32">
2194 <dim>384</dim>
2195 <dim>1536</dim>
2196 </port>
2197 </output>
2198 </layer>
2199 <layer id="162" name="Constant_299960" type="Const" version="opset1">
2200 <data element_type="f32" shape="384, 1" offset="13857216" size="1536" />
2201 <output>
2202 <port id="0" precision="FP32">
2203 <dim>384</dim>
2204 <dim>1</dim>
2205 </port>
2206 </output>
2207 </layer>
2208 <layer id="163" name="__module.bert.encoder.layer.0.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2209 <data auto_broadcast="numpy" />
2210 <input>
2211 <port id="0" precision="FP32">
2212 <dim>384</dim>
2213 <dim>1536</dim>
2214 </port>
2215 <port id="1" precision="FP32">
2216 <dim>384</dim>
2217 <dim>1</dim>
2218 </port>
2219 </input>
2220 <output>
2221 <port id="2" precision="FP32">
2222 <dim>384</dim>
2223 <dim>1536</dim>
2224 </port>
2225 </output>
2226 </layer>
2227 <layer id="164" name="__module.bert.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2228 <data transpose_a="false" transpose_b="true" />
2229 <input>
2230 <port id="0" precision="FP32">
2231 <dim>-1</dim>
2232 <dim>-1</dim>
2233 <dim>1536</dim>
2234 </port>
2235 <port id="1" precision="FP32">
2236 <dim>384</dim>
2237 <dim>1536</dim>
2238 </port>
2239 </input>
2240 <output>
2241 <port id="2" precision="FP32">
2242 <dim>-1</dim>
2243 <dim>-1</dim>
2244 <dim>384</dim>
2245 </port>
2246 </output>
2247 </layer>
2248 <layer id="165" name="Constant_90312" type="Const" version="opset1">
2249 <data element_type="f32" shape="1, 1, 384" offset="13858752" size="1536" />
2250 <output>
2251 <port id="0" precision="FP32">
2252 <dim>1</dim>
2253 <dim>1</dim>
2254 <dim>384</dim>
2255 </port>
2256 </output>
2257 </layer>
2258 <layer id="166" name="__module.bert.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
2259 <data auto_broadcast="numpy" />
2260 <input>
2261 <port id="0" precision="FP32">
2262 <dim>-1</dim>
2263 <dim>-1</dim>
2264 <dim>384</dim>
2265 </port>
2266 <port id="1" precision="FP32">
2267 <dim>1</dim>
2268 <dim>1</dim>
2269 <dim>384</dim>
2270 </port>
2271 </input>
2272 <output>
2273 <port id="2" precision="FP32" names="145,input.5">
2274 <dim>-1</dim>
2275 <dim>-1</dim>
2276 <dim>384</dim>
2277 </port>
2278 </output>
2279 </layer>
2280 <layer id="167" name="__module.bert.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
2281 <data auto_broadcast="numpy" />
2282 <input>
2283 <port id="0" precision="FP32">
2284 <dim>-1</dim>
2285 <dim>-1</dim>
2286 <dim>384</dim>
2287 </port>
2288 <port id="1" precision="FP32">
2289 <dim>-1</dim>
2290 <dim>-1</dim>
2291 <dim>384</dim>
2292 </port>
2293 </input>
2294 <output>
2295 <port id="2" precision="FP32" names="147">
2296 <dim>-1</dim>
2297 <dim>-1</dim>
2298 <dim>384</dim>
2299 </port>
2300 </output>
2301 </layer>
2302 <layer id="168" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2303 <data element_type="i32" shape="1" offset="12046084" size="4" />
2304 <output>
2305 <port id="0" precision="I32">
2306 <dim>1</dim>
2307 </port>
2308 </output>
2309 </layer>
2310 <layer id="169" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2311 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2312 <input>
2313 <port id="0" precision="FP32">
2314 <dim>-1</dim>
2315 <dim>-1</dim>
2316 <dim>384</dim>
2317 </port>
2318 <port id="1" precision="I32">
2319 <dim>1</dim>
2320 </port>
2321 </input>
2322 <output>
2323 <port id="2" precision="FP32">
2324 <dim>-1</dim>
2325 <dim>-1</dim>
2326 <dim>384</dim>
2327 </port>
2328 </output>
2329 </layer>
2330 <layer id="170" name="Constant_90313" type="Const" version="opset1">
2331 <data element_type="f32" shape="1, 1, 384" offset="13860288" size="1536" />
2332 <output>
2333 <port id="0" precision="FP32">
2334 <dim>1</dim>
2335 <dim>1</dim>
2336 <dim>384</dim>
2337 </port>
2338 </output>
2339 </layer>
2340 <layer id="171" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2341 <data auto_broadcast="numpy" />
2342 <input>
2343 <port id="0" precision="FP32">
2344 <dim>-1</dim>
2345 <dim>-1</dim>
2346 <dim>384</dim>
2347 </port>
2348 <port id="1" precision="FP32">
2349 <dim>1</dim>
2350 <dim>1</dim>
2351 <dim>384</dim>
2352 </port>
2353 </input>
2354 <output>
2355 <port id="2" precision="FP32">
2356 <dim>-1</dim>
2357 <dim>-1</dim>
2358 <dim>384</dim>
2359 </port>
2360 </output>
2361 </layer>
2362 <layer id="172" name="Constant_90314" type="Const" version="opset1">
2363 <data element_type="f32" shape="1, 1, 384" offset="13861824" size="1536" />
2364 <output>
2365 <port id="0" precision="FP32">
2366 <dim>1</dim>
2367 <dim>1</dim>
2368 <dim>384</dim>
2369 </port>
2370 </output>
2371 </layer>
2372 <layer id="173" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2373 <data auto_broadcast="numpy" />
2374 <input>
2375 <port id="0" precision="FP32">
2376 <dim>-1</dim>
2377 <dim>-1</dim>
2378 <dim>384</dim>
2379 </port>
2380 <port id="1" precision="FP32">
2381 <dim>1</dim>
2382 <dim>1</dim>
2383 <dim>384</dim>
2384 </port>
2385 </input>
2386 <output>
2387 <port id="2" precision="FP32" names="151,hidden_states.7">
2388 <dim>-1</dim>
2389 <dim>-1</dim>
2390 <dim>384</dim>
2391 </port>
2392 </output>
2393 </layer>
2394 <layer id="174" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2395 <data element_type="f32" shape="1, 1, 384" offset="13863360" size="1536" />
2396 <output>
2397 <port id="0" precision="FP32">
2398 <dim>1</dim>
2399 <dim>1</dim>
2400 <dim>384</dim>
2401 </port>
2402 </output>
2403 </layer>
2404 <layer id="175" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2405 <data auto_broadcast="numpy" />
2406 <input>
2407 <port id="0" precision="FP32">
2408 <dim>-1</dim>
2409 <dim>-1</dim>
2410 <dim>384</dim>
2411 </port>
2412 <port id="1" precision="FP32">
2413 <dim>1</dim>
2414 <dim>1</dim>
2415 <dim>384</dim>
2416 </port>
2417 </input>
2418 <output>
2419 <port id="2" precision="FP32">
2420 <dim>-1</dim>
2421 <dim>-1</dim>
2422 <dim>384</dim>
2423 </port>
2424 </output>
2425 </layer>
2426 <layer id="176" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2427 <data element_type="f32" shape="" offset="13864896" size="4" />
2428 <output>
2429 <port id="0" precision="FP32" />
2430 </output>
2431 </layer>
2432 <layer id="177" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2433 <data element_type="f32" shape="" offset="13864900" size="4" />
2434 <output>
2435 <port id="0" precision="FP32" />
2436 </output>
2437 </layer>
2438 <layer id="178" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2439 <data element_type="f32" shape="" offset="13864896" size="4" />
2440 <output>
2441 <port id="0" precision="FP32" />
2442 </output>
2443 </layer>
2444 <layer id="179" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2445 <data element_type="f32" shape="" offset="13864900" size="4" />
2446 <output>
2447 <port id="0" precision="FP32" />
2448 </output>
2449 </layer>
2450 <layer id="180" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2451 <data levels="256" auto_broadcast="numpy" />
2452 <input>
2453 <port id="0" precision="FP32">
2454 <dim>-1</dim>
2455 <dim>-1</dim>
2456 <dim>384</dim>
2457 </port>
2458 <port id="1" precision="FP32" />
2459 <port id="2" precision="FP32" />
2460 <port id="3" precision="FP32" />
2461 <port id="4" precision="FP32" />
2462 </input>
2463 <output>
2464 <port id="5" precision="FP32">
2465 <dim>-1</dim>
2466 <dim>-1</dim>
2467 <dim>384</dim>
2468 </port>
2469 </output>
2470 </layer>
2471 <layer id="181" name="Constant_299962" type="Const" version="opset1">
2472 <data element_type="i8" shape="384, 384" offset="13864904" size="147456" />
2473 <output>
2474 <port id="0" precision="I8">
2475 <dim>384</dim>
2476 <dim>384</dim>
2477 </port>
2478 </output>
2479 </layer>
2480 <layer id="182" name="Convert_299963" type="Convert" version="opset1">
2481 <data destination_type="f32" />
2482 <input>
2483 <port id="0" precision="I8">
2484 <dim>384</dim>
2485 <dim>384</dim>
2486 </port>
2487 </input>
2488 <output>
2489 <port id="1" precision="FP32">
2490 <dim>384</dim>
2491 <dim>384</dim>
2492 </port>
2493 </output>
2494 </layer>
2495 <layer id="183" name="Constant_299964" type="Const" version="opset1">
2496 <data element_type="f32" shape="384, 1" offset="14012360" size="1536" />
2497 <output>
2498 <port id="0" precision="FP32">
2499 <dim>384</dim>
2500 <dim>1</dim>
2501 </port>
2502 </output>
2503 </layer>
2504 <layer id="184" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2505 <data auto_broadcast="numpy" />
2506 <input>
2507 <port id="0" precision="FP32">
2508 <dim>384</dim>
2509 <dim>384</dim>
2510 </port>
2511 <port id="1" precision="FP32">
2512 <dim>384</dim>
2513 <dim>1</dim>
2514 </port>
2515 </input>
2516 <output>
2517 <port id="2" precision="FP32">
2518 <dim>384</dim>
2519 <dim>384</dim>
2520 </port>
2521 </output>
2522 </layer>
2523 <layer id="185" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2524 <data transpose_a="false" transpose_b="true" />
2525 <input>
2526 <port id="0" precision="FP32">
2527 <dim>-1</dim>
2528 <dim>-1</dim>
2529 <dim>384</dim>
2530 </port>
2531 <port id="1" precision="FP32">
2532 <dim>384</dim>
2533 <dim>384</dim>
2534 </port>
2535 </input>
2536 <output>
2537 <port id="2" precision="FP32">
2538 <dim>-1</dim>
2539 <dim>-1</dim>
2540 <dim>384</dim>
2541 </port>
2542 </output>
2543 </layer>
2544 <layer id="186" name="Constant_90315" type="Const" version="opset1">
2545 <data element_type="f32" shape="1, 1, 384" offset="14013896" size="1536" />
2546 <output>
2547 <port id="0" precision="FP32">
2548 <dim>1</dim>
2549 <dim>1</dim>
2550 <dim>384</dim>
2551 </port>
2552 </output>
2553 </layer>
2554 <layer id="187" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2555 <data auto_broadcast="numpy" />
2556 <input>
2557 <port id="0" precision="FP32">
2558 <dim>-1</dim>
2559 <dim>-1</dim>
2560 <dim>384</dim>
2561 </port>
2562 <port id="1" precision="FP32">
2563 <dim>1</dim>
2564 <dim>1</dim>
2565 <dim>384</dim>
2566 </port>
2567 </input>
2568 <output>
2569 <port id="2" precision="FP32" names="164,x.13">
2570 <dim>-1</dim>
2571 <dim>-1</dim>
2572 <dim>384</dim>
2573 </port>
2574 </output>
2575 </layer>
2576 <layer id="188" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2577 <data element_type="f32" shape="" offset="14015432" size="4" />
2578 <output>
2579 <port id="0" precision="FP32" />
2580 </output>
2581 </layer>
2582 <layer id="189" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2583 <data element_type="f32" shape="" offset="14015436" size="4" />
2584 <output>
2585 <port id="0" precision="FP32" />
2586 </output>
2587 </layer>
2588 <layer id="190" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2589 <data element_type="f32" shape="" offset="14015432" size="4" />
2590 <output>
2591 <port id="0" precision="FP32" />
2592 </output>
2593 </layer>
2594 <layer id="191" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2595 <data element_type="f32" shape="" offset="14015436" size="4" />
2596 <output>
2597 <port id="0" precision="FP32" />
2598 </output>
2599 </layer>
2600 <layer id="192" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2601 <data levels="256" auto_broadcast="numpy" />
2602 <input>
2603 <port id="0" precision="FP32">
2604 <dim>-1</dim>
2605 <dim>-1</dim>
2606 <dim>384</dim>
2607 </port>
2608 <port id="1" precision="FP32" />
2609 <port id="2" precision="FP32" />
2610 <port id="3" precision="FP32" />
2611 <port id="4" precision="FP32" />
2612 </input>
2613 <output>
2614 <port id="5" precision="FP32">
2615 <dim>-1</dim>
2616 <dim>-1</dim>
2617 <dim>384</dim>
2618 </port>
2619 </output>
2620 </layer>
2621 <layer id="193" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2622 <data element_type="i64" shape="4" offset="12201240" size="32" />
2623 <output>
2624 <port id="0" precision="I64">
2625 <dim>4</dim>
2626 </port>
2627 </output>
2628 </layer>
2629 <layer id="194" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2630 <data special_zero="true" />
2631 <input>
2632 <port id="0" precision="FP32">
2633 <dim>-1</dim>
2634 <dim>-1</dim>
2635 <dim>384</dim>
2636 </port>
2637 <port id="1" precision="I64">
2638 <dim>4</dim>
2639 </port>
2640 </input>
2641 <output>
2642 <port id="2" precision="FP32" names="168,x.15">
2643 <dim>-1</dim>
2644 <dim>-1</dim>
2645 <dim>12</dim>
2646 <dim>32</dim>
2647 </port>
2648 </output>
2649 </layer>
2650 <layer id="195" name="Constant_84681" type="Const" version="opset1">
2651 <data element_type="i64" shape="4" offset="12201272" size="32" />
2652 <output>
2653 <port id="0" precision="I64" names="169">
2654 <dim>4</dim>
2655 </port>
2656 </output>
2657 </layer>
2658 <layer id="196" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2659 <input>
2660 <port id="0" precision="FP32">
2661 <dim>-1</dim>
2662 <dim>-1</dim>
2663 <dim>12</dim>
2664 <dim>32</dim>
2665 </port>
2666 <port id="1" precision="I64">
2667 <dim>4</dim>
2668 </port>
2669 </input>
2670 <output>
2671 <port id="2" precision="FP32" names="170">
2672 <dim>-1</dim>
2673 <dim>12</dim>
2674 <dim>-1</dim>
2675 <dim>32</dim>
2676 </port>
2677 </output>
2678 </layer>
2679 <layer id="197" name="Constant_299966" type="Const" version="opset1">
2680 <data element_type="i8" shape="384, 384" offset="14015440" size="147456" />
2681 <output>
2682 <port id="0" precision="I8">
2683 <dim>384</dim>
2684 <dim>384</dim>
2685 </port>
2686 </output>
2687 </layer>
2688 <layer id="198" name="Convert_299967" type="Convert" version="opset1">
2689 <data destination_type="f32" />
2690 <input>
2691 <port id="0" precision="I8">
2692 <dim>384</dim>
2693 <dim>384</dim>
2694 </port>
2695 </input>
2696 <output>
2697 <port id="1" precision="FP32">
2698 <dim>384</dim>
2699 <dim>384</dim>
2700 </port>
2701 </output>
2702 </layer>
2703 <layer id="199" name="Constant_299968" type="Const" version="opset1">
2704 <data element_type="f32" shape="384, 1" offset="14162896" size="1536" />
2705 <output>
2706 <port id="0" precision="FP32">
2707 <dim>384</dim>
2708 <dim>1</dim>
2709 </port>
2710 </output>
2711 </layer>
2712 <layer id="200" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2713 <data auto_broadcast="numpy" />
2714 <input>
2715 <port id="0" precision="FP32">
2716 <dim>384</dim>
2717 <dim>384</dim>
2718 </port>
2719 <port id="1" precision="FP32">
2720 <dim>384</dim>
2721 <dim>1</dim>
2722 </port>
2723 </input>
2724 <output>
2725 <port id="2" precision="FP32">
2726 <dim>384</dim>
2727 <dim>384</dim>
2728 </port>
2729 </output>
2730 </layer>
2731 <layer id="201" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2732 <data transpose_a="false" transpose_b="true" />
2733 <input>
2734 <port id="0" precision="FP32">
2735 <dim>-1</dim>
2736 <dim>-1</dim>
2737 <dim>384</dim>
2738 </port>
2739 <port id="1" precision="FP32">
2740 <dim>384</dim>
2741 <dim>384</dim>
2742 </port>
2743 </input>
2744 <output>
2745 <port id="2" precision="FP32">
2746 <dim>-1</dim>
2747 <dim>-1</dim>
2748 <dim>384</dim>
2749 </port>
2750 </output>
2751 </layer>
2752 <layer id="202" name="Constant_90316" type="Const" version="opset1">
2753 <data element_type="f32" shape="1, 1, 384" offset="14164432" size="1536" />
2754 <output>
2755 <port id="0" precision="FP32">
2756 <dim>1</dim>
2757 <dim>1</dim>
2758 <dim>384</dim>
2759 </port>
2760 </output>
2761 </layer>
2762 <layer id="203" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2763 <data auto_broadcast="numpy" />
2764 <input>
2765 <port id="0" precision="FP32">
2766 <dim>-1</dim>
2767 <dim>-1</dim>
2768 <dim>384</dim>
2769 </port>
2770 <port id="1" precision="FP32">
2771 <dim>1</dim>
2772 <dim>1</dim>
2773 <dim>384</dim>
2774 </port>
2775 </input>
2776 <output>
2777 <port id="2" precision="FP32" names="173,x.17">
2778 <dim>-1</dim>
2779 <dim>-1</dim>
2780 <dim>384</dim>
2781 </port>
2782 </output>
2783 </layer>
2784 <layer id="204" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2785 <data element_type="f32" shape="" offset="14165968" size="4" />
2786 <output>
2787 <port id="0" precision="FP32" />
2788 </output>
2789 </layer>
2790 <layer id="205" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2791 <data element_type="f32" shape="" offset="14165972" size="4" />
2792 <output>
2793 <port id="0" precision="FP32" />
2794 </output>
2795 </layer>
2796 <layer id="206" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2797 <data element_type="f32" shape="" offset="14165968" size="4" />
2798 <output>
2799 <port id="0" precision="FP32" />
2800 </output>
2801 </layer>
2802 <layer id="207" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2803 <data element_type="f32" shape="" offset="14165972" size="4" />
2804 <output>
2805 <port id="0" precision="FP32" />
2806 </output>
2807 </layer>
2808 <layer id="208" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2809 <data levels="256" auto_broadcast="numpy" />
2810 <input>
2811 <port id="0" precision="FP32">
2812 <dim>-1</dim>
2813 <dim>-1</dim>
2814 <dim>384</dim>
2815 </port>
2816 <port id="1" precision="FP32" />
2817 <port id="2" precision="FP32" />
2818 <port id="3" precision="FP32" />
2819 <port id="4" precision="FP32" />
2820 </input>
2821 <output>
2822 <port id="5" precision="FP32">
2823 <dim>-1</dim>
2824 <dim>-1</dim>
2825 <dim>384</dim>
2826 </port>
2827 </output>
2828 </layer>
2829 <layer id="209" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2830 <data element_type="i64" shape="4" offset="12201240" size="32" />
2831 <output>
2832 <port id="0" precision="I64">
2833 <dim>4</dim>
2834 </port>
2835 </output>
2836 </layer>
2837 <layer id="210" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2838 <data special_zero="true" />
2839 <input>
2840 <port id="0" precision="FP32">
2841 <dim>-1</dim>
2842 <dim>-1</dim>
2843 <dim>384</dim>
2844 </port>
2845 <port id="1" precision="I64">
2846 <dim>4</dim>
2847 </port>
2848 </input>
2849 <output>
2850 <port id="2" precision="FP32" names="177,x.19">
2851 <dim>-1</dim>
2852 <dim>-1</dim>
2853 <dim>12</dim>
2854 <dim>32</dim>
2855 </port>
2856 </output>
2857 </layer>
2858 <layer id="211" name="Constant_84704" type="Const" version="opset1">
2859 <data element_type="i64" shape="4" offset="12201272" size="32" />
2860 <output>
2861 <port id="0" precision="I64" names="178">
2862 <dim>4</dim>
2863 </port>
2864 </output>
2865 </layer>
2866 <layer id="212" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2867 <input>
2868 <port id="0" precision="FP32">
2869 <dim>-1</dim>
2870 <dim>-1</dim>
2871 <dim>12</dim>
2872 <dim>32</dim>
2873 </port>
2874 <port id="1" precision="I64">
2875 <dim>4</dim>
2876 </port>
2877 </input>
2878 <output>
2879 <port id="2" precision="FP32" names="179">
2880 <dim>-1</dim>
2881 <dim>12</dim>
2882 <dim>-1</dim>
2883 <dim>32</dim>
2884 </port>
2885 </output>
2886 </layer>
2887 <layer id="213" name="Constant_299970" type="Const" version="opset1">
2888 <data element_type="i8" shape="384, 384" offset="14165976" size="147456" />
2889 <output>
2890 <port id="0" precision="I8">
2891 <dim>384</dim>
2892 <dim>384</dim>
2893 </port>
2894 </output>
2895 </layer>
2896 <layer id="214" name="Convert_299971" type="Convert" version="opset1">
2897 <data destination_type="f32" />
2898 <input>
2899 <port id="0" precision="I8">
2900 <dim>384</dim>
2901 <dim>384</dim>
2902 </port>
2903 </input>
2904 <output>
2905 <port id="1" precision="FP32">
2906 <dim>384</dim>
2907 <dim>384</dim>
2908 </port>
2909 </output>
2910 </layer>
2911 <layer id="215" name="Constant_299972" type="Const" version="opset1">
2912 <data element_type="f32" shape="384, 1" offset="14313432" size="1536" />
2913 <output>
2914 <port id="0" precision="FP32">
2915 <dim>384</dim>
2916 <dim>1</dim>
2917 </port>
2918 </output>
2919 </layer>
2920 <layer id="216" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2921 <data auto_broadcast="numpy" />
2922 <input>
2923 <port id="0" precision="FP32">
2924 <dim>384</dim>
2925 <dim>384</dim>
2926 </port>
2927 <port id="1" precision="FP32">
2928 <dim>384</dim>
2929 <dim>1</dim>
2930 </port>
2931 </input>
2932 <output>
2933 <port id="2" precision="FP32">
2934 <dim>384</dim>
2935 <dim>384</dim>
2936 </port>
2937 </output>
2938 </layer>
2939 <layer id="217" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
2940 <data transpose_a="false" transpose_b="true" />
2941 <input>
2942 <port id="0" precision="FP32">
2943 <dim>-1</dim>
2944 <dim>-1</dim>
2945 <dim>384</dim>
2946 </port>
2947 <port id="1" precision="FP32">
2948 <dim>384</dim>
2949 <dim>384</dim>
2950 </port>
2951 </input>
2952 <output>
2953 <port id="2" precision="FP32">
2954 <dim>-1</dim>
2955 <dim>-1</dim>
2956 <dim>384</dim>
2957 </port>
2958 </output>
2959 </layer>
2960 <layer id="218" name="Constant_90317" type="Const" version="opset1">
2961 <data element_type="f32" shape="1, 1, 384" offset="14314968" size="1536" />
2962 <output>
2963 <port id="0" precision="FP32">
2964 <dim>1</dim>
2965 <dim>1</dim>
2966 <dim>384</dim>
2967 </port>
2968 </output>
2969 </layer>
2970 <layer id="219" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
2971 <data auto_broadcast="numpy" />
2972 <input>
2973 <port id="0" precision="FP32">
2974 <dim>-1</dim>
2975 <dim>-1</dim>
2976 <dim>384</dim>
2977 </port>
2978 <port id="1" precision="FP32">
2979 <dim>1</dim>
2980 <dim>1</dim>
2981 <dim>384</dim>
2982 </port>
2983 </input>
2984 <output>
2985 <port id="2" precision="FP32" names="182,x.21">
2986 <dim>-1</dim>
2987 <dim>-1</dim>
2988 <dim>384</dim>
2989 </port>
2990 </output>
2991 </layer>
2992 <layer id="220" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
2993 <data element_type="i64" shape="4" offset="12201240" size="32" />
2994 <output>
2995 <port id="0" precision="I64">
2996 <dim>4</dim>
2997 </port>
2998 </output>
2999 </layer>
3000 <layer id="221" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3001 <data special_zero="true" />
3002 <input>
3003 <port id="0" precision="FP32">
3004 <dim>-1</dim>
3005 <dim>-1</dim>
3006 <dim>384</dim>
3007 </port>
3008 <port id="1" precision="I64">
3009 <dim>4</dim>
3010 </port>
3011 </input>
3012 <output>
3013 <port id="2" precision="FP32" names="186,x.23">
3014 <dim>-1</dim>
3015 <dim>-1</dim>
3016 <dim>12</dim>
3017 <dim>32</dim>
3018 </port>
3019 </output>
3020 </layer>
3021 <layer id="222" name="Constant_84727" type="Const" version="opset1">
3022 <data element_type="i64" shape="4" offset="12201272" size="32" />
3023 <output>
3024 <port id="0" precision="I64" names="187">
3025 <dim>4</dim>
3026 </port>
3027 </output>
3028 </layer>
3029 <layer id="223" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3030 <input>
3031 <port id="0" precision="FP32">
3032 <dim>-1</dim>
3033 <dim>-1</dim>
3034 <dim>12</dim>
3035 <dim>32</dim>
3036 </port>
3037 <port id="1" precision="I64">
3038 <dim>4</dim>
3039 </port>
3040 </input>
3041 <output>
3042 <port id="2" precision="FP32" names="188">
3043 <dim>-1</dim>
3044 <dim>12</dim>
3045 <dim>-1</dim>
3046 <dim>32</dim>
3047 </port>
3048 </output>
3049 </layer>
3050 <layer id="224" name="__module.bert.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3051 <data causal="false" />
3052 <input>
3053 <port id="0" precision="FP32">
3054 <dim>-1</dim>
3055 <dim>12</dim>
3056 <dim>-1</dim>
3057 <dim>32</dim>
3058 </port>
3059 <port id="1" precision="FP32">
3060 <dim>-1</dim>
3061 <dim>12</dim>
3062 <dim>-1</dim>
3063 <dim>32</dim>
3064 </port>
3065 <port id="2" precision="FP32">
3066 <dim>-1</dim>
3067 <dim>12</dim>
3068 <dim>-1</dim>
3069 <dim>32</dim>
3070 </port>
3071 <port id="3" precision="FP32">
3072 <dim>-1</dim>
3073 <dim>1</dim>
3074 <dim>-1</dim>
3075 <dim>-1</dim>
3076 </port>
3077 </input>
3078 <output>
3079 <port id="4" precision="FP32" names="189,attn_output.5">
3080 <dim>-1</dim>
3081 <dim>12</dim>
3082 <dim>-1</dim>
3083 <dim>32</dim>
3084 </port>
3085 </output>
3086 </layer>
3087 <layer id="225" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3088 <data element_type="i32" shape="4" offset="12502400" size="16" />
3089 <output>
3090 <port id="0" precision="I32">
3091 <dim>4</dim>
3092 </port>
3093 </output>
3094 </layer>
3095 <layer id="226" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3096 <input>
3097 <port id="0" precision="FP32">
3098 <dim>-1</dim>
3099 <dim>12</dim>
3100 <dim>-1</dim>
3101 <dim>32</dim>
3102 </port>
3103 <port id="1" precision="I32">
3104 <dim>4</dim>
3105 </port>
3106 </input>
3107 <output>
3108 <port id="2" precision="FP32" names="190,attn_output.7">
3109 <dim>-1</dim>
3110 <dim>-1</dim>
3111 <dim>12</dim>
3112 <dim>32</dim>
3113 </port>
3114 </output>
3115 </layer>
3116 <layer id="227" name="Constant_90444" type="Const" version="opset1">
3117 <data element_type="i64" shape="3" offset="12502416" size="24" />
3118 <output>
3119 <port id="0" precision="I64">
3120 <dim>3</dim>
3121 </port>
3122 </output>
3123 </layer>
3124 <layer id="228" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3125 <data special_zero="true" />
3126 <input>
3127 <port id="0" precision="FP32">
3128 <dim>-1</dim>
3129 <dim>-1</dim>
3130 <dim>12</dim>
3131 <dim>32</dim>
3132 </port>
3133 <port id="1" precision="I64">
3134 <dim>3</dim>
3135 </port>
3136 </input>
3137 <output>
3138 <port id="2" precision="FP32" names="192">
3139 <dim>-1</dim>
3140 <dim>-1</dim>
3141 <dim>384</dim>
3142 </port>
3143 </output>
3144 </layer>
3145 <layer id="229" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3146 <data element_type="f32" shape="1, 1, 384" offset="14316504" size="1536" />
3147 <output>
3148 <port id="0" precision="FP32">
3149 <dim>1</dim>
3150 <dim>1</dim>
3151 <dim>384</dim>
3152 </port>
3153 </output>
3154 </layer>
3155 <layer id="230" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3156 <data auto_broadcast="numpy" />
3157 <input>
3158 <port id="0" precision="FP32">
3159 <dim>-1</dim>
3160 <dim>-1</dim>
3161 <dim>384</dim>
3162 </port>
3163 <port id="1" precision="FP32">
3164 <dim>1</dim>
3165 <dim>1</dim>
3166 <dim>384</dim>
3167 </port>
3168 </input>
3169 <output>
3170 <port id="2" precision="FP32">
3171 <dim>-1</dim>
3172 <dim>-1</dim>
3173 <dim>384</dim>
3174 </port>
3175 </output>
3176 </layer>
3177 <layer id="231" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3178 <data element_type="f32" shape="" offset="14318040" size="4" />
3179 <output>
3180 <port id="0" precision="FP32" />
3181 </output>
3182 </layer>
3183 <layer id="232" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3184 <data element_type="f32" shape="" offset="14318044" size="4" />
3185 <output>
3186 <port id="0" precision="FP32" />
3187 </output>
3188 </layer>
3189 <layer id="233" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3190 <data element_type="f32" shape="" offset="14318040" size="4" />
3191 <output>
3192 <port id="0" precision="FP32" />
3193 </output>
3194 </layer>
3195 <layer id="234" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3196 <data element_type="f32" shape="" offset="14318044" size="4" />
3197 <output>
3198 <port id="0" precision="FP32" />
3199 </output>
3200 </layer>
3201 <layer id="235" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3202 <data levels="256" auto_broadcast="numpy" />
3203 <input>
3204 <port id="0" precision="FP32">
3205 <dim>-1</dim>
3206 <dim>-1</dim>
3207 <dim>384</dim>
3208 </port>
3209 <port id="1" precision="FP32" />
3210 <port id="2" precision="FP32" />
3211 <port id="3" precision="FP32" />
3212 <port id="4" precision="FP32" />
3213 </input>
3214 <output>
3215 <port id="5" precision="FP32">
3216 <dim>-1</dim>
3217 <dim>-1</dim>
3218 <dim>384</dim>
3219 </port>
3220 </output>
3221 </layer>
3222 <layer id="236" name="Constant_299974" type="Const" version="opset1">
3223 <data element_type="i8" shape="384, 384" offset="14318048" size="147456" />
3224 <output>
3225 <port id="0" precision="I8">
3226 <dim>384</dim>
3227 <dim>384</dim>
3228 </port>
3229 </output>
3230 </layer>
3231 <layer id="237" name="Convert_299975" type="Convert" version="opset1">
3232 <data destination_type="f32" />
3233 <input>
3234 <port id="0" precision="I8">
3235 <dim>384</dim>
3236 <dim>384</dim>
3237 </port>
3238 </input>
3239 <output>
3240 <port id="1" precision="FP32">
3241 <dim>384</dim>
3242 <dim>384</dim>
3243 </port>
3244 </output>
3245 </layer>
3246 <layer id="238" name="Constant_299976" type="Const" version="opset1">
3247 <data element_type="f32" shape="384, 1" offset="14465504" size="1536" />
3248 <output>
3249 <port id="0" precision="FP32">
3250 <dim>384</dim>
3251 <dim>1</dim>
3252 </port>
3253 </output>
3254 </layer>
3255 <layer id="239" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3256 <data auto_broadcast="numpy" />
3257 <input>
3258 <port id="0" precision="FP32">
3259 <dim>384</dim>
3260 <dim>384</dim>
3261 </port>
3262 <port id="1" precision="FP32">
3263 <dim>384</dim>
3264 <dim>1</dim>
3265 </port>
3266 </input>
3267 <output>
3268 <port id="2" precision="FP32">
3269 <dim>384</dim>
3270 <dim>384</dim>
3271 </port>
3272 </output>
3273 </layer>
3274 <layer id="240" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3275 <data transpose_a="false" transpose_b="true" />
3276 <input>
3277 <port id="0" precision="FP32">
3278 <dim>-1</dim>
3279 <dim>-1</dim>
3280 <dim>384</dim>
3281 </port>
3282 <port id="1" precision="FP32">
3283 <dim>384</dim>
3284 <dim>384</dim>
3285 </port>
3286 </input>
3287 <output>
3288 <port id="2" precision="FP32">
3289 <dim>-1</dim>
3290 <dim>-1</dim>
3291 <dim>384</dim>
3292 </port>
3293 </output>
3294 </layer>
3295 <layer id="241" name="Constant_90318" type="Const" version="opset1">
3296 <data element_type="f32" shape="1, 1, 384" offset="14467040" size="1536" />
3297 <output>
3298 <port id="0" precision="FP32">
3299 <dim>1</dim>
3300 <dim>1</dim>
3301 <dim>384</dim>
3302 </port>
3303 </output>
3304 </layer>
3305 <layer id="242" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3306 <data auto_broadcast="numpy" />
3307 <input>
3308 <port id="0" precision="FP32">
3309 <dim>-1</dim>
3310 <dim>-1</dim>
3311 <dim>384</dim>
3312 </port>
3313 <port id="1" precision="FP32">
3314 <dim>1</dim>
3315 <dim>1</dim>
3316 <dim>384</dim>
3317 </port>
3318 </input>
3319 <output>
3320 <port id="2" precision="FP32" names="197,input.7">
3321 <dim>-1</dim>
3322 <dim>-1</dim>
3323 <dim>384</dim>
3324 </port>
3325 </output>
3326 </layer>
3327 <layer id="243" name="__module.bert.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
3328 <data auto_broadcast="numpy" />
3329 <input>
3330 <port id="0" precision="FP32">
3331 <dim>-1</dim>
3332 <dim>-1</dim>
3333 <dim>384</dim>
3334 </port>
3335 <port id="1" precision="FP32">
3336 <dim>-1</dim>
3337 <dim>-1</dim>
3338 <dim>384</dim>
3339 </port>
3340 </input>
3341 <output>
3342 <port id="2" precision="FP32" names="199">
3343 <dim>-1</dim>
3344 <dim>-1</dim>
3345 <dim>384</dim>
3346 </port>
3347 </output>
3348 </layer>
3349 <layer id="244" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3350 <data element_type="i32" shape="1" offset="12046084" size="4" />
3351 <output>
3352 <port id="0" precision="I32">
3353 <dim>1</dim>
3354 </port>
3355 </output>
3356 </layer>
3357 <layer id="245" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3358 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3359 <input>
3360 <port id="0" precision="FP32">
3361 <dim>-1</dim>
3362 <dim>-1</dim>
3363 <dim>384</dim>
3364 </port>
3365 <port id="1" precision="I32">
3366 <dim>1</dim>
3367 </port>
3368 </input>
3369 <output>
3370 <port id="2" precision="FP32">
3371 <dim>-1</dim>
3372 <dim>-1</dim>
3373 <dim>384</dim>
3374 </port>
3375 </output>
3376 </layer>
3377 <layer id="246" name="Constant_90319" type="Const" version="opset1">
3378 <data element_type="f32" shape="1, 1, 384" offset="14468576" size="1536" />
3379 <output>
3380 <port id="0" precision="FP32">
3381 <dim>1</dim>
3382 <dim>1</dim>
3383 <dim>384</dim>
3384 </port>
3385 </output>
3386 </layer>
3387 <layer id="247" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3388 <data auto_broadcast="numpy" />
3389 <input>
3390 <port id="0" precision="FP32">
3391 <dim>-1</dim>
3392 <dim>-1</dim>
3393 <dim>384</dim>
3394 </port>
3395 <port id="1" precision="FP32">
3396 <dim>1</dim>
3397 <dim>1</dim>
3398 <dim>384</dim>
3399 </port>
3400 </input>
3401 <output>
3402 <port id="2" precision="FP32">
3403 <dim>-1</dim>
3404 <dim>-1</dim>
3405 <dim>384</dim>
3406 </port>
3407 </output>
3408 </layer>
3409 <layer id="248" name="Constant_90320" type="Const" version="opset1">
3410 <data element_type="f32" shape="1, 1, 384" offset="14470112" size="1536" />
3411 <output>
3412 <port id="0" precision="FP32">
3413 <dim>1</dim>
3414 <dim>1</dim>
3415 <dim>384</dim>
3416 </port>
3417 </output>
3418 </layer>
3419 <layer id="249" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3420 <data auto_broadcast="numpy" />
3421 <input>
3422 <port id="0" precision="FP32">
3423 <dim>-1</dim>
3424 <dim>-1</dim>
3425 <dim>384</dim>
3426 </port>
3427 <port id="1" precision="FP32">
3428 <dim>1</dim>
3429 <dim>1</dim>
3430 <dim>384</dim>
3431 </port>
3432 </input>
3433 <output>
3434 <port id="2" precision="FP32" names="203,input_tensor.3">
3435 <dim>-1</dim>
3436 <dim>-1</dim>
3437 <dim>384</dim>
3438 </port>
3439 </output>
3440 </layer>
3441 <layer id="250" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3442 <data element_type="f32" shape="1, 1, 384" offset="14471648" size="1536" />
3443 <output>
3444 <port id="0" precision="FP32">
3445 <dim>1</dim>
3446 <dim>1</dim>
3447 <dim>384</dim>
3448 </port>
3449 </output>
3450 </layer>
3451 <layer id="251" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3452 <data auto_broadcast="numpy" />
3453 <input>
3454 <port id="0" precision="FP32">
3455 <dim>-1</dim>
3456 <dim>-1</dim>
3457 <dim>384</dim>
3458 </port>
3459 <port id="1" precision="FP32">
3460 <dim>1</dim>
3461 <dim>1</dim>
3462 <dim>384</dim>
3463 </port>
3464 </input>
3465 <output>
3466 <port id="2" precision="FP32">
3467 <dim>-1</dim>
3468 <dim>-1</dim>
3469 <dim>384</dim>
3470 </port>
3471 </output>
3472 </layer>
3473 <layer id="252" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3474 <data element_type="f32" shape="" offset="14473184" size="4" />
3475 <output>
3476 <port id="0" precision="FP32" />
3477 </output>
3478 </layer>
3479 <layer id="253" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3480 <data element_type="f32" shape="" offset="14473188" size="4" />
3481 <output>
3482 <port id="0" precision="FP32" />
3483 </output>
3484 </layer>
3485 <layer id="254" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3486 <data element_type="f32" shape="" offset="14473184" size="4" />
3487 <output>
3488 <port id="0" precision="FP32" />
3489 </output>
3490 </layer>
3491 <layer id="255" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3492 <data element_type="f32" shape="" offset="14473188" size="4" />
3493 <output>
3494 <port id="0" precision="FP32" />
3495 </output>
3496 </layer>
3497 <layer id="256" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3498 <data levels="256" auto_broadcast="numpy" />
3499 <input>
3500 <port id="0" precision="FP32">
3501 <dim>-1</dim>
3502 <dim>-1</dim>
3503 <dim>384</dim>
3504 </port>
3505 <port id="1" precision="FP32" />
3506 <port id="2" precision="FP32" />
3507 <port id="3" precision="FP32" />
3508 <port id="4" precision="FP32" />
3509 </input>
3510 <output>
3511 <port id="5" precision="FP32">
3512 <dim>-1</dim>
3513 <dim>-1</dim>
3514 <dim>384</dim>
3515 </port>
3516 </output>
3517 </layer>
3518 <layer id="257" name="Constant_299978" type="Const" version="opset1">
3519 <data element_type="i8" shape="1536, 384" offset="14473192" size="589824" />
3520 <output>
3521 <port id="0" precision="I8">
3522 <dim>1536</dim>
3523 <dim>384</dim>
3524 </port>
3525 </output>
3526 </layer>
3527 <layer id="258" name="Convert_299979" type="Convert" version="opset1">
3528 <data destination_type="f32" />
3529 <input>
3530 <port id="0" precision="I8">
3531 <dim>1536</dim>
3532 <dim>384</dim>
3533 </port>
3534 </input>
3535 <output>
3536 <port id="1" precision="FP32">
3537 <dim>1536</dim>
3538 <dim>384</dim>
3539 </port>
3540 </output>
3541 </layer>
3542 <layer id="259" name="Constant_299980" type="Const" version="opset1">
3543 <data element_type="f32" shape="1536, 1" offset="15063016" size="6144" />
3544 <output>
3545 <port id="0" precision="FP32">
3546 <dim>1536</dim>
3547 <dim>1</dim>
3548 </port>
3549 </output>
3550 </layer>
3551 <layer id="260" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3552 <data auto_broadcast="numpy" />
3553 <input>
3554 <port id="0" precision="FP32">
3555 <dim>1536</dim>
3556 <dim>384</dim>
3557 </port>
3558 <port id="1" precision="FP32">
3559 <dim>1536</dim>
3560 <dim>1</dim>
3561 </port>
3562 </input>
3563 <output>
3564 <port id="2" precision="FP32">
3565 <dim>1536</dim>
3566 <dim>384</dim>
3567 </port>
3568 </output>
3569 </layer>
3570 <layer id="261" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3571 <data transpose_a="false" transpose_b="true" />
3572 <input>
3573 <port id="0" precision="FP32">
3574 <dim>-1</dim>
3575 <dim>-1</dim>
3576 <dim>384</dim>
3577 </port>
3578 <port id="1" precision="FP32">
3579 <dim>1536</dim>
3580 <dim>384</dim>
3581 </port>
3582 </input>
3583 <output>
3584 <port id="2" precision="FP32">
3585 <dim>-1</dim>
3586 <dim>-1</dim>
3587 <dim>1536</dim>
3588 </port>
3589 </output>
3590 </layer>
3591 <layer id="262" name="Constant_90321" type="Const" version="opset1">
3592 <data element_type="f32" shape="1, 1, 1536" offset="15069160" size="6144" />
3593 <output>
3594 <port id="0" precision="FP32">
3595 <dim>1</dim>
3596 <dim>1</dim>
3597 <dim>1536</dim>
3598 </port>
3599 </output>
3600 </layer>
3601 <layer id="263" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3602 <data auto_broadcast="numpy" />
3603 <input>
3604 <port id="0" precision="FP32">
3605 <dim>-1</dim>
3606 <dim>-1</dim>
3607 <dim>1536</dim>
3608 </port>
3609 <port id="1" precision="FP32">
3610 <dim>1</dim>
3611 <dim>1</dim>
3612 <dim>1536</dim>
3613 </port>
3614 </input>
3615 <output>
3616 <port id="2" precision="FP32" names="207">
3617 <dim>-1</dim>
3618 <dim>-1</dim>
3619 <dim>1536</dim>
3620 </port>
3621 </output>
3622 </layer>
3623 <layer id="264" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3624 <data approximation_mode="ERF" />
3625 <input>
3626 <port id="0" precision="FP32">
3627 <dim>-1</dim>
3628 <dim>-1</dim>
3629 <dim>1536</dim>
3630 </port>
3631 </input>
3632 <output>
3633 <port id="1" precision="FP32" names="208">
3634 <dim>-1</dim>
3635 <dim>-1</dim>
3636 <dim>1536</dim>
3637 </port>
3638 </output>
3639 </layer>
3640 <layer id="265" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3641 <data element_type="f32" shape="1, 1, 1536" offset="15075304" size="6144" />
3642 <output>
3643 <port id="0" precision="FP32">
3644 <dim>1</dim>
3645 <dim>1</dim>
3646 <dim>1536</dim>
3647 </port>
3648 </output>
3649 </layer>
3650 <layer id="266" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3651 <data auto_broadcast="numpy" />
3652 <input>
3653 <port id="0" precision="FP32">
3654 <dim>-1</dim>
3655 <dim>-1</dim>
3656 <dim>1536</dim>
3657 </port>
3658 <port id="1" precision="FP32">
3659 <dim>1</dim>
3660 <dim>1</dim>
3661 <dim>1536</dim>
3662 </port>
3663 </input>
3664 <output>
3665 <port id="2" precision="FP32">
3666 <dim>-1</dim>
3667 <dim>-1</dim>
3668 <dim>1536</dim>
3669 </port>
3670 </output>
3671 </layer>
3672 <layer id="267" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3673 <data element_type="f32" shape="" offset="15081448" size="4" />
3674 <output>
3675 <port id="0" precision="FP32" />
3676 </output>
3677 </layer>
3678 <layer id="268" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3679 <data element_type="f32" shape="" offset="15081452" size="4" />
3680 <output>
3681 <port id="0" precision="FP32" />
3682 </output>
3683 </layer>
3684 <layer id="269" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3685 <data element_type="f32" shape="" offset="15081448" size="4" />
3686 <output>
3687 <port id="0" precision="FP32" />
3688 </output>
3689 </layer>
3690 <layer id="270" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3691 <data element_type="f32" shape="" offset="15081452" size="4" />
3692 <output>
3693 <port id="0" precision="FP32" />
3694 </output>
3695 </layer>
3696 <layer id="271" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3697 <data levels="256" auto_broadcast="numpy" />
3698 <input>
3699 <port id="0" precision="FP32">
3700 <dim>-1</dim>
3701 <dim>-1</dim>
3702 <dim>1536</dim>
3703 </port>
3704 <port id="1" precision="FP32" />
3705 <port id="2" precision="FP32" />
3706 <port id="3" precision="FP32" />
3707 <port id="4" precision="FP32" />
3708 </input>
3709 <output>
3710 <port id="5" precision="FP32">
3711 <dim>-1</dim>
3712 <dim>-1</dim>
3713 <dim>1536</dim>
3714 </port>
3715 </output>
3716 </layer>
3717 <layer id="272" name="Constant_299982" type="Const" version="opset1">
3718 <data element_type="i8" shape="384, 1536" offset="15081456" size="589824" />
3719 <output>
3720 <port id="0" precision="I8">
3721 <dim>384</dim>
3722 <dim>1536</dim>
3723 </port>
3724 </output>
3725 </layer>
3726 <layer id="273" name="Convert_299983" type="Convert" version="opset1">
3727 <data destination_type="f32" />
3728 <input>
3729 <port id="0" precision="I8">
3730 <dim>384</dim>
3731 <dim>1536</dim>
3732 </port>
3733 </input>
3734 <output>
3735 <port id="1" precision="FP32">
3736 <dim>384</dim>
3737 <dim>1536</dim>
3738 </port>
3739 </output>
3740 </layer>
3741 <layer id="274" name="Constant_299984" type="Const" version="opset1">
3742 <data element_type="f32" shape="384, 1" offset="15671280" size="1536" />
3743 <output>
3744 <port id="0" precision="FP32">
3745 <dim>384</dim>
3746 <dim>1</dim>
3747 </port>
3748 </output>
3749 </layer>
3750 <layer id="275" name="__module.bert.encoder.layer.1.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3751 <data auto_broadcast="numpy" />
3752 <input>
3753 <port id="0" precision="FP32">
3754 <dim>384</dim>
3755 <dim>1536</dim>
3756 </port>
3757 <port id="1" precision="FP32">
3758 <dim>384</dim>
3759 <dim>1</dim>
3760 </port>
3761 </input>
3762 <output>
3763 <port id="2" precision="FP32">
3764 <dim>384</dim>
3765 <dim>1536</dim>
3766 </port>
3767 </output>
3768 </layer>
3769 <layer id="276" name="__module.bert.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3770 <data transpose_a="false" transpose_b="true" />
3771 <input>
3772 <port id="0" precision="FP32">
3773 <dim>-1</dim>
3774 <dim>-1</dim>
3775 <dim>1536</dim>
3776 </port>
3777 <port id="1" precision="FP32">
3778 <dim>384</dim>
3779 <dim>1536</dim>
3780 </port>
3781 </input>
3782 <output>
3783 <port id="2" precision="FP32">
3784 <dim>-1</dim>
3785 <dim>-1</dim>
3786 <dim>384</dim>
3787 </port>
3788 </output>
3789 </layer>
3790 <layer id="277" name="Constant_90322" type="Const" version="opset1">
3791 <data element_type="f32" shape="1, 1, 384" offset="15672816" size="1536" />
3792 <output>
3793 <port id="0" precision="FP32">
3794 <dim>1</dim>
3795 <dim>1</dim>
3796 <dim>384</dim>
3797 </port>
3798 </output>
3799 </layer>
3800 <layer id="278" name="__module.bert.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
3801 <data auto_broadcast="numpy" />
3802 <input>
3803 <port id="0" precision="FP32">
3804 <dim>-1</dim>
3805 <dim>-1</dim>
3806 <dim>384</dim>
3807 </port>
3808 <port id="1" precision="FP32">
3809 <dim>1</dim>
3810 <dim>1</dim>
3811 <dim>384</dim>
3812 </port>
3813 </input>
3814 <output>
3815 <port id="2" precision="FP32" names="213,input.9">
3816 <dim>-1</dim>
3817 <dim>-1</dim>
3818 <dim>384</dim>
3819 </port>
3820 </output>
3821 </layer>
3822 <layer id="279" name="__module.bert.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
3823 <data auto_broadcast="numpy" />
3824 <input>
3825 <port id="0" precision="FP32">
3826 <dim>-1</dim>
3827 <dim>-1</dim>
3828 <dim>384</dim>
3829 </port>
3830 <port id="1" precision="FP32">
3831 <dim>-1</dim>
3832 <dim>-1</dim>
3833 <dim>384</dim>
3834 </port>
3835 </input>
3836 <output>
3837 <port id="2" precision="FP32" names="215">
3838 <dim>-1</dim>
3839 <dim>-1</dim>
3840 <dim>384</dim>
3841 </port>
3842 </output>
3843 </layer>
3844 <layer id="280" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3845 <data element_type="i32" shape="1" offset="12046084" size="4" />
3846 <output>
3847 <port id="0" precision="I32">
3848 <dim>1</dim>
3849 </port>
3850 </output>
3851 </layer>
3852 <layer id="281" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3853 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3854 <input>
3855 <port id="0" precision="FP32">
3856 <dim>-1</dim>
3857 <dim>-1</dim>
3858 <dim>384</dim>
3859 </port>
3860 <port id="1" precision="I32">
3861 <dim>1</dim>
3862 </port>
3863 </input>
3864 <output>
3865 <port id="2" precision="FP32">
3866 <dim>-1</dim>
3867 <dim>-1</dim>
3868 <dim>384</dim>
3869 </port>
3870 </output>
3871 </layer>
3872 <layer id="282" name="Constant_90323" type="Const" version="opset1">
3873 <data element_type="f32" shape="1, 1, 384" offset="15674352" size="1536" />
3874 <output>
3875 <port id="0" precision="FP32">
3876 <dim>1</dim>
3877 <dim>1</dim>
3878 <dim>384</dim>
3879 </port>
3880 </output>
3881 </layer>
3882 <layer id="283" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3883 <data auto_broadcast="numpy" />
3884 <input>
3885 <port id="0" precision="FP32">
3886 <dim>-1</dim>
3887 <dim>-1</dim>
3888 <dim>384</dim>
3889 </port>
3890 <port id="1" precision="FP32">
3891 <dim>1</dim>
3892 <dim>1</dim>
3893 <dim>384</dim>
3894 </port>
3895 </input>
3896 <output>
3897 <port id="2" precision="FP32">
3898 <dim>-1</dim>
3899 <dim>-1</dim>
3900 <dim>384</dim>
3901 </port>
3902 </output>
3903 </layer>
3904 <layer id="284" name="Constant_90324" type="Const" version="opset1">
3905 <data element_type="f32" shape="1, 1, 384" offset="15675888" size="1536" />
3906 <output>
3907 <port id="0" precision="FP32">
3908 <dim>1</dim>
3909 <dim>1</dim>
3910 <dim>384</dim>
3911 </port>
3912 </output>
3913 </layer>
3914 <layer id="285" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3915 <data auto_broadcast="numpy" />
3916 <input>
3917 <port id="0" precision="FP32">
3918 <dim>-1</dim>
3919 <dim>-1</dim>
3920 <dim>384</dim>
3921 </port>
3922 <port id="1" precision="FP32">
3923 <dim>1</dim>
3924 <dim>1</dim>
3925 <dim>384</dim>
3926 </port>
3927 </input>
3928 <output>
3929 <port id="2" precision="FP32" names="219,hidden_states.13">
3930 <dim>-1</dim>
3931 <dim>-1</dim>
3932 <dim>384</dim>
3933 </port>
3934 </output>
3935 </layer>
3936 <layer id="286" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3937 <data element_type="f32" shape="1, 1, 384" offset="15677424" size="1536" />
3938 <output>
3939 <port id="0" precision="FP32">
3940 <dim>1</dim>
3941 <dim>1</dim>
3942 <dim>384</dim>
3943 </port>
3944 </output>
3945 </layer>
3946 <layer id="287" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3947 <data auto_broadcast="numpy" />
3948 <input>
3949 <port id="0" precision="FP32">
3950 <dim>-1</dim>
3951 <dim>-1</dim>
3952 <dim>384</dim>
3953 </port>
3954 <port id="1" precision="FP32">
3955 <dim>1</dim>
3956 <dim>1</dim>
3957 <dim>384</dim>
3958 </port>
3959 </input>
3960 <output>
3961 <port id="2" precision="FP32">
3962 <dim>-1</dim>
3963 <dim>-1</dim>
3964 <dim>384</dim>
3965 </port>
3966 </output>
3967 </layer>
3968 <layer id="288" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3969 <data element_type="f32" shape="" offset="15678960" size="4" />
3970 <output>
3971 <port id="0" precision="FP32" />
3972 </output>
3973 </layer>
3974 <layer id="289" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3975 <data element_type="f32" shape="" offset="15678964" size="4" />
3976 <output>
3977 <port id="0" precision="FP32" />
3978 </output>
3979 </layer>
3980 <layer id="290" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3981 <data element_type="f32" shape="" offset="15678960" size="4" />
3982 <output>
3983 <port id="0" precision="FP32" />
3984 </output>
3985 </layer>
3986 <layer id="291" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3987 <data element_type="f32" shape="" offset="15678964" size="4" />
3988 <output>
3989 <port id="0" precision="FP32" />
3990 </output>
3991 </layer>
3992 <layer id="292" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3993 <data levels="256" auto_broadcast="numpy" />
3994 <input>
3995 <port id="0" precision="FP32">
3996 <dim>-1</dim>
3997 <dim>-1</dim>
3998 <dim>384</dim>
3999 </port>
4000 <port id="1" precision="FP32" />
4001 <port id="2" precision="FP32" />
4002 <port id="3" precision="FP32" />
4003 <port id="4" precision="FP32" />
4004 </input>
4005 <output>
4006 <port id="5" precision="FP32">
4007 <dim>-1</dim>
4008 <dim>-1</dim>
4009 <dim>384</dim>
4010 </port>
4011 </output>
4012 </layer>
4013 <layer id="293" name="Constant_299986" type="Const" version="opset1">
4014 <data element_type="i8" shape="384, 384" offset="15678968" size="147456" />
4015 <output>
4016 <port id="0" precision="I8">
4017 <dim>384</dim>
4018 <dim>384</dim>
4019 </port>
4020 </output>
4021 </layer>
4022 <layer id="294" name="Convert_299987" type="Convert" version="opset1">
4023 <data destination_type="f32" />
4024 <input>
4025 <port id="0" precision="I8">
4026 <dim>384</dim>
4027 <dim>384</dim>
4028 </port>
4029 </input>
4030 <output>
4031 <port id="1" precision="FP32">
4032 <dim>384</dim>
4033 <dim>384</dim>
4034 </port>
4035 </output>
4036 </layer>
4037 <layer id="295" name="Constant_299988" type="Const" version="opset1">
4038 <data element_type="f32" shape="384, 1" offset="15826424" size="1536" />
4039 <output>
4040 <port id="0" precision="FP32">
4041 <dim>384</dim>
4042 <dim>1</dim>
4043 </port>
4044 </output>
4045 </layer>
4046 <layer id="296" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4047 <data auto_broadcast="numpy" />
4048 <input>
4049 <port id="0" precision="FP32">
4050 <dim>384</dim>
4051 <dim>384</dim>
4052 </port>
4053 <port id="1" precision="FP32">
4054 <dim>384</dim>
4055 <dim>1</dim>
4056 </port>
4057 </input>
4058 <output>
4059 <port id="2" precision="FP32">
4060 <dim>384</dim>
4061 <dim>384</dim>
4062 </port>
4063 </output>
4064 </layer>
4065 <layer id="297" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
4066 <data transpose_a="false" transpose_b="true" />
4067 <input>
4068 <port id="0" precision="FP32">
4069 <dim>-1</dim>
4070 <dim>-1</dim>
4071 <dim>384</dim>
4072 </port>
4073 <port id="1" precision="FP32">
4074 <dim>384</dim>
4075 <dim>384</dim>
4076 </port>
4077 </input>
4078 <output>
4079 <port id="2" precision="FP32">
4080 <dim>-1</dim>
4081 <dim>-1</dim>
4082 <dim>384</dim>
4083 </port>
4084 </output>
4085 </layer>
4086 <layer id="298" name="Constant_90325" type="Const" version="opset1">
4087 <data element_type="f32" shape="1, 1, 384" offset="15827960" size="1536" />
4088 <output>
4089 <port id="0" precision="FP32">
4090 <dim>1</dim>
4091 <dim>1</dim>
4092 <dim>384</dim>
4093 </port>
4094 </output>
4095 </layer>
4096 <layer id="299" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
4097 <data auto_broadcast="numpy" />
4098 <input>
4099 <port id="0" precision="FP32">
4100 <dim>-1</dim>
4101 <dim>-1</dim>
4102 <dim>384</dim>
4103 </port>
4104 <port id="1" precision="FP32">
4105 <dim>1</dim>
4106 <dim>1</dim>
4107 <dim>384</dim>
4108 </port>
4109 </input>
4110 <output>
4111 <port id="2" precision="FP32" names="232,x.25">
4112 <dim>-1</dim>
4113 <dim>-1</dim>
4114 <dim>384</dim>
4115 </port>
4116 </output>
4117 </layer>
4118 <layer id="300" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4119 <data element_type="f32" shape="" offset="15829496" size="4" />
4120 <output>
4121 <port id="0" precision="FP32" />
4122 </output>
4123 </layer>
4124 <layer id="301" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4125 <data element_type="f32" shape="" offset="15829500" size="4" />
4126 <output>
4127 <port id="0" precision="FP32" />
4128 </output>
4129 </layer>
4130 <layer id="302" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4131 <data element_type="f32" shape="" offset="15829496" size="4" />
4132 <output>
4133 <port id="0" precision="FP32" />
4134 </output>
4135 </layer>
4136 <layer id="303" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4137 <data element_type="f32" shape="" offset="15829500" size="4" />
4138 <output>
4139 <port id="0" precision="FP32" />
4140 </output>
4141 </layer>
4142 <layer id="304" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4143 <data levels="256" auto_broadcast="numpy" />
4144 <input>
4145 <port id="0" precision="FP32">
4146 <dim>-1</dim>
4147 <dim>-1</dim>
4148 <dim>384</dim>
4149 </port>
4150 <port id="1" precision="FP32" />
4151 <port id="2" precision="FP32" />
4152 <port id="3" precision="FP32" />
4153 <port id="4" precision="FP32" />
4154 </input>
4155 <output>
4156 <port id="5" precision="FP32">
4157 <dim>-1</dim>
4158 <dim>-1</dim>
4159 <dim>384</dim>
4160 </port>
4161 </output>
4162 </layer>
4163 <layer id="305" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
4164 <data element_type="i64" shape="4" offset="12201240" size="32" />
4165 <output>
4166 <port id="0" precision="I64">
4167 <dim>4</dim>
4168 </port>
4169 </output>
4170 </layer>
4171 <layer id="306" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
4172 <data special_zero="true" />
4173 <input>
4174 <port id="0" precision="FP32">
4175 <dim>-1</dim>
4176 <dim>-1</dim>
4177 <dim>384</dim>
4178 </port>
4179 <port id="1" precision="I64">
4180 <dim>4</dim>
4181 </port>
4182 </input>
4183 <output>
4184 <port id="2" precision="FP32" names="236,x.27">
4185 <dim>-1</dim>
4186 <dim>-1</dim>
4187 <dim>12</dim>
4188 <dim>32</dim>
4189 </port>
4190 </output>
4191 </layer>
4192 <layer id="307" name="Constant_84904" type="Const" version="opset1">
4193 <data element_type="i64" shape="4" offset="12201272" size="32" />
4194 <output>
4195 <port id="0" precision="I64" names="237">
4196 <dim>4</dim>
4197 </port>
4198 </output>
4199 </layer>
4200 <layer id="308" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
4201 <input>
4202 <port id="0" precision="FP32">
4203 <dim>-1</dim>
4204 <dim>-1</dim>
4205 <dim>12</dim>
4206 <dim>32</dim>
4207 </port>
4208 <port id="1" precision="I64">
4209 <dim>4</dim>
4210 </port>
4211 </input>
4212 <output>
4213 <port id="2" precision="FP32" names="238">
4214 <dim>-1</dim>
4215 <dim>12</dim>
4216 <dim>-1</dim>
4217 <dim>32</dim>
4218 </port>
4219 </output>
4220 </layer>
4221 <layer id="309" name="Constant_299990" type="Const" version="opset1">
4222 <data element_type="i8" shape="384, 384" offset="15829504" size="147456" />
4223 <output>
4224 <port id="0" precision="I8">
4225 <dim>384</dim>
4226 <dim>384</dim>
4227 </port>
4228 </output>
4229 </layer>
4230 <layer id="310" name="Convert_299991" type="Convert" version="opset1">
4231 <data destination_type="f32" />
4232 <input>
4233 <port id="0" precision="I8">
4234 <dim>384</dim>
4235 <dim>384</dim>
4236 </port>
4237 </input>
4238 <output>
4239 <port id="1" precision="FP32">
4240 <dim>384</dim>
4241 <dim>384</dim>
4242 </port>
4243 </output>
4244 </layer>
4245 <layer id="311" name="Constant_299992" type="Const" version="opset1">
4246 <data element_type="f32" shape="384, 1" offset="15976960" size="1536" />
4247 <output>
4248 <port id="0" precision="FP32">
4249 <dim>384</dim>
4250 <dim>1</dim>
4251 </port>
4252 </output>
4253 </layer>
4254 <layer id="312" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4255 <data auto_broadcast="numpy" />
4256 <input>
4257 <port id="0" precision="FP32">
4258 <dim>384</dim>
4259 <dim>384</dim>
4260 </port>
4261 <port id="1" precision="FP32">
4262 <dim>384</dim>
4263 <dim>1</dim>
4264 </port>
4265 </input>
4266 <output>
4267 <port id="2" precision="FP32">
4268 <dim>384</dim>
4269 <dim>384</dim>
4270 </port>
4271 </output>
4272 </layer>
4273 <layer id="313" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
4274 <data transpose_a="false" transpose_b="true" />
4275 <input>
4276 <port id="0" precision="FP32">
4277 <dim>-1</dim>
4278 <dim>-1</dim>
4279 <dim>384</dim>
4280 </port>
4281 <port id="1" precision="FP32">
4282 <dim>384</dim>
4283 <dim>384</dim>
4284 </port>
4285 </input>
4286 <output>
4287 <port id="2" precision="FP32">
4288 <dim>-1</dim>
4289 <dim>-1</dim>
4290 <dim>384</dim>
4291 </port>
4292 </output>
4293 </layer>
4294 <layer id="314" name="Constant_90326" type="Const" version="opset1">
4295 <data element_type="f32" shape="1, 1, 384" offset="15978496" size="1536" />
4296 <output>
4297 <port id="0" precision="FP32">
4298 <dim>1</dim>
4299 <dim>1</dim>
4300 <dim>384</dim>
4301 </port>
4302 </output>
4303 </layer>
4304 <layer id="315" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
4305 <data auto_broadcast="numpy" />
4306 <input>
4307 <port id="0" precision="FP32">
4308 <dim>-1</dim>
4309 <dim>-1</dim>
4310 <dim>384</dim>
4311 </port>
4312 <port id="1" precision="FP32">
4313 <dim>1</dim>
4314 <dim>1</dim>
4315 <dim>384</dim>
4316 </port>
4317 </input>
4318 <output>
4319 <port id="2" precision="FP32" names="241,x.29">
4320 <dim>-1</dim>
4321 <dim>-1</dim>
4322 <dim>384</dim>
4323 </port>
4324 </output>
4325 </layer>
4326 <layer id="316" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4327 <data element_type="f32" shape="" offset="15980032" size="4" />
4328 <output>
4329 <port id="0" precision="FP32" />
4330 </output>
4331 </layer>
4332 <layer id="317" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4333 <data element_type="f32" shape="" offset="15980036" size="4" />
4334 <output>
4335 <port id="0" precision="FP32" />
4336 </output>
4337 </layer>
4338 <layer id="318" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4339 <data element_type="f32" shape="" offset="15980032" size="4" />
4340 <output>
4341 <port id="0" precision="FP32" />
4342 </output>
4343 </layer>
4344 <layer id="319" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4345 <data element_type="f32" shape="" offset="15980036" size="4" />
4346 <output>
4347 <port id="0" precision="FP32" />
4348 </output>
4349 </layer>
4350 <layer id="320" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4351 <data levels="256" auto_broadcast="numpy" />
4352 <input>
4353 <port id="0" precision="FP32">
4354 <dim>-1</dim>
4355 <dim>-1</dim>
4356 <dim>384</dim>
4357 </port>
4358 <port id="1" precision="FP32" />
4359 <port id="2" precision="FP32" />
4360 <port id="3" precision="FP32" />
4361 <port id="4" precision="FP32" />
4362 </input>
4363 <output>
4364 <port id="5" precision="FP32">
4365 <dim>-1</dim>
4366 <dim>-1</dim>
4367 <dim>384</dim>
4368 </port>
4369 </output>
4370 </layer>
4371 <layer id="321" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
4372 <data element_type="i64" shape="4" offset="12201240" size="32" />
4373 <output>
4374 <port id="0" precision="I64">
4375 <dim>4</dim>
4376 </port>
4377 </output>
4378 </layer>
4379 <layer id="322" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
4380 <data special_zero="true" />
4381 <input>
4382 <port id="0" precision="FP32">
4383 <dim>-1</dim>
4384 <dim>-1</dim>
4385 <dim>384</dim>
4386 </port>
4387 <port id="1" precision="I64">
4388 <dim>4</dim>
4389 </port>
4390 </input>
4391 <output>
4392 <port id="2" precision="FP32" names="245,x.31">
4393 <dim>-1</dim>
4394 <dim>-1</dim>
4395 <dim>12</dim>
4396 <dim>32</dim>
4397 </port>
4398 </output>
4399 </layer>
4400 <layer id="323" name="Constant_84927" type="Const" version="opset1">
4401 <data element_type="i64" shape="4" offset="12201272" size="32" />
4402 <output>
4403 <port id="0" precision="I64" names="246">
4404 <dim>4</dim>
4405 </port>
4406 </output>
4407 </layer>
4408 <layer id="324" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
4409 <input>
4410 <port id="0" precision="FP32">
4411 <dim>-1</dim>
4412 <dim>-1</dim>
4413 <dim>12</dim>
4414 <dim>32</dim>
4415 </port>
4416 <port id="1" precision="I64">
4417 <dim>4</dim>
4418 </port>
4419 </input>
4420 <output>
4421 <port id="2" precision="FP32" names="247">
4422 <dim>-1</dim>
4423 <dim>12</dim>
4424 <dim>-1</dim>
4425 <dim>32</dim>
4426 </port>
4427 </output>
4428 </layer>
4429 <layer id="325" name="Constant_299994" type="Const" version="opset1">
4430 <data element_type="i8" shape="384, 384" offset="15980040" size="147456" />
4431 <output>
4432 <port id="0" precision="I8">
4433 <dim>384</dim>
4434 <dim>384</dim>
4435 </port>
4436 </output>
4437 </layer>
4438 <layer id="326" name="Convert_299995" type="Convert" version="opset1">
4439 <data destination_type="f32" />
4440 <input>
4441 <port id="0" precision="I8">
4442 <dim>384</dim>
4443 <dim>384</dim>
4444 </port>
4445 </input>
4446 <output>
4447 <port id="1" precision="FP32">
4448 <dim>384</dim>
4449 <dim>384</dim>
4450 </port>
4451 </output>
4452 </layer>
4453 <layer id="327" name="Constant_299996" type="Const" version="opset1">
4454 <data element_type="f32" shape="384, 1" offset="16127496" size="1536" />
4455 <output>
4456 <port id="0" precision="FP32">
4457 <dim>384</dim>
4458 <dim>1</dim>
4459 </port>
4460 </output>
4461 </layer>
4462 <layer id="328" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4463 <data auto_broadcast="numpy" />
4464 <input>
4465 <port id="0" precision="FP32">
4466 <dim>384</dim>
4467 <dim>384</dim>
4468 </port>
4469 <port id="1" precision="FP32">
4470 <dim>384</dim>
4471 <dim>1</dim>
4472 </port>
4473 </input>
4474 <output>
4475 <port id="2" precision="FP32">
4476 <dim>384</dim>
4477 <dim>384</dim>
4478 </port>
4479 </output>
4480 </layer>
4481 <layer id="329" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
4482 <data transpose_a="false" transpose_b="true" />
4483 <input>
4484 <port id="0" precision="FP32">
4485 <dim>-1</dim>
4486 <dim>-1</dim>
4487 <dim>384</dim>
4488 </port>
4489 <port id="1" precision="FP32">
4490 <dim>384</dim>
4491 <dim>384</dim>
4492 </port>
4493 </input>
4494 <output>
4495 <port id="2" precision="FP32">
4496 <dim>-1</dim>
4497 <dim>-1</dim>
4498 <dim>384</dim>
4499 </port>
4500 </output>
4501 </layer>
4502 <layer id="330" name="Constant_90327" type="Const" version="opset1">
4503 <data element_type="f32" shape="1, 1, 384" offset="16129032" size="1536" />
4504 <output>
4505 <port id="0" precision="FP32">
4506 <dim>1</dim>
4507 <dim>1</dim>
4508 <dim>384</dim>
4509 </port>
4510 </output>
4511 </layer>
4512 <layer id="331" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
4513 <data auto_broadcast="numpy" />
4514 <input>
4515 <port id="0" precision="FP32">
4516 <dim>-1</dim>
4517 <dim>-1</dim>
4518 <dim>384</dim>
4519 </port>
4520 <port id="1" precision="FP32">
4521 <dim>1</dim>
4522 <dim>1</dim>
4523 <dim>384</dim>
4524 </port>
4525 </input>
4526 <output>
4527 <port id="2" precision="FP32" names="250,x.33">
4528 <dim>-1</dim>
4529 <dim>-1</dim>
4530 <dim>384</dim>
4531 </port>
4532 </output>
4533 </layer>
4534 <layer id="332" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
4535 <data element_type="i64" shape="4" offset="12201240" size="32" />
4536 <output>
4537 <port id="0" precision="I64">
4538 <dim>4</dim>
4539 </port>
4540 </output>
4541 </layer>
4542 <layer id="333" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
4543 <data special_zero="true" />
4544 <input>
4545 <port id="0" precision="FP32">
4546 <dim>-1</dim>
4547 <dim>-1</dim>
4548 <dim>384</dim>
4549 </port>
4550 <port id="1" precision="I64">
4551 <dim>4</dim>
4552 </port>
4553 </input>
4554 <output>
4555 <port id="2" precision="FP32" names="254,x.35">
4556 <dim>-1</dim>
4557 <dim>-1</dim>
4558 <dim>12</dim>
4559 <dim>32</dim>
4560 </port>
4561 </output>
4562 </layer>
4563 <layer id="334" name="Constant_84950" type="Const" version="opset1">
4564 <data element_type="i64" shape="4" offset="12201272" size="32" />
4565 <output>
4566 <port id="0" precision="I64" names="255">
4567 <dim>4</dim>
4568 </port>
4569 </output>
4570 </layer>
4571 <layer id="335" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
4572 <input>
4573 <port id="0" precision="FP32">
4574 <dim>-1</dim>
4575 <dim>-1</dim>
4576 <dim>12</dim>
4577 <dim>32</dim>
4578 </port>
4579 <port id="1" precision="I64">
4580 <dim>4</dim>
4581 </port>
4582 </input>
4583 <output>
4584 <port id="2" precision="FP32" names="256">
4585 <dim>-1</dim>
4586 <dim>12</dim>
4587 <dim>-1</dim>
4588 <dim>32</dim>
4589 </port>
4590 </output>
4591 </layer>
4592 <layer id="336" name="__module.bert.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
4593 <data causal="false" />
4594 <input>
4595 <port id="0" precision="FP32">
4596 <dim>-1</dim>
4597 <dim>12</dim>
4598 <dim>-1</dim>
4599 <dim>32</dim>
4600 </port>
4601 <port id="1" precision="FP32">
4602 <dim>-1</dim>
4603 <dim>12</dim>
4604 <dim>-1</dim>
4605 <dim>32</dim>
4606 </port>
4607 <port id="2" precision="FP32">
4608 <dim>-1</dim>
4609 <dim>12</dim>
4610 <dim>-1</dim>
4611 <dim>32</dim>
4612 </port>
4613 <port id="3" precision="FP32">
4614 <dim>-1</dim>
4615 <dim>1</dim>
4616 <dim>-1</dim>
4617 <dim>-1</dim>
4618 </port>
4619 </input>
4620 <output>
4621 <port id="4" precision="FP32" names="257,attn_output.9">
4622 <dim>-1</dim>
4623 <dim>12</dim>
4624 <dim>-1</dim>
4625 <dim>32</dim>
4626 </port>
4627 </output>
4628 </layer>
4629 <layer id="337" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
4630 <data element_type="i32" shape="4" offset="12502400" size="16" />
4631 <output>
4632 <port id="0" precision="I32">
4633 <dim>4</dim>
4634 </port>
4635 </output>
4636 </layer>
4637 <layer id="338" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
4638 <input>
4639 <port id="0" precision="FP32">
4640 <dim>-1</dim>
4641 <dim>12</dim>
4642 <dim>-1</dim>
4643 <dim>32</dim>
4644 </port>
4645 <port id="1" precision="I32">
4646 <dim>4</dim>
4647 </port>
4648 </input>
4649 <output>
4650 <port id="2" precision="FP32" names="258,attn_output.11">
4651 <dim>-1</dim>
4652 <dim>-1</dim>
4653 <dim>12</dim>
4654 <dim>32</dim>
4655 </port>
4656 </output>
4657 </layer>
4658 <layer id="339" name="Constant_90445" type="Const" version="opset1">
4659 <data element_type="i64" shape="3" offset="12502416" size="24" />
4660 <output>
4661 <port id="0" precision="I64">
4662 <dim>3</dim>
4663 </port>
4664 </output>
4665 </layer>
4666 <layer id="340" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
4667 <data special_zero="true" />
4668 <input>
4669 <port id="0" precision="FP32">
4670 <dim>-1</dim>
4671 <dim>-1</dim>
4672 <dim>12</dim>
4673 <dim>32</dim>
4674 </port>
4675 <port id="1" precision="I64">
4676 <dim>3</dim>
4677 </port>
4678 </input>
4679 <output>
4680 <port id="2" precision="FP32" names="260">
4681 <dim>-1</dim>
4682 <dim>-1</dim>
4683 <dim>384</dim>
4684 </port>
4685 </output>
4686 </layer>
4687 <layer id="341" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4688 <data element_type="f32" shape="1, 1, 384" offset="16130568" size="1536" />
4689 <output>
4690 <port id="0" precision="FP32">
4691 <dim>1</dim>
4692 <dim>1</dim>
4693 <dim>384</dim>
4694 </port>
4695 </output>
4696 </layer>
4697 <layer id="342" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4698 <data auto_broadcast="numpy" />
4699 <input>
4700 <port id="0" precision="FP32">
4701 <dim>-1</dim>
4702 <dim>-1</dim>
4703 <dim>384</dim>
4704 </port>
4705 <port id="1" precision="FP32">
4706 <dim>1</dim>
4707 <dim>1</dim>
4708 <dim>384</dim>
4709 </port>
4710 </input>
4711 <output>
4712 <port id="2" precision="FP32">
4713 <dim>-1</dim>
4714 <dim>-1</dim>
4715 <dim>384</dim>
4716 </port>
4717 </output>
4718 </layer>
4719 <layer id="343" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
4720 <data element_type="f32" shape="" offset="16132104" size="4" />
4721 <output>
4722 <port id="0" precision="FP32" />
4723 </output>
4724 </layer>
4725 <layer id="344" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
4726 <data element_type="f32" shape="" offset="16132108" size="4" />
4727 <output>
4728 <port id="0" precision="FP32" />
4729 </output>
4730 </layer>
4731 <layer id="345" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
4732 <data element_type="f32" shape="" offset="16132104" size="4" />
4733 <output>
4734 <port id="0" precision="FP32" />
4735 </output>
4736 </layer>
4737 <layer id="346" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
4738 <data element_type="f32" shape="" offset="16132108" size="4" />
4739 <output>
4740 <port id="0" precision="FP32" />
4741 </output>
4742 </layer>
4743 <layer id="347" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
4744 <data levels="256" auto_broadcast="numpy" />
4745 <input>
4746 <port id="0" precision="FP32">
4747 <dim>-1</dim>
4748 <dim>-1</dim>
4749 <dim>384</dim>
4750 </port>
4751 <port id="1" precision="FP32" />
4752 <port id="2" precision="FP32" />
4753 <port id="3" precision="FP32" />
4754 <port id="4" precision="FP32" />
4755 </input>
4756 <output>
4757 <port id="5" precision="FP32">
4758 <dim>-1</dim>
4759 <dim>-1</dim>
4760 <dim>384</dim>
4761 </port>
4762 </output>
4763 </layer>
4764 <layer id="348" name="Constant_299998" type="Const" version="opset1">
4765 <data element_type="i8" shape="384, 384" offset="16132112" size="147456" />
4766 <output>
4767 <port id="0" precision="I8">
4768 <dim>384</dim>
4769 <dim>384</dim>
4770 </port>
4771 </output>
4772 </layer>
4773 <layer id="349" name="Convert_299999" type="Convert" version="opset1">
4774 <data destination_type="f32" />
4775 <input>
4776 <port id="0" precision="I8">
4777 <dim>384</dim>
4778 <dim>384</dim>
4779 </port>
4780 </input>
4781 <output>
4782 <port id="1" precision="FP32">
4783 <dim>384</dim>
4784 <dim>384</dim>
4785 </port>
4786 </output>
4787 </layer>
4788 <layer id="350" name="Constant_300000" type="Const" version="opset1">
4789 <data element_type="f32" shape="384, 1" offset="16279568" size="1536" />
4790 <output>
4791 <port id="0" precision="FP32">
4792 <dim>384</dim>
4793 <dim>1</dim>
4794 </port>
4795 </output>
4796 </layer>
4797 <layer id="351" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4798 <data auto_broadcast="numpy" />
4799 <input>
4800 <port id="0" precision="FP32">
4801 <dim>384</dim>
4802 <dim>384</dim>
4803 </port>
4804 <port id="1" precision="FP32">
4805 <dim>384</dim>
4806 <dim>1</dim>
4807 </port>
4808 </input>
4809 <output>
4810 <port id="2" precision="FP32">
4811 <dim>384</dim>
4812 <dim>384</dim>
4813 </port>
4814 </output>
4815 </layer>
4816 <layer id="352" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4817 <data transpose_a="false" transpose_b="true" />
4818 <input>
4819 <port id="0" precision="FP32">
4820 <dim>-1</dim>
4821 <dim>-1</dim>
4822 <dim>384</dim>
4823 </port>
4824 <port id="1" precision="FP32">
4825 <dim>384</dim>
4826 <dim>384</dim>
4827 </port>
4828 </input>
4829 <output>
4830 <port id="2" precision="FP32">
4831 <dim>-1</dim>
4832 <dim>-1</dim>
4833 <dim>384</dim>
4834 </port>
4835 </output>
4836 </layer>
4837 <layer id="353" name="Constant_90328" type="Const" version="opset1">
4838 <data element_type="f32" shape="1, 1, 384" offset="16281104" size="1536" />
4839 <output>
4840 <port id="0" precision="FP32">
4841 <dim>1</dim>
4842 <dim>1</dim>
4843 <dim>384</dim>
4844 </port>
4845 </output>
4846 </layer>
4847 <layer id="354" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
4848 <data auto_broadcast="numpy" />
4849 <input>
4850 <port id="0" precision="FP32">
4851 <dim>-1</dim>
4852 <dim>-1</dim>
4853 <dim>384</dim>
4854 </port>
4855 <port id="1" precision="FP32">
4856 <dim>1</dim>
4857 <dim>1</dim>
4858 <dim>384</dim>
4859 </port>
4860 </input>
4861 <output>
4862 <port id="2" precision="FP32" names="265,input.11">
4863 <dim>-1</dim>
4864 <dim>-1</dim>
4865 <dim>384</dim>
4866 </port>
4867 </output>
4868 </layer>
4869 <layer id="355" name="__module.bert.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
4870 <data auto_broadcast="numpy" />
4871 <input>
4872 <port id="0" precision="FP32">
4873 <dim>-1</dim>
4874 <dim>-1</dim>
4875 <dim>384</dim>
4876 </port>
4877 <port id="1" precision="FP32">
4878 <dim>-1</dim>
4879 <dim>-1</dim>
4880 <dim>384</dim>
4881 </port>
4882 </input>
4883 <output>
4884 <port id="2" precision="FP32" names="267">
4885 <dim>-1</dim>
4886 <dim>-1</dim>
4887 <dim>384</dim>
4888 </port>
4889 </output>
4890 </layer>
4891 <layer id="356" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4892 <data element_type="i32" shape="1" offset="12046084" size="4" />
4893 <output>
4894 <port id="0" precision="I32">
4895 <dim>1</dim>
4896 </port>
4897 </output>
4898 </layer>
4899 <layer id="357" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4900 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4901 <input>
4902 <port id="0" precision="FP32">
4903 <dim>-1</dim>
4904 <dim>-1</dim>
4905 <dim>384</dim>
4906 </port>
4907 <port id="1" precision="I32">
4908 <dim>1</dim>
4909 </port>
4910 </input>
4911 <output>
4912 <port id="2" precision="FP32">
4913 <dim>-1</dim>
4914 <dim>-1</dim>
4915 <dim>384</dim>
4916 </port>
4917 </output>
4918 </layer>
4919 <layer id="358" name="Constant_90329" type="Const" version="opset1">
4920 <data element_type="f32" shape="1, 1, 384" offset="16282640" size="1536" />
4921 <output>
4922 <port id="0" precision="FP32">
4923 <dim>1</dim>
4924 <dim>1</dim>
4925 <dim>384</dim>
4926 </port>
4927 </output>
4928 </layer>
4929 <layer id="359" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4930 <data auto_broadcast="numpy" />
4931 <input>
4932 <port id="0" precision="FP32">
4933 <dim>-1</dim>
4934 <dim>-1</dim>
4935 <dim>384</dim>
4936 </port>
4937 <port id="1" precision="FP32">
4938 <dim>1</dim>
4939 <dim>1</dim>
4940 <dim>384</dim>
4941 </port>
4942 </input>
4943 <output>
4944 <port id="2" precision="FP32">
4945 <dim>-1</dim>
4946 <dim>-1</dim>
4947 <dim>384</dim>
4948 </port>
4949 </output>
4950 </layer>
4951 <layer id="360" name="Constant_90330" type="Const" version="opset1">
4952 <data element_type="f32" shape="1, 1, 384" offset="16284176" size="1536" />
4953 <output>
4954 <port id="0" precision="FP32">
4955 <dim>1</dim>
4956 <dim>1</dim>
4957 <dim>384</dim>
4958 </port>
4959 </output>
4960 </layer>
4961 <layer id="361" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4962 <data auto_broadcast="numpy" />
4963 <input>
4964 <port id="0" precision="FP32">
4965 <dim>-1</dim>
4966 <dim>-1</dim>
4967 <dim>384</dim>
4968 </port>
4969 <port id="1" precision="FP32">
4970 <dim>1</dim>
4971 <dim>1</dim>
4972 <dim>384</dim>
4973 </port>
4974 </input>
4975 <output>
4976 <port id="2" precision="FP32" names="271,input_tensor.5">
4977 <dim>-1</dim>
4978 <dim>-1</dim>
4979 <dim>384</dim>
4980 </port>
4981 </output>
4982 </layer>
4983 <layer id="362" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4984 <data element_type="f32" shape="1, 1, 384" offset="16285712" size="1536" />
4985 <output>
4986 <port id="0" precision="FP32">
4987 <dim>1</dim>
4988 <dim>1</dim>
4989 <dim>384</dim>
4990 </port>
4991 </output>
4992 </layer>
4993 <layer id="363" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4994 <data auto_broadcast="numpy" />
4995 <input>
4996 <port id="0" precision="FP32">
4997 <dim>-1</dim>
4998 <dim>-1</dim>
4999 <dim>384</dim>
5000 </port>
5001 <port id="1" precision="FP32">
5002 <dim>1</dim>
5003 <dim>1</dim>
5004 <dim>384</dim>
5005 </port>
5006 </input>
5007 <output>
5008 <port id="2" precision="FP32">
5009 <dim>-1</dim>
5010 <dim>-1</dim>
5011 <dim>384</dim>
5012 </port>
5013 </output>
5014 </layer>
5015 <layer id="364" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5016 <data element_type="f32" shape="" offset="16287248" size="4" />
5017 <output>
5018 <port id="0" precision="FP32" />
5019 </output>
5020 </layer>
5021 <layer id="365" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5022 <data element_type="f32" shape="" offset="16287252" size="4" />
5023 <output>
5024 <port id="0" precision="FP32" />
5025 </output>
5026 </layer>
5027 <layer id="366" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5028 <data element_type="f32" shape="" offset="16287248" size="4" />
5029 <output>
5030 <port id="0" precision="FP32" />
5031 </output>
5032 </layer>
5033 <layer id="367" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5034 <data element_type="f32" shape="" offset="16287252" size="4" />
5035 <output>
5036 <port id="0" precision="FP32" />
5037 </output>
5038 </layer>
5039 <layer id="368" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5040 <data levels="256" auto_broadcast="numpy" />
5041 <input>
5042 <port id="0" precision="FP32">
5043 <dim>-1</dim>
5044 <dim>-1</dim>
5045 <dim>384</dim>
5046 </port>
5047 <port id="1" precision="FP32" />
5048 <port id="2" precision="FP32" />
5049 <port id="3" precision="FP32" />
5050 <port id="4" precision="FP32" />
5051 </input>
5052 <output>
5053 <port id="5" precision="FP32">
5054 <dim>-1</dim>
5055 <dim>-1</dim>
5056 <dim>384</dim>
5057 </port>
5058 </output>
5059 </layer>
5060 <layer id="369" name="Constant_300002" type="Const" version="opset1">
5061 <data element_type="i8" shape="1536, 384" offset="16287256" size="589824" />
5062 <output>
5063 <port id="0" precision="I8">
5064 <dim>1536</dim>
5065 <dim>384</dim>
5066 </port>
5067 </output>
5068 </layer>
5069 <layer id="370" name="Convert_300003" type="Convert" version="opset1">
5070 <data destination_type="f32" />
5071 <input>
5072 <port id="0" precision="I8">
5073 <dim>1536</dim>
5074 <dim>384</dim>
5075 </port>
5076 </input>
5077 <output>
5078 <port id="1" precision="FP32">
5079 <dim>1536</dim>
5080 <dim>384</dim>
5081 </port>
5082 </output>
5083 </layer>
5084 <layer id="371" name="Constant_300004" type="Const" version="opset1">
5085 <data element_type="f32" shape="1536, 1" offset="16877080" size="6144" />
5086 <output>
5087 <port id="0" precision="FP32">
5088 <dim>1536</dim>
5089 <dim>1</dim>
5090 </port>
5091 </output>
5092 </layer>
5093 <layer id="372" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5094 <data auto_broadcast="numpy" />
5095 <input>
5096 <port id="0" precision="FP32">
5097 <dim>1536</dim>
5098 <dim>384</dim>
5099 </port>
5100 <port id="1" precision="FP32">
5101 <dim>1536</dim>
5102 <dim>1</dim>
5103 </port>
5104 </input>
5105 <output>
5106 <port id="2" precision="FP32">
5107 <dim>1536</dim>
5108 <dim>384</dim>
5109 </port>
5110 </output>
5111 </layer>
5112 <layer id="373" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5113 <data transpose_a="false" transpose_b="true" />
5114 <input>
5115 <port id="0" precision="FP32">
5116 <dim>-1</dim>
5117 <dim>-1</dim>
5118 <dim>384</dim>
5119 </port>
5120 <port id="1" precision="FP32">
5121 <dim>1536</dim>
5122 <dim>384</dim>
5123 </port>
5124 </input>
5125 <output>
5126 <port id="2" precision="FP32">
5127 <dim>-1</dim>
5128 <dim>-1</dim>
5129 <dim>1536</dim>
5130 </port>
5131 </output>
5132 </layer>
5133 <layer id="374" name="Constant_90331" type="Const" version="opset1">
5134 <data element_type="f32" shape="1, 1, 1536" offset="16883224" size="6144" />
5135 <output>
5136 <port id="0" precision="FP32">
5137 <dim>1</dim>
5138 <dim>1</dim>
5139 <dim>1536</dim>
5140 </port>
5141 </output>
5142 </layer>
5143 <layer id="375" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5144 <data auto_broadcast="numpy" />
5145 <input>
5146 <port id="0" precision="FP32">
5147 <dim>-1</dim>
5148 <dim>-1</dim>
5149 <dim>1536</dim>
5150 </port>
5151 <port id="1" precision="FP32">
5152 <dim>1</dim>
5153 <dim>1</dim>
5154 <dim>1536</dim>
5155 </port>
5156 </input>
5157 <output>
5158 <port id="2" precision="FP32" names="275">
5159 <dim>-1</dim>
5160 <dim>-1</dim>
5161 <dim>1536</dim>
5162 </port>
5163 </output>
5164 </layer>
5165 <layer id="376" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5166 <data approximation_mode="ERF" />
5167 <input>
5168 <port id="0" precision="FP32">
5169 <dim>-1</dim>
5170 <dim>-1</dim>
5171 <dim>1536</dim>
5172 </port>
5173 </input>
5174 <output>
5175 <port id="1" precision="FP32" names="276">
5176 <dim>-1</dim>
5177 <dim>-1</dim>
5178 <dim>1536</dim>
5179 </port>
5180 </output>
5181 </layer>
5182 <layer id="377" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5183 <data element_type="f32" shape="1, 1, 1536" offset="16889368" size="6144" />
5184 <output>
5185 <port id="0" precision="FP32">
5186 <dim>1</dim>
5187 <dim>1</dim>
5188 <dim>1536</dim>
5189 </port>
5190 </output>
5191 </layer>
5192 <layer id="378" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5193 <data auto_broadcast="numpy" />
5194 <input>
5195 <port id="0" precision="FP32">
5196 <dim>-1</dim>
5197 <dim>-1</dim>
5198 <dim>1536</dim>
5199 </port>
5200 <port id="1" precision="FP32">
5201 <dim>1</dim>
5202 <dim>1</dim>
5203 <dim>1536</dim>
5204 </port>
5205 </input>
5206 <output>
5207 <port id="2" precision="FP32">
5208 <dim>-1</dim>
5209 <dim>-1</dim>
5210 <dim>1536</dim>
5211 </port>
5212 </output>
5213 </layer>
5214 <layer id="379" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5215 <data element_type="f32" shape="" offset="16895512" size="4" />
5216 <output>
5217 <port id="0" precision="FP32" />
5218 </output>
5219 </layer>
5220 <layer id="380" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5221 <data element_type="f32" shape="" offset="16895516" size="4" />
5222 <output>
5223 <port id="0" precision="FP32" />
5224 </output>
5225 </layer>
5226 <layer id="381" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5227 <data element_type="f32" shape="" offset="16895512" size="4" />
5228 <output>
5229 <port id="0" precision="FP32" />
5230 </output>
5231 </layer>
5232 <layer id="382" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5233 <data element_type="f32" shape="" offset="16895516" size="4" />
5234 <output>
5235 <port id="0" precision="FP32" />
5236 </output>
5237 </layer>
5238 <layer id="383" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5239 <data levels="256" auto_broadcast="numpy" />
5240 <input>
5241 <port id="0" precision="FP32">
5242 <dim>-1</dim>
5243 <dim>-1</dim>
5244 <dim>1536</dim>
5245 </port>
5246 <port id="1" precision="FP32" />
5247 <port id="2" precision="FP32" />
5248 <port id="3" precision="FP32" />
5249 <port id="4" precision="FP32" />
5250 </input>
5251 <output>
5252 <port id="5" precision="FP32">
5253 <dim>-1</dim>
5254 <dim>-1</dim>
5255 <dim>1536</dim>
5256 </port>
5257 </output>
5258 </layer>
5259 <layer id="384" name="Constant_300006" type="Const" version="opset1">
5260 <data element_type="i8" shape="384, 1536" offset="16895520" size="589824" />
5261 <output>
5262 <port id="0" precision="I8">
5263 <dim>384</dim>
5264 <dim>1536</dim>
5265 </port>
5266 </output>
5267 </layer>
5268 <layer id="385" name="Convert_300007" type="Convert" version="opset1">
5269 <data destination_type="f32" />
5270 <input>
5271 <port id="0" precision="I8">
5272 <dim>384</dim>
5273 <dim>1536</dim>
5274 </port>
5275 </input>
5276 <output>
5277 <port id="1" precision="FP32">
5278 <dim>384</dim>
5279 <dim>1536</dim>
5280 </port>
5281 </output>
5282 </layer>
5283 <layer id="386" name="Constant_300008" type="Const" version="opset1">
5284 <data element_type="f32" shape="384, 1" offset="17485344" size="1536" />
5285 <output>
5286 <port id="0" precision="FP32">
5287 <dim>384</dim>
5288 <dim>1</dim>
5289 </port>
5290 </output>
5291 </layer>
5292 <layer id="387" name="__module.bert.encoder.layer.2.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5293 <data auto_broadcast="numpy" />
5294 <input>
5295 <port id="0" precision="FP32">
5296 <dim>384</dim>
5297 <dim>1536</dim>
5298 </port>
5299 <port id="1" precision="FP32">
5300 <dim>384</dim>
5301 <dim>1</dim>
5302 </port>
5303 </input>
5304 <output>
5305 <port id="2" precision="FP32">
5306 <dim>384</dim>
5307 <dim>1536</dim>
5308 </port>
5309 </output>
5310 </layer>
5311 <layer id="388" name="__module.bert.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5312 <data transpose_a="false" transpose_b="true" />
5313 <input>
5314 <port id="0" precision="FP32">
5315 <dim>-1</dim>
5316 <dim>-1</dim>
5317 <dim>1536</dim>
5318 </port>
5319 <port id="1" precision="FP32">
5320 <dim>384</dim>
5321 <dim>1536</dim>
5322 </port>
5323 </input>
5324 <output>
5325 <port id="2" precision="FP32">
5326 <dim>-1</dim>
5327 <dim>-1</dim>
5328 <dim>384</dim>
5329 </port>
5330 </output>
5331 </layer>
5332 <layer id="389" name="Constant_90332" type="Const" version="opset1">
5333 <data element_type="f32" shape="1, 1, 384" offset="17486880" size="1536" />
5334 <output>
5335 <port id="0" precision="FP32">
5336 <dim>1</dim>
5337 <dim>1</dim>
5338 <dim>384</dim>
5339 </port>
5340 </output>
5341 </layer>
5342 <layer id="390" name="__module.bert.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
5343 <data auto_broadcast="numpy" />
5344 <input>
5345 <port id="0" precision="FP32">
5346 <dim>-1</dim>
5347 <dim>-1</dim>
5348 <dim>384</dim>
5349 </port>
5350 <port id="1" precision="FP32">
5351 <dim>1</dim>
5352 <dim>1</dim>
5353 <dim>384</dim>
5354 </port>
5355 </input>
5356 <output>
5357 <port id="2" precision="FP32" names="281,input.13">
5358 <dim>-1</dim>
5359 <dim>-1</dim>
5360 <dim>384</dim>
5361 </port>
5362 </output>
5363 </layer>
5364 <layer id="391" name="__module.bert.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
5365 <data auto_broadcast="numpy" />
5366 <input>
5367 <port id="0" precision="FP32">
5368 <dim>-1</dim>
5369 <dim>-1</dim>
5370 <dim>384</dim>
5371 </port>
5372 <port id="1" precision="FP32">
5373 <dim>-1</dim>
5374 <dim>-1</dim>
5375 <dim>384</dim>
5376 </port>
5377 </input>
5378 <output>
5379 <port id="2" precision="FP32" names="283">
5380 <dim>-1</dim>
5381 <dim>-1</dim>
5382 <dim>384</dim>
5383 </port>
5384 </output>
5385 </layer>
5386 <layer id="392" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5387 <data element_type="i32" shape="1" offset="12046084" size="4" />
5388 <output>
5389 <port id="0" precision="I32">
5390 <dim>1</dim>
5391 </port>
5392 </output>
5393 </layer>
5394 <layer id="393" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5395 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5396 <input>
5397 <port id="0" precision="FP32">
5398 <dim>-1</dim>
5399 <dim>-1</dim>
5400 <dim>384</dim>
5401 </port>
5402 <port id="1" precision="I32">
5403 <dim>1</dim>
5404 </port>
5405 </input>
5406 <output>
5407 <port id="2" precision="FP32">
5408 <dim>-1</dim>
5409 <dim>-1</dim>
5410 <dim>384</dim>
5411 </port>
5412 </output>
5413 </layer>
5414 <layer id="394" name="Constant_90333" type="Const" version="opset1">
5415 <data element_type="f32" shape="1, 1, 384" offset="17488416" size="1536" />
5416 <output>
5417 <port id="0" precision="FP32">
5418 <dim>1</dim>
5419 <dim>1</dim>
5420 <dim>384</dim>
5421 </port>
5422 </output>
5423 </layer>
5424 <layer id="395" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5425 <data auto_broadcast="numpy" />
5426 <input>
5427 <port id="0" precision="FP32">
5428 <dim>-1</dim>
5429 <dim>-1</dim>
5430 <dim>384</dim>
5431 </port>
5432 <port id="1" precision="FP32">
5433 <dim>1</dim>
5434 <dim>1</dim>
5435 <dim>384</dim>
5436 </port>
5437 </input>
5438 <output>
5439 <port id="2" precision="FP32">
5440 <dim>-1</dim>
5441 <dim>-1</dim>
5442 <dim>384</dim>
5443 </port>
5444 </output>
5445 </layer>
5446 <layer id="396" name="Constant_90334" type="Const" version="opset1">
5447 <data element_type="f32" shape="1, 1, 384" offset="17489952" size="1536" />
5448 <output>
5449 <port id="0" precision="FP32">
5450 <dim>1</dim>
5451 <dim>1</dim>
5452 <dim>384</dim>
5453 </port>
5454 </output>
5455 </layer>
5456 <layer id="397" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5457 <data auto_broadcast="numpy" />
5458 <input>
5459 <port id="0" precision="FP32">
5460 <dim>-1</dim>
5461 <dim>-1</dim>
5462 <dim>384</dim>
5463 </port>
5464 <port id="1" precision="FP32">
5465 <dim>1</dim>
5466 <dim>1</dim>
5467 <dim>384</dim>
5468 </port>
5469 </input>
5470 <output>
5471 <port id="2" precision="FP32" names="287,hidden_states.19">
5472 <dim>-1</dim>
5473 <dim>-1</dim>
5474 <dim>384</dim>
5475 </port>
5476 </output>
5477 </layer>
5478 <layer id="398" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5479 <data element_type="f32" shape="1, 1, 384" offset="17491488" size="1536" />
5480 <output>
5481 <port id="0" precision="FP32">
5482 <dim>1</dim>
5483 <dim>1</dim>
5484 <dim>384</dim>
5485 </port>
5486 </output>
5487 </layer>
5488 <layer id="399" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5489 <data auto_broadcast="numpy" />
5490 <input>
5491 <port id="0" precision="FP32">
5492 <dim>-1</dim>
5493 <dim>-1</dim>
5494 <dim>384</dim>
5495 </port>
5496 <port id="1" precision="FP32">
5497 <dim>1</dim>
5498 <dim>1</dim>
5499 <dim>384</dim>
5500 </port>
5501 </input>
5502 <output>
5503 <port id="2" precision="FP32">
5504 <dim>-1</dim>
5505 <dim>-1</dim>
5506 <dim>384</dim>
5507 </port>
5508 </output>
5509 </layer>
5510 <layer id="400" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5511 <data element_type="f32" shape="" offset="17493024" size="4" />
5512 <output>
5513 <port id="0" precision="FP32" />
5514 </output>
5515 </layer>
5516 <layer id="401" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5517 <data element_type="f32" shape="" offset="17493028" size="4" />
5518 <output>
5519 <port id="0" precision="FP32" />
5520 </output>
5521 </layer>
5522 <layer id="402" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5523 <data element_type="f32" shape="" offset="17493024" size="4" />
5524 <output>
5525 <port id="0" precision="FP32" />
5526 </output>
5527 </layer>
5528 <layer id="403" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5529 <data element_type="f32" shape="" offset="17493028" size="4" />
5530 <output>
5531 <port id="0" precision="FP32" />
5532 </output>
5533 </layer>
5534 <layer id="404" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5535 <data levels="256" auto_broadcast="numpy" />
5536 <input>
5537 <port id="0" precision="FP32">
5538 <dim>-1</dim>
5539 <dim>-1</dim>
5540 <dim>384</dim>
5541 </port>
5542 <port id="1" precision="FP32" />
5543 <port id="2" precision="FP32" />
5544 <port id="3" precision="FP32" />
5545 <port id="4" precision="FP32" />
5546 </input>
5547 <output>
5548 <port id="5" precision="FP32">
5549 <dim>-1</dim>
5550 <dim>-1</dim>
5551 <dim>384</dim>
5552 </port>
5553 </output>
5554 </layer>
5555 <layer id="405" name="Constant_300010" type="Const" version="opset1">
5556 <data element_type="i8" shape="384, 384" offset="17493032" size="147456" />
5557 <output>
5558 <port id="0" precision="I8">
5559 <dim>384</dim>
5560 <dim>384</dim>
5561 </port>
5562 </output>
5563 </layer>
5564 <layer id="406" name="Convert_300011" type="Convert" version="opset1">
5565 <data destination_type="f32" />
5566 <input>
5567 <port id="0" precision="I8">
5568 <dim>384</dim>
5569 <dim>384</dim>
5570 </port>
5571 </input>
5572 <output>
5573 <port id="1" precision="FP32">
5574 <dim>384</dim>
5575 <dim>384</dim>
5576 </port>
5577 </output>
5578 </layer>
5579 <layer id="407" name="Constant_300012" type="Const" version="opset1">
5580 <data element_type="f32" shape="384, 1" offset="17640488" size="1536" />
5581 <output>
5582 <port id="0" precision="FP32">
5583 <dim>384</dim>
5584 <dim>1</dim>
5585 </port>
5586 </output>
5587 </layer>
5588 <layer id="408" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5589 <data auto_broadcast="numpy" />
5590 <input>
5591 <port id="0" precision="FP32">
5592 <dim>384</dim>
5593 <dim>384</dim>
5594 </port>
5595 <port id="1" precision="FP32">
5596 <dim>384</dim>
5597 <dim>1</dim>
5598 </port>
5599 </input>
5600 <output>
5601 <port id="2" precision="FP32">
5602 <dim>384</dim>
5603 <dim>384</dim>
5604 </port>
5605 </output>
5606 </layer>
5607 <layer id="409" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
5608 <data transpose_a="false" transpose_b="true" />
5609 <input>
5610 <port id="0" precision="FP32">
5611 <dim>-1</dim>
5612 <dim>-1</dim>
5613 <dim>384</dim>
5614 </port>
5615 <port id="1" precision="FP32">
5616 <dim>384</dim>
5617 <dim>384</dim>
5618 </port>
5619 </input>
5620 <output>
5621 <port id="2" precision="FP32">
5622 <dim>-1</dim>
5623 <dim>-1</dim>
5624 <dim>384</dim>
5625 </port>
5626 </output>
5627 </layer>
5628 <layer id="410" name="Constant_90335" type="Const" version="opset1">
5629 <data element_type="f32" shape="1, 1, 384" offset="17642024" size="1536" />
5630 <output>
5631 <port id="0" precision="FP32">
5632 <dim>1</dim>
5633 <dim>1</dim>
5634 <dim>384</dim>
5635 </port>
5636 </output>
5637 </layer>
5638 <layer id="411" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
5639 <data auto_broadcast="numpy" />
5640 <input>
5641 <port id="0" precision="FP32">
5642 <dim>-1</dim>
5643 <dim>-1</dim>
5644 <dim>384</dim>
5645 </port>
5646 <port id="1" precision="FP32">
5647 <dim>1</dim>
5648 <dim>1</dim>
5649 <dim>384</dim>
5650 </port>
5651 </input>
5652 <output>
5653 <port id="2" precision="FP32" names="300,x.37">
5654 <dim>-1</dim>
5655 <dim>-1</dim>
5656 <dim>384</dim>
5657 </port>
5658 </output>
5659 </layer>
5660 <layer id="412" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5661 <data element_type="f32" shape="" offset="17643560" size="4" />
5662 <output>
5663 <port id="0" precision="FP32" />
5664 </output>
5665 </layer>
5666 <layer id="413" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
5667 <data element_type="f32" shape="" offset="17643564" size="4" />
5668 <output>
5669 <port id="0" precision="FP32" />
5670 </output>
5671 </layer>
5672 <layer id="414" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
5673 <data element_type="f32" shape="" offset="17643560" size="4" />
5674 <output>
5675 <port id="0" precision="FP32" />
5676 </output>
5677 </layer>
5678 <layer id="415" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
5679 <data element_type="f32" shape="" offset="17643564" size="4" />
5680 <output>
5681 <port id="0" precision="FP32" />
5682 </output>
5683 </layer>
5684 <layer id="416" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
5685 <data levels="256" auto_broadcast="numpy" />
5686 <input>
5687 <port id="0" precision="FP32">
5688 <dim>-1</dim>
5689 <dim>-1</dim>
5690 <dim>384</dim>
5691 </port>
5692 <port id="1" precision="FP32" />
5693 <port id="2" precision="FP32" />
5694 <port id="3" precision="FP32" />
5695 <port id="4" precision="FP32" />
5696 </input>
5697 <output>
5698 <port id="5" precision="FP32">
5699 <dim>-1</dim>
5700 <dim>-1</dim>
5701 <dim>384</dim>
5702 </port>
5703 </output>
5704 </layer>
5705 <layer id="417" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
5706 <data element_type="i64" shape="4" offset="12201240" size="32" />
5707 <output>
5708 <port id="0" precision="I64">
5709 <dim>4</dim>
5710 </port>
5711 </output>
5712 </layer>
5713 <layer id="418" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
5714 <data special_zero="true" />
5715 <input>
5716 <port id="0" precision="FP32">
5717 <dim>-1</dim>
5718 <dim>-1</dim>
5719 <dim>384</dim>
5720 </port>
5721 <port id="1" precision="I64">
5722 <dim>4</dim>
5723 </port>
5724 </input>
5725 <output>
5726 <port id="2" precision="FP32" names="304,x.39">
5727 <dim>-1</dim>
5728 <dim>-1</dim>
5729 <dim>12</dim>
5730 <dim>32</dim>
5731 </port>
5732 </output>
5733 </layer>
5734 <layer id="419" name="Constant_85127" type="Const" version="opset1">
5735 <data element_type="i64" shape="4" offset="12201272" size="32" />
5736 <output>
5737 <port id="0" precision="I64" names="305">
5738 <dim>4</dim>
5739 </port>
5740 </output>
5741 </layer>
5742 <layer id="420" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
5743 <input>
5744 <port id="0" precision="FP32">
5745 <dim>-1</dim>
5746 <dim>-1</dim>
5747 <dim>12</dim>
5748 <dim>32</dim>
5749 </port>
5750 <port id="1" precision="I64">
5751 <dim>4</dim>
5752 </port>
5753 </input>
5754 <output>
5755 <port id="2" precision="FP32" names="306">
5756 <dim>-1</dim>
5757 <dim>12</dim>
5758 <dim>-1</dim>
5759 <dim>32</dim>
5760 </port>
5761 </output>
5762 </layer>
5763 <layer id="421" name="Constant_300014" type="Const" version="opset1">
5764 <data element_type="i8" shape="384, 384" offset="17643568" size="147456" />
5765 <output>
5766 <port id="0" precision="I8">
5767 <dim>384</dim>
5768 <dim>384</dim>
5769 </port>
5770 </output>
5771 </layer>
5772 <layer id="422" name="Convert_300015" type="Convert" version="opset1">
5773 <data destination_type="f32" />
5774 <input>
5775 <port id="0" precision="I8">
5776 <dim>384</dim>
5777 <dim>384</dim>
5778 </port>
5779 </input>
5780 <output>
5781 <port id="1" precision="FP32">
5782 <dim>384</dim>
5783 <dim>384</dim>
5784 </port>
5785 </output>
5786 </layer>
5787 <layer id="423" name="Constant_300016" type="Const" version="opset1">
5788 <data element_type="f32" shape="384, 1" offset="17791024" size="1536" />
5789 <output>
5790 <port id="0" precision="FP32">
5791 <dim>384</dim>
5792 <dim>1</dim>
5793 </port>
5794 </output>
5795 </layer>
5796 <layer id="424" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5797 <data auto_broadcast="numpy" />
5798 <input>
5799 <port id="0" precision="FP32">
5800 <dim>384</dim>
5801 <dim>384</dim>
5802 </port>
5803 <port id="1" precision="FP32">
5804 <dim>384</dim>
5805 <dim>1</dim>
5806 </port>
5807 </input>
5808 <output>
5809 <port id="2" precision="FP32">
5810 <dim>384</dim>
5811 <dim>384</dim>
5812 </port>
5813 </output>
5814 </layer>
5815 <layer id="425" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
5816 <data transpose_a="false" transpose_b="true" />
5817 <input>
5818 <port id="0" precision="FP32">
5819 <dim>-1</dim>
5820 <dim>-1</dim>
5821 <dim>384</dim>
5822 </port>
5823 <port id="1" precision="FP32">
5824 <dim>384</dim>
5825 <dim>384</dim>
5826 </port>
5827 </input>
5828 <output>
5829 <port id="2" precision="FP32">
5830 <dim>-1</dim>
5831 <dim>-1</dim>
5832 <dim>384</dim>
5833 </port>
5834 </output>
5835 </layer>
5836 <layer id="426" name="Constant_90336" type="Const" version="opset1">
5837 <data element_type="f32" shape="1, 1, 384" offset="17792560" size="1536" />
5838 <output>
5839 <port id="0" precision="FP32">
5840 <dim>1</dim>
5841 <dim>1</dim>
5842 <dim>384</dim>
5843 </port>
5844 </output>
5845 </layer>
5846 <layer id="427" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
5847 <data auto_broadcast="numpy" />
5848 <input>
5849 <port id="0" precision="FP32">
5850 <dim>-1</dim>
5851 <dim>-1</dim>
5852 <dim>384</dim>
5853 </port>
5854 <port id="1" precision="FP32">
5855 <dim>1</dim>
5856 <dim>1</dim>
5857 <dim>384</dim>
5858 </port>
5859 </input>
5860 <output>
5861 <port id="2" precision="FP32" names="309,x.41">
5862 <dim>-1</dim>
5863 <dim>-1</dim>
5864 <dim>384</dim>
5865 </port>
5866 </output>
5867 </layer>
5868 <layer id="428" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5869 <data element_type="f32" shape="" offset="17794096" size="4" />
5870 <output>
5871 <port id="0" precision="FP32" />
5872 </output>
5873 </layer>
5874 <layer id="429" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
5875 <data element_type="f32" shape="" offset="17794100" size="4" />
5876 <output>
5877 <port id="0" precision="FP32" />
5878 </output>
5879 </layer>
5880 <layer id="430" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
5881 <data element_type="f32" shape="" offset="17794096" size="4" />
5882 <output>
5883 <port id="0" precision="FP32" />
5884 </output>
5885 </layer>
5886 <layer id="431" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
5887 <data element_type="f32" shape="" offset="17794100" size="4" />
5888 <output>
5889 <port id="0" precision="FP32" />
5890 </output>
5891 </layer>
5892 <layer id="432" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
5893 <data levels="256" auto_broadcast="numpy" />
5894 <input>
5895 <port id="0" precision="FP32">
5896 <dim>-1</dim>
5897 <dim>-1</dim>
5898 <dim>384</dim>
5899 </port>
5900 <port id="1" precision="FP32" />
5901 <port id="2" precision="FP32" />
5902 <port id="3" precision="FP32" />
5903 <port id="4" precision="FP32" />
5904 </input>
5905 <output>
5906 <port id="5" precision="FP32">
5907 <dim>-1</dim>
5908 <dim>-1</dim>
5909 <dim>384</dim>
5910 </port>
5911 </output>
5912 </layer>
5913 <layer id="433" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
5914 <data element_type="i64" shape="4" offset="12201240" size="32" />
5915 <output>
5916 <port id="0" precision="I64">
5917 <dim>4</dim>
5918 </port>
5919 </output>
5920 </layer>
5921 <layer id="434" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
5922 <data special_zero="true" />
5923 <input>
5924 <port id="0" precision="FP32">
5925 <dim>-1</dim>
5926 <dim>-1</dim>
5927 <dim>384</dim>
5928 </port>
5929 <port id="1" precision="I64">
5930 <dim>4</dim>
5931 </port>
5932 </input>
5933 <output>
5934 <port id="2" precision="FP32" names="313,x.43">
5935 <dim>-1</dim>
5936 <dim>-1</dim>
5937 <dim>12</dim>
5938 <dim>32</dim>
5939 </port>
5940 </output>
5941 </layer>
5942 <layer id="435" name="Constant_85150" type="Const" version="opset1">
5943 <data element_type="i64" shape="4" offset="12201272" size="32" />
5944 <output>
5945 <port id="0" precision="I64" names="314">
5946 <dim>4</dim>
5947 </port>
5948 </output>
5949 </layer>
5950 <layer id="436" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
5951 <input>
5952 <port id="0" precision="FP32">
5953 <dim>-1</dim>
5954 <dim>-1</dim>
5955 <dim>12</dim>
5956 <dim>32</dim>
5957 </port>
5958 <port id="1" precision="I64">
5959 <dim>4</dim>
5960 </port>
5961 </input>
5962 <output>
5963 <port id="2" precision="FP32" names="315">
5964 <dim>-1</dim>
5965 <dim>12</dim>
5966 <dim>-1</dim>
5967 <dim>32</dim>
5968 </port>
5969 </output>
5970 </layer>
5971 <layer id="437" name="Constant_300018" type="Const" version="opset1">
5972 <data element_type="i8" shape="384, 384" offset="17794104" size="147456" />
5973 <output>
5974 <port id="0" precision="I8">
5975 <dim>384</dim>
5976 <dim>384</dim>
5977 </port>
5978 </output>
5979 </layer>
5980 <layer id="438" name="Convert_300019" type="Convert" version="opset1">
5981 <data destination_type="f32" />
5982 <input>
5983 <port id="0" precision="I8">
5984 <dim>384</dim>
5985 <dim>384</dim>
5986 </port>
5987 </input>
5988 <output>
5989 <port id="1" precision="FP32">
5990 <dim>384</dim>
5991 <dim>384</dim>
5992 </port>
5993 </output>
5994 </layer>
5995 <layer id="439" name="Constant_300020" type="Const" version="opset1">
5996 <data element_type="f32" shape="384, 1" offset="17941560" size="1536" />
5997 <output>
5998 <port id="0" precision="FP32">
5999 <dim>384</dim>
6000 <dim>1</dim>
6001 </port>
6002 </output>
6003 </layer>
6004 <layer id="440" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6005 <data auto_broadcast="numpy" />
6006 <input>
6007 <port id="0" precision="FP32">
6008 <dim>384</dim>
6009 <dim>384</dim>
6010 </port>
6011 <port id="1" precision="FP32">
6012 <dim>384</dim>
6013 <dim>1</dim>
6014 </port>
6015 </input>
6016 <output>
6017 <port id="2" precision="FP32">
6018 <dim>384</dim>
6019 <dim>384</dim>
6020 </port>
6021 </output>
6022 </layer>
6023 <layer id="441" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
6024 <data transpose_a="false" transpose_b="true" />
6025 <input>
6026 <port id="0" precision="FP32">
6027 <dim>-1</dim>
6028 <dim>-1</dim>
6029 <dim>384</dim>
6030 </port>
6031 <port id="1" precision="FP32">
6032 <dim>384</dim>
6033 <dim>384</dim>
6034 </port>
6035 </input>
6036 <output>
6037 <port id="2" precision="FP32">
6038 <dim>-1</dim>
6039 <dim>-1</dim>
6040 <dim>384</dim>
6041 </port>
6042 </output>
6043 </layer>
6044 <layer id="442" name="Constant_90337" type="Const" version="opset1">
6045 <data element_type="f32" shape="1, 1, 384" offset="17943096" size="1536" />
6046 <output>
6047 <port id="0" precision="FP32">
6048 <dim>1</dim>
6049 <dim>1</dim>
6050 <dim>384</dim>
6051 </port>
6052 </output>
6053 </layer>
6054 <layer id="443" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
6055 <data auto_broadcast="numpy" />
6056 <input>
6057 <port id="0" precision="FP32">
6058 <dim>-1</dim>
6059 <dim>-1</dim>
6060 <dim>384</dim>
6061 </port>
6062 <port id="1" precision="FP32">
6063 <dim>1</dim>
6064 <dim>1</dim>
6065 <dim>384</dim>
6066 </port>
6067 </input>
6068 <output>
6069 <port id="2" precision="FP32" names="318,x.45">
6070 <dim>-1</dim>
6071 <dim>-1</dim>
6072 <dim>384</dim>
6073 </port>
6074 </output>
6075 </layer>
6076 <layer id="444" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
6077 <data element_type="i64" shape="4" offset="12201240" size="32" />
6078 <output>
6079 <port id="0" precision="I64">
6080 <dim>4</dim>
6081 </port>
6082 </output>
6083 </layer>
6084 <layer id="445" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
6085 <data special_zero="true" />
6086 <input>
6087 <port id="0" precision="FP32">
6088 <dim>-1</dim>
6089 <dim>-1</dim>
6090 <dim>384</dim>
6091 </port>
6092 <port id="1" precision="I64">
6093 <dim>4</dim>
6094 </port>
6095 </input>
6096 <output>
6097 <port id="2" precision="FP32" names="322,x.47">
6098 <dim>-1</dim>
6099 <dim>-1</dim>
6100 <dim>12</dim>
6101 <dim>32</dim>
6102 </port>
6103 </output>
6104 </layer>
6105 <layer id="446" name="Constant_85173" type="Const" version="opset1">
6106 <data element_type="i64" shape="4" offset="12201272" size="32" />
6107 <output>
6108 <port id="0" precision="I64" names="323">
6109 <dim>4</dim>
6110 </port>
6111 </output>
6112 </layer>
6113 <layer id="447" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
6114 <input>
6115 <port id="0" precision="FP32">
6116 <dim>-1</dim>
6117 <dim>-1</dim>
6118 <dim>12</dim>
6119 <dim>32</dim>
6120 </port>
6121 <port id="1" precision="I64">
6122 <dim>4</dim>
6123 </port>
6124 </input>
6125 <output>
6126 <port id="2" precision="FP32" names="324">
6127 <dim>-1</dim>
6128 <dim>12</dim>
6129 <dim>-1</dim>
6130 <dim>32</dim>
6131 </port>
6132 </output>
6133 </layer>
6134 <layer id="448" name="__module.bert.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
6135 <data causal="false" />
6136 <input>
6137 <port id="0" precision="FP32">
6138 <dim>-1</dim>
6139 <dim>12</dim>
6140 <dim>-1</dim>
6141 <dim>32</dim>
6142 </port>
6143 <port id="1" precision="FP32">
6144 <dim>-1</dim>
6145 <dim>12</dim>
6146 <dim>-1</dim>
6147 <dim>32</dim>
6148 </port>
6149 <port id="2" precision="FP32">
6150 <dim>-1</dim>
6151 <dim>12</dim>
6152 <dim>-1</dim>
6153 <dim>32</dim>
6154 </port>
6155 <port id="3" precision="FP32">
6156 <dim>-1</dim>
6157 <dim>1</dim>
6158 <dim>-1</dim>
6159 <dim>-1</dim>
6160 </port>
6161 </input>
6162 <output>
6163 <port id="4" precision="FP32" names="325,attn_output.13">
6164 <dim>-1</dim>
6165 <dim>12</dim>
6166 <dim>-1</dim>
6167 <dim>32</dim>
6168 </port>
6169 </output>
6170 </layer>
6171 <layer id="449" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
6172 <data element_type="i32" shape="4" offset="12502400" size="16" />
6173 <output>
6174 <port id="0" precision="I32">
6175 <dim>4</dim>
6176 </port>
6177 </output>
6178 </layer>
6179 <layer id="450" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
6180 <input>
6181 <port id="0" precision="FP32">
6182 <dim>-1</dim>
6183 <dim>12</dim>
6184 <dim>-1</dim>
6185 <dim>32</dim>
6186 </port>
6187 <port id="1" precision="I32">
6188 <dim>4</dim>
6189 </port>
6190 </input>
6191 <output>
6192 <port id="2" precision="FP32" names="326,attn_output.15">
6193 <dim>-1</dim>
6194 <dim>-1</dim>
6195 <dim>12</dim>
6196 <dim>32</dim>
6197 </port>
6198 </output>
6199 </layer>
6200 <layer id="451" name="Constant_90446" type="Const" version="opset1">
6201 <data element_type="i64" shape="3" offset="12502416" size="24" />
6202 <output>
6203 <port id="0" precision="I64">
6204 <dim>3</dim>
6205 </port>
6206 </output>
6207 </layer>
6208 <layer id="452" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
6209 <data special_zero="true" />
6210 <input>
6211 <port id="0" precision="FP32">
6212 <dim>-1</dim>
6213 <dim>-1</dim>
6214 <dim>12</dim>
6215 <dim>32</dim>
6216 </port>
6217 <port id="1" precision="I64">
6218 <dim>3</dim>
6219 </port>
6220 </input>
6221 <output>
6222 <port id="2" precision="FP32" names="328">
6223 <dim>-1</dim>
6224 <dim>-1</dim>
6225 <dim>384</dim>
6226 </port>
6227 </output>
6228 </layer>
6229 <layer id="453" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6230 <data element_type="f32" shape="1, 1, 384" offset="17944632" size="1536" />
6231 <output>
6232 <port id="0" precision="FP32">
6233 <dim>1</dim>
6234 <dim>1</dim>
6235 <dim>384</dim>
6236 </port>
6237 </output>
6238 </layer>
6239 <layer id="454" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6240 <data auto_broadcast="numpy" />
6241 <input>
6242 <port id="0" precision="FP32">
6243 <dim>-1</dim>
6244 <dim>-1</dim>
6245 <dim>384</dim>
6246 </port>
6247 <port id="1" precision="FP32">
6248 <dim>1</dim>
6249 <dim>1</dim>
6250 <dim>384</dim>
6251 </port>
6252 </input>
6253 <output>
6254 <port id="2" precision="FP32">
6255 <dim>-1</dim>
6256 <dim>-1</dim>
6257 <dim>384</dim>
6258 </port>
6259 </output>
6260 </layer>
6261 <layer id="455" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6262 <data element_type="f32" shape="" offset="17946168" size="4" />
6263 <output>
6264 <port id="0" precision="FP32" />
6265 </output>
6266 </layer>
6267 <layer id="456" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6268 <data element_type="f32" shape="" offset="17946172" size="4" />
6269 <output>
6270 <port id="0" precision="FP32" />
6271 </output>
6272 </layer>
6273 <layer id="457" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6274 <data element_type="f32" shape="" offset="17946168" size="4" />
6275 <output>
6276 <port id="0" precision="FP32" />
6277 </output>
6278 </layer>
6279 <layer id="458" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6280 <data element_type="f32" shape="" offset="17946172" size="4" />
6281 <output>
6282 <port id="0" precision="FP32" />
6283 </output>
6284 </layer>
6285 <layer id="459" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6286 <data levels="256" auto_broadcast="numpy" />
6287 <input>
6288 <port id="0" precision="FP32">
6289 <dim>-1</dim>
6290 <dim>-1</dim>
6291 <dim>384</dim>
6292 </port>
6293 <port id="1" precision="FP32" />
6294 <port id="2" precision="FP32" />
6295 <port id="3" precision="FP32" />
6296 <port id="4" precision="FP32" />
6297 </input>
6298 <output>
6299 <port id="5" precision="FP32">
6300 <dim>-1</dim>
6301 <dim>-1</dim>
6302 <dim>384</dim>
6303 </port>
6304 </output>
6305 </layer>
6306 <layer id="460" name="Constant_300022" type="Const" version="opset1">
6307 <data element_type="i8" shape="384, 384" offset="17946176" size="147456" />
6308 <output>
6309 <port id="0" precision="I8">
6310 <dim>384</dim>
6311 <dim>384</dim>
6312 </port>
6313 </output>
6314 </layer>
6315 <layer id="461" name="Convert_300023" type="Convert" version="opset1">
6316 <data destination_type="f32" />
6317 <input>
6318 <port id="0" precision="I8">
6319 <dim>384</dim>
6320 <dim>384</dim>
6321 </port>
6322 </input>
6323 <output>
6324 <port id="1" precision="FP32">
6325 <dim>384</dim>
6326 <dim>384</dim>
6327 </port>
6328 </output>
6329 </layer>
6330 <layer id="462" name="Constant_300024" type="Const" version="opset1">
6331 <data element_type="f32" shape="384, 1" offset="18093632" size="1536" />
6332 <output>
6333 <port id="0" precision="FP32">
6334 <dim>384</dim>
6335 <dim>1</dim>
6336 </port>
6337 </output>
6338 </layer>
6339 <layer id="463" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6340 <data auto_broadcast="numpy" />
6341 <input>
6342 <port id="0" precision="FP32">
6343 <dim>384</dim>
6344 <dim>384</dim>
6345 </port>
6346 <port id="1" precision="FP32">
6347 <dim>384</dim>
6348 <dim>1</dim>
6349 </port>
6350 </input>
6351 <output>
6352 <port id="2" precision="FP32">
6353 <dim>384</dim>
6354 <dim>384</dim>
6355 </port>
6356 </output>
6357 </layer>
6358 <layer id="464" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6359 <data transpose_a="false" transpose_b="true" />
6360 <input>
6361 <port id="0" precision="FP32">
6362 <dim>-1</dim>
6363 <dim>-1</dim>
6364 <dim>384</dim>
6365 </port>
6366 <port id="1" precision="FP32">
6367 <dim>384</dim>
6368 <dim>384</dim>
6369 </port>
6370 </input>
6371 <output>
6372 <port id="2" precision="FP32">
6373 <dim>-1</dim>
6374 <dim>-1</dim>
6375 <dim>384</dim>
6376 </port>
6377 </output>
6378 </layer>
6379 <layer id="465" name="Constant_90338" type="Const" version="opset1">
6380 <data element_type="f32" shape="1, 1, 384" offset="18095168" size="1536" />
6381 <output>
6382 <port id="0" precision="FP32">
6383 <dim>1</dim>
6384 <dim>1</dim>
6385 <dim>384</dim>
6386 </port>
6387 </output>
6388 </layer>
6389 <layer id="466" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
6390 <data auto_broadcast="numpy" />
6391 <input>
6392 <port id="0" precision="FP32">
6393 <dim>-1</dim>
6394 <dim>-1</dim>
6395 <dim>384</dim>
6396 </port>
6397 <port id="1" precision="FP32">
6398 <dim>1</dim>
6399 <dim>1</dim>
6400 <dim>384</dim>
6401 </port>
6402 </input>
6403 <output>
6404 <port id="2" precision="FP32" names="333,input.15">
6405 <dim>-1</dim>
6406 <dim>-1</dim>
6407 <dim>384</dim>
6408 </port>
6409 </output>
6410 </layer>
6411 <layer id="467" name="__module.bert.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
6412 <data auto_broadcast="numpy" />
6413 <input>
6414 <port id="0" precision="FP32">
6415 <dim>-1</dim>
6416 <dim>-1</dim>
6417 <dim>384</dim>
6418 </port>
6419 <port id="1" precision="FP32">
6420 <dim>-1</dim>
6421 <dim>-1</dim>
6422 <dim>384</dim>
6423 </port>
6424 </input>
6425 <output>
6426 <port id="2" precision="FP32" names="335">
6427 <dim>-1</dim>
6428 <dim>-1</dim>
6429 <dim>384</dim>
6430 </port>
6431 </output>
6432 </layer>
6433 <layer id="468" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6434 <data element_type="i32" shape="1" offset="12046084" size="4" />
6435 <output>
6436 <port id="0" precision="I32">
6437 <dim>1</dim>
6438 </port>
6439 </output>
6440 </layer>
6441 <layer id="469" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6442 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6443 <input>
6444 <port id="0" precision="FP32">
6445 <dim>-1</dim>
6446 <dim>-1</dim>
6447 <dim>384</dim>
6448 </port>
6449 <port id="1" precision="I32">
6450 <dim>1</dim>
6451 </port>
6452 </input>
6453 <output>
6454 <port id="2" precision="FP32">
6455 <dim>-1</dim>
6456 <dim>-1</dim>
6457 <dim>384</dim>
6458 </port>
6459 </output>
6460 </layer>
6461 <layer id="470" name="Constant_90339" type="Const" version="opset1">
6462 <data element_type="f32" shape="1, 1, 384" offset="18096704" size="1536" />
6463 <output>
6464 <port id="0" precision="FP32">
6465 <dim>1</dim>
6466 <dim>1</dim>
6467 <dim>384</dim>
6468 </port>
6469 </output>
6470 </layer>
6471 <layer id="471" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6472 <data auto_broadcast="numpy" />
6473 <input>
6474 <port id="0" precision="FP32">
6475 <dim>-1</dim>
6476 <dim>-1</dim>
6477 <dim>384</dim>
6478 </port>
6479 <port id="1" precision="FP32">
6480 <dim>1</dim>
6481 <dim>1</dim>
6482 <dim>384</dim>
6483 </port>
6484 </input>
6485 <output>
6486 <port id="2" precision="FP32">
6487 <dim>-1</dim>
6488 <dim>-1</dim>
6489 <dim>384</dim>
6490 </port>
6491 </output>
6492 </layer>
6493 <layer id="472" name="Constant_90340" type="Const" version="opset1">
6494 <data element_type="f32" shape="1, 1, 384" offset="18098240" size="1536" />
6495 <output>
6496 <port id="0" precision="FP32">
6497 <dim>1</dim>
6498 <dim>1</dim>
6499 <dim>384</dim>
6500 </port>
6501 </output>
6502 </layer>
6503 <layer id="473" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6504 <data auto_broadcast="numpy" />
6505 <input>
6506 <port id="0" precision="FP32">
6507 <dim>-1</dim>
6508 <dim>-1</dim>
6509 <dim>384</dim>
6510 </port>
6511 <port id="1" precision="FP32">
6512 <dim>1</dim>
6513 <dim>1</dim>
6514 <dim>384</dim>
6515 </port>
6516 </input>
6517 <output>
6518 <port id="2" precision="FP32" names="339,input_tensor.7">
6519 <dim>-1</dim>
6520 <dim>-1</dim>
6521 <dim>384</dim>
6522 </port>
6523 </output>
6524 </layer>
6525 <layer id="474" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6526 <data element_type="f32" shape="1, 1, 384" offset="18099776" size="1536" />
6527 <output>
6528 <port id="0" precision="FP32">
6529 <dim>1</dim>
6530 <dim>1</dim>
6531 <dim>384</dim>
6532 </port>
6533 </output>
6534 </layer>
6535 <layer id="475" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6536 <data auto_broadcast="numpy" />
6537 <input>
6538 <port id="0" precision="FP32">
6539 <dim>-1</dim>
6540 <dim>-1</dim>
6541 <dim>384</dim>
6542 </port>
6543 <port id="1" precision="FP32">
6544 <dim>1</dim>
6545 <dim>1</dim>
6546 <dim>384</dim>
6547 </port>
6548 </input>
6549 <output>
6550 <port id="2" precision="FP32">
6551 <dim>-1</dim>
6552 <dim>-1</dim>
6553 <dim>384</dim>
6554 </port>
6555 </output>
6556 </layer>
6557 <layer id="476" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6558 <data element_type="f32" shape="" offset="18101312" size="4" />
6559 <output>
6560 <port id="0" precision="FP32" />
6561 </output>
6562 </layer>
6563 <layer id="477" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6564 <data element_type="f32" shape="" offset="18101316" size="4" />
6565 <output>
6566 <port id="0" precision="FP32" />
6567 </output>
6568 </layer>
6569 <layer id="478" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6570 <data element_type="f32" shape="" offset="18101312" size="4" />
6571 <output>
6572 <port id="0" precision="FP32" />
6573 </output>
6574 </layer>
6575 <layer id="479" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6576 <data element_type="f32" shape="" offset="18101316" size="4" />
6577 <output>
6578 <port id="0" precision="FP32" />
6579 </output>
6580 </layer>
6581 <layer id="480" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6582 <data levels="256" auto_broadcast="numpy" />
6583 <input>
6584 <port id="0" precision="FP32">
6585 <dim>-1</dim>
6586 <dim>-1</dim>
6587 <dim>384</dim>
6588 </port>
6589 <port id="1" precision="FP32" />
6590 <port id="2" precision="FP32" />
6591 <port id="3" precision="FP32" />
6592 <port id="4" precision="FP32" />
6593 </input>
6594 <output>
6595 <port id="5" precision="FP32">
6596 <dim>-1</dim>
6597 <dim>-1</dim>
6598 <dim>384</dim>
6599 </port>
6600 </output>
6601 </layer>
6602 <layer id="481" name="Constant_300026" type="Const" version="opset1">
6603 <data element_type="i8" shape="1536, 384" offset="18101320" size="589824" />
6604 <output>
6605 <port id="0" precision="I8">
6606 <dim>1536</dim>
6607 <dim>384</dim>
6608 </port>
6609 </output>
6610 </layer>
6611 <layer id="482" name="Convert_300027" type="Convert" version="opset1">
6612 <data destination_type="f32" />
6613 <input>
6614 <port id="0" precision="I8">
6615 <dim>1536</dim>
6616 <dim>384</dim>
6617 </port>
6618 </input>
6619 <output>
6620 <port id="1" precision="FP32">
6621 <dim>1536</dim>
6622 <dim>384</dim>
6623 </port>
6624 </output>
6625 </layer>
6626 <layer id="483" name="Constant_300028" type="Const" version="opset1">
6627 <data element_type="f32" shape="1536, 1" offset="18691144" size="6144" />
6628 <output>
6629 <port id="0" precision="FP32">
6630 <dim>1536</dim>
6631 <dim>1</dim>
6632 </port>
6633 </output>
6634 </layer>
6635 <layer id="484" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6636 <data auto_broadcast="numpy" />
6637 <input>
6638 <port id="0" precision="FP32">
6639 <dim>1536</dim>
6640 <dim>384</dim>
6641 </port>
6642 <port id="1" precision="FP32">
6643 <dim>1536</dim>
6644 <dim>1</dim>
6645 </port>
6646 </input>
6647 <output>
6648 <port id="2" precision="FP32">
6649 <dim>1536</dim>
6650 <dim>384</dim>
6651 </port>
6652 </output>
6653 </layer>
6654 <layer id="485" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6655 <data transpose_a="false" transpose_b="true" />
6656 <input>
6657 <port id="0" precision="FP32">
6658 <dim>-1</dim>
6659 <dim>-1</dim>
6660 <dim>384</dim>
6661 </port>
6662 <port id="1" precision="FP32">
6663 <dim>1536</dim>
6664 <dim>384</dim>
6665 </port>
6666 </input>
6667 <output>
6668 <port id="2" precision="FP32">
6669 <dim>-1</dim>
6670 <dim>-1</dim>
6671 <dim>1536</dim>
6672 </port>
6673 </output>
6674 </layer>
6675 <layer id="486" name="Constant_90341" type="Const" version="opset1">
6676 <data element_type="f32" shape="1, 1, 1536" offset="18697288" size="6144" />
6677 <output>
6678 <port id="0" precision="FP32">
6679 <dim>1</dim>
6680 <dim>1</dim>
6681 <dim>1536</dim>
6682 </port>
6683 </output>
6684 </layer>
6685 <layer id="487" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
6686 <data auto_broadcast="numpy" />
6687 <input>
6688 <port id="0" precision="FP32">
6689 <dim>-1</dim>
6690 <dim>-1</dim>
6691 <dim>1536</dim>
6692 </port>
6693 <port id="1" precision="FP32">
6694 <dim>1</dim>
6695 <dim>1</dim>
6696 <dim>1536</dim>
6697 </port>
6698 </input>
6699 <output>
6700 <port id="2" precision="FP32" names="343">
6701 <dim>-1</dim>
6702 <dim>-1</dim>
6703 <dim>1536</dim>
6704 </port>
6705 </output>
6706 </layer>
6707 <layer id="488" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
6708 <data approximation_mode="ERF" />
6709 <input>
6710 <port id="0" precision="FP32">
6711 <dim>-1</dim>
6712 <dim>-1</dim>
6713 <dim>1536</dim>
6714 </port>
6715 </input>
6716 <output>
6717 <port id="1" precision="FP32" names="344">
6718 <dim>-1</dim>
6719 <dim>-1</dim>
6720 <dim>1536</dim>
6721 </port>
6722 </output>
6723 </layer>
6724 <layer id="489" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6725 <data element_type="f32" shape="1, 1, 1536" offset="18703432" size="6144" />
6726 <output>
6727 <port id="0" precision="FP32">
6728 <dim>1</dim>
6729 <dim>1</dim>
6730 <dim>1536</dim>
6731 </port>
6732 </output>
6733 </layer>
6734 <layer id="490" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6735 <data auto_broadcast="numpy" />
6736 <input>
6737 <port id="0" precision="FP32">
6738 <dim>-1</dim>
6739 <dim>-1</dim>
6740 <dim>1536</dim>
6741 </port>
6742 <port id="1" precision="FP32">
6743 <dim>1</dim>
6744 <dim>1</dim>
6745 <dim>1536</dim>
6746 </port>
6747 </input>
6748 <output>
6749 <port id="2" precision="FP32">
6750 <dim>-1</dim>
6751 <dim>-1</dim>
6752 <dim>1536</dim>
6753 </port>
6754 </output>
6755 </layer>
6756 <layer id="491" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6757 <data element_type="f32" shape="" offset="18709576" size="4" />
6758 <output>
6759 <port id="0" precision="FP32" />
6760 </output>
6761 </layer>
6762 <layer id="492" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6763 <data element_type="f32" shape="" offset="18709580" size="4" />
6764 <output>
6765 <port id="0" precision="FP32" />
6766 </output>
6767 </layer>
6768 <layer id="493" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6769 <data element_type="f32" shape="" offset="18709576" size="4" />
6770 <output>
6771 <port id="0" precision="FP32" />
6772 </output>
6773 </layer>
6774 <layer id="494" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6775 <data element_type="f32" shape="" offset="18709580" size="4" />
6776 <output>
6777 <port id="0" precision="FP32" />
6778 </output>
6779 </layer>
6780 <layer id="495" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6781 <data levels="256" auto_broadcast="numpy" />
6782 <input>
6783 <port id="0" precision="FP32">
6784 <dim>-1</dim>
6785 <dim>-1</dim>
6786 <dim>1536</dim>
6787 </port>
6788 <port id="1" precision="FP32" />
6789 <port id="2" precision="FP32" />
6790 <port id="3" precision="FP32" />
6791 <port id="4" precision="FP32" />
6792 </input>
6793 <output>
6794 <port id="5" precision="FP32">
6795 <dim>-1</dim>
6796 <dim>-1</dim>
6797 <dim>1536</dim>
6798 </port>
6799 </output>
6800 </layer>
6801 <layer id="496" name="Constant_300030" type="Const" version="opset1">
6802 <data element_type="i8" shape="384, 1536" offset="18709584" size="589824" />
6803 <output>
6804 <port id="0" precision="I8">
6805 <dim>384</dim>
6806 <dim>1536</dim>
6807 </port>
6808 </output>
6809 </layer>
6810 <layer id="497" name="Convert_300031" type="Convert" version="opset1">
6811 <data destination_type="f32" />
6812 <input>
6813 <port id="0" precision="I8">
6814 <dim>384</dim>
6815 <dim>1536</dim>
6816 </port>
6817 </input>
6818 <output>
6819 <port id="1" precision="FP32">
6820 <dim>384</dim>
6821 <dim>1536</dim>
6822 </port>
6823 </output>
6824 </layer>
6825 <layer id="498" name="Constant_300032" type="Const" version="opset1">
6826 <data element_type="f32" shape="384, 1" offset="19299408" size="1536" />
6827 <output>
6828 <port id="0" precision="FP32">
6829 <dim>384</dim>
6830 <dim>1</dim>
6831 </port>
6832 </output>
6833 </layer>
6834 <layer id="499" name="__module.bert.encoder.layer.3.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6835 <data auto_broadcast="numpy" />
6836 <input>
6837 <port id="0" precision="FP32">
6838 <dim>384</dim>
6839 <dim>1536</dim>
6840 </port>
6841 <port id="1" precision="FP32">
6842 <dim>384</dim>
6843 <dim>1</dim>
6844 </port>
6845 </input>
6846 <output>
6847 <port id="2" precision="FP32">
6848 <dim>384</dim>
6849 <dim>1536</dim>
6850 </port>
6851 </output>
6852 </layer>
6853 <layer id="500" name="__module.bert.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6854 <data transpose_a="false" transpose_b="true" />
6855 <input>
6856 <port id="0" precision="FP32">
6857 <dim>-1</dim>
6858 <dim>-1</dim>
6859 <dim>1536</dim>
6860 </port>
6861 <port id="1" precision="FP32">
6862 <dim>384</dim>
6863 <dim>1536</dim>
6864 </port>
6865 </input>
6866 <output>
6867 <port id="2" precision="FP32">
6868 <dim>-1</dim>
6869 <dim>-1</dim>
6870 <dim>384</dim>
6871 </port>
6872 </output>
6873 </layer>
6874 <layer id="501" name="Constant_90342" type="Const" version="opset1">
6875 <data element_type="f32" shape="1, 1, 384" offset="19300944" size="1536" />
6876 <output>
6877 <port id="0" precision="FP32">
6878 <dim>1</dim>
6879 <dim>1</dim>
6880 <dim>384</dim>
6881 </port>
6882 </output>
6883 </layer>
6884 <layer id="502" name="__module.bert.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
6885 <data auto_broadcast="numpy" />
6886 <input>
6887 <port id="0" precision="FP32">
6888 <dim>-1</dim>
6889 <dim>-1</dim>
6890 <dim>384</dim>
6891 </port>
6892 <port id="1" precision="FP32">
6893 <dim>1</dim>
6894 <dim>1</dim>
6895 <dim>384</dim>
6896 </port>
6897 </input>
6898 <output>
6899 <port id="2" precision="FP32" names="349,input.17">
6900 <dim>-1</dim>
6901 <dim>-1</dim>
6902 <dim>384</dim>
6903 </port>
6904 </output>
6905 </layer>
6906 <layer id="503" name="__module.bert.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
6907 <data auto_broadcast="numpy" />
6908 <input>
6909 <port id="0" precision="FP32">
6910 <dim>-1</dim>
6911 <dim>-1</dim>
6912 <dim>384</dim>
6913 </port>
6914 <port id="1" precision="FP32">
6915 <dim>-1</dim>
6916 <dim>-1</dim>
6917 <dim>384</dim>
6918 </port>
6919 </input>
6920 <output>
6921 <port id="2" precision="FP32" names="351">
6922 <dim>-1</dim>
6923 <dim>-1</dim>
6924 <dim>384</dim>
6925 </port>
6926 </output>
6927 </layer>
6928 <layer id="504" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6929 <data element_type="i32" shape="1" offset="12046084" size="4" />
6930 <output>
6931 <port id="0" precision="I32">
6932 <dim>1</dim>
6933 </port>
6934 </output>
6935 </layer>
6936 <layer id="505" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6937 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6938 <input>
6939 <port id="0" precision="FP32">
6940 <dim>-1</dim>
6941 <dim>-1</dim>
6942 <dim>384</dim>
6943 </port>
6944 <port id="1" precision="I32">
6945 <dim>1</dim>
6946 </port>
6947 </input>
6948 <output>
6949 <port id="2" precision="FP32">
6950 <dim>-1</dim>
6951 <dim>-1</dim>
6952 <dim>384</dim>
6953 </port>
6954 </output>
6955 </layer>
6956 <layer id="506" name="Constant_90343" type="Const" version="opset1">
6957 <data element_type="f32" shape="1, 1, 384" offset="19302480" size="1536" />
6958 <output>
6959 <port id="0" precision="FP32">
6960 <dim>1</dim>
6961 <dim>1</dim>
6962 <dim>384</dim>
6963 </port>
6964 </output>
6965 </layer>
6966 <layer id="507" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6967 <data auto_broadcast="numpy" />
6968 <input>
6969 <port id="0" precision="FP32">
6970 <dim>-1</dim>
6971 <dim>-1</dim>
6972 <dim>384</dim>
6973 </port>
6974 <port id="1" precision="FP32">
6975 <dim>1</dim>
6976 <dim>1</dim>
6977 <dim>384</dim>
6978 </port>
6979 </input>
6980 <output>
6981 <port id="2" precision="FP32">
6982 <dim>-1</dim>
6983 <dim>-1</dim>
6984 <dim>384</dim>
6985 </port>
6986 </output>
6987 </layer>
6988 <layer id="508" name="Constant_90344" type="Const" version="opset1">
6989 <data element_type="f32" shape="1, 1, 384" offset="19304016" size="1536" />
6990 <output>
6991 <port id="0" precision="FP32">
6992 <dim>1</dim>
6993 <dim>1</dim>
6994 <dim>384</dim>
6995 </port>
6996 </output>
6997 </layer>
6998 <layer id="509" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6999 <data auto_broadcast="numpy" />
7000 <input>
7001 <port id="0" precision="FP32">
7002 <dim>-1</dim>
7003 <dim>-1</dim>
7004 <dim>384</dim>
7005 </port>
7006 <port id="1" precision="FP32">
7007 <dim>1</dim>
7008 <dim>1</dim>
7009 <dim>384</dim>
7010 </port>
7011 </input>
7012 <output>
7013 <port id="2" precision="FP32" names="355,hidden_states.25">
7014 <dim>-1</dim>
7015 <dim>-1</dim>
7016 <dim>384</dim>
7017 </port>
7018 </output>
7019 </layer>
7020 <layer id="510" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7021 <data element_type="f32" shape="1, 1, 384" offset="19305552" size="1536" />
7022 <output>
7023 <port id="0" precision="FP32">
7024 <dim>1</dim>
7025 <dim>1</dim>
7026 <dim>384</dim>
7027 </port>
7028 </output>
7029 </layer>
7030 <layer id="511" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7031 <data auto_broadcast="numpy" />
7032 <input>
7033 <port id="0" precision="FP32">
7034 <dim>-1</dim>
7035 <dim>-1</dim>
7036 <dim>384</dim>
7037 </port>
7038 <port id="1" precision="FP32">
7039 <dim>1</dim>
7040 <dim>1</dim>
7041 <dim>384</dim>
7042 </port>
7043 </input>
7044 <output>
7045 <port id="2" precision="FP32">
7046 <dim>-1</dim>
7047 <dim>-1</dim>
7048 <dim>384</dim>
7049 </port>
7050 </output>
7051 </layer>
7052 <layer id="512" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7053 <data element_type="f32" shape="" offset="19307088" size="4" />
7054 <output>
7055 <port id="0" precision="FP32" />
7056 </output>
7057 </layer>
7058 <layer id="513" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7059 <data element_type="f32" shape="" offset="19307092" size="4" />
7060 <output>
7061 <port id="0" precision="FP32" />
7062 </output>
7063 </layer>
7064 <layer id="514" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7065 <data element_type="f32" shape="" offset="19307088" size="4" />
7066 <output>
7067 <port id="0" precision="FP32" />
7068 </output>
7069 </layer>
7070 <layer id="515" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7071 <data element_type="f32" shape="" offset="19307092" size="4" />
7072 <output>
7073 <port id="0" precision="FP32" />
7074 </output>
7075 </layer>
7076 <layer id="516" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7077 <data levels="256" auto_broadcast="numpy" />
7078 <input>
7079 <port id="0" precision="FP32">
7080 <dim>-1</dim>
7081 <dim>-1</dim>
7082 <dim>384</dim>
7083 </port>
7084 <port id="1" precision="FP32" />
7085 <port id="2" precision="FP32" />
7086 <port id="3" precision="FP32" />
7087 <port id="4" precision="FP32" />
7088 </input>
7089 <output>
7090 <port id="5" precision="FP32">
7091 <dim>-1</dim>
7092 <dim>-1</dim>
7093 <dim>384</dim>
7094 </port>
7095 </output>
7096 </layer>
7097 <layer id="517" name="Constant_300034" type="Const" version="opset1">
7098 <data element_type="i8" shape="384, 384" offset="19307096" size="147456" />
7099 <output>
7100 <port id="0" precision="I8">
7101 <dim>384</dim>
7102 <dim>384</dim>
7103 </port>
7104 </output>
7105 </layer>
7106 <layer id="518" name="Convert_300035" type="Convert" version="opset1">
7107 <data destination_type="f32" />
7108 <input>
7109 <port id="0" precision="I8">
7110 <dim>384</dim>
7111 <dim>384</dim>
7112 </port>
7113 </input>
7114 <output>
7115 <port id="1" precision="FP32">
7116 <dim>384</dim>
7117 <dim>384</dim>
7118 </port>
7119 </output>
7120 </layer>
7121 <layer id="519" name="Constant_300036" type="Const" version="opset1">
7122 <data element_type="f32" shape="384, 1" offset="19454552" size="1536" />
7123 <output>
7124 <port id="0" precision="FP32">
7125 <dim>384</dim>
7126 <dim>1</dim>
7127 </port>
7128 </output>
7129 </layer>
7130 <layer id="520" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7131 <data auto_broadcast="numpy" />
7132 <input>
7133 <port id="0" precision="FP32">
7134 <dim>384</dim>
7135 <dim>384</dim>
7136 </port>
7137 <port id="1" precision="FP32">
7138 <dim>384</dim>
7139 <dim>1</dim>
7140 </port>
7141 </input>
7142 <output>
7143 <port id="2" precision="FP32">
7144 <dim>384</dim>
7145 <dim>384</dim>
7146 </port>
7147 </output>
7148 </layer>
7149 <layer id="521" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
7150 <data transpose_a="false" transpose_b="true" />
7151 <input>
7152 <port id="0" precision="FP32">
7153 <dim>-1</dim>
7154 <dim>-1</dim>
7155 <dim>384</dim>
7156 </port>
7157 <port id="1" precision="FP32">
7158 <dim>384</dim>
7159 <dim>384</dim>
7160 </port>
7161 </input>
7162 <output>
7163 <port id="2" precision="FP32">
7164 <dim>-1</dim>
7165 <dim>-1</dim>
7166 <dim>384</dim>
7167 </port>
7168 </output>
7169 </layer>
7170 <layer id="522" name="Constant_90345" type="Const" version="opset1">
7171 <data element_type="f32" shape="1, 1, 384" offset="19456088" size="1536" />
7172 <output>
7173 <port id="0" precision="FP32">
7174 <dim>1</dim>
7175 <dim>1</dim>
7176 <dim>384</dim>
7177 </port>
7178 </output>
7179 </layer>
7180 <layer id="523" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add" type="Add" version="opset1">
7181 <data auto_broadcast="numpy" />
7182 <input>
7183 <port id="0" precision="FP32">
7184 <dim>-1</dim>
7185 <dim>-1</dim>
7186 <dim>384</dim>
7187 </port>
7188 <port id="1" precision="FP32">
7189 <dim>1</dim>
7190 <dim>1</dim>
7191 <dim>384</dim>
7192 </port>
7193 </input>
7194 <output>
7195 <port id="2" precision="FP32" names="368,x.49">
7196 <dim>-1</dim>
7197 <dim>-1</dim>
7198 <dim>384</dim>
7199 </port>
7200 </output>
7201 </layer>
7202 <layer id="524" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
7203 <data element_type="f32" shape="" offset="19457624" size="4" />
7204 <output>
7205 <port id="0" precision="FP32" />
7206 </output>
7207 </layer>
7208 <layer id="525" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
7209 <data element_type="f32" shape="" offset="19457628" size="4" />
7210 <output>
7211 <port id="0" precision="FP32" />
7212 </output>
7213 </layer>
7214 <layer id="526" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
7215 <data element_type="f32" shape="" offset="19457624" size="4" />
7216 <output>
7217 <port id="0" precision="FP32" />
7218 </output>
7219 </layer>
7220 <layer id="527" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
7221 <data element_type="f32" shape="" offset="19457628" size="4" />
7222 <output>
7223 <port id="0" precision="FP32" />
7224 </output>
7225 </layer>
7226 <layer id="528" name="__module.bert.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
7227 <data levels="256" auto_broadcast="numpy" />
7228 <input>
7229 <port id="0" precision="FP32">
7230 <dim>-1</dim>
7231 <dim>-1</dim>
7232 <dim>384</dim>
7233 </port>
7234 <port id="1" precision="FP32" />
7235 <port id="2" precision="FP32" />
7236 <port id="3" precision="FP32" />
7237 <port id="4" precision="FP32" />
7238 </input>
7239 <output>
7240 <port id="5" precision="FP32">
7241 <dim>-1</dim>
7242 <dim>-1</dim>
7243 <dim>384</dim>
7244 </port>
7245 </output>
7246 </layer>
7247 <layer id="529" name="__module.bert.encoder.layer.4.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
7248 <data element_type="i64" shape="4" offset="12201240" size="32" />
7249 <output>
7250 <port id="0" precision="I64">
7251 <dim>4</dim>
7252 </port>
7253 </output>
7254 </layer>
7255 <layer id="530" name="__module.bert.encoder.layer.4.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
7256 <data special_zero="true" />
7257 <input>
7258 <port id="0" precision="FP32">
7259 <dim>-1</dim>
7260 <dim>-1</dim>
7261 <dim>384</dim>
7262 </port>
7263 <port id="1" precision="I64">
7264 <dim>4</dim>
7265 </port>
7266 </input>
7267 <output>
7268 <port id="2" precision="FP32" names="372,x.51">
7269 <dim>-1</dim>
7270 <dim>-1</dim>
7271 <dim>12</dim>
7272 <dim>32</dim>
7273 </port>
7274 </output>
7275 </layer>
7276 <layer id="531" name="Constant_85350" type="Const" version="opset1">
7277 <data element_type="i64" shape="4" offset="12201272" size="32" />
7278 <output>
7279 <port id="0" precision="I64" names="373">
7280 <dim>4</dim>
7281 </port>
7282 </output>
7283 </layer>
7284 <layer id="532" name="__module.bert.encoder.layer.4.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
7285 <input>
7286 <port id="0" precision="FP32">
7287 <dim>-1</dim>
7288 <dim>-1</dim>
7289 <dim>12</dim>
7290 <dim>32</dim>
7291 </port>
7292 <port id="1" precision="I64">
7293 <dim>4</dim>
7294 </port>
7295 </input>
7296 <output>
7297 <port id="2" precision="FP32" names="374">
7298 <dim>-1</dim>
7299 <dim>12</dim>
7300 <dim>-1</dim>
7301 <dim>32</dim>
7302 </port>
7303 </output>
7304 </layer>
7305 <layer id="533" name="Constant_300038" type="Const" version="opset1">
7306 <data element_type="i8" shape="384, 384" offset="19457632" size="147456" />
7307 <output>
7308 <port id="0" precision="I8">
7309 <dim>384</dim>
7310 <dim>384</dim>
7311 </port>
7312 </output>
7313 </layer>
7314 <layer id="534" name="Convert_300039" type="Convert" version="opset1">
7315 <data destination_type="f32" />
7316 <input>
7317 <port id="0" precision="I8">
7318 <dim>384</dim>
7319 <dim>384</dim>
7320 </port>
7321 </input>
7322 <output>
7323 <port id="1" precision="FP32">
7324 <dim>384</dim>
7325 <dim>384</dim>
7326 </port>
7327 </output>
7328 </layer>
7329 <layer id="535" name="Constant_300040" type="Const" version="opset1">
7330 <data element_type="f32" shape="384, 1" offset="19605088" size="1536" />
7331 <output>
7332 <port id="0" precision="FP32">
7333 <dim>384</dim>
7334 <dim>1</dim>
7335 </port>
7336 </output>
7337 </layer>
7338 <layer id="536" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7339 <data auto_broadcast="numpy" />
7340 <input>
7341 <port id="0" precision="FP32">
7342 <dim>384</dim>
7343 <dim>384</dim>
7344 </port>
7345 <port id="1" precision="FP32">
7346 <dim>384</dim>
7347 <dim>1</dim>
7348 </port>
7349 </input>
7350 <output>
7351 <port id="2" precision="FP32">
7352 <dim>384</dim>
7353 <dim>384</dim>
7354 </port>
7355 </output>
7356 </layer>
7357 <layer id="537" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
7358 <data transpose_a="false" transpose_b="true" />
7359 <input>
7360 <port id="0" precision="FP32">
7361 <dim>-1</dim>
7362 <dim>-1</dim>
7363 <dim>384</dim>
7364 </port>
7365 <port id="1" precision="FP32">
7366 <dim>384</dim>
7367 <dim>384</dim>
7368 </port>
7369 </input>
7370 <output>
7371 <port id="2" precision="FP32">
7372 <dim>-1</dim>
7373 <dim>-1</dim>
7374 <dim>384</dim>
7375 </port>
7376 </output>
7377 </layer>
7378 <layer id="538" name="Constant_90346" type="Const" version="opset1">
7379 <data element_type="f32" shape="1, 1, 384" offset="19606624" size="1536" />
7380 <output>
7381 <port id="0" precision="FP32">
7382 <dim>1</dim>
7383 <dim>1</dim>
7384 <dim>384</dim>
7385 </port>
7386 </output>
7387 </layer>
7388 <layer id="539" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add" type="Add" version="opset1">
7389 <data auto_broadcast="numpy" />
7390 <input>
7391 <port id="0" precision="FP32">
7392 <dim>-1</dim>
7393 <dim>-1</dim>
7394 <dim>384</dim>
7395 </port>
7396 <port id="1" precision="FP32">
7397 <dim>1</dim>
7398 <dim>1</dim>
7399 <dim>384</dim>
7400 </port>
7401 </input>
7402 <output>
7403 <port id="2" precision="FP32" names="377,x.53">
7404 <dim>-1</dim>
7405 <dim>-1</dim>
7406 <dim>384</dim>
7407 </port>
7408 </output>
7409 </layer>
7410 <layer id="540" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
7411 <data element_type="f32" shape="" offset="19608160" size="4" />
7412 <output>
7413 <port id="0" precision="FP32" />
7414 </output>
7415 </layer>
7416 <layer id="541" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
7417 <data element_type="f32" shape="" offset="19608164" size="4" />
7418 <output>
7419 <port id="0" precision="FP32" />
7420 </output>
7421 </layer>
7422 <layer id="542" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
7423 <data element_type="f32" shape="" offset="19608160" size="4" />
7424 <output>
7425 <port id="0" precision="FP32" />
7426 </output>
7427 </layer>
7428 <layer id="543" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
7429 <data element_type="f32" shape="" offset="19608164" size="4" />
7430 <output>
7431 <port id="0" precision="FP32" />
7432 </output>
7433 </layer>
7434 <layer id="544" name="__module.bert.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
7435 <data levels="256" auto_broadcast="numpy" />
7436 <input>
7437 <port id="0" precision="FP32">
7438 <dim>-1</dim>
7439 <dim>-1</dim>
7440 <dim>384</dim>
7441 </port>
7442 <port id="1" precision="FP32" />
7443 <port id="2" precision="FP32" />
7444 <port id="3" precision="FP32" />
7445 <port id="4" precision="FP32" />
7446 </input>
7447 <output>
7448 <port id="5" precision="FP32">
7449 <dim>-1</dim>
7450 <dim>-1</dim>
7451 <dim>384</dim>
7452 </port>
7453 </output>
7454 </layer>
7455 <layer id="545" name="__module.bert.encoder.layer.4.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
7456 <data element_type="i64" shape="4" offset="12201240" size="32" />
7457 <output>
7458 <port id="0" precision="I64">
7459 <dim>4</dim>
7460 </port>
7461 </output>
7462 </layer>
7463 <layer id="546" name="__module.bert.encoder.layer.4.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
7464 <data special_zero="true" />
7465 <input>
7466 <port id="0" precision="FP32">
7467 <dim>-1</dim>
7468 <dim>-1</dim>
7469 <dim>384</dim>
7470 </port>
7471 <port id="1" precision="I64">
7472 <dim>4</dim>
7473 </port>
7474 </input>
7475 <output>
7476 <port id="2" precision="FP32" names="381,x.55">
7477 <dim>-1</dim>
7478 <dim>-1</dim>
7479 <dim>12</dim>
7480 <dim>32</dim>
7481 </port>
7482 </output>
7483 </layer>
7484 <layer id="547" name="Constant_85373" type="Const" version="opset1">
7485 <data element_type="i64" shape="4" offset="12201272" size="32" />
7486 <output>
7487 <port id="0" precision="I64" names="382">
7488 <dim>4</dim>
7489 </port>
7490 </output>
7491 </layer>
7492 <layer id="548" name="__module.bert.encoder.layer.4.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
7493 <input>
7494 <port id="0" precision="FP32">
7495 <dim>-1</dim>
7496 <dim>-1</dim>
7497 <dim>12</dim>
7498 <dim>32</dim>
7499 </port>
7500 <port id="1" precision="I64">
7501 <dim>4</dim>
7502 </port>
7503 </input>
7504 <output>
7505 <port id="2" precision="FP32" names="383">
7506 <dim>-1</dim>
7507 <dim>12</dim>
7508 <dim>-1</dim>
7509 <dim>32</dim>
7510 </port>
7511 </output>
7512 </layer>
7513 <layer id="549" name="Constant_300042" type="Const" version="opset1">
7514 <data element_type="i8" shape="384, 384" offset="19608168" size="147456" />
7515 <output>
7516 <port id="0" precision="I8">
7517 <dim>384</dim>
7518 <dim>384</dim>
7519 </port>
7520 </output>
7521 </layer>
7522 <layer id="550" name="Convert_300043" type="Convert" version="opset1">
7523 <data destination_type="f32" />
7524 <input>
7525 <port id="0" precision="I8">
7526 <dim>384</dim>
7527 <dim>384</dim>
7528 </port>
7529 </input>
7530 <output>
7531 <port id="1" precision="FP32">
7532 <dim>384</dim>
7533 <dim>384</dim>
7534 </port>
7535 </output>
7536 </layer>
7537 <layer id="551" name="Constant_300044" type="Const" version="opset1">
7538 <data element_type="f32" shape="384, 1" offset="19755624" size="1536" />
7539 <output>
7540 <port id="0" precision="FP32">
7541 <dim>384</dim>
7542 <dim>1</dim>
7543 </port>
7544 </output>
7545 </layer>
7546 <layer id="552" name="__module.bert.encoder.layer.4.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7547 <data auto_broadcast="numpy" />
7548 <input>
7549 <port id="0" precision="FP32">
7550 <dim>384</dim>
7551 <dim>384</dim>
7552 </port>
7553 <port id="1" precision="FP32">
7554 <dim>384</dim>
7555 <dim>1</dim>
7556 </port>
7557 </input>
7558 <output>
7559 <port id="2" precision="FP32">
7560 <dim>384</dim>
7561 <dim>384</dim>
7562 </port>
7563 </output>
7564 </layer>
7565 <layer id="553" name="__module.bert.encoder.layer.4.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
7566 <data transpose_a="false" transpose_b="true" />
7567 <input>
7568 <port id="0" precision="FP32">
7569 <dim>-1</dim>
7570 <dim>-1</dim>
7571 <dim>384</dim>
7572 </port>
7573 <port id="1" precision="FP32">
7574 <dim>384</dim>
7575 <dim>384</dim>
7576 </port>
7577 </input>
7578 <output>
7579 <port id="2" precision="FP32">
7580 <dim>-1</dim>
7581 <dim>-1</dim>
7582 <dim>384</dim>
7583 </port>
7584 </output>
7585 </layer>
7586 <layer id="554" name="Constant_90347" type="Const" version="opset1">
7587 <data element_type="f32" shape="1, 1, 384" offset="19757160" size="1536" />
7588 <output>
7589 <port id="0" precision="FP32">
7590 <dim>1</dim>
7591 <dim>1</dim>
7592 <dim>384</dim>
7593 </port>
7594 </output>
7595 </layer>
7596 <layer id="555" name="__module.bert.encoder.layer.4.attention.self.value/aten::linear/Add" type="Add" version="opset1">
7597 <data auto_broadcast="numpy" />
7598 <input>
7599 <port id="0" precision="FP32">
7600 <dim>-1</dim>
7601 <dim>-1</dim>
7602 <dim>384</dim>
7603 </port>
7604 <port id="1" precision="FP32">
7605 <dim>1</dim>
7606 <dim>1</dim>
7607 <dim>384</dim>
7608 </port>
7609 </input>
7610 <output>
7611 <port id="2" precision="FP32" names="386,x.57">
7612 <dim>-1</dim>
7613 <dim>-1</dim>
7614 <dim>384</dim>
7615 </port>
7616 </output>
7617 </layer>
7618 <layer id="556" name="__module.bert.encoder.layer.4.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
7619 <data element_type="i64" shape="4" offset="12201240" size="32" />
7620 <output>
7621 <port id="0" precision="I64">
7622 <dim>4</dim>
7623 </port>
7624 </output>
7625 </layer>
7626 <layer id="557" name="__module.bert.encoder.layer.4.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
7627 <data special_zero="true" />
7628 <input>
7629 <port id="0" precision="FP32">
7630 <dim>-1</dim>
7631 <dim>-1</dim>
7632 <dim>384</dim>
7633 </port>
7634 <port id="1" precision="I64">
7635 <dim>4</dim>
7636 </port>
7637 </input>
7638 <output>
7639 <port id="2" precision="FP32" names="390,x.59">
7640 <dim>-1</dim>
7641 <dim>-1</dim>
7642 <dim>12</dim>
7643 <dim>32</dim>
7644 </port>
7645 </output>
7646 </layer>
7647 <layer id="558" name="Constant_85396" type="Const" version="opset1">
7648 <data element_type="i64" shape="4" offset="12201272" size="32" />
7649 <output>
7650 <port id="0" precision="I64" names="391">
7651 <dim>4</dim>
7652 </port>
7653 </output>
7654 </layer>
7655 <layer id="559" name="__module.bert.encoder.layer.4.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
7656 <input>
7657 <port id="0" precision="FP32">
7658 <dim>-1</dim>
7659 <dim>-1</dim>
7660 <dim>12</dim>
7661 <dim>32</dim>
7662 </port>
7663 <port id="1" precision="I64">
7664 <dim>4</dim>
7665 </port>
7666 </input>
7667 <output>
7668 <port id="2" precision="FP32" names="392">
7669 <dim>-1</dim>
7670 <dim>12</dim>
7671 <dim>-1</dim>
7672 <dim>32</dim>
7673 </port>
7674 </output>
7675 </layer>
7676 <layer id="560" name="__module.bert.encoder.layer.4.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
7677 <data causal="false" />
7678 <input>
7679 <port id="0" precision="FP32">
7680 <dim>-1</dim>
7681 <dim>12</dim>
7682 <dim>-1</dim>
7683 <dim>32</dim>
7684 </port>
7685 <port id="1" precision="FP32">
7686 <dim>-1</dim>
7687 <dim>12</dim>
7688 <dim>-1</dim>
7689 <dim>32</dim>
7690 </port>
7691 <port id="2" precision="FP32">
7692 <dim>-1</dim>
7693 <dim>12</dim>
7694 <dim>-1</dim>
7695 <dim>32</dim>
7696 </port>
7697 <port id="3" precision="FP32">
7698 <dim>-1</dim>
7699 <dim>1</dim>
7700 <dim>-1</dim>
7701 <dim>-1</dim>
7702 </port>
7703 </input>
7704 <output>
7705 <port id="4" precision="FP32" names="393,attn_output.17">
7706 <dim>-1</dim>
7707 <dim>12</dim>
7708 <dim>-1</dim>
7709 <dim>32</dim>
7710 </port>
7711 </output>
7712 </layer>
7713 <layer id="561" name="__module.bert.encoder.layer.4.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
7714 <data element_type="i32" shape="4" offset="12502400" size="16" />
7715 <output>
7716 <port id="0" precision="I32">
7717 <dim>4</dim>
7718 </port>
7719 </output>
7720 </layer>
7721 <layer id="562" name="__module.bert.encoder.layer.4.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
7722 <input>
7723 <port id="0" precision="FP32">
7724 <dim>-1</dim>
7725 <dim>12</dim>
7726 <dim>-1</dim>
7727 <dim>32</dim>
7728 </port>
7729 <port id="1" precision="I32">
7730 <dim>4</dim>
7731 </port>
7732 </input>
7733 <output>
7734 <port id="2" precision="FP32" names="394,attn_output.19">
7735 <dim>-1</dim>
7736 <dim>-1</dim>
7737 <dim>12</dim>
7738 <dim>32</dim>
7739 </port>
7740 </output>
7741 </layer>
7742 <layer id="563" name="Constant_90447" type="Const" version="opset1">
7743 <data element_type="i64" shape="3" offset="12502416" size="24" />
7744 <output>
7745 <port id="0" precision="I64">
7746 <dim>3</dim>
7747 </port>
7748 </output>
7749 </layer>
7750 <layer id="564" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
7751 <data special_zero="true" />
7752 <input>
7753 <port id="0" precision="FP32">
7754 <dim>-1</dim>
7755 <dim>-1</dim>
7756 <dim>12</dim>
7757 <dim>32</dim>
7758 </port>
7759 <port id="1" precision="I64">
7760 <dim>3</dim>
7761 </port>
7762 </input>
7763 <output>
7764 <port id="2" precision="FP32" names="396">
7765 <dim>-1</dim>
7766 <dim>-1</dim>
7767 <dim>384</dim>
7768 </port>
7769 </output>
7770 </layer>
7771 <layer id="565" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7772 <data element_type="f32" shape="1, 1, 384" offset="19758696" size="1536" />
7773 <output>
7774 <port id="0" precision="FP32">
7775 <dim>1</dim>
7776 <dim>1</dim>
7777 <dim>384</dim>
7778 </port>
7779 </output>
7780 </layer>
7781 <layer id="566" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7782 <data auto_broadcast="numpy" />
7783 <input>
7784 <port id="0" precision="FP32">
7785 <dim>-1</dim>
7786 <dim>-1</dim>
7787 <dim>384</dim>
7788 </port>
7789 <port id="1" precision="FP32">
7790 <dim>1</dim>
7791 <dim>1</dim>
7792 <dim>384</dim>
7793 </port>
7794 </input>
7795 <output>
7796 <port id="2" precision="FP32">
7797 <dim>-1</dim>
7798 <dim>-1</dim>
7799 <dim>384</dim>
7800 </port>
7801 </output>
7802 </layer>
7803 <layer id="567" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7804 <data element_type="f32" shape="" offset="19760232" size="4" />
7805 <output>
7806 <port id="0" precision="FP32" />
7807 </output>
7808 </layer>
7809 <layer id="568" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7810 <data element_type="f32" shape="" offset="19760236" size="4" />
7811 <output>
7812 <port id="0" precision="FP32" />
7813 </output>
7814 </layer>
7815 <layer id="569" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7816 <data element_type="f32" shape="" offset="19760232" size="4" />
7817 <output>
7818 <port id="0" precision="FP32" />
7819 </output>
7820 </layer>
7821 <layer id="570" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7822 <data element_type="f32" shape="" offset="19760236" size="4" />
7823 <output>
7824 <port id="0" precision="FP32" />
7825 </output>
7826 </layer>
7827 <layer id="571" name="__module.bert.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7828 <data levels="256" auto_broadcast="numpy" />
7829 <input>
7830 <port id="0" precision="FP32">
7831 <dim>-1</dim>
7832 <dim>-1</dim>
7833 <dim>384</dim>
7834 </port>
7835 <port id="1" precision="FP32" />
7836 <port id="2" precision="FP32" />
7837 <port id="3" precision="FP32" />
7838 <port id="4" precision="FP32" />
7839 </input>
7840 <output>
7841 <port id="5" precision="FP32">
7842 <dim>-1</dim>
7843 <dim>-1</dim>
7844 <dim>384</dim>
7845 </port>
7846 </output>
7847 </layer>
7848 <layer id="572" name="Constant_300046" type="Const" version="opset1">
7849 <data element_type="i8" shape="384, 384" offset="19760240" size="147456" />
7850 <output>
7851 <port id="0" precision="I8">
7852 <dim>384</dim>
7853 <dim>384</dim>
7854 </port>
7855 </output>
7856 </layer>
7857 <layer id="573" name="Convert_300047" type="Convert" version="opset1">
7858 <data destination_type="f32" />
7859 <input>
7860 <port id="0" precision="I8">
7861 <dim>384</dim>
7862 <dim>384</dim>
7863 </port>
7864 </input>
7865 <output>
7866 <port id="1" precision="FP32">
7867 <dim>384</dim>
7868 <dim>384</dim>
7869 </port>
7870 </output>
7871 </layer>
7872 <layer id="574" name="Constant_300048" type="Const" version="opset1">
7873 <data element_type="f32" shape="384, 1" offset="19907696" size="1536" />
7874 <output>
7875 <port id="0" precision="FP32">
7876 <dim>384</dim>
7877 <dim>1</dim>
7878 </port>
7879 </output>
7880 </layer>
7881 <layer id="575" name="__module.bert.encoder.layer.4.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7882 <data auto_broadcast="numpy" />
7883 <input>
7884 <port id="0" precision="FP32">
7885 <dim>384</dim>
7886 <dim>384</dim>
7887 </port>
7888 <port id="1" precision="FP32">
7889 <dim>384</dim>
7890 <dim>1</dim>
7891 </port>
7892 </input>
7893 <output>
7894 <port id="2" precision="FP32">
7895 <dim>384</dim>
7896 <dim>384</dim>
7897 </port>
7898 </output>
7899 </layer>
7900 <layer id="576" name="__module.bert.encoder.layer.4.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
7901 <data transpose_a="false" transpose_b="true" />
7902 <input>
7903 <port id="0" precision="FP32">
7904 <dim>-1</dim>
7905 <dim>-1</dim>
7906 <dim>384</dim>
7907 </port>
7908 <port id="1" precision="FP32">
7909 <dim>384</dim>
7910 <dim>384</dim>
7911 </port>
7912 </input>
7913 <output>
7914 <port id="2" precision="FP32">
7915 <dim>-1</dim>
7916 <dim>-1</dim>
7917 <dim>384</dim>
7918 </port>
7919 </output>
7920 </layer>
7921 <layer id="577" name="Constant_90348" type="Const" version="opset1">
7922 <data element_type="f32" shape="1, 1, 384" offset="19909232" size="1536" />
7923 <output>
7924 <port id="0" precision="FP32">
7925 <dim>1</dim>
7926 <dim>1</dim>
7927 <dim>384</dim>
7928 </port>
7929 </output>
7930 </layer>
7931 <layer id="578" name="__module.bert.encoder.layer.4.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
7932 <data auto_broadcast="numpy" />
7933 <input>
7934 <port id="0" precision="FP32">
7935 <dim>-1</dim>
7936 <dim>-1</dim>
7937 <dim>384</dim>
7938 </port>
7939 <port id="1" precision="FP32">
7940 <dim>1</dim>
7941 <dim>1</dim>
7942 <dim>384</dim>
7943 </port>
7944 </input>
7945 <output>
7946 <port id="2" precision="FP32" names="401,input.19">
7947 <dim>-1</dim>
7948 <dim>-1</dim>
7949 <dim>384</dim>
7950 </port>
7951 </output>
7952 </layer>
7953 <layer id="579" name="__module.bert.encoder.layer.4.attention.output/aten::add/Add" type="Add" version="opset1">
7954 <data auto_broadcast="numpy" />
7955 <input>
7956 <port id="0" precision="FP32">
7957 <dim>-1</dim>
7958 <dim>-1</dim>
7959 <dim>384</dim>
7960 </port>
7961 <port id="1" precision="FP32">
7962 <dim>-1</dim>
7963 <dim>-1</dim>
7964 <dim>384</dim>
7965 </port>
7966 </input>
7967 <output>
7968 <port id="2" precision="FP32" names="403">
7969 <dim>-1</dim>
7970 <dim>-1</dim>
7971 <dim>384</dim>
7972 </port>
7973 </output>
7974 </layer>
7975 <layer id="580" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
7976 <data element_type="i32" shape="1" offset="12046084" size="4" />
7977 <output>
7978 <port id="0" precision="I32">
7979 <dim>1</dim>
7980 </port>
7981 </output>
7982 </layer>
7983 <layer id="581" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
7984 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
7985 <input>
7986 <port id="0" precision="FP32">
7987 <dim>-1</dim>
7988 <dim>-1</dim>
7989 <dim>384</dim>
7990 </port>
7991 <port id="1" precision="I32">
7992 <dim>1</dim>
7993 </port>
7994 </input>
7995 <output>
7996 <port id="2" precision="FP32">
7997 <dim>-1</dim>
7998 <dim>-1</dim>
7999 <dim>384</dim>
8000 </port>
8001 </output>
8002 </layer>
8003 <layer id="582" name="Constant_90349" type="Const" version="opset1">
8004 <data element_type="f32" shape="1, 1, 384" offset="19910768" size="1536" />
8005 <output>
8006 <port id="0" precision="FP32">
8007 <dim>1</dim>
8008 <dim>1</dim>
8009 <dim>384</dim>
8010 </port>
8011 </output>
8012 </layer>
8013 <layer id="583" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
8014 <data auto_broadcast="numpy" />
8015 <input>
8016 <port id="0" precision="FP32">
8017 <dim>-1</dim>
8018 <dim>-1</dim>
8019 <dim>384</dim>
8020 </port>
8021 <port id="1" precision="FP32">
8022 <dim>1</dim>
8023 <dim>1</dim>
8024 <dim>384</dim>
8025 </port>
8026 </input>
8027 <output>
8028 <port id="2" precision="FP32">
8029 <dim>-1</dim>
8030 <dim>-1</dim>
8031 <dim>384</dim>
8032 </port>
8033 </output>
8034 </layer>
8035 <layer id="584" name="Constant_90350" type="Const" version="opset1">
8036 <data element_type="f32" shape="1, 1, 384" offset="19912304" size="1536" />
8037 <output>
8038 <port id="0" precision="FP32">
8039 <dim>1</dim>
8040 <dim>1</dim>
8041 <dim>384</dim>
8042 </port>
8043 </output>
8044 </layer>
8045 <layer id="585" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
8046 <data auto_broadcast="numpy" />
8047 <input>
8048 <port id="0" precision="FP32">
8049 <dim>-1</dim>
8050 <dim>-1</dim>
8051 <dim>384</dim>
8052 </port>
8053 <port id="1" precision="FP32">
8054 <dim>1</dim>
8055 <dim>1</dim>
8056 <dim>384</dim>
8057 </port>
8058 </input>
8059 <output>
8060 <port id="2" precision="FP32" names="407,input_tensor.9">
8061 <dim>-1</dim>
8062 <dim>-1</dim>
8063 <dim>384</dim>
8064 </port>
8065 </output>
8066 </layer>
8067 <layer id="586" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8068 <data element_type="f32" shape="1, 1, 384" offset="19913840" size="1536" />
8069 <output>
8070 <port id="0" precision="FP32">
8071 <dim>1</dim>
8072 <dim>1</dim>
8073 <dim>384</dim>
8074 </port>
8075 </output>
8076 </layer>
8077 <layer id="587" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8078 <data auto_broadcast="numpy" />
8079 <input>
8080 <port id="0" precision="FP32">
8081 <dim>-1</dim>
8082 <dim>-1</dim>
8083 <dim>384</dim>
8084 </port>
8085 <port id="1" precision="FP32">
8086 <dim>1</dim>
8087 <dim>1</dim>
8088 <dim>384</dim>
8089 </port>
8090 </input>
8091 <output>
8092 <port id="2" precision="FP32">
8093 <dim>-1</dim>
8094 <dim>-1</dim>
8095 <dim>384</dim>
8096 </port>
8097 </output>
8098 </layer>
8099 <layer id="588" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8100 <data element_type="f32" shape="" offset="19915376" size="4" />
8101 <output>
8102 <port id="0" precision="FP32" />
8103 </output>
8104 </layer>
8105 <layer id="589" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8106 <data element_type="f32" shape="" offset="19915380" size="4" />
8107 <output>
8108 <port id="0" precision="FP32" />
8109 </output>
8110 </layer>
8111 <layer id="590" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8112 <data element_type="f32" shape="" offset="19915376" size="4" />
8113 <output>
8114 <port id="0" precision="FP32" />
8115 </output>
8116 </layer>
8117 <layer id="591" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8118 <data element_type="f32" shape="" offset="19915380" size="4" />
8119 <output>
8120 <port id="0" precision="FP32" />
8121 </output>
8122 </layer>
8123 <layer id="592" name="__module.bert.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8124 <data levels="256" auto_broadcast="numpy" />
8125 <input>
8126 <port id="0" precision="FP32">
8127 <dim>-1</dim>
8128 <dim>-1</dim>
8129 <dim>384</dim>
8130 </port>
8131 <port id="1" precision="FP32" />
8132 <port id="2" precision="FP32" />
8133 <port id="3" precision="FP32" />
8134 <port id="4" precision="FP32" />
8135 </input>
8136 <output>
8137 <port id="5" precision="FP32">
8138 <dim>-1</dim>
8139 <dim>-1</dim>
8140 <dim>384</dim>
8141 </port>
8142 </output>
8143 </layer>
8144 <layer id="593" name="Constant_300050" type="Const" version="opset1">
8145 <data element_type="i8" shape="1536, 384" offset="19915384" size="589824" />
8146 <output>
8147 <port id="0" precision="I8">
8148 <dim>1536</dim>
8149 <dim>384</dim>
8150 </port>
8151 </output>
8152 </layer>
8153 <layer id="594" name="Convert_300051" type="Convert" version="opset1">
8154 <data destination_type="f32" />
8155 <input>
8156 <port id="0" precision="I8">
8157 <dim>1536</dim>
8158 <dim>384</dim>
8159 </port>
8160 </input>
8161 <output>
8162 <port id="1" precision="FP32">
8163 <dim>1536</dim>
8164 <dim>384</dim>
8165 </port>
8166 </output>
8167 </layer>
8168 <layer id="595" name="Constant_300052" type="Const" version="opset1">
8169 <data element_type="f32" shape="1536, 1" offset="20505208" size="6144" />
8170 <output>
8171 <port id="0" precision="FP32">
8172 <dim>1536</dim>
8173 <dim>1</dim>
8174 </port>
8175 </output>
8176 </layer>
8177 <layer id="596" name="__module.bert.encoder.layer.4.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8178 <data auto_broadcast="numpy" />
8179 <input>
8180 <port id="0" precision="FP32">
8181 <dim>1536</dim>
8182 <dim>384</dim>
8183 </port>
8184 <port id="1" precision="FP32">
8185 <dim>1536</dim>
8186 <dim>1</dim>
8187 </port>
8188 </input>
8189 <output>
8190 <port id="2" precision="FP32">
8191 <dim>1536</dim>
8192 <dim>384</dim>
8193 </port>
8194 </output>
8195 </layer>
8196 <layer id="597" name="__module.bert.encoder.layer.4.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8197 <data transpose_a="false" transpose_b="true" />
8198 <input>
8199 <port id="0" precision="FP32">
8200 <dim>-1</dim>
8201 <dim>-1</dim>
8202 <dim>384</dim>
8203 </port>
8204 <port id="1" precision="FP32">
8205 <dim>1536</dim>
8206 <dim>384</dim>
8207 </port>
8208 </input>
8209 <output>
8210 <port id="2" precision="FP32">
8211 <dim>-1</dim>
8212 <dim>-1</dim>
8213 <dim>1536</dim>
8214 </port>
8215 </output>
8216 </layer>
8217 <layer id="598" name="Constant_90351" type="Const" version="opset1">
8218 <data element_type="f32" shape="1, 1, 1536" offset="20511352" size="6144" />
8219 <output>
8220 <port id="0" precision="FP32">
8221 <dim>1</dim>
8222 <dim>1</dim>
8223 <dim>1536</dim>
8224 </port>
8225 </output>
8226 </layer>
8227 <layer id="599" name="__module.bert.encoder.layer.4.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
8228 <data auto_broadcast="numpy" />
8229 <input>
8230 <port id="0" precision="FP32">
8231 <dim>-1</dim>
8232 <dim>-1</dim>
8233 <dim>1536</dim>
8234 </port>
8235 <port id="1" precision="FP32">
8236 <dim>1</dim>
8237 <dim>1</dim>
8238 <dim>1536</dim>
8239 </port>
8240 </input>
8241 <output>
8242 <port id="2" precision="FP32" names="411">
8243 <dim>-1</dim>
8244 <dim>-1</dim>
8245 <dim>1536</dim>
8246 </port>
8247 </output>
8248 </layer>
8249 <layer id="600" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
8250 <data approximation_mode="ERF" />
8251 <input>
8252 <port id="0" precision="FP32">
8253 <dim>-1</dim>
8254 <dim>-1</dim>
8255 <dim>1536</dim>
8256 </port>
8257 </input>
8258 <output>
8259 <port id="1" precision="FP32" names="412">
8260 <dim>-1</dim>
8261 <dim>-1</dim>
8262 <dim>1536</dim>
8263 </port>
8264 </output>
8265 </layer>
8266 <layer id="601" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8267 <data element_type="f32" shape="1, 1, 1536" offset="20517496" size="6144" />
8268 <output>
8269 <port id="0" precision="FP32">
8270 <dim>1</dim>
8271 <dim>1</dim>
8272 <dim>1536</dim>
8273 </port>
8274 </output>
8275 </layer>
8276 <layer id="602" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8277 <data auto_broadcast="numpy" />
8278 <input>
8279 <port id="0" precision="FP32">
8280 <dim>-1</dim>
8281 <dim>-1</dim>
8282 <dim>1536</dim>
8283 </port>
8284 <port id="1" precision="FP32">
8285 <dim>1</dim>
8286 <dim>1</dim>
8287 <dim>1536</dim>
8288 </port>
8289 </input>
8290 <output>
8291 <port id="2" precision="FP32">
8292 <dim>-1</dim>
8293 <dim>-1</dim>
8294 <dim>1536</dim>
8295 </port>
8296 </output>
8297 </layer>
8298 <layer id="603" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8299 <data element_type="f32" shape="" offset="20523640" size="4" />
8300 <output>
8301 <port id="0" precision="FP32" />
8302 </output>
8303 </layer>
8304 <layer id="604" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8305 <data element_type="f32" shape="" offset="20523644" size="4" />
8306 <output>
8307 <port id="0" precision="FP32" />
8308 </output>
8309 </layer>
8310 <layer id="605" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8311 <data element_type="f32" shape="" offset="20523640" size="4" />
8312 <output>
8313 <port id="0" precision="FP32" />
8314 </output>
8315 </layer>
8316 <layer id="606" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8317 <data element_type="f32" shape="" offset="20523644" size="4" />
8318 <output>
8319 <port id="0" precision="FP32" />
8320 </output>
8321 </layer>
8322 <layer id="607" name="__module.bert.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8323 <data levels="256" auto_broadcast="numpy" />
8324 <input>
8325 <port id="0" precision="FP32">
8326 <dim>-1</dim>
8327 <dim>-1</dim>
8328 <dim>1536</dim>
8329 </port>
8330 <port id="1" precision="FP32" />
8331 <port id="2" precision="FP32" />
8332 <port id="3" precision="FP32" />
8333 <port id="4" precision="FP32" />
8334 </input>
8335 <output>
8336 <port id="5" precision="FP32">
8337 <dim>-1</dim>
8338 <dim>-1</dim>
8339 <dim>1536</dim>
8340 </port>
8341 </output>
8342 </layer>
8343 <layer id="608" name="Constant_300054" type="Const" version="opset1">
8344 <data element_type="i8" shape="384, 1536" offset="20523648" size="589824" />
8345 <output>
8346 <port id="0" precision="I8">
8347 <dim>384</dim>
8348 <dim>1536</dim>
8349 </port>
8350 </output>
8351 </layer>
8352 <layer id="609" name="Convert_300055" type="Convert" version="opset1">
8353 <data destination_type="f32" />
8354 <input>
8355 <port id="0" precision="I8">
8356 <dim>384</dim>
8357 <dim>1536</dim>
8358 </port>
8359 </input>
8360 <output>
8361 <port id="1" precision="FP32">
8362 <dim>384</dim>
8363 <dim>1536</dim>
8364 </port>
8365 </output>
8366 </layer>
8367 <layer id="610" name="Constant_300056" type="Const" version="opset1">
8368 <data element_type="f32" shape="384, 1" offset="21113472" size="1536" />
8369 <output>
8370 <port id="0" precision="FP32">
8371 <dim>384</dim>
8372 <dim>1</dim>
8373 </port>
8374 </output>
8375 </layer>
8376 <layer id="611" name="__module.bert.encoder.layer.4.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8377 <data auto_broadcast="numpy" />
8378 <input>
8379 <port id="0" precision="FP32">
8380 <dim>384</dim>
8381 <dim>1536</dim>
8382 </port>
8383 <port id="1" precision="FP32">
8384 <dim>384</dim>
8385 <dim>1</dim>
8386 </port>
8387 </input>
8388 <output>
8389 <port id="2" precision="FP32">
8390 <dim>384</dim>
8391 <dim>1536</dim>
8392 </port>
8393 </output>
8394 </layer>
8395 <layer id="612" name="__module.bert.encoder.layer.4.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8396 <data transpose_a="false" transpose_b="true" />
8397 <input>
8398 <port id="0" precision="FP32">
8399 <dim>-1</dim>
8400 <dim>-1</dim>
8401 <dim>1536</dim>
8402 </port>
8403 <port id="1" precision="FP32">
8404 <dim>384</dim>
8405 <dim>1536</dim>
8406 </port>
8407 </input>
8408 <output>
8409 <port id="2" precision="FP32">
8410 <dim>-1</dim>
8411 <dim>-1</dim>
8412 <dim>384</dim>
8413 </port>
8414 </output>
8415 </layer>
8416 <layer id="613" name="Constant_90352" type="Const" version="opset1">
8417 <data element_type="f32" shape="1, 1, 384" offset="21115008" size="1536" />
8418 <output>
8419 <port id="0" precision="FP32">
8420 <dim>1</dim>
8421 <dim>1</dim>
8422 <dim>384</dim>
8423 </port>
8424 </output>
8425 </layer>
8426 <layer id="614" name="__module.bert.encoder.layer.4.output.dense/aten::linear/Add" type="Add" version="opset1">
8427 <data auto_broadcast="numpy" />
8428 <input>
8429 <port id="0" precision="FP32">
8430 <dim>-1</dim>
8431 <dim>-1</dim>
8432 <dim>384</dim>
8433 </port>
8434 <port id="1" precision="FP32">
8435 <dim>1</dim>
8436 <dim>1</dim>
8437 <dim>384</dim>
8438 </port>
8439 </input>
8440 <output>
8441 <port id="2" precision="FP32" names="417,input.21">
8442 <dim>-1</dim>
8443 <dim>-1</dim>
8444 <dim>384</dim>
8445 </port>
8446 </output>
8447 </layer>
8448 <layer id="615" name="__module.bert.encoder.layer.4.output/aten::add/Add" type="Add" version="opset1">
8449 <data auto_broadcast="numpy" />
8450 <input>
8451 <port id="0" precision="FP32">
8452 <dim>-1</dim>
8453 <dim>-1</dim>
8454 <dim>384</dim>
8455 </port>
8456 <port id="1" precision="FP32">
8457 <dim>-1</dim>
8458 <dim>-1</dim>
8459 <dim>384</dim>
8460 </port>
8461 </input>
8462 <output>
8463 <port id="2" precision="FP32" names="419">
8464 <dim>-1</dim>
8465 <dim>-1</dim>
8466 <dim>384</dim>
8467 </port>
8468 </output>
8469 </layer>
8470 <layer id="616" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
8471 <data element_type="i32" shape="1" offset="12046084" size="4" />
8472 <output>
8473 <port id="0" precision="I32">
8474 <dim>1</dim>
8475 </port>
8476 </output>
8477 </layer>
8478 <layer id="617" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
8479 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
8480 <input>
8481 <port id="0" precision="FP32">
8482 <dim>-1</dim>
8483 <dim>-1</dim>
8484 <dim>384</dim>
8485 </port>
8486 <port id="1" precision="I32">
8487 <dim>1</dim>
8488 </port>
8489 </input>
8490 <output>
8491 <port id="2" precision="FP32">
8492 <dim>-1</dim>
8493 <dim>-1</dim>
8494 <dim>384</dim>
8495 </port>
8496 </output>
8497 </layer>
8498 <layer id="618" name="Constant_90353" type="Const" version="opset1">
8499 <data element_type="f32" shape="1, 1, 384" offset="21116544" size="1536" />
8500 <output>
8501 <port id="0" precision="FP32">
8502 <dim>1</dim>
8503 <dim>1</dim>
8504 <dim>384</dim>
8505 </port>
8506 </output>
8507 </layer>
8508 <layer id="619" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
8509 <data auto_broadcast="numpy" />
8510 <input>
8511 <port id="0" precision="FP32">
8512 <dim>-1</dim>
8513 <dim>-1</dim>
8514 <dim>384</dim>
8515 </port>
8516 <port id="1" precision="FP32">
8517 <dim>1</dim>
8518 <dim>1</dim>
8519 <dim>384</dim>
8520 </port>
8521 </input>
8522 <output>
8523 <port id="2" precision="FP32">
8524 <dim>-1</dim>
8525 <dim>-1</dim>
8526 <dim>384</dim>
8527 </port>
8528 </output>
8529 </layer>
8530 <layer id="620" name="Constant_90354" type="Const" version="opset1">
8531 <data element_type="f32" shape="1, 1, 384" offset="21118080" size="1536" />
8532 <output>
8533 <port id="0" precision="FP32">
8534 <dim>1</dim>
8535 <dim>1</dim>
8536 <dim>384</dim>
8537 </port>
8538 </output>
8539 </layer>
8540 <layer id="621" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
8541 <data auto_broadcast="numpy" />
8542 <input>
8543 <port id="0" precision="FP32">
8544 <dim>-1</dim>
8545 <dim>-1</dim>
8546 <dim>384</dim>
8547 </port>
8548 <port id="1" precision="FP32">
8549 <dim>1</dim>
8550 <dim>1</dim>
8551 <dim>384</dim>
8552 </port>
8553 </input>
8554 <output>
8555 <port id="2" precision="FP32" names="423,hidden_states.31">
8556 <dim>-1</dim>
8557 <dim>-1</dim>
8558 <dim>384</dim>
8559 </port>
8560 </output>
8561 </layer>
8562 <layer id="622" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8563 <data element_type="f32" shape="1, 1, 384" offset="21119616" size="1536" />
8564 <output>
8565 <port id="0" precision="FP32">
8566 <dim>1</dim>
8567 <dim>1</dim>
8568 <dim>384</dim>
8569 </port>
8570 </output>
8571 </layer>
8572 <layer id="623" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8573 <data auto_broadcast="numpy" />
8574 <input>
8575 <port id="0" precision="FP32">
8576 <dim>-1</dim>
8577 <dim>-1</dim>
8578 <dim>384</dim>
8579 </port>
8580 <port id="1" precision="FP32">
8581 <dim>1</dim>
8582 <dim>1</dim>
8583 <dim>384</dim>
8584 </port>
8585 </input>
8586 <output>
8587 <port id="2" precision="FP32">
8588 <dim>-1</dim>
8589 <dim>-1</dim>
8590 <dim>384</dim>
8591 </port>
8592 </output>
8593 </layer>
8594 <layer id="624" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8595 <data element_type="f32" shape="" offset="21121152" size="4" />
8596 <output>
8597 <port id="0" precision="FP32" />
8598 </output>
8599 </layer>
8600 <layer id="625" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8601 <data element_type="f32" shape="" offset="21121156" size="4" />
8602 <output>
8603 <port id="0" precision="FP32" />
8604 </output>
8605 </layer>
8606 <layer id="626" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8607 <data element_type="f32" shape="" offset="21121152" size="4" />
8608 <output>
8609 <port id="0" precision="FP32" />
8610 </output>
8611 </layer>
8612 <layer id="627" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8613 <data element_type="f32" shape="" offset="21121156" size="4" />
8614 <output>
8615 <port id="0" precision="FP32" />
8616 </output>
8617 </layer>
8618 <layer id="628" name="__module.bert.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8619 <data levels="256" auto_broadcast="numpy" />
8620 <input>
8621 <port id="0" precision="FP32">
8622 <dim>-1</dim>
8623 <dim>-1</dim>
8624 <dim>384</dim>
8625 </port>
8626 <port id="1" precision="FP32" />
8627 <port id="2" precision="FP32" />
8628 <port id="3" precision="FP32" />
8629 <port id="4" precision="FP32" />
8630 </input>
8631 <output>
8632 <port id="5" precision="FP32">
8633 <dim>-1</dim>
8634 <dim>-1</dim>
8635 <dim>384</dim>
8636 </port>
8637 </output>
8638 </layer>
8639 <layer id="629" name="Constant_300058" type="Const" version="opset1">
8640 <data element_type="i8" shape="384, 384" offset="21121160" size="147456" />
8641 <output>
8642 <port id="0" precision="I8">
8643 <dim>384</dim>
8644 <dim>384</dim>
8645 </port>
8646 </output>
8647 </layer>
8648 <layer id="630" name="Convert_300059" type="Convert" version="opset1">
8649 <data destination_type="f32" />
8650 <input>
8651 <port id="0" precision="I8">
8652 <dim>384</dim>
8653 <dim>384</dim>
8654 </port>
8655 </input>
8656 <output>
8657 <port id="1" precision="FP32">
8658 <dim>384</dim>
8659 <dim>384</dim>
8660 </port>
8661 </output>
8662 </layer>
8663 <layer id="631" name="Constant_300060" type="Const" version="opset1">
8664 <data element_type="f32" shape="384, 1" offset="21268616" size="1536" />
8665 <output>
8666 <port id="0" precision="FP32">
8667 <dim>384</dim>
8668 <dim>1</dim>
8669 </port>
8670 </output>
8671 </layer>
8672 <layer id="632" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8673 <data auto_broadcast="numpy" />
8674 <input>
8675 <port id="0" precision="FP32">
8676 <dim>384</dim>
8677 <dim>384</dim>
8678 </port>
8679 <port id="1" precision="FP32">
8680 <dim>384</dim>
8681 <dim>1</dim>
8682 </port>
8683 </input>
8684 <output>
8685 <port id="2" precision="FP32">
8686 <dim>384</dim>
8687 <dim>384</dim>
8688 </port>
8689 </output>
8690 </layer>
8691 <layer id="633" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
8692 <data transpose_a="false" transpose_b="true" />
8693 <input>
8694 <port id="0" precision="FP32">
8695 <dim>-1</dim>
8696 <dim>-1</dim>
8697 <dim>384</dim>
8698 </port>
8699 <port id="1" precision="FP32">
8700 <dim>384</dim>
8701 <dim>384</dim>
8702 </port>
8703 </input>
8704 <output>
8705 <port id="2" precision="FP32">
8706 <dim>-1</dim>
8707 <dim>-1</dim>
8708 <dim>384</dim>
8709 </port>
8710 </output>
8711 </layer>
8712 <layer id="634" name="Constant_90355" type="Const" version="opset1">
8713 <data element_type="f32" shape="1, 1, 384" offset="21270152" size="1536" />
8714 <output>
8715 <port id="0" precision="FP32">
8716 <dim>1</dim>
8717 <dim>1</dim>
8718 <dim>384</dim>
8719 </port>
8720 </output>
8721 </layer>
8722 <layer id="635" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add" type="Add" version="opset1">
8723 <data auto_broadcast="numpy" />
8724 <input>
8725 <port id="0" precision="FP32">
8726 <dim>-1</dim>
8727 <dim>-1</dim>
8728 <dim>384</dim>
8729 </port>
8730 <port id="1" precision="FP32">
8731 <dim>1</dim>
8732 <dim>1</dim>
8733 <dim>384</dim>
8734 </port>
8735 </input>
8736 <output>
8737 <port id="2" precision="FP32" names="436,x.61">
8738 <dim>-1</dim>
8739 <dim>-1</dim>
8740 <dim>384</dim>
8741 </port>
8742 </output>
8743 </layer>
8744 <layer id="636" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
8745 <data element_type="f32" shape="" offset="21271688" size="4" />
8746 <output>
8747 <port id="0" precision="FP32" />
8748 </output>
8749 </layer>
8750 <layer id="637" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
8751 <data element_type="f32" shape="" offset="21271692" size="4" />
8752 <output>
8753 <port id="0" precision="FP32" />
8754 </output>
8755 </layer>
8756 <layer id="638" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
8757 <data element_type="f32" shape="" offset="21271688" size="4" />
8758 <output>
8759 <port id="0" precision="FP32" />
8760 </output>
8761 </layer>
8762 <layer id="639" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
8763 <data element_type="f32" shape="" offset="21271692" size="4" />
8764 <output>
8765 <port id="0" precision="FP32" />
8766 </output>
8767 </layer>
8768 <layer id="640" name="__module.bert.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
8769 <data levels="256" auto_broadcast="numpy" />
8770 <input>
8771 <port id="0" precision="FP32">
8772 <dim>-1</dim>
8773 <dim>-1</dim>
8774 <dim>384</dim>
8775 </port>
8776 <port id="1" precision="FP32" />
8777 <port id="2" precision="FP32" />
8778 <port id="3" precision="FP32" />
8779 <port id="4" precision="FP32" />
8780 </input>
8781 <output>
8782 <port id="5" precision="FP32">
8783 <dim>-1</dim>
8784 <dim>-1</dim>
8785 <dim>384</dim>
8786 </port>
8787 </output>
8788 </layer>
8789 <layer id="641" name="__module.bert.encoder.layer.5.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
8790 <data element_type="i64" shape="4" offset="12201240" size="32" />
8791 <output>
8792 <port id="0" precision="I64">
8793 <dim>4</dim>
8794 </port>
8795 </output>
8796 </layer>
8797 <layer id="642" name="__module.bert.encoder.layer.5.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
8798 <data special_zero="true" />
8799 <input>
8800 <port id="0" precision="FP32">
8801 <dim>-1</dim>
8802 <dim>-1</dim>
8803 <dim>384</dim>
8804 </port>
8805 <port id="1" precision="I64">
8806 <dim>4</dim>
8807 </port>
8808 </input>
8809 <output>
8810 <port id="2" precision="FP32" names="440,x.63">
8811 <dim>-1</dim>
8812 <dim>-1</dim>
8813 <dim>12</dim>
8814 <dim>32</dim>
8815 </port>
8816 </output>
8817 </layer>
8818 <layer id="643" name="Constant_85573" type="Const" version="opset1">
8819 <data element_type="i64" shape="4" offset="12201272" size="32" />
8820 <output>
8821 <port id="0" precision="I64" names="441">
8822 <dim>4</dim>
8823 </port>
8824 </output>
8825 </layer>
8826 <layer id="644" name="__module.bert.encoder.layer.5.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
8827 <input>
8828 <port id="0" precision="FP32">
8829 <dim>-1</dim>
8830 <dim>-1</dim>
8831 <dim>12</dim>
8832 <dim>32</dim>
8833 </port>
8834 <port id="1" precision="I64">
8835 <dim>4</dim>
8836 </port>
8837 </input>
8838 <output>
8839 <port id="2" precision="FP32" names="442">
8840 <dim>-1</dim>
8841 <dim>12</dim>
8842 <dim>-1</dim>
8843 <dim>32</dim>
8844 </port>
8845 </output>
8846 </layer>
8847 <layer id="645" name="Constant_300062" type="Const" version="opset1">
8848 <data element_type="i8" shape="384, 384" offset="21271696" size="147456" />
8849 <output>
8850 <port id="0" precision="I8">
8851 <dim>384</dim>
8852 <dim>384</dim>
8853 </port>
8854 </output>
8855 </layer>
8856 <layer id="646" name="Convert_300063" type="Convert" version="opset1">
8857 <data destination_type="f32" />
8858 <input>
8859 <port id="0" precision="I8">
8860 <dim>384</dim>
8861 <dim>384</dim>
8862 </port>
8863 </input>
8864 <output>
8865 <port id="1" precision="FP32">
8866 <dim>384</dim>
8867 <dim>384</dim>
8868 </port>
8869 </output>
8870 </layer>
8871 <layer id="647" name="Constant_300064" type="Const" version="opset1">
8872 <data element_type="f32" shape="384, 1" offset="21419152" size="1536" />
8873 <output>
8874 <port id="0" precision="FP32">
8875 <dim>384</dim>
8876 <dim>1</dim>
8877 </port>
8878 </output>
8879 </layer>
8880 <layer id="648" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8881 <data auto_broadcast="numpy" />
8882 <input>
8883 <port id="0" precision="FP32">
8884 <dim>384</dim>
8885 <dim>384</dim>
8886 </port>
8887 <port id="1" precision="FP32">
8888 <dim>384</dim>
8889 <dim>1</dim>
8890 </port>
8891 </input>
8892 <output>
8893 <port id="2" precision="FP32">
8894 <dim>384</dim>
8895 <dim>384</dim>
8896 </port>
8897 </output>
8898 </layer>
8899 <layer id="649" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
8900 <data transpose_a="false" transpose_b="true" />
8901 <input>
8902 <port id="0" precision="FP32">
8903 <dim>-1</dim>
8904 <dim>-1</dim>
8905 <dim>384</dim>
8906 </port>
8907 <port id="1" precision="FP32">
8908 <dim>384</dim>
8909 <dim>384</dim>
8910 </port>
8911 </input>
8912 <output>
8913 <port id="2" precision="FP32">
8914 <dim>-1</dim>
8915 <dim>-1</dim>
8916 <dim>384</dim>
8917 </port>
8918 </output>
8919 </layer>
8920 <layer id="650" name="Constant_90356" type="Const" version="opset1">
8921 <data element_type="f32" shape="1, 1, 384" offset="21420688" size="1536" />
8922 <output>
8923 <port id="0" precision="FP32">
8924 <dim>1</dim>
8925 <dim>1</dim>
8926 <dim>384</dim>
8927 </port>
8928 </output>
8929 </layer>
8930 <layer id="651" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add" type="Add" version="opset1">
8931 <data auto_broadcast="numpy" />
8932 <input>
8933 <port id="0" precision="FP32">
8934 <dim>-1</dim>
8935 <dim>-1</dim>
8936 <dim>384</dim>
8937 </port>
8938 <port id="1" precision="FP32">
8939 <dim>1</dim>
8940 <dim>1</dim>
8941 <dim>384</dim>
8942 </port>
8943 </input>
8944 <output>
8945 <port id="2" precision="FP32" names="445,x.65">
8946 <dim>-1</dim>
8947 <dim>-1</dim>
8948 <dim>384</dim>
8949 </port>
8950 </output>
8951 </layer>
8952 <layer id="652" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
8953 <data element_type="f32" shape="" offset="21422224" size="4" />
8954 <output>
8955 <port id="0" precision="FP32" />
8956 </output>
8957 </layer>
8958 <layer id="653" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
8959 <data element_type="f32" shape="" offset="21422228" size="4" />
8960 <output>
8961 <port id="0" precision="FP32" />
8962 </output>
8963 </layer>
8964 <layer id="654" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
8965 <data element_type="f32" shape="" offset="21422224" size="4" />
8966 <output>
8967 <port id="0" precision="FP32" />
8968 </output>
8969 </layer>
8970 <layer id="655" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
8971 <data element_type="f32" shape="" offset="21422228" size="4" />
8972 <output>
8973 <port id="0" precision="FP32" />
8974 </output>
8975 </layer>
8976 <layer id="656" name="__module.bert.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
8977 <data levels="256" auto_broadcast="numpy" />
8978 <input>
8979 <port id="0" precision="FP32">
8980 <dim>-1</dim>
8981 <dim>-1</dim>
8982 <dim>384</dim>
8983 </port>
8984 <port id="1" precision="FP32" />
8985 <port id="2" precision="FP32" />
8986 <port id="3" precision="FP32" />
8987 <port id="4" precision="FP32" />
8988 </input>
8989 <output>
8990 <port id="5" precision="FP32">
8991 <dim>-1</dim>
8992 <dim>-1</dim>
8993 <dim>384</dim>
8994 </port>
8995 </output>
8996 </layer>
8997 <layer id="657" name="__module.bert.encoder.layer.5.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
8998 <data element_type="i64" shape="4" offset="12201240" size="32" />
8999 <output>
9000 <port id="0" precision="I64">
9001 <dim>4</dim>
9002 </port>
9003 </output>
9004 </layer>
9005 <layer id="658" name="__module.bert.encoder.layer.5.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
9006 <data special_zero="true" />
9007 <input>
9008 <port id="0" precision="FP32">
9009 <dim>-1</dim>
9010 <dim>-1</dim>
9011 <dim>384</dim>
9012 </port>
9013 <port id="1" precision="I64">
9014 <dim>4</dim>
9015 </port>
9016 </input>
9017 <output>
9018 <port id="2" precision="FP32" names="449,x.67">
9019 <dim>-1</dim>
9020 <dim>-1</dim>
9021 <dim>12</dim>
9022 <dim>32</dim>
9023 </port>
9024 </output>
9025 </layer>
9026 <layer id="659" name="Constant_85596" type="Const" version="opset1">
9027 <data element_type="i64" shape="4" offset="12201272" size="32" />
9028 <output>
9029 <port id="0" precision="I64" names="450">
9030 <dim>4</dim>
9031 </port>
9032 </output>
9033 </layer>
9034 <layer id="660" name="__module.bert.encoder.layer.5.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
9035 <input>
9036 <port id="0" precision="FP32">
9037 <dim>-1</dim>
9038 <dim>-1</dim>
9039 <dim>12</dim>
9040 <dim>32</dim>
9041 </port>
9042 <port id="1" precision="I64">
9043 <dim>4</dim>
9044 </port>
9045 </input>
9046 <output>
9047 <port id="2" precision="FP32" names="451">
9048 <dim>-1</dim>
9049 <dim>12</dim>
9050 <dim>-1</dim>
9051 <dim>32</dim>
9052 </port>
9053 </output>
9054 </layer>
9055 <layer id="661" name="Constant_300066" type="Const" version="opset1">
9056 <data element_type="i8" shape="384, 384" offset="21422232" size="147456" />
9057 <output>
9058 <port id="0" precision="I8">
9059 <dim>384</dim>
9060 <dim>384</dim>
9061 </port>
9062 </output>
9063 </layer>
9064 <layer id="662" name="Convert_300067" type="Convert" version="opset1">
9065 <data destination_type="f32" />
9066 <input>
9067 <port id="0" precision="I8">
9068 <dim>384</dim>
9069 <dim>384</dim>
9070 </port>
9071 </input>
9072 <output>
9073 <port id="1" precision="FP32">
9074 <dim>384</dim>
9075 <dim>384</dim>
9076 </port>
9077 </output>
9078 </layer>
9079 <layer id="663" name="Constant_300068" type="Const" version="opset1">
9080 <data element_type="f32" shape="384, 1" offset="21569688" size="1536" />
9081 <output>
9082 <port id="0" precision="FP32">
9083 <dim>384</dim>
9084 <dim>1</dim>
9085 </port>
9086 </output>
9087 </layer>
9088 <layer id="664" name="__module.bert.encoder.layer.5.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9089 <data auto_broadcast="numpy" />
9090 <input>
9091 <port id="0" precision="FP32">
9092 <dim>384</dim>
9093 <dim>384</dim>
9094 </port>
9095 <port id="1" precision="FP32">
9096 <dim>384</dim>
9097 <dim>1</dim>
9098 </port>
9099 </input>
9100 <output>
9101 <port id="2" precision="FP32">
9102 <dim>384</dim>
9103 <dim>384</dim>
9104 </port>
9105 </output>
9106 </layer>
9107 <layer id="665" name="__module.bert.encoder.layer.5.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
9108 <data transpose_a="false" transpose_b="true" />
9109 <input>
9110 <port id="0" precision="FP32">
9111 <dim>-1</dim>
9112 <dim>-1</dim>
9113 <dim>384</dim>
9114 </port>
9115 <port id="1" precision="FP32">
9116 <dim>384</dim>
9117 <dim>384</dim>
9118 </port>
9119 </input>
9120 <output>
9121 <port id="2" precision="FP32">
9122 <dim>-1</dim>
9123 <dim>-1</dim>
9124 <dim>384</dim>
9125 </port>
9126 </output>
9127 </layer>
9128 <layer id="666" name="Constant_90357" type="Const" version="opset1">
9129 <data element_type="f32" shape="1, 1, 384" offset="21571224" size="1536" />
9130 <output>
9131 <port id="0" precision="FP32">
9132 <dim>1</dim>
9133 <dim>1</dim>
9134 <dim>384</dim>
9135 </port>
9136 </output>
9137 </layer>
9138 <layer id="667" name="__module.bert.encoder.layer.5.attention.self.value/aten::linear/Add" type="Add" version="opset1">
9139 <data auto_broadcast="numpy" />
9140 <input>
9141 <port id="0" precision="FP32">
9142 <dim>-1</dim>
9143 <dim>-1</dim>
9144 <dim>384</dim>
9145 </port>
9146 <port id="1" precision="FP32">
9147 <dim>1</dim>
9148 <dim>1</dim>
9149 <dim>384</dim>
9150 </port>
9151 </input>
9152 <output>
9153 <port id="2" precision="FP32" names="454,x.69">
9154 <dim>-1</dim>
9155 <dim>-1</dim>
9156 <dim>384</dim>
9157 </port>
9158 </output>
9159 </layer>
9160 <layer id="668" name="__module.bert.encoder.layer.5.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
9161 <data element_type="i64" shape="4" offset="12201240" size="32" />
9162 <output>
9163 <port id="0" precision="I64">
9164 <dim>4</dim>
9165 </port>
9166 </output>
9167 </layer>
9168 <layer id="669" name="__module.bert.encoder.layer.5.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
9169 <data special_zero="true" />
9170 <input>
9171 <port id="0" precision="FP32">
9172 <dim>-1</dim>
9173 <dim>-1</dim>
9174 <dim>384</dim>
9175 </port>
9176 <port id="1" precision="I64">
9177 <dim>4</dim>
9178 </port>
9179 </input>
9180 <output>
9181 <port id="2" precision="FP32" names="458,x">
9182 <dim>-1</dim>
9183 <dim>-1</dim>
9184 <dim>12</dim>
9185 <dim>32</dim>
9186 </port>
9187 </output>
9188 </layer>
9189 <layer id="670" name="Constant_85619" type="Const" version="opset1">
9190 <data element_type="i64" shape="4" offset="12201272" size="32" />
9191 <output>
9192 <port id="0" precision="I64" names="459">
9193 <dim>4</dim>
9194 </port>
9195 </output>
9196 </layer>
9197 <layer id="671" name="__module.bert.encoder.layer.5.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
9198 <input>
9199 <port id="0" precision="FP32">
9200 <dim>-1</dim>
9201 <dim>-1</dim>
9202 <dim>12</dim>
9203 <dim>32</dim>
9204 </port>
9205 <port id="1" precision="I64">
9206 <dim>4</dim>
9207 </port>
9208 </input>
9209 <output>
9210 <port id="2" precision="FP32" names="460">
9211 <dim>-1</dim>
9212 <dim>12</dim>
9213 <dim>-1</dim>
9214 <dim>32</dim>
9215 </port>
9216 </output>
9217 </layer>
9218 <layer id="672" name="__module.bert.encoder.layer.5.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
9219 <data causal="false" />
9220 <input>
9221 <port id="0" precision="FP32">
9222 <dim>-1</dim>
9223 <dim>12</dim>
9224 <dim>-1</dim>
9225 <dim>32</dim>
9226 </port>
9227 <port id="1" precision="FP32">
9228 <dim>-1</dim>
9229 <dim>12</dim>
9230 <dim>-1</dim>
9231 <dim>32</dim>
9232 </port>
9233 <port id="2" precision="FP32">
9234 <dim>-1</dim>
9235 <dim>12</dim>
9236 <dim>-1</dim>
9237 <dim>32</dim>
9238 </port>
9239 <port id="3" precision="FP32">
9240 <dim>-1</dim>
9241 <dim>1</dim>
9242 <dim>-1</dim>
9243 <dim>-1</dim>
9244 </port>
9245 </input>
9246 <output>
9247 <port id="4" precision="FP32" names="461,attn_output.21">
9248 <dim>-1</dim>
9249 <dim>12</dim>
9250 <dim>-1</dim>
9251 <dim>32</dim>
9252 </port>
9253 </output>
9254 </layer>
9255 <layer id="673" name="__module.bert.encoder.layer.5.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
9256 <data element_type="i32" shape="4" offset="12502400" size="16" />
9257 <output>
9258 <port id="0" precision="I32">
9259 <dim>4</dim>
9260 </port>
9261 </output>
9262 </layer>
9263 <layer id="674" name="__module.bert.encoder.layer.5.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
9264 <input>
9265 <port id="0" precision="FP32">
9266 <dim>-1</dim>
9267 <dim>12</dim>
9268 <dim>-1</dim>
9269 <dim>32</dim>
9270 </port>
9271 <port id="1" precision="I32">
9272 <dim>4</dim>
9273 </port>
9274 </input>
9275 <output>
9276 <port id="2" precision="FP32" names="462,attn_output">
9277 <dim>-1</dim>
9278 <dim>-1</dim>
9279 <dim>12</dim>
9280 <dim>32</dim>
9281 </port>
9282 </output>
9283 </layer>
9284 <layer id="675" name="Constant_90448" type="Const" version="opset1">
9285 <data element_type="i64" shape="3" offset="12502416" size="24" />
9286 <output>
9287 <port id="0" precision="I64">
9288 <dim>3</dim>
9289 </port>
9290 </output>
9291 </layer>
9292 <layer id="676" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
9293 <data special_zero="true" />
9294 <input>
9295 <port id="0" precision="FP32">
9296 <dim>-1</dim>
9297 <dim>-1</dim>
9298 <dim>12</dim>
9299 <dim>32</dim>
9300 </port>
9301 <port id="1" precision="I64">
9302 <dim>3</dim>
9303 </port>
9304 </input>
9305 <output>
9306 <port id="2" precision="FP32" names="464">
9307 <dim>-1</dim>
9308 <dim>-1</dim>
9309 <dim>384</dim>
9310 </port>
9311 </output>
9312 </layer>
9313 <layer id="677" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9314 <data element_type="f32" shape="1, 1, 384" offset="21572760" size="1536" />
9315 <output>
9316 <port id="0" precision="FP32">
9317 <dim>1</dim>
9318 <dim>1</dim>
9319 <dim>384</dim>
9320 </port>
9321 </output>
9322 </layer>
9323 <layer id="678" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9324 <data auto_broadcast="numpy" />
9325 <input>
9326 <port id="0" precision="FP32">
9327 <dim>-1</dim>
9328 <dim>-1</dim>
9329 <dim>384</dim>
9330 </port>
9331 <port id="1" precision="FP32">
9332 <dim>1</dim>
9333 <dim>1</dim>
9334 <dim>384</dim>
9335 </port>
9336 </input>
9337 <output>
9338 <port id="2" precision="FP32">
9339 <dim>-1</dim>
9340 <dim>-1</dim>
9341 <dim>384</dim>
9342 </port>
9343 </output>
9344 </layer>
9345 <layer id="679" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9346 <data element_type="f32" shape="" offset="21574296" size="4" />
9347 <output>
9348 <port id="0" precision="FP32" />
9349 </output>
9350 </layer>
9351 <layer id="680" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9352 <data element_type="f32" shape="" offset="21574300" size="4" />
9353 <output>
9354 <port id="0" precision="FP32" />
9355 </output>
9356 </layer>
9357 <layer id="681" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9358 <data element_type="f32" shape="" offset="21574296" size="4" />
9359 <output>
9360 <port id="0" precision="FP32" />
9361 </output>
9362 </layer>
9363 <layer id="682" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9364 <data element_type="f32" shape="" offset="21574300" size="4" />
9365 <output>
9366 <port id="0" precision="FP32" />
9367 </output>
9368 </layer>
9369 <layer id="683" name="__module.bert.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9370 <data levels="256" auto_broadcast="numpy" />
9371 <input>
9372 <port id="0" precision="FP32">
9373 <dim>-1</dim>
9374 <dim>-1</dim>
9375 <dim>384</dim>
9376 </port>
9377 <port id="1" precision="FP32" />
9378 <port id="2" precision="FP32" />
9379 <port id="3" precision="FP32" />
9380 <port id="4" precision="FP32" />
9381 </input>
9382 <output>
9383 <port id="5" precision="FP32">
9384 <dim>-1</dim>
9385 <dim>-1</dim>
9386 <dim>384</dim>
9387 </port>
9388 </output>
9389 </layer>
9390 <layer id="684" name="Constant_300070" type="Const" version="opset1">
9391 <data element_type="i8" shape="384, 384" offset="21574304" size="147456" />
9392 <output>
9393 <port id="0" precision="I8">
9394 <dim>384</dim>
9395 <dim>384</dim>
9396 </port>
9397 </output>
9398 </layer>
9399 <layer id="685" name="Convert_300071" type="Convert" version="opset1">
9400 <data destination_type="f32" />
9401 <input>
9402 <port id="0" precision="I8">
9403 <dim>384</dim>
9404 <dim>384</dim>
9405 </port>
9406 </input>
9407 <output>
9408 <port id="1" precision="FP32">
9409 <dim>384</dim>
9410 <dim>384</dim>
9411 </port>
9412 </output>
9413 </layer>
9414 <layer id="686" name="Constant_300072" type="Const" version="opset1">
9415 <data element_type="f32" shape="384, 1" offset="21721760" size="1536" />
9416 <output>
9417 <port id="0" precision="FP32">
9418 <dim>384</dim>
9419 <dim>1</dim>
9420 </port>
9421 </output>
9422 </layer>
9423 <layer id="687" name="__module.bert.encoder.layer.5.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9424 <data auto_broadcast="numpy" />
9425 <input>
9426 <port id="0" precision="FP32">
9427 <dim>384</dim>
9428 <dim>384</dim>
9429 </port>
9430 <port id="1" precision="FP32">
9431 <dim>384</dim>
9432 <dim>1</dim>
9433 </port>
9434 </input>
9435 <output>
9436 <port id="2" precision="FP32">
9437 <dim>384</dim>
9438 <dim>384</dim>
9439 </port>
9440 </output>
9441 </layer>
9442 <layer id="688" name="__module.bert.encoder.layer.5.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
9443 <data transpose_a="false" transpose_b="true" />
9444 <input>
9445 <port id="0" precision="FP32">
9446 <dim>-1</dim>
9447 <dim>-1</dim>
9448 <dim>384</dim>
9449 </port>
9450 <port id="1" precision="FP32">
9451 <dim>384</dim>
9452 <dim>384</dim>
9453 </port>
9454 </input>
9455 <output>
9456 <port id="2" precision="FP32">
9457 <dim>-1</dim>
9458 <dim>-1</dim>
9459 <dim>384</dim>
9460 </port>
9461 </output>
9462 </layer>
9463 <layer id="689" name="Constant_90358" type="Const" version="opset1">
9464 <data element_type="f32" shape="1, 1, 384" offset="21723296" size="1536" />
9465 <output>
9466 <port id="0" precision="FP32">
9467 <dim>1</dim>
9468 <dim>1</dim>
9469 <dim>384</dim>
9470 </port>
9471 </output>
9472 </layer>
9473 <layer id="690" name="__module.bert.encoder.layer.5.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
9474 <data auto_broadcast="numpy" />
9475 <input>
9476 <port id="0" precision="FP32">
9477 <dim>-1</dim>
9478 <dim>-1</dim>
9479 <dim>384</dim>
9480 </port>
9481 <port id="1" precision="FP32">
9482 <dim>1</dim>
9483 <dim>1</dim>
9484 <dim>384</dim>
9485 </port>
9486 </input>
9487 <output>
9488 <port id="2" precision="FP32" names="469,input.23">
9489 <dim>-1</dim>
9490 <dim>-1</dim>
9491 <dim>384</dim>
9492 </port>
9493 </output>
9494 </layer>
9495 <layer id="691" name="__module.bert.encoder.layer.5.attention.output/aten::add/Add" type="Add" version="opset1">
9496 <data auto_broadcast="numpy" />
9497 <input>
9498 <port id="0" precision="FP32">
9499 <dim>-1</dim>
9500 <dim>-1</dim>
9501 <dim>384</dim>
9502 </port>
9503 <port id="1" precision="FP32">
9504 <dim>-1</dim>
9505 <dim>-1</dim>
9506 <dim>384</dim>
9507 </port>
9508 </input>
9509 <output>
9510 <port id="2" precision="FP32" names="471">
9511 <dim>-1</dim>
9512 <dim>-1</dim>
9513 <dim>384</dim>
9514 </port>
9515 </output>
9516 </layer>
9517 <layer id="692" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
9518 <data element_type="i32" shape="1" offset="12046084" size="4" />
9519 <output>
9520 <port id="0" precision="I32">
9521 <dim>1</dim>
9522 </port>
9523 </output>
9524 </layer>
9525 <layer id="693" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
9526 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
9527 <input>
9528 <port id="0" precision="FP32">
9529 <dim>-1</dim>
9530 <dim>-1</dim>
9531 <dim>384</dim>
9532 </port>
9533 <port id="1" precision="I32">
9534 <dim>1</dim>
9535 </port>
9536 </input>
9537 <output>
9538 <port id="2" precision="FP32">
9539 <dim>-1</dim>
9540 <dim>-1</dim>
9541 <dim>384</dim>
9542 </port>
9543 </output>
9544 </layer>
9545 <layer id="694" name="Constant_90359" type="Const" version="opset1">
9546 <data element_type="f32" shape="1, 1, 384" offset="21724832" size="1536" />
9547 <output>
9548 <port id="0" precision="FP32">
9549 <dim>1</dim>
9550 <dim>1</dim>
9551 <dim>384</dim>
9552 </port>
9553 </output>
9554 </layer>
9555 <layer id="695" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
9556 <data auto_broadcast="numpy" />
9557 <input>
9558 <port id="0" precision="FP32">
9559 <dim>-1</dim>
9560 <dim>-1</dim>
9561 <dim>384</dim>
9562 </port>
9563 <port id="1" precision="FP32">
9564 <dim>1</dim>
9565 <dim>1</dim>
9566 <dim>384</dim>
9567 </port>
9568 </input>
9569 <output>
9570 <port id="2" precision="FP32">
9571 <dim>-1</dim>
9572 <dim>-1</dim>
9573 <dim>384</dim>
9574 </port>
9575 </output>
9576 </layer>
9577 <layer id="696" name="Constant_90360" type="Const" version="opset1">
9578 <data element_type="f32" shape="1, 1, 384" offset="21726368" size="1536" />
9579 <output>
9580 <port id="0" precision="FP32">
9581 <dim>1</dim>
9582 <dim>1</dim>
9583 <dim>384</dim>
9584 </port>
9585 </output>
9586 </layer>
9587 <layer id="697" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
9588 <data auto_broadcast="numpy" />
9589 <input>
9590 <port id="0" precision="FP32">
9591 <dim>-1</dim>
9592 <dim>-1</dim>
9593 <dim>384</dim>
9594 </port>
9595 <port id="1" precision="FP32">
9596 <dim>1</dim>
9597 <dim>1</dim>
9598 <dim>384</dim>
9599 </port>
9600 </input>
9601 <output>
9602 <port id="2" precision="FP32" names="475,input_tensor">
9603 <dim>-1</dim>
9604 <dim>-1</dim>
9605 <dim>384</dim>
9606 </port>
9607 </output>
9608 </layer>
9609 <layer id="698" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9610 <data element_type="f32" shape="1, 1, 384" offset="21727904" size="1536" />
9611 <output>
9612 <port id="0" precision="FP32">
9613 <dim>1</dim>
9614 <dim>1</dim>
9615 <dim>384</dim>
9616 </port>
9617 </output>
9618 </layer>
9619 <layer id="699" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9620 <data auto_broadcast="numpy" />
9621 <input>
9622 <port id="0" precision="FP32">
9623 <dim>-1</dim>
9624 <dim>-1</dim>
9625 <dim>384</dim>
9626 </port>
9627 <port id="1" precision="FP32">
9628 <dim>1</dim>
9629 <dim>1</dim>
9630 <dim>384</dim>
9631 </port>
9632 </input>
9633 <output>
9634 <port id="2" precision="FP32">
9635 <dim>-1</dim>
9636 <dim>-1</dim>
9637 <dim>384</dim>
9638 </port>
9639 </output>
9640 </layer>
9641 <layer id="700" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9642 <data element_type="f32" shape="" offset="21729440" size="4" />
9643 <output>
9644 <port id="0" precision="FP32" />
9645 </output>
9646 </layer>
9647 <layer id="701" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9648 <data element_type="f32" shape="" offset="21729444" size="4" />
9649 <output>
9650 <port id="0" precision="FP32" />
9651 </output>
9652 </layer>
9653 <layer id="702" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9654 <data element_type="f32" shape="" offset="21729440" size="4" />
9655 <output>
9656 <port id="0" precision="FP32" />
9657 </output>
9658 </layer>
9659 <layer id="703" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9660 <data element_type="f32" shape="" offset="21729444" size="4" />
9661 <output>
9662 <port id="0" precision="FP32" />
9663 </output>
9664 </layer>
9665 <layer id="704" name="__module.bert.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9666 <data levels="256" auto_broadcast="numpy" />
9667 <input>
9668 <port id="0" precision="FP32">
9669 <dim>-1</dim>
9670 <dim>-1</dim>
9671 <dim>384</dim>
9672 </port>
9673 <port id="1" precision="FP32" />
9674 <port id="2" precision="FP32" />
9675 <port id="3" precision="FP32" />
9676 <port id="4" precision="FP32" />
9677 </input>
9678 <output>
9679 <port id="5" precision="FP32">
9680 <dim>-1</dim>
9681 <dim>-1</dim>
9682 <dim>384</dim>
9683 </port>
9684 </output>
9685 </layer>
9686 <layer id="705" name="Constant_300074" type="Const" version="opset1">
9687 <data element_type="i8" shape="1536, 384" offset="21729448" size="589824" />
9688 <output>
9689 <port id="0" precision="I8">
9690 <dim>1536</dim>
9691 <dim>384</dim>
9692 </port>
9693 </output>
9694 </layer>
9695 <layer id="706" name="Convert_300075" type="Convert" version="opset1">
9696 <data destination_type="f32" />
9697 <input>
9698 <port id="0" precision="I8">
9699 <dim>1536</dim>
9700 <dim>384</dim>
9701 </port>
9702 </input>
9703 <output>
9704 <port id="1" precision="FP32">
9705 <dim>1536</dim>
9706 <dim>384</dim>
9707 </port>
9708 </output>
9709 </layer>
9710 <layer id="707" name="Constant_300076" type="Const" version="opset1">
9711 <data element_type="f32" shape="1536, 1" offset="22319272" size="6144" />
9712 <output>
9713 <port id="0" precision="FP32">
9714 <dim>1536</dim>
9715 <dim>1</dim>
9716 </port>
9717 </output>
9718 </layer>
9719 <layer id="708" name="__module.bert.encoder.layer.5.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9720 <data auto_broadcast="numpy" />
9721 <input>
9722 <port id="0" precision="FP32">
9723 <dim>1536</dim>
9724 <dim>384</dim>
9725 </port>
9726 <port id="1" precision="FP32">
9727 <dim>1536</dim>
9728 <dim>1</dim>
9729 </port>
9730 </input>
9731 <output>
9732 <port id="2" precision="FP32">
9733 <dim>1536</dim>
9734 <dim>384</dim>
9735 </port>
9736 </output>
9737 </layer>
9738 <layer id="709" name="__module.bert.encoder.layer.5.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
9739 <data transpose_a="false" transpose_b="true" />
9740 <input>
9741 <port id="0" precision="FP32">
9742 <dim>-1</dim>
9743 <dim>-1</dim>
9744 <dim>384</dim>
9745 </port>
9746 <port id="1" precision="FP32">
9747 <dim>1536</dim>
9748 <dim>384</dim>
9749 </port>
9750 </input>
9751 <output>
9752 <port id="2" precision="FP32">
9753 <dim>-1</dim>
9754 <dim>-1</dim>
9755 <dim>1536</dim>
9756 </port>
9757 </output>
9758 </layer>
9759 <layer id="710" name="Constant_90361" type="Const" version="opset1">
9760 <data element_type="f32" shape="1, 1, 1536" offset="22325416" size="6144" />
9761 <output>
9762 <port id="0" precision="FP32">
9763 <dim>1</dim>
9764 <dim>1</dim>
9765 <dim>1536</dim>
9766 </port>
9767 </output>
9768 </layer>
9769 <layer id="711" name="__module.bert.encoder.layer.5.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
9770 <data auto_broadcast="numpy" />
9771 <input>
9772 <port id="0" precision="FP32">
9773 <dim>-1</dim>
9774 <dim>-1</dim>
9775 <dim>1536</dim>
9776 </port>
9777 <port id="1" precision="FP32">
9778 <dim>1</dim>
9779 <dim>1</dim>
9780 <dim>1536</dim>
9781 </port>
9782 </input>
9783 <output>
9784 <port id="2" precision="FP32" names="479">
9785 <dim>-1</dim>
9786 <dim>-1</dim>
9787 <dim>1536</dim>
9788 </port>
9789 </output>
9790 </layer>
9791 <layer id="712" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
9792 <data approximation_mode="ERF" />
9793 <input>
9794 <port id="0" precision="FP32">
9795 <dim>-1</dim>
9796 <dim>-1</dim>
9797 <dim>1536</dim>
9798 </port>
9799 </input>
9800 <output>
9801 <port id="1" precision="FP32" names="480">
9802 <dim>-1</dim>
9803 <dim>-1</dim>
9804 <dim>1536</dim>
9805 </port>
9806 </output>
9807 </layer>
9808 <layer id="713" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9809 <data element_type="f32" shape="1, 1, 1536" offset="22331560" size="6144" />
9810 <output>
9811 <port id="0" precision="FP32">
9812 <dim>1</dim>
9813 <dim>1</dim>
9814 <dim>1536</dim>
9815 </port>
9816 </output>
9817 </layer>
9818 <layer id="714" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9819 <data auto_broadcast="numpy" />
9820 <input>
9821 <port id="0" precision="FP32">
9822 <dim>-1</dim>
9823 <dim>-1</dim>
9824 <dim>1536</dim>
9825 </port>
9826 <port id="1" precision="FP32">
9827 <dim>1</dim>
9828 <dim>1</dim>
9829 <dim>1536</dim>
9830 </port>
9831 </input>
9832 <output>
9833 <port id="2" precision="FP32">
9834 <dim>-1</dim>
9835 <dim>-1</dim>
9836 <dim>1536</dim>
9837 </port>
9838 </output>
9839 </layer>
9840 <layer id="715" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9841 <data element_type="f32" shape="" offset="22337704" size="4" />
9842 <output>
9843 <port id="0" precision="FP32" />
9844 </output>
9845 </layer>
9846 <layer id="716" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9847 <data element_type="f32" shape="" offset="22337708" size="4" />
9848 <output>
9849 <port id="0" precision="FP32" />
9850 </output>
9851 </layer>
9852 <layer id="717" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9853 <data element_type="f32" shape="" offset="22337704" size="4" />
9854 <output>
9855 <port id="0" precision="FP32" />
9856 </output>
9857 </layer>
9858 <layer id="718" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9859 <data element_type="f32" shape="" offset="22337708" size="4" />
9860 <output>
9861 <port id="0" precision="FP32" />
9862 </output>
9863 </layer>
9864 <layer id="719" name="__module.bert.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9865 <data levels="256" auto_broadcast="numpy" />
9866 <input>
9867 <port id="0" precision="FP32">
9868 <dim>-1</dim>
9869 <dim>-1</dim>
9870 <dim>1536</dim>
9871 </port>
9872 <port id="1" precision="FP32" />
9873 <port id="2" precision="FP32" />
9874 <port id="3" precision="FP32" />
9875 <port id="4" precision="FP32" />
9876 </input>
9877 <output>
9878 <port id="5" precision="FP32">
9879 <dim>-1</dim>
9880 <dim>-1</dim>
9881 <dim>1536</dim>
9882 </port>
9883 </output>
9884 </layer>
9885 <layer id="720" name="Constant_300078" type="Const" version="opset1">
9886 <data element_type="i8" shape="384, 1536" offset="22337712" size="589824" />
9887 <output>
9888 <port id="0" precision="I8">
9889 <dim>384</dim>
9890 <dim>1536</dim>
9891 </port>
9892 </output>
9893 </layer>
9894 <layer id="721" name="Convert_300079" type="Convert" version="opset1">
9895 <data destination_type="f32" />
9896 <input>
9897 <port id="0" precision="I8">
9898 <dim>384</dim>
9899 <dim>1536</dim>
9900 </port>
9901 </input>
9902 <output>
9903 <port id="1" precision="FP32">
9904 <dim>384</dim>
9905 <dim>1536</dim>
9906 </port>
9907 </output>
9908 </layer>
9909 <layer id="722" name="Constant_300080" type="Const" version="opset1">
9910 <data element_type="f32" shape="384, 1" offset="22927536" size="1536" />
9911 <output>
9912 <port id="0" precision="FP32">
9913 <dim>384</dim>
9914 <dim>1</dim>
9915 </port>
9916 </output>
9917 </layer>
9918 <layer id="723" name="__module.bert.encoder.layer.5.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9919 <data auto_broadcast="numpy" />
9920 <input>
9921 <port id="0" precision="FP32">
9922 <dim>384</dim>
9923 <dim>1536</dim>
9924 </port>
9925 <port id="1" precision="FP32">
9926 <dim>384</dim>
9927 <dim>1</dim>
9928 </port>
9929 </input>
9930 <output>
9931 <port id="2" precision="FP32">
9932 <dim>384</dim>
9933 <dim>1536</dim>
9934 </port>
9935 </output>
9936 </layer>
9937 <layer id="724" name="__module.bert.encoder.layer.5.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
9938 <data transpose_a="false" transpose_b="true" />
9939 <input>
9940 <port id="0" precision="FP32">
9941 <dim>-1</dim>
9942 <dim>-1</dim>
9943 <dim>1536</dim>
9944 </port>
9945 <port id="1" precision="FP32">
9946 <dim>384</dim>
9947 <dim>1536</dim>
9948 </port>
9949 </input>
9950 <output>
9951 <port id="2" precision="FP32">
9952 <dim>-1</dim>
9953 <dim>-1</dim>
9954 <dim>384</dim>
9955 </port>
9956 </output>
9957 </layer>
9958 <layer id="725" name="Constant_90362" type="Const" version="opset1">
9959 <data element_type="f32" shape="1, 1, 384" offset="22929072" size="1536" />
9960 <output>
9961 <port id="0" precision="FP32">
9962 <dim>1</dim>
9963 <dim>1</dim>
9964 <dim>384</dim>
9965 </port>
9966 </output>
9967 </layer>
9968 <layer id="726" name="__module.bert.encoder.layer.5.output.dense/aten::linear/Add" type="Add" version="opset1">
9969 <data auto_broadcast="numpy" />
9970 <input>
9971 <port id="0" precision="FP32">
9972 <dim>-1</dim>
9973 <dim>-1</dim>
9974 <dim>384</dim>
9975 </port>
9976 <port id="1" precision="FP32">
9977 <dim>1</dim>
9978 <dim>1</dim>
9979 <dim>384</dim>
9980 </port>
9981 </input>
9982 <output>
9983 <port id="2" precision="FP32" names="485,input.25">
9984 <dim>-1</dim>
9985 <dim>-1</dim>
9986 <dim>384</dim>
9987 </port>
9988 </output>
9989 </layer>
9990 <layer id="727" name="__module.bert.encoder.layer.5.output/aten::add/Add" type="Add" version="opset1">
9991 <data auto_broadcast="numpy" />
9992 <input>
9993 <port id="0" precision="FP32">
9994 <dim>-1</dim>
9995 <dim>-1</dim>
9996 <dim>384</dim>
9997 </port>
9998 <port id="1" precision="FP32">
9999 <dim>-1</dim>
10000 <dim>-1</dim>
10001 <dim>384</dim>
10002 </port>
10003 </input>
10004 <output>
10005 <port id="2" precision="FP32" names="487">
10006 <dim>-1</dim>
10007 <dim>-1</dim>
10008 <dim>384</dim>
10009 </port>
10010 </output>
10011 </layer>
10012 <layer id="728" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
10013 <data element_type="i32" shape="1" offset="12046084" size="4" />
10014 <output>
10015 <port id="0" precision="I32">
10016 <dim>1</dim>
10017 </port>
10018 </output>
10019 </layer>
10020 <layer id="729" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
10021 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
10022 <input>
10023 <port id="0" precision="FP32">
10024 <dim>-1</dim>
10025 <dim>-1</dim>
10026 <dim>384</dim>
10027 </port>
10028 <port id="1" precision="I32">
10029 <dim>1</dim>
10030 </port>
10031 </input>
10032 <output>
10033 <port id="2" precision="FP32">
10034 <dim>-1</dim>
10035 <dim>-1</dim>
10036 <dim>384</dim>
10037 </port>
10038 </output>
10039 </layer>
10040 <layer id="730" name="Constant_90363" type="Const" version="opset1">
10041 <data element_type="f32" shape="1, 1, 384" offset="22930608" size="1536" />
10042 <output>
10043 <port id="0" precision="FP32">
10044 <dim>1</dim>
10045 <dim>1</dim>
10046 <dim>384</dim>
10047 </port>
10048 </output>
10049 </layer>
10050 <layer id="731" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
10051 <data auto_broadcast="numpy" />
10052 <input>
10053 <port id="0" precision="FP32">
10054 <dim>-1</dim>
10055 <dim>-1</dim>
10056 <dim>384</dim>
10057 </port>
10058 <port id="1" precision="FP32">
10059 <dim>1</dim>
10060 <dim>1</dim>
10061 <dim>384</dim>
10062 </port>
10063 </input>
10064 <output>
10065 <port id="2" precision="FP32">
10066 <dim>-1</dim>
10067 <dim>-1</dim>
10068 <dim>384</dim>
10069 </port>
10070 </output>
10071 </layer>
10072 <layer id="732" name="Constant_90364" type="Const" version="opset1">
10073 <data element_type="f32" shape="1, 1, 384" offset="22932144" size="1536" />
10074 <output>
10075 <port id="0" precision="FP32">
10076 <dim>1</dim>
10077 <dim>1</dim>
10078 <dim>384</dim>
10079 </port>
10080 </output>
10081 </layer>
10082 <layer id="733" name="__module.bert.encoder.layer.5.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
10083 <data auto_broadcast="numpy" />
10084 <input>
10085 <port id="0" precision="FP32">
10086 <dim>-1</dim>
10087 <dim>-1</dim>
10088 <dim>384</dim>
10089 </port>
10090 <port id="1" precision="FP32">
10091 <dim>1</dim>
10092 <dim>1</dim>
10093 <dim>384</dim>
10094 </port>
10095 </input>
10096 <output>
10097 <port id="2" precision="FP32" names="491,493,hidden_states">
10098 <dim>-1</dim>
10099 <dim>-1</dim>
10100 <dim>384</dim>
10101 </port>
10102 </output>
10103 </layer>
10104 <layer id="734" name="23" type="Const" version="opset1">
10105 <data element_type="i64" shape="" offset="12046068" size="8" />
10106 <output>
10107 <port id="0" precision="I64" names="23" />
10108 </output>
10109 </layer>
10110 <layer id="735" name="__module.bert.pooler/aten::select/Gather" type="Gather" version="opset8">
10111 <data batch_dims="0" />
10112 <input>
10113 <port id="0" precision="FP32">
10114 <dim>-1</dim>
10115 <dim>-1</dim>
10116 <dim>384</dim>
10117 </port>
10118 <port id="1" precision="I64" />
10119 <port id="2" precision="I64" />
10120 </input>
10121 <output>
10122 <port id="3" precision="FP32" names="494">
10123 <dim>-1</dim>
10124 <dim>384</dim>
10125 </port>
10126 </output>
10127 </layer>
10128 <layer id="736" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
10129 <data element_type="f32" shape="1, 384" offset="22933680" size="1536" />
10130 <output>
10131 <port id="0" precision="FP32">
10132 <dim>1</dim>
10133 <dim>384</dim>
10134 </port>
10135 </output>
10136 </layer>
10137 <layer id="737" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
10138 <data auto_broadcast="numpy" />
10139 <input>
10140 <port id="0" precision="FP32">
10141 <dim>-1</dim>
10142 <dim>384</dim>
10143 </port>
10144 <port id="1" precision="FP32">
10145 <dim>1</dim>
10146 <dim>384</dim>
10147 </port>
10148 </input>
10149 <output>
10150 <port id="2" precision="FP32">
10151 <dim>-1</dim>
10152 <dim>384</dim>
10153 </port>
10154 </output>
10155 </layer>
10156 <layer id="738" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
10157 <data element_type="f32" shape="" offset="22935216" size="4" />
10158 <output>
10159 <port id="0" precision="FP32" />
10160 </output>
10161 </layer>
10162 <layer id="739" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
10163 <data element_type="f32" shape="" offset="22935220" size="4" />
10164 <output>
10165 <port id="0" precision="FP32" />
10166 </output>
10167 </layer>
10168 <layer id="740" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
10169 <data element_type="f32" shape="" offset="22935216" size="4" />
10170 <output>
10171 <port id="0" precision="FP32" />
10172 </output>
10173 </layer>
10174 <layer id="741" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
10175 <data element_type="f32" shape="" offset="22935220" size="4" />
10176 <output>
10177 <port id="0" precision="FP32" />
10178 </output>
10179 </layer>
10180 <layer id="742" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
10181 <data levels="256" auto_broadcast="numpy" />
10182 <input>
10183 <port id="0" precision="FP32">
10184 <dim>-1</dim>
10185 <dim>384</dim>
10186 </port>
10187 <port id="1" precision="FP32" />
10188 <port id="2" precision="FP32" />
10189 <port id="3" precision="FP32" />
10190 <port id="4" precision="FP32" />
10191 </input>
10192 <output>
10193 <port id="5" precision="FP32">
10194 <dim>-1</dim>
10195 <dim>384</dim>
10196 </port>
10197 </output>
10198 </layer>
10199 <layer id="743" name="Constant_300082" type="Const" version="opset1">
10200 <data element_type="i8" shape="384, 384" offset="22935224" size="147456" />
10201 <output>
10202 <port id="0" precision="I8">
10203 <dim>384</dim>
10204 <dim>384</dim>
10205 </port>
10206 </output>
10207 </layer>
10208 <layer id="744" name="Convert_300083" type="Convert" version="opset1">
10209 <data destination_type="f32" />
10210 <input>
10211 <port id="0" precision="I8">
10212 <dim>384</dim>
10213 <dim>384</dim>
10214 </port>
10215 </input>
10216 <output>
10217 <port id="1" precision="FP32">
10218 <dim>384</dim>
10219 <dim>384</dim>
10220 </port>
10221 </output>
10222 </layer>
10223 <layer id="745" name="Constant_300084" type="Const" version="opset1">
10224 <data element_type="f32" shape="384, 1" offset="23082680" size="1536" />
10225 <output>
10226 <port id="0" precision="FP32">
10227 <dim>384</dim>
10228 <dim>1</dim>
10229 </port>
10230 </output>
10231 </layer>
10232 <layer id="746" name="__module.bert.pooler.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
10233 <data auto_broadcast="numpy" />
10234 <input>
10235 <port id="0" precision="FP32">
10236 <dim>384</dim>
10237 <dim>384</dim>
10238 </port>
10239 <port id="1" precision="FP32">
10240 <dim>384</dim>
10241 <dim>1</dim>
10242 </port>
10243 </input>
10244 <output>
10245 <port id="2" precision="FP32">
10246 <dim>384</dim>
10247 <dim>384</dim>
10248 </port>
10249 </output>
10250 </layer>
10251 <layer id="747" name="__module.bert.pooler.dense/aten::linear/MatMul" type="MatMul" version="opset1">
10252 <data transpose_a="false" transpose_b="true" />
10253 <input>
10254 <port id="0" precision="FP32">
10255 <dim>-1</dim>
10256 <dim>384</dim>
10257 </port>
10258 <port id="1" precision="FP32">
10259 <dim>384</dim>
10260 <dim>384</dim>
10261 </port>
10262 </input>
10263 <output>
10264 <port id="2" precision="FP32">
10265 <dim>-1</dim>
10266 <dim>384</dim>
10267 </port>
10268 </output>
10269 </layer>
10270 <layer id="748" name="Constant_90365" type="Const" version="opset1">
10271 <data element_type="f32" shape="1, 384" offset="23084216" size="1536" />
10272 <output>
10273 <port id="0" precision="FP32">
10274 <dim>1</dim>
10275 <dim>384</dim>
10276 </port>
10277 </output>
10278 </layer>
10279 <layer id="749" name="__module.bert.pooler.dense/aten::linear/Add" type="Add" version="opset1">
10280 <data auto_broadcast="numpy" />
10281 <input>
10282 <port id="0" precision="FP32">
10283 <dim>-1</dim>
10284 <dim>384</dim>
10285 </port>
10286 <port id="1" precision="FP32">
10287 <dim>1</dim>
10288 <dim>384</dim>
10289 </port>
10290 </input>
10291 <output>
10292 <port id="2" precision="FP32" names="497">
10293 <dim>-1</dim>
10294 <dim>384</dim>
10295 </port>
10296 </output>
10297 </layer>
10298 <layer id="750" name="__module.bert.pooler.activation/aten::tanh/Tanh" type="Tanh" version="opset1">
10299 <input>
10300 <port id="0" precision="FP32">
10301 <dim>-1</dim>
10302 <dim>384</dim>
10303 </port>
10304 </input>
10305 <output>
10306 <port id="1" precision="FP32" names="498,input">
10307 <dim>-1</dim>
10308 <dim>384</dim>
10309 </port>
10310 </output>
10311 </layer>
10312 <layer id="751" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
10313 <data element_type="f32" shape="1, 384" offset="23085752" size="1536" />
10314 <output>
10315 <port id="0" precision="FP32">
10316 <dim>1</dim>
10317 <dim>384</dim>
10318 </port>
10319 </output>
10320 </layer>
10321 <layer id="752" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
10322 <data auto_broadcast="numpy" />
10323 <input>
10324 <port id="0" precision="FP32">
10325 <dim>-1</dim>
10326 <dim>384</dim>
10327 </port>
10328 <port id="1" precision="FP32">
10329 <dim>1</dim>
10330 <dim>384</dim>
10331 </port>
10332 </input>
10333 <output>
10334 <port id="2" precision="FP32">
10335 <dim>-1</dim>
10336 <dim>384</dim>
10337 </port>
10338 </output>
10339 </layer>
10340 <layer id="753" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
10341 <data element_type="f32" shape="" offset="23087288" size="4" />
10342 <output>
10343 <port id="0" precision="FP32" />
10344 </output>
10345 </layer>
10346 <layer id="754" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
10347 <data element_type="f32" shape="" offset="23087292" size="4" />
10348 <output>
10349 <port id="0" precision="FP32" />
10350 </output>
10351 </layer>
10352 <layer id="755" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
10353 <data element_type="f32" shape="" offset="23087288" size="4" />
10354 <output>
10355 <port id="0" precision="FP32" />
10356 </output>
10357 </layer>
10358 <layer id="756" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
10359 <data element_type="f32" shape="" offset="23087292" size="4" />
10360 <output>
10361 <port id="0" precision="FP32" />
10362 </output>
10363 </layer>
10364 <layer id="757" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
10365 <data levels="256" auto_broadcast="numpy" />
10366 <input>
10367 <port id="0" precision="FP32">
10368 <dim>-1</dim>
10369 <dim>384</dim>
10370 </port>
10371 <port id="1" precision="FP32" />
10372 <port id="2" precision="FP32" />
10373 <port id="3" precision="FP32" />
10374 <port id="4" precision="FP32" />
10375 </input>
10376 <output>
10377 <port id="5" precision="FP32">
10378 <dim>-1</dim>
10379 <dim>384</dim>
10380 </port>
10381 </output>
10382 </layer>
10383 <layer id="758" name="Constant_300086" type="Const" version="opset1">
10384 <data element_type="i8" shape="1, 384" offset="23087296" size="384" />
10385 <output>
10386 <port id="0" precision="I8">
10387 <dim>1</dim>
10388 <dim>384</dim>
10389 </port>
10390 </output>
10391 </layer>
10392 <layer id="759" name="Convert_300087" type="Convert" version="opset1">
10393 <data destination_type="f32" />
10394 <input>
10395 <port id="0" precision="I8">
10396 <dim>1</dim>
10397 <dim>384</dim>
10398 </port>
10399 </input>
10400 <output>
10401 <port id="1" precision="FP32">
10402 <dim>1</dim>
10403 <dim>384</dim>
10404 </port>
10405 </output>
10406 </layer>
10407 <layer id="760" name="Constant_300088" type="Const" version="opset1">
10408 <data element_type="f32" shape="1, 1" offset="23087680" size="4" />
10409 <output>
10410 <port id="0" precision="FP32">
10411 <dim>1</dim>
10412 <dim>1</dim>
10413 </port>
10414 </output>
10415 </layer>
10416 <layer id="761" name="__module.classifier/aten::linear/Add/fq_weights_1" type="Multiply" version="opset1">
10417 <data auto_broadcast="numpy" />
10418 <input>
10419 <port id="0" precision="FP32">
10420 <dim>1</dim>
10421 <dim>384</dim>
10422 </port>
10423 <port id="1" precision="FP32">
10424 <dim>1</dim>
10425 <dim>1</dim>
10426 </port>
10427 </input>
10428 <output>
10429 <port id="2" precision="FP32">
10430 <dim>1</dim>
10431 <dim>384</dim>
10432 </port>
10433 </output>
10434 </layer>
10435 <layer id="762" name="__module.classifier/aten::linear/Add" type="MatMul" version="opset1">
10436 <data transpose_a="false" transpose_b="true" />
10437 <input>
10438 <port id="0" precision="FP32">
10439 <dim>-1</dim>
10440 <dim>384</dim>
10441 </port>
10442 <port id="1" precision="FP32">
10443 <dim>1</dim>
10444 <dim>384</dim>
10445 </port>
10446 </input>
10447 <output>
10448 <port id="2" precision="FP32" names="logits">
10449 <dim>-1</dim>
10450 <dim>1</dim>
10451 </port>
10452 </output>
10453 </layer>
10454 <layer id="763" name="Result_86857" type="Result" version="opset1">
10455 <input>
10456 <port id="0" precision="FP32">
10457 <dim>-1</dim>
10458 <dim>1</dim>
10459 </port>
10460 </input>
10461 </layer>
10462 </layers>
10463 <edges>
10464 <edge from-layer="0" from-port="0" to-layer="14" to-port="0" />
10465 <edge from-layer="1" from-port="0" to-layer="93" to-port="0" />
10466 <edge from-layer="2" from-port="0" to-layer="7" to-port="0" />
10467 <edge from-layer="2" from-port="0" to-layer="24" to-port="0" />
10468 <edge from-layer="3" from-port="0" to-layer="4" to-port="0" />
10469 <edge from-layer="4" from-port="1" to-layer="6" to-port="0" />
10470 <edge from-layer="5" from-port="0" to-layer="6" to-port="1" />
10471 <edge from-layer="6" from-port="2" to-layer="9" to-port="0" />
10472 <edge from-layer="7" from-port="1" to-layer="9" to-port="1" />
10473 <edge from-layer="8" from-port="0" to-layer="9" to-port="2" />
10474 <edge from-layer="9" from-port="3" to-layer="17" to-port="0" />
10475 <edge from-layer="10" from-port="0" to-layer="11" to-port="0" />
10476 <edge from-layer="11" from-port="1" to-layer="13" to-port="0" />
10477 <edge from-layer="12" from-port="0" to-layer="13" to-port="1" />
10478 <edge from-layer="13" from-port="2" to-layer="16" to-port="0" />
10479 <edge from-layer="14" from-port="1" to-layer="16" to-port="1" />
10480 <edge from-layer="15" from-port="0" to-layer="16" to-port="2" />
10481 <edge from-layer="16" from-port="3" to-layer="17" to-port="1" />
10482 <edge from-layer="17" from-port="2" to-layer="34" to-port="0" />
10483 <edge from-layer="18" from-port="0" to-layer="19" to-port="0" />
10484 <edge from-layer="19" from-port="1" to-layer="21" to-port="0" />
10485 <edge from-layer="20" from-port="0" to-layer="21" to-port="1" />
10486 <edge from-layer="21" from-port="2" to-layer="33" to-port="0" />
10487 <edge from-layer="22" from-port="0" to-layer="30" to-port="0" />
10488 <edge from-layer="23" from-port="0" to-layer="30" to-port="1" />
10489 <edge from-layer="24" from-port="1" to-layer="102" to-port="0" />
10490 <edge from-layer="24" from-port="1" to-layer="27" to-port="0" />
10491 <edge from-layer="24" from-port="1" to-layer="98" to-port="0" />
10492 <edge from-layer="25" from-port="0" to-layer="27" to-port="1" />
10493 <edge from-layer="26" from-port="0" to-layer="27" to-port="2" />
10494 <edge from-layer="27" from-port="3" to-layer="30" to-port="2" />
10495 <edge from-layer="28" from-port="0" to-layer="30" to-port="3" />
10496 <edge from-layer="29" from-port="0" to-layer="30" to-port="4" />
10497 <edge from-layer="30" from-port="5" to-layer="31" to-port="0" />
10498 <edge from-layer="31" from-port="1" to-layer="33" to-port="1" />
10499 <edge from-layer="32" from-port="0" to-layer="33" to-port="2" />
10500 <edge from-layer="33" from-port="3" to-layer="34" to-port="1" />
10501 <edge from-layer="34" from-port="2" to-layer="36" to-port="0" />
10502 <edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
10503 <edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
10504 <edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
10505 <edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
10506 <edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
10507 <edge from-layer="40" from-port="2" to-layer="42" to-port="0" />
10508 <edge from-layer="40" from-port="2" to-layer="131" to-port="1" />
10509 <edge from-layer="41" from-port="0" to-layer="42" to-port="1" />
10510 <edge from-layer="42" from-port="2" to-layer="47" to-port="0" />
10511 <edge from-layer="43" from-port="0" to-layer="47" to-port="1" />
10512 <edge from-layer="44" from-port="0" to-layer="47" to-port="2" />
10513 <edge from-layer="45" from-port="0" to-layer="47" to-port="3" />
10514 <edge from-layer="46" from-port="0" to-layer="47" to-port="4" />
10515 <edge from-layer="47" from-port="5" to-layer="52" to-port="0" />
10516 <edge from-layer="47" from-port="5" to-layer="68" to-port="0" />
10517 <edge from-layer="47" from-port="5" to-layer="84" to-port="0" />
10518 <edge from-layer="48" from-port="0" to-layer="49" to-port="0" />
10519 <edge from-layer="49" from-port="1" to-layer="51" to-port="0" />
10520 <edge from-layer="50" from-port="0" to-layer="51" to-port="1" />
10521 <edge from-layer="51" from-port="2" to-layer="52" to-port="1" />
10522 <edge from-layer="52" from-port="2" to-layer="54" to-port="0" />
10523 <edge from-layer="53" from-port="0" to-layer="54" to-port="1" />
10524 <edge from-layer="54" from-port="2" to-layer="59" to-port="0" />
10525 <edge from-layer="55" from-port="0" to-layer="59" to-port="1" />
10526 <edge from-layer="56" from-port="0" to-layer="59" to-port="2" />
10527 <edge from-layer="57" from-port="0" to-layer="59" to-port="3" />
10528 <edge from-layer="58" from-port="0" to-layer="59" to-port="4" />
10529 <edge from-layer="59" from-port="5" to-layer="61" to-port="0" />
10530 <edge from-layer="60" from-port="0" to-layer="61" to-port="1" />
10531 <edge from-layer="61" from-port="2" to-layer="63" to-port="0" />
10532 <edge from-layer="62" from-port="0" to-layer="63" to-port="1" />
10533 <edge from-layer="63" from-port="2" to-layer="112" to-port="0" />
10534 <edge from-layer="64" from-port="0" to-layer="65" to-port="0" />
10535 <edge from-layer="65" from-port="1" to-layer="67" to-port="0" />
10536 <edge from-layer="66" from-port="0" to-layer="67" to-port="1" />
10537 <edge from-layer="67" from-port="2" to-layer="68" to-port="1" />
10538 <edge from-layer="68" from-port="2" to-layer="70" to-port="0" />
10539 <edge from-layer="69" from-port="0" to-layer="70" to-port="1" />
10540 <edge from-layer="70" from-port="2" to-layer="75" to-port="0" />
10541 <edge from-layer="71" from-port="0" to-layer="75" to-port="1" />
10542 <edge from-layer="72" from-port="0" to-layer="75" to-port="2" />
10543 <edge from-layer="73" from-port="0" to-layer="75" to-port="3" />
10544 <edge from-layer="74" from-port="0" to-layer="75" to-port="4" />
10545 <edge from-layer="75" from-port="5" to-layer="77" to-port="0" />
10546 <edge from-layer="76" from-port="0" to-layer="77" to-port="1" />
10547 <edge from-layer="77" from-port="2" to-layer="79" to-port="0" />
10548 <edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
10549 <edge from-layer="79" from-port="2" to-layer="112" to-port="1" />
10550 <edge from-layer="80" from-port="0" to-layer="81" to-port="0" />
10551 <edge from-layer="81" from-port="1" to-layer="83" to-port="0" />
10552 <edge from-layer="82" from-port="0" to-layer="83" to-port="1" />
10553 <edge from-layer="83" from-port="2" to-layer="84" to-port="1" />
10554 <edge from-layer="84" from-port="2" to-layer="86" to-port="0" />
10555 <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
10556 <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
10557 <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
10558 <edge from-layer="88" from-port="2" to-layer="90" to-port="0" />
10559 <edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
10560 <edge from-layer="90" from-port="2" to-layer="112" to-port="2" />
10561 <edge from-layer="91" from-port="0" to-layer="108" to-port="0" />
10562 <edge from-layer="92" from-port="0" to-layer="735" to-port="2" />
10563 <edge from-layer="92" from-port="0" to-layer="93" to-port="1" />
10564 <edge from-layer="93" from-port="2" to-layer="95" to-port="0" />
10565 <edge from-layer="94" from-port="0" to-layer="95" to-port="1" />
10566 <edge from-layer="95" from-port="2" to-layer="104" to-port="0" />
10567 <edge from-layer="96" from-port="0" to-layer="98" to-port="1" />
10568 <edge from-layer="97" from-port="0" to-layer="98" to-port="2" />
10569 <edge from-layer="98" from-port="3" to-layer="103" to-port="0" />
10570 <edge from-layer="99" from-port="0" to-layer="103" to-port="1" />
10571 <edge from-layer="100" from-port="0" to-layer="102" to-port="1" />
10572 <edge from-layer="101" from-port="0" to-layer="102" to-port="2" />
10573 <edge from-layer="102" from-port="3" to-layer="103" to-port="2" />
10574 <edge from-layer="103" from-port="3" to-layer="104" to-port="1" />
10575 <edge from-layer="104" from-port="2" to-layer="105" to-port="0" />
10576 <edge from-layer="105" from-port="1" to-layer="107" to-port="0" />
10577 <edge from-layer="106" from-port="0" to-layer="107" to-port="1" />
10578 <edge from-layer="107" from-port="2" to-layer="108" to-port="1" />
10579 <edge from-layer="108" from-port="2" to-layer="111" to-port="2" />
10580 <edge from-layer="108" from-port="2" to-layer="109" to-port="0" />
10581 <edge from-layer="109" from-port="1" to-layer="111" to-port="0" />
10582 <edge from-layer="110" from-port="0" to-layer="111" to-port="1" />
10583 <edge from-layer="111" from-port="3" to-layer="112" to-port="3" />
10584 <edge from-layer="111" from-port="3" to-layer="560" to-port="3" />
10585 <edge from-layer="111" from-port="3" to-layer="672" to-port="3" />
10586 <edge from-layer="111" from-port="3" to-layer="224" to-port="3" />
10587 <edge from-layer="111" from-port="3" to-layer="336" to-port="3" />
10588 <edge from-layer="111" from-port="3" to-layer="448" to-port="3" />
10589 <edge from-layer="112" from-port="4" to-layer="114" to-port="0" />
10590 <edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
10591 <edge from-layer="114" from-port="2" to-layer="116" to-port="0" />
10592 <edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
10593 <edge from-layer="116" from-port="2" to-layer="118" to-port="0" />
10594 <edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
10595 <edge from-layer="118" from-port="2" to-layer="123" to-port="0" />
10596 <edge from-layer="119" from-port="0" to-layer="123" to-port="1" />
10597 <edge from-layer="120" from-port="0" to-layer="123" to-port="2" />
10598 <edge from-layer="121" from-port="0" to-layer="123" to-port="3" />
10599 <edge from-layer="122" from-port="0" to-layer="123" to-port="4" />
10600 <edge from-layer="123" from-port="5" to-layer="128" to-port="0" />
10601 <edge from-layer="124" from-port="0" to-layer="125" to-port="0" />
10602 <edge from-layer="125" from-port="1" to-layer="127" to-port="0" />
10603 <edge from-layer="126" from-port="0" to-layer="127" to-port="1" />
10604 <edge from-layer="127" from-port="2" to-layer="128" to-port="1" />
10605 <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
10606 <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
10607 <edge from-layer="130" from-port="2" to-layer="131" to-port="0" />
10608 <edge from-layer="131" from-port="2" to-layer="133" to-port="0" />
10609 <edge from-layer="132" from-port="0" to-layer="133" to-port="1" />
10610 <edge from-layer="133" from-port="2" to-layer="135" to-port="0" />
10611 <edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
10612 <edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
10613 <edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
10614 <edge from-layer="137" from-port="2" to-layer="167" to-port="1" />
10615 <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
10616 <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
10617 <edge from-layer="139" from-port="2" to-layer="144" to-port="0" />
10618 <edge from-layer="140" from-port="0" to-layer="144" to-port="1" />
10619 <edge from-layer="141" from-port="0" to-layer="144" to-port="2" />
10620 <edge from-layer="142" from-port="0" to-layer="144" to-port="3" />
10621 <edge from-layer="143" from-port="0" to-layer="144" to-port="4" />
10622 <edge from-layer="144" from-port="5" to-layer="149" to-port="0" />
10623 <edge from-layer="145" from-port="0" to-layer="146" to-port="0" />
10624 <edge from-layer="146" from-port="1" to-layer="148" to-port="0" />
10625 <edge from-layer="147" from-port="0" to-layer="148" to-port="1" />
10626 <edge from-layer="148" from-port="2" to-layer="149" to-port="1" />
10627 <edge from-layer="149" from-port="2" to-layer="151" to-port="0" />
10628 <edge from-layer="150" from-port="0" to-layer="151" to-port="1" />
10629 <edge from-layer="151" from-port="2" to-layer="152" to-port="0" />
10630 <edge from-layer="152" from-port="1" to-layer="154" to-port="0" />
10631 <edge from-layer="153" from-port="0" to-layer="154" to-port="1" />
10632 <edge from-layer="154" from-port="2" to-layer="159" to-port="0" />
10633 <edge from-layer="155" from-port="0" to-layer="159" to-port="1" />
10634 <edge from-layer="156" from-port="0" to-layer="159" to-port="2" />
10635 <edge from-layer="157" from-port="0" to-layer="159" to-port="3" />
10636 <edge from-layer="158" from-port="0" to-layer="159" to-port="4" />
10637 <edge from-layer="159" from-port="5" to-layer="164" to-port="0" />
10638 <edge from-layer="160" from-port="0" to-layer="161" to-port="0" />
10639 <edge from-layer="161" from-port="1" to-layer="163" to-port="0" />
10640 <edge from-layer="162" from-port="0" to-layer="163" to-port="1" />
10641 <edge from-layer="163" from-port="2" to-layer="164" to-port="1" />
10642 <edge from-layer="164" from-port="2" to-layer="166" to-port="0" />
10643 <edge from-layer="165" from-port="0" to-layer="166" to-port="1" />
10644 <edge from-layer="166" from-port="2" to-layer="167" to-port="0" />
10645 <edge from-layer="167" from-port="2" to-layer="169" to-port="0" />
10646 <edge from-layer="168" from-port="0" to-layer="169" to-port="1" />
10647 <edge from-layer="169" from-port="2" to-layer="171" to-port="0" />
10648 <edge from-layer="170" from-port="0" to-layer="171" to-port="1" />
10649 <edge from-layer="171" from-port="2" to-layer="173" to-port="0" />
10650 <edge from-layer="172" from-port="0" to-layer="173" to-port="1" />
10651 <edge from-layer="173" from-port="2" to-layer="175" to-port="0" />
10652 <edge from-layer="173" from-port="2" to-layer="243" to-port="1" />
10653 <edge from-layer="174" from-port="0" to-layer="175" to-port="1" />
10654 <edge from-layer="175" from-port="2" to-layer="180" to-port="0" />
10655 <edge from-layer="176" from-port="0" to-layer="180" to-port="1" />
10656 <edge from-layer="177" from-port="0" to-layer="180" to-port="2" />
10657 <edge from-layer="178" from-port="0" to-layer="180" to-port="3" />
10658 <edge from-layer="179" from-port="0" to-layer="180" to-port="4" />
10659 <edge from-layer="180" from-port="5" to-layer="185" to-port="0" />
10660 <edge from-layer="180" from-port="5" to-layer="201" to-port="0" />
10661 <edge from-layer="180" from-port="5" to-layer="217" to-port="0" />
10662 <edge from-layer="181" from-port="0" to-layer="182" to-port="0" />
10663 <edge from-layer="182" from-port="1" to-layer="184" to-port="0" />
10664 <edge from-layer="183" from-port="0" to-layer="184" to-port="1" />
10665 <edge from-layer="184" from-port="2" to-layer="185" to-port="1" />
10666 <edge from-layer="185" from-port="2" to-layer="187" to-port="0" />
10667 <edge from-layer="186" from-port="0" to-layer="187" to-port="1" />
10668 <edge from-layer="187" from-port="2" to-layer="192" to-port="0" />
10669 <edge from-layer="188" from-port="0" to-layer="192" to-port="1" />
10670 <edge from-layer="189" from-port="0" to-layer="192" to-port="2" />
10671 <edge from-layer="190" from-port="0" to-layer="192" to-port="3" />
10672 <edge from-layer="191" from-port="0" to-layer="192" to-port="4" />
10673 <edge from-layer="192" from-port="5" to-layer="194" to-port="0" />
10674 <edge from-layer="193" from-port="0" to-layer="194" to-port="1" />
10675 <edge from-layer="194" from-port="2" to-layer="196" to-port="0" />
10676 <edge from-layer="195" from-port="0" to-layer="196" to-port="1" />
10677 <edge from-layer="196" from-port="2" to-layer="224" to-port="0" />
10678 <edge from-layer="197" from-port="0" to-layer="198" to-port="0" />
10679 <edge from-layer="198" from-port="1" to-layer="200" to-port="0" />
10680 <edge from-layer="199" from-port="0" to-layer="200" to-port="1" />
10681 <edge from-layer="200" from-port="2" to-layer="201" to-port="1" />
10682 <edge from-layer="201" from-port="2" to-layer="203" to-port="0" />
10683 <edge from-layer="202" from-port="0" to-layer="203" to-port="1" />
10684 <edge from-layer="203" from-port="2" to-layer="208" to-port="0" />
10685 <edge from-layer="204" from-port="0" to-layer="208" to-port="1" />
10686 <edge from-layer="205" from-port="0" to-layer="208" to-port="2" />
10687 <edge from-layer="206" from-port="0" to-layer="208" to-port="3" />
10688 <edge from-layer="207" from-port="0" to-layer="208" to-port="4" />
10689 <edge from-layer="208" from-port="5" to-layer="210" to-port="0" />
10690 <edge from-layer="209" from-port="0" to-layer="210" to-port="1" />
10691 <edge from-layer="210" from-port="2" to-layer="212" to-port="0" />
10692 <edge from-layer="211" from-port="0" to-layer="212" to-port="1" />
10693 <edge from-layer="212" from-port="2" to-layer="224" to-port="1" />
10694 <edge from-layer="213" from-port="0" to-layer="214" to-port="0" />
10695 <edge from-layer="214" from-port="1" to-layer="216" to-port="0" />
10696 <edge from-layer="215" from-port="0" to-layer="216" to-port="1" />
10697 <edge from-layer="216" from-port="2" to-layer="217" to-port="1" />
10698 <edge from-layer="217" from-port="2" to-layer="219" to-port="0" />
10699 <edge from-layer="218" from-port="0" to-layer="219" to-port="1" />
10700 <edge from-layer="219" from-port="2" to-layer="221" to-port="0" />
10701 <edge from-layer="220" from-port="0" to-layer="221" to-port="1" />
10702 <edge from-layer="221" from-port="2" to-layer="223" to-port="0" />
10703 <edge from-layer="222" from-port="0" to-layer="223" to-port="1" />
10704 <edge from-layer="223" from-port="2" to-layer="224" to-port="2" />
10705 <edge from-layer="224" from-port="4" to-layer="226" to-port="0" />
10706 <edge from-layer="225" from-port="0" to-layer="226" to-port="1" />
10707 <edge from-layer="226" from-port="2" to-layer="228" to-port="0" />
10708 <edge from-layer="227" from-port="0" to-layer="228" to-port="1" />
10709 <edge from-layer="228" from-port="2" to-layer="230" to-port="0" />
10710 <edge from-layer="229" from-port="0" to-layer="230" to-port="1" />
10711 <edge from-layer="230" from-port="2" to-layer="235" to-port="0" />
10712 <edge from-layer="231" from-port="0" to-layer="235" to-port="1" />
10713 <edge from-layer="232" from-port="0" to-layer="235" to-port="2" />
10714 <edge from-layer="233" from-port="0" to-layer="235" to-port="3" />
10715 <edge from-layer="234" from-port="0" to-layer="235" to-port="4" />
10716 <edge from-layer="235" from-port="5" to-layer="240" to-port="0" />
10717 <edge from-layer="236" from-port="0" to-layer="237" to-port="0" />
10718 <edge from-layer="237" from-port="1" to-layer="239" to-port="0" />
10719 <edge from-layer="238" from-port="0" to-layer="239" to-port="1" />
10720 <edge from-layer="239" from-port="2" to-layer="240" to-port="1" />
10721 <edge from-layer="240" from-port="2" to-layer="242" to-port="0" />
10722 <edge from-layer="241" from-port="0" to-layer="242" to-port="1" />
10723 <edge from-layer="242" from-port="2" to-layer="243" to-port="0" />
10724 <edge from-layer="243" from-port="2" to-layer="245" to-port="0" />
10725 <edge from-layer="244" from-port="0" to-layer="245" to-port="1" />
10726 <edge from-layer="245" from-port="2" to-layer="247" to-port="0" />
10727 <edge from-layer="246" from-port="0" to-layer="247" to-port="1" />
10728 <edge from-layer="247" from-port="2" to-layer="249" to-port="0" />
10729 <edge from-layer="248" from-port="0" to-layer="249" to-port="1" />
10730 <edge from-layer="249" from-port="2" to-layer="279" to-port="1" />
10731 <edge from-layer="249" from-port="2" to-layer="251" to-port="0" />
10732 <edge from-layer="250" from-port="0" to-layer="251" to-port="1" />
10733 <edge from-layer="251" from-port="2" to-layer="256" to-port="0" />
10734 <edge from-layer="252" from-port="0" to-layer="256" to-port="1" />
10735 <edge from-layer="253" from-port="0" to-layer="256" to-port="2" />
10736 <edge from-layer="254" from-port="0" to-layer="256" to-port="3" />
10737 <edge from-layer="255" from-port="0" to-layer="256" to-port="4" />
10738 <edge from-layer="256" from-port="5" to-layer="261" to-port="0" />
10739 <edge from-layer="257" from-port="0" to-layer="258" to-port="0" />
10740 <edge from-layer="258" from-port="1" to-layer="260" to-port="0" />
10741 <edge from-layer="259" from-port="0" to-layer="260" to-port="1" />
10742 <edge from-layer="260" from-port="2" to-layer="261" to-port="1" />
10743 <edge from-layer="261" from-port="2" to-layer="263" to-port="0" />
10744 <edge from-layer="262" from-port="0" to-layer="263" to-port="1" />
10745 <edge from-layer="263" from-port="2" to-layer="264" to-port="0" />
10746 <edge from-layer="264" from-port="1" to-layer="266" to-port="0" />
10747 <edge from-layer="265" from-port="0" to-layer="266" to-port="1" />
10748 <edge from-layer="266" from-port="2" to-layer="271" to-port="0" />
10749 <edge from-layer="267" from-port="0" to-layer="271" to-port="1" />
10750 <edge from-layer="268" from-port="0" to-layer="271" to-port="2" />
10751 <edge from-layer="269" from-port="0" to-layer="271" to-port="3" />
10752 <edge from-layer="270" from-port="0" to-layer="271" to-port="4" />
10753 <edge from-layer="271" from-port="5" to-layer="276" to-port="0" />
10754 <edge from-layer="272" from-port="0" to-layer="273" to-port="0" />
10755 <edge from-layer="273" from-port="1" to-layer="275" to-port="0" />
10756 <edge from-layer="274" from-port="0" to-layer="275" to-port="1" />
10757 <edge from-layer="275" from-port="2" to-layer="276" to-port="1" />
10758 <edge from-layer="276" from-port="2" to-layer="278" to-port="0" />
10759 <edge from-layer="277" from-port="0" to-layer="278" to-port="1" />
10760 <edge from-layer="278" from-port="2" to-layer="279" to-port="0" />
10761 <edge from-layer="279" from-port="2" to-layer="281" to-port="0" />
10762 <edge from-layer="280" from-port="0" to-layer="281" to-port="1" />
10763 <edge from-layer="281" from-port="2" to-layer="283" to-port="0" />
10764 <edge from-layer="282" from-port="0" to-layer="283" to-port="1" />
10765 <edge from-layer="283" from-port="2" to-layer="285" to-port="0" />
10766 <edge from-layer="284" from-port="0" to-layer="285" to-port="1" />
10767 <edge from-layer="285" from-port="2" to-layer="287" to-port="0" />
10768 <edge from-layer="285" from-port="2" to-layer="355" to-port="1" />
10769 <edge from-layer="286" from-port="0" to-layer="287" to-port="1" />
10770 <edge from-layer="287" from-port="2" to-layer="292" to-port="0" />
10771 <edge from-layer="288" from-port="0" to-layer="292" to-port="1" />
10772 <edge from-layer="289" from-port="0" to-layer="292" to-port="2" />
10773 <edge from-layer="290" from-port="0" to-layer="292" to-port="3" />
10774 <edge from-layer="291" from-port="0" to-layer="292" to-port="4" />
10775 <edge from-layer="292" from-port="5" to-layer="297" to-port="0" />
10776 <edge from-layer="292" from-port="5" to-layer="313" to-port="0" />
10777 <edge from-layer="292" from-port="5" to-layer="329" to-port="0" />
10778 <edge from-layer="293" from-port="0" to-layer="294" to-port="0" />
10779 <edge from-layer="294" from-port="1" to-layer="296" to-port="0" />
10780 <edge from-layer="295" from-port="0" to-layer="296" to-port="1" />
10781 <edge from-layer="296" from-port="2" to-layer="297" to-port="1" />
10782 <edge from-layer="297" from-port="2" to-layer="299" to-port="0" />
10783 <edge from-layer="298" from-port="0" to-layer="299" to-port="1" />
10784 <edge from-layer="299" from-port="2" to-layer="304" to-port="0" />
10785 <edge from-layer="300" from-port="0" to-layer="304" to-port="1" />
10786 <edge from-layer="301" from-port="0" to-layer="304" to-port="2" />
10787 <edge from-layer="302" from-port="0" to-layer="304" to-port="3" />
10788 <edge from-layer="303" from-port="0" to-layer="304" to-port="4" />
10789 <edge from-layer="304" from-port="5" to-layer="306" to-port="0" />
10790 <edge from-layer="305" from-port="0" to-layer="306" to-port="1" />
10791 <edge from-layer="306" from-port="2" to-layer="308" to-port="0" />
10792 <edge from-layer="307" from-port="0" to-layer="308" to-port="1" />
10793 <edge from-layer="308" from-port="2" to-layer="336" to-port="0" />
10794 <edge from-layer="309" from-port="0" to-layer="310" to-port="0" />
10795 <edge from-layer="310" from-port="1" to-layer="312" to-port="0" />
10796 <edge from-layer="311" from-port="0" to-layer="312" to-port="1" />
10797 <edge from-layer="312" from-port="2" to-layer="313" to-port="1" />
10798 <edge from-layer="313" from-port="2" to-layer="315" to-port="0" />
10799 <edge from-layer="314" from-port="0" to-layer="315" to-port="1" />
10800 <edge from-layer="315" from-port="2" to-layer="320" to-port="0" />
10801 <edge from-layer="316" from-port="0" to-layer="320" to-port="1" />
10802 <edge from-layer="317" from-port="0" to-layer="320" to-port="2" />
10803 <edge from-layer="318" from-port="0" to-layer="320" to-port="3" />
10804 <edge from-layer="319" from-port="0" to-layer="320" to-port="4" />
10805 <edge from-layer="320" from-port="5" to-layer="322" to-port="0" />
10806 <edge from-layer="321" from-port="0" to-layer="322" to-port="1" />
10807 <edge from-layer="322" from-port="2" to-layer="324" to-port="0" />
10808 <edge from-layer="323" from-port="0" to-layer="324" to-port="1" />
10809 <edge from-layer="324" from-port="2" to-layer="336" to-port="1" />
10810 <edge from-layer="325" from-port="0" to-layer="326" to-port="0" />
10811 <edge from-layer="326" from-port="1" to-layer="328" to-port="0" />
10812 <edge from-layer="327" from-port="0" to-layer="328" to-port="1" />
10813 <edge from-layer="328" from-port="2" to-layer="329" to-port="1" />
10814 <edge from-layer="329" from-port="2" to-layer="331" to-port="0" />
10815 <edge from-layer="330" from-port="0" to-layer="331" to-port="1" />
10816 <edge from-layer="331" from-port="2" to-layer="333" to-port="0" />
10817 <edge from-layer="332" from-port="0" to-layer="333" to-port="1" />
10818 <edge from-layer="333" from-port="2" to-layer="335" to-port="0" />
10819 <edge from-layer="334" from-port="0" to-layer="335" to-port="1" />
10820 <edge from-layer="335" from-port="2" to-layer="336" to-port="2" />
10821 <edge from-layer="336" from-port="4" to-layer="338" to-port="0" />
10822 <edge from-layer="337" from-port="0" to-layer="338" to-port="1" />
10823 <edge from-layer="338" from-port="2" to-layer="340" to-port="0" />
10824 <edge from-layer="339" from-port="0" to-layer="340" to-port="1" />
10825 <edge from-layer="340" from-port="2" to-layer="342" to-port="0" />
10826 <edge from-layer="341" from-port="0" to-layer="342" to-port="1" />
10827 <edge from-layer="342" from-port="2" to-layer="347" to-port="0" />
10828 <edge from-layer="343" from-port="0" to-layer="347" to-port="1" />
10829 <edge from-layer="344" from-port="0" to-layer="347" to-port="2" />
10830 <edge from-layer="345" from-port="0" to-layer="347" to-port="3" />
10831 <edge from-layer="346" from-port="0" to-layer="347" to-port="4" />
10832 <edge from-layer="347" from-port="5" to-layer="352" to-port="0" />
10833 <edge from-layer="348" from-port="0" to-layer="349" to-port="0" />
10834 <edge from-layer="349" from-port="1" to-layer="351" to-port="0" />
10835 <edge from-layer="350" from-port="0" to-layer="351" to-port="1" />
10836 <edge from-layer="351" from-port="2" to-layer="352" to-port="1" />
10837 <edge from-layer="352" from-port="2" to-layer="354" to-port="0" />
10838 <edge from-layer="353" from-port="0" to-layer="354" to-port="1" />
10839 <edge from-layer="354" from-port="2" to-layer="355" to-port="0" />
10840 <edge from-layer="355" from-port="2" to-layer="357" to-port="0" />
10841 <edge from-layer="356" from-port="0" to-layer="357" to-port="1" />
10842 <edge from-layer="357" from-port="2" to-layer="359" to-port="0" />
10843 <edge from-layer="358" from-port="0" to-layer="359" to-port="1" />
10844 <edge from-layer="359" from-port="2" to-layer="361" to-port="0" />
10845 <edge from-layer="360" from-port="0" to-layer="361" to-port="1" />
10846 <edge from-layer="361" from-port="2" to-layer="363" to-port="0" />
10847 <edge from-layer="361" from-port="2" to-layer="391" to-port="1" />
10848 <edge from-layer="362" from-port="0" to-layer="363" to-port="1" />
10849 <edge from-layer="363" from-port="2" to-layer="368" to-port="0" />
10850 <edge from-layer="364" from-port="0" to-layer="368" to-port="1" />
10851 <edge from-layer="365" from-port="0" to-layer="368" to-port="2" />
10852 <edge from-layer="366" from-port="0" to-layer="368" to-port="3" />
10853 <edge from-layer="367" from-port="0" to-layer="368" to-port="4" />
10854 <edge from-layer="368" from-port="5" to-layer="373" to-port="0" />
10855 <edge from-layer="369" from-port="0" to-layer="370" to-port="0" />
10856 <edge from-layer="370" from-port="1" to-layer="372" to-port="0" />
10857 <edge from-layer="371" from-port="0" to-layer="372" to-port="1" />
10858 <edge from-layer="372" from-port="2" to-layer="373" to-port="1" />
10859 <edge from-layer="373" from-port="2" to-layer="375" to-port="0" />
10860 <edge from-layer="374" from-port="0" to-layer="375" to-port="1" />
10861 <edge from-layer="375" from-port="2" to-layer="376" to-port="0" />
10862 <edge from-layer="376" from-port="1" to-layer="378" to-port="0" />
10863 <edge from-layer="377" from-port="0" to-layer="378" to-port="1" />
10864 <edge from-layer="378" from-port="2" to-layer="383" to-port="0" />
10865 <edge from-layer="379" from-port="0" to-layer="383" to-port="1" />
10866 <edge from-layer="380" from-port="0" to-layer="383" to-port="2" />
10867 <edge from-layer="381" from-port="0" to-layer="383" to-port="3" />
10868 <edge from-layer="382" from-port="0" to-layer="383" to-port="4" />
10869 <edge from-layer="383" from-port="5" to-layer="388" to-port="0" />
10870 <edge from-layer="384" from-port="0" to-layer="385" to-port="0" />
10871 <edge from-layer="385" from-port="1" to-layer="387" to-port="0" />
10872 <edge from-layer="386" from-port="0" to-layer="387" to-port="1" />
10873 <edge from-layer="387" from-port="2" to-layer="388" to-port="1" />
10874 <edge from-layer="388" from-port="2" to-layer="390" to-port="0" />
10875 <edge from-layer="389" from-port="0" to-layer="390" to-port="1" />
10876 <edge from-layer="390" from-port="2" to-layer="391" to-port="0" />
10877 <edge from-layer="391" from-port="2" to-layer="393" to-port="0" />
10878 <edge from-layer="392" from-port="0" to-layer="393" to-port="1" />
10879 <edge from-layer="393" from-port="2" to-layer="395" to-port="0" />
10880 <edge from-layer="394" from-port="0" to-layer="395" to-port="1" />
10881 <edge from-layer="395" from-port="2" to-layer="397" to-port="0" />
10882 <edge from-layer="396" from-port="0" to-layer="397" to-port="1" />
10883 <edge from-layer="397" from-port="2" to-layer="399" to-port="0" />
10884 <edge from-layer="397" from-port="2" to-layer="467" to-port="1" />
10885 <edge from-layer="398" from-port="0" to-layer="399" to-port="1" />
10886 <edge from-layer="399" from-port="2" to-layer="404" to-port="0" />
10887 <edge from-layer="400" from-port="0" to-layer="404" to-port="1" />
10888 <edge from-layer="401" from-port="0" to-layer="404" to-port="2" />
10889 <edge from-layer="402" from-port="0" to-layer="404" to-port="3" />
10890 <edge from-layer="403" from-port="0" to-layer="404" to-port="4" />
10891 <edge from-layer="404" from-port="5" to-layer="409" to-port="0" />
10892 <edge from-layer="404" from-port="5" to-layer="441" to-port="0" />
10893 <edge from-layer="404" from-port="5" to-layer="425" to-port="0" />
10894 <edge from-layer="405" from-port="0" to-layer="406" to-port="0" />
10895 <edge from-layer="406" from-port="1" to-layer="408" to-port="0" />
10896 <edge from-layer="407" from-port="0" to-layer="408" to-port="1" />
10897 <edge from-layer="408" from-port="2" to-layer="409" to-port="1" />
10898 <edge from-layer="409" from-port="2" to-layer="411" to-port="0" />
10899 <edge from-layer="410" from-port="0" to-layer="411" to-port="1" />
10900 <edge from-layer="411" from-port="2" to-layer="416" to-port="0" />
10901 <edge from-layer="412" from-port="0" to-layer="416" to-port="1" />
10902 <edge from-layer="413" from-port="0" to-layer="416" to-port="2" />
10903 <edge from-layer="414" from-port="0" to-layer="416" to-port="3" />
10904 <edge from-layer="415" from-port="0" to-layer="416" to-port="4" />
10905 <edge from-layer="416" from-port="5" to-layer="418" to-port="0" />
10906 <edge from-layer="417" from-port="0" to-layer="418" to-port="1" />
10907 <edge from-layer="418" from-port="2" to-layer="420" to-port="0" />
10908 <edge from-layer="419" from-port="0" to-layer="420" to-port="1" />
10909 <edge from-layer="420" from-port="2" to-layer="448" to-port="0" />
10910 <edge from-layer="421" from-port="0" to-layer="422" to-port="0" />
10911 <edge from-layer="422" from-port="1" to-layer="424" to-port="0" />
10912 <edge from-layer="423" from-port="0" to-layer="424" to-port="1" />
10913 <edge from-layer="424" from-port="2" to-layer="425" to-port="1" />
10914 <edge from-layer="425" from-port="2" to-layer="427" to-port="0" />
10915 <edge from-layer="426" from-port="0" to-layer="427" to-port="1" />
10916 <edge from-layer="427" from-port="2" to-layer="432" to-port="0" />
10917 <edge from-layer="428" from-port="0" to-layer="432" to-port="1" />
10918 <edge from-layer="429" from-port="0" to-layer="432" to-port="2" />
10919 <edge from-layer="430" from-port="0" to-layer="432" to-port="3" />
10920 <edge from-layer="431" from-port="0" to-layer="432" to-port="4" />
10921 <edge from-layer="432" from-port="5" to-layer="434" to-port="0" />
10922 <edge from-layer="433" from-port="0" to-layer="434" to-port="1" />
10923 <edge from-layer="434" from-port="2" to-layer="436" to-port="0" />
10924 <edge from-layer="435" from-port="0" to-layer="436" to-port="1" />
10925 <edge from-layer="436" from-port="2" to-layer="448" to-port="1" />
10926 <edge from-layer="437" from-port="0" to-layer="438" to-port="0" />
10927 <edge from-layer="438" from-port="1" to-layer="440" to-port="0" />
10928 <edge from-layer="439" from-port="0" to-layer="440" to-port="1" />
10929 <edge from-layer="440" from-port="2" to-layer="441" to-port="1" />
10930 <edge from-layer="441" from-port="2" to-layer="443" to-port="0" />
10931 <edge from-layer="442" from-port="0" to-layer="443" to-port="1" />
10932 <edge from-layer="443" from-port="2" to-layer="445" to-port="0" />
10933 <edge from-layer="444" from-port="0" to-layer="445" to-port="1" />
10934 <edge from-layer="445" from-port="2" to-layer="447" to-port="0" />
10935 <edge from-layer="446" from-port="0" to-layer="447" to-port="1" />
10936 <edge from-layer="447" from-port="2" to-layer="448" to-port="2" />
10937 <edge from-layer="448" from-port="4" to-layer="450" to-port="0" />
10938 <edge from-layer="449" from-port="0" to-layer="450" to-port="1" />
10939 <edge from-layer="450" from-port="2" to-layer="452" to-port="0" />
10940 <edge from-layer="451" from-port="0" to-layer="452" to-port="1" />
10941 <edge from-layer="452" from-port="2" to-layer="454" to-port="0" />
10942 <edge from-layer="453" from-port="0" to-layer="454" to-port="1" />
10943 <edge from-layer="454" from-port="2" to-layer="459" to-port="0" />
10944 <edge from-layer="455" from-port="0" to-layer="459" to-port="1" />
10945 <edge from-layer="456" from-port="0" to-layer="459" to-port="2" />
10946 <edge from-layer="457" from-port="0" to-layer="459" to-port="3" />
10947 <edge from-layer="458" from-port="0" to-layer="459" to-port="4" />
10948 <edge from-layer="459" from-port="5" to-layer="464" to-port="0" />
10949 <edge from-layer="460" from-port="0" to-layer="461" to-port="0" />
10950 <edge from-layer="461" from-port="1" to-layer="463" to-port="0" />
10951 <edge from-layer="462" from-port="0" to-layer="463" to-port="1" />
10952 <edge from-layer="463" from-port="2" to-layer="464" to-port="1" />
10953 <edge from-layer="464" from-port="2" to-layer="466" to-port="0" />
10954 <edge from-layer="465" from-port="0" to-layer="466" to-port="1" />
10955 <edge from-layer="466" from-port="2" to-layer="467" to-port="0" />
10956 <edge from-layer="467" from-port="2" to-layer="469" to-port="0" />
10957 <edge from-layer="468" from-port="0" to-layer="469" to-port="1" />
10958 <edge from-layer="469" from-port="2" to-layer="471" to-port="0" />
10959 <edge from-layer="470" from-port="0" to-layer="471" to-port="1" />
10960 <edge from-layer="471" from-port="2" to-layer="473" to-port="0" />
10961 <edge from-layer="472" from-port="0" to-layer="473" to-port="1" />
10962 <edge from-layer="473" from-port="2" to-layer="475" to-port="0" />
10963 <edge from-layer="473" from-port="2" to-layer="503" to-port="1" />
10964 <edge from-layer="474" from-port="0" to-layer="475" to-port="1" />
10965 <edge from-layer="475" from-port="2" to-layer="480" to-port="0" />
10966 <edge from-layer="476" from-port="0" to-layer="480" to-port="1" />
10967 <edge from-layer="477" from-port="0" to-layer="480" to-port="2" />
10968 <edge from-layer="478" from-port="0" to-layer="480" to-port="3" />
10969 <edge from-layer="479" from-port="0" to-layer="480" to-port="4" />
10970 <edge from-layer="480" from-port="5" to-layer="485" to-port="0" />
10971 <edge from-layer="481" from-port="0" to-layer="482" to-port="0" />
10972 <edge from-layer="482" from-port="1" to-layer="484" to-port="0" />
10973 <edge from-layer="483" from-port="0" to-layer="484" to-port="1" />
10974 <edge from-layer="484" from-port="2" to-layer="485" to-port="1" />
10975 <edge from-layer="485" from-port="2" to-layer="487" to-port="0" />
10976 <edge from-layer="486" from-port="0" to-layer="487" to-port="1" />
10977 <edge from-layer="487" from-port="2" to-layer="488" to-port="0" />
10978 <edge from-layer="488" from-port="1" to-layer="490" to-port="0" />
10979 <edge from-layer="489" from-port="0" to-layer="490" to-port="1" />
10980 <edge from-layer="490" from-port="2" to-layer="495" to-port="0" />
10981 <edge from-layer="491" from-port="0" to-layer="495" to-port="1" />
10982 <edge from-layer="492" from-port="0" to-layer="495" to-port="2" />
10983 <edge from-layer="493" from-port="0" to-layer="495" to-port="3" />
10984 <edge from-layer="494" from-port="0" to-layer="495" to-port="4" />
10985 <edge from-layer="495" from-port="5" to-layer="500" to-port="0" />
10986 <edge from-layer="496" from-port="0" to-layer="497" to-port="0" />
10987 <edge from-layer="497" from-port="1" to-layer="499" to-port="0" />
10988 <edge from-layer="498" from-port="0" to-layer="499" to-port="1" />
10989 <edge from-layer="499" from-port="2" to-layer="500" to-port="1" />
10990 <edge from-layer="500" from-port="2" to-layer="502" to-port="0" />
10991 <edge from-layer="501" from-port="0" to-layer="502" to-port="1" />
10992 <edge from-layer="502" from-port="2" to-layer="503" to-port="0" />
10993 <edge from-layer="503" from-port="2" to-layer="505" to-port="0" />
10994 <edge from-layer="504" from-port="0" to-layer="505" to-port="1" />
10995 <edge from-layer="505" from-port="2" to-layer="507" to-port="0" />
10996 <edge from-layer="506" from-port="0" to-layer="507" to-port="1" />
10997 <edge from-layer="507" from-port="2" to-layer="509" to-port="0" />
10998 <edge from-layer="508" from-port="0" to-layer="509" to-port="1" />
10999 <edge from-layer="509" from-port="2" to-layer="511" to-port="0" />
11000 <edge from-layer="509" from-port="2" to-layer="579" to-port="1" />
11001 <edge from-layer="510" from-port="0" to-layer="511" to-port="1" />
11002 <edge from-layer="511" from-port="2" to-layer="516" to-port="0" />
11003 <edge from-layer="512" from-port="0" to-layer="516" to-port="1" />
11004 <edge from-layer="513" from-port="0" to-layer="516" to-port="2" />
11005 <edge from-layer="514" from-port="0" to-layer="516" to-port="3" />
11006 <edge from-layer="515" from-port="0" to-layer="516" to-port="4" />
11007 <edge from-layer="516" from-port="5" to-layer="521" to-port="0" />
11008 <edge from-layer="516" from-port="5" to-layer="537" to-port="0" />
11009 <edge from-layer="516" from-port="5" to-layer="553" to-port="0" />
11010 <edge from-layer="517" from-port="0" to-layer="518" to-port="0" />
11011 <edge from-layer="518" from-port="1" to-layer="520" to-port="0" />
11012 <edge from-layer="519" from-port="0" to-layer="520" to-port="1" />
11013 <edge from-layer="520" from-port="2" to-layer="521" to-port="1" />
11014 <edge from-layer="521" from-port="2" to-layer="523" to-port="0" />
11015 <edge from-layer="522" from-port="0" to-layer="523" to-port="1" />
11016 <edge from-layer="523" from-port="2" to-layer="528" to-port="0" />
11017 <edge from-layer="524" from-port="0" to-layer="528" to-port="1" />
11018 <edge from-layer="525" from-port="0" to-layer="528" to-port="2" />
11019 <edge from-layer="526" from-port="0" to-layer="528" to-port="3" />
11020 <edge from-layer="527" from-port="0" to-layer="528" to-port="4" />
11021 <edge from-layer="528" from-port="5" to-layer="530" to-port="0" />
11022 <edge from-layer="529" from-port="0" to-layer="530" to-port="1" />
11023 <edge from-layer="530" from-port="2" to-layer="532" to-port="0" />
11024 <edge from-layer="531" from-port="0" to-layer="532" to-port="1" />
11025 <edge from-layer="532" from-port="2" to-layer="560" to-port="0" />
11026 <edge from-layer="533" from-port="0" to-layer="534" to-port="0" />
11027 <edge from-layer="534" from-port="1" to-layer="536" to-port="0" />
11028 <edge from-layer="535" from-port="0" to-layer="536" to-port="1" />
11029 <edge from-layer="536" from-port="2" to-layer="537" to-port="1" />
11030 <edge from-layer="537" from-port="2" to-layer="539" to-port="0" />
11031 <edge from-layer="538" from-port="0" to-layer="539" to-port="1" />
11032 <edge from-layer="539" from-port="2" to-layer="544" to-port="0" />
11033 <edge from-layer="540" from-port="0" to-layer="544" to-port="1" />
11034 <edge from-layer="541" from-port="0" to-layer="544" to-port="2" />
11035 <edge from-layer="542" from-port="0" to-layer="544" to-port="3" />
11036 <edge from-layer="543" from-port="0" to-layer="544" to-port="4" />
11037 <edge from-layer="544" from-port="5" to-layer="546" to-port="0" />
11038 <edge from-layer="545" from-port="0" to-layer="546" to-port="1" />
11039 <edge from-layer="546" from-port="2" to-layer="548" to-port="0" />
11040 <edge from-layer="547" from-port="0" to-layer="548" to-port="1" />
11041 <edge from-layer="548" from-port="2" to-layer="560" to-port="1" />
11042 <edge from-layer="549" from-port="0" to-layer="550" to-port="0" />
11043 <edge from-layer="550" from-port="1" to-layer="552" to-port="0" />
11044 <edge from-layer="551" from-port="0" to-layer="552" to-port="1" />
11045 <edge from-layer="552" from-port="2" to-layer="553" to-port="1" />
11046 <edge from-layer="553" from-port="2" to-layer="555" to-port="0" />
11047 <edge from-layer="554" from-port="0" to-layer="555" to-port="1" />
11048 <edge from-layer="555" from-port="2" to-layer="557" to-port="0" />
11049 <edge from-layer="556" from-port="0" to-layer="557" to-port="1" />
11050 <edge from-layer="557" from-port="2" to-layer="559" to-port="0" />
11051 <edge from-layer="558" from-port="0" to-layer="559" to-port="1" />
11052 <edge from-layer="559" from-port="2" to-layer="560" to-port="2" />
11053 <edge from-layer="560" from-port="4" to-layer="562" to-port="0" />
11054 <edge from-layer="561" from-port="0" to-layer="562" to-port="1" />
11055 <edge from-layer="562" from-port="2" to-layer="564" to-port="0" />
11056 <edge from-layer="563" from-port="0" to-layer="564" to-port="1" />
11057 <edge from-layer="564" from-port="2" to-layer="566" to-port="0" />
11058 <edge from-layer="565" from-port="0" to-layer="566" to-port="1" />
11059 <edge from-layer="566" from-port="2" to-layer="571" to-port="0" />
11060 <edge from-layer="567" from-port="0" to-layer="571" to-port="1" />
11061 <edge from-layer="568" from-port="0" to-layer="571" to-port="2" />
11062 <edge from-layer="569" from-port="0" to-layer="571" to-port="3" />
11063 <edge from-layer="570" from-port="0" to-layer="571" to-port="4" />
11064 <edge from-layer="571" from-port="5" to-layer="576" to-port="0" />
11065 <edge from-layer="572" from-port="0" to-layer="573" to-port="0" />
11066 <edge from-layer="573" from-port="1" to-layer="575" to-port="0" />
11067 <edge from-layer="574" from-port="0" to-layer="575" to-port="1" />
11068 <edge from-layer="575" from-port="2" to-layer="576" to-port="1" />
11069 <edge from-layer="576" from-port="2" to-layer="578" to-port="0" />
11070 <edge from-layer="577" from-port="0" to-layer="578" to-port="1" />
11071 <edge from-layer="578" from-port="2" to-layer="579" to-port="0" />
11072 <edge from-layer="579" from-port="2" to-layer="581" to-port="0" />
11073 <edge from-layer="580" from-port="0" to-layer="581" to-port="1" />
11074 <edge from-layer="581" from-port="2" to-layer="583" to-port="0" />
11075 <edge from-layer="582" from-port="0" to-layer="583" to-port="1" />
11076 <edge from-layer="583" from-port="2" to-layer="585" to-port="0" />
11077 <edge from-layer="584" from-port="0" to-layer="585" to-port="1" />
11078 <edge from-layer="585" from-port="2" to-layer="587" to-port="0" />
11079 <edge from-layer="585" from-port="2" to-layer="615" to-port="1" />
11080 <edge from-layer="586" from-port="0" to-layer="587" to-port="1" />
11081 <edge from-layer="587" from-port="2" to-layer="592" to-port="0" />
11082 <edge from-layer="588" from-port="0" to-layer="592" to-port="1" />
11083 <edge from-layer="589" from-port="0" to-layer="592" to-port="2" />
11084 <edge from-layer="590" from-port="0" to-layer="592" to-port="3" />
11085 <edge from-layer="591" from-port="0" to-layer="592" to-port="4" />
11086 <edge from-layer="592" from-port="5" to-layer="597" to-port="0" />
11087 <edge from-layer="593" from-port="0" to-layer="594" to-port="0" />
11088 <edge from-layer="594" from-port="1" to-layer="596" to-port="0" />
11089 <edge from-layer="595" from-port="0" to-layer="596" to-port="1" />
11090 <edge from-layer="596" from-port="2" to-layer="597" to-port="1" />
11091 <edge from-layer="597" from-port="2" to-layer="599" to-port="0" />
11092 <edge from-layer="598" from-port="0" to-layer="599" to-port="1" />
11093 <edge from-layer="599" from-port="2" to-layer="600" to-port="0" />
11094 <edge from-layer="600" from-port="1" to-layer="602" to-port="0" />
11095 <edge from-layer="601" from-port="0" to-layer="602" to-port="1" />
11096 <edge from-layer="602" from-port="2" to-layer="607" to-port="0" />
11097 <edge from-layer="603" from-port="0" to-layer="607" to-port="1" />
11098 <edge from-layer="604" from-port="0" to-layer="607" to-port="2" />
11099 <edge from-layer="605" from-port="0" to-layer="607" to-port="3" />
11100 <edge from-layer="606" from-port="0" to-layer="607" to-port="4" />
11101 <edge from-layer="607" from-port="5" to-layer="612" to-port="0" />
11102 <edge from-layer="608" from-port="0" to-layer="609" to-port="0" />
11103 <edge from-layer="609" from-port="1" to-layer="611" to-port="0" />
11104 <edge from-layer="610" from-port="0" to-layer="611" to-port="1" />
11105 <edge from-layer="611" from-port="2" to-layer="612" to-port="1" />
11106 <edge from-layer="612" from-port="2" to-layer="614" to-port="0" />
11107 <edge from-layer="613" from-port="0" to-layer="614" to-port="1" />
11108 <edge from-layer="614" from-port="2" to-layer="615" to-port="0" />
11109 <edge from-layer="615" from-port="2" to-layer="617" to-port="0" />
11110 <edge from-layer="616" from-port="0" to-layer="617" to-port="1" />
11111 <edge from-layer="617" from-port="2" to-layer="619" to-port="0" />
11112 <edge from-layer="618" from-port="0" to-layer="619" to-port="1" />
11113 <edge from-layer="619" from-port="2" to-layer="621" to-port="0" />
11114 <edge from-layer="620" from-port="0" to-layer="621" to-port="1" />
11115 <edge from-layer="621" from-port="2" to-layer="623" to-port="0" />
11116 <edge from-layer="621" from-port="2" to-layer="691" to-port="1" />
11117 <edge from-layer="622" from-port="0" to-layer="623" to-port="1" />
11118 <edge from-layer="623" from-port="2" to-layer="628" to-port="0" />
11119 <edge from-layer="624" from-port="0" to-layer="628" to-port="1" />
11120 <edge from-layer="625" from-port="0" to-layer="628" to-port="2" />
11121 <edge from-layer="626" from-port="0" to-layer="628" to-port="3" />
11122 <edge from-layer="627" from-port="0" to-layer="628" to-port="4" />
11123 <edge from-layer="628" from-port="5" to-layer="633" to-port="0" />
11124 <edge from-layer="628" from-port="5" to-layer="665" to-port="0" />
11125 <edge from-layer="628" from-port="5" to-layer="649" to-port="0" />
11126 <edge from-layer="629" from-port="0" to-layer="630" to-port="0" />
11127 <edge from-layer="630" from-port="1" to-layer="632" to-port="0" />
11128 <edge from-layer="631" from-port="0" to-layer="632" to-port="1" />
11129 <edge from-layer="632" from-port="2" to-layer="633" to-port="1" />
11130 <edge from-layer="633" from-port="2" to-layer="635" to-port="0" />
11131 <edge from-layer="634" from-port="0" to-layer="635" to-port="1" />
11132 <edge from-layer="635" from-port="2" to-layer="640" to-port="0" />
11133 <edge from-layer="636" from-port="0" to-layer="640" to-port="1" />
11134 <edge from-layer="637" from-port="0" to-layer="640" to-port="2" />
11135 <edge from-layer="638" from-port="0" to-layer="640" to-port="3" />
11136 <edge from-layer="639" from-port="0" to-layer="640" to-port="4" />
11137 <edge from-layer="640" from-port="5" to-layer="642" to-port="0" />
11138 <edge from-layer="641" from-port="0" to-layer="642" to-port="1" />
11139 <edge from-layer="642" from-port="2" to-layer="644" to-port="0" />
11140 <edge from-layer="643" from-port="0" to-layer="644" to-port="1" />
11141 <edge from-layer="644" from-port="2" to-layer="672" to-port="0" />
11142 <edge from-layer="645" from-port="0" to-layer="646" to-port="0" />
11143 <edge from-layer="646" from-port="1" to-layer="648" to-port="0" />
11144 <edge from-layer="647" from-port="0" to-layer="648" to-port="1" />
11145 <edge from-layer="648" from-port="2" to-layer="649" to-port="1" />
11146 <edge from-layer="649" from-port="2" to-layer="651" to-port="0" />
11147 <edge from-layer="650" from-port="0" to-layer="651" to-port="1" />
11148 <edge from-layer="651" from-port="2" to-layer="656" to-port="0" />
11149 <edge from-layer="652" from-port="0" to-layer="656" to-port="1" />
11150 <edge from-layer="653" from-port="0" to-layer="656" to-port="2" />
11151 <edge from-layer="654" from-port="0" to-layer="656" to-port="3" />
11152 <edge from-layer="655" from-port="0" to-layer="656" to-port="4" />
11153 <edge from-layer="656" from-port="5" to-layer="658" to-port="0" />
11154 <edge from-layer="657" from-port="0" to-layer="658" to-port="1" />
11155 <edge from-layer="658" from-port="2" to-layer="660" to-port="0" />
11156 <edge from-layer="659" from-port="0" to-layer="660" to-port="1" />
11157 <edge from-layer="660" from-port="2" to-layer="672" to-port="1" />
11158 <edge from-layer="661" from-port="0" to-layer="662" to-port="0" />
11159 <edge from-layer="662" from-port="1" to-layer="664" to-port="0" />
11160 <edge from-layer="663" from-port="0" to-layer="664" to-port="1" />
11161 <edge from-layer="664" from-port="2" to-layer="665" to-port="1" />
11162 <edge from-layer="665" from-port="2" to-layer="667" to-port="0" />
11163 <edge from-layer="666" from-port="0" to-layer="667" to-port="1" />
11164 <edge from-layer="667" from-port="2" to-layer="669" to-port="0" />
11165 <edge from-layer="668" from-port="0" to-layer="669" to-port="1" />
11166 <edge from-layer="669" from-port="2" to-layer="671" to-port="0" />
11167 <edge from-layer="670" from-port="0" to-layer="671" to-port="1" />
11168 <edge from-layer="671" from-port="2" to-layer="672" to-port="2" />
11169 <edge from-layer="672" from-port="4" to-layer="674" to-port="0" />
11170 <edge from-layer="673" from-port="0" to-layer="674" to-port="1" />
11171 <edge from-layer="674" from-port="2" to-layer="676" to-port="0" />
11172 <edge from-layer="675" from-port="0" to-layer="676" to-port="1" />
11173 <edge from-layer="676" from-port="2" to-layer="678" to-port="0" />
11174 <edge from-layer="677" from-port="0" to-layer="678" to-port="1" />
11175 <edge from-layer="678" from-port="2" to-layer="683" to-port="0" />
11176 <edge from-layer="679" from-port="0" to-layer="683" to-port="1" />
11177 <edge from-layer="680" from-port="0" to-layer="683" to-port="2" />
11178 <edge from-layer="681" from-port="0" to-layer="683" to-port="3" />
11179 <edge from-layer="682" from-port="0" to-layer="683" to-port="4" />
11180 <edge from-layer="683" from-port="5" to-layer="688" to-port="0" />
11181 <edge from-layer="684" from-port="0" to-layer="685" to-port="0" />
11182 <edge from-layer="685" from-port="1" to-layer="687" to-port="0" />
11183 <edge from-layer="686" from-port="0" to-layer="687" to-port="1" />
11184 <edge from-layer="687" from-port="2" to-layer="688" to-port="1" />
11185 <edge from-layer="688" from-port="2" to-layer="690" to-port="0" />
11186 <edge from-layer="689" from-port="0" to-layer="690" to-port="1" />
11187 <edge from-layer="690" from-port="2" to-layer="691" to-port="0" />
11188 <edge from-layer="691" from-port="2" to-layer="693" to-port="0" />
11189 <edge from-layer="692" from-port="0" to-layer="693" to-port="1" />
11190 <edge from-layer="693" from-port="2" to-layer="695" to-port="0" />
11191 <edge from-layer="694" from-port="0" to-layer="695" to-port="1" />
11192 <edge from-layer="695" from-port="2" to-layer="697" to-port="0" />
11193 <edge from-layer="696" from-port="0" to-layer="697" to-port="1" />
11194 <edge from-layer="697" from-port="2" to-layer="699" to-port="0" />
11195 <edge from-layer="697" from-port="2" to-layer="727" to-port="1" />
11196 <edge from-layer="698" from-port="0" to-layer="699" to-port="1" />
11197 <edge from-layer="699" from-port="2" to-layer="704" to-port="0" />
11198 <edge from-layer="700" from-port="0" to-layer="704" to-port="1" />
11199 <edge from-layer="701" from-port="0" to-layer="704" to-port="2" />
11200 <edge from-layer="702" from-port="0" to-layer="704" to-port="3" />
11201 <edge from-layer="703" from-port="0" to-layer="704" to-port="4" />
11202 <edge from-layer="704" from-port="5" to-layer="709" to-port="0" />
11203 <edge from-layer="705" from-port="0" to-layer="706" to-port="0" />
11204 <edge from-layer="706" from-port="1" to-layer="708" to-port="0" />
11205 <edge from-layer="707" from-port="0" to-layer="708" to-port="1" />
11206 <edge from-layer="708" from-port="2" to-layer="709" to-port="1" />
11207 <edge from-layer="709" from-port="2" to-layer="711" to-port="0" />
11208 <edge from-layer="710" from-port="0" to-layer="711" to-port="1" />
11209 <edge from-layer="711" from-port="2" to-layer="712" to-port="0" />
11210 <edge from-layer="712" from-port="1" to-layer="714" to-port="0" />
11211 <edge from-layer="713" from-port="0" to-layer="714" to-port="1" />
11212 <edge from-layer="714" from-port="2" to-layer="719" to-port="0" />
11213 <edge from-layer="715" from-port="0" to-layer="719" to-port="1" />
11214 <edge from-layer="716" from-port="0" to-layer="719" to-port="2" />
11215 <edge from-layer="717" from-port="0" to-layer="719" to-port="3" />
11216 <edge from-layer="718" from-port="0" to-layer="719" to-port="4" />
11217 <edge from-layer="719" from-port="5" to-layer="724" to-port="0" />
11218 <edge from-layer="720" from-port="0" to-layer="721" to-port="0" />
11219 <edge from-layer="721" from-port="1" to-layer="723" to-port="0" />
11220 <edge from-layer="722" from-port="0" to-layer="723" to-port="1" />
11221 <edge from-layer="723" from-port="2" to-layer="724" to-port="1" />
11222 <edge from-layer="724" from-port="2" to-layer="726" to-port="0" />
11223 <edge from-layer="725" from-port="0" to-layer="726" to-port="1" />
11224 <edge from-layer="726" from-port="2" to-layer="727" to-port="0" />
11225 <edge from-layer="727" from-port="2" to-layer="729" to-port="0" />
11226 <edge from-layer="728" from-port="0" to-layer="729" to-port="1" />
11227 <edge from-layer="729" from-port="2" to-layer="731" to-port="0" />
11228 <edge from-layer="730" from-port="0" to-layer="731" to-port="1" />
11229 <edge from-layer="731" from-port="2" to-layer="733" to-port="0" />
11230 <edge from-layer="732" from-port="0" to-layer="733" to-port="1" />
11231 <edge from-layer="733" from-port="2" to-layer="735" to-port="0" />
11232 <edge from-layer="734" from-port="0" to-layer="735" to-port="1" />
11233 <edge from-layer="735" from-port="3" to-layer="737" to-port="0" />
11234 <edge from-layer="736" from-port="0" to-layer="737" to-port="1" />
11235 <edge from-layer="737" from-port="2" to-layer="742" to-port="0" />
11236 <edge from-layer="738" from-port="0" to-layer="742" to-port="1" />
11237 <edge from-layer="739" from-port="0" to-layer="742" to-port="2" />
11238 <edge from-layer="740" from-port="0" to-layer="742" to-port="3" />
11239 <edge from-layer="741" from-port="0" to-layer="742" to-port="4" />
11240 <edge from-layer="742" from-port="5" to-layer="747" to-port="0" />
11241 <edge from-layer="743" from-port="0" to-layer="744" to-port="0" />
11242 <edge from-layer="744" from-port="1" to-layer="746" to-port="0" />
11243 <edge from-layer="745" from-port="0" to-layer="746" to-port="1" />
11244 <edge from-layer="746" from-port="2" to-layer="747" to-port="1" />
11245 <edge from-layer="747" from-port="2" to-layer="749" to-port="0" />
11246 <edge from-layer="748" from-port="0" to-layer="749" to-port="1" />
11247 <edge from-layer="749" from-port="2" to-layer="750" to-port="0" />
11248 <edge from-layer="750" from-port="1" to-layer="752" to-port="0" />
11249 <edge from-layer="751" from-port="0" to-layer="752" to-port="1" />
11250 <edge from-layer="752" from-port="2" to-layer="757" to-port="0" />
11251 <edge from-layer="753" from-port="0" to-layer="757" to-port="1" />
11252 <edge from-layer="754" from-port="0" to-layer="757" to-port="2" />
11253 <edge from-layer="755" from-port="0" to-layer="757" to-port="3" />
11254 <edge from-layer="756" from-port="0" to-layer="757" to-port="4" />
11255 <edge from-layer="757" from-port="5" to-layer="762" to-port="0" />
11256 <edge from-layer="758" from-port="0" to-layer="759" to-port="0" />
11257 <edge from-layer="759" from-port="1" to-layer="761" to-port="0" />
11258 <edge from-layer="760" from-port="0" to-layer="761" to-port="1" />
11259 <edge from-layer="761" from-port="2" to-layer="762" to-port="1" />
11260 <edge from-layer="762" from-port="2" to-layer="763" to-port="0" />
11261 </edges>
11262 <rt_info>
11263 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
11264 <conversion_parameters>
11265 <framework value="pytorch" />
11266 <is_python_object value="True" />
11267 </conversion_parameters>
11268 <nncf>
11269 <friendly_names_were_updated value="True" />
11270 <quantization>
11271 <advanced_parameters value="{'overflow_fix': 'disable', 'quantize_outputs': False, 'inplace_statistics': True, 'disable_channel_alignment': True, 'disable_bias_correction': False, 'batchwise_statistics': None, 'activations_quantization_params': None, 'weights_quantization_params': None, 'activations_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'weights_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'bias_correction_params': {'apply_for_all_nodes': False, 'threshold': None}, 'smooth_quant_alphas': {'convolution': -1, 'matmul': 0.95}, 'smooth_quant_alpha': None, 'backend_params': {}}" />
11272 <fast_bias_correction value="True" />
11273 <ignored_scope>
11274 <types value="['GroupNormalization']" />
11275 </ignored_scope>
11276 <model_type value="transformer" />
11277 <preset value="mixed" />
11278 <subset_size value="300" />
11279 <target_device value="ANY" />
11280 </quantization>
11281 </nncf>
11282 <optimum>
11283 <optimum_intel_version value="1.20.1" />
11284 <optimum_version value="1.24.0" />
11285 <pytorch_version value="2.6.0+cu124" />
11286 <transformers_version value="4.52.0.dev0" />
11287 </optimum>
11288 </rt_info>
11289 </net>
11290