openvino/openvino_model_qint8_quantized.xml
255.8 KB · 7972 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model385" version="11">
3 <layers>
4 <layer id="2" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="1" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="0" name="token_type_ids" type="Parameter" version="opset1">
23 <data shape="?,?" element_type="i64" />
24 <output>
25 <port id="0" precision="I64" names="token_type_ids">
26 <dim>-1</dim>
27 <dim>-1</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="Constant_445498" type="Const" version="opset1">
32 <data element_type="i8" shape="30522, 384" offset="0" size="11720448" />
33 <output>
34 <port id="0" precision="I8">
35 <dim>30522</dim>
36 <dim>384</dim>
37 </port>
38 </output>
39 </layer>
40 <layer id="4" name="Convert_445499" type="Convert" version="opset1">
41 <data destination_type="f32" />
42 <input>
43 <port id="0" precision="I8">
44 <dim>30522</dim>
45 <dim>384</dim>
46 </port>
47 </input>
48 <output>
49 <port id="1" precision="FP32">
50 <dim>30522</dim>
51 <dim>384</dim>
52 </port>
53 </output>
54 </layer>
55 <layer id="5" name="Constant_445500" type="Const" version="opset1">
56 <data element_type="f32" shape="30522, 1" offset="11720448" size="122088" />
57 <output>
58 <port id="0" precision="FP32">
59 <dim>30522</dim>
60 <dim>1</dim>
61 </port>
62 </output>
63 </layer>
64 <layer id="6" name="__module.bert.embeddings.word_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
65 <data auto_broadcast="numpy" />
66 <input>
67 <port id="0" precision="FP32">
68 <dim>30522</dim>
69 <dim>384</dim>
70 </port>
71 <port id="1" precision="FP32">
72 <dim>30522</dim>
73 <dim>1</dim>
74 </port>
75 </input>
76 <output>
77 <port id="2" precision="FP32">
78 <dim>30522</dim>
79 <dim>384</dim>
80 </port>
81 </output>
82 </layer>
83 <layer id="7" name="__module.bert.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
84 <data destination_type="i32" />
85 <input>
86 <port id="0" precision="I64">
87 <dim>-1</dim>
88 <dim>-1</dim>
89 </port>
90 </input>
91 <output>
92 <port id="1" precision="I32">
93 <dim>-1</dim>
94 <dim>-1</dim>
95 </port>
96 </output>
97 </layer>
98 <layer id="8" name="__module.bert.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
99 <data element_type="i32" shape="" offset="11842536" size="4" />
100 <output>
101 <port id="0" precision="I32" />
102 </output>
103 </layer>
104 <layer id="9" name="__module.bert.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
105 <data batch_dims="0" />
106 <input>
107 <port id="0" precision="FP32">
108 <dim>30522</dim>
109 <dim>384</dim>
110 </port>
111 <port id="1" precision="I32">
112 <dim>-1</dim>
113 <dim>-1</dim>
114 </port>
115 <port id="2" precision="I32" />
116 </input>
117 <output>
118 <port id="3" precision="FP32" names="48,inputs_embeds">
119 <dim>-1</dim>
120 <dim>-1</dim>
121 <dim>384</dim>
122 </port>
123 </output>
124 </layer>
125 <layer id="10" name="Constant_445502" type="Const" version="opset1">
126 <data element_type="i8" shape="2, 384" offset="11842540" size="768" />
127 <output>
128 <port id="0" precision="I8">
129 <dim>2</dim>
130 <dim>384</dim>
131 </port>
132 </output>
133 </layer>
134 <layer id="11" name="Convert_445503" type="Convert" version="opset1">
135 <data destination_type="f32" />
136 <input>
137 <port id="0" precision="I8">
138 <dim>2</dim>
139 <dim>384</dim>
140 </port>
141 </input>
142 <output>
143 <port id="1" precision="FP32">
144 <dim>2</dim>
145 <dim>384</dim>
146 </port>
147 </output>
148 </layer>
149 <layer id="12" name="Constant_445504" type="Const" version="opset1">
150 <data element_type="f32" shape="2, 1" offset="11843308" size="8" />
151 <output>
152 <port id="0" precision="FP32">
153 <dim>2</dim>
154 <dim>1</dim>
155 </port>
156 </output>
157 </layer>
158 <layer id="13" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
159 <data auto_broadcast="numpy" />
160 <input>
161 <port id="0" precision="FP32">
162 <dim>2</dim>
163 <dim>384</dim>
164 </port>
165 <port id="1" precision="FP32">
166 <dim>2</dim>
167 <dim>1</dim>
168 </port>
169 </input>
170 <output>
171 <port id="2" precision="FP32">
172 <dim>2</dim>
173 <dim>384</dim>
174 </port>
175 </output>
176 </layer>
177 <layer id="14" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
178 <data destination_type="i32" />
179 <input>
180 <port id="0" precision="I64">
181 <dim>-1</dim>
182 <dim>-1</dim>
183 </port>
184 </input>
185 <output>
186 <port id="1" precision="I32">
187 <dim>-1</dim>
188 <dim>-1</dim>
189 </port>
190 </output>
191 </layer>
192 <layer id="15" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
193 <data element_type="i32" shape="" offset="11842536" size="4" />
194 <output>
195 <port id="0" precision="I32" />
196 </output>
197 </layer>
198 <layer id="16" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
199 <data batch_dims="0" />
200 <input>
201 <port id="0" precision="FP32">
202 <dim>2</dim>
203 <dim>384</dim>
204 </port>
205 <port id="1" precision="I32">
206 <dim>-1</dim>
207 <dim>-1</dim>
208 </port>
209 <port id="2" precision="I32" />
210 </input>
211 <output>
212 <port id="3" precision="FP32" names="50,token_type_embeddings.1">
213 <dim>-1</dim>
214 <dim>-1</dim>
215 <dim>384</dim>
216 </port>
217 </output>
218 </layer>
219 <layer id="17" name="__module.bert.embeddings/aten::add/Add" type="Add" version="opset1">
220 <data auto_broadcast="numpy" />
221 <input>
222 <port id="0" precision="FP32">
223 <dim>-1</dim>
224 <dim>-1</dim>
225 <dim>384</dim>
226 </port>
227 <port id="1" precision="FP32">
228 <dim>-1</dim>
229 <dim>-1</dim>
230 <dim>384</dim>
231 </port>
232 </input>
233 <output>
234 <port id="2" precision="FP32" names="51_1">
235 <dim>-1</dim>
236 <dim>-1</dim>
237 <dim>384</dim>
238 </port>
239 </output>
240 </layer>
241 <layer id="18" name="Constant_445506" type="Const" version="opset1">
242 <data element_type="i8" shape="512, 384" offset="11843316" size="196608" />
243 <output>
244 <port id="0" precision="I8">
245 <dim>512</dim>
246 <dim>384</dim>
247 </port>
248 </output>
249 </layer>
250 <layer id="19" name="Convert_445507" type="Convert" version="opset1">
251 <data destination_type="f32" />
252 <input>
253 <port id="0" precision="I8">
254 <dim>512</dim>
255 <dim>384</dim>
256 </port>
257 </input>
258 <output>
259 <port id="1" precision="FP32">
260 <dim>512</dim>
261 <dim>384</dim>
262 </port>
263 </output>
264 </layer>
265 <layer id="20" name="Constant_445508" type="Const" version="opset1">
266 <data element_type="f32" shape="512, 1" offset="12039924" size="2048" />
267 <output>
268 <port id="0" precision="FP32">
269 <dim>512</dim>
270 <dim>1</dim>
271 </port>
272 </output>
273 </layer>
274 <layer id="21" name="__module.bert.embeddings.position_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
275 <data auto_broadcast="numpy" />
276 <input>
277 <port id="0" precision="FP32">
278 <dim>512</dim>
279 <dim>384</dim>
280 </port>
281 <port id="1" precision="FP32">
282 <dim>512</dim>
283 <dim>1</dim>
284 </port>
285 </input>
286 <output>
287 <port id="2" precision="FP32">
288 <dim>512</dim>
289 <dim>384</dim>
290 </port>
291 </output>
292 </layer>
293 <layer id="22" name="__module.bert.embeddings/aten::slice/Slice" type="Const" version="opset1">
294 <data element_type="i64" shape="1, 512" offset="12041972" size="4096" />
295 <output>
296 <port id="0" precision="I64" names="45">
297 <dim>1</dim>
298 <dim>512</dim>
299 </port>
300 </output>
301 </layer>
302 <layer id="23" name="__module.bert.embeddings/aten::slice/Reshape" type="Const" version="opset1">
303 <data element_type="i64" shape="1" offset="12046068" size="8" />
304 <output>
305 <port id="0" precision="I64">
306 <dim>1</dim>
307 </port>
308 </output>
309 </layer>
310 <layer id="24" name="ShapeOf_305961" type="ShapeOf" version="opset3">
311 <data output_type="i64" />
312 <input>
313 <port id="0" precision="I64">
314 <dim>-1</dim>
315 <dim>-1</dim>
316 </port>
317 </input>
318 <output>
319 <port id="1" precision="I64">
320 <dim>2</dim>
321 </port>
322 </output>
323 </layer>
324 <layer id="25" name="Constant_306078" type="Const" version="opset1">
325 <data element_type="i64" shape="1" offset="12046076" size="8" />
326 <output>
327 <port id="0" precision="I64">
328 <dim>1</dim>
329 </port>
330 </output>
331 </layer>
332 <layer id="26" name="Constant_305963" type="Const" version="opset1">
333 <data element_type="i64" shape="" offset="12046068" size="8" />
334 <output>
335 <port id="0" precision="I64" />
336 </output>
337 </layer>
338 <layer id="27" name="Gather_305964" type="Gather" version="opset8">
339 <data batch_dims="0" />
340 <input>
341 <port id="0" precision="I64">
342 <dim>2</dim>
343 </port>
344 <port id="1" precision="I64">
345 <dim>1</dim>
346 </port>
347 <port id="2" precision="I64" />
348 </input>
349 <output>
350 <port id="3" precision="I64" names="35,41,43,44,61">
351 <dim>1</dim>
352 </port>
353 </output>
354 </layer>
355 <layer id="28" name="__module.bert.embeddings/aten::slice/Reshape_2" type="Const" version="opset1">
356 <data element_type="i64" shape="1" offset="12046076" size="8" />
357 <output>
358 <port id="0" precision="I64">
359 <dim>1</dim>
360 </port>
361 </output>
362 </layer>
363 <layer id="29" name="__module.bert.embeddings/aten::slice/Reshape_3" type="Const" version="opset1">
364 <data element_type="i64" shape="1" offset="12046076" size="8" />
365 <output>
366 <port id="0" precision="I64">
367 <dim>1</dim>
368 </port>
369 </output>
370 </layer>
371 <layer id="30" name="__module.bert.embeddings/aten::slice/Slice_1" type="Slice" version="opset8">
372 <input>
373 <port id="0" precision="I64">
374 <dim>1</dim>
375 <dim>512</dim>
376 </port>
377 <port id="1" precision="I64">
378 <dim>1</dim>
379 </port>
380 <port id="2" precision="I64">
381 <dim>1</dim>
382 </port>
383 <port id="3" precision="I64">
384 <dim>1</dim>
385 </port>
386 <port id="4" precision="I64">
387 <dim>1</dim>
388 </port>
389 </input>
390 <output>
391 <port id="5" precision="I64" names="46">
392 <dim>1</dim>
393 <dim>-1</dim>
394 </port>
395 </output>
396 </layer>
397 <layer id="31" name="__module.bert.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
398 <data destination_type="i32" />
399 <input>
400 <port id="0" precision="I64">
401 <dim>1</dim>
402 <dim>-1</dim>
403 </port>
404 </input>
405 <output>
406 <port id="1" precision="I32">
407 <dim>1</dim>
408 <dim>-1</dim>
409 </port>
410 </output>
411 </layer>
412 <layer id="32" name="__module.bert.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
413 <data element_type="i32" shape="" offset="11842536" size="4" />
414 <output>
415 <port id="0" precision="I32" />
416 </output>
417 </layer>
418 <layer id="33" name="__module.bert.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
419 <data batch_dims="0" />
420 <input>
421 <port id="0" precision="FP32">
422 <dim>512</dim>
423 <dim>384</dim>
424 </port>
425 <port id="1" precision="I32">
426 <dim>1</dim>
427 <dim>-1</dim>
428 </port>
429 <port id="2" precision="I32" />
430 </input>
431 <output>
432 <port id="3" precision="FP32" names="53,position_embeddings.1">
433 <dim>1</dim>
434 <dim>-1</dim>
435 <dim>384</dim>
436 </port>
437 </output>
438 </layer>
439 <layer id="34" name="__module.bert.embeddings/aten::add_/Add" type="Add" version="opset1">
440 <data auto_broadcast="numpy" />
441 <input>
442 <port id="0" precision="FP32">
443 <dim>-1</dim>
444 <dim>-1</dim>
445 <dim>384</dim>
446 </port>
447 <port id="1" precision="FP32">
448 <dim>1</dim>
449 <dim>-1</dim>
450 <dim>384</dim>
451 </port>
452 </input>
453 <output>
454 <port id="2" precision="FP32" names="51,embeddings.1">
455 <dim>-1</dim>
456 <dim>-1</dim>
457 <dim>384</dim>
458 </port>
459 </output>
460 </layer>
461 <layer id="35" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
462 <data element_type="i32" shape="1" offset="12046084" size="4" />
463 <output>
464 <port id="0" precision="I32">
465 <dim>1</dim>
466 </port>
467 </output>
468 </layer>
469 <layer id="36" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
470 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
471 <input>
472 <port id="0" precision="FP32">
473 <dim>-1</dim>
474 <dim>-1</dim>
475 <dim>384</dim>
476 </port>
477 <port id="1" precision="I32">
478 <dim>1</dim>
479 </port>
480 </input>
481 <output>
482 <port id="2" precision="FP32">
483 <dim>-1</dim>
484 <dim>-1</dim>
485 <dim>384</dim>
486 </port>
487 </output>
488 </layer>
489 <layer id="37" name="Constant_305869" type="Const" version="opset1">
490 <data element_type="f32" shape="1, 1, 384" offset="12046088" size="1536" />
491 <output>
492 <port id="0" precision="FP32">
493 <dim>1</dim>
494 <dim>1</dim>
495 <dim>384</dim>
496 </port>
497 </output>
498 </layer>
499 <layer id="38" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
500 <data auto_broadcast="numpy" />
501 <input>
502 <port id="0" precision="FP32">
503 <dim>-1</dim>
504 <dim>-1</dim>
505 <dim>384</dim>
506 </port>
507 <port id="1" precision="FP32">
508 <dim>1</dim>
509 <dim>1</dim>
510 <dim>384</dim>
511 </port>
512 </input>
513 <output>
514 <port id="2" precision="FP32">
515 <dim>-1</dim>
516 <dim>-1</dim>
517 <dim>384</dim>
518 </port>
519 </output>
520 </layer>
521 <layer id="39" name="Constant_305870" type="Const" version="opset1">
522 <data element_type="f32" shape="1, 1, 384" offset="12047624" size="1536" />
523 <output>
524 <port id="0" precision="FP32">
525 <dim>1</dim>
526 <dim>1</dim>
527 <dim>384</dim>
528 </port>
529 </output>
530 </layer>
531 <layer id="40" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
532 <data auto_broadcast="numpy" />
533 <input>
534 <port id="0" precision="FP32">
535 <dim>-1</dim>
536 <dim>-1</dim>
537 <dim>384</dim>
538 </port>
539 <port id="1" precision="FP32">
540 <dim>1</dim>
541 <dim>1</dim>
542 <dim>384</dim>
543 </port>
544 </input>
545 <output>
546 <port id="2" precision="FP32" names="58,input.1">
547 <dim>-1</dim>
548 <dim>-1</dim>
549 <dim>384</dim>
550 </port>
551 </output>
552 </layer>
553 <layer id="41" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
554 <data element_type="f32" shape="1, 1, 384" offset="12049160" size="1536" />
555 <output>
556 <port id="0" precision="FP32">
557 <dim>1</dim>
558 <dim>1</dim>
559 <dim>384</dim>
560 </port>
561 </output>
562 </layer>
563 <layer id="42" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
564 <data auto_broadcast="numpy" />
565 <input>
566 <port id="0" precision="FP32">
567 <dim>-1</dim>
568 <dim>-1</dim>
569 <dim>384</dim>
570 </port>
571 <port id="1" precision="FP32">
572 <dim>1</dim>
573 <dim>1</dim>
574 <dim>384</dim>
575 </port>
576 </input>
577 <output>
578 <port id="2" precision="FP32">
579 <dim>-1</dim>
580 <dim>-1</dim>
581 <dim>384</dim>
582 </port>
583 </output>
584 </layer>
585 <layer id="43" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
586 <data element_type="f32" shape="" offset="12050696" size="4" />
587 <output>
588 <port id="0" precision="FP32" />
589 </output>
590 </layer>
591 <layer id="44" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
592 <data element_type="f32" shape="" offset="12050700" size="4" />
593 <output>
594 <port id="0" precision="FP32" />
595 </output>
596 </layer>
597 <layer id="45" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
598 <data element_type="f32" shape="" offset="12050696" size="4" />
599 <output>
600 <port id="0" precision="FP32" />
601 </output>
602 </layer>
603 <layer id="46" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
604 <data element_type="f32" shape="" offset="12050700" size="4" />
605 <output>
606 <port id="0" precision="FP32" />
607 </output>
608 </layer>
609 <layer id="47" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
610 <data levels="256" auto_broadcast="numpy" />
611 <input>
612 <port id="0" precision="FP32">
613 <dim>-1</dim>
614 <dim>-1</dim>
615 <dim>384</dim>
616 </port>
617 <port id="1" precision="FP32" />
618 <port id="2" precision="FP32" />
619 <port id="3" precision="FP32" />
620 <port id="4" precision="FP32" />
621 </input>
622 <output>
623 <port id="5" precision="FP32">
624 <dim>-1</dim>
625 <dim>-1</dim>
626 <dim>384</dim>
627 </port>
628 </output>
629 </layer>
630 <layer id="48" name="Constant_445510" type="Const" version="opset1">
631 <data element_type="i8" shape="384, 384" offset="12050704" size="147456" />
632 <output>
633 <port id="0" precision="I8">
634 <dim>384</dim>
635 <dim>384</dim>
636 </port>
637 </output>
638 </layer>
639 <layer id="49" name="Convert_445511" type="Convert" version="opset1">
640 <data destination_type="f32" />
641 <input>
642 <port id="0" precision="I8">
643 <dim>384</dim>
644 <dim>384</dim>
645 </port>
646 </input>
647 <output>
648 <port id="1" precision="FP32">
649 <dim>384</dim>
650 <dim>384</dim>
651 </port>
652 </output>
653 </layer>
654 <layer id="50" name="Constant_445512" type="Const" version="opset1">
655 <data element_type="f32" shape="384, 1" offset="12198160" size="1536" />
656 <output>
657 <port id="0" precision="FP32">
658 <dim>384</dim>
659 <dim>1</dim>
660 </port>
661 </output>
662 </layer>
663 <layer id="51" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
664 <data auto_broadcast="numpy" />
665 <input>
666 <port id="0" precision="FP32">
667 <dim>384</dim>
668 <dim>384</dim>
669 </port>
670 <port id="1" precision="FP32">
671 <dim>384</dim>
672 <dim>1</dim>
673 </port>
674 </input>
675 <output>
676 <port id="2" precision="FP32">
677 <dim>384</dim>
678 <dim>384</dim>
679 </port>
680 </output>
681 </layer>
682 <layer id="52" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
683 <data transpose_a="false" transpose_b="true" />
684 <input>
685 <port id="0" precision="FP32">
686 <dim>-1</dim>
687 <dim>-1</dim>
688 <dim>384</dim>
689 </port>
690 <port id="1" precision="FP32">
691 <dim>384</dim>
692 <dim>384</dim>
693 </port>
694 </input>
695 <output>
696 <port id="2" precision="FP32">
697 <dim>-1</dim>
698 <dim>-1</dim>
699 <dim>384</dim>
700 </port>
701 </output>
702 </layer>
703 <layer id="53" name="Constant_305871" type="Const" version="opset1">
704 <data element_type="f32" shape="1, 1, 384" offset="12199696" size="1536" />
705 <output>
706 <port id="0" precision="FP32">
707 <dim>1</dim>
708 <dim>1</dim>
709 <dim>384</dim>
710 </port>
711 </output>
712 </layer>
713 <layer id="54" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
714 <data auto_broadcast="numpy" />
715 <input>
716 <port id="0" precision="FP32">
717 <dim>-1</dim>
718 <dim>-1</dim>
719 <dim>384</dim>
720 </port>
721 <port id="1" precision="FP32">
722 <dim>1</dim>
723 <dim>1</dim>
724 <dim>384</dim>
725 </port>
726 </input>
727 <output>
728 <port id="2" precision="FP32" names="92,x.1">
729 <dim>-1</dim>
730 <dim>-1</dim>
731 <dim>384</dim>
732 </port>
733 </output>
734 </layer>
735 <layer id="55" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
736 <data element_type="f32" shape="" offset="12201232" size="4" />
737 <output>
738 <port id="0" precision="FP32" />
739 </output>
740 </layer>
741 <layer id="56" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
742 <data element_type="f32" shape="" offset="12201236" size="4" />
743 <output>
744 <port id="0" precision="FP32" />
745 </output>
746 </layer>
747 <layer id="57" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
748 <data element_type="f32" shape="" offset="12201232" size="4" />
749 <output>
750 <port id="0" precision="FP32" />
751 </output>
752 </layer>
753 <layer id="58" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
754 <data element_type="f32" shape="" offset="12201236" size="4" />
755 <output>
756 <port id="0" precision="FP32" />
757 </output>
758 </layer>
759 <layer id="59" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
760 <data levels="256" auto_broadcast="numpy" />
761 <input>
762 <port id="0" precision="FP32">
763 <dim>-1</dim>
764 <dim>-1</dim>
765 <dim>384</dim>
766 </port>
767 <port id="1" precision="FP32" />
768 <port id="2" precision="FP32" />
769 <port id="3" precision="FP32" />
770 <port id="4" precision="FP32" />
771 </input>
772 <output>
773 <port id="5" precision="FP32">
774 <dim>-1</dim>
775 <dim>-1</dim>
776 <dim>384</dim>
777 </port>
778 </output>
779 </layer>
780 <layer id="60" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
781 <data element_type="i64" shape="4" offset="12201240" size="32" />
782 <output>
783 <port id="0" precision="I64">
784 <dim>4</dim>
785 </port>
786 </output>
787 </layer>
788 <layer id="61" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
789 <data special_zero="true" />
790 <input>
791 <port id="0" precision="FP32">
792 <dim>-1</dim>
793 <dim>-1</dim>
794 <dim>384</dim>
795 </port>
796 <port id="1" precision="I64">
797 <dim>4</dim>
798 </port>
799 </input>
800 <output>
801 <port id="2" precision="FP32" names="96,x.3">
802 <dim>-1</dim>
803 <dim>-1</dim>
804 <dim>12</dim>
805 <dim>32</dim>
806 </port>
807 </output>
808 </layer>
809 <layer id="62" name="Constant_301060" type="Const" version="opset1">
810 <data element_type="i64" shape="4" offset="12201272" size="32" />
811 <output>
812 <port id="0" precision="I64" names="97">
813 <dim>4</dim>
814 </port>
815 </output>
816 </layer>
817 <layer id="63" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
818 <input>
819 <port id="0" precision="FP32">
820 <dim>-1</dim>
821 <dim>-1</dim>
822 <dim>12</dim>
823 <dim>32</dim>
824 </port>
825 <port id="1" precision="I64">
826 <dim>4</dim>
827 </port>
828 </input>
829 <output>
830 <port id="2" precision="FP32" names="98">
831 <dim>-1</dim>
832 <dim>12</dim>
833 <dim>-1</dim>
834 <dim>32</dim>
835 </port>
836 </output>
837 </layer>
838 <layer id="64" name="Constant_445514" type="Const" version="opset1">
839 <data element_type="i8" shape="384, 384" offset="12201304" size="147456" />
840 <output>
841 <port id="0" precision="I8">
842 <dim>384</dim>
843 <dim>384</dim>
844 </port>
845 </output>
846 </layer>
847 <layer id="65" name="Convert_445515" type="Convert" version="opset1">
848 <data destination_type="f32" />
849 <input>
850 <port id="0" precision="I8">
851 <dim>384</dim>
852 <dim>384</dim>
853 </port>
854 </input>
855 <output>
856 <port id="1" precision="FP32">
857 <dim>384</dim>
858 <dim>384</dim>
859 </port>
860 </output>
861 </layer>
862 <layer id="66" name="Constant_445516" type="Const" version="opset1">
863 <data element_type="f32" shape="384, 1" offset="12348760" size="1536" />
864 <output>
865 <port id="0" precision="FP32">
866 <dim>384</dim>
867 <dim>1</dim>
868 </port>
869 </output>
870 </layer>
871 <layer id="67" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
872 <data auto_broadcast="numpy" />
873 <input>
874 <port id="0" precision="FP32">
875 <dim>384</dim>
876 <dim>384</dim>
877 </port>
878 <port id="1" precision="FP32">
879 <dim>384</dim>
880 <dim>1</dim>
881 </port>
882 </input>
883 <output>
884 <port id="2" precision="FP32">
885 <dim>384</dim>
886 <dim>384</dim>
887 </port>
888 </output>
889 </layer>
890 <layer id="68" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
891 <data transpose_a="false" transpose_b="true" />
892 <input>
893 <port id="0" precision="FP32">
894 <dim>-1</dim>
895 <dim>-1</dim>
896 <dim>384</dim>
897 </port>
898 <port id="1" precision="FP32">
899 <dim>384</dim>
900 <dim>384</dim>
901 </port>
902 </input>
903 <output>
904 <port id="2" precision="FP32">
905 <dim>-1</dim>
906 <dim>-1</dim>
907 <dim>384</dim>
908 </port>
909 </output>
910 </layer>
911 <layer id="69" name="Constant_305872" type="Const" version="opset1">
912 <data element_type="f32" shape="1, 1, 384" offset="12350296" size="1536" />
913 <output>
914 <port id="0" precision="FP32">
915 <dim>1</dim>
916 <dim>1</dim>
917 <dim>384</dim>
918 </port>
919 </output>
920 </layer>
921 <layer id="70" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
922 <data auto_broadcast="numpy" />
923 <input>
924 <port id="0" precision="FP32">
925 <dim>-1</dim>
926 <dim>-1</dim>
927 <dim>384</dim>
928 </port>
929 <port id="1" precision="FP32">
930 <dim>1</dim>
931 <dim>1</dim>
932 <dim>384</dim>
933 </port>
934 </input>
935 <output>
936 <port id="2" precision="FP32" names="101,x.5">
937 <dim>-1</dim>
938 <dim>-1</dim>
939 <dim>384</dim>
940 </port>
941 </output>
942 </layer>
943 <layer id="71" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
944 <data element_type="f32" shape="" offset="12351832" size="4" />
945 <output>
946 <port id="0" precision="FP32" />
947 </output>
948 </layer>
949 <layer id="72" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
950 <data element_type="f32" shape="" offset="12351836" size="4" />
951 <output>
952 <port id="0" precision="FP32" />
953 </output>
954 </layer>
955 <layer id="73" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
956 <data element_type="f32" shape="" offset="12351832" size="4" />
957 <output>
958 <port id="0" precision="FP32" />
959 </output>
960 </layer>
961 <layer id="74" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
962 <data element_type="f32" shape="" offset="12351836" size="4" />
963 <output>
964 <port id="0" precision="FP32" />
965 </output>
966 </layer>
967 <layer id="75" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
968 <data levels="256" auto_broadcast="numpy" />
969 <input>
970 <port id="0" precision="FP32">
971 <dim>-1</dim>
972 <dim>-1</dim>
973 <dim>384</dim>
974 </port>
975 <port id="1" precision="FP32" />
976 <port id="2" precision="FP32" />
977 <port id="3" precision="FP32" />
978 <port id="4" precision="FP32" />
979 </input>
980 <output>
981 <port id="5" precision="FP32">
982 <dim>-1</dim>
983 <dim>-1</dim>
984 <dim>384</dim>
985 </port>
986 </output>
987 </layer>
988 <layer id="76" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
989 <data element_type="i64" shape="4" offset="12201240" size="32" />
990 <output>
991 <port id="0" precision="I64">
992 <dim>4</dim>
993 </port>
994 </output>
995 </layer>
996 <layer id="77" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
997 <data special_zero="true" />
998 <input>
999 <port id="0" precision="FP32">
1000 <dim>-1</dim>
1001 <dim>-1</dim>
1002 <dim>384</dim>
1003 </port>
1004 <port id="1" precision="I64">
1005 <dim>4</dim>
1006 </port>
1007 </input>
1008 <output>
1009 <port id="2" precision="FP32" names="105,x.7">
1010 <dim>-1</dim>
1011 <dim>-1</dim>
1012 <dim>12</dim>
1013 <dim>32</dim>
1014 </port>
1015 </output>
1016 </layer>
1017 <layer id="78" name="Constant_301085" type="Const" version="opset1">
1018 <data element_type="i64" shape="4" offset="12201272" size="32" />
1019 <output>
1020 <port id="0" precision="I64" names="106">
1021 <dim>4</dim>
1022 </port>
1023 </output>
1024 </layer>
1025 <layer id="79" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1026 <input>
1027 <port id="0" precision="FP32">
1028 <dim>-1</dim>
1029 <dim>-1</dim>
1030 <dim>12</dim>
1031 <dim>32</dim>
1032 </port>
1033 <port id="1" precision="I64">
1034 <dim>4</dim>
1035 </port>
1036 </input>
1037 <output>
1038 <port id="2" precision="FP32" names="107">
1039 <dim>-1</dim>
1040 <dim>12</dim>
1041 <dim>-1</dim>
1042 <dim>32</dim>
1043 </port>
1044 </output>
1045 </layer>
1046 <layer id="80" name="Constant_445518" type="Const" version="opset1">
1047 <data element_type="i8" shape="384, 384" offset="12351840" size="147456" />
1048 <output>
1049 <port id="0" precision="I8">
1050 <dim>384</dim>
1051 <dim>384</dim>
1052 </port>
1053 </output>
1054 </layer>
1055 <layer id="81" name="Convert_445519" type="Convert" version="opset1">
1056 <data destination_type="f32" />
1057 <input>
1058 <port id="0" precision="I8">
1059 <dim>384</dim>
1060 <dim>384</dim>
1061 </port>
1062 </input>
1063 <output>
1064 <port id="1" precision="FP32">
1065 <dim>384</dim>
1066 <dim>384</dim>
1067 </port>
1068 </output>
1069 </layer>
1070 <layer id="82" name="Constant_445520" type="Const" version="opset1">
1071 <data element_type="f32" shape="384, 1" offset="12499296" size="1536" />
1072 <output>
1073 <port id="0" precision="FP32">
1074 <dim>384</dim>
1075 <dim>1</dim>
1076 </port>
1077 </output>
1078 </layer>
1079 <layer id="83" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1080 <data auto_broadcast="numpy" />
1081 <input>
1082 <port id="0" precision="FP32">
1083 <dim>384</dim>
1084 <dim>384</dim>
1085 </port>
1086 <port id="1" precision="FP32">
1087 <dim>384</dim>
1088 <dim>1</dim>
1089 </port>
1090 </input>
1091 <output>
1092 <port id="2" precision="FP32">
1093 <dim>384</dim>
1094 <dim>384</dim>
1095 </port>
1096 </output>
1097 </layer>
1098 <layer id="84" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1099 <data transpose_a="false" transpose_b="true" />
1100 <input>
1101 <port id="0" precision="FP32">
1102 <dim>-1</dim>
1103 <dim>-1</dim>
1104 <dim>384</dim>
1105 </port>
1106 <port id="1" precision="FP32">
1107 <dim>384</dim>
1108 <dim>384</dim>
1109 </port>
1110 </input>
1111 <output>
1112 <port id="2" precision="FP32">
1113 <dim>-1</dim>
1114 <dim>-1</dim>
1115 <dim>384</dim>
1116 </port>
1117 </output>
1118 </layer>
1119 <layer id="85" name="Constant_305873" type="Const" version="opset1">
1120 <data element_type="f32" shape="1, 1, 384" offset="12500832" size="1536" />
1121 <output>
1122 <port id="0" precision="FP32">
1123 <dim>1</dim>
1124 <dim>1</dim>
1125 <dim>384</dim>
1126 </port>
1127 </output>
1128 </layer>
1129 <layer id="86" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
1130 <data auto_broadcast="numpy" />
1131 <input>
1132 <port id="0" precision="FP32">
1133 <dim>-1</dim>
1134 <dim>-1</dim>
1135 <dim>384</dim>
1136 </port>
1137 <port id="1" precision="FP32">
1138 <dim>1</dim>
1139 <dim>1</dim>
1140 <dim>384</dim>
1141 </port>
1142 </input>
1143 <output>
1144 <port id="2" precision="FP32" names="110,x.9">
1145 <dim>-1</dim>
1146 <dim>-1</dim>
1147 <dim>384</dim>
1148 </port>
1149 </output>
1150 </layer>
1151 <layer id="87" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
1152 <data element_type="i64" shape="4" offset="12201240" size="32" />
1153 <output>
1154 <port id="0" precision="I64">
1155 <dim>4</dim>
1156 </port>
1157 </output>
1158 </layer>
1159 <layer id="88" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
1160 <data special_zero="true" />
1161 <input>
1162 <port id="0" precision="FP32">
1163 <dim>-1</dim>
1164 <dim>-1</dim>
1165 <dim>384</dim>
1166 </port>
1167 <port id="1" precision="I64">
1168 <dim>4</dim>
1169 </port>
1170 </input>
1171 <output>
1172 <port id="2" precision="FP32" names="114,x.11">
1173 <dim>-1</dim>
1174 <dim>-1</dim>
1175 <dim>12</dim>
1176 <dim>32</dim>
1177 </port>
1178 </output>
1179 </layer>
1180 <layer id="89" name="Constant_301110" type="Const" version="opset1">
1181 <data element_type="i64" shape="4" offset="12201272" size="32" />
1182 <output>
1183 <port id="0" precision="I64" names="115">
1184 <dim>4</dim>
1185 </port>
1186 </output>
1187 </layer>
1188 <layer id="90" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
1189 <input>
1190 <port id="0" precision="FP32">
1191 <dim>-1</dim>
1192 <dim>-1</dim>
1193 <dim>12</dim>
1194 <dim>32</dim>
1195 </port>
1196 <port id="1" precision="I64">
1197 <dim>4</dim>
1198 </port>
1199 </input>
1200 <output>
1201 <port id="2" precision="FP32" names="116">
1202 <dim>-1</dim>
1203 <dim>12</dim>
1204 <dim>-1</dim>
1205 <dim>32</dim>
1206 </port>
1207 </output>
1208 </layer>
1209 <layer id="91" name="Constant_305875" type="Const" version="opset1">
1210 <data element_type="f32" shape="1, 1, 1, 1" offset="12502368" size="4" />
1211 <output>
1212 <port id="0" precision="FP32">
1213 <dim>1</dim>
1214 <dim>1</dim>
1215 <dim>1</dim>
1216 <dim>1</dim>
1217 </port>
1218 </output>
1219 </layer>
1220 <layer id="92" name="31" type="Const" version="opset1">
1221 <data element_type="i64" shape="" offset="12046076" size="8" />
1222 <output>
1223 <port id="0" precision="I64" names="31" />
1224 </output>
1225 </layer>
1226 <layer id="93" name="__module.bert/aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
1227 <input>
1228 <port id="0" precision="I64">
1229 <dim>-1</dim>
1230 <dim>-1</dim>
1231 </port>
1232 <port id="1" precision="I64" />
1233 </input>
1234 <output>
1235 <port id="2" precision="I64" names="63">
1236 <dim>-1</dim>
1237 <dim>1</dim>
1238 <dim>-1</dim>
1239 </port>
1240 </output>
1241 </layer>
1242 <layer id="94" name="21" type="Const" version="opset1">
1243 <data element_type="i64" shape="" offset="12502372" size="8" />
1244 <output>
1245 <port id="0" precision="I64" names="21" />
1246 </output>
1247 </layer>
1248 <layer id="95" name="__module.bert/aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
1249 <input>
1250 <port id="0" precision="I64">
1251 <dim>-1</dim>
1252 <dim>1</dim>
1253 <dim>-1</dim>
1254 </port>
1255 <port id="1" precision="I64" />
1256 </input>
1257 <output>
1258 <port id="2" precision="I64" names="64,65">
1259 <dim>-1</dim>
1260 <dim>1</dim>
1261 <dim>1</dim>
1262 <dim>-1</dim>
1263 </port>
1264 </output>
1265 </layer>
1266 <layer id="96" name="Constant_305970" type="Const" version="opset1">
1267 <data element_type="i64" shape="1" offset="12046068" size="8" />
1268 <output>
1269 <port id="0" precision="I64">
1270 <dim>1</dim>
1271 </port>
1272 </output>
1273 </layer>
1274 <layer id="97" name="Constant_305971" type="Const" version="opset1">
1275 <data element_type="i64" shape="" offset="12046068" size="8" />
1276 <output>
1277 <port id="0" precision="I64" />
1278 </output>
1279 </layer>
1280 <layer id="98" name="Gather_305972" type="Gather" version="opset8">
1281 <data batch_dims="0" />
1282 <input>
1283 <port id="0" precision="I64">
1284 <dim>2</dim>
1285 </port>
1286 <port id="1" precision="I64">
1287 <dim>1</dim>
1288 </port>
1289 <port id="2" precision="I64" />
1290 </input>
1291 <output>
1292 <port id="3" precision="I64" names="60">
1293 <dim>1</dim>
1294 </port>
1295 </output>
1296 </layer>
1297 <layer id="99" name="Constant_305305" type="Const" version="opset1">
1298 <data element_type="i64" shape="1" offset="12046076" size="8" />
1299 <output>
1300 <port id="0" precision="I64">
1301 <dim>1</dim>
1302 </port>
1303 </output>
1304 </layer>
1305 <layer id="100" name="Constant_306080" type="Const" version="opset1">
1306 <data element_type="i64" shape="2" offset="12502380" size="16" />
1307 <output>
1308 <port id="0" precision="I64">
1309 <dim>2</dim>
1310 </port>
1311 </output>
1312 </layer>
1313 <layer id="101" name="Constant_306081" type="Const" version="opset1">
1314 <data element_type="i64" shape="" offset="12046068" size="8" />
1315 <output>
1316 <port id="0" precision="I64" />
1317 </output>
1318 </layer>
1319 <layer id="102" name="Gather_306082" type="Gather" version="opset8">
1320 <data batch_dims="0" />
1321 <input>
1322 <port id="0" precision="I64">
1323 <dim>2</dim>
1324 </port>
1325 <port id="1" precision="I64">
1326 <dim>2</dim>
1327 </port>
1328 <port id="2" precision="I64" />
1329 </input>
1330 <output>
1331 <port id="3" precision="I64">
1332 <dim>2</dim>
1333 </port>
1334 </output>
1335 </layer>
1336 <layer id="103" name="__module.bert/prim::ListConstruct/Concat" type="Concat" version="opset1">
1337 <data axis="0" />
1338 <input>
1339 <port id="0" precision="I64">
1340 <dim>1</dim>
1341 </port>
1342 <port id="1" precision="I64">
1343 <dim>1</dim>
1344 </port>
1345 <port id="2" precision="I64">
1346 <dim>2</dim>
1347 </port>
1348 </input>
1349 <output>
1350 <port id="3" precision="I64" names="66">
1351 <dim>4</dim>
1352 </port>
1353 </output>
1354 </layer>
1355 <layer id="104" name="__module.bert/aten::expand/Broadcast" type="Broadcast" version="opset3">
1356 <data mode="bidirectional" />
1357 <input>
1358 <port id="0" precision="I64">
1359 <dim>-1</dim>
1360 <dim>1</dim>
1361 <dim>1</dim>
1362 <dim>-1</dim>
1363 </port>
1364 <port id="1" precision="I64">
1365 <dim>4</dim>
1366 </port>
1367 </input>
1368 <output>
1369 <port id="2" precision="I64" names="67">
1370 <dim>-1</dim>
1371 <dim>1</dim>
1372 <dim>-1</dim>
1373 <dim>-1</dim>
1374 </port>
1375 </output>
1376 </layer>
1377 <layer id="105" name="__module.bert/aten::to/Convert" type="Convert" version="opset1">
1378 <data destination_type="f32" />
1379 <input>
1380 <port id="0" precision="I64">
1381 <dim>-1</dim>
1382 <dim>1</dim>
1383 <dim>-1</dim>
1384 <dim>-1</dim>
1385 </port>
1386 </input>
1387 <output>
1388 <port id="1" precision="FP32" names="68">
1389 <dim>-1</dim>
1390 <dim>1</dim>
1391 <dim>-1</dim>
1392 <dim>-1</dim>
1393 </port>
1394 </output>
1395 </layer>
1396 <layer id="106" name="Constant_305874" type="Const" version="opset1">
1397 <data element_type="f32" shape="1, 1, 1, 1" offset="12502368" size="4" />
1398 <output>
1399 <port id="0" precision="FP32">
1400 <dim>1</dim>
1401 <dim>1</dim>
1402 <dim>1</dim>
1403 <dim>1</dim>
1404 </port>
1405 </output>
1406 </layer>
1407 <layer id="107" name="__module.bert/aten::rsub/Multiply" type="Multiply" version="opset1">
1408 <data auto_broadcast="numpy" />
1409 <input>
1410 <port id="0" precision="FP32">
1411 <dim>-1</dim>
1412 <dim>1</dim>
1413 <dim>-1</dim>
1414 <dim>-1</dim>
1415 </port>
1416 <port id="1" precision="FP32">
1417 <dim>1</dim>
1418 <dim>1</dim>
1419 <dim>1</dim>
1420 <dim>1</dim>
1421 </port>
1422 </input>
1423 <output>
1424 <port id="2" precision="FP32">
1425 <dim>-1</dim>
1426 <dim>1</dim>
1427 <dim>-1</dim>
1428 <dim>-1</dim>
1429 </port>
1430 </output>
1431 </layer>
1432 <layer id="108" name="__module.bert/aten::rsub/Subtract" type="Subtract" version="opset1">
1433 <data auto_broadcast="numpy" />
1434 <input>
1435 <port id="0" precision="FP32">
1436 <dim>1</dim>
1437 <dim>1</dim>
1438 <dim>1</dim>
1439 <dim>1</dim>
1440 </port>
1441 <port id="1" precision="FP32">
1442 <dim>-1</dim>
1443 <dim>1</dim>
1444 <dim>-1</dim>
1445 <dim>-1</dim>
1446 </port>
1447 </input>
1448 <output>
1449 <port id="2" precision="FP32" names="69,inverted_mask">
1450 <dim>-1</dim>
1451 <dim>1</dim>
1452 <dim>-1</dim>
1453 <dim>-1</dim>
1454 </port>
1455 </output>
1456 </layer>
1457 <layer id="109" name="__module.bert/aten::to/Convert_1" type="Convert" version="opset1">
1458 <data destination_type="boolean" />
1459 <input>
1460 <port id="0" precision="FP32">
1461 <dim>-1</dim>
1462 <dim>1</dim>
1463 <dim>-1</dim>
1464 <dim>-1</dim>
1465 </port>
1466 </input>
1467 <output>
1468 <port id="1" precision="BOOL" names="70">
1469 <dim>-1</dim>
1470 <dim>1</dim>
1471 <dim>-1</dim>
1472 <dim>-1</dim>
1473 </port>
1474 </output>
1475 </layer>
1476 <layer id="110" name="__module.bert/aten::masked_fill/ConvertLike" type="Const" version="opset1">
1477 <data element_type="f32" shape="" offset="12502396" size="4" />
1478 <output>
1479 <port id="0" precision="FP32" />
1480 </output>
1481 </layer>
1482 <layer id="111" name="__module.bert/aten::masked_fill/Select" type="Select" version="opset1">
1483 <data auto_broadcast="numpy" />
1484 <input>
1485 <port id="0" precision="BOOL">
1486 <dim>-1</dim>
1487 <dim>1</dim>
1488 <dim>-1</dim>
1489 <dim>-1</dim>
1490 </port>
1491 <port id="1" precision="FP32" />
1492 <port id="2" precision="FP32">
1493 <dim>-1</dim>
1494 <dim>1</dim>
1495 <dim>-1</dim>
1496 <dim>-1</dim>
1497 </port>
1498 </input>
1499 <output>
1500 <port id="3" precision="FP32" names="71">
1501 <dim>-1</dim>
1502 <dim>1</dim>
1503 <dim>-1</dim>
1504 <dim>-1</dim>
1505 </port>
1506 </output>
1507 </layer>
1508 <layer id="112" name="__module.bert.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1509 <data causal="false" />
1510 <input>
1511 <port id="0" precision="FP32">
1512 <dim>-1</dim>
1513 <dim>12</dim>
1514 <dim>-1</dim>
1515 <dim>32</dim>
1516 </port>
1517 <port id="1" precision="FP32">
1518 <dim>-1</dim>
1519 <dim>12</dim>
1520 <dim>-1</dim>
1521 <dim>32</dim>
1522 </port>
1523 <port id="2" precision="FP32">
1524 <dim>-1</dim>
1525 <dim>12</dim>
1526 <dim>-1</dim>
1527 <dim>32</dim>
1528 </port>
1529 <port id="3" precision="FP32">
1530 <dim>-1</dim>
1531 <dim>1</dim>
1532 <dim>-1</dim>
1533 <dim>-1</dim>
1534 </port>
1535 </input>
1536 <output>
1537 <port id="4" precision="FP32" names="117,attn_output.1">
1538 <dim>-1</dim>
1539 <dim>12</dim>
1540 <dim>-1</dim>
1541 <dim>32</dim>
1542 </port>
1543 </output>
1544 </layer>
1545 <layer id="113" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1546 <data element_type="i32" shape="4" offset="12502400" size="16" />
1547 <output>
1548 <port id="0" precision="I32">
1549 <dim>4</dim>
1550 </port>
1551 </output>
1552 </layer>
1553 <layer id="114" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1554 <input>
1555 <port id="0" precision="FP32">
1556 <dim>-1</dim>
1557 <dim>12</dim>
1558 <dim>-1</dim>
1559 <dim>32</dim>
1560 </port>
1561 <port id="1" precision="I32">
1562 <dim>4</dim>
1563 </port>
1564 </input>
1565 <output>
1566 <port id="2" precision="FP32" names="118,attn_output.3">
1567 <dim>-1</dim>
1568 <dim>-1</dim>
1569 <dim>12</dim>
1570 <dim>32</dim>
1571 </port>
1572 </output>
1573 </layer>
1574 <layer id="115" name="Constant_305977" type="Const" version="opset1">
1575 <data element_type="i64" shape="3" offset="12502416" size="24" />
1576 <output>
1577 <port id="0" precision="I64">
1578 <dim>3</dim>
1579 </port>
1580 </output>
1581 </layer>
1582 <layer id="116" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1583 <data special_zero="true" />
1584 <input>
1585 <port id="0" precision="FP32">
1586 <dim>-1</dim>
1587 <dim>-1</dim>
1588 <dim>12</dim>
1589 <dim>32</dim>
1590 </port>
1591 <port id="1" precision="I64">
1592 <dim>3</dim>
1593 </port>
1594 </input>
1595 <output>
1596 <port id="2" precision="FP32" names="120">
1597 <dim>-1</dim>
1598 <dim>-1</dim>
1599 <dim>384</dim>
1600 </port>
1601 </output>
1602 </layer>
1603 <layer id="117" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1604 <data element_type="f32" shape="1, 1, 384" offset="12502440" size="1536" />
1605 <output>
1606 <port id="0" precision="FP32">
1607 <dim>1</dim>
1608 <dim>1</dim>
1609 <dim>384</dim>
1610 </port>
1611 </output>
1612 </layer>
1613 <layer id="118" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1614 <data auto_broadcast="numpy" />
1615 <input>
1616 <port id="0" precision="FP32">
1617 <dim>-1</dim>
1618 <dim>-1</dim>
1619 <dim>384</dim>
1620 </port>
1621 <port id="1" precision="FP32">
1622 <dim>1</dim>
1623 <dim>1</dim>
1624 <dim>384</dim>
1625 </port>
1626 </input>
1627 <output>
1628 <port id="2" precision="FP32">
1629 <dim>-1</dim>
1630 <dim>-1</dim>
1631 <dim>384</dim>
1632 </port>
1633 </output>
1634 </layer>
1635 <layer id="119" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
1636 <data element_type="f32" shape="" offset="12503976" size="4" />
1637 <output>
1638 <port id="0" precision="FP32" />
1639 </output>
1640 </layer>
1641 <layer id="120" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
1642 <data element_type="f32" shape="" offset="12503980" size="4" />
1643 <output>
1644 <port id="0" precision="FP32" />
1645 </output>
1646 </layer>
1647 <layer id="121" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
1648 <data element_type="f32" shape="" offset="12503976" size="4" />
1649 <output>
1650 <port id="0" precision="FP32" />
1651 </output>
1652 </layer>
1653 <layer id="122" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
1654 <data element_type="f32" shape="" offset="12503980" size="4" />
1655 <output>
1656 <port id="0" precision="FP32" />
1657 </output>
1658 </layer>
1659 <layer id="123" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
1660 <data levels="256" auto_broadcast="numpy" />
1661 <input>
1662 <port id="0" precision="FP32">
1663 <dim>-1</dim>
1664 <dim>-1</dim>
1665 <dim>384</dim>
1666 </port>
1667 <port id="1" precision="FP32" />
1668 <port id="2" precision="FP32" />
1669 <port id="3" precision="FP32" />
1670 <port id="4" precision="FP32" />
1671 </input>
1672 <output>
1673 <port id="5" precision="FP32">
1674 <dim>-1</dim>
1675 <dim>-1</dim>
1676 <dim>384</dim>
1677 </port>
1678 </output>
1679 </layer>
1680 <layer id="124" name="Constant_445522" type="Const" version="opset1">
1681 <data element_type="i8" shape="384, 384" offset="12503984" size="147456" />
1682 <output>
1683 <port id="0" precision="I8">
1684 <dim>384</dim>
1685 <dim>384</dim>
1686 </port>
1687 </output>
1688 </layer>
1689 <layer id="125" name="Convert_445523" type="Convert" version="opset1">
1690 <data destination_type="f32" />
1691 <input>
1692 <port id="0" precision="I8">
1693 <dim>384</dim>
1694 <dim>384</dim>
1695 </port>
1696 </input>
1697 <output>
1698 <port id="1" precision="FP32">
1699 <dim>384</dim>
1700 <dim>384</dim>
1701 </port>
1702 </output>
1703 </layer>
1704 <layer id="126" name="Constant_445524" type="Const" version="opset1">
1705 <data element_type="f32" shape="384, 1" offset="12651440" size="1536" />
1706 <output>
1707 <port id="0" precision="FP32">
1708 <dim>384</dim>
1709 <dim>1</dim>
1710 </port>
1711 </output>
1712 </layer>
1713 <layer id="127" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1714 <data auto_broadcast="numpy" />
1715 <input>
1716 <port id="0" precision="FP32">
1717 <dim>384</dim>
1718 <dim>384</dim>
1719 </port>
1720 <port id="1" precision="FP32">
1721 <dim>384</dim>
1722 <dim>1</dim>
1723 </port>
1724 </input>
1725 <output>
1726 <port id="2" precision="FP32">
1727 <dim>384</dim>
1728 <dim>384</dim>
1729 </port>
1730 </output>
1731 </layer>
1732 <layer id="128" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1733 <data transpose_a="false" transpose_b="true" />
1734 <input>
1735 <port id="0" precision="FP32">
1736 <dim>-1</dim>
1737 <dim>-1</dim>
1738 <dim>384</dim>
1739 </port>
1740 <port id="1" precision="FP32">
1741 <dim>384</dim>
1742 <dim>384</dim>
1743 </port>
1744 </input>
1745 <output>
1746 <port id="2" precision="FP32">
1747 <dim>-1</dim>
1748 <dim>-1</dim>
1749 <dim>384</dim>
1750 </port>
1751 </output>
1752 </layer>
1753 <layer id="129" name="Constant_305876" type="Const" version="opset1">
1754 <data element_type="f32" shape="1, 1, 384" offset="12652976" size="1536" />
1755 <output>
1756 <port id="0" precision="FP32">
1757 <dim>1</dim>
1758 <dim>1</dim>
1759 <dim>384</dim>
1760 </port>
1761 </output>
1762 </layer>
1763 <layer id="130" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1764 <data auto_broadcast="numpy" />
1765 <input>
1766 <port id="0" precision="FP32">
1767 <dim>-1</dim>
1768 <dim>-1</dim>
1769 <dim>384</dim>
1770 </port>
1771 <port id="1" precision="FP32">
1772 <dim>1</dim>
1773 <dim>1</dim>
1774 <dim>384</dim>
1775 </port>
1776 </input>
1777 <output>
1778 <port id="2" precision="FP32" names="125,input.3">
1779 <dim>-1</dim>
1780 <dim>-1</dim>
1781 <dim>384</dim>
1782 </port>
1783 </output>
1784 </layer>
1785 <layer id="131" name="__module.bert.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1786 <data auto_broadcast="numpy" />
1787 <input>
1788 <port id="0" precision="FP32">
1789 <dim>-1</dim>
1790 <dim>-1</dim>
1791 <dim>384</dim>
1792 </port>
1793 <port id="1" precision="FP32">
1794 <dim>-1</dim>
1795 <dim>-1</dim>
1796 <dim>384</dim>
1797 </port>
1798 </input>
1799 <output>
1800 <port id="2" precision="FP32" names="127">
1801 <dim>-1</dim>
1802 <dim>-1</dim>
1803 <dim>384</dim>
1804 </port>
1805 </output>
1806 </layer>
1807 <layer id="132" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1808 <data element_type="i32" shape="1" offset="12046084" size="4" />
1809 <output>
1810 <port id="0" precision="I32">
1811 <dim>1</dim>
1812 </port>
1813 </output>
1814 </layer>
1815 <layer id="133" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1816 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1817 <input>
1818 <port id="0" precision="FP32">
1819 <dim>-1</dim>
1820 <dim>-1</dim>
1821 <dim>384</dim>
1822 </port>
1823 <port id="1" precision="I32">
1824 <dim>1</dim>
1825 </port>
1826 </input>
1827 <output>
1828 <port id="2" precision="FP32">
1829 <dim>-1</dim>
1830 <dim>-1</dim>
1831 <dim>384</dim>
1832 </port>
1833 </output>
1834 </layer>
1835 <layer id="134" name="Constant_305877" type="Const" version="opset1">
1836 <data element_type="f32" shape="1, 1, 384" offset="12654512" size="1536" />
1837 <output>
1838 <port id="0" precision="FP32">
1839 <dim>1</dim>
1840 <dim>1</dim>
1841 <dim>384</dim>
1842 </port>
1843 </output>
1844 </layer>
1845 <layer id="135" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1846 <data auto_broadcast="numpy" />
1847 <input>
1848 <port id="0" precision="FP32">
1849 <dim>-1</dim>
1850 <dim>-1</dim>
1851 <dim>384</dim>
1852 </port>
1853 <port id="1" precision="FP32">
1854 <dim>1</dim>
1855 <dim>1</dim>
1856 <dim>384</dim>
1857 </port>
1858 </input>
1859 <output>
1860 <port id="2" precision="FP32">
1861 <dim>-1</dim>
1862 <dim>-1</dim>
1863 <dim>384</dim>
1864 </port>
1865 </output>
1866 </layer>
1867 <layer id="136" name="Constant_305878" type="Const" version="opset1">
1868 <data element_type="f32" shape="1, 1, 384" offset="12656048" size="1536" />
1869 <output>
1870 <port id="0" precision="FP32">
1871 <dim>1</dim>
1872 <dim>1</dim>
1873 <dim>384</dim>
1874 </port>
1875 </output>
1876 </layer>
1877 <layer id="137" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1878 <data auto_broadcast="numpy" />
1879 <input>
1880 <port id="0" precision="FP32">
1881 <dim>-1</dim>
1882 <dim>-1</dim>
1883 <dim>384</dim>
1884 </port>
1885 <port id="1" precision="FP32">
1886 <dim>1</dim>
1887 <dim>1</dim>
1888 <dim>384</dim>
1889 </port>
1890 </input>
1891 <output>
1892 <port id="2" precision="FP32" names="131,input_tensor.1">
1893 <dim>-1</dim>
1894 <dim>-1</dim>
1895 <dim>384</dim>
1896 </port>
1897 </output>
1898 </layer>
1899 <layer id="138" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1900 <data element_type="f32" shape="1, 1, 384" offset="12657584" size="1536" />
1901 <output>
1902 <port id="0" precision="FP32">
1903 <dim>1</dim>
1904 <dim>1</dim>
1905 <dim>384</dim>
1906 </port>
1907 </output>
1908 </layer>
1909 <layer id="139" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1910 <data auto_broadcast="numpy" />
1911 <input>
1912 <port id="0" precision="FP32">
1913 <dim>-1</dim>
1914 <dim>-1</dim>
1915 <dim>384</dim>
1916 </port>
1917 <port id="1" precision="FP32">
1918 <dim>1</dim>
1919 <dim>1</dim>
1920 <dim>384</dim>
1921 </port>
1922 </input>
1923 <output>
1924 <port id="2" precision="FP32">
1925 <dim>-1</dim>
1926 <dim>-1</dim>
1927 <dim>384</dim>
1928 </port>
1929 </output>
1930 </layer>
1931 <layer id="140" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
1932 <data element_type="f32" shape="" offset="12659120" size="4" />
1933 <output>
1934 <port id="0" precision="FP32" />
1935 </output>
1936 </layer>
1937 <layer id="141" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
1938 <data element_type="f32" shape="" offset="12659124" size="4" />
1939 <output>
1940 <port id="0" precision="FP32" />
1941 </output>
1942 </layer>
1943 <layer id="142" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
1944 <data element_type="f32" shape="" offset="12659120" size="4" />
1945 <output>
1946 <port id="0" precision="FP32" />
1947 </output>
1948 </layer>
1949 <layer id="143" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
1950 <data element_type="f32" shape="" offset="12659124" size="4" />
1951 <output>
1952 <port id="0" precision="FP32" />
1953 </output>
1954 </layer>
1955 <layer id="144" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
1956 <data levels="256" auto_broadcast="numpy" />
1957 <input>
1958 <port id="0" precision="FP32">
1959 <dim>-1</dim>
1960 <dim>-1</dim>
1961 <dim>384</dim>
1962 </port>
1963 <port id="1" precision="FP32" />
1964 <port id="2" precision="FP32" />
1965 <port id="3" precision="FP32" />
1966 <port id="4" precision="FP32" />
1967 </input>
1968 <output>
1969 <port id="5" precision="FP32">
1970 <dim>-1</dim>
1971 <dim>-1</dim>
1972 <dim>384</dim>
1973 </port>
1974 </output>
1975 </layer>
1976 <layer id="145" name="Constant_445526" type="Const" version="opset1">
1977 <data element_type="i8" shape="1536, 384" offset="12659128" size="589824" />
1978 <output>
1979 <port id="0" precision="I8">
1980 <dim>1536</dim>
1981 <dim>384</dim>
1982 </port>
1983 </output>
1984 </layer>
1985 <layer id="146" name="Convert_445527" type="Convert" version="opset1">
1986 <data destination_type="f32" />
1987 <input>
1988 <port id="0" precision="I8">
1989 <dim>1536</dim>
1990 <dim>384</dim>
1991 </port>
1992 </input>
1993 <output>
1994 <port id="1" precision="FP32">
1995 <dim>1536</dim>
1996 <dim>384</dim>
1997 </port>
1998 </output>
1999 </layer>
2000 <layer id="147" name="Constant_445528" type="Const" version="opset1">
2001 <data element_type="f32" shape="1536, 1" offset="13248952" size="6144" />
2002 <output>
2003 <port id="0" precision="FP32">
2004 <dim>1536</dim>
2005 <dim>1</dim>
2006 </port>
2007 </output>
2008 </layer>
2009 <layer id="148" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2010 <data auto_broadcast="numpy" />
2011 <input>
2012 <port id="0" precision="FP32">
2013 <dim>1536</dim>
2014 <dim>384</dim>
2015 </port>
2016 <port id="1" precision="FP32">
2017 <dim>1536</dim>
2018 <dim>1</dim>
2019 </port>
2020 </input>
2021 <output>
2022 <port id="2" precision="FP32">
2023 <dim>1536</dim>
2024 <dim>384</dim>
2025 </port>
2026 </output>
2027 </layer>
2028 <layer id="149" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2029 <data transpose_a="false" transpose_b="true" />
2030 <input>
2031 <port id="0" precision="FP32">
2032 <dim>-1</dim>
2033 <dim>-1</dim>
2034 <dim>384</dim>
2035 </port>
2036 <port id="1" precision="FP32">
2037 <dim>1536</dim>
2038 <dim>384</dim>
2039 </port>
2040 </input>
2041 <output>
2042 <port id="2" precision="FP32">
2043 <dim>-1</dim>
2044 <dim>-1</dim>
2045 <dim>1536</dim>
2046 </port>
2047 </output>
2048 </layer>
2049 <layer id="150" name="Constant_305879" type="Const" version="opset1">
2050 <data element_type="f32" shape="1, 1, 1536" offset="13255096" size="6144" />
2051 <output>
2052 <port id="0" precision="FP32">
2053 <dim>1</dim>
2054 <dim>1</dim>
2055 <dim>1536</dim>
2056 </port>
2057 </output>
2058 </layer>
2059 <layer id="151" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2060 <data auto_broadcast="numpy" />
2061 <input>
2062 <port id="0" precision="FP32">
2063 <dim>-1</dim>
2064 <dim>-1</dim>
2065 <dim>1536</dim>
2066 </port>
2067 <port id="1" precision="FP32">
2068 <dim>1</dim>
2069 <dim>1</dim>
2070 <dim>1536</dim>
2071 </port>
2072 </input>
2073 <output>
2074 <port id="2" precision="FP32" names="135">
2075 <dim>-1</dim>
2076 <dim>-1</dim>
2077 <dim>1536</dim>
2078 </port>
2079 </output>
2080 </layer>
2081 <layer id="152" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2082 <data approximation_mode="ERF" />
2083 <input>
2084 <port id="0" precision="FP32">
2085 <dim>-1</dim>
2086 <dim>-1</dim>
2087 <dim>1536</dim>
2088 </port>
2089 </input>
2090 <output>
2091 <port id="1" precision="FP32" names="136">
2092 <dim>-1</dim>
2093 <dim>-1</dim>
2094 <dim>1536</dim>
2095 </port>
2096 </output>
2097 </layer>
2098 <layer id="153" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2099 <data element_type="f32" shape="1, 1, 1536" offset="13261240" size="6144" />
2100 <output>
2101 <port id="0" precision="FP32">
2102 <dim>1</dim>
2103 <dim>1</dim>
2104 <dim>1536</dim>
2105 </port>
2106 </output>
2107 </layer>
2108 <layer id="154" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2109 <data auto_broadcast="numpy" />
2110 <input>
2111 <port id="0" precision="FP32">
2112 <dim>-1</dim>
2113 <dim>-1</dim>
2114 <dim>1536</dim>
2115 </port>
2116 <port id="1" precision="FP32">
2117 <dim>1</dim>
2118 <dim>1</dim>
2119 <dim>1536</dim>
2120 </port>
2121 </input>
2122 <output>
2123 <port id="2" precision="FP32">
2124 <dim>-1</dim>
2125 <dim>-1</dim>
2126 <dim>1536</dim>
2127 </port>
2128 </output>
2129 </layer>
2130 <layer id="155" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2131 <data element_type="f32" shape="" offset="13267384" size="4" />
2132 <output>
2133 <port id="0" precision="FP32" />
2134 </output>
2135 </layer>
2136 <layer id="156" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2137 <data element_type="f32" shape="" offset="13267388" size="4" />
2138 <output>
2139 <port id="0" precision="FP32" />
2140 </output>
2141 </layer>
2142 <layer id="157" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2143 <data element_type="f32" shape="" offset="13267384" size="4" />
2144 <output>
2145 <port id="0" precision="FP32" />
2146 </output>
2147 </layer>
2148 <layer id="158" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2149 <data element_type="f32" shape="" offset="13267388" size="4" />
2150 <output>
2151 <port id="0" precision="FP32" />
2152 </output>
2153 </layer>
2154 <layer id="159" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2155 <data levels="256" auto_broadcast="numpy" />
2156 <input>
2157 <port id="0" precision="FP32">
2158 <dim>-1</dim>
2159 <dim>-1</dim>
2160 <dim>1536</dim>
2161 </port>
2162 <port id="1" precision="FP32" />
2163 <port id="2" precision="FP32" />
2164 <port id="3" precision="FP32" />
2165 <port id="4" precision="FP32" />
2166 </input>
2167 <output>
2168 <port id="5" precision="FP32">
2169 <dim>-1</dim>
2170 <dim>-1</dim>
2171 <dim>1536</dim>
2172 </port>
2173 </output>
2174 </layer>
2175 <layer id="160" name="Constant_445530" type="Const" version="opset1">
2176 <data element_type="i8" shape="384, 1536" offset="13267392" size="589824" />
2177 <output>
2178 <port id="0" precision="I8">
2179 <dim>384</dim>
2180 <dim>1536</dim>
2181 </port>
2182 </output>
2183 </layer>
2184 <layer id="161" name="Convert_445531" type="Convert" version="opset1">
2185 <data destination_type="f32" />
2186 <input>
2187 <port id="0" precision="I8">
2188 <dim>384</dim>
2189 <dim>1536</dim>
2190 </port>
2191 </input>
2192 <output>
2193 <port id="1" precision="FP32">
2194 <dim>384</dim>
2195 <dim>1536</dim>
2196 </port>
2197 </output>
2198 </layer>
2199 <layer id="162" name="Constant_445532" type="Const" version="opset1">
2200 <data element_type="f32" shape="384, 1" offset="13857216" size="1536" />
2201 <output>
2202 <port id="0" precision="FP32">
2203 <dim>384</dim>
2204 <dim>1</dim>
2205 </port>
2206 </output>
2207 </layer>
2208 <layer id="163" name="__module.bert.encoder.layer.0.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2209 <data auto_broadcast="numpy" />
2210 <input>
2211 <port id="0" precision="FP32">
2212 <dim>384</dim>
2213 <dim>1536</dim>
2214 </port>
2215 <port id="1" precision="FP32">
2216 <dim>384</dim>
2217 <dim>1</dim>
2218 </port>
2219 </input>
2220 <output>
2221 <port id="2" precision="FP32">
2222 <dim>384</dim>
2223 <dim>1536</dim>
2224 </port>
2225 </output>
2226 </layer>
2227 <layer id="164" name="__module.bert.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2228 <data transpose_a="false" transpose_b="true" />
2229 <input>
2230 <port id="0" precision="FP32">
2231 <dim>-1</dim>
2232 <dim>-1</dim>
2233 <dim>1536</dim>
2234 </port>
2235 <port id="1" precision="FP32">
2236 <dim>384</dim>
2237 <dim>1536</dim>
2238 </port>
2239 </input>
2240 <output>
2241 <port id="2" precision="FP32">
2242 <dim>-1</dim>
2243 <dim>-1</dim>
2244 <dim>384</dim>
2245 </port>
2246 </output>
2247 </layer>
2248 <layer id="165" name="Constant_305880" type="Const" version="opset1">
2249 <data element_type="f32" shape="1, 1, 384" offset="13858752" size="1536" />
2250 <output>
2251 <port id="0" precision="FP32">
2252 <dim>1</dim>
2253 <dim>1</dim>
2254 <dim>384</dim>
2255 </port>
2256 </output>
2257 </layer>
2258 <layer id="166" name="__module.bert.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
2259 <data auto_broadcast="numpy" />
2260 <input>
2261 <port id="0" precision="FP32">
2262 <dim>-1</dim>
2263 <dim>-1</dim>
2264 <dim>384</dim>
2265 </port>
2266 <port id="1" precision="FP32">
2267 <dim>1</dim>
2268 <dim>1</dim>
2269 <dim>384</dim>
2270 </port>
2271 </input>
2272 <output>
2273 <port id="2" precision="FP32" names="141,input.5">
2274 <dim>-1</dim>
2275 <dim>-1</dim>
2276 <dim>384</dim>
2277 </port>
2278 </output>
2279 </layer>
2280 <layer id="167" name="__module.bert.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
2281 <data auto_broadcast="numpy" />
2282 <input>
2283 <port id="0" precision="FP32">
2284 <dim>-1</dim>
2285 <dim>-1</dim>
2286 <dim>384</dim>
2287 </port>
2288 <port id="1" precision="FP32">
2289 <dim>-1</dim>
2290 <dim>-1</dim>
2291 <dim>384</dim>
2292 </port>
2293 </input>
2294 <output>
2295 <port id="2" precision="FP32" names="143">
2296 <dim>-1</dim>
2297 <dim>-1</dim>
2298 <dim>384</dim>
2299 </port>
2300 </output>
2301 </layer>
2302 <layer id="168" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2303 <data element_type="i32" shape="1" offset="12046084" size="4" />
2304 <output>
2305 <port id="0" precision="I32">
2306 <dim>1</dim>
2307 </port>
2308 </output>
2309 </layer>
2310 <layer id="169" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2311 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2312 <input>
2313 <port id="0" precision="FP32">
2314 <dim>-1</dim>
2315 <dim>-1</dim>
2316 <dim>384</dim>
2317 </port>
2318 <port id="1" precision="I32">
2319 <dim>1</dim>
2320 </port>
2321 </input>
2322 <output>
2323 <port id="2" precision="FP32">
2324 <dim>-1</dim>
2325 <dim>-1</dim>
2326 <dim>384</dim>
2327 </port>
2328 </output>
2329 </layer>
2330 <layer id="170" name="Constant_305881" type="Const" version="opset1">
2331 <data element_type="f32" shape="1, 1, 384" offset="13860288" size="1536" />
2332 <output>
2333 <port id="0" precision="FP32">
2334 <dim>1</dim>
2335 <dim>1</dim>
2336 <dim>384</dim>
2337 </port>
2338 </output>
2339 </layer>
2340 <layer id="171" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2341 <data auto_broadcast="numpy" />
2342 <input>
2343 <port id="0" precision="FP32">
2344 <dim>-1</dim>
2345 <dim>-1</dim>
2346 <dim>384</dim>
2347 </port>
2348 <port id="1" precision="FP32">
2349 <dim>1</dim>
2350 <dim>1</dim>
2351 <dim>384</dim>
2352 </port>
2353 </input>
2354 <output>
2355 <port id="2" precision="FP32">
2356 <dim>-1</dim>
2357 <dim>-1</dim>
2358 <dim>384</dim>
2359 </port>
2360 </output>
2361 </layer>
2362 <layer id="172" name="Constant_305882" type="Const" version="opset1">
2363 <data element_type="f32" shape="1, 1, 384" offset="13861824" size="1536" />
2364 <output>
2365 <port id="0" precision="FP32">
2366 <dim>1</dim>
2367 <dim>1</dim>
2368 <dim>384</dim>
2369 </port>
2370 </output>
2371 </layer>
2372 <layer id="173" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2373 <data auto_broadcast="numpy" />
2374 <input>
2375 <port id="0" precision="FP32">
2376 <dim>-1</dim>
2377 <dim>-1</dim>
2378 <dim>384</dim>
2379 </port>
2380 <port id="1" precision="FP32">
2381 <dim>1</dim>
2382 <dim>1</dim>
2383 <dim>384</dim>
2384 </port>
2385 </input>
2386 <output>
2387 <port id="2" precision="FP32" names="147,hidden_states.7">
2388 <dim>-1</dim>
2389 <dim>-1</dim>
2390 <dim>384</dim>
2391 </port>
2392 </output>
2393 </layer>
2394 <layer id="174" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2395 <data element_type="f32" shape="1, 1, 384" offset="13863360" size="1536" />
2396 <output>
2397 <port id="0" precision="FP32">
2398 <dim>1</dim>
2399 <dim>1</dim>
2400 <dim>384</dim>
2401 </port>
2402 </output>
2403 </layer>
2404 <layer id="175" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2405 <data auto_broadcast="numpy" />
2406 <input>
2407 <port id="0" precision="FP32">
2408 <dim>-1</dim>
2409 <dim>-1</dim>
2410 <dim>384</dim>
2411 </port>
2412 <port id="1" precision="FP32">
2413 <dim>1</dim>
2414 <dim>1</dim>
2415 <dim>384</dim>
2416 </port>
2417 </input>
2418 <output>
2419 <port id="2" precision="FP32">
2420 <dim>-1</dim>
2421 <dim>-1</dim>
2422 <dim>384</dim>
2423 </port>
2424 </output>
2425 </layer>
2426 <layer id="176" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2427 <data element_type="f32" shape="" offset="13864896" size="4" />
2428 <output>
2429 <port id="0" precision="FP32" />
2430 </output>
2431 </layer>
2432 <layer id="177" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2433 <data element_type="f32" shape="" offset="13864900" size="4" />
2434 <output>
2435 <port id="0" precision="FP32" />
2436 </output>
2437 </layer>
2438 <layer id="178" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2439 <data element_type="f32" shape="" offset="13864896" size="4" />
2440 <output>
2441 <port id="0" precision="FP32" />
2442 </output>
2443 </layer>
2444 <layer id="179" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2445 <data element_type="f32" shape="" offset="13864900" size="4" />
2446 <output>
2447 <port id="0" precision="FP32" />
2448 </output>
2449 </layer>
2450 <layer id="180" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2451 <data levels="256" auto_broadcast="numpy" />
2452 <input>
2453 <port id="0" precision="FP32">
2454 <dim>-1</dim>
2455 <dim>-1</dim>
2456 <dim>384</dim>
2457 </port>
2458 <port id="1" precision="FP32" />
2459 <port id="2" precision="FP32" />
2460 <port id="3" precision="FP32" />
2461 <port id="4" precision="FP32" />
2462 </input>
2463 <output>
2464 <port id="5" precision="FP32">
2465 <dim>-1</dim>
2466 <dim>-1</dim>
2467 <dim>384</dim>
2468 </port>
2469 </output>
2470 </layer>
2471 <layer id="181" name="Constant_445534" type="Const" version="opset1">
2472 <data element_type="i8" shape="384, 384" offset="13864904" size="147456" />
2473 <output>
2474 <port id="0" precision="I8">
2475 <dim>384</dim>
2476 <dim>384</dim>
2477 </port>
2478 </output>
2479 </layer>
2480 <layer id="182" name="Convert_445535" type="Convert" version="opset1">
2481 <data destination_type="f32" />
2482 <input>
2483 <port id="0" precision="I8">
2484 <dim>384</dim>
2485 <dim>384</dim>
2486 </port>
2487 </input>
2488 <output>
2489 <port id="1" precision="FP32">
2490 <dim>384</dim>
2491 <dim>384</dim>
2492 </port>
2493 </output>
2494 </layer>
2495 <layer id="183" name="Constant_445536" type="Const" version="opset1">
2496 <data element_type="f32" shape="384, 1" offset="14012360" size="1536" />
2497 <output>
2498 <port id="0" precision="FP32">
2499 <dim>384</dim>
2500 <dim>1</dim>
2501 </port>
2502 </output>
2503 </layer>
2504 <layer id="184" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2505 <data auto_broadcast="numpy" />
2506 <input>
2507 <port id="0" precision="FP32">
2508 <dim>384</dim>
2509 <dim>384</dim>
2510 </port>
2511 <port id="1" precision="FP32">
2512 <dim>384</dim>
2513 <dim>1</dim>
2514 </port>
2515 </input>
2516 <output>
2517 <port id="2" precision="FP32">
2518 <dim>384</dim>
2519 <dim>384</dim>
2520 </port>
2521 </output>
2522 </layer>
2523 <layer id="185" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2524 <data transpose_a="false" transpose_b="true" />
2525 <input>
2526 <port id="0" precision="FP32">
2527 <dim>-1</dim>
2528 <dim>-1</dim>
2529 <dim>384</dim>
2530 </port>
2531 <port id="1" precision="FP32">
2532 <dim>384</dim>
2533 <dim>384</dim>
2534 </port>
2535 </input>
2536 <output>
2537 <port id="2" precision="FP32">
2538 <dim>-1</dim>
2539 <dim>-1</dim>
2540 <dim>384</dim>
2541 </port>
2542 </output>
2543 </layer>
2544 <layer id="186" name="Constant_305883" type="Const" version="opset1">
2545 <data element_type="f32" shape="1, 1, 384" offset="14013896" size="1536" />
2546 <output>
2547 <port id="0" precision="FP32">
2548 <dim>1</dim>
2549 <dim>1</dim>
2550 <dim>384</dim>
2551 </port>
2552 </output>
2553 </layer>
2554 <layer id="187" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2555 <data auto_broadcast="numpy" />
2556 <input>
2557 <port id="0" precision="FP32">
2558 <dim>-1</dim>
2559 <dim>-1</dim>
2560 <dim>384</dim>
2561 </port>
2562 <port id="1" precision="FP32">
2563 <dim>1</dim>
2564 <dim>1</dim>
2565 <dim>384</dim>
2566 </port>
2567 </input>
2568 <output>
2569 <port id="2" precision="FP32" names="160,x.13">
2570 <dim>-1</dim>
2571 <dim>-1</dim>
2572 <dim>384</dim>
2573 </port>
2574 </output>
2575 </layer>
2576 <layer id="188" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2577 <data element_type="f32" shape="" offset="14015432" size="4" />
2578 <output>
2579 <port id="0" precision="FP32" />
2580 </output>
2581 </layer>
2582 <layer id="189" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2583 <data element_type="f32" shape="" offset="14015436" size="4" />
2584 <output>
2585 <port id="0" precision="FP32" />
2586 </output>
2587 </layer>
2588 <layer id="190" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2589 <data element_type="f32" shape="" offset="14015432" size="4" />
2590 <output>
2591 <port id="0" precision="FP32" />
2592 </output>
2593 </layer>
2594 <layer id="191" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2595 <data element_type="f32" shape="" offset="14015436" size="4" />
2596 <output>
2597 <port id="0" precision="FP32" />
2598 </output>
2599 </layer>
2600 <layer id="192" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2601 <data levels="256" auto_broadcast="numpy" />
2602 <input>
2603 <port id="0" precision="FP32">
2604 <dim>-1</dim>
2605 <dim>-1</dim>
2606 <dim>384</dim>
2607 </port>
2608 <port id="1" precision="FP32" />
2609 <port id="2" precision="FP32" />
2610 <port id="3" precision="FP32" />
2611 <port id="4" precision="FP32" />
2612 </input>
2613 <output>
2614 <port id="5" precision="FP32">
2615 <dim>-1</dim>
2616 <dim>-1</dim>
2617 <dim>384</dim>
2618 </port>
2619 </output>
2620 </layer>
2621 <layer id="193" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2622 <data element_type="i64" shape="4" offset="12201240" size="32" />
2623 <output>
2624 <port id="0" precision="I64">
2625 <dim>4</dim>
2626 </port>
2627 </output>
2628 </layer>
2629 <layer id="194" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2630 <data special_zero="true" />
2631 <input>
2632 <port id="0" precision="FP32">
2633 <dim>-1</dim>
2634 <dim>-1</dim>
2635 <dim>384</dim>
2636 </port>
2637 <port id="1" precision="I64">
2638 <dim>4</dim>
2639 </port>
2640 </input>
2641 <output>
2642 <port id="2" precision="FP32" names="164,x.15">
2643 <dim>-1</dim>
2644 <dim>-1</dim>
2645 <dim>12</dim>
2646 <dim>32</dim>
2647 </port>
2648 </output>
2649 </layer>
2650 <layer id="195" name="Constant_301289" type="Const" version="opset1">
2651 <data element_type="i64" shape="4" offset="12201272" size="32" />
2652 <output>
2653 <port id="0" precision="I64" names="165">
2654 <dim>4</dim>
2655 </port>
2656 </output>
2657 </layer>
2658 <layer id="196" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2659 <input>
2660 <port id="0" precision="FP32">
2661 <dim>-1</dim>
2662 <dim>-1</dim>
2663 <dim>12</dim>
2664 <dim>32</dim>
2665 </port>
2666 <port id="1" precision="I64">
2667 <dim>4</dim>
2668 </port>
2669 </input>
2670 <output>
2671 <port id="2" precision="FP32" names="166">
2672 <dim>-1</dim>
2673 <dim>12</dim>
2674 <dim>-1</dim>
2675 <dim>32</dim>
2676 </port>
2677 </output>
2678 </layer>
2679 <layer id="197" name="Constant_445538" type="Const" version="opset1">
2680 <data element_type="i8" shape="384, 384" offset="14015440" size="147456" />
2681 <output>
2682 <port id="0" precision="I8">
2683 <dim>384</dim>
2684 <dim>384</dim>
2685 </port>
2686 </output>
2687 </layer>
2688 <layer id="198" name="Convert_445539" type="Convert" version="opset1">
2689 <data destination_type="f32" />
2690 <input>
2691 <port id="0" precision="I8">
2692 <dim>384</dim>
2693 <dim>384</dim>
2694 </port>
2695 </input>
2696 <output>
2697 <port id="1" precision="FP32">
2698 <dim>384</dim>
2699 <dim>384</dim>
2700 </port>
2701 </output>
2702 </layer>
2703 <layer id="199" name="Constant_445540" type="Const" version="opset1">
2704 <data element_type="f32" shape="384, 1" offset="14162896" size="1536" />
2705 <output>
2706 <port id="0" precision="FP32">
2707 <dim>384</dim>
2708 <dim>1</dim>
2709 </port>
2710 </output>
2711 </layer>
2712 <layer id="200" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2713 <data auto_broadcast="numpy" />
2714 <input>
2715 <port id="0" precision="FP32">
2716 <dim>384</dim>
2717 <dim>384</dim>
2718 </port>
2719 <port id="1" precision="FP32">
2720 <dim>384</dim>
2721 <dim>1</dim>
2722 </port>
2723 </input>
2724 <output>
2725 <port id="2" precision="FP32">
2726 <dim>384</dim>
2727 <dim>384</dim>
2728 </port>
2729 </output>
2730 </layer>
2731 <layer id="201" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2732 <data transpose_a="false" transpose_b="true" />
2733 <input>
2734 <port id="0" precision="FP32">
2735 <dim>-1</dim>
2736 <dim>-1</dim>
2737 <dim>384</dim>
2738 </port>
2739 <port id="1" precision="FP32">
2740 <dim>384</dim>
2741 <dim>384</dim>
2742 </port>
2743 </input>
2744 <output>
2745 <port id="2" precision="FP32">
2746 <dim>-1</dim>
2747 <dim>-1</dim>
2748 <dim>384</dim>
2749 </port>
2750 </output>
2751 </layer>
2752 <layer id="202" name="Constant_305884" type="Const" version="opset1">
2753 <data element_type="f32" shape="1, 1, 384" offset="14164432" size="1536" />
2754 <output>
2755 <port id="0" precision="FP32">
2756 <dim>1</dim>
2757 <dim>1</dim>
2758 <dim>384</dim>
2759 </port>
2760 </output>
2761 </layer>
2762 <layer id="203" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2763 <data auto_broadcast="numpy" />
2764 <input>
2765 <port id="0" precision="FP32">
2766 <dim>-1</dim>
2767 <dim>-1</dim>
2768 <dim>384</dim>
2769 </port>
2770 <port id="1" precision="FP32">
2771 <dim>1</dim>
2772 <dim>1</dim>
2773 <dim>384</dim>
2774 </port>
2775 </input>
2776 <output>
2777 <port id="2" precision="FP32" names="169,x.17">
2778 <dim>-1</dim>
2779 <dim>-1</dim>
2780 <dim>384</dim>
2781 </port>
2782 </output>
2783 </layer>
2784 <layer id="204" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2785 <data element_type="f32" shape="" offset="14165968" size="4" />
2786 <output>
2787 <port id="0" precision="FP32" />
2788 </output>
2789 </layer>
2790 <layer id="205" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2791 <data element_type="f32" shape="" offset="14165972" size="4" />
2792 <output>
2793 <port id="0" precision="FP32" />
2794 </output>
2795 </layer>
2796 <layer id="206" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2797 <data element_type="f32" shape="" offset="14165968" size="4" />
2798 <output>
2799 <port id="0" precision="FP32" />
2800 </output>
2801 </layer>
2802 <layer id="207" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2803 <data element_type="f32" shape="" offset="14165972" size="4" />
2804 <output>
2805 <port id="0" precision="FP32" />
2806 </output>
2807 </layer>
2808 <layer id="208" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2809 <data levels="256" auto_broadcast="numpy" />
2810 <input>
2811 <port id="0" precision="FP32">
2812 <dim>-1</dim>
2813 <dim>-1</dim>
2814 <dim>384</dim>
2815 </port>
2816 <port id="1" precision="FP32" />
2817 <port id="2" precision="FP32" />
2818 <port id="3" precision="FP32" />
2819 <port id="4" precision="FP32" />
2820 </input>
2821 <output>
2822 <port id="5" precision="FP32">
2823 <dim>-1</dim>
2824 <dim>-1</dim>
2825 <dim>384</dim>
2826 </port>
2827 </output>
2828 </layer>
2829 <layer id="209" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2830 <data element_type="i64" shape="4" offset="12201240" size="32" />
2831 <output>
2832 <port id="0" precision="I64">
2833 <dim>4</dim>
2834 </port>
2835 </output>
2836 </layer>
2837 <layer id="210" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2838 <data special_zero="true" />
2839 <input>
2840 <port id="0" precision="FP32">
2841 <dim>-1</dim>
2842 <dim>-1</dim>
2843 <dim>384</dim>
2844 </port>
2845 <port id="1" precision="I64">
2846 <dim>4</dim>
2847 </port>
2848 </input>
2849 <output>
2850 <port id="2" precision="FP32" names="173,x.19">
2851 <dim>-1</dim>
2852 <dim>-1</dim>
2853 <dim>12</dim>
2854 <dim>32</dim>
2855 </port>
2856 </output>
2857 </layer>
2858 <layer id="211" name="Constant_301312" type="Const" version="opset1">
2859 <data element_type="i64" shape="4" offset="12201272" size="32" />
2860 <output>
2861 <port id="0" precision="I64" names="174">
2862 <dim>4</dim>
2863 </port>
2864 </output>
2865 </layer>
2866 <layer id="212" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2867 <input>
2868 <port id="0" precision="FP32">
2869 <dim>-1</dim>
2870 <dim>-1</dim>
2871 <dim>12</dim>
2872 <dim>32</dim>
2873 </port>
2874 <port id="1" precision="I64">
2875 <dim>4</dim>
2876 </port>
2877 </input>
2878 <output>
2879 <port id="2" precision="FP32" names="175">
2880 <dim>-1</dim>
2881 <dim>12</dim>
2882 <dim>-1</dim>
2883 <dim>32</dim>
2884 </port>
2885 </output>
2886 </layer>
2887 <layer id="213" name="Constant_445542" type="Const" version="opset1">
2888 <data element_type="i8" shape="384, 384" offset="14165976" size="147456" />
2889 <output>
2890 <port id="0" precision="I8">
2891 <dim>384</dim>
2892 <dim>384</dim>
2893 </port>
2894 </output>
2895 </layer>
2896 <layer id="214" name="Convert_445543" type="Convert" version="opset1">
2897 <data destination_type="f32" />
2898 <input>
2899 <port id="0" precision="I8">
2900 <dim>384</dim>
2901 <dim>384</dim>
2902 </port>
2903 </input>
2904 <output>
2905 <port id="1" precision="FP32">
2906 <dim>384</dim>
2907 <dim>384</dim>
2908 </port>
2909 </output>
2910 </layer>
2911 <layer id="215" name="Constant_445544" type="Const" version="opset1">
2912 <data element_type="f32" shape="384, 1" offset="14313432" size="1536" />
2913 <output>
2914 <port id="0" precision="FP32">
2915 <dim>384</dim>
2916 <dim>1</dim>
2917 </port>
2918 </output>
2919 </layer>
2920 <layer id="216" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2921 <data auto_broadcast="numpy" />
2922 <input>
2923 <port id="0" precision="FP32">
2924 <dim>384</dim>
2925 <dim>384</dim>
2926 </port>
2927 <port id="1" precision="FP32">
2928 <dim>384</dim>
2929 <dim>1</dim>
2930 </port>
2931 </input>
2932 <output>
2933 <port id="2" precision="FP32">
2934 <dim>384</dim>
2935 <dim>384</dim>
2936 </port>
2937 </output>
2938 </layer>
2939 <layer id="217" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
2940 <data transpose_a="false" transpose_b="true" />
2941 <input>
2942 <port id="0" precision="FP32">
2943 <dim>-1</dim>
2944 <dim>-1</dim>
2945 <dim>384</dim>
2946 </port>
2947 <port id="1" precision="FP32">
2948 <dim>384</dim>
2949 <dim>384</dim>
2950 </port>
2951 </input>
2952 <output>
2953 <port id="2" precision="FP32">
2954 <dim>-1</dim>
2955 <dim>-1</dim>
2956 <dim>384</dim>
2957 </port>
2958 </output>
2959 </layer>
2960 <layer id="218" name="Constant_305885" type="Const" version="opset1">
2961 <data element_type="f32" shape="1, 1, 384" offset="14314968" size="1536" />
2962 <output>
2963 <port id="0" precision="FP32">
2964 <dim>1</dim>
2965 <dim>1</dim>
2966 <dim>384</dim>
2967 </port>
2968 </output>
2969 </layer>
2970 <layer id="219" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
2971 <data auto_broadcast="numpy" />
2972 <input>
2973 <port id="0" precision="FP32">
2974 <dim>-1</dim>
2975 <dim>-1</dim>
2976 <dim>384</dim>
2977 </port>
2978 <port id="1" precision="FP32">
2979 <dim>1</dim>
2980 <dim>1</dim>
2981 <dim>384</dim>
2982 </port>
2983 </input>
2984 <output>
2985 <port id="2" precision="FP32" names="178,x.21">
2986 <dim>-1</dim>
2987 <dim>-1</dim>
2988 <dim>384</dim>
2989 </port>
2990 </output>
2991 </layer>
2992 <layer id="220" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
2993 <data element_type="i64" shape="4" offset="12201240" size="32" />
2994 <output>
2995 <port id="0" precision="I64">
2996 <dim>4</dim>
2997 </port>
2998 </output>
2999 </layer>
3000 <layer id="221" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3001 <data special_zero="true" />
3002 <input>
3003 <port id="0" precision="FP32">
3004 <dim>-1</dim>
3005 <dim>-1</dim>
3006 <dim>384</dim>
3007 </port>
3008 <port id="1" precision="I64">
3009 <dim>4</dim>
3010 </port>
3011 </input>
3012 <output>
3013 <port id="2" precision="FP32" names="182,x.23">
3014 <dim>-1</dim>
3015 <dim>-1</dim>
3016 <dim>12</dim>
3017 <dim>32</dim>
3018 </port>
3019 </output>
3020 </layer>
3021 <layer id="222" name="Constant_301335" type="Const" version="opset1">
3022 <data element_type="i64" shape="4" offset="12201272" size="32" />
3023 <output>
3024 <port id="0" precision="I64" names="183">
3025 <dim>4</dim>
3026 </port>
3027 </output>
3028 </layer>
3029 <layer id="223" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3030 <input>
3031 <port id="0" precision="FP32">
3032 <dim>-1</dim>
3033 <dim>-1</dim>
3034 <dim>12</dim>
3035 <dim>32</dim>
3036 </port>
3037 <port id="1" precision="I64">
3038 <dim>4</dim>
3039 </port>
3040 </input>
3041 <output>
3042 <port id="2" precision="FP32" names="184">
3043 <dim>-1</dim>
3044 <dim>12</dim>
3045 <dim>-1</dim>
3046 <dim>32</dim>
3047 </port>
3048 </output>
3049 </layer>
3050 <layer id="224" name="__module.bert.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3051 <data causal="false" />
3052 <input>
3053 <port id="0" precision="FP32">
3054 <dim>-1</dim>
3055 <dim>12</dim>
3056 <dim>-1</dim>
3057 <dim>32</dim>
3058 </port>
3059 <port id="1" precision="FP32">
3060 <dim>-1</dim>
3061 <dim>12</dim>
3062 <dim>-1</dim>
3063 <dim>32</dim>
3064 </port>
3065 <port id="2" precision="FP32">
3066 <dim>-1</dim>
3067 <dim>12</dim>
3068 <dim>-1</dim>
3069 <dim>32</dim>
3070 </port>
3071 <port id="3" precision="FP32">
3072 <dim>-1</dim>
3073 <dim>1</dim>
3074 <dim>-1</dim>
3075 <dim>-1</dim>
3076 </port>
3077 </input>
3078 <output>
3079 <port id="4" precision="FP32" names="185,attn_output.5">
3080 <dim>-1</dim>
3081 <dim>12</dim>
3082 <dim>-1</dim>
3083 <dim>32</dim>
3084 </port>
3085 </output>
3086 </layer>
3087 <layer id="225" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3088 <data element_type="i32" shape="4" offset="12502400" size="16" />
3089 <output>
3090 <port id="0" precision="I32">
3091 <dim>4</dim>
3092 </port>
3093 </output>
3094 </layer>
3095 <layer id="226" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3096 <input>
3097 <port id="0" precision="FP32">
3098 <dim>-1</dim>
3099 <dim>12</dim>
3100 <dim>-1</dim>
3101 <dim>32</dim>
3102 </port>
3103 <port id="1" precision="I32">
3104 <dim>4</dim>
3105 </port>
3106 </input>
3107 <output>
3108 <port id="2" precision="FP32" names="186,attn_output.7">
3109 <dim>-1</dim>
3110 <dim>-1</dim>
3111 <dim>12</dim>
3112 <dim>32</dim>
3113 </port>
3114 </output>
3115 </layer>
3116 <layer id="227" name="Constant_305978" type="Const" version="opset1">
3117 <data element_type="i64" shape="3" offset="12502416" size="24" />
3118 <output>
3119 <port id="0" precision="I64">
3120 <dim>3</dim>
3121 </port>
3122 </output>
3123 </layer>
3124 <layer id="228" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3125 <data special_zero="true" />
3126 <input>
3127 <port id="0" precision="FP32">
3128 <dim>-1</dim>
3129 <dim>-1</dim>
3130 <dim>12</dim>
3131 <dim>32</dim>
3132 </port>
3133 <port id="1" precision="I64">
3134 <dim>3</dim>
3135 </port>
3136 </input>
3137 <output>
3138 <port id="2" precision="FP32" names="188">
3139 <dim>-1</dim>
3140 <dim>-1</dim>
3141 <dim>384</dim>
3142 </port>
3143 </output>
3144 </layer>
3145 <layer id="229" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3146 <data element_type="f32" shape="1, 1, 384" offset="14316504" size="1536" />
3147 <output>
3148 <port id="0" precision="FP32">
3149 <dim>1</dim>
3150 <dim>1</dim>
3151 <dim>384</dim>
3152 </port>
3153 </output>
3154 </layer>
3155 <layer id="230" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3156 <data auto_broadcast="numpy" />
3157 <input>
3158 <port id="0" precision="FP32">
3159 <dim>-1</dim>
3160 <dim>-1</dim>
3161 <dim>384</dim>
3162 </port>
3163 <port id="1" precision="FP32">
3164 <dim>1</dim>
3165 <dim>1</dim>
3166 <dim>384</dim>
3167 </port>
3168 </input>
3169 <output>
3170 <port id="2" precision="FP32">
3171 <dim>-1</dim>
3172 <dim>-1</dim>
3173 <dim>384</dim>
3174 </port>
3175 </output>
3176 </layer>
3177 <layer id="231" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3178 <data element_type="f32" shape="" offset="14318040" size="4" />
3179 <output>
3180 <port id="0" precision="FP32" />
3181 </output>
3182 </layer>
3183 <layer id="232" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3184 <data element_type="f32" shape="" offset="14318044" size="4" />
3185 <output>
3186 <port id="0" precision="FP32" />
3187 </output>
3188 </layer>
3189 <layer id="233" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3190 <data element_type="f32" shape="" offset="14318040" size="4" />
3191 <output>
3192 <port id="0" precision="FP32" />
3193 </output>
3194 </layer>
3195 <layer id="234" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3196 <data element_type="f32" shape="" offset="14318044" size="4" />
3197 <output>
3198 <port id="0" precision="FP32" />
3199 </output>
3200 </layer>
3201 <layer id="235" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3202 <data levels="256" auto_broadcast="numpy" />
3203 <input>
3204 <port id="0" precision="FP32">
3205 <dim>-1</dim>
3206 <dim>-1</dim>
3207 <dim>384</dim>
3208 </port>
3209 <port id="1" precision="FP32" />
3210 <port id="2" precision="FP32" />
3211 <port id="3" precision="FP32" />
3212 <port id="4" precision="FP32" />
3213 </input>
3214 <output>
3215 <port id="5" precision="FP32">
3216 <dim>-1</dim>
3217 <dim>-1</dim>
3218 <dim>384</dim>
3219 </port>
3220 </output>
3221 </layer>
3222 <layer id="236" name="Constant_445546" type="Const" version="opset1">
3223 <data element_type="i8" shape="384, 384" offset="14318048" size="147456" />
3224 <output>
3225 <port id="0" precision="I8">
3226 <dim>384</dim>
3227 <dim>384</dim>
3228 </port>
3229 </output>
3230 </layer>
3231 <layer id="237" name="Convert_445547" type="Convert" version="opset1">
3232 <data destination_type="f32" />
3233 <input>
3234 <port id="0" precision="I8">
3235 <dim>384</dim>
3236 <dim>384</dim>
3237 </port>
3238 </input>
3239 <output>
3240 <port id="1" precision="FP32">
3241 <dim>384</dim>
3242 <dim>384</dim>
3243 </port>
3244 </output>
3245 </layer>
3246 <layer id="238" name="Constant_445548" type="Const" version="opset1">
3247 <data element_type="f32" shape="384, 1" offset="14465504" size="1536" />
3248 <output>
3249 <port id="0" precision="FP32">
3250 <dim>384</dim>
3251 <dim>1</dim>
3252 </port>
3253 </output>
3254 </layer>
3255 <layer id="239" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3256 <data auto_broadcast="numpy" />
3257 <input>
3258 <port id="0" precision="FP32">
3259 <dim>384</dim>
3260 <dim>384</dim>
3261 </port>
3262 <port id="1" precision="FP32">
3263 <dim>384</dim>
3264 <dim>1</dim>
3265 </port>
3266 </input>
3267 <output>
3268 <port id="2" precision="FP32">
3269 <dim>384</dim>
3270 <dim>384</dim>
3271 </port>
3272 </output>
3273 </layer>
3274 <layer id="240" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3275 <data transpose_a="false" transpose_b="true" />
3276 <input>
3277 <port id="0" precision="FP32">
3278 <dim>-1</dim>
3279 <dim>-1</dim>
3280 <dim>384</dim>
3281 </port>
3282 <port id="1" precision="FP32">
3283 <dim>384</dim>
3284 <dim>384</dim>
3285 </port>
3286 </input>
3287 <output>
3288 <port id="2" precision="FP32">
3289 <dim>-1</dim>
3290 <dim>-1</dim>
3291 <dim>384</dim>
3292 </port>
3293 </output>
3294 </layer>
3295 <layer id="241" name="Constant_305886" type="Const" version="opset1">
3296 <data element_type="f32" shape="1, 1, 384" offset="14467040" size="1536" />
3297 <output>
3298 <port id="0" precision="FP32">
3299 <dim>1</dim>
3300 <dim>1</dim>
3301 <dim>384</dim>
3302 </port>
3303 </output>
3304 </layer>
3305 <layer id="242" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3306 <data auto_broadcast="numpy" />
3307 <input>
3308 <port id="0" precision="FP32">
3309 <dim>-1</dim>
3310 <dim>-1</dim>
3311 <dim>384</dim>
3312 </port>
3313 <port id="1" precision="FP32">
3314 <dim>1</dim>
3315 <dim>1</dim>
3316 <dim>384</dim>
3317 </port>
3318 </input>
3319 <output>
3320 <port id="2" precision="FP32" names="193,input.7">
3321 <dim>-1</dim>
3322 <dim>-1</dim>
3323 <dim>384</dim>
3324 </port>
3325 </output>
3326 </layer>
3327 <layer id="243" name="__module.bert.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
3328 <data auto_broadcast="numpy" />
3329 <input>
3330 <port id="0" precision="FP32">
3331 <dim>-1</dim>
3332 <dim>-1</dim>
3333 <dim>384</dim>
3334 </port>
3335 <port id="1" precision="FP32">
3336 <dim>-1</dim>
3337 <dim>-1</dim>
3338 <dim>384</dim>
3339 </port>
3340 </input>
3341 <output>
3342 <port id="2" precision="FP32" names="195">
3343 <dim>-1</dim>
3344 <dim>-1</dim>
3345 <dim>384</dim>
3346 </port>
3347 </output>
3348 </layer>
3349 <layer id="244" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3350 <data element_type="i32" shape="1" offset="12046084" size="4" />
3351 <output>
3352 <port id="0" precision="I32">
3353 <dim>1</dim>
3354 </port>
3355 </output>
3356 </layer>
3357 <layer id="245" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3358 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3359 <input>
3360 <port id="0" precision="FP32">
3361 <dim>-1</dim>
3362 <dim>-1</dim>
3363 <dim>384</dim>
3364 </port>
3365 <port id="1" precision="I32">
3366 <dim>1</dim>
3367 </port>
3368 </input>
3369 <output>
3370 <port id="2" precision="FP32">
3371 <dim>-1</dim>
3372 <dim>-1</dim>
3373 <dim>384</dim>
3374 </port>
3375 </output>
3376 </layer>
3377 <layer id="246" name="Constant_305887" type="Const" version="opset1">
3378 <data element_type="f32" shape="1, 1, 384" offset="14468576" size="1536" />
3379 <output>
3380 <port id="0" precision="FP32">
3381 <dim>1</dim>
3382 <dim>1</dim>
3383 <dim>384</dim>
3384 </port>
3385 </output>
3386 </layer>
3387 <layer id="247" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3388 <data auto_broadcast="numpy" />
3389 <input>
3390 <port id="0" precision="FP32">
3391 <dim>-1</dim>
3392 <dim>-1</dim>
3393 <dim>384</dim>
3394 </port>
3395 <port id="1" precision="FP32">
3396 <dim>1</dim>
3397 <dim>1</dim>
3398 <dim>384</dim>
3399 </port>
3400 </input>
3401 <output>
3402 <port id="2" precision="FP32">
3403 <dim>-1</dim>
3404 <dim>-1</dim>
3405 <dim>384</dim>
3406 </port>
3407 </output>
3408 </layer>
3409 <layer id="248" name="Constant_305888" type="Const" version="opset1">
3410 <data element_type="f32" shape="1, 1, 384" offset="14470112" size="1536" />
3411 <output>
3412 <port id="0" precision="FP32">
3413 <dim>1</dim>
3414 <dim>1</dim>
3415 <dim>384</dim>
3416 </port>
3417 </output>
3418 </layer>
3419 <layer id="249" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3420 <data auto_broadcast="numpy" />
3421 <input>
3422 <port id="0" precision="FP32">
3423 <dim>-1</dim>
3424 <dim>-1</dim>
3425 <dim>384</dim>
3426 </port>
3427 <port id="1" precision="FP32">
3428 <dim>1</dim>
3429 <dim>1</dim>
3430 <dim>384</dim>
3431 </port>
3432 </input>
3433 <output>
3434 <port id="2" precision="FP32" names="199,input_tensor.3">
3435 <dim>-1</dim>
3436 <dim>-1</dim>
3437 <dim>384</dim>
3438 </port>
3439 </output>
3440 </layer>
3441 <layer id="250" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3442 <data element_type="f32" shape="1, 1, 384" offset="14471648" size="1536" />
3443 <output>
3444 <port id="0" precision="FP32">
3445 <dim>1</dim>
3446 <dim>1</dim>
3447 <dim>384</dim>
3448 </port>
3449 </output>
3450 </layer>
3451 <layer id="251" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3452 <data auto_broadcast="numpy" />
3453 <input>
3454 <port id="0" precision="FP32">
3455 <dim>-1</dim>
3456 <dim>-1</dim>
3457 <dim>384</dim>
3458 </port>
3459 <port id="1" precision="FP32">
3460 <dim>1</dim>
3461 <dim>1</dim>
3462 <dim>384</dim>
3463 </port>
3464 </input>
3465 <output>
3466 <port id="2" precision="FP32">
3467 <dim>-1</dim>
3468 <dim>-1</dim>
3469 <dim>384</dim>
3470 </port>
3471 </output>
3472 </layer>
3473 <layer id="252" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3474 <data element_type="f32" shape="" offset="14473184" size="4" />
3475 <output>
3476 <port id="0" precision="FP32" />
3477 </output>
3478 </layer>
3479 <layer id="253" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3480 <data element_type="f32" shape="" offset="14473188" size="4" />
3481 <output>
3482 <port id="0" precision="FP32" />
3483 </output>
3484 </layer>
3485 <layer id="254" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3486 <data element_type="f32" shape="" offset="14473184" size="4" />
3487 <output>
3488 <port id="0" precision="FP32" />
3489 </output>
3490 </layer>
3491 <layer id="255" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3492 <data element_type="f32" shape="" offset="14473188" size="4" />
3493 <output>
3494 <port id="0" precision="FP32" />
3495 </output>
3496 </layer>
3497 <layer id="256" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3498 <data levels="256" auto_broadcast="numpy" />
3499 <input>
3500 <port id="0" precision="FP32">
3501 <dim>-1</dim>
3502 <dim>-1</dim>
3503 <dim>384</dim>
3504 </port>
3505 <port id="1" precision="FP32" />
3506 <port id="2" precision="FP32" />
3507 <port id="3" precision="FP32" />
3508 <port id="4" precision="FP32" />
3509 </input>
3510 <output>
3511 <port id="5" precision="FP32">
3512 <dim>-1</dim>
3513 <dim>-1</dim>
3514 <dim>384</dim>
3515 </port>
3516 </output>
3517 </layer>
3518 <layer id="257" name="Constant_445550" type="Const" version="opset1">
3519 <data element_type="i8" shape="1536, 384" offset="14473192" size="589824" />
3520 <output>
3521 <port id="0" precision="I8">
3522 <dim>1536</dim>
3523 <dim>384</dim>
3524 </port>
3525 </output>
3526 </layer>
3527 <layer id="258" name="Convert_445551" type="Convert" version="opset1">
3528 <data destination_type="f32" />
3529 <input>
3530 <port id="0" precision="I8">
3531 <dim>1536</dim>
3532 <dim>384</dim>
3533 </port>
3534 </input>
3535 <output>
3536 <port id="1" precision="FP32">
3537 <dim>1536</dim>
3538 <dim>384</dim>
3539 </port>
3540 </output>
3541 </layer>
3542 <layer id="259" name="Constant_445552" type="Const" version="opset1">
3543 <data element_type="f32" shape="1536, 1" offset="15063016" size="6144" />
3544 <output>
3545 <port id="0" precision="FP32">
3546 <dim>1536</dim>
3547 <dim>1</dim>
3548 </port>
3549 </output>
3550 </layer>
3551 <layer id="260" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3552 <data auto_broadcast="numpy" />
3553 <input>
3554 <port id="0" precision="FP32">
3555 <dim>1536</dim>
3556 <dim>384</dim>
3557 </port>
3558 <port id="1" precision="FP32">
3559 <dim>1536</dim>
3560 <dim>1</dim>
3561 </port>
3562 </input>
3563 <output>
3564 <port id="2" precision="FP32">
3565 <dim>1536</dim>
3566 <dim>384</dim>
3567 </port>
3568 </output>
3569 </layer>
3570 <layer id="261" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3571 <data transpose_a="false" transpose_b="true" />
3572 <input>
3573 <port id="0" precision="FP32">
3574 <dim>-1</dim>
3575 <dim>-1</dim>
3576 <dim>384</dim>
3577 </port>
3578 <port id="1" precision="FP32">
3579 <dim>1536</dim>
3580 <dim>384</dim>
3581 </port>
3582 </input>
3583 <output>
3584 <port id="2" precision="FP32">
3585 <dim>-1</dim>
3586 <dim>-1</dim>
3587 <dim>1536</dim>
3588 </port>
3589 </output>
3590 </layer>
3591 <layer id="262" name="Constant_305889" type="Const" version="opset1">
3592 <data element_type="f32" shape="1, 1, 1536" offset="15069160" size="6144" />
3593 <output>
3594 <port id="0" precision="FP32">
3595 <dim>1</dim>
3596 <dim>1</dim>
3597 <dim>1536</dim>
3598 </port>
3599 </output>
3600 </layer>
3601 <layer id="263" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3602 <data auto_broadcast="numpy" />
3603 <input>
3604 <port id="0" precision="FP32">
3605 <dim>-1</dim>
3606 <dim>-1</dim>
3607 <dim>1536</dim>
3608 </port>
3609 <port id="1" precision="FP32">
3610 <dim>1</dim>
3611 <dim>1</dim>
3612 <dim>1536</dim>
3613 </port>
3614 </input>
3615 <output>
3616 <port id="2" precision="FP32" names="203">
3617 <dim>-1</dim>
3618 <dim>-1</dim>
3619 <dim>1536</dim>
3620 </port>
3621 </output>
3622 </layer>
3623 <layer id="264" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3624 <data approximation_mode="ERF" />
3625 <input>
3626 <port id="0" precision="FP32">
3627 <dim>-1</dim>
3628 <dim>-1</dim>
3629 <dim>1536</dim>
3630 </port>
3631 </input>
3632 <output>
3633 <port id="1" precision="FP32" names="204">
3634 <dim>-1</dim>
3635 <dim>-1</dim>
3636 <dim>1536</dim>
3637 </port>
3638 </output>
3639 </layer>
3640 <layer id="265" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3641 <data element_type="f32" shape="1, 1, 1536" offset="15075304" size="6144" />
3642 <output>
3643 <port id="0" precision="FP32">
3644 <dim>1</dim>
3645 <dim>1</dim>
3646 <dim>1536</dim>
3647 </port>
3648 </output>
3649 </layer>
3650 <layer id="266" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3651 <data auto_broadcast="numpy" />
3652 <input>
3653 <port id="0" precision="FP32">
3654 <dim>-1</dim>
3655 <dim>-1</dim>
3656 <dim>1536</dim>
3657 </port>
3658 <port id="1" precision="FP32">
3659 <dim>1</dim>
3660 <dim>1</dim>
3661 <dim>1536</dim>
3662 </port>
3663 </input>
3664 <output>
3665 <port id="2" precision="FP32">
3666 <dim>-1</dim>
3667 <dim>-1</dim>
3668 <dim>1536</dim>
3669 </port>
3670 </output>
3671 </layer>
3672 <layer id="267" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3673 <data element_type="f32" shape="" offset="15081448" size="4" />
3674 <output>
3675 <port id="0" precision="FP32" />
3676 </output>
3677 </layer>
3678 <layer id="268" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3679 <data element_type="f32" shape="" offset="15081452" size="4" />
3680 <output>
3681 <port id="0" precision="FP32" />
3682 </output>
3683 </layer>
3684 <layer id="269" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3685 <data element_type="f32" shape="" offset="15081448" size="4" />
3686 <output>
3687 <port id="0" precision="FP32" />
3688 </output>
3689 </layer>
3690 <layer id="270" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3691 <data element_type="f32" shape="" offset="15081452" size="4" />
3692 <output>
3693 <port id="0" precision="FP32" />
3694 </output>
3695 </layer>
3696 <layer id="271" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3697 <data levels="256" auto_broadcast="numpy" />
3698 <input>
3699 <port id="0" precision="FP32">
3700 <dim>-1</dim>
3701 <dim>-1</dim>
3702 <dim>1536</dim>
3703 </port>
3704 <port id="1" precision="FP32" />
3705 <port id="2" precision="FP32" />
3706 <port id="3" precision="FP32" />
3707 <port id="4" precision="FP32" />
3708 </input>
3709 <output>
3710 <port id="5" precision="FP32">
3711 <dim>-1</dim>
3712 <dim>-1</dim>
3713 <dim>1536</dim>
3714 </port>
3715 </output>
3716 </layer>
3717 <layer id="272" name="Constant_445554" type="Const" version="opset1">
3718 <data element_type="i8" shape="384, 1536" offset="15081456" size="589824" />
3719 <output>
3720 <port id="0" precision="I8">
3721 <dim>384</dim>
3722 <dim>1536</dim>
3723 </port>
3724 </output>
3725 </layer>
3726 <layer id="273" name="Convert_445555" type="Convert" version="opset1">
3727 <data destination_type="f32" />
3728 <input>
3729 <port id="0" precision="I8">
3730 <dim>384</dim>
3731 <dim>1536</dim>
3732 </port>
3733 </input>
3734 <output>
3735 <port id="1" precision="FP32">
3736 <dim>384</dim>
3737 <dim>1536</dim>
3738 </port>
3739 </output>
3740 </layer>
3741 <layer id="274" name="Constant_445556" type="Const" version="opset1">
3742 <data element_type="f32" shape="384, 1" offset="15671280" size="1536" />
3743 <output>
3744 <port id="0" precision="FP32">
3745 <dim>384</dim>
3746 <dim>1</dim>
3747 </port>
3748 </output>
3749 </layer>
3750 <layer id="275" name="__module.bert.encoder.layer.1.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3751 <data auto_broadcast="numpy" />
3752 <input>
3753 <port id="0" precision="FP32">
3754 <dim>384</dim>
3755 <dim>1536</dim>
3756 </port>
3757 <port id="1" precision="FP32">
3758 <dim>384</dim>
3759 <dim>1</dim>
3760 </port>
3761 </input>
3762 <output>
3763 <port id="2" precision="FP32">
3764 <dim>384</dim>
3765 <dim>1536</dim>
3766 </port>
3767 </output>
3768 </layer>
3769 <layer id="276" name="__module.bert.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3770 <data transpose_a="false" transpose_b="true" />
3771 <input>
3772 <port id="0" precision="FP32">
3773 <dim>-1</dim>
3774 <dim>-1</dim>
3775 <dim>1536</dim>
3776 </port>
3777 <port id="1" precision="FP32">
3778 <dim>384</dim>
3779 <dim>1536</dim>
3780 </port>
3781 </input>
3782 <output>
3783 <port id="2" precision="FP32">
3784 <dim>-1</dim>
3785 <dim>-1</dim>
3786 <dim>384</dim>
3787 </port>
3788 </output>
3789 </layer>
3790 <layer id="277" name="Constant_305890" type="Const" version="opset1">
3791 <data element_type="f32" shape="1, 1, 384" offset="15672816" size="1536" />
3792 <output>
3793 <port id="0" precision="FP32">
3794 <dim>1</dim>
3795 <dim>1</dim>
3796 <dim>384</dim>
3797 </port>
3798 </output>
3799 </layer>
3800 <layer id="278" name="__module.bert.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
3801 <data auto_broadcast="numpy" />
3802 <input>
3803 <port id="0" precision="FP32">
3804 <dim>-1</dim>
3805 <dim>-1</dim>
3806 <dim>384</dim>
3807 </port>
3808 <port id="1" precision="FP32">
3809 <dim>1</dim>
3810 <dim>1</dim>
3811 <dim>384</dim>
3812 </port>
3813 </input>
3814 <output>
3815 <port id="2" precision="FP32" names="209,input.9">
3816 <dim>-1</dim>
3817 <dim>-1</dim>
3818 <dim>384</dim>
3819 </port>
3820 </output>
3821 </layer>
3822 <layer id="279" name="__module.bert.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
3823 <data auto_broadcast="numpy" />
3824 <input>
3825 <port id="0" precision="FP32">
3826 <dim>-1</dim>
3827 <dim>-1</dim>
3828 <dim>384</dim>
3829 </port>
3830 <port id="1" precision="FP32">
3831 <dim>-1</dim>
3832 <dim>-1</dim>
3833 <dim>384</dim>
3834 </port>
3835 </input>
3836 <output>
3837 <port id="2" precision="FP32" names="211">
3838 <dim>-1</dim>
3839 <dim>-1</dim>
3840 <dim>384</dim>
3841 </port>
3842 </output>
3843 </layer>
3844 <layer id="280" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3845 <data element_type="i32" shape="1" offset="12046084" size="4" />
3846 <output>
3847 <port id="0" precision="I32">
3848 <dim>1</dim>
3849 </port>
3850 </output>
3851 </layer>
3852 <layer id="281" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3853 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3854 <input>
3855 <port id="0" precision="FP32">
3856 <dim>-1</dim>
3857 <dim>-1</dim>
3858 <dim>384</dim>
3859 </port>
3860 <port id="1" precision="I32">
3861 <dim>1</dim>
3862 </port>
3863 </input>
3864 <output>
3865 <port id="2" precision="FP32">
3866 <dim>-1</dim>
3867 <dim>-1</dim>
3868 <dim>384</dim>
3869 </port>
3870 </output>
3871 </layer>
3872 <layer id="282" name="Constant_305891" type="Const" version="opset1">
3873 <data element_type="f32" shape="1, 1, 384" offset="15674352" size="1536" />
3874 <output>
3875 <port id="0" precision="FP32">
3876 <dim>1</dim>
3877 <dim>1</dim>
3878 <dim>384</dim>
3879 </port>
3880 </output>
3881 </layer>
3882 <layer id="283" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3883 <data auto_broadcast="numpy" />
3884 <input>
3885 <port id="0" precision="FP32">
3886 <dim>-1</dim>
3887 <dim>-1</dim>
3888 <dim>384</dim>
3889 </port>
3890 <port id="1" precision="FP32">
3891 <dim>1</dim>
3892 <dim>1</dim>
3893 <dim>384</dim>
3894 </port>
3895 </input>
3896 <output>
3897 <port id="2" precision="FP32">
3898 <dim>-1</dim>
3899 <dim>-1</dim>
3900 <dim>384</dim>
3901 </port>
3902 </output>
3903 </layer>
3904 <layer id="284" name="Constant_305892" type="Const" version="opset1">
3905 <data element_type="f32" shape="1, 1, 384" offset="15675888" size="1536" />
3906 <output>
3907 <port id="0" precision="FP32">
3908 <dim>1</dim>
3909 <dim>1</dim>
3910 <dim>384</dim>
3911 </port>
3912 </output>
3913 </layer>
3914 <layer id="285" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3915 <data auto_broadcast="numpy" />
3916 <input>
3917 <port id="0" precision="FP32">
3918 <dim>-1</dim>
3919 <dim>-1</dim>
3920 <dim>384</dim>
3921 </port>
3922 <port id="1" precision="FP32">
3923 <dim>1</dim>
3924 <dim>1</dim>
3925 <dim>384</dim>
3926 </port>
3927 </input>
3928 <output>
3929 <port id="2" precision="FP32" names="215,hidden_states.13">
3930 <dim>-1</dim>
3931 <dim>-1</dim>
3932 <dim>384</dim>
3933 </port>
3934 </output>
3935 </layer>
3936 <layer id="286" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3937 <data element_type="f32" shape="1, 1, 384" offset="15677424" size="1536" />
3938 <output>
3939 <port id="0" precision="FP32">
3940 <dim>1</dim>
3941 <dim>1</dim>
3942 <dim>384</dim>
3943 </port>
3944 </output>
3945 </layer>
3946 <layer id="287" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3947 <data auto_broadcast="numpy" />
3948 <input>
3949 <port id="0" precision="FP32">
3950 <dim>-1</dim>
3951 <dim>-1</dim>
3952 <dim>384</dim>
3953 </port>
3954 <port id="1" precision="FP32">
3955 <dim>1</dim>
3956 <dim>1</dim>
3957 <dim>384</dim>
3958 </port>
3959 </input>
3960 <output>
3961 <port id="2" precision="FP32">
3962 <dim>-1</dim>
3963 <dim>-1</dim>
3964 <dim>384</dim>
3965 </port>
3966 </output>
3967 </layer>
3968 <layer id="288" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3969 <data element_type="f32" shape="" offset="15678960" size="4" />
3970 <output>
3971 <port id="0" precision="FP32" />
3972 </output>
3973 </layer>
3974 <layer id="289" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3975 <data element_type="f32" shape="" offset="15678964" size="4" />
3976 <output>
3977 <port id="0" precision="FP32" />
3978 </output>
3979 </layer>
3980 <layer id="290" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3981 <data element_type="f32" shape="" offset="15678960" size="4" />
3982 <output>
3983 <port id="0" precision="FP32" />
3984 </output>
3985 </layer>
3986 <layer id="291" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3987 <data element_type="f32" shape="" offset="15678964" size="4" />
3988 <output>
3989 <port id="0" precision="FP32" />
3990 </output>
3991 </layer>
3992 <layer id="292" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3993 <data levels="256" auto_broadcast="numpy" />
3994 <input>
3995 <port id="0" precision="FP32">
3996 <dim>-1</dim>
3997 <dim>-1</dim>
3998 <dim>384</dim>
3999 </port>
4000 <port id="1" precision="FP32" />
4001 <port id="2" precision="FP32" />
4002 <port id="3" precision="FP32" />
4003 <port id="4" precision="FP32" />
4004 </input>
4005 <output>
4006 <port id="5" precision="FP32">
4007 <dim>-1</dim>
4008 <dim>-1</dim>
4009 <dim>384</dim>
4010 </port>
4011 </output>
4012 </layer>
4013 <layer id="293" name="Constant_445558" type="Const" version="opset1">
4014 <data element_type="i8" shape="384, 384" offset="15678968" size="147456" />
4015 <output>
4016 <port id="0" precision="I8">
4017 <dim>384</dim>
4018 <dim>384</dim>
4019 </port>
4020 </output>
4021 </layer>
4022 <layer id="294" name="Convert_445559" type="Convert" version="opset1">
4023 <data destination_type="f32" />
4024 <input>
4025 <port id="0" precision="I8">
4026 <dim>384</dim>
4027 <dim>384</dim>
4028 </port>
4029 </input>
4030 <output>
4031 <port id="1" precision="FP32">
4032 <dim>384</dim>
4033 <dim>384</dim>
4034 </port>
4035 </output>
4036 </layer>
4037 <layer id="295" name="Constant_445560" type="Const" version="opset1">
4038 <data element_type="f32" shape="384, 1" offset="15826424" size="1536" />
4039 <output>
4040 <port id="0" precision="FP32">
4041 <dim>384</dim>
4042 <dim>1</dim>
4043 </port>
4044 </output>
4045 </layer>
4046 <layer id="296" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4047 <data auto_broadcast="numpy" />
4048 <input>
4049 <port id="0" precision="FP32">
4050 <dim>384</dim>
4051 <dim>384</dim>
4052 </port>
4053 <port id="1" precision="FP32">
4054 <dim>384</dim>
4055 <dim>1</dim>
4056 </port>
4057 </input>
4058 <output>
4059 <port id="2" precision="FP32">
4060 <dim>384</dim>
4061 <dim>384</dim>
4062 </port>
4063 </output>
4064 </layer>
4065 <layer id="297" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
4066 <data transpose_a="false" transpose_b="true" />
4067 <input>
4068 <port id="0" precision="FP32">
4069 <dim>-1</dim>
4070 <dim>-1</dim>
4071 <dim>384</dim>
4072 </port>
4073 <port id="1" precision="FP32">
4074 <dim>384</dim>
4075 <dim>384</dim>
4076 </port>
4077 </input>
4078 <output>
4079 <port id="2" precision="FP32">
4080 <dim>-1</dim>
4081 <dim>-1</dim>
4082 <dim>384</dim>
4083 </port>
4084 </output>
4085 </layer>
4086 <layer id="298" name="Constant_305893" type="Const" version="opset1">
4087 <data element_type="f32" shape="1, 1, 384" offset="15827960" size="1536" />
4088 <output>
4089 <port id="0" precision="FP32">
4090 <dim>1</dim>
4091 <dim>1</dim>
4092 <dim>384</dim>
4093 </port>
4094 </output>
4095 </layer>
4096 <layer id="299" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
4097 <data auto_broadcast="numpy" />
4098 <input>
4099 <port id="0" precision="FP32">
4100 <dim>-1</dim>
4101 <dim>-1</dim>
4102 <dim>384</dim>
4103 </port>
4104 <port id="1" precision="FP32">
4105 <dim>1</dim>
4106 <dim>1</dim>
4107 <dim>384</dim>
4108 </port>
4109 </input>
4110 <output>
4111 <port id="2" precision="FP32" names="228,x.25">
4112 <dim>-1</dim>
4113 <dim>-1</dim>
4114 <dim>384</dim>
4115 </port>
4116 </output>
4117 </layer>
4118 <layer id="300" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4119 <data element_type="f32" shape="" offset="15829496" size="4" />
4120 <output>
4121 <port id="0" precision="FP32" />
4122 </output>
4123 </layer>
4124 <layer id="301" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4125 <data element_type="f32" shape="" offset="15829500" size="4" />
4126 <output>
4127 <port id="0" precision="FP32" />
4128 </output>
4129 </layer>
4130 <layer id="302" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4131 <data element_type="f32" shape="" offset="15829496" size="4" />
4132 <output>
4133 <port id="0" precision="FP32" />
4134 </output>
4135 </layer>
4136 <layer id="303" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4137 <data element_type="f32" shape="" offset="15829500" size="4" />
4138 <output>
4139 <port id="0" precision="FP32" />
4140 </output>
4141 </layer>
4142 <layer id="304" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4143 <data levels="256" auto_broadcast="numpy" />
4144 <input>
4145 <port id="0" precision="FP32">
4146 <dim>-1</dim>
4147 <dim>-1</dim>
4148 <dim>384</dim>
4149 </port>
4150 <port id="1" precision="FP32" />
4151 <port id="2" precision="FP32" />
4152 <port id="3" precision="FP32" />
4153 <port id="4" precision="FP32" />
4154 </input>
4155 <output>
4156 <port id="5" precision="FP32">
4157 <dim>-1</dim>
4158 <dim>-1</dim>
4159 <dim>384</dim>
4160 </port>
4161 </output>
4162 </layer>
4163 <layer id="305" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
4164 <data element_type="i64" shape="4" offset="12201240" size="32" />
4165 <output>
4166 <port id="0" precision="I64">
4167 <dim>4</dim>
4168 </port>
4169 </output>
4170 </layer>
4171 <layer id="306" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
4172 <data special_zero="true" />
4173 <input>
4174 <port id="0" precision="FP32">
4175 <dim>-1</dim>
4176 <dim>-1</dim>
4177 <dim>384</dim>
4178 </port>
4179 <port id="1" precision="I64">
4180 <dim>4</dim>
4181 </port>
4182 </input>
4183 <output>
4184 <port id="2" precision="FP32" names="232,x.27">
4185 <dim>-1</dim>
4186 <dim>-1</dim>
4187 <dim>12</dim>
4188 <dim>32</dim>
4189 </port>
4190 </output>
4191 </layer>
4192 <layer id="307" name="Constant_301512" type="Const" version="opset1">
4193 <data element_type="i64" shape="4" offset="12201272" size="32" />
4194 <output>
4195 <port id="0" precision="I64" names="233">
4196 <dim>4</dim>
4197 </port>
4198 </output>
4199 </layer>
4200 <layer id="308" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
4201 <input>
4202 <port id="0" precision="FP32">
4203 <dim>-1</dim>
4204 <dim>-1</dim>
4205 <dim>12</dim>
4206 <dim>32</dim>
4207 </port>
4208 <port id="1" precision="I64">
4209 <dim>4</dim>
4210 </port>
4211 </input>
4212 <output>
4213 <port id="2" precision="FP32" names="234">
4214 <dim>-1</dim>
4215 <dim>12</dim>
4216 <dim>-1</dim>
4217 <dim>32</dim>
4218 </port>
4219 </output>
4220 </layer>
4221 <layer id="309" name="Constant_445562" type="Const" version="opset1">
4222 <data element_type="i8" shape="384, 384" offset="15829504" size="147456" />
4223 <output>
4224 <port id="0" precision="I8">
4225 <dim>384</dim>
4226 <dim>384</dim>
4227 </port>
4228 </output>
4229 </layer>
4230 <layer id="310" name="Convert_445563" type="Convert" version="opset1">
4231 <data destination_type="f32" />
4232 <input>
4233 <port id="0" precision="I8">
4234 <dim>384</dim>
4235 <dim>384</dim>
4236 </port>
4237 </input>
4238 <output>
4239 <port id="1" precision="FP32">
4240 <dim>384</dim>
4241 <dim>384</dim>
4242 </port>
4243 </output>
4244 </layer>
4245 <layer id="311" name="Constant_445564" type="Const" version="opset1">
4246 <data element_type="f32" shape="384, 1" offset="15976960" size="1536" />
4247 <output>
4248 <port id="0" precision="FP32">
4249 <dim>384</dim>
4250 <dim>1</dim>
4251 </port>
4252 </output>
4253 </layer>
4254 <layer id="312" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4255 <data auto_broadcast="numpy" />
4256 <input>
4257 <port id="0" precision="FP32">
4258 <dim>384</dim>
4259 <dim>384</dim>
4260 </port>
4261 <port id="1" precision="FP32">
4262 <dim>384</dim>
4263 <dim>1</dim>
4264 </port>
4265 </input>
4266 <output>
4267 <port id="2" precision="FP32">
4268 <dim>384</dim>
4269 <dim>384</dim>
4270 </port>
4271 </output>
4272 </layer>
4273 <layer id="313" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
4274 <data transpose_a="false" transpose_b="true" />
4275 <input>
4276 <port id="0" precision="FP32">
4277 <dim>-1</dim>
4278 <dim>-1</dim>
4279 <dim>384</dim>
4280 </port>
4281 <port id="1" precision="FP32">
4282 <dim>384</dim>
4283 <dim>384</dim>
4284 </port>
4285 </input>
4286 <output>
4287 <port id="2" precision="FP32">
4288 <dim>-1</dim>
4289 <dim>-1</dim>
4290 <dim>384</dim>
4291 </port>
4292 </output>
4293 </layer>
4294 <layer id="314" name="Constant_305894" type="Const" version="opset1">
4295 <data element_type="f32" shape="1, 1, 384" offset="15978496" size="1536" />
4296 <output>
4297 <port id="0" precision="FP32">
4298 <dim>1</dim>
4299 <dim>1</dim>
4300 <dim>384</dim>
4301 </port>
4302 </output>
4303 </layer>
4304 <layer id="315" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
4305 <data auto_broadcast="numpy" />
4306 <input>
4307 <port id="0" precision="FP32">
4308 <dim>-1</dim>
4309 <dim>-1</dim>
4310 <dim>384</dim>
4311 </port>
4312 <port id="1" precision="FP32">
4313 <dim>1</dim>
4314 <dim>1</dim>
4315 <dim>384</dim>
4316 </port>
4317 </input>
4318 <output>
4319 <port id="2" precision="FP32" names="237,x.29">
4320 <dim>-1</dim>
4321 <dim>-1</dim>
4322 <dim>384</dim>
4323 </port>
4324 </output>
4325 </layer>
4326 <layer id="316" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4327 <data element_type="f32" shape="" offset="15980032" size="4" />
4328 <output>
4329 <port id="0" precision="FP32" />
4330 </output>
4331 </layer>
4332 <layer id="317" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4333 <data element_type="f32" shape="" offset="15980036" size="4" />
4334 <output>
4335 <port id="0" precision="FP32" />
4336 </output>
4337 </layer>
4338 <layer id="318" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4339 <data element_type="f32" shape="" offset="15980032" size="4" />
4340 <output>
4341 <port id="0" precision="FP32" />
4342 </output>
4343 </layer>
4344 <layer id="319" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4345 <data element_type="f32" shape="" offset="15980036" size="4" />
4346 <output>
4347 <port id="0" precision="FP32" />
4348 </output>
4349 </layer>
4350 <layer id="320" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4351 <data levels="256" auto_broadcast="numpy" />
4352 <input>
4353 <port id="0" precision="FP32">
4354 <dim>-1</dim>
4355 <dim>-1</dim>
4356 <dim>384</dim>
4357 </port>
4358 <port id="1" precision="FP32" />
4359 <port id="2" precision="FP32" />
4360 <port id="3" precision="FP32" />
4361 <port id="4" precision="FP32" />
4362 </input>
4363 <output>
4364 <port id="5" precision="FP32">
4365 <dim>-1</dim>
4366 <dim>-1</dim>
4367 <dim>384</dim>
4368 </port>
4369 </output>
4370 </layer>
4371 <layer id="321" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
4372 <data element_type="i64" shape="4" offset="12201240" size="32" />
4373 <output>
4374 <port id="0" precision="I64">
4375 <dim>4</dim>
4376 </port>
4377 </output>
4378 </layer>
4379 <layer id="322" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
4380 <data special_zero="true" />
4381 <input>
4382 <port id="0" precision="FP32">
4383 <dim>-1</dim>
4384 <dim>-1</dim>
4385 <dim>384</dim>
4386 </port>
4387 <port id="1" precision="I64">
4388 <dim>4</dim>
4389 </port>
4390 </input>
4391 <output>
4392 <port id="2" precision="FP32" names="241,x.31">
4393 <dim>-1</dim>
4394 <dim>-1</dim>
4395 <dim>12</dim>
4396 <dim>32</dim>
4397 </port>
4398 </output>
4399 </layer>
4400 <layer id="323" name="Constant_301535" type="Const" version="opset1">
4401 <data element_type="i64" shape="4" offset="12201272" size="32" />
4402 <output>
4403 <port id="0" precision="I64" names="242">
4404 <dim>4</dim>
4405 </port>
4406 </output>
4407 </layer>
4408 <layer id="324" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
4409 <input>
4410 <port id="0" precision="FP32">
4411 <dim>-1</dim>
4412 <dim>-1</dim>
4413 <dim>12</dim>
4414 <dim>32</dim>
4415 </port>
4416 <port id="1" precision="I64">
4417 <dim>4</dim>
4418 </port>
4419 </input>
4420 <output>
4421 <port id="2" precision="FP32" names="243">
4422 <dim>-1</dim>
4423 <dim>12</dim>
4424 <dim>-1</dim>
4425 <dim>32</dim>
4426 </port>
4427 </output>
4428 </layer>
4429 <layer id="325" name="Constant_445566" type="Const" version="opset1">
4430 <data element_type="i8" shape="384, 384" offset="15980040" size="147456" />
4431 <output>
4432 <port id="0" precision="I8">
4433 <dim>384</dim>
4434 <dim>384</dim>
4435 </port>
4436 </output>
4437 </layer>
4438 <layer id="326" name="Convert_445567" type="Convert" version="opset1">
4439 <data destination_type="f32" />
4440 <input>
4441 <port id="0" precision="I8">
4442 <dim>384</dim>
4443 <dim>384</dim>
4444 </port>
4445 </input>
4446 <output>
4447 <port id="1" precision="FP32">
4448 <dim>384</dim>
4449 <dim>384</dim>
4450 </port>
4451 </output>
4452 </layer>
4453 <layer id="327" name="Constant_445568" type="Const" version="opset1">
4454 <data element_type="f32" shape="384, 1" offset="16127496" size="1536" />
4455 <output>
4456 <port id="0" precision="FP32">
4457 <dim>384</dim>
4458 <dim>1</dim>
4459 </port>
4460 </output>
4461 </layer>
4462 <layer id="328" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4463 <data auto_broadcast="numpy" />
4464 <input>
4465 <port id="0" precision="FP32">
4466 <dim>384</dim>
4467 <dim>384</dim>
4468 </port>
4469 <port id="1" precision="FP32">
4470 <dim>384</dim>
4471 <dim>1</dim>
4472 </port>
4473 </input>
4474 <output>
4475 <port id="2" precision="FP32">
4476 <dim>384</dim>
4477 <dim>384</dim>
4478 </port>
4479 </output>
4480 </layer>
4481 <layer id="329" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
4482 <data transpose_a="false" transpose_b="true" />
4483 <input>
4484 <port id="0" precision="FP32">
4485 <dim>-1</dim>
4486 <dim>-1</dim>
4487 <dim>384</dim>
4488 </port>
4489 <port id="1" precision="FP32">
4490 <dim>384</dim>
4491 <dim>384</dim>
4492 </port>
4493 </input>
4494 <output>
4495 <port id="2" precision="FP32">
4496 <dim>-1</dim>
4497 <dim>-1</dim>
4498 <dim>384</dim>
4499 </port>
4500 </output>
4501 </layer>
4502 <layer id="330" name="Constant_305895" type="Const" version="opset1">
4503 <data element_type="f32" shape="1, 1, 384" offset="16129032" size="1536" />
4504 <output>
4505 <port id="0" precision="FP32">
4506 <dim>1</dim>
4507 <dim>1</dim>
4508 <dim>384</dim>
4509 </port>
4510 </output>
4511 </layer>
4512 <layer id="331" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
4513 <data auto_broadcast="numpy" />
4514 <input>
4515 <port id="0" precision="FP32">
4516 <dim>-1</dim>
4517 <dim>-1</dim>
4518 <dim>384</dim>
4519 </port>
4520 <port id="1" precision="FP32">
4521 <dim>1</dim>
4522 <dim>1</dim>
4523 <dim>384</dim>
4524 </port>
4525 </input>
4526 <output>
4527 <port id="2" precision="FP32" names="246,x.33">
4528 <dim>-1</dim>
4529 <dim>-1</dim>
4530 <dim>384</dim>
4531 </port>
4532 </output>
4533 </layer>
4534 <layer id="332" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
4535 <data element_type="i64" shape="4" offset="12201240" size="32" />
4536 <output>
4537 <port id="0" precision="I64">
4538 <dim>4</dim>
4539 </port>
4540 </output>
4541 </layer>
4542 <layer id="333" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
4543 <data special_zero="true" />
4544 <input>
4545 <port id="0" precision="FP32">
4546 <dim>-1</dim>
4547 <dim>-1</dim>
4548 <dim>384</dim>
4549 </port>
4550 <port id="1" precision="I64">
4551 <dim>4</dim>
4552 </port>
4553 </input>
4554 <output>
4555 <port id="2" precision="FP32" names="250,x.35">
4556 <dim>-1</dim>
4557 <dim>-1</dim>
4558 <dim>12</dim>
4559 <dim>32</dim>
4560 </port>
4561 </output>
4562 </layer>
4563 <layer id="334" name="Constant_301558" type="Const" version="opset1">
4564 <data element_type="i64" shape="4" offset="12201272" size="32" />
4565 <output>
4566 <port id="0" precision="I64" names="251">
4567 <dim>4</dim>
4568 </port>
4569 </output>
4570 </layer>
4571 <layer id="335" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
4572 <input>
4573 <port id="0" precision="FP32">
4574 <dim>-1</dim>
4575 <dim>-1</dim>
4576 <dim>12</dim>
4577 <dim>32</dim>
4578 </port>
4579 <port id="1" precision="I64">
4580 <dim>4</dim>
4581 </port>
4582 </input>
4583 <output>
4584 <port id="2" precision="FP32" names="252">
4585 <dim>-1</dim>
4586 <dim>12</dim>
4587 <dim>-1</dim>
4588 <dim>32</dim>
4589 </port>
4590 </output>
4591 </layer>
4592 <layer id="336" name="__module.bert.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
4593 <data causal="false" />
4594 <input>
4595 <port id="0" precision="FP32">
4596 <dim>-1</dim>
4597 <dim>12</dim>
4598 <dim>-1</dim>
4599 <dim>32</dim>
4600 </port>
4601 <port id="1" precision="FP32">
4602 <dim>-1</dim>
4603 <dim>12</dim>
4604 <dim>-1</dim>
4605 <dim>32</dim>
4606 </port>
4607 <port id="2" precision="FP32">
4608 <dim>-1</dim>
4609 <dim>12</dim>
4610 <dim>-1</dim>
4611 <dim>32</dim>
4612 </port>
4613 <port id="3" precision="FP32">
4614 <dim>-1</dim>
4615 <dim>1</dim>
4616 <dim>-1</dim>
4617 <dim>-1</dim>
4618 </port>
4619 </input>
4620 <output>
4621 <port id="4" precision="FP32" names="253,attn_output.9">
4622 <dim>-1</dim>
4623 <dim>12</dim>
4624 <dim>-1</dim>
4625 <dim>32</dim>
4626 </port>
4627 </output>
4628 </layer>
4629 <layer id="337" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
4630 <data element_type="i32" shape="4" offset="12502400" size="16" />
4631 <output>
4632 <port id="0" precision="I32">
4633 <dim>4</dim>
4634 </port>
4635 </output>
4636 </layer>
4637 <layer id="338" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
4638 <input>
4639 <port id="0" precision="FP32">
4640 <dim>-1</dim>
4641 <dim>12</dim>
4642 <dim>-1</dim>
4643 <dim>32</dim>
4644 </port>
4645 <port id="1" precision="I32">
4646 <dim>4</dim>
4647 </port>
4648 </input>
4649 <output>
4650 <port id="2" precision="FP32" names="254,attn_output.11">
4651 <dim>-1</dim>
4652 <dim>-1</dim>
4653 <dim>12</dim>
4654 <dim>32</dim>
4655 </port>
4656 </output>
4657 </layer>
4658 <layer id="339" name="Constant_305979" type="Const" version="opset1">
4659 <data element_type="i64" shape="3" offset="12502416" size="24" />
4660 <output>
4661 <port id="0" precision="I64">
4662 <dim>3</dim>
4663 </port>
4664 </output>
4665 </layer>
4666 <layer id="340" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
4667 <data special_zero="true" />
4668 <input>
4669 <port id="0" precision="FP32">
4670 <dim>-1</dim>
4671 <dim>-1</dim>
4672 <dim>12</dim>
4673 <dim>32</dim>
4674 </port>
4675 <port id="1" precision="I64">
4676 <dim>3</dim>
4677 </port>
4678 </input>
4679 <output>
4680 <port id="2" precision="FP32" names="256">
4681 <dim>-1</dim>
4682 <dim>-1</dim>
4683 <dim>384</dim>
4684 </port>
4685 </output>
4686 </layer>
4687 <layer id="341" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4688 <data element_type="f32" shape="1, 1, 384" offset="16130568" size="1536" />
4689 <output>
4690 <port id="0" precision="FP32">
4691 <dim>1</dim>
4692 <dim>1</dim>
4693 <dim>384</dim>
4694 </port>
4695 </output>
4696 </layer>
4697 <layer id="342" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4698 <data auto_broadcast="numpy" />
4699 <input>
4700 <port id="0" precision="FP32">
4701 <dim>-1</dim>
4702 <dim>-1</dim>
4703 <dim>384</dim>
4704 </port>
4705 <port id="1" precision="FP32">
4706 <dim>1</dim>
4707 <dim>1</dim>
4708 <dim>384</dim>
4709 </port>
4710 </input>
4711 <output>
4712 <port id="2" precision="FP32">
4713 <dim>-1</dim>
4714 <dim>-1</dim>
4715 <dim>384</dim>
4716 </port>
4717 </output>
4718 </layer>
4719 <layer id="343" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
4720 <data element_type="f32" shape="" offset="16132104" size="4" />
4721 <output>
4722 <port id="0" precision="FP32" />
4723 </output>
4724 </layer>
4725 <layer id="344" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
4726 <data element_type="f32" shape="" offset="16132108" size="4" />
4727 <output>
4728 <port id="0" precision="FP32" />
4729 </output>
4730 </layer>
4731 <layer id="345" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
4732 <data element_type="f32" shape="" offset="16132104" size="4" />
4733 <output>
4734 <port id="0" precision="FP32" />
4735 </output>
4736 </layer>
4737 <layer id="346" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
4738 <data element_type="f32" shape="" offset="16132108" size="4" />
4739 <output>
4740 <port id="0" precision="FP32" />
4741 </output>
4742 </layer>
4743 <layer id="347" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
4744 <data levels="256" auto_broadcast="numpy" />
4745 <input>
4746 <port id="0" precision="FP32">
4747 <dim>-1</dim>
4748 <dim>-1</dim>
4749 <dim>384</dim>
4750 </port>
4751 <port id="1" precision="FP32" />
4752 <port id="2" precision="FP32" />
4753 <port id="3" precision="FP32" />
4754 <port id="4" precision="FP32" />
4755 </input>
4756 <output>
4757 <port id="5" precision="FP32">
4758 <dim>-1</dim>
4759 <dim>-1</dim>
4760 <dim>384</dim>
4761 </port>
4762 </output>
4763 </layer>
4764 <layer id="348" name="Constant_445570" type="Const" version="opset1">
4765 <data element_type="i8" shape="384, 384" offset="16132112" size="147456" />
4766 <output>
4767 <port id="0" precision="I8">
4768 <dim>384</dim>
4769 <dim>384</dim>
4770 </port>
4771 </output>
4772 </layer>
4773 <layer id="349" name="Convert_445571" type="Convert" version="opset1">
4774 <data destination_type="f32" />
4775 <input>
4776 <port id="0" precision="I8">
4777 <dim>384</dim>
4778 <dim>384</dim>
4779 </port>
4780 </input>
4781 <output>
4782 <port id="1" precision="FP32">
4783 <dim>384</dim>
4784 <dim>384</dim>
4785 </port>
4786 </output>
4787 </layer>
4788 <layer id="350" name="Constant_445572" type="Const" version="opset1">
4789 <data element_type="f32" shape="384, 1" offset="16279568" size="1536" />
4790 <output>
4791 <port id="0" precision="FP32">
4792 <dim>384</dim>
4793 <dim>1</dim>
4794 </port>
4795 </output>
4796 </layer>
4797 <layer id="351" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4798 <data auto_broadcast="numpy" />
4799 <input>
4800 <port id="0" precision="FP32">
4801 <dim>384</dim>
4802 <dim>384</dim>
4803 </port>
4804 <port id="1" precision="FP32">
4805 <dim>384</dim>
4806 <dim>1</dim>
4807 </port>
4808 </input>
4809 <output>
4810 <port id="2" precision="FP32">
4811 <dim>384</dim>
4812 <dim>384</dim>
4813 </port>
4814 </output>
4815 </layer>
4816 <layer id="352" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4817 <data transpose_a="false" transpose_b="true" />
4818 <input>
4819 <port id="0" precision="FP32">
4820 <dim>-1</dim>
4821 <dim>-1</dim>
4822 <dim>384</dim>
4823 </port>
4824 <port id="1" precision="FP32">
4825 <dim>384</dim>
4826 <dim>384</dim>
4827 </port>
4828 </input>
4829 <output>
4830 <port id="2" precision="FP32">
4831 <dim>-1</dim>
4832 <dim>-1</dim>
4833 <dim>384</dim>
4834 </port>
4835 </output>
4836 </layer>
4837 <layer id="353" name="Constant_305896" type="Const" version="opset1">
4838 <data element_type="f32" shape="1, 1, 384" offset="16281104" size="1536" />
4839 <output>
4840 <port id="0" precision="FP32">
4841 <dim>1</dim>
4842 <dim>1</dim>
4843 <dim>384</dim>
4844 </port>
4845 </output>
4846 </layer>
4847 <layer id="354" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
4848 <data auto_broadcast="numpy" />
4849 <input>
4850 <port id="0" precision="FP32">
4851 <dim>-1</dim>
4852 <dim>-1</dim>
4853 <dim>384</dim>
4854 </port>
4855 <port id="1" precision="FP32">
4856 <dim>1</dim>
4857 <dim>1</dim>
4858 <dim>384</dim>
4859 </port>
4860 </input>
4861 <output>
4862 <port id="2" precision="FP32" names="261,input.11">
4863 <dim>-1</dim>
4864 <dim>-1</dim>
4865 <dim>384</dim>
4866 </port>
4867 </output>
4868 </layer>
4869 <layer id="355" name="__module.bert.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
4870 <data auto_broadcast="numpy" />
4871 <input>
4872 <port id="0" precision="FP32">
4873 <dim>-1</dim>
4874 <dim>-1</dim>
4875 <dim>384</dim>
4876 </port>
4877 <port id="1" precision="FP32">
4878 <dim>-1</dim>
4879 <dim>-1</dim>
4880 <dim>384</dim>
4881 </port>
4882 </input>
4883 <output>
4884 <port id="2" precision="FP32" names="263">
4885 <dim>-1</dim>
4886 <dim>-1</dim>
4887 <dim>384</dim>
4888 </port>
4889 </output>
4890 </layer>
4891 <layer id="356" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4892 <data element_type="i32" shape="1" offset="12046084" size="4" />
4893 <output>
4894 <port id="0" precision="I32">
4895 <dim>1</dim>
4896 </port>
4897 </output>
4898 </layer>
4899 <layer id="357" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4900 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4901 <input>
4902 <port id="0" precision="FP32">
4903 <dim>-1</dim>
4904 <dim>-1</dim>
4905 <dim>384</dim>
4906 </port>
4907 <port id="1" precision="I32">
4908 <dim>1</dim>
4909 </port>
4910 </input>
4911 <output>
4912 <port id="2" precision="FP32">
4913 <dim>-1</dim>
4914 <dim>-1</dim>
4915 <dim>384</dim>
4916 </port>
4917 </output>
4918 </layer>
4919 <layer id="358" name="Constant_305897" type="Const" version="opset1">
4920 <data element_type="f32" shape="1, 1, 384" offset="16282640" size="1536" />
4921 <output>
4922 <port id="0" precision="FP32">
4923 <dim>1</dim>
4924 <dim>1</dim>
4925 <dim>384</dim>
4926 </port>
4927 </output>
4928 </layer>
4929 <layer id="359" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4930 <data auto_broadcast="numpy" />
4931 <input>
4932 <port id="0" precision="FP32">
4933 <dim>-1</dim>
4934 <dim>-1</dim>
4935 <dim>384</dim>
4936 </port>
4937 <port id="1" precision="FP32">
4938 <dim>1</dim>
4939 <dim>1</dim>
4940 <dim>384</dim>
4941 </port>
4942 </input>
4943 <output>
4944 <port id="2" precision="FP32">
4945 <dim>-1</dim>
4946 <dim>-1</dim>
4947 <dim>384</dim>
4948 </port>
4949 </output>
4950 </layer>
4951 <layer id="360" name="Constant_305898" type="Const" version="opset1">
4952 <data element_type="f32" shape="1, 1, 384" offset="16284176" size="1536" />
4953 <output>
4954 <port id="0" precision="FP32">
4955 <dim>1</dim>
4956 <dim>1</dim>
4957 <dim>384</dim>
4958 </port>
4959 </output>
4960 </layer>
4961 <layer id="361" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4962 <data auto_broadcast="numpy" />
4963 <input>
4964 <port id="0" precision="FP32">
4965 <dim>-1</dim>
4966 <dim>-1</dim>
4967 <dim>384</dim>
4968 </port>
4969 <port id="1" precision="FP32">
4970 <dim>1</dim>
4971 <dim>1</dim>
4972 <dim>384</dim>
4973 </port>
4974 </input>
4975 <output>
4976 <port id="2" precision="FP32" names="267,input_tensor.5">
4977 <dim>-1</dim>
4978 <dim>-1</dim>
4979 <dim>384</dim>
4980 </port>
4981 </output>
4982 </layer>
4983 <layer id="362" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4984 <data element_type="f32" shape="1, 1, 384" offset="16285712" size="1536" />
4985 <output>
4986 <port id="0" precision="FP32">
4987 <dim>1</dim>
4988 <dim>1</dim>
4989 <dim>384</dim>
4990 </port>
4991 </output>
4992 </layer>
4993 <layer id="363" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4994 <data auto_broadcast="numpy" />
4995 <input>
4996 <port id="0" precision="FP32">
4997 <dim>-1</dim>
4998 <dim>-1</dim>
4999 <dim>384</dim>
5000 </port>
5001 <port id="1" precision="FP32">
5002 <dim>1</dim>
5003 <dim>1</dim>
5004 <dim>384</dim>
5005 </port>
5006 </input>
5007 <output>
5008 <port id="2" precision="FP32">
5009 <dim>-1</dim>
5010 <dim>-1</dim>
5011 <dim>384</dim>
5012 </port>
5013 </output>
5014 </layer>
5015 <layer id="364" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5016 <data element_type="f32" shape="" offset="16287248" size="4" />
5017 <output>
5018 <port id="0" precision="FP32" />
5019 </output>
5020 </layer>
5021 <layer id="365" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5022 <data element_type="f32" shape="" offset="16287252" size="4" />
5023 <output>
5024 <port id="0" precision="FP32" />
5025 </output>
5026 </layer>
5027 <layer id="366" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5028 <data element_type="f32" shape="" offset="16287248" size="4" />
5029 <output>
5030 <port id="0" precision="FP32" />
5031 </output>
5032 </layer>
5033 <layer id="367" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5034 <data element_type="f32" shape="" offset="16287252" size="4" />
5035 <output>
5036 <port id="0" precision="FP32" />
5037 </output>
5038 </layer>
5039 <layer id="368" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5040 <data levels="256" auto_broadcast="numpy" />
5041 <input>
5042 <port id="0" precision="FP32">
5043 <dim>-1</dim>
5044 <dim>-1</dim>
5045 <dim>384</dim>
5046 </port>
5047 <port id="1" precision="FP32" />
5048 <port id="2" precision="FP32" />
5049 <port id="3" precision="FP32" />
5050 <port id="4" precision="FP32" />
5051 </input>
5052 <output>
5053 <port id="5" precision="FP32">
5054 <dim>-1</dim>
5055 <dim>-1</dim>
5056 <dim>384</dim>
5057 </port>
5058 </output>
5059 </layer>
5060 <layer id="369" name="Constant_445574" type="Const" version="opset1">
5061 <data element_type="i8" shape="1536, 384" offset="16287256" size="589824" />
5062 <output>
5063 <port id="0" precision="I8">
5064 <dim>1536</dim>
5065 <dim>384</dim>
5066 </port>
5067 </output>
5068 </layer>
5069 <layer id="370" name="Convert_445575" type="Convert" version="opset1">
5070 <data destination_type="f32" />
5071 <input>
5072 <port id="0" precision="I8">
5073 <dim>1536</dim>
5074 <dim>384</dim>
5075 </port>
5076 </input>
5077 <output>
5078 <port id="1" precision="FP32">
5079 <dim>1536</dim>
5080 <dim>384</dim>
5081 </port>
5082 </output>
5083 </layer>
5084 <layer id="371" name="Constant_445576" type="Const" version="opset1">
5085 <data element_type="f32" shape="1536, 1" offset="16877080" size="6144" />
5086 <output>
5087 <port id="0" precision="FP32">
5088 <dim>1536</dim>
5089 <dim>1</dim>
5090 </port>
5091 </output>
5092 </layer>
5093 <layer id="372" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5094 <data auto_broadcast="numpy" />
5095 <input>
5096 <port id="0" precision="FP32">
5097 <dim>1536</dim>
5098 <dim>384</dim>
5099 </port>
5100 <port id="1" precision="FP32">
5101 <dim>1536</dim>
5102 <dim>1</dim>
5103 </port>
5104 </input>
5105 <output>
5106 <port id="2" precision="FP32">
5107 <dim>1536</dim>
5108 <dim>384</dim>
5109 </port>
5110 </output>
5111 </layer>
5112 <layer id="373" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5113 <data transpose_a="false" transpose_b="true" />
5114 <input>
5115 <port id="0" precision="FP32">
5116 <dim>-1</dim>
5117 <dim>-1</dim>
5118 <dim>384</dim>
5119 </port>
5120 <port id="1" precision="FP32">
5121 <dim>1536</dim>
5122 <dim>384</dim>
5123 </port>
5124 </input>
5125 <output>
5126 <port id="2" precision="FP32">
5127 <dim>-1</dim>
5128 <dim>-1</dim>
5129 <dim>1536</dim>
5130 </port>
5131 </output>
5132 </layer>
5133 <layer id="374" name="Constant_305899" type="Const" version="opset1">
5134 <data element_type="f32" shape="1, 1, 1536" offset="16883224" size="6144" />
5135 <output>
5136 <port id="0" precision="FP32">
5137 <dim>1</dim>
5138 <dim>1</dim>
5139 <dim>1536</dim>
5140 </port>
5141 </output>
5142 </layer>
5143 <layer id="375" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5144 <data auto_broadcast="numpy" />
5145 <input>
5146 <port id="0" precision="FP32">
5147 <dim>-1</dim>
5148 <dim>-1</dim>
5149 <dim>1536</dim>
5150 </port>
5151 <port id="1" precision="FP32">
5152 <dim>1</dim>
5153 <dim>1</dim>
5154 <dim>1536</dim>
5155 </port>
5156 </input>
5157 <output>
5158 <port id="2" precision="FP32" names="271">
5159 <dim>-1</dim>
5160 <dim>-1</dim>
5161 <dim>1536</dim>
5162 </port>
5163 </output>
5164 </layer>
5165 <layer id="376" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5166 <data approximation_mode="ERF" />
5167 <input>
5168 <port id="0" precision="FP32">
5169 <dim>-1</dim>
5170 <dim>-1</dim>
5171 <dim>1536</dim>
5172 </port>
5173 </input>
5174 <output>
5175 <port id="1" precision="FP32" names="272">
5176 <dim>-1</dim>
5177 <dim>-1</dim>
5178 <dim>1536</dim>
5179 </port>
5180 </output>
5181 </layer>
5182 <layer id="377" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5183 <data element_type="f32" shape="1, 1, 1536" offset="16889368" size="6144" />
5184 <output>
5185 <port id="0" precision="FP32">
5186 <dim>1</dim>
5187 <dim>1</dim>
5188 <dim>1536</dim>
5189 </port>
5190 </output>
5191 </layer>
5192 <layer id="378" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5193 <data auto_broadcast="numpy" />
5194 <input>
5195 <port id="0" precision="FP32">
5196 <dim>-1</dim>
5197 <dim>-1</dim>
5198 <dim>1536</dim>
5199 </port>
5200 <port id="1" precision="FP32">
5201 <dim>1</dim>
5202 <dim>1</dim>
5203 <dim>1536</dim>
5204 </port>
5205 </input>
5206 <output>
5207 <port id="2" precision="FP32">
5208 <dim>-1</dim>
5209 <dim>-1</dim>
5210 <dim>1536</dim>
5211 </port>
5212 </output>
5213 </layer>
5214 <layer id="379" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5215 <data element_type="f32" shape="" offset="16895512" size="4" />
5216 <output>
5217 <port id="0" precision="FP32" />
5218 </output>
5219 </layer>
5220 <layer id="380" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5221 <data element_type="f32" shape="" offset="16895516" size="4" />
5222 <output>
5223 <port id="0" precision="FP32" />
5224 </output>
5225 </layer>
5226 <layer id="381" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5227 <data element_type="f32" shape="" offset="16895512" size="4" />
5228 <output>
5229 <port id="0" precision="FP32" />
5230 </output>
5231 </layer>
5232 <layer id="382" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5233 <data element_type="f32" shape="" offset="16895516" size="4" />
5234 <output>
5235 <port id="0" precision="FP32" />
5236 </output>
5237 </layer>
5238 <layer id="383" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5239 <data levels="256" auto_broadcast="numpy" />
5240 <input>
5241 <port id="0" precision="FP32">
5242 <dim>-1</dim>
5243 <dim>-1</dim>
5244 <dim>1536</dim>
5245 </port>
5246 <port id="1" precision="FP32" />
5247 <port id="2" precision="FP32" />
5248 <port id="3" precision="FP32" />
5249 <port id="4" precision="FP32" />
5250 </input>
5251 <output>
5252 <port id="5" precision="FP32">
5253 <dim>-1</dim>
5254 <dim>-1</dim>
5255 <dim>1536</dim>
5256 </port>
5257 </output>
5258 </layer>
5259 <layer id="384" name="Constant_445578" type="Const" version="opset1">
5260 <data element_type="i8" shape="384, 1536" offset="16895520" size="589824" />
5261 <output>
5262 <port id="0" precision="I8">
5263 <dim>384</dim>
5264 <dim>1536</dim>
5265 </port>
5266 </output>
5267 </layer>
5268 <layer id="385" name="Convert_445579" type="Convert" version="opset1">
5269 <data destination_type="f32" />
5270 <input>
5271 <port id="0" precision="I8">
5272 <dim>384</dim>
5273 <dim>1536</dim>
5274 </port>
5275 </input>
5276 <output>
5277 <port id="1" precision="FP32">
5278 <dim>384</dim>
5279 <dim>1536</dim>
5280 </port>
5281 </output>
5282 </layer>
5283 <layer id="386" name="Constant_445580" type="Const" version="opset1">
5284 <data element_type="f32" shape="384, 1" offset="17485344" size="1536" />
5285 <output>
5286 <port id="0" precision="FP32">
5287 <dim>384</dim>
5288 <dim>1</dim>
5289 </port>
5290 </output>
5291 </layer>
5292 <layer id="387" name="__module.bert.encoder.layer.2.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5293 <data auto_broadcast="numpy" />
5294 <input>
5295 <port id="0" precision="FP32">
5296 <dim>384</dim>
5297 <dim>1536</dim>
5298 </port>
5299 <port id="1" precision="FP32">
5300 <dim>384</dim>
5301 <dim>1</dim>
5302 </port>
5303 </input>
5304 <output>
5305 <port id="2" precision="FP32">
5306 <dim>384</dim>
5307 <dim>1536</dim>
5308 </port>
5309 </output>
5310 </layer>
5311 <layer id="388" name="__module.bert.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5312 <data transpose_a="false" transpose_b="true" />
5313 <input>
5314 <port id="0" precision="FP32">
5315 <dim>-1</dim>
5316 <dim>-1</dim>
5317 <dim>1536</dim>
5318 </port>
5319 <port id="1" precision="FP32">
5320 <dim>384</dim>
5321 <dim>1536</dim>
5322 </port>
5323 </input>
5324 <output>
5325 <port id="2" precision="FP32">
5326 <dim>-1</dim>
5327 <dim>-1</dim>
5328 <dim>384</dim>
5329 </port>
5330 </output>
5331 </layer>
5332 <layer id="389" name="Constant_305900" type="Const" version="opset1">
5333 <data element_type="f32" shape="1, 1, 384" offset="17486880" size="1536" />
5334 <output>
5335 <port id="0" precision="FP32">
5336 <dim>1</dim>
5337 <dim>1</dim>
5338 <dim>384</dim>
5339 </port>
5340 </output>
5341 </layer>
5342 <layer id="390" name="__module.bert.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
5343 <data auto_broadcast="numpy" />
5344 <input>
5345 <port id="0" precision="FP32">
5346 <dim>-1</dim>
5347 <dim>-1</dim>
5348 <dim>384</dim>
5349 </port>
5350 <port id="1" precision="FP32">
5351 <dim>1</dim>
5352 <dim>1</dim>
5353 <dim>384</dim>
5354 </port>
5355 </input>
5356 <output>
5357 <port id="2" precision="FP32" names="277,input.13">
5358 <dim>-1</dim>
5359 <dim>-1</dim>
5360 <dim>384</dim>
5361 </port>
5362 </output>
5363 </layer>
5364 <layer id="391" name="__module.bert.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
5365 <data auto_broadcast="numpy" />
5366 <input>
5367 <port id="0" precision="FP32">
5368 <dim>-1</dim>
5369 <dim>-1</dim>
5370 <dim>384</dim>
5371 </port>
5372 <port id="1" precision="FP32">
5373 <dim>-1</dim>
5374 <dim>-1</dim>
5375 <dim>384</dim>
5376 </port>
5377 </input>
5378 <output>
5379 <port id="2" precision="FP32" names="279">
5380 <dim>-1</dim>
5381 <dim>-1</dim>
5382 <dim>384</dim>
5383 </port>
5384 </output>
5385 </layer>
5386 <layer id="392" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5387 <data element_type="i32" shape="1" offset="12046084" size="4" />
5388 <output>
5389 <port id="0" precision="I32">
5390 <dim>1</dim>
5391 </port>
5392 </output>
5393 </layer>
5394 <layer id="393" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5395 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5396 <input>
5397 <port id="0" precision="FP32">
5398 <dim>-1</dim>
5399 <dim>-1</dim>
5400 <dim>384</dim>
5401 </port>
5402 <port id="1" precision="I32">
5403 <dim>1</dim>
5404 </port>
5405 </input>
5406 <output>
5407 <port id="2" precision="FP32">
5408 <dim>-1</dim>
5409 <dim>-1</dim>
5410 <dim>384</dim>
5411 </port>
5412 </output>
5413 </layer>
5414 <layer id="394" name="Constant_305901" type="Const" version="opset1">
5415 <data element_type="f32" shape="1, 1, 384" offset="17488416" size="1536" />
5416 <output>
5417 <port id="0" precision="FP32">
5418 <dim>1</dim>
5419 <dim>1</dim>
5420 <dim>384</dim>
5421 </port>
5422 </output>
5423 </layer>
5424 <layer id="395" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5425 <data auto_broadcast="numpy" />
5426 <input>
5427 <port id="0" precision="FP32">
5428 <dim>-1</dim>
5429 <dim>-1</dim>
5430 <dim>384</dim>
5431 </port>
5432 <port id="1" precision="FP32">
5433 <dim>1</dim>
5434 <dim>1</dim>
5435 <dim>384</dim>
5436 </port>
5437 </input>
5438 <output>
5439 <port id="2" precision="FP32">
5440 <dim>-1</dim>
5441 <dim>-1</dim>
5442 <dim>384</dim>
5443 </port>
5444 </output>
5445 </layer>
5446 <layer id="396" name="Constant_305902" type="Const" version="opset1">
5447 <data element_type="f32" shape="1, 1, 384" offset="17489952" size="1536" />
5448 <output>
5449 <port id="0" precision="FP32">
5450 <dim>1</dim>
5451 <dim>1</dim>
5452 <dim>384</dim>
5453 </port>
5454 </output>
5455 </layer>
5456 <layer id="397" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5457 <data auto_broadcast="numpy" />
5458 <input>
5459 <port id="0" precision="FP32">
5460 <dim>-1</dim>
5461 <dim>-1</dim>
5462 <dim>384</dim>
5463 </port>
5464 <port id="1" precision="FP32">
5465 <dim>1</dim>
5466 <dim>1</dim>
5467 <dim>384</dim>
5468 </port>
5469 </input>
5470 <output>
5471 <port id="2" precision="FP32" names="283,hidden_states.19">
5472 <dim>-1</dim>
5473 <dim>-1</dim>
5474 <dim>384</dim>
5475 </port>
5476 </output>
5477 </layer>
5478 <layer id="398" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5479 <data element_type="f32" shape="1, 1, 384" offset="17491488" size="1536" />
5480 <output>
5481 <port id="0" precision="FP32">
5482 <dim>1</dim>
5483 <dim>1</dim>
5484 <dim>384</dim>
5485 </port>
5486 </output>
5487 </layer>
5488 <layer id="399" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5489 <data auto_broadcast="numpy" />
5490 <input>
5491 <port id="0" precision="FP32">
5492 <dim>-1</dim>
5493 <dim>-1</dim>
5494 <dim>384</dim>
5495 </port>
5496 <port id="1" precision="FP32">
5497 <dim>1</dim>
5498 <dim>1</dim>
5499 <dim>384</dim>
5500 </port>
5501 </input>
5502 <output>
5503 <port id="2" precision="FP32">
5504 <dim>-1</dim>
5505 <dim>-1</dim>
5506 <dim>384</dim>
5507 </port>
5508 </output>
5509 </layer>
5510 <layer id="400" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5511 <data element_type="f32" shape="" offset="17493024" size="4" />
5512 <output>
5513 <port id="0" precision="FP32" />
5514 </output>
5515 </layer>
5516 <layer id="401" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5517 <data element_type="f32" shape="" offset="17493028" size="4" />
5518 <output>
5519 <port id="0" precision="FP32" />
5520 </output>
5521 </layer>
5522 <layer id="402" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5523 <data element_type="f32" shape="" offset="17493024" size="4" />
5524 <output>
5525 <port id="0" precision="FP32" />
5526 </output>
5527 </layer>
5528 <layer id="403" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5529 <data element_type="f32" shape="" offset="17493028" size="4" />
5530 <output>
5531 <port id="0" precision="FP32" />
5532 </output>
5533 </layer>
5534 <layer id="404" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5535 <data levels="256" auto_broadcast="numpy" />
5536 <input>
5537 <port id="0" precision="FP32">
5538 <dim>-1</dim>
5539 <dim>-1</dim>
5540 <dim>384</dim>
5541 </port>
5542 <port id="1" precision="FP32" />
5543 <port id="2" precision="FP32" />
5544 <port id="3" precision="FP32" />
5545 <port id="4" precision="FP32" />
5546 </input>
5547 <output>
5548 <port id="5" precision="FP32">
5549 <dim>-1</dim>
5550 <dim>-1</dim>
5551 <dim>384</dim>
5552 </port>
5553 </output>
5554 </layer>
5555 <layer id="405" name="Constant_445582" type="Const" version="opset1">
5556 <data element_type="i8" shape="384, 384" offset="17493032" size="147456" />
5557 <output>
5558 <port id="0" precision="I8">
5559 <dim>384</dim>
5560 <dim>384</dim>
5561 </port>
5562 </output>
5563 </layer>
5564 <layer id="406" name="Convert_445583" type="Convert" version="opset1">
5565 <data destination_type="f32" />
5566 <input>
5567 <port id="0" precision="I8">
5568 <dim>384</dim>
5569 <dim>384</dim>
5570 </port>
5571 </input>
5572 <output>
5573 <port id="1" precision="FP32">
5574 <dim>384</dim>
5575 <dim>384</dim>
5576 </port>
5577 </output>
5578 </layer>
5579 <layer id="407" name="Constant_445584" type="Const" version="opset1">
5580 <data element_type="f32" shape="384, 1" offset="17640488" size="1536" />
5581 <output>
5582 <port id="0" precision="FP32">
5583 <dim>384</dim>
5584 <dim>1</dim>
5585 </port>
5586 </output>
5587 </layer>
5588 <layer id="408" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5589 <data auto_broadcast="numpy" />
5590 <input>
5591 <port id="0" precision="FP32">
5592 <dim>384</dim>
5593 <dim>384</dim>
5594 </port>
5595 <port id="1" precision="FP32">
5596 <dim>384</dim>
5597 <dim>1</dim>
5598 </port>
5599 </input>
5600 <output>
5601 <port id="2" precision="FP32">
5602 <dim>384</dim>
5603 <dim>384</dim>
5604 </port>
5605 </output>
5606 </layer>
5607 <layer id="409" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
5608 <data transpose_a="false" transpose_b="true" />
5609 <input>
5610 <port id="0" precision="FP32">
5611 <dim>-1</dim>
5612 <dim>-1</dim>
5613 <dim>384</dim>
5614 </port>
5615 <port id="1" precision="FP32">
5616 <dim>384</dim>
5617 <dim>384</dim>
5618 </port>
5619 </input>
5620 <output>
5621 <port id="2" precision="FP32">
5622 <dim>-1</dim>
5623 <dim>-1</dim>
5624 <dim>384</dim>
5625 </port>
5626 </output>
5627 </layer>
5628 <layer id="410" name="Constant_305903" type="Const" version="opset1">
5629 <data element_type="f32" shape="1, 1, 384" offset="17642024" size="1536" />
5630 <output>
5631 <port id="0" precision="FP32">
5632 <dim>1</dim>
5633 <dim>1</dim>
5634 <dim>384</dim>
5635 </port>
5636 </output>
5637 </layer>
5638 <layer id="411" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
5639 <data auto_broadcast="numpy" />
5640 <input>
5641 <port id="0" precision="FP32">
5642 <dim>-1</dim>
5643 <dim>-1</dim>
5644 <dim>384</dim>
5645 </port>
5646 <port id="1" precision="FP32">
5647 <dim>1</dim>
5648 <dim>1</dim>
5649 <dim>384</dim>
5650 </port>
5651 </input>
5652 <output>
5653 <port id="2" precision="FP32" names="296,x.37">
5654 <dim>-1</dim>
5655 <dim>-1</dim>
5656 <dim>384</dim>
5657 </port>
5658 </output>
5659 </layer>
5660 <layer id="412" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5661 <data element_type="f32" shape="" offset="17643560" size="4" />
5662 <output>
5663 <port id="0" precision="FP32" />
5664 </output>
5665 </layer>
5666 <layer id="413" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
5667 <data element_type="f32" shape="" offset="17643564" size="4" />
5668 <output>
5669 <port id="0" precision="FP32" />
5670 </output>
5671 </layer>
5672 <layer id="414" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
5673 <data element_type="f32" shape="" offset="17643560" size="4" />
5674 <output>
5675 <port id="0" precision="FP32" />
5676 </output>
5677 </layer>
5678 <layer id="415" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
5679 <data element_type="f32" shape="" offset="17643564" size="4" />
5680 <output>
5681 <port id="0" precision="FP32" />
5682 </output>
5683 </layer>
5684 <layer id="416" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
5685 <data levels="256" auto_broadcast="numpy" />
5686 <input>
5687 <port id="0" precision="FP32">
5688 <dim>-1</dim>
5689 <dim>-1</dim>
5690 <dim>384</dim>
5691 </port>
5692 <port id="1" precision="FP32" />
5693 <port id="2" precision="FP32" />
5694 <port id="3" precision="FP32" />
5695 <port id="4" precision="FP32" />
5696 </input>
5697 <output>
5698 <port id="5" precision="FP32">
5699 <dim>-1</dim>
5700 <dim>-1</dim>
5701 <dim>384</dim>
5702 </port>
5703 </output>
5704 </layer>
5705 <layer id="417" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
5706 <data element_type="i64" shape="4" offset="12201240" size="32" />
5707 <output>
5708 <port id="0" precision="I64">
5709 <dim>4</dim>
5710 </port>
5711 </output>
5712 </layer>
5713 <layer id="418" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
5714 <data special_zero="true" />
5715 <input>
5716 <port id="0" precision="FP32">
5717 <dim>-1</dim>
5718 <dim>-1</dim>
5719 <dim>384</dim>
5720 </port>
5721 <port id="1" precision="I64">
5722 <dim>4</dim>
5723 </port>
5724 </input>
5725 <output>
5726 <port id="2" precision="FP32" names="300,x.39">
5727 <dim>-1</dim>
5728 <dim>-1</dim>
5729 <dim>12</dim>
5730 <dim>32</dim>
5731 </port>
5732 </output>
5733 </layer>
5734 <layer id="419" name="Constant_301735" type="Const" version="opset1">
5735 <data element_type="i64" shape="4" offset="12201272" size="32" />
5736 <output>
5737 <port id="0" precision="I64" names="301">
5738 <dim>4</dim>
5739 </port>
5740 </output>
5741 </layer>
5742 <layer id="420" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
5743 <input>
5744 <port id="0" precision="FP32">
5745 <dim>-1</dim>
5746 <dim>-1</dim>
5747 <dim>12</dim>
5748 <dim>32</dim>
5749 </port>
5750 <port id="1" precision="I64">
5751 <dim>4</dim>
5752 </port>
5753 </input>
5754 <output>
5755 <port id="2" precision="FP32" names="302">
5756 <dim>-1</dim>
5757 <dim>12</dim>
5758 <dim>-1</dim>
5759 <dim>32</dim>
5760 </port>
5761 </output>
5762 </layer>
5763 <layer id="421" name="Constant_445586" type="Const" version="opset1">
5764 <data element_type="i8" shape="384, 384" offset="17643568" size="147456" />
5765 <output>
5766 <port id="0" precision="I8">
5767 <dim>384</dim>
5768 <dim>384</dim>
5769 </port>
5770 </output>
5771 </layer>
5772 <layer id="422" name="Convert_445587" type="Convert" version="opset1">
5773 <data destination_type="f32" />
5774 <input>
5775 <port id="0" precision="I8">
5776 <dim>384</dim>
5777 <dim>384</dim>
5778 </port>
5779 </input>
5780 <output>
5781 <port id="1" precision="FP32">
5782 <dim>384</dim>
5783 <dim>384</dim>
5784 </port>
5785 </output>
5786 </layer>
5787 <layer id="423" name="Constant_445588" type="Const" version="opset1">
5788 <data element_type="f32" shape="384, 1" offset="17791024" size="1536" />
5789 <output>
5790 <port id="0" precision="FP32">
5791 <dim>384</dim>
5792 <dim>1</dim>
5793 </port>
5794 </output>
5795 </layer>
5796 <layer id="424" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5797 <data auto_broadcast="numpy" />
5798 <input>
5799 <port id="0" precision="FP32">
5800 <dim>384</dim>
5801 <dim>384</dim>
5802 </port>
5803 <port id="1" precision="FP32">
5804 <dim>384</dim>
5805 <dim>1</dim>
5806 </port>
5807 </input>
5808 <output>
5809 <port id="2" precision="FP32">
5810 <dim>384</dim>
5811 <dim>384</dim>
5812 </port>
5813 </output>
5814 </layer>
5815 <layer id="425" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
5816 <data transpose_a="false" transpose_b="true" />
5817 <input>
5818 <port id="0" precision="FP32">
5819 <dim>-1</dim>
5820 <dim>-1</dim>
5821 <dim>384</dim>
5822 </port>
5823 <port id="1" precision="FP32">
5824 <dim>384</dim>
5825 <dim>384</dim>
5826 </port>
5827 </input>
5828 <output>
5829 <port id="2" precision="FP32">
5830 <dim>-1</dim>
5831 <dim>-1</dim>
5832 <dim>384</dim>
5833 </port>
5834 </output>
5835 </layer>
5836 <layer id="426" name="Constant_305904" type="Const" version="opset1">
5837 <data element_type="f32" shape="1, 1, 384" offset="17792560" size="1536" />
5838 <output>
5839 <port id="0" precision="FP32">
5840 <dim>1</dim>
5841 <dim>1</dim>
5842 <dim>384</dim>
5843 </port>
5844 </output>
5845 </layer>
5846 <layer id="427" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
5847 <data auto_broadcast="numpy" />
5848 <input>
5849 <port id="0" precision="FP32">
5850 <dim>-1</dim>
5851 <dim>-1</dim>
5852 <dim>384</dim>
5853 </port>
5854 <port id="1" precision="FP32">
5855 <dim>1</dim>
5856 <dim>1</dim>
5857 <dim>384</dim>
5858 </port>
5859 </input>
5860 <output>
5861 <port id="2" precision="FP32" names="305,x.41">
5862 <dim>-1</dim>
5863 <dim>-1</dim>
5864 <dim>384</dim>
5865 </port>
5866 </output>
5867 </layer>
5868 <layer id="428" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5869 <data element_type="f32" shape="" offset="17794096" size="4" />
5870 <output>
5871 <port id="0" precision="FP32" />
5872 </output>
5873 </layer>
5874 <layer id="429" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
5875 <data element_type="f32" shape="" offset="17794100" size="4" />
5876 <output>
5877 <port id="0" precision="FP32" />
5878 </output>
5879 </layer>
5880 <layer id="430" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
5881 <data element_type="f32" shape="" offset="17794096" size="4" />
5882 <output>
5883 <port id="0" precision="FP32" />
5884 </output>
5885 </layer>
5886 <layer id="431" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
5887 <data element_type="f32" shape="" offset="17794100" size="4" />
5888 <output>
5889 <port id="0" precision="FP32" />
5890 </output>
5891 </layer>
5892 <layer id="432" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
5893 <data levels="256" auto_broadcast="numpy" />
5894 <input>
5895 <port id="0" precision="FP32">
5896 <dim>-1</dim>
5897 <dim>-1</dim>
5898 <dim>384</dim>
5899 </port>
5900 <port id="1" precision="FP32" />
5901 <port id="2" precision="FP32" />
5902 <port id="3" precision="FP32" />
5903 <port id="4" precision="FP32" />
5904 </input>
5905 <output>
5906 <port id="5" precision="FP32">
5907 <dim>-1</dim>
5908 <dim>-1</dim>
5909 <dim>384</dim>
5910 </port>
5911 </output>
5912 </layer>
5913 <layer id="433" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
5914 <data element_type="i64" shape="4" offset="12201240" size="32" />
5915 <output>
5916 <port id="0" precision="I64">
5917 <dim>4</dim>
5918 </port>
5919 </output>
5920 </layer>
5921 <layer id="434" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
5922 <data special_zero="true" />
5923 <input>
5924 <port id="0" precision="FP32">
5925 <dim>-1</dim>
5926 <dim>-1</dim>
5927 <dim>384</dim>
5928 </port>
5929 <port id="1" precision="I64">
5930 <dim>4</dim>
5931 </port>
5932 </input>
5933 <output>
5934 <port id="2" precision="FP32" names="309,x.43">
5935 <dim>-1</dim>
5936 <dim>-1</dim>
5937 <dim>12</dim>
5938 <dim>32</dim>
5939 </port>
5940 </output>
5941 </layer>
5942 <layer id="435" name="Constant_301758" type="Const" version="opset1">
5943 <data element_type="i64" shape="4" offset="12201272" size="32" />
5944 <output>
5945 <port id="0" precision="I64" names="310">
5946 <dim>4</dim>
5947 </port>
5948 </output>
5949 </layer>
5950 <layer id="436" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
5951 <input>
5952 <port id="0" precision="FP32">
5953 <dim>-1</dim>
5954 <dim>-1</dim>
5955 <dim>12</dim>
5956 <dim>32</dim>
5957 </port>
5958 <port id="1" precision="I64">
5959 <dim>4</dim>
5960 </port>
5961 </input>
5962 <output>
5963 <port id="2" precision="FP32" names="311">
5964 <dim>-1</dim>
5965 <dim>12</dim>
5966 <dim>-1</dim>
5967 <dim>32</dim>
5968 </port>
5969 </output>
5970 </layer>
5971 <layer id="437" name="Constant_445590" type="Const" version="opset1">
5972 <data element_type="i8" shape="384, 384" offset="17794104" size="147456" />
5973 <output>
5974 <port id="0" precision="I8">
5975 <dim>384</dim>
5976 <dim>384</dim>
5977 </port>
5978 </output>
5979 </layer>
5980 <layer id="438" name="Convert_445591" type="Convert" version="opset1">
5981 <data destination_type="f32" />
5982 <input>
5983 <port id="0" precision="I8">
5984 <dim>384</dim>
5985 <dim>384</dim>
5986 </port>
5987 </input>
5988 <output>
5989 <port id="1" precision="FP32">
5990 <dim>384</dim>
5991 <dim>384</dim>
5992 </port>
5993 </output>
5994 </layer>
5995 <layer id="439" name="Constant_445592" type="Const" version="opset1">
5996 <data element_type="f32" shape="384, 1" offset="17941560" size="1536" />
5997 <output>
5998 <port id="0" precision="FP32">
5999 <dim>384</dim>
6000 <dim>1</dim>
6001 </port>
6002 </output>
6003 </layer>
6004 <layer id="440" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6005 <data auto_broadcast="numpy" />
6006 <input>
6007 <port id="0" precision="FP32">
6008 <dim>384</dim>
6009 <dim>384</dim>
6010 </port>
6011 <port id="1" precision="FP32">
6012 <dim>384</dim>
6013 <dim>1</dim>
6014 </port>
6015 </input>
6016 <output>
6017 <port id="2" precision="FP32">
6018 <dim>384</dim>
6019 <dim>384</dim>
6020 </port>
6021 </output>
6022 </layer>
6023 <layer id="441" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
6024 <data transpose_a="false" transpose_b="true" />
6025 <input>
6026 <port id="0" precision="FP32">
6027 <dim>-1</dim>
6028 <dim>-1</dim>
6029 <dim>384</dim>
6030 </port>
6031 <port id="1" precision="FP32">
6032 <dim>384</dim>
6033 <dim>384</dim>
6034 </port>
6035 </input>
6036 <output>
6037 <port id="2" precision="FP32">
6038 <dim>-1</dim>
6039 <dim>-1</dim>
6040 <dim>384</dim>
6041 </port>
6042 </output>
6043 </layer>
6044 <layer id="442" name="Constant_305905" type="Const" version="opset1">
6045 <data element_type="f32" shape="1, 1, 384" offset="17943096" size="1536" />
6046 <output>
6047 <port id="0" precision="FP32">
6048 <dim>1</dim>
6049 <dim>1</dim>
6050 <dim>384</dim>
6051 </port>
6052 </output>
6053 </layer>
6054 <layer id="443" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
6055 <data auto_broadcast="numpy" />
6056 <input>
6057 <port id="0" precision="FP32">
6058 <dim>-1</dim>
6059 <dim>-1</dim>
6060 <dim>384</dim>
6061 </port>
6062 <port id="1" precision="FP32">
6063 <dim>1</dim>
6064 <dim>1</dim>
6065 <dim>384</dim>
6066 </port>
6067 </input>
6068 <output>
6069 <port id="2" precision="FP32" names="314,x.45">
6070 <dim>-1</dim>
6071 <dim>-1</dim>
6072 <dim>384</dim>
6073 </port>
6074 </output>
6075 </layer>
6076 <layer id="444" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
6077 <data element_type="i64" shape="4" offset="12201240" size="32" />
6078 <output>
6079 <port id="0" precision="I64">
6080 <dim>4</dim>
6081 </port>
6082 </output>
6083 </layer>
6084 <layer id="445" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
6085 <data special_zero="true" />
6086 <input>
6087 <port id="0" precision="FP32">
6088 <dim>-1</dim>
6089 <dim>-1</dim>
6090 <dim>384</dim>
6091 </port>
6092 <port id="1" precision="I64">
6093 <dim>4</dim>
6094 </port>
6095 </input>
6096 <output>
6097 <port id="2" precision="FP32" names="318,x">
6098 <dim>-1</dim>
6099 <dim>-1</dim>
6100 <dim>12</dim>
6101 <dim>32</dim>
6102 </port>
6103 </output>
6104 </layer>
6105 <layer id="446" name="Constant_301781" type="Const" version="opset1">
6106 <data element_type="i64" shape="4" offset="12201272" size="32" />
6107 <output>
6108 <port id="0" precision="I64" names="319">
6109 <dim>4</dim>
6110 </port>
6111 </output>
6112 </layer>
6113 <layer id="447" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
6114 <input>
6115 <port id="0" precision="FP32">
6116 <dim>-1</dim>
6117 <dim>-1</dim>
6118 <dim>12</dim>
6119 <dim>32</dim>
6120 </port>
6121 <port id="1" precision="I64">
6122 <dim>4</dim>
6123 </port>
6124 </input>
6125 <output>
6126 <port id="2" precision="FP32" names="320">
6127 <dim>-1</dim>
6128 <dim>12</dim>
6129 <dim>-1</dim>
6130 <dim>32</dim>
6131 </port>
6132 </output>
6133 </layer>
6134 <layer id="448" name="__module.bert.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
6135 <data causal="false" />
6136 <input>
6137 <port id="0" precision="FP32">
6138 <dim>-1</dim>
6139 <dim>12</dim>
6140 <dim>-1</dim>
6141 <dim>32</dim>
6142 </port>
6143 <port id="1" precision="FP32">
6144 <dim>-1</dim>
6145 <dim>12</dim>
6146 <dim>-1</dim>
6147 <dim>32</dim>
6148 </port>
6149 <port id="2" precision="FP32">
6150 <dim>-1</dim>
6151 <dim>12</dim>
6152 <dim>-1</dim>
6153 <dim>32</dim>
6154 </port>
6155 <port id="3" precision="FP32">
6156 <dim>-1</dim>
6157 <dim>1</dim>
6158 <dim>-1</dim>
6159 <dim>-1</dim>
6160 </port>
6161 </input>
6162 <output>
6163 <port id="4" precision="FP32" names="321,attn_output.13">
6164 <dim>-1</dim>
6165 <dim>12</dim>
6166 <dim>-1</dim>
6167 <dim>32</dim>
6168 </port>
6169 </output>
6170 </layer>
6171 <layer id="449" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
6172 <data element_type="i32" shape="4" offset="12502400" size="16" />
6173 <output>
6174 <port id="0" precision="I32">
6175 <dim>4</dim>
6176 </port>
6177 </output>
6178 </layer>
6179 <layer id="450" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
6180 <input>
6181 <port id="0" precision="FP32">
6182 <dim>-1</dim>
6183 <dim>12</dim>
6184 <dim>-1</dim>
6185 <dim>32</dim>
6186 </port>
6187 <port id="1" precision="I32">
6188 <dim>4</dim>
6189 </port>
6190 </input>
6191 <output>
6192 <port id="2" precision="FP32" names="322,attn_output">
6193 <dim>-1</dim>
6194 <dim>-1</dim>
6195 <dim>12</dim>
6196 <dim>32</dim>
6197 </port>
6198 </output>
6199 </layer>
6200 <layer id="451" name="Constant_305980" type="Const" version="opset1">
6201 <data element_type="i64" shape="3" offset="12502416" size="24" />
6202 <output>
6203 <port id="0" precision="I64">
6204 <dim>3</dim>
6205 </port>
6206 </output>
6207 </layer>
6208 <layer id="452" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
6209 <data special_zero="true" />
6210 <input>
6211 <port id="0" precision="FP32">
6212 <dim>-1</dim>
6213 <dim>-1</dim>
6214 <dim>12</dim>
6215 <dim>32</dim>
6216 </port>
6217 <port id="1" precision="I64">
6218 <dim>3</dim>
6219 </port>
6220 </input>
6221 <output>
6222 <port id="2" precision="FP32" names="324">
6223 <dim>-1</dim>
6224 <dim>-1</dim>
6225 <dim>384</dim>
6226 </port>
6227 </output>
6228 </layer>
6229 <layer id="453" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6230 <data element_type="f32" shape="1, 1, 384" offset="17944632" size="1536" />
6231 <output>
6232 <port id="0" precision="FP32">
6233 <dim>1</dim>
6234 <dim>1</dim>
6235 <dim>384</dim>
6236 </port>
6237 </output>
6238 </layer>
6239 <layer id="454" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6240 <data auto_broadcast="numpy" />
6241 <input>
6242 <port id="0" precision="FP32">
6243 <dim>-1</dim>
6244 <dim>-1</dim>
6245 <dim>384</dim>
6246 </port>
6247 <port id="1" precision="FP32">
6248 <dim>1</dim>
6249 <dim>1</dim>
6250 <dim>384</dim>
6251 </port>
6252 </input>
6253 <output>
6254 <port id="2" precision="FP32">
6255 <dim>-1</dim>
6256 <dim>-1</dim>
6257 <dim>384</dim>
6258 </port>
6259 </output>
6260 </layer>
6261 <layer id="455" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6262 <data element_type="f32" shape="" offset="17946168" size="4" />
6263 <output>
6264 <port id="0" precision="FP32" />
6265 </output>
6266 </layer>
6267 <layer id="456" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6268 <data element_type="f32" shape="" offset="17946172" size="4" />
6269 <output>
6270 <port id="0" precision="FP32" />
6271 </output>
6272 </layer>
6273 <layer id="457" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6274 <data element_type="f32" shape="" offset="17946168" size="4" />
6275 <output>
6276 <port id="0" precision="FP32" />
6277 </output>
6278 </layer>
6279 <layer id="458" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6280 <data element_type="f32" shape="" offset="17946172" size="4" />
6281 <output>
6282 <port id="0" precision="FP32" />
6283 </output>
6284 </layer>
6285 <layer id="459" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6286 <data levels="256" auto_broadcast="numpy" />
6287 <input>
6288 <port id="0" precision="FP32">
6289 <dim>-1</dim>
6290 <dim>-1</dim>
6291 <dim>384</dim>
6292 </port>
6293 <port id="1" precision="FP32" />
6294 <port id="2" precision="FP32" />
6295 <port id="3" precision="FP32" />
6296 <port id="4" precision="FP32" />
6297 </input>
6298 <output>
6299 <port id="5" precision="FP32">
6300 <dim>-1</dim>
6301 <dim>-1</dim>
6302 <dim>384</dim>
6303 </port>
6304 </output>
6305 </layer>
6306 <layer id="460" name="Constant_445594" type="Const" version="opset1">
6307 <data element_type="i8" shape="384, 384" offset="17946176" size="147456" />
6308 <output>
6309 <port id="0" precision="I8">
6310 <dim>384</dim>
6311 <dim>384</dim>
6312 </port>
6313 </output>
6314 </layer>
6315 <layer id="461" name="Convert_445595" type="Convert" version="opset1">
6316 <data destination_type="f32" />
6317 <input>
6318 <port id="0" precision="I8">
6319 <dim>384</dim>
6320 <dim>384</dim>
6321 </port>
6322 </input>
6323 <output>
6324 <port id="1" precision="FP32">
6325 <dim>384</dim>
6326 <dim>384</dim>
6327 </port>
6328 </output>
6329 </layer>
6330 <layer id="462" name="Constant_445596" type="Const" version="opset1">
6331 <data element_type="f32" shape="384, 1" offset="18093632" size="1536" />
6332 <output>
6333 <port id="0" precision="FP32">
6334 <dim>384</dim>
6335 <dim>1</dim>
6336 </port>
6337 </output>
6338 </layer>
6339 <layer id="463" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6340 <data auto_broadcast="numpy" />
6341 <input>
6342 <port id="0" precision="FP32">
6343 <dim>384</dim>
6344 <dim>384</dim>
6345 </port>
6346 <port id="1" precision="FP32">
6347 <dim>384</dim>
6348 <dim>1</dim>
6349 </port>
6350 </input>
6351 <output>
6352 <port id="2" precision="FP32">
6353 <dim>384</dim>
6354 <dim>384</dim>
6355 </port>
6356 </output>
6357 </layer>
6358 <layer id="464" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6359 <data transpose_a="false" transpose_b="true" />
6360 <input>
6361 <port id="0" precision="FP32">
6362 <dim>-1</dim>
6363 <dim>-1</dim>
6364 <dim>384</dim>
6365 </port>
6366 <port id="1" precision="FP32">
6367 <dim>384</dim>
6368 <dim>384</dim>
6369 </port>
6370 </input>
6371 <output>
6372 <port id="2" precision="FP32">
6373 <dim>-1</dim>
6374 <dim>-1</dim>
6375 <dim>384</dim>
6376 </port>
6377 </output>
6378 </layer>
6379 <layer id="465" name="Constant_305906" type="Const" version="opset1">
6380 <data element_type="f32" shape="1, 1, 384" offset="18095168" size="1536" />
6381 <output>
6382 <port id="0" precision="FP32">
6383 <dim>1</dim>
6384 <dim>1</dim>
6385 <dim>384</dim>
6386 </port>
6387 </output>
6388 </layer>
6389 <layer id="466" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
6390 <data auto_broadcast="numpy" />
6391 <input>
6392 <port id="0" precision="FP32">
6393 <dim>-1</dim>
6394 <dim>-1</dim>
6395 <dim>384</dim>
6396 </port>
6397 <port id="1" precision="FP32">
6398 <dim>1</dim>
6399 <dim>1</dim>
6400 <dim>384</dim>
6401 </port>
6402 </input>
6403 <output>
6404 <port id="2" precision="FP32" names="329,input.15">
6405 <dim>-1</dim>
6406 <dim>-1</dim>
6407 <dim>384</dim>
6408 </port>
6409 </output>
6410 </layer>
6411 <layer id="467" name="__module.bert.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
6412 <data auto_broadcast="numpy" />
6413 <input>
6414 <port id="0" precision="FP32">
6415 <dim>-1</dim>
6416 <dim>-1</dim>
6417 <dim>384</dim>
6418 </port>
6419 <port id="1" precision="FP32">
6420 <dim>-1</dim>
6421 <dim>-1</dim>
6422 <dim>384</dim>
6423 </port>
6424 </input>
6425 <output>
6426 <port id="2" precision="FP32" names="331">
6427 <dim>-1</dim>
6428 <dim>-1</dim>
6429 <dim>384</dim>
6430 </port>
6431 </output>
6432 </layer>
6433 <layer id="468" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6434 <data element_type="i32" shape="1" offset="12046084" size="4" />
6435 <output>
6436 <port id="0" precision="I32">
6437 <dim>1</dim>
6438 </port>
6439 </output>
6440 </layer>
6441 <layer id="469" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6442 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6443 <input>
6444 <port id="0" precision="FP32">
6445 <dim>-1</dim>
6446 <dim>-1</dim>
6447 <dim>384</dim>
6448 </port>
6449 <port id="1" precision="I32">
6450 <dim>1</dim>
6451 </port>
6452 </input>
6453 <output>
6454 <port id="2" precision="FP32">
6455 <dim>-1</dim>
6456 <dim>-1</dim>
6457 <dim>384</dim>
6458 </port>
6459 </output>
6460 </layer>
6461 <layer id="470" name="Constant_305907" type="Const" version="opset1">
6462 <data element_type="f32" shape="1, 1, 384" offset="18096704" size="1536" />
6463 <output>
6464 <port id="0" precision="FP32">
6465 <dim>1</dim>
6466 <dim>1</dim>
6467 <dim>384</dim>
6468 </port>
6469 </output>
6470 </layer>
6471 <layer id="471" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6472 <data auto_broadcast="numpy" />
6473 <input>
6474 <port id="0" precision="FP32">
6475 <dim>-1</dim>
6476 <dim>-1</dim>
6477 <dim>384</dim>
6478 </port>
6479 <port id="1" precision="FP32">
6480 <dim>1</dim>
6481 <dim>1</dim>
6482 <dim>384</dim>
6483 </port>
6484 </input>
6485 <output>
6486 <port id="2" precision="FP32">
6487 <dim>-1</dim>
6488 <dim>-1</dim>
6489 <dim>384</dim>
6490 </port>
6491 </output>
6492 </layer>
6493 <layer id="472" name="Constant_305908" type="Const" version="opset1">
6494 <data element_type="f32" shape="1, 1, 384" offset="18098240" size="1536" />
6495 <output>
6496 <port id="0" precision="FP32">
6497 <dim>1</dim>
6498 <dim>1</dim>
6499 <dim>384</dim>
6500 </port>
6501 </output>
6502 </layer>
6503 <layer id="473" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6504 <data auto_broadcast="numpy" />
6505 <input>
6506 <port id="0" precision="FP32">
6507 <dim>-1</dim>
6508 <dim>-1</dim>
6509 <dim>384</dim>
6510 </port>
6511 <port id="1" precision="FP32">
6512 <dim>1</dim>
6513 <dim>1</dim>
6514 <dim>384</dim>
6515 </port>
6516 </input>
6517 <output>
6518 <port id="2" precision="FP32" names="335,input_tensor">
6519 <dim>-1</dim>
6520 <dim>-1</dim>
6521 <dim>384</dim>
6522 </port>
6523 </output>
6524 </layer>
6525 <layer id="474" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6526 <data element_type="f32" shape="1, 1, 384" offset="18099776" size="1536" />
6527 <output>
6528 <port id="0" precision="FP32">
6529 <dim>1</dim>
6530 <dim>1</dim>
6531 <dim>384</dim>
6532 </port>
6533 </output>
6534 </layer>
6535 <layer id="475" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6536 <data auto_broadcast="numpy" />
6537 <input>
6538 <port id="0" precision="FP32">
6539 <dim>-1</dim>
6540 <dim>-1</dim>
6541 <dim>384</dim>
6542 </port>
6543 <port id="1" precision="FP32">
6544 <dim>1</dim>
6545 <dim>1</dim>
6546 <dim>384</dim>
6547 </port>
6548 </input>
6549 <output>
6550 <port id="2" precision="FP32">
6551 <dim>-1</dim>
6552 <dim>-1</dim>
6553 <dim>384</dim>
6554 </port>
6555 </output>
6556 </layer>
6557 <layer id="476" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6558 <data element_type="f32" shape="" offset="18101312" size="4" />
6559 <output>
6560 <port id="0" precision="FP32" />
6561 </output>
6562 </layer>
6563 <layer id="477" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6564 <data element_type="f32" shape="" offset="18101316" size="4" />
6565 <output>
6566 <port id="0" precision="FP32" />
6567 </output>
6568 </layer>
6569 <layer id="478" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6570 <data element_type="f32" shape="" offset="18101312" size="4" />
6571 <output>
6572 <port id="0" precision="FP32" />
6573 </output>
6574 </layer>
6575 <layer id="479" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6576 <data element_type="f32" shape="" offset="18101316" size="4" />
6577 <output>
6578 <port id="0" precision="FP32" />
6579 </output>
6580 </layer>
6581 <layer id="480" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6582 <data levels="256" auto_broadcast="numpy" />
6583 <input>
6584 <port id="0" precision="FP32">
6585 <dim>-1</dim>
6586 <dim>-1</dim>
6587 <dim>384</dim>
6588 </port>
6589 <port id="1" precision="FP32" />
6590 <port id="2" precision="FP32" />
6591 <port id="3" precision="FP32" />
6592 <port id="4" precision="FP32" />
6593 </input>
6594 <output>
6595 <port id="5" precision="FP32">
6596 <dim>-1</dim>
6597 <dim>-1</dim>
6598 <dim>384</dim>
6599 </port>
6600 </output>
6601 </layer>
6602 <layer id="481" name="Constant_445598" type="Const" version="opset1">
6603 <data element_type="i8" shape="1536, 384" offset="18101320" size="589824" />
6604 <output>
6605 <port id="0" precision="I8">
6606 <dim>1536</dim>
6607 <dim>384</dim>
6608 </port>
6609 </output>
6610 </layer>
6611 <layer id="482" name="Convert_445599" type="Convert" version="opset1">
6612 <data destination_type="f32" />
6613 <input>
6614 <port id="0" precision="I8">
6615 <dim>1536</dim>
6616 <dim>384</dim>
6617 </port>
6618 </input>
6619 <output>
6620 <port id="1" precision="FP32">
6621 <dim>1536</dim>
6622 <dim>384</dim>
6623 </port>
6624 </output>
6625 </layer>
6626 <layer id="483" name="Constant_445600" type="Const" version="opset1">
6627 <data element_type="f32" shape="1536, 1" offset="18691144" size="6144" />
6628 <output>
6629 <port id="0" precision="FP32">
6630 <dim>1536</dim>
6631 <dim>1</dim>
6632 </port>
6633 </output>
6634 </layer>
6635 <layer id="484" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6636 <data auto_broadcast="numpy" />
6637 <input>
6638 <port id="0" precision="FP32">
6639 <dim>1536</dim>
6640 <dim>384</dim>
6641 </port>
6642 <port id="1" precision="FP32">
6643 <dim>1536</dim>
6644 <dim>1</dim>
6645 </port>
6646 </input>
6647 <output>
6648 <port id="2" precision="FP32">
6649 <dim>1536</dim>
6650 <dim>384</dim>
6651 </port>
6652 </output>
6653 </layer>
6654 <layer id="485" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6655 <data transpose_a="false" transpose_b="true" />
6656 <input>
6657 <port id="0" precision="FP32">
6658 <dim>-1</dim>
6659 <dim>-1</dim>
6660 <dim>384</dim>
6661 </port>
6662 <port id="1" precision="FP32">
6663 <dim>1536</dim>
6664 <dim>384</dim>
6665 </port>
6666 </input>
6667 <output>
6668 <port id="2" precision="FP32">
6669 <dim>-1</dim>
6670 <dim>-1</dim>
6671 <dim>1536</dim>
6672 </port>
6673 </output>
6674 </layer>
6675 <layer id="486" name="Constant_305909" type="Const" version="opset1">
6676 <data element_type="f32" shape="1, 1, 1536" offset="18697288" size="6144" />
6677 <output>
6678 <port id="0" precision="FP32">
6679 <dim>1</dim>
6680 <dim>1</dim>
6681 <dim>1536</dim>
6682 </port>
6683 </output>
6684 </layer>
6685 <layer id="487" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
6686 <data auto_broadcast="numpy" />
6687 <input>
6688 <port id="0" precision="FP32">
6689 <dim>-1</dim>
6690 <dim>-1</dim>
6691 <dim>1536</dim>
6692 </port>
6693 <port id="1" precision="FP32">
6694 <dim>1</dim>
6695 <dim>1</dim>
6696 <dim>1536</dim>
6697 </port>
6698 </input>
6699 <output>
6700 <port id="2" precision="FP32" names="339">
6701 <dim>-1</dim>
6702 <dim>-1</dim>
6703 <dim>1536</dim>
6704 </port>
6705 </output>
6706 </layer>
6707 <layer id="488" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
6708 <data approximation_mode="ERF" />
6709 <input>
6710 <port id="0" precision="FP32">
6711 <dim>-1</dim>
6712 <dim>-1</dim>
6713 <dim>1536</dim>
6714 </port>
6715 </input>
6716 <output>
6717 <port id="1" precision="FP32" names="340">
6718 <dim>-1</dim>
6719 <dim>-1</dim>
6720 <dim>1536</dim>
6721 </port>
6722 </output>
6723 </layer>
6724 <layer id="489" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6725 <data element_type="f32" shape="1, 1, 1536" offset="18703432" size="6144" />
6726 <output>
6727 <port id="0" precision="FP32">
6728 <dim>1</dim>
6729 <dim>1</dim>
6730 <dim>1536</dim>
6731 </port>
6732 </output>
6733 </layer>
6734 <layer id="490" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6735 <data auto_broadcast="numpy" />
6736 <input>
6737 <port id="0" precision="FP32">
6738 <dim>-1</dim>
6739 <dim>-1</dim>
6740 <dim>1536</dim>
6741 </port>
6742 <port id="1" precision="FP32">
6743 <dim>1</dim>
6744 <dim>1</dim>
6745 <dim>1536</dim>
6746 </port>
6747 </input>
6748 <output>
6749 <port id="2" precision="FP32">
6750 <dim>-1</dim>
6751 <dim>-1</dim>
6752 <dim>1536</dim>
6753 </port>
6754 </output>
6755 </layer>
6756 <layer id="491" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6757 <data element_type="f32" shape="" offset="18709576" size="4" />
6758 <output>
6759 <port id="0" precision="FP32" />
6760 </output>
6761 </layer>
6762 <layer id="492" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6763 <data element_type="f32" shape="" offset="18709580" size="4" />
6764 <output>
6765 <port id="0" precision="FP32" />
6766 </output>
6767 </layer>
6768 <layer id="493" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6769 <data element_type="f32" shape="" offset="18709576" size="4" />
6770 <output>
6771 <port id="0" precision="FP32" />
6772 </output>
6773 </layer>
6774 <layer id="494" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6775 <data element_type="f32" shape="" offset="18709580" size="4" />
6776 <output>
6777 <port id="0" precision="FP32" />
6778 </output>
6779 </layer>
6780 <layer id="495" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6781 <data levels="256" auto_broadcast="numpy" />
6782 <input>
6783 <port id="0" precision="FP32">
6784 <dim>-1</dim>
6785 <dim>-1</dim>
6786 <dim>1536</dim>
6787 </port>
6788 <port id="1" precision="FP32" />
6789 <port id="2" precision="FP32" />
6790 <port id="3" precision="FP32" />
6791 <port id="4" precision="FP32" />
6792 </input>
6793 <output>
6794 <port id="5" precision="FP32">
6795 <dim>-1</dim>
6796 <dim>-1</dim>
6797 <dim>1536</dim>
6798 </port>
6799 </output>
6800 </layer>
6801 <layer id="496" name="Constant_445602" type="Const" version="opset1">
6802 <data element_type="i8" shape="384, 1536" offset="18709584" size="589824" />
6803 <output>
6804 <port id="0" precision="I8">
6805 <dim>384</dim>
6806 <dim>1536</dim>
6807 </port>
6808 </output>
6809 </layer>
6810 <layer id="497" name="Convert_445603" type="Convert" version="opset1">
6811 <data destination_type="f32" />
6812 <input>
6813 <port id="0" precision="I8">
6814 <dim>384</dim>
6815 <dim>1536</dim>
6816 </port>
6817 </input>
6818 <output>
6819 <port id="1" precision="FP32">
6820 <dim>384</dim>
6821 <dim>1536</dim>
6822 </port>
6823 </output>
6824 </layer>
6825 <layer id="498" name="Constant_445604" type="Const" version="opset1">
6826 <data element_type="f32" shape="384, 1" offset="19299408" size="1536" />
6827 <output>
6828 <port id="0" precision="FP32">
6829 <dim>384</dim>
6830 <dim>1</dim>
6831 </port>
6832 </output>
6833 </layer>
6834 <layer id="499" name="__module.bert.encoder.layer.3.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6835 <data auto_broadcast="numpy" />
6836 <input>
6837 <port id="0" precision="FP32">
6838 <dim>384</dim>
6839 <dim>1536</dim>
6840 </port>
6841 <port id="1" precision="FP32">
6842 <dim>384</dim>
6843 <dim>1</dim>
6844 </port>
6845 </input>
6846 <output>
6847 <port id="2" precision="FP32">
6848 <dim>384</dim>
6849 <dim>1536</dim>
6850 </port>
6851 </output>
6852 </layer>
6853 <layer id="500" name="__module.bert.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6854 <data transpose_a="false" transpose_b="true" />
6855 <input>
6856 <port id="0" precision="FP32">
6857 <dim>-1</dim>
6858 <dim>-1</dim>
6859 <dim>1536</dim>
6860 </port>
6861 <port id="1" precision="FP32">
6862 <dim>384</dim>
6863 <dim>1536</dim>
6864 </port>
6865 </input>
6866 <output>
6867 <port id="2" precision="FP32">
6868 <dim>-1</dim>
6869 <dim>-1</dim>
6870 <dim>384</dim>
6871 </port>
6872 </output>
6873 </layer>
6874 <layer id="501" name="Constant_305910" type="Const" version="opset1">
6875 <data element_type="f32" shape="1, 1, 384" offset="19300944" size="1536" />
6876 <output>
6877 <port id="0" precision="FP32">
6878 <dim>1</dim>
6879 <dim>1</dim>
6880 <dim>384</dim>
6881 </port>
6882 </output>
6883 </layer>
6884 <layer id="502" name="__module.bert.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
6885 <data auto_broadcast="numpy" />
6886 <input>
6887 <port id="0" precision="FP32">
6888 <dim>-1</dim>
6889 <dim>-1</dim>
6890 <dim>384</dim>
6891 </port>
6892 <port id="1" precision="FP32">
6893 <dim>1</dim>
6894 <dim>1</dim>
6895 <dim>384</dim>
6896 </port>
6897 </input>
6898 <output>
6899 <port id="2" precision="FP32" names="345,input.17">
6900 <dim>-1</dim>
6901 <dim>-1</dim>
6902 <dim>384</dim>
6903 </port>
6904 </output>
6905 </layer>
6906 <layer id="503" name="__module.bert.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
6907 <data auto_broadcast="numpy" />
6908 <input>
6909 <port id="0" precision="FP32">
6910 <dim>-1</dim>
6911 <dim>-1</dim>
6912 <dim>384</dim>
6913 </port>
6914 <port id="1" precision="FP32">
6915 <dim>-1</dim>
6916 <dim>-1</dim>
6917 <dim>384</dim>
6918 </port>
6919 </input>
6920 <output>
6921 <port id="2" precision="FP32" names="347">
6922 <dim>-1</dim>
6923 <dim>-1</dim>
6924 <dim>384</dim>
6925 </port>
6926 </output>
6927 </layer>
6928 <layer id="504" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6929 <data element_type="i32" shape="1" offset="12046084" size="4" />
6930 <output>
6931 <port id="0" precision="I32">
6932 <dim>1</dim>
6933 </port>
6934 </output>
6935 </layer>
6936 <layer id="505" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6937 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6938 <input>
6939 <port id="0" precision="FP32">
6940 <dim>-1</dim>
6941 <dim>-1</dim>
6942 <dim>384</dim>
6943 </port>
6944 <port id="1" precision="I32">
6945 <dim>1</dim>
6946 </port>
6947 </input>
6948 <output>
6949 <port id="2" precision="FP32">
6950 <dim>-1</dim>
6951 <dim>-1</dim>
6952 <dim>384</dim>
6953 </port>
6954 </output>
6955 </layer>
6956 <layer id="506" name="Constant_305911" type="Const" version="opset1">
6957 <data element_type="f32" shape="1, 1, 384" offset="19302480" size="1536" />
6958 <output>
6959 <port id="0" precision="FP32">
6960 <dim>1</dim>
6961 <dim>1</dim>
6962 <dim>384</dim>
6963 </port>
6964 </output>
6965 </layer>
6966 <layer id="507" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6967 <data auto_broadcast="numpy" />
6968 <input>
6969 <port id="0" precision="FP32">
6970 <dim>-1</dim>
6971 <dim>-1</dim>
6972 <dim>384</dim>
6973 </port>
6974 <port id="1" precision="FP32">
6975 <dim>1</dim>
6976 <dim>1</dim>
6977 <dim>384</dim>
6978 </port>
6979 </input>
6980 <output>
6981 <port id="2" precision="FP32">
6982 <dim>-1</dim>
6983 <dim>-1</dim>
6984 <dim>384</dim>
6985 </port>
6986 </output>
6987 </layer>
6988 <layer id="508" name="Constant_305912" type="Const" version="opset1">
6989 <data element_type="f32" shape="1, 1, 384" offset="19304016" size="1536" />
6990 <output>
6991 <port id="0" precision="FP32">
6992 <dim>1</dim>
6993 <dim>1</dim>
6994 <dim>384</dim>
6995 </port>
6996 </output>
6997 </layer>
6998 <layer id="509" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6999 <data auto_broadcast="numpy" />
7000 <input>
7001 <port id="0" precision="FP32">
7002 <dim>-1</dim>
7003 <dim>-1</dim>
7004 <dim>384</dim>
7005 </port>
7006 <port id="1" precision="FP32">
7007 <dim>1</dim>
7008 <dim>1</dim>
7009 <dim>384</dim>
7010 </port>
7011 </input>
7012 <output>
7013 <port id="2" precision="FP32" names="351,353,hidden_states">
7014 <dim>-1</dim>
7015 <dim>-1</dim>
7016 <dim>384</dim>
7017 </port>
7018 </output>
7019 </layer>
7020 <layer id="510" name="23" type="Const" version="opset1">
7021 <data element_type="i64" shape="" offset="12046068" size="8" />
7022 <output>
7023 <port id="0" precision="I64" names="23" />
7024 </output>
7025 </layer>
7026 <layer id="511" name="__module.bert.pooler/aten::select/Gather" type="Gather" version="opset8">
7027 <data batch_dims="0" />
7028 <input>
7029 <port id="0" precision="FP32">
7030 <dim>-1</dim>
7031 <dim>-1</dim>
7032 <dim>384</dim>
7033 </port>
7034 <port id="1" precision="I64" />
7035 <port id="2" precision="I64" />
7036 </input>
7037 <output>
7038 <port id="3" precision="FP32" names="354">
7039 <dim>-1</dim>
7040 <dim>384</dim>
7041 </port>
7042 </output>
7043 </layer>
7044 <layer id="512" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7045 <data element_type="f32" shape="1, 384" offset="19305552" size="1536" />
7046 <output>
7047 <port id="0" precision="FP32">
7048 <dim>1</dim>
7049 <dim>384</dim>
7050 </port>
7051 </output>
7052 </layer>
7053 <layer id="513" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7054 <data auto_broadcast="numpy" />
7055 <input>
7056 <port id="0" precision="FP32">
7057 <dim>-1</dim>
7058 <dim>384</dim>
7059 </port>
7060 <port id="1" precision="FP32">
7061 <dim>1</dim>
7062 <dim>384</dim>
7063 </port>
7064 </input>
7065 <output>
7066 <port id="2" precision="FP32">
7067 <dim>-1</dim>
7068 <dim>384</dim>
7069 </port>
7070 </output>
7071 </layer>
7072 <layer id="514" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7073 <data element_type="f32" shape="" offset="19307088" size="4" />
7074 <output>
7075 <port id="0" precision="FP32" />
7076 </output>
7077 </layer>
7078 <layer id="515" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7079 <data element_type="f32" shape="" offset="19307092" size="4" />
7080 <output>
7081 <port id="0" precision="FP32" />
7082 </output>
7083 </layer>
7084 <layer id="516" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7085 <data element_type="f32" shape="" offset="19307088" size="4" />
7086 <output>
7087 <port id="0" precision="FP32" />
7088 </output>
7089 </layer>
7090 <layer id="517" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7091 <data element_type="f32" shape="" offset="19307092" size="4" />
7092 <output>
7093 <port id="0" precision="FP32" />
7094 </output>
7095 </layer>
7096 <layer id="518" name="__module.bert.pooler/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7097 <data levels="256" auto_broadcast="numpy" />
7098 <input>
7099 <port id="0" precision="FP32">
7100 <dim>-1</dim>
7101 <dim>384</dim>
7102 </port>
7103 <port id="1" precision="FP32" />
7104 <port id="2" precision="FP32" />
7105 <port id="3" precision="FP32" />
7106 <port id="4" precision="FP32" />
7107 </input>
7108 <output>
7109 <port id="5" precision="FP32">
7110 <dim>-1</dim>
7111 <dim>384</dim>
7112 </port>
7113 </output>
7114 </layer>
7115 <layer id="519" name="Constant_445606" type="Const" version="opset1">
7116 <data element_type="i8" shape="384, 384" offset="19307096" size="147456" />
7117 <output>
7118 <port id="0" precision="I8">
7119 <dim>384</dim>
7120 <dim>384</dim>
7121 </port>
7122 </output>
7123 </layer>
7124 <layer id="520" name="Convert_445607" type="Convert" version="opset1">
7125 <data destination_type="f32" />
7126 <input>
7127 <port id="0" precision="I8">
7128 <dim>384</dim>
7129 <dim>384</dim>
7130 </port>
7131 </input>
7132 <output>
7133 <port id="1" precision="FP32">
7134 <dim>384</dim>
7135 <dim>384</dim>
7136 </port>
7137 </output>
7138 </layer>
7139 <layer id="521" name="Constant_445608" type="Const" version="opset1">
7140 <data element_type="f32" shape="384, 1" offset="19454552" size="1536" />
7141 <output>
7142 <port id="0" precision="FP32">
7143 <dim>384</dim>
7144 <dim>1</dim>
7145 </port>
7146 </output>
7147 </layer>
7148 <layer id="522" name="__module.bert.pooler.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7149 <data auto_broadcast="numpy" />
7150 <input>
7151 <port id="0" precision="FP32">
7152 <dim>384</dim>
7153 <dim>384</dim>
7154 </port>
7155 <port id="1" precision="FP32">
7156 <dim>384</dim>
7157 <dim>1</dim>
7158 </port>
7159 </input>
7160 <output>
7161 <port id="2" precision="FP32">
7162 <dim>384</dim>
7163 <dim>384</dim>
7164 </port>
7165 </output>
7166 </layer>
7167 <layer id="523" name="__module.bert.pooler.dense/aten::linear/MatMul" type="MatMul" version="opset1">
7168 <data transpose_a="false" transpose_b="true" />
7169 <input>
7170 <port id="0" precision="FP32">
7171 <dim>-1</dim>
7172 <dim>384</dim>
7173 </port>
7174 <port id="1" precision="FP32">
7175 <dim>384</dim>
7176 <dim>384</dim>
7177 </port>
7178 </input>
7179 <output>
7180 <port id="2" precision="FP32">
7181 <dim>-1</dim>
7182 <dim>384</dim>
7183 </port>
7184 </output>
7185 </layer>
7186 <layer id="524" name="Constant_305913" type="Const" version="opset1">
7187 <data element_type="f32" shape="1, 384" offset="19456088" size="1536" />
7188 <output>
7189 <port id="0" precision="FP32">
7190 <dim>1</dim>
7191 <dim>384</dim>
7192 </port>
7193 </output>
7194 </layer>
7195 <layer id="525" name="__module.bert.pooler.dense/aten::linear/Add" type="Add" version="opset1">
7196 <data auto_broadcast="numpy" />
7197 <input>
7198 <port id="0" precision="FP32">
7199 <dim>-1</dim>
7200 <dim>384</dim>
7201 </port>
7202 <port id="1" precision="FP32">
7203 <dim>1</dim>
7204 <dim>384</dim>
7205 </port>
7206 </input>
7207 <output>
7208 <port id="2" precision="FP32" names="357">
7209 <dim>-1</dim>
7210 <dim>384</dim>
7211 </port>
7212 </output>
7213 </layer>
7214 <layer id="526" name="__module.bert.pooler.activation/aten::tanh/Tanh" type="Tanh" version="opset1">
7215 <input>
7216 <port id="0" precision="FP32">
7217 <dim>-1</dim>
7218 <dim>384</dim>
7219 </port>
7220 </input>
7221 <output>
7222 <port id="1" precision="FP32" names="358,input">
7223 <dim>-1</dim>
7224 <dim>384</dim>
7225 </port>
7226 </output>
7227 </layer>
7228 <layer id="527" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7229 <data element_type="f32" shape="1, 384" offset="19457624" size="1536" />
7230 <output>
7231 <port id="0" precision="FP32">
7232 <dim>1</dim>
7233 <dim>384</dim>
7234 </port>
7235 </output>
7236 </layer>
7237 <layer id="528" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7238 <data auto_broadcast="numpy" />
7239 <input>
7240 <port id="0" precision="FP32">
7241 <dim>-1</dim>
7242 <dim>384</dim>
7243 </port>
7244 <port id="1" precision="FP32">
7245 <dim>1</dim>
7246 <dim>384</dim>
7247 </port>
7248 </input>
7249 <output>
7250 <port id="2" precision="FP32">
7251 <dim>-1</dim>
7252 <dim>384</dim>
7253 </port>
7254 </output>
7255 </layer>
7256 <layer id="529" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7257 <data element_type="f32" shape="" offset="19459160" size="4" />
7258 <output>
7259 <port id="0" precision="FP32" />
7260 </output>
7261 </layer>
7262 <layer id="530" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7263 <data element_type="f32" shape="" offset="19459164" size="4" />
7264 <output>
7265 <port id="0" precision="FP32" />
7266 </output>
7267 </layer>
7268 <layer id="531" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7269 <data element_type="f32" shape="" offset="19459160" size="4" />
7270 <output>
7271 <port id="0" precision="FP32" />
7272 </output>
7273 </layer>
7274 <layer id="532" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7275 <data element_type="f32" shape="" offset="19459164" size="4" />
7276 <output>
7277 <port id="0" precision="FP32" />
7278 </output>
7279 </layer>
7280 <layer id="533" name="__module.bert.pooler.activation/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7281 <data levels="256" auto_broadcast="numpy" />
7282 <input>
7283 <port id="0" precision="FP32">
7284 <dim>-1</dim>
7285 <dim>384</dim>
7286 </port>
7287 <port id="1" precision="FP32" />
7288 <port id="2" precision="FP32" />
7289 <port id="3" precision="FP32" />
7290 <port id="4" precision="FP32" />
7291 </input>
7292 <output>
7293 <port id="5" precision="FP32">
7294 <dim>-1</dim>
7295 <dim>384</dim>
7296 </port>
7297 </output>
7298 </layer>
7299 <layer id="534" name="Constant_445610" type="Const" version="opset1">
7300 <data element_type="i8" shape="1, 384" offset="19459168" size="384" />
7301 <output>
7302 <port id="0" precision="I8">
7303 <dim>1</dim>
7304 <dim>384</dim>
7305 </port>
7306 </output>
7307 </layer>
7308 <layer id="535" name="Convert_445611" type="Convert" version="opset1">
7309 <data destination_type="f32" />
7310 <input>
7311 <port id="0" precision="I8">
7312 <dim>1</dim>
7313 <dim>384</dim>
7314 </port>
7315 </input>
7316 <output>
7317 <port id="1" precision="FP32">
7318 <dim>1</dim>
7319 <dim>384</dim>
7320 </port>
7321 </output>
7322 </layer>
7323 <layer id="536" name="Constant_445612" type="Const" version="opset1">
7324 <data element_type="f32" shape="1, 1" offset="19459552" size="4" />
7325 <output>
7326 <port id="0" precision="FP32">
7327 <dim>1</dim>
7328 <dim>1</dim>
7329 </port>
7330 </output>
7331 </layer>
7332 <layer id="537" name="__module.classifier/aten::linear/Add/fq_weights_1" type="Multiply" version="opset1">
7333 <data auto_broadcast="numpy" />
7334 <input>
7335 <port id="0" precision="FP32">
7336 <dim>1</dim>
7337 <dim>384</dim>
7338 </port>
7339 <port id="1" precision="FP32">
7340 <dim>1</dim>
7341 <dim>1</dim>
7342 </port>
7343 </input>
7344 <output>
7345 <port id="2" precision="FP32">
7346 <dim>1</dim>
7347 <dim>384</dim>
7348 </port>
7349 </output>
7350 </layer>
7351 <layer id="538" name="__module.classifier/aten::linear/Add" type="MatMul" version="opset1">
7352 <data transpose_a="false" transpose_b="true" />
7353 <input>
7354 <port id="0" precision="FP32">
7355 <dim>-1</dim>
7356 <dim>384</dim>
7357 </port>
7358 <port id="1" precision="FP32">
7359 <dim>1</dim>
7360 <dim>384</dim>
7361 </port>
7362 </input>
7363 <output>
7364 <port id="2" precision="FP32" names="logits">
7365 <dim>-1</dim>
7366 <dim>1</dim>
7367 </port>
7368 </output>
7369 </layer>
7370 <layer id="539" name="Result_302809" type="Result" version="opset1">
7371 <input>
7372 <port id="0" precision="FP32">
7373 <dim>-1</dim>
7374 <dim>1</dim>
7375 </port>
7376 </input>
7377 </layer>
7378 </layers>
7379 <edges>
7380 <edge from-layer="0" from-port="0" to-layer="14" to-port="0" />
7381 <edge from-layer="1" from-port="0" to-layer="93" to-port="0" />
7382 <edge from-layer="2" from-port="0" to-layer="7" to-port="0" />
7383 <edge from-layer="2" from-port="0" to-layer="24" to-port="0" />
7384 <edge from-layer="3" from-port="0" to-layer="4" to-port="0" />
7385 <edge from-layer="4" from-port="1" to-layer="6" to-port="0" />
7386 <edge from-layer="5" from-port="0" to-layer="6" to-port="1" />
7387 <edge from-layer="6" from-port="2" to-layer="9" to-port="0" />
7388 <edge from-layer="7" from-port="1" to-layer="9" to-port="1" />
7389 <edge from-layer="8" from-port="0" to-layer="9" to-port="2" />
7390 <edge from-layer="9" from-port="3" to-layer="17" to-port="0" />
7391 <edge from-layer="10" from-port="0" to-layer="11" to-port="0" />
7392 <edge from-layer="11" from-port="1" to-layer="13" to-port="0" />
7393 <edge from-layer="12" from-port="0" to-layer="13" to-port="1" />
7394 <edge from-layer="13" from-port="2" to-layer="16" to-port="0" />
7395 <edge from-layer="14" from-port="1" to-layer="16" to-port="1" />
7396 <edge from-layer="15" from-port="0" to-layer="16" to-port="2" />
7397 <edge from-layer="16" from-port="3" to-layer="17" to-port="1" />
7398 <edge from-layer="17" from-port="2" to-layer="34" to-port="0" />
7399 <edge from-layer="18" from-port="0" to-layer="19" to-port="0" />
7400 <edge from-layer="19" from-port="1" to-layer="21" to-port="0" />
7401 <edge from-layer="20" from-port="0" to-layer="21" to-port="1" />
7402 <edge from-layer="21" from-port="2" to-layer="33" to-port="0" />
7403 <edge from-layer="22" from-port="0" to-layer="30" to-port="0" />
7404 <edge from-layer="23" from-port="0" to-layer="30" to-port="1" />
7405 <edge from-layer="24" from-port="1" to-layer="27" to-port="0" />
7406 <edge from-layer="24" from-port="1" to-layer="98" to-port="0" />
7407 <edge from-layer="24" from-port="1" to-layer="102" to-port="0" />
7408 <edge from-layer="25" from-port="0" to-layer="27" to-port="1" />
7409 <edge from-layer="26" from-port="0" to-layer="27" to-port="2" />
7410 <edge from-layer="27" from-port="3" to-layer="30" to-port="2" />
7411 <edge from-layer="28" from-port="0" to-layer="30" to-port="3" />
7412 <edge from-layer="29" from-port="0" to-layer="30" to-port="4" />
7413 <edge from-layer="30" from-port="5" to-layer="31" to-port="0" />
7414 <edge from-layer="31" from-port="1" to-layer="33" to-port="1" />
7415 <edge from-layer="32" from-port="0" to-layer="33" to-port="2" />
7416 <edge from-layer="33" from-port="3" to-layer="34" to-port="1" />
7417 <edge from-layer="34" from-port="2" to-layer="36" to-port="0" />
7418 <edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
7419 <edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
7420 <edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
7421 <edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
7422 <edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
7423 <edge from-layer="40" from-port="2" to-layer="42" to-port="0" />
7424 <edge from-layer="40" from-port="2" to-layer="131" to-port="1" />
7425 <edge from-layer="41" from-port="0" to-layer="42" to-port="1" />
7426 <edge from-layer="42" from-port="2" to-layer="47" to-port="0" />
7427 <edge from-layer="43" from-port="0" to-layer="47" to-port="1" />
7428 <edge from-layer="44" from-port="0" to-layer="47" to-port="2" />
7429 <edge from-layer="45" from-port="0" to-layer="47" to-port="3" />
7430 <edge from-layer="46" from-port="0" to-layer="47" to-port="4" />
7431 <edge from-layer="47" from-port="5" to-layer="52" to-port="0" />
7432 <edge from-layer="47" from-port="5" to-layer="68" to-port="0" />
7433 <edge from-layer="47" from-port="5" to-layer="84" to-port="0" />
7434 <edge from-layer="48" from-port="0" to-layer="49" to-port="0" />
7435 <edge from-layer="49" from-port="1" to-layer="51" to-port="0" />
7436 <edge from-layer="50" from-port="0" to-layer="51" to-port="1" />
7437 <edge from-layer="51" from-port="2" to-layer="52" to-port="1" />
7438 <edge from-layer="52" from-port="2" to-layer="54" to-port="0" />
7439 <edge from-layer="53" from-port="0" to-layer="54" to-port="1" />
7440 <edge from-layer="54" from-port="2" to-layer="59" to-port="0" />
7441 <edge from-layer="55" from-port="0" to-layer="59" to-port="1" />
7442 <edge from-layer="56" from-port="0" to-layer="59" to-port="2" />
7443 <edge from-layer="57" from-port="0" to-layer="59" to-port="3" />
7444 <edge from-layer="58" from-port="0" to-layer="59" to-port="4" />
7445 <edge from-layer="59" from-port="5" to-layer="61" to-port="0" />
7446 <edge from-layer="60" from-port="0" to-layer="61" to-port="1" />
7447 <edge from-layer="61" from-port="2" to-layer="63" to-port="0" />
7448 <edge from-layer="62" from-port="0" to-layer="63" to-port="1" />
7449 <edge from-layer="63" from-port="2" to-layer="112" to-port="0" />
7450 <edge from-layer="64" from-port="0" to-layer="65" to-port="0" />
7451 <edge from-layer="65" from-port="1" to-layer="67" to-port="0" />
7452 <edge from-layer="66" from-port="0" to-layer="67" to-port="1" />
7453 <edge from-layer="67" from-port="2" to-layer="68" to-port="1" />
7454 <edge from-layer="68" from-port="2" to-layer="70" to-port="0" />
7455 <edge from-layer="69" from-port="0" to-layer="70" to-port="1" />
7456 <edge from-layer="70" from-port="2" to-layer="75" to-port="0" />
7457 <edge from-layer="71" from-port="0" to-layer="75" to-port="1" />
7458 <edge from-layer="72" from-port="0" to-layer="75" to-port="2" />
7459 <edge from-layer="73" from-port="0" to-layer="75" to-port="3" />
7460 <edge from-layer="74" from-port="0" to-layer="75" to-port="4" />
7461 <edge from-layer="75" from-port="5" to-layer="77" to-port="0" />
7462 <edge from-layer="76" from-port="0" to-layer="77" to-port="1" />
7463 <edge from-layer="77" from-port="2" to-layer="79" to-port="0" />
7464 <edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
7465 <edge from-layer="79" from-port="2" to-layer="112" to-port="1" />
7466 <edge from-layer="80" from-port="0" to-layer="81" to-port="0" />
7467 <edge from-layer="81" from-port="1" to-layer="83" to-port="0" />
7468 <edge from-layer="82" from-port="0" to-layer="83" to-port="1" />
7469 <edge from-layer="83" from-port="2" to-layer="84" to-port="1" />
7470 <edge from-layer="84" from-port="2" to-layer="86" to-port="0" />
7471 <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
7472 <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
7473 <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
7474 <edge from-layer="88" from-port="2" to-layer="90" to-port="0" />
7475 <edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
7476 <edge from-layer="90" from-port="2" to-layer="112" to-port="2" />
7477 <edge from-layer="91" from-port="0" to-layer="108" to-port="0" />
7478 <edge from-layer="92" from-port="0" to-layer="511" to-port="2" />
7479 <edge from-layer="92" from-port="0" to-layer="93" to-port="1" />
7480 <edge from-layer="93" from-port="2" to-layer="95" to-port="0" />
7481 <edge from-layer="94" from-port="0" to-layer="95" to-port="1" />
7482 <edge from-layer="95" from-port="2" to-layer="104" to-port="0" />
7483 <edge from-layer="96" from-port="0" to-layer="98" to-port="1" />
7484 <edge from-layer="97" from-port="0" to-layer="98" to-port="2" />
7485 <edge from-layer="98" from-port="3" to-layer="103" to-port="0" />
7486 <edge from-layer="99" from-port="0" to-layer="103" to-port="1" />
7487 <edge from-layer="100" from-port="0" to-layer="102" to-port="1" />
7488 <edge from-layer="101" from-port="0" to-layer="102" to-port="2" />
7489 <edge from-layer="102" from-port="3" to-layer="103" to-port="2" />
7490 <edge from-layer="103" from-port="3" to-layer="104" to-port="1" />
7491 <edge from-layer="104" from-port="2" to-layer="105" to-port="0" />
7492 <edge from-layer="105" from-port="1" to-layer="107" to-port="0" />
7493 <edge from-layer="106" from-port="0" to-layer="107" to-port="1" />
7494 <edge from-layer="107" from-port="2" to-layer="108" to-port="1" />
7495 <edge from-layer="108" from-port="2" to-layer="111" to-port="2" />
7496 <edge from-layer="108" from-port="2" to-layer="109" to-port="0" />
7497 <edge from-layer="109" from-port="1" to-layer="111" to-port="0" />
7498 <edge from-layer="110" from-port="0" to-layer="111" to-port="1" />
7499 <edge from-layer="111" from-port="3" to-layer="112" to-port="3" />
7500 <edge from-layer="111" from-port="3" to-layer="336" to-port="3" />
7501 <edge from-layer="111" from-port="3" to-layer="224" to-port="3" />
7502 <edge from-layer="111" from-port="3" to-layer="448" to-port="3" />
7503 <edge from-layer="112" from-port="4" to-layer="114" to-port="0" />
7504 <edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
7505 <edge from-layer="114" from-port="2" to-layer="116" to-port="0" />
7506 <edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
7507 <edge from-layer="116" from-port="2" to-layer="118" to-port="0" />
7508 <edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
7509 <edge from-layer="118" from-port="2" to-layer="123" to-port="0" />
7510 <edge from-layer="119" from-port="0" to-layer="123" to-port="1" />
7511 <edge from-layer="120" from-port="0" to-layer="123" to-port="2" />
7512 <edge from-layer="121" from-port="0" to-layer="123" to-port="3" />
7513 <edge from-layer="122" from-port="0" to-layer="123" to-port="4" />
7514 <edge from-layer="123" from-port="5" to-layer="128" to-port="0" />
7515 <edge from-layer="124" from-port="0" to-layer="125" to-port="0" />
7516 <edge from-layer="125" from-port="1" to-layer="127" to-port="0" />
7517 <edge from-layer="126" from-port="0" to-layer="127" to-port="1" />
7518 <edge from-layer="127" from-port="2" to-layer="128" to-port="1" />
7519 <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
7520 <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
7521 <edge from-layer="130" from-port="2" to-layer="131" to-port="0" />
7522 <edge from-layer="131" from-port="2" to-layer="133" to-port="0" />
7523 <edge from-layer="132" from-port="0" to-layer="133" to-port="1" />
7524 <edge from-layer="133" from-port="2" to-layer="135" to-port="0" />
7525 <edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
7526 <edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
7527 <edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
7528 <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
7529 <edge from-layer="137" from-port="2" to-layer="167" to-port="1" />
7530 <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
7531 <edge from-layer="139" from-port="2" to-layer="144" to-port="0" />
7532 <edge from-layer="140" from-port="0" to-layer="144" to-port="1" />
7533 <edge from-layer="141" from-port="0" to-layer="144" to-port="2" />
7534 <edge from-layer="142" from-port="0" to-layer="144" to-port="3" />
7535 <edge from-layer="143" from-port="0" to-layer="144" to-port="4" />
7536 <edge from-layer="144" from-port="5" to-layer="149" to-port="0" />
7537 <edge from-layer="145" from-port="0" to-layer="146" to-port="0" />
7538 <edge from-layer="146" from-port="1" to-layer="148" to-port="0" />
7539 <edge from-layer="147" from-port="0" to-layer="148" to-port="1" />
7540 <edge from-layer="148" from-port="2" to-layer="149" to-port="1" />
7541 <edge from-layer="149" from-port="2" to-layer="151" to-port="0" />
7542 <edge from-layer="150" from-port="0" to-layer="151" to-port="1" />
7543 <edge from-layer="151" from-port="2" to-layer="152" to-port="0" />
7544 <edge from-layer="152" from-port="1" to-layer="154" to-port="0" />
7545 <edge from-layer="153" from-port="0" to-layer="154" to-port="1" />
7546 <edge from-layer="154" from-port="2" to-layer="159" to-port="0" />
7547 <edge from-layer="155" from-port="0" to-layer="159" to-port="1" />
7548 <edge from-layer="156" from-port="0" to-layer="159" to-port="2" />
7549 <edge from-layer="157" from-port="0" to-layer="159" to-port="3" />
7550 <edge from-layer="158" from-port="0" to-layer="159" to-port="4" />
7551 <edge from-layer="159" from-port="5" to-layer="164" to-port="0" />
7552 <edge from-layer="160" from-port="0" to-layer="161" to-port="0" />
7553 <edge from-layer="161" from-port="1" to-layer="163" to-port="0" />
7554 <edge from-layer="162" from-port="0" to-layer="163" to-port="1" />
7555 <edge from-layer="163" from-port="2" to-layer="164" to-port="1" />
7556 <edge from-layer="164" from-port="2" to-layer="166" to-port="0" />
7557 <edge from-layer="165" from-port="0" to-layer="166" to-port="1" />
7558 <edge from-layer="166" from-port="2" to-layer="167" to-port="0" />
7559 <edge from-layer="167" from-port="2" to-layer="169" to-port="0" />
7560 <edge from-layer="168" from-port="0" to-layer="169" to-port="1" />
7561 <edge from-layer="169" from-port="2" to-layer="171" to-port="0" />
7562 <edge from-layer="170" from-port="0" to-layer="171" to-port="1" />
7563 <edge from-layer="171" from-port="2" to-layer="173" to-port="0" />
7564 <edge from-layer="172" from-port="0" to-layer="173" to-port="1" />
7565 <edge from-layer="173" from-port="2" to-layer="175" to-port="0" />
7566 <edge from-layer="173" from-port="2" to-layer="243" to-port="1" />
7567 <edge from-layer="174" from-port="0" to-layer="175" to-port="1" />
7568 <edge from-layer="175" from-port="2" to-layer="180" to-port="0" />
7569 <edge from-layer="176" from-port="0" to-layer="180" to-port="1" />
7570 <edge from-layer="177" from-port="0" to-layer="180" to-port="2" />
7571 <edge from-layer="178" from-port="0" to-layer="180" to-port="3" />
7572 <edge from-layer="179" from-port="0" to-layer="180" to-port="4" />
7573 <edge from-layer="180" from-port="5" to-layer="217" to-port="0" />
7574 <edge from-layer="180" from-port="5" to-layer="185" to-port="0" />
7575 <edge from-layer="180" from-port="5" to-layer="201" to-port="0" />
7576 <edge from-layer="181" from-port="0" to-layer="182" to-port="0" />
7577 <edge from-layer="182" from-port="1" to-layer="184" to-port="0" />
7578 <edge from-layer="183" from-port="0" to-layer="184" to-port="1" />
7579 <edge from-layer="184" from-port="2" to-layer="185" to-port="1" />
7580 <edge from-layer="185" from-port="2" to-layer="187" to-port="0" />
7581 <edge from-layer="186" from-port="0" to-layer="187" to-port="1" />
7582 <edge from-layer="187" from-port="2" to-layer="192" to-port="0" />
7583 <edge from-layer="188" from-port="0" to-layer="192" to-port="1" />
7584 <edge from-layer="189" from-port="0" to-layer="192" to-port="2" />
7585 <edge from-layer="190" from-port="0" to-layer="192" to-port="3" />
7586 <edge from-layer="191" from-port="0" to-layer="192" to-port="4" />
7587 <edge from-layer="192" from-port="5" to-layer="194" to-port="0" />
7588 <edge from-layer="193" from-port="0" to-layer="194" to-port="1" />
7589 <edge from-layer="194" from-port="2" to-layer="196" to-port="0" />
7590 <edge from-layer="195" from-port="0" to-layer="196" to-port="1" />
7591 <edge from-layer="196" from-port="2" to-layer="224" to-port="0" />
7592 <edge from-layer="197" from-port="0" to-layer="198" to-port="0" />
7593 <edge from-layer="198" from-port="1" to-layer="200" to-port="0" />
7594 <edge from-layer="199" from-port="0" to-layer="200" to-port="1" />
7595 <edge from-layer="200" from-port="2" to-layer="201" to-port="1" />
7596 <edge from-layer="201" from-port="2" to-layer="203" to-port="0" />
7597 <edge from-layer="202" from-port="0" to-layer="203" to-port="1" />
7598 <edge from-layer="203" from-port="2" to-layer="208" to-port="0" />
7599 <edge from-layer="204" from-port="0" to-layer="208" to-port="1" />
7600 <edge from-layer="205" from-port="0" to-layer="208" to-port="2" />
7601 <edge from-layer="206" from-port="0" to-layer="208" to-port="3" />
7602 <edge from-layer="207" from-port="0" to-layer="208" to-port="4" />
7603 <edge from-layer="208" from-port="5" to-layer="210" to-port="0" />
7604 <edge from-layer="209" from-port="0" to-layer="210" to-port="1" />
7605 <edge from-layer="210" from-port="2" to-layer="212" to-port="0" />
7606 <edge from-layer="211" from-port="0" to-layer="212" to-port="1" />
7607 <edge from-layer="212" from-port="2" to-layer="224" to-port="1" />
7608 <edge from-layer="213" from-port="0" to-layer="214" to-port="0" />
7609 <edge from-layer="214" from-port="1" to-layer="216" to-port="0" />
7610 <edge from-layer="215" from-port="0" to-layer="216" to-port="1" />
7611 <edge from-layer="216" from-port="2" to-layer="217" to-port="1" />
7612 <edge from-layer="217" from-port="2" to-layer="219" to-port="0" />
7613 <edge from-layer="218" from-port="0" to-layer="219" to-port="1" />
7614 <edge from-layer="219" from-port="2" to-layer="221" to-port="0" />
7615 <edge from-layer="220" from-port="0" to-layer="221" to-port="1" />
7616 <edge from-layer="221" from-port="2" to-layer="223" to-port="0" />
7617 <edge from-layer="222" from-port="0" to-layer="223" to-port="1" />
7618 <edge from-layer="223" from-port="2" to-layer="224" to-port="2" />
7619 <edge from-layer="224" from-port="4" to-layer="226" to-port="0" />
7620 <edge from-layer="225" from-port="0" to-layer="226" to-port="1" />
7621 <edge from-layer="226" from-port="2" to-layer="228" to-port="0" />
7622 <edge from-layer="227" from-port="0" to-layer="228" to-port="1" />
7623 <edge from-layer="228" from-port="2" to-layer="230" to-port="0" />
7624 <edge from-layer="229" from-port="0" to-layer="230" to-port="1" />
7625 <edge from-layer="230" from-port="2" to-layer="235" to-port="0" />
7626 <edge from-layer="231" from-port="0" to-layer="235" to-port="1" />
7627 <edge from-layer="232" from-port="0" to-layer="235" to-port="2" />
7628 <edge from-layer="233" from-port="0" to-layer="235" to-port="3" />
7629 <edge from-layer="234" from-port="0" to-layer="235" to-port="4" />
7630 <edge from-layer="235" from-port="5" to-layer="240" to-port="0" />
7631 <edge from-layer="236" from-port="0" to-layer="237" to-port="0" />
7632 <edge from-layer="237" from-port="1" to-layer="239" to-port="0" />
7633 <edge from-layer="238" from-port="0" to-layer="239" to-port="1" />
7634 <edge from-layer="239" from-port="2" to-layer="240" to-port="1" />
7635 <edge from-layer="240" from-port="2" to-layer="242" to-port="0" />
7636 <edge from-layer="241" from-port="0" to-layer="242" to-port="1" />
7637 <edge from-layer="242" from-port="2" to-layer="243" to-port="0" />
7638 <edge from-layer="243" from-port="2" to-layer="245" to-port="0" />
7639 <edge from-layer="244" from-port="0" to-layer="245" to-port="1" />
7640 <edge from-layer="245" from-port="2" to-layer="247" to-port="0" />
7641 <edge from-layer="246" from-port="0" to-layer="247" to-port="1" />
7642 <edge from-layer="247" from-port="2" to-layer="249" to-port="0" />
7643 <edge from-layer="248" from-port="0" to-layer="249" to-port="1" />
7644 <edge from-layer="249" from-port="2" to-layer="251" to-port="0" />
7645 <edge from-layer="249" from-port="2" to-layer="279" to-port="1" />
7646 <edge from-layer="250" from-port="0" to-layer="251" to-port="1" />
7647 <edge from-layer="251" from-port="2" to-layer="256" to-port="0" />
7648 <edge from-layer="252" from-port="0" to-layer="256" to-port="1" />
7649 <edge from-layer="253" from-port="0" to-layer="256" to-port="2" />
7650 <edge from-layer="254" from-port="0" to-layer="256" to-port="3" />
7651 <edge from-layer="255" from-port="0" to-layer="256" to-port="4" />
7652 <edge from-layer="256" from-port="5" to-layer="261" to-port="0" />
7653 <edge from-layer="257" from-port="0" to-layer="258" to-port="0" />
7654 <edge from-layer="258" from-port="1" to-layer="260" to-port="0" />
7655 <edge from-layer="259" from-port="0" to-layer="260" to-port="1" />
7656 <edge from-layer="260" from-port="2" to-layer="261" to-port="1" />
7657 <edge from-layer="261" from-port="2" to-layer="263" to-port="0" />
7658 <edge from-layer="262" from-port="0" to-layer="263" to-port="1" />
7659 <edge from-layer="263" from-port="2" to-layer="264" to-port="0" />
7660 <edge from-layer="264" from-port="1" to-layer="266" to-port="0" />
7661 <edge from-layer="265" from-port="0" to-layer="266" to-port="1" />
7662 <edge from-layer="266" from-port="2" to-layer="271" to-port="0" />
7663 <edge from-layer="267" from-port="0" to-layer="271" to-port="1" />
7664 <edge from-layer="268" from-port="0" to-layer="271" to-port="2" />
7665 <edge from-layer="269" from-port="0" to-layer="271" to-port="3" />
7666 <edge from-layer="270" from-port="0" to-layer="271" to-port="4" />
7667 <edge from-layer="271" from-port="5" to-layer="276" to-port="0" />
7668 <edge from-layer="272" from-port="0" to-layer="273" to-port="0" />
7669 <edge from-layer="273" from-port="1" to-layer="275" to-port="0" />
7670 <edge from-layer="274" from-port="0" to-layer="275" to-port="1" />
7671 <edge from-layer="275" from-port="2" to-layer="276" to-port="1" />
7672 <edge from-layer="276" from-port="2" to-layer="278" to-port="0" />
7673 <edge from-layer="277" from-port="0" to-layer="278" to-port="1" />
7674 <edge from-layer="278" from-port="2" to-layer="279" to-port="0" />
7675 <edge from-layer="279" from-port="2" to-layer="281" to-port="0" />
7676 <edge from-layer="280" from-port="0" to-layer="281" to-port="1" />
7677 <edge from-layer="281" from-port="2" to-layer="283" to-port="0" />
7678 <edge from-layer="282" from-port="0" to-layer="283" to-port="1" />
7679 <edge from-layer="283" from-port="2" to-layer="285" to-port="0" />
7680 <edge from-layer="284" from-port="0" to-layer="285" to-port="1" />
7681 <edge from-layer="285" from-port="2" to-layer="287" to-port="0" />
7682 <edge from-layer="285" from-port="2" to-layer="355" to-port="1" />
7683 <edge from-layer="286" from-port="0" to-layer="287" to-port="1" />
7684 <edge from-layer="287" from-port="2" to-layer="292" to-port="0" />
7685 <edge from-layer="288" from-port="0" to-layer="292" to-port="1" />
7686 <edge from-layer="289" from-port="0" to-layer="292" to-port="2" />
7687 <edge from-layer="290" from-port="0" to-layer="292" to-port="3" />
7688 <edge from-layer="291" from-port="0" to-layer="292" to-port="4" />
7689 <edge from-layer="292" from-port="5" to-layer="329" to-port="0" />
7690 <edge from-layer="292" from-port="5" to-layer="313" to-port="0" />
7691 <edge from-layer="292" from-port="5" to-layer="297" to-port="0" />
7692 <edge from-layer="293" from-port="0" to-layer="294" to-port="0" />
7693 <edge from-layer="294" from-port="1" to-layer="296" to-port="0" />
7694 <edge from-layer="295" from-port="0" to-layer="296" to-port="1" />
7695 <edge from-layer="296" from-port="2" to-layer="297" to-port="1" />
7696 <edge from-layer="297" from-port="2" to-layer="299" to-port="0" />
7697 <edge from-layer="298" from-port="0" to-layer="299" to-port="1" />
7698 <edge from-layer="299" from-port="2" to-layer="304" to-port="0" />
7699 <edge from-layer="300" from-port="0" to-layer="304" to-port="1" />
7700 <edge from-layer="301" from-port="0" to-layer="304" to-port="2" />
7701 <edge from-layer="302" from-port="0" to-layer="304" to-port="3" />
7702 <edge from-layer="303" from-port="0" to-layer="304" to-port="4" />
7703 <edge from-layer="304" from-port="5" to-layer="306" to-port="0" />
7704 <edge from-layer="305" from-port="0" to-layer="306" to-port="1" />
7705 <edge from-layer="306" from-port="2" to-layer="308" to-port="0" />
7706 <edge from-layer="307" from-port="0" to-layer="308" to-port="1" />
7707 <edge from-layer="308" from-port="2" to-layer="336" to-port="0" />
7708 <edge from-layer="309" from-port="0" to-layer="310" to-port="0" />
7709 <edge from-layer="310" from-port="1" to-layer="312" to-port="0" />
7710 <edge from-layer="311" from-port="0" to-layer="312" to-port="1" />
7711 <edge from-layer="312" from-port="2" to-layer="313" to-port="1" />
7712 <edge from-layer="313" from-port="2" to-layer="315" to-port="0" />
7713 <edge from-layer="314" from-port="0" to-layer="315" to-port="1" />
7714 <edge from-layer="315" from-port="2" to-layer="320" to-port="0" />
7715 <edge from-layer="316" from-port="0" to-layer="320" to-port="1" />
7716 <edge from-layer="317" from-port="0" to-layer="320" to-port="2" />
7717 <edge from-layer="318" from-port="0" to-layer="320" to-port="3" />
7718 <edge from-layer="319" from-port="0" to-layer="320" to-port="4" />
7719 <edge from-layer="320" from-port="5" to-layer="322" to-port="0" />
7720 <edge from-layer="321" from-port="0" to-layer="322" to-port="1" />
7721 <edge from-layer="322" from-port="2" to-layer="324" to-port="0" />
7722 <edge from-layer="323" from-port="0" to-layer="324" to-port="1" />
7723 <edge from-layer="324" from-port="2" to-layer="336" to-port="1" />
7724 <edge from-layer="325" from-port="0" to-layer="326" to-port="0" />
7725 <edge from-layer="326" from-port="1" to-layer="328" to-port="0" />
7726 <edge from-layer="327" from-port="0" to-layer="328" to-port="1" />
7727 <edge from-layer="328" from-port="2" to-layer="329" to-port="1" />
7728 <edge from-layer="329" from-port="2" to-layer="331" to-port="0" />
7729 <edge from-layer="330" from-port="0" to-layer="331" to-port="1" />
7730 <edge from-layer="331" from-port="2" to-layer="333" to-port="0" />
7731 <edge from-layer="332" from-port="0" to-layer="333" to-port="1" />
7732 <edge from-layer="333" from-port="2" to-layer="335" to-port="0" />
7733 <edge from-layer="334" from-port="0" to-layer="335" to-port="1" />
7734 <edge from-layer="335" from-port="2" to-layer="336" to-port="2" />
7735 <edge from-layer="336" from-port="4" to-layer="338" to-port="0" />
7736 <edge from-layer="337" from-port="0" to-layer="338" to-port="1" />
7737 <edge from-layer="338" from-port="2" to-layer="340" to-port="0" />
7738 <edge from-layer="339" from-port="0" to-layer="340" to-port="1" />
7739 <edge from-layer="340" from-port="2" to-layer="342" to-port="0" />
7740 <edge from-layer="341" from-port="0" to-layer="342" to-port="1" />
7741 <edge from-layer="342" from-port="2" to-layer="347" to-port="0" />
7742 <edge from-layer="343" from-port="0" to-layer="347" to-port="1" />
7743 <edge from-layer="344" from-port="0" to-layer="347" to-port="2" />
7744 <edge from-layer="345" from-port="0" to-layer="347" to-port="3" />
7745 <edge from-layer="346" from-port="0" to-layer="347" to-port="4" />
7746 <edge from-layer="347" from-port="5" to-layer="352" to-port="0" />
7747 <edge from-layer="348" from-port="0" to-layer="349" to-port="0" />
7748 <edge from-layer="349" from-port="1" to-layer="351" to-port="0" />
7749 <edge from-layer="350" from-port="0" to-layer="351" to-port="1" />
7750 <edge from-layer="351" from-port="2" to-layer="352" to-port="1" />
7751 <edge from-layer="352" from-port="2" to-layer="354" to-port="0" />
7752 <edge from-layer="353" from-port="0" to-layer="354" to-port="1" />
7753 <edge from-layer="354" from-port="2" to-layer="355" to-port="0" />
7754 <edge from-layer="355" from-port="2" to-layer="357" to-port="0" />
7755 <edge from-layer="356" from-port="0" to-layer="357" to-port="1" />
7756 <edge from-layer="357" from-port="2" to-layer="359" to-port="0" />
7757 <edge from-layer="358" from-port="0" to-layer="359" to-port="1" />
7758 <edge from-layer="359" from-port="2" to-layer="361" to-port="0" />
7759 <edge from-layer="360" from-port="0" to-layer="361" to-port="1" />
7760 <edge from-layer="361" from-port="2" to-layer="363" to-port="0" />
7761 <edge from-layer="361" from-port="2" to-layer="391" to-port="1" />
7762 <edge from-layer="362" from-port="0" to-layer="363" to-port="1" />
7763 <edge from-layer="363" from-port="2" to-layer="368" to-port="0" />
7764 <edge from-layer="364" from-port="0" to-layer="368" to-port="1" />
7765 <edge from-layer="365" from-port="0" to-layer="368" to-port="2" />
7766 <edge from-layer="366" from-port="0" to-layer="368" to-port="3" />
7767 <edge from-layer="367" from-port="0" to-layer="368" to-port="4" />
7768 <edge from-layer="368" from-port="5" to-layer="373" to-port="0" />
7769 <edge from-layer="369" from-port="0" to-layer="370" to-port="0" />
7770 <edge from-layer="370" from-port="1" to-layer="372" to-port="0" />
7771 <edge from-layer="371" from-port="0" to-layer="372" to-port="1" />
7772 <edge from-layer="372" from-port="2" to-layer="373" to-port="1" />
7773 <edge from-layer="373" from-port="2" to-layer="375" to-port="0" />
7774 <edge from-layer="374" from-port="0" to-layer="375" to-port="1" />
7775 <edge from-layer="375" from-port="2" to-layer="376" to-port="0" />
7776 <edge from-layer="376" from-port="1" to-layer="378" to-port="0" />
7777 <edge from-layer="377" from-port="0" to-layer="378" to-port="1" />
7778 <edge from-layer="378" from-port="2" to-layer="383" to-port="0" />
7779 <edge from-layer="379" from-port="0" to-layer="383" to-port="1" />
7780 <edge from-layer="380" from-port="0" to-layer="383" to-port="2" />
7781 <edge from-layer="381" from-port="0" to-layer="383" to-port="3" />
7782 <edge from-layer="382" from-port="0" to-layer="383" to-port="4" />
7783 <edge from-layer="383" from-port="5" to-layer="388" to-port="0" />
7784 <edge from-layer="384" from-port="0" to-layer="385" to-port="0" />
7785 <edge from-layer="385" from-port="1" to-layer="387" to-port="0" />
7786 <edge from-layer="386" from-port="0" to-layer="387" to-port="1" />
7787 <edge from-layer="387" from-port="2" to-layer="388" to-port="1" />
7788 <edge from-layer="388" from-port="2" to-layer="390" to-port="0" />
7789 <edge from-layer="389" from-port="0" to-layer="390" to-port="1" />
7790 <edge from-layer="390" from-port="2" to-layer="391" to-port="0" />
7791 <edge from-layer="391" from-port="2" to-layer="393" to-port="0" />
7792 <edge from-layer="392" from-port="0" to-layer="393" to-port="1" />
7793 <edge from-layer="393" from-port="2" to-layer="395" to-port="0" />
7794 <edge from-layer="394" from-port="0" to-layer="395" to-port="1" />
7795 <edge from-layer="395" from-port="2" to-layer="397" to-port="0" />
7796 <edge from-layer="396" from-port="0" to-layer="397" to-port="1" />
7797 <edge from-layer="397" from-port="2" to-layer="399" to-port="0" />
7798 <edge from-layer="397" from-port="2" to-layer="467" to-port="1" />
7799 <edge from-layer="398" from-port="0" to-layer="399" to-port="1" />
7800 <edge from-layer="399" from-port="2" to-layer="404" to-port="0" />
7801 <edge from-layer="400" from-port="0" to-layer="404" to-port="1" />
7802 <edge from-layer="401" from-port="0" to-layer="404" to-port="2" />
7803 <edge from-layer="402" from-port="0" to-layer="404" to-port="3" />
7804 <edge from-layer="403" from-port="0" to-layer="404" to-port="4" />
7805 <edge from-layer="404" from-port="5" to-layer="409" to-port="0" />
7806 <edge from-layer="404" from-port="5" to-layer="425" to-port="0" />
7807 <edge from-layer="404" from-port="5" to-layer="441" to-port="0" />
7808 <edge from-layer="405" from-port="0" to-layer="406" to-port="0" />
7809 <edge from-layer="406" from-port="1" to-layer="408" to-port="0" />
7810 <edge from-layer="407" from-port="0" to-layer="408" to-port="1" />
7811 <edge from-layer="408" from-port="2" to-layer="409" to-port="1" />
7812 <edge from-layer="409" from-port="2" to-layer="411" to-port="0" />
7813 <edge from-layer="410" from-port="0" to-layer="411" to-port="1" />
7814 <edge from-layer="411" from-port="2" to-layer="416" to-port="0" />
7815 <edge from-layer="412" from-port="0" to-layer="416" to-port="1" />
7816 <edge from-layer="413" from-port="0" to-layer="416" to-port="2" />
7817 <edge from-layer="414" from-port="0" to-layer="416" to-port="3" />
7818 <edge from-layer="415" from-port="0" to-layer="416" to-port="4" />
7819 <edge from-layer="416" from-port="5" to-layer="418" to-port="0" />
7820 <edge from-layer="417" from-port="0" to-layer="418" to-port="1" />
7821 <edge from-layer="418" from-port="2" to-layer="420" to-port="0" />
7822 <edge from-layer="419" from-port="0" to-layer="420" to-port="1" />
7823 <edge from-layer="420" from-port="2" to-layer="448" to-port="0" />
7824 <edge from-layer="421" from-port="0" to-layer="422" to-port="0" />
7825 <edge from-layer="422" from-port="1" to-layer="424" to-port="0" />
7826 <edge from-layer="423" from-port="0" to-layer="424" to-port="1" />
7827 <edge from-layer="424" from-port="2" to-layer="425" to-port="1" />
7828 <edge from-layer="425" from-port="2" to-layer="427" to-port="0" />
7829 <edge from-layer="426" from-port="0" to-layer="427" to-port="1" />
7830 <edge from-layer="427" from-port="2" to-layer="432" to-port="0" />
7831 <edge from-layer="428" from-port="0" to-layer="432" to-port="1" />
7832 <edge from-layer="429" from-port="0" to-layer="432" to-port="2" />
7833 <edge from-layer="430" from-port="0" to-layer="432" to-port="3" />
7834 <edge from-layer="431" from-port="0" to-layer="432" to-port="4" />
7835 <edge from-layer="432" from-port="5" to-layer="434" to-port="0" />
7836 <edge from-layer="433" from-port="0" to-layer="434" to-port="1" />
7837 <edge from-layer="434" from-port="2" to-layer="436" to-port="0" />
7838 <edge from-layer="435" from-port="0" to-layer="436" to-port="1" />
7839 <edge from-layer="436" from-port="2" to-layer="448" to-port="1" />
7840 <edge from-layer="437" from-port="0" to-layer="438" to-port="0" />
7841 <edge from-layer="438" from-port="1" to-layer="440" to-port="0" />
7842 <edge from-layer="439" from-port="0" to-layer="440" to-port="1" />
7843 <edge from-layer="440" from-port="2" to-layer="441" to-port="1" />
7844 <edge from-layer="441" from-port="2" to-layer="443" to-port="0" />
7845 <edge from-layer="442" from-port="0" to-layer="443" to-port="1" />
7846 <edge from-layer="443" from-port="2" to-layer="445" to-port="0" />
7847 <edge from-layer="444" from-port="0" to-layer="445" to-port="1" />
7848 <edge from-layer="445" from-port="2" to-layer="447" to-port="0" />
7849 <edge from-layer="446" from-port="0" to-layer="447" to-port="1" />
7850 <edge from-layer="447" from-port="2" to-layer="448" to-port="2" />
7851 <edge from-layer="448" from-port="4" to-layer="450" to-port="0" />
7852 <edge from-layer="449" from-port="0" to-layer="450" to-port="1" />
7853 <edge from-layer="450" from-port="2" to-layer="452" to-port="0" />
7854 <edge from-layer="451" from-port="0" to-layer="452" to-port="1" />
7855 <edge from-layer="452" from-port="2" to-layer="454" to-port="0" />
7856 <edge from-layer="453" from-port="0" to-layer="454" to-port="1" />
7857 <edge from-layer="454" from-port="2" to-layer="459" to-port="0" />
7858 <edge from-layer="455" from-port="0" to-layer="459" to-port="1" />
7859 <edge from-layer="456" from-port="0" to-layer="459" to-port="2" />
7860 <edge from-layer="457" from-port="0" to-layer="459" to-port="3" />
7861 <edge from-layer="458" from-port="0" to-layer="459" to-port="4" />
7862 <edge from-layer="459" from-port="5" to-layer="464" to-port="0" />
7863 <edge from-layer="460" from-port="0" to-layer="461" to-port="0" />
7864 <edge from-layer="461" from-port="1" to-layer="463" to-port="0" />
7865 <edge from-layer="462" from-port="0" to-layer="463" to-port="1" />
7866 <edge from-layer="463" from-port="2" to-layer="464" to-port="1" />
7867 <edge from-layer="464" from-port="2" to-layer="466" to-port="0" />
7868 <edge from-layer="465" from-port="0" to-layer="466" to-port="1" />
7869 <edge from-layer="466" from-port="2" to-layer="467" to-port="0" />
7870 <edge from-layer="467" from-port="2" to-layer="469" to-port="0" />
7871 <edge from-layer="468" from-port="0" to-layer="469" to-port="1" />
7872 <edge from-layer="469" from-port="2" to-layer="471" to-port="0" />
7873 <edge from-layer="470" from-port="0" to-layer="471" to-port="1" />
7874 <edge from-layer="471" from-port="2" to-layer="473" to-port="0" />
7875 <edge from-layer="472" from-port="0" to-layer="473" to-port="1" />
7876 <edge from-layer="473" from-port="2" to-layer="475" to-port="0" />
7877 <edge from-layer="473" from-port="2" to-layer="503" to-port="1" />
7878 <edge from-layer="474" from-port="0" to-layer="475" to-port="1" />
7879 <edge from-layer="475" from-port="2" to-layer="480" to-port="0" />
7880 <edge from-layer="476" from-port="0" to-layer="480" to-port="1" />
7881 <edge from-layer="477" from-port="0" to-layer="480" to-port="2" />
7882 <edge from-layer="478" from-port="0" to-layer="480" to-port="3" />
7883 <edge from-layer="479" from-port="0" to-layer="480" to-port="4" />
7884 <edge from-layer="480" from-port="5" to-layer="485" to-port="0" />
7885 <edge from-layer="481" from-port="0" to-layer="482" to-port="0" />
7886 <edge from-layer="482" from-port="1" to-layer="484" to-port="0" />
7887 <edge from-layer="483" from-port="0" to-layer="484" to-port="1" />
7888 <edge from-layer="484" from-port="2" to-layer="485" to-port="1" />
7889 <edge from-layer="485" from-port="2" to-layer="487" to-port="0" />
7890 <edge from-layer="486" from-port="0" to-layer="487" to-port="1" />
7891 <edge from-layer="487" from-port="2" to-layer="488" to-port="0" />
7892 <edge from-layer="488" from-port="1" to-layer="490" to-port="0" />
7893 <edge from-layer="489" from-port="0" to-layer="490" to-port="1" />
7894 <edge from-layer="490" from-port="2" to-layer="495" to-port="0" />
7895 <edge from-layer="491" from-port="0" to-layer="495" to-port="1" />
7896 <edge from-layer="492" from-port="0" to-layer="495" to-port="2" />
7897 <edge from-layer="493" from-port="0" to-layer="495" to-port="3" />
7898 <edge from-layer="494" from-port="0" to-layer="495" to-port="4" />
7899 <edge from-layer="495" from-port="5" to-layer="500" to-port="0" />
7900 <edge from-layer="496" from-port="0" to-layer="497" to-port="0" />
7901 <edge from-layer="497" from-port="1" to-layer="499" to-port="0" />
7902 <edge from-layer="498" from-port="0" to-layer="499" to-port="1" />
7903 <edge from-layer="499" from-port="2" to-layer="500" to-port="1" />
7904 <edge from-layer="500" from-port="2" to-layer="502" to-port="0" />
7905 <edge from-layer="501" from-port="0" to-layer="502" to-port="1" />
7906 <edge from-layer="502" from-port="2" to-layer="503" to-port="0" />
7907 <edge from-layer="503" from-port="2" to-layer="505" to-port="0" />
7908 <edge from-layer="504" from-port="0" to-layer="505" to-port="1" />
7909 <edge from-layer="505" from-port="2" to-layer="507" to-port="0" />
7910 <edge from-layer="506" from-port="0" to-layer="507" to-port="1" />
7911 <edge from-layer="507" from-port="2" to-layer="509" to-port="0" />
7912 <edge from-layer="508" from-port="0" to-layer="509" to-port="1" />
7913 <edge from-layer="509" from-port="2" to-layer="511" to-port="0" />
7914 <edge from-layer="510" from-port="0" to-layer="511" to-port="1" />
7915 <edge from-layer="511" from-port="3" to-layer="513" to-port="0" />
7916 <edge from-layer="512" from-port="0" to-layer="513" to-port="1" />
7917 <edge from-layer="513" from-port="2" to-layer="518" to-port="0" />
7918 <edge from-layer="514" from-port="0" to-layer="518" to-port="1" />
7919 <edge from-layer="515" from-port="0" to-layer="518" to-port="2" />
7920 <edge from-layer="516" from-port="0" to-layer="518" to-port="3" />
7921 <edge from-layer="517" from-port="0" to-layer="518" to-port="4" />
7922 <edge from-layer="518" from-port="5" to-layer="523" to-port="0" />
7923 <edge from-layer="519" from-port="0" to-layer="520" to-port="0" />
7924 <edge from-layer="520" from-port="1" to-layer="522" to-port="0" />
7925 <edge from-layer="521" from-port="0" to-layer="522" to-port="1" />
7926 <edge from-layer="522" from-port="2" to-layer="523" to-port="1" />
7927 <edge from-layer="523" from-port="2" to-layer="525" to-port="0" />
7928 <edge from-layer="524" from-port="0" to-layer="525" to-port="1" />
7929 <edge from-layer="525" from-port="2" to-layer="526" to-port="0" />
7930 <edge from-layer="526" from-port="1" to-layer="528" to-port="0" />
7931 <edge from-layer="527" from-port="0" to-layer="528" to-port="1" />
7932 <edge from-layer="528" from-port="2" to-layer="533" to-port="0" />
7933 <edge from-layer="529" from-port="0" to-layer="533" to-port="1" />
7934 <edge from-layer="530" from-port="0" to-layer="533" to-port="2" />
7935 <edge from-layer="531" from-port="0" to-layer="533" to-port="3" />
7936 <edge from-layer="532" from-port="0" to-layer="533" to-port="4" />
7937 <edge from-layer="533" from-port="5" to-layer="538" to-port="0" />
7938 <edge from-layer="534" from-port="0" to-layer="535" to-port="0" />
7939 <edge from-layer="535" from-port="1" to-layer="537" to-port="0" />
7940 <edge from-layer="536" from-port="0" to-layer="537" to-port="1" />
7941 <edge from-layer="537" from-port="2" to-layer="538" to-port="1" />
7942 <edge from-layer="538" from-port="2" to-layer="539" to-port="0" />
7943 </edges>
7944 <rt_info>
7945 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
7946 <conversion_parameters>
7947 <framework value="pytorch" />
7948 <is_python_object value="True" />
7949 </conversion_parameters>
7950 <nncf>
7951 <friendly_names_were_updated value="True" />
7952 <quantization>
7953 <advanced_parameters value="{'overflow_fix': 'disable', 'quantize_outputs': False, 'inplace_statistics': True, 'disable_channel_alignment': True, 'disable_bias_correction': False, 'batchwise_statistics': None, 'activations_quantization_params': None, 'weights_quantization_params': None, 'activations_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'weights_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'bias_correction_params': {'apply_for_all_nodes': False, 'threshold': None}, 'smooth_quant_alphas': {'convolution': -1, 'matmul': 0.95}, 'smooth_quant_alpha': None, 'backend_params': {}}" />
7954 <fast_bias_correction value="True" />
7955 <ignored_scope>
7956 <types value="['GroupNormalization']" />
7957 </ignored_scope>
7958 <model_type value="transformer" />
7959 <preset value="mixed" />
7960 <subset_size value="300" />
7961 <target_device value="ANY" />
7962 </quantization>
7963 </nncf>
7964 <optimum>
7965 <optimum_intel_version value="1.20.1" />
7966 <optimum_version value="1.24.0" />
7967 <pytorch_version value="2.6.0+cu124" />
7968 <transformers_version value="4.52.0.dev0" />
7969 </optimum>
7970 </rt_info>
7971 </net>
7972