openvino/openvino_model_qint8_quantized.xml
369.5 KB · 11468 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model0" version="11">
3 <layers>
4 <layer id="1" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="0" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="2" name="Constant_220482" type="Const" version="opset1">
23 <data element_type="i8" shape="50265, 768" offset="0" size="38603520" />
24 <output>
25 <port id="0" precision="I8">
26 <dim>50265</dim>
27 <dim>768</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="Convert_220483" type="Convert" version="opset1">
32 <data destination_type="f32" />
33 <input>
34 <port id="0" precision="I8">
35 <dim>50265</dim>
36 <dim>768</dim>
37 </port>
38 </input>
39 <output>
40 <port id="1" precision="FP32">
41 <dim>50265</dim>
42 <dim>768</dim>
43 </port>
44 </output>
45 </layer>
46 <layer id="4" name="Constant_220484" type="Const" version="opset1">
47 <data element_type="f32" shape="50265, 1" offset="38603520" size="201060" />
48 <output>
49 <port id="0" precision="FP32">
50 <dim>50265</dim>
51 <dim>1</dim>
52 </port>
53 </output>
54 </layer>
55 <layer id="5" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
56 <data auto_broadcast="numpy" />
57 <input>
58 <port id="0" precision="FP32">
59 <dim>50265</dim>
60 <dim>768</dim>
61 </port>
62 <port id="1" precision="FP32">
63 <dim>50265</dim>
64 <dim>1</dim>
65 </port>
66 </input>
67 <output>
68 <port id="2" precision="FP32">
69 <dim>50265</dim>
70 <dim>768</dim>
71 </port>
72 </output>
73 </layer>
74 <layer id="6" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
75 <data destination_type="i32" />
76 <input>
77 <port id="0" precision="I64">
78 <dim>-1</dim>
79 <dim>-1</dim>
80 </port>
81 </input>
82 <output>
83 <port id="1" precision="I32">
84 <dim>-1</dim>
85 <dim>-1</dim>
86 </port>
87 </output>
88 </layer>
89 <layer id="7" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
90 <data element_type="i32" shape="" offset="38804580" size="4" />
91 <output>
92 <port id="0" precision="I32" />
93 </output>
94 </layer>
95 <layer id="8" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
96 <data batch_dims="0" />
97 <input>
98 <port id="0" precision="FP32">
99 <dim>50265</dim>
100 <dim>768</dim>
101 </port>
102 <port id="1" precision="I32">
103 <dim>-1</dim>
104 <dim>-1</dim>
105 </port>
106 <port id="2" precision="I32" />
107 </input>
108 <output>
109 <port id="3" precision="FP32" names="54,inputs_embeds">
110 <dim>-1</dim>
111 <dim>-1</dim>
112 <dim>768</dim>
113 </port>
114 </output>
115 </layer>
116 <layer id="9" name="Constant_220486" type="Const" version="opset1">
117 <data element_type="i8" shape="1, 768" offset="38804584" size="768" />
118 <output>
119 <port id="0" precision="I8">
120 <dim>1</dim>
121 <dim>768</dim>
122 </port>
123 </output>
124 </layer>
125 <layer id="10" name="Convert_220487" type="Convert" version="opset1">
126 <data destination_type="f32" />
127 <input>
128 <port id="0" precision="I8">
129 <dim>1</dim>
130 <dim>768</dim>
131 </port>
132 </input>
133 <output>
134 <port id="1" precision="FP32">
135 <dim>1</dim>
136 <dim>768</dim>
137 </port>
138 </output>
139 </layer>
140 <layer id="11" name="Constant_220488" type="Const" version="opset1">
141 <data element_type="f32" shape="1, 1" offset="38805352" size="4" />
142 <output>
143 <port id="0" precision="FP32">
144 <dim>1</dim>
145 <dim>1</dim>
146 </port>
147 </output>
148 </layer>
149 <layer id="12" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
150 <data auto_broadcast="numpy" />
151 <input>
152 <port id="0" precision="FP32">
153 <dim>1</dim>
154 <dim>768</dim>
155 </port>
156 <port id="1" precision="FP32">
157 <dim>1</dim>
158 <dim>1</dim>
159 </port>
160 </input>
161 <output>
162 <port id="2" precision="FP32">
163 <dim>1</dim>
164 <dim>768</dim>
165 </port>
166 </output>
167 </layer>
168 <layer id="13" name="__module.roberta/aten::slice/Slice" type="Const" version="opset1">
169 <data element_type="i64" shape="1, 514" offset="38805356" size="4112" />
170 <output>
171 <port id="0" precision="I64" names="37">
172 <dim>1</dim>
173 <dim>514</dim>
174 </port>
175 </output>
176 </layer>
177 <layer id="14" name="__module.roberta/aten::slice/Reshape" type="Const" version="opset1">
178 <data element_type="i64" shape="1" offset="38809468" size="8" />
179 <output>
180 <port id="0" precision="I64">
181 <dim>1</dim>
182 </port>
183 </output>
184 </layer>
185 <layer id="15" name="__module.roberta/aten::size/ShapeOf_1" type="ShapeOf" version="opset3">
186 <data output_type="i64" />
187 <input>
188 <port id="0" precision="I64">
189 <dim>-1</dim>
190 <dim>-1</dim>
191 </port>
192 </input>
193 <output>
194 <port id="1" precision="I64" names="39">
195 <dim>2</dim>
196 </port>
197 </output>
198 </layer>
199 <layer id="16" name="Constant_5485" type="Const" version="opset1">
200 <data element_type="i64" shape="1" offset="38809476" size="8" />
201 <output>
202 <port id="0" precision="I64">
203 <dim>1</dim>
204 </port>
205 </output>
206 </layer>
207 <layer id="17" name="Constant_5486" type="Const" version="opset1">
208 <data element_type="i64" shape="" offset="38809468" size="8" />
209 <output>
210 <port id="0" precision="I64" />
211 </output>
212 </layer>
213 <layer id="18" name="Gather_5487" type="Gather" version="opset8">
214 <data batch_dims="0" />
215 <input>
216 <port id="0" precision="I64">
217 <dim>2</dim>
218 </port>
219 <port id="1" precision="I64">
220 <dim>1</dim>
221 </port>
222 <port id="2" precision="I64" />
223 </input>
224 <output>
225 <port id="3" precision="I64" names="36,67">
226 <dim>1</dim>
227 </port>
228 </output>
229 </layer>
230 <layer id="19" name="__module.roberta/aten::slice/Reshape_2" type="Const" version="opset1">
231 <data element_type="i64" shape="1" offset="38809476" size="8" />
232 <output>
233 <port id="0" precision="I64">
234 <dim>1</dim>
235 </port>
236 </output>
237 </layer>
238 <layer id="20" name="__module.roberta/aten::slice/Reshape_3" type="Const" version="opset1">
239 <data element_type="i64" shape="1" offset="38809476" size="8" />
240 <output>
241 <port id="0" precision="I64">
242 <dim>1</dim>
243 </port>
244 </output>
245 </layer>
246 <layer id="21" name="__module.roberta/aten::slice/Slice_1" type="Slice" version="opset8">
247 <input>
248 <port id="0" precision="I64">
249 <dim>1</dim>
250 <dim>514</dim>
251 </port>
252 <port id="1" precision="I64">
253 <dim>1</dim>
254 </port>
255 <port id="2" precision="I64">
256 <dim>1</dim>
257 </port>
258 <port id="3" precision="I64">
259 <dim>1</dim>
260 </port>
261 <port id="4" precision="I64">
262 <dim>1</dim>
263 </port>
264 </input>
265 <output>
266 <port id="5" precision="I64" names="38,buffered_token_type_ids">
267 <dim>1</dim>
268 <dim>-1</dim>
269 </port>
270 </output>
271 </layer>
272 <layer id="22" name="__module.roberta/aten::expand/Broadcast" type="Broadcast" version="opset3">
273 <data mode="bidirectional" />
274 <input>
275 <port id="0" precision="I64">
276 <dim>1</dim>
277 <dim>-1</dim>
278 </port>
279 <port id="1" precision="I64">
280 <dim>2</dim>
281 </port>
282 </input>
283 <output>
284 <port id="2" precision="I64" names="40">
285 <dim>-1</dim>
286 <dim>-1</dim>
287 </port>
288 </output>
289 </layer>
290 <layer id="23" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
291 <data destination_type="i32" />
292 <input>
293 <port id="0" precision="I64">
294 <dim>-1</dim>
295 <dim>-1</dim>
296 </port>
297 </input>
298 <output>
299 <port id="1" precision="I32">
300 <dim>-1</dim>
301 <dim>-1</dim>
302 </port>
303 </output>
304 </layer>
305 <layer id="24" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
306 <data element_type="i32" shape="" offset="38804580" size="4" />
307 <output>
308 <port id="0" precision="I32" />
309 </output>
310 </layer>
311 <layer id="25" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
312 <data batch_dims="0" />
313 <input>
314 <port id="0" precision="FP32">
315 <dim>1</dim>
316 <dim>768</dim>
317 </port>
318 <port id="1" precision="I32">
319 <dim>-1</dim>
320 <dim>-1</dim>
321 </port>
322 <port id="2" precision="I32" />
323 </input>
324 <output>
325 <port id="3" precision="FP32" names="56,token_type_embeddings.1">
326 <dim>-1</dim>
327 <dim>-1</dim>
328 <dim>768</dim>
329 </port>
330 </output>
331 </layer>
332 <layer id="26" name="__module.roberta.embeddings/aten::add/Add" type="Add" version="opset1">
333 <data auto_broadcast="numpy" />
334 <input>
335 <port id="0" precision="FP32">
336 <dim>-1</dim>
337 <dim>-1</dim>
338 <dim>768</dim>
339 </port>
340 <port id="1" precision="FP32">
341 <dim>-1</dim>
342 <dim>-1</dim>
343 <dim>768</dim>
344 </port>
345 </input>
346 <output>
347 <port id="2" precision="FP32" names="57_1">
348 <dim>-1</dim>
349 <dim>-1</dim>
350 <dim>768</dim>
351 </port>
352 </output>
353 </layer>
354 <layer id="27" name="Constant_220490" type="Const" version="opset1">
355 <data element_type="i8" shape="514, 768" offset="38809484" size="394752" />
356 <output>
357 <port id="0" precision="I8">
358 <dim>514</dim>
359 <dim>768</dim>
360 </port>
361 </output>
362 </layer>
363 <layer id="28" name="Convert_220491" type="Convert" version="opset1">
364 <data destination_type="f32" />
365 <input>
366 <port id="0" precision="I8">
367 <dim>514</dim>
368 <dim>768</dim>
369 </port>
370 </input>
371 <output>
372 <port id="1" precision="FP32">
373 <dim>514</dim>
374 <dim>768</dim>
375 </port>
376 </output>
377 </layer>
378 <layer id="29" name="Constant_220492" type="Const" version="opset1">
379 <data element_type="f32" shape="514, 1" offset="39204236" size="2056" />
380 <output>
381 <port id="0" precision="FP32">
382 <dim>514</dim>
383 <dim>1</dim>
384 </port>
385 </output>
386 </layer>
387 <layer id="30" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Gather/fq_weights_0" type="Multiply" version="opset1">
388 <data auto_broadcast="numpy" />
389 <input>
390 <port id="0" precision="FP32">
391 <dim>514</dim>
392 <dim>768</dim>
393 </port>
394 <port id="1" precision="FP32">
395 <dim>514</dim>
396 <dim>1</dim>
397 </port>
398 </input>
399 <output>
400 <port id="2" precision="FP32">
401 <dim>514</dim>
402 <dim>768</dim>
403 </port>
404 </output>
405 </layer>
406 <layer id="31" name="Constant_5356" type="Const" version="opset1">
407 <data element_type="i64" shape="1, 1" offset="38809476" size="8" />
408 <output>
409 <port id="0" precision="I64">
410 <dim>1</dim>
411 <dim>1</dim>
412 </port>
413 </output>
414 </layer>
415 <layer id="32" name="__module.roberta.embeddings/aten::ne/NotEqual" type="NotEqual" version="opset1">
416 <data auto_broadcast="numpy" />
417 <input>
418 <port id="0" precision="I64">
419 <dim>-1</dim>
420 <dim>-1</dim>
421 </port>
422 <port id="1" precision="I64">
423 <dim>1</dim>
424 <dim>1</dim>
425 </port>
426 </input>
427 <output>
428 <port id="2" precision="BOOL" names="45">
429 <dim>-1</dim>
430 <dim>-1</dim>
431 </port>
432 </output>
433 </layer>
434 <layer id="33" name="__module.roberta.embeddings/aten::to/Convert" type="Convert" version="opset1">
435 <data destination_type="i32" />
436 <input>
437 <port id="0" precision="BOOL">
438 <dim>-1</dim>
439 <dim>-1</dim>
440 </port>
441 </input>
442 <output>
443 <port id="1" precision="I32" names="46,mask">
444 <dim>-1</dim>
445 <dim>-1</dim>
446 </port>
447 </output>
448 </layer>
449 <layer id="34" name="29" type="Const" version="opset1">
450 <data element_type="i64" shape="" offset="38809476" size="8" />
451 <output>
452 <port id="0" precision="I64" names="29" />
453 </output>
454 </layer>
455 <layer id="35" name="__module.roberta.embeddings/aten::cumsum/CumSum" type="CumSum" version="opset3">
456 <data exclusive="false" reverse="false" />
457 <input>
458 <port id="0" precision="I32">
459 <dim>-1</dim>
460 <dim>-1</dim>
461 </port>
462 <port id="1" precision="I64" />
463 </input>
464 <output>
465 <port id="2" precision="I32" names="47,48,49">
466 <dim>-1</dim>
467 <dim>-1</dim>
468 </port>
469 </output>
470 </layer>
471 <layer id="36" name="__module.roberta.embeddings/aten::mul/Multiply" type="Multiply" version="opset1">
472 <data auto_broadcast="numpy" />
473 <input>
474 <port id="0" precision="I32">
475 <dim>-1</dim>
476 <dim>-1</dim>
477 </port>
478 <port id="1" precision="I32">
479 <dim>-1</dim>
480 <dim>-1</dim>
481 </port>
482 </input>
483 <output>
484 <port id="2" precision="I32" names="50,incremental_indices">
485 <dim>-1</dim>
486 <dim>-1</dim>
487 </port>
488 </output>
489 </layer>
490 <layer id="37" name="__module.roberta.embeddings/aten::to/Convert_1" type="Convert" version="opset1">
491 <data destination_type="i64" />
492 <input>
493 <port id="0" precision="I32">
494 <dim>-1</dim>
495 <dim>-1</dim>
496 </port>
497 </input>
498 <output>
499 <port id="1" precision="I64" names="51">
500 <dim>-1</dim>
501 <dim>-1</dim>
502 </port>
503 </output>
504 </layer>
505 <layer id="38" name="Constant_5357" type="Const" version="opset1">
506 <data element_type="i64" shape="1, 1" offset="38809476" size="8" />
507 <output>
508 <port id="0" precision="I64">
509 <dim>1</dim>
510 <dim>1</dim>
511 </port>
512 </output>
513 </layer>
514 <layer id="39" name="__module.roberta.embeddings/aten::add/Add_2" type="Add" version="opset1">
515 <data auto_broadcast="numpy" />
516 <input>
517 <port id="0" precision="I64">
518 <dim>-1</dim>
519 <dim>-1</dim>
520 </port>
521 <port id="1" precision="I64">
522 <dim>1</dim>
523 <dim>1</dim>
524 </port>
525 </input>
526 <output>
527 <port id="2" precision="I64" names="52">
528 <dim>-1</dim>
529 <dim>-1</dim>
530 </port>
531 </output>
532 </layer>
533 <layer id="40" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
534 <data destination_type="i32" />
535 <input>
536 <port id="0" precision="I64">
537 <dim>-1</dim>
538 <dim>-1</dim>
539 </port>
540 </input>
541 <output>
542 <port id="1" precision="I32">
543 <dim>-1</dim>
544 <dim>-1</dim>
545 </port>
546 </output>
547 </layer>
548 <layer id="41" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
549 <data element_type="i32" shape="" offset="38804580" size="4" />
550 <output>
551 <port id="0" precision="I32" />
552 </output>
553 </layer>
554 <layer id="42" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
555 <data batch_dims="0" />
556 <input>
557 <port id="0" precision="FP32">
558 <dim>514</dim>
559 <dim>768</dim>
560 </port>
561 <port id="1" precision="I32">
562 <dim>-1</dim>
563 <dim>-1</dim>
564 </port>
565 <port id="2" precision="I32" />
566 </input>
567 <output>
568 <port id="3" precision="FP32" names="59,position_embeddings.1">
569 <dim>-1</dim>
570 <dim>-1</dim>
571 <dim>768</dim>
572 </port>
573 </output>
574 </layer>
575 <layer id="43" name="__module.roberta.embeddings/aten::add_/Add" type="Add" version="opset1">
576 <data auto_broadcast="numpy" />
577 <input>
578 <port id="0" precision="FP32">
579 <dim>-1</dim>
580 <dim>-1</dim>
581 <dim>768</dim>
582 </port>
583 <port id="1" precision="FP32">
584 <dim>-1</dim>
585 <dim>-1</dim>
586 <dim>768</dim>
587 </port>
588 </input>
589 <output>
590 <port id="2" precision="FP32" names="57,embeddings.1">
591 <dim>-1</dim>
592 <dim>-1</dim>
593 <dim>768</dim>
594 </port>
595 </output>
596 </layer>
597 <layer id="44" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
598 <data element_type="i32" shape="1" offset="39206292" size="4" />
599 <output>
600 <port id="0" precision="I32">
601 <dim>1</dim>
602 </port>
603 </output>
604 </layer>
605 <layer id="45" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
606 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
607 <input>
608 <port id="0" precision="FP32">
609 <dim>-1</dim>
610 <dim>-1</dim>
611 <dim>768</dim>
612 </port>
613 <port id="1" precision="I32">
614 <dim>1</dim>
615 </port>
616 </input>
617 <output>
618 <port id="2" precision="FP32">
619 <dim>-1</dim>
620 <dim>-1</dim>
621 <dim>768</dim>
622 </port>
623 </output>
624 </layer>
625 <layer id="46" name="Constant_5358" type="Const" version="opset1">
626 <data element_type="f32" shape="1, 1, 768" offset="39206296" size="3072" />
627 <output>
628 <port id="0" precision="FP32">
629 <dim>1</dim>
630 <dim>1</dim>
631 <dim>768</dim>
632 </port>
633 </output>
634 </layer>
635 <layer id="47" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
636 <data auto_broadcast="numpy" />
637 <input>
638 <port id="0" precision="FP32">
639 <dim>-1</dim>
640 <dim>-1</dim>
641 <dim>768</dim>
642 </port>
643 <port id="1" precision="FP32">
644 <dim>1</dim>
645 <dim>1</dim>
646 <dim>768</dim>
647 </port>
648 </input>
649 <output>
650 <port id="2" precision="FP32">
651 <dim>-1</dim>
652 <dim>-1</dim>
653 <dim>768</dim>
654 </port>
655 </output>
656 </layer>
657 <layer id="48" name="Constant_5359" type="Const" version="opset1">
658 <data element_type="f32" shape="1, 1, 768" offset="39209368" size="3072" />
659 <output>
660 <port id="0" precision="FP32">
661 <dim>1</dim>
662 <dim>1</dim>
663 <dim>768</dim>
664 </port>
665 </output>
666 </layer>
667 <layer id="49" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
668 <data auto_broadcast="numpy" />
669 <input>
670 <port id="0" precision="FP32">
671 <dim>-1</dim>
672 <dim>-1</dim>
673 <dim>768</dim>
674 </port>
675 <port id="1" precision="FP32">
676 <dim>1</dim>
677 <dim>1</dim>
678 <dim>768</dim>
679 </port>
680 </input>
681 <output>
682 <port id="2" precision="FP32" names="64,input.1">
683 <dim>-1</dim>
684 <dim>-1</dim>
685 <dim>768</dim>
686 </port>
687 </output>
688 </layer>
689 <layer id="50" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
690 <data element_type="f32" shape="1, 1, 768" offset="39212440" size="3072" />
691 <output>
692 <port id="0" precision="FP32">
693 <dim>1</dim>
694 <dim>1</dim>
695 <dim>768</dim>
696 </port>
697 </output>
698 </layer>
699 <layer id="51" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
700 <data auto_broadcast="numpy" />
701 <input>
702 <port id="0" precision="FP32">
703 <dim>-1</dim>
704 <dim>-1</dim>
705 <dim>768</dim>
706 </port>
707 <port id="1" precision="FP32">
708 <dim>1</dim>
709 <dim>1</dim>
710 <dim>768</dim>
711 </port>
712 </input>
713 <output>
714 <port id="2" precision="FP32">
715 <dim>-1</dim>
716 <dim>-1</dim>
717 <dim>768</dim>
718 </port>
719 </output>
720 </layer>
721 <layer id="52" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
722 <data element_type="f32" shape="" offset="39215512" size="4" />
723 <output>
724 <port id="0" precision="FP32" />
725 </output>
726 </layer>
727 <layer id="53" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
728 <data element_type="f32" shape="" offset="39215516" size="4" />
729 <output>
730 <port id="0" precision="FP32" />
731 </output>
732 </layer>
733 <layer id="54" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
734 <data element_type="f32" shape="" offset="39215512" size="4" />
735 <output>
736 <port id="0" precision="FP32" />
737 </output>
738 </layer>
739 <layer id="55" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
740 <data element_type="f32" shape="" offset="39215516" size="4" />
741 <output>
742 <port id="0" precision="FP32" />
743 </output>
744 </layer>
745 <layer id="56" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
746 <data levels="256" auto_broadcast="numpy" />
747 <input>
748 <port id="0" precision="FP32">
749 <dim>-1</dim>
750 <dim>-1</dim>
751 <dim>768</dim>
752 </port>
753 <port id="1" precision="FP32" />
754 <port id="2" precision="FP32" />
755 <port id="3" precision="FP32" />
756 <port id="4" precision="FP32" />
757 </input>
758 <output>
759 <port id="5" precision="FP32">
760 <dim>-1</dim>
761 <dim>-1</dim>
762 <dim>768</dim>
763 </port>
764 </output>
765 </layer>
766 <layer id="57" name="Constant_220494" type="Const" version="opset1">
767 <data element_type="i8" shape="768, 768" offset="39215520" size="589824" />
768 <output>
769 <port id="0" precision="I8">
770 <dim>768</dim>
771 <dim>768</dim>
772 </port>
773 </output>
774 </layer>
775 <layer id="58" name="Convert_220495" type="Convert" version="opset1">
776 <data destination_type="f32" />
777 <input>
778 <port id="0" precision="I8">
779 <dim>768</dim>
780 <dim>768</dim>
781 </port>
782 </input>
783 <output>
784 <port id="1" precision="FP32">
785 <dim>768</dim>
786 <dim>768</dim>
787 </port>
788 </output>
789 </layer>
790 <layer id="59" name="Constant_220496" type="Const" version="opset1">
791 <data element_type="f32" shape="768, 1" offset="39805344" size="3072" />
792 <output>
793 <port id="0" precision="FP32">
794 <dim>768</dim>
795 <dim>1</dim>
796 </port>
797 </output>
798 </layer>
799 <layer id="60" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
800 <data auto_broadcast="numpy" />
801 <input>
802 <port id="0" precision="FP32">
803 <dim>768</dim>
804 <dim>768</dim>
805 </port>
806 <port id="1" precision="FP32">
807 <dim>768</dim>
808 <dim>1</dim>
809 </port>
810 </input>
811 <output>
812 <port id="2" precision="FP32">
813 <dim>768</dim>
814 <dim>768</dim>
815 </port>
816 </output>
817 </layer>
818 <layer id="61" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
819 <data transpose_a="false" transpose_b="true" />
820 <input>
821 <port id="0" precision="FP32">
822 <dim>-1</dim>
823 <dim>-1</dim>
824 <dim>768</dim>
825 </port>
826 <port id="1" precision="FP32">
827 <dim>768</dim>
828 <dim>768</dim>
829 </port>
830 </input>
831 <output>
832 <port id="2" precision="FP32">
833 <dim>-1</dim>
834 <dim>-1</dim>
835 <dim>768</dim>
836 </port>
837 </output>
838 </layer>
839 <layer id="62" name="Constant_5360" type="Const" version="opset1">
840 <data element_type="f32" shape="1, 1, 768" offset="39808416" size="3072" />
841 <output>
842 <port id="0" precision="FP32">
843 <dim>1</dim>
844 <dim>1</dim>
845 <dim>768</dim>
846 </port>
847 </output>
848 </layer>
849 <layer id="63" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
850 <data auto_broadcast="numpy" />
851 <input>
852 <port id="0" precision="FP32">
853 <dim>-1</dim>
854 <dim>-1</dim>
855 <dim>768</dim>
856 </port>
857 <port id="1" precision="FP32">
858 <dim>1</dim>
859 <dim>1</dim>
860 <dim>768</dim>
861 </port>
862 </input>
863 <output>
864 <port id="2" precision="FP32" names="102,x.1">
865 <dim>-1</dim>
866 <dim>-1</dim>
867 <dim>768</dim>
868 </port>
869 </output>
870 </layer>
871 <layer id="64" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
872 <data element_type="f32" shape="" offset="39811488" size="4" />
873 <output>
874 <port id="0" precision="FP32" />
875 </output>
876 </layer>
877 <layer id="65" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
878 <data element_type="f32" shape="" offset="39811492" size="4" />
879 <output>
880 <port id="0" precision="FP32" />
881 </output>
882 </layer>
883 <layer id="66" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
884 <data element_type="f32" shape="" offset="39811488" size="4" />
885 <output>
886 <port id="0" precision="FP32" />
887 </output>
888 </layer>
889 <layer id="67" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
890 <data element_type="f32" shape="" offset="39811492" size="4" />
891 <output>
892 <port id="0" precision="FP32" />
893 </output>
894 </layer>
895 <layer id="68" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
896 <data levels="256" auto_broadcast="numpy" />
897 <input>
898 <port id="0" precision="FP32">
899 <dim>-1</dim>
900 <dim>-1</dim>
901 <dim>768</dim>
902 </port>
903 <port id="1" precision="FP32" />
904 <port id="2" precision="FP32" />
905 <port id="3" precision="FP32" />
906 <port id="4" precision="FP32" />
907 </input>
908 <output>
909 <port id="5" precision="FP32">
910 <dim>-1</dim>
911 <dim>-1</dim>
912 <dim>768</dim>
913 </port>
914 </output>
915 </layer>
916 <layer id="69" name="__module.roberta.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
917 <data element_type="i64" shape="4" offset="39811496" size="32" />
918 <output>
919 <port id="0" precision="I64">
920 <dim>4</dim>
921 </port>
922 </output>
923 </layer>
924 <layer id="70" name="__module.roberta.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
925 <data special_zero="true" />
926 <input>
927 <port id="0" precision="FP32">
928 <dim>-1</dim>
929 <dim>-1</dim>
930 <dim>768</dim>
931 </port>
932 <port id="1" precision="I64">
933 <dim>4</dim>
934 </port>
935 </input>
936 <output>
937 <port id="2" precision="FP32" names="106,x.3">
938 <dim>-1</dim>
939 <dim>-1</dim>
940 <dim>12</dim>
941 <dim>64</dim>
942 </port>
943 </output>
944 </layer>
945 <layer id="71" name="Constant_229" type="Const" version="opset1">
946 <data element_type="i64" shape="4" offset="39811528" size="32" />
947 <output>
948 <port id="0" precision="I64" names="107">
949 <dim>4</dim>
950 </port>
951 </output>
952 </layer>
953 <layer id="72" name="__module.roberta.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
954 <input>
955 <port id="0" precision="FP32">
956 <dim>-1</dim>
957 <dim>-1</dim>
958 <dim>12</dim>
959 <dim>64</dim>
960 </port>
961 <port id="1" precision="I64">
962 <dim>4</dim>
963 </port>
964 </input>
965 <output>
966 <port id="2" precision="FP32" names="108">
967 <dim>-1</dim>
968 <dim>12</dim>
969 <dim>-1</dim>
970 <dim>64</dim>
971 </port>
972 </output>
973 </layer>
974 <layer id="73" name="Constant_220498" type="Const" version="opset1">
975 <data element_type="i8" shape="768, 768" offset="39811560" size="589824" />
976 <output>
977 <port id="0" precision="I8">
978 <dim>768</dim>
979 <dim>768</dim>
980 </port>
981 </output>
982 </layer>
983 <layer id="74" name="Convert_220499" type="Convert" version="opset1">
984 <data destination_type="f32" />
985 <input>
986 <port id="0" precision="I8">
987 <dim>768</dim>
988 <dim>768</dim>
989 </port>
990 </input>
991 <output>
992 <port id="1" precision="FP32">
993 <dim>768</dim>
994 <dim>768</dim>
995 </port>
996 </output>
997 </layer>
998 <layer id="75" name="Constant_220500" type="Const" version="opset1">
999 <data element_type="f32" shape="768, 1" offset="40401384" size="3072" />
1000 <output>
1001 <port id="0" precision="FP32">
1002 <dim>768</dim>
1003 <dim>1</dim>
1004 </port>
1005 </output>
1006 </layer>
1007 <layer id="76" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1008 <data auto_broadcast="numpy" />
1009 <input>
1010 <port id="0" precision="FP32">
1011 <dim>768</dim>
1012 <dim>768</dim>
1013 </port>
1014 <port id="1" precision="FP32">
1015 <dim>768</dim>
1016 <dim>1</dim>
1017 </port>
1018 </input>
1019 <output>
1020 <port id="2" precision="FP32">
1021 <dim>768</dim>
1022 <dim>768</dim>
1023 </port>
1024 </output>
1025 </layer>
1026 <layer id="77" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
1027 <data transpose_a="false" transpose_b="true" />
1028 <input>
1029 <port id="0" precision="FP32">
1030 <dim>-1</dim>
1031 <dim>-1</dim>
1032 <dim>768</dim>
1033 </port>
1034 <port id="1" precision="FP32">
1035 <dim>768</dim>
1036 <dim>768</dim>
1037 </port>
1038 </input>
1039 <output>
1040 <port id="2" precision="FP32">
1041 <dim>-1</dim>
1042 <dim>-1</dim>
1043 <dim>768</dim>
1044 </port>
1045 </output>
1046 </layer>
1047 <layer id="78" name="Constant_5361" type="Const" version="opset1">
1048 <data element_type="f32" shape="1, 1, 768" offset="40404456" size="3072" />
1049 <output>
1050 <port id="0" precision="FP32">
1051 <dim>1</dim>
1052 <dim>1</dim>
1053 <dim>768</dim>
1054 </port>
1055 </output>
1056 </layer>
1057 <layer id="79" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
1058 <data auto_broadcast="numpy" />
1059 <input>
1060 <port id="0" precision="FP32">
1061 <dim>-1</dim>
1062 <dim>-1</dim>
1063 <dim>768</dim>
1064 </port>
1065 <port id="1" precision="FP32">
1066 <dim>1</dim>
1067 <dim>1</dim>
1068 <dim>768</dim>
1069 </port>
1070 </input>
1071 <output>
1072 <port id="2" precision="FP32" names="111,x.5">
1073 <dim>-1</dim>
1074 <dim>-1</dim>
1075 <dim>768</dim>
1076 </port>
1077 </output>
1078 </layer>
1079 <layer id="80" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
1080 <data element_type="f32" shape="" offset="40407528" size="4" />
1081 <output>
1082 <port id="0" precision="FP32" />
1083 </output>
1084 </layer>
1085 <layer id="81" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
1086 <data element_type="f32" shape="" offset="40407532" size="4" />
1087 <output>
1088 <port id="0" precision="FP32" />
1089 </output>
1090 </layer>
1091 <layer id="82" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
1092 <data element_type="f32" shape="" offset="40407528" size="4" />
1093 <output>
1094 <port id="0" precision="FP32" />
1095 </output>
1096 </layer>
1097 <layer id="83" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
1098 <data element_type="f32" shape="" offset="40407532" size="4" />
1099 <output>
1100 <port id="0" precision="FP32" />
1101 </output>
1102 </layer>
1103 <layer id="84" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
1104 <data levels="256" auto_broadcast="numpy" />
1105 <input>
1106 <port id="0" precision="FP32">
1107 <dim>-1</dim>
1108 <dim>-1</dim>
1109 <dim>768</dim>
1110 </port>
1111 <port id="1" precision="FP32" />
1112 <port id="2" precision="FP32" />
1113 <port id="3" precision="FP32" />
1114 <port id="4" precision="FP32" />
1115 </input>
1116 <output>
1117 <port id="5" precision="FP32">
1118 <dim>-1</dim>
1119 <dim>-1</dim>
1120 <dim>768</dim>
1121 </port>
1122 </output>
1123 </layer>
1124 <layer id="85" name="__module.roberta.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
1125 <data element_type="i64" shape="4" offset="39811496" size="32" />
1126 <output>
1127 <port id="0" precision="I64">
1128 <dim>4</dim>
1129 </port>
1130 </output>
1131 </layer>
1132 <layer id="86" name="__module.roberta.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
1133 <data special_zero="true" />
1134 <input>
1135 <port id="0" precision="FP32">
1136 <dim>-1</dim>
1137 <dim>-1</dim>
1138 <dim>768</dim>
1139 </port>
1140 <port id="1" precision="I64">
1141 <dim>4</dim>
1142 </port>
1143 </input>
1144 <output>
1145 <port id="2" precision="FP32" names="115,x.7">
1146 <dim>-1</dim>
1147 <dim>-1</dim>
1148 <dim>12</dim>
1149 <dim>64</dim>
1150 </port>
1151 </output>
1152 </layer>
1153 <layer id="87" name="Constant_252" type="Const" version="opset1">
1154 <data element_type="i64" shape="4" offset="39811528" size="32" />
1155 <output>
1156 <port id="0" precision="I64" names="116">
1157 <dim>4</dim>
1158 </port>
1159 </output>
1160 </layer>
1161 <layer id="88" name="__module.roberta.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1162 <input>
1163 <port id="0" precision="FP32">
1164 <dim>-1</dim>
1165 <dim>-1</dim>
1166 <dim>12</dim>
1167 <dim>64</dim>
1168 </port>
1169 <port id="1" precision="I64">
1170 <dim>4</dim>
1171 </port>
1172 </input>
1173 <output>
1174 <port id="2" precision="FP32" names="117">
1175 <dim>-1</dim>
1176 <dim>12</dim>
1177 <dim>-1</dim>
1178 <dim>64</dim>
1179 </port>
1180 </output>
1181 </layer>
1182 <layer id="89" name="Constant_220502" type="Const" version="opset1">
1183 <data element_type="i8" shape="768, 768" offset="40407536" size="589824" />
1184 <output>
1185 <port id="0" precision="I8">
1186 <dim>768</dim>
1187 <dim>768</dim>
1188 </port>
1189 </output>
1190 </layer>
1191 <layer id="90" name="Convert_220503" type="Convert" version="opset1">
1192 <data destination_type="f32" />
1193 <input>
1194 <port id="0" precision="I8">
1195 <dim>768</dim>
1196 <dim>768</dim>
1197 </port>
1198 </input>
1199 <output>
1200 <port id="1" precision="FP32">
1201 <dim>768</dim>
1202 <dim>768</dim>
1203 </port>
1204 </output>
1205 </layer>
1206 <layer id="91" name="Constant_220504" type="Const" version="opset1">
1207 <data element_type="f32" shape="768, 1" offset="40997360" size="3072" />
1208 <output>
1209 <port id="0" precision="FP32">
1210 <dim>768</dim>
1211 <dim>1</dim>
1212 </port>
1213 </output>
1214 </layer>
1215 <layer id="92" name="__module.roberta.encoder.layer.0.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1216 <data auto_broadcast="numpy" />
1217 <input>
1218 <port id="0" precision="FP32">
1219 <dim>768</dim>
1220 <dim>768</dim>
1221 </port>
1222 <port id="1" precision="FP32">
1223 <dim>768</dim>
1224 <dim>1</dim>
1225 </port>
1226 </input>
1227 <output>
1228 <port id="2" precision="FP32">
1229 <dim>768</dim>
1230 <dim>768</dim>
1231 </port>
1232 </output>
1233 </layer>
1234 <layer id="93" name="__module.roberta.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1235 <data transpose_a="false" transpose_b="true" />
1236 <input>
1237 <port id="0" precision="FP32">
1238 <dim>-1</dim>
1239 <dim>-1</dim>
1240 <dim>768</dim>
1241 </port>
1242 <port id="1" precision="FP32">
1243 <dim>768</dim>
1244 <dim>768</dim>
1245 </port>
1246 </input>
1247 <output>
1248 <port id="2" precision="FP32">
1249 <dim>-1</dim>
1250 <dim>-1</dim>
1251 <dim>768</dim>
1252 </port>
1253 </output>
1254 </layer>
1255 <layer id="94" name="Constant_5362" type="Const" version="opset1">
1256 <data element_type="f32" shape="1, 1, 768" offset="41000432" size="3072" />
1257 <output>
1258 <port id="0" precision="FP32">
1259 <dim>1</dim>
1260 <dim>1</dim>
1261 <dim>768</dim>
1262 </port>
1263 </output>
1264 </layer>
1265 <layer id="95" name="__module.roberta.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
1266 <data auto_broadcast="numpy" />
1267 <input>
1268 <port id="0" precision="FP32">
1269 <dim>-1</dim>
1270 <dim>-1</dim>
1271 <dim>768</dim>
1272 </port>
1273 <port id="1" precision="FP32">
1274 <dim>1</dim>
1275 <dim>1</dim>
1276 <dim>768</dim>
1277 </port>
1278 </input>
1279 <output>
1280 <port id="2" precision="FP32" names="120,x.9">
1281 <dim>-1</dim>
1282 <dim>-1</dim>
1283 <dim>768</dim>
1284 </port>
1285 </output>
1286 </layer>
1287 <layer id="96" name="__module.roberta.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
1288 <data element_type="i64" shape="4" offset="39811496" size="32" />
1289 <output>
1290 <port id="0" precision="I64">
1291 <dim>4</dim>
1292 </port>
1293 </output>
1294 </layer>
1295 <layer id="97" name="__module.roberta.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
1296 <data special_zero="true" />
1297 <input>
1298 <port id="0" precision="FP32">
1299 <dim>-1</dim>
1300 <dim>-1</dim>
1301 <dim>768</dim>
1302 </port>
1303 <port id="1" precision="I64">
1304 <dim>4</dim>
1305 </port>
1306 </input>
1307 <output>
1308 <port id="2" precision="FP32" names="124,x.11">
1309 <dim>-1</dim>
1310 <dim>-1</dim>
1311 <dim>12</dim>
1312 <dim>64</dim>
1313 </port>
1314 </output>
1315 </layer>
1316 <layer id="98" name="Constant_275" type="Const" version="opset1">
1317 <data element_type="i64" shape="4" offset="39811528" size="32" />
1318 <output>
1319 <port id="0" precision="I64" names="125">
1320 <dim>4</dim>
1321 </port>
1322 </output>
1323 </layer>
1324 <layer id="99" name="__module.roberta.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
1325 <input>
1326 <port id="0" precision="FP32">
1327 <dim>-1</dim>
1328 <dim>-1</dim>
1329 <dim>12</dim>
1330 <dim>64</dim>
1331 </port>
1332 <port id="1" precision="I64">
1333 <dim>4</dim>
1334 </port>
1335 </input>
1336 <output>
1337 <port id="2" precision="FP32" names="126">
1338 <dim>-1</dim>
1339 <dim>12</dim>
1340 <dim>-1</dim>
1341 <dim>64</dim>
1342 </port>
1343 </output>
1344 </layer>
1345 <layer id="100" name="Constant_5364" type="Const" version="opset1">
1346 <data element_type="f32" shape="1, 1, 1, 1" offset="41003504" size="4" />
1347 <output>
1348 <port id="0" precision="FP32">
1349 <dim>1</dim>
1350 <dim>1</dim>
1351 <dim>1</dim>
1352 <dim>1</dim>
1353 </port>
1354 </output>
1355 </layer>
1356 <layer id="101" name="__module.roberta/aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
1357 <input>
1358 <port id="0" precision="I64">
1359 <dim>-1</dim>
1360 <dim>-1</dim>
1361 </port>
1362 <port id="1" precision="I64" />
1363 </input>
1364 <output>
1365 <port id="2" precision="I64" names="69">
1366 <dim>-1</dim>
1367 <dim>1</dim>
1368 <dim>-1</dim>
1369 </port>
1370 </output>
1371 </layer>
1372 <layer id="102" name="16" type="Const" version="opset1">
1373 <data element_type="i64" shape="" offset="41003508" size="8" />
1374 <output>
1375 <port id="0" precision="I64" names="16" />
1376 </output>
1377 </layer>
1378 <layer id="103" name="__module.roberta/aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
1379 <input>
1380 <port id="0" precision="I64">
1381 <dim>-1</dim>
1382 <dim>1</dim>
1383 <dim>-1</dim>
1384 </port>
1385 <port id="1" precision="I64" />
1386 </input>
1387 <output>
1388 <port id="2" precision="I64" names="70,71">
1389 <dim>-1</dim>
1390 <dim>1</dim>
1391 <dim>1</dim>
1392 <dim>-1</dim>
1393 </port>
1394 </output>
1395 </layer>
1396 <layer id="104" name="Constant_5490" type="Const" version="opset1">
1397 <data element_type="i64" shape="1" offset="38809468" size="8" />
1398 <output>
1399 <port id="0" precision="I64">
1400 <dim>1</dim>
1401 </port>
1402 </output>
1403 </layer>
1404 <layer id="105" name="Constant_5491" type="Const" version="opset1">
1405 <data element_type="i64" shape="" offset="38809468" size="8" />
1406 <output>
1407 <port id="0" precision="I64" />
1408 </output>
1409 </layer>
1410 <layer id="106" name="Gather_5492" type="Gather" version="opset8">
1411 <data batch_dims="0" />
1412 <input>
1413 <port id="0" precision="I64">
1414 <dim>2</dim>
1415 </port>
1416 <port id="1" precision="I64">
1417 <dim>1</dim>
1418 </port>
1419 <port id="2" precision="I64" />
1420 </input>
1421 <output>
1422 <port id="3" precision="I64" names="66">
1423 <dim>1</dim>
1424 </port>
1425 </output>
1426 </layer>
1427 <layer id="107" name="Constant_4638" type="Const" version="opset1">
1428 <data element_type="i64" shape="1" offset="38809476" size="8" />
1429 <output>
1430 <port id="0" precision="I64">
1431 <dim>1</dim>
1432 </port>
1433 </output>
1434 </layer>
1435 <layer id="108" name="Constant_5596" type="Const" version="opset1">
1436 <data element_type="i64" shape="2" offset="41003516" size="16" />
1437 <output>
1438 <port id="0" precision="I64">
1439 <dim>2</dim>
1440 </port>
1441 </output>
1442 </layer>
1443 <layer id="109" name="Constant_5597" type="Const" version="opset1">
1444 <data element_type="i64" shape="" offset="38809468" size="8" />
1445 <output>
1446 <port id="0" precision="I64" />
1447 </output>
1448 </layer>
1449 <layer id="110" name="Gather_5598" type="Gather" version="opset8">
1450 <data batch_dims="0" />
1451 <input>
1452 <port id="0" precision="I64">
1453 <dim>2</dim>
1454 </port>
1455 <port id="1" precision="I64">
1456 <dim>2</dim>
1457 </port>
1458 <port id="2" precision="I64" />
1459 </input>
1460 <output>
1461 <port id="3" precision="I64">
1462 <dim>2</dim>
1463 </port>
1464 </output>
1465 </layer>
1466 <layer id="111" name="__module.roberta/prim::ListConstruct/Concat_1" type="Concat" version="opset1">
1467 <data axis="0" />
1468 <input>
1469 <port id="0" precision="I64">
1470 <dim>1</dim>
1471 </port>
1472 <port id="1" precision="I64">
1473 <dim>1</dim>
1474 </port>
1475 <port id="2" precision="I64">
1476 <dim>2</dim>
1477 </port>
1478 </input>
1479 <output>
1480 <port id="3" precision="I64" names="72">
1481 <dim>4</dim>
1482 </port>
1483 </output>
1484 </layer>
1485 <layer id="112" name="__module.roberta/aten::expand/Broadcast_1" type="Broadcast" version="opset3">
1486 <data mode="bidirectional" />
1487 <input>
1488 <port id="0" precision="I64">
1489 <dim>-1</dim>
1490 <dim>1</dim>
1491 <dim>1</dim>
1492 <dim>-1</dim>
1493 </port>
1494 <port id="1" precision="I64">
1495 <dim>4</dim>
1496 </port>
1497 </input>
1498 <output>
1499 <port id="2" precision="I64" names="73">
1500 <dim>-1</dim>
1501 <dim>1</dim>
1502 <dim>-1</dim>
1503 <dim>-1</dim>
1504 </port>
1505 </output>
1506 </layer>
1507 <layer id="113" name="__module.roberta/aten::to/Convert" type="Convert" version="opset1">
1508 <data destination_type="f32" />
1509 <input>
1510 <port id="0" precision="I64">
1511 <dim>-1</dim>
1512 <dim>1</dim>
1513 <dim>-1</dim>
1514 <dim>-1</dim>
1515 </port>
1516 </input>
1517 <output>
1518 <port id="1" precision="FP32" names="74">
1519 <dim>-1</dim>
1520 <dim>1</dim>
1521 <dim>-1</dim>
1522 <dim>-1</dim>
1523 </port>
1524 </output>
1525 </layer>
1526 <layer id="114" name="Constant_5363" type="Const" version="opset1">
1527 <data element_type="f32" shape="1, 1, 1, 1" offset="41003504" size="4" />
1528 <output>
1529 <port id="0" precision="FP32">
1530 <dim>1</dim>
1531 <dim>1</dim>
1532 <dim>1</dim>
1533 <dim>1</dim>
1534 </port>
1535 </output>
1536 </layer>
1537 <layer id="115" name="__module.roberta/aten::rsub/Multiply" type="Multiply" version="opset1">
1538 <data auto_broadcast="numpy" />
1539 <input>
1540 <port id="0" precision="FP32">
1541 <dim>-1</dim>
1542 <dim>1</dim>
1543 <dim>-1</dim>
1544 <dim>-1</dim>
1545 </port>
1546 <port id="1" precision="FP32">
1547 <dim>1</dim>
1548 <dim>1</dim>
1549 <dim>1</dim>
1550 <dim>1</dim>
1551 </port>
1552 </input>
1553 <output>
1554 <port id="2" precision="FP32">
1555 <dim>-1</dim>
1556 <dim>1</dim>
1557 <dim>-1</dim>
1558 <dim>-1</dim>
1559 </port>
1560 </output>
1561 </layer>
1562 <layer id="116" name="__module.roberta/aten::rsub/Subtract" type="Subtract" version="opset1">
1563 <data auto_broadcast="numpy" />
1564 <input>
1565 <port id="0" precision="FP32">
1566 <dim>1</dim>
1567 <dim>1</dim>
1568 <dim>1</dim>
1569 <dim>1</dim>
1570 </port>
1571 <port id="1" precision="FP32">
1572 <dim>-1</dim>
1573 <dim>1</dim>
1574 <dim>-1</dim>
1575 <dim>-1</dim>
1576 </port>
1577 </input>
1578 <output>
1579 <port id="2" precision="FP32" names="75,inverted_mask">
1580 <dim>-1</dim>
1581 <dim>1</dim>
1582 <dim>-1</dim>
1583 <dim>-1</dim>
1584 </port>
1585 </output>
1586 </layer>
1587 <layer id="117" name="__module.roberta/aten::to/Convert_1" type="Convert" version="opset1">
1588 <data destination_type="boolean" />
1589 <input>
1590 <port id="0" precision="FP32">
1591 <dim>-1</dim>
1592 <dim>1</dim>
1593 <dim>-1</dim>
1594 <dim>-1</dim>
1595 </port>
1596 </input>
1597 <output>
1598 <port id="1" precision="BOOL" names="76">
1599 <dim>-1</dim>
1600 <dim>1</dim>
1601 <dim>-1</dim>
1602 <dim>-1</dim>
1603 </port>
1604 </output>
1605 </layer>
1606 <layer id="118" name="__module.roberta/aten::masked_fill/ConvertLike" type="Const" version="opset1">
1607 <data element_type="f32" shape="" offset="41003532" size="4" />
1608 <output>
1609 <port id="0" precision="FP32" />
1610 </output>
1611 </layer>
1612 <layer id="119" name="__module.roberta/aten::masked_fill/Select" type="Select" version="opset1">
1613 <data auto_broadcast="numpy" />
1614 <input>
1615 <port id="0" precision="BOOL">
1616 <dim>-1</dim>
1617 <dim>1</dim>
1618 <dim>-1</dim>
1619 <dim>-1</dim>
1620 </port>
1621 <port id="1" precision="FP32" />
1622 <port id="2" precision="FP32">
1623 <dim>-1</dim>
1624 <dim>1</dim>
1625 <dim>-1</dim>
1626 <dim>-1</dim>
1627 </port>
1628 </input>
1629 <output>
1630 <port id="3" precision="FP32" names="77">
1631 <dim>-1</dim>
1632 <dim>1</dim>
1633 <dim>-1</dim>
1634 <dim>-1</dim>
1635 </port>
1636 </output>
1637 </layer>
1638 <layer id="120" name="__module.roberta.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1639 <data causal="false" />
1640 <input>
1641 <port id="0" precision="FP32">
1642 <dim>-1</dim>
1643 <dim>12</dim>
1644 <dim>-1</dim>
1645 <dim>64</dim>
1646 </port>
1647 <port id="1" precision="FP32">
1648 <dim>-1</dim>
1649 <dim>12</dim>
1650 <dim>-1</dim>
1651 <dim>64</dim>
1652 </port>
1653 <port id="2" precision="FP32">
1654 <dim>-1</dim>
1655 <dim>12</dim>
1656 <dim>-1</dim>
1657 <dim>64</dim>
1658 </port>
1659 <port id="3" precision="FP32">
1660 <dim>-1</dim>
1661 <dim>1</dim>
1662 <dim>-1</dim>
1663 <dim>-1</dim>
1664 </port>
1665 </input>
1666 <output>
1667 <port id="4" precision="FP32" names="127,attn_output.1">
1668 <dim>-1</dim>
1669 <dim>12</dim>
1670 <dim>-1</dim>
1671 <dim>64</dim>
1672 </port>
1673 </output>
1674 </layer>
1675 <layer id="121" name="__module.roberta.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1676 <data element_type="i32" shape="4" offset="41003536" size="16" />
1677 <output>
1678 <port id="0" precision="I32">
1679 <dim>4</dim>
1680 </port>
1681 </output>
1682 </layer>
1683 <layer id="122" name="__module.roberta.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1684 <input>
1685 <port id="0" precision="FP32">
1686 <dim>-1</dim>
1687 <dim>12</dim>
1688 <dim>-1</dim>
1689 <dim>64</dim>
1690 </port>
1691 <port id="1" precision="I32">
1692 <dim>4</dim>
1693 </port>
1694 </input>
1695 <output>
1696 <port id="2" precision="FP32" names="128,attn_output.3">
1697 <dim>-1</dim>
1698 <dim>-1</dim>
1699 <dim>12</dim>
1700 <dim>64</dim>
1701 </port>
1702 </output>
1703 </layer>
1704 <layer id="123" name="Constant_5494" type="Const" version="opset1">
1705 <data element_type="i64" shape="3" offset="41003552" size="24" />
1706 <output>
1707 <port id="0" precision="I64">
1708 <dim>3</dim>
1709 </port>
1710 </output>
1711 </layer>
1712 <layer id="124" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1713 <data special_zero="true" />
1714 <input>
1715 <port id="0" precision="FP32">
1716 <dim>-1</dim>
1717 <dim>-1</dim>
1718 <dim>12</dim>
1719 <dim>64</dim>
1720 </port>
1721 <port id="1" precision="I64">
1722 <dim>3</dim>
1723 </port>
1724 </input>
1725 <output>
1726 <port id="2" precision="FP32" names="130">
1727 <dim>-1</dim>
1728 <dim>-1</dim>
1729 <dim>768</dim>
1730 </port>
1731 </output>
1732 </layer>
1733 <layer id="125" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
1734 <data element_type="f32" shape="1, 1, 768" offset="41003576" size="3072" />
1735 <output>
1736 <port id="0" precision="FP32">
1737 <dim>1</dim>
1738 <dim>1</dim>
1739 <dim>768</dim>
1740 </port>
1741 </output>
1742 </layer>
1743 <layer id="126" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
1744 <data auto_broadcast="numpy" />
1745 <input>
1746 <port id="0" precision="FP32">
1747 <dim>-1</dim>
1748 <dim>-1</dim>
1749 <dim>768</dim>
1750 </port>
1751 <port id="1" precision="FP32">
1752 <dim>1</dim>
1753 <dim>1</dim>
1754 <dim>768</dim>
1755 </port>
1756 </input>
1757 <output>
1758 <port id="2" precision="FP32">
1759 <dim>-1</dim>
1760 <dim>-1</dim>
1761 <dim>768</dim>
1762 </port>
1763 </output>
1764 </layer>
1765 <layer id="127" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
1766 <data element_type="f32" shape="" offset="41006648" size="4" />
1767 <output>
1768 <port id="0" precision="FP32" />
1769 </output>
1770 </layer>
1771 <layer id="128" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
1772 <data element_type="f32" shape="" offset="41006652" size="4" />
1773 <output>
1774 <port id="0" precision="FP32" />
1775 </output>
1776 </layer>
1777 <layer id="129" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
1778 <data element_type="f32" shape="" offset="41006648" size="4" />
1779 <output>
1780 <port id="0" precision="FP32" />
1781 </output>
1782 </layer>
1783 <layer id="130" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
1784 <data element_type="f32" shape="" offset="41006652" size="4" />
1785 <output>
1786 <port id="0" precision="FP32" />
1787 </output>
1788 </layer>
1789 <layer id="131" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
1790 <data levels="256" auto_broadcast="numpy" />
1791 <input>
1792 <port id="0" precision="FP32">
1793 <dim>-1</dim>
1794 <dim>-1</dim>
1795 <dim>768</dim>
1796 </port>
1797 <port id="1" precision="FP32" />
1798 <port id="2" precision="FP32" />
1799 <port id="3" precision="FP32" />
1800 <port id="4" precision="FP32" />
1801 </input>
1802 <output>
1803 <port id="5" precision="FP32">
1804 <dim>-1</dim>
1805 <dim>-1</dim>
1806 <dim>768</dim>
1807 </port>
1808 </output>
1809 </layer>
1810 <layer id="132" name="Constant_220506" type="Const" version="opset1">
1811 <data element_type="i8" shape="768, 768" offset="41006656" size="589824" />
1812 <output>
1813 <port id="0" precision="I8">
1814 <dim>768</dim>
1815 <dim>768</dim>
1816 </port>
1817 </output>
1818 </layer>
1819 <layer id="133" name="Convert_220507" type="Convert" version="opset1">
1820 <data destination_type="f32" />
1821 <input>
1822 <port id="0" precision="I8">
1823 <dim>768</dim>
1824 <dim>768</dim>
1825 </port>
1826 </input>
1827 <output>
1828 <port id="1" precision="FP32">
1829 <dim>768</dim>
1830 <dim>768</dim>
1831 </port>
1832 </output>
1833 </layer>
1834 <layer id="134" name="Constant_220508" type="Const" version="opset1">
1835 <data element_type="f32" shape="768, 1" offset="41596480" size="3072" />
1836 <output>
1837 <port id="0" precision="FP32">
1838 <dim>768</dim>
1839 <dim>1</dim>
1840 </port>
1841 </output>
1842 </layer>
1843 <layer id="135" name="__module.roberta.encoder.layer.0.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
1844 <data auto_broadcast="numpy" />
1845 <input>
1846 <port id="0" precision="FP32">
1847 <dim>768</dim>
1848 <dim>768</dim>
1849 </port>
1850 <port id="1" precision="FP32">
1851 <dim>768</dim>
1852 <dim>1</dim>
1853 </port>
1854 </input>
1855 <output>
1856 <port id="2" precision="FP32">
1857 <dim>768</dim>
1858 <dim>768</dim>
1859 </port>
1860 </output>
1861 </layer>
1862 <layer id="136" name="__module.roberta.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1863 <data transpose_a="false" transpose_b="true" />
1864 <input>
1865 <port id="0" precision="FP32">
1866 <dim>-1</dim>
1867 <dim>-1</dim>
1868 <dim>768</dim>
1869 </port>
1870 <port id="1" precision="FP32">
1871 <dim>768</dim>
1872 <dim>768</dim>
1873 </port>
1874 </input>
1875 <output>
1876 <port id="2" precision="FP32">
1877 <dim>-1</dim>
1878 <dim>-1</dim>
1879 <dim>768</dim>
1880 </port>
1881 </output>
1882 </layer>
1883 <layer id="137" name="Constant_5365" type="Const" version="opset1">
1884 <data element_type="f32" shape="1, 1, 768" offset="41599552" size="3072" />
1885 <output>
1886 <port id="0" precision="FP32">
1887 <dim>1</dim>
1888 <dim>1</dim>
1889 <dim>768</dim>
1890 </port>
1891 </output>
1892 </layer>
1893 <layer id="138" name="__module.roberta.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1894 <data auto_broadcast="numpy" />
1895 <input>
1896 <port id="0" precision="FP32">
1897 <dim>-1</dim>
1898 <dim>-1</dim>
1899 <dim>768</dim>
1900 </port>
1901 <port id="1" precision="FP32">
1902 <dim>1</dim>
1903 <dim>1</dim>
1904 <dim>768</dim>
1905 </port>
1906 </input>
1907 <output>
1908 <port id="2" precision="FP32" names="135,input.3">
1909 <dim>-1</dim>
1910 <dim>-1</dim>
1911 <dim>768</dim>
1912 </port>
1913 </output>
1914 </layer>
1915 <layer id="139" name="__module.roberta.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1916 <data auto_broadcast="numpy" />
1917 <input>
1918 <port id="0" precision="FP32">
1919 <dim>-1</dim>
1920 <dim>-1</dim>
1921 <dim>768</dim>
1922 </port>
1923 <port id="1" precision="FP32">
1924 <dim>-1</dim>
1925 <dim>-1</dim>
1926 <dim>768</dim>
1927 </port>
1928 </input>
1929 <output>
1930 <port id="2" precision="FP32" names="137">
1931 <dim>-1</dim>
1932 <dim>-1</dim>
1933 <dim>768</dim>
1934 </port>
1935 </output>
1936 </layer>
1937 <layer id="140" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1938 <data element_type="i32" shape="1" offset="39206292" size="4" />
1939 <output>
1940 <port id="0" precision="I32">
1941 <dim>1</dim>
1942 </port>
1943 </output>
1944 </layer>
1945 <layer id="141" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1946 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1947 <input>
1948 <port id="0" precision="FP32">
1949 <dim>-1</dim>
1950 <dim>-1</dim>
1951 <dim>768</dim>
1952 </port>
1953 <port id="1" precision="I32">
1954 <dim>1</dim>
1955 </port>
1956 </input>
1957 <output>
1958 <port id="2" precision="FP32">
1959 <dim>-1</dim>
1960 <dim>-1</dim>
1961 <dim>768</dim>
1962 </port>
1963 </output>
1964 </layer>
1965 <layer id="142" name="Constant_5366" type="Const" version="opset1">
1966 <data element_type="f32" shape="1, 1, 768" offset="41602624" size="3072" />
1967 <output>
1968 <port id="0" precision="FP32">
1969 <dim>1</dim>
1970 <dim>1</dim>
1971 <dim>768</dim>
1972 </port>
1973 </output>
1974 </layer>
1975 <layer id="143" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1976 <data auto_broadcast="numpy" />
1977 <input>
1978 <port id="0" precision="FP32">
1979 <dim>-1</dim>
1980 <dim>-1</dim>
1981 <dim>768</dim>
1982 </port>
1983 <port id="1" precision="FP32">
1984 <dim>1</dim>
1985 <dim>1</dim>
1986 <dim>768</dim>
1987 </port>
1988 </input>
1989 <output>
1990 <port id="2" precision="FP32">
1991 <dim>-1</dim>
1992 <dim>-1</dim>
1993 <dim>768</dim>
1994 </port>
1995 </output>
1996 </layer>
1997 <layer id="144" name="Constant_5367" type="Const" version="opset1">
1998 <data element_type="f32" shape="1, 1, 768" offset="41605696" size="3072" />
1999 <output>
2000 <port id="0" precision="FP32">
2001 <dim>1</dim>
2002 <dim>1</dim>
2003 <dim>768</dim>
2004 </port>
2005 </output>
2006 </layer>
2007 <layer id="145" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2008 <data auto_broadcast="numpy" />
2009 <input>
2010 <port id="0" precision="FP32">
2011 <dim>-1</dim>
2012 <dim>-1</dim>
2013 <dim>768</dim>
2014 </port>
2015 <port id="1" precision="FP32">
2016 <dim>1</dim>
2017 <dim>1</dim>
2018 <dim>768</dim>
2019 </port>
2020 </input>
2021 <output>
2022 <port id="2" precision="FP32" names="141,input_tensor.1">
2023 <dim>-1</dim>
2024 <dim>-1</dim>
2025 <dim>768</dim>
2026 </port>
2027 </output>
2028 </layer>
2029 <layer id="146" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2030 <data element_type="f32" shape="1, 1, 768" offset="41608768" size="3072" />
2031 <output>
2032 <port id="0" precision="FP32">
2033 <dim>1</dim>
2034 <dim>1</dim>
2035 <dim>768</dim>
2036 </port>
2037 </output>
2038 </layer>
2039 <layer id="147" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2040 <data auto_broadcast="numpy" />
2041 <input>
2042 <port id="0" precision="FP32">
2043 <dim>-1</dim>
2044 <dim>-1</dim>
2045 <dim>768</dim>
2046 </port>
2047 <port id="1" precision="FP32">
2048 <dim>1</dim>
2049 <dim>1</dim>
2050 <dim>768</dim>
2051 </port>
2052 </input>
2053 <output>
2054 <port id="2" precision="FP32">
2055 <dim>-1</dim>
2056 <dim>-1</dim>
2057 <dim>768</dim>
2058 </port>
2059 </output>
2060 </layer>
2061 <layer id="148" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2062 <data element_type="f32" shape="" offset="41611840" size="4" />
2063 <output>
2064 <port id="0" precision="FP32" />
2065 </output>
2066 </layer>
2067 <layer id="149" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2068 <data element_type="f32" shape="" offset="41611844" size="4" />
2069 <output>
2070 <port id="0" precision="FP32" />
2071 </output>
2072 </layer>
2073 <layer id="150" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2074 <data element_type="f32" shape="" offset="41611840" size="4" />
2075 <output>
2076 <port id="0" precision="FP32" />
2077 </output>
2078 </layer>
2079 <layer id="151" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2080 <data element_type="f32" shape="" offset="41611844" size="4" />
2081 <output>
2082 <port id="0" precision="FP32" />
2083 </output>
2084 </layer>
2085 <layer id="152" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2086 <data levels="256" auto_broadcast="numpy" />
2087 <input>
2088 <port id="0" precision="FP32">
2089 <dim>-1</dim>
2090 <dim>-1</dim>
2091 <dim>768</dim>
2092 </port>
2093 <port id="1" precision="FP32" />
2094 <port id="2" precision="FP32" />
2095 <port id="3" precision="FP32" />
2096 <port id="4" precision="FP32" />
2097 </input>
2098 <output>
2099 <port id="5" precision="FP32">
2100 <dim>-1</dim>
2101 <dim>-1</dim>
2102 <dim>768</dim>
2103 </port>
2104 </output>
2105 </layer>
2106 <layer id="153" name="Constant_220510" type="Const" version="opset1">
2107 <data element_type="i8" shape="3072, 768" offset="41611848" size="2359296" />
2108 <output>
2109 <port id="0" precision="I8">
2110 <dim>3072</dim>
2111 <dim>768</dim>
2112 </port>
2113 </output>
2114 </layer>
2115 <layer id="154" name="Convert_220511" type="Convert" version="opset1">
2116 <data destination_type="f32" />
2117 <input>
2118 <port id="0" precision="I8">
2119 <dim>3072</dim>
2120 <dim>768</dim>
2121 </port>
2122 </input>
2123 <output>
2124 <port id="1" precision="FP32">
2125 <dim>3072</dim>
2126 <dim>768</dim>
2127 </port>
2128 </output>
2129 </layer>
2130 <layer id="155" name="Constant_220512" type="Const" version="opset1">
2131 <data element_type="f32" shape="3072, 1" offset="43971144" size="12288" />
2132 <output>
2133 <port id="0" precision="FP32">
2134 <dim>3072</dim>
2135 <dim>1</dim>
2136 </port>
2137 </output>
2138 </layer>
2139 <layer id="156" name="__module.roberta.encoder.layer.0.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2140 <data auto_broadcast="numpy" />
2141 <input>
2142 <port id="0" precision="FP32">
2143 <dim>3072</dim>
2144 <dim>768</dim>
2145 </port>
2146 <port id="1" precision="FP32">
2147 <dim>3072</dim>
2148 <dim>1</dim>
2149 </port>
2150 </input>
2151 <output>
2152 <port id="2" precision="FP32">
2153 <dim>3072</dim>
2154 <dim>768</dim>
2155 </port>
2156 </output>
2157 </layer>
2158 <layer id="157" name="__module.roberta.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2159 <data transpose_a="false" transpose_b="true" />
2160 <input>
2161 <port id="0" precision="FP32">
2162 <dim>-1</dim>
2163 <dim>-1</dim>
2164 <dim>768</dim>
2165 </port>
2166 <port id="1" precision="FP32">
2167 <dim>3072</dim>
2168 <dim>768</dim>
2169 </port>
2170 </input>
2171 <output>
2172 <port id="2" precision="FP32">
2173 <dim>-1</dim>
2174 <dim>-1</dim>
2175 <dim>3072</dim>
2176 </port>
2177 </output>
2178 </layer>
2179 <layer id="158" name="Constant_5368" type="Const" version="opset1">
2180 <data element_type="f32" shape="1, 1, 3072" offset="43983432" size="12288" />
2181 <output>
2182 <port id="0" precision="FP32">
2183 <dim>1</dim>
2184 <dim>1</dim>
2185 <dim>3072</dim>
2186 </port>
2187 </output>
2188 </layer>
2189 <layer id="159" name="__module.roberta.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2190 <data auto_broadcast="numpy" />
2191 <input>
2192 <port id="0" precision="FP32">
2193 <dim>-1</dim>
2194 <dim>-1</dim>
2195 <dim>3072</dim>
2196 </port>
2197 <port id="1" precision="FP32">
2198 <dim>1</dim>
2199 <dim>1</dim>
2200 <dim>3072</dim>
2201 </port>
2202 </input>
2203 <output>
2204 <port id="2" precision="FP32" names="145">
2205 <dim>-1</dim>
2206 <dim>-1</dim>
2207 <dim>3072</dim>
2208 </port>
2209 </output>
2210 </layer>
2211 <layer id="160" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2212 <data approximation_mode="ERF" />
2213 <input>
2214 <port id="0" precision="FP32">
2215 <dim>-1</dim>
2216 <dim>-1</dim>
2217 <dim>3072</dim>
2218 </port>
2219 </input>
2220 <output>
2221 <port id="1" precision="FP32" names="146">
2222 <dim>-1</dim>
2223 <dim>-1</dim>
2224 <dim>3072</dim>
2225 </port>
2226 </output>
2227 </layer>
2228 <layer id="161" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2229 <data element_type="f32" shape="1, 1, 3072" offset="43995720" size="12288" />
2230 <output>
2231 <port id="0" precision="FP32">
2232 <dim>1</dim>
2233 <dim>1</dim>
2234 <dim>3072</dim>
2235 </port>
2236 </output>
2237 </layer>
2238 <layer id="162" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2239 <data auto_broadcast="numpy" />
2240 <input>
2241 <port id="0" precision="FP32">
2242 <dim>-1</dim>
2243 <dim>-1</dim>
2244 <dim>3072</dim>
2245 </port>
2246 <port id="1" precision="FP32">
2247 <dim>1</dim>
2248 <dim>1</dim>
2249 <dim>3072</dim>
2250 </port>
2251 </input>
2252 <output>
2253 <port id="2" precision="FP32">
2254 <dim>-1</dim>
2255 <dim>-1</dim>
2256 <dim>3072</dim>
2257 </port>
2258 </output>
2259 </layer>
2260 <layer id="163" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2261 <data element_type="f32" shape="" offset="44008008" size="4" />
2262 <output>
2263 <port id="0" precision="FP32" />
2264 </output>
2265 </layer>
2266 <layer id="164" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2267 <data element_type="f32" shape="" offset="44008012" size="4" />
2268 <output>
2269 <port id="0" precision="FP32" />
2270 </output>
2271 </layer>
2272 <layer id="165" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2273 <data element_type="f32" shape="" offset="44008008" size="4" />
2274 <output>
2275 <port id="0" precision="FP32" />
2276 </output>
2277 </layer>
2278 <layer id="166" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2279 <data element_type="f32" shape="" offset="44008012" size="4" />
2280 <output>
2281 <port id="0" precision="FP32" />
2282 </output>
2283 </layer>
2284 <layer id="167" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2285 <data levels="256" auto_broadcast="numpy" />
2286 <input>
2287 <port id="0" precision="FP32">
2288 <dim>-1</dim>
2289 <dim>-1</dim>
2290 <dim>3072</dim>
2291 </port>
2292 <port id="1" precision="FP32" />
2293 <port id="2" precision="FP32" />
2294 <port id="3" precision="FP32" />
2295 <port id="4" precision="FP32" />
2296 </input>
2297 <output>
2298 <port id="5" precision="FP32">
2299 <dim>-1</dim>
2300 <dim>-1</dim>
2301 <dim>3072</dim>
2302 </port>
2303 </output>
2304 </layer>
2305 <layer id="168" name="Constant_220514" type="Const" version="opset1">
2306 <data element_type="i8" shape="768, 3072" offset="44008016" size="2359296" />
2307 <output>
2308 <port id="0" precision="I8">
2309 <dim>768</dim>
2310 <dim>3072</dim>
2311 </port>
2312 </output>
2313 </layer>
2314 <layer id="169" name="Convert_220515" type="Convert" version="opset1">
2315 <data destination_type="f32" />
2316 <input>
2317 <port id="0" precision="I8">
2318 <dim>768</dim>
2319 <dim>3072</dim>
2320 </port>
2321 </input>
2322 <output>
2323 <port id="1" precision="FP32">
2324 <dim>768</dim>
2325 <dim>3072</dim>
2326 </port>
2327 </output>
2328 </layer>
2329 <layer id="170" name="Constant_220516" type="Const" version="opset1">
2330 <data element_type="f32" shape="768, 1" offset="46367312" size="3072" />
2331 <output>
2332 <port id="0" precision="FP32">
2333 <dim>768</dim>
2334 <dim>1</dim>
2335 </port>
2336 </output>
2337 </layer>
2338 <layer id="171" name="__module.roberta.encoder.layer.0.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2339 <data auto_broadcast="numpy" />
2340 <input>
2341 <port id="0" precision="FP32">
2342 <dim>768</dim>
2343 <dim>3072</dim>
2344 </port>
2345 <port id="1" precision="FP32">
2346 <dim>768</dim>
2347 <dim>1</dim>
2348 </port>
2349 </input>
2350 <output>
2351 <port id="2" precision="FP32">
2352 <dim>768</dim>
2353 <dim>3072</dim>
2354 </port>
2355 </output>
2356 </layer>
2357 <layer id="172" name="__module.roberta.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2358 <data transpose_a="false" transpose_b="true" />
2359 <input>
2360 <port id="0" precision="FP32">
2361 <dim>-1</dim>
2362 <dim>-1</dim>
2363 <dim>3072</dim>
2364 </port>
2365 <port id="1" precision="FP32">
2366 <dim>768</dim>
2367 <dim>3072</dim>
2368 </port>
2369 </input>
2370 <output>
2371 <port id="2" precision="FP32">
2372 <dim>-1</dim>
2373 <dim>-1</dim>
2374 <dim>768</dim>
2375 </port>
2376 </output>
2377 </layer>
2378 <layer id="173" name="Constant_5369" type="Const" version="opset1">
2379 <data element_type="f32" shape="1, 1, 768" offset="46370384" size="3072" />
2380 <output>
2381 <port id="0" precision="FP32">
2382 <dim>1</dim>
2383 <dim>1</dim>
2384 <dim>768</dim>
2385 </port>
2386 </output>
2387 </layer>
2388 <layer id="174" name="__module.roberta.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
2389 <data auto_broadcast="numpy" />
2390 <input>
2391 <port id="0" precision="FP32">
2392 <dim>-1</dim>
2393 <dim>-1</dim>
2394 <dim>768</dim>
2395 </port>
2396 <port id="1" precision="FP32">
2397 <dim>1</dim>
2398 <dim>1</dim>
2399 <dim>768</dim>
2400 </port>
2401 </input>
2402 <output>
2403 <port id="2" precision="FP32" names="151,input.5">
2404 <dim>-1</dim>
2405 <dim>-1</dim>
2406 <dim>768</dim>
2407 </port>
2408 </output>
2409 </layer>
2410 <layer id="175" name="__module.roberta.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
2411 <data auto_broadcast="numpy" />
2412 <input>
2413 <port id="0" precision="FP32">
2414 <dim>-1</dim>
2415 <dim>-1</dim>
2416 <dim>768</dim>
2417 </port>
2418 <port id="1" precision="FP32">
2419 <dim>-1</dim>
2420 <dim>-1</dim>
2421 <dim>768</dim>
2422 </port>
2423 </input>
2424 <output>
2425 <port id="2" precision="FP32" names="153">
2426 <dim>-1</dim>
2427 <dim>-1</dim>
2428 <dim>768</dim>
2429 </port>
2430 </output>
2431 </layer>
2432 <layer id="176" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2433 <data element_type="i32" shape="1" offset="39206292" size="4" />
2434 <output>
2435 <port id="0" precision="I32">
2436 <dim>1</dim>
2437 </port>
2438 </output>
2439 </layer>
2440 <layer id="177" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2441 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2442 <input>
2443 <port id="0" precision="FP32">
2444 <dim>-1</dim>
2445 <dim>-1</dim>
2446 <dim>768</dim>
2447 </port>
2448 <port id="1" precision="I32">
2449 <dim>1</dim>
2450 </port>
2451 </input>
2452 <output>
2453 <port id="2" precision="FP32">
2454 <dim>-1</dim>
2455 <dim>-1</dim>
2456 <dim>768</dim>
2457 </port>
2458 </output>
2459 </layer>
2460 <layer id="178" name="Constant_5370" type="Const" version="opset1">
2461 <data element_type="f32" shape="1, 1, 768" offset="46373456" size="3072" />
2462 <output>
2463 <port id="0" precision="FP32">
2464 <dim>1</dim>
2465 <dim>1</dim>
2466 <dim>768</dim>
2467 </port>
2468 </output>
2469 </layer>
2470 <layer id="179" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2471 <data auto_broadcast="numpy" />
2472 <input>
2473 <port id="0" precision="FP32">
2474 <dim>-1</dim>
2475 <dim>-1</dim>
2476 <dim>768</dim>
2477 </port>
2478 <port id="1" precision="FP32">
2479 <dim>1</dim>
2480 <dim>1</dim>
2481 <dim>768</dim>
2482 </port>
2483 </input>
2484 <output>
2485 <port id="2" precision="FP32">
2486 <dim>-1</dim>
2487 <dim>-1</dim>
2488 <dim>768</dim>
2489 </port>
2490 </output>
2491 </layer>
2492 <layer id="180" name="Constant_5371" type="Const" version="opset1">
2493 <data element_type="f32" shape="1, 1, 768" offset="46376528" size="3072" />
2494 <output>
2495 <port id="0" precision="FP32">
2496 <dim>1</dim>
2497 <dim>1</dim>
2498 <dim>768</dim>
2499 </port>
2500 </output>
2501 </layer>
2502 <layer id="181" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2503 <data auto_broadcast="numpy" />
2504 <input>
2505 <port id="0" precision="FP32">
2506 <dim>-1</dim>
2507 <dim>-1</dim>
2508 <dim>768</dim>
2509 </port>
2510 <port id="1" precision="FP32">
2511 <dim>1</dim>
2512 <dim>1</dim>
2513 <dim>768</dim>
2514 </port>
2515 </input>
2516 <output>
2517 <port id="2" precision="FP32" names="157,hidden_states.7">
2518 <dim>-1</dim>
2519 <dim>-1</dim>
2520 <dim>768</dim>
2521 </port>
2522 </output>
2523 </layer>
2524 <layer id="182" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
2525 <data element_type="f32" shape="1, 1, 768" offset="46379600" size="3072" />
2526 <output>
2527 <port id="0" precision="FP32">
2528 <dim>1</dim>
2529 <dim>1</dim>
2530 <dim>768</dim>
2531 </port>
2532 </output>
2533 </layer>
2534 <layer id="183" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
2535 <data auto_broadcast="numpy" />
2536 <input>
2537 <port id="0" precision="FP32">
2538 <dim>-1</dim>
2539 <dim>-1</dim>
2540 <dim>768</dim>
2541 </port>
2542 <port id="1" precision="FP32">
2543 <dim>1</dim>
2544 <dim>1</dim>
2545 <dim>768</dim>
2546 </port>
2547 </input>
2548 <output>
2549 <port id="2" precision="FP32">
2550 <dim>-1</dim>
2551 <dim>-1</dim>
2552 <dim>768</dim>
2553 </port>
2554 </output>
2555 </layer>
2556 <layer id="184" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
2557 <data element_type="f32" shape="" offset="46382672" size="4" />
2558 <output>
2559 <port id="0" precision="FP32" />
2560 </output>
2561 </layer>
2562 <layer id="185" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
2563 <data element_type="f32" shape="" offset="46382676" size="4" />
2564 <output>
2565 <port id="0" precision="FP32" />
2566 </output>
2567 </layer>
2568 <layer id="186" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
2569 <data element_type="f32" shape="" offset="46382672" size="4" />
2570 <output>
2571 <port id="0" precision="FP32" />
2572 </output>
2573 </layer>
2574 <layer id="187" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
2575 <data element_type="f32" shape="" offset="46382676" size="4" />
2576 <output>
2577 <port id="0" precision="FP32" />
2578 </output>
2579 </layer>
2580 <layer id="188" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
2581 <data levels="256" auto_broadcast="numpy" />
2582 <input>
2583 <port id="0" precision="FP32">
2584 <dim>-1</dim>
2585 <dim>-1</dim>
2586 <dim>768</dim>
2587 </port>
2588 <port id="1" precision="FP32" />
2589 <port id="2" precision="FP32" />
2590 <port id="3" precision="FP32" />
2591 <port id="4" precision="FP32" />
2592 </input>
2593 <output>
2594 <port id="5" precision="FP32">
2595 <dim>-1</dim>
2596 <dim>-1</dim>
2597 <dim>768</dim>
2598 </port>
2599 </output>
2600 </layer>
2601 <layer id="189" name="Constant_220518" type="Const" version="opset1">
2602 <data element_type="i8" shape="768, 768" offset="46382680" size="589824" />
2603 <output>
2604 <port id="0" precision="I8">
2605 <dim>768</dim>
2606 <dim>768</dim>
2607 </port>
2608 </output>
2609 </layer>
2610 <layer id="190" name="Convert_220519" type="Convert" version="opset1">
2611 <data destination_type="f32" />
2612 <input>
2613 <port id="0" precision="I8">
2614 <dim>768</dim>
2615 <dim>768</dim>
2616 </port>
2617 </input>
2618 <output>
2619 <port id="1" precision="FP32">
2620 <dim>768</dim>
2621 <dim>768</dim>
2622 </port>
2623 </output>
2624 </layer>
2625 <layer id="191" name="Constant_220520" type="Const" version="opset1">
2626 <data element_type="f32" shape="768, 1" offset="46972504" size="3072" />
2627 <output>
2628 <port id="0" precision="FP32">
2629 <dim>768</dim>
2630 <dim>1</dim>
2631 </port>
2632 </output>
2633 </layer>
2634 <layer id="192" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2635 <data auto_broadcast="numpy" />
2636 <input>
2637 <port id="0" precision="FP32">
2638 <dim>768</dim>
2639 <dim>768</dim>
2640 </port>
2641 <port id="1" precision="FP32">
2642 <dim>768</dim>
2643 <dim>1</dim>
2644 </port>
2645 </input>
2646 <output>
2647 <port id="2" precision="FP32">
2648 <dim>768</dim>
2649 <dim>768</dim>
2650 </port>
2651 </output>
2652 </layer>
2653 <layer id="193" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2654 <data transpose_a="false" transpose_b="true" />
2655 <input>
2656 <port id="0" precision="FP32">
2657 <dim>-1</dim>
2658 <dim>-1</dim>
2659 <dim>768</dim>
2660 </port>
2661 <port id="1" precision="FP32">
2662 <dim>768</dim>
2663 <dim>768</dim>
2664 </port>
2665 </input>
2666 <output>
2667 <port id="2" precision="FP32">
2668 <dim>-1</dim>
2669 <dim>-1</dim>
2670 <dim>768</dim>
2671 </port>
2672 </output>
2673 </layer>
2674 <layer id="194" name="Constant_5372" type="Const" version="opset1">
2675 <data element_type="f32" shape="1, 1, 768" offset="46975576" size="3072" />
2676 <output>
2677 <port id="0" precision="FP32">
2678 <dim>1</dim>
2679 <dim>1</dim>
2680 <dim>768</dim>
2681 </port>
2682 </output>
2683 </layer>
2684 <layer id="195" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2685 <data auto_broadcast="numpy" />
2686 <input>
2687 <port id="0" precision="FP32">
2688 <dim>-1</dim>
2689 <dim>-1</dim>
2690 <dim>768</dim>
2691 </port>
2692 <port id="1" precision="FP32">
2693 <dim>1</dim>
2694 <dim>1</dim>
2695 <dim>768</dim>
2696 </port>
2697 </input>
2698 <output>
2699 <port id="2" precision="FP32" names="170,x.13">
2700 <dim>-1</dim>
2701 <dim>-1</dim>
2702 <dim>768</dim>
2703 </port>
2704 </output>
2705 </layer>
2706 <layer id="196" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2707 <data element_type="f32" shape="" offset="46978648" size="4" />
2708 <output>
2709 <port id="0" precision="FP32" />
2710 </output>
2711 </layer>
2712 <layer id="197" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2713 <data element_type="f32" shape="" offset="46978652" size="4" />
2714 <output>
2715 <port id="0" precision="FP32" />
2716 </output>
2717 </layer>
2718 <layer id="198" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2719 <data element_type="f32" shape="" offset="46978648" size="4" />
2720 <output>
2721 <port id="0" precision="FP32" />
2722 </output>
2723 </layer>
2724 <layer id="199" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2725 <data element_type="f32" shape="" offset="46978652" size="4" />
2726 <output>
2727 <port id="0" precision="FP32" />
2728 </output>
2729 </layer>
2730 <layer id="200" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2731 <data levels="256" auto_broadcast="numpy" />
2732 <input>
2733 <port id="0" precision="FP32">
2734 <dim>-1</dim>
2735 <dim>-1</dim>
2736 <dim>768</dim>
2737 </port>
2738 <port id="1" precision="FP32" />
2739 <port id="2" precision="FP32" />
2740 <port id="3" precision="FP32" />
2741 <port id="4" precision="FP32" />
2742 </input>
2743 <output>
2744 <port id="5" precision="FP32">
2745 <dim>-1</dim>
2746 <dim>-1</dim>
2747 <dim>768</dim>
2748 </port>
2749 </output>
2750 </layer>
2751 <layer id="201" name="__module.roberta.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2752 <data element_type="i64" shape="4" offset="39811496" size="32" />
2753 <output>
2754 <port id="0" precision="I64">
2755 <dim>4</dim>
2756 </port>
2757 </output>
2758 </layer>
2759 <layer id="202" name="__module.roberta.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2760 <data special_zero="true" />
2761 <input>
2762 <port id="0" precision="FP32">
2763 <dim>-1</dim>
2764 <dim>-1</dim>
2765 <dim>768</dim>
2766 </port>
2767 <port id="1" precision="I64">
2768 <dim>4</dim>
2769 </port>
2770 </input>
2771 <output>
2772 <port id="2" precision="FP32" names="174,x.15">
2773 <dim>-1</dim>
2774 <dim>-1</dim>
2775 <dim>12</dim>
2776 <dim>64</dim>
2777 </port>
2778 </output>
2779 </layer>
2780 <layer id="203" name="Constant_452" type="Const" version="opset1">
2781 <data element_type="i64" shape="4" offset="39811528" size="32" />
2782 <output>
2783 <port id="0" precision="I64" names="175">
2784 <dim>4</dim>
2785 </port>
2786 </output>
2787 </layer>
2788 <layer id="204" name="__module.roberta.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2789 <input>
2790 <port id="0" precision="FP32">
2791 <dim>-1</dim>
2792 <dim>-1</dim>
2793 <dim>12</dim>
2794 <dim>64</dim>
2795 </port>
2796 <port id="1" precision="I64">
2797 <dim>4</dim>
2798 </port>
2799 </input>
2800 <output>
2801 <port id="2" precision="FP32" names="176">
2802 <dim>-1</dim>
2803 <dim>12</dim>
2804 <dim>-1</dim>
2805 <dim>64</dim>
2806 </port>
2807 </output>
2808 </layer>
2809 <layer id="205" name="Constant_220522" type="Const" version="opset1">
2810 <data element_type="i8" shape="768, 768" offset="46978656" size="589824" />
2811 <output>
2812 <port id="0" precision="I8">
2813 <dim>768</dim>
2814 <dim>768</dim>
2815 </port>
2816 </output>
2817 </layer>
2818 <layer id="206" name="Convert_220523" type="Convert" version="opset1">
2819 <data destination_type="f32" />
2820 <input>
2821 <port id="0" precision="I8">
2822 <dim>768</dim>
2823 <dim>768</dim>
2824 </port>
2825 </input>
2826 <output>
2827 <port id="1" precision="FP32">
2828 <dim>768</dim>
2829 <dim>768</dim>
2830 </port>
2831 </output>
2832 </layer>
2833 <layer id="207" name="Constant_220524" type="Const" version="opset1">
2834 <data element_type="f32" shape="768, 1" offset="47568480" size="3072" />
2835 <output>
2836 <port id="0" precision="FP32">
2837 <dim>768</dim>
2838 <dim>1</dim>
2839 </port>
2840 </output>
2841 </layer>
2842 <layer id="208" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
2843 <data auto_broadcast="numpy" />
2844 <input>
2845 <port id="0" precision="FP32">
2846 <dim>768</dim>
2847 <dim>768</dim>
2848 </port>
2849 <port id="1" precision="FP32">
2850 <dim>768</dim>
2851 <dim>1</dim>
2852 </port>
2853 </input>
2854 <output>
2855 <port id="2" precision="FP32">
2856 <dim>768</dim>
2857 <dim>768</dim>
2858 </port>
2859 </output>
2860 </layer>
2861 <layer id="209" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2862 <data transpose_a="false" transpose_b="true" />
2863 <input>
2864 <port id="0" precision="FP32">
2865 <dim>-1</dim>
2866 <dim>-1</dim>
2867 <dim>768</dim>
2868 </port>
2869 <port id="1" precision="FP32">
2870 <dim>768</dim>
2871 <dim>768</dim>
2872 </port>
2873 </input>
2874 <output>
2875 <port id="2" precision="FP32">
2876 <dim>-1</dim>
2877 <dim>-1</dim>
2878 <dim>768</dim>
2879 </port>
2880 </output>
2881 </layer>
2882 <layer id="210" name="Constant_5373" type="Const" version="opset1">
2883 <data element_type="f32" shape="1, 1, 768" offset="47571552" size="3072" />
2884 <output>
2885 <port id="0" precision="FP32">
2886 <dim>1</dim>
2887 <dim>1</dim>
2888 <dim>768</dim>
2889 </port>
2890 </output>
2891 </layer>
2892 <layer id="211" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2893 <data auto_broadcast="numpy" />
2894 <input>
2895 <port id="0" precision="FP32">
2896 <dim>-1</dim>
2897 <dim>-1</dim>
2898 <dim>768</dim>
2899 </port>
2900 <port id="1" precision="FP32">
2901 <dim>1</dim>
2902 <dim>1</dim>
2903 <dim>768</dim>
2904 </port>
2905 </input>
2906 <output>
2907 <port id="2" precision="FP32" names="179,x.17">
2908 <dim>-1</dim>
2909 <dim>-1</dim>
2910 <dim>768</dim>
2911 </port>
2912 </output>
2913 </layer>
2914 <layer id="212" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
2915 <data element_type="f32" shape="" offset="47574624" size="4" />
2916 <output>
2917 <port id="0" precision="FP32" />
2918 </output>
2919 </layer>
2920 <layer id="213" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
2921 <data element_type="f32" shape="" offset="47574628" size="4" />
2922 <output>
2923 <port id="0" precision="FP32" />
2924 </output>
2925 </layer>
2926 <layer id="214" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
2927 <data element_type="f32" shape="" offset="47574624" size="4" />
2928 <output>
2929 <port id="0" precision="FP32" />
2930 </output>
2931 </layer>
2932 <layer id="215" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
2933 <data element_type="f32" shape="" offset="47574628" size="4" />
2934 <output>
2935 <port id="0" precision="FP32" />
2936 </output>
2937 </layer>
2938 <layer id="216" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
2939 <data levels="256" auto_broadcast="numpy" />
2940 <input>
2941 <port id="0" precision="FP32">
2942 <dim>-1</dim>
2943 <dim>-1</dim>
2944 <dim>768</dim>
2945 </port>
2946 <port id="1" precision="FP32" />
2947 <port id="2" precision="FP32" />
2948 <port id="3" precision="FP32" />
2949 <port id="4" precision="FP32" />
2950 </input>
2951 <output>
2952 <port id="5" precision="FP32">
2953 <dim>-1</dim>
2954 <dim>-1</dim>
2955 <dim>768</dim>
2956 </port>
2957 </output>
2958 </layer>
2959 <layer id="217" name="__module.roberta.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2960 <data element_type="i64" shape="4" offset="39811496" size="32" />
2961 <output>
2962 <port id="0" precision="I64">
2963 <dim>4</dim>
2964 </port>
2965 </output>
2966 </layer>
2967 <layer id="218" name="__module.roberta.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2968 <data special_zero="true" />
2969 <input>
2970 <port id="0" precision="FP32">
2971 <dim>-1</dim>
2972 <dim>-1</dim>
2973 <dim>768</dim>
2974 </port>
2975 <port id="1" precision="I64">
2976 <dim>4</dim>
2977 </port>
2978 </input>
2979 <output>
2980 <port id="2" precision="FP32" names="183,x.19">
2981 <dim>-1</dim>
2982 <dim>-1</dim>
2983 <dim>12</dim>
2984 <dim>64</dim>
2985 </port>
2986 </output>
2987 </layer>
2988 <layer id="219" name="Constant_475" type="Const" version="opset1">
2989 <data element_type="i64" shape="4" offset="39811528" size="32" />
2990 <output>
2991 <port id="0" precision="I64" names="184">
2992 <dim>4</dim>
2993 </port>
2994 </output>
2995 </layer>
2996 <layer id="220" name="__module.roberta.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2997 <input>
2998 <port id="0" precision="FP32">
2999 <dim>-1</dim>
3000 <dim>-1</dim>
3001 <dim>12</dim>
3002 <dim>64</dim>
3003 </port>
3004 <port id="1" precision="I64">
3005 <dim>4</dim>
3006 </port>
3007 </input>
3008 <output>
3009 <port id="2" precision="FP32" names="185">
3010 <dim>-1</dim>
3011 <dim>12</dim>
3012 <dim>-1</dim>
3013 <dim>64</dim>
3014 </port>
3015 </output>
3016 </layer>
3017 <layer id="221" name="Constant_220526" type="Const" version="opset1">
3018 <data element_type="i8" shape="768, 768" offset="47574632" size="589824" />
3019 <output>
3020 <port id="0" precision="I8">
3021 <dim>768</dim>
3022 <dim>768</dim>
3023 </port>
3024 </output>
3025 </layer>
3026 <layer id="222" name="Convert_220527" type="Convert" version="opset1">
3027 <data destination_type="f32" />
3028 <input>
3029 <port id="0" precision="I8">
3030 <dim>768</dim>
3031 <dim>768</dim>
3032 </port>
3033 </input>
3034 <output>
3035 <port id="1" precision="FP32">
3036 <dim>768</dim>
3037 <dim>768</dim>
3038 </port>
3039 </output>
3040 </layer>
3041 <layer id="223" name="Constant_220528" type="Const" version="opset1">
3042 <data element_type="f32" shape="768, 1" offset="48164456" size="3072" />
3043 <output>
3044 <port id="0" precision="FP32">
3045 <dim>768</dim>
3046 <dim>1</dim>
3047 </port>
3048 </output>
3049 </layer>
3050 <layer id="224" name="__module.roberta.encoder.layer.1.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3051 <data auto_broadcast="numpy" />
3052 <input>
3053 <port id="0" precision="FP32">
3054 <dim>768</dim>
3055 <dim>768</dim>
3056 </port>
3057 <port id="1" precision="FP32">
3058 <dim>768</dim>
3059 <dim>1</dim>
3060 </port>
3061 </input>
3062 <output>
3063 <port id="2" precision="FP32">
3064 <dim>768</dim>
3065 <dim>768</dim>
3066 </port>
3067 </output>
3068 </layer>
3069 <layer id="225" name="__module.roberta.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
3070 <data transpose_a="false" transpose_b="true" />
3071 <input>
3072 <port id="0" precision="FP32">
3073 <dim>-1</dim>
3074 <dim>-1</dim>
3075 <dim>768</dim>
3076 </port>
3077 <port id="1" precision="FP32">
3078 <dim>768</dim>
3079 <dim>768</dim>
3080 </port>
3081 </input>
3082 <output>
3083 <port id="2" precision="FP32">
3084 <dim>-1</dim>
3085 <dim>-1</dim>
3086 <dim>768</dim>
3087 </port>
3088 </output>
3089 </layer>
3090 <layer id="226" name="Constant_5374" type="Const" version="opset1">
3091 <data element_type="f32" shape="1, 1, 768" offset="48167528" size="3072" />
3092 <output>
3093 <port id="0" precision="FP32">
3094 <dim>1</dim>
3095 <dim>1</dim>
3096 <dim>768</dim>
3097 </port>
3098 </output>
3099 </layer>
3100 <layer id="227" name="__module.roberta.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
3101 <data auto_broadcast="numpy" />
3102 <input>
3103 <port id="0" precision="FP32">
3104 <dim>-1</dim>
3105 <dim>-1</dim>
3106 <dim>768</dim>
3107 </port>
3108 <port id="1" precision="FP32">
3109 <dim>1</dim>
3110 <dim>1</dim>
3111 <dim>768</dim>
3112 </port>
3113 </input>
3114 <output>
3115 <port id="2" precision="FP32" names="188,x.21">
3116 <dim>-1</dim>
3117 <dim>-1</dim>
3118 <dim>768</dim>
3119 </port>
3120 </output>
3121 </layer>
3122 <layer id="228" name="__module.roberta.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
3123 <data element_type="i64" shape="4" offset="39811496" size="32" />
3124 <output>
3125 <port id="0" precision="I64">
3126 <dim>4</dim>
3127 </port>
3128 </output>
3129 </layer>
3130 <layer id="229" name="__module.roberta.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3131 <data special_zero="true" />
3132 <input>
3133 <port id="0" precision="FP32">
3134 <dim>-1</dim>
3135 <dim>-1</dim>
3136 <dim>768</dim>
3137 </port>
3138 <port id="1" precision="I64">
3139 <dim>4</dim>
3140 </port>
3141 </input>
3142 <output>
3143 <port id="2" precision="FP32" names="192,x.23">
3144 <dim>-1</dim>
3145 <dim>-1</dim>
3146 <dim>12</dim>
3147 <dim>64</dim>
3148 </port>
3149 </output>
3150 </layer>
3151 <layer id="230" name="Constant_498" type="Const" version="opset1">
3152 <data element_type="i64" shape="4" offset="39811528" size="32" />
3153 <output>
3154 <port id="0" precision="I64" names="193">
3155 <dim>4</dim>
3156 </port>
3157 </output>
3158 </layer>
3159 <layer id="231" name="__module.roberta.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3160 <input>
3161 <port id="0" precision="FP32">
3162 <dim>-1</dim>
3163 <dim>-1</dim>
3164 <dim>12</dim>
3165 <dim>64</dim>
3166 </port>
3167 <port id="1" precision="I64">
3168 <dim>4</dim>
3169 </port>
3170 </input>
3171 <output>
3172 <port id="2" precision="FP32" names="194">
3173 <dim>-1</dim>
3174 <dim>12</dim>
3175 <dim>-1</dim>
3176 <dim>64</dim>
3177 </port>
3178 </output>
3179 </layer>
3180 <layer id="232" name="__module.roberta.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3181 <data causal="false" />
3182 <input>
3183 <port id="0" precision="FP32">
3184 <dim>-1</dim>
3185 <dim>12</dim>
3186 <dim>-1</dim>
3187 <dim>64</dim>
3188 </port>
3189 <port id="1" precision="FP32">
3190 <dim>-1</dim>
3191 <dim>12</dim>
3192 <dim>-1</dim>
3193 <dim>64</dim>
3194 </port>
3195 <port id="2" precision="FP32">
3196 <dim>-1</dim>
3197 <dim>12</dim>
3198 <dim>-1</dim>
3199 <dim>64</dim>
3200 </port>
3201 <port id="3" precision="FP32">
3202 <dim>-1</dim>
3203 <dim>1</dim>
3204 <dim>-1</dim>
3205 <dim>-1</dim>
3206 </port>
3207 </input>
3208 <output>
3209 <port id="4" precision="FP32" names="195,attn_output.5">
3210 <dim>-1</dim>
3211 <dim>12</dim>
3212 <dim>-1</dim>
3213 <dim>64</dim>
3214 </port>
3215 </output>
3216 </layer>
3217 <layer id="233" name="__module.roberta.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3218 <data element_type="i32" shape="4" offset="41003536" size="16" />
3219 <output>
3220 <port id="0" precision="I32">
3221 <dim>4</dim>
3222 </port>
3223 </output>
3224 </layer>
3225 <layer id="234" name="__module.roberta.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3226 <input>
3227 <port id="0" precision="FP32">
3228 <dim>-1</dim>
3229 <dim>12</dim>
3230 <dim>-1</dim>
3231 <dim>64</dim>
3232 </port>
3233 <port id="1" precision="I32">
3234 <dim>4</dim>
3235 </port>
3236 </input>
3237 <output>
3238 <port id="2" precision="FP32" names="196,attn_output.7">
3239 <dim>-1</dim>
3240 <dim>-1</dim>
3241 <dim>12</dim>
3242 <dim>64</dim>
3243 </port>
3244 </output>
3245 </layer>
3246 <layer id="235" name="Constant_5495" type="Const" version="opset1">
3247 <data element_type="i64" shape="3" offset="41003552" size="24" />
3248 <output>
3249 <port id="0" precision="I64">
3250 <dim>3</dim>
3251 </port>
3252 </output>
3253 </layer>
3254 <layer id="236" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3255 <data special_zero="true" />
3256 <input>
3257 <port id="0" precision="FP32">
3258 <dim>-1</dim>
3259 <dim>-1</dim>
3260 <dim>12</dim>
3261 <dim>64</dim>
3262 </port>
3263 <port id="1" precision="I64">
3264 <dim>3</dim>
3265 </port>
3266 </input>
3267 <output>
3268 <port id="2" precision="FP32" names="198">
3269 <dim>-1</dim>
3270 <dim>-1</dim>
3271 <dim>768</dim>
3272 </port>
3273 </output>
3274 </layer>
3275 <layer id="237" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3276 <data element_type="f32" shape="1, 1, 768" offset="48170600" size="3072" />
3277 <output>
3278 <port id="0" precision="FP32">
3279 <dim>1</dim>
3280 <dim>1</dim>
3281 <dim>768</dim>
3282 </port>
3283 </output>
3284 </layer>
3285 <layer id="238" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3286 <data auto_broadcast="numpy" />
3287 <input>
3288 <port id="0" precision="FP32">
3289 <dim>-1</dim>
3290 <dim>-1</dim>
3291 <dim>768</dim>
3292 </port>
3293 <port id="1" precision="FP32">
3294 <dim>1</dim>
3295 <dim>1</dim>
3296 <dim>768</dim>
3297 </port>
3298 </input>
3299 <output>
3300 <port id="2" precision="FP32">
3301 <dim>-1</dim>
3302 <dim>-1</dim>
3303 <dim>768</dim>
3304 </port>
3305 </output>
3306 </layer>
3307 <layer id="239" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3308 <data element_type="f32" shape="" offset="48173672" size="4" />
3309 <output>
3310 <port id="0" precision="FP32" />
3311 </output>
3312 </layer>
3313 <layer id="240" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3314 <data element_type="f32" shape="" offset="48173676" size="4" />
3315 <output>
3316 <port id="0" precision="FP32" />
3317 </output>
3318 </layer>
3319 <layer id="241" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3320 <data element_type="f32" shape="" offset="48173672" size="4" />
3321 <output>
3322 <port id="0" precision="FP32" />
3323 </output>
3324 </layer>
3325 <layer id="242" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3326 <data element_type="f32" shape="" offset="48173676" size="4" />
3327 <output>
3328 <port id="0" precision="FP32" />
3329 </output>
3330 </layer>
3331 <layer id="243" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3332 <data levels="256" auto_broadcast="numpy" />
3333 <input>
3334 <port id="0" precision="FP32">
3335 <dim>-1</dim>
3336 <dim>-1</dim>
3337 <dim>768</dim>
3338 </port>
3339 <port id="1" precision="FP32" />
3340 <port id="2" precision="FP32" />
3341 <port id="3" precision="FP32" />
3342 <port id="4" precision="FP32" />
3343 </input>
3344 <output>
3345 <port id="5" precision="FP32">
3346 <dim>-1</dim>
3347 <dim>-1</dim>
3348 <dim>768</dim>
3349 </port>
3350 </output>
3351 </layer>
3352 <layer id="244" name="Constant_220530" type="Const" version="opset1">
3353 <data element_type="i8" shape="768, 768" offset="48173680" size="589824" />
3354 <output>
3355 <port id="0" precision="I8">
3356 <dim>768</dim>
3357 <dim>768</dim>
3358 </port>
3359 </output>
3360 </layer>
3361 <layer id="245" name="Convert_220531" type="Convert" version="opset1">
3362 <data destination_type="f32" />
3363 <input>
3364 <port id="0" precision="I8">
3365 <dim>768</dim>
3366 <dim>768</dim>
3367 </port>
3368 </input>
3369 <output>
3370 <port id="1" precision="FP32">
3371 <dim>768</dim>
3372 <dim>768</dim>
3373 </port>
3374 </output>
3375 </layer>
3376 <layer id="246" name="Constant_220532" type="Const" version="opset1">
3377 <data element_type="f32" shape="768, 1" offset="48763504" size="3072" />
3378 <output>
3379 <port id="0" precision="FP32">
3380 <dim>768</dim>
3381 <dim>1</dim>
3382 </port>
3383 </output>
3384 </layer>
3385 <layer id="247" name="__module.roberta.encoder.layer.1.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3386 <data auto_broadcast="numpy" />
3387 <input>
3388 <port id="0" precision="FP32">
3389 <dim>768</dim>
3390 <dim>768</dim>
3391 </port>
3392 <port id="1" precision="FP32">
3393 <dim>768</dim>
3394 <dim>1</dim>
3395 </port>
3396 </input>
3397 <output>
3398 <port id="2" precision="FP32">
3399 <dim>768</dim>
3400 <dim>768</dim>
3401 </port>
3402 </output>
3403 </layer>
3404 <layer id="248" name="__module.roberta.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3405 <data transpose_a="false" transpose_b="true" />
3406 <input>
3407 <port id="0" precision="FP32">
3408 <dim>-1</dim>
3409 <dim>-1</dim>
3410 <dim>768</dim>
3411 </port>
3412 <port id="1" precision="FP32">
3413 <dim>768</dim>
3414 <dim>768</dim>
3415 </port>
3416 </input>
3417 <output>
3418 <port id="2" precision="FP32">
3419 <dim>-1</dim>
3420 <dim>-1</dim>
3421 <dim>768</dim>
3422 </port>
3423 </output>
3424 </layer>
3425 <layer id="249" name="Constant_5375" type="Const" version="opset1">
3426 <data element_type="f32" shape="1, 1, 768" offset="48766576" size="3072" />
3427 <output>
3428 <port id="0" precision="FP32">
3429 <dim>1</dim>
3430 <dim>1</dim>
3431 <dim>768</dim>
3432 </port>
3433 </output>
3434 </layer>
3435 <layer id="250" name="__module.roberta.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3436 <data auto_broadcast="numpy" />
3437 <input>
3438 <port id="0" precision="FP32">
3439 <dim>-1</dim>
3440 <dim>-1</dim>
3441 <dim>768</dim>
3442 </port>
3443 <port id="1" precision="FP32">
3444 <dim>1</dim>
3445 <dim>1</dim>
3446 <dim>768</dim>
3447 </port>
3448 </input>
3449 <output>
3450 <port id="2" precision="FP32" names="203,input.7">
3451 <dim>-1</dim>
3452 <dim>-1</dim>
3453 <dim>768</dim>
3454 </port>
3455 </output>
3456 </layer>
3457 <layer id="251" name="__module.roberta.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
3458 <data auto_broadcast="numpy" />
3459 <input>
3460 <port id="0" precision="FP32">
3461 <dim>-1</dim>
3462 <dim>-1</dim>
3463 <dim>768</dim>
3464 </port>
3465 <port id="1" precision="FP32">
3466 <dim>-1</dim>
3467 <dim>-1</dim>
3468 <dim>768</dim>
3469 </port>
3470 </input>
3471 <output>
3472 <port id="2" precision="FP32" names="205">
3473 <dim>-1</dim>
3474 <dim>-1</dim>
3475 <dim>768</dim>
3476 </port>
3477 </output>
3478 </layer>
3479 <layer id="252" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3480 <data element_type="i32" shape="1" offset="39206292" size="4" />
3481 <output>
3482 <port id="0" precision="I32">
3483 <dim>1</dim>
3484 </port>
3485 </output>
3486 </layer>
3487 <layer id="253" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3488 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3489 <input>
3490 <port id="0" precision="FP32">
3491 <dim>-1</dim>
3492 <dim>-1</dim>
3493 <dim>768</dim>
3494 </port>
3495 <port id="1" precision="I32">
3496 <dim>1</dim>
3497 </port>
3498 </input>
3499 <output>
3500 <port id="2" precision="FP32">
3501 <dim>-1</dim>
3502 <dim>-1</dim>
3503 <dim>768</dim>
3504 </port>
3505 </output>
3506 </layer>
3507 <layer id="254" name="Constant_5376" type="Const" version="opset1">
3508 <data element_type="f32" shape="1, 1, 768" offset="48769648" size="3072" />
3509 <output>
3510 <port id="0" precision="FP32">
3511 <dim>1</dim>
3512 <dim>1</dim>
3513 <dim>768</dim>
3514 </port>
3515 </output>
3516 </layer>
3517 <layer id="255" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3518 <data auto_broadcast="numpy" />
3519 <input>
3520 <port id="0" precision="FP32">
3521 <dim>-1</dim>
3522 <dim>-1</dim>
3523 <dim>768</dim>
3524 </port>
3525 <port id="1" precision="FP32">
3526 <dim>1</dim>
3527 <dim>1</dim>
3528 <dim>768</dim>
3529 </port>
3530 </input>
3531 <output>
3532 <port id="2" precision="FP32">
3533 <dim>-1</dim>
3534 <dim>-1</dim>
3535 <dim>768</dim>
3536 </port>
3537 </output>
3538 </layer>
3539 <layer id="256" name="Constant_5377" type="Const" version="opset1">
3540 <data element_type="f32" shape="1, 1, 768" offset="48772720" size="3072" />
3541 <output>
3542 <port id="0" precision="FP32">
3543 <dim>1</dim>
3544 <dim>1</dim>
3545 <dim>768</dim>
3546 </port>
3547 </output>
3548 </layer>
3549 <layer id="257" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3550 <data auto_broadcast="numpy" />
3551 <input>
3552 <port id="0" precision="FP32">
3553 <dim>-1</dim>
3554 <dim>-1</dim>
3555 <dim>768</dim>
3556 </port>
3557 <port id="1" precision="FP32">
3558 <dim>1</dim>
3559 <dim>1</dim>
3560 <dim>768</dim>
3561 </port>
3562 </input>
3563 <output>
3564 <port id="2" precision="FP32" names="209,input_tensor.3">
3565 <dim>-1</dim>
3566 <dim>-1</dim>
3567 <dim>768</dim>
3568 </port>
3569 </output>
3570 </layer>
3571 <layer id="258" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3572 <data element_type="f32" shape="1, 1, 768" offset="48775792" size="3072" />
3573 <output>
3574 <port id="0" precision="FP32">
3575 <dim>1</dim>
3576 <dim>1</dim>
3577 <dim>768</dim>
3578 </port>
3579 </output>
3580 </layer>
3581 <layer id="259" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3582 <data auto_broadcast="numpy" />
3583 <input>
3584 <port id="0" precision="FP32">
3585 <dim>-1</dim>
3586 <dim>-1</dim>
3587 <dim>768</dim>
3588 </port>
3589 <port id="1" precision="FP32">
3590 <dim>1</dim>
3591 <dim>1</dim>
3592 <dim>768</dim>
3593 </port>
3594 </input>
3595 <output>
3596 <port id="2" precision="FP32">
3597 <dim>-1</dim>
3598 <dim>-1</dim>
3599 <dim>768</dim>
3600 </port>
3601 </output>
3602 </layer>
3603 <layer id="260" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3604 <data element_type="f32" shape="" offset="48778864" size="4" />
3605 <output>
3606 <port id="0" precision="FP32" />
3607 </output>
3608 </layer>
3609 <layer id="261" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3610 <data element_type="f32" shape="" offset="48778868" size="4" />
3611 <output>
3612 <port id="0" precision="FP32" />
3613 </output>
3614 </layer>
3615 <layer id="262" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3616 <data element_type="f32" shape="" offset="48778864" size="4" />
3617 <output>
3618 <port id="0" precision="FP32" />
3619 </output>
3620 </layer>
3621 <layer id="263" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3622 <data element_type="f32" shape="" offset="48778868" size="4" />
3623 <output>
3624 <port id="0" precision="FP32" />
3625 </output>
3626 </layer>
3627 <layer id="264" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3628 <data levels="256" auto_broadcast="numpy" />
3629 <input>
3630 <port id="0" precision="FP32">
3631 <dim>-1</dim>
3632 <dim>-1</dim>
3633 <dim>768</dim>
3634 </port>
3635 <port id="1" precision="FP32" />
3636 <port id="2" precision="FP32" />
3637 <port id="3" precision="FP32" />
3638 <port id="4" precision="FP32" />
3639 </input>
3640 <output>
3641 <port id="5" precision="FP32">
3642 <dim>-1</dim>
3643 <dim>-1</dim>
3644 <dim>768</dim>
3645 </port>
3646 </output>
3647 </layer>
3648 <layer id="265" name="Constant_220534" type="Const" version="opset1">
3649 <data element_type="i8" shape="3072, 768" offset="48778872" size="2359296" />
3650 <output>
3651 <port id="0" precision="I8">
3652 <dim>3072</dim>
3653 <dim>768</dim>
3654 </port>
3655 </output>
3656 </layer>
3657 <layer id="266" name="Convert_220535" type="Convert" version="opset1">
3658 <data destination_type="f32" />
3659 <input>
3660 <port id="0" precision="I8">
3661 <dim>3072</dim>
3662 <dim>768</dim>
3663 </port>
3664 </input>
3665 <output>
3666 <port id="1" precision="FP32">
3667 <dim>3072</dim>
3668 <dim>768</dim>
3669 </port>
3670 </output>
3671 </layer>
3672 <layer id="267" name="Constant_220536" type="Const" version="opset1">
3673 <data element_type="f32" shape="3072, 1" offset="51138168" size="12288" />
3674 <output>
3675 <port id="0" precision="FP32">
3676 <dim>3072</dim>
3677 <dim>1</dim>
3678 </port>
3679 </output>
3680 </layer>
3681 <layer id="268" name="__module.roberta.encoder.layer.1.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3682 <data auto_broadcast="numpy" />
3683 <input>
3684 <port id="0" precision="FP32">
3685 <dim>3072</dim>
3686 <dim>768</dim>
3687 </port>
3688 <port id="1" precision="FP32">
3689 <dim>3072</dim>
3690 <dim>1</dim>
3691 </port>
3692 </input>
3693 <output>
3694 <port id="2" precision="FP32">
3695 <dim>3072</dim>
3696 <dim>768</dim>
3697 </port>
3698 </output>
3699 </layer>
3700 <layer id="269" name="__module.roberta.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3701 <data transpose_a="false" transpose_b="true" />
3702 <input>
3703 <port id="0" precision="FP32">
3704 <dim>-1</dim>
3705 <dim>-1</dim>
3706 <dim>768</dim>
3707 </port>
3708 <port id="1" precision="FP32">
3709 <dim>3072</dim>
3710 <dim>768</dim>
3711 </port>
3712 </input>
3713 <output>
3714 <port id="2" precision="FP32">
3715 <dim>-1</dim>
3716 <dim>-1</dim>
3717 <dim>3072</dim>
3718 </port>
3719 </output>
3720 </layer>
3721 <layer id="270" name="Constant_5378" type="Const" version="opset1">
3722 <data element_type="f32" shape="1, 1, 3072" offset="51150456" size="12288" />
3723 <output>
3724 <port id="0" precision="FP32">
3725 <dim>1</dim>
3726 <dim>1</dim>
3727 <dim>3072</dim>
3728 </port>
3729 </output>
3730 </layer>
3731 <layer id="271" name="__module.roberta.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3732 <data auto_broadcast="numpy" />
3733 <input>
3734 <port id="0" precision="FP32">
3735 <dim>-1</dim>
3736 <dim>-1</dim>
3737 <dim>3072</dim>
3738 </port>
3739 <port id="1" precision="FP32">
3740 <dim>1</dim>
3741 <dim>1</dim>
3742 <dim>3072</dim>
3743 </port>
3744 </input>
3745 <output>
3746 <port id="2" precision="FP32" names="213">
3747 <dim>-1</dim>
3748 <dim>-1</dim>
3749 <dim>3072</dim>
3750 </port>
3751 </output>
3752 </layer>
3753 <layer id="272" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3754 <data approximation_mode="ERF" />
3755 <input>
3756 <port id="0" precision="FP32">
3757 <dim>-1</dim>
3758 <dim>-1</dim>
3759 <dim>3072</dim>
3760 </port>
3761 </input>
3762 <output>
3763 <port id="1" precision="FP32" names="214">
3764 <dim>-1</dim>
3765 <dim>-1</dim>
3766 <dim>3072</dim>
3767 </port>
3768 </output>
3769 </layer>
3770 <layer id="273" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
3771 <data element_type="f32" shape="1, 1, 3072" offset="51162744" size="12288" />
3772 <output>
3773 <port id="0" precision="FP32">
3774 <dim>1</dim>
3775 <dim>1</dim>
3776 <dim>3072</dim>
3777 </port>
3778 </output>
3779 </layer>
3780 <layer id="274" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
3781 <data auto_broadcast="numpy" />
3782 <input>
3783 <port id="0" precision="FP32">
3784 <dim>-1</dim>
3785 <dim>-1</dim>
3786 <dim>3072</dim>
3787 </port>
3788 <port id="1" precision="FP32">
3789 <dim>1</dim>
3790 <dim>1</dim>
3791 <dim>3072</dim>
3792 </port>
3793 </input>
3794 <output>
3795 <port id="2" precision="FP32">
3796 <dim>-1</dim>
3797 <dim>-1</dim>
3798 <dim>3072</dim>
3799 </port>
3800 </output>
3801 </layer>
3802 <layer id="275" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
3803 <data element_type="f32" shape="" offset="51175032" size="4" />
3804 <output>
3805 <port id="0" precision="FP32" />
3806 </output>
3807 </layer>
3808 <layer id="276" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
3809 <data element_type="f32" shape="" offset="51175036" size="4" />
3810 <output>
3811 <port id="0" precision="FP32" />
3812 </output>
3813 </layer>
3814 <layer id="277" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
3815 <data element_type="f32" shape="" offset="51175032" size="4" />
3816 <output>
3817 <port id="0" precision="FP32" />
3818 </output>
3819 </layer>
3820 <layer id="278" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
3821 <data element_type="f32" shape="" offset="51175036" size="4" />
3822 <output>
3823 <port id="0" precision="FP32" />
3824 </output>
3825 </layer>
3826 <layer id="279" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
3827 <data levels="256" auto_broadcast="numpy" />
3828 <input>
3829 <port id="0" precision="FP32">
3830 <dim>-1</dim>
3831 <dim>-1</dim>
3832 <dim>3072</dim>
3833 </port>
3834 <port id="1" precision="FP32" />
3835 <port id="2" precision="FP32" />
3836 <port id="3" precision="FP32" />
3837 <port id="4" precision="FP32" />
3838 </input>
3839 <output>
3840 <port id="5" precision="FP32">
3841 <dim>-1</dim>
3842 <dim>-1</dim>
3843 <dim>3072</dim>
3844 </port>
3845 </output>
3846 </layer>
3847 <layer id="280" name="Constant_220538" type="Const" version="opset1">
3848 <data element_type="i8" shape="768, 3072" offset="51175040" size="2359296" />
3849 <output>
3850 <port id="0" precision="I8">
3851 <dim>768</dim>
3852 <dim>3072</dim>
3853 </port>
3854 </output>
3855 </layer>
3856 <layer id="281" name="Convert_220539" type="Convert" version="opset1">
3857 <data destination_type="f32" />
3858 <input>
3859 <port id="0" precision="I8">
3860 <dim>768</dim>
3861 <dim>3072</dim>
3862 </port>
3863 </input>
3864 <output>
3865 <port id="1" precision="FP32">
3866 <dim>768</dim>
3867 <dim>3072</dim>
3868 </port>
3869 </output>
3870 </layer>
3871 <layer id="282" name="Constant_220540" type="Const" version="opset1">
3872 <data element_type="f32" shape="768, 1" offset="53534336" size="3072" />
3873 <output>
3874 <port id="0" precision="FP32">
3875 <dim>768</dim>
3876 <dim>1</dim>
3877 </port>
3878 </output>
3879 </layer>
3880 <layer id="283" name="__module.roberta.encoder.layer.1.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
3881 <data auto_broadcast="numpy" />
3882 <input>
3883 <port id="0" precision="FP32">
3884 <dim>768</dim>
3885 <dim>3072</dim>
3886 </port>
3887 <port id="1" precision="FP32">
3888 <dim>768</dim>
3889 <dim>1</dim>
3890 </port>
3891 </input>
3892 <output>
3893 <port id="2" precision="FP32">
3894 <dim>768</dim>
3895 <dim>3072</dim>
3896 </port>
3897 </output>
3898 </layer>
3899 <layer id="284" name="__module.roberta.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3900 <data transpose_a="false" transpose_b="true" />
3901 <input>
3902 <port id="0" precision="FP32">
3903 <dim>-1</dim>
3904 <dim>-1</dim>
3905 <dim>3072</dim>
3906 </port>
3907 <port id="1" precision="FP32">
3908 <dim>768</dim>
3909 <dim>3072</dim>
3910 </port>
3911 </input>
3912 <output>
3913 <port id="2" precision="FP32">
3914 <dim>-1</dim>
3915 <dim>-1</dim>
3916 <dim>768</dim>
3917 </port>
3918 </output>
3919 </layer>
3920 <layer id="285" name="Constant_5379" type="Const" version="opset1">
3921 <data element_type="f32" shape="1, 1, 768" offset="53537408" size="3072" />
3922 <output>
3923 <port id="0" precision="FP32">
3924 <dim>1</dim>
3925 <dim>1</dim>
3926 <dim>768</dim>
3927 </port>
3928 </output>
3929 </layer>
3930 <layer id="286" name="__module.roberta.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
3931 <data auto_broadcast="numpy" />
3932 <input>
3933 <port id="0" precision="FP32">
3934 <dim>-1</dim>
3935 <dim>-1</dim>
3936 <dim>768</dim>
3937 </port>
3938 <port id="1" precision="FP32">
3939 <dim>1</dim>
3940 <dim>1</dim>
3941 <dim>768</dim>
3942 </port>
3943 </input>
3944 <output>
3945 <port id="2" precision="FP32" names="219,input.9">
3946 <dim>-1</dim>
3947 <dim>-1</dim>
3948 <dim>768</dim>
3949 </port>
3950 </output>
3951 </layer>
3952 <layer id="287" name="__module.roberta.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
3953 <data auto_broadcast="numpy" />
3954 <input>
3955 <port id="0" precision="FP32">
3956 <dim>-1</dim>
3957 <dim>-1</dim>
3958 <dim>768</dim>
3959 </port>
3960 <port id="1" precision="FP32">
3961 <dim>-1</dim>
3962 <dim>-1</dim>
3963 <dim>768</dim>
3964 </port>
3965 </input>
3966 <output>
3967 <port id="2" precision="FP32" names="221">
3968 <dim>-1</dim>
3969 <dim>-1</dim>
3970 <dim>768</dim>
3971 </port>
3972 </output>
3973 </layer>
3974 <layer id="288" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3975 <data element_type="i32" shape="1" offset="39206292" size="4" />
3976 <output>
3977 <port id="0" precision="I32">
3978 <dim>1</dim>
3979 </port>
3980 </output>
3981 </layer>
3982 <layer id="289" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3983 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3984 <input>
3985 <port id="0" precision="FP32">
3986 <dim>-1</dim>
3987 <dim>-1</dim>
3988 <dim>768</dim>
3989 </port>
3990 <port id="1" precision="I32">
3991 <dim>1</dim>
3992 </port>
3993 </input>
3994 <output>
3995 <port id="2" precision="FP32">
3996 <dim>-1</dim>
3997 <dim>-1</dim>
3998 <dim>768</dim>
3999 </port>
4000 </output>
4001 </layer>
4002 <layer id="290" name="Constant_5380" type="Const" version="opset1">
4003 <data element_type="f32" shape="1, 1, 768" offset="53540480" size="3072" />
4004 <output>
4005 <port id="0" precision="FP32">
4006 <dim>1</dim>
4007 <dim>1</dim>
4008 <dim>768</dim>
4009 </port>
4010 </output>
4011 </layer>
4012 <layer id="291" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4013 <data auto_broadcast="numpy" />
4014 <input>
4015 <port id="0" precision="FP32">
4016 <dim>-1</dim>
4017 <dim>-1</dim>
4018 <dim>768</dim>
4019 </port>
4020 <port id="1" precision="FP32">
4021 <dim>1</dim>
4022 <dim>1</dim>
4023 <dim>768</dim>
4024 </port>
4025 </input>
4026 <output>
4027 <port id="2" precision="FP32">
4028 <dim>-1</dim>
4029 <dim>-1</dim>
4030 <dim>768</dim>
4031 </port>
4032 </output>
4033 </layer>
4034 <layer id="292" name="Constant_5381" type="Const" version="opset1">
4035 <data element_type="f32" shape="1, 1, 768" offset="53543552" size="3072" />
4036 <output>
4037 <port id="0" precision="FP32">
4038 <dim>1</dim>
4039 <dim>1</dim>
4040 <dim>768</dim>
4041 </port>
4042 </output>
4043 </layer>
4044 <layer id="293" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4045 <data auto_broadcast="numpy" />
4046 <input>
4047 <port id="0" precision="FP32">
4048 <dim>-1</dim>
4049 <dim>-1</dim>
4050 <dim>768</dim>
4051 </port>
4052 <port id="1" precision="FP32">
4053 <dim>1</dim>
4054 <dim>1</dim>
4055 <dim>768</dim>
4056 </port>
4057 </input>
4058 <output>
4059 <port id="2" precision="FP32" names="225,hidden_states.13">
4060 <dim>-1</dim>
4061 <dim>-1</dim>
4062 <dim>768</dim>
4063 </port>
4064 </output>
4065 </layer>
4066 <layer id="294" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4067 <data element_type="f32" shape="1, 1, 768" offset="53546624" size="3072" />
4068 <output>
4069 <port id="0" precision="FP32">
4070 <dim>1</dim>
4071 <dim>1</dim>
4072 <dim>768</dim>
4073 </port>
4074 </output>
4075 </layer>
4076 <layer id="295" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4077 <data auto_broadcast="numpy" />
4078 <input>
4079 <port id="0" precision="FP32">
4080 <dim>-1</dim>
4081 <dim>-1</dim>
4082 <dim>768</dim>
4083 </port>
4084 <port id="1" precision="FP32">
4085 <dim>1</dim>
4086 <dim>1</dim>
4087 <dim>768</dim>
4088 </port>
4089 </input>
4090 <output>
4091 <port id="2" precision="FP32">
4092 <dim>-1</dim>
4093 <dim>-1</dim>
4094 <dim>768</dim>
4095 </port>
4096 </output>
4097 </layer>
4098 <layer id="296" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
4099 <data element_type="f32" shape="" offset="53549696" size="4" />
4100 <output>
4101 <port id="0" precision="FP32" />
4102 </output>
4103 </layer>
4104 <layer id="297" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
4105 <data element_type="f32" shape="" offset="53549700" size="4" />
4106 <output>
4107 <port id="0" precision="FP32" />
4108 </output>
4109 </layer>
4110 <layer id="298" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
4111 <data element_type="f32" shape="" offset="53549696" size="4" />
4112 <output>
4113 <port id="0" precision="FP32" />
4114 </output>
4115 </layer>
4116 <layer id="299" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
4117 <data element_type="f32" shape="" offset="53549700" size="4" />
4118 <output>
4119 <port id="0" precision="FP32" />
4120 </output>
4121 </layer>
4122 <layer id="300" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
4123 <data levels="256" auto_broadcast="numpy" />
4124 <input>
4125 <port id="0" precision="FP32">
4126 <dim>-1</dim>
4127 <dim>-1</dim>
4128 <dim>768</dim>
4129 </port>
4130 <port id="1" precision="FP32" />
4131 <port id="2" precision="FP32" />
4132 <port id="3" precision="FP32" />
4133 <port id="4" precision="FP32" />
4134 </input>
4135 <output>
4136 <port id="5" precision="FP32">
4137 <dim>-1</dim>
4138 <dim>-1</dim>
4139 <dim>768</dim>
4140 </port>
4141 </output>
4142 </layer>
4143 <layer id="301" name="Constant_220542" type="Const" version="opset1">
4144 <data element_type="i8" shape="768, 768" offset="53549704" size="589824" />
4145 <output>
4146 <port id="0" precision="I8">
4147 <dim>768</dim>
4148 <dim>768</dim>
4149 </port>
4150 </output>
4151 </layer>
4152 <layer id="302" name="Convert_220543" type="Convert" version="opset1">
4153 <data destination_type="f32" />
4154 <input>
4155 <port id="0" precision="I8">
4156 <dim>768</dim>
4157 <dim>768</dim>
4158 </port>
4159 </input>
4160 <output>
4161 <port id="1" precision="FP32">
4162 <dim>768</dim>
4163 <dim>768</dim>
4164 </port>
4165 </output>
4166 </layer>
4167 <layer id="303" name="Constant_220544" type="Const" version="opset1">
4168 <data element_type="f32" shape="768, 1" offset="54139528" size="3072" />
4169 <output>
4170 <port id="0" precision="FP32">
4171 <dim>768</dim>
4172 <dim>1</dim>
4173 </port>
4174 </output>
4175 </layer>
4176 <layer id="304" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4177 <data auto_broadcast="numpy" />
4178 <input>
4179 <port id="0" precision="FP32">
4180 <dim>768</dim>
4181 <dim>768</dim>
4182 </port>
4183 <port id="1" precision="FP32">
4184 <dim>768</dim>
4185 <dim>1</dim>
4186 </port>
4187 </input>
4188 <output>
4189 <port id="2" precision="FP32">
4190 <dim>768</dim>
4191 <dim>768</dim>
4192 </port>
4193 </output>
4194 </layer>
4195 <layer id="305" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
4196 <data transpose_a="false" transpose_b="true" />
4197 <input>
4198 <port id="0" precision="FP32">
4199 <dim>-1</dim>
4200 <dim>-1</dim>
4201 <dim>768</dim>
4202 </port>
4203 <port id="1" precision="FP32">
4204 <dim>768</dim>
4205 <dim>768</dim>
4206 </port>
4207 </input>
4208 <output>
4209 <port id="2" precision="FP32">
4210 <dim>-1</dim>
4211 <dim>-1</dim>
4212 <dim>768</dim>
4213 </port>
4214 </output>
4215 </layer>
4216 <layer id="306" name="Constant_5382" type="Const" version="opset1">
4217 <data element_type="f32" shape="1, 1, 768" offset="54142600" size="3072" />
4218 <output>
4219 <port id="0" precision="FP32">
4220 <dim>1</dim>
4221 <dim>1</dim>
4222 <dim>768</dim>
4223 </port>
4224 </output>
4225 </layer>
4226 <layer id="307" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
4227 <data auto_broadcast="numpy" />
4228 <input>
4229 <port id="0" precision="FP32">
4230 <dim>-1</dim>
4231 <dim>-1</dim>
4232 <dim>768</dim>
4233 </port>
4234 <port id="1" precision="FP32">
4235 <dim>1</dim>
4236 <dim>1</dim>
4237 <dim>768</dim>
4238 </port>
4239 </input>
4240 <output>
4241 <port id="2" precision="FP32" names="238,x.25">
4242 <dim>-1</dim>
4243 <dim>-1</dim>
4244 <dim>768</dim>
4245 </port>
4246 </output>
4247 </layer>
4248 <layer id="308" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4249 <data element_type="f32" shape="" offset="54145672" size="4" />
4250 <output>
4251 <port id="0" precision="FP32" />
4252 </output>
4253 </layer>
4254 <layer id="309" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4255 <data element_type="f32" shape="" offset="54145676" size="4" />
4256 <output>
4257 <port id="0" precision="FP32" />
4258 </output>
4259 </layer>
4260 <layer id="310" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4261 <data element_type="f32" shape="" offset="54145672" size="4" />
4262 <output>
4263 <port id="0" precision="FP32" />
4264 </output>
4265 </layer>
4266 <layer id="311" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4267 <data element_type="f32" shape="" offset="54145676" size="4" />
4268 <output>
4269 <port id="0" precision="FP32" />
4270 </output>
4271 </layer>
4272 <layer id="312" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4273 <data levels="256" auto_broadcast="numpy" />
4274 <input>
4275 <port id="0" precision="FP32">
4276 <dim>-1</dim>
4277 <dim>-1</dim>
4278 <dim>768</dim>
4279 </port>
4280 <port id="1" precision="FP32" />
4281 <port id="2" precision="FP32" />
4282 <port id="3" precision="FP32" />
4283 <port id="4" precision="FP32" />
4284 </input>
4285 <output>
4286 <port id="5" precision="FP32">
4287 <dim>-1</dim>
4288 <dim>-1</dim>
4289 <dim>768</dim>
4290 </port>
4291 </output>
4292 </layer>
4293 <layer id="313" name="__module.roberta.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
4294 <data element_type="i64" shape="4" offset="39811496" size="32" />
4295 <output>
4296 <port id="0" precision="I64">
4297 <dim>4</dim>
4298 </port>
4299 </output>
4300 </layer>
4301 <layer id="314" name="__module.roberta.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
4302 <data special_zero="true" />
4303 <input>
4304 <port id="0" precision="FP32">
4305 <dim>-1</dim>
4306 <dim>-1</dim>
4307 <dim>768</dim>
4308 </port>
4309 <port id="1" precision="I64">
4310 <dim>4</dim>
4311 </port>
4312 </input>
4313 <output>
4314 <port id="2" precision="FP32" names="242,x.27">
4315 <dim>-1</dim>
4316 <dim>-1</dim>
4317 <dim>12</dim>
4318 <dim>64</dim>
4319 </port>
4320 </output>
4321 </layer>
4322 <layer id="315" name="Constant_675" type="Const" version="opset1">
4323 <data element_type="i64" shape="4" offset="39811528" size="32" />
4324 <output>
4325 <port id="0" precision="I64" names="243">
4326 <dim>4</dim>
4327 </port>
4328 </output>
4329 </layer>
4330 <layer id="316" name="__module.roberta.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
4331 <input>
4332 <port id="0" precision="FP32">
4333 <dim>-1</dim>
4334 <dim>-1</dim>
4335 <dim>12</dim>
4336 <dim>64</dim>
4337 </port>
4338 <port id="1" precision="I64">
4339 <dim>4</dim>
4340 </port>
4341 </input>
4342 <output>
4343 <port id="2" precision="FP32" names="244">
4344 <dim>-1</dim>
4345 <dim>12</dim>
4346 <dim>-1</dim>
4347 <dim>64</dim>
4348 </port>
4349 </output>
4350 </layer>
4351 <layer id="317" name="Constant_220546" type="Const" version="opset1">
4352 <data element_type="i8" shape="768, 768" offset="54145680" size="589824" />
4353 <output>
4354 <port id="0" precision="I8">
4355 <dim>768</dim>
4356 <dim>768</dim>
4357 </port>
4358 </output>
4359 </layer>
4360 <layer id="318" name="Convert_220547" type="Convert" version="opset1">
4361 <data destination_type="f32" />
4362 <input>
4363 <port id="0" precision="I8">
4364 <dim>768</dim>
4365 <dim>768</dim>
4366 </port>
4367 </input>
4368 <output>
4369 <port id="1" precision="FP32">
4370 <dim>768</dim>
4371 <dim>768</dim>
4372 </port>
4373 </output>
4374 </layer>
4375 <layer id="319" name="Constant_220548" type="Const" version="opset1">
4376 <data element_type="f32" shape="768, 1" offset="54735504" size="3072" />
4377 <output>
4378 <port id="0" precision="FP32">
4379 <dim>768</dim>
4380 <dim>1</dim>
4381 </port>
4382 </output>
4383 </layer>
4384 <layer id="320" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4385 <data auto_broadcast="numpy" />
4386 <input>
4387 <port id="0" precision="FP32">
4388 <dim>768</dim>
4389 <dim>768</dim>
4390 </port>
4391 <port id="1" precision="FP32">
4392 <dim>768</dim>
4393 <dim>1</dim>
4394 </port>
4395 </input>
4396 <output>
4397 <port id="2" precision="FP32">
4398 <dim>768</dim>
4399 <dim>768</dim>
4400 </port>
4401 </output>
4402 </layer>
4403 <layer id="321" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
4404 <data transpose_a="false" transpose_b="true" />
4405 <input>
4406 <port id="0" precision="FP32">
4407 <dim>-1</dim>
4408 <dim>-1</dim>
4409 <dim>768</dim>
4410 </port>
4411 <port id="1" precision="FP32">
4412 <dim>768</dim>
4413 <dim>768</dim>
4414 </port>
4415 </input>
4416 <output>
4417 <port id="2" precision="FP32">
4418 <dim>-1</dim>
4419 <dim>-1</dim>
4420 <dim>768</dim>
4421 </port>
4422 </output>
4423 </layer>
4424 <layer id="322" name="Constant_5383" type="Const" version="opset1">
4425 <data element_type="f32" shape="1, 1, 768" offset="54738576" size="3072" />
4426 <output>
4427 <port id="0" precision="FP32">
4428 <dim>1</dim>
4429 <dim>1</dim>
4430 <dim>768</dim>
4431 </port>
4432 </output>
4433 </layer>
4434 <layer id="323" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
4435 <data auto_broadcast="numpy" />
4436 <input>
4437 <port id="0" precision="FP32">
4438 <dim>-1</dim>
4439 <dim>-1</dim>
4440 <dim>768</dim>
4441 </port>
4442 <port id="1" precision="FP32">
4443 <dim>1</dim>
4444 <dim>1</dim>
4445 <dim>768</dim>
4446 </port>
4447 </input>
4448 <output>
4449 <port id="2" precision="FP32" names="247,x.29">
4450 <dim>-1</dim>
4451 <dim>-1</dim>
4452 <dim>768</dim>
4453 </port>
4454 </output>
4455 </layer>
4456 <layer id="324" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
4457 <data element_type="f32" shape="" offset="54741648" size="4" />
4458 <output>
4459 <port id="0" precision="FP32" />
4460 </output>
4461 </layer>
4462 <layer id="325" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
4463 <data element_type="f32" shape="" offset="54741652" size="4" />
4464 <output>
4465 <port id="0" precision="FP32" />
4466 </output>
4467 </layer>
4468 <layer id="326" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
4469 <data element_type="f32" shape="" offset="54741648" size="4" />
4470 <output>
4471 <port id="0" precision="FP32" />
4472 </output>
4473 </layer>
4474 <layer id="327" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
4475 <data element_type="f32" shape="" offset="54741652" size="4" />
4476 <output>
4477 <port id="0" precision="FP32" />
4478 </output>
4479 </layer>
4480 <layer id="328" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
4481 <data levels="256" auto_broadcast="numpy" />
4482 <input>
4483 <port id="0" precision="FP32">
4484 <dim>-1</dim>
4485 <dim>-1</dim>
4486 <dim>768</dim>
4487 </port>
4488 <port id="1" precision="FP32" />
4489 <port id="2" precision="FP32" />
4490 <port id="3" precision="FP32" />
4491 <port id="4" precision="FP32" />
4492 </input>
4493 <output>
4494 <port id="5" precision="FP32">
4495 <dim>-1</dim>
4496 <dim>-1</dim>
4497 <dim>768</dim>
4498 </port>
4499 </output>
4500 </layer>
4501 <layer id="329" name="__module.roberta.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
4502 <data element_type="i64" shape="4" offset="39811496" size="32" />
4503 <output>
4504 <port id="0" precision="I64">
4505 <dim>4</dim>
4506 </port>
4507 </output>
4508 </layer>
4509 <layer id="330" name="__module.roberta.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
4510 <data special_zero="true" />
4511 <input>
4512 <port id="0" precision="FP32">
4513 <dim>-1</dim>
4514 <dim>-1</dim>
4515 <dim>768</dim>
4516 </port>
4517 <port id="1" precision="I64">
4518 <dim>4</dim>
4519 </port>
4520 </input>
4521 <output>
4522 <port id="2" precision="FP32" names="251,x.31">
4523 <dim>-1</dim>
4524 <dim>-1</dim>
4525 <dim>12</dim>
4526 <dim>64</dim>
4527 </port>
4528 </output>
4529 </layer>
4530 <layer id="331" name="Constant_698" type="Const" version="opset1">
4531 <data element_type="i64" shape="4" offset="39811528" size="32" />
4532 <output>
4533 <port id="0" precision="I64" names="252">
4534 <dim>4</dim>
4535 </port>
4536 </output>
4537 </layer>
4538 <layer id="332" name="__module.roberta.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
4539 <input>
4540 <port id="0" precision="FP32">
4541 <dim>-1</dim>
4542 <dim>-1</dim>
4543 <dim>12</dim>
4544 <dim>64</dim>
4545 </port>
4546 <port id="1" precision="I64">
4547 <dim>4</dim>
4548 </port>
4549 </input>
4550 <output>
4551 <port id="2" precision="FP32" names="253">
4552 <dim>-1</dim>
4553 <dim>12</dim>
4554 <dim>-1</dim>
4555 <dim>64</dim>
4556 </port>
4557 </output>
4558 </layer>
4559 <layer id="333" name="Constant_220550" type="Const" version="opset1">
4560 <data element_type="i8" shape="768, 768" offset="54741656" size="589824" />
4561 <output>
4562 <port id="0" precision="I8">
4563 <dim>768</dim>
4564 <dim>768</dim>
4565 </port>
4566 </output>
4567 </layer>
4568 <layer id="334" name="Convert_220551" type="Convert" version="opset1">
4569 <data destination_type="f32" />
4570 <input>
4571 <port id="0" precision="I8">
4572 <dim>768</dim>
4573 <dim>768</dim>
4574 </port>
4575 </input>
4576 <output>
4577 <port id="1" precision="FP32">
4578 <dim>768</dim>
4579 <dim>768</dim>
4580 </port>
4581 </output>
4582 </layer>
4583 <layer id="335" name="Constant_220552" type="Const" version="opset1">
4584 <data element_type="f32" shape="768, 1" offset="55331480" size="3072" />
4585 <output>
4586 <port id="0" precision="FP32">
4587 <dim>768</dim>
4588 <dim>1</dim>
4589 </port>
4590 </output>
4591 </layer>
4592 <layer id="336" name="__module.roberta.encoder.layer.2.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4593 <data auto_broadcast="numpy" />
4594 <input>
4595 <port id="0" precision="FP32">
4596 <dim>768</dim>
4597 <dim>768</dim>
4598 </port>
4599 <port id="1" precision="FP32">
4600 <dim>768</dim>
4601 <dim>1</dim>
4602 </port>
4603 </input>
4604 <output>
4605 <port id="2" precision="FP32">
4606 <dim>768</dim>
4607 <dim>768</dim>
4608 </port>
4609 </output>
4610 </layer>
4611 <layer id="337" name="__module.roberta.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
4612 <data transpose_a="false" transpose_b="true" />
4613 <input>
4614 <port id="0" precision="FP32">
4615 <dim>-1</dim>
4616 <dim>-1</dim>
4617 <dim>768</dim>
4618 </port>
4619 <port id="1" precision="FP32">
4620 <dim>768</dim>
4621 <dim>768</dim>
4622 </port>
4623 </input>
4624 <output>
4625 <port id="2" precision="FP32">
4626 <dim>-1</dim>
4627 <dim>-1</dim>
4628 <dim>768</dim>
4629 </port>
4630 </output>
4631 </layer>
4632 <layer id="338" name="Constant_5384" type="Const" version="opset1">
4633 <data element_type="f32" shape="1, 1, 768" offset="55334552" size="3072" />
4634 <output>
4635 <port id="0" precision="FP32">
4636 <dim>1</dim>
4637 <dim>1</dim>
4638 <dim>768</dim>
4639 </port>
4640 </output>
4641 </layer>
4642 <layer id="339" name="__module.roberta.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
4643 <data auto_broadcast="numpy" />
4644 <input>
4645 <port id="0" precision="FP32">
4646 <dim>-1</dim>
4647 <dim>-1</dim>
4648 <dim>768</dim>
4649 </port>
4650 <port id="1" precision="FP32">
4651 <dim>1</dim>
4652 <dim>1</dim>
4653 <dim>768</dim>
4654 </port>
4655 </input>
4656 <output>
4657 <port id="2" precision="FP32" names="256,x.33">
4658 <dim>-1</dim>
4659 <dim>-1</dim>
4660 <dim>768</dim>
4661 </port>
4662 </output>
4663 </layer>
4664 <layer id="340" name="__module.roberta.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
4665 <data element_type="i64" shape="4" offset="39811496" size="32" />
4666 <output>
4667 <port id="0" precision="I64">
4668 <dim>4</dim>
4669 </port>
4670 </output>
4671 </layer>
4672 <layer id="341" name="__module.roberta.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
4673 <data special_zero="true" />
4674 <input>
4675 <port id="0" precision="FP32">
4676 <dim>-1</dim>
4677 <dim>-1</dim>
4678 <dim>768</dim>
4679 </port>
4680 <port id="1" precision="I64">
4681 <dim>4</dim>
4682 </port>
4683 </input>
4684 <output>
4685 <port id="2" precision="FP32" names="260,x.35">
4686 <dim>-1</dim>
4687 <dim>-1</dim>
4688 <dim>12</dim>
4689 <dim>64</dim>
4690 </port>
4691 </output>
4692 </layer>
4693 <layer id="342" name="Constant_721" type="Const" version="opset1">
4694 <data element_type="i64" shape="4" offset="39811528" size="32" />
4695 <output>
4696 <port id="0" precision="I64" names="261">
4697 <dim>4</dim>
4698 </port>
4699 </output>
4700 </layer>
4701 <layer id="343" name="__module.roberta.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
4702 <input>
4703 <port id="0" precision="FP32">
4704 <dim>-1</dim>
4705 <dim>-1</dim>
4706 <dim>12</dim>
4707 <dim>64</dim>
4708 </port>
4709 <port id="1" precision="I64">
4710 <dim>4</dim>
4711 </port>
4712 </input>
4713 <output>
4714 <port id="2" precision="FP32" names="262">
4715 <dim>-1</dim>
4716 <dim>12</dim>
4717 <dim>-1</dim>
4718 <dim>64</dim>
4719 </port>
4720 </output>
4721 </layer>
4722 <layer id="344" name="__module.roberta.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
4723 <data causal="false" />
4724 <input>
4725 <port id="0" precision="FP32">
4726 <dim>-1</dim>
4727 <dim>12</dim>
4728 <dim>-1</dim>
4729 <dim>64</dim>
4730 </port>
4731 <port id="1" precision="FP32">
4732 <dim>-1</dim>
4733 <dim>12</dim>
4734 <dim>-1</dim>
4735 <dim>64</dim>
4736 </port>
4737 <port id="2" precision="FP32">
4738 <dim>-1</dim>
4739 <dim>12</dim>
4740 <dim>-1</dim>
4741 <dim>64</dim>
4742 </port>
4743 <port id="3" precision="FP32">
4744 <dim>-1</dim>
4745 <dim>1</dim>
4746 <dim>-1</dim>
4747 <dim>-1</dim>
4748 </port>
4749 </input>
4750 <output>
4751 <port id="4" precision="FP32" names="263,attn_output.9">
4752 <dim>-1</dim>
4753 <dim>12</dim>
4754 <dim>-1</dim>
4755 <dim>64</dim>
4756 </port>
4757 </output>
4758 </layer>
4759 <layer id="345" name="__module.roberta.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
4760 <data element_type="i32" shape="4" offset="41003536" size="16" />
4761 <output>
4762 <port id="0" precision="I32">
4763 <dim>4</dim>
4764 </port>
4765 </output>
4766 </layer>
4767 <layer id="346" name="__module.roberta.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
4768 <input>
4769 <port id="0" precision="FP32">
4770 <dim>-1</dim>
4771 <dim>12</dim>
4772 <dim>-1</dim>
4773 <dim>64</dim>
4774 </port>
4775 <port id="1" precision="I32">
4776 <dim>4</dim>
4777 </port>
4778 </input>
4779 <output>
4780 <port id="2" precision="FP32" names="264,attn_output.11">
4781 <dim>-1</dim>
4782 <dim>-1</dim>
4783 <dim>12</dim>
4784 <dim>64</dim>
4785 </port>
4786 </output>
4787 </layer>
4788 <layer id="347" name="Constant_5496" type="Const" version="opset1">
4789 <data element_type="i64" shape="3" offset="41003552" size="24" />
4790 <output>
4791 <port id="0" precision="I64">
4792 <dim>3</dim>
4793 </port>
4794 </output>
4795 </layer>
4796 <layer id="348" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
4797 <data special_zero="true" />
4798 <input>
4799 <port id="0" precision="FP32">
4800 <dim>-1</dim>
4801 <dim>-1</dim>
4802 <dim>12</dim>
4803 <dim>64</dim>
4804 </port>
4805 <port id="1" precision="I64">
4806 <dim>3</dim>
4807 </port>
4808 </input>
4809 <output>
4810 <port id="2" precision="FP32" names="266">
4811 <dim>-1</dim>
4812 <dim>-1</dim>
4813 <dim>768</dim>
4814 </port>
4815 </output>
4816 </layer>
4817 <layer id="349" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
4818 <data element_type="f32" shape="1, 1, 768" offset="55337624" size="3072" />
4819 <output>
4820 <port id="0" precision="FP32">
4821 <dim>1</dim>
4822 <dim>1</dim>
4823 <dim>768</dim>
4824 </port>
4825 </output>
4826 </layer>
4827 <layer id="350" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
4828 <data auto_broadcast="numpy" />
4829 <input>
4830 <port id="0" precision="FP32">
4831 <dim>-1</dim>
4832 <dim>-1</dim>
4833 <dim>768</dim>
4834 </port>
4835 <port id="1" precision="FP32">
4836 <dim>1</dim>
4837 <dim>1</dim>
4838 <dim>768</dim>
4839 </port>
4840 </input>
4841 <output>
4842 <port id="2" precision="FP32">
4843 <dim>-1</dim>
4844 <dim>-1</dim>
4845 <dim>768</dim>
4846 </port>
4847 </output>
4848 </layer>
4849 <layer id="351" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
4850 <data element_type="f32" shape="" offset="55340696" size="4" />
4851 <output>
4852 <port id="0" precision="FP32" />
4853 </output>
4854 </layer>
4855 <layer id="352" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
4856 <data element_type="f32" shape="" offset="55340700" size="4" />
4857 <output>
4858 <port id="0" precision="FP32" />
4859 </output>
4860 </layer>
4861 <layer id="353" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
4862 <data element_type="f32" shape="" offset="55340696" size="4" />
4863 <output>
4864 <port id="0" precision="FP32" />
4865 </output>
4866 </layer>
4867 <layer id="354" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
4868 <data element_type="f32" shape="" offset="55340700" size="4" />
4869 <output>
4870 <port id="0" precision="FP32" />
4871 </output>
4872 </layer>
4873 <layer id="355" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
4874 <data levels="256" auto_broadcast="numpy" />
4875 <input>
4876 <port id="0" precision="FP32">
4877 <dim>-1</dim>
4878 <dim>-1</dim>
4879 <dim>768</dim>
4880 </port>
4881 <port id="1" precision="FP32" />
4882 <port id="2" precision="FP32" />
4883 <port id="3" precision="FP32" />
4884 <port id="4" precision="FP32" />
4885 </input>
4886 <output>
4887 <port id="5" precision="FP32">
4888 <dim>-1</dim>
4889 <dim>-1</dim>
4890 <dim>768</dim>
4891 </port>
4892 </output>
4893 </layer>
4894 <layer id="356" name="Constant_220554" type="Const" version="opset1">
4895 <data element_type="i8" shape="768, 768" offset="55340704" size="589824" />
4896 <output>
4897 <port id="0" precision="I8">
4898 <dim>768</dim>
4899 <dim>768</dim>
4900 </port>
4901 </output>
4902 </layer>
4903 <layer id="357" name="Convert_220555" type="Convert" version="opset1">
4904 <data destination_type="f32" />
4905 <input>
4906 <port id="0" precision="I8">
4907 <dim>768</dim>
4908 <dim>768</dim>
4909 </port>
4910 </input>
4911 <output>
4912 <port id="1" precision="FP32">
4913 <dim>768</dim>
4914 <dim>768</dim>
4915 </port>
4916 </output>
4917 </layer>
4918 <layer id="358" name="Constant_220556" type="Const" version="opset1">
4919 <data element_type="f32" shape="768, 1" offset="55930528" size="3072" />
4920 <output>
4921 <port id="0" precision="FP32">
4922 <dim>768</dim>
4923 <dim>1</dim>
4924 </port>
4925 </output>
4926 </layer>
4927 <layer id="359" name="__module.roberta.encoder.layer.2.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
4928 <data auto_broadcast="numpy" />
4929 <input>
4930 <port id="0" precision="FP32">
4931 <dim>768</dim>
4932 <dim>768</dim>
4933 </port>
4934 <port id="1" precision="FP32">
4935 <dim>768</dim>
4936 <dim>1</dim>
4937 </port>
4938 </input>
4939 <output>
4940 <port id="2" precision="FP32">
4941 <dim>768</dim>
4942 <dim>768</dim>
4943 </port>
4944 </output>
4945 </layer>
4946 <layer id="360" name="__module.roberta.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4947 <data transpose_a="false" transpose_b="true" />
4948 <input>
4949 <port id="0" precision="FP32">
4950 <dim>-1</dim>
4951 <dim>-1</dim>
4952 <dim>768</dim>
4953 </port>
4954 <port id="1" precision="FP32">
4955 <dim>768</dim>
4956 <dim>768</dim>
4957 </port>
4958 </input>
4959 <output>
4960 <port id="2" precision="FP32">
4961 <dim>-1</dim>
4962 <dim>-1</dim>
4963 <dim>768</dim>
4964 </port>
4965 </output>
4966 </layer>
4967 <layer id="361" name="Constant_5385" type="Const" version="opset1">
4968 <data element_type="f32" shape="1, 1, 768" offset="55933600" size="3072" />
4969 <output>
4970 <port id="0" precision="FP32">
4971 <dim>1</dim>
4972 <dim>1</dim>
4973 <dim>768</dim>
4974 </port>
4975 </output>
4976 </layer>
4977 <layer id="362" name="__module.roberta.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
4978 <data auto_broadcast="numpy" />
4979 <input>
4980 <port id="0" precision="FP32">
4981 <dim>-1</dim>
4982 <dim>-1</dim>
4983 <dim>768</dim>
4984 </port>
4985 <port id="1" precision="FP32">
4986 <dim>1</dim>
4987 <dim>1</dim>
4988 <dim>768</dim>
4989 </port>
4990 </input>
4991 <output>
4992 <port id="2" precision="FP32" names="271,input.11">
4993 <dim>-1</dim>
4994 <dim>-1</dim>
4995 <dim>768</dim>
4996 </port>
4997 </output>
4998 </layer>
4999 <layer id="363" name="__module.roberta.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
5000 <data auto_broadcast="numpy" />
5001 <input>
5002 <port id="0" precision="FP32">
5003 <dim>-1</dim>
5004 <dim>-1</dim>
5005 <dim>768</dim>
5006 </port>
5007 <port id="1" precision="FP32">
5008 <dim>-1</dim>
5009 <dim>-1</dim>
5010 <dim>768</dim>
5011 </port>
5012 </input>
5013 <output>
5014 <port id="2" precision="FP32" names="273">
5015 <dim>-1</dim>
5016 <dim>-1</dim>
5017 <dim>768</dim>
5018 </port>
5019 </output>
5020 </layer>
5021 <layer id="364" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5022 <data element_type="i32" shape="1" offset="39206292" size="4" />
5023 <output>
5024 <port id="0" precision="I32">
5025 <dim>1</dim>
5026 </port>
5027 </output>
5028 </layer>
5029 <layer id="365" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5030 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5031 <input>
5032 <port id="0" precision="FP32">
5033 <dim>-1</dim>
5034 <dim>-1</dim>
5035 <dim>768</dim>
5036 </port>
5037 <port id="1" precision="I32">
5038 <dim>1</dim>
5039 </port>
5040 </input>
5041 <output>
5042 <port id="2" precision="FP32">
5043 <dim>-1</dim>
5044 <dim>-1</dim>
5045 <dim>768</dim>
5046 </port>
5047 </output>
5048 </layer>
5049 <layer id="366" name="Constant_5386" type="Const" version="opset1">
5050 <data element_type="f32" shape="1, 1, 768" offset="55936672" size="3072" />
5051 <output>
5052 <port id="0" precision="FP32">
5053 <dim>1</dim>
5054 <dim>1</dim>
5055 <dim>768</dim>
5056 </port>
5057 </output>
5058 </layer>
5059 <layer id="367" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5060 <data auto_broadcast="numpy" />
5061 <input>
5062 <port id="0" precision="FP32">
5063 <dim>-1</dim>
5064 <dim>-1</dim>
5065 <dim>768</dim>
5066 </port>
5067 <port id="1" precision="FP32">
5068 <dim>1</dim>
5069 <dim>1</dim>
5070 <dim>768</dim>
5071 </port>
5072 </input>
5073 <output>
5074 <port id="2" precision="FP32">
5075 <dim>-1</dim>
5076 <dim>-1</dim>
5077 <dim>768</dim>
5078 </port>
5079 </output>
5080 </layer>
5081 <layer id="368" name="Constant_5387" type="Const" version="opset1">
5082 <data element_type="f32" shape="1, 1, 768" offset="55939744" size="3072" />
5083 <output>
5084 <port id="0" precision="FP32">
5085 <dim>1</dim>
5086 <dim>1</dim>
5087 <dim>768</dim>
5088 </port>
5089 </output>
5090 </layer>
5091 <layer id="369" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5092 <data auto_broadcast="numpy" />
5093 <input>
5094 <port id="0" precision="FP32">
5095 <dim>-1</dim>
5096 <dim>-1</dim>
5097 <dim>768</dim>
5098 </port>
5099 <port id="1" precision="FP32">
5100 <dim>1</dim>
5101 <dim>1</dim>
5102 <dim>768</dim>
5103 </port>
5104 </input>
5105 <output>
5106 <port id="2" precision="FP32" names="277,input_tensor.5">
5107 <dim>-1</dim>
5108 <dim>-1</dim>
5109 <dim>768</dim>
5110 </port>
5111 </output>
5112 </layer>
5113 <layer id="370" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5114 <data element_type="f32" shape="1, 1, 768" offset="55942816" size="3072" />
5115 <output>
5116 <port id="0" precision="FP32">
5117 <dim>1</dim>
5118 <dim>1</dim>
5119 <dim>768</dim>
5120 </port>
5121 </output>
5122 </layer>
5123 <layer id="371" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5124 <data auto_broadcast="numpy" />
5125 <input>
5126 <port id="0" precision="FP32">
5127 <dim>-1</dim>
5128 <dim>-1</dim>
5129 <dim>768</dim>
5130 </port>
5131 <port id="1" precision="FP32">
5132 <dim>1</dim>
5133 <dim>1</dim>
5134 <dim>768</dim>
5135 </port>
5136 </input>
5137 <output>
5138 <port id="2" precision="FP32">
5139 <dim>-1</dim>
5140 <dim>-1</dim>
5141 <dim>768</dim>
5142 </port>
5143 </output>
5144 </layer>
5145 <layer id="372" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5146 <data element_type="f32" shape="" offset="55945888" size="4" />
5147 <output>
5148 <port id="0" precision="FP32" />
5149 </output>
5150 </layer>
5151 <layer id="373" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5152 <data element_type="f32" shape="" offset="55945892" size="4" />
5153 <output>
5154 <port id="0" precision="FP32" />
5155 </output>
5156 </layer>
5157 <layer id="374" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5158 <data element_type="f32" shape="" offset="55945888" size="4" />
5159 <output>
5160 <port id="0" precision="FP32" />
5161 </output>
5162 </layer>
5163 <layer id="375" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5164 <data element_type="f32" shape="" offset="55945892" size="4" />
5165 <output>
5166 <port id="0" precision="FP32" />
5167 </output>
5168 </layer>
5169 <layer id="376" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5170 <data levels="256" auto_broadcast="numpy" />
5171 <input>
5172 <port id="0" precision="FP32">
5173 <dim>-1</dim>
5174 <dim>-1</dim>
5175 <dim>768</dim>
5176 </port>
5177 <port id="1" precision="FP32" />
5178 <port id="2" precision="FP32" />
5179 <port id="3" precision="FP32" />
5180 <port id="4" precision="FP32" />
5181 </input>
5182 <output>
5183 <port id="5" precision="FP32">
5184 <dim>-1</dim>
5185 <dim>-1</dim>
5186 <dim>768</dim>
5187 </port>
5188 </output>
5189 </layer>
5190 <layer id="377" name="Constant_220558" type="Const" version="opset1">
5191 <data element_type="i8" shape="3072, 768" offset="55945896" size="2359296" />
5192 <output>
5193 <port id="0" precision="I8">
5194 <dim>3072</dim>
5195 <dim>768</dim>
5196 </port>
5197 </output>
5198 </layer>
5199 <layer id="378" name="Convert_220559" type="Convert" version="opset1">
5200 <data destination_type="f32" />
5201 <input>
5202 <port id="0" precision="I8">
5203 <dim>3072</dim>
5204 <dim>768</dim>
5205 </port>
5206 </input>
5207 <output>
5208 <port id="1" precision="FP32">
5209 <dim>3072</dim>
5210 <dim>768</dim>
5211 </port>
5212 </output>
5213 </layer>
5214 <layer id="379" name="Constant_220560" type="Const" version="opset1">
5215 <data element_type="f32" shape="3072, 1" offset="58305192" size="12288" />
5216 <output>
5217 <port id="0" precision="FP32">
5218 <dim>3072</dim>
5219 <dim>1</dim>
5220 </port>
5221 </output>
5222 </layer>
5223 <layer id="380" name="__module.roberta.encoder.layer.2.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5224 <data auto_broadcast="numpy" />
5225 <input>
5226 <port id="0" precision="FP32">
5227 <dim>3072</dim>
5228 <dim>768</dim>
5229 </port>
5230 <port id="1" precision="FP32">
5231 <dim>3072</dim>
5232 <dim>1</dim>
5233 </port>
5234 </input>
5235 <output>
5236 <port id="2" precision="FP32">
5237 <dim>3072</dim>
5238 <dim>768</dim>
5239 </port>
5240 </output>
5241 </layer>
5242 <layer id="381" name="__module.roberta.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5243 <data transpose_a="false" transpose_b="true" />
5244 <input>
5245 <port id="0" precision="FP32">
5246 <dim>-1</dim>
5247 <dim>-1</dim>
5248 <dim>768</dim>
5249 </port>
5250 <port id="1" precision="FP32">
5251 <dim>3072</dim>
5252 <dim>768</dim>
5253 </port>
5254 </input>
5255 <output>
5256 <port id="2" precision="FP32">
5257 <dim>-1</dim>
5258 <dim>-1</dim>
5259 <dim>3072</dim>
5260 </port>
5261 </output>
5262 </layer>
5263 <layer id="382" name="Constant_5388" type="Const" version="opset1">
5264 <data element_type="f32" shape="1, 1, 3072" offset="58317480" size="12288" />
5265 <output>
5266 <port id="0" precision="FP32">
5267 <dim>1</dim>
5268 <dim>1</dim>
5269 <dim>3072</dim>
5270 </port>
5271 </output>
5272 </layer>
5273 <layer id="383" name="__module.roberta.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5274 <data auto_broadcast="numpy" />
5275 <input>
5276 <port id="0" precision="FP32">
5277 <dim>-1</dim>
5278 <dim>-1</dim>
5279 <dim>3072</dim>
5280 </port>
5281 <port id="1" precision="FP32">
5282 <dim>1</dim>
5283 <dim>1</dim>
5284 <dim>3072</dim>
5285 </port>
5286 </input>
5287 <output>
5288 <port id="2" precision="FP32" names="281">
5289 <dim>-1</dim>
5290 <dim>-1</dim>
5291 <dim>3072</dim>
5292 </port>
5293 </output>
5294 </layer>
5295 <layer id="384" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5296 <data approximation_mode="ERF" />
5297 <input>
5298 <port id="0" precision="FP32">
5299 <dim>-1</dim>
5300 <dim>-1</dim>
5301 <dim>3072</dim>
5302 </port>
5303 </input>
5304 <output>
5305 <port id="1" precision="FP32" names="282">
5306 <dim>-1</dim>
5307 <dim>-1</dim>
5308 <dim>3072</dim>
5309 </port>
5310 </output>
5311 </layer>
5312 <layer id="385" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5313 <data element_type="f32" shape="1, 1, 3072" offset="58329768" size="12288" />
5314 <output>
5315 <port id="0" precision="FP32">
5316 <dim>1</dim>
5317 <dim>1</dim>
5318 <dim>3072</dim>
5319 </port>
5320 </output>
5321 </layer>
5322 <layer id="386" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5323 <data auto_broadcast="numpy" />
5324 <input>
5325 <port id="0" precision="FP32">
5326 <dim>-1</dim>
5327 <dim>-1</dim>
5328 <dim>3072</dim>
5329 </port>
5330 <port id="1" precision="FP32">
5331 <dim>1</dim>
5332 <dim>1</dim>
5333 <dim>3072</dim>
5334 </port>
5335 </input>
5336 <output>
5337 <port id="2" precision="FP32">
5338 <dim>-1</dim>
5339 <dim>-1</dim>
5340 <dim>3072</dim>
5341 </port>
5342 </output>
5343 </layer>
5344 <layer id="387" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5345 <data element_type="f32" shape="" offset="58342056" size="4" />
5346 <output>
5347 <port id="0" precision="FP32" />
5348 </output>
5349 </layer>
5350 <layer id="388" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5351 <data element_type="f32" shape="" offset="58342060" size="4" />
5352 <output>
5353 <port id="0" precision="FP32" />
5354 </output>
5355 </layer>
5356 <layer id="389" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5357 <data element_type="f32" shape="" offset="58342056" size="4" />
5358 <output>
5359 <port id="0" precision="FP32" />
5360 </output>
5361 </layer>
5362 <layer id="390" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5363 <data element_type="f32" shape="" offset="58342060" size="4" />
5364 <output>
5365 <port id="0" precision="FP32" />
5366 </output>
5367 </layer>
5368 <layer id="391" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5369 <data levels="256" auto_broadcast="numpy" />
5370 <input>
5371 <port id="0" precision="FP32">
5372 <dim>-1</dim>
5373 <dim>-1</dim>
5374 <dim>3072</dim>
5375 </port>
5376 <port id="1" precision="FP32" />
5377 <port id="2" precision="FP32" />
5378 <port id="3" precision="FP32" />
5379 <port id="4" precision="FP32" />
5380 </input>
5381 <output>
5382 <port id="5" precision="FP32">
5383 <dim>-1</dim>
5384 <dim>-1</dim>
5385 <dim>3072</dim>
5386 </port>
5387 </output>
5388 </layer>
5389 <layer id="392" name="Constant_220562" type="Const" version="opset1">
5390 <data element_type="i8" shape="768, 3072" offset="58342064" size="2359296" />
5391 <output>
5392 <port id="0" precision="I8">
5393 <dim>768</dim>
5394 <dim>3072</dim>
5395 </port>
5396 </output>
5397 </layer>
5398 <layer id="393" name="Convert_220563" type="Convert" version="opset1">
5399 <data destination_type="f32" />
5400 <input>
5401 <port id="0" precision="I8">
5402 <dim>768</dim>
5403 <dim>3072</dim>
5404 </port>
5405 </input>
5406 <output>
5407 <port id="1" precision="FP32">
5408 <dim>768</dim>
5409 <dim>3072</dim>
5410 </port>
5411 </output>
5412 </layer>
5413 <layer id="394" name="Constant_220564" type="Const" version="opset1">
5414 <data element_type="f32" shape="768, 1" offset="60701360" size="3072" />
5415 <output>
5416 <port id="0" precision="FP32">
5417 <dim>768</dim>
5418 <dim>1</dim>
5419 </port>
5420 </output>
5421 </layer>
5422 <layer id="395" name="__module.roberta.encoder.layer.2.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5423 <data auto_broadcast="numpy" />
5424 <input>
5425 <port id="0" precision="FP32">
5426 <dim>768</dim>
5427 <dim>3072</dim>
5428 </port>
5429 <port id="1" precision="FP32">
5430 <dim>768</dim>
5431 <dim>1</dim>
5432 </port>
5433 </input>
5434 <output>
5435 <port id="2" precision="FP32">
5436 <dim>768</dim>
5437 <dim>3072</dim>
5438 </port>
5439 </output>
5440 </layer>
5441 <layer id="396" name="__module.roberta.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5442 <data transpose_a="false" transpose_b="true" />
5443 <input>
5444 <port id="0" precision="FP32">
5445 <dim>-1</dim>
5446 <dim>-1</dim>
5447 <dim>3072</dim>
5448 </port>
5449 <port id="1" precision="FP32">
5450 <dim>768</dim>
5451 <dim>3072</dim>
5452 </port>
5453 </input>
5454 <output>
5455 <port id="2" precision="FP32">
5456 <dim>-1</dim>
5457 <dim>-1</dim>
5458 <dim>768</dim>
5459 </port>
5460 </output>
5461 </layer>
5462 <layer id="397" name="Constant_5389" type="Const" version="opset1">
5463 <data element_type="f32" shape="1, 1, 768" offset="60704432" size="3072" />
5464 <output>
5465 <port id="0" precision="FP32">
5466 <dim>1</dim>
5467 <dim>1</dim>
5468 <dim>768</dim>
5469 </port>
5470 </output>
5471 </layer>
5472 <layer id="398" name="__module.roberta.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
5473 <data auto_broadcast="numpy" />
5474 <input>
5475 <port id="0" precision="FP32">
5476 <dim>-1</dim>
5477 <dim>-1</dim>
5478 <dim>768</dim>
5479 </port>
5480 <port id="1" precision="FP32">
5481 <dim>1</dim>
5482 <dim>1</dim>
5483 <dim>768</dim>
5484 </port>
5485 </input>
5486 <output>
5487 <port id="2" precision="FP32" names="287,input.13">
5488 <dim>-1</dim>
5489 <dim>-1</dim>
5490 <dim>768</dim>
5491 </port>
5492 </output>
5493 </layer>
5494 <layer id="399" name="__module.roberta.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
5495 <data auto_broadcast="numpy" />
5496 <input>
5497 <port id="0" precision="FP32">
5498 <dim>-1</dim>
5499 <dim>-1</dim>
5500 <dim>768</dim>
5501 </port>
5502 <port id="1" precision="FP32">
5503 <dim>-1</dim>
5504 <dim>-1</dim>
5505 <dim>768</dim>
5506 </port>
5507 </input>
5508 <output>
5509 <port id="2" precision="FP32" names="289">
5510 <dim>-1</dim>
5511 <dim>-1</dim>
5512 <dim>768</dim>
5513 </port>
5514 </output>
5515 </layer>
5516 <layer id="400" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5517 <data element_type="i32" shape="1" offset="39206292" size="4" />
5518 <output>
5519 <port id="0" precision="I32">
5520 <dim>1</dim>
5521 </port>
5522 </output>
5523 </layer>
5524 <layer id="401" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5525 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5526 <input>
5527 <port id="0" precision="FP32">
5528 <dim>-1</dim>
5529 <dim>-1</dim>
5530 <dim>768</dim>
5531 </port>
5532 <port id="1" precision="I32">
5533 <dim>1</dim>
5534 </port>
5535 </input>
5536 <output>
5537 <port id="2" precision="FP32">
5538 <dim>-1</dim>
5539 <dim>-1</dim>
5540 <dim>768</dim>
5541 </port>
5542 </output>
5543 </layer>
5544 <layer id="402" name="Constant_5390" type="Const" version="opset1">
5545 <data element_type="f32" shape="1, 1, 768" offset="60707504" size="3072" />
5546 <output>
5547 <port id="0" precision="FP32">
5548 <dim>1</dim>
5549 <dim>1</dim>
5550 <dim>768</dim>
5551 </port>
5552 </output>
5553 </layer>
5554 <layer id="403" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5555 <data auto_broadcast="numpy" />
5556 <input>
5557 <port id="0" precision="FP32">
5558 <dim>-1</dim>
5559 <dim>-1</dim>
5560 <dim>768</dim>
5561 </port>
5562 <port id="1" precision="FP32">
5563 <dim>1</dim>
5564 <dim>1</dim>
5565 <dim>768</dim>
5566 </port>
5567 </input>
5568 <output>
5569 <port id="2" precision="FP32">
5570 <dim>-1</dim>
5571 <dim>-1</dim>
5572 <dim>768</dim>
5573 </port>
5574 </output>
5575 </layer>
5576 <layer id="404" name="Constant_5391" type="Const" version="opset1">
5577 <data element_type="f32" shape="1, 1, 768" offset="60710576" size="3072" />
5578 <output>
5579 <port id="0" precision="FP32">
5580 <dim>1</dim>
5581 <dim>1</dim>
5582 <dim>768</dim>
5583 </port>
5584 </output>
5585 </layer>
5586 <layer id="405" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5587 <data auto_broadcast="numpy" />
5588 <input>
5589 <port id="0" precision="FP32">
5590 <dim>-1</dim>
5591 <dim>-1</dim>
5592 <dim>768</dim>
5593 </port>
5594 <port id="1" precision="FP32">
5595 <dim>1</dim>
5596 <dim>1</dim>
5597 <dim>768</dim>
5598 </port>
5599 </input>
5600 <output>
5601 <port id="2" precision="FP32" names="293,hidden_states.19">
5602 <dim>-1</dim>
5603 <dim>-1</dim>
5604 <dim>768</dim>
5605 </port>
5606 </output>
5607 </layer>
5608 <layer id="406" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
5609 <data element_type="f32" shape="1, 1, 768" offset="60713648" size="3072" />
5610 <output>
5611 <port id="0" precision="FP32">
5612 <dim>1</dim>
5613 <dim>1</dim>
5614 <dim>768</dim>
5615 </port>
5616 </output>
5617 </layer>
5618 <layer id="407" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
5619 <data auto_broadcast="numpy" />
5620 <input>
5621 <port id="0" precision="FP32">
5622 <dim>-1</dim>
5623 <dim>-1</dim>
5624 <dim>768</dim>
5625 </port>
5626 <port id="1" precision="FP32">
5627 <dim>1</dim>
5628 <dim>1</dim>
5629 <dim>768</dim>
5630 </port>
5631 </input>
5632 <output>
5633 <port id="2" precision="FP32">
5634 <dim>-1</dim>
5635 <dim>-1</dim>
5636 <dim>768</dim>
5637 </port>
5638 </output>
5639 </layer>
5640 <layer id="408" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
5641 <data element_type="f32" shape="" offset="60716720" size="4" />
5642 <output>
5643 <port id="0" precision="FP32" />
5644 </output>
5645 </layer>
5646 <layer id="409" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
5647 <data element_type="f32" shape="" offset="60716724" size="4" />
5648 <output>
5649 <port id="0" precision="FP32" />
5650 </output>
5651 </layer>
5652 <layer id="410" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
5653 <data element_type="f32" shape="" offset="60716720" size="4" />
5654 <output>
5655 <port id="0" precision="FP32" />
5656 </output>
5657 </layer>
5658 <layer id="411" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
5659 <data element_type="f32" shape="" offset="60716724" size="4" />
5660 <output>
5661 <port id="0" precision="FP32" />
5662 </output>
5663 </layer>
5664 <layer id="412" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
5665 <data levels="256" auto_broadcast="numpy" />
5666 <input>
5667 <port id="0" precision="FP32">
5668 <dim>-1</dim>
5669 <dim>-1</dim>
5670 <dim>768</dim>
5671 </port>
5672 <port id="1" precision="FP32" />
5673 <port id="2" precision="FP32" />
5674 <port id="3" precision="FP32" />
5675 <port id="4" precision="FP32" />
5676 </input>
5677 <output>
5678 <port id="5" precision="FP32">
5679 <dim>-1</dim>
5680 <dim>-1</dim>
5681 <dim>768</dim>
5682 </port>
5683 </output>
5684 </layer>
5685 <layer id="413" name="Constant_220566" type="Const" version="opset1">
5686 <data element_type="i8" shape="768, 768" offset="60716728" size="589824" />
5687 <output>
5688 <port id="0" precision="I8">
5689 <dim>768</dim>
5690 <dim>768</dim>
5691 </port>
5692 </output>
5693 </layer>
5694 <layer id="414" name="Convert_220567" type="Convert" version="opset1">
5695 <data destination_type="f32" />
5696 <input>
5697 <port id="0" precision="I8">
5698 <dim>768</dim>
5699 <dim>768</dim>
5700 </port>
5701 </input>
5702 <output>
5703 <port id="1" precision="FP32">
5704 <dim>768</dim>
5705 <dim>768</dim>
5706 </port>
5707 </output>
5708 </layer>
5709 <layer id="415" name="Constant_220568" type="Const" version="opset1">
5710 <data element_type="f32" shape="768, 1" offset="61306552" size="3072" />
5711 <output>
5712 <port id="0" precision="FP32">
5713 <dim>768</dim>
5714 <dim>1</dim>
5715 </port>
5716 </output>
5717 </layer>
5718 <layer id="416" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5719 <data auto_broadcast="numpy" />
5720 <input>
5721 <port id="0" precision="FP32">
5722 <dim>768</dim>
5723 <dim>768</dim>
5724 </port>
5725 <port id="1" precision="FP32">
5726 <dim>768</dim>
5727 <dim>1</dim>
5728 </port>
5729 </input>
5730 <output>
5731 <port id="2" precision="FP32">
5732 <dim>768</dim>
5733 <dim>768</dim>
5734 </port>
5735 </output>
5736 </layer>
5737 <layer id="417" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
5738 <data transpose_a="false" transpose_b="true" />
5739 <input>
5740 <port id="0" precision="FP32">
5741 <dim>-1</dim>
5742 <dim>-1</dim>
5743 <dim>768</dim>
5744 </port>
5745 <port id="1" precision="FP32">
5746 <dim>768</dim>
5747 <dim>768</dim>
5748 </port>
5749 </input>
5750 <output>
5751 <port id="2" precision="FP32">
5752 <dim>-1</dim>
5753 <dim>-1</dim>
5754 <dim>768</dim>
5755 </port>
5756 </output>
5757 </layer>
5758 <layer id="418" name="Constant_5392" type="Const" version="opset1">
5759 <data element_type="f32" shape="1, 1, 768" offset="61309624" size="3072" />
5760 <output>
5761 <port id="0" precision="FP32">
5762 <dim>1</dim>
5763 <dim>1</dim>
5764 <dim>768</dim>
5765 </port>
5766 </output>
5767 </layer>
5768 <layer id="419" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
5769 <data auto_broadcast="numpy" />
5770 <input>
5771 <port id="0" precision="FP32">
5772 <dim>-1</dim>
5773 <dim>-1</dim>
5774 <dim>768</dim>
5775 </port>
5776 <port id="1" precision="FP32">
5777 <dim>1</dim>
5778 <dim>1</dim>
5779 <dim>768</dim>
5780 </port>
5781 </input>
5782 <output>
5783 <port id="2" precision="FP32" names="306,x.37">
5784 <dim>-1</dim>
5785 <dim>-1</dim>
5786 <dim>768</dim>
5787 </port>
5788 </output>
5789 </layer>
5790 <layer id="420" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5791 <data element_type="f32" shape="" offset="61312696" size="4" />
5792 <output>
5793 <port id="0" precision="FP32" />
5794 </output>
5795 </layer>
5796 <layer id="421" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
5797 <data element_type="f32" shape="" offset="61312700" size="4" />
5798 <output>
5799 <port id="0" precision="FP32" />
5800 </output>
5801 </layer>
5802 <layer id="422" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
5803 <data element_type="f32" shape="" offset="61312696" size="4" />
5804 <output>
5805 <port id="0" precision="FP32" />
5806 </output>
5807 </layer>
5808 <layer id="423" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
5809 <data element_type="f32" shape="" offset="61312700" size="4" />
5810 <output>
5811 <port id="0" precision="FP32" />
5812 </output>
5813 </layer>
5814 <layer id="424" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
5815 <data levels="256" auto_broadcast="numpy" />
5816 <input>
5817 <port id="0" precision="FP32">
5818 <dim>-1</dim>
5819 <dim>-1</dim>
5820 <dim>768</dim>
5821 </port>
5822 <port id="1" precision="FP32" />
5823 <port id="2" precision="FP32" />
5824 <port id="3" precision="FP32" />
5825 <port id="4" precision="FP32" />
5826 </input>
5827 <output>
5828 <port id="5" precision="FP32">
5829 <dim>-1</dim>
5830 <dim>-1</dim>
5831 <dim>768</dim>
5832 </port>
5833 </output>
5834 </layer>
5835 <layer id="425" name="__module.roberta.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
5836 <data element_type="i64" shape="4" offset="39811496" size="32" />
5837 <output>
5838 <port id="0" precision="I64">
5839 <dim>4</dim>
5840 </port>
5841 </output>
5842 </layer>
5843 <layer id="426" name="__module.roberta.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
5844 <data special_zero="true" />
5845 <input>
5846 <port id="0" precision="FP32">
5847 <dim>-1</dim>
5848 <dim>-1</dim>
5849 <dim>768</dim>
5850 </port>
5851 <port id="1" precision="I64">
5852 <dim>4</dim>
5853 </port>
5854 </input>
5855 <output>
5856 <port id="2" precision="FP32" names="310,x.39">
5857 <dim>-1</dim>
5858 <dim>-1</dim>
5859 <dim>12</dim>
5860 <dim>64</dim>
5861 </port>
5862 </output>
5863 </layer>
5864 <layer id="427" name="Constant_898" type="Const" version="opset1">
5865 <data element_type="i64" shape="4" offset="39811528" size="32" />
5866 <output>
5867 <port id="0" precision="I64" names="311">
5868 <dim>4</dim>
5869 </port>
5870 </output>
5871 </layer>
5872 <layer id="428" name="__module.roberta.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
5873 <input>
5874 <port id="0" precision="FP32">
5875 <dim>-1</dim>
5876 <dim>-1</dim>
5877 <dim>12</dim>
5878 <dim>64</dim>
5879 </port>
5880 <port id="1" precision="I64">
5881 <dim>4</dim>
5882 </port>
5883 </input>
5884 <output>
5885 <port id="2" precision="FP32" names="312">
5886 <dim>-1</dim>
5887 <dim>12</dim>
5888 <dim>-1</dim>
5889 <dim>64</dim>
5890 </port>
5891 </output>
5892 </layer>
5893 <layer id="429" name="Constant_220570" type="Const" version="opset1">
5894 <data element_type="i8" shape="768, 768" offset="61312704" size="589824" />
5895 <output>
5896 <port id="0" precision="I8">
5897 <dim>768</dim>
5898 <dim>768</dim>
5899 </port>
5900 </output>
5901 </layer>
5902 <layer id="430" name="Convert_220571" type="Convert" version="opset1">
5903 <data destination_type="f32" />
5904 <input>
5905 <port id="0" precision="I8">
5906 <dim>768</dim>
5907 <dim>768</dim>
5908 </port>
5909 </input>
5910 <output>
5911 <port id="1" precision="FP32">
5912 <dim>768</dim>
5913 <dim>768</dim>
5914 </port>
5915 </output>
5916 </layer>
5917 <layer id="431" name="Constant_220572" type="Const" version="opset1">
5918 <data element_type="f32" shape="768, 1" offset="61902528" size="3072" />
5919 <output>
5920 <port id="0" precision="FP32">
5921 <dim>768</dim>
5922 <dim>1</dim>
5923 </port>
5924 </output>
5925 </layer>
5926 <layer id="432" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
5927 <data auto_broadcast="numpy" />
5928 <input>
5929 <port id="0" precision="FP32">
5930 <dim>768</dim>
5931 <dim>768</dim>
5932 </port>
5933 <port id="1" precision="FP32">
5934 <dim>768</dim>
5935 <dim>1</dim>
5936 </port>
5937 </input>
5938 <output>
5939 <port id="2" precision="FP32">
5940 <dim>768</dim>
5941 <dim>768</dim>
5942 </port>
5943 </output>
5944 </layer>
5945 <layer id="433" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
5946 <data transpose_a="false" transpose_b="true" />
5947 <input>
5948 <port id="0" precision="FP32">
5949 <dim>-1</dim>
5950 <dim>-1</dim>
5951 <dim>768</dim>
5952 </port>
5953 <port id="1" precision="FP32">
5954 <dim>768</dim>
5955 <dim>768</dim>
5956 </port>
5957 </input>
5958 <output>
5959 <port id="2" precision="FP32">
5960 <dim>-1</dim>
5961 <dim>-1</dim>
5962 <dim>768</dim>
5963 </port>
5964 </output>
5965 </layer>
5966 <layer id="434" name="Constant_5393" type="Const" version="opset1">
5967 <data element_type="f32" shape="1, 1, 768" offset="61905600" size="3072" />
5968 <output>
5969 <port id="0" precision="FP32">
5970 <dim>1</dim>
5971 <dim>1</dim>
5972 <dim>768</dim>
5973 </port>
5974 </output>
5975 </layer>
5976 <layer id="435" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
5977 <data auto_broadcast="numpy" />
5978 <input>
5979 <port id="0" precision="FP32">
5980 <dim>-1</dim>
5981 <dim>-1</dim>
5982 <dim>768</dim>
5983 </port>
5984 <port id="1" precision="FP32">
5985 <dim>1</dim>
5986 <dim>1</dim>
5987 <dim>768</dim>
5988 </port>
5989 </input>
5990 <output>
5991 <port id="2" precision="FP32" names="315,x.41">
5992 <dim>-1</dim>
5993 <dim>-1</dim>
5994 <dim>768</dim>
5995 </port>
5996 </output>
5997 </layer>
5998 <layer id="436" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
5999 <data element_type="f32" shape="" offset="61908672" size="4" />
6000 <output>
6001 <port id="0" precision="FP32" />
6002 </output>
6003 </layer>
6004 <layer id="437" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
6005 <data element_type="f32" shape="" offset="61908676" size="4" />
6006 <output>
6007 <port id="0" precision="FP32" />
6008 </output>
6009 </layer>
6010 <layer id="438" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
6011 <data element_type="f32" shape="" offset="61908672" size="4" />
6012 <output>
6013 <port id="0" precision="FP32" />
6014 </output>
6015 </layer>
6016 <layer id="439" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
6017 <data element_type="f32" shape="" offset="61908676" size="4" />
6018 <output>
6019 <port id="0" precision="FP32" />
6020 </output>
6021 </layer>
6022 <layer id="440" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
6023 <data levels="256" auto_broadcast="numpy" />
6024 <input>
6025 <port id="0" precision="FP32">
6026 <dim>-1</dim>
6027 <dim>-1</dim>
6028 <dim>768</dim>
6029 </port>
6030 <port id="1" precision="FP32" />
6031 <port id="2" precision="FP32" />
6032 <port id="3" precision="FP32" />
6033 <port id="4" precision="FP32" />
6034 </input>
6035 <output>
6036 <port id="5" precision="FP32">
6037 <dim>-1</dim>
6038 <dim>-1</dim>
6039 <dim>768</dim>
6040 </port>
6041 </output>
6042 </layer>
6043 <layer id="441" name="__module.roberta.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
6044 <data element_type="i64" shape="4" offset="39811496" size="32" />
6045 <output>
6046 <port id="0" precision="I64">
6047 <dim>4</dim>
6048 </port>
6049 </output>
6050 </layer>
6051 <layer id="442" name="__module.roberta.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
6052 <data special_zero="true" />
6053 <input>
6054 <port id="0" precision="FP32">
6055 <dim>-1</dim>
6056 <dim>-1</dim>
6057 <dim>768</dim>
6058 </port>
6059 <port id="1" precision="I64">
6060 <dim>4</dim>
6061 </port>
6062 </input>
6063 <output>
6064 <port id="2" precision="FP32" names="319,x.43">
6065 <dim>-1</dim>
6066 <dim>-1</dim>
6067 <dim>12</dim>
6068 <dim>64</dim>
6069 </port>
6070 </output>
6071 </layer>
6072 <layer id="443" name="Constant_921" type="Const" version="opset1">
6073 <data element_type="i64" shape="4" offset="39811528" size="32" />
6074 <output>
6075 <port id="0" precision="I64" names="320">
6076 <dim>4</dim>
6077 </port>
6078 </output>
6079 </layer>
6080 <layer id="444" name="__module.roberta.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
6081 <input>
6082 <port id="0" precision="FP32">
6083 <dim>-1</dim>
6084 <dim>-1</dim>
6085 <dim>12</dim>
6086 <dim>64</dim>
6087 </port>
6088 <port id="1" precision="I64">
6089 <dim>4</dim>
6090 </port>
6091 </input>
6092 <output>
6093 <port id="2" precision="FP32" names="321">
6094 <dim>-1</dim>
6095 <dim>12</dim>
6096 <dim>-1</dim>
6097 <dim>64</dim>
6098 </port>
6099 </output>
6100 </layer>
6101 <layer id="445" name="Constant_220574" type="Const" version="opset1">
6102 <data element_type="i8" shape="768, 768" offset="61908680" size="589824" />
6103 <output>
6104 <port id="0" precision="I8">
6105 <dim>768</dim>
6106 <dim>768</dim>
6107 </port>
6108 </output>
6109 </layer>
6110 <layer id="446" name="Convert_220575" type="Convert" version="opset1">
6111 <data destination_type="f32" />
6112 <input>
6113 <port id="0" precision="I8">
6114 <dim>768</dim>
6115 <dim>768</dim>
6116 </port>
6117 </input>
6118 <output>
6119 <port id="1" precision="FP32">
6120 <dim>768</dim>
6121 <dim>768</dim>
6122 </port>
6123 </output>
6124 </layer>
6125 <layer id="447" name="Constant_220576" type="Const" version="opset1">
6126 <data element_type="f32" shape="768, 1" offset="62498504" size="3072" />
6127 <output>
6128 <port id="0" precision="FP32">
6129 <dim>768</dim>
6130 <dim>1</dim>
6131 </port>
6132 </output>
6133 </layer>
6134 <layer id="448" name="__module.roberta.encoder.layer.3.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6135 <data auto_broadcast="numpy" />
6136 <input>
6137 <port id="0" precision="FP32">
6138 <dim>768</dim>
6139 <dim>768</dim>
6140 </port>
6141 <port id="1" precision="FP32">
6142 <dim>768</dim>
6143 <dim>1</dim>
6144 </port>
6145 </input>
6146 <output>
6147 <port id="2" precision="FP32">
6148 <dim>768</dim>
6149 <dim>768</dim>
6150 </port>
6151 </output>
6152 </layer>
6153 <layer id="449" name="__module.roberta.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
6154 <data transpose_a="false" transpose_b="true" />
6155 <input>
6156 <port id="0" precision="FP32">
6157 <dim>-1</dim>
6158 <dim>-1</dim>
6159 <dim>768</dim>
6160 </port>
6161 <port id="1" precision="FP32">
6162 <dim>768</dim>
6163 <dim>768</dim>
6164 </port>
6165 </input>
6166 <output>
6167 <port id="2" precision="FP32">
6168 <dim>-1</dim>
6169 <dim>-1</dim>
6170 <dim>768</dim>
6171 </port>
6172 </output>
6173 </layer>
6174 <layer id="450" name="Constant_5394" type="Const" version="opset1">
6175 <data element_type="f32" shape="1, 1, 768" offset="62501576" size="3072" />
6176 <output>
6177 <port id="0" precision="FP32">
6178 <dim>1</dim>
6179 <dim>1</dim>
6180 <dim>768</dim>
6181 </port>
6182 </output>
6183 </layer>
6184 <layer id="451" name="__module.roberta.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
6185 <data auto_broadcast="numpy" />
6186 <input>
6187 <port id="0" precision="FP32">
6188 <dim>-1</dim>
6189 <dim>-1</dim>
6190 <dim>768</dim>
6191 </port>
6192 <port id="1" precision="FP32">
6193 <dim>1</dim>
6194 <dim>1</dim>
6195 <dim>768</dim>
6196 </port>
6197 </input>
6198 <output>
6199 <port id="2" precision="FP32" names="324,x.45">
6200 <dim>-1</dim>
6201 <dim>-1</dim>
6202 <dim>768</dim>
6203 </port>
6204 </output>
6205 </layer>
6206 <layer id="452" name="__module.roberta.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
6207 <data element_type="i64" shape="4" offset="39811496" size="32" />
6208 <output>
6209 <port id="0" precision="I64">
6210 <dim>4</dim>
6211 </port>
6212 </output>
6213 </layer>
6214 <layer id="453" name="__module.roberta.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
6215 <data special_zero="true" />
6216 <input>
6217 <port id="0" precision="FP32">
6218 <dim>-1</dim>
6219 <dim>-1</dim>
6220 <dim>768</dim>
6221 </port>
6222 <port id="1" precision="I64">
6223 <dim>4</dim>
6224 </port>
6225 </input>
6226 <output>
6227 <port id="2" precision="FP32" names="328,x.47">
6228 <dim>-1</dim>
6229 <dim>-1</dim>
6230 <dim>12</dim>
6231 <dim>64</dim>
6232 </port>
6233 </output>
6234 </layer>
6235 <layer id="454" name="Constant_944" type="Const" version="opset1">
6236 <data element_type="i64" shape="4" offset="39811528" size="32" />
6237 <output>
6238 <port id="0" precision="I64" names="329">
6239 <dim>4</dim>
6240 </port>
6241 </output>
6242 </layer>
6243 <layer id="455" name="__module.roberta.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
6244 <input>
6245 <port id="0" precision="FP32">
6246 <dim>-1</dim>
6247 <dim>-1</dim>
6248 <dim>12</dim>
6249 <dim>64</dim>
6250 </port>
6251 <port id="1" precision="I64">
6252 <dim>4</dim>
6253 </port>
6254 </input>
6255 <output>
6256 <port id="2" precision="FP32" names="330">
6257 <dim>-1</dim>
6258 <dim>12</dim>
6259 <dim>-1</dim>
6260 <dim>64</dim>
6261 </port>
6262 </output>
6263 </layer>
6264 <layer id="456" name="__module.roberta.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
6265 <data causal="false" />
6266 <input>
6267 <port id="0" precision="FP32">
6268 <dim>-1</dim>
6269 <dim>12</dim>
6270 <dim>-1</dim>
6271 <dim>64</dim>
6272 </port>
6273 <port id="1" precision="FP32">
6274 <dim>-1</dim>
6275 <dim>12</dim>
6276 <dim>-1</dim>
6277 <dim>64</dim>
6278 </port>
6279 <port id="2" precision="FP32">
6280 <dim>-1</dim>
6281 <dim>12</dim>
6282 <dim>-1</dim>
6283 <dim>64</dim>
6284 </port>
6285 <port id="3" precision="FP32">
6286 <dim>-1</dim>
6287 <dim>1</dim>
6288 <dim>-1</dim>
6289 <dim>-1</dim>
6290 </port>
6291 </input>
6292 <output>
6293 <port id="4" precision="FP32" names="331,attn_output.13">
6294 <dim>-1</dim>
6295 <dim>12</dim>
6296 <dim>-1</dim>
6297 <dim>64</dim>
6298 </port>
6299 </output>
6300 </layer>
6301 <layer id="457" name="__module.roberta.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
6302 <data element_type="i32" shape="4" offset="41003536" size="16" />
6303 <output>
6304 <port id="0" precision="I32">
6305 <dim>4</dim>
6306 </port>
6307 </output>
6308 </layer>
6309 <layer id="458" name="__module.roberta.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
6310 <input>
6311 <port id="0" precision="FP32">
6312 <dim>-1</dim>
6313 <dim>12</dim>
6314 <dim>-1</dim>
6315 <dim>64</dim>
6316 </port>
6317 <port id="1" precision="I32">
6318 <dim>4</dim>
6319 </port>
6320 </input>
6321 <output>
6322 <port id="2" precision="FP32" names="332,attn_output.15">
6323 <dim>-1</dim>
6324 <dim>-1</dim>
6325 <dim>12</dim>
6326 <dim>64</dim>
6327 </port>
6328 </output>
6329 </layer>
6330 <layer id="459" name="Constant_5497" type="Const" version="opset1">
6331 <data element_type="i64" shape="3" offset="41003552" size="24" />
6332 <output>
6333 <port id="0" precision="I64">
6334 <dim>3</dim>
6335 </port>
6336 </output>
6337 </layer>
6338 <layer id="460" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
6339 <data special_zero="true" />
6340 <input>
6341 <port id="0" precision="FP32">
6342 <dim>-1</dim>
6343 <dim>-1</dim>
6344 <dim>12</dim>
6345 <dim>64</dim>
6346 </port>
6347 <port id="1" precision="I64">
6348 <dim>3</dim>
6349 </port>
6350 </input>
6351 <output>
6352 <port id="2" precision="FP32" names="334">
6353 <dim>-1</dim>
6354 <dim>-1</dim>
6355 <dim>768</dim>
6356 </port>
6357 </output>
6358 </layer>
6359 <layer id="461" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6360 <data element_type="f32" shape="1, 1, 768" offset="62504648" size="3072" />
6361 <output>
6362 <port id="0" precision="FP32">
6363 <dim>1</dim>
6364 <dim>1</dim>
6365 <dim>768</dim>
6366 </port>
6367 </output>
6368 </layer>
6369 <layer id="462" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6370 <data auto_broadcast="numpy" />
6371 <input>
6372 <port id="0" precision="FP32">
6373 <dim>-1</dim>
6374 <dim>-1</dim>
6375 <dim>768</dim>
6376 </port>
6377 <port id="1" precision="FP32">
6378 <dim>1</dim>
6379 <dim>1</dim>
6380 <dim>768</dim>
6381 </port>
6382 </input>
6383 <output>
6384 <port id="2" precision="FP32">
6385 <dim>-1</dim>
6386 <dim>-1</dim>
6387 <dim>768</dim>
6388 </port>
6389 </output>
6390 </layer>
6391 <layer id="463" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6392 <data element_type="f32" shape="" offset="62507720" size="4" />
6393 <output>
6394 <port id="0" precision="FP32" />
6395 </output>
6396 </layer>
6397 <layer id="464" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6398 <data element_type="f32" shape="" offset="62507724" size="4" />
6399 <output>
6400 <port id="0" precision="FP32" />
6401 </output>
6402 </layer>
6403 <layer id="465" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6404 <data element_type="f32" shape="" offset="62507720" size="4" />
6405 <output>
6406 <port id="0" precision="FP32" />
6407 </output>
6408 </layer>
6409 <layer id="466" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6410 <data element_type="f32" shape="" offset="62507724" size="4" />
6411 <output>
6412 <port id="0" precision="FP32" />
6413 </output>
6414 </layer>
6415 <layer id="467" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6416 <data levels="256" auto_broadcast="numpy" />
6417 <input>
6418 <port id="0" precision="FP32">
6419 <dim>-1</dim>
6420 <dim>-1</dim>
6421 <dim>768</dim>
6422 </port>
6423 <port id="1" precision="FP32" />
6424 <port id="2" precision="FP32" />
6425 <port id="3" precision="FP32" />
6426 <port id="4" precision="FP32" />
6427 </input>
6428 <output>
6429 <port id="5" precision="FP32">
6430 <dim>-1</dim>
6431 <dim>-1</dim>
6432 <dim>768</dim>
6433 </port>
6434 </output>
6435 </layer>
6436 <layer id="468" name="Constant_220578" type="Const" version="opset1">
6437 <data element_type="i8" shape="768, 768" offset="62507728" size="589824" />
6438 <output>
6439 <port id="0" precision="I8">
6440 <dim>768</dim>
6441 <dim>768</dim>
6442 </port>
6443 </output>
6444 </layer>
6445 <layer id="469" name="Convert_220579" type="Convert" version="opset1">
6446 <data destination_type="f32" />
6447 <input>
6448 <port id="0" precision="I8">
6449 <dim>768</dim>
6450 <dim>768</dim>
6451 </port>
6452 </input>
6453 <output>
6454 <port id="1" precision="FP32">
6455 <dim>768</dim>
6456 <dim>768</dim>
6457 </port>
6458 </output>
6459 </layer>
6460 <layer id="470" name="Constant_220580" type="Const" version="opset1">
6461 <data element_type="f32" shape="768, 1" offset="63097552" size="3072" />
6462 <output>
6463 <port id="0" precision="FP32">
6464 <dim>768</dim>
6465 <dim>1</dim>
6466 </port>
6467 </output>
6468 </layer>
6469 <layer id="471" name="__module.roberta.encoder.layer.3.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6470 <data auto_broadcast="numpy" />
6471 <input>
6472 <port id="0" precision="FP32">
6473 <dim>768</dim>
6474 <dim>768</dim>
6475 </port>
6476 <port id="1" precision="FP32">
6477 <dim>768</dim>
6478 <dim>1</dim>
6479 </port>
6480 </input>
6481 <output>
6482 <port id="2" precision="FP32">
6483 <dim>768</dim>
6484 <dim>768</dim>
6485 </port>
6486 </output>
6487 </layer>
6488 <layer id="472" name="__module.roberta.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6489 <data transpose_a="false" transpose_b="true" />
6490 <input>
6491 <port id="0" precision="FP32">
6492 <dim>-1</dim>
6493 <dim>-1</dim>
6494 <dim>768</dim>
6495 </port>
6496 <port id="1" precision="FP32">
6497 <dim>768</dim>
6498 <dim>768</dim>
6499 </port>
6500 </input>
6501 <output>
6502 <port id="2" precision="FP32">
6503 <dim>-1</dim>
6504 <dim>-1</dim>
6505 <dim>768</dim>
6506 </port>
6507 </output>
6508 </layer>
6509 <layer id="473" name="Constant_5395" type="Const" version="opset1">
6510 <data element_type="f32" shape="1, 1, 768" offset="63100624" size="3072" />
6511 <output>
6512 <port id="0" precision="FP32">
6513 <dim>1</dim>
6514 <dim>1</dim>
6515 <dim>768</dim>
6516 </port>
6517 </output>
6518 </layer>
6519 <layer id="474" name="__module.roberta.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
6520 <data auto_broadcast="numpy" />
6521 <input>
6522 <port id="0" precision="FP32">
6523 <dim>-1</dim>
6524 <dim>-1</dim>
6525 <dim>768</dim>
6526 </port>
6527 <port id="1" precision="FP32">
6528 <dim>1</dim>
6529 <dim>1</dim>
6530 <dim>768</dim>
6531 </port>
6532 </input>
6533 <output>
6534 <port id="2" precision="FP32" names="339,input.15">
6535 <dim>-1</dim>
6536 <dim>-1</dim>
6537 <dim>768</dim>
6538 </port>
6539 </output>
6540 </layer>
6541 <layer id="475" name="__module.roberta.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
6542 <data auto_broadcast="numpy" />
6543 <input>
6544 <port id="0" precision="FP32">
6545 <dim>-1</dim>
6546 <dim>-1</dim>
6547 <dim>768</dim>
6548 </port>
6549 <port id="1" precision="FP32">
6550 <dim>-1</dim>
6551 <dim>-1</dim>
6552 <dim>768</dim>
6553 </port>
6554 </input>
6555 <output>
6556 <port id="2" precision="FP32" names="341">
6557 <dim>-1</dim>
6558 <dim>-1</dim>
6559 <dim>768</dim>
6560 </port>
6561 </output>
6562 </layer>
6563 <layer id="476" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6564 <data element_type="i32" shape="1" offset="39206292" size="4" />
6565 <output>
6566 <port id="0" precision="I32">
6567 <dim>1</dim>
6568 </port>
6569 </output>
6570 </layer>
6571 <layer id="477" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6572 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6573 <input>
6574 <port id="0" precision="FP32">
6575 <dim>-1</dim>
6576 <dim>-1</dim>
6577 <dim>768</dim>
6578 </port>
6579 <port id="1" precision="I32">
6580 <dim>1</dim>
6581 </port>
6582 </input>
6583 <output>
6584 <port id="2" precision="FP32">
6585 <dim>-1</dim>
6586 <dim>-1</dim>
6587 <dim>768</dim>
6588 </port>
6589 </output>
6590 </layer>
6591 <layer id="478" name="Constant_5396" type="Const" version="opset1">
6592 <data element_type="f32" shape="1, 1, 768" offset="63103696" size="3072" />
6593 <output>
6594 <port id="0" precision="FP32">
6595 <dim>1</dim>
6596 <dim>1</dim>
6597 <dim>768</dim>
6598 </port>
6599 </output>
6600 </layer>
6601 <layer id="479" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6602 <data auto_broadcast="numpy" />
6603 <input>
6604 <port id="0" precision="FP32">
6605 <dim>-1</dim>
6606 <dim>-1</dim>
6607 <dim>768</dim>
6608 </port>
6609 <port id="1" precision="FP32">
6610 <dim>1</dim>
6611 <dim>1</dim>
6612 <dim>768</dim>
6613 </port>
6614 </input>
6615 <output>
6616 <port id="2" precision="FP32">
6617 <dim>-1</dim>
6618 <dim>-1</dim>
6619 <dim>768</dim>
6620 </port>
6621 </output>
6622 </layer>
6623 <layer id="480" name="Constant_5397" type="Const" version="opset1">
6624 <data element_type="f32" shape="1, 1, 768" offset="63106768" size="3072" />
6625 <output>
6626 <port id="0" precision="FP32">
6627 <dim>1</dim>
6628 <dim>1</dim>
6629 <dim>768</dim>
6630 </port>
6631 </output>
6632 </layer>
6633 <layer id="481" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6634 <data auto_broadcast="numpy" />
6635 <input>
6636 <port id="0" precision="FP32">
6637 <dim>-1</dim>
6638 <dim>-1</dim>
6639 <dim>768</dim>
6640 </port>
6641 <port id="1" precision="FP32">
6642 <dim>1</dim>
6643 <dim>1</dim>
6644 <dim>768</dim>
6645 </port>
6646 </input>
6647 <output>
6648 <port id="2" precision="FP32" names="345,input_tensor.7">
6649 <dim>-1</dim>
6650 <dim>-1</dim>
6651 <dim>768</dim>
6652 </port>
6653 </output>
6654 </layer>
6655 <layer id="482" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6656 <data element_type="f32" shape="1, 1, 768" offset="63109840" size="3072" />
6657 <output>
6658 <port id="0" precision="FP32">
6659 <dim>1</dim>
6660 <dim>1</dim>
6661 <dim>768</dim>
6662 </port>
6663 </output>
6664 </layer>
6665 <layer id="483" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6666 <data auto_broadcast="numpy" />
6667 <input>
6668 <port id="0" precision="FP32">
6669 <dim>-1</dim>
6670 <dim>-1</dim>
6671 <dim>768</dim>
6672 </port>
6673 <port id="1" precision="FP32">
6674 <dim>1</dim>
6675 <dim>1</dim>
6676 <dim>768</dim>
6677 </port>
6678 </input>
6679 <output>
6680 <port id="2" precision="FP32">
6681 <dim>-1</dim>
6682 <dim>-1</dim>
6683 <dim>768</dim>
6684 </port>
6685 </output>
6686 </layer>
6687 <layer id="484" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6688 <data element_type="f32" shape="" offset="63112912" size="4" />
6689 <output>
6690 <port id="0" precision="FP32" />
6691 </output>
6692 </layer>
6693 <layer id="485" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6694 <data element_type="f32" shape="" offset="63112916" size="4" />
6695 <output>
6696 <port id="0" precision="FP32" />
6697 </output>
6698 </layer>
6699 <layer id="486" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6700 <data element_type="f32" shape="" offset="63112912" size="4" />
6701 <output>
6702 <port id="0" precision="FP32" />
6703 </output>
6704 </layer>
6705 <layer id="487" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6706 <data element_type="f32" shape="" offset="63112916" size="4" />
6707 <output>
6708 <port id="0" precision="FP32" />
6709 </output>
6710 </layer>
6711 <layer id="488" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6712 <data levels="256" auto_broadcast="numpy" />
6713 <input>
6714 <port id="0" precision="FP32">
6715 <dim>-1</dim>
6716 <dim>-1</dim>
6717 <dim>768</dim>
6718 </port>
6719 <port id="1" precision="FP32" />
6720 <port id="2" precision="FP32" />
6721 <port id="3" precision="FP32" />
6722 <port id="4" precision="FP32" />
6723 </input>
6724 <output>
6725 <port id="5" precision="FP32">
6726 <dim>-1</dim>
6727 <dim>-1</dim>
6728 <dim>768</dim>
6729 </port>
6730 </output>
6731 </layer>
6732 <layer id="489" name="Constant_220582" type="Const" version="opset1">
6733 <data element_type="i8" shape="3072, 768" offset="63112920" size="2359296" />
6734 <output>
6735 <port id="0" precision="I8">
6736 <dim>3072</dim>
6737 <dim>768</dim>
6738 </port>
6739 </output>
6740 </layer>
6741 <layer id="490" name="Convert_220583" type="Convert" version="opset1">
6742 <data destination_type="f32" />
6743 <input>
6744 <port id="0" precision="I8">
6745 <dim>3072</dim>
6746 <dim>768</dim>
6747 </port>
6748 </input>
6749 <output>
6750 <port id="1" precision="FP32">
6751 <dim>3072</dim>
6752 <dim>768</dim>
6753 </port>
6754 </output>
6755 </layer>
6756 <layer id="491" name="Constant_220584" type="Const" version="opset1">
6757 <data element_type="f32" shape="3072, 1" offset="65472216" size="12288" />
6758 <output>
6759 <port id="0" precision="FP32">
6760 <dim>3072</dim>
6761 <dim>1</dim>
6762 </port>
6763 </output>
6764 </layer>
6765 <layer id="492" name="__module.roberta.encoder.layer.3.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6766 <data auto_broadcast="numpy" />
6767 <input>
6768 <port id="0" precision="FP32">
6769 <dim>3072</dim>
6770 <dim>768</dim>
6771 </port>
6772 <port id="1" precision="FP32">
6773 <dim>3072</dim>
6774 <dim>1</dim>
6775 </port>
6776 </input>
6777 <output>
6778 <port id="2" precision="FP32">
6779 <dim>3072</dim>
6780 <dim>768</dim>
6781 </port>
6782 </output>
6783 </layer>
6784 <layer id="493" name="__module.roberta.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6785 <data transpose_a="false" transpose_b="true" />
6786 <input>
6787 <port id="0" precision="FP32">
6788 <dim>-1</dim>
6789 <dim>-1</dim>
6790 <dim>768</dim>
6791 </port>
6792 <port id="1" precision="FP32">
6793 <dim>3072</dim>
6794 <dim>768</dim>
6795 </port>
6796 </input>
6797 <output>
6798 <port id="2" precision="FP32">
6799 <dim>-1</dim>
6800 <dim>-1</dim>
6801 <dim>3072</dim>
6802 </port>
6803 </output>
6804 </layer>
6805 <layer id="494" name="Constant_5398" type="Const" version="opset1">
6806 <data element_type="f32" shape="1, 1, 3072" offset="65484504" size="12288" />
6807 <output>
6808 <port id="0" precision="FP32">
6809 <dim>1</dim>
6810 <dim>1</dim>
6811 <dim>3072</dim>
6812 </port>
6813 </output>
6814 </layer>
6815 <layer id="495" name="__module.roberta.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
6816 <data auto_broadcast="numpy" />
6817 <input>
6818 <port id="0" precision="FP32">
6819 <dim>-1</dim>
6820 <dim>-1</dim>
6821 <dim>3072</dim>
6822 </port>
6823 <port id="1" precision="FP32">
6824 <dim>1</dim>
6825 <dim>1</dim>
6826 <dim>3072</dim>
6827 </port>
6828 </input>
6829 <output>
6830 <port id="2" precision="FP32" names="349">
6831 <dim>-1</dim>
6832 <dim>-1</dim>
6833 <dim>3072</dim>
6834 </port>
6835 </output>
6836 </layer>
6837 <layer id="496" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
6838 <data approximation_mode="ERF" />
6839 <input>
6840 <port id="0" precision="FP32">
6841 <dim>-1</dim>
6842 <dim>-1</dim>
6843 <dim>3072</dim>
6844 </port>
6845 </input>
6846 <output>
6847 <port id="1" precision="FP32" names="350">
6848 <dim>-1</dim>
6849 <dim>-1</dim>
6850 <dim>3072</dim>
6851 </port>
6852 </output>
6853 </layer>
6854 <layer id="497" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
6855 <data element_type="f32" shape="1, 1, 3072" offset="65496792" size="12288" />
6856 <output>
6857 <port id="0" precision="FP32">
6858 <dim>1</dim>
6859 <dim>1</dim>
6860 <dim>3072</dim>
6861 </port>
6862 </output>
6863 </layer>
6864 <layer id="498" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
6865 <data auto_broadcast="numpy" />
6866 <input>
6867 <port id="0" precision="FP32">
6868 <dim>-1</dim>
6869 <dim>-1</dim>
6870 <dim>3072</dim>
6871 </port>
6872 <port id="1" precision="FP32">
6873 <dim>1</dim>
6874 <dim>1</dim>
6875 <dim>3072</dim>
6876 </port>
6877 </input>
6878 <output>
6879 <port id="2" precision="FP32">
6880 <dim>-1</dim>
6881 <dim>-1</dim>
6882 <dim>3072</dim>
6883 </port>
6884 </output>
6885 </layer>
6886 <layer id="499" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
6887 <data element_type="f32" shape="" offset="65509080" size="4" />
6888 <output>
6889 <port id="0" precision="FP32" />
6890 </output>
6891 </layer>
6892 <layer id="500" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
6893 <data element_type="f32" shape="" offset="65509084" size="4" />
6894 <output>
6895 <port id="0" precision="FP32" />
6896 </output>
6897 </layer>
6898 <layer id="501" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
6899 <data element_type="f32" shape="" offset="65509080" size="4" />
6900 <output>
6901 <port id="0" precision="FP32" />
6902 </output>
6903 </layer>
6904 <layer id="502" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
6905 <data element_type="f32" shape="" offset="65509084" size="4" />
6906 <output>
6907 <port id="0" precision="FP32" />
6908 </output>
6909 </layer>
6910 <layer id="503" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
6911 <data levels="256" auto_broadcast="numpy" />
6912 <input>
6913 <port id="0" precision="FP32">
6914 <dim>-1</dim>
6915 <dim>-1</dim>
6916 <dim>3072</dim>
6917 </port>
6918 <port id="1" precision="FP32" />
6919 <port id="2" precision="FP32" />
6920 <port id="3" precision="FP32" />
6921 <port id="4" precision="FP32" />
6922 </input>
6923 <output>
6924 <port id="5" precision="FP32">
6925 <dim>-1</dim>
6926 <dim>-1</dim>
6927 <dim>3072</dim>
6928 </port>
6929 </output>
6930 </layer>
6931 <layer id="504" name="Constant_220586" type="Const" version="opset1">
6932 <data element_type="i8" shape="768, 3072" offset="65509088" size="2359296" />
6933 <output>
6934 <port id="0" precision="I8">
6935 <dim>768</dim>
6936 <dim>3072</dim>
6937 </port>
6938 </output>
6939 </layer>
6940 <layer id="505" name="Convert_220587" type="Convert" version="opset1">
6941 <data destination_type="f32" />
6942 <input>
6943 <port id="0" precision="I8">
6944 <dim>768</dim>
6945 <dim>3072</dim>
6946 </port>
6947 </input>
6948 <output>
6949 <port id="1" precision="FP32">
6950 <dim>768</dim>
6951 <dim>3072</dim>
6952 </port>
6953 </output>
6954 </layer>
6955 <layer id="506" name="Constant_220588" type="Const" version="opset1">
6956 <data element_type="f32" shape="768, 1" offset="67868384" size="3072" />
6957 <output>
6958 <port id="0" precision="FP32">
6959 <dim>768</dim>
6960 <dim>1</dim>
6961 </port>
6962 </output>
6963 </layer>
6964 <layer id="507" name="__module.roberta.encoder.layer.3.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
6965 <data auto_broadcast="numpy" />
6966 <input>
6967 <port id="0" precision="FP32">
6968 <dim>768</dim>
6969 <dim>3072</dim>
6970 </port>
6971 <port id="1" precision="FP32">
6972 <dim>768</dim>
6973 <dim>1</dim>
6974 </port>
6975 </input>
6976 <output>
6977 <port id="2" precision="FP32">
6978 <dim>768</dim>
6979 <dim>3072</dim>
6980 </port>
6981 </output>
6982 </layer>
6983 <layer id="508" name="__module.roberta.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6984 <data transpose_a="false" transpose_b="true" />
6985 <input>
6986 <port id="0" precision="FP32">
6987 <dim>-1</dim>
6988 <dim>-1</dim>
6989 <dim>3072</dim>
6990 </port>
6991 <port id="1" precision="FP32">
6992 <dim>768</dim>
6993 <dim>3072</dim>
6994 </port>
6995 </input>
6996 <output>
6997 <port id="2" precision="FP32">
6998 <dim>-1</dim>
6999 <dim>-1</dim>
7000 <dim>768</dim>
7001 </port>
7002 </output>
7003 </layer>
7004 <layer id="509" name="Constant_5399" type="Const" version="opset1">
7005 <data element_type="f32" shape="1, 1, 768" offset="67871456" size="3072" />
7006 <output>
7007 <port id="0" precision="FP32">
7008 <dim>1</dim>
7009 <dim>1</dim>
7010 <dim>768</dim>
7011 </port>
7012 </output>
7013 </layer>
7014 <layer id="510" name="__module.roberta.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
7015 <data auto_broadcast="numpy" />
7016 <input>
7017 <port id="0" precision="FP32">
7018 <dim>-1</dim>
7019 <dim>-1</dim>
7020 <dim>768</dim>
7021 </port>
7022 <port id="1" precision="FP32">
7023 <dim>1</dim>
7024 <dim>1</dim>
7025 <dim>768</dim>
7026 </port>
7027 </input>
7028 <output>
7029 <port id="2" precision="FP32" names="355,input.17">
7030 <dim>-1</dim>
7031 <dim>-1</dim>
7032 <dim>768</dim>
7033 </port>
7034 </output>
7035 </layer>
7036 <layer id="511" name="__module.roberta.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
7037 <data auto_broadcast="numpy" />
7038 <input>
7039 <port id="0" precision="FP32">
7040 <dim>-1</dim>
7041 <dim>-1</dim>
7042 <dim>768</dim>
7043 </port>
7044 <port id="1" precision="FP32">
7045 <dim>-1</dim>
7046 <dim>-1</dim>
7047 <dim>768</dim>
7048 </port>
7049 </input>
7050 <output>
7051 <port id="2" precision="FP32" names="357">
7052 <dim>-1</dim>
7053 <dim>-1</dim>
7054 <dim>768</dim>
7055 </port>
7056 </output>
7057 </layer>
7058 <layer id="512" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
7059 <data element_type="i32" shape="1" offset="39206292" size="4" />
7060 <output>
7061 <port id="0" precision="I32">
7062 <dim>1</dim>
7063 </port>
7064 </output>
7065 </layer>
7066 <layer id="513" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
7067 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
7068 <input>
7069 <port id="0" precision="FP32">
7070 <dim>-1</dim>
7071 <dim>-1</dim>
7072 <dim>768</dim>
7073 </port>
7074 <port id="1" precision="I32">
7075 <dim>1</dim>
7076 </port>
7077 </input>
7078 <output>
7079 <port id="2" precision="FP32">
7080 <dim>-1</dim>
7081 <dim>-1</dim>
7082 <dim>768</dim>
7083 </port>
7084 </output>
7085 </layer>
7086 <layer id="514" name="Constant_5400" type="Const" version="opset1">
7087 <data element_type="f32" shape="1, 1, 768" offset="67874528" size="3072" />
7088 <output>
7089 <port id="0" precision="FP32">
7090 <dim>1</dim>
7091 <dim>1</dim>
7092 <dim>768</dim>
7093 </port>
7094 </output>
7095 </layer>
7096 <layer id="515" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
7097 <data auto_broadcast="numpy" />
7098 <input>
7099 <port id="0" precision="FP32">
7100 <dim>-1</dim>
7101 <dim>-1</dim>
7102 <dim>768</dim>
7103 </port>
7104 <port id="1" precision="FP32">
7105 <dim>1</dim>
7106 <dim>1</dim>
7107 <dim>768</dim>
7108 </port>
7109 </input>
7110 <output>
7111 <port id="2" precision="FP32">
7112 <dim>-1</dim>
7113 <dim>-1</dim>
7114 <dim>768</dim>
7115 </port>
7116 </output>
7117 </layer>
7118 <layer id="516" name="Constant_5401" type="Const" version="opset1">
7119 <data element_type="f32" shape="1, 1, 768" offset="67877600" size="3072" />
7120 <output>
7121 <port id="0" precision="FP32">
7122 <dim>1</dim>
7123 <dim>1</dim>
7124 <dim>768</dim>
7125 </port>
7126 </output>
7127 </layer>
7128 <layer id="517" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
7129 <data auto_broadcast="numpy" />
7130 <input>
7131 <port id="0" precision="FP32">
7132 <dim>-1</dim>
7133 <dim>-1</dim>
7134 <dim>768</dim>
7135 </port>
7136 <port id="1" precision="FP32">
7137 <dim>1</dim>
7138 <dim>1</dim>
7139 <dim>768</dim>
7140 </port>
7141 </input>
7142 <output>
7143 <port id="2" precision="FP32" names="361,hidden_states.25">
7144 <dim>-1</dim>
7145 <dim>-1</dim>
7146 <dim>768</dim>
7147 </port>
7148 </output>
7149 </layer>
7150 <layer id="518" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7151 <data element_type="f32" shape="1, 1, 768" offset="67880672" size="3072" />
7152 <output>
7153 <port id="0" precision="FP32">
7154 <dim>1</dim>
7155 <dim>1</dim>
7156 <dim>768</dim>
7157 </port>
7158 </output>
7159 </layer>
7160 <layer id="519" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7161 <data auto_broadcast="numpy" />
7162 <input>
7163 <port id="0" precision="FP32">
7164 <dim>-1</dim>
7165 <dim>-1</dim>
7166 <dim>768</dim>
7167 </port>
7168 <port id="1" precision="FP32">
7169 <dim>1</dim>
7170 <dim>1</dim>
7171 <dim>768</dim>
7172 </port>
7173 </input>
7174 <output>
7175 <port id="2" precision="FP32">
7176 <dim>-1</dim>
7177 <dim>-1</dim>
7178 <dim>768</dim>
7179 </port>
7180 </output>
7181 </layer>
7182 <layer id="520" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7183 <data element_type="f32" shape="" offset="67883744" size="4" />
7184 <output>
7185 <port id="0" precision="FP32" />
7186 </output>
7187 </layer>
7188 <layer id="521" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7189 <data element_type="f32" shape="" offset="67883748" size="4" />
7190 <output>
7191 <port id="0" precision="FP32" />
7192 </output>
7193 </layer>
7194 <layer id="522" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7195 <data element_type="f32" shape="" offset="67883744" size="4" />
7196 <output>
7197 <port id="0" precision="FP32" />
7198 </output>
7199 </layer>
7200 <layer id="523" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7201 <data element_type="f32" shape="" offset="67883748" size="4" />
7202 <output>
7203 <port id="0" precision="FP32" />
7204 </output>
7205 </layer>
7206 <layer id="524" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7207 <data levels="256" auto_broadcast="numpy" />
7208 <input>
7209 <port id="0" precision="FP32">
7210 <dim>-1</dim>
7211 <dim>-1</dim>
7212 <dim>768</dim>
7213 </port>
7214 <port id="1" precision="FP32" />
7215 <port id="2" precision="FP32" />
7216 <port id="3" precision="FP32" />
7217 <port id="4" precision="FP32" />
7218 </input>
7219 <output>
7220 <port id="5" precision="FP32">
7221 <dim>-1</dim>
7222 <dim>-1</dim>
7223 <dim>768</dim>
7224 </port>
7225 </output>
7226 </layer>
7227 <layer id="525" name="Constant_220590" type="Const" version="opset1">
7228 <data element_type="i8" shape="768, 768" offset="67883752" size="589824" />
7229 <output>
7230 <port id="0" precision="I8">
7231 <dim>768</dim>
7232 <dim>768</dim>
7233 </port>
7234 </output>
7235 </layer>
7236 <layer id="526" name="Convert_220591" type="Convert" version="opset1">
7237 <data destination_type="f32" />
7238 <input>
7239 <port id="0" precision="I8">
7240 <dim>768</dim>
7241 <dim>768</dim>
7242 </port>
7243 </input>
7244 <output>
7245 <port id="1" precision="FP32">
7246 <dim>768</dim>
7247 <dim>768</dim>
7248 </port>
7249 </output>
7250 </layer>
7251 <layer id="527" name="Constant_220592" type="Const" version="opset1">
7252 <data element_type="f32" shape="768, 1" offset="68473576" size="3072" />
7253 <output>
7254 <port id="0" precision="FP32">
7255 <dim>768</dim>
7256 <dim>1</dim>
7257 </port>
7258 </output>
7259 </layer>
7260 <layer id="528" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7261 <data auto_broadcast="numpy" />
7262 <input>
7263 <port id="0" precision="FP32">
7264 <dim>768</dim>
7265 <dim>768</dim>
7266 </port>
7267 <port id="1" precision="FP32">
7268 <dim>768</dim>
7269 <dim>1</dim>
7270 </port>
7271 </input>
7272 <output>
7273 <port id="2" precision="FP32">
7274 <dim>768</dim>
7275 <dim>768</dim>
7276 </port>
7277 </output>
7278 </layer>
7279 <layer id="529" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
7280 <data transpose_a="false" transpose_b="true" />
7281 <input>
7282 <port id="0" precision="FP32">
7283 <dim>-1</dim>
7284 <dim>-1</dim>
7285 <dim>768</dim>
7286 </port>
7287 <port id="1" precision="FP32">
7288 <dim>768</dim>
7289 <dim>768</dim>
7290 </port>
7291 </input>
7292 <output>
7293 <port id="2" precision="FP32">
7294 <dim>-1</dim>
7295 <dim>-1</dim>
7296 <dim>768</dim>
7297 </port>
7298 </output>
7299 </layer>
7300 <layer id="530" name="Constant_5402" type="Const" version="opset1">
7301 <data element_type="f32" shape="1, 1, 768" offset="68476648" size="3072" />
7302 <output>
7303 <port id="0" precision="FP32">
7304 <dim>1</dim>
7305 <dim>1</dim>
7306 <dim>768</dim>
7307 </port>
7308 </output>
7309 </layer>
7310 <layer id="531" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add" type="Add" version="opset1">
7311 <data auto_broadcast="numpy" />
7312 <input>
7313 <port id="0" precision="FP32">
7314 <dim>-1</dim>
7315 <dim>-1</dim>
7316 <dim>768</dim>
7317 </port>
7318 <port id="1" precision="FP32">
7319 <dim>1</dim>
7320 <dim>1</dim>
7321 <dim>768</dim>
7322 </port>
7323 </input>
7324 <output>
7325 <port id="2" precision="FP32" names="374,x.49">
7326 <dim>-1</dim>
7327 <dim>-1</dim>
7328 <dim>768</dim>
7329 </port>
7330 </output>
7331 </layer>
7332 <layer id="532" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
7333 <data element_type="f32" shape="" offset="68479720" size="4" />
7334 <output>
7335 <port id="0" precision="FP32" />
7336 </output>
7337 </layer>
7338 <layer id="533" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
7339 <data element_type="f32" shape="" offset="68479724" size="4" />
7340 <output>
7341 <port id="0" precision="FP32" />
7342 </output>
7343 </layer>
7344 <layer id="534" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
7345 <data element_type="f32" shape="" offset="68479720" size="4" />
7346 <output>
7347 <port id="0" precision="FP32" />
7348 </output>
7349 </layer>
7350 <layer id="535" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
7351 <data element_type="f32" shape="" offset="68479724" size="4" />
7352 <output>
7353 <port id="0" precision="FP32" />
7354 </output>
7355 </layer>
7356 <layer id="536" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
7357 <data levels="256" auto_broadcast="numpy" />
7358 <input>
7359 <port id="0" precision="FP32">
7360 <dim>-1</dim>
7361 <dim>-1</dim>
7362 <dim>768</dim>
7363 </port>
7364 <port id="1" precision="FP32" />
7365 <port id="2" precision="FP32" />
7366 <port id="3" precision="FP32" />
7367 <port id="4" precision="FP32" />
7368 </input>
7369 <output>
7370 <port id="5" precision="FP32">
7371 <dim>-1</dim>
7372 <dim>-1</dim>
7373 <dim>768</dim>
7374 </port>
7375 </output>
7376 </layer>
7377 <layer id="537" name="__module.roberta.encoder.layer.4.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
7378 <data element_type="i64" shape="4" offset="39811496" size="32" />
7379 <output>
7380 <port id="0" precision="I64">
7381 <dim>4</dim>
7382 </port>
7383 </output>
7384 </layer>
7385 <layer id="538" name="__module.roberta.encoder.layer.4.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
7386 <data special_zero="true" />
7387 <input>
7388 <port id="0" precision="FP32">
7389 <dim>-1</dim>
7390 <dim>-1</dim>
7391 <dim>768</dim>
7392 </port>
7393 <port id="1" precision="I64">
7394 <dim>4</dim>
7395 </port>
7396 </input>
7397 <output>
7398 <port id="2" precision="FP32" names="378,x.51">
7399 <dim>-1</dim>
7400 <dim>-1</dim>
7401 <dim>12</dim>
7402 <dim>64</dim>
7403 </port>
7404 </output>
7405 </layer>
7406 <layer id="539" name="Constant_1121" type="Const" version="opset1">
7407 <data element_type="i64" shape="4" offset="39811528" size="32" />
7408 <output>
7409 <port id="0" precision="I64" names="379">
7410 <dim>4</dim>
7411 </port>
7412 </output>
7413 </layer>
7414 <layer id="540" name="__module.roberta.encoder.layer.4.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
7415 <input>
7416 <port id="0" precision="FP32">
7417 <dim>-1</dim>
7418 <dim>-1</dim>
7419 <dim>12</dim>
7420 <dim>64</dim>
7421 </port>
7422 <port id="1" precision="I64">
7423 <dim>4</dim>
7424 </port>
7425 </input>
7426 <output>
7427 <port id="2" precision="FP32" names="380">
7428 <dim>-1</dim>
7429 <dim>12</dim>
7430 <dim>-1</dim>
7431 <dim>64</dim>
7432 </port>
7433 </output>
7434 </layer>
7435 <layer id="541" name="Constant_220594" type="Const" version="opset1">
7436 <data element_type="i8" shape="768, 768" offset="68479728" size="589824" />
7437 <output>
7438 <port id="0" precision="I8">
7439 <dim>768</dim>
7440 <dim>768</dim>
7441 </port>
7442 </output>
7443 </layer>
7444 <layer id="542" name="Convert_220595" type="Convert" version="opset1">
7445 <data destination_type="f32" />
7446 <input>
7447 <port id="0" precision="I8">
7448 <dim>768</dim>
7449 <dim>768</dim>
7450 </port>
7451 </input>
7452 <output>
7453 <port id="1" precision="FP32">
7454 <dim>768</dim>
7455 <dim>768</dim>
7456 </port>
7457 </output>
7458 </layer>
7459 <layer id="543" name="Constant_220596" type="Const" version="opset1">
7460 <data element_type="f32" shape="768, 1" offset="69069552" size="3072" />
7461 <output>
7462 <port id="0" precision="FP32">
7463 <dim>768</dim>
7464 <dim>1</dim>
7465 </port>
7466 </output>
7467 </layer>
7468 <layer id="544" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7469 <data auto_broadcast="numpy" />
7470 <input>
7471 <port id="0" precision="FP32">
7472 <dim>768</dim>
7473 <dim>768</dim>
7474 </port>
7475 <port id="1" precision="FP32">
7476 <dim>768</dim>
7477 <dim>1</dim>
7478 </port>
7479 </input>
7480 <output>
7481 <port id="2" precision="FP32">
7482 <dim>768</dim>
7483 <dim>768</dim>
7484 </port>
7485 </output>
7486 </layer>
7487 <layer id="545" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
7488 <data transpose_a="false" transpose_b="true" />
7489 <input>
7490 <port id="0" precision="FP32">
7491 <dim>-1</dim>
7492 <dim>-1</dim>
7493 <dim>768</dim>
7494 </port>
7495 <port id="1" precision="FP32">
7496 <dim>768</dim>
7497 <dim>768</dim>
7498 </port>
7499 </input>
7500 <output>
7501 <port id="2" precision="FP32">
7502 <dim>-1</dim>
7503 <dim>-1</dim>
7504 <dim>768</dim>
7505 </port>
7506 </output>
7507 </layer>
7508 <layer id="546" name="Constant_5403" type="Const" version="opset1">
7509 <data element_type="f32" shape="1, 1, 768" offset="69072624" size="3072" />
7510 <output>
7511 <port id="0" precision="FP32">
7512 <dim>1</dim>
7513 <dim>1</dim>
7514 <dim>768</dim>
7515 </port>
7516 </output>
7517 </layer>
7518 <layer id="547" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add" type="Add" version="opset1">
7519 <data auto_broadcast="numpy" />
7520 <input>
7521 <port id="0" precision="FP32">
7522 <dim>-1</dim>
7523 <dim>-1</dim>
7524 <dim>768</dim>
7525 </port>
7526 <port id="1" precision="FP32">
7527 <dim>1</dim>
7528 <dim>1</dim>
7529 <dim>768</dim>
7530 </port>
7531 </input>
7532 <output>
7533 <port id="2" precision="FP32" names="383,x.53">
7534 <dim>-1</dim>
7535 <dim>-1</dim>
7536 <dim>768</dim>
7537 </port>
7538 </output>
7539 </layer>
7540 <layer id="548" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
7541 <data element_type="f32" shape="" offset="69075696" size="4" />
7542 <output>
7543 <port id="0" precision="FP32" />
7544 </output>
7545 </layer>
7546 <layer id="549" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
7547 <data element_type="f32" shape="" offset="69075700" size="4" />
7548 <output>
7549 <port id="0" precision="FP32" />
7550 </output>
7551 </layer>
7552 <layer id="550" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
7553 <data element_type="f32" shape="" offset="69075696" size="4" />
7554 <output>
7555 <port id="0" precision="FP32" />
7556 </output>
7557 </layer>
7558 <layer id="551" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
7559 <data element_type="f32" shape="" offset="69075700" size="4" />
7560 <output>
7561 <port id="0" precision="FP32" />
7562 </output>
7563 </layer>
7564 <layer id="552" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
7565 <data levels="256" auto_broadcast="numpy" />
7566 <input>
7567 <port id="0" precision="FP32">
7568 <dim>-1</dim>
7569 <dim>-1</dim>
7570 <dim>768</dim>
7571 </port>
7572 <port id="1" precision="FP32" />
7573 <port id="2" precision="FP32" />
7574 <port id="3" precision="FP32" />
7575 <port id="4" precision="FP32" />
7576 </input>
7577 <output>
7578 <port id="5" precision="FP32">
7579 <dim>-1</dim>
7580 <dim>-1</dim>
7581 <dim>768</dim>
7582 </port>
7583 </output>
7584 </layer>
7585 <layer id="553" name="__module.roberta.encoder.layer.4.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
7586 <data element_type="i64" shape="4" offset="39811496" size="32" />
7587 <output>
7588 <port id="0" precision="I64">
7589 <dim>4</dim>
7590 </port>
7591 </output>
7592 </layer>
7593 <layer id="554" name="__module.roberta.encoder.layer.4.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
7594 <data special_zero="true" />
7595 <input>
7596 <port id="0" precision="FP32">
7597 <dim>-1</dim>
7598 <dim>-1</dim>
7599 <dim>768</dim>
7600 </port>
7601 <port id="1" precision="I64">
7602 <dim>4</dim>
7603 </port>
7604 </input>
7605 <output>
7606 <port id="2" precision="FP32" names="387,x.55">
7607 <dim>-1</dim>
7608 <dim>-1</dim>
7609 <dim>12</dim>
7610 <dim>64</dim>
7611 </port>
7612 </output>
7613 </layer>
7614 <layer id="555" name="Constant_1144" type="Const" version="opset1">
7615 <data element_type="i64" shape="4" offset="39811528" size="32" />
7616 <output>
7617 <port id="0" precision="I64" names="388">
7618 <dim>4</dim>
7619 </port>
7620 </output>
7621 </layer>
7622 <layer id="556" name="__module.roberta.encoder.layer.4.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
7623 <input>
7624 <port id="0" precision="FP32">
7625 <dim>-1</dim>
7626 <dim>-1</dim>
7627 <dim>12</dim>
7628 <dim>64</dim>
7629 </port>
7630 <port id="1" precision="I64">
7631 <dim>4</dim>
7632 </port>
7633 </input>
7634 <output>
7635 <port id="2" precision="FP32" names="389">
7636 <dim>-1</dim>
7637 <dim>12</dim>
7638 <dim>-1</dim>
7639 <dim>64</dim>
7640 </port>
7641 </output>
7642 </layer>
7643 <layer id="557" name="Constant_220598" type="Const" version="opset1">
7644 <data element_type="i8" shape="768, 768" offset="69075704" size="589824" />
7645 <output>
7646 <port id="0" precision="I8">
7647 <dim>768</dim>
7648 <dim>768</dim>
7649 </port>
7650 </output>
7651 </layer>
7652 <layer id="558" name="Convert_220599" type="Convert" version="opset1">
7653 <data destination_type="f32" />
7654 <input>
7655 <port id="0" precision="I8">
7656 <dim>768</dim>
7657 <dim>768</dim>
7658 </port>
7659 </input>
7660 <output>
7661 <port id="1" precision="FP32">
7662 <dim>768</dim>
7663 <dim>768</dim>
7664 </port>
7665 </output>
7666 </layer>
7667 <layer id="559" name="Constant_220600" type="Const" version="opset1">
7668 <data element_type="f32" shape="768, 1" offset="69665528" size="3072" />
7669 <output>
7670 <port id="0" precision="FP32">
7671 <dim>768</dim>
7672 <dim>1</dim>
7673 </port>
7674 </output>
7675 </layer>
7676 <layer id="560" name="__module.roberta.encoder.layer.4.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
7677 <data auto_broadcast="numpy" />
7678 <input>
7679 <port id="0" precision="FP32">
7680 <dim>768</dim>
7681 <dim>768</dim>
7682 </port>
7683 <port id="1" precision="FP32">
7684 <dim>768</dim>
7685 <dim>1</dim>
7686 </port>
7687 </input>
7688 <output>
7689 <port id="2" precision="FP32">
7690 <dim>768</dim>
7691 <dim>768</dim>
7692 </port>
7693 </output>
7694 </layer>
7695 <layer id="561" name="__module.roberta.encoder.layer.4.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
7696 <data transpose_a="false" transpose_b="true" />
7697 <input>
7698 <port id="0" precision="FP32">
7699 <dim>-1</dim>
7700 <dim>-1</dim>
7701 <dim>768</dim>
7702 </port>
7703 <port id="1" precision="FP32">
7704 <dim>768</dim>
7705 <dim>768</dim>
7706 </port>
7707 </input>
7708 <output>
7709 <port id="2" precision="FP32">
7710 <dim>-1</dim>
7711 <dim>-1</dim>
7712 <dim>768</dim>
7713 </port>
7714 </output>
7715 </layer>
7716 <layer id="562" name="Constant_5404" type="Const" version="opset1">
7717 <data element_type="f32" shape="1, 1, 768" offset="69668600" size="3072" />
7718 <output>
7719 <port id="0" precision="FP32">
7720 <dim>1</dim>
7721 <dim>1</dim>
7722 <dim>768</dim>
7723 </port>
7724 </output>
7725 </layer>
7726 <layer id="563" name="__module.roberta.encoder.layer.4.attention.self.value/aten::linear/Add" type="Add" version="opset1">
7727 <data auto_broadcast="numpy" />
7728 <input>
7729 <port id="0" precision="FP32">
7730 <dim>-1</dim>
7731 <dim>-1</dim>
7732 <dim>768</dim>
7733 </port>
7734 <port id="1" precision="FP32">
7735 <dim>1</dim>
7736 <dim>1</dim>
7737 <dim>768</dim>
7738 </port>
7739 </input>
7740 <output>
7741 <port id="2" precision="FP32" names="392,x.57">
7742 <dim>-1</dim>
7743 <dim>-1</dim>
7744 <dim>768</dim>
7745 </port>
7746 </output>
7747 </layer>
7748 <layer id="564" name="__module.roberta.encoder.layer.4.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
7749 <data element_type="i64" shape="4" offset="39811496" size="32" />
7750 <output>
7751 <port id="0" precision="I64">
7752 <dim>4</dim>
7753 </port>
7754 </output>
7755 </layer>
7756 <layer id="565" name="__module.roberta.encoder.layer.4.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
7757 <data special_zero="true" />
7758 <input>
7759 <port id="0" precision="FP32">
7760 <dim>-1</dim>
7761 <dim>-1</dim>
7762 <dim>768</dim>
7763 </port>
7764 <port id="1" precision="I64">
7765 <dim>4</dim>
7766 </port>
7767 </input>
7768 <output>
7769 <port id="2" precision="FP32" names="396,x.59">
7770 <dim>-1</dim>
7771 <dim>-1</dim>
7772 <dim>12</dim>
7773 <dim>64</dim>
7774 </port>
7775 </output>
7776 </layer>
7777 <layer id="566" name="Constant_1167" type="Const" version="opset1">
7778 <data element_type="i64" shape="4" offset="39811528" size="32" />
7779 <output>
7780 <port id="0" precision="I64" names="397">
7781 <dim>4</dim>
7782 </port>
7783 </output>
7784 </layer>
7785 <layer id="567" name="__module.roberta.encoder.layer.4.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
7786 <input>
7787 <port id="0" precision="FP32">
7788 <dim>-1</dim>
7789 <dim>-1</dim>
7790 <dim>12</dim>
7791 <dim>64</dim>
7792 </port>
7793 <port id="1" precision="I64">
7794 <dim>4</dim>
7795 </port>
7796 </input>
7797 <output>
7798 <port id="2" precision="FP32" names="398">
7799 <dim>-1</dim>
7800 <dim>12</dim>
7801 <dim>-1</dim>
7802 <dim>64</dim>
7803 </port>
7804 </output>
7805 </layer>
7806 <layer id="568" name="__module.roberta.encoder.layer.4.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
7807 <data causal="false" />
7808 <input>
7809 <port id="0" precision="FP32">
7810 <dim>-1</dim>
7811 <dim>12</dim>
7812 <dim>-1</dim>
7813 <dim>64</dim>
7814 </port>
7815 <port id="1" precision="FP32">
7816 <dim>-1</dim>
7817 <dim>12</dim>
7818 <dim>-1</dim>
7819 <dim>64</dim>
7820 </port>
7821 <port id="2" precision="FP32">
7822 <dim>-1</dim>
7823 <dim>12</dim>
7824 <dim>-1</dim>
7825 <dim>64</dim>
7826 </port>
7827 <port id="3" precision="FP32">
7828 <dim>-1</dim>
7829 <dim>1</dim>
7830 <dim>-1</dim>
7831 <dim>-1</dim>
7832 </port>
7833 </input>
7834 <output>
7835 <port id="4" precision="FP32" names="399,attn_output.17">
7836 <dim>-1</dim>
7837 <dim>12</dim>
7838 <dim>-1</dim>
7839 <dim>64</dim>
7840 </port>
7841 </output>
7842 </layer>
7843 <layer id="569" name="__module.roberta.encoder.layer.4.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
7844 <data element_type="i32" shape="4" offset="41003536" size="16" />
7845 <output>
7846 <port id="0" precision="I32">
7847 <dim>4</dim>
7848 </port>
7849 </output>
7850 </layer>
7851 <layer id="570" name="__module.roberta.encoder.layer.4.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
7852 <input>
7853 <port id="0" precision="FP32">
7854 <dim>-1</dim>
7855 <dim>12</dim>
7856 <dim>-1</dim>
7857 <dim>64</dim>
7858 </port>
7859 <port id="1" precision="I32">
7860 <dim>4</dim>
7861 </port>
7862 </input>
7863 <output>
7864 <port id="2" precision="FP32" names="400,attn_output.19">
7865 <dim>-1</dim>
7866 <dim>-1</dim>
7867 <dim>12</dim>
7868 <dim>64</dim>
7869 </port>
7870 </output>
7871 </layer>
7872 <layer id="571" name="Constant_5498" type="Const" version="opset1">
7873 <data element_type="i64" shape="3" offset="41003552" size="24" />
7874 <output>
7875 <port id="0" precision="I64">
7876 <dim>3</dim>
7877 </port>
7878 </output>
7879 </layer>
7880 <layer id="572" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
7881 <data special_zero="true" />
7882 <input>
7883 <port id="0" precision="FP32">
7884 <dim>-1</dim>
7885 <dim>-1</dim>
7886 <dim>12</dim>
7887 <dim>64</dim>
7888 </port>
7889 <port id="1" precision="I64">
7890 <dim>3</dim>
7891 </port>
7892 </input>
7893 <output>
7894 <port id="2" precision="FP32" names="402">
7895 <dim>-1</dim>
7896 <dim>-1</dim>
7897 <dim>768</dim>
7898 </port>
7899 </output>
7900 </layer>
7901 <layer id="573" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
7902 <data element_type="f32" shape="1, 1, 768" offset="69671672" size="3072" />
7903 <output>
7904 <port id="0" precision="FP32">
7905 <dim>1</dim>
7906 <dim>1</dim>
7907 <dim>768</dim>
7908 </port>
7909 </output>
7910 </layer>
7911 <layer id="574" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
7912 <data auto_broadcast="numpy" />
7913 <input>
7914 <port id="0" precision="FP32">
7915 <dim>-1</dim>
7916 <dim>-1</dim>
7917 <dim>768</dim>
7918 </port>
7919 <port id="1" precision="FP32">
7920 <dim>1</dim>
7921 <dim>1</dim>
7922 <dim>768</dim>
7923 </port>
7924 </input>
7925 <output>
7926 <port id="2" precision="FP32">
7927 <dim>-1</dim>
7928 <dim>-1</dim>
7929 <dim>768</dim>
7930 </port>
7931 </output>
7932 </layer>
7933 <layer id="575" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
7934 <data element_type="f32" shape="" offset="69674744" size="4" />
7935 <output>
7936 <port id="0" precision="FP32" />
7937 </output>
7938 </layer>
7939 <layer id="576" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
7940 <data element_type="f32" shape="" offset="69674748" size="4" />
7941 <output>
7942 <port id="0" precision="FP32" />
7943 </output>
7944 </layer>
7945 <layer id="577" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
7946 <data element_type="f32" shape="" offset="69674744" size="4" />
7947 <output>
7948 <port id="0" precision="FP32" />
7949 </output>
7950 </layer>
7951 <layer id="578" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
7952 <data element_type="f32" shape="" offset="69674748" size="4" />
7953 <output>
7954 <port id="0" precision="FP32" />
7955 </output>
7956 </layer>
7957 <layer id="579" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
7958 <data levels="256" auto_broadcast="numpy" />
7959 <input>
7960 <port id="0" precision="FP32">
7961 <dim>-1</dim>
7962 <dim>-1</dim>
7963 <dim>768</dim>
7964 </port>
7965 <port id="1" precision="FP32" />
7966 <port id="2" precision="FP32" />
7967 <port id="3" precision="FP32" />
7968 <port id="4" precision="FP32" />
7969 </input>
7970 <output>
7971 <port id="5" precision="FP32">
7972 <dim>-1</dim>
7973 <dim>-1</dim>
7974 <dim>768</dim>
7975 </port>
7976 </output>
7977 </layer>
7978 <layer id="580" name="Constant_220602" type="Const" version="opset1">
7979 <data element_type="i8" shape="768, 768" offset="69674752" size="589824" />
7980 <output>
7981 <port id="0" precision="I8">
7982 <dim>768</dim>
7983 <dim>768</dim>
7984 </port>
7985 </output>
7986 </layer>
7987 <layer id="581" name="Convert_220603" type="Convert" version="opset1">
7988 <data destination_type="f32" />
7989 <input>
7990 <port id="0" precision="I8">
7991 <dim>768</dim>
7992 <dim>768</dim>
7993 </port>
7994 </input>
7995 <output>
7996 <port id="1" precision="FP32">
7997 <dim>768</dim>
7998 <dim>768</dim>
7999 </port>
8000 </output>
8001 </layer>
8002 <layer id="582" name="Constant_220604" type="Const" version="opset1">
8003 <data element_type="f32" shape="768, 1" offset="70264576" size="3072" />
8004 <output>
8005 <port id="0" precision="FP32">
8006 <dim>768</dim>
8007 <dim>1</dim>
8008 </port>
8009 </output>
8010 </layer>
8011 <layer id="583" name="__module.roberta.encoder.layer.4.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8012 <data auto_broadcast="numpy" />
8013 <input>
8014 <port id="0" precision="FP32">
8015 <dim>768</dim>
8016 <dim>768</dim>
8017 </port>
8018 <port id="1" precision="FP32">
8019 <dim>768</dim>
8020 <dim>1</dim>
8021 </port>
8022 </input>
8023 <output>
8024 <port id="2" precision="FP32">
8025 <dim>768</dim>
8026 <dim>768</dim>
8027 </port>
8028 </output>
8029 </layer>
8030 <layer id="584" name="__module.roberta.encoder.layer.4.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8031 <data transpose_a="false" transpose_b="true" />
8032 <input>
8033 <port id="0" precision="FP32">
8034 <dim>-1</dim>
8035 <dim>-1</dim>
8036 <dim>768</dim>
8037 </port>
8038 <port id="1" precision="FP32">
8039 <dim>768</dim>
8040 <dim>768</dim>
8041 </port>
8042 </input>
8043 <output>
8044 <port id="2" precision="FP32">
8045 <dim>-1</dim>
8046 <dim>-1</dim>
8047 <dim>768</dim>
8048 </port>
8049 </output>
8050 </layer>
8051 <layer id="585" name="Constant_5405" type="Const" version="opset1">
8052 <data element_type="f32" shape="1, 1, 768" offset="70267648" size="3072" />
8053 <output>
8054 <port id="0" precision="FP32">
8055 <dim>1</dim>
8056 <dim>1</dim>
8057 <dim>768</dim>
8058 </port>
8059 </output>
8060 </layer>
8061 <layer id="586" name="__module.roberta.encoder.layer.4.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
8062 <data auto_broadcast="numpy" />
8063 <input>
8064 <port id="0" precision="FP32">
8065 <dim>-1</dim>
8066 <dim>-1</dim>
8067 <dim>768</dim>
8068 </port>
8069 <port id="1" precision="FP32">
8070 <dim>1</dim>
8071 <dim>1</dim>
8072 <dim>768</dim>
8073 </port>
8074 </input>
8075 <output>
8076 <port id="2" precision="FP32" names="407,input.19">
8077 <dim>-1</dim>
8078 <dim>-1</dim>
8079 <dim>768</dim>
8080 </port>
8081 </output>
8082 </layer>
8083 <layer id="587" name="__module.roberta.encoder.layer.4.attention.output/aten::add/Add" type="Add" version="opset1">
8084 <data auto_broadcast="numpy" />
8085 <input>
8086 <port id="0" precision="FP32">
8087 <dim>-1</dim>
8088 <dim>-1</dim>
8089 <dim>768</dim>
8090 </port>
8091 <port id="1" precision="FP32">
8092 <dim>-1</dim>
8093 <dim>-1</dim>
8094 <dim>768</dim>
8095 </port>
8096 </input>
8097 <output>
8098 <port id="2" precision="FP32" names="409">
8099 <dim>-1</dim>
8100 <dim>-1</dim>
8101 <dim>768</dim>
8102 </port>
8103 </output>
8104 </layer>
8105 <layer id="588" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
8106 <data element_type="i32" shape="1" offset="39206292" size="4" />
8107 <output>
8108 <port id="0" precision="I32">
8109 <dim>1</dim>
8110 </port>
8111 </output>
8112 </layer>
8113 <layer id="589" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
8114 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
8115 <input>
8116 <port id="0" precision="FP32">
8117 <dim>-1</dim>
8118 <dim>-1</dim>
8119 <dim>768</dim>
8120 </port>
8121 <port id="1" precision="I32">
8122 <dim>1</dim>
8123 </port>
8124 </input>
8125 <output>
8126 <port id="2" precision="FP32">
8127 <dim>-1</dim>
8128 <dim>-1</dim>
8129 <dim>768</dim>
8130 </port>
8131 </output>
8132 </layer>
8133 <layer id="590" name="Constant_5406" type="Const" version="opset1">
8134 <data element_type="f32" shape="1, 1, 768" offset="70270720" size="3072" />
8135 <output>
8136 <port id="0" precision="FP32">
8137 <dim>1</dim>
8138 <dim>1</dim>
8139 <dim>768</dim>
8140 </port>
8141 </output>
8142 </layer>
8143 <layer id="591" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
8144 <data auto_broadcast="numpy" />
8145 <input>
8146 <port id="0" precision="FP32">
8147 <dim>-1</dim>
8148 <dim>-1</dim>
8149 <dim>768</dim>
8150 </port>
8151 <port id="1" precision="FP32">
8152 <dim>1</dim>
8153 <dim>1</dim>
8154 <dim>768</dim>
8155 </port>
8156 </input>
8157 <output>
8158 <port id="2" precision="FP32">
8159 <dim>-1</dim>
8160 <dim>-1</dim>
8161 <dim>768</dim>
8162 </port>
8163 </output>
8164 </layer>
8165 <layer id="592" name="Constant_5407" type="Const" version="opset1">
8166 <data element_type="f32" shape="1, 1, 768" offset="70273792" size="3072" />
8167 <output>
8168 <port id="0" precision="FP32">
8169 <dim>1</dim>
8170 <dim>1</dim>
8171 <dim>768</dim>
8172 </port>
8173 </output>
8174 </layer>
8175 <layer id="593" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
8176 <data auto_broadcast="numpy" />
8177 <input>
8178 <port id="0" precision="FP32">
8179 <dim>-1</dim>
8180 <dim>-1</dim>
8181 <dim>768</dim>
8182 </port>
8183 <port id="1" precision="FP32">
8184 <dim>1</dim>
8185 <dim>1</dim>
8186 <dim>768</dim>
8187 </port>
8188 </input>
8189 <output>
8190 <port id="2" precision="FP32" names="413,input_tensor.9">
8191 <dim>-1</dim>
8192 <dim>-1</dim>
8193 <dim>768</dim>
8194 </port>
8195 </output>
8196 </layer>
8197 <layer id="594" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8198 <data element_type="f32" shape="1, 1, 768" offset="70276864" size="3072" />
8199 <output>
8200 <port id="0" precision="FP32">
8201 <dim>1</dim>
8202 <dim>1</dim>
8203 <dim>768</dim>
8204 </port>
8205 </output>
8206 </layer>
8207 <layer id="595" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8208 <data auto_broadcast="numpy" />
8209 <input>
8210 <port id="0" precision="FP32">
8211 <dim>-1</dim>
8212 <dim>-1</dim>
8213 <dim>768</dim>
8214 </port>
8215 <port id="1" precision="FP32">
8216 <dim>1</dim>
8217 <dim>1</dim>
8218 <dim>768</dim>
8219 </port>
8220 </input>
8221 <output>
8222 <port id="2" precision="FP32">
8223 <dim>-1</dim>
8224 <dim>-1</dim>
8225 <dim>768</dim>
8226 </port>
8227 </output>
8228 </layer>
8229 <layer id="596" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8230 <data element_type="f32" shape="" offset="70279936" size="4" />
8231 <output>
8232 <port id="0" precision="FP32" />
8233 </output>
8234 </layer>
8235 <layer id="597" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8236 <data element_type="f32" shape="" offset="70279940" size="4" />
8237 <output>
8238 <port id="0" precision="FP32" />
8239 </output>
8240 </layer>
8241 <layer id="598" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8242 <data element_type="f32" shape="" offset="70279936" size="4" />
8243 <output>
8244 <port id="0" precision="FP32" />
8245 </output>
8246 </layer>
8247 <layer id="599" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8248 <data element_type="f32" shape="" offset="70279940" size="4" />
8249 <output>
8250 <port id="0" precision="FP32" />
8251 </output>
8252 </layer>
8253 <layer id="600" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8254 <data levels="256" auto_broadcast="numpy" />
8255 <input>
8256 <port id="0" precision="FP32">
8257 <dim>-1</dim>
8258 <dim>-1</dim>
8259 <dim>768</dim>
8260 </port>
8261 <port id="1" precision="FP32" />
8262 <port id="2" precision="FP32" />
8263 <port id="3" precision="FP32" />
8264 <port id="4" precision="FP32" />
8265 </input>
8266 <output>
8267 <port id="5" precision="FP32">
8268 <dim>-1</dim>
8269 <dim>-1</dim>
8270 <dim>768</dim>
8271 </port>
8272 </output>
8273 </layer>
8274 <layer id="601" name="Constant_220606" type="Const" version="opset1">
8275 <data element_type="i8" shape="3072, 768" offset="70279944" size="2359296" />
8276 <output>
8277 <port id="0" precision="I8">
8278 <dim>3072</dim>
8279 <dim>768</dim>
8280 </port>
8281 </output>
8282 </layer>
8283 <layer id="602" name="Convert_220607" type="Convert" version="opset1">
8284 <data destination_type="f32" />
8285 <input>
8286 <port id="0" precision="I8">
8287 <dim>3072</dim>
8288 <dim>768</dim>
8289 </port>
8290 </input>
8291 <output>
8292 <port id="1" precision="FP32">
8293 <dim>3072</dim>
8294 <dim>768</dim>
8295 </port>
8296 </output>
8297 </layer>
8298 <layer id="603" name="Constant_220608" type="Const" version="opset1">
8299 <data element_type="f32" shape="3072, 1" offset="72639240" size="12288" />
8300 <output>
8301 <port id="0" precision="FP32">
8302 <dim>3072</dim>
8303 <dim>1</dim>
8304 </port>
8305 </output>
8306 </layer>
8307 <layer id="604" name="__module.roberta.encoder.layer.4.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8308 <data auto_broadcast="numpy" />
8309 <input>
8310 <port id="0" precision="FP32">
8311 <dim>3072</dim>
8312 <dim>768</dim>
8313 </port>
8314 <port id="1" precision="FP32">
8315 <dim>3072</dim>
8316 <dim>1</dim>
8317 </port>
8318 </input>
8319 <output>
8320 <port id="2" precision="FP32">
8321 <dim>3072</dim>
8322 <dim>768</dim>
8323 </port>
8324 </output>
8325 </layer>
8326 <layer id="605" name="__module.roberta.encoder.layer.4.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8327 <data transpose_a="false" transpose_b="true" />
8328 <input>
8329 <port id="0" precision="FP32">
8330 <dim>-1</dim>
8331 <dim>-1</dim>
8332 <dim>768</dim>
8333 </port>
8334 <port id="1" precision="FP32">
8335 <dim>3072</dim>
8336 <dim>768</dim>
8337 </port>
8338 </input>
8339 <output>
8340 <port id="2" precision="FP32">
8341 <dim>-1</dim>
8342 <dim>-1</dim>
8343 <dim>3072</dim>
8344 </port>
8345 </output>
8346 </layer>
8347 <layer id="606" name="Constant_5408" type="Const" version="opset1">
8348 <data element_type="f32" shape="1, 1, 3072" offset="72651528" size="12288" />
8349 <output>
8350 <port id="0" precision="FP32">
8351 <dim>1</dim>
8352 <dim>1</dim>
8353 <dim>3072</dim>
8354 </port>
8355 </output>
8356 </layer>
8357 <layer id="607" name="__module.roberta.encoder.layer.4.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
8358 <data auto_broadcast="numpy" />
8359 <input>
8360 <port id="0" precision="FP32">
8361 <dim>-1</dim>
8362 <dim>-1</dim>
8363 <dim>3072</dim>
8364 </port>
8365 <port id="1" precision="FP32">
8366 <dim>1</dim>
8367 <dim>1</dim>
8368 <dim>3072</dim>
8369 </port>
8370 </input>
8371 <output>
8372 <port id="2" precision="FP32" names="417">
8373 <dim>-1</dim>
8374 <dim>-1</dim>
8375 <dim>3072</dim>
8376 </port>
8377 </output>
8378 </layer>
8379 <layer id="608" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
8380 <data approximation_mode="ERF" />
8381 <input>
8382 <port id="0" precision="FP32">
8383 <dim>-1</dim>
8384 <dim>-1</dim>
8385 <dim>3072</dim>
8386 </port>
8387 </input>
8388 <output>
8389 <port id="1" precision="FP32" names="418">
8390 <dim>-1</dim>
8391 <dim>-1</dim>
8392 <dim>3072</dim>
8393 </port>
8394 </output>
8395 </layer>
8396 <layer id="609" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8397 <data element_type="f32" shape="1, 1, 3072" offset="72663816" size="12288" />
8398 <output>
8399 <port id="0" precision="FP32">
8400 <dim>1</dim>
8401 <dim>1</dim>
8402 <dim>3072</dim>
8403 </port>
8404 </output>
8405 </layer>
8406 <layer id="610" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8407 <data auto_broadcast="numpy" />
8408 <input>
8409 <port id="0" precision="FP32">
8410 <dim>-1</dim>
8411 <dim>-1</dim>
8412 <dim>3072</dim>
8413 </port>
8414 <port id="1" precision="FP32">
8415 <dim>1</dim>
8416 <dim>1</dim>
8417 <dim>3072</dim>
8418 </port>
8419 </input>
8420 <output>
8421 <port id="2" precision="FP32">
8422 <dim>-1</dim>
8423 <dim>-1</dim>
8424 <dim>3072</dim>
8425 </port>
8426 </output>
8427 </layer>
8428 <layer id="611" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8429 <data element_type="f32" shape="" offset="72676104" size="4" />
8430 <output>
8431 <port id="0" precision="FP32" />
8432 </output>
8433 </layer>
8434 <layer id="612" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8435 <data element_type="f32" shape="" offset="72676108" size="4" />
8436 <output>
8437 <port id="0" precision="FP32" />
8438 </output>
8439 </layer>
8440 <layer id="613" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8441 <data element_type="f32" shape="" offset="72676104" size="4" />
8442 <output>
8443 <port id="0" precision="FP32" />
8444 </output>
8445 </layer>
8446 <layer id="614" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8447 <data element_type="f32" shape="" offset="72676108" size="4" />
8448 <output>
8449 <port id="0" precision="FP32" />
8450 </output>
8451 </layer>
8452 <layer id="615" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8453 <data levels="256" auto_broadcast="numpy" />
8454 <input>
8455 <port id="0" precision="FP32">
8456 <dim>-1</dim>
8457 <dim>-1</dim>
8458 <dim>3072</dim>
8459 </port>
8460 <port id="1" precision="FP32" />
8461 <port id="2" precision="FP32" />
8462 <port id="3" precision="FP32" />
8463 <port id="4" precision="FP32" />
8464 </input>
8465 <output>
8466 <port id="5" precision="FP32">
8467 <dim>-1</dim>
8468 <dim>-1</dim>
8469 <dim>3072</dim>
8470 </port>
8471 </output>
8472 </layer>
8473 <layer id="616" name="Constant_220610" type="Const" version="opset1">
8474 <data element_type="i8" shape="768, 3072" offset="72676112" size="2359296" />
8475 <output>
8476 <port id="0" precision="I8">
8477 <dim>768</dim>
8478 <dim>3072</dim>
8479 </port>
8480 </output>
8481 </layer>
8482 <layer id="617" name="Convert_220611" type="Convert" version="opset1">
8483 <data destination_type="f32" />
8484 <input>
8485 <port id="0" precision="I8">
8486 <dim>768</dim>
8487 <dim>3072</dim>
8488 </port>
8489 </input>
8490 <output>
8491 <port id="1" precision="FP32">
8492 <dim>768</dim>
8493 <dim>3072</dim>
8494 </port>
8495 </output>
8496 </layer>
8497 <layer id="618" name="Constant_220612" type="Const" version="opset1">
8498 <data element_type="f32" shape="768, 1" offset="75035408" size="3072" />
8499 <output>
8500 <port id="0" precision="FP32">
8501 <dim>768</dim>
8502 <dim>1</dim>
8503 </port>
8504 </output>
8505 </layer>
8506 <layer id="619" name="__module.roberta.encoder.layer.4.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8507 <data auto_broadcast="numpy" />
8508 <input>
8509 <port id="0" precision="FP32">
8510 <dim>768</dim>
8511 <dim>3072</dim>
8512 </port>
8513 <port id="1" precision="FP32">
8514 <dim>768</dim>
8515 <dim>1</dim>
8516 </port>
8517 </input>
8518 <output>
8519 <port id="2" precision="FP32">
8520 <dim>768</dim>
8521 <dim>3072</dim>
8522 </port>
8523 </output>
8524 </layer>
8525 <layer id="620" name="__module.roberta.encoder.layer.4.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
8526 <data transpose_a="false" transpose_b="true" />
8527 <input>
8528 <port id="0" precision="FP32">
8529 <dim>-1</dim>
8530 <dim>-1</dim>
8531 <dim>3072</dim>
8532 </port>
8533 <port id="1" precision="FP32">
8534 <dim>768</dim>
8535 <dim>3072</dim>
8536 </port>
8537 </input>
8538 <output>
8539 <port id="2" precision="FP32">
8540 <dim>-1</dim>
8541 <dim>-1</dim>
8542 <dim>768</dim>
8543 </port>
8544 </output>
8545 </layer>
8546 <layer id="621" name="Constant_5409" type="Const" version="opset1">
8547 <data element_type="f32" shape="1, 1, 768" offset="75038480" size="3072" />
8548 <output>
8549 <port id="0" precision="FP32">
8550 <dim>1</dim>
8551 <dim>1</dim>
8552 <dim>768</dim>
8553 </port>
8554 </output>
8555 </layer>
8556 <layer id="622" name="__module.roberta.encoder.layer.4.output.dense/aten::linear/Add" type="Add" version="opset1">
8557 <data auto_broadcast="numpy" />
8558 <input>
8559 <port id="0" precision="FP32">
8560 <dim>-1</dim>
8561 <dim>-1</dim>
8562 <dim>768</dim>
8563 </port>
8564 <port id="1" precision="FP32">
8565 <dim>1</dim>
8566 <dim>1</dim>
8567 <dim>768</dim>
8568 </port>
8569 </input>
8570 <output>
8571 <port id="2" precision="FP32" names="423,input.21">
8572 <dim>-1</dim>
8573 <dim>-1</dim>
8574 <dim>768</dim>
8575 </port>
8576 </output>
8577 </layer>
8578 <layer id="623" name="__module.roberta.encoder.layer.4.output/aten::add/Add" type="Add" version="opset1">
8579 <data auto_broadcast="numpy" />
8580 <input>
8581 <port id="0" precision="FP32">
8582 <dim>-1</dim>
8583 <dim>-1</dim>
8584 <dim>768</dim>
8585 </port>
8586 <port id="1" precision="FP32">
8587 <dim>-1</dim>
8588 <dim>-1</dim>
8589 <dim>768</dim>
8590 </port>
8591 </input>
8592 <output>
8593 <port id="2" precision="FP32" names="425">
8594 <dim>-1</dim>
8595 <dim>-1</dim>
8596 <dim>768</dim>
8597 </port>
8598 </output>
8599 </layer>
8600 <layer id="624" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
8601 <data element_type="i32" shape="1" offset="39206292" size="4" />
8602 <output>
8603 <port id="0" precision="I32">
8604 <dim>1</dim>
8605 </port>
8606 </output>
8607 </layer>
8608 <layer id="625" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
8609 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
8610 <input>
8611 <port id="0" precision="FP32">
8612 <dim>-1</dim>
8613 <dim>-1</dim>
8614 <dim>768</dim>
8615 </port>
8616 <port id="1" precision="I32">
8617 <dim>1</dim>
8618 </port>
8619 </input>
8620 <output>
8621 <port id="2" precision="FP32">
8622 <dim>-1</dim>
8623 <dim>-1</dim>
8624 <dim>768</dim>
8625 </port>
8626 </output>
8627 </layer>
8628 <layer id="626" name="Constant_5410" type="Const" version="opset1">
8629 <data element_type="f32" shape="1, 1, 768" offset="75041552" size="3072" />
8630 <output>
8631 <port id="0" precision="FP32">
8632 <dim>1</dim>
8633 <dim>1</dim>
8634 <dim>768</dim>
8635 </port>
8636 </output>
8637 </layer>
8638 <layer id="627" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
8639 <data auto_broadcast="numpy" />
8640 <input>
8641 <port id="0" precision="FP32">
8642 <dim>-1</dim>
8643 <dim>-1</dim>
8644 <dim>768</dim>
8645 </port>
8646 <port id="1" precision="FP32">
8647 <dim>1</dim>
8648 <dim>1</dim>
8649 <dim>768</dim>
8650 </port>
8651 </input>
8652 <output>
8653 <port id="2" precision="FP32">
8654 <dim>-1</dim>
8655 <dim>-1</dim>
8656 <dim>768</dim>
8657 </port>
8658 </output>
8659 </layer>
8660 <layer id="628" name="Constant_5411" type="Const" version="opset1">
8661 <data element_type="f32" shape="1, 1, 768" offset="75044624" size="3072" />
8662 <output>
8663 <port id="0" precision="FP32">
8664 <dim>1</dim>
8665 <dim>1</dim>
8666 <dim>768</dim>
8667 </port>
8668 </output>
8669 </layer>
8670 <layer id="629" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
8671 <data auto_broadcast="numpy" />
8672 <input>
8673 <port id="0" precision="FP32">
8674 <dim>-1</dim>
8675 <dim>-1</dim>
8676 <dim>768</dim>
8677 </port>
8678 <port id="1" precision="FP32">
8679 <dim>1</dim>
8680 <dim>1</dim>
8681 <dim>768</dim>
8682 </port>
8683 </input>
8684 <output>
8685 <port id="2" precision="FP32" names="429,hidden_states.31">
8686 <dim>-1</dim>
8687 <dim>-1</dim>
8688 <dim>768</dim>
8689 </port>
8690 </output>
8691 </layer>
8692 <layer id="630" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
8693 <data element_type="f32" shape="1, 1, 768" offset="75047696" size="3072" />
8694 <output>
8695 <port id="0" precision="FP32">
8696 <dim>1</dim>
8697 <dim>1</dim>
8698 <dim>768</dim>
8699 </port>
8700 </output>
8701 </layer>
8702 <layer id="631" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
8703 <data auto_broadcast="numpy" />
8704 <input>
8705 <port id="0" precision="FP32">
8706 <dim>-1</dim>
8707 <dim>-1</dim>
8708 <dim>768</dim>
8709 </port>
8710 <port id="1" precision="FP32">
8711 <dim>1</dim>
8712 <dim>1</dim>
8713 <dim>768</dim>
8714 </port>
8715 </input>
8716 <output>
8717 <port id="2" precision="FP32">
8718 <dim>-1</dim>
8719 <dim>-1</dim>
8720 <dim>768</dim>
8721 </port>
8722 </output>
8723 </layer>
8724 <layer id="632" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
8725 <data element_type="f32" shape="" offset="75050768" size="4" />
8726 <output>
8727 <port id="0" precision="FP32" />
8728 </output>
8729 </layer>
8730 <layer id="633" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
8731 <data element_type="f32" shape="" offset="75050772" size="4" />
8732 <output>
8733 <port id="0" precision="FP32" />
8734 </output>
8735 </layer>
8736 <layer id="634" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
8737 <data element_type="f32" shape="" offset="75050768" size="4" />
8738 <output>
8739 <port id="0" precision="FP32" />
8740 </output>
8741 </layer>
8742 <layer id="635" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
8743 <data element_type="f32" shape="" offset="75050772" size="4" />
8744 <output>
8745 <port id="0" precision="FP32" />
8746 </output>
8747 </layer>
8748 <layer id="636" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
8749 <data levels="256" auto_broadcast="numpy" />
8750 <input>
8751 <port id="0" precision="FP32">
8752 <dim>-1</dim>
8753 <dim>-1</dim>
8754 <dim>768</dim>
8755 </port>
8756 <port id="1" precision="FP32" />
8757 <port id="2" precision="FP32" />
8758 <port id="3" precision="FP32" />
8759 <port id="4" precision="FP32" />
8760 </input>
8761 <output>
8762 <port id="5" precision="FP32">
8763 <dim>-1</dim>
8764 <dim>-1</dim>
8765 <dim>768</dim>
8766 </port>
8767 </output>
8768 </layer>
8769 <layer id="637" name="Constant_220614" type="Const" version="opset1">
8770 <data element_type="i8" shape="768, 768" offset="75050776" size="589824" />
8771 <output>
8772 <port id="0" precision="I8">
8773 <dim>768</dim>
8774 <dim>768</dim>
8775 </port>
8776 </output>
8777 </layer>
8778 <layer id="638" name="Convert_220615" type="Convert" version="opset1">
8779 <data destination_type="f32" />
8780 <input>
8781 <port id="0" precision="I8">
8782 <dim>768</dim>
8783 <dim>768</dim>
8784 </port>
8785 </input>
8786 <output>
8787 <port id="1" precision="FP32">
8788 <dim>768</dim>
8789 <dim>768</dim>
8790 </port>
8791 </output>
8792 </layer>
8793 <layer id="639" name="Constant_220616" type="Const" version="opset1">
8794 <data element_type="f32" shape="768, 1" offset="75640600" size="3072" />
8795 <output>
8796 <port id="0" precision="FP32">
8797 <dim>768</dim>
8798 <dim>1</dim>
8799 </port>
8800 </output>
8801 </layer>
8802 <layer id="640" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
8803 <data auto_broadcast="numpy" />
8804 <input>
8805 <port id="0" precision="FP32">
8806 <dim>768</dim>
8807 <dim>768</dim>
8808 </port>
8809 <port id="1" precision="FP32">
8810 <dim>768</dim>
8811 <dim>1</dim>
8812 </port>
8813 </input>
8814 <output>
8815 <port id="2" precision="FP32">
8816 <dim>768</dim>
8817 <dim>768</dim>
8818 </port>
8819 </output>
8820 </layer>
8821 <layer id="641" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
8822 <data transpose_a="false" transpose_b="true" />
8823 <input>
8824 <port id="0" precision="FP32">
8825 <dim>-1</dim>
8826 <dim>-1</dim>
8827 <dim>768</dim>
8828 </port>
8829 <port id="1" precision="FP32">
8830 <dim>768</dim>
8831 <dim>768</dim>
8832 </port>
8833 </input>
8834 <output>
8835 <port id="2" precision="FP32">
8836 <dim>-1</dim>
8837 <dim>-1</dim>
8838 <dim>768</dim>
8839 </port>
8840 </output>
8841 </layer>
8842 <layer id="642" name="Constant_5412" type="Const" version="opset1">
8843 <data element_type="f32" shape="1, 1, 768" offset="75643672" size="3072" />
8844 <output>
8845 <port id="0" precision="FP32">
8846 <dim>1</dim>
8847 <dim>1</dim>
8848 <dim>768</dim>
8849 </port>
8850 </output>
8851 </layer>
8852 <layer id="643" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add" type="Add" version="opset1">
8853 <data auto_broadcast="numpy" />
8854 <input>
8855 <port id="0" precision="FP32">
8856 <dim>-1</dim>
8857 <dim>-1</dim>
8858 <dim>768</dim>
8859 </port>
8860 <port id="1" precision="FP32">
8861 <dim>1</dim>
8862 <dim>1</dim>
8863 <dim>768</dim>
8864 </port>
8865 </input>
8866 <output>
8867 <port id="2" precision="FP32" names="442,x.61">
8868 <dim>-1</dim>
8869 <dim>-1</dim>
8870 <dim>768</dim>
8871 </port>
8872 </output>
8873 </layer>
8874 <layer id="644" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
8875 <data element_type="f32" shape="" offset="75646744" size="4" />
8876 <output>
8877 <port id="0" precision="FP32" />
8878 </output>
8879 </layer>
8880 <layer id="645" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
8881 <data element_type="f32" shape="" offset="75646748" size="4" />
8882 <output>
8883 <port id="0" precision="FP32" />
8884 </output>
8885 </layer>
8886 <layer id="646" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
8887 <data element_type="f32" shape="" offset="75646744" size="4" />
8888 <output>
8889 <port id="0" precision="FP32" />
8890 </output>
8891 </layer>
8892 <layer id="647" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
8893 <data element_type="f32" shape="" offset="75646748" size="4" />
8894 <output>
8895 <port id="0" precision="FP32" />
8896 </output>
8897 </layer>
8898 <layer id="648" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
8899 <data levels="256" auto_broadcast="numpy" />
8900 <input>
8901 <port id="0" precision="FP32">
8902 <dim>-1</dim>
8903 <dim>-1</dim>
8904 <dim>768</dim>
8905 </port>
8906 <port id="1" precision="FP32" />
8907 <port id="2" precision="FP32" />
8908 <port id="3" precision="FP32" />
8909 <port id="4" precision="FP32" />
8910 </input>
8911 <output>
8912 <port id="5" precision="FP32">
8913 <dim>-1</dim>
8914 <dim>-1</dim>
8915 <dim>768</dim>
8916 </port>
8917 </output>
8918 </layer>
8919 <layer id="649" name="__module.roberta.encoder.layer.5.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
8920 <data element_type="i64" shape="4" offset="39811496" size="32" />
8921 <output>
8922 <port id="0" precision="I64">
8923 <dim>4</dim>
8924 </port>
8925 </output>
8926 </layer>
8927 <layer id="650" name="__module.roberta.encoder.layer.5.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
8928 <data special_zero="true" />
8929 <input>
8930 <port id="0" precision="FP32">
8931 <dim>-1</dim>
8932 <dim>-1</dim>
8933 <dim>768</dim>
8934 </port>
8935 <port id="1" precision="I64">
8936 <dim>4</dim>
8937 </port>
8938 </input>
8939 <output>
8940 <port id="2" precision="FP32" names="446,x.63">
8941 <dim>-1</dim>
8942 <dim>-1</dim>
8943 <dim>12</dim>
8944 <dim>64</dim>
8945 </port>
8946 </output>
8947 </layer>
8948 <layer id="651" name="Constant_1344" type="Const" version="opset1">
8949 <data element_type="i64" shape="4" offset="39811528" size="32" />
8950 <output>
8951 <port id="0" precision="I64" names="447">
8952 <dim>4</dim>
8953 </port>
8954 </output>
8955 </layer>
8956 <layer id="652" name="__module.roberta.encoder.layer.5.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
8957 <input>
8958 <port id="0" precision="FP32">
8959 <dim>-1</dim>
8960 <dim>-1</dim>
8961 <dim>12</dim>
8962 <dim>64</dim>
8963 </port>
8964 <port id="1" precision="I64">
8965 <dim>4</dim>
8966 </port>
8967 </input>
8968 <output>
8969 <port id="2" precision="FP32" names="448">
8970 <dim>-1</dim>
8971 <dim>12</dim>
8972 <dim>-1</dim>
8973 <dim>64</dim>
8974 </port>
8975 </output>
8976 </layer>
8977 <layer id="653" name="Constant_220618" type="Const" version="opset1">
8978 <data element_type="i8" shape="768, 768" offset="75646752" size="589824" />
8979 <output>
8980 <port id="0" precision="I8">
8981 <dim>768</dim>
8982 <dim>768</dim>
8983 </port>
8984 </output>
8985 </layer>
8986 <layer id="654" name="Convert_220619" type="Convert" version="opset1">
8987 <data destination_type="f32" />
8988 <input>
8989 <port id="0" precision="I8">
8990 <dim>768</dim>
8991 <dim>768</dim>
8992 </port>
8993 </input>
8994 <output>
8995 <port id="1" precision="FP32">
8996 <dim>768</dim>
8997 <dim>768</dim>
8998 </port>
8999 </output>
9000 </layer>
9001 <layer id="655" name="Constant_220620" type="Const" version="opset1">
9002 <data element_type="f32" shape="768, 1" offset="76236576" size="3072" />
9003 <output>
9004 <port id="0" precision="FP32">
9005 <dim>768</dim>
9006 <dim>1</dim>
9007 </port>
9008 </output>
9009 </layer>
9010 <layer id="656" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9011 <data auto_broadcast="numpy" />
9012 <input>
9013 <port id="0" precision="FP32">
9014 <dim>768</dim>
9015 <dim>768</dim>
9016 </port>
9017 <port id="1" precision="FP32">
9018 <dim>768</dim>
9019 <dim>1</dim>
9020 </port>
9021 </input>
9022 <output>
9023 <port id="2" precision="FP32">
9024 <dim>768</dim>
9025 <dim>768</dim>
9026 </port>
9027 </output>
9028 </layer>
9029 <layer id="657" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
9030 <data transpose_a="false" transpose_b="true" />
9031 <input>
9032 <port id="0" precision="FP32">
9033 <dim>-1</dim>
9034 <dim>-1</dim>
9035 <dim>768</dim>
9036 </port>
9037 <port id="1" precision="FP32">
9038 <dim>768</dim>
9039 <dim>768</dim>
9040 </port>
9041 </input>
9042 <output>
9043 <port id="2" precision="FP32">
9044 <dim>-1</dim>
9045 <dim>-1</dim>
9046 <dim>768</dim>
9047 </port>
9048 </output>
9049 </layer>
9050 <layer id="658" name="Constant_5413" type="Const" version="opset1">
9051 <data element_type="f32" shape="1, 1, 768" offset="76239648" size="3072" />
9052 <output>
9053 <port id="0" precision="FP32">
9054 <dim>1</dim>
9055 <dim>1</dim>
9056 <dim>768</dim>
9057 </port>
9058 </output>
9059 </layer>
9060 <layer id="659" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add" type="Add" version="opset1">
9061 <data auto_broadcast="numpy" />
9062 <input>
9063 <port id="0" precision="FP32">
9064 <dim>-1</dim>
9065 <dim>-1</dim>
9066 <dim>768</dim>
9067 </port>
9068 <port id="1" precision="FP32">
9069 <dim>1</dim>
9070 <dim>1</dim>
9071 <dim>768</dim>
9072 </port>
9073 </input>
9074 <output>
9075 <port id="2" precision="FP32" names="451,x.65">
9076 <dim>-1</dim>
9077 <dim>-1</dim>
9078 <dim>768</dim>
9079 </port>
9080 </output>
9081 </layer>
9082 <layer id="660" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/input_low" type="Const" version="opset1">
9083 <data element_type="f32" shape="" offset="76242720" size="4" />
9084 <output>
9085 <port id="0" precision="FP32" />
9086 </output>
9087 </layer>
9088 <layer id="661" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/input_high" type="Const" version="opset1">
9089 <data element_type="f32" shape="" offset="76242724" size="4" />
9090 <output>
9091 <port id="0" precision="FP32" />
9092 </output>
9093 </layer>
9094 <layer id="662" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/output_low" type="Const" version="opset1">
9095 <data element_type="f32" shape="" offset="76242720" size="4" />
9096 <output>
9097 <port id="0" precision="FP32" />
9098 </output>
9099 </layer>
9100 <layer id="663" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0/output_high" type="Const" version="opset1">
9101 <data element_type="f32" shape="" offset="76242724" size="4" />
9102 <output>
9103 <port id="0" precision="FP32" />
9104 </output>
9105 </layer>
9106 <layer id="664" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add/fq_output_0" type="FakeQuantize" version="opset1">
9107 <data levels="256" auto_broadcast="numpy" />
9108 <input>
9109 <port id="0" precision="FP32">
9110 <dim>-1</dim>
9111 <dim>-1</dim>
9112 <dim>768</dim>
9113 </port>
9114 <port id="1" precision="FP32" />
9115 <port id="2" precision="FP32" />
9116 <port id="3" precision="FP32" />
9117 <port id="4" precision="FP32" />
9118 </input>
9119 <output>
9120 <port id="5" precision="FP32">
9121 <dim>-1</dim>
9122 <dim>-1</dim>
9123 <dim>768</dim>
9124 </port>
9125 </output>
9126 </layer>
9127 <layer id="665" name="__module.roberta.encoder.layer.5.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
9128 <data element_type="i64" shape="4" offset="39811496" size="32" />
9129 <output>
9130 <port id="0" precision="I64">
9131 <dim>4</dim>
9132 </port>
9133 </output>
9134 </layer>
9135 <layer id="666" name="__module.roberta.encoder.layer.5.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
9136 <data special_zero="true" />
9137 <input>
9138 <port id="0" precision="FP32">
9139 <dim>-1</dim>
9140 <dim>-1</dim>
9141 <dim>768</dim>
9142 </port>
9143 <port id="1" precision="I64">
9144 <dim>4</dim>
9145 </port>
9146 </input>
9147 <output>
9148 <port id="2" precision="FP32" names="455,x.67">
9149 <dim>-1</dim>
9150 <dim>-1</dim>
9151 <dim>12</dim>
9152 <dim>64</dim>
9153 </port>
9154 </output>
9155 </layer>
9156 <layer id="667" name="Constant_1367" type="Const" version="opset1">
9157 <data element_type="i64" shape="4" offset="39811528" size="32" />
9158 <output>
9159 <port id="0" precision="I64" names="456">
9160 <dim>4</dim>
9161 </port>
9162 </output>
9163 </layer>
9164 <layer id="668" name="__module.roberta.encoder.layer.5.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
9165 <input>
9166 <port id="0" precision="FP32">
9167 <dim>-1</dim>
9168 <dim>-1</dim>
9169 <dim>12</dim>
9170 <dim>64</dim>
9171 </port>
9172 <port id="1" precision="I64">
9173 <dim>4</dim>
9174 </port>
9175 </input>
9176 <output>
9177 <port id="2" precision="FP32" names="457">
9178 <dim>-1</dim>
9179 <dim>12</dim>
9180 <dim>-1</dim>
9181 <dim>64</dim>
9182 </port>
9183 </output>
9184 </layer>
9185 <layer id="669" name="Constant_220622" type="Const" version="opset1">
9186 <data element_type="i8" shape="768, 768" offset="76242728" size="589824" />
9187 <output>
9188 <port id="0" precision="I8">
9189 <dim>768</dim>
9190 <dim>768</dim>
9191 </port>
9192 </output>
9193 </layer>
9194 <layer id="670" name="Convert_220623" type="Convert" version="opset1">
9195 <data destination_type="f32" />
9196 <input>
9197 <port id="0" precision="I8">
9198 <dim>768</dim>
9199 <dim>768</dim>
9200 </port>
9201 </input>
9202 <output>
9203 <port id="1" precision="FP32">
9204 <dim>768</dim>
9205 <dim>768</dim>
9206 </port>
9207 </output>
9208 </layer>
9209 <layer id="671" name="Constant_220624" type="Const" version="opset1">
9210 <data element_type="f32" shape="768, 1" offset="76832552" size="3072" />
9211 <output>
9212 <port id="0" precision="FP32">
9213 <dim>768</dim>
9214 <dim>1</dim>
9215 </port>
9216 </output>
9217 </layer>
9218 <layer id="672" name="__module.roberta.encoder.layer.5.attention.self.value/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9219 <data auto_broadcast="numpy" />
9220 <input>
9221 <port id="0" precision="FP32">
9222 <dim>768</dim>
9223 <dim>768</dim>
9224 </port>
9225 <port id="1" precision="FP32">
9226 <dim>768</dim>
9227 <dim>1</dim>
9228 </port>
9229 </input>
9230 <output>
9231 <port id="2" precision="FP32">
9232 <dim>768</dim>
9233 <dim>768</dim>
9234 </port>
9235 </output>
9236 </layer>
9237 <layer id="673" name="__module.roberta.encoder.layer.5.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
9238 <data transpose_a="false" transpose_b="true" />
9239 <input>
9240 <port id="0" precision="FP32">
9241 <dim>-1</dim>
9242 <dim>-1</dim>
9243 <dim>768</dim>
9244 </port>
9245 <port id="1" precision="FP32">
9246 <dim>768</dim>
9247 <dim>768</dim>
9248 </port>
9249 </input>
9250 <output>
9251 <port id="2" precision="FP32">
9252 <dim>-1</dim>
9253 <dim>-1</dim>
9254 <dim>768</dim>
9255 </port>
9256 </output>
9257 </layer>
9258 <layer id="674" name="Constant_5414" type="Const" version="opset1">
9259 <data element_type="f32" shape="1, 1, 768" offset="76835624" size="3072" />
9260 <output>
9261 <port id="0" precision="FP32">
9262 <dim>1</dim>
9263 <dim>1</dim>
9264 <dim>768</dim>
9265 </port>
9266 </output>
9267 </layer>
9268 <layer id="675" name="__module.roberta.encoder.layer.5.attention.self.value/aten::linear/Add" type="Add" version="opset1">
9269 <data auto_broadcast="numpy" />
9270 <input>
9271 <port id="0" precision="FP32">
9272 <dim>-1</dim>
9273 <dim>-1</dim>
9274 <dim>768</dim>
9275 </port>
9276 <port id="1" precision="FP32">
9277 <dim>1</dim>
9278 <dim>1</dim>
9279 <dim>768</dim>
9280 </port>
9281 </input>
9282 <output>
9283 <port id="2" precision="FP32" names="460,x.69">
9284 <dim>-1</dim>
9285 <dim>-1</dim>
9286 <dim>768</dim>
9287 </port>
9288 </output>
9289 </layer>
9290 <layer id="676" name="__module.roberta.encoder.layer.5.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
9291 <data element_type="i64" shape="4" offset="39811496" size="32" />
9292 <output>
9293 <port id="0" precision="I64">
9294 <dim>4</dim>
9295 </port>
9296 </output>
9297 </layer>
9298 <layer id="677" name="__module.roberta.encoder.layer.5.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
9299 <data special_zero="true" />
9300 <input>
9301 <port id="0" precision="FP32">
9302 <dim>-1</dim>
9303 <dim>-1</dim>
9304 <dim>768</dim>
9305 </port>
9306 <port id="1" precision="I64">
9307 <dim>4</dim>
9308 </port>
9309 </input>
9310 <output>
9311 <port id="2" precision="FP32" names="464,x">
9312 <dim>-1</dim>
9313 <dim>-1</dim>
9314 <dim>12</dim>
9315 <dim>64</dim>
9316 </port>
9317 </output>
9318 </layer>
9319 <layer id="678" name="Constant_1390" type="Const" version="opset1">
9320 <data element_type="i64" shape="4" offset="39811528" size="32" />
9321 <output>
9322 <port id="0" precision="I64" names="465">
9323 <dim>4</dim>
9324 </port>
9325 </output>
9326 </layer>
9327 <layer id="679" name="__module.roberta.encoder.layer.5.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
9328 <input>
9329 <port id="0" precision="FP32">
9330 <dim>-1</dim>
9331 <dim>-1</dim>
9332 <dim>12</dim>
9333 <dim>64</dim>
9334 </port>
9335 <port id="1" precision="I64">
9336 <dim>4</dim>
9337 </port>
9338 </input>
9339 <output>
9340 <port id="2" precision="FP32" names="466">
9341 <dim>-1</dim>
9342 <dim>12</dim>
9343 <dim>-1</dim>
9344 <dim>64</dim>
9345 </port>
9346 </output>
9347 </layer>
9348 <layer id="680" name="__module.roberta.encoder.layer.5.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
9349 <data causal="false" />
9350 <input>
9351 <port id="0" precision="FP32">
9352 <dim>-1</dim>
9353 <dim>12</dim>
9354 <dim>-1</dim>
9355 <dim>64</dim>
9356 </port>
9357 <port id="1" precision="FP32">
9358 <dim>-1</dim>
9359 <dim>12</dim>
9360 <dim>-1</dim>
9361 <dim>64</dim>
9362 </port>
9363 <port id="2" precision="FP32">
9364 <dim>-1</dim>
9365 <dim>12</dim>
9366 <dim>-1</dim>
9367 <dim>64</dim>
9368 </port>
9369 <port id="3" precision="FP32">
9370 <dim>-1</dim>
9371 <dim>1</dim>
9372 <dim>-1</dim>
9373 <dim>-1</dim>
9374 </port>
9375 </input>
9376 <output>
9377 <port id="4" precision="FP32" names="467,attn_output.21">
9378 <dim>-1</dim>
9379 <dim>12</dim>
9380 <dim>-1</dim>
9381 <dim>64</dim>
9382 </port>
9383 </output>
9384 </layer>
9385 <layer id="681" name="__module.roberta.encoder.layer.5.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
9386 <data element_type="i32" shape="4" offset="41003536" size="16" />
9387 <output>
9388 <port id="0" precision="I32">
9389 <dim>4</dim>
9390 </port>
9391 </output>
9392 </layer>
9393 <layer id="682" name="__module.roberta.encoder.layer.5.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
9394 <input>
9395 <port id="0" precision="FP32">
9396 <dim>-1</dim>
9397 <dim>12</dim>
9398 <dim>-1</dim>
9399 <dim>64</dim>
9400 </port>
9401 <port id="1" precision="I32">
9402 <dim>4</dim>
9403 </port>
9404 </input>
9405 <output>
9406 <port id="2" precision="FP32" names="468,attn_output">
9407 <dim>-1</dim>
9408 <dim>-1</dim>
9409 <dim>12</dim>
9410 <dim>64</dim>
9411 </port>
9412 </output>
9413 </layer>
9414 <layer id="683" name="Constant_5499" type="Const" version="opset1">
9415 <data element_type="i64" shape="3" offset="41003552" size="24" />
9416 <output>
9417 <port id="0" precision="I64">
9418 <dim>3</dim>
9419 </port>
9420 </output>
9421 </layer>
9422 <layer id="684" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
9423 <data special_zero="true" />
9424 <input>
9425 <port id="0" precision="FP32">
9426 <dim>-1</dim>
9427 <dim>-1</dim>
9428 <dim>12</dim>
9429 <dim>64</dim>
9430 </port>
9431 <port id="1" precision="I64">
9432 <dim>3</dim>
9433 </port>
9434 </input>
9435 <output>
9436 <port id="2" precision="FP32" names="470">
9437 <dim>-1</dim>
9438 <dim>-1</dim>
9439 <dim>768</dim>
9440 </port>
9441 </output>
9442 </layer>
9443 <layer id="685" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9444 <data element_type="f32" shape="1, 1, 768" offset="76838696" size="3072" />
9445 <output>
9446 <port id="0" precision="FP32">
9447 <dim>1</dim>
9448 <dim>1</dim>
9449 <dim>768</dim>
9450 </port>
9451 </output>
9452 </layer>
9453 <layer id="686" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9454 <data auto_broadcast="numpy" />
9455 <input>
9456 <port id="0" precision="FP32">
9457 <dim>-1</dim>
9458 <dim>-1</dim>
9459 <dim>768</dim>
9460 </port>
9461 <port id="1" precision="FP32">
9462 <dim>1</dim>
9463 <dim>1</dim>
9464 <dim>768</dim>
9465 </port>
9466 </input>
9467 <output>
9468 <port id="2" precision="FP32">
9469 <dim>-1</dim>
9470 <dim>-1</dim>
9471 <dim>768</dim>
9472 </port>
9473 </output>
9474 </layer>
9475 <layer id="687" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9476 <data element_type="f32" shape="" offset="76841768" size="4" />
9477 <output>
9478 <port id="0" precision="FP32" />
9479 </output>
9480 </layer>
9481 <layer id="688" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9482 <data element_type="f32" shape="" offset="76841772" size="4" />
9483 <output>
9484 <port id="0" precision="FP32" />
9485 </output>
9486 </layer>
9487 <layer id="689" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9488 <data element_type="f32" shape="" offset="76841768" size="4" />
9489 <output>
9490 <port id="0" precision="FP32" />
9491 </output>
9492 </layer>
9493 <layer id="690" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9494 <data element_type="f32" shape="" offset="76841772" size="4" />
9495 <output>
9496 <port id="0" precision="FP32" />
9497 </output>
9498 </layer>
9499 <layer id="691" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9500 <data levels="256" auto_broadcast="numpy" />
9501 <input>
9502 <port id="0" precision="FP32">
9503 <dim>-1</dim>
9504 <dim>-1</dim>
9505 <dim>768</dim>
9506 </port>
9507 <port id="1" precision="FP32" />
9508 <port id="2" precision="FP32" />
9509 <port id="3" precision="FP32" />
9510 <port id="4" precision="FP32" />
9511 </input>
9512 <output>
9513 <port id="5" precision="FP32">
9514 <dim>-1</dim>
9515 <dim>-1</dim>
9516 <dim>768</dim>
9517 </port>
9518 </output>
9519 </layer>
9520 <layer id="692" name="Constant_220626" type="Const" version="opset1">
9521 <data element_type="i8" shape="768, 768" offset="76841776" size="589824" />
9522 <output>
9523 <port id="0" precision="I8">
9524 <dim>768</dim>
9525 <dim>768</dim>
9526 </port>
9527 </output>
9528 </layer>
9529 <layer id="693" name="Convert_220627" type="Convert" version="opset1">
9530 <data destination_type="f32" />
9531 <input>
9532 <port id="0" precision="I8">
9533 <dim>768</dim>
9534 <dim>768</dim>
9535 </port>
9536 </input>
9537 <output>
9538 <port id="1" precision="FP32">
9539 <dim>768</dim>
9540 <dim>768</dim>
9541 </port>
9542 </output>
9543 </layer>
9544 <layer id="694" name="Constant_220628" type="Const" version="opset1">
9545 <data element_type="f32" shape="768, 1" offset="77431600" size="3072" />
9546 <output>
9547 <port id="0" precision="FP32">
9548 <dim>768</dim>
9549 <dim>1</dim>
9550 </port>
9551 </output>
9552 </layer>
9553 <layer id="695" name="__module.roberta.encoder.layer.5.attention.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9554 <data auto_broadcast="numpy" />
9555 <input>
9556 <port id="0" precision="FP32">
9557 <dim>768</dim>
9558 <dim>768</dim>
9559 </port>
9560 <port id="1" precision="FP32">
9561 <dim>768</dim>
9562 <dim>1</dim>
9563 </port>
9564 </input>
9565 <output>
9566 <port id="2" precision="FP32">
9567 <dim>768</dim>
9568 <dim>768</dim>
9569 </port>
9570 </output>
9571 </layer>
9572 <layer id="696" name="__module.roberta.encoder.layer.5.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
9573 <data transpose_a="false" transpose_b="true" />
9574 <input>
9575 <port id="0" precision="FP32">
9576 <dim>-1</dim>
9577 <dim>-1</dim>
9578 <dim>768</dim>
9579 </port>
9580 <port id="1" precision="FP32">
9581 <dim>768</dim>
9582 <dim>768</dim>
9583 </port>
9584 </input>
9585 <output>
9586 <port id="2" precision="FP32">
9587 <dim>-1</dim>
9588 <dim>-1</dim>
9589 <dim>768</dim>
9590 </port>
9591 </output>
9592 </layer>
9593 <layer id="697" name="Constant_5415" type="Const" version="opset1">
9594 <data element_type="f32" shape="1, 1, 768" offset="77434672" size="3072" />
9595 <output>
9596 <port id="0" precision="FP32">
9597 <dim>1</dim>
9598 <dim>1</dim>
9599 <dim>768</dim>
9600 </port>
9601 </output>
9602 </layer>
9603 <layer id="698" name="__module.roberta.encoder.layer.5.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
9604 <data auto_broadcast="numpy" />
9605 <input>
9606 <port id="0" precision="FP32">
9607 <dim>-1</dim>
9608 <dim>-1</dim>
9609 <dim>768</dim>
9610 </port>
9611 <port id="1" precision="FP32">
9612 <dim>1</dim>
9613 <dim>1</dim>
9614 <dim>768</dim>
9615 </port>
9616 </input>
9617 <output>
9618 <port id="2" precision="FP32" names="475,input.23">
9619 <dim>-1</dim>
9620 <dim>-1</dim>
9621 <dim>768</dim>
9622 </port>
9623 </output>
9624 </layer>
9625 <layer id="699" name="__module.roberta.encoder.layer.5.attention.output/aten::add/Add" type="Add" version="opset1">
9626 <data auto_broadcast="numpy" />
9627 <input>
9628 <port id="0" precision="FP32">
9629 <dim>-1</dim>
9630 <dim>-1</dim>
9631 <dim>768</dim>
9632 </port>
9633 <port id="1" precision="FP32">
9634 <dim>-1</dim>
9635 <dim>-1</dim>
9636 <dim>768</dim>
9637 </port>
9638 </input>
9639 <output>
9640 <port id="2" precision="FP32" names="477">
9641 <dim>-1</dim>
9642 <dim>-1</dim>
9643 <dim>768</dim>
9644 </port>
9645 </output>
9646 </layer>
9647 <layer id="700" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
9648 <data element_type="i32" shape="1" offset="39206292" size="4" />
9649 <output>
9650 <port id="0" precision="I32">
9651 <dim>1</dim>
9652 </port>
9653 </output>
9654 </layer>
9655 <layer id="701" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
9656 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
9657 <input>
9658 <port id="0" precision="FP32">
9659 <dim>-1</dim>
9660 <dim>-1</dim>
9661 <dim>768</dim>
9662 </port>
9663 <port id="1" precision="I32">
9664 <dim>1</dim>
9665 </port>
9666 </input>
9667 <output>
9668 <port id="2" precision="FP32">
9669 <dim>-1</dim>
9670 <dim>-1</dim>
9671 <dim>768</dim>
9672 </port>
9673 </output>
9674 </layer>
9675 <layer id="702" name="Constant_5416" type="Const" version="opset1">
9676 <data element_type="f32" shape="1, 1, 768" offset="77437744" size="3072" />
9677 <output>
9678 <port id="0" precision="FP32">
9679 <dim>1</dim>
9680 <dim>1</dim>
9681 <dim>768</dim>
9682 </port>
9683 </output>
9684 </layer>
9685 <layer id="703" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
9686 <data auto_broadcast="numpy" />
9687 <input>
9688 <port id="0" precision="FP32">
9689 <dim>-1</dim>
9690 <dim>-1</dim>
9691 <dim>768</dim>
9692 </port>
9693 <port id="1" precision="FP32">
9694 <dim>1</dim>
9695 <dim>1</dim>
9696 <dim>768</dim>
9697 </port>
9698 </input>
9699 <output>
9700 <port id="2" precision="FP32">
9701 <dim>-1</dim>
9702 <dim>-1</dim>
9703 <dim>768</dim>
9704 </port>
9705 </output>
9706 </layer>
9707 <layer id="704" name="Constant_5417" type="Const" version="opset1">
9708 <data element_type="f32" shape="1, 1, 768" offset="77440816" size="3072" />
9709 <output>
9710 <port id="0" precision="FP32">
9711 <dim>1</dim>
9712 <dim>1</dim>
9713 <dim>768</dim>
9714 </port>
9715 </output>
9716 </layer>
9717 <layer id="705" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
9718 <data auto_broadcast="numpy" />
9719 <input>
9720 <port id="0" precision="FP32">
9721 <dim>-1</dim>
9722 <dim>-1</dim>
9723 <dim>768</dim>
9724 </port>
9725 <port id="1" precision="FP32">
9726 <dim>1</dim>
9727 <dim>1</dim>
9728 <dim>768</dim>
9729 </port>
9730 </input>
9731 <output>
9732 <port id="2" precision="FP32" names="481,input_tensor">
9733 <dim>-1</dim>
9734 <dim>-1</dim>
9735 <dim>768</dim>
9736 </port>
9737 </output>
9738 </layer>
9739 <layer id="706" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9740 <data element_type="f32" shape="1, 1, 768" offset="77443888" size="3072" />
9741 <output>
9742 <port id="0" precision="FP32">
9743 <dim>1</dim>
9744 <dim>1</dim>
9745 <dim>768</dim>
9746 </port>
9747 </output>
9748 </layer>
9749 <layer id="707" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9750 <data auto_broadcast="numpy" />
9751 <input>
9752 <port id="0" precision="FP32">
9753 <dim>-1</dim>
9754 <dim>-1</dim>
9755 <dim>768</dim>
9756 </port>
9757 <port id="1" precision="FP32">
9758 <dim>1</dim>
9759 <dim>1</dim>
9760 <dim>768</dim>
9761 </port>
9762 </input>
9763 <output>
9764 <port id="2" precision="FP32">
9765 <dim>-1</dim>
9766 <dim>-1</dim>
9767 <dim>768</dim>
9768 </port>
9769 </output>
9770 </layer>
9771 <layer id="708" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9772 <data element_type="f32" shape="" offset="77446960" size="4" />
9773 <output>
9774 <port id="0" precision="FP32" />
9775 </output>
9776 </layer>
9777 <layer id="709" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9778 <data element_type="f32" shape="" offset="77446964" size="4" />
9779 <output>
9780 <port id="0" precision="FP32" />
9781 </output>
9782 </layer>
9783 <layer id="710" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9784 <data element_type="f32" shape="" offset="77446960" size="4" />
9785 <output>
9786 <port id="0" precision="FP32" />
9787 </output>
9788 </layer>
9789 <layer id="711" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9790 <data element_type="f32" shape="" offset="77446964" size="4" />
9791 <output>
9792 <port id="0" precision="FP32" />
9793 </output>
9794 </layer>
9795 <layer id="712" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9796 <data levels="256" auto_broadcast="numpy" />
9797 <input>
9798 <port id="0" precision="FP32">
9799 <dim>-1</dim>
9800 <dim>-1</dim>
9801 <dim>768</dim>
9802 </port>
9803 <port id="1" precision="FP32" />
9804 <port id="2" precision="FP32" />
9805 <port id="3" precision="FP32" />
9806 <port id="4" precision="FP32" />
9807 </input>
9808 <output>
9809 <port id="5" precision="FP32">
9810 <dim>-1</dim>
9811 <dim>-1</dim>
9812 <dim>768</dim>
9813 </port>
9814 </output>
9815 </layer>
9816 <layer id="713" name="Constant_220630" type="Const" version="opset1">
9817 <data element_type="i8" shape="3072, 768" offset="77446968" size="2359296" />
9818 <output>
9819 <port id="0" precision="I8">
9820 <dim>3072</dim>
9821 <dim>768</dim>
9822 </port>
9823 </output>
9824 </layer>
9825 <layer id="714" name="Convert_220631" type="Convert" version="opset1">
9826 <data destination_type="f32" />
9827 <input>
9828 <port id="0" precision="I8">
9829 <dim>3072</dim>
9830 <dim>768</dim>
9831 </port>
9832 </input>
9833 <output>
9834 <port id="1" precision="FP32">
9835 <dim>3072</dim>
9836 <dim>768</dim>
9837 </port>
9838 </output>
9839 </layer>
9840 <layer id="715" name="Constant_220632" type="Const" version="opset1">
9841 <data element_type="f32" shape="3072, 1" offset="79806264" size="12288" />
9842 <output>
9843 <port id="0" precision="FP32">
9844 <dim>3072</dim>
9845 <dim>1</dim>
9846 </port>
9847 </output>
9848 </layer>
9849 <layer id="716" name="__module.roberta.encoder.layer.5.intermediate.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
9850 <data auto_broadcast="numpy" />
9851 <input>
9852 <port id="0" precision="FP32">
9853 <dim>3072</dim>
9854 <dim>768</dim>
9855 </port>
9856 <port id="1" precision="FP32">
9857 <dim>3072</dim>
9858 <dim>1</dim>
9859 </port>
9860 </input>
9861 <output>
9862 <port id="2" precision="FP32">
9863 <dim>3072</dim>
9864 <dim>768</dim>
9865 </port>
9866 </output>
9867 </layer>
9868 <layer id="717" name="__module.roberta.encoder.layer.5.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
9869 <data transpose_a="false" transpose_b="true" />
9870 <input>
9871 <port id="0" precision="FP32">
9872 <dim>-1</dim>
9873 <dim>-1</dim>
9874 <dim>768</dim>
9875 </port>
9876 <port id="1" precision="FP32">
9877 <dim>3072</dim>
9878 <dim>768</dim>
9879 </port>
9880 </input>
9881 <output>
9882 <port id="2" precision="FP32">
9883 <dim>-1</dim>
9884 <dim>-1</dim>
9885 <dim>3072</dim>
9886 </port>
9887 </output>
9888 </layer>
9889 <layer id="718" name="Constant_5418" type="Const" version="opset1">
9890 <data element_type="f32" shape="1, 1, 3072" offset="79818552" size="12288" />
9891 <output>
9892 <port id="0" precision="FP32">
9893 <dim>1</dim>
9894 <dim>1</dim>
9895 <dim>3072</dim>
9896 </port>
9897 </output>
9898 </layer>
9899 <layer id="719" name="__module.roberta.encoder.layer.5.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
9900 <data auto_broadcast="numpy" />
9901 <input>
9902 <port id="0" precision="FP32">
9903 <dim>-1</dim>
9904 <dim>-1</dim>
9905 <dim>3072</dim>
9906 </port>
9907 <port id="1" precision="FP32">
9908 <dim>1</dim>
9909 <dim>1</dim>
9910 <dim>3072</dim>
9911 </port>
9912 </input>
9913 <output>
9914 <port id="2" precision="FP32" names="485">
9915 <dim>-1</dim>
9916 <dim>-1</dim>
9917 <dim>3072</dim>
9918 </port>
9919 </output>
9920 </layer>
9921 <layer id="720" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
9922 <data approximation_mode="ERF" />
9923 <input>
9924 <port id="0" precision="FP32">
9925 <dim>-1</dim>
9926 <dim>-1</dim>
9927 <dim>3072</dim>
9928 </port>
9929 </input>
9930 <output>
9931 <port id="1" precision="FP32" names="486">
9932 <dim>-1</dim>
9933 <dim>-1</dim>
9934 <dim>3072</dim>
9935 </port>
9936 </output>
9937 </layer>
9938 <layer id="721" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
9939 <data element_type="f32" shape="1, 1, 3072" offset="79830840" size="12288" />
9940 <output>
9941 <port id="0" precision="FP32">
9942 <dim>1</dim>
9943 <dim>1</dim>
9944 <dim>3072</dim>
9945 </port>
9946 </output>
9947 </layer>
9948 <layer id="722" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
9949 <data auto_broadcast="numpy" />
9950 <input>
9951 <port id="0" precision="FP32">
9952 <dim>-1</dim>
9953 <dim>-1</dim>
9954 <dim>3072</dim>
9955 </port>
9956 <port id="1" precision="FP32">
9957 <dim>1</dim>
9958 <dim>1</dim>
9959 <dim>3072</dim>
9960 </port>
9961 </input>
9962 <output>
9963 <port id="2" precision="FP32">
9964 <dim>-1</dim>
9965 <dim>-1</dim>
9966 <dim>3072</dim>
9967 </port>
9968 </output>
9969 </layer>
9970 <layer id="723" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
9971 <data element_type="f32" shape="" offset="79843128" size="4" />
9972 <output>
9973 <port id="0" precision="FP32" />
9974 </output>
9975 </layer>
9976 <layer id="724" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
9977 <data element_type="f32" shape="" offset="79843132" size="4" />
9978 <output>
9979 <port id="0" precision="FP32" />
9980 </output>
9981 </layer>
9982 <layer id="725" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
9983 <data element_type="f32" shape="" offset="79843128" size="4" />
9984 <output>
9985 <port id="0" precision="FP32" />
9986 </output>
9987 </layer>
9988 <layer id="726" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
9989 <data element_type="f32" shape="" offset="79843132" size="4" />
9990 <output>
9991 <port id="0" precision="FP32" />
9992 </output>
9993 </layer>
9994 <layer id="727" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
9995 <data levels="256" auto_broadcast="numpy" />
9996 <input>
9997 <port id="0" precision="FP32">
9998 <dim>-1</dim>
9999 <dim>-1</dim>
10000 <dim>3072</dim>
10001 </port>
10002 <port id="1" precision="FP32" />
10003 <port id="2" precision="FP32" />
10004 <port id="3" precision="FP32" />
10005 <port id="4" precision="FP32" />
10006 </input>
10007 <output>
10008 <port id="5" precision="FP32">
10009 <dim>-1</dim>
10010 <dim>-1</dim>
10011 <dim>3072</dim>
10012 </port>
10013 </output>
10014 </layer>
10015 <layer id="728" name="Constant_220634" type="Const" version="opset1">
10016 <data element_type="i8" shape="768, 3072" offset="79843136" size="2359296" />
10017 <output>
10018 <port id="0" precision="I8">
10019 <dim>768</dim>
10020 <dim>3072</dim>
10021 </port>
10022 </output>
10023 </layer>
10024 <layer id="729" name="Convert_220635" type="Convert" version="opset1">
10025 <data destination_type="f32" />
10026 <input>
10027 <port id="0" precision="I8">
10028 <dim>768</dim>
10029 <dim>3072</dim>
10030 </port>
10031 </input>
10032 <output>
10033 <port id="1" precision="FP32">
10034 <dim>768</dim>
10035 <dim>3072</dim>
10036 </port>
10037 </output>
10038 </layer>
10039 <layer id="730" name="Constant_220636" type="Const" version="opset1">
10040 <data element_type="f32" shape="768, 1" offset="82202432" size="3072" />
10041 <output>
10042 <port id="0" precision="FP32">
10043 <dim>768</dim>
10044 <dim>1</dim>
10045 </port>
10046 </output>
10047 </layer>
10048 <layer id="731" name="__module.roberta.encoder.layer.5.output.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
10049 <data auto_broadcast="numpy" />
10050 <input>
10051 <port id="0" precision="FP32">
10052 <dim>768</dim>
10053 <dim>3072</dim>
10054 </port>
10055 <port id="1" precision="FP32">
10056 <dim>768</dim>
10057 <dim>1</dim>
10058 </port>
10059 </input>
10060 <output>
10061 <port id="2" precision="FP32">
10062 <dim>768</dim>
10063 <dim>3072</dim>
10064 </port>
10065 </output>
10066 </layer>
10067 <layer id="732" name="__module.roberta.encoder.layer.5.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
10068 <data transpose_a="false" transpose_b="true" />
10069 <input>
10070 <port id="0" precision="FP32">
10071 <dim>-1</dim>
10072 <dim>-1</dim>
10073 <dim>3072</dim>
10074 </port>
10075 <port id="1" precision="FP32">
10076 <dim>768</dim>
10077 <dim>3072</dim>
10078 </port>
10079 </input>
10080 <output>
10081 <port id="2" precision="FP32">
10082 <dim>-1</dim>
10083 <dim>-1</dim>
10084 <dim>768</dim>
10085 </port>
10086 </output>
10087 </layer>
10088 <layer id="733" name="Constant_5419" type="Const" version="opset1">
10089 <data element_type="f32" shape="1, 1, 768" offset="82205504" size="3072" />
10090 <output>
10091 <port id="0" precision="FP32">
10092 <dim>1</dim>
10093 <dim>1</dim>
10094 <dim>768</dim>
10095 </port>
10096 </output>
10097 </layer>
10098 <layer id="734" name="__module.roberta.encoder.layer.5.output.dense/aten::linear/Add" type="Add" version="opset1">
10099 <data auto_broadcast="numpy" />
10100 <input>
10101 <port id="0" precision="FP32">
10102 <dim>-1</dim>
10103 <dim>-1</dim>
10104 <dim>768</dim>
10105 </port>
10106 <port id="1" precision="FP32">
10107 <dim>1</dim>
10108 <dim>1</dim>
10109 <dim>768</dim>
10110 </port>
10111 </input>
10112 <output>
10113 <port id="2" precision="FP32" names="491,input.25">
10114 <dim>-1</dim>
10115 <dim>-1</dim>
10116 <dim>768</dim>
10117 </port>
10118 </output>
10119 </layer>
10120 <layer id="735" name="__module.roberta.encoder.layer.5.output/aten::add/Add" type="Add" version="opset1">
10121 <data auto_broadcast="numpy" />
10122 <input>
10123 <port id="0" precision="FP32">
10124 <dim>-1</dim>
10125 <dim>-1</dim>
10126 <dim>768</dim>
10127 </port>
10128 <port id="1" precision="FP32">
10129 <dim>-1</dim>
10130 <dim>-1</dim>
10131 <dim>768</dim>
10132 </port>
10133 </input>
10134 <output>
10135 <port id="2" precision="FP32" names="493">
10136 <dim>-1</dim>
10137 <dim>-1</dim>
10138 <dim>768</dim>
10139 </port>
10140 </output>
10141 </layer>
10142 <layer id="736" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
10143 <data element_type="i32" shape="1" offset="39206292" size="4" />
10144 <output>
10145 <port id="0" precision="I32">
10146 <dim>1</dim>
10147 </port>
10148 </output>
10149 </layer>
10150 <layer id="737" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
10151 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
10152 <input>
10153 <port id="0" precision="FP32">
10154 <dim>-1</dim>
10155 <dim>-1</dim>
10156 <dim>768</dim>
10157 </port>
10158 <port id="1" precision="I32">
10159 <dim>1</dim>
10160 </port>
10161 </input>
10162 <output>
10163 <port id="2" precision="FP32">
10164 <dim>-1</dim>
10165 <dim>-1</dim>
10166 <dim>768</dim>
10167 </port>
10168 </output>
10169 </layer>
10170 <layer id="738" name="Constant_5420" type="Const" version="opset1">
10171 <data element_type="f32" shape="1, 1, 768" offset="82208576" size="3072" />
10172 <output>
10173 <port id="0" precision="FP32">
10174 <dim>1</dim>
10175 <dim>1</dim>
10176 <dim>768</dim>
10177 </port>
10178 </output>
10179 </layer>
10180 <layer id="739" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
10181 <data auto_broadcast="numpy" />
10182 <input>
10183 <port id="0" precision="FP32">
10184 <dim>-1</dim>
10185 <dim>-1</dim>
10186 <dim>768</dim>
10187 </port>
10188 <port id="1" precision="FP32">
10189 <dim>1</dim>
10190 <dim>1</dim>
10191 <dim>768</dim>
10192 </port>
10193 </input>
10194 <output>
10195 <port id="2" precision="FP32">
10196 <dim>-1</dim>
10197 <dim>-1</dim>
10198 <dim>768</dim>
10199 </port>
10200 </output>
10201 </layer>
10202 <layer id="740" name="Constant_5421" type="Const" version="opset1">
10203 <data element_type="f32" shape="1, 1, 768" offset="82211648" size="3072" />
10204 <output>
10205 <port id="0" precision="FP32">
10206 <dim>1</dim>
10207 <dim>1</dim>
10208 <dim>768</dim>
10209 </port>
10210 </output>
10211 </layer>
10212 <layer id="741" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
10213 <data auto_broadcast="numpy" />
10214 <input>
10215 <port id="0" precision="FP32">
10216 <dim>-1</dim>
10217 <dim>-1</dim>
10218 <dim>768</dim>
10219 </port>
10220 <port id="1" precision="FP32">
10221 <dim>1</dim>
10222 <dim>1</dim>
10223 <dim>768</dim>
10224 </port>
10225 </input>
10226 <output>
10227 <port id="2" precision="FP32" names="497,506,features">
10228 <dim>-1</dim>
10229 <dim>-1</dim>
10230 <dim>768</dim>
10231 </port>
10232 </output>
10233 </layer>
10234 <layer id="742" name="502" type="Const" version="opset1">
10235 <data element_type="i64" shape="" offset="38809468" size="8" />
10236 <output>
10237 <port id="0" precision="I64" names="502" />
10238 </output>
10239 </layer>
10240 <layer id="743" name="500" type="Const" version="opset1">
10241 <data element_type="i64" shape="" offset="38809476" size="8" />
10242 <output>
10243 <port id="0" precision="I64" names="500" />
10244 </output>
10245 </layer>
10246 <layer id="744" name="__module.classifier/aten::select/Gather" type="Gather" version="opset8">
10247 <data batch_dims="0" />
10248 <input>
10249 <port id="0" precision="FP32">
10250 <dim>-1</dim>
10251 <dim>-1</dim>
10252 <dim>768</dim>
10253 </port>
10254 <port id="1" precision="I64" />
10255 <port id="2" precision="I64" />
10256 </input>
10257 <output>
10258 <port id="3" precision="FP32" names="507,508,input.27">
10259 <dim>-1</dim>
10260 <dim>768</dim>
10261 </port>
10262 </output>
10263 </layer>
10264 <layer id="745" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
10265 <data element_type="f32" shape="1, 768" offset="82214720" size="3072" />
10266 <output>
10267 <port id="0" precision="FP32">
10268 <dim>1</dim>
10269 <dim>768</dim>
10270 </port>
10271 </output>
10272 </layer>
10273 <layer id="746" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
10274 <data auto_broadcast="numpy" />
10275 <input>
10276 <port id="0" precision="FP32">
10277 <dim>-1</dim>
10278 <dim>768</dim>
10279 </port>
10280 <port id="1" precision="FP32">
10281 <dim>1</dim>
10282 <dim>768</dim>
10283 </port>
10284 </input>
10285 <output>
10286 <port id="2" precision="FP32">
10287 <dim>-1</dim>
10288 <dim>768</dim>
10289 </port>
10290 </output>
10291 </layer>
10292 <layer id="747" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
10293 <data element_type="f32" shape="" offset="82217792" size="4" />
10294 <output>
10295 <port id="0" precision="FP32" />
10296 </output>
10297 </layer>
10298 <layer id="748" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
10299 <data element_type="f32" shape="" offset="82217796" size="4" />
10300 <output>
10301 <port id="0" precision="FP32" />
10302 </output>
10303 </layer>
10304 <layer id="749" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
10305 <data element_type="f32" shape="" offset="82217792" size="4" />
10306 <output>
10307 <port id="0" precision="FP32" />
10308 </output>
10309 </layer>
10310 <layer id="750" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
10311 <data element_type="f32" shape="" offset="82217796" size="4" />
10312 <output>
10313 <port id="0" precision="FP32" />
10314 </output>
10315 </layer>
10316 <layer id="751" name="__module.classifier/aten::select/Gather_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
10317 <data levels="256" auto_broadcast="numpy" />
10318 <input>
10319 <port id="0" precision="FP32">
10320 <dim>-1</dim>
10321 <dim>768</dim>
10322 </port>
10323 <port id="1" precision="FP32" />
10324 <port id="2" precision="FP32" />
10325 <port id="3" precision="FP32" />
10326 <port id="4" precision="FP32" />
10327 </input>
10328 <output>
10329 <port id="5" precision="FP32">
10330 <dim>-1</dim>
10331 <dim>768</dim>
10332 </port>
10333 </output>
10334 </layer>
10335 <layer id="752" name="Constant_220638" type="Const" version="opset1">
10336 <data element_type="i8" shape="768, 768" offset="82217800" size="589824" />
10337 <output>
10338 <port id="0" precision="I8">
10339 <dim>768</dim>
10340 <dim>768</dim>
10341 </port>
10342 </output>
10343 </layer>
10344 <layer id="753" name="Convert_220639" type="Convert" version="opset1">
10345 <data destination_type="f32" />
10346 <input>
10347 <port id="0" precision="I8">
10348 <dim>768</dim>
10349 <dim>768</dim>
10350 </port>
10351 </input>
10352 <output>
10353 <port id="1" precision="FP32">
10354 <dim>768</dim>
10355 <dim>768</dim>
10356 </port>
10357 </output>
10358 </layer>
10359 <layer id="754" name="Constant_220640" type="Const" version="opset1">
10360 <data element_type="f32" shape="768, 1" offset="82807624" size="3072" />
10361 <output>
10362 <port id="0" precision="FP32">
10363 <dim>768</dim>
10364 <dim>1</dim>
10365 </port>
10366 </output>
10367 </layer>
10368 <layer id="755" name="__module.classifier.dense/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
10369 <data auto_broadcast="numpy" />
10370 <input>
10371 <port id="0" precision="FP32">
10372 <dim>768</dim>
10373 <dim>768</dim>
10374 </port>
10375 <port id="1" precision="FP32">
10376 <dim>768</dim>
10377 <dim>1</dim>
10378 </port>
10379 </input>
10380 <output>
10381 <port id="2" precision="FP32">
10382 <dim>768</dim>
10383 <dim>768</dim>
10384 </port>
10385 </output>
10386 </layer>
10387 <layer id="756" name="__module.classifier.dense/aten::linear/MatMul" type="MatMul" version="opset1">
10388 <data transpose_a="false" transpose_b="true" />
10389 <input>
10390 <port id="0" precision="FP32">
10391 <dim>-1</dim>
10392 <dim>768</dim>
10393 </port>
10394 <port id="1" precision="FP32">
10395 <dim>768</dim>
10396 <dim>768</dim>
10397 </port>
10398 </input>
10399 <output>
10400 <port id="2" precision="FP32">
10401 <dim>-1</dim>
10402 <dim>768</dim>
10403 </port>
10404 </output>
10405 </layer>
10406 <layer id="757" name="Constant_5422" type="Const" version="opset1">
10407 <data element_type="f32" shape="1, 768" offset="82810696" size="3072" />
10408 <output>
10409 <port id="0" precision="FP32">
10410 <dim>1</dim>
10411 <dim>768</dim>
10412 </port>
10413 </output>
10414 </layer>
10415 <layer id="758" name="__module.classifier.dense/aten::linear/Add" type="Add" version="opset1">
10416 <data auto_broadcast="numpy" />
10417 <input>
10418 <port id="0" precision="FP32">
10419 <dim>-1</dim>
10420 <dim>768</dim>
10421 </port>
10422 <port id="1" precision="FP32">
10423 <dim>1</dim>
10424 <dim>768</dim>
10425 </port>
10426 </input>
10427 <output>
10428 <port id="2" precision="FP32" names="512">
10429 <dim>-1</dim>
10430 <dim>768</dim>
10431 </port>
10432 </output>
10433 </layer>
10434 <layer id="759" name="__module.classifier/aten::tanh/Tanh" type="Tanh" version="opset1">
10435 <input>
10436 <port id="0" precision="FP32">
10437 <dim>-1</dim>
10438 <dim>768</dim>
10439 </port>
10440 </input>
10441 <output>
10442 <port id="1" precision="FP32" names="513,input">
10443 <dim>-1</dim>
10444 <dim>768</dim>
10445 </port>
10446 </output>
10447 </layer>
10448 <layer id="760" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant/scale" type="Const" version="opset1">
10449 <data element_type="f32" shape="1, 768" offset="82813768" size="3072" />
10450 <output>
10451 <port id="0" precision="FP32">
10452 <dim>1</dim>
10453 <dim>768</dim>
10454 </port>
10455 </output>
10456 </layer>
10457 <layer id="761" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant" type="Multiply" version="opset1">
10458 <data auto_broadcast="numpy" />
10459 <input>
10460 <port id="0" precision="FP32">
10461 <dim>-1</dim>
10462 <dim>768</dim>
10463 </port>
10464 <port id="1" precision="FP32">
10465 <dim>1</dim>
10466 <dim>768</dim>
10467 </port>
10468 </input>
10469 <output>
10470 <port id="2" precision="FP32">
10471 <dim>-1</dim>
10472 <dim>768</dim>
10473 </port>
10474 </output>
10475 </layer>
10476 <layer id="762" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/input_low" type="Const" version="opset1">
10477 <data element_type="f32" shape="" offset="82816840" size="4" />
10478 <output>
10479 <port id="0" precision="FP32" />
10480 </output>
10481 </layer>
10482 <layer id="763" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/input_high" type="Const" version="opset1">
10483 <data element_type="f32" shape="" offset="82816844" size="4" />
10484 <output>
10485 <port id="0" precision="FP32" />
10486 </output>
10487 </layer>
10488 <layer id="764" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/output_low" type="Const" version="opset1">
10489 <data element_type="f32" shape="" offset="82816840" size="4" />
10490 <output>
10491 <port id="0" precision="FP32" />
10492 </output>
10493 </layer>
10494 <layer id="765" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0/output_high" type="Const" version="opset1">
10495 <data element_type="f32" shape="" offset="82816844" size="4" />
10496 <output>
10497 <port id="0" precision="FP32" />
10498 </output>
10499 </layer>
10500 <layer id="766" name="__module.classifier/aten::tanh/Tanh_0_0/nncf_smooth_quant/fq_output_0" type="FakeQuantize" version="opset1">
10501 <data levels="256" auto_broadcast="numpy" />
10502 <input>
10503 <port id="0" precision="FP32">
10504 <dim>-1</dim>
10505 <dim>768</dim>
10506 </port>
10507 <port id="1" precision="FP32" />
10508 <port id="2" precision="FP32" />
10509 <port id="3" precision="FP32" />
10510 <port id="4" precision="FP32" />
10511 </input>
10512 <output>
10513 <port id="5" precision="FP32">
10514 <dim>-1</dim>
10515 <dim>768</dim>
10516 </port>
10517 </output>
10518 </layer>
10519 <layer id="767" name="Constant_220642" type="Const" version="opset1">
10520 <data element_type="i8" shape="3, 768" offset="82816848" size="2304" />
10521 <output>
10522 <port id="0" precision="I8">
10523 <dim>3</dim>
10524 <dim>768</dim>
10525 </port>
10526 </output>
10527 </layer>
10528 <layer id="768" name="Convert_220643" type="Convert" version="opset1">
10529 <data destination_type="f32" />
10530 <input>
10531 <port id="0" precision="I8">
10532 <dim>3</dim>
10533 <dim>768</dim>
10534 </port>
10535 </input>
10536 <output>
10537 <port id="1" precision="FP32">
10538 <dim>3</dim>
10539 <dim>768</dim>
10540 </port>
10541 </output>
10542 </layer>
10543 <layer id="769" name="Constant_220644" type="Const" version="opset1">
10544 <data element_type="f32" shape="3, 1" offset="82819152" size="12" />
10545 <output>
10546 <port id="0" precision="FP32">
10547 <dim>3</dim>
10548 <dim>1</dim>
10549 </port>
10550 </output>
10551 </layer>
10552 <layer id="770" name="__module.classifier.out_proj/aten::linear/MatMul/fq_weights_1" type="Multiply" version="opset1">
10553 <data auto_broadcast="numpy" />
10554 <input>
10555 <port id="0" precision="FP32">
10556 <dim>3</dim>
10557 <dim>768</dim>
10558 </port>
10559 <port id="1" precision="FP32">
10560 <dim>3</dim>
10561 <dim>1</dim>
10562 </port>
10563 </input>
10564 <output>
10565 <port id="2" precision="FP32">
10566 <dim>3</dim>
10567 <dim>768</dim>
10568 </port>
10569 </output>
10570 </layer>
10571 <layer id="771" name="__module.classifier.out_proj/aten::linear/MatMul" type="MatMul" version="opset1">
10572 <data transpose_a="false" transpose_b="true" />
10573 <input>
10574 <port id="0" precision="FP32">
10575 <dim>-1</dim>
10576 <dim>768</dim>
10577 </port>
10578 <port id="1" precision="FP32">
10579 <dim>3</dim>
10580 <dim>768</dim>
10581 </port>
10582 </input>
10583 <output>
10584 <port id="2" precision="FP32">
10585 <dim>-1</dim>
10586 <dim>3</dim>
10587 </port>
10588 </output>
10589 </layer>
10590 <layer id="772" name="Constant_5423" type="Const" version="opset1">
10591 <data element_type="f32" shape="1, 3" offset="82819164" size="12" />
10592 <output>
10593 <port id="0" precision="FP32">
10594 <dim>1</dim>
10595 <dim>3</dim>
10596 </port>
10597 </output>
10598 </layer>
10599 <layer id="773" name="__module.classifier.out_proj/aten::linear/Add" type="Add" version="opset1">
10600 <data auto_broadcast="numpy" />
10601 <input>
10602 <port id="0" precision="FP32">
10603 <dim>-1</dim>
10604 <dim>3</dim>
10605 </port>
10606 <port id="1" precision="FP32">
10607 <dim>1</dim>
10608 <dim>3</dim>
10609 </port>
10610 </input>
10611 <output>
10612 <port id="2" precision="FP32" names="logits">
10613 <dim>-1</dim>
10614 <dim>3</dim>
10615 </port>
10616 </output>
10617 </layer>
10618 <layer id="774" name="Result_2399" type="Result" version="opset1">
10619 <input>
10620 <port id="0" precision="FP32">
10621 <dim>-1</dim>
10622 <dim>3</dim>
10623 </port>
10624 </input>
10625 </layer>
10626 </layers>
10627 <edges>
10628 <edge from-layer="0" from-port="0" to-layer="101" to-port="0" />
10629 <edge from-layer="1" from-port="0" to-layer="6" to-port="0" />
10630 <edge from-layer="1" from-port="0" to-layer="15" to-port="0" />
10631 <edge from-layer="1" from-port="0" to-layer="32" to-port="0" />
10632 <edge from-layer="2" from-port="0" to-layer="3" to-port="0" />
10633 <edge from-layer="3" from-port="1" to-layer="5" to-port="0" />
10634 <edge from-layer="4" from-port="0" to-layer="5" to-port="1" />
10635 <edge from-layer="5" from-port="2" to-layer="8" to-port="0" />
10636 <edge from-layer="6" from-port="1" to-layer="8" to-port="1" />
10637 <edge from-layer="7" from-port="0" to-layer="8" to-port="2" />
10638 <edge from-layer="8" from-port="3" to-layer="26" to-port="0" />
10639 <edge from-layer="9" from-port="0" to-layer="10" to-port="0" />
10640 <edge from-layer="10" from-port="1" to-layer="12" to-port="0" />
10641 <edge from-layer="11" from-port="0" to-layer="12" to-port="1" />
10642 <edge from-layer="12" from-port="2" to-layer="25" to-port="0" />
10643 <edge from-layer="13" from-port="0" to-layer="21" to-port="0" />
10644 <edge from-layer="14" from-port="0" to-layer="21" to-port="1" />
10645 <edge from-layer="15" from-port="1" to-layer="18" to-port="0" />
10646 <edge from-layer="15" from-port="1" to-layer="22" to-port="1" />
10647 <edge from-layer="15" from-port="1" to-layer="110" to-port="0" />
10648 <edge from-layer="15" from-port="1" to-layer="106" to-port="0" />
10649 <edge from-layer="16" from-port="0" to-layer="18" to-port="1" />
10650 <edge from-layer="17" from-port="0" to-layer="18" to-port="2" />
10651 <edge from-layer="18" from-port="3" to-layer="21" to-port="2" />
10652 <edge from-layer="19" from-port="0" to-layer="21" to-port="3" />
10653 <edge from-layer="20" from-port="0" to-layer="21" to-port="4" />
10654 <edge from-layer="21" from-port="5" to-layer="22" to-port="0" />
10655 <edge from-layer="22" from-port="2" to-layer="23" to-port="0" />
10656 <edge from-layer="23" from-port="1" to-layer="25" to-port="1" />
10657 <edge from-layer="24" from-port="0" to-layer="25" to-port="2" />
10658 <edge from-layer="25" from-port="3" to-layer="26" to-port="1" />
10659 <edge from-layer="26" from-port="2" to-layer="43" to-port="0" />
10660 <edge from-layer="27" from-port="0" to-layer="28" to-port="0" />
10661 <edge from-layer="28" from-port="1" to-layer="30" to-port="0" />
10662 <edge from-layer="29" from-port="0" to-layer="30" to-port="1" />
10663 <edge from-layer="30" from-port="2" to-layer="42" to-port="0" />
10664 <edge from-layer="31" from-port="0" to-layer="32" to-port="1" />
10665 <edge from-layer="32" from-port="2" to-layer="33" to-port="0" />
10666 <edge from-layer="33" from-port="1" to-layer="35" to-port="0" />
10667 <edge from-layer="33" from-port="1" to-layer="36" to-port="1" />
10668 <edge from-layer="34" from-port="0" to-layer="35" to-port="1" />
10669 <edge from-layer="34" from-port="0" to-layer="101" to-port="1" />
10670 <edge from-layer="35" from-port="2" to-layer="36" to-port="0" />
10671 <edge from-layer="36" from-port="2" to-layer="37" to-port="0" />
10672 <edge from-layer="37" from-port="1" to-layer="39" to-port="0" />
10673 <edge from-layer="38" from-port="0" to-layer="39" to-port="1" />
10674 <edge from-layer="39" from-port="2" to-layer="40" to-port="0" />
10675 <edge from-layer="40" from-port="1" to-layer="42" to-port="1" />
10676 <edge from-layer="41" from-port="0" to-layer="42" to-port="2" />
10677 <edge from-layer="42" from-port="3" to-layer="43" to-port="1" />
10678 <edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
10679 <edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
10680 <edge from-layer="45" from-port="2" to-layer="47" to-port="0" />
10681 <edge from-layer="46" from-port="0" to-layer="47" to-port="1" />
10682 <edge from-layer="47" from-port="2" to-layer="49" to-port="0" />
10683 <edge from-layer="48" from-port="0" to-layer="49" to-port="1" />
10684 <edge from-layer="49" from-port="2" to-layer="51" to-port="0" />
10685 <edge from-layer="49" from-port="2" to-layer="139" to-port="1" />
10686 <edge from-layer="50" from-port="0" to-layer="51" to-port="1" />
10687 <edge from-layer="51" from-port="2" to-layer="56" to-port="0" />
10688 <edge from-layer="52" from-port="0" to-layer="56" to-port="1" />
10689 <edge from-layer="53" from-port="0" to-layer="56" to-port="2" />
10690 <edge from-layer="54" from-port="0" to-layer="56" to-port="3" />
10691 <edge from-layer="55" from-port="0" to-layer="56" to-port="4" />
10692 <edge from-layer="56" from-port="5" to-layer="61" to-port="0" />
10693 <edge from-layer="56" from-port="5" to-layer="77" to-port="0" />
10694 <edge from-layer="56" from-port="5" to-layer="93" to-port="0" />
10695 <edge from-layer="57" from-port="0" to-layer="58" to-port="0" />
10696 <edge from-layer="58" from-port="1" to-layer="60" to-port="0" />
10697 <edge from-layer="59" from-port="0" to-layer="60" to-port="1" />
10698 <edge from-layer="60" from-port="2" to-layer="61" to-port="1" />
10699 <edge from-layer="61" from-port="2" to-layer="63" to-port="0" />
10700 <edge from-layer="62" from-port="0" to-layer="63" to-port="1" />
10701 <edge from-layer="63" from-port="2" to-layer="68" to-port="0" />
10702 <edge from-layer="64" from-port="0" to-layer="68" to-port="1" />
10703 <edge from-layer="65" from-port="0" to-layer="68" to-port="2" />
10704 <edge from-layer="66" from-port="0" to-layer="68" to-port="3" />
10705 <edge from-layer="67" from-port="0" to-layer="68" to-port="4" />
10706 <edge from-layer="68" from-port="5" to-layer="70" to-port="0" />
10707 <edge from-layer="69" from-port="0" to-layer="70" to-port="1" />
10708 <edge from-layer="70" from-port="2" to-layer="72" to-port="0" />
10709 <edge from-layer="71" from-port="0" to-layer="72" to-port="1" />
10710 <edge from-layer="72" from-port="2" to-layer="120" to-port="0" />
10711 <edge from-layer="73" from-port="0" to-layer="74" to-port="0" />
10712 <edge from-layer="74" from-port="1" to-layer="76" to-port="0" />
10713 <edge from-layer="75" from-port="0" to-layer="76" to-port="1" />
10714 <edge from-layer="76" from-port="2" to-layer="77" to-port="1" />
10715 <edge from-layer="77" from-port="2" to-layer="79" to-port="0" />
10716 <edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
10717 <edge from-layer="79" from-port="2" to-layer="84" to-port="0" />
10718 <edge from-layer="80" from-port="0" to-layer="84" to-port="1" />
10719 <edge from-layer="81" from-port="0" to-layer="84" to-port="2" />
10720 <edge from-layer="82" from-port="0" to-layer="84" to-port="3" />
10721 <edge from-layer="83" from-port="0" to-layer="84" to-port="4" />
10722 <edge from-layer="84" from-port="5" to-layer="86" to-port="0" />
10723 <edge from-layer="85" from-port="0" to-layer="86" to-port="1" />
10724 <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
10725 <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
10726 <edge from-layer="88" from-port="2" to-layer="120" to-port="1" />
10727 <edge from-layer="89" from-port="0" to-layer="90" to-port="0" />
10728 <edge from-layer="90" from-port="1" to-layer="92" to-port="0" />
10729 <edge from-layer="91" from-port="0" to-layer="92" to-port="1" />
10730 <edge from-layer="92" from-port="2" to-layer="93" to-port="1" />
10731 <edge from-layer="93" from-port="2" to-layer="95" to-port="0" />
10732 <edge from-layer="94" from-port="0" to-layer="95" to-port="1" />
10733 <edge from-layer="95" from-port="2" to-layer="97" to-port="0" />
10734 <edge from-layer="96" from-port="0" to-layer="97" to-port="1" />
10735 <edge from-layer="97" from-port="2" to-layer="99" to-port="0" />
10736 <edge from-layer="98" from-port="0" to-layer="99" to-port="1" />
10737 <edge from-layer="99" from-port="2" to-layer="120" to-port="2" />
10738 <edge from-layer="100" from-port="0" to-layer="116" to-port="0" />
10739 <edge from-layer="101" from-port="2" to-layer="103" to-port="0" />
10740 <edge from-layer="102" from-port="0" to-layer="103" to-port="1" />
10741 <edge from-layer="103" from-port="2" to-layer="112" to-port="0" />
10742 <edge from-layer="104" from-port="0" to-layer="106" to-port="1" />
10743 <edge from-layer="105" from-port="0" to-layer="106" to-port="2" />
10744 <edge from-layer="106" from-port="3" to-layer="111" to-port="0" />
10745 <edge from-layer="107" from-port="0" to-layer="111" to-port="1" />
10746 <edge from-layer="108" from-port="0" to-layer="110" to-port="1" />
10747 <edge from-layer="109" from-port="0" to-layer="110" to-port="2" />
10748 <edge from-layer="110" from-port="3" to-layer="111" to-port="2" />
10749 <edge from-layer="111" from-port="3" to-layer="112" to-port="1" />
10750 <edge from-layer="112" from-port="2" to-layer="113" to-port="0" />
10751 <edge from-layer="113" from-port="1" to-layer="115" to-port="0" />
10752 <edge from-layer="114" from-port="0" to-layer="115" to-port="1" />
10753 <edge from-layer="115" from-port="2" to-layer="116" to-port="1" />
10754 <edge from-layer="116" from-port="2" to-layer="117" to-port="0" />
10755 <edge from-layer="116" from-port="2" to-layer="119" to-port="2" />
10756 <edge from-layer="117" from-port="1" to-layer="119" to-port="0" />
10757 <edge from-layer="118" from-port="0" to-layer="119" to-port="1" />
10758 <edge from-layer="119" from-port="3" to-layer="120" to-port="3" />
10759 <edge from-layer="119" from-port="3" to-layer="456" to-port="3" />
10760 <edge from-layer="119" from-port="3" to-layer="568" to-port="3" />
10761 <edge from-layer="119" from-port="3" to-layer="232" to-port="3" />
10762 <edge from-layer="119" from-port="3" to-layer="680" to-port="3" />
10763 <edge from-layer="119" from-port="3" to-layer="344" to-port="3" />
10764 <edge from-layer="120" from-port="4" to-layer="122" to-port="0" />
10765 <edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
10766 <edge from-layer="122" from-port="2" to-layer="124" to-port="0" />
10767 <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
10768 <edge from-layer="124" from-port="2" to-layer="126" to-port="0" />
10769 <edge from-layer="125" from-port="0" to-layer="126" to-port="1" />
10770 <edge from-layer="126" from-port="2" to-layer="131" to-port="0" />
10771 <edge from-layer="127" from-port="0" to-layer="131" to-port="1" />
10772 <edge from-layer="128" from-port="0" to-layer="131" to-port="2" />
10773 <edge from-layer="129" from-port="0" to-layer="131" to-port="3" />
10774 <edge from-layer="130" from-port="0" to-layer="131" to-port="4" />
10775 <edge from-layer="131" from-port="5" to-layer="136" to-port="0" />
10776 <edge from-layer="132" from-port="0" to-layer="133" to-port="0" />
10777 <edge from-layer="133" from-port="1" to-layer="135" to-port="0" />
10778 <edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
10779 <edge from-layer="135" from-port="2" to-layer="136" to-port="1" />
10780 <edge from-layer="136" from-port="2" to-layer="138" to-port="0" />
10781 <edge from-layer="137" from-port="0" to-layer="138" to-port="1" />
10782 <edge from-layer="138" from-port="2" to-layer="139" to-port="0" />
10783 <edge from-layer="139" from-port="2" to-layer="141" to-port="0" />
10784 <edge from-layer="140" from-port="0" to-layer="141" to-port="1" />
10785 <edge from-layer="141" from-port="2" to-layer="143" to-port="0" />
10786 <edge from-layer="142" from-port="0" to-layer="143" to-port="1" />
10787 <edge from-layer="143" from-port="2" to-layer="145" to-port="0" />
10788 <edge from-layer="144" from-port="0" to-layer="145" to-port="1" />
10789 <edge from-layer="145" from-port="2" to-layer="147" to-port="0" />
10790 <edge from-layer="145" from-port="2" to-layer="175" to-port="1" />
10791 <edge from-layer="146" from-port="0" to-layer="147" to-port="1" />
10792 <edge from-layer="147" from-port="2" to-layer="152" to-port="0" />
10793 <edge from-layer="148" from-port="0" to-layer="152" to-port="1" />
10794 <edge from-layer="149" from-port="0" to-layer="152" to-port="2" />
10795 <edge from-layer="150" from-port="0" to-layer="152" to-port="3" />
10796 <edge from-layer="151" from-port="0" to-layer="152" to-port="4" />
10797 <edge from-layer="152" from-port="5" to-layer="157" to-port="0" />
10798 <edge from-layer="153" from-port="0" to-layer="154" to-port="0" />
10799 <edge from-layer="154" from-port="1" to-layer="156" to-port="0" />
10800 <edge from-layer="155" from-port="0" to-layer="156" to-port="1" />
10801 <edge from-layer="156" from-port="2" to-layer="157" to-port="1" />
10802 <edge from-layer="157" from-port="2" to-layer="159" to-port="0" />
10803 <edge from-layer="158" from-port="0" to-layer="159" to-port="1" />
10804 <edge from-layer="159" from-port="2" to-layer="160" to-port="0" />
10805 <edge from-layer="160" from-port="1" to-layer="162" to-port="0" />
10806 <edge from-layer="161" from-port="0" to-layer="162" to-port="1" />
10807 <edge from-layer="162" from-port="2" to-layer="167" to-port="0" />
10808 <edge from-layer="163" from-port="0" to-layer="167" to-port="1" />
10809 <edge from-layer="164" from-port="0" to-layer="167" to-port="2" />
10810 <edge from-layer="165" from-port="0" to-layer="167" to-port="3" />
10811 <edge from-layer="166" from-port="0" to-layer="167" to-port="4" />
10812 <edge from-layer="167" from-port="5" to-layer="172" to-port="0" />
10813 <edge from-layer="168" from-port="0" to-layer="169" to-port="0" />
10814 <edge from-layer="169" from-port="1" to-layer="171" to-port="0" />
10815 <edge from-layer="170" from-port="0" to-layer="171" to-port="1" />
10816 <edge from-layer="171" from-port="2" to-layer="172" to-port="1" />
10817 <edge from-layer="172" from-port="2" to-layer="174" to-port="0" />
10818 <edge from-layer="173" from-port="0" to-layer="174" to-port="1" />
10819 <edge from-layer="174" from-port="2" to-layer="175" to-port="0" />
10820 <edge from-layer="175" from-port="2" to-layer="177" to-port="0" />
10821 <edge from-layer="176" from-port="0" to-layer="177" to-port="1" />
10822 <edge from-layer="177" from-port="2" to-layer="179" to-port="0" />
10823 <edge from-layer="178" from-port="0" to-layer="179" to-port="1" />
10824 <edge from-layer="179" from-port="2" to-layer="181" to-port="0" />
10825 <edge from-layer="180" from-port="0" to-layer="181" to-port="1" />
10826 <edge from-layer="181" from-port="2" to-layer="183" to-port="0" />
10827 <edge from-layer="181" from-port="2" to-layer="251" to-port="1" />
10828 <edge from-layer="182" from-port="0" to-layer="183" to-port="1" />
10829 <edge from-layer="183" from-port="2" to-layer="188" to-port="0" />
10830 <edge from-layer="184" from-port="0" to-layer="188" to-port="1" />
10831 <edge from-layer="185" from-port="0" to-layer="188" to-port="2" />
10832 <edge from-layer="186" from-port="0" to-layer="188" to-port="3" />
10833 <edge from-layer="187" from-port="0" to-layer="188" to-port="4" />
10834 <edge from-layer="188" from-port="5" to-layer="193" to-port="0" />
10835 <edge from-layer="188" from-port="5" to-layer="225" to-port="0" />
10836 <edge from-layer="188" from-port="5" to-layer="209" to-port="0" />
10837 <edge from-layer="189" from-port="0" to-layer="190" to-port="0" />
10838 <edge from-layer="190" from-port="1" to-layer="192" to-port="0" />
10839 <edge from-layer="191" from-port="0" to-layer="192" to-port="1" />
10840 <edge from-layer="192" from-port="2" to-layer="193" to-port="1" />
10841 <edge from-layer="193" from-port="2" to-layer="195" to-port="0" />
10842 <edge from-layer="194" from-port="0" to-layer="195" to-port="1" />
10843 <edge from-layer="195" from-port="2" to-layer="200" to-port="0" />
10844 <edge from-layer="196" from-port="0" to-layer="200" to-port="1" />
10845 <edge from-layer="197" from-port="0" to-layer="200" to-port="2" />
10846 <edge from-layer="198" from-port="0" to-layer="200" to-port="3" />
10847 <edge from-layer="199" from-port="0" to-layer="200" to-port="4" />
10848 <edge from-layer="200" from-port="5" to-layer="202" to-port="0" />
10849 <edge from-layer="201" from-port="0" to-layer="202" to-port="1" />
10850 <edge from-layer="202" from-port="2" to-layer="204" to-port="0" />
10851 <edge from-layer="203" from-port="0" to-layer="204" to-port="1" />
10852 <edge from-layer="204" from-port="2" to-layer="232" to-port="0" />
10853 <edge from-layer="205" from-port="0" to-layer="206" to-port="0" />
10854 <edge from-layer="206" from-port="1" to-layer="208" to-port="0" />
10855 <edge from-layer="207" from-port="0" to-layer="208" to-port="1" />
10856 <edge from-layer="208" from-port="2" to-layer="209" to-port="1" />
10857 <edge from-layer="209" from-port="2" to-layer="211" to-port="0" />
10858 <edge from-layer="210" from-port="0" to-layer="211" to-port="1" />
10859 <edge from-layer="211" from-port="2" to-layer="216" to-port="0" />
10860 <edge from-layer="212" from-port="0" to-layer="216" to-port="1" />
10861 <edge from-layer="213" from-port="0" to-layer="216" to-port="2" />
10862 <edge from-layer="214" from-port="0" to-layer="216" to-port="3" />
10863 <edge from-layer="215" from-port="0" to-layer="216" to-port="4" />
10864 <edge from-layer="216" from-port="5" to-layer="218" to-port="0" />
10865 <edge from-layer="217" from-port="0" to-layer="218" to-port="1" />
10866 <edge from-layer="218" from-port="2" to-layer="220" to-port="0" />
10867 <edge from-layer="219" from-port="0" to-layer="220" to-port="1" />
10868 <edge from-layer="220" from-port="2" to-layer="232" to-port="1" />
10869 <edge from-layer="221" from-port="0" to-layer="222" to-port="0" />
10870 <edge from-layer="222" from-port="1" to-layer="224" to-port="0" />
10871 <edge from-layer="223" from-port="0" to-layer="224" to-port="1" />
10872 <edge from-layer="224" from-port="2" to-layer="225" to-port="1" />
10873 <edge from-layer="225" from-port="2" to-layer="227" to-port="0" />
10874 <edge from-layer="226" from-port="0" to-layer="227" to-port="1" />
10875 <edge from-layer="227" from-port="2" to-layer="229" to-port="0" />
10876 <edge from-layer="228" from-port="0" to-layer="229" to-port="1" />
10877 <edge from-layer="229" from-port="2" to-layer="231" to-port="0" />
10878 <edge from-layer="230" from-port="0" to-layer="231" to-port="1" />
10879 <edge from-layer="231" from-port="2" to-layer="232" to-port="2" />
10880 <edge from-layer="232" from-port="4" to-layer="234" to-port="0" />
10881 <edge from-layer="233" from-port="0" to-layer="234" to-port="1" />
10882 <edge from-layer="234" from-port="2" to-layer="236" to-port="0" />
10883 <edge from-layer="235" from-port="0" to-layer="236" to-port="1" />
10884 <edge from-layer="236" from-port="2" to-layer="238" to-port="0" />
10885 <edge from-layer="237" from-port="0" to-layer="238" to-port="1" />
10886 <edge from-layer="238" from-port="2" to-layer="243" to-port="0" />
10887 <edge from-layer="239" from-port="0" to-layer="243" to-port="1" />
10888 <edge from-layer="240" from-port="0" to-layer="243" to-port="2" />
10889 <edge from-layer="241" from-port="0" to-layer="243" to-port="3" />
10890 <edge from-layer="242" from-port="0" to-layer="243" to-port="4" />
10891 <edge from-layer="243" from-port="5" to-layer="248" to-port="0" />
10892 <edge from-layer="244" from-port="0" to-layer="245" to-port="0" />
10893 <edge from-layer="245" from-port="1" to-layer="247" to-port="0" />
10894 <edge from-layer="246" from-port="0" to-layer="247" to-port="1" />
10895 <edge from-layer="247" from-port="2" to-layer="248" to-port="1" />
10896 <edge from-layer="248" from-port="2" to-layer="250" to-port="0" />
10897 <edge from-layer="249" from-port="0" to-layer="250" to-port="1" />
10898 <edge from-layer="250" from-port="2" to-layer="251" to-port="0" />
10899 <edge from-layer="251" from-port="2" to-layer="253" to-port="0" />
10900 <edge from-layer="252" from-port="0" to-layer="253" to-port="1" />
10901 <edge from-layer="253" from-port="2" to-layer="255" to-port="0" />
10902 <edge from-layer="254" from-port="0" to-layer="255" to-port="1" />
10903 <edge from-layer="255" from-port="2" to-layer="257" to-port="0" />
10904 <edge from-layer="256" from-port="0" to-layer="257" to-port="1" />
10905 <edge from-layer="257" from-port="2" to-layer="259" to-port="0" />
10906 <edge from-layer="257" from-port="2" to-layer="287" to-port="1" />
10907 <edge from-layer="258" from-port="0" to-layer="259" to-port="1" />
10908 <edge from-layer="259" from-port="2" to-layer="264" to-port="0" />
10909 <edge from-layer="260" from-port="0" to-layer="264" to-port="1" />
10910 <edge from-layer="261" from-port="0" to-layer="264" to-port="2" />
10911 <edge from-layer="262" from-port="0" to-layer="264" to-port="3" />
10912 <edge from-layer="263" from-port="0" to-layer="264" to-port="4" />
10913 <edge from-layer="264" from-port="5" to-layer="269" to-port="0" />
10914 <edge from-layer="265" from-port="0" to-layer="266" to-port="0" />
10915 <edge from-layer="266" from-port="1" to-layer="268" to-port="0" />
10916 <edge from-layer="267" from-port="0" to-layer="268" to-port="1" />
10917 <edge from-layer="268" from-port="2" to-layer="269" to-port="1" />
10918 <edge from-layer="269" from-port="2" to-layer="271" to-port="0" />
10919 <edge from-layer="270" from-port="0" to-layer="271" to-port="1" />
10920 <edge from-layer="271" from-port="2" to-layer="272" to-port="0" />
10921 <edge from-layer="272" from-port="1" to-layer="274" to-port="0" />
10922 <edge from-layer="273" from-port="0" to-layer="274" to-port="1" />
10923 <edge from-layer="274" from-port="2" to-layer="279" to-port="0" />
10924 <edge from-layer="275" from-port="0" to-layer="279" to-port="1" />
10925 <edge from-layer="276" from-port="0" to-layer="279" to-port="2" />
10926 <edge from-layer="277" from-port="0" to-layer="279" to-port="3" />
10927 <edge from-layer="278" from-port="0" to-layer="279" to-port="4" />
10928 <edge from-layer="279" from-port="5" to-layer="284" to-port="0" />
10929 <edge from-layer="280" from-port="0" to-layer="281" to-port="0" />
10930 <edge from-layer="281" from-port="1" to-layer="283" to-port="0" />
10931 <edge from-layer="282" from-port="0" to-layer="283" to-port="1" />
10932 <edge from-layer="283" from-port="2" to-layer="284" to-port="1" />
10933 <edge from-layer="284" from-port="2" to-layer="286" to-port="0" />
10934 <edge from-layer="285" from-port="0" to-layer="286" to-port="1" />
10935 <edge from-layer="286" from-port="2" to-layer="287" to-port="0" />
10936 <edge from-layer="287" from-port="2" to-layer="289" to-port="0" />
10937 <edge from-layer="288" from-port="0" to-layer="289" to-port="1" />
10938 <edge from-layer="289" from-port="2" to-layer="291" to-port="0" />
10939 <edge from-layer="290" from-port="0" to-layer="291" to-port="1" />
10940 <edge from-layer="291" from-port="2" to-layer="293" to-port="0" />
10941 <edge from-layer="292" from-port="0" to-layer="293" to-port="1" />
10942 <edge from-layer="293" from-port="2" to-layer="363" to-port="1" />
10943 <edge from-layer="293" from-port="2" to-layer="295" to-port="0" />
10944 <edge from-layer="294" from-port="0" to-layer="295" to-port="1" />
10945 <edge from-layer="295" from-port="2" to-layer="300" to-port="0" />
10946 <edge from-layer="296" from-port="0" to-layer="300" to-port="1" />
10947 <edge from-layer="297" from-port="0" to-layer="300" to-port="2" />
10948 <edge from-layer="298" from-port="0" to-layer="300" to-port="3" />
10949 <edge from-layer="299" from-port="0" to-layer="300" to-port="4" />
10950 <edge from-layer="300" from-port="5" to-layer="305" to-port="0" />
10951 <edge from-layer="300" from-port="5" to-layer="337" to-port="0" />
10952 <edge from-layer="300" from-port="5" to-layer="321" to-port="0" />
10953 <edge from-layer="301" from-port="0" to-layer="302" to-port="0" />
10954 <edge from-layer="302" from-port="1" to-layer="304" to-port="0" />
10955 <edge from-layer="303" from-port="0" to-layer="304" to-port="1" />
10956 <edge from-layer="304" from-port="2" to-layer="305" to-port="1" />
10957 <edge from-layer="305" from-port="2" to-layer="307" to-port="0" />
10958 <edge from-layer="306" from-port="0" to-layer="307" to-port="1" />
10959 <edge from-layer="307" from-port="2" to-layer="312" to-port="0" />
10960 <edge from-layer="308" from-port="0" to-layer="312" to-port="1" />
10961 <edge from-layer="309" from-port="0" to-layer="312" to-port="2" />
10962 <edge from-layer="310" from-port="0" to-layer="312" to-port="3" />
10963 <edge from-layer="311" from-port="0" to-layer="312" to-port="4" />
10964 <edge from-layer="312" from-port="5" to-layer="314" to-port="0" />
10965 <edge from-layer="313" from-port="0" to-layer="314" to-port="1" />
10966 <edge from-layer="314" from-port="2" to-layer="316" to-port="0" />
10967 <edge from-layer="315" from-port="0" to-layer="316" to-port="1" />
10968 <edge from-layer="316" from-port="2" to-layer="344" to-port="0" />
10969 <edge from-layer="317" from-port="0" to-layer="318" to-port="0" />
10970 <edge from-layer="318" from-port="1" to-layer="320" to-port="0" />
10971 <edge from-layer="319" from-port="0" to-layer="320" to-port="1" />
10972 <edge from-layer="320" from-port="2" to-layer="321" to-port="1" />
10973 <edge from-layer="321" from-port="2" to-layer="323" to-port="0" />
10974 <edge from-layer="322" from-port="0" to-layer="323" to-port="1" />
10975 <edge from-layer="323" from-port="2" to-layer="328" to-port="0" />
10976 <edge from-layer="324" from-port="0" to-layer="328" to-port="1" />
10977 <edge from-layer="325" from-port="0" to-layer="328" to-port="2" />
10978 <edge from-layer="326" from-port="0" to-layer="328" to-port="3" />
10979 <edge from-layer="327" from-port="0" to-layer="328" to-port="4" />
10980 <edge from-layer="328" from-port="5" to-layer="330" to-port="0" />
10981 <edge from-layer="329" from-port="0" to-layer="330" to-port="1" />
10982 <edge from-layer="330" from-port="2" to-layer="332" to-port="0" />
10983 <edge from-layer="331" from-port="0" to-layer="332" to-port="1" />
10984 <edge from-layer="332" from-port="2" to-layer="344" to-port="1" />
10985 <edge from-layer="333" from-port="0" to-layer="334" to-port="0" />
10986 <edge from-layer="334" from-port="1" to-layer="336" to-port="0" />
10987 <edge from-layer="335" from-port="0" to-layer="336" to-port="1" />
10988 <edge from-layer="336" from-port="2" to-layer="337" to-port="1" />
10989 <edge from-layer="337" from-port="2" to-layer="339" to-port="0" />
10990 <edge from-layer="338" from-port="0" to-layer="339" to-port="1" />
10991 <edge from-layer="339" from-port="2" to-layer="341" to-port="0" />
10992 <edge from-layer="340" from-port="0" to-layer="341" to-port="1" />
10993 <edge from-layer="341" from-port="2" to-layer="343" to-port="0" />
10994 <edge from-layer="342" from-port="0" to-layer="343" to-port="1" />
10995 <edge from-layer="343" from-port="2" to-layer="344" to-port="2" />
10996 <edge from-layer="344" from-port="4" to-layer="346" to-port="0" />
10997 <edge from-layer="345" from-port="0" to-layer="346" to-port="1" />
10998 <edge from-layer="346" from-port="2" to-layer="348" to-port="0" />
10999 <edge from-layer="347" from-port="0" to-layer="348" to-port="1" />
11000 <edge from-layer="348" from-port="2" to-layer="350" to-port="0" />
11001 <edge from-layer="349" from-port="0" to-layer="350" to-port="1" />
11002 <edge from-layer="350" from-port="2" to-layer="355" to-port="0" />
11003 <edge from-layer="351" from-port="0" to-layer="355" to-port="1" />
11004 <edge from-layer="352" from-port="0" to-layer="355" to-port="2" />
11005 <edge from-layer="353" from-port="0" to-layer="355" to-port="3" />
11006 <edge from-layer="354" from-port="0" to-layer="355" to-port="4" />
11007 <edge from-layer="355" from-port="5" to-layer="360" to-port="0" />
11008 <edge from-layer="356" from-port="0" to-layer="357" to-port="0" />
11009 <edge from-layer="357" from-port="1" to-layer="359" to-port="0" />
11010 <edge from-layer="358" from-port="0" to-layer="359" to-port="1" />
11011 <edge from-layer="359" from-port="2" to-layer="360" to-port="1" />
11012 <edge from-layer="360" from-port="2" to-layer="362" to-port="0" />
11013 <edge from-layer="361" from-port="0" to-layer="362" to-port="1" />
11014 <edge from-layer="362" from-port="2" to-layer="363" to-port="0" />
11015 <edge from-layer="363" from-port="2" to-layer="365" to-port="0" />
11016 <edge from-layer="364" from-port="0" to-layer="365" to-port="1" />
11017 <edge from-layer="365" from-port="2" to-layer="367" to-port="0" />
11018 <edge from-layer="366" from-port="0" to-layer="367" to-port="1" />
11019 <edge from-layer="367" from-port="2" to-layer="369" to-port="0" />
11020 <edge from-layer="368" from-port="0" to-layer="369" to-port="1" />
11021 <edge from-layer="369" from-port="2" to-layer="371" to-port="0" />
11022 <edge from-layer="369" from-port="2" to-layer="399" to-port="1" />
11023 <edge from-layer="370" from-port="0" to-layer="371" to-port="1" />
11024 <edge from-layer="371" from-port="2" to-layer="376" to-port="0" />
11025 <edge from-layer="372" from-port="0" to-layer="376" to-port="1" />
11026 <edge from-layer="373" from-port="0" to-layer="376" to-port="2" />
11027 <edge from-layer="374" from-port="0" to-layer="376" to-port="3" />
11028 <edge from-layer="375" from-port="0" to-layer="376" to-port="4" />
11029 <edge from-layer="376" from-port="5" to-layer="381" to-port="0" />
11030 <edge from-layer="377" from-port="0" to-layer="378" to-port="0" />
11031 <edge from-layer="378" from-port="1" to-layer="380" to-port="0" />
11032 <edge from-layer="379" from-port="0" to-layer="380" to-port="1" />
11033 <edge from-layer="380" from-port="2" to-layer="381" to-port="1" />
11034 <edge from-layer="381" from-port="2" to-layer="383" to-port="0" />
11035 <edge from-layer="382" from-port="0" to-layer="383" to-port="1" />
11036 <edge from-layer="383" from-port="2" to-layer="384" to-port="0" />
11037 <edge from-layer="384" from-port="1" to-layer="386" to-port="0" />
11038 <edge from-layer="385" from-port="0" to-layer="386" to-port="1" />
11039 <edge from-layer="386" from-port="2" to-layer="391" to-port="0" />
11040 <edge from-layer="387" from-port="0" to-layer="391" to-port="1" />
11041 <edge from-layer="388" from-port="0" to-layer="391" to-port="2" />
11042 <edge from-layer="389" from-port="0" to-layer="391" to-port="3" />
11043 <edge from-layer="390" from-port="0" to-layer="391" to-port="4" />
11044 <edge from-layer="391" from-port="5" to-layer="396" to-port="0" />
11045 <edge from-layer="392" from-port="0" to-layer="393" to-port="0" />
11046 <edge from-layer="393" from-port="1" to-layer="395" to-port="0" />
11047 <edge from-layer="394" from-port="0" to-layer="395" to-port="1" />
11048 <edge from-layer="395" from-port="2" to-layer="396" to-port="1" />
11049 <edge from-layer="396" from-port="2" to-layer="398" to-port="0" />
11050 <edge from-layer="397" from-port="0" to-layer="398" to-port="1" />
11051 <edge from-layer="398" from-port="2" to-layer="399" to-port="0" />
11052 <edge from-layer="399" from-port="2" to-layer="401" to-port="0" />
11053 <edge from-layer="400" from-port="0" to-layer="401" to-port="1" />
11054 <edge from-layer="401" from-port="2" to-layer="403" to-port="0" />
11055 <edge from-layer="402" from-port="0" to-layer="403" to-port="1" />
11056 <edge from-layer="403" from-port="2" to-layer="405" to-port="0" />
11057 <edge from-layer="404" from-port="0" to-layer="405" to-port="1" />
11058 <edge from-layer="405" from-port="2" to-layer="407" to-port="0" />
11059 <edge from-layer="405" from-port="2" to-layer="475" to-port="1" />
11060 <edge from-layer="406" from-port="0" to-layer="407" to-port="1" />
11061 <edge from-layer="407" from-port="2" to-layer="412" to-port="0" />
11062 <edge from-layer="408" from-port="0" to-layer="412" to-port="1" />
11063 <edge from-layer="409" from-port="0" to-layer="412" to-port="2" />
11064 <edge from-layer="410" from-port="0" to-layer="412" to-port="3" />
11065 <edge from-layer="411" from-port="0" to-layer="412" to-port="4" />
11066 <edge from-layer="412" from-port="5" to-layer="417" to-port="0" />
11067 <edge from-layer="412" from-port="5" to-layer="433" to-port="0" />
11068 <edge from-layer="412" from-port="5" to-layer="449" to-port="0" />
11069 <edge from-layer="413" from-port="0" to-layer="414" to-port="0" />
11070 <edge from-layer="414" from-port="1" to-layer="416" to-port="0" />
11071 <edge from-layer="415" from-port="0" to-layer="416" to-port="1" />
11072 <edge from-layer="416" from-port="2" to-layer="417" to-port="1" />
11073 <edge from-layer="417" from-port="2" to-layer="419" to-port="0" />
11074 <edge from-layer="418" from-port="0" to-layer="419" to-port="1" />
11075 <edge from-layer="419" from-port="2" to-layer="424" to-port="0" />
11076 <edge from-layer="420" from-port="0" to-layer="424" to-port="1" />
11077 <edge from-layer="421" from-port="0" to-layer="424" to-port="2" />
11078 <edge from-layer="422" from-port="0" to-layer="424" to-port="3" />
11079 <edge from-layer="423" from-port="0" to-layer="424" to-port="4" />
11080 <edge from-layer="424" from-port="5" to-layer="426" to-port="0" />
11081 <edge from-layer="425" from-port="0" to-layer="426" to-port="1" />
11082 <edge from-layer="426" from-port="2" to-layer="428" to-port="0" />
11083 <edge from-layer="427" from-port="0" to-layer="428" to-port="1" />
11084 <edge from-layer="428" from-port="2" to-layer="456" to-port="0" />
11085 <edge from-layer="429" from-port="0" to-layer="430" to-port="0" />
11086 <edge from-layer="430" from-port="1" to-layer="432" to-port="0" />
11087 <edge from-layer="431" from-port="0" to-layer="432" to-port="1" />
11088 <edge from-layer="432" from-port="2" to-layer="433" to-port="1" />
11089 <edge from-layer="433" from-port="2" to-layer="435" to-port="0" />
11090 <edge from-layer="434" from-port="0" to-layer="435" to-port="1" />
11091 <edge from-layer="435" from-port="2" to-layer="440" to-port="0" />
11092 <edge from-layer="436" from-port="0" to-layer="440" to-port="1" />
11093 <edge from-layer="437" from-port="0" to-layer="440" to-port="2" />
11094 <edge from-layer="438" from-port="0" to-layer="440" to-port="3" />
11095 <edge from-layer="439" from-port="0" to-layer="440" to-port="4" />
11096 <edge from-layer="440" from-port="5" to-layer="442" to-port="0" />
11097 <edge from-layer="441" from-port="0" to-layer="442" to-port="1" />
11098 <edge from-layer="442" from-port="2" to-layer="444" to-port="0" />
11099 <edge from-layer="443" from-port="0" to-layer="444" to-port="1" />
11100 <edge from-layer="444" from-port="2" to-layer="456" to-port="1" />
11101 <edge from-layer="445" from-port="0" to-layer="446" to-port="0" />
11102 <edge from-layer="446" from-port="1" to-layer="448" to-port="0" />
11103 <edge from-layer="447" from-port="0" to-layer="448" to-port="1" />
11104 <edge from-layer="448" from-port="2" to-layer="449" to-port="1" />
11105 <edge from-layer="449" from-port="2" to-layer="451" to-port="0" />
11106 <edge from-layer="450" from-port="0" to-layer="451" to-port="1" />
11107 <edge from-layer="451" from-port="2" to-layer="453" to-port="0" />
11108 <edge from-layer="452" from-port="0" to-layer="453" to-port="1" />
11109 <edge from-layer="453" from-port="2" to-layer="455" to-port="0" />
11110 <edge from-layer="454" from-port="0" to-layer="455" to-port="1" />
11111 <edge from-layer="455" from-port="2" to-layer="456" to-port="2" />
11112 <edge from-layer="456" from-port="4" to-layer="458" to-port="0" />
11113 <edge from-layer="457" from-port="0" to-layer="458" to-port="1" />
11114 <edge from-layer="458" from-port="2" to-layer="460" to-port="0" />
11115 <edge from-layer="459" from-port="0" to-layer="460" to-port="1" />
11116 <edge from-layer="460" from-port="2" to-layer="462" to-port="0" />
11117 <edge from-layer="461" from-port="0" to-layer="462" to-port="1" />
11118 <edge from-layer="462" from-port="2" to-layer="467" to-port="0" />
11119 <edge from-layer="463" from-port="0" to-layer="467" to-port="1" />
11120 <edge from-layer="464" from-port="0" to-layer="467" to-port="2" />
11121 <edge from-layer="465" from-port="0" to-layer="467" to-port="3" />
11122 <edge from-layer="466" from-port="0" to-layer="467" to-port="4" />
11123 <edge from-layer="467" from-port="5" to-layer="472" to-port="0" />
11124 <edge from-layer="468" from-port="0" to-layer="469" to-port="0" />
11125 <edge from-layer="469" from-port="1" to-layer="471" to-port="0" />
11126 <edge from-layer="470" from-port="0" to-layer="471" to-port="1" />
11127 <edge from-layer="471" from-port="2" to-layer="472" to-port="1" />
11128 <edge from-layer="472" from-port="2" to-layer="474" to-port="0" />
11129 <edge from-layer="473" from-port="0" to-layer="474" to-port="1" />
11130 <edge from-layer="474" from-port="2" to-layer="475" to-port="0" />
11131 <edge from-layer="475" from-port="2" to-layer="477" to-port="0" />
11132 <edge from-layer="476" from-port="0" to-layer="477" to-port="1" />
11133 <edge from-layer="477" from-port="2" to-layer="479" to-port="0" />
11134 <edge from-layer="478" from-port="0" to-layer="479" to-port="1" />
11135 <edge from-layer="479" from-port="2" to-layer="481" to-port="0" />
11136 <edge from-layer="480" from-port="0" to-layer="481" to-port="1" />
11137 <edge from-layer="481" from-port="2" to-layer="483" to-port="0" />
11138 <edge from-layer="481" from-port="2" to-layer="511" to-port="1" />
11139 <edge from-layer="482" from-port="0" to-layer="483" to-port="1" />
11140 <edge from-layer="483" from-port="2" to-layer="488" to-port="0" />
11141 <edge from-layer="484" from-port="0" to-layer="488" to-port="1" />
11142 <edge from-layer="485" from-port="0" to-layer="488" to-port="2" />
11143 <edge from-layer="486" from-port="0" to-layer="488" to-port="3" />
11144 <edge from-layer="487" from-port="0" to-layer="488" to-port="4" />
11145 <edge from-layer="488" from-port="5" to-layer="493" to-port="0" />
11146 <edge from-layer="489" from-port="0" to-layer="490" to-port="0" />
11147 <edge from-layer="490" from-port="1" to-layer="492" to-port="0" />
11148 <edge from-layer="491" from-port="0" to-layer="492" to-port="1" />
11149 <edge from-layer="492" from-port="2" to-layer="493" to-port="1" />
11150 <edge from-layer="493" from-port="2" to-layer="495" to-port="0" />
11151 <edge from-layer="494" from-port="0" to-layer="495" to-port="1" />
11152 <edge from-layer="495" from-port="2" to-layer="496" to-port="0" />
11153 <edge from-layer="496" from-port="1" to-layer="498" to-port="0" />
11154 <edge from-layer="497" from-port="0" to-layer="498" to-port="1" />
11155 <edge from-layer="498" from-port="2" to-layer="503" to-port="0" />
11156 <edge from-layer="499" from-port="0" to-layer="503" to-port="1" />
11157 <edge from-layer="500" from-port="0" to-layer="503" to-port="2" />
11158 <edge from-layer="501" from-port="0" to-layer="503" to-port="3" />
11159 <edge from-layer="502" from-port="0" to-layer="503" to-port="4" />
11160 <edge from-layer="503" from-port="5" to-layer="508" to-port="0" />
11161 <edge from-layer="504" from-port="0" to-layer="505" to-port="0" />
11162 <edge from-layer="505" from-port="1" to-layer="507" to-port="0" />
11163 <edge from-layer="506" from-port="0" to-layer="507" to-port="1" />
11164 <edge from-layer="507" from-port="2" to-layer="508" to-port="1" />
11165 <edge from-layer="508" from-port="2" to-layer="510" to-port="0" />
11166 <edge from-layer="509" from-port="0" to-layer="510" to-port="1" />
11167 <edge from-layer="510" from-port="2" to-layer="511" to-port="0" />
11168 <edge from-layer="511" from-port="2" to-layer="513" to-port="0" />
11169 <edge from-layer="512" from-port="0" to-layer="513" to-port="1" />
11170 <edge from-layer="513" from-port="2" to-layer="515" to-port="0" />
11171 <edge from-layer="514" from-port="0" to-layer="515" to-port="1" />
11172 <edge from-layer="515" from-port="2" to-layer="517" to-port="0" />
11173 <edge from-layer="516" from-port="0" to-layer="517" to-port="1" />
11174 <edge from-layer="517" from-port="2" to-layer="587" to-port="1" />
11175 <edge from-layer="517" from-port="2" to-layer="519" to-port="0" />
11176 <edge from-layer="518" from-port="0" to-layer="519" to-port="1" />
11177 <edge from-layer="519" from-port="2" to-layer="524" to-port="0" />
11178 <edge from-layer="520" from-port="0" to-layer="524" to-port="1" />
11179 <edge from-layer="521" from-port="0" to-layer="524" to-port="2" />
11180 <edge from-layer="522" from-port="0" to-layer="524" to-port="3" />
11181 <edge from-layer="523" from-port="0" to-layer="524" to-port="4" />
11182 <edge from-layer="524" from-port="5" to-layer="529" to-port="0" />
11183 <edge from-layer="524" from-port="5" to-layer="545" to-port="0" />
11184 <edge from-layer="524" from-port="5" to-layer="561" to-port="0" />
11185 <edge from-layer="525" from-port="0" to-layer="526" to-port="0" />
11186 <edge from-layer="526" from-port="1" to-layer="528" to-port="0" />
11187 <edge from-layer="527" from-port="0" to-layer="528" to-port="1" />
11188 <edge from-layer="528" from-port="2" to-layer="529" to-port="1" />
11189 <edge from-layer="529" from-port="2" to-layer="531" to-port="0" />
11190 <edge from-layer="530" from-port="0" to-layer="531" to-port="1" />
11191 <edge from-layer="531" from-port="2" to-layer="536" to-port="0" />
11192 <edge from-layer="532" from-port="0" to-layer="536" to-port="1" />
11193 <edge from-layer="533" from-port="0" to-layer="536" to-port="2" />
11194 <edge from-layer="534" from-port="0" to-layer="536" to-port="3" />
11195 <edge from-layer="535" from-port="0" to-layer="536" to-port="4" />
11196 <edge from-layer="536" from-port="5" to-layer="538" to-port="0" />
11197 <edge from-layer="537" from-port="0" to-layer="538" to-port="1" />
11198 <edge from-layer="538" from-port="2" to-layer="540" to-port="0" />
11199 <edge from-layer="539" from-port="0" to-layer="540" to-port="1" />
11200 <edge from-layer="540" from-port="2" to-layer="568" to-port="0" />
11201 <edge from-layer="541" from-port="0" to-layer="542" to-port="0" />
11202 <edge from-layer="542" from-port="1" to-layer="544" to-port="0" />
11203 <edge from-layer="543" from-port="0" to-layer="544" to-port="1" />
11204 <edge from-layer="544" from-port="2" to-layer="545" to-port="1" />
11205 <edge from-layer="545" from-port="2" to-layer="547" to-port="0" />
11206 <edge from-layer="546" from-port="0" to-layer="547" to-port="1" />
11207 <edge from-layer="547" from-port="2" to-layer="552" to-port="0" />
11208 <edge from-layer="548" from-port="0" to-layer="552" to-port="1" />
11209 <edge from-layer="549" from-port="0" to-layer="552" to-port="2" />
11210 <edge from-layer="550" from-port="0" to-layer="552" to-port="3" />
11211 <edge from-layer="551" from-port="0" to-layer="552" to-port="4" />
11212 <edge from-layer="552" from-port="5" to-layer="554" to-port="0" />
11213 <edge from-layer="553" from-port="0" to-layer="554" to-port="1" />
11214 <edge from-layer="554" from-port="2" to-layer="556" to-port="0" />
11215 <edge from-layer="555" from-port="0" to-layer="556" to-port="1" />
11216 <edge from-layer="556" from-port="2" to-layer="568" to-port="1" />
11217 <edge from-layer="557" from-port="0" to-layer="558" to-port="0" />
11218 <edge from-layer="558" from-port="1" to-layer="560" to-port="0" />
11219 <edge from-layer="559" from-port="0" to-layer="560" to-port="1" />
11220 <edge from-layer="560" from-port="2" to-layer="561" to-port="1" />
11221 <edge from-layer="561" from-port="2" to-layer="563" to-port="0" />
11222 <edge from-layer="562" from-port="0" to-layer="563" to-port="1" />
11223 <edge from-layer="563" from-port="2" to-layer="565" to-port="0" />
11224 <edge from-layer="564" from-port="0" to-layer="565" to-port="1" />
11225 <edge from-layer="565" from-port="2" to-layer="567" to-port="0" />
11226 <edge from-layer="566" from-port="0" to-layer="567" to-port="1" />
11227 <edge from-layer="567" from-port="2" to-layer="568" to-port="2" />
11228 <edge from-layer="568" from-port="4" to-layer="570" to-port="0" />
11229 <edge from-layer="569" from-port="0" to-layer="570" to-port="1" />
11230 <edge from-layer="570" from-port="2" to-layer="572" to-port="0" />
11231 <edge from-layer="571" from-port="0" to-layer="572" to-port="1" />
11232 <edge from-layer="572" from-port="2" to-layer="574" to-port="0" />
11233 <edge from-layer="573" from-port="0" to-layer="574" to-port="1" />
11234 <edge from-layer="574" from-port="2" to-layer="579" to-port="0" />
11235 <edge from-layer="575" from-port="0" to-layer="579" to-port="1" />
11236 <edge from-layer="576" from-port="0" to-layer="579" to-port="2" />
11237 <edge from-layer="577" from-port="0" to-layer="579" to-port="3" />
11238 <edge from-layer="578" from-port="0" to-layer="579" to-port="4" />
11239 <edge from-layer="579" from-port="5" to-layer="584" to-port="0" />
11240 <edge from-layer="580" from-port="0" to-layer="581" to-port="0" />
11241 <edge from-layer="581" from-port="1" to-layer="583" to-port="0" />
11242 <edge from-layer="582" from-port="0" to-layer="583" to-port="1" />
11243 <edge from-layer="583" from-port="2" to-layer="584" to-port="1" />
11244 <edge from-layer="584" from-port="2" to-layer="586" to-port="0" />
11245 <edge from-layer="585" from-port="0" to-layer="586" to-port="1" />
11246 <edge from-layer="586" from-port="2" to-layer="587" to-port="0" />
11247 <edge from-layer="587" from-port="2" to-layer="589" to-port="0" />
11248 <edge from-layer="588" from-port="0" to-layer="589" to-port="1" />
11249 <edge from-layer="589" from-port="2" to-layer="591" to-port="0" />
11250 <edge from-layer="590" from-port="0" to-layer="591" to-port="1" />
11251 <edge from-layer="591" from-port="2" to-layer="593" to-port="0" />
11252 <edge from-layer="592" from-port="0" to-layer="593" to-port="1" />
11253 <edge from-layer="593" from-port="2" to-layer="595" to-port="0" />
11254 <edge from-layer="593" from-port="2" to-layer="623" to-port="1" />
11255 <edge from-layer="594" from-port="0" to-layer="595" to-port="1" />
11256 <edge from-layer="595" from-port="2" to-layer="600" to-port="0" />
11257 <edge from-layer="596" from-port="0" to-layer="600" to-port="1" />
11258 <edge from-layer="597" from-port="0" to-layer="600" to-port="2" />
11259 <edge from-layer="598" from-port="0" to-layer="600" to-port="3" />
11260 <edge from-layer="599" from-port="0" to-layer="600" to-port="4" />
11261 <edge from-layer="600" from-port="5" to-layer="605" to-port="0" />
11262 <edge from-layer="601" from-port="0" to-layer="602" to-port="0" />
11263 <edge from-layer="602" from-port="1" to-layer="604" to-port="0" />
11264 <edge from-layer="603" from-port="0" to-layer="604" to-port="1" />
11265 <edge from-layer="604" from-port="2" to-layer="605" to-port="1" />
11266 <edge from-layer="605" from-port="2" to-layer="607" to-port="0" />
11267 <edge from-layer="606" from-port="0" to-layer="607" to-port="1" />
11268 <edge from-layer="607" from-port="2" to-layer="608" to-port="0" />
11269 <edge from-layer="608" from-port="1" to-layer="610" to-port="0" />
11270 <edge from-layer="609" from-port="0" to-layer="610" to-port="1" />
11271 <edge from-layer="610" from-port="2" to-layer="615" to-port="0" />
11272 <edge from-layer="611" from-port="0" to-layer="615" to-port="1" />
11273 <edge from-layer="612" from-port="0" to-layer="615" to-port="2" />
11274 <edge from-layer="613" from-port="0" to-layer="615" to-port="3" />
11275 <edge from-layer="614" from-port="0" to-layer="615" to-port="4" />
11276 <edge from-layer="615" from-port="5" to-layer="620" to-port="0" />
11277 <edge from-layer="616" from-port="0" to-layer="617" to-port="0" />
11278 <edge from-layer="617" from-port="1" to-layer="619" to-port="0" />
11279 <edge from-layer="618" from-port="0" to-layer="619" to-port="1" />
11280 <edge from-layer="619" from-port="2" to-layer="620" to-port="1" />
11281 <edge from-layer="620" from-port="2" to-layer="622" to-port="0" />
11282 <edge from-layer="621" from-port="0" to-layer="622" to-port="1" />
11283 <edge from-layer="622" from-port="2" to-layer="623" to-port="0" />
11284 <edge from-layer="623" from-port="2" to-layer="625" to-port="0" />
11285 <edge from-layer="624" from-port="0" to-layer="625" to-port="1" />
11286 <edge from-layer="625" from-port="2" to-layer="627" to-port="0" />
11287 <edge from-layer="626" from-port="0" to-layer="627" to-port="1" />
11288 <edge from-layer="627" from-port="2" to-layer="629" to-port="0" />
11289 <edge from-layer="628" from-port="0" to-layer="629" to-port="1" />
11290 <edge from-layer="629" from-port="2" to-layer="631" to-port="0" />
11291 <edge from-layer="629" from-port="2" to-layer="699" to-port="1" />
11292 <edge from-layer="630" from-port="0" to-layer="631" to-port="1" />
11293 <edge from-layer="631" from-port="2" to-layer="636" to-port="0" />
11294 <edge from-layer="632" from-port="0" to-layer="636" to-port="1" />
11295 <edge from-layer="633" from-port="0" to-layer="636" to-port="2" />
11296 <edge from-layer="634" from-port="0" to-layer="636" to-port="3" />
11297 <edge from-layer="635" from-port="0" to-layer="636" to-port="4" />
11298 <edge from-layer="636" from-port="5" to-layer="641" to-port="0" />
11299 <edge from-layer="636" from-port="5" to-layer="657" to-port="0" />
11300 <edge from-layer="636" from-port="5" to-layer="673" to-port="0" />
11301 <edge from-layer="637" from-port="0" to-layer="638" to-port="0" />
11302 <edge from-layer="638" from-port="1" to-layer="640" to-port="0" />
11303 <edge from-layer="639" from-port="0" to-layer="640" to-port="1" />
11304 <edge from-layer="640" from-port="2" to-layer="641" to-port="1" />
11305 <edge from-layer="641" from-port="2" to-layer="643" to-port="0" />
11306 <edge from-layer="642" from-port="0" to-layer="643" to-port="1" />
11307 <edge from-layer="643" from-port="2" to-layer="648" to-port="0" />
11308 <edge from-layer="644" from-port="0" to-layer="648" to-port="1" />
11309 <edge from-layer="645" from-port="0" to-layer="648" to-port="2" />
11310 <edge from-layer="646" from-port="0" to-layer="648" to-port="3" />
11311 <edge from-layer="647" from-port="0" to-layer="648" to-port="4" />
11312 <edge from-layer="648" from-port="5" to-layer="650" to-port="0" />
11313 <edge from-layer="649" from-port="0" to-layer="650" to-port="1" />
11314 <edge from-layer="650" from-port="2" to-layer="652" to-port="0" />
11315 <edge from-layer="651" from-port="0" to-layer="652" to-port="1" />
11316 <edge from-layer="652" from-port="2" to-layer="680" to-port="0" />
11317 <edge from-layer="653" from-port="0" to-layer="654" to-port="0" />
11318 <edge from-layer="654" from-port="1" to-layer="656" to-port="0" />
11319 <edge from-layer="655" from-port="0" to-layer="656" to-port="1" />
11320 <edge from-layer="656" from-port="2" to-layer="657" to-port="1" />
11321 <edge from-layer="657" from-port="2" to-layer="659" to-port="0" />
11322 <edge from-layer="658" from-port="0" to-layer="659" to-port="1" />
11323 <edge from-layer="659" from-port="2" to-layer="664" to-port="0" />
11324 <edge from-layer="660" from-port="0" to-layer="664" to-port="1" />
11325 <edge from-layer="661" from-port="0" to-layer="664" to-port="2" />
11326 <edge from-layer="662" from-port="0" to-layer="664" to-port="3" />
11327 <edge from-layer="663" from-port="0" to-layer="664" to-port="4" />
11328 <edge from-layer="664" from-port="5" to-layer="666" to-port="0" />
11329 <edge from-layer="665" from-port="0" to-layer="666" to-port="1" />
11330 <edge from-layer="666" from-port="2" to-layer="668" to-port="0" />
11331 <edge from-layer="667" from-port="0" to-layer="668" to-port="1" />
11332 <edge from-layer="668" from-port="2" to-layer="680" to-port="1" />
11333 <edge from-layer="669" from-port="0" to-layer="670" to-port="0" />
11334 <edge from-layer="670" from-port="1" to-layer="672" to-port="0" />
11335 <edge from-layer="671" from-port="0" to-layer="672" to-port="1" />
11336 <edge from-layer="672" from-port="2" to-layer="673" to-port="1" />
11337 <edge from-layer="673" from-port="2" to-layer="675" to-port="0" />
11338 <edge from-layer="674" from-port="0" to-layer="675" to-port="1" />
11339 <edge from-layer="675" from-port="2" to-layer="677" to-port="0" />
11340 <edge from-layer="676" from-port="0" to-layer="677" to-port="1" />
11341 <edge from-layer="677" from-port="2" to-layer="679" to-port="0" />
11342 <edge from-layer="678" from-port="0" to-layer="679" to-port="1" />
11343 <edge from-layer="679" from-port="2" to-layer="680" to-port="2" />
11344 <edge from-layer="680" from-port="4" to-layer="682" to-port="0" />
11345 <edge from-layer="681" from-port="0" to-layer="682" to-port="1" />
11346 <edge from-layer="682" from-port="2" to-layer="684" to-port="0" />
11347 <edge from-layer="683" from-port="0" to-layer="684" to-port="1" />
11348 <edge from-layer="684" from-port="2" to-layer="686" to-port="0" />
11349 <edge from-layer="685" from-port="0" to-layer="686" to-port="1" />
11350 <edge from-layer="686" from-port="2" to-layer="691" to-port="0" />
11351 <edge from-layer="687" from-port="0" to-layer="691" to-port="1" />
11352 <edge from-layer="688" from-port="0" to-layer="691" to-port="2" />
11353 <edge from-layer="689" from-port="0" to-layer="691" to-port="3" />
11354 <edge from-layer="690" from-port="0" to-layer="691" to-port="4" />
11355 <edge from-layer="691" from-port="5" to-layer="696" to-port="0" />
11356 <edge from-layer="692" from-port="0" to-layer="693" to-port="0" />
11357 <edge from-layer="693" from-port="1" to-layer="695" to-port="0" />
11358 <edge from-layer="694" from-port="0" to-layer="695" to-port="1" />
11359 <edge from-layer="695" from-port="2" to-layer="696" to-port="1" />
11360 <edge from-layer="696" from-port="2" to-layer="698" to-port="0" />
11361 <edge from-layer="697" from-port="0" to-layer="698" to-port="1" />
11362 <edge from-layer="698" from-port="2" to-layer="699" to-port="0" />
11363 <edge from-layer="699" from-port="2" to-layer="701" to-port="0" />
11364 <edge from-layer="700" from-port="0" to-layer="701" to-port="1" />
11365 <edge from-layer="701" from-port="2" to-layer="703" to-port="0" />
11366 <edge from-layer="702" from-port="0" to-layer="703" to-port="1" />
11367 <edge from-layer="703" from-port="2" to-layer="705" to-port="0" />
11368 <edge from-layer="704" from-port="0" to-layer="705" to-port="1" />
11369 <edge from-layer="705" from-port="2" to-layer="707" to-port="0" />
11370 <edge from-layer="705" from-port="2" to-layer="735" to-port="1" />
11371 <edge from-layer="706" from-port="0" to-layer="707" to-port="1" />
11372 <edge from-layer="707" from-port="2" to-layer="712" to-port="0" />
11373 <edge from-layer="708" from-port="0" to-layer="712" to-port="1" />
11374 <edge from-layer="709" from-port="0" to-layer="712" to-port="2" />
11375 <edge from-layer="710" from-port="0" to-layer="712" to-port="3" />
11376 <edge from-layer="711" from-port="0" to-layer="712" to-port="4" />
11377 <edge from-layer="712" from-port="5" to-layer="717" to-port="0" />
11378 <edge from-layer="713" from-port="0" to-layer="714" to-port="0" />
11379 <edge from-layer="714" from-port="1" to-layer="716" to-port="0" />
11380 <edge from-layer="715" from-port="0" to-layer="716" to-port="1" />
11381 <edge from-layer="716" from-port="2" to-layer="717" to-port="1" />
11382 <edge from-layer="717" from-port="2" to-layer="719" to-port="0" />
11383 <edge from-layer="718" from-port="0" to-layer="719" to-port="1" />
11384 <edge from-layer="719" from-port="2" to-layer="720" to-port="0" />
11385 <edge from-layer="720" from-port="1" to-layer="722" to-port="0" />
11386 <edge from-layer="721" from-port="0" to-layer="722" to-port="1" />
11387 <edge from-layer="722" from-port="2" to-layer="727" to-port="0" />
11388 <edge from-layer="723" from-port="0" to-layer="727" to-port="1" />
11389 <edge from-layer="724" from-port="0" to-layer="727" to-port="2" />
11390 <edge from-layer="725" from-port="0" to-layer="727" to-port="3" />
11391 <edge from-layer="726" from-port="0" to-layer="727" to-port="4" />
11392 <edge from-layer="727" from-port="5" to-layer="732" to-port="0" />
11393 <edge from-layer="728" from-port="0" to-layer="729" to-port="0" />
11394 <edge from-layer="729" from-port="1" to-layer="731" to-port="0" />
11395 <edge from-layer="730" from-port="0" to-layer="731" to-port="1" />
11396 <edge from-layer="731" from-port="2" to-layer="732" to-port="1" />
11397 <edge from-layer="732" from-port="2" to-layer="734" to-port="0" />
11398 <edge from-layer="733" from-port="0" to-layer="734" to-port="1" />
11399 <edge from-layer="734" from-port="2" to-layer="735" to-port="0" />
11400 <edge from-layer="735" from-port="2" to-layer="737" to-port="0" />
11401 <edge from-layer="736" from-port="0" to-layer="737" to-port="1" />
11402 <edge from-layer="737" from-port="2" to-layer="739" to-port="0" />
11403 <edge from-layer="738" from-port="0" to-layer="739" to-port="1" />
11404 <edge from-layer="739" from-port="2" to-layer="741" to-port="0" />
11405 <edge from-layer="740" from-port="0" to-layer="741" to-port="1" />
11406 <edge from-layer="741" from-port="2" to-layer="744" to-port="0" />
11407 <edge from-layer="742" from-port="0" to-layer="744" to-port="1" />
11408 <edge from-layer="743" from-port="0" to-layer="744" to-port="2" />
11409 <edge from-layer="744" from-port="3" to-layer="746" to-port="0" />
11410 <edge from-layer="745" from-port="0" to-layer="746" to-port="1" />
11411 <edge from-layer="746" from-port="2" to-layer="751" to-port="0" />
11412 <edge from-layer="747" from-port="0" to-layer="751" to-port="1" />
11413 <edge from-layer="748" from-port="0" to-layer="751" to-port="2" />
11414 <edge from-layer="749" from-port="0" to-layer="751" to-port="3" />
11415 <edge from-layer="750" from-port="0" to-layer="751" to-port="4" />
11416 <edge from-layer="751" from-port="5" to-layer="756" to-port="0" />
11417 <edge from-layer="752" from-port="0" to-layer="753" to-port="0" />
11418 <edge from-layer="753" from-port="1" to-layer="755" to-port="0" />
11419 <edge from-layer="754" from-port="0" to-layer="755" to-port="1" />
11420 <edge from-layer="755" from-port="2" to-layer="756" to-port="1" />
11421 <edge from-layer="756" from-port="2" to-layer="758" to-port="0" />
11422 <edge from-layer="757" from-port="0" to-layer="758" to-port="1" />
11423 <edge from-layer="758" from-port="2" to-layer="759" to-port="0" />
11424 <edge from-layer="759" from-port="1" to-layer="761" to-port="0" />
11425 <edge from-layer="760" from-port="0" to-layer="761" to-port="1" />
11426 <edge from-layer="761" from-port="2" to-layer="766" to-port="0" />
11427 <edge from-layer="762" from-port="0" to-layer="766" to-port="1" />
11428 <edge from-layer="763" from-port="0" to-layer="766" to-port="2" />
11429 <edge from-layer="764" from-port="0" to-layer="766" to-port="3" />
11430 <edge from-layer="765" from-port="0" to-layer="766" to-port="4" />
11431 <edge from-layer="766" from-port="5" to-layer="771" to-port="0" />
11432 <edge from-layer="767" from-port="0" to-layer="768" to-port="0" />
11433 <edge from-layer="768" from-port="1" to-layer="770" to-port="0" />
11434 <edge from-layer="769" from-port="0" to-layer="770" to-port="1" />
11435 <edge from-layer="770" from-port="2" to-layer="771" to-port="1" />
11436 <edge from-layer="771" from-port="2" to-layer="773" to-port="0" />
11437 <edge from-layer="772" from-port="0" to-layer="773" to-port="1" />
11438 <edge from-layer="773" from-port="2" to-layer="774" to-port="0" />
11439 </edges>
11440 <rt_info>
11441 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
11442 <conversion_parameters>
11443 <framework value="pytorch" />
11444 <is_python_object value="True" />
11445 </conversion_parameters>
11446 <nncf>
11447 <friendly_names_were_updated value="True" />
11448 <quantization>
11449 <advanced_parameters value="{'overflow_fix': 'disable', 'quantize_outputs': False, 'inplace_statistics': True, 'disable_channel_alignment': True, 'disable_bias_correction': False, 'batchwise_statistics': None, 'activations_quantization_params': None, 'weights_quantization_params': None, 'activations_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'weights_range_estimator_params': {'min': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}, 'max': {'statistics_type': None, 'aggregator_type': None, 'clipping_value': None, 'quantile_outlier_prob': 0.0001}}, 'bias_correction_params': {'apply_for_all_nodes': False, 'threshold': None}, 'smooth_quant_alphas': {'convolution': -1, 'matmul': 0.95}, 'smooth_quant_alpha': None, 'backend_params': {}}" />
11450 <fast_bias_correction value="True" />
11451 <ignored_scope>
11452 <types value="['GroupNormalization']" />
11453 </ignored_scope>
11454 <model_type value="transformer" />
11455 <preset value="mixed" />
11456 <subset_size value="300" />
11457 <target_device value="ANY" />
11458 </quantization>
11459 </nncf>
11460 <optimum>
11461 <optimum_intel_version value="1.20.1" />
11462 <optimum_version value="1.24.0" />
11463 <pytorch_version value="2.6.0+cu124" />
11464 <transformers_version value="4.52.0.dev0" />
11465 </optimum>
11466 </rt_info>
11467 </net>
11468