openvino/openvino_model.xml
206.8 KB · 6796 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model0" version="11">
3 <layers>
4 <layer id="1" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="0" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="2" name="self.roberta.embeddings.word_embeddings.weight" type="Const" version="opset1">
23 <data element_type="f32" shape="50265, 768" offset="0" size="154414080" />
24 <output>
25 <port id="0" precision="FP32" names="self.roberta.embeddings.word_embeddings.weight">
26 <dim>50265</dim>
27 <dim>768</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
32 <data destination_type="i32" />
33 <input>
34 <port id="0" precision="I64">
35 <dim>-1</dim>
36 <dim>-1</dim>
37 </port>
38 </input>
39 <output>
40 <port id="1" precision="I32">
41 <dim>-1</dim>
42 <dim>-1</dim>
43 </port>
44 </output>
45 </layer>
46 <layer id="4" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
47 <data element_type="i32" shape="" offset="154414080" size="4" />
48 <output>
49 <port id="0" precision="I32" />
50 </output>
51 </layer>
52 <layer id="5" name="__module.roberta.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
53 <data batch_dims="0" />
54 <input>
55 <port id="0" precision="FP32">
56 <dim>50265</dim>
57 <dim>768</dim>
58 </port>
59 <port id="1" precision="I32">
60 <dim>-1</dim>
61 <dim>-1</dim>
62 </port>
63 <port id="2" precision="I32" />
64 </input>
65 <output>
66 <port id="3" precision="FP32" names="54,inputs_embeds">
67 <dim>-1</dim>
68 <dim>-1</dim>
69 <dim>768</dim>
70 </port>
71 </output>
72 </layer>
73 <layer id="6" name="self.roberta.embeddings.token_type_embeddings.weight" type="Const" version="opset1">
74 <data element_type="f32" shape="1, 768" offset="154414084" size="3072" />
75 <output>
76 <port id="0" precision="FP32" names="self.roberta.embeddings.token_type_embeddings.weight">
77 <dim>1</dim>
78 <dim>768</dim>
79 </port>
80 </output>
81 </layer>
82 <layer id="7" name="__module.roberta/aten::slice/Slice" type="Const" version="opset1">
83 <data element_type="i64" shape="1, 514" offset="154417156" size="4112" />
84 <output>
85 <port id="0" precision="I64" names="37">
86 <dim>1</dim>
87 <dim>514</dim>
88 </port>
89 </output>
90 </layer>
91 <layer id="8" name="__module.roberta/aten::slice/Reshape" type="Const" version="opset1">
92 <data element_type="i64" shape="1" offset="154421268" size="8" />
93 <output>
94 <port id="0" precision="I64">
95 <dim>1</dim>
96 </port>
97 </output>
98 </layer>
99 <layer id="9" name="__module.roberta/aten::size/ShapeOf_1" type="ShapeOf" version="opset3">
100 <data output_type="i64" />
101 <input>
102 <port id="0" precision="I64">
103 <dim>-1</dim>
104 <dim>-1</dim>
105 </port>
106 </input>
107 <output>
108 <port id="1" precision="I64" names="39">
109 <dim>2</dim>
110 </port>
111 </output>
112 </layer>
113 <layer id="10" name="Constant_5485" type="Const" version="opset1">
114 <data element_type="i64" shape="1" offset="154421276" size="8" />
115 <output>
116 <port id="0" precision="I64">
117 <dim>1</dim>
118 </port>
119 </output>
120 </layer>
121 <layer id="11" name="Constant_5486" type="Const" version="opset1">
122 <data element_type="i64" shape="" offset="154421268" size="8" />
123 <output>
124 <port id="0" precision="I64" />
125 </output>
126 </layer>
127 <layer id="12" name="Gather_5487" type="Gather" version="opset8">
128 <data batch_dims="0" />
129 <input>
130 <port id="0" precision="I64">
131 <dim>2</dim>
132 </port>
133 <port id="1" precision="I64">
134 <dim>1</dim>
135 </port>
136 <port id="2" precision="I64" />
137 </input>
138 <output>
139 <port id="3" precision="I64" names="36,67">
140 <dim>1</dim>
141 </port>
142 </output>
143 </layer>
144 <layer id="13" name="__module.roberta/aten::slice/Reshape_2" type="Const" version="opset1">
145 <data element_type="i64" shape="1" offset="154421276" size="8" />
146 <output>
147 <port id="0" precision="I64">
148 <dim>1</dim>
149 </port>
150 </output>
151 </layer>
152 <layer id="14" name="__module.roberta/aten::slice/Reshape_3" type="Const" version="opset1">
153 <data element_type="i64" shape="1" offset="154421276" size="8" />
154 <output>
155 <port id="0" precision="I64">
156 <dim>1</dim>
157 </port>
158 </output>
159 </layer>
160 <layer id="15" name="__module.roberta/aten::slice/Slice_1" type="Slice" version="opset8">
161 <input>
162 <port id="0" precision="I64">
163 <dim>1</dim>
164 <dim>514</dim>
165 </port>
166 <port id="1" precision="I64">
167 <dim>1</dim>
168 </port>
169 <port id="2" precision="I64">
170 <dim>1</dim>
171 </port>
172 <port id="3" precision="I64">
173 <dim>1</dim>
174 </port>
175 <port id="4" precision="I64">
176 <dim>1</dim>
177 </port>
178 </input>
179 <output>
180 <port id="5" precision="I64" names="38,buffered_token_type_ids">
181 <dim>1</dim>
182 <dim>-1</dim>
183 </port>
184 </output>
185 </layer>
186 <layer id="16" name="__module.roberta/aten::expand/Broadcast" type="Broadcast" version="opset3">
187 <data mode="bidirectional" />
188 <input>
189 <port id="0" precision="I64">
190 <dim>1</dim>
191 <dim>-1</dim>
192 </port>
193 <port id="1" precision="I64">
194 <dim>2</dim>
195 </port>
196 </input>
197 <output>
198 <port id="2" precision="I64" names="40">
199 <dim>-1</dim>
200 <dim>-1</dim>
201 </port>
202 </output>
203 </layer>
204 <layer id="17" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
205 <data destination_type="i32" />
206 <input>
207 <port id="0" precision="I64">
208 <dim>-1</dim>
209 <dim>-1</dim>
210 </port>
211 </input>
212 <output>
213 <port id="1" precision="I32">
214 <dim>-1</dim>
215 <dim>-1</dim>
216 </port>
217 </output>
218 </layer>
219 <layer id="18" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
220 <data element_type="i32" shape="" offset="154414080" size="4" />
221 <output>
222 <port id="0" precision="I32" />
223 </output>
224 </layer>
225 <layer id="19" name="__module.roberta.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
226 <data batch_dims="0" />
227 <input>
228 <port id="0" precision="FP32">
229 <dim>1</dim>
230 <dim>768</dim>
231 </port>
232 <port id="1" precision="I32">
233 <dim>-1</dim>
234 <dim>-1</dim>
235 </port>
236 <port id="2" precision="I32" />
237 </input>
238 <output>
239 <port id="3" precision="FP32" names="56,token_type_embeddings.1">
240 <dim>-1</dim>
241 <dim>-1</dim>
242 <dim>768</dim>
243 </port>
244 </output>
245 </layer>
246 <layer id="20" name="__module.roberta.embeddings/aten::add/Add" type="Add" version="opset1">
247 <data auto_broadcast="numpy" />
248 <input>
249 <port id="0" precision="FP32">
250 <dim>-1</dim>
251 <dim>-1</dim>
252 <dim>768</dim>
253 </port>
254 <port id="1" precision="FP32">
255 <dim>-1</dim>
256 <dim>-1</dim>
257 <dim>768</dim>
258 </port>
259 </input>
260 <output>
261 <port id="2" precision="FP32" names="57_1">
262 <dim>-1</dim>
263 <dim>-1</dim>
264 <dim>768</dim>
265 </port>
266 </output>
267 </layer>
268 <layer id="21" name="self.roberta.embeddings.position_embeddings.weight" type="Const" version="opset1">
269 <data element_type="f32" shape="514, 768" offset="154421284" size="1579008" />
270 <output>
271 <port id="0" precision="FP32" names="self.roberta.embeddings.position_embeddings.weight">
272 <dim>514</dim>
273 <dim>768</dim>
274 </port>
275 </output>
276 </layer>
277 <layer id="22" name="Constant_5356" type="Const" version="opset1">
278 <data element_type="i64" shape="1, 1" offset="154421276" size="8" />
279 <output>
280 <port id="0" precision="I64">
281 <dim>1</dim>
282 <dim>1</dim>
283 </port>
284 </output>
285 </layer>
286 <layer id="23" name="__module.roberta.embeddings/aten::ne/NotEqual" type="NotEqual" version="opset1">
287 <data auto_broadcast="numpy" />
288 <input>
289 <port id="0" precision="I64">
290 <dim>-1</dim>
291 <dim>-1</dim>
292 </port>
293 <port id="1" precision="I64">
294 <dim>1</dim>
295 <dim>1</dim>
296 </port>
297 </input>
298 <output>
299 <port id="2" precision="BOOL" names="45">
300 <dim>-1</dim>
301 <dim>-1</dim>
302 </port>
303 </output>
304 </layer>
305 <layer id="24" name="__module.roberta.embeddings/aten::to/Convert" type="Convert" version="opset1">
306 <data destination_type="i32" />
307 <input>
308 <port id="0" precision="BOOL">
309 <dim>-1</dim>
310 <dim>-1</dim>
311 </port>
312 </input>
313 <output>
314 <port id="1" precision="I32" names="46,mask">
315 <dim>-1</dim>
316 <dim>-1</dim>
317 </port>
318 </output>
319 </layer>
320 <layer id="25" name="29" type="Const" version="opset1">
321 <data element_type="i64" shape="" offset="154421276" size="8" />
322 <output>
323 <port id="0" precision="I64" names="29" />
324 </output>
325 </layer>
326 <layer id="26" name="__module.roberta.embeddings/aten::cumsum/CumSum" type="CumSum" version="opset3">
327 <data exclusive="false" reverse="false" />
328 <input>
329 <port id="0" precision="I32">
330 <dim>-1</dim>
331 <dim>-1</dim>
332 </port>
333 <port id="1" precision="I64" />
334 </input>
335 <output>
336 <port id="2" precision="I32" names="47,48,49">
337 <dim>-1</dim>
338 <dim>-1</dim>
339 </port>
340 </output>
341 </layer>
342 <layer id="27" name="__module.roberta.embeddings/aten::mul/Multiply" type="Multiply" version="opset1">
343 <data auto_broadcast="numpy" />
344 <input>
345 <port id="0" precision="I32">
346 <dim>-1</dim>
347 <dim>-1</dim>
348 </port>
349 <port id="1" precision="I32">
350 <dim>-1</dim>
351 <dim>-1</dim>
352 </port>
353 </input>
354 <output>
355 <port id="2" precision="I32" names="50,incremental_indices">
356 <dim>-1</dim>
357 <dim>-1</dim>
358 </port>
359 </output>
360 </layer>
361 <layer id="28" name="__module.roberta.embeddings/aten::to/Convert_1" type="Convert" version="opset1">
362 <data destination_type="i64" />
363 <input>
364 <port id="0" precision="I32">
365 <dim>-1</dim>
366 <dim>-1</dim>
367 </port>
368 </input>
369 <output>
370 <port id="1" precision="I64" names="51">
371 <dim>-1</dim>
372 <dim>-1</dim>
373 </port>
374 </output>
375 </layer>
376 <layer id="29" name="Constant_5357" type="Const" version="opset1">
377 <data element_type="i64" shape="1, 1" offset="154421276" size="8" />
378 <output>
379 <port id="0" precision="I64">
380 <dim>1</dim>
381 <dim>1</dim>
382 </port>
383 </output>
384 </layer>
385 <layer id="30" name="__module.roberta.embeddings/aten::add/Add_2" type="Add" version="opset1">
386 <data auto_broadcast="numpy" />
387 <input>
388 <port id="0" precision="I64">
389 <dim>-1</dim>
390 <dim>-1</dim>
391 </port>
392 <port id="1" precision="I64">
393 <dim>1</dim>
394 <dim>1</dim>
395 </port>
396 </input>
397 <output>
398 <port id="2" precision="I64" names="52">
399 <dim>-1</dim>
400 <dim>-1</dim>
401 </port>
402 </output>
403 </layer>
404 <layer id="31" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
405 <data destination_type="i32" />
406 <input>
407 <port id="0" precision="I64">
408 <dim>-1</dim>
409 <dim>-1</dim>
410 </port>
411 </input>
412 <output>
413 <port id="1" precision="I32">
414 <dim>-1</dim>
415 <dim>-1</dim>
416 </port>
417 </output>
418 </layer>
419 <layer id="32" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
420 <data element_type="i32" shape="" offset="154414080" size="4" />
421 <output>
422 <port id="0" precision="I32" />
423 </output>
424 </layer>
425 <layer id="33" name="__module.roberta.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
426 <data batch_dims="0" />
427 <input>
428 <port id="0" precision="FP32">
429 <dim>514</dim>
430 <dim>768</dim>
431 </port>
432 <port id="1" precision="I32">
433 <dim>-1</dim>
434 <dim>-1</dim>
435 </port>
436 <port id="2" precision="I32" />
437 </input>
438 <output>
439 <port id="3" precision="FP32" names="59,position_embeddings.1">
440 <dim>-1</dim>
441 <dim>-1</dim>
442 <dim>768</dim>
443 </port>
444 </output>
445 </layer>
446 <layer id="34" name="__module.roberta.embeddings/aten::add_/Add" type="Add" version="opset1">
447 <data auto_broadcast="numpy" />
448 <input>
449 <port id="0" precision="FP32">
450 <dim>-1</dim>
451 <dim>-1</dim>
452 <dim>768</dim>
453 </port>
454 <port id="1" precision="FP32">
455 <dim>-1</dim>
456 <dim>-1</dim>
457 <dim>768</dim>
458 </port>
459 </input>
460 <output>
461 <port id="2" precision="FP32" names="57,embeddings.1">
462 <dim>-1</dim>
463 <dim>-1</dim>
464 <dim>768</dim>
465 </port>
466 </output>
467 </layer>
468 <layer id="35" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
469 <data element_type="i32" shape="1" offset="156000292" size="4" />
470 <output>
471 <port id="0" precision="I32">
472 <dim>1</dim>
473 </port>
474 </output>
475 </layer>
476 <layer id="36" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
477 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
478 <input>
479 <port id="0" precision="FP32">
480 <dim>-1</dim>
481 <dim>-1</dim>
482 <dim>768</dim>
483 </port>
484 <port id="1" precision="I32">
485 <dim>1</dim>
486 </port>
487 </input>
488 <output>
489 <port id="2" precision="FP32">
490 <dim>-1</dim>
491 <dim>-1</dim>
492 <dim>768</dim>
493 </port>
494 </output>
495 </layer>
496 <layer id="37" name="Constant_5358" type="Const" version="opset1">
497 <data element_type="f32" shape="1, 1, 768" offset="156000296" size="3072" />
498 <output>
499 <port id="0" precision="FP32">
500 <dim>1</dim>
501 <dim>1</dim>
502 <dim>768</dim>
503 </port>
504 </output>
505 </layer>
506 <layer id="38" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
507 <data auto_broadcast="numpy" />
508 <input>
509 <port id="0" precision="FP32">
510 <dim>-1</dim>
511 <dim>-1</dim>
512 <dim>768</dim>
513 </port>
514 <port id="1" precision="FP32">
515 <dim>1</dim>
516 <dim>1</dim>
517 <dim>768</dim>
518 </port>
519 </input>
520 <output>
521 <port id="2" precision="FP32">
522 <dim>-1</dim>
523 <dim>-1</dim>
524 <dim>768</dim>
525 </port>
526 </output>
527 </layer>
528 <layer id="39" name="Constant_5359" type="Const" version="opset1">
529 <data element_type="f32" shape="1, 1, 768" offset="156003368" size="3072" />
530 <output>
531 <port id="0" precision="FP32">
532 <dim>1</dim>
533 <dim>1</dim>
534 <dim>768</dim>
535 </port>
536 </output>
537 </layer>
538 <layer id="40" name="__module.roberta.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
539 <data auto_broadcast="numpy" />
540 <input>
541 <port id="0" precision="FP32">
542 <dim>-1</dim>
543 <dim>-1</dim>
544 <dim>768</dim>
545 </port>
546 <port id="1" precision="FP32">
547 <dim>1</dim>
548 <dim>1</dim>
549 <dim>768</dim>
550 </port>
551 </input>
552 <output>
553 <port id="2" precision="FP32" names="64,input.1">
554 <dim>-1</dim>
555 <dim>-1</dim>
556 <dim>768</dim>
557 </port>
558 </output>
559 </layer>
560 <layer id="41" name="self.roberta.encoder.layer.0.attention.self.query.weight" type="Const" version="opset1">
561 <data element_type="f32" shape="768, 768" offset="156006440" size="2359296" />
562 <output>
563 <port id="0" precision="FP32" names="self.roberta.encoder.layer.0.attention.self.query.weight">
564 <dim>768</dim>
565 <dim>768</dim>
566 </port>
567 </output>
568 </layer>
569 <layer id="42" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
570 <data transpose_a="false" transpose_b="true" />
571 <input>
572 <port id="0" precision="FP32">
573 <dim>-1</dim>
574 <dim>-1</dim>
575 <dim>768</dim>
576 </port>
577 <port id="1" precision="FP32">
578 <dim>768</dim>
579 <dim>768</dim>
580 </port>
581 </input>
582 <output>
583 <port id="2" precision="FP32">
584 <dim>-1</dim>
585 <dim>-1</dim>
586 <dim>768</dim>
587 </port>
588 </output>
589 </layer>
590 <layer id="43" name="Constant_5360" type="Const" version="opset1">
591 <data element_type="f32" shape="1, 1, 768" offset="158365736" size="3072" />
592 <output>
593 <port id="0" precision="FP32">
594 <dim>1</dim>
595 <dim>1</dim>
596 <dim>768</dim>
597 </port>
598 </output>
599 </layer>
600 <layer id="44" name="__module.roberta.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
601 <data auto_broadcast="numpy" />
602 <input>
603 <port id="0" precision="FP32">
604 <dim>-1</dim>
605 <dim>-1</dim>
606 <dim>768</dim>
607 </port>
608 <port id="1" precision="FP32">
609 <dim>1</dim>
610 <dim>1</dim>
611 <dim>768</dim>
612 </port>
613 </input>
614 <output>
615 <port id="2" precision="FP32" names="102,x.1">
616 <dim>-1</dim>
617 <dim>-1</dim>
618 <dim>768</dim>
619 </port>
620 </output>
621 </layer>
622 <layer id="45" name="__module.roberta.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
623 <data element_type="i64" shape="4" offset="158368808" size="32" />
624 <output>
625 <port id="0" precision="I64">
626 <dim>4</dim>
627 </port>
628 </output>
629 </layer>
630 <layer id="46" name="__module.roberta.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
631 <data special_zero="true" />
632 <input>
633 <port id="0" precision="FP32">
634 <dim>-1</dim>
635 <dim>-1</dim>
636 <dim>768</dim>
637 </port>
638 <port id="1" precision="I64">
639 <dim>4</dim>
640 </port>
641 </input>
642 <output>
643 <port id="2" precision="FP32" names="106,x.3">
644 <dim>-1</dim>
645 <dim>-1</dim>
646 <dim>12</dim>
647 <dim>64</dim>
648 </port>
649 </output>
650 </layer>
651 <layer id="47" name="Constant_229" type="Const" version="opset1">
652 <data element_type="i64" shape="4" offset="158368840" size="32" />
653 <output>
654 <port id="0" precision="I64" names="107">
655 <dim>4</dim>
656 </port>
657 </output>
658 </layer>
659 <layer id="48" name="__module.roberta.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
660 <input>
661 <port id="0" precision="FP32">
662 <dim>-1</dim>
663 <dim>-1</dim>
664 <dim>12</dim>
665 <dim>64</dim>
666 </port>
667 <port id="1" precision="I64">
668 <dim>4</dim>
669 </port>
670 </input>
671 <output>
672 <port id="2" precision="FP32" names="108">
673 <dim>-1</dim>
674 <dim>12</dim>
675 <dim>-1</dim>
676 <dim>64</dim>
677 </port>
678 </output>
679 </layer>
680 <layer id="49" name="self.roberta.encoder.layer.0.attention.self.key.weight" type="Const" version="opset1">
681 <data element_type="f32" shape="768, 768" offset="158368872" size="2359296" />
682 <output>
683 <port id="0" precision="FP32" names="self.roberta.encoder.layer.0.attention.self.key.weight">
684 <dim>768</dim>
685 <dim>768</dim>
686 </port>
687 </output>
688 </layer>
689 <layer id="50" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
690 <data transpose_a="false" transpose_b="true" />
691 <input>
692 <port id="0" precision="FP32">
693 <dim>-1</dim>
694 <dim>-1</dim>
695 <dim>768</dim>
696 </port>
697 <port id="1" precision="FP32">
698 <dim>768</dim>
699 <dim>768</dim>
700 </port>
701 </input>
702 <output>
703 <port id="2" precision="FP32">
704 <dim>-1</dim>
705 <dim>-1</dim>
706 <dim>768</dim>
707 </port>
708 </output>
709 </layer>
710 <layer id="51" name="Constant_5361" type="Const" version="opset1">
711 <data element_type="f32" shape="1, 1, 768" offset="160728168" size="3072" />
712 <output>
713 <port id="0" precision="FP32">
714 <dim>1</dim>
715 <dim>1</dim>
716 <dim>768</dim>
717 </port>
718 </output>
719 </layer>
720 <layer id="52" name="__module.roberta.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
721 <data auto_broadcast="numpy" />
722 <input>
723 <port id="0" precision="FP32">
724 <dim>-1</dim>
725 <dim>-1</dim>
726 <dim>768</dim>
727 </port>
728 <port id="1" precision="FP32">
729 <dim>1</dim>
730 <dim>1</dim>
731 <dim>768</dim>
732 </port>
733 </input>
734 <output>
735 <port id="2" precision="FP32" names="111,x.5">
736 <dim>-1</dim>
737 <dim>-1</dim>
738 <dim>768</dim>
739 </port>
740 </output>
741 </layer>
742 <layer id="53" name="__module.roberta.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
743 <data element_type="i64" shape="4" offset="158368808" size="32" />
744 <output>
745 <port id="0" precision="I64">
746 <dim>4</dim>
747 </port>
748 </output>
749 </layer>
750 <layer id="54" name="__module.roberta.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
751 <data special_zero="true" />
752 <input>
753 <port id="0" precision="FP32">
754 <dim>-1</dim>
755 <dim>-1</dim>
756 <dim>768</dim>
757 </port>
758 <port id="1" precision="I64">
759 <dim>4</dim>
760 </port>
761 </input>
762 <output>
763 <port id="2" precision="FP32" names="115,x.7">
764 <dim>-1</dim>
765 <dim>-1</dim>
766 <dim>12</dim>
767 <dim>64</dim>
768 </port>
769 </output>
770 </layer>
771 <layer id="55" name="Constant_252" type="Const" version="opset1">
772 <data element_type="i64" shape="4" offset="158368840" size="32" />
773 <output>
774 <port id="0" precision="I64" names="116">
775 <dim>4</dim>
776 </port>
777 </output>
778 </layer>
779 <layer id="56" name="__module.roberta.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
780 <input>
781 <port id="0" precision="FP32">
782 <dim>-1</dim>
783 <dim>-1</dim>
784 <dim>12</dim>
785 <dim>64</dim>
786 </port>
787 <port id="1" precision="I64">
788 <dim>4</dim>
789 </port>
790 </input>
791 <output>
792 <port id="2" precision="FP32" names="117">
793 <dim>-1</dim>
794 <dim>12</dim>
795 <dim>-1</dim>
796 <dim>64</dim>
797 </port>
798 </output>
799 </layer>
800 <layer id="57" name="self.roberta.encoder.layer.0.attention.self.value.weight" type="Const" version="opset1">
801 <data element_type="f32" shape="768, 768" offset="160731240" size="2359296" />
802 <output>
803 <port id="0" precision="FP32" names="self.roberta.encoder.layer.0.attention.self.value.weight">
804 <dim>768</dim>
805 <dim>768</dim>
806 </port>
807 </output>
808 </layer>
809 <layer id="58" name="__module.roberta.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
810 <data transpose_a="false" transpose_b="true" />
811 <input>
812 <port id="0" precision="FP32">
813 <dim>-1</dim>
814 <dim>-1</dim>
815 <dim>768</dim>
816 </port>
817 <port id="1" precision="FP32">
818 <dim>768</dim>
819 <dim>768</dim>
820 </port>
821 </input>
822 <output>
823 <port id="2" precision="FP32">
824 <dim>-1</dim>
825 <dim>-1</dim>
826 <dim>768</dim>
827 </port>
828 </output>
829 </layer>
830 <layer id="59" name="Constant_5362" type="Const" version="opset1">
831 <data element_type="f32" shape="1, 1, 768" offset="163090536" size="3072" />
832 <output>
833 <port id="0" precision="FP32">
834 <dim>1</dim>
835 <dim>1</dim>
836 <dim>768</dim>
837 </port>
838 </output>
839 </layer>
840 <layer id="60" name="__module.roberta.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
841 <data auto_broadcast="numpy" />
842 <input>
843 <port id="0" precision="FP32">
844 <dim>-1</dim>
845 <dim>-1</dim>
846 <dim>768</dim>
847 </port>
848 <port id="1" precision="FP32">
849 <dim>1</dim>
850 <dim>1</dim>
851 <dim>768</dim>
852 </port>
853 </input>
854 <output>
855 <port id="2" precision="FP32" names="120,x.9">
856 <dim>-1</dim>
857 <dim>-1</dim>
858 <dim>768</dim>
859 </port>
860 </output>
861 </layer>
862 <layer id="61" name="__module.roberta.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
863 <data element_type="i64" shape="4" offset="158368808" size="32" />
864 <output>
865 <port id="0" precision="I64">
866 <dim>4</dim>
867 </port>
868 </output>
869 </layer>
870 <layer id="62" name="__module.roberta.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
871 <data special_zero="true" />
872 <input>
873 <port id="0" precision="FP32">
874 <dim>-1</dim>
875 <dim>-1</dim>
876 <dim>768</dim>
877 </port>
878 <port id="1" precision="I64">
879 <dim>4</dim>
880 </port>
881 </input>
882 <output>
883 <port id="2" precision="FP32" names="124,x.11">
884 <dim>-1</dim>
885 <dim>-1</dim>
886 <dim>12</dim>
887 <dim>64</dim>
888 </port>
889 </output>
890 </layer>
891 <layer id="63" name="Constant_275" type="Const" version="opset1">
892 <data element_type="i64" shape="4" offset="158368840" size="32" />
893 <output>
894 <port id="0" precision="I64" names="125">
895 <dim>4</dim>
896 </port>
897 </output>
898 </layer>
899 <layer id="64" name="__module.roberta.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
900 <input>
901 <port id="0" precision="FP32">
902 <dim>-1</dim>
903 <dim>-1</dim>
904 <dim>12</dim>
905 <dim>64</dim>
906 </port>
907 <port id="1" precision="I64">
908 <dim>4</dim>
909 </port>
910 </input>
911 <output>
912 <port id="2" precision="FP32" names="126">
913 <dim>-1</dim>
914 <dim>12</dim>
915 <dim>-1</dim>
916 <dim>64</dim>
917 </port>
918 </output>
919 </layer>
920 <layer id="65" name="Constant_5364" type="Const" version="opset1">
921 <data element_type="f32" shape="1, 1, 1, 1" offset="163093608" size="4" />
922 <output>
923 <port id="0" precision="FP32">
924 <dim>1</dim>
925 <dim>1</dim>
926 <dim>1</dim>
927 <dim>1</dim>
928 </port>
929 </output>
930 </layer>
931 <layer id="66" name="__module.roberta/aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
932 <input>
933 <port id="0" precision="I64">
934 <dim>-1</dim>
935 <dim>-1</dim>
936 </port>
937 <port id="1" precision="I64" />
938 </input>
939 <output>
940 <port id="2" precision="I64" names="69">
941 <dim>-1</dim>
942 <dim>1</dim>
943 <dim>-1</dim>
944 </port>
945 </output>
946 </layer>
947 <layer id="67" name="16" type="Const" version="opset1">
948 <data element_type="i64" shape="" offset="163093612" size="8" />
949 <output>
950 <port id="0" precision="I64" names="16" />
951 </output>
952 </layer>
953 <layer id="68" name="__module.roberta/aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
954 <input>
955 <port id="0" precision="I64">
956 <dim>-1</dim>
957 <dim>1</dim>
958 <dim>-1</dim>
959 </port>
960 <port id="1" precision="I64" />
961 </input>
962 <output>
963 <port id="2" precision="I64" names="70,71">
964 <dim>-1</dim>
965 <dim>1</dim>
966 <dim>1</dim>
967 <dim>-1</dim>
968 </port>
969 </output>
970 </layer>
971 <layer id="69" name="Constant_5490" type="Const" version="opset1">
972 <data element_type="i64" shape="1" offset="154421268" size="8" />
973 <output>
974 <port id="0" precision="I64">
975 <dim>1</dim>
976 </port>
977 </output>
978 </layer>
979 <layer id="70" name="Constant_5491" type="Const" version="opset1">
980 <data element_type="i64" shape="" offset="154421268" size="8" />
981 <output>
982 <port id="0" precision="I64" />
983 </output>
984 </layer>
985 <layer id="71" name="Gather_5492" type="Gather" version="opset8">
986 <data batch_dims="0" />
987 <input>
988 <port id="0" precision="I64">
989 <dim>2</dim>
990 </port>
991 <port id="1" precision="I64">
992 <dim>1</dim>
993 </port>
994 <port id="2" precision="I64" />
995 </input>
996 <output>
997 <port id="3" precision="I64" names="66">
998 <dim>1</dim>
999 </port>
1000 </output>
1001 </layer>
1002 <layer id="72" name="Constant_4638" type="Const" version="opset1">
1003 <data element_type="i64" shape="1" offset="154421276" size="8" />
1004 <output>
1005 <port id="0" precision="I64">
1006 <dim>1</dim>
1007 </port>
1008 </output>
1009 </layer>
1010 <layer id="73" name="Constant_5596" type="Const" version="opset1">
1011 <data element_type="i64" shape="2" offset="163093620" size="16" />
1012 <output>
1013 <port id="0" precision="I64">
1014 <dim>2</dim>
1015 </port>
1016 </output>
1017 </layer>
1018 <layer id="74" name="Constant_5597" type="Const" version="opset1">
1019 <data element_type="i64" shape="" offset="154421268" size="8" />
1020 <output>
1021 <port id="0" precision="I64" />
1022 </output>
1023 </layer>
1024 <layer id="75" name="Gather_5598" type="Gather" version="opset8">
1025 <data batch_dims="0" />
1026 <input>
1027 <port id="0" precision="I64">
1028 <dim>2</dim>
1029 </port>
1030 <port id="1" precision="I64">
1031 <dim>2</dim>
1032 </port>
1033 <port id="2" precision="I64" />
1034 </input>
1035 <output>
1036 <port id="3" precision="I64">
1037 <dim>2</dim>
1038 </port>
1039 </output>
1040 </layer>
1041 <layer id="76" name="__module.roberta/prim::ListConstruct/Concat_1" type="Concat" version="opset1">
1042 <data axis="0" />
1043 <input>
1044 <port id="0" precision="I64">
1045 <dim>1</dim>
1046 </port>
1047 <port id="1" precision="I64">
1048 <dim>1</dim>
1049 </port>
1050 <port id="2" precision="I64">
1051 <dim>2</dim>
1052 </port>
1053 </input>
1054 <output>
1055 <port id="3" precision="I64" names="72">
1056 <dim>4</dim>
1057 </port>
1058 </output>
1059 </layer>
1060 <layer id="77" name="__module.roberta/aten::expand/Broadcast_1" type="Broadcast" version="opset3">
1061 <data mode="bidirectional" />
1062 <input>
1063 <port id="0" precision="I64">
1064 <dim>-1</dim>
1065 <dim>1</dim>
1066 <dim>1</dim>
1067 <dim>-1</dim>
1068 </port>
1069 <port id="1" precision="I64">
1070 <dim>4</dim>
1071 </port>
1072 </input>
1073 <output>
1074 <port id="2" precision="I64" names="73">
1075 <dim>-1</dim>
1076 <dim>1</dim>
1077 <dim>-1</dim>
1078 <dim>-1</dim>
1079 </port>
1080 </output>
1081 </layer>
1082 <layer id="78" name="__module.roberta/aten::to/Convert" type="Convert" version="opset1">
1083 <data destination_type="f32" />
1084 <input>
1085 <port id="0" precision="I64">
1086 <dim>-1</dim>
1087 <dim>1</dim>
1088 <dim>-1</dim>
1089 <dim>-1</dim>
1090 </port>
1091 </input>
1092 <output>
1093 <port id="1" precision="FP32" names="74">
1094 <dim>-1</dim>
1095 <dim>1</dim>
1096 <dim>-1</dim>
1097 <dim>-1</dim>
1098 </port>
1099 </output>
1100 </layer>
1101 <layer id="79" name="Constant_5363" type="Const" version="opset1">
1102 <data element_type="f32" shape="1, 1, 1, 1" offset="163093608" size="4" />
1103 <output>
1104 <port id="0" precision="FP32">
1105 <dim>1</dim>
1106 <dim>1</dim>
1107 <dim>1</dim>
1108 <dim>1</dim>
1109 </port>
1110 </output>
1111 </layer>
1112 <layer id="80" name="__module.roberta/aten::rsub/Multiply" type="Multiply" version="opset1">
1113 <data auto_broadcast="numpy" />
1114 <input>
1115 <port id="0" precision="FP32">
1116 <dim>-1</dim>
1117 <dim>1</dim>
1118 <dim>-1</dim>
1119 <dim>-1</dim>
1120 </port>
1121 <port id="1" precision="FP32">
1122 <dim>1</dim>
1123 <dim>1</dim>
1124 <dim>1</dim>
1125 <dim>1</dim>
1126 </port>
1127 </input>
1128 <output>
1129 <port id="2" precision="FP32">
1130 <dim>-1</dim>
1131 <dim>1</dim>
1132 <dim>-1</dim>
1133 <dim>-1</dim>
1134 </port>
1135 </output>
1136 </layer>
1137 <layer id="81" name="__module.roberta/aten::rsub/Subtract" type="Subtract" version="opset1">
1138 <data auto_broadcast="numpy" />
1139 <input>
1140 <port id="0" precision="FP32">
1141 <dim>1</dim>
1142 <dim>1</dim>
1143 <dim>1</dim>
1144 <dim>1</dim>
1145 </port>
1146 <port id="1" precision="FP32">
1147 <dim>-1</dim>
1148 <dim>1</dim>
1149 <dim>-1</dim>
1150 <dim>-1</dim>
1151 </port>
1152 </input>
1153 <output>
1154 <port id="2" precision="FP32" names="75,inverted_mask">
1155 <dim>-1</dim>
1156 <dim>1</dim>
1157 <dim>-1</dim>
1158 <dim>-1</dim>
1159 </port>
1160 </output>
1161 </layer>
1162 <layer id="82" name="__module.roberta/aten::to/Convert_1" type="Convert" version="opset1">
1163 <data destination_type="boolean" />
1164 <input>
1165 <port id="0" precision="FP32">
1166 <dim>-1</dim>
1167 <dim>1</dim>
1168 <dim>-1</dim>
1169 <dim>-1</dim>
1170 </port>
1171 </input>
1172 <output>
1173 <port id="1" precision="BOOL" names="76">
1174 <dim>-1</dim>
1175 <dim>1</dim>
1176 <dim>-1</dim>
1177 <dim>-1</dim>
1178 </port>
1179 </output>
1180 </layer>
1181 <layer id="83" name="__module.roberta/aten::masked_fill/ConvertLike" type="Const" version="opset1">
1182 <data element_type="f32" shape="" offset="163093636" size="4" />
1183 <output>
1184 <port id="0" precision="FP32" />
1185 </output>
1186 </layer>
1187 <layer id="84" name="__module.roberta/aten::masked_fill/Select" type="Select" version="opset1">
1188 <data auto_broadcast="numpy" />
1189 <input>
1190 <port id="0" precision="BOOL">
1191 <dim>-1</dim>
1192 <dim>1</dim>
1193 <dim>-1</dim>
1194 <dim>-1</dim>
1195 </port>
1196 <port id="1" precision="FP32" />
1197 <port id="2" precision="FP32">
1198 <dim>-1</dim>
1199 <dim>1</dim>
1200 <dim>-1</dim>
1201 <dim>-1</dim>
1202 </port>
1203 </input>
1204 <output>
1205 <port id="3" precision="FP32" names="77">
1206 <dim>-1</dim>
1207 <dim>1</dim>
1208 <dim>-1</dim>
1209 <dim>-1</dim>
1210 </port>
1211 </output>
1212 </layer>
1213 <layer id="85" name="__module.roberta.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1214 <data causal="false" />
1215 <input>
1216 <port id="0" precision="FP32">
1217 <dim>-1</dim>
1218 <dim>12</dim>
1219 <dim>-1</dim>
1220 <dim>64</dim>
1221 </port>
1222 <port id="1" precision="FP32">
1223 <dim>-1</dim>
1224 <dim>12</dim>
1225 <dim>-1</dim>
1226 <dim>64</dim>
1227 </port>
1228 <port id="2" precision="FP32">
1229 <dim>-1</dim>
1230 <dim>12</dim>
1231 <dim>-1</dim>
1232 <dim>64</dim>
1233 </port>
1234 <port id="3" precision="FP32">
1235 <dim>-1</dim>
1236 <dim>1</dim>
1237 <dim>-1</dim>
1238 <dim>-1</dim>
1239 </port>
1240 </input>
1241 <output>
1242 <port id="4" precision="FP32" names="127,attn_output.1">
1243 <dim>-1</dim>
1244 <dim>12</dim>
1245 <dim>-1</dim>
1246 <dim>64</dim>
1247 </port>
1248 </output>
1249 </layer>
1250 <layer id="86" name="__module.roberta.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1251 <data element_type="i32" shape="4" offset="163093640" size="16" />
1252 <output>
1253 <port id="0" precision="I32">
1254 <dim>4</dim>
1255 </port>
1256 </output>
1257 </layer>
1258 <layer id="87" name="__module.roberta.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1259 <input>
1260 <port id="0" precision="FP32">
1261 <dim>-1</dim>
1262 <dim>12</dim>
1263 <dim>-1</dim>
1264 <dim>64</dim>
1265 </port>
1266 <port id="1" precision="I32">
1267 <dim>4</dim>
1268 </port>
1269 </input>
1270 <output>
1271 <port id="2" precision="FP32" names="128,attn_output.3">
1272 <dim>-1</dim>
1273 <dim>-1</dim>
1274 <dim>12</dim>
1275 <dim>64</dim>
1276 </port>
1277 </output>
1278 </layer>
1279 <layer id="88" name="Constant_5494" type="Const" version="opset1">
1280 <data element_type="i64" shape="3" offset="163093656" size="24" />
1281 <output>
1282 <port id="0" precision="I64">
1283 <dim>3</dim>
1284 </port>
1285 </output>
1286 </layer>
1287 <layer id="89" name="__module.roberta.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1288 <data special_zero="true" />
1289 <input>
1290 <port id="0" precision="FP32">
1291 <dim>-1</dim>
1292 <dim>-1</dim>
1293 <dim>12</dim>
1294 <dim>64</dim>
1295 </port>
1296 <port id="1" precision="I64">
1297 <dim>3</dim>
1298 </port>
1299 </input>
1300 <output>
1301 <port id="2" precision="FP32" names="130">
1302 <dim>-1</dim>
1303 <dim>-1</dim>
1304 <dim>768</dim>
1305 </port>
1306 </output>
1307 </layer>
1308 <layer id="90" name="self.roberta.encoder.layer.0.attention.output.dense.weight" type="Const" version="opset1">
1309 <data element_type="f32" shape="768, 768" offset="163093680" size="2359296" />
1310 <output>
1311 <port id="0" precision="FP32" names="self.roberta.encoder.layer.0.attention.output.dense.weight">
1312 <dim>768</dim>
1313 <dim>768</dim>
1314 </port>
1315 </output>
1316 </layer>
1317 <layer id="91" name="__module.roberta.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1318 <data transpose_a="false" transpose_b="true" />
1319 <input>
1320 <port id="0" precision="FP32">
1321 <dim>-1</dim>
1322 <dim>-1</dim>
1323 <dim>768</dim>
1324 </port>
1325 <port id="1" precision="FP32">
1326 <dim>768</dim>
1327 <dim>768</dim>
1328 </port>
1329 </input>
1330 <output>
1331 <port id="2" precision="FP32">
1332 <dim>-1</dim>
1333 <dim>-1</dim>
1334 <dim>768</dim>
1335 </port>
1336 </output>
1337 </layer>
1338 <layer id="92" name="Constant_5365" type="Const" version="opset1">
1339 <data element_type="f32" shape="1, 1, 768" offset="165452976" size="3072" />
1340 <output>
1341 <port id="0" precision="FP32">
1342 <dim>1</dim>
1343 <dim>1</dim>
1344 <dim>768</dim>
1345 </port>
1346 </output>
1347 </layer>
1348 <layer id="93" name="__module.roberta.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1349 <data auto_broadcast="numpy" />
1350 <input>
1351 <port id="0" precision="FP32">
1352 <dim>-1</dim>
1353 <dim>-1</dim>
1354 <dim>768</dim>
1355 </port>
1356 <port id="1" precision="FP32">
1357 <dim>1</dim>
1358 <dim>1</dim>
1359 <dim>768</dim>
1360 </port>
1361 </input>
1362 <output>
1363 <port id="2" precision="FP32" names="135,input.3">
1364 <dim>-1</dim>
1365 <dim>-1</dim>
1366 <dim>768</dim>
1367 </port>
1368 </output>
1369 </layer>
1370 <layer id="94" name="__module.roberta.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1371 <data auto_broadcast="numpy" />
1372 <input>
1373 <port id="0" precision="FP32">
1374 <dim>-1</dim>
1375 <dim>-1</dim>
1376 <dim>768</dim>
1377 </port>
1378 <port id="1" precision="FP32">
1379 <dim>-1</dim>
1380 <dim>-1</dim>
1381 <dim>768</dim>
1382 </port>
1383 </input>
1384 <output>
1385 <port id="2" precision="FP32" names="137">
1386 <dim>-1</dim>
1387 <dim>-1</dim>
1388 <dim>768</dim>
1389 </port>
1390 </output>
1391 </layer>
1392 <layer id="95" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1393 <data element_type="i32" shape="1" offset="156000292" size="4" />
1394 <output>
1395 <port id="0" precision="I32">
1396 <dim>1</dim>
1397 </port>
1398 </output>
1399 </layer>
1400 <layer id="96" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1401 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1402 <input>
1403 <port id="0" precision="FP32">
1404 <dim>-1</dim>
1405 <dim>-1</dim>
1406 <dim>768</dim>
1407 </port>
1408 <port id="1" precision="I32">
1409 <dim>1</dim>
1410 </port>
1411 </input>
1412 <output>
1413 <port id="2" precision="FP32">
1414 <dim>-1</dim>
1415 <dim>-1</dim>
1416 <dim>768</dim>
1417 </port>
1418 </output>
1419 </layer>
1420 <layer id="97" name="Constant_5366" type="Const" version="opset1">
1421 <data element_type="f32" shape="1, 1, 768" offset="165456048" size="3072" />
1422 <output>
1423 <port id="0" precision="FP32">
1424 <dim>1</dim>
1425 <dim>1</dim>
1426 <dim>768</dim>
1427 </port>
1428 </output>
1429 </layer>
1430 <layer id="98" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1431 <data auto_broadcast="numpy" />
1432 <input>
1433 <port id="0" precision="FP32">
1434 <dim>-1</dim>
1435 <dim>-1</dim>
1436 <dim>768</dim>
1437 </port>
1438 <port id="1" precision="FP32">
1439 <dim>1</dim>
1440 <dim>1</dim>
1441 <dim>768</dim>
1442 </port>
1443 </input>
1444 <output>
1445 <port id="2" precision="FP32">
1446 <dim>-1</dim>
1447 <dim>-1</dim>
1448 <dim>768</dim>
1449 </port>
1450 </output>
1451 </layer>
1452 <layer id="99" name="Constant_5367" type="Const" version="opset1">
1453 <data element_type="f32" shape="1, 1, 768" offset="165459120" size="3072" />
1454 <output>
1455 <port id="0" precision="FP32">
1456 <dim>1</dim>
1457 <dim>1</dim>
1458 <dim>768</dim>
1459 </port>
1460 </output>
1461 </layer>
1462 <layer id="100" name="__module.roberta.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1463 <data auto_broadcast="numpy" />
1464 <input>
1465 <port id="0" precision="FP32">
1466 <dim>-1</dim>
1467 <dim>-1</dim>
1468 <dim>768</dim>
1469 </port>
1470 <port id="1" precision="FP32">
1471 <dim>1</dim>
1472 <dim>1</dim>
1473 <dim>768</dim>
1474 </port>
1475 </input>
1476 <output>
1477 <port id="2" precision="FP32" names="141,input_tensor.1">
1478 <dim>-1</dim>
1479 <dim>-1</dim>
1480 <dim>768</dim>
1481 </port>
1482 </output>
1483 </layer>
1484 <layer id="101" name="self.roberta.encoder.layer.0.intermediate.dense.weight" type="Const" version="opset1">
1485 <data element_type="f32" shape="3072, 768" offset="165462192" size="9437184" />
1486 <output>
1487 <port id="0" precision="FP32" names="self.roberta.encoder.layer.0.intermediate.dense.weight">
1488 <dim>3072</dim>
1489 <dim>768</dim>
1490 </port>
1491 </output>
1492 </layer>
1493 <layer id="102" name="__module.roberta.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1494 <data transpose_a="false" transpose_b="true" />
1495 <input>
1496 <port id="0" precision="FP32">
1497 <dim>-1</dim>
1498 <dim>-1</dim>
1499 <dim>768</dim>
1500 </port>
1501 <port id="1" precision="FP32">
1502 <dim>3072</dim>
1503 <dim>768</dim>
1504 </port>
1505 </input>
1506 <output>
1507 <port id="2" precision="FP32">
1508 <dim>-1</dim>
1509 <dim>-1</dim>
1510 <dim>3072</dim>
1511 </port>
1512 </output>
1513 </layer>
1514 <layer id="103" name="Constant_5368" type="Const" version="opset1">
1515 <data element_type="f32" shape="1, 1, 3072" offset="174899376" size="12288" />
1516 <output>
1517 <port id="0" precision="FP32">
1518 <dim>1</dim>
1519 <dim>1</dim>
1520 <dim>3072</dim>
1521 </port>
1522 </output>
1523 </layer>
1524 <layer id="104" name="__module.roberta.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
1525 <data auto_broadcast="numpy" />
1526 <input>
1527 <port id="0" precision="FP32">
1528 <dim>-1</dim>
1529 <dim>-1</dim>
1530 <dim>3072</dim>
1531 </port>
1532 <port id="1" precision="FP32">
1533 <dim>1</dim>
1534 <dim>1</dim>
1535 <dim>3072</dim>
1536 </port>
1537 </input>
1538 <output>
1539 <port id="2" precision="FP32" names="145">
1540 <dim>-1</dim>
1541 <dim>-1</dim>
1542 <dim>3072</dim>
1543 </port>
1544 </output>
1545 </layer>
1546 <layer id="105" name="__module.roberta.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
1547 <data approximation_mode="ERF" />
1548 <input>
1549 <port id="0" precision="FP32">
1550 <dim>-1</dim>
1551 <dim>-1</dim>
1552 <dim>3072</dim>
1553 </port>
1554 </input>
1555 <output>
1556 <port id="1" precision="FP32" names="146">
1557 <dim>-1</dim>
1558 <dim>-1</dim>
1559 <dim>3072</dim>
1560 </port>
1561 </output>
1562 </layer>
1563 <layer id="106" name="self.roberta.encoder.layer.0.output.dense.weight" type="Const" version="opset1">
1564 <data element_type="f32" shape="768, 3072" offset="174911664" size="9437184" />
1565 <output>
1566 <port id="0" precision="FP32" names="self.roberta.encoder.layer.0.output.dense.weight">
1567 <dim>768</dim>
1568 <dim>3072</dim>
1569 </port>
1570 </output>
1571 </layer>
1572 <layer id="107" name="__module.roberta.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1573 <data transpose_a="false" transpose_b="true" />
1574 <input>
1575 <port id="0" precision="FP32">
1576 <dim>-1</dim>
1577 <dim>-1</dim>
1578 <dim>3072</dim>
1579 </port>
1580 <port id="1" precision="FP32">
1581 <dim>768</dim>
1582 <dim>3072</dim>
1583 </port>
1584 </input>
1585 <output>
1586 <port id="2" precision="FP32">
1587 <dim>-1</dim>
1588 <dim>-1</dim>
1589 <dim>768</dim>
1590 </port>
1591 </output>
1592 </layer>
1593 <layer id="108" name="Constant_5369" type="Const" version="opset1">
1594 <data element_type="f32" shape="1, 1, 768" offset="184348848" size="3072" />
1595 <output>
1596 <port id="0" precision="FP32">
1597 <dim>1</dim>
1598 <dim>1</dim>
1599 <dim>768</dim>
1600 </port>
1601 </output>
1602 </layer>
1603 <layer id="109" name="__module.roberta.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
1604 <data auto_broadcast="numpy" />
1605 <input>
1606 <port id="0" precision="FP32">
1607 <dim>-1</dim>
1608 <dim>-1</dim>
1609 <dim>768</dim>
1610 </port>
1611 <port id="1" precision="FP32">
1612 <dim>1</dim>
1613 <dim>1</dim>
1614 <dim>768</dim>
1615 </port>
1616 </input>
1617 <output>
1618 <port id="2" precision="FP32" names="151,input.5">
1619 <dim>-1</dim>
1620 <dim>-1</dim>
1621 <dim>768</dim>
1622 </port>
1623 </output>
1624 </layer>
1625 <layer id="110" name="__module.roberta.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
1626 <data auto_broadcast="numpy" />
1627 <input>
1628 <port id="0" precision="FP32">
1629 <dim>-1</dim>
1630 <dim>-1</dim>
1631 <dim>768</dim>
1632 </port>
1633 <port id="1" precision="FP32">
1634 <dim>-1</dim>
1635 <dim>-1</dim>
1636 <dim>768</dim>
1637 </port>
1638 </input>
1639 <output>
1640 <port id="2" precision="FP32" names="153">
1641 <dim>-1</dim>
1642 <dim>-1</dim>
1643 <dim>768</dim>
1644 </port>
1645 </output>
1646 </layer>
1647 <layer id="111" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1648 <data element_type="i32" shape="1" offset="156000292" size="4" />
1649 <output>
1650 <port id="0" precision="I32">
1651 <dim>1</dim>
1652 </port>
1653 </output>
1654 </layer>
1655 <layer id="112" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1656 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1657 <input>
1658 <port id="0" precision="FP32">
1659 <dim>-1</dim>
1660 <dim>-1</dim>
1661 <dim>768</dim>
1662 </port>
1663 <port id="1" precision="I32">
1664 <dim>1</dim>
1665 </port>
1666 </input>
1667 <output>
1668 <port id="2" precision="FP32">
1669 <dim>-1</dim>
1670 <dim>-1</dim>
1671 <dim>768</dim>
1672 </port>
1673 </output>
1674 </layer>
1675 <layer id="113" name="Constant_5370" type="Const" version="opset1">
1676 <data element_type="f32" shape="1, 1, 768" offset="184351920" size="3072" />
1677 <output>
1678 <port id="0" precision="FP32">
1679 <dim>1</dim>
1680 <dim>1</dim>
1681 <dim>768</dim>
1682 </port>
1683 </output>
1684 </layer>
1685 <layer id="114" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1686 <data auto_broadcast="numpy" />
1687 <input>
1688 <port id="0" precision="FP32">
1689 <dim>-1</dim>
1690 <dim>-1</dim>
1691 <dim>768</dim>
1692 </port>
1693 <port id="1" precision="FP32">
1694 <dim>1</dim>
1695 <dim>1</dim>
1696 <dim>768</dim>
1697 </port>
1698 </input>
1699 <output>
1700 <port id="2" precision="FP32">
1701 <dim>-1</dim>
1702 <dim>-1</dim>
1703 <dim>768</dim>
1704 </port>
1705 </output>
1706 </layer>
1707 <layer id="115" name="Constant_5371" type="Const" version="opset1">
1708 <data element_type="f32" shape="1, 1, 768" offset="184354992" size="3072" />
1709 <output>
1710 <port id="0" precision="FP32">
1711 <dim>1</dim>
1712 <dim>1</dim>
1713 <dim>768</dim>
1714 </port>
1715 </output>
1716 </layer>
1717 <layer id="116" name="__module.roberta.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1718 <data auto_broadcast="numpy" />
1719 <input>
1720 <port id="0" precision="FP32">
1721 <dim>-1</dim>
1722 <dim>-1</dim>
1723 <dim>768</dim>
1724 </port>
1725 <port id="1" precision="FP32">
1726 <dim>1</dim>
1727 <dim>1</dim>
1728 <dim>768</dim>
1729 </port>
1730 </input>
1731 <output>
1732 <port id="2" precision="FP32" names="157,hidden_states.7">
1733 <dim>-1</dim>
1734 <dim>-1</dim>
1735 <dim>768</dim>
1736 </port>
1737 </output>
1738 </layer>
1739 <layer id="117" name="self.roberta.encoder.layer.1.attention.self.query.weight" type="Const" version="opset1">
1740 <data element_type="f32" shape="768, 768" offset="184358064" size="2359296" />
1741 <output>
1742 <port id="0" precision="FP32" names="self.roberta.encoder.layer.1.attention.self.query.weight">
1743 <dim>768</dim>
1744 <dim>768</dim>
1745 </port>
1746 </output>
1747 </layer>
1748 <layer id="118" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
1749 <data transpose_a="false" transpose_b="true" />
1750 <input>
1751 <port id="0" precision="FP32">
1752 <dim>-1</dim>
1753 <dim>-1</dim>
1754 <dim>768</dim>
1755 </port>
1756 <port id="1" precision="FP32">
1757 <dim>768</dim>
1758 <dim>768</dim>
1759 </port>
1760 </input>
1761 <output>
1762 <port id="2" precision="FP32">
1763 <dim>-1</dim>
1764 <dim>-1</dim>
1765 <dim>768</dim>
1766 </port>
1767 </output>
1768 </layer>
1769 <layer id="119" name="Constant_5372" type="Const" version="opset1">
1770 <data element_type="f32" shape="1, 1, 768" offset="186717360" size="3072" />
1771 <output>
1772 <port id="0" precision="FP32">
1773 <dim>1</dim>
1774 <dim>1</dim>
1775 <dim>768</dim>
1776 </port>
1777 </output>
1778 </layer>
1779 <layer id="120" name="__module.roberta.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
1780 <data auto_broadcast="numpy" />
1781 <input>
1782 <port id="0" precision="FP32">
1783 <dim>-1</dim>
1784 <dim>-1</dim>
1785 <dim>768</dim>
1786 </port>
1787 <port id="1" precision="FP32">
1788 <dim>1</dim>
1789 <dim>1</dim>
1790 <dim>768</dim>
1791 </port>
1792 </input>
1793 <output>
1794 <port id="2" precision="FP32" names="170,x.13">
1795 <dim>-1</dim>
1796 <dim>-1</dim>
1797 <dim>768</dim>
1798 </port>
1799 </output>
1800 </layer>
1801 <layer id="121" name="__module.roberta.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
1802 <data element_type="i64" shape="4" offset="158368808" size="32" />
1803 <output>
1804 <port id="0" precision="I64">
1805 <dim>4</dim>
1806 </port>
1807 </output>
1808 </layer>
1809 <layer id="122" name="__module.roberta.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
1810 <data special_zero="true" />
1811 <input>
1812 <port id="0" precision="FP32">
1813 <dim>-1</dim>
1814 <dim>-1</dim>
1815 <dim>768</dim>
1816 </port>
1817 <port id="1" precision="I64">
1818 <dim>4</dim>
1819 </port>
1820 </input>
1821 <output>
1822 <port id="2" precision="FP32" names="174,x.15">
1823 <dim>-1</dim>
1824 <dim>-1</dim>
1825 <dim>12</dim>
1826 <dim>64</dim>
1827 </port>
1828 </output>
1829 </layer>
1830 <layer id="123" name="Constant_452" type="Const" version="opset1">
1831 <data element_type="i64" shape="4" offset="158368840" size="32" />
1832 <output>
1833 <port id="0" precision="I64" names="175">
1834 <dim>4</dim>
1835 </port>
1836 </output>
1837 </layer>
1838 <layer id="124" name="__module.roberta.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
1839 <input>
1840 <port id="0" precision="FP32">
1841 <dim>-1</dim>
1842 <dim>-1</dim>
1843 <dim>12</dim>
1844 <dim>64</dim>
1845 </port>
1846 <port id="1" precision="I64">
1847 <dim>4</dim>
1848 </port>
1849 </input>
1850 <output>
1851 <port id="2" precision="FP32" names="176">
1852 <dim>-1</dim>
1853 <dim>12</dim>
1854 <dim>-1</dim>
1855 <dim>64</dim>
1856 </port>
1857 </output>
1858 </layer>
1859 <layer id="125" name="self.roberta.encoder.layer.1.attention.self.key.weight" type="Const" version="opset1">
1860 <data element_type="f32" shape="768, 768" offset="186720432" size="2359296" />
1861 <output>
1862 <port id="0" precision="FP32" names="self.roberta.encoder.layer.1.attention.self.key.weight">
1863 <dim>768</dim>
1864 <dim>768</dim>
1865 </port>
1866 </output>
1867 </layer>
1868 <layer id="126" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
1869 <data transpose_a="false" transpose_b="true" />
1870 <input>
1871 <port id="0" precision="FP32">
1872 <dim>-1</dim>
1873 <dim>-1</dim>
1874 <dim>768</dim>
1875 </port>
1876 <port id="1" precision="FP32">
1877 <dim>768</dim>
1878 <dim>768</dim>
1879 </port>
1880 </input>
1881 <output>
1882 <port id="2" precision="FP32">
1883 <dim>-1</dim>
1884 <dim>-1</dim>
1885 <dim>768</dim>
1886 </port>
1887 </output>
1888 </layer>
1889 <layer id="127" name="Constant_5373" type="Const" version="opset1">
1890 <data element_type="f32" shape="1, 1, 768" offset="189079728" size="3072" />
1891 <output>
1892 <port id="0" precision="FP32">
1893 <dim>1</dim>
1894 <dim>1</dim>
1895 <dim>768</dim>
1896 </port>
1897 </output>
1898 </layer>
1899 <layer id="128" name="__module.roberta.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
1900 <data auto_broadcast="numpy" />
1901 <input>
1902 <port id="0" precision="FP32">
1903 <dim>-1</dim>
1904 <dim>-1</dim>
1905 <dim>768</dim>
1906 </port>
1907 <port id="1" precision="FP32">
1908 <dim>1</dim>
1909 <dim>1</dim>
1910 <dim>768</dim>
1911 </port>
1912 </input>
1913 <output>
1914 <port id="2" precision="FP32" names="179,x.17">
1915 <dim>-1</dim>
1916 <dim>-1</dim>
1917 <dim>768</dim>
1918 </port>
1919 </output>
1920 </layer>
1921 <layer id="129" name="__module.roberta.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
1922 <data element_type="i64" shape="4" offset="158368808" size="32" />
1923 <output>
1924 <port id="0" precision="I64">
1925 <dim>4</dim>
1926 </port>
1927 </output>
1928 </layer>
1929 <layer id="130" name="__module.roberta.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
1930 <data special_zero="true" />
1931 <input>
1932 <port id="0" precision="FP32">
1933 <dim>-1</dim>
1934 <dim>-1</dim>
1935 <dim>768</dim>
1936 </port>
1937 <port id="1" precision="I64">
1938 <dim>4</dim>
1939 </port>
1940 </input>
1941 <output>
1942 <port id="2" precision="FP32" names="183,x.19">
1943 <dim>-1</dim>
1944 <dim>-1</dim>
1945 <dim>12</dim>
1946 <dim>64</dim>
1947 </port>
1948 </output>
1949 </layer>
1950 <layer id="131" name="Constant_475" type="Const" version="opset1">
1951 <data element_type="i64" shape="4" offset="158368840" size="32" />
1952 <output>
1953 <port id="0" precision="I64" names="184">
1954 <dim>4</dim>
1955 </port>
1956 </output>
1957 </layer>
1958 <layer id="132" name="__module.roberta.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1959 <input>
1960 <port id="0" precision="FP32">
1961 <dim>-1</dim>
1962 <dim>-1</dim>
1963 <dim>12</dim>
1964 <dim>64</dim>
1965 </port>
1966 <port id="1" precision="I64">
1967 <dim>4</dim>
1968 </port>
1969 </input>
1970 <output>
1971 <port id="2" precision="FP32" names="185">
1972 <dim>-1</dim>
1973 <dim>12</dim>
1974 <dim>-1</dim>
1975 <dim>64</dim>
1976 </port>
1977 </output>
1978 </layer>
1979 <layer id="133" name="self.roberta.encoder.layer.1.attention.self.value.weight" type="Const" version="opset1">
1980 <data element_type="f32" shape="768, 768" offset="189082800" size="2359296" />
1981 <output>
1982 <port id="0" precision="FP32" names="self.roberta.encoder.layer.1.attention.self.value.weight">
1983 <dim>768</dim>
1984 <dim>768</dim>
1985 </port>
1986 </output>
1987 </layer>
1988 <layer id="134" name="__module.roberta.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1989 <data transpose_a="false" transpose_b="true" />
1990 <input>
1991 <port id="0" precision="FP32">
1992 <dim>-1</dim>
1993 <dim>-1</dim>
1994 <dim>768</dim>
1995 </port>
1996 <port id="1" precision="FP32">
1997 <dim>768</dim>
1998 <dim>768</dim>
1999 </port>
2000 </input>
2001 <output>
2002 <port id="2" precision="FP32">
2003 <dim>-1</dim>
2004 <dim>-1</dim>
2005 <dim>768</dim>
2006 </port>
2007 </output>
2008 </layer>
2009 <layer id="135" name="Constant_5374" type="Const" version="opset1">
2010 <data element_type="f32" shape="1, 1, 768" offset="191442096" size="3072" />
2011 <output>
2012 <port id="0" precision="FP32">
2013 <dim>1</dim>
2014 <dim>1</dim>
2015 <dim>768</dim>
2016 </port>
2017 </output>
2018 </layer>
2019 <layer id="136" name="__module.roberta.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
2020 <data auto_broadcast="numpy" />
2021 <input>
2022 <port id="0" precision="FP32">
2023 <dim>-1</dim>
2024 <dim>-1</dim>
2025 <dim>768</dim>
2026 </port>
2027 <port id="1" precision="FP32">
2028 <dim>1</dim>
2029 <dim>1</dim>
2030 <dim>768</dim>
2031 </port>
2032 </input>
2033 <output>
2034 <port id="2" precision="FP32" names="188,x.21">
2035 <dim>-1</dim>
2036 <dim>-1</dim>
2037 <dim>768</dim>
2038 </port>
2039 </output>
2040 </layer>
2041 <layer id="137" name="__module.roberta.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
2042 <data element_type="i64" shape="4" offset="158368808" size="32" />
2043 <output>
2044 <port id="0" precision="I64">
2045 <dim>4</dim>
2046 </port>
2047 </output>
2048 </layer>
2049 <layer id="138" name="__module.roberta.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
2050 <data special_zero="true" />
2051 <input>
2052 <port id="0" precision="FP32">
2053 <dim>-1</dim>
2054 <dim>-1</dim>
2055 <dim>768</dim>
2056 </port>
2057 <port id="1" precision="I64">
2058 <dim>4</dim>
2059 </port>
2060 </input>
2061 <output>
2062 <port id="2" precision="FP32" names="192,x.23">
2063 <dim>-1</dim>
2064 <dim>-1</dim>
2065 <dim>12</dim>
2066 <dim>64</dim>
2067 </port>
2068 </output>
2069 </layer>
2070 <layer id="139" name="Constant_498" type="Const" version="opset1">
2071 <data element_type="i64" shape="4" offset="158368840" size="32" />
2072 <output>
2073 <port id="0" precision="I64" names="193">
2074 <dim>4</dim>
2075 </port>
2076 </output>
2077 </layer>
2078 <layer id="140" name="__module.roberta.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
2079 <input>
2080 <port id="0" precision="FP32">
2081 <dim>-1</dim>
2082 <dim>-1</dim>
2083 <dim>12</dim>
2084 <dim>64</dim>
2085 </port>
2086 <port id="1" precision="I64">
2087 <dim>4</dim>
2088 </port>
2089 </input>
2090 <output>
2091 <port id="2" precision="FP32" names="194">
2092 <dim>-1</dim>
2093 <dim>12</dim>
2094 <dim>-1</dim>
2095 <dim>64</dim>
2096 </port>
2097 </output>
2098 </layer>
2099 <layer id="141" name="__module.roberta.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
2100 <data causal="false" />
2101 <input>
2102 <port id="0" precision="FP32">
2103 <dim>-1</dim>
2104 <dim>12</dim>
2105 <dim>-1</dim>
2106 <dim>64</dim>
2107 </port>
2108 <port id="1" precision="FP32">
2109 <dim>-1</dim>
2110 <dim>12</dim>
2111 <dim>-1</dim>
2112 <dim>64</dim>
2113 </port>
2114 <port id="2" precision="FP32">
2115 <dim>-1</dim>
2116 <dim>12</dim>
2117 <dim>-1</dim>
2118 <dim>64</dim>
2119 </port>
2120 <port id="3" precision="FP32">
2121 <dim>-1</dim>
2122 <dim>1</dim>
2123 <dim>-1</dim>
2124 <dim>-1</dim>
2125 </port>
2126 </input>
2127 <output>
2128 <port id="4" precision="FP32" names="195,attn_output.5">
2129 <dim>-1</dim>
2130 <dim>12</dim>
2131 <dim>-1</dim>
2132 <dim>64</dim>
2133 </port>
2134 </output>
2135 </layer>
2136 <layer id="142" name="__module.roberta.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
2137 <data element_type="i32" shape="4" offset="163093640" size="16" />
2138 <output>
2139 <port id="0" precision="I32">
2140 <dim>4</dim>
2141 </port>
2142 </output>
2143 </layer>
2144 <layer id="143" name="__module.roberta.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
2145 <input>
2146 <port id="0" precision="FP32">
2147 <dim>-1</dim>
2148 <dim>12</dim>
2149 <dim>-1</dim>
2150 <dim>64</dim>
2151 </port>
2152 <port id="1" precision="I32">
2153 <dim>4</dim>
2154 </port>
2155 </input>
2156 <output>
2157 <port id="2" precision="FP32" names="196,attn_output.7">
2158 <dim>-1</dim>
2159 <dim>-1</dim>
2160 <dim>12</dim>
2161 <dim>64</dim>
2162 </port>
2163 </output>
2164 </layer>
2165 <layer id="144" name="Constant_5495" type="Const" version="opset1">
2166 <data element_type="i64" shape="3" offset="163093656" size="24" />
2167 <output>
2168 <port id="0" precision="I64">
2169 <dim>3</dim>
2170 </port>
2171 </output>
2172 </layer>
2173 <layer id="145" name="__module.roberta.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
2174 <data special_zero="true" />
2175 <input>
2176 <port id="0" precision="FP32">
2177 <dim>-1</dim>
2178 <dim>-1</dim>
2179 <dim>12</dim>
2180 <dim>64</dim>
2181 </port>
2182 <port id="1" precision="I64">
2183 <dim>3</dim>
2184 </port>
2185 </input>
2186 <output>
2187 <port id="2" precision="FP32" names="198">
2188 <dim>-1</dim>
2189 <dim>-1</dim>
2190 <dim>768</dim>
2191 </port>
2192 </output>
2193 </layer>
2194 <layer id="146" name="self.roberta.encoder.layer.1.attention.output.dense.weight" type="Const" version="opset1">
2195 <data element_type="f32" shape="768, 768" offset="191445168" size="2359296" />
2196 <output>
2197 <port id="0" precision="FP32" names="self.roberta.encoder.layer.1.attention.output.dense.weight">
2198 <dim>768</dim>
2199 <dim>768</dim>
2200 </port>
2201 </output>
2202 </layer>
2203 <layer id="147" name="__module.roberta.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2204 <data transpose_a="false" transpose_b="true" />
2205 <input>
2206 <port id="0" precision="FP32">
2207 <dim>-1</dim>
2208 <dim>-1</dim>
2209 <dim>768</dim>
2210 </port>
2211 <port id="1" precision="FP32">
2212 <dim>768</dim>
2213 <dim>768</dim>
2214 </port>
2215 </input>
2216 <output>
2217 <port id="2" precision="FP32">
2218 <dim>-1</dim>
2219 <dim>-1</dim>
2220 <dim>768</dim>
2221 </port>
2222 </output>
2223 </layer>
2224 <layer id="148" name="Constant_5375" type="Const" version="opset1">
2225 <data element_type="f32" shape="1, 1, 768" offset="193804464" size="3072" />
2226 <output>
2227 <port id="0" precision="FP32">
2228 <dim>1</dim>
2229 <dim>1</dim>
2230 <dim>768</dim>
2231 </port>
2232 </output>
2233 </layer>
2234 <layer id="149" name="__module.roberta.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
2235 <data auto_broadcast="numpy" />
2236 <input>
2237 <port id="0" precision="FP32">
2238 <dim>-1</dim>
2239 <dim>-1</dim>
2240 <dim>768</dim>
2241 </port>
2242 <port id="1" precision="FP32">
2243 <dim>1</dim>
2244 <dim>1</dim>
2245 <dim>768</dim>
2246 </port>
2247 </input>
2248 <output>
2249 <port id="2" precision="FP32" names="203,input.7">
2250 <dim>-1</dim>
2251 <dim>-1</dim>
2252 <dim>768</dim>
2253 </port>
2254 </output>
2255 </layer>
2256 <layer id="150" name="__module.roberta.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
2257 <data auto_broadcast="numpy" />
2258 <input>
2259 <port id="0" precision="FP32">
2260 <dim>-1</dim>
2261 <dim>-1</dim>
2262 <dim>768</dim>
2263 </port>
2264 <port id="1" precision="FP32">
2265 <dim>-1</dim>
2266 <dim>-1</dim>
2267 <dim>768</dim>
2268 </port>
2269 </input>
2270 <output>
2271 <port id="2" precision="FP32" names="205">
2272 <dim>-1</dim>
2273 <dim>-1</dim>
2274 <dim>768</dim>
2275 </port>
2276 </output>
2277 </layer>
2278 <layer id="151" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2279 <data element_type="i32" shape="1" offset="156000292" size="4" />
2280 <output>
2281 <port id="0" precision="I32">
2282 <dim>1</dim>
2283 </port>
2284 </output>
2285 </layer>
2286 <layer id="152" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2287 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2288 <input>
2289 <port id="0" precision="FP32">
2290 <dim>-1</dim>
2291 <dim>-1</dim>
2292 <dim>768</dim>
2293 </port>
2294 <port id="1" precision="I32">
2295 <dim>1</dim>
2296 </port>
2297 </input>
2298 <output>
2299 <port id="2" precision="FP32">
2300 <dim>-1</dim>
2301 <dim>-1</dim>
2302 <dim>768</dim>
2303 </port>
2304 </output>
2305 </layer>
2306 <layer id="153" name="Constant_5376" type="Const" version="opset1">
2307 <data element_type="f32" shape="1, 1, 768" offset="193807536" size="3072" />
2308 <output>
2309 <port id="0" precision="FP32">
2310 <dim>1</dim>
2311 <dim>1</dim>
2312 <dim>768</dim>
2313 </port>
2314 </output>
2315 </layer>
2316 <layer id="154" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2317 <data auto_broadcast="numpy" />
2318 <input>
2319 <port id="0" precision="FP32">
2320 <dim>-1</dim>
2321 <dim>-1</dim>
2322 <dim>768</dim>
2323 </port>
2324 <port id="1" precision="FP32">
2325 <dim>1</dim>
2326 <dim>1</dim>
2327 <dim>768</dim>
2328 </port>
2329 </input>
2330 <output>
2331 <port id="2" precision="FP32">
2332 <dim>-1</dim>
2333 <dim>-1</dim>
2334 <dim>768</dim>
2335 </port>
2336 </output>
2337 </layer>
2338 <layer id="155" name="Constant_5377" type="Const" version="opset1">
2339 <data element_type="f32" shape="1, 1, 768" offset="193810608" size="3072" />
2340 <output>
2341 <port id="0" precision="FP32">
2342 <dim>1</dim>
2343 <dim>1</dim>
2344 <dim>768</dim>
2345 </port>
2346 </output>
2347 </layer>
2348 <layer id="156" name="__module.roberta.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2349 <data auto_broadcast="numpy" />
2350 <input>
2351 <port id="0" precision="FP32">
2352 <dim>-1</dim>
2353 <dim>-1</dim>
2354 <dim>768</dim>
2355 </port>
2356 <port id="1" precision="FP32">
2357 <dim>1</dim>
2358 <dim>1</dim>
2359 <dim>768</dim>
2360 </port>
2361 </input>
2362 <output>
2363 <port id="2" precision="FP32" names="209,input_tensor.3">
2364 <dim>-1</dim>
2365 <dim>-1</dim>
2366 <dim>768</dim>
2367 </port>
2368 </output>
2369 </layer>
2370 <layer id="157" name="self.roberta.encoder.layer.1.intermediate.dense.weight" type="Const" version="opset1">
2371 <data element_type="f32" shape="3072, 768" offset="193813680" size="9437184" />
2372 <output>
2373 <port id="0" precision="FP32" names="self.roberta.encoder.layer.1.intermediate.dense.weight">
2374 <dim>3072</dim>
2375 <dim>768</dim>
2376 </port>
2377 </output>
2378 </layer>
2379 <layer id="158" name="__module.roberta.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2380 <data transpose_a="false" transpose_b="true" />
2381 <input>
2382 <port id="0" precision="FP32">
2383 <dim>-1</dim>
2384 <dim>-1</dim>
2385 <dim>768</dim>
2386 </port>
2387 <port id="1" precision="FP32">
2388 <dim>3072</dim>
2389 <dim>768</dim>
2390 </port>
2391 </input>
2392 <output>
2393 <port id="2" precision="FP32">
2394 <dim>-1</dim>
2395 <dim>-1</dim>
2396 <dim>3072</dim>
2397 </port>
2398 </output>
2399 </layer>
2400 <layer id="159" name="Constant_5378" type="Const" version="opset1">
2401 <data element_type="f32" shape="1, 1, 3072" offset="203250864" size="12288" />
2402 <output>
2403 <port id="0" precision="FP32">
2404 <dim>1</dim>
2405 <dim>1</dim>
2406 <dim>3072</dim>
2407 </port>
2408 </output>
2409 </layer>
2410 <layer id="160" name="__module.roberta.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2411 <data auto_broadcast="numpy" />
2412 <input>
2413 <port id="0" precision="FP32">
2414 <dim>-1</dim>
2415 <dim>-1</dim>
2416 <dim>3072</dim>
2417 </port>
2418 <port id="1" precision="FP32">
2419 <dim>1</dim>
2420 <dim>1</dim>
2421 <dim>3072</dim>
2422 </port>
2423 </input>
2424 <output>
2425 <port id="2" precision="FP32" names="213">
2426 <dim>-1</dim>
2427 <dim>-1</dim>
2428 <dim>3072</dim>
2429 </port>
2430 </output>
2431 </layer>
2432 <layer id="161" name="__module.roberta.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2433 <data approximation_mode="ERF" />
2434 <input>
2435 <port id="0" precision="FP32">
2436 <dim>-1</dim>
2437 <dim>-1</dim>
2438 <dim>3072</dim>
2439 </port>
2440 </input>
2441 <output>
2442 <port id="1" precision="FP32" names="214">
2443 <dim>-1</dim>
2444 <dim>-1</dim>
2445 <dim>3072</dim>
2446 </port>
2447 </output>
2448 </layer>
2449 <layer id="162" name="self.roberta.encoder.layer.1.output.dense.weight" type="Const" version="opset1">
2450 <data element_type="f32" shape="768, 3072" offset="203263152" size="9437184" />
2451 <output>
2452 <port id="0" precision="FP32" names="self.roberta.encoder.layer.1.output.dense.weight">
2453 <dim>768</dim>
2454 <dim>3072</dim>
2455 </port>
2456 </output>
2457 </layer>
2458 <layer id="163" name="__module.roberta.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2459 <data transpose_a="false" transpose_b="true" />
2460 <input>
2461 <port id="0" precision="FP32">
2462 <dim>-1</dim>
2463 <dim>-1</dim>
2464 <dim>3072</dim>
2465 </port>
2466 <port id="1" precision="FP32">
2467 <dim>768</dim>
2468 <dim>3072</dim>
2469 </port>
2470 </input>
2471 <output>
2472 <port id="2" precision="FP32">
2473 <dim>-1</dim>
2474 <dim>-1</dim>
2475 <dim>768</dim>
2476 </port>
2477 </output>
2478 </layer>
2479 <layer id="164" name="Constant_5379" type="Const" version="opset1">
2480 <data element_type="f32" shape="1, 1, 768" offset="212700336" size="3072" />
2481 <output>
2482 <port id="0" precision="FP32">
2483 <dim>1</dim>
2484 <dim>1</dim>
2485 <dim>768</dim>
2486 </port>
2487 </output>
2488 </layer>
2489 <layer id="165" name="__module.roberta.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
2490 <data auto_broadcast="numpy" />
2491 <input>
2492 <port id="0" precision="FP32">
2493 <dim>-1</dim>
2494 <dim>-1</dim>
2495 <dim>768</dim>
2496 </port>
2497 <port id="1" precision="FP32">
2498 <dim>1</dim>
2499 <dim>1</dim>
2500 <dim>768</dim>
2501 </port>
2502 </input>
2503 <output>
2504 <port id="2" precision="FP32" names="219,input.9">
2505 <dim>-1</dim>
2506 <dim>-1</dim>
2507 <dim>768</dim>
2508 </port>
2509 </output>
2510 </layer>
2511 <layer id="166" name="__module.roberta.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
2512 <data auto_broadcast="numpy" />
2513 <input>
2514 <port id="0" precision="FP32">
2515 <dim>-1</dim>
2516 <dim>-1</dim>
2517 <dim>768</dim>
2518 </port>
2519 <port id="1" precision="FP32">
2520 <dim>-1</dim>
2521 <dim>-1</dim>
2522 <dim>768</dim>
2523 </port>
2524 </input>
2525 <output>
2526 <port id="2" precision="FP32" names="221">
2527 <dim>-1</dim>
2528 <dim>-1</dim>
2529 <dim>768</dim>
2530 </port>
2531 </output>
2532 </layer>
2533 <layer id="167" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2534 <data element_type="i32" shape="1" offset="156000292" size="4" />
2535 <output>
2536 <port id="0" precision="I32">
2537 <dim>1</dim>
2538 </port>
2539 </output>
2540 </layer>
2541 <layer id="168" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2542 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2543 <input>
2544 <port id="0" precision="FP32">
2545 <dim>-1</dim>
2546 <dim>-1</dim>
2547 <dim>768</dim>
2548 </port>
2549 <port id="1" precision="I32">
2550 <dim>1</dim>
2551 </port>
2552 </input>
2553 <output>
2554 <port id="2" precision="FP32">
2555 <dim>-1</dim>
2556 <dim>-1</dim>
2557 <dim>768</dim>
2558 </port>
2559 </output>
2560 </layer>
2561 <layer id="169" name="Constant_5380" type="Const" version="opset1">
2562 <data element_type="f32" shape="1, 1, 768" offset="212703408" size="3072" />
2563 <output>
2564 <port id="0" precision="FP32">
2565 <dim>1</dim>
2566 <dim>1</dim>
2567 <dim>768</dim>
2568 </port>
2569 </output>
2570 </layer>
2571 <layer id="170" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2572 <data auto_broadcast="numpy" />
2573 <input>
2574 <port id="0" precision="FP32">
2575 <dim>-1</dim>
2576 <dim>-1</dim>
2577 <dim>768</dim>
2578 </port>
2579 <port id="1" precision="FP32">
2580 <dim>1</dim>
2581 <dim>1</dim>
2582 <dim>768</dim>
2583 </port>
2584 </input>
2585 <output>
2586 <port id="2" precision="FP32">
2587 <dim>-1</dim>
2588 <dim>-1</dim>
2589 <dim>768</dim>
2590 </port>
2591 </output>
2592 </layer>
2593 <layer id="171" name="Constant_5381" type="Const" version="opset1">
2594 <data element_type="f32" shape="1, 1, 768" offset="212706480" size="3072" />
2595 <output>
2596 <port id="0" precision="FP32">
2597 <dim>1</dim>
2598 <dim>1</dim>
2599 <dim>768</dim>
2600 </port>
2601 </output>
2602 </layer>
2603 <layer id="172" name="__module.roberta.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2604 <data auto_broadcast="numpy" />
2605 <input>
2606 <port id="0" precision="FP32">
2607 <dim>-1</dim>
2608 <dim>-1</dim>
2609 <dim>768</dim>
2610 </port>
2611 <port id="1" precision="FP32">
2612 <dim>1</dim>
2613 <dim>1</dim>
2614 <dim>768</dim>
2615 </port>
2616 </input>
2617 <output>
2618 <port id="2" precision="FP32" names="225,hidden_states.13">
2619 <dim>-1</dim>
2620 <dim>-1</dim>
2621 <dim>768</dim>
2622 </port>
2623 </output>
2624 </layer>
2625 <layer id="173" name="self.roberta.encoder.layer.2.attention.self.query.weight" type="Const" version="opset1">
2626 <data element_type="f32" shape="768, 768" offset="212709552" size="2359296" />
2627 <output>
2628 <port id="0" precision="FP32" names="self.roberta.encoder.layer.2.attention.self.query.weight">
2629 <dim>768</dim>
2630 <dim>768</dim>
2631 </port>
2632 </output>
2633 </layer>
2634 <layer id="174" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2635 <data transpose_a="false" transpose_b="true" />
2636 <input>
2637 <port id="0" precision="FP32">
2638 <dim>-1</dim>
2639 <dim>-1</dim>
2640 <dim>768</dim>
2641 </port>
2642 <port id="1" precision="FP32">
2643 <dim>768</dim>
2644 <dim>768</dim>
2645 </port>
2646 </input>
2647 <output>
2648 <port id="2" precision="FP32">
2649 <dim>-1</dim>
2650 <dim>-1</dim>
2651 <dim>768</dim>
2652 </port>
2653 </output>
2654 </layer>
2655 <layer id="175" name="Constant_5382" type="Const" version="opset1">
2656 <data element_type="f32" shape="1, 1, 768" offset="215068848" size="3072" />
2657 <output>
2658 <port id="0" precision="FP32">
2659 <dim>1</dim>
2660 <dim>1</dim>
2661 <dim>768</dim>
2662 </port>
2663 </output>
2664 </layer>
2665 <layer id="176" name="__module.roberta.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2666 <data auto_broadcast="numpy" />
2667 <input>
2668 <port id="0" precision="FP32">
2669 <dim>-1</dim>
2670 <dim>-1</dim>
2671 <dim>768</dim>
2672 </port>
2673 <port id="1" precision="FP32">
2674 <dim>1</dim>
2675 <dim>1</dim>
2676 <dim>768</dim>
2677 </port>
2678 </input>
2679 <output>
2680 <port id="2" precision="FP32" names="238,x.25">
2681 <dim>-1</dim>
2682 <dim>-1</dim>
2683 <dim>768</dim>
2684 </port>
2685 </output>
2686 </layer>
2687 <layer id="177" name="__module.roberta.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2688 <data element_type="i64" shape="4" offset="158368808" size="32" />
2689 <output>
2690 <port id="0" precision="I64">
2691 <dim>4</dim>
2692 </port>
2693 </output>
2694 </layer>
2695 <layer id="178" name="__module.roberta.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2696 <data special_zero="true" />
2697 <input>
2698 <port id="0" precision="FP32">
2699 <dim>-1</dim>
2700 <dim>-1</dim>
2701 <dim>768</dim>
2702 </port>
2703 <port id="1" precision="I64">
2704 <dim>4</dim>
2705 </port>
2706 </input>
2707 <output>
2708 <port id="2" precision="FP32" names="242,x.27">
2709 <dim>-1</dim>
2710 <dim>-1</dim>
2711 <dim>12</dim>
2712 <dim>64</dim>
2713 </port>
2714 </output>
2715 </layer>
2716 <layer id="179" name="Constant_675" type="Const" version="opset1">
2717 <data element_type="i64" shape="4" offset="158368840" size="32" />
2718 <output>
2719 <port id="0" precision="I64" names="243">
2720 <dim>4</dim>
2721 </port>
2722 </output>
2723 </layer>
2724 <layer id="180" name="__module.roberta.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2725 <input>
2726 <port id="0" precision="FP32">
2727 <dim>-1</dim>
2728 <dim>-1</dim>
2729 <dim>12</dim>
2730 <dim>64</dim>
2731 </port>
2732 <port id="1" precision="I64">
2733 <dim>4</dim>
2734 </port>
2735 </input>
2736 <output>
2737 <port id="2" precision="FP32" names="244">
2738 <dim>-1</dim>
2739 <dim>12</dim>
2740 <dim>-1</dim>
2741 <dim>64</dim>
2742 </port>
2743 </output>
2744 </layer>
2745 <layer id="181" name="self.roberta.encoder.layer.2.attention.self.key.weight" type="Const" version="opset1">
2746 <data element_type="f32" shape="768, 768" offset="215071920" size="2359296" />
2747 <output>
2748 <port id="0" precision="FP32" names="self.roberta.encoder.layer.2.attention.self.key.weight">
2749 <dim>768</dim>
2750 <dim>768</dim>
2751 </port>
2752 </output>
2753 </layer>
2754 <layer id="182" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2755 <data transpose_a="false" transpose_b="true" />
2756 <input>
2757 <port id="0" precision="FP32">
2758 <dim>-1</dim>
2759 <dim>-1</dim>
2760 <dim>768</dim>
2761 </port>
2762 <port id="1" precision="FP32">
2763 <dim>768</dim>
2764 <dim>768</dim>
2765 </port>
2766 </input>
2767 <output>
2768 <port id="2" precision="FP32">
2769 <dim>-1</dim>
2770 <dim>-1</dim>
2771 <dim>768</dim>
2772 </port>
2773 </output>
2774 </layer>
2775 <layer id="183" name="Constant_5383" type="Const" version="opset1">
2776 <data element_type="f32" shape="1, 1, 768" offset="217431216" size="3072" />
2777 <output>
2778 <port id="0" precision="FP32">
2779 <dim>1</dim>
2780 <dim>1</dim>
2781 <dim>768</dim>
2782 </port>
2783 </output>
2784 </layer>
2785 <layer id="184" name="__module.roberta.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2786 <data auto_broadcast="numpy" />
2787 <input>
2788 <port id="0" precision="FP32">
2789 <dim>-1</dim>
2790 <dim>-1</dim>
2791 <dim>768</dim>
2792 </port>
2793 <port id="1" precision="FP32">
2794 <dim>1</dim>
2795 <dim>1</dim>
2796 <dim>768</dim>
2797 </port>
2798 </input>
2799 <output>
2800 <port id="2" precision="FP32" names="247,x.29">
2801 <dim>-1</dim>
2802 <dim>-1</dim>
2803 <dim>768</dim>
2804 </port>
2805 </output>
2806 </layer>
2807 <layer id="185" name="__module.roberta.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2808 <data element_type="i64" shape="4" offset="158368808" size="32" />
2809 <output>
2810 <port id="0" precision="I64">
2811 <dim>4</dim>
2812 </port>
2813 </output>
2814 </layer>
2815 <layer id="186" name="__module.roberta.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2816 <data special_zero="true" />
2817 <input>
2818 <port id="0" precision="FP32">
2819 <dim>-1</dim>
2820 <dim>-1</dim>
2821 <dim>768</dim>
2822 </port>
2823 <port id="1" precision="I64">
2824 <dim>4</dim>
2825 </port>
2826 </input>
2827 <output>
2828 <port id="2" precision="FP32" names="251,x.31">
2829 <dim>-1</dim>
2830 <dim>-1</dim>
2831 <dim>12</dim>
2832 <dim>64</dim>
2833 </port>
2834 </output>
2835 </layer>
2836 <layer id="187" name="Constant_698" type="Const" version="opset1">
2837 <data element_type="i64" shape="4" offset="158368840" size="32" />
2838 <output>
2839 <port id="0" precision="I64" names="252">
2840 <dim>4</dim>
2841 </port>
2842 </output>
2843 </layer>
2844 <layer id="188" name="__module.roberta.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2845 <input>
2846 <port id="0" precision="FP32">
2847 <dim>-1</dim>
2848 <dim>-1</dim>
2849 <dim>12</dim>
2850 <dim>64</dim>
2851 </port>
2852 <port id="1" precision="I64">
2853 <dim>4</dim>
2854 </port>
2855 </input>
2856 <output>
2857 <port id="2" precision="FP32" names="253">
2858 <dim>-1</dim>
2859 <dim>12</dim>
2860 <dim>-1</dim>
2861 <dim>64</dim>
2862 </port>
2863 </output>
2864 </layer>
2865 <layer id="189" name="self.roberta.encoder.layer.2.attention.self.value.weight" type="Const" version="opset1">
2866 <data element_type="f32" shape="768, 768" offset="217434288" size="2359296" />
2867 <output>
2868 <port id="0" precision="FP32" names="self.roberta.encoder.layer.2.attention.self.value.weight">
2869 <dim>768</dim>
2870 <dim>768</dim>
2871 </port>
2872 </output>
2873 </layer>
2874 <layer id="190" name="__module.roberta.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
2875 <data transpose_a="false" transpose_b="true" />
2876 <input>
2877 <port id="0" precision="FP32">
2878 <dim>-1</dim>
2879 <dim>-1</dim>
2880 <dim>768</dim>
2881 </port>
2882 <port id="1" precision="FP32">
2883 <dim>768</dim>
2884 <dim>768</dim>
2885 </port>
2886 </input>
2887 <output>
2888 <port id="2" precision="FP32">
2889 <dim>-1</dim>
2890 <dim>-1</dim>
2891 <dim>768</dim>
2892 </port>
2893 </output>
2894 </layer>
2895 <layer id="191" name="Constant_5384" type="Const" version="opset1">
2896 <data element_type="f32" shape="1, 1, 768" offset="219793584" size="3072" />
2897 <output>
2898 <port id="0" precision="FP32">
2899 <dim>1</dim>
2900 <dim>1</dim>
2901 <dim>768</dim>
2902 </port>
2903 </output>
2904 </layer>
2905 <layer id="192" name="__module.roberta.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
2906 <data auto_broadcast="numpy" />
2907 <input>
2908 <port id="0" precision="FP32">
2909 <dim>-1</dim>
2910 <dim>-1</dim>
2911 <dim>768</dim>
2912 </port>
2913 <port id="1" precision="FP32">
2914 <dim>1</dim>
2915 <dim>1</dim>
2916 <dim>768</dim>
2917 </port>
2918 </input>
2919 <output>
2920 <port id="2" precision="FP32" names="256,x.33">
2921 <dim>-1</dim>
2922 <dim>-1</dim>
2923 <dim>768</dim>
2924 </port>
2925 </output>
2926 </layer>
2927 <layer id="193" name="__module.roberta.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
2928 <data element_type="i64" shape="4" offset="158368808" size="32" />
2929 <output>
2930 <port id="0" precision="I64">
2931 <dim>4</dim>
2932 </port>
2933 </output>
2934 </layer>
2935 <layer id="194" name="__module.roberta.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
2936 <data special_zero="true" />
2937 <input>
2938 <port id="0" precision="FP32">
2939 <dim>-1</dim>
2940 <dim>-1</dim>
2941 <dim>768</dim>
2942 </port>
2943 <port id="1" precision="I64">
2944 <dim>4</dim>
2945 </port>
2946 </input>
2947 <output>
2948 <port id="2" precision="FP32" names="260,x.35">
2949 <dim>-1</dim>
2950 <dim>-1</dim>
2951 <dim>12</dim>
2952 <dim>64</dim>
2953 </port>
2954 </output>
2955 </layer>
2956 <layer id="195" name="Constant_721" type="Const" version="opset1">
2957 <data element_type="i64" shape="4" offset="158368840" size="32" />
2958 <output>
2959 <port id="0" precision="I64" names="261">
2960 <dim>4</dim>
2961 </port>
2962 </output>
2963 </layer>
2964 <layer id="196" name="__module.roberta.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
2965 <input>
2966 <port id="0" precision="FP32">
2967 <dim>-1</dim>
2968 <dim>-1</dim>
2969 <dim>12</dim>
2970 <dim>64</dim>
2971 </port>
2972 <port id="1" precision="I64">
2973 <dim>4</dim>
2974 </port>
2975 </input>
2976 <output>
2977 <port id="2" precision="FP32" names="262">
2978 <dim>-1</dim>
2979 <dim>12</dim>
2980 <dim>-1</dim>
2981 <dim>64</dim>
2982 </port>
2983 </output>
2984 </layer>
2985 <layer id="197" name="__module.roberta.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
2986 <data causal="false" />
2987 <input>
2988 <port id="0" precision="FP32">
2989 <dim>-1</dim>
2990 <dim>12</dim>
2991 <dim>-1</dim>
2992 <dim>64</dim>
2993 </port>
2994 <port id="1" precision="FP32">
2995 <dim>-1</dim>
2996 <dim>12</dim>
2997 <dim>-1</dim>
2998 <dim>64</dim>
2999 </port>
3000 <port id="2" precision="FP32">
3001 <dim>-1</dim>
3002 <dim>12</dim>
3003 <dim>-1</dim>
3004 <dim>64</dim>
3005 </port>
3006 <port id="3" precision="FP32">
3007 <dim>-1</dim>
3008 <dim>1</dim>
3009 <dim>-1</dim>
3010 <dim>-1</dim>
3011 </port>
3012 </input>
3013 <output>
3014 <port id="4" precision="FP32" names="263,attn_output.9">
3015 <dim>-1</dim>
3016 <dim>12</dim>
3017 <dim>-1</dim>
3018 <dim>64</dim>
3019 </port>
3020 </output>
3021 </layer>
3022 <layer id="198" name="__module.roberta.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3023 <data element_type="i32" shape="4" offset="163093640" size="16" />
3024 <output>
3025 <port id="0" precision="I32">
3026 <dim>4</dim>
3027 </port>
3028 </output>
3029 </layer>
3030 <layer id="199" name="__module.roberta.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3031 <input>
3032 <port id="0" precision="FP32">
3033 <dim>-1</dim>
3034 <dim>12</dim>
3035 <dim>-1</dim>
3036 <dim>64</dim>
3037 </port>
3038 <port id="1" precision="I32">
3039 <dim>4</dim>
3040 </port>
3041 </input>
3042 <output>
3043 <port id="2" precision="FP32" names="264,attn_output.11">
3044 <dim>-1</dim>
3045 <dim>-1</dim>
3046 <dim>12</dim>
3047 <dim>64</dim>
3048 </port>
3049 </output>
3050 </layer>
3051 <layer id="200" name="Constant_5496" type="Const" version="opset1">
3052 <data element_type="i64" shape="3" offset="163093656" size="24" />
3053 <output>
3054 <port id="0" precision="I64">
3055 <dim>3</dim>
3056 </port>
3057 </output>
3058 </layer>
3059 <layer id="201" name="__module.roberta.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3060 <data special_zero="true" />
3061 <input>
3062 <port id="0" precision="FP32">
3063 <dim>-1</dim>
3064 <dim>-1</dim>
3065 <dim>12</dim>
3066 <dim>64</dim>
3067 </port>
3068 <port id="1" precision="I64">
3069 <dim>3</dim>
3070 </port>
3071 </input>
3072 <output>
3073 <port id="2" precision="FP32" names="266">
3074 <dim>-1</dim>
3075 <dim>-1</dim>
3076 <dim>768</dim>
3077 </port>
3078 </output>
3079 </layer>
3080 <layer id="202" name="self.roberta.encoder.layer.2.attention.output.dense.weight" type="Const" version="opset1">
3081 <data element_type="f32" shape="768, 768" offset="219796656" size="2359296" />
3082 <output>
3083 <port id="0" precision="FP32" names="self.roberta.encoder.layer.2.attention.output.dense.weight">
3084 <dim>768</dim>
3085 <dim>768</dim>
3086 </port>
3087 </output>
3088 </layer>
3089 <layer id="203" name="__module.roberta.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3090 <data transpose_a="false" transpose_b="true" />
3091 <input>
3092 <port id="0" precision="FP32">
3093 <dim>-1</dim>
3094 <dim>-1</dim>
3095 <dim>768</dim>
3096 </port>
3097 <port id="1" precision="FP32">
3098 <dim>768</dim>
3099 <dim>768</dim>
3100 </port>
3101 </input>
3102 <output>
3103 <port id="2" precision="FP32">
3104 <dim>-1</dim>
3105 <dim>-1</dim>
3106 <dim>768</dim>
3107 </port>
3108 </output>
3109 </layer>
3110 <layer id="204" name="Constant_5385" type="Const" version="opset1">
3111 <data element_type="f32" shape="1, 1, 768" offset="222155952" size="3072" />
3112 <output>
3113 <port id="0" precision="FP32">
3114 <dim>1</dim>
3115 <dim>1</dim>
3116 <dim>768</dim>
3117 </port>
3118 </output>
3119 </layer>
3120 <layer id="205" name="__module.roberta.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3121 <data auto_broadcast="numpy" />
3122 <input>
3123 <port id="0" precision="FP32">
3124 <dim>-1</dim>
3125 <dim>-1</dim>
3126 <dim>768</dim>
3127 </port>
3128 <port id="1" precision="FP32">
3129 <dim>1</dim>
3130 <dim>1</dim>
3131 <dim>768</dim>
3132 </port>
3133 </input>
3134 <output>
3135 <port id="2" precision="FP32" names="271,input.11">
3136 <dim>-1</dim>
3137 <dim>-1</dim>
3138 <dim>768</dim>
3139 </port>
3140 </output>
3141 </layer>
3142 <layer id="206" name="__module.roberta.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
3143 <data auto_broadcast="numpy" />
3144 <input>
3145 <port id="0" precision="FP32">
3146 <dim>-1</dim>
3147 <dim>-1</dim>
3148 <dim>768</dim>
3149 </port>
3150 <port id="1" precision="FP32">
3151 <dim>-1</dim>
3152 <dim>-1</dim>
3153 <dim>768</dim>
3154 </port>
3155 </input>
3156 <output>
3157 <port id="2" precision="FP32" names="273">
3158 <dim>-1</dim>
3159 <dim>-1</dim>
3160 <dim>768</dim>
3161 </port>
3162 </output>
3163 </layer>
3164 <layer id="207" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3165 <data element_type="i32" shape="1" offset="156000292" size="4" />
3166 <output>
3167 <port id="0" precision="I32">
3168 <dim>1</dim>
3169 </port>
3170 </output>
3171 </layer>
3172 <layer id="208" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3173 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3174 <input>
3175 <port id="0" precision="FP32">
3176 <dim>-1</dim>
3177 <dim>-1</dim>
3178 <dim>768</dim>
3179 </port>
3180 <port id="1" precision="I32">
3181 <dim>1</dim>
3182 </port>
3183 </input>
3184 <output>
3185 <port id="2" precision="FP32">
3186 <dim>-1</dim>
3187 <dim>-1</dim>
3188 <dim>768</dim>
3189 </port>
3190 </output>
3191 </layer>
3192 <layer id="209" name="Constant_5386" type="Const" version="opset1">
3193 <data element_type="f32" shape="1, 1, 768" offset="222159024" size="3072" />
3194 <output>
3195 <port id="0" precision="FP32">
3196 <dim>1</dim>
3197 <dim>1</dim>
3198 <dim>768</dim>
3199 </port>
3200 </output>
3201 </layer>
3202 <layer id="210" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3203 <data auto_broadcast="numpy" />
3204 <input>
3205 <port id="0" precision="FP32">
3206 <dim>-1</dim>
3207 <dim>-1</dim>
3208 <dim>768</dim>
3209 </port>
3210 <port id="1" precision="FP32">
3211 <dim>1</dim>
3212 <dim>1</dim>
3213 <dim>768</dim>
3214 </port>
3215 </input>
3216 <output>
3217 <port id="2" precision="FP32">
3218 <dim>-1</dim>
3219 <dim>-1</dim>
3220 <dim>768</dim>
3221 </port>
3222 </output>
3223 </layer>
3224 <layer id="211" name="Constant_5387" type="Const" version="opset1">
3225 <data element_type="f32" shape="1, 1, 768" offset="222162096" size="3072" />
3226 <output>
3227 <port id="0" precision="FP32">
3228 <dim>1</dim>
3229 <dim>1</dim>
3230 <dim>768</dim>
3231 </port>
3232 </output>
3233 </layer>
3234 <layer id="212" name="__module.roberta.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3235 <data auto_broadcast="numpy" />
3236 <input>
3237 <port id="0" precision="FP32">
3238 <dim>-1</dim>
3239 <dim>-1</dim>
3240 <dim>768</dim>
3241 </port>
3242 <port id="1" precision="FP32">
3243 <dim>1</dim>
3244 <dim>1</dim>
3245 <dim>768</dim>
3246 </port>
3247 </input>
3248 <output>
3249 <port id="2" precision="FP32" names="277,input_tensor.5">
3250 <dim>-1</dim>
3251 <dim>-1</dim>
3252 <dim>768</dim>
3253 </port>
3254 </output>
3255 </layer>
3256 <layer id="213" name="self.roberta.encoder.layer.2.intermediate.dense.weight" type="Const" version="opset1">
3257 <data element_type="f32" shape="3072, 768" offset="222165168" size="9437184" />
3258 <output>
3259 <port id="0" precision="FP32" names="self.roberta.encoder.layer.2.intermediate.dense.weight">
3260 <dim>3072</dim>
3261 <dim>768</dim>
3262 </port>
3263 </output>
3264 </layer>
3265 <layer id="214" name="__module.roberta.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3266 <data transpose_a="false" transpose_b="true" />
3267 <input>
3268 <port id="0" precision="FP32">
3269 <dim>-1</dim>
3270 <dim>-1</dim>
3271 <dim>768</dim>
3272 </port>
3273 <port id="1" precision="FP32">
3274 <dim>3072</dim>
3275 <dim>768</dim>
3276 </port>
3277 </input>
3278 <output>
3279 <port id="2" precision="FP32">
3280 <dim>-1</dim>
3281 <dim>-1</dim>
3282 <dim>3072</dim>
3283 </port>
3284 </output>
3285 </layer>
3286 <layer id="215" name="Constant_5388" type="Const" version="opset1">
3287 <data element_type="f32" shape="1, 1, 3072" offset="231602352" size="12288" />
3288 <output>
3289 <port id="0" precision="FP32">
3290 <dim>1</dim>
3291 <dim>1</dim>
3292 <dim>3072</dim>
3293 </port>
3294 </output>
3295 </layer>
3296 <layer id="216" name="__module.roberta.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3297 <data auto_broadcast="numpy" />
3298 <input>
3299 <port id="0" precision="FP32">
3300 <dim>-1</dim>
3301 <dim>-1</dim>
3302 <dim>3072</dim>
3303 </port>
3304 <port id="1" precision="FP32">
3305 <dim>1</dim>
3306 <dim>1</dim>
3307 <dim>3072</dim>
3308 </port>
3309 </input>
3310 <output>
3311 <port id="2" precision="FP32" names="281">
3312 <dim>-1</dim>
3313 <dim>-1</dim>
3314 <dim>3072</dim>
3315 </port>
3316 </output>
3317 </layer>
3318 <layer id="217" name="__module.roberta.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3319 <data approximation_mode="ERF" />
3320 <input>
3321 <port id="0" precision="FP32">
3322 <dim>-1</dim>
3323 <dim>-1</dim>
3324 <dim>3072</dim>
3325 </port>
3326 </input>
3327 <output>
3328 <port id="1" precision="FP32" names="282">
3329 <dim>-1</dim>
3330 <dim>-1</dim>
3331 <dim>3072</dim>
3332 </port>
3333 </output>
3334 </layer>
3335 <layer id="218" name="self.roberta.encoder.layer.2.output.dense.weight" type="Const" version="opset1">
3336 <data element_type="f32" shape="768, 3072" offset="231614640" size="9437184" />
3337 <output>
3338 <port id="0" precision="FP32" names="self.roberta.encoder.layer.2.output.dense.weight">
3339 <dim>768</dim>
3340 <dim>3072</dim>
3341 </port>
3342 </output>
3343 </layer>
3344 <layer id="219" name="__module.roberta.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3345 <data transpose_a="false" transpose_b="true" />
3346 <input>
3347 <port id="0" precision="FP32">
3348 <dim>-1</dim>
3349 <dim>-1</dim>
3350 <dim>3072</dim>
3351 </port>
3352 <port id="1" precision="FP32">
3353 <dim>768</dim>
3354 <dim>3072</dim>
3355 </port>
3356 </input>
3357 <output>
3358 <port id="2" precision="FP32">
3359 <dim>-1</dim>
3360 <dim>-1</dim>
3361 <dim>768</dim>
3362 </port>
3363 </output>
3364 </layer>
3365 <layer id="220" name="Constant_5389" type="Const" version="opset1">
3366 <data element_type="f32" shape="1, 1, 768" offset="241051824" size="3072" />
3367 <output>
3368 <port id="0" precision="FP32">
3369 <dim>1</dim>
3370 <dim>1</dim>
3371 <dim>768</dim>
3372 </port>
3373 </output>
3374 </layer>
3375 <layer id="221" name="__module.roberta.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
3376 <data auto_broadcast="numpy" />
3377 <input>
3378 <port id="0" precision="FP32">
3379 <dim>-1</dim>
3380 <dim>-1</dim>
3381 <dim>768</dim>
3382 </port>
3383 <port id="1" precision="FP32">
3384 <dim>1</dim>
3385 <dim>1</dim>
3386 <dim>768</dim>
3387 </port>
3388 </input>
3389 <output>
3390 <port id="2" precision="FP32" names="287,input.13">
3391 <dim>-1</dim>
3392 <dim>-1</dim>
3393 <dim>768</dim>
3394 </port>
3395 </output>
3396 </layer>
3397 <layer id="222" name="__module.roberta.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
3398 <data auto_broadcast="numpy" />
3399 <input>
3400 <port id="0" precision="FP32">
3401 <dim>-1</dim>
3402 <dim>-1</dim>
3403 <dim>768</dim>
3404 </port>
3405 <port id="1" precision="FP32">
3406 <dim>-1</dim>
3407 <dim>-1</dim>
3408 <dim>768</dim>
3409 </port>
3410 </input>
3411 <output>
3412 <port id="2" precision="FP32" names="289">
3413 <dim>-1</dim>
3414 <dim>-1</dim>
3415 <dim>768</dim>
3416 </port>
3417 </output>
3418 </layer>
3419 <layer id="223" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3420 <data element_type="i32" shape="1" offset="156000292" size="4" />
3421 <output>
3422 <port id="0" precision="I32">
3423 <dim>1</dim>
3424 </port>
3425 </output>
3426 </layer>
3427 <layer id="224" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3428 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3429 <input>
3430 <port id="0" precision="FP32">
3431 <dim>-1</dim>
3432 <dim>-1</dim>
3433 <dim>768</dim>
3434 </port>
3435 <port id="1" precision="I32">
3436 <dim>1</dim>
3437 </port>
3438 </input>
3439 <output>
3440 <port id="2" precision="FP32">
3441 <dim>-1</dim>
3442 <dim>-1</dim>
3443 <dim>768</dim>
3444 </port>
3445 </output>
3446 </layer>
3447 <layer id="225" name="Constant_5390" type="Const" version="opset1">
3448 <data element_type="f32" shape="1, 1, 768" offset="241054896" size="3072" />
3449 <output>
3450 <port id="0" precision="FP32">
3451 <dim>1</dim>
3452 <dim>1</dim>
3453 <dim>768</dim>
3454 </port>
3455 </output>
3456 </layer>
3457 <layer id="226" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3458 <data auto_broadcast="numpy" />
3459 <input>
3460 <port id="0" precision="FP32">
3461 <dim>-1</dim>
3462 <dim>-1</dim>
3463 <dim>768</dim>
3464 </port>
3465 <port id="1" precision="FP32">
3466 <dim>1</dim>
3467 <dim>1</dim>
3468 <dim>768</dim>
3469 </port>
3470 </input>
3471 <output>
3472 <port id="2" precision="FP32">
3473 <dim>-1</dim>
3474 <dim>-1</dim>
3475 <dim>768</dim>
3476 </port>
3477 </output>
3478 </layer>
3479 <layer id="227" name="Constant_5391" type="Const" version="opset1">
3480 <data element_type="f32" shape="1, 1, 768" offset="241057968" size="3072" />
3481 <output>
3482 <port id="0" precision="FP32">
3483 <dim>1</dim>
3484 <dim>1</dim>
3485 <dim>768</dim>
3486 </port>
3487 </output>
3488 </layer>
3489 <layer id="228" name="__module.roberta.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3490 <data auto_broadcast="numpy" />
3491 <input>
3492 <port id="0" precision="FP32">
3493 <dim>-1</dim>
3494 <dim>-1</dim>
3495 <dim>768</dim>
3496 </port>
3497 <port id="1" precision="FP32">
3498 <dim>1</dim>
3499 <dim>1</dim>
3500 <dim>768</dim>
3501 </port>
3502 </input>
3503 <output>
3504 <port id="2" precision="FP32" names="293,hidden_states.19">
3505 <dim>-1</dim>
3506 <dim>-1</dim>
3507 <dim>768</dim>
3508 </port>
3509 </output>
3510 </layer>
3511 <layer id="229" name="self.roberta.encoder.layer.3.attention.self.query.weight" type="Const" version="opset1">
3512 <data element_type="f32" shape="768, 768" offset="241061040" size="2359296" />
3513 <output>
3514 <port id="0" precision="FP32" names="self.roberta.encoder.layer.3.attention.self.query.weight">
3515 <dim>768</dim>
3516 <dim>768</dim>
3517 </port>
3518 </output>
3519 </layer>
3520 <layer id="230" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
3521 <data transpose_a="false" transpose_b="true" />
3522 <input>
3523 <port id="0" precision="FP32">
3524 <dim>-1</dim>
3525 <dim>-1</dim>
3526 <dim>768</dim>
3527 </port>
3528 <port id="1" precision="FP32">
3529 <dim>768</dim>
3530 <dim>768</dim>
3531 </port>
3532 </input>
3533 <output>
3534 <port id="2" precision="FP32">
3535 <dim>-1</dim>
3536 <dim>-1</dim>
3537 <dim>768</dim>
3538 </port>
3539 </output>
3540 </layer>
3541 <layer id="231" name="Constant_5392" type="Const" version="opset1">
3542 <data element_type="f32" shape="1, 1, 768" offset="243420336" size="3072" />
3543 <output>
3544 <port id="0" precision="FP32">
3545 <dim>1</dim>
3546 <dim>1</dim>
3547 <dim>768</dim>
3548 </port>
3549 </output>
3550 </layer>
3551 <layer id="232" name="__module.roberta.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
3552 <data auto_broadcast="numpy" />
3553 <input>
3554 <port id="0" precision="FP32">
3555 <dim>-1</dim>
3556 <dim>-1</dim>
3557 <dim>768</dim>
3558 </port>
3559 <port id="1" precision="FP32">
3560 <dim>1</dim>
3561 <dim>1</dim>
3562 <dim>768</dim>
3563 </port>
3564 </input>
3565 <output>
3566 <port id="2" precision="FP32" names="306,x.37">
3567 <dim>-1</dim>
3568 <dim>-1</dim>
3569 <dim>768</dim>
3570 </port>
3571 </output>
3572 </layer>
3573 <layer id="233" name="__module.roberta.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
3574 <data element_type="i64" shape="4" offset="158368808" size="32" />
3575 <output>
3576 <port id="0" precision="I64">
3577 <dim>4</dim>
3578 </port>
3579 </output>
3580 </layer>
3581 <layer id="234" name="__module.roberta.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
3582 <data special_zero="true" />
3583 <input>
3584 <port id="0" precision="FP32">
3585 <dim>-1</dim>
3586 <dim>-1</dim>
3587 <dim>768</dim>
3588 </port>
3589 <port id="1" precision="I64">
3590 <dim>4</dim>
3591 </port>
3592 </input>
3593 <output>
3594 <port id="2" precision="FP32" names="310,x.39">
3595 <dim>-1</dim>
3596 <dim>-1</dim>
3597 <dim>12</dim>
3598 <dim>64</dim>
3599 </port>
3600 </output>
3601 </layer>
3602 <layer id="235" name="Constant_898" type="Const" version="opset1">
3603 <data element_type="i64" shape="4" offset="158368840" size="32" />
3604 <output>
3605 <port id="0" precision="I64" names="311">
3606 <dim>4</dim>
3607 </port>
3608 </output>
3609 </layer>
3610 <layer id="236" name="__module.roberta.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
3611 <input>
3612 <port id="0" precision="FP32">
3613 <dim>-1</dim>
3614 <dim>-1</dim>
3615 <dim>12</dim>
3616 <dim>64</dim>
3617 </port>
3618 <port id="1" precision="I64">
3619 <dim>4</dim>
3620 </port>
3621 </input>
3622 <output>
3623 <port id="2" precision="FP32" names="312">
3624 <dim>-1</dim>
3625 <dim>12</dim>
3626 <dim>-1</dim>
3627 <dim>64</dim>
3628 </port>
3629 </output>
3630 </layer>
3631 <layer id="237" name="self.roberta.encoder.layer.3.attention.self.key.weight" type="Const" version="opset1">
3632 <data element_type="f32" shape="768, 768" offset="243423408" size="2359296" />
3633 <output>
3634 <port id="0" precision="FP32" names="self.roberta.encoder.layer.3.attention.self.key.weight">
3635 <dim>768</dim>
3636 <dim>768</dim>
3637 </port>
3638 </output>
3639 </layer>
3640 <layer id="238" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
3641 <data transpose_a="false" transpose_b="true" />
3642 <input>
3643 <port id="0" precision="FP32">
3644 <dim>-1</dim>
3645 <dim>-1</dim>
3646 <dim>768</dim>
3647 </port>
3648 <port id="1" precision="FP32">
3649 <dim>768</dim>
3650 <dim>768</dim>
3651 </port>
3652 </input>
3653 <output>
3654 <port id="2" precision="FP32">
3655 <dim>-1</dim>
3656 <dim>-1</dim>
3657 <dim>768</dim>
3658 </port>
3659 </output>
3660 </layer>
3661 <layer id="239" name="Constant_5393" type="Const" version="opset1">
3662 <data element_type="f32" shape="1, 1, 768" offset="245782704" size="3072" />
3663 <output>
3664 <port id="0" precision="FP32">
3665 <dim>1</dim>
3666 <dim>1</dim>
3667 <dim>768</dim>
3668 </port>
3669 </output>
3670 </layer>
3671 <layer id="240" name="__module.roberta.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
3672 <data auto_broadcast="numpy" />
3673 <input>
3674 <port id="0" precision="FP32">
3675 <dim>-1</dim>
3676 <dim>-1</dim>
3677 <dim>768</dim>
3678 </port>
3679 <port id="1" precision="FP32">
3680 <dim>1</dim>
3681 <dim>1</dim>
3682 <dim>768</dim>
3683 </port>
3684 </input>
3685 <output>
3686 <port id="2" precision="FP32" names="315,x.41">
3687 <dim>-1</dim>
3688 <dim>-1</dim>
3689 <dim>768</dim>
3690 </port>
3691 </output>
3692 </layer>
3693 <layer id="241" name="__module.roberta.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
3694 <data element_type="i64" shape="4" offset="158368808" size="32" />
3695 <output>
3696 <port id="0" precision="I64">
3697 <dim>4</dim>
3698 </port>
3699 </output>
3700 </layer>
3701 <layer id="242" name="__module.roberta.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
3702 <data special_zero="true" />
3703 <input>
3704 <port id="0" precision="FP32">
3705 <dim>-1</dim>
3706 <dim>-1</dim>
3707 <dim>768</dim>
3708 </port>
3709 <port id="1" precision="I64">
3710 <dim>4</dim>
3711 </port>
3712 </input>
3713 <output>
3714 <port id="2" precision="FP32" names="319,x.43">
3715 <dim>-1</dim>
3716 <dim>-1</dim>
3717 <dim>12</dim>
3718 <dim>64</dim>
3719 </port>
3720 </output>
3721 </layer>
3722 <layer id="243" name="Constant_921" type="Const" version="opset1">
3723 <data element_type="i64" shape="4" offset="158368840" size="32" />
3724 <output>
3725 <port id="0" precision="I64" names="320">
3726 <dim>4</dim>
3727 </port>
3728 </output>
3729 </layer>
3730 <layer id="244" name="__module.roberta.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
3731 <input>
3732 <port id="0" precision="FP32">
3733 <dim>-1</dim>
3734 <dim>-1</dim>
3735 <dim>12</dim>
3736 <dim>64</dim>
3737 </port>
3738 <port id="1" precision="I64">
3739 <dim>4</dim>
3740 </port>
3741 </input>
3742 <output>
3743 <port id="2" precision="FP32" names="321">
3744 <dim>-1</dim>
3745 <dim>12</dim>
3746 <dim>-1</dim>
3747 <dim>64</dim>
3748 </port>
3749 </output>
3750 </layer>
3751 <layer id="245" name="self.roberta.encoder.layer.3.attention.self.value.weight" type="Const" version="opset1">
3752 <data element_type="f32" shape="768, 768" offset="245785776" size="2359296" />
3753 <output>
3754 <port id="0" precision="FP32" names="self.roberta.encoder.layer.3.attention.self.value.weight">
3755 <dim>768</dim>
3756 <dim>768</dim>
3757 </port>
3758 </output>
3759 </layer>
3760 <layer id="246" name="__module.roberta.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
3761 <data transpose_a="false" transpose_b="true" />
3762 <input>
3763 <port id="0" precision="FP32">
3764 <dim>-1</dim>
3765 <dim>-1</dim>
3766 <dim>768</dim>
3767 </port>
3768 <port id="1" precision="FP32">
3769 <dim>768</dim>
3770 <dim>768</dim>
3771 </port>
3772 </input>
3773 <output>
3774 <port id="2" precision="FP32">
3775 <dim>-1</dim>
3776 <dim>-1</dim>
3777 <dim>768</dim>
3778 </port>
3779 </output>
3780 </layer>
3781 <layer id="247" name="Constant_5394" type="Const" version="opset1">
3782 <data element_type="f32" shape="1, 1, 768" offset="248145072" size="3072" />
3783 <output>
3784 <port id="0" precision="FP32">
3785 <dim>1</dim>
3786 <dim>1</dim>
3787 <dim>768</dim>
3788 </port>
3789 </output>
3790 </layer>
3791 <layer id="248" name="__module.roberta.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
3792 <data auto_broadcast="numpy" />
3793 <input>
3794 <port id="0" precision="FP32">
3795 <dim>-1</dim>
3796 <dim>-1</dim>
3797 <dim>768</dim>
3798 </port>
3799 <port id="1" precision="FP32">
3800 <dim>1</dim>
3801 <dim>1</dim>
3802 <dim>768</dim>
3803 </port>
3804 </input>
3805 <output>
3806 <port id="2" precision="FP32" names="324,x.45">
3807 <dim>-1</dim>
3808 <dim>-1</dim>
3809 <dim>768</dim>
3810 </port>
3811 </output>
3812 </layer>
3813 <layer id="249" name="__module.roberta.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
3814 <data element_type="i64" shape="4" offset="158368808" size="32" />
3815 <output>
3816 <port id="0" precision="I64">
3817 <dim>4</dim>
3818 </port>
3819 </output>
3820 </layer>
3821 <layer id="250" name="__module.roberta.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3822 <data special_zero="true" />
3823 <input>
3824 <port id="0" precision="FP32">
3825 <dim>-1</dim>
3826 <dim>-1</dim>
3827 <dim>768</dim>
3828 </port>
3829 <port id="1" precision="I64">
3830 <dim>4</dim>
3831 </port>
3832 </input>
3833 <output>
3834 <port id="2" precision="FP32" names="328,x.47">
3835 <dim>-1</dim>
3836 <dim>-1</dim>
3837 <dim>12</dim>
3838 <dim>64</dim>
3839 </port>
3840 </output>
3841 </layer>
3842 <layer id="251" name="Constant_944" type="Const" version="opset1">
3843 <data element_type="i64" shape="4" offset="158368840" size="32" />
3844 <output>
3845 <port id="0" precision="I64" names="329">
3846 <dim>4</dim>
3847 </port>
3848 </output>
3849 </layer>
3850 <layer id="252" name="__module.roberta.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3851 <input>
3852 <port id="0" precision="FP32">
3853 <dim>-1</dim>
3854 <dim>-1</dim>
3855 <dim>12</dim>
3856 <dim>64</dim>
3857 </port>
3858 <port id="1" precision="I64">
3859 <dim>4</dim>
3860 </port>
3861 </input>
3862 <output>
3863 <port id="2" precision="FP32" names="330">
3864 <dim>-1</dim>
3865 <dim>12</dim>
3866 <dim>-1</dim>
3867 <dim>64</dim>
3868 </port>
3869 </output>
3870 </layer>
3871 <layer id="253" name="__module.roberta.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3872 <data causal="false" />
3873 <input>
3874 <port id="0" precision="FP32">
3875 <dim>-1</dim>
3876 <dim>12</dim>
3877 <dim>-1</dim>
3878 <dim>64</dim>
3879 </port>
3880 <port id="1" precision="FP32">
3881 <dim>-1</dim>
3882 <dim>12</dim>
3883 <dim>-1</dim>
3884 <dim>64</dim>
3885 </port>
3886 <port id="2" precision="FP32">
3887 <dim>-1</dim>
3888 <dim>12</dim>
3889 <dim>-1</dim>
3890 <dim>64</dim>
3891 </port>
3892 <port id="3" precision="FP32">
3893 <dim>-1</dim>
3894 <dim>1</dim>
3895 <dim>-1</dim>
3896 <dim>-1</dim>
3897 </port>
3898 </input>
3899 <output>
3900 <port id="4" precision="FP32" names="331,attn_output.13">
3901 <dim>-1</dim>
3902 <dim>12</dim>
3903 <dim>-1</dim>
3904 <dim>64</dim>
3905 </port>
3906 </output>
3907 </layer>
3908 <layer id="254" name="__module.roberta.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3909 <data element_type="i32" shape="4" offset="163093640" size="16" />
3910 <output>
3911 <port id="0" precision="I32">
3912 <dim>4</dim>
3913 </port>
3914 </output>
3915 </layer>
3916 <layer id="255" name="__module.roberta.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3917 <input>
3918 <port id="0" precision="FP32">
3919 <dim>-1</dim>
3920 <dim>12</dim>
3921 <dim>-1</dim>
3922 <dim>64</dim>
3923 </port>
3924 <port id="1" precision="I32">
3925 <dim>4</dim>
3926 </port>
3927 </input>
3928 <output>
3929 <port id="2" precision="FP32" names="332,attn_output.15">
3930 <dim>-1</dim>
3931 <dim>-1</dim>
3932 <dim>12</dim>
3933 <dim>64</dim>
3934 </port>
3935 </output>
3936 </layer>
3937 <layer id="256" name="Constant_5497" type="Const" version="opset1">
3938 <data element_type="i64" shape="3" offset="163093656" size="24" />
3939 <output>
3940 <port id="0" precision="I64">
3941 <dim>3</dim>
3942 </port>
3943 </output>
3944 </layer>
3945 <layer id="257" name="__module.roberta.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3946 <data special_zero="true" />
3947 <input>
3948 <port id="0" precision="FP32">
3949 <dim>-1</dim>
3950 <dim>-1</dim>
3951 <dim>12</dim>
3952 <dim>64</dim>
3953 </port>
3954 <port id="1" precision="I64">
3955 <dim>3</dim>
3956 </port>
3957 </input>
3958 <output>
3959 <port id="2" precision="FP32" names="334">
3960 <dim>-1</dim>
3961 <dim>-1</dim>
3962 <dim>768</dim>
3963 </port>
3964 </output>
3965 </layer>
3966 <layer id="258" name="self.roberta.encoder.layer.3.attention.output.dense.weight" type="Const" version="opset1">
3967 <data element_type="f32" shape="768, 768" offset="248148144" size="2359296" />
3968 <output>
3969 <port id="0" precision="FP32" names="self.roberta.encoder.layer.3.attention.output.dense.weight">
3970 <dim>768</dim>
3971 <dim>768</dim>
3972 </port>
3973 </output>
3974 </layer>
3975 <layer id="259" name="__module.roberta.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3976 <data transpose_a="false" transpose_b="true" />
3977 <input>
3978 <port id="0" precision="FP32">
3979 <dim>-1</dim>
3980 <dim>-1</dim>
3981 <dim>768</dim>
3982 </port>
3983 <port id="1" precision="FP32">
3984 <dim>768</dim>
3985 <dim>768</dim>
3986 </port>
3987 </input>
3988 <output>
3989 <port id="2" precision="FP32">
3990 <dim>-1</dim>
3991 <dim>-1</dim>
3992 <dim>768</dim>
3993 </port>
3994 </output>
3995 </layer>
3996 <layer id="260" name="Constant_5395" type="Const" version="opset1">
3997 <data element_type="f32" shape="1, 1, 768" offset="250507440" size="3072" />
3998 <output>
3999 <port id="0" precision="FP32">
4000 <dim>1</dim>
4001 <dim>1</dim>
4002 <dim>768</dim>
4003 </port>
4004 </output>
4005 </layer>
4006 <layer id="261" name="__module.roberta.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
4007 <data auto_broadcast="numpy" />
4008 <input>
4009 <port id="0" precision="FP32">
4010 <dim>-1</dim>
4011 <dim>-1</dim>
4012 <dim>768</dim>
4013 </port>
4014 <port id="1" precision="FP32">
4015 <dim>1</dim>
4016 <dim>1</dim>
4017 <dim>768</dim>
4018 </port>
4019 </input>
4020 <output>
4021 <port id="2" precision="FP32" names="339,input.15">
4022 <dim>-1</dim>
4023 <dim>-1</dim>
4024 <dim>768</dim>
4025 </port>
4026 </output>
4027 </layer>
4028 <layer id="262" name="__module.roberta.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
4029 <data auto_broadcast="numpy" />
4030 <input>
4031 <port id="0" precision="FP32">
4032 <dim>-1</dim>
4033 <dim>-1</dim>
4034 <dim>768</dim>
4035 </port>
4036 <port id="1" precision="FP32">
4037 <dim>-1</dim>
4038 <dim>-1</dim>
4039 <dim>768</dim>
4040 </port>
4041 </input>
4042 <output>
4043 <port id="2" precision="FP32" names="341">
4044 <dim>-1</dim>
4045 <dim>-1</dim>
4046 <dim>768</dim>
4047 </port>
4048 </output>
4049 </layer>
4050 <layer id="263" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4051 <data element_type="i32" shape="1" offset="156000292" size="4" />
4052 <output>
4053 <port id="0" precision="I32">
4054 <dim>1</dim>
4055 </port>
4056 </output>
4057 </layer>
4058 <layer id="264" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4059 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4060 <input>
4061 <port id="0" precision="FP32">
4062 <dim>-1</dim>
4063 <dim>-1</dim>
4064 <dim>768</dim>
4065 </port>
4066 <port id="1" precision="I32">
4067 <dim>1</dim>
4068 </port>
4069 </input>
4070 <output>
4071 <port id="2" precision="FP32">
4072 <dim>-1</dim>
4073 <dim>-1</dim>
4074 <dim>768</dim>
4075 </port>
4076 </output>
4077 </layer>
4078 <layer id="265" name="Constant_5396" type="Const" version="opset1">
4079 <data element_type="f32" shape="1, 1, 768" offset="250510512" size="3072" />
4080 <output>
4081 <port id="0" precision="FP32">
4082 <dim>1</dim>
4083 <dim>1</dim>
4084 <dim>768</dim>
4085 </port>
4086 </output>
4087 </layer>
4088 <layer id="266" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4089 <data auto_broadcast="numpy" />
4090 <input>
4091 <port id="0" precision="FP32">
4092 <dim>-1</dim>
4093 <dim>-1</dim>
4094 <dim>768</dim>
4095 </port>
4096 <port id="1" precision="FP32">
4097 <dim>1</dim>
4098 <dim>1</dim>
4099 <dim>768</dim>
4100 </port>
4101 </input>
4102 <output>
4103 <port id="2" precision="FP32">
4104 <dim>-1</dim>
4105 <dim>-1</dim>
4106 <dim>768</dim>
4107 </port>
4108 </output>
4109 </layer>
4110 <layer id="267" name="Constant_5397" type="Const" version="opset1">
4111 <data element_type="f32" shape="1, 1, 768" offset="250513584" size="3072" />
4112 <output>
4113 <port id="0" precision="FP32">
4114 <dim>1</dim>
4115 <dim>1</dim>
4116 <dim>768</dim>
4117 </port>
4118 </output>
4119 </layer>
4120 <layer id="268" name="__module.roberta.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4121 <data auto_broadcast="numpy" />
4122 <input>
4123 <port id="0" precision="FP32">
4124 <dim>-1</dim>
4125 <dim>-1</dim>
4126 <dim>768</dim>
4127 </port>
4128 <port id="1" precision="FP32">
4129 <dim>1</dim>
4130 <dim>1</dim>
4131 <dim>768</dim>
4132 </port>
4133 </input>
4134 <output>
4135 <port id="2" precision="FP32" names="345,input_tensor.7">
4136 <dim>-1</dim>
4137 <dim>-1</dim>
4138 <dim>768</dim>
4139 </port>
4140 </output>
4141 </layer>
4142 <layer id="269" name="self.roberta.encoder.layer.3.intermediate.dense.weight" type="Const" version="opset1">
4143 <data element_type="f32" shape="3072, 768" offset="250516656" size="9437184" />
4144 <output>
4145 <port id="0" precision="FP32" names="self.roberta.encoder.layer.3.intermediate.dense.weight">
4146 <dim>3072</dim>
4147 <dim>768</dim>
4148 </port>
4149 </output>
4150 </layer>
4151 <layer id="270" name="__module.roberta.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4152 <data transpose_a="false" transpose_b="true" />
4153 <input>
4154 <port id="0" precision="FP32">
4155 <dim>-1</dim>
4156 <dim>-1</dim>
4157 <dim>768</dim>
4158 </port>
4159 <port id="1" precision="FP32">
4160 <dim>3072</dim>
4161 <dim>768</dim>
4162 </port>
4163 </input>
4164 <output>
4165 <port id="2" precision="FP32">
4166 <dim>-1</dim>
4167 <dim>-1</dim>
4168 <dim>3072</dim>
4169 </port>
4170 </output>
4171 </layer>
4172 <layer id="271" name="Constant_5398" type="Const" version="opset1">
4173 <data element_type="f32" shape="1, 1, 3072" offset="259953840" size="12288" />
4174 <output>
4175 <port id="0" precision="FP32">
4176 <dim>1</dim>
4177 <dim>1</dim>
4178 <dim>3072</dim>
4179 </port>
4180 </output>
4181 </layer>
4182 <layer id="272" name="__module.roberta.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
4183 <data auto_broadcast="numpy" />
4184 <input>
4185 <port id="0" precision="FP32">
4186 <dim>-1</dim>
4187 <dim>-1</dim>
4188 <dim>3072</dim>
4189 </port>
4190 <port id="1" precision="FP32">
4191 <dim>1</dim>
4192 <dim>1</dim>
4193 <dim>3072</dim>
4194 </port>
4195 </input>
4196 <output>
4197 <port id="2" precision="FP32" names="349">
4198 <dim>-1</dim>
4199 <dim>-1</dim>
4200 <dim>3072</dim>
4201 </port>
4202 </output>
4203 </layer>
4204 <layer id="273" name="__module.roberta.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
4205 <data approximation_mode="ERF" />
4206 <input>
4207 <port id="0" precision="FP32">
4208 <dim>-1</dim>
4209 <dim>-1</dim>
4210 <dim>3072</dim>
4211 </port>
4212 </input>
4213 <output>
4214 <port id="1" precision="FP32" names="350">
4215 <dim>-1</dim>
4216 <dim>-1</dim>
4217 <dim>3072</dim>
4218 </port>
4219 </output>
4220 </layer>
4221 <layer id="274" name="self.roberta.encoder.layer.3.output.dense.weight" type="Const" version="opset1">
4222 <data element_type="f32" shape="768, 3072" offset="259966128" size="9437184" />
4223 <output>
4224 <port id="0" precision="FP32" names="self.roberta.encoder.layer.3.output.dense.weight">
4225 <dim>768</dim>
4226 <dim>3072</dim>
4227 </port>
4228 </output>
4229 </layer>
4230 <layer id="275" name="__module.roberta.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4231 <data transpose_a="false" transpose_b="true" />
4232 <input>
4233 <port id="0" precision="FP32">
4234 <dim>-1</dim>
4235 <dim>-1</dim>
4236 <dim>3072</dim>
4237 </port>
4238 <port id="1" precision="FP32">
4239 <dim>768</dim>
4240 <dim>3072</dim>
4241 </port>
4242 </input>
4243 <output>
4244 <port id="2" precision="FP32">
4245 <dim>-1</dim>
4246 <dim>-1</dim>
4247 <dim>768</dim>
4248 </port>
4249 </output>
4250 </layer>
4251 <layer id="276" name="Constant_5399" type="Const" version="opset1">
4252 <data element_type="f32" shape="1, 1, 768" offset="269403312" size="3072" />
4253 <output>
4254 <port id="0" precision="FP32">
4255 <dim>1</dim>
4256 <dim>1</dim>
4257 <dim>768</dim>
4258 </port>
4259 </output>
4260 </layer>
4261 <layer id="277" name="__module.roberta.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
4262 <data auto_broadcast="numpy" />
4263 <input>
4264 <port id="0" precision="FP32">
4265 <dim>-1</dim>
4266 <dim>-1</dim>
4267 <dim>768</dim>
4268 </port>
4269 <port id="1" precision="FP32">
4270 <dim>1</dim>
4271 <dim>1</dim>
4272 <dim>768</dim>
4273 </port>
4274 </input>
4275 <output>
4276 <port id="2" precision="FP32" names="355,input.17">
4277 <dim>-1</dim>
4278 <dim>-1</dim>
4279 <dim>768</dim>
4280 </port>
4281 </output>
4282 </layer>
4283 <layer id="278" name="__module.roberta.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
4284 <data auto_broadcast="numpy" />
4285 <input>
4286 <port id="0" precision="FP32">
4287 <dim>-1</dim>
4288 <dim>-1</dim>
4289 <dim>768</dim>
4290 </port>
4291 <port id="1" precision="FP32">
4292 <dim>-1</dim>
4293 <dim>-1</dim>
4294 <dim>768</dim>
4295 </port>
4296 </input>
4297 <output>
4298 <port id="2" precision="FP32" names="357">
4299 <dim>-1</dim>
4300 <dim>-1</dim>
4301 <dim>768</dim>
4302 </port>
4303 </output>
4304 </layer>
4305 <layer id="279" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4306 <data element_type="i32" shape="1" offset="156000292" size="4" />
4307 <output>
4308 <port id="0" precision="I32">
4309 <dim>1</dim>
4310 </port>
4311 </output>
4312 </layer>
4313 <layer id="280" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4314 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4315 <input>
4316 <port id="0" precision="FP32">
4317 <dim>-1</dim>
4318 <dim>-1</dim>
4319 <dim>768</dim>
4320 </port>
4321 <port id="1" precision="I32">
4322 <dim>1</dim>
4323 </port>
4324 </input>
4325 <output>
4326 <port id="2" precision="FP32">
4327 <dim>-1</dim>
4328 <dim>-1</dim>
4329 <dim>768</dim>
4330 </port>
4331 </output>
4332 </layer>
4333 <layer id="281" name="Constant_5400" type="Const" version="opset1">
4334 <data element_type="f32" shape="1, 1, 768" offset="269406384" size="3072" />
4335 <output>
4336 <port id="0" precision="FP32">
4337 <dim>1</dim>
4338 <dim>1</dim>
4339 <dim>768</dim>
4340 </port>
4341 </output>
4342 </layer>
4343 <layer id="282" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4344 <data auto_broadcast="numpy" />
4345 <input>
4346 <port id="0" precision="FP32">
4347 <dim>-1</dim>
4348 <dim>-1</dim>
4349 <dim>768</dim>
4350 </port>
4351 <port id="1" precision="FP32">
4352 <dim>1</dim>
4353 <dim>1</dim>
4354 <dim>768</dim>
4355 </port>
4356 </input>
4357 <output>
4358 <port id="2" precision="FP32">
4359 <dim>-1</dim>
4360 <dim>-1</dim>
4361 <dim>768</dim>
4362 </port>
4363 </output>
4364 </layer>
4365 <layer id="283" name="Constant_5401" type="Const" version="opset1">
4366 <data element_type="f32" shape="1, 1, 768" offset="269409456" size="3072" />
4367 <output>
4368 <port id="0" precision="FP32">
4369 <dim>1</dim>
4370 <dim>1</dim>
4371 <dim>768</dim>
4372 </port>
4373 </output>
4374 </layer>
4375 <layer id="284" name="__module.roberta.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4376 <data auto_broadcast="numpy" />
4377 <input>
4378 <port id="0" precision="FP32">
4379 <dim>-1</dim>
4380 <dim>-1</dim>
4381 <dim>768</dim>
4382 </port>
4383 <port id="1" precision="FP32">
4384 <dim>1</dim>
4385 <dim>1</dim>
4386 <dim>768</dim>
4387 </port>
4388 </input>
4389 <output>
4390 <port id="2" precision="FP32" names="361,hidden_states.25">
4391 <dim>-1</dim>
4392 <dim>-1</dim>
4393 <dim>768</dim>
4394 </port>
4395 </output>
4396 </layer>
4397 <layer id="285" name="self.roberta.encoder.layer.4.attention.self.query.weight" type="Const" version="opset1">
4398 <data element_type="f32" shape="768, 768" offset="269412528" size="2359296" />
4399 <output>
4400 <port id="0" precision="FP32" names="self.roberta.encoder.layer.4.attention.self.query.weight">
4401 <dim>768</dim>
4402 <dim>768</dim>
4403 </port>
4404 </output>
4405 </layer>
4406 <layer id="286" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
4407 <data transpose_a="false" transpose_b="true" />
4408 <input>
4409 <port id="0" precision="FP32">
4410 <dim>-1</dim>
4411 <dim>-1</dim>
4412 <dim>768</dim>
4413 </port>
4414 <port id="1" precision="FP32">
4415 <dim>768</dim>
4416 <dim>768</dim>
4417 </port>
4418 </input>
4419 <output>
4420 <port id="2" precision="FP32">
4421 <dim>-1</dim>
4422 <dim>-1</dim>
4423 <dim>768</dim>
4424 </port>
4425 </output>
4426 </layer>
4427 <layer id="287" name="Constant_5402" type="Const" version="opset1">
4428 <data element_type="f32" shape="1, 1, 768" offset="271771824" size="3072" />
4429 <output>
4430 <port id="0" precision="FP32">
4431 <dim>1</dim>
4432 <dim>1</dim>
4433 <dim>768</dim>
4434 </port>
4435 </output>
4436 </layer>
4437 <layer id="288" name="__module.roberta.encoder.layer.4.attention.self.query/aten::linear/Add" type="Add" version="opset1">
4438 <data auto_broadcast="numpy" />
4439 <input>
4440 <port id="0" precision="FP32">
4441 <dim>-1</dim>
4442 <dim>-1</dim>
4443 <dim>768</dim>
4444 </port>
4445 <port id="1" precision="FP32">
4446 <dim>1</dim>
4447 <dim>1</dim>
4448 <dim>768</dim>
4449 </port>
4450 </input>
4451 <output>
4452 <port id="2" precision="FP32" names="374,x.49">
4453 <dim>-1</dim>
4454 <dim>-1</dim>
4455 <dim>768</dim>
4456 </port>
4457 </output>
4458 </layer>
4459 <layer id="289" name="__module.roberta.encoder.layer.4.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
4460 <data element_type="i64" shape="4" offset="158368808" size="32" />
4461 <output>
4462 <port id="0" precision="I64">
4463 <dim>4</dim>
4464 </port>
4465 </output>
4466 </layer>
4467 <layer id="290" name="__module.roberta.encoder.layer.4.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
4468 <data special_zero="true" />
4469 <input>
4470 <port id="0" precision="FP32">
4471 <dim>-1</dim>
4472 <dim>-1</dim>
4473 <dim>768</dim>
4474 </port>
4475 <port id="1" precision="I64">
4476 <dim>4</dim>
4477 </port>
4478 </input>
4479 <output>
4480 <port id="2" precision="FP32" names="378,x.51">
4481 <dim>-1</dim>
4482 <dim>-1</dim>
4483 <dim>12</dim>
4484 <dim>64</dim>
4485 </port>
4486 </output>
4487 </layer>
4488 <layer id="291" name="Constant_1121" type="Const" version="opset1">
4489 <data element_type="i64" shape="4" offset="158368840" size="32" />
4490 <output>
4491 <port id="0" precision="I64" names="379">
4492 <dim>4</dim>
4493 </port>
4494 </output>
4495 </layer>
4496 <layer id="292" name="__module.roberta.encoder.layer.4.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
4497 <input>
4498 <port id="0" precision="FP32">
4499 <dim>-1</dim>
4500 <dim>-1</dim>
4501 <dim>12</dim>
4502 <dim>64</dim>
4503 </port>
4504 <port id="1" precision="I64">
4505 <dim>4</dim>
4506 </port>
4507 </input>
4508 <output>
4509 <port id="2" precision="FP32" names="380">
4510 <dim>-1</dim>
4511 <dim>12</dim>
4512 <dim>-1</dim>
4513 <dim>64</dim>
4514 </port>
4515 </output>
4516 </layer>
4517 <layer id="293" name="self.roberta.encoder.layer.4.attention.self.key.weight" type="Const" version="opset1">
4518 <data element_type="f32" shape="768, 768" offset="271774896" size="2359296" />
4519 <output>
4520 <port id="0" precision="FP32" names="self.roberta.encoder.layer.4.attention.self.key.weight">
4521 <dim>768</dim>
4522 <dim>768</dim>
4523 </port>
4524 </output>
4525 </layer>
4526 <layer id="294" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
4527 <data transpose_a="false" transpose_b="true" />
4528 <input>
4529 <port id="0" precision="FP32">
4530 <dim>-1</dim>
4531 <dim>-1</dim>
4532 <dim>768</dim>
4533 </port>
4534 <port id="1" precision="FP32">
4535 <dim>768</dim>
4536 <dim>768</dim>
4537 </port>
4538 </input>
4539 <output>
4540 <port id="2" precision="FP32">
4541 <dim>-1</dim>
4542 <dim>-1</dim>
4543 <dim>768</dim>
4544 </port>
4545 </output>
4546 </layer>
4547 <layer id="295" name="Constant_5403" type="Const" version="opset1">
4548 <data element_type="f32" shape="1, 1, 768" offset="274134192" size="3072" />
4549 <output>
4550 <port id="0" precision="FP32">
4551 <dim>1</dim>
4552 <dim>1</dim>
4553 <dim>768</dim>
4554 </port>
4555 </output>
4556 </layer>
4557 <layer id="296" name="__module.roberta.encoder.layer.4.attention.self.key/aten::linear/Add" type="Add" version="opset1">
4558 <data auto_broadcast="numpy" />
4559 <input>
4560 <port id="0" precision="FP32">
4561 <dim>-1</dim>
4562 <dim>-1</dim>
4563 <dim>768</dim>
4564 </port>
4565 <port id="1" precision="FP32">
4566 <dim>1</dim>
4567 <dim>1</dim>
4568 <dim>768</dim>
4569 </port>
4570 </input>
4571 <output>
4572 <port id="2" precision="FP32" names="383,x.53">
4573 <dim>-1</dim>
4574 <dim>-1</dim>
4575 <dim>768</dim>
4576 </port>
4577 </output>
4578 </layer>
4579 <layer id="297" name="__module.roberta.encoder.layer.4.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
4580 <data element_type="i64" shape="4" offset="158368808" size="32" />
4581 <output>
4582 <port id="0" precision="I64">
4583 <dim>4</dim>
4584 </port>
4585 </output>
4586 </layer>
4587 <layer id="298" name="__module.roberta.encoder.layer.4.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
4588 <data special_zero="true" />
4589 <input>
4590 <port id="0" precision="FP32">
4591 <dim>-1</dim>
4592 <dim>-1</dim>
4593 <dim>768</dim>
4594 </port>
4595 <port id="1" precision="I64">
4596 <dim>4</dim>
4597 </port>
4598 </input>
4599 <output>
4600 <port id="2" precision="FP32" names="387,x.55">
4601 <dim>-1</dim>
4602 <dim>-1</dim>
4603 <dim>12</dim>
4604 <dim>64</dim>
4605 </port>
4606 </output>
4607 </layer>
4608 <layer id="299" name="Constant_1144" type="Const" version="opset1">
4609 <data element_type="i64" shape="4" offset="158368840" size="32" />
4610 <output>
4611 <port id="0" precision="I64" names="388">
4612 <dim>4</dim>
4613 </port>
4614 </output>
4615 </layer>
4616 <layer id="300" name="__module.roberta.encoder.layer.4.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
4617 <input>
4618 <port id="0" precision="FP32">
4619 <dim>-1</dim>
4620 <dim>-1</dim>
4621 <dim>12</dim>
4622 <dim>64</dim>
4623 </port>
4624 <port id="1" precision="I64">
4625 <dim>4</dim>
4626 </port>
4627 </input>
4628 <output>
4629 <port id="2" precision="FP32" names="389">
4630 <dim>-1</dim>
4631 <dim>12</dim>
4632 <dim>-1</dim>
4633 <dim>64</dim>
4634 </port>
4635 </output>
4636 </layer>
4637 <layer id="301" name="self.roberta.encoder.layer.4.attention.self.value.weight" type="Const" version="opset1">
4638 <data element_type="f32" shape="768, 768" offset="274137264" size="2359296" />
4639 <output>
4640 <port id="0" precision="FP32" names="self.roberta.encoder.layer.4.attention.self.value.weight">
4641 <dim>768</dim>
4642 <dim>768</dim>
4643 </port>
4644 </output>
4645 </layer>
4646 <layer id="302" name="__module.roberta.encoder.layer.4.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
4647 <data transpose_a="false" transpose_b="true" />
4648 <input>
4649 <port id="0" precision="FP32">
4650 <dim>-1</dim>
4651 <dim>-1</dim>
4652 <dim>768</dim>
4653 </port>
4654 <port id="1" precision="FP32">
4655 <dim>768</dim>
4656 <dim>768</dim>
4657 </port>
4658 </input>
4659 <output>
4660 <port id="2" precision="FP32">
4661 <dim>-1</dim>
4662 <dim>-1</dim>
4663 <dim>768</dim>
4664 </port>
4665 </output>
4666 </layer>
4667 <layer id="303" name="Constant_5404" type="Const" version="opset1">
4668 <data element_type="f32" shape="1, 1, 768" offset="276496560" size="3072" />
4669 <output>
4670 <port id="0" precision="FP32">
4671 <dim>1</dim>
4672 <dim>1</dim>
4673 <dim>768</dim>
4674 </port>
4675 </output>
4676 </layer>
4677 <layer id="304" name="__module.roberta.encoder.layer.4.attention.self.value/aten::linear/Add" type="Add" version="opset1">
4678 <data auto_broadcast="numpy" />
4679 <input>
4680 <port id="0" precision="FP32">
4681 <dim>-1</dim>
4682 <dim>-1</dim>
4683 <dim>768</dim>
4684 </port>
4685 <port id="1" precision="FP32">
4686 <dim>1</dim>
4687 <dim>1</dim>
4688 <dim>768</dim>
4689 </port>
4690 </input>
4691 <output>
4692 <port id="2" precision="FP32" names="392,x.57">
4693 <dim>-1</dim>
4694 <dim>-1</dim>
4695 <dim>768</dim>
4696 </port>
4697 </output>
4698 </layer>
4699 <layer id="305" name="__module.roberta.encoder.layer.4.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
4700 <data element_type="i64" shape="4" offset="158368808" size="32" />
4701 <output>
4702 <port id="0" precision="I64">
4703 <dim>4</dim>
4704 </port>
4705 </output>
4706 </layer>
4707 <layer id="306" name="__module.roberta.encoder.layer.4.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
4708 <data special_zero="true" />
4709 <input>
4710 <port id="0" precision="FP32">
4711 <dim>-1</dim>
4712 <dim>-1</dim>
4713 <dim>768</dim>
4714 </port>
4715 <port id="1" precision="I64">
4716 <dim>4</dim>
4717 </port>
4718 </input>
4719 <output>
4720 <port id="2" precision="FP32" names="396,x.59">
4721 <dim>-1</dim>
4722 <dim>-1</dim>
4723 <dim>12</dim>
4724 <dim>64</dim>
4725 </port>
4726 </output>
4727 </layer>
4728 <layer id="307" name="Constant_1167" type="Const" version="opset1">
4729 <data element_type="i64" shape="4" offset="158368840" size="32" />
4730 <output>
4731 <port id="0" precision="I64" names="397">
4732 <dim>4</dim>
4733 </port>
4734 </output>
4735 </layer>
4736 <layer id="308" name="__module.roberta.encoder.layer.4.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
4737 <input>
4738 <port id="0" precision="FP32">
4739 <dim>-1</dim>
4740 <dim>-1</dim>
4741 <dim>12</dim>
4742 <dim>64</dim>
4743 </port>
4744 <port id="1" precision="I64">
4745 <dim>4</dim>
4746 </port>
4747 </input>
4748 <output>
4749 <port id="2" precision="FP32" names="398">
4750 <dim>-1</dim>
4751 <dim>12</dim>
4752 <dim>-1</dim>
4753 <dim>64</dim>
4754 </port>
4755 </output>
4756 </layer>
4757 <layer id="309" name="__module.roberta.encoder.layer.4.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
4758 <data causal="false" />
4759 <input>
4760 <port id="0" precision="FP32">
4761 <dim>-1</dim>
4762 <dim>12</dim>
4763 <dim>-1</dim>
4764 <dim>64</dim>
4765 </port>
4766 <port id="1" precision="FP32">
4767 <dim>-1</dim>
4768 <dim>12</dim>
4769 <dim>-1</dim>
4770 <dim>64</dim>
4771 </port>
4772 <port id="2" precision="FP32">
4773 <dim>-1</dim>
4774 <dim>12</dim>
4775 <dim>-1</dim>
4776 <dim>64</dim>
4777 </port>
4778 <port id="3" precision="FP32">
4779 <dim>-1</dim>
4780 <dim>1</dim>
4781 <dim>-1</dim>
4782 <dim>-1</dim>
4783 </port>
4784 </input>
4785 <output>
4786 <port id="4" precision="FP32" names="399,attn_output.17">
4787 <dim>-1</dim>
4788 <dim>12</dim>
4789 <dim>-1</dim>
4790 <dim>64</dim>
4791 </port>
4792 </output>
4793 </layer>
4794 <layer id="310" name="__module.roberta.encoder.layer.4.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
4795 <data element_type="i32" shape="4" offset="163093640" size="16" />
4796 <output>
4797 <port id="0" precision="I32">
4798 <dim>4</dim>
4799 </port>
4800 </output>
4801 </layer>
4802 <layer id="311" name="__module.roberta.encoder.layer.4.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
4803 <input>
4804 <port id="0" precision="FP32">
4805 <dim>-1</dim>
4806 <dim>12</dim>
4807 <dim>-1</dim>
4808 <dim>64</dim>
4809 </port>
4810 <port id="1" precision="I32">
4811 <dim>4</dim>
4812 </port>
4813 </input>
4814 <output>
4815 <port id="2" precision="FP32" names="400,attn_output.19">
4816 <dim>-1</dim>
4817 <dim>-1</dim>
4818 <dim>12</dim>
4819 <dim>64</dim>
4820 </port>
4821 </output>
4822 </layer>
4823 <layer id="312" name="Constant_5498" type="Const" version="opset1">
4824 <data element_type="i64" shape="3" offset="163093656" size="24" />
4825 <output>
4826 <port id="0" precision="I64">
4827 <dim>3</dim>
4828 </port>
4829 </output>
4830 </layer>
4831 <layer id="313" name="__module.roberta.encoder.layer.4.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
4832 <data special_zero="true" />
4833 <input>
4834 <port id="0" precision="FP32">
4835 <dim>-1</dim>
4836 <dim>-1</dim>
4837 <dim>12</dim>
4838 <dim>64</dim>
4839 </port>
4840 <port id="1" precision="I64">
4841 <dim>3</dim>
4842 </port>
4843 </input>
4844 <output>
4845 <port id="2" precision="FP32" names="402">
4846 <dim>-1</dim>
4847 <dim>-1</dim>
4848 <dim>768</dim>
4849 </port>
4850 </output>
4851 </layer>
4852 <layer id="314" name="self.roberta.encoder.layer.4.attention.output.dense.weight" type="Const" version="opset1">
4853 <data element_type="f32" shape="768, 768" offset="276499632" size="2359296" />
4854 <output>
4855 <port id="0" precision="FP32" names="self.roberta.encoder.layer.4.attention.output.dense.weight">
4856 <dim>768</dim>
4857 <dim>768</dim>
4858 </port>
4859 </output>
4860 </layer>
4861 <layer id="315" name="__module.roberta.encoder.layer.4.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4862 <data transpose_a="false" transpose_b="true" />
4863 <input>
4864 <port id="0" precision="FP32">
4865 <dim>-1</dim>
4866 <dim>-1</dim>
4867 <dim>768</dim>
4868 </port>
4869 <port id="1" precision="FP32">
4870 <dim>768</dim>
4871 <dim>768</dim>
4872 </port>
4873 </input>
4874 <output>
4875 <port id="2" precision="FP32">
4876 <dim>-1</dim>
4877 <dim>-1</dim>
4878 <dim>768</dim>
4879 </port>
4880 </output>
4881 </layer>
4882 <layer id="316" name="Constant_5405" type="Const" version="opset1">
4883 <data element_type="f32" shape="1, 1, 768" offset="278858928" size="3072" />
4884 <output>
4885 <port id="0" precision="FP32">
4886 <dim>1</dim>
4887 <dim>1</dim>
4888 <dim>768</dim>
4889 </port>
4890 </output>
4891 </layer>
4892 <layer id="317" name="__module.roberta.encoder.layer.4.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
4893 <data auto_broadcast="numpy" />
4894 <input>
4895 <port id="0" precision="FP32">
4896 <dim>-1</dim>
4897 <dim>-1</dim>
4898 <dim>768</dim>
4899 </port>
4900 <port id="1" precision="FP32">
4901 <dim>1</dim>
4902 <dim>1</dim>
4903 <dim>768</dim>
4904 </port>
4905 </input>
4906 <output>
4907 <port id="2" precision="FP32" names="407,input.19">
4908 <dim>-1</dim>
4909 <dim>-1</dim>
4910 <dim>768</dim>
4911 </port>
4912 </output>
4913 </layer>
4914 <layer id="318" name="__module.roberta.encoder.layer.4.attention.output/aten::add/Add" type="Add" version="opset1">
4915 <data auto_broadcast="numpy" />
4916 <input>
4917 <port id="0" precision="FP32">
4918 <dim>-1</dim>
4919 <dim>-1</dim>
4920 <dim>768</dim>
4921 </port>
4922 <port id="1" precision="FP32">
4923 <dim>-1</dim>
4924 <dim>-1</dim>
4925 <dim>768</dim>
4926 </port>
4927 </input>
4928 <output>
4929 <port id="2" precision="FP32" names="409">
4930 <dim>-1</dim>
4931 <dim>-1</dim>
4932 <dim>768</dim>
4933 </port>
4934 </output>
4935 </layer>
4936 <layer id="319" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4937 <data element_type="i32" shape="1" offset="156000292" size="4" />
4938 <output>
4939 <port id="0" precision="I32">
4940 <dim>1</dim>
4941 </port>
4942 </output>
4943 </layer>
4944 <layer id="320" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4945 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4946 <input>
4947 <port id="0" precision="FP32">
4948 <dim>-1</dim>
4949 <dim>-1</dim>
4950 <dim>768</dim>
4951 </port>
4952 <port id="1" precision="I32">
4953 <dim>1</dim>
4954 </port>
4955 </input>
4956 <output>
4957 <port id="2" precision="FP32">
4958 <dim>-1</dim>
4959 <dim>-1</dim>
4960 <dim>768</dim>
4961 </port>
4962 </output>
4963 </layer>
4964 <layer id="321" name="Constant_5406" type="Const" version="opset1">
4965 <data element_type="f32" shape="1, 1, 768" offset="278862000" size="3072" />
4966 <output>
4967 <port id="0" precision="FP32">
4968 <dim>1</dim>
4969 <dim>1</dim>
4970 <dim>768</dim>
4971 </port>
4972 </output>
4973 </layer>
4974 <layer id="322" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4975 <data auto_broadcast="numpy" />
4976 <input>
4977 <port id="0" precision="FP32">
4978 <dim>-1</dim>
4979 <dim>-1</dim>
4980 <dim>768</dim>
4981 </port>
4982 <port id="1" precision="FP32">
4983 <dim>1</dim>
4984 <dim>1</dim>
4985 <dim>768</dim>
4986 </port>
4987 </input>
4988 <output>
4989 <port id="2" precision="FP32">
4990 <dim>-1</dim>
4991 <dim>-1</dim>
4992 <dim>768</dim>
4993 </port>
4994 </output>
4995 </layer>
4996 <layer id="323" name="Constant_5407" type="Const" version="opset1">
4997 <data element_type="f32" shape="1, 1, 768" offset="278865072" size="3072" />
4998 <output>
4999 <port id="0" precision="FP32">
5000 <dim>1</dim>
5001 <dim>1</dim>
5002 <dim>768</dim>
5003 </port>
5004 </output>
5005 </layer>
5006 <layer id="324" name="__module.roberta.encoder.layer.4.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5007 <data auto_broadcast="numpy" />
5008 <input>
5009 <port id="0" precision="FP32">
5010 <dim>-1</dim>
5011 <dim>-1</dim>
5012 <dim>768</dim>
5013 </port>
5014 <port id="1" precision="FP32">
5015 <dim>1</dim>
5016 <dim>1</dim>
5017 <dim>768</dim>
5018 </port>
5019 </input>
5020 <output>
5021 <port id="2" precision="FP32" names="413,input_tensor.9">
5022 <dim>-1</dim>
5023 <dim>-1</dim>
5024 <dim>768</dim>
5025 </port>
5026 </output>
5027 </layer>
5028 <layer id="325" name="self.roberta.encoder.layer.4.intermediate.dense.weight" type="Const" version="opset1">
5029 <data element_type="f32" shape="3072, 768" offset="278868144" size="9437184" />
5030 <output>
5031 <port id="0" precision="FP32" names="self.roberta.encoder.layer.4.intermediate.dense.weight">
5032 <dim>3072</dim>
5033 <dim>768</dim>
5034 </port>
5035 </output>
5036 </layer>
5037 <layer id="326" name="__module.roberta.encoder.layer.4.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5038 <data transpose_a="false" transpose_b="true" />
5039 <input>
5040 <port id="0" precision="FP32">
5041 <dim>-1</dim>
5042 <dim>-1</dim>
5043 <dim>768</dim>
5044 </port>
5045 <port id="1" precision="FP32">
5046 <dim>3072</dim>
5047 <dim>768</dim>
5048 </port>
5049 </input>
5050 <output>
5051 <port id="2" precision="FP32">
5052 <dim>-1</dim>
5053 <dim>-1</dim>
5054 <dim>3072</dim>
5055 </port>
5056 </output>
5057 </layer>
5058 <layer id="327" name="Constant_5408" type="Const" version="opset1">
5059 <data element_type="f32" shape="1, 1, 3072" offset="288305328" size="12288" />
5060 <output>
5061 <port id="0" precision="FP32">
5062 <dim>1</dim>
5063 <dim>1</dim>
5064 <dim>3072</dim>
5065 </port>
5066 </output>
5067 </layer>
5068 <layer id="328" name="__module.roberta.encoder.layer.4.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5069 <data auto_broadcast="numpy" />
5070 <input>
5071 <port id="0" precision="FP32">
5072 <dim>-1</dim>
5073 <dim>-1</dim>
5074 <dim>3072</dim>
5075 </port>
5076 <port id="1" precision="FP32">
5077 <dim>1</dim>
5078 <dim>1</dim>
5079 <dim>3072</dim>
5080 </port>
5081 </input>
5082 <output>
5083 <port id="2" precision="FP32" names="417">
5084 <dim>-1</dim>
5085 <dim>-1</dim>
5086 <dim>3072</dim>
5087 </port>
5088 </output>
5089 </layer>
5090 <layer id="329" name="__module.roberta.encoder.layer.4.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5091 <data approximation_mode="ERF" />
5092 <input>
5093 <port id="0" precision="FP32">
5094 <dim>-1</dim>
5095 <dim>-1</dim>
5096 <dim>3072</dim>
5097 </port>
5098 </input>
5099 <output>
5100 <port id="1" precision="FP32" names="418">
5101 <dim>-1</dim>
5102 <dim>-1</dim>
5103 <dim>3072</dim>
5104 </port>
5105 </output>
5106 </layer>
5107 <layer id="330" name="self.roberta.encoder.layer.4.output.dense.weight" type="Const" version="opset1">
5108 <data element_type="f32" shape="768, 3072" offset="288317616" size="9437184" />
5109 <output>
5110 <port id="0" precision="FP32" names="self.roberta.encoder.layer.4.output.dense.weight">
5111 <dim>768</dim>
5112 <dim>3072</dim>
5113 </port>
5114 </output>
5115 </layer>
5116 <layer id="331" name="__module.roberta.encoder.layer.4.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5117 <data transpose_a="false" transpose_b="true" />
5118 <input>
5119 <port id="0" precision="FP32">
5120 <dim>-1</dim>
5121 <dim>-1</dim>
5122 <dim>3072</dim>
5123 </port>
5124 <port id="1" precision="FP32">
5125 <dim>768</dim>
5126 <dim>3072</dim>
5127 </port>
5128 </input>
5129 <output>
5130 <port id="2" precision="FP32">
5131 <dim>-1</dim>
5132 <dim>-1</dim>
5133 <dim>768</dim>
5134 </port>
5135 </output>
5136 </layer>
5137 <layer id="332" name="Constant_5409" type="Const" version="opset1">
5138 <data element_type="f32" shape="1, 1, 768" offset="297754800" size="3072" />
5139 <output>
5140 <port id="0" precision="FP32">
5141 <dim>1</dim>
5142 <dim>1</dim>
5143 <dim>768</dim>
5144 </port>
5145 </output>
5146 </layer>
5147 <layer id="333" name="__module.roberta.encoder.layer.4.output.dense/aten::linear/Add" type="Add" version="opset1">
5148 <data auto_broadcast="numpy" />
5149 <input>
5150 <port id="0" precision="FP32">
5151 <dim>-1</dim>
5152 <dim>-1</dim>
5153 <dim>768</dim>
5154 </port>
5155 <port id="1" precision="FP32">
5156 <dim>1</dim>
5157 <dim>1</dim>
5158 <dim>768</dim>
5159 </port>
5160 </input>
5161 <output>
5162 <port id="2" precision="FP32" names="423,input.21">
5163 <dim>-1</dim>
5164 <dim>-1</dim>
5165 <dim>768</dim>
5166 </port>
5167 </output>
5168 </layer>
5169 <layer id="334" name="__module.roberta.encoder.layer.4.output/aten::add/Add" type="Add" version="opset1">
5170 <data auto_broadcast="numpy" />
5171 <input>
5172 <port id="0" precision="FP32">
5173 <dim>-1</dim>
5174 <dim>-1</dim>
5175 <dim>768</dim>
5176 </port>
5177 <port id="1" precision="FP32">
5178 <dim>-1</dim>
5179 <dim>-1</dim>
5180 <dim>768</dim>
5181 </port>
5182 </input>
5183 <output>
5184 <port id="2" precision="FP32" names="425">
5185 <dim>-1</dim>
5186 <dim>-1</dim>
5187 <dim>768</dim>
5188 </port>
5189 </output>
5190 </layer>
5191 <layer id="335" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5192 <data element_type="i32" shape="1" offset="156000292" size="4" />
5193 <output>
5194 <port id="0" precision="I32">
5195 <dim>1</dim>
5196 </port>
5197 </output>
5198 </layer>
5199 <layer id="336" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5200 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5201 <input>
5202 <port id="0" precision="FP32">
5203 <dim>-1</dim>
5204 <dim>-1</dim>
5205 <dim>768</dim>
5206 </port>
5207 <port id="1" precision="I32">
5208 <dim>1</dim>
5209 </port>
5210 </input>
5211 <output>
5212 <port id="2" precision="FP32">
5213 <dim>-1</dim>
5214 <dim>-1</dim>
5215 <dim>768</dim>
5216 </port>
5217 </output>
5218 </layer>
5219 <layer id="337" name="Constant_5410" type="Const" version="opset1">
5220 <data element_type="f32" shape="1, 1, 768" offset="297757872" size="3072" />
5221 <output>
5222 <port id="0" precision="FP32">
5223 <dim>1</dim>
5224 <dim>1</dim>
5225 <dim>768</dim>
5226 </port>
5227 </output>
5228 </layer>
5229 <layer id="338" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5230 <data auto_broadcast="numpy" />
5231 <input>
5232 <port id="0" precision="FP32">
5233 <dim>-1</dim>
5234 <dim>-1</dim>
5235 <dim>768</dim>
5236 </port>
5237 <port id="1" precision="FP32">
5238 <dim>1</dim>
5239 <dim>1</dim>
5240 <dim>768</dim>
5241 </port>
5242 </input>
5243 <output>
5244 <port id="2" precision="FP32">
5245 <dim>-1</dim>
5246 <dim>-1</dim>
5247 <dim>768</dim>
5248 </port>
5249 </output>
5250 </layer>
5251 <layer id="339" name="Constant_5411" type="Const" version="opset1">
5252 <data element_type="f32" shape="1, 1, 768" offset="297760944" size="3072" />
5253 <output>
5254 <port id="0" precision="FP32">
5255 <dim>1</dim>
5256 <dim>1</dim>
5257 <dim>768</dim>
5258 </port>
5259 </output>
5260 </layer>
5261 <layer id="340" name="__module.roberta.encoder.layer.4.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5262 <data auto_broadcast="numpy" />
5263 <input>
5264 <port id="0" precision="FP32">
5265 <dim>-1</dim>
5266 <dim>-1</dim>
5267 <dim>768</dim>
5268 </port>
5269 <port id="1" precision="FP32">
5270 <dim>1</dim>
5271 <dim>1</dim>
5272 <dim>768</dim>
5273 </port>
5274 </input>
5275 <output>
5276 <port id="2" precision="FP32" names="429,hidden_states.31">
5277 <dim>-1</dim>
5278 <dim>-1</dim>
5279 <dim>768</dim>
5280 </port>
5281 </output>
5282 </layer>
5283 <layer id="341" name="self.roberta.encoder.layer.5.attention.self.query.weight" type="Const" version="opset1">
5284 <data element_type="f32" shape="768, 768" offset="297764016" size="2359296" />
5285 <output>
5286 <port id="0" precision="FP32" names="self.roberta.encoder.layer.5.attention.self.query.weight">
5287 <dim>768</dim>
5288 <dim>768</dim>
5289 </port>
5290 </output>
5291 </layer>
5292 <layer id="342" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
5293 <data transpose_a="false" transpose_b="true" />
5294 <input>
5295 <port id="0" precision="FP32">
5296 <dim>-1</dim>
5297 <dim>-1</dim>
5298 <dim>768</dim>
5299 </port>
5300 <port id="1" precision="FP32">
5301 <dim>768</dim>
5302 <dim>768</dim>
5303 </port>
5304 </input>
5305 <output>
5306 <port id="2" precision="FP32">
5307 <dim>-1</dim>
5308 <dim>-1</dim>
5309 <dim>768</dim>
5310 </port>
5311 </output>
5312 </layer>
5313 <layer id="343" name="Constant_5412" type="Const" version="opset1">
5314 <data element_type="f32" shape="1, 1, 768" offset="300123312" size="3072" />
5315 <output>
5316 <port id="0" precision="FP32">
5317 <dim>1</dim>
5318 <dim>1</dim>
5319 <dim>768</dim>
5320 </port>
5321 </output>
5322 </layer>
5323 <layer id="344" name="__module.roberta.encoder.layer.5.attention.self.query/aten::linear/Add" type="Add" version="opset1">
5324 <data auto_broadcast="numpy" />
5325 <input>
5326 <port id="0" precision="FP32">
5327 <dim>-1</dim>
5328 <dim>-1</dim>
5329 <dim>768</dim>
5330 </port>
5331 <port id="1" precision="FP32">
5332 <dim>1</dim>
5333 <dim>1</dim>
5334 <dim>768</dim>
5335 </port>
5336 </input>
5337 <output>
5338 <port id="2" precision="FP32" names="442,x.61">
5339 <dim>-1</dim>
5340 <dim>-1</dim>
5341 <dim>768</dim>
5342 </port>
5343 </output>
5344 </layer>
5345 <layer id="345" name="__module.roberta.encoder.layer.5.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
5346 <data element_type="i64" shape="4" offset="158368808" size="32" />
5347 <output>
5348 <port id="0" precision="I64">
5349 <dim>4</dim>
5350 </port>
5351 </output>
5352 </layer>
5353 <layer id="346" name="__module.roberta.encoder.layer.5.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
5354 <data special_zero="true" />
5355 <input>
5356 <port id="0" precision="FP32">
5357 <dim>-1</dim>
5358 <dim>-1</dim>
5359 <dim>768</dim>
5360 </port>
5361 <port id="1" precision="I64">
5362 <dim>4</dim>
5363 </port>
5364 </input>
5365 <output>
5366 <port id="2" precision="FP32" names="446,x.63">
5367 <dim>-1</dim>
5368 <dim>-1</dim>
5369 <dim>12</dim>
5370 <dim>64</dim>
5371 </port>
5372 </output>
5373 </layer>
5374 <layer id="347" name="Constant_1344" type="Const" version="opset1">
5375 <data element_type="i64" shape="4" offset="158368840" size="32" />
5376 <output>
5377 <port id="0" precision="I64" names="447">
5378 <dim>4</dim>
5379 </port>
5380 </output>
5381 </layer>
5382 <layer id="348" name="__module.roberta.encoder.layer.5.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
5383 <input>
5384 <port id="0" precision="FP32">
5385 <dim>-1</dim>
5386 <dim>-1</dim>
5387 <dim>12</dim>
5388 <dim>64</dim>
5389 </port>
5390 <port id="1" precision="I64">
5391 <dim>4</dim>
5392 </port>
5393 </input>
5394 <output>
5395 <port id="2" precision="FP32" names="448">
5396 <dim>-1</dim>
5397 <dim>12</dim>
5398 <dim>-1</dim>
5399 <dim>64</dim>
5400 </port>
5401 </output>
5402 </layer>
5403 <layer id="349" name="self.roberta.encoder.layer.5.attention.self.key.weight" type="Const" version="opset1">
5404 <data element_type="f32" shape="768, 768" offset="300126384" size="2359296" />
5405 <output>
5406 <port id="0" precision="FP32" names="self.roberta.encoder.layer.5.attention.self.key.weight">
5407 <dim>768</dim>
5408 <dim>768</dim>
5409 </port>
5410 </output>
5411 </layer>
5412 <layer id="350" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
5413 <data transpose_a="false" transpose_b="true" />
5414 <input>
5415 <port id="0" precision="FP32">
5416 <dim>-1</dim>
5417 <dim>-1</dim>
5418 <dim>768</dim>
5419 </port>
5420 <port id="1" precision="FP32">
5421 <dim>768</dim>
5422 <dim>768</dim>
5423 </port>
5424 </input>
5425 <output>
5426 <port id="2" precision="FP32">
5427 <dim>-1</dim>
5428 <dim>-1</dim>
5429 <dim>768</dim>
5430 </port>
5431 </output>
5432 </layer>
5433 <layer id="351" name="Constant_5413" type="Const" version="opset1">
5434 <data element_type="f32" shape="1, 1, 768" offset="302485680" size="3072" />
5435 <output>
5436 <port id="0" precision="FP32">
5437 <dim>1</dim>
5438 <dim>1</dim>
5439 <dim>768</dim>
5440 </port>
5441 </output>
5442 </layer>
5443 <layer id="352" name="__module.roberta.encoder.layer.5.attention.self.key/aten::linear/Add" type="Add" version="opset1">
5444 <data auto_broadcast="numpy" />
5445 <input>
5446 <port id="0" precision="FP32">
5447 <dim>-1</dim>
5448 <dim>-1</dim>
5449 <dim>768</dim>
5450 </port>
5451 <port id="1" precision="FP32">
5452 <dim>1</dim>
5453 <dim>1</dim>
5454 <dim>768</dim>
5455 </port>
5456 </input>
5457 <output>
5458 <port id="2" precision="FP32" names="451,x.65">
5459 <dim>-1</dim>
5460 <dim>-1</dim>
5461 <dim>768</dim>
5462 </port>
5463 </output>
5464 </layer>
5465 <layer id="353" name="__module.roberta.encoder.layer.5.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
5466 <data element_type="i64" shape="4" offset="158368808" size="32" />
5467 <output>
5468 <port id="0" precision="I64">
5469 <dim>4</dim>
5470 </port>
5471 </output>
5472 </layer>
5473 <layer id="354" name="__module.roberta.encoder.layer.5.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
5474 <data special_zero="true" />
5475 <input>
5476 <port id="0" precision="FP32">
5477 <dim>-1</dim>
5478 <dim>-1</dim>
5479 <dim>768</dim>
5480 </port>
5481 <port id="1" precision="I64">
5482 <dim>4</dim>
5483 </port>
5484 </input>
5485 <output>
5486 <port id="2" precision="FP32" names="455,x.67">
5487 <dim>-1</dim>
5488 <dim>-1</dim>
5489 <dim>12</dim>
5490 <dim>64</dim>
5491 </port>
5492 </output>
5493 </layer>
5494 <layer id="355" name="Constant_1367" type="Const" version="opset1">
5495 <data element_type="i64" shape="4" offset="158368840" size="32" />
5496 <output>
5497 <port id="0" precision="I64" names="456">
5498 <dim>4</dim>
5499 </port>
5500 </output>
5501 </layer>
5502 <layer id="356" name="__module.roberta.encoder.layer.5.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
5503 <input>
5504 <port id="0" precision="FP32">
5505 <dim>-1</dim>
5506 <dim>-1</dim>
5507 <dim>12</dim>
5508 <dim>64</dim>
5509 </port>
5510 <port id="1" precision="I64">
5511 <dim>4</dim>
5512 </port>
5513 </input>
5514 <output>
5515 <port id="2" precision="FP32" names="457">
5516 <dim>-1</dim>
5517 <dim>12</dim>
5518 <dim>-1</dim>
5519 <dim>64</dim>
5520 </port>
5521 </output>
5522 </layer>
5523 <layer id="357" name="self.roberta.encoder.layer.5.attention.self.value.weight" type="Const" version="opset1">
5524 <data element_type="f32" shape="768, 768" offset="302488752" size="2359296" />
5525 <output>
5526 <port id="0" precision="FP32" names="self.roberta.encoder.layer.5.attention.self.value.weight">
5527 <dim>768</dim>
5528 <dim>768</dim>
5529 </port>
5530 </output>
5531 </layer>
5532 <layer id="358" name="__module.roberta.encoder.layer.5.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
5533 <data transpose_a="false" transpose_b="true" />
5534 <input>
5535 <port id="0" precision="FP32">
5536 <dim>-1</dim>
5537 <dim>-1</dim>
5538 <dim>768</dim>
5539 </port>
5540 <port id="1" precision="FP32">
5541 <dim>768</dim>
5542 <dim>768</dim>
5543 </port>
5544 </input>
5545 <output>
5546 <port id="2" precision="FP32">
5547 <dim>-1</dim>
5548 <dim>-1</dim>
5549 <dim>768</dim>
5550 </port>
5551 </output>
5552 </layer>
5553 <layer id="359" name="Constant_5414" type="Const" version="opset1">
5554 <data element_type="f32" shape="1, 1, 768" offset="304848048" size="3072" />
5555 <output>
5556 <port id="0" precision="FP32">
5557 <dim>1</dim>
5558 <dim>1</dim>
5559 <dim>768</dim>
5560 </port>
5561 </output>
5562 </layer>
5563 <layer id="360" name="__module.roberta.encoder.layer.5.attention.self.value/aten::linear/Add" type="Add" version="opset1">
5564 <data auto_broadcast="numpy" />
5565 <input>
5566 <port id="0" precision="FP32">
5567 <dim>-1</dim>
5568 <dim>-1</dim>
5569 <dim>768</dim>
5570 </port>
5571 <port id="1" precision="FP32">
5572 <dim>1</dim>
5573 <dim>1</dim>
5574 <dim>768</dim>
5575 </port>
5576 </input>
5577 <output>
5578 <port id="2" precision="FP32" names="460,x.69">
5579 <dim>-1</dim>
5580 <dim>-1</dim>
5581 <dim>768</dim>
5582 </port>
5583 </output>
5584 </layer>
5585 <layer id="361" name="__module.roberta.encoder.layer.5.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
5586 <data element_type="i64" shape="4" offset="158368808" size="32" />
5587 <output>
5588 <port id="0" precision="I64">
5589 <dim>4</dim>
5590 </port>
5591 </output>
5592 </layer>
5593 <layer id="362" name="__module.roberta.encoder.layer.5.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
5594 <data special_zero="true" />
5595 <input>
5596 <port id="0" precision="FP32">
5597 <dim>-1</dim>
5598 <dim>-1</dim>
5599 <dim>768</dim>
5600 </port>
5601 <port id="1" precision="I64">
5602 <dim>4</dim>
5603 </port>
5604 </input>
5605 <output>
5606 <port id="2" precision="FP32" names="464,x">
5607 <dim>-1</dim>
5608 <dim>-1</dim>
5609 <dim>12</dim>
5610 <dim>64</dim>
5611 </port>
5612 </output>
5613 </layer>
5614 <layer id="363" name="Constant_1390" type="Const" version="opset1">
5615 <data element_type="i64" shape="4" offset="158368840" size="32" />
5616 <output>
5617 <port id="0" precision="I64" names="465">
5618 <dim>4</dim>
5619 </port>
5620 </output>
5621 </layer>
5622 <layer id="364" name="__module.roberta.encoder.layer.5.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
5623 <input>
5624 <port id="0" precision="FP32">
5625 <dim>-1</dim>
5626 <dim>-1</dim>
5627 <dim>12</dim>
5628 <dim>64</dim>
5629 </port>
5630 <port id="1" precision="I64">
5631 <dim>4</dim>
5632 </port>
5633 </input>
5634 <output>
5635 <port id="2" precision="FP32" names="466">
5636 <dim>-1</dim>
5637 <dim>12</dim>
5638 <dim>-1</dim>
5639 <dim>64</dim>
5640 </port>
5641 </output>
5642 </layer>
5643 <layer id="365" name="__module.roberta.encoder.layer.5.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
5644 <data causal="false" />
5645 <input>
5646 <port id="0" precision="FP32">
5647 <dim>-1</dim>
5648 <dim>12</dim>
5649 <dim>-1</dim>
5650 <dim>64</dim>
5651 </port>
5652 <port id="1" precision="FP32">
5653 <dim>-1</dim>
5654 <dim>12</dim>
5655 <dim>-1</dim>
5656 <dim>64</dim>
5657 </port>
5658 <port id="2" precision="FP32">
5659 <dim>-1</dim>
5660 <dim>12</dim>
5661 <dim>-1</dim>
5662 <dim>64</dim>
5663 </port>
5664 <port id="3" precision="FP32">
5665 <dim>-1</dim>
5666 <dim>1</dim>
5667 <dim>-1</dim>
5668 <dim>-1</dim>
5669 </port>
5670 </input>
5671 <output>
5672 <port id="4" precision="FP32" names="467,attn_output.21">
5673 <dim>-1</dim>
5674 <dim>12</dim>
5675 <dim>-1</dim>
5676 <dim>64</dim>
5677 </port>
5678 </output>
5679 </layer>
5680 <layer id="366" name="__module.roberta.encoder.layer.5.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
5681 <data element_type="i32" shape="4" offset="163093640" size="16" />
5682 <output>
5683 <port id="0" precision="I32">
5684 <dim>4</dim>
5685 </port>
5686 </output>
5687 </layer>
5688 <layer id="367" name="__module.roberta.encoder.layer.5.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
5689 <input>
5690 <port id="0" precision="FP32">
5691 <dim>-1</dim>
5692 <dim>12</dim>
5693 <dim>-1</dim>
5694 <dim>64</dim>
5695 </port>
5696 <port id="1" precision="I32">
5697 <dim>4</dim>
5698 </port>
5699 </input>
5700 <output>
5701 <port id="2" precision="FP32" names="468,attn_output">
5702 <dim>-1</dim>
5703 <dim>-1</dim>
5704 <dim>12</dim>
5705 <dim>64</dim>
5706 </port>
5707 </output>
5708 </layer>
5709 <layer id="368" name="Constant_5499" type="Const" version="opset1">
5710 <data element_type="i64" shape="3" offset="163093656" size="24" />
5711 <output>
5712 <port id="0" precision="I64">
5713 <dim>3</dim>
5714 </port>
5715 </output>
5716 </layer>
5717 <layer id="369" name="__module.roberta.encoder.layer.5.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
5718 <data special_zero="true" />
5719 <input>
5720 <port id="0" precision="FP32">
5721 <dim>-1</dim>
5722 <dim>-1</dim>
5723 <dim>12</dim>
5724 <dim>64</dim>
5725 </port>
5726 <port id="1" precision="I64">
5727 <dim>3</dim>
5728 </port>
5729 </input>
5730 <output>
5731 <port id="2" precision="FP32" names="470">
5732 <dim>-1</dim>
5733 <dim>-1</dim>
5734 <dim>768</dim>
5735 </port>
5736 </output>
5737 </layer>
5738 <layer id="370" name="self.roberta.encoder.layer.5.attention.output.dense.weight" type="Const" version="opset1">
5739 <data element_type="f32" shape="768, 768" offset="304851120" size="2359296" />
5740 <output>
5741 <port id="0" precision="FP32" names="self.roberta.encoder.layer.5.attention.output.dense.weight">
5742 <dim>768</dim>
5743 <dim>768</dim>
5744 </port>
5745 </output>
5746 </layer>
5747 <layer id="371" name="__module.roberta.encoder.layer.5.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5748 <data transpose_a="false" transpose_b="true" />
5749 <input>
5750 <port id="0" precision="FP32">
5751 <dim>-1</dim>
5752 <dim>-1</dim>
5753 <dim>768</dim>
5754 </port>
5755 <port id="1" precision="FP32">
5756 <dim>768</dim>
5757 <dim>768</dim>
5758 </port>
5759 </input>
5760 <output>
5761 <port id="2" precision="FP32">
5762 <dim>-1</dim>
5763 <dim>-1</dim>
5764 <dim>768</dim>
5765 </port>
5766 </output>
5767 </layer>
5768 <layer id="372" name="Constant_5415" type="Const" version="opset1">
5769 <data element_type="f32" shape="1, 1, 768" offset="307210416" size="3072" />
5770 <output>
5771 <port id="0" precision="FP32">
5772 <dim>1</dim>
5773 <dim>1</dim>
5774 <dim>768</dim>
5775 </port>
5776 </output>
5777 </layer>
5778 <layer id="373" name="__module.roberta.encoder.layer.5.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
5779 <data auto_broadcast="numpy" />
5780 <input>
5781 <port id="0" precision="FP32">
5782 <dim>-1</dim>
5783 <dim>-1</dim>
5784 <dim>768</dim>
5785 </port>
5786 <port id="1" precision="FP32">
5787 <dim>1</dim>
5788 <dim>1</dim>
5789 <dim>768</dim>
5790 </port>
5791 </input>
5792 <output>
5793 <port id="2" precision="FP32" names="475,input.23">
5794 <dim>-1</dim>
5795 <dim>-1</dim>
5796 <dim>768</dim>
5797 </port>
5798 </output>
5799 </layer>
5800 <layer id="374" name="__module.roberta.encoder.layer.5.attention.output/aten::add/Add" type="Add" version="opset1">
5801 <data auto_broadcast="numpy" />
5802 <input>
5803 <port id="0" precision="FP32">
5804 <dim>-1</dim>
5805 <dim>-1</dim>
5806 <dim>768</dim>
5807 </port>
5808 <port id="1" precision="FP32">
5809 <dim>-1</dim>
5810 <dim>-1</dim>
5811 <dim>768</dim>
5812 </port>
5813 </input>
5814 <output>
5815 <port id="2" precision="FP32" names="477">
5816 <dim>-1</dim>
5817 <dim>-1</dim>
5818 <dim>768</dim>
5819 </port>
5820 </output>
5821 </layer>
5822 <layer id="375" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
5823 <data element_type="i32" shape="1" offset="156000292" size="4" />
5824 <output>
5825 <port id="0" precision="I32">
5826 <dim>1</dim>
5827 </port>
5828 </output>
5829 </layer>
5830 <layer id="376" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
5831 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
5832 <input>
5833 <port id="0" precision="FP32">
5834 <dim>-1</dim>
5835 <dim>-1</dim>
5836 <dim>768</dim>
5837 </port>
5838 <port id="1" precision="I32">
5839 <dim>1</dim>
5840 </port>
5841 </input>
5842 <output>
5843 <port id="2" precision="FP32">
5844 <dim>-1</dim>
5845 <dim>-1</dim>
5846 <dim>768</dim>
5847 </port>
5848 </output>
5849 </layer>
5850 <layer id="377" name="Constant_5416" type="Const" version="opset1">
5851 <data element_type="f32" shape="1, 1, 768" offset="307213488" size="3072" />
5852 <output>
5853 <port id="0" precision="FP32">
5854 <dim>1</dim>
5855 <dim>1</dim>
5856 <dim>768</dim>
5857 </port>
5858 </output>
5859 </layer>
5860 <layer id="378" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
5861 <data auto_broadcast="numpy" />
5862 <input>
5863 <port id="0" precision="FP32">
5864 <dim>-1</dim>
5865 <dim>-1</dim>
5866 <dim>768</dim>
5867 </port>
5868 <port id="1" precision="FP32">
5869 <dim>1</dim>
5870 <dim>1</dim>
5871 <dim>768</dim>
5872 </port>
5873 </input>
5874 <output>
5875 <port id="2" precision="FP32">
5876 <dim>-1</dim>
5877 <dim>-1</dim>
5878 <dim>768</dim>
5879 </port>
5880 </output>
5881 </layer>
5882 <layer id="379" name="Constant_5417" type="Const" version="opset1">
5883 <data element_type="f32" shape="1, 1, 768" offset="307216560" size="3072" />
5884 <output>
5885 <port id="0" precision="FP32">
5886 <dim>1</dim>
5887 <dim>1</dim>
5888 <dim>768</dim>
5889 </port>
5890 </output>
5891 </layer>
5892 <layer id="380" name="__module.roberta.encoder.layer.5.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
5893 <data auto_broadcast="numpy" />
5894 <input>
5895 <port id="0" precision="FP32">
5896 <dim>-1</dim>
5897 <dim>-1</dim>
5898 <dim>768</dim>
5899 </port>
5900 <port id="1" precision="FP32">
5901 <dim>1</dim>
5902 <dim>1</dim>
5903 <dim>768</dim>
5904 </port>
5905 </input>
5906 <output>
5907 <port id="2" precision="FP32" names="481,input_tensor">
5908 <dim>-1</dim>
5909 <dim>-1</dim>
5910 <dim>768</dim>
5911 </port>
5912 </output>
5913 </layer>
5914 <layer id="381" name="self.roberta.encoder.layer.5.intermediate.dense.weight" type="Const" version="opset1">
5915 <data element_type="f32" shape="3072, 768" offset="307219632" size="9437184" />
5916 <output>
5917 <port id="0" precision="FP32" names="self.roberta.encoder.layer.5.intermediate.dense.weight">
5918 <dim>3072</dim>
5919 <dim>768</dim>
5920 </port>
5921 </output>
5922 </layer>
5923 <layer id="382" name="__module.roberta.encoder.layer.5.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
5924 <data transpose_a="false" transpose_b="true" />
5925 <input>
5926 <port id="0" precision="FP32">
5927 <dim>-1</dim>
5928 <dim>-1</dim>
5929 <dim>768</dim>
5930 </port>
5931 <port id="1" precision="FP32">
5932 <dim>3072</dim>
5933 <dim>768</dim>
5934 </port>
5935 </input>
5936 <output>
5937 <port id="2" precision="FP32">
5938 <dim>-1</dim>
5939 <dim>-1</dim>
5940 <dim>3072</dim>
5941 </port>
5942 </output>
5943 </layer>
5944 <layer id="383" name="Constant_5418" type="Const" version="opset1">
5945 <data element_type="f32" shape="1, 1, 3072" offset="316656816" size="12288" />
5946 <output>
5947 <port id="0" precision="FP32">
5948 <dim>1</dim>
5949 <dim>1</dim>
5950 <dim>3072</dim>
5951 </port>
5952 </output>
5953 </layer>
5954 <layer id="384" name="__module.roberta.encoder.layer.5.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
5955 <data auto_broadcast="numpy" />
5956 <input>
5957 <port id="0" precision="FP32">
5958 <dim>-1</dim>
5959 <dim>-1</dim>
5960 <dim>3072</dim>
5961 </port>
5962 <port id="1" precision="FP32">
5963 <dim>1</dim>
5964 <dim>1</dim>
5965 <dim>3072</dim>
5966 </port>
5967 </input>
5968 <output>
5969 <port id="2" precision="FP32" names="485">
5970 <dim>-1</dim>
5971 <dim>-1</dim>
5972 <dim>3072</dim>
5973 </port>
5974 </output>
5975 </layer>
5976 <layer id="385" name="__module.roberta.encoder.layer.5.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
5977 <data approximation_mode="ERF" />
5978 <input>
5979 <port id="0" precision="FP32">
5980 <dim>-1</dim>
5981 <dim>-1</dim>
5982 <dim>3072</dim>
5983 </port>
5984 </input>
5985 <output>
5986 <port id="1" precision="FP32" names="486">
5987 <dim>-1</dim>
5988 <dim>-1</dim>
5989 <dim>3072</dim>
5990 </port>
5991 </output>
5992 </layer>
5993 <layer id="386" name="self.roberta.encoder.layer.5.output.dense.weight" type="Const" version="opset1">
5994 <data element_type="f32" shape="768, 3072" offset="316669104" size="9437184" />
5995 <output>
5996 <port id="0" precision="FP32" names="self.roberta.encoder.layer.5.output.dense.weight">
5997 <dim>768</dim>
5998 <dim>3072</dim>
5999 </port>
6000 </output>
6001 </layer>
6002 <layer id="387" name="__module.roberta.encoder.layer.5.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6003 <data transpose_a="false" transpose_b="true" />
6004 <input>
6005 <port id="0" precision="FP32">
6006 <dim>-1</dim>
6007 <dim>-1</dim>
6008 <dim>3072</dim>
6009 </port>
6010 <port id="1" precision="FP32">
6011 <dim>768</dim>
6012 <dim>3072</dim>
6013 </port>
6014 </input>
6015 <output>
6016 <port id="2" precision="FP32">
6017 <dim>-1</dim>
6018 <dim>-1</dim>
6019 <dim>768</dim>
6020 </port>
6021 </output>
6022 </layer>
6023 <layer id="388" name="Constant_5419" type="Const" version="opset1">
6024 <data element_type="f32" shape="1, 1, 768" offset="326106288" size="3072" />
6025 <output>
6026 <port id="0" precision="FP32">
6027 <dim>1</dim>
6028 <dim>1</dim>
6029 <dim>768</dim>
6030 </port>
6031 </output>
6032 </layer>
6033 <layer id="389" name="__module.roberta.encoder.layer.5.output.dense/aten::linear/Add" type="Add" version="opset1">
6034 <data auto_broadcast="numpy" />
6035 <input>
6036 <port id="0" precision="FP32">
6037 <dim>-1</dim>
6038 <dim>-1</dim>
6039 <dim>768</dim>
6040 </port>
6041 <port id="1" precision="FP32">
6042 <dim>1</dim>
6043 <dim>1</dim>
6044 <dim>768</dim>
6045 </port>
6046 </input>
6047 <output>
6048 <port id="2" precision="FP32" names="491,input.25">
6049 <dim>-1</dim>
6050 <dim>-1</dim>
6051 <dim>768</dim>
6052 </port>
6053 </output>
6054 </layer>
6055 <layer id="390" name="__module.roberta.encoder.layer.5.output/aten::add/Add" type="Add" version="opset1">
6056 <data auto_broadcast="numpy" />
6057 <input>
6058 <port id="0" precision="FP32">
6059 <dim>-1</dim>
6060 <dim>-1</dim>
6061 <dim>768</dim>
6062 </port>
6063 <port id="1" precision="FP32">
6064 <dim>-1</dim>
6065 <dim>-1</dim>
6066 <dim>768</dim>
6067 </port>
6068 </input>
6069 <output>
6070 <port id="2" precision="FP32" names="493">
6071 <dim>-1</dim>
6072 <dim>-1</dim>
6073 <dim>768</dim>
6074 </port>
6075 </output>
6076 </layer>
6077 <layer id="391" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
6078 <data element_type="i32" shape="1" offset="156000292" size="4" />
6079 <output>
6080 <port id="0" precision="I32">
6081 <dim>1</dim>
6082 </port>
6083 </output>
6084 </layer>
6085 <layer id="392" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
6086 <data eps="9.9999997473787516e-06" normalize_variance="true" eps_mode="INSIDE_SQRT" />
6087 <input>
6088 <port id="0" precision="FP32">
6089 <dim>-1</dim>
6090 <dim>-1</dim>
6091 <dim>768</dim>
6092 </port>
6093 <port id="1" precision="I32">
6094 <dim>1</dim>
6095 </port>
6096 </input>
6097 <output>
6098 <port id="2" precision="FP32">
6099 <dim>-1</dim>
6100 <dim>-1</dim>
6101 <dim>768</dim>
6102 </port>
6103 </output>
6104 </layer>
6105 <layer id="393" name="Constant_5420" type="Const" version="opset1">
6106 <data element_type="f32" shape="1, 1, 768" offset="326109360" size="3072" />
6107 <output>
6108 <port id="0" precision="FP32">
6109 <dim>1</dim>
6110 <dim>1</dim>
6111 <dim>768</dim>
6112 </port>
6113 </output>
6114 </layer>
6115 <layer id="394" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
6116 <data auto_broadcast="numpy" />
6117 <input>
6118 <port id="0" precision="FP32">
6119 <dim>-1</dim>
6120 <dim>-1</dim>
6121 <dim>768</dim>
6122 </port>
6123 <port id="1" precision="FP32">
6124 <dim>1</dim>
6125 <dim>1</dim>
6126 <dim>768</dim>
6127 </port>
6128 </input>
6129 <output>
6130 <port id="2" precision="FP32">
6131 <dim>-1</dim>
6132 <dim>-1</dim>
6133 <dim>768</dim>
6134 </port>
6135 </output>
6136 </layer>
6137 <layer id="395" name="Constant_5421" type="Const" version="opset1">
6138 <data element_type="f32" shape="1, 1, 768" offset="326112432" size="3072" />
6139 <output>
6140 <port id="0" precision="FP32">
6141 <dim>1</dim>
6142 <dim>1</dim>
6143 <dim>768</dim>
6144 </port>
6145 </output>
6146 </layer>
6147 <layer id="396" name="__module.roberta.encoder.layer.5.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
6148 <data auto_broadcast="numpy" />
6149 <input>
6150 <port id="0" precision="FP32">
6151 <dim>-1</dim>
6152 <dim>-1</dim>
6153 <dim>768</dim>
6154 </port>
6155 <port id="1" precision="FP32">
6156 <dim>1</dim>
6157 <dim>1</dim>
6158 <dim>768</dim>
6159 </port>
6160 </input>
6161 <output>
6162 <port id="2" precision="FP32" names="497,506,features">
6163 <dim>-1</dim>
6164 <dim>-1</dim>
6165 <dim>768</dim>
6166 </port>
6167 </output>
6168 </layer>
6169 <layer id="397" name="502" type="Const" version="opset1">
6170 <data element_type="i64" shape="" offset="154421268" size="8" />
6171 <output>
6172 <port id="0" precision="I64" names="502" />
6173 </output>
6174 </layer>
6175 <layer id="398" name="500" type="Const" version="opset1">
6176 <data element_type="i64" shape="" offset="154421276" size="8" />
6177 <output>
6178 <port id="0" precision="I64" names="500" />
6179 </output>
6180 </layer>
6181 <layer id="399" name="__module.classifier/aten::select/Gather" type="Gather" version="opset8">
6182 <data batch_dims="0" />
6183 <input>
6184 <port id="0" precision="FP32">
6185 <dim>-1</dim>
6186 <dim>-1</dim>
6187 <dim>768</dim>
6188 </port>
6189 <port id="1" precision="I64" />
6190 <port id="2" precision="I64" />
6191 </input>
6192 <output>
6193 <port id="3" precision="FP32" names="507,508,input.27">
6194 <dim>-1</dim>
6195 <dim>768</dim>
6196 </port>
6197 </output>
6198 </layer>
6199 <layer id="400" name="self.classifier.dense.weight" type="Const" version="opset1">
6200 <data element_type="f32" shape="768, 768" offset="326115504" size="2359296" />
6201 <output>
6202 <port id="0" precision="FP32" names="self.classifier.dense.weight">
6203 <dim>768</dim>
6204 <dim>768</dim>
6205 </port>
6206 </output>
6207 </layer>
6208 <layer id="401" name="__module.classifier.dense/aten::linear/MatMul" type="MatMul" version="opset1">
6209 <data transpose_a="false" transpose_b="true" />
6210 <input>
6211 <port id="0" precision="FP32">
6212 <dim>-1</dim>
6213 <dim>768</dim>
6214 </port>
6215 <port id="1" precision="FP32">
6216 <dim>768</dim>
6217 <dim>768</dim>
6218 </port>
6219 </input>
6220 <output>
6221 <port id="2" precision="FP32">
6222 <dim>-1</dim>
6223 <dim>768</dim>
6224 </port>
6225 </output>
6226 </layer>
6227 <layer id="402" name="Constant_5422" type="Const" version="opset1">
6228 <data element_type="f32" shape="1, 768" offset="328474800" size="3072" />
6229 <output>
6230 <port id="0" precision="FP32">
6231 <dim>1</dim>
6232 <dim>768</dim>
6233 </port>
6234 </output>
6235 </layer>
6236 <layer id="403" name="__module.classifier.dense/aten::linear/Add" type="Add" version="opset1">
6237 <data auto_broadcast="numpy" />
6238 <input>
6239 <port id="0" precision="FP32">
6240 <dim>-1</dim>
6241 <dim>768</dim>
6242 </port>
6243 <port id="1" precision="FP32">
6244 <dim>1</dim>
6245 <dim>768</dim>
6246 </port>
6247 </input>
6248 <output>
6249 <port id="2" precision="FP32" names="512">
6250 <dim>-1</dim>
6251 <dim>768</dim>
6252 </port>
6253 </output>
6254 </layer>
6255 <layer id="404" name="__module.classifier/aten::tanh/Tanh" type="Tanh" version="opset1">
6256 <input>
6257 <port id="0" precision="FP32">
6258 <dim>-1</dim>
6259 <dim>768</dim>
6260 </port>
6261 </input>
6262 <output>
6263 <port id="1" precision="FP32" names="513,input">
6264 <dim>-1</dim>
6265 <dim>768</dim>
6266 </port>
6267 </output>
6268 </layer>
6269 <layer id="405" name="self.classifier.out_proj.weight" type="Const" version="opset1">
6270 <data element_type="f32" shape="3, 768" offset="328477872" size="9216" />
6271 <output>
6272 <port id="0" precision="FP32" names="self.classifier.out_proj.weight">
6273 <dim>3</dim>
6274 <dim>768</dim>
6275 </port>
6276 </output>
6277 </layer>
6278 <layer id="406" name="__module.classifier.out_proj/aten::linear/MatMul" type="MatMul" version="opset1">
6279 <data transpose_a="false" transpose_b="true" />
6280 <input>
6281 <port id="0" precision="FP32">
6282 <dim>-1</dim>
6283 <dim>768</dim>
6284 </port>
6285 <port id="1" precision="FP32">
6286 <dim>3</dim>
6287 <dim>768</dim>
6288 </port>
6289 </input>
6290 <output>
6291 <port id="2" precision="FP32">
6292 <dim>-1</dim>
6293 <dim>3</dim>
6294 </port>
6295 </output>
6296 </layer>
6297 <layer id="407" name="Constant_5423" type="Const" version="opset1">
6298 <data element_type="f32" shape="1, 3" offset="328487088" size="12" />
6299 <output>
6300 <port id="0" precision="FP32">
6301 <dim>1</dim>
6302 <dim>3</dim>
6303 </port>
6304 </output>
6305 </layer>
6306 <layer id="408" name="__module.classifier.out_proj/aten::linear/Add" type="Add" version="opset1">
6307 <data auto_broadcast="numpy" />
6308 <input>
6309 <port id="0" precision="FP32">
6310 <dim>-1</dim>
6311 <dim>3</dim>
6312 </port>
6313 <port id="1" precision="FP32">
6314 <dim>1</dim>
6315 <dim>3</dim>
6316 </port>
6317 </input>
6318 <output>
6319 <port id="2" precision="FP32" names="logits">
6320 <dim>-1</dim>
6321 <dim>3</dim>
6322 </port>
6323 </output>
6324 </layer>
6325 <layer id="409" name="Result_2399" type="Result" version="opset1">
6326 <input>
6327 <port id="0" precision="FP32">
6328 <dim>-1</dim>
6329 <dim>3</dim>
6330 </port>
6331 </input>
6332 </layer>
6333 </layers>
6334 <edges>
6335 <edge from-layer="0" from-port="0" to-layer="66" to-port="0" />
6336 <edge from-layer="1" from-port="0" to-layer="3" to-port="0" />
6337 <edge from-layer="1" from-port="0" to-layer="9" to-port="0" />
6338 <edge from-layer="1" from-port="0" to-layer="23" to-port="0" />
6339 <edge from-layer="2" from-port="0" to-layer="5" to-port="0" />
6340 <edge from-layer="3" from-port="1" to-layer="5" to-port="1" />
6341 <edge from-layer="4" from-port="0" to-layer="5" to-port="2" />
6342 <edge from-layer="5" from-port="3" to-layer="20" to-port="0" />
6343 <edge from-layer="6" from-port="0" to-layer="19" to-port="0" />
6344 <edge from-layer="7" from-port="0" to-layer="15" to-port="0" />
6345 <edge from-layer="8" from-port="0" to-layer="15" to-port="1" />
6346 <edge from-layer="9" from-port="1" to-layer="12" to-port="0" />
6347 <edge from-layer="9" from-port="1" to-layer="16" to-port="1" />
6348 <edge from-layer="9" from-port="1" to-layer="71" to-port="0" />
6349 <edge from-layer="9" from-port="1" to-layer="75" to-port="0" />
6350 <edge from-layer="10" from-port="0" to-layer="12" to-port="1" />
6351 <edge from-layer="11" from-port="0" to-layer="12" to-port="2" />
6352 <edge from-layer="12" from-port="3" to-layer="15" to-port="2" />
6353 <edge from-layer="13" from-port="0" to-layer="15" to-port="3" />
6354 <edge from-layer="14" from-port="0" to-layer="15" to-port="4" />
6355 <edge from-layer="15" from-port="5" to-layer="16" to-port="0" />
6356 <edge from-layer="16" from-port="2" to-layer="17" to-port="0" />
6357 <edge from-layer="17" from-port="1" to-layer="19" to-port="1" />
6358 <edge from-layer="18" from-port="0" to-layer="19" to-port="2" />
6359 <edge from-layer="19" from-port="3" to-layer="20" to-port="1" />
6360 <edge from-layer="20" from-port="2" to-layer="34" to-port="0" />
6361 <edge from-layer="21" from-port="0" to-layer="33" to-port="0" />
6362 <edge from-layer="22" from-port="0" to-layer="23" to-port="1" />
6363 <edge from-layer="23" from-port="2" to-layer="24" to-port="0" />
6364 <edge from-layer="24" from-port="1" to-layer="27" to-port="1" />
6365 <edge from-layer="24" from-port="1" to-layer="26" to-port="0" />
6366 <edge from-layer="25" from-port="0" to-layer="26" to-port="1" />
6367 <edge from-layer="25" from-port="0" to-layer="66" to-port="1" />
6368 <edge from-layer="26" from-port="2" to-layer="27" to-port="0" />
6369 <edge from-layer="27" from-port="2" to-layer="28" to-port="0" />
6370 <edge from-layer="28" from-port="1" to-layer="30" to-port="0" />
6371 <edge from-layer="29" from-port="0" to-layer="30" to-port="1" />
6372 <edge from-layer="30" from-port="2" to-layer="31" to-port="0" />
6373 <edge from-layer="31" from-port="1" to-layer="33" to-port="1" />
6374 <edge from-layer="32" from-port="0" to-layer="33" to-port="2" />
6375 <edge from-layer="33" from-port="3" to-layer="34" to-port="1" />
6376 <edge from-layer="34" from-port="2" to-layer="36" to-port="0" />
6377 <edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
6378 <edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
6379 <edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
6380 <edge from-layer="38" from-port="2" to-layer="40" to-port="0" />
6381 <edge from-layer="39" from-port="0" to-layer="40" to-port="1" />
6382 <edge from-layer="40" from-port="2" to-layer="42" to-port="0" />
6383 <edge from-layer="40" from-port="2" to-layer="58" to-port="0" />
6384 <edge from-layer="40" from-port="2" to-layer="94" to-port="1" />
6385 <edge from-layer="40" from-port="2" to-layer="50" to-port="0" />
6386 <edge from-layer="41" from-port="0" to-layer="42" to-port="1" />
6387 <edge from-layer="42" from-port="2" to-layer="44" to-port="0" />
6388 <edge from-layer="43" from-port="0" to-layer="44" to-port="1" />
6389 <edge from-layer="44" from-port="2" to-layer="46" to-port="0" />
6390 <edge from-layer="45" from-port="0" to-layer="46" to-port="1" />
6391 <edge from-layer="46" from-port="2" to-layer="48" to-port="0" />
6392 <edge from-layer="47" from-port="0" to-layer="48" to-port="1" />
6393 <edge from-layer="48" from-port="2" to-layer="85" to-port="0" />
6394 <edge from-layer="49" from-port="0" to-layer="50" to-port="1" />
6395 <edge from-layer="50" from-port="2" to-layer="52" to-port="0" />
6396 <edge from-layer="51" from-port="0" to-layer="52" to-port="1" />
6397 <edge from-layer="52" from-port="2" to-layer="54" to-port="0" />
6398 <edge from-layer="53" from-port="0" to-layer="54" to-port="1" />
6399 <edge from-layer="54" from-port="2" to-layer="56" to-port="0" />
6400 <edge from-layer="55" from-port="0" to-layer="56" to-port="1" />
6401 <edge from-layer="56" from-port="2" to-layer="85" to-port="1" />
6402 <edge from-layer="57" from-port="0" to-layer="58" to-port="1" />
6403 <edge from-layer="58" from-port="2" to-layer="60" to-port="0" />
6404 <edge from-layer="59" from-port="0" to-layer="60" to-port="1" />
6405 <edge from-layer="60" from-port="2" to-layer="62" to-port="0" />
6406 <edge from-layer="61" from-port="0" to-layer="62" to-port="1" />
6407 <edge from-layer="62" from-port="2" to-layer="64" to-port="0" />
6408 <edge from-layer="63" from-port="0" to-layer="64" to-port="1" />
6409 <edge from-layer="64" from-port="2" to-layer="85" to-port="2" />
6410 <edge from-layer="65" from-port="0" to-layer="81" to-port="0" />
6411 <edge from-layer="66" from-port="2" to-layer="68" to-port="0" />
6412 <edge from-layer="67" from-port="0" to-layer="68" to-port="1" />
6413 <edge from-layer="68" from-port="2" to-layer="77" to-port="0" />
6414 <edge from-layer="69" from-port="0" to-layer="71" to-port="1" />
6415 <edge from-layer="70" from-port="0" to-layer="71" to-port="2" />
6416 <edge from-layer="71" from-port="3" to-layer="76" to-port="0" />
6417 <edge from-layer="72" from-port="0" to-layer="76" to-port="1" />
6418 <edge from-layer="73" from-port="0" to-layer="75" to-port="1" />
6419 <edge from-layer="74" from-port="0" to-layer="75" to-port="2" />
6420 <edge from-layer="75" from-port="3" to-layer="76" to-port="2" />
6421 <edge from-layer="76" from-port="3" to-layer="77" to-port="1" />
6422 <edge from-layer="77" from-port="2" to-layer="78" to-port="0" />
6423 <edge from-layer="78" from-port="1" to-layer="80" to-port="0" />
6424 <edge from-layer="79" from-port="0" to-layer="80" to-port="1" />
6425 <edge from-layer="80" from-port="2" to-layer="81" to-port="1" />
6426 <edge from-layer="81" from-port="2" to-layer="82" to-port="0" />
6427 <edge from-layer="81" from-port="2" to-layer="84" to-port="2" />
6428 <edge from-layer="82" from-port="1" to-layer="84" to-port="0" />
6429 <edge from-layer="83" from-port="0" to-layer="84" to-port="1" />
6430 <edge from-layer="84" from-port="3" to-layer="85" to-port="3" />
6431 <edge from-layer="84" from-port="3" to-layer="309" to-port="3" />
6432 <edge from-layer="84" from-port="3" to-layer="141" to-port="3" />
6433 <edge from-layer="84" from-port="3" to-layer="365" to-port="3" />
6434 <edge from-layer="84" from-port="3" to-layer="197" to-port="3" />
6435 <edge from-layer="84" from-port="3" to-layer="253" to-port="3" />
6436 <edge from-layer="85" from-port="4" to-layer="87" to-port="0" />
6437 <edge from-layer="86" from-port="0" to-layer="87" to-port="1" />
6438 <edge from-layer="87" from-port="2" to-layer="89" to-port="0" />
6439 <edge from-layer="88" from-port="0" to-layer="89" to-port="1" />
6440 <edge from-layer="89" from-port="2" to-layer="91" to-port="0" />
6441 <edge from-layer="90" from-port="0" to-layer="91" to-port="1" />
6442 <edge from-layer="91" from-port="2" to-layer="93" to-port="0" />
6443 <edge from-layer="92" from-port="0" to-layer="93" to-port="1" />
6444 <edge from-layer="93" from-port="2" to-layer="94" to-port="0" />
6445 <edge from-layer="94" from-port="2" to-layer="96" to-port="0" />
6446 <edge from-layer="95" from-port="0" to-layer="96" to-port="1" />
6447 <edge from-layer="96" from-port="2" to-layer="98" to-port="0" />
6448 <edge from-layer="97" from-port="0" to-layer="98" to-port="1" />
6449 <edge from-layer="98" from-port="2" to-layer="100" to-port="0" />
6450 <edge from-layer="99" from-port="0" to-layer="100" to-port="1" />
6451 <edge from-layer="100" from-port="2" to-layer="102" to-port="0" />
6452 <edge from-layer="100" from-port="2" to-layer="110" to-port="1" />
6453 <edge from-layer="101" from-port="0" to-layer="102" to-port="1" />
6454 <edge from-layer="102" from-port="2" to-layer="104" to-port="0" />
6455 <edge from-layer="103" from-port="0" to-layer="104" to-port="1" />
6456 <edge from-layer="104" from-port="2" to-layer="105" to-port="0" />
6457 <edge from-layer="105" from-port="1" to-layer="107" to-port="0" />
6458 <edge from-layer="106" from-port="0" to-layer="107" to-port="1" />
6459 <edge from-layer="107" from-port="2" to-layer="109" to-port="0" />
6460 <edge from-layer="108" from-port="0" to-layer="109" to-port="1" />
6461 <edge from-layer="109" from-port="2" to-layer="110" to-port="0" />
6462 <edge from-layer="110" from-port="2" to-layer="112" to-port="0" />
6463 <edge from-layer="111" from-port="0" to-layer="112" to-port="1" />
6464 <edge from-layer="112" from-port="2" to-layer="114" to-port="0" />
6465 <edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
6466 <edge from-layer="114" from-port="2" to-layer="116" to-port="0" />
6467 <edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
6468 <edge from-layer="116" from-port="2" to-layer="118" to-port="0" />
6469 <edge from-layer="116" from-port="2" to-layer="126" to-port="0" />
6470 <edge from-layer="116" from-port="2" to-layer="134" to-port="0" />
6471 <edge from-layer="116" from-port="2" to-layer="150" to-port="1" />
6472 <edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
6473 <edge from-layer="118" from-port="2" to-layer="120" to-port="0" />
6474 <edge from-layer="119" from-port="0" to-layer="120" to-port="1" />
6475 <edge from-layer="120" from-port="2" to-layer="122" to-port="0" />
6476 <edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
6477 <edge from-layer="122" from-port="2" to-layer="124" to-port="0" />
6478 <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
6479 <edge from-layer="124" from-port="2" to-layer="141" to-port="0" />
6480 <edge from-layer="125" from-port="0" to-layer="126" to-port="1" />
6481 <edge from-layer="126" from-port="2" to-layer="128" to-port="0" />
6482 <edge from-layer="127" from-port="0" to-layer="128" to-port="1" />
6483 <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
6484 <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
6485 <edge from-layer="130" from-port="2" to-layer="132" to-port="0" />
6486 <edge from-layer="131" from-port="0" to-layer="132" to-port="1" />
6487 <edge from-layer="132" from-port="2" to-layer="141" to-port="1" />
6488 <edge from-layer="133" from-port="0" to-layer="134" to-port="1" />
6489 <edge from-layer="134" from-port="2" to-layer="136" to-port="0" />
6490 <edge from-layer="135" from-port="0" to-layer="136" to-port="1" />
6491 <edge from-layer="136" from-port="2" to-layer="138" to-port="0" />
6492 <edge from-layer="137" from-port="0" to-layer="138" to-port="1" />
6493 <edge from-layer="138" from-port="2" to-layer="140" to-port="0" />
6494 <edge from-layer="139" from-port="0" to-layer="140" to-port="1" />
6495 <edge from-layer="140" from-port="2" to-layer="141" to-port="2" />
6496 <edge from-layer="141" from-port="4" to-layer="143" to-port="0" />
6497 <edge from-layer="142" from-port="0" to-layer="143" to-port="1" />
6498 <edge from-layer="143" from-port="2" to-layer="145" to-port="0" />
6499 <edge from-layer="144" from-port="0" to-layer="145" to-port="1" />
6500 <edge from-layer="145" from-port="2" to-layer="147" to-port="0" />
6501 <edge from-layer="146" from-port="0" to-layer="147" to-port="1" />
6502 <edge from-layer="147" from-port="2" to-layer="149" to-port="0" />
6503 <edge from-layer="148" from-port="0" to-layer="149" to-port="1" />
6504 <edge from-layer="149" from-port="2" to-layer="150" to-port="0" />
6505 <edge from-layer="150" from-port="2" to-layer="152" to-port="0" />
6506 <edge from-layer="151" from-port="0" to-layer="152" to-port="1" />
6507 <edge from-layer="152" from-port="2" to-layer="154" to-port="0" />
6508 <edge from-layer="153" from-port="0" to-layer="154" to-port="1" />
6509 <edge from-layer="154" from-port="2" to-layer="156" to-port="0" />
6510 <edge from-layer="155" from-port="0" to-layer="156" to-port="1" />
6511 <edge from-layer="156" from-port="2" to-layer="166" to-port="1" />
6512 <edge from-layer="156" from-port="2" to-layer="158" to-port="0" />
6513 <edge from-layer="157" from-port="0" to-layer="158" to-port="1" />
6514 <edge from-layer="158" from-port="2" to-layer="160" to-port="0" />
6515 <edge from-layer="159" from-port="0" to-layer="160" to-port="1" />
6516 <edge from-layer="160" from-port="2" to-layer="161" to-port="0" />
6517 <edge from-layer="161" from-port="1" to-layer="163" to-port="0" />
6518 <edge from-layer="162" from-port="0" to-layer="163" to-port="1" />
6519 <edge from-layer="163" from-port="2" to-layer="165" to-port="0" />
6520 <edge from-layer="164" from-port="0" to-layer="165" to-port="1" />
6521 <edge from-layer="165" from-port="2" to-layer="166" to-port="0" />
6522 <edge from-layer="166" from-port="2" to-layer="168" to-port="0" />
6523 <edge from-layer="167" from-port="0" to-layer="168" to-port="1" />
6524 <edge from-layer="168" from-port="2" to-layer="170" to-port="0" />
6525 <edge from-layer="169" from-port="0" to-layer="170" to-port="1" />
6526 <edge from-layer="170" from-port="2" to-layer="172" to-port="0" />
6527 <edge from-layer="171" from-port="0" to-layer="172" to-port="1" />
6528 <edge from-layer="172" from-port="2" to-layer="174" to-port="0" />
6529 <edge from-layer="172" from-port="2" to-layer="182" to-port="0" />
6530 <edge from-layer="172" from-port="2" to-layer="190" to-port="0" />
6531 <edge from-layer="172" from-port="2" to-layer="206" to-port="1" />
6532 <edge from-layer="173" from-port="0" to-layer="174" to-port="1" />
6533 <edge from-layer="174" from-port="2" to-layer="176" to-port="0" />
6534 <edge from-layer="175" from-port="0" to-layer="176" to-port="1" />
6535 <edge from-layer="176" from-port="2" to-layer="178" to-port="0" />
6536 <edge from-layer="177" from-port="0" to-layer="178" to-port="1" />
6537 <edge from-layer="178" from-port="2" to-layer="180" to-port="0" />
6538 <edge from-layer="179" from-port="0" to-layer="180" to-port="1" />
6539 <edge from-layer="180" from-port="2" to-layer="197" to-port="0" />
6540 <edge from-layer="181" from-port="0" to-layer="182" to-port="1" />
6541 <edge from-layer="182" from-port="2" to-layer="184" to-port="0" />
6542 <edge from-layer="183" from-port="0" to-layer="184" to-port="1" />
6543 <edge from-layer="184" from-port="2" to-layer="186" to-port="0" />
6544 <edge from-layer="185" from-port="0" to-layer="186" to-port="1" />
6545 <edge from-layer="186" from-port="2" to-layer="188" to-port="0" />
6546 <edge from-layer="187" from-port="0" to-layer="188" to-port="1" />
6547 <edge from-layer="188" from-port="2" to-layer="197" to-port="1" />
6548 <edge from-layer="189" from-port="0" to-layer="190" to-port="1" />
6549 <edge from-layer="190" from-port="2" to-layer="192" to-port="0" />
6550 <edge from-layer="191" from-port="0" to-layer="192" to-port="1" />
6551 <edge from-layer="192" from-port="2" to-layer="194" to-port="0" />
6552 <edge from-layer="193" from-port="0" to-layer="194" to-port="1" />
6553 <edge from-layer="194" from-port="2" to-layer="196" to-port="0" />
6554 <edge from-layer="195" from-port="0" to-layer="196" to-port="1" />
6555 <edge from-layer="196" from-port="2" to-layer="197" to-port="2" />
6556 <edge from-layer="197" from-port="4" to-layer="199" to-port="0" />
6557 <edge from-layer="198" from-port="0" to-layer="199" to-port="1" />
6558 <edge from-layer="199" from-port="2" to-layer="201" to-port="0" />
6559 <edge from-layer="200" from-port="0" to-layer="201" to-port="1" />
6560 <edge from-layer="201" from-port="2" to-layer="203" to-port="0" />
6561 <edge from-layer="202" from-port="0" to-layer="203" to-port="1" />
6562 <edge from-layer="203" from-port="2" to-layer="205" to-port="0" />
6563 <edge from-layer="204" from-port="0" to-layer="205" to-port="1" />
6564 <edge from-layer="205" from-port="2" to-layer="206" to-port="0" />
6565 <edge from-layer="206" from-port="2" to-layer="208" to-port="0" />
6566 <edge from-layer="207" from-port="0" to-layer="208" to-port="1" />
6567 <edge from-layer="208" from-port="2" to-layer="210" to-port="0" />
6568 <edge from-layer="209" from-port="0" to-layer="210" to-port="1" />
6569 <edge from-layer="210" from-port="2" to-layer="212" to-port="0" />
6570 <edge from-layer="211" from-port="0" to-layer="212" to-port="1" />
6571 <edge from-layer="212" from-port="2" to-layer="214" to-port="0" />
6572 <edge from-layer="212" from-port="2" to-layer="222" to-port="1" />
6573 <edge from-layer="213" from-port="0" to-layer="214" to-port="1" />
6574 <edge from-layer="214" from-port="2" to-layer="216" to-port="0" />
6575 <edge from-layer="215" from-port="0" to-layer="216" to-port="1" />
6576 <edge from-layer="216" from-port="2" to-layer="217" to-port="0" />
6577 <edge from-layer="217" from-port="1" to-layer="219" to-port="0" />
6578 <edge from-layer="218" from-port="0" to-layer="219" to-port="1" />
6579 <edge from-layer="219" from-port="2" to-layer="221" to-port="0" />
6580 <edge from-layer="220" from-port="0" to-layer="221" to-port="1" />
6581 <edge from-layer="221" from-port="2" to-layer="222" to-port="0" />
6582 <edge from-layer="222" from-port="2" to-layer="224" to-port="0" />
6583 <edge from-layer="223" from-port="0" to-layer="224" to-port="1" />
6584 <edge from-layer="224" from-port="2" to-layer="226" to-port="0" />
6585 <edge from-layer="225" from-port="0" to-layer="226" to-port="1" />
6586 <edge from-layer="226" from-port="2" to-layer="228" to-port="0" />
6587 <edge from-layer="227" from-port="0" to-layer="228" to-port="1" />
6588 <edge from-layer="228" from-port="2" to-layer="230" to-port="0" />
6589 <edge from-layer="228" from-port="2" to-layer="238" to-port="0" />
6590 <edge from-layer="228" from-port="2" to-layer="246" to-port="0" />
6591 <edge from-layer="228" from-port="2" to-layer="262" to-port="1" />
6592 <edge from-layer="229" from-port="0" to-layer="230" to-port="1" />
6593 <edge from-layer="230" from-port="2" to-layer="232" to-port="0" />
6594 <edge from-layer="231" from-port="0" to-layer="232" to-port="1" />
6595 <edge from-layer="232" from-port="2" to-layer="234" to-port="0" />
6596 <edge from-layer="233" from-port="0" to-layer="234" to-port="1" />
6597 <edge from-layer="234" from-port="2" to-layer="236" to-port="0" />
6598 <edge from-layer="235" from-port="0" to-layer="236" to-port="1" />
6599 <edge from-layer="236" from-port="2" to-layer="253" to-port="0" />
6600 <edge from-layer="237" from-port="0" to-layer="238" to-port="1" />
6601 <edge from-layer="238" from-port="2" to-layer="240" to-port="0" />
6602 <edge from-layer="239" from-port="0" to-layer="240" to-port="1" />
6603 <edge from-layer="240" from-port="2" to-layer="242" to-port="0" />
6604 <edge from-layer="241" from-port="0" to-layer="242" to-port="1" />
6605 <edge from-layer="242" from-port="2" to-layer="244" to-port="0" />
6606 <edge from-layer="243" from-port="0" to-layer="244" to-port="1" />
6607 <edge from-layer="244" from-port="2" to-layer="253" to-port="1" />
6608 <edge from-layer="245" from-port="0" to-layer="246" to-port="1" />
6609 <edge from-layer="246" from-port="2" to-layer="248" to-port="0" />
6610 <edge from-layer="247" from-port="0" to-layer="248" to-port="1" />
6611 <edge from-layer="248" from-port="2" to-layer="250" to-port="0" />
6612 <edge from-layer="249" from-port="0" to-layer="250" to-port="1" />
6613 <edge from-layer="250" from-port="2" to-layer="252" to-port="0" />
6614 <edge from-layer="251" from-port="0" to-layer="252" to-port="1" />
6615 <edge from-layer="252" from-port="2" to-layer="253" to-port="2" />
6616 <edge from-layer="253" from-port="4" to-layer="255" to-port="0" />
6617 <edge from-layer="254" from-port="0" to-layer="255" to-port="1" />
6618 <edge from-layer="255" from-port="2" to-layer="257" to-port="0" />
6619 <edge from-layer="256" from-port="0" to-layer="257" to-port="1" />
6620 <edge from-layer="257" from-port="2" to-layer="259" to-port="0" />
6621 <edge from-layer="258" from-port="0" to-layer="259" to-port="1" />
6622 <edge from-layer="259" from-port="2" to-layer="261" to-port="0" />
6623 <edge from-layer="260" from-port="0" to-layer="261" to-port="1" />
6624 <edge from-layer="261" from-port="2" to-layer="262" to-port="0" />
6625 <edge from-layer="262" from-port="2" to-layer="264" to-port="0" />
6626 <edge from-layer="263" from-port="0" to-layer="264" to-port="1" />
6627 <edge from-layer="264" from-port="2" to-layer="266" to-port="0" />
6628 <edge from-layer="265" from-port="0" to-layer="266" to-port="1" />
6629 <edge from-layer="266" from-port="2" to-layer="268" to-port="0" />
6630 <edge from-layer="267" from-port="0" to-layer="268" to-port="1" />
6631 <edge from-layer="268" from-port="2" to-layer="270" to-port="0" />
6632 <edge from-layer="268" from-port="2" to-layer="278" to-port="1" />
6633 <edge from-layer="269" from-port="0" to-layer="270" to-port="1" />
6634 <edge from-layer="270" from-port="2" to-layer="272" to-port="0" />
6635 <edge from-layer="271" from-port="0" to-layer="272" to-port="1" />
6636 <edge from-layer="272" from-port="2" to-layer="273" to-port="0" />
6637 <edge from-layer="273" from-port="1" to-layer="275" to-port="0" />
6638 <edge from-layer="274" from-port="0" to-layer="275" to-port="1" />
6639 <edge from-layer="275" from-port="2" to-layer="277" to-port="0" />
6640 <edge from-layer="276" from-port="0" to-layer="277" to-port="1" />
6641 <edge from-layer="277" from-port="2" to-layer="278" to-port="0" />
6642 <edge from-layer="278" from-port="2" to-layer="280" to-port="0" />
6643 <edge from-layer="279" from-port="0" to-layer="280" to-port="1" />
6644 <edge from-layer="280" from-port="2" to-layer="282" to-port="0" />
6645 <edge from-layer="281" from-port="0" to-layer="282" to-port="1" />
6646 <edge from-layer="282" from-port="2" to-layer="284" to-port="0" />
6647 <edge from-layer="283" from-port="0" to-layer="284" to-port="1" />
6648 <edge from-layer="284" from-port="2" to-layer="294" to-port="0" />
6649 <edge from-layer="284" from-port="2" to-layer="286" to-port="0" />
6650 <edge from-layer="284" from-port="2" to-layer="302" to-port="0" />
6651 <edge from-layer="284" from-port="2" to-layer="318" to-port="1" />
6652 <edge from-layer="285" from-port="0" to-layer="286" to-port="1" />
6653 <edge from-layer="286" from-port="2" to-layer="288" to-port="0" />
6654 <edge from-layer="287" from-port="0" to-layer="288" to-port="1" />
6655 <edge from-layer="288" from-port="2" to-layer="290" to-port="0" />
6656 <edge from-layer="289" from-port="0" to-layer="290" to-port="1" />
6657 <edge from-layer="290" from-port="2" to-layer="292" to-port="0" />
6658 <edge from-layer="291" from-port="0" to-layer="292" to-port="1" />
6659 <edge from-layer="292" from-port="2" to-layer="309" to-port="0" />
6660 <edge from-layer="293" from-port="0" to-layer="294" to-port="1" />
6661 <edge from-layer="294" from-port="2" to-layer="296" to-port="0" />
6662 <edge from-layer="295" from-port="0" to-layer="296" to-port="1" />
6663 <edge from-layer="296" from-port="2" to-layer="298" to-port="0" />
6664 <edge from-layer="297" from-port="0" to-layer="298" to-port="1" />
6665 <edge from-layer="298" from-port="2" to-layer="300" to-port="0" />
6666 <edge from-layer="299" from-port="0" to-layer="300" to-port="1" />
6667 <edge from-layer="300" from-port="2" to-layer="309" to-port="1" />
6668 <edge from-layer="301" from-port="0" to-layer="302" to-port="1" />
6669 <edge from-layer="302" from-port="2" to-layer="304" to-port="0" />
6670 <edge from-layer="303" from-port="0" to-layer="304" to-port="1" />
6671 <edge from-layer="304" from-port="2" to-layer="306" to-port="0" />
6672 <edge from-layer="305" from-port="0" to-layer="306" to-port="1" />
6673 <edge from-layer="306" from-port="2" to-layer="308" to-port="0" />
6674 <edge from-layer="307" from-port="0" to-layer="308" to-port="1" />
6675 <edge from-layer="308" from-port="2" to-layer="309" to-port="2" />
6676 <edge from-layer="309" from-port="4" to-layer="311" to-port="0" />
6677 <edge from-layer="310" from-port="0" to-layer="311" to-port="1" />
6678 <edge from-layer="311" from-port="2" to-layer="313" to-port="0" />
6679 <edge from-layer="312" from-port="0" to-layer="313" to-port="1" />
6680 <edge from-layer="313" from-port="2" to-layer="315" to-port="0" />
6681 <edge from-layer="314" from-port="0" to-layer="315" to-port="1" />
6682 <edge from-layer="315" from-port="2" to-layer="317" to-port="0" />
6683 <edge from-layer="316" from-port="0" to-layer="317" to-port="1" />
6684 <edge from-layer="317" from-port="2" to-layer="318" to-port="0" />
6685 <edge from-layer="318" from-port="2" to-layer="320" to-port="0" />
6686 <edge from-layer="319" from-port="0" to-layer="320" to-port="1" />
6687 <edge from-layer="320" from-port="2" to-layer="322" to-port="0" />
6688 <edge from-layer="321" from-port="0" to-layer="322" to-port="1" />
6689 <edge from-layer="322" from-port="2" to-layer="324" to-port="0" />
6690 <edge from-layer="323" from-port="0" to-layer="324" to-port="1" />
6691 <edge from-layer="324" from-port="2" to-layer="326" to-port="0" />
6692 <edge from-layer="324" from-port="2" to-layer="334" to-port="1" />
6693 <edge from-layer="325" from-port="0" to-layer="326" to-port="1" />
6694 <edge from-layer="326" from-port="2" to-layer="328" to-port="0" />
6695 <edge from-layer="327" from-port="0" to-layer="328" to-port="1" />
6696 <edge from-layer="328" from-port="2" to-layer="329" to-port="0" />
6697 <edge from-layer="329" from-port="1" to-layer="331" to-port="0" />
6698 <edge from-layer="330" from-port="0" to-layer="331" to-port="1" />
6699 <edge from-layer="331" from-port="2" to-layer="333" to-port="0" />
6700 <edge from-layer="332" from-port="0" to-layer="333" to-port="1" />
6701 <edge from-layer="333" from-port="2" to-layer="334" to-port="0" />
6702 <edge from-layer="334" from-port="2" to-layer="336" to-port="0" />
6703 <edge from-layer="335" from-port="0" to-layer="336" to-port="1" />
6704 <edge from-layer="336" from-port="2" to-layer="338" to-port="0" />
6705 <edge from-layer="337" from-port="0" to-layer="338" to-port="1" />
6706 <edge from-layer="338" from-port="2" to-layer="340" to-port="0" />
6707 <edge from-layer="339" from-port="0" to-layer="340" to-port="1" />
6708 <edge from-layer="340" from-port="2" to-layer="342" to-port="0" />
6709 <edge from-layer="340" from-port="2" to-layer="350" to-port="0" />
6710 <edge from-layer="340" from-port="2" to-layer="374" to-port="1" />
6711 <edge from-layer="340" from-port="2" to-layer="358" to-port="0" />
6712 <edge from-layer="341" from-port="0" to-layer="342" to-port="1" />
6713 <edge from-layer="342" from-port="2" to-layer="344" to-port="0" />
6714 <edge from-layer="343" from-port="0" to-layer="344" to-port="1" />
6715 <edge from-layer="344" from-port="2" to-layer="346" to-port="0" />
6716 <edge from-layer="345" from-port="0" to-layer="346" to-port="1" />
6717 <edge from-layer="346" from-port="2" to-layer="348" to-port="0" />
6718 <edge from-layer="347" from-port="0" to-layer="348" to-port="1" />
6719 <edge from-layer="348" from-port="2" to-layer="365" to-port="0" />
6720 <edge from-layer="349" from-port="0" to-layer="350" to-port="1" />
6721 <edge from-layer="350" from-port="2" to-layer="352" to-port="0" />
6722 <edge from-layer="351" from-port="0" to-layer="352" to-port="1" />
6723 <edge from-layer="352" from-port="2" to-layer="354" to-port="0" />
6724 <edge from-layer="353" from-port="0" to-layer="354" to-port="1" />
6725 <edge from-layer="354" from-port="2" to-layer="356" to-port="0" />
6726 <edge from-layer="355" from-port="0" to-layer="356" to-port="1" />
6727 <edge from-layer="356" from-port="2" to-layer="365" to-port="1" />
6728 <edge from-layer="357" from-port="0" to-layer="358" to-port="1" />
6729 <edge from-layer="358" from-port="2" to-layer="360" to-port="0" />
6730 <edge from-layer="359" from-port="0" to-layer="360" to-port="1" />
6731 <edge from-layer="360" from-port="2" to-layer="362" to-port="0" />
6732 <edge from-layer="361" from-port="0" to-layer="362" to-port="1" />
6733 <edge from-layer="362" from-port="2" to-layer="364" to-port="0" />
6734 <edge from-layer="363" from-port="0" to-layer="364" to-port="1" />
6735 <edge from-layer="364" from-port="2" to-layer="365" to-port="2" />
6736 <edge from-layer="365" from-port="4" to-layer="367" to-port="0" />
6737 <edge from-layer="366" from-port="0" to-layer="367" to-port="1" />
6738 <edge from-layer="367" from-port="2" to-layer="369" to-port="0" />
6739 <edge from-layer="368" from-port="0" to-layer="369" to-port="1" />
6740 <edge from-layer="369" from-port="2" to-layer="371" to-port="0" />
6741 <edge from-layer="370" from-port="0" to-layer="371" to-port="1" />
6742 <edge from-layer="371" from-port="2" to-layer="373" to-port="0" />
6743 <edge from-layer="372" from-port="0" to-layer="373" to-port="1" />
6744 <edge from-layer="373" from-port="2" to-layer="374" to-port="0" />
6745 <edge from-layer="374" from-port="2" to-layer="376" to-port="0" />
6746 <edge from-layer="375" from-port="0" to-layer="376" to-port="1" />
6747 <edge from-layer="376" from-port="2" to-layer="378" to-port="0" />
6748 <edge from-layer="377" from-port="0" to-layer="378" to-port="1" />
6749 <edge from-layer="378" from-port="2" to-layer="380" to-port="0" />
6750 <edge from-layer="379" from-port="0" to-layer="380" to-port="1" />
6751 <edge from-layer="380" from-port="2" to-layer="382" to-port="0" />
6752 <edge from-layer="380" from-port="2" to-layer="390" to-port="1" />
6753 <edge from-layer="381" from-port="0" to-layer="382" to-port="1" />
6754 <edge from-layer="382" from-port="2" to-layer="384" to-port="0" />
6755 <edge from-layer="383" from-port="0" to-layer="384" to-port="1" />
6756 <edge from-layer="384" from-port="2" to-layer="385" to-port="0" />
6757 <edge from-layer="385" from-port="1" to-layer="387" to-port="0" />
6758 <edge from-layer="386" from-port="0" to-layer="387" to-port="1" />
6759 <edge from-layer="387" from-port="2" to-layer="389" to-port="0" />
6760 <edge from-layer="388" from-port="0" to-layer="389" to-port="1" />
6761 <edge from-layer="389" from-port="2" to-layer="390" to-port="0" />
6762 <edge from-layer="390" from-port="2" to-layer="392" to-port="0" />
6763 <edge from-layer="391" from-port="0" to-layer="392" to-port="1" />
6764 <edge from-layer="392" from-port="2" to-layer="394" to-port="0" />
6765 <edge from-layer="393" from-port="0" to-layer="394" to-port="1" />
6766 <edge from-layer="394" from-port="2" to-layer="396" to-port="0" />
6767 <edge from-layer="395" from-port="0" to-layer="396" to-port="1" />
6768 <edge from-layer="396" from-port="2" to-layer="399" to-port="0" />
6769 <edge from-layer="397" from-port="0" to-layer="399" to-port="1" />
6770 <edge from-layer="398" from-port="0" to-layer="399" to-port="2" />
6771 <edge from-layer="399" from-port="3" to-layer="401" to-port="0" />
6772 <edge from-layer="400" from-port="0" to-layer="401" to-port="1" />
6773 <edge from-layer="401" from-port="2" to-layer="403" to-port="0" />
6774 <edge from-layer="402" from-port="0" to-layer="403" to-port="1" />
6775 <edge from-layer="403" from-port="2" to-layer="404" to-port="0" />
6776 <edge from-layer="404" from-port="1" to-layer="406" to-port="0" />
6777 <edge from-layer="405" from-port="0" to-layer="406" to-port="1" />
6778 <edge from-layer="406" from-port="2" to-layer="408" to-port="0" />
6779 <edge from-layer="407" from-port="0" to-layer="408" to-port="1" />
6780 <edge from-layer="408" from-port="2" to-layer="409" to-port="0" />
6781 </edges>
6782 <rt_info>
6783 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
6784 <conversion_parameters>
6785 <framework value="pytorch" />
6786 <is_python_object value="True" />
6787 </conversion_parameters>
6788 <optimum>
6789 <optimum_intel_version value="1.20.1" />
6790 <optimum_version value="1.24.0" />
6791 <pytorch_version value="2.6.0+cu124" />
6792 <transformers_version value="4.52.0.dev0" />
6793 </optimum>
6794 </rt_info>
6795 </net>
6796