openvino/openvino_model.xml
143.0 KB · 4724 lines · xml Raw
1 <?xml version="1.0"?>
2 <net name="Model385" version="11">
3 <layers>
4 <layer id="2" name="input_ids" type="Parameter" version="opset1">
5 <data shape="?,?" element_type="i64" />
6 <output>
7 <port id="0" precision="I64" names="input_ids">
8 <dim>-1</dim>
9 <dim>-1</dim>
10 </port>
11 </output>
12 </layer>
13 <layer id="1" name="attention_mask" type="Parameter" version="opset1">
14 <data shape="?,?" element_type="i64" />
15 <output>
16 <port id="0" precision="I64" names="attention_mask">
17 <dim>-1</dim>
18 <dim>-1</dim>
19 </port>
20 </output>
21 </layer>
22 <layer id="0" name="token_type_ids" type="Parameter" version="opset1">
23 <data shape="?,?" element_type="i64" />
24 <output>
25 <port id="0" precision="I64" names="token_type_ids">
26 <dim>-1</dim>
27 <dim>-1</dim>
28 </port>
29 </output>
30 </layer>
31 <layer id="3" name="self.bert.embeddings.word_embeddings.weight" type="Const" version="opset1">
32 <data element_type="f32" shape="30522, 384" offset="0" size="46881792" />
33 <output>
34 <port id="0" precision="FP32" names="self.bert.embeddings.word_embeddings.weight">
35 <dim>30522</dim>
36 <dim>384</dim>
37 </port>
38 </output>
39 </layer>
40 <layer id="4" name="__module.bert.embeddings.word_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
41 <data destination_type="i32" />
42 <input>
43 <port id="0" precision="I64">
44 <dim>-1</dim>
45 <dim>-1</dim>
46 </port>
47 </input>
48 <output>
49 <port id="1" precision="I32">
50 <dim>-1</dim>
51 <dim>-1</dim>
52 </port>
53 </output>
54 </layer>
55 <layer id="5" name="__module.bert.embeddings.word_embeddings/aten::embedding/Constant" type="Const" version="opset1">
56 <data element_type="i32" shape="" offset="46881792" size="4" />
57 <output>
58 <port id="0" precision="I32" />
59 </output>
60 </layer>
61 <layer id="6" name="__module.bert.embeddings.word_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
62 <data batch_dims="0" />
63 <input>
64 <port id="0" precision="FP32">
65 <dim>30522</dim>
66 <dim>384</dim>
67 </port>
68 <port id="1" precision="I32">
69 <dim>-1</dim>
70 <dim>-1</dim>
71 </port>
72 <port id="2" precision="I32" />
73 </input>
74 <output>
75 <port id="3" precision="FP32" names="48,inputs_embeds">
76 <dim>-1</dim>
77 <dim>-1</dim>
78 <dim>384</dim>
79 </port>
80 </output>
81 </layer>
82 <layer id="7" name="self.bert.embeddings.token_type_embeddings.weight" type="Const" version="opset1">
83 <data element_type="f32" shape="2, 384" offset="46881796" size="3072" />
84 <output>
85 <port id="0" precision="FP32" names="self.bert.embeddings.token_type_embeddings.weight">
86 <dim>2</dim>
87 <dim>384</dim>
88 </port>
89 </output>
90 </layer>
91 <layer id="8" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
92 <data destination_type="i32" />
93 <input>
94 <port id="0" precision="I64">
95 <dim>-1</dim>
96 <dim>-1</dim>
97 </port>
98 </input>
99 <output>
100 <port id="1" precision="I32">
101 <dim>-1</dim>
102 <dim>-1</dim>
103 </port>
104 </output>
105 </layer>
106 <layer id="9" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Constant" type="Const" version="opset1">
107 <data element_type="i32" shape="" offset="46881792" size="4" />
108 <output>
109 <port id="0" precision="I32" />
110 </output>
111 </layer>
112 <layer id="10" name="__module.bert.embeddings.token_type_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
113 <data batch_dims="0" />
114 <input>
115 <port id="0" precision="FP32">
116 <dim>2</dim>
117 <dim>384</dim>
118 </port>
119 <port id="1" precision="I32">
120 <dim>-1</dim>
121 <dim>-1</dim>
122 </port>
123 <port id="2" precision="I32" />
124 </input>
125 <output>
126 <port id="3" precision="FP32" names="50,token_type_embeddings.1">
127 <dim>-1</dim>
128 <dim>-1</dim>
129 <dim>384</dim>
130 </port>
131 </output>
132 </layer>
133 <layer id="11" name="__module.bert.embeddings/aten::add/Add" type="Add" version="opset1">
134 <data auto_broadcast="numpy" />
135 <input>
136 <port id="0" precision="FP32">
137 <dim>-1</dim>
138 <dim>-1</dim>
139 <dim>384</dim>
140 </port>
141 <port id="1" precision="FP32">
142 <dim>-1</dim>
143 <dim>-1</dim>
144 <dim>384</dim>
145 </port>
146 </input>
147 <output>
148 <port id="2" precision="FP32" names="51_1">
149 <dim>-1</dim>
150 <dim>-1</dim>
151 <dim>384</dim>
152 </port>
153 </output>
154 </layer>
155 <layer id="12" name="self.bert.embeddings.position_embeddings.weight" type="Const" version="opset1">
156 <data element_type="f32" shape="512, 384" offset="46884868" size="786432" />
157 <output>
158 <port id="0" precision="FP32" names="self.bert.embeddings.position_embeddings.weight">
159 <dim>512</dim>
160 <dim>384</dim>
161 </port>
162 </output>
163 </layer>
164 <layer id="13" name="__module.bert.embeddings/aten::slice/Slice" type="Const" version="opset1">
165 <data element_type="i64" shape="1, 512" offset="47671300" size="4096" />
166 <output>
167 <port id="0" precision="I64" names="45">
168 <dim>1</dim>
169 <dim>512</dim>
170 </port>
171 </output>
172 </layer>
173 <layer id="14" name="__module.bert.embeddings/aten::slice/Reshape" type="Const" version="opset1">
174 <data element_type="i64" shape="1" offset="47675396" size="8" />
175 <output>
176 <port id="0" precision="I64">
177 <dim>1</dim>
178 </port>
179 </output>
180 </layer>
181 <layer id="15" name="ShapeOf_305961" type="ShapeOf" version="opset3">
182 <data output_type="i64" />
183 <input>
184 <port id="0" precision="I64">
185 <dim>-1</dim>
186 <dim>-1</dim>
187 </port>
188 </input>
189 <output>
190 <port id="1" precision="I64">
191 <dim>2</dim>
192 </port>
193 </output>
194 </layer>
195 <layer id="16" name="Constant_306078" type="Const" version="opset1">
196 <data element_type="i64" shape="1" offset="47675404" size="8" />
197 <output>
198 <port id="0" precision="I64">
199 <dim>1</dim>
200 </port>
201 </output>
202 </layer>
203 <layer id="17" name="Constant_305963" type="Const" version="opset1">
204 <data element_type="i64" shape="" offset="47675396" size="8" />
205 <output>
206 <port id="0" precision="I64" />
207 </output>
208 </layer>
209 <layer id="18" name="Gather_305964" type="Gather" version="opset8">
210 <data batch_dims="0" />
211 <input>
212 <port id="0" precision="I64">
213 <dim>2</dim>
214 </port>
215 <port id="1" precision="I64">
216 <dim>1</dim>
217 </port>
218 <port id="2" precision="I64" />
219 </input>
220 <output>
221 <port id="3" precision="I64" names="35,41,43,44,61">
222 <dim>1</dim>
223 </port>
224 </output>
225 </layer>
226 <layer id="19" name="__module.bert.embeddings/aten::slice/Reshape_2" type="Const" version="opset1">
227 <data element_type="i64" shape="1" offset="47675404" size="8" />
228 <output>
229 <port id="0" precision="I64">
230 <dim>1</dim>
231 </port>
232 </output>
233 </layer>
234 <layer id="20" name="__module.bert.embeddings/aten::slice/Reshape_3" type="Const" version="opset1">
235 <data element_type="i64" shape="1" offset="47675404" size="8" />
236 <output>
237 <port id="0" precision="I64">
238 <dim>1</dim>
239 </port>
240 </output>
241 </layer>
242 <layer id="21" name="__module.bert.embeddings/aten::slice/Slice_1" type="Slice" version="opset8">
243 <input>
244 <port id="0" precision="I64">
245 <dim>1</dim>
246 <dim>512</dim>
247 </port>
248 <port id="1" precision="I64">
249 <dim>1</dim>
250 </port>
251 <port id="2" precision="I64">
252 <dim>1</dim>
253 </port>
254 <port id="3" precision="I64">
255 <dim>1</dim>
256 </port>
257 <port id="4" precision="I64">
258 <dim>1</dim>
259 </port>
260 </input>
261 <output>
262 <port id="5" precision="I64" names="46">
263 <dim>1</dim>
264 <dim>-1</dim>
265 </port>
266 </output>
267 </layer>
268 <layer id="22" name="__module.bert.embeddings.position_embeddings/aten::embedding/Convert" type="Convert" version="opset1">
269 <data destination_type="i32" />
270 <input>
271 <port id="0" precision="I64">
272 <dim>1</dim>
273 <dim>-1</dim>
274 </port>
275 </input>
276 <output>
277 <port id="1" precision="I32">
278 <dim>1</dim>
279 <dim>-1</dim>
280 </port>
281 </output>
282 </layer>
283 <layer id="23" name="__module.bert.embeddings.position_embeddings/aten::embedding/Constant" type="Const" version="opset1">
284 <data element_type="i32" shape="" offset="46881792" size="4" />
285 <output>
286 <port id="0" precision="I32" />
287 </output>
288 </layer>
289 <layer id="24" name="__module.bert.embeddings.position_embeddings/aten::embedding/Gather" type="Gather" version="opset8">
290 <data batch_dims="0" />
291 <input>
292 <port id="0" precision="FP32">
293 <dim>512</dim>
294 <dim>384</dim>
295 </port>
296 <port id="1" precision="I32">
297 <dim>1</dim>
298 <dim>-1</dim>
299 </port>
300 <port id="2" precision="I32" />
301 </input>
302 <output>
303 <port id="3" precision="FP32" names="53,position_embeddings.1">
304 <dim>1</dim>
305 <dim>-1</dim>
306 <dim>384</dim>
307 </port>
308 </output>
309 </layer>
310 <layer id="25" name="__module.bert.embeddings/aten::add_/Add" type="Add" version="opset1">
311 <data auto_broadcast="numpy" />
312 <input>
313 <port id="0" precision="FP32">
314 <dim>-1</dim>
315 <dim>-1</dim>
316 <dim>384</dim>
317 </port>
318 <port id="1" precision="FP32">
319 <dim>1</dim>
320 <dim>-1</dim>
321 <dim>384</dim>
322 </port>
323 </input>
324 <output>
325 <port id="2" precision="FP32" names="51,embeddings.1">
326 <dim>-1</dim>
327 <dim>-1</dim>
328 <dim>384</dim>
329 </port>
330 </output>
331 </layer>
332 <layer id="26" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
333 <data element_type="i32" shape="1" offset="47675412" size="4" />
334 <output>
335 <port id="0" precision="I32">
336 <dim>1</dim>
337 </port>
338 </output>
339 </layer>
340 <layer id="27" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
341 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
342 <input>
343 <port id="0" precision="FP32">
344 <dim>-1</dim>
345 <dim>-1</dim>
346 <dim>384</dim>
347 </port>
348 <port id="1" precision="I32">
349 <dim>1</dim>
350 </port>
351 </input>
352 <output>
353 <port id="2" precision="FP32">
354 <dim>-1</dim>
355 <dim>-1</dim>
356 <dim>384</dim>
357 </port>
358 </output>
359 </layer>
360 <layer id="28" name="Constant_305869" type="Const" version="opset1">
361 <data element_type="f32" shape="1, 1, 384" offset="47675416" size="1536" />
362 <output>
363 <port id="0" precision="FP32">
364 <dim>1</dim>
365 <dim>1</dim>
366 <dim>384</dim>
367 </port>
368 </output>
369 </layer>
370 <layer id="29" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
371 <data auto_broadcast="numpy" />
372 <input>
373 <port id="0" precision="FP32">
374 <dim>-1</dim>
375 <dim>-1</dim>
376 <dim>384</dim>
377 </port>
378 <port id="1" precision="FP32">
379 <dim>1</dim>
380 <dim>1</dim>
381 <dim>384</dim>
382 </port>
383 </input>
384 <output>
385 <port id="2" precision="FP32">
386 <dim>-1</dim>
387 <dim>-1</dim>
388 <dim>384</dim>
389 </port>
390 </output>
391 </layer>
392 <layer id="30" name="Constant_305870" type="Const" version="opset1">
393 <data element_type="f32" shape="1, 1, 384" offset="47676952" size="1536" />
394 <output>
395 <port id="0" precision="FP32">
396 <dim>1</dim>
397 <dim>1</dim>
398 <dim>384</dim>
399 </port>
400 </output>
401 </layer>
402 <layer id="31" name="__module.bert.embeddings.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
403 <data auto_broadcast="numpy" />
404 <input>
405 <port id="0" precision="FP32">
406 <dim>-1</dim>
407 <dim>-1</dim>
408 <dim>384</dim>
409 </port>
410 <port id="1" precision="FP32">
411 <dim>1</dim>
412 <dim>1</dim>
413 <dim>384</dim>
414 </port>
415 </input>
416 <output>
417 <port id="2" precision="FP32" names="58,input.1">
418 <dim>-1</dim>
419 <dim>-1</dim>
420 <dim>384</dim>
421 </port>
422 </output>
423 </layer>
424 <layer id="32" name="self.bert.encoder.layer.0.attention.self.query.weight" type="Const" version="opset1">
425 <data element_type="f32" shape="384, 384" offset="47678488" size="589824" />
426 <output>
427 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.self.query.weight">
428 <dim>384</dim>
429 <dim>384</dim>
430 </port>
431 </output>
432 </layer>
433 <layer id="33" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
434 <data transpose_a="false" transpose_b="true" />
435 <input>
436 <port id="0" precision="FP32">
437 <dim>-1</dim>
438 <dim>-1</dim>
439 <dim>384</dim>
440 </port>
441 <port id="1" precision="FP32">
442 <dim>384</dim>
443 <dim>384</dim>
444 </port>
445 </input>
446 <output>
447 <port id="2" precision="FP32">
448 <dim>-1</dim>
449 <dim>-1</dim>
450 <dim>384</dim>
451 </port>
452 </output>
453 </layer>
454 <layer id="34" name="Constant_305871" type="Const" version="opset1">
455 <data element_type="f32" shape="1, 1, 384" offset="48268312" size="1536" />
456 <output>
457 <port id="0" precision="FP32">
458 <dim>1</dim>
459 <dim>1</dim>
460 <dim>384</dim>
461 </port>
462 </output>
463 </layer>
464 <layer id="35" name="__module.bert.encoder.layer.0.attention.self.query/aten::linear/Add" type="Add" version="opset1">
465 <data auto_broadcast="numpy" />
466 <input>
467 <port id="0" precision="FP32">
468 <dim>-1</dim>
469 <dim>-1</dim>
470 <dim>384</dim>
471 </port>
472 <port id="1" precision="FP32">
473 <dim>1</dim>
474 <dim>1</dim>
475 <dim>384</dim>
476 </port>
477 </input>
478 <output>
479 <port id="2" precision="FP32" names="92,x.1">
480 <dim>-1</dim>
481 <dim>-1</dim>
482 <dim>384</dim>
483 </port>
484 </output>
485 </layer>
486 <layer id="36" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
487 <data element_type="i64" shape="4" offset="48269848" size="32" />
488 <output>
489 <port id="0" precision="I64">
490 <dim>4</dim>
491 </port>
492 </output>
493 </layer>
494 <layer id="37" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
495 <data special_zero="true" />
496 <input>
497 <port id="0" precision="FP32">
498 <dim>-1</dim>
499 <dim>-1</dim>
500 <dim>384</dim>
501 </port>
502 <port id="1" precision="I64">
503 <dim>4</dim>
504 </port>
505 </input>
506 <output>
507 <port id="2" precision="FP32" names="96,x.3">
508 <dim>-1</dim>
509 <dim>-1</dim>
510 <dim>12</dim>
511 <dim>32</dim>
512 </port>
513 </output>
514 </layer>
515 <layer id="38" name="Constant_301060" type="Const" version="opset1">
516 <data element_type="i64" shape="4" offset="48269880" size="32" />
517 <output>
518 <port id="0" precision="I64" names="97">
519 <dim>4</dim>
520 </port>
521 </output>
522 </layer>
523 <layer id="39" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
524 <input>
525 <port id="0" precision="FP32">
526 <dim>-1</dim>
527 <dim>-1</dim>
528 <dim>12</dim>
529 <dim>32</dim>
530 </port>
531 <port id="1" precision="I64">
532 <dim>4</dim>
533 </port>
534 </input>
535 <output>
536 <port id="2" precision="FP32" names="98">
537 <dim>-1</dim>
538 <dim>12</dim>
539 <dim>-1</dim>
540 <dim>32</dim>
541 </port>
542 </output>
543 </layer>
544 <layer id="40" name="self.bert.encoder.layer.0.attention.self.key.weight" type="Const" version="opset1">
545 <data element_type="f32" shape="384, 384" offset="48269912" size="589824" />
546 <output>
547 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.self.key.weight">
548 <dim>384</dim>
549 <dim>384</dim>
550 </port>
551 </output>
552 </layer>
553 <layer id="41" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
554 <data transpose_a="false" transpose_b="true" />
555 <input>
556 <port id="0" precision="FP32">
557 <dim>-1</dim>
558 <dim>-1</dim>
559 <dim>384</dim>
560 </port>
561 <port id="1" precision="FP32">
562 <dim>384</dim>
563 <dim>384</dim>
564 </port>
565 </input>
566 <output>
567 <port id="2" precision="FP32">
568 <dim>-1</dim>
569 <dim>-1</dim>
570 <dim>384</dim>
571 </port>
572 </output>
573 </layer>
574 <layer id="42" name="Constant_305872" type="Const" version="opset1">
575 <data element_type="f32" shape="1, 1, 384" offset="48859736" size="1536" />
576 <output>
577 <port id="0" precision="FP32">
578 <dim>1</dim>
579 <dim>1</dim>
580 <dim>384</dim>
581 </port>
582 </output>
583 </layer>
584 <layer id="43" name="__module.bert.encoder.layer.0.attention.self.key/aten::linear/Add" type="Add" version="opset1">
585 <data auto_broadcast="numpy" />
586 <input>
587 <port id="0" precision="FP32">
588 <dim>-1</dim>
589 <dim>-1</dim>
590 <dim>384</dim>
591 </port>
592 <port id="1" precision="FP32">
593 <dim>1</dim>
594 <dim>1</dim>
595 <dim>384</dim>
596 </port>
597 </input>
598 <output>
599 <port id="2" precision="FP32" names="101,x.5">
600 <dim>-1</dim>
601 <dim>-1</dim>
602 <dim>384</dim>
603 </port>
604 </output>
605 </layer>
606 <layer id="44" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
607 <data element_type="i64" shape="4" offset="48269848" size="32" />
608 <output>
609 <port id="0" precision="I64">
610 <dim>4</dim>
611 </port>
612 </output>
613 </layer>
614 <layer id="45" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
615 <data special_zero="true" />
616 <input>
617 <port id="0" precision="FP32">
618 <dim>-1</dim>
619 <dim>-1</dim>
620 <dim>384</dim>
621 </port>
622 <port id="1" precision="I64">
623 <dim>4</dim>
624 </port>
625 </input>
626 <output>
627 <port id="2" precision="FP32" names="105,x.7">
628 <dim>-1</dim>
629 <dim>-1</dim>
630 <dim>12</dim>
631 <dim>32</dim>
632 </port>
633 </output>
634 </layer>
635 <layer id="46" name="Constant_301085" type="Const" version="opset1">
636 <data element_type="i64" shape="4" offset="48269880" size="32" />
637 <output>
638 <port id="0" precision="I64" names="106">
639 <dim>4</dim>
640 </port>
641 </output>
642 </layer>
643 <layer id="47" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
644 <input>
645 <port id="0" precision="FP32">
646 <dim>-1</dim>
647 <dim>-1</dim>
648 <dim>12</dim>
649 <dim>32</dim>
650 </port>
651 <port id="1" precision="I64">
652 <dim>4</dim>
653 </port>
654 </input>
655 <output>
656 <port id="2" precision="FP32" names="107">
657 <dim>-1</dim>
658 <dim>12</dim>
659 <dim>-1</dim>
660 <dim>32</dim>
661 </port>
662 </output>
663 </layer>
664 <layer id="48" name="self.bert.encoder.layer.0.attention.self.value.weight" type="Const" version="opset1">
665 <data element_type="f32" shape="384, 384" offset="48861272" size="589824" />
666 <output>
667 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.self.value.weight">
668 <dim>384</dim>
669 <dim>384</dim>
670 </port>
671 </output>
672 </layer>
673 <layer id="49" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
674 <data transpose_a="false" transpose_b="true" />
675 <input>
676 <port id="0" precision="FP32">
677 <dim>-1</dim>
678 <dim>-1</dim>
679 <dim>384</dim>
680 </port>
681 <port id="1" precision="FP32">
682 <dim>384</dim>
683 <dim>384</dim>
684 </port>
685 </input>
686 <output>
687 <port id="2" precision="FP32">
688 <dim>-1</dim>
689 <dim>-1</dim>
690 <dim>384</dim>
691 </port>
692 </output>
693 </layer>
694 <layer id="50" name="Constant_305873" type="Const" version="opset1">
695 <data element_type="f32" shape="1, 1, 384" offset="49451096" size="1536" />
696 <output>
697 <port id="0" precision="FP32">
698 <dim>1</dim>
699 <dim>1</dim>
700 <dim>384</dim>
701 </port>
702 </output>
703 </layer>
704 <layer id="51" name="__module.bert.encoder.layer.0.attention.self.value/aten::linear/Add" type="Add" version="opset1">
705 <data auto_broadcast="numpy" />
706 <input>
707 <port id="0" precision="FP32">
708 <dim>-1</dim>
709 <dim>-1</dim>
710 <dim>384</dim>
711 </port>
712 <port id="1" precision="FP32">
713 <dim>1</dim>
714 <dim>1</dim>
715 <dim>384</dim>
716 </port>
717 </input>
718 <output>
719 <port id="2" precision="FP32" names="110,x.9">
720 <dim>-1</dim>
721 <dim>-1</dim>
722 <dim>384</dim>
723 </port>
724 </output>
725 </layer>
726 <layer id="52" name="__module.bert.encoder.layer.0.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
727 <data element_type="i64" shape="4" offset="48269848" size="32" />
728 <output>
729 <port id="0" precision="I64">
730 <dim>4</dim>
731 </port>
732 </output>
733 </layer>
734 <layer id="53" name="__module.bert.encoder.layer.0.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
735 <data special_zero="true" />
736 <input>
737 <port id="0" precision="FP32">
738 <dim>-1</dim>
739 <dim>-1</dim>
740 <dim>384</dim>
741 </port>
742 <port id="1" precision="I64">
743 <dim>4</dim>
744 </port>
745 </input>
746 <output>
747 <port id="2" precision="FP32" names="114,x.11">
748 <dim>-1</dim>
749 <dim>-1</dim>
750 <dim>12</dim>
751 <dim>32</dim>
752 </port>
753 </output>
754 </layer>
755 <layer id="54" name="Constant_301110" type="Const" version="opset1">
756 <data element_type="i64" shape="4" offset="48269880" size="32" />
757 <output>
758 <port id="0" precision="I64" names="115">
759 <dim>4</dim>
760 </port>
761 </output>
762 </layer>
763 <layer id="55" name="__module.bert.encoder.layer.0.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
764 <input>
765 <port id="0" precision="FP32">
766 <dim>-1</dim>
767 <dim>-1</dim>
768 <dim>12</dim>
769 <dim>32</dim>
770 </port>
771 <port id="1" precision="I64">
772 <dim>4</dim>
773 </port>
774 </input>
775 <output>
776 <port id="2" precision="FP32" names="116">
777 <dim>-1</dim>
778 <dim>12</dim>
779 <dim>-1</dim>
780 <dim>32</dim>
781 </port>
782 </output>
783 </layer>
784 <layer id="56" name="Constant_305875" type="Const" version="opset1">
785 <data element_type="f32" shape="1, 1, 1, 1" offset="49452632" size="4" />
786 <output>
787 <port id="0" precision="FP32">
788 <dim>1</dim>
789 <dim>1</dim>
790 <dim>1</dim>
791 <dim>1</dim>
792 </port>
793 </output>
794 </layer>
795 <layer id="57" name="31" type="Const" version="opset1">
796 <data element_type="i64" shape="" offset="47675404" size="8" />
797 <output>
798 <port id="0" precision="I64" names="31" />
799 </output>
800 </layer>
801 <layer id="58" name="__module.bert/aten::unsqueeze/Unsqueeze" type="Unsqueeze" version="opset1">
802 <input>
803 <port id="0" precision="I64">
804 <dim>-1</dim>
805 <dim>-1</dim>
806 </port>
807 <port id="1" precision="I64" />
808 </input>
809 <output>
810 <port id="2" precision="I64" names="63">
811 <dim>-1</dim>
812 <dim>1</dim>
813 <dim>-1</dim>
814 </port>
815 </output>
816 </layer>
817 <layer id="59" name="21" type="Const" version="opset1">
818 <data element_type="i64" shape="" offset="49452636" size="8" />
819 <output>
820 <port id="0" precision="I64" names="21" />
821 </output>
822 </layer>
823 <layer id="60" name="__module.bert/aten::unsqueeze/Unsqueeze_1" type="Unsqueeze" version="opset1">
824 <input>
825 <port id="0" precision="I64">
826 <dim>-1</dim>
827 <dim>1</dim>
828 <dim>-1</dim>
829 </port>
830 <port id="1" precision="I64" />
831 </input>
832 <output>
833 <port id="2" precision="I64" names="64,65">
834 <dim>-1</dim>
835 <dim>1</dim>
836 <dim>1</dim>
837 <dim>-1</dim>
838 </port>
839 </output>
840 </layer>
841 <layer id="61" name="Constant_305970" type="Const" version="opset1">
842 <data element_type="i64" shape="1" offset="47675396" size="8" />
843 <output>
844 <port id="0" precision="I64">
845 <dim>1</dim>
846 </port>
847 </output>
848 </layer>
849 <layer id="62" name="Constant_305971" type="Const" version="opset1">
850 <data element_type="i64" shape="" offset="47675396" size="8" />
851 <output>
852 <port id="0" precision="I64" />
853 </output>
854 </layer>
855 <layer id="63" name="Gather_305972" type="Gather" version="opset8">
856 <data batch_dims="0" />
857 <input>
858 <port id="0" precision="I64">
859 <dim>2</dim>
860 </port>
861 <port id="1" precision="I64">
862 <dim>1</dim>
863 </port>
864 <port id="2" precision="I64" />
865 </input>
866 <output>
867 <port id="3" precision="I64" names="60">
868 <dim>1</dim>
869 </port>
870 </output>
871 </layer>
872 <layer id="64" name="Constant_305305" type="Const" version="opset1">
873 <data element_type="i64" shape="1" offset="47675404" size="8" />
874 <output>
875 <port id="0" precision="I64">
876 <dim>1</dim>
877 </port>
878 </output>
879 </layer>
880 <layer id="65" name="Constant_306080" type="Const" version="opset1">
881 <data element_type="i64" shape="2" offset="49452644" size="16" />
882 <output>
883 <port id="0" precision="I64">
884 <dim>2</dim>
885 </port>
886 </output>
887 </layer>
888 <layer id="66" name="Constant_306081" type="Const" version="opset1">
889 <data element_type="i64" shape="" offset="47675396" size="8" />
890 <output>
891 <port id="0" precision="I64" />
892 </output>
893 </layer>
894 <layer id="67" name="Gather_306082" type="Gather" version="opset8">
895 <data batch_dims="0" />
896 <input>
897 <port id="0" precision="I64">
898 <dim>2</dim>
899 </port>
900 <port id="1" precision="I64">
901 <dim>2</dim>
902 </port>
903 <port id="2" precision="I64" />
904 </input>
905 <output>
906 <port id="3" precision="I64">
907 <dim>2</dim>
908 </port>
909 </output>
910 </layer>
911 <layer id="68" name="__module.bert/prim::ListConstruct/Concat" type="Concat" version="opset1">
912 <data axis="0" />
913 <input>
914 <port id="0" precision="I64">
915 <dim>1</dim>
916 </port>
917 <port id="1" precision="I64">
918 <dim>1</dim>
919 </port>
920 <port id="2" precision="I64">
921 <dim>2</dim>
922 </port>
923 </input>
924 <output>
925 <port id="3" precision="I64" names="66">
926 <dim>4</dim>
927 </port>
928 </output>
929 </layer>
930 <layer id="69" name="__module.bert/aten::expand/Broadcast" type="Broadcast" version="opset3">
931 <data mode="bidirectional" />
932 <input>
933 <port id="0" precision="I64">
934 <dim>-1</dim>
935 <dim>1</dim>
936 <dim>1</dim>
937 <dim>-1</dim>
938 </port>
939 <port id="1" precision="I64">
940 <dim>4</dim>
941 </port>
942 </input>
943 <output>
944 <port id="2" precision="I64" names="67">
945 <dim>-1</dim>
946 <dim>1</dim>
947 <dim>-1</dim>
948 <dim>-1</dim>
949 </port>
950 </output>
951 </layer>
952 <layer id="70" name="__module.bert/aten::to/Convert" type="Convert" version="opset1">
953 <data destination_type="f32" />
954 <input>
955 <port id="0" precision="I64">
956 <dim>-1</dim>
957 <dim>1</dim>
958 <dim>-1</dim>
959 <dim>-1</dim>
960 </port>
961 </input>
962 <output>
963 <port id="1" precision="FP32" names="68">
964 <dim>-1</dim>
965 <dim>1</dim>
966 <dim>-1</dim>
967 <dim>-1</dim>
968 </port>
969 </output>
970 </layer>
971 <layer id="71" name="Constant_305874" type="Const" version="opset1">
972 <data element_type="f32" shape="1, 1, 1, 1" offset="49452632" size="4" />
973 <output>
974 <port id="0" precision="FP32">
975 <dim>1</dim>
976 <dim>1</dim>
977 <dim>1</dim>
978 <dim>1</dim>
979 </port>
980 </output>
981 </layer>
982 <layer id="72" name="__module.bert/aten::rsub/Multiply" type="Multiply" version="opset1">
983 <data auto_broadcast="numpy" />
984 <input>
985 <port id="0" precision="FP32">
986 <dim>-1</dim>
987 <dim>1</dim>
988 <dim>-1</dim>
989 <dim>-1</dim>
990 </port>
991 <port id="1" precision="FP32">
992 <dim>1</dim>
993 <dim>1</dim>
994 <dim>1</dim>
995 <dim>1</dim>
996 </port>
997 </input>
998 <output>
999 <port id="2" precision="FP32">
1000 <dim>-1</dim>
1001 <dim>1</dim>
1002 <dim>-1</dim>
1003 <dim>-1</dim>
1004 </port>
1005 </output>
1006 </layer>
1007 <layer id="73" name="__module.bert/aten::rsub/Subtract" type="Subtract" version="opset1">
1008 <data auto_broadcast="numpy" />
1009 <input>
1010 <port id="0" precision="FP32">
1011 <dim>1</dim>
1012 <dim>1</dim>
1013 <dim>1</dim>
1014 <dim>1</dim>
1015 </port>
1016 <port id="1" precision="FP32">
1017 <dim>-1</dim>
1018 <dim>1</dim>
1019 <dim>-1</dim>
1020 <dim>-1</dim>
1021 </port>
1022 </input>
1023 <output>
1024 <port id="2" precision="FP32" names="69,inverted_mask">
1025 <dim>-1</dim>
1026 <dim>1</dim>
1027 <dim>-1</dim>
1028 <dim>-1</dim>
1029 </port>
1030 </output>
1031 </layer>
1032 <layer id="74" name="__module.bert/aten::to/Convert_1" type="Convert" version="opset1">
1033 <data destination_type="boolean" />
1034 <input>
1035 <port id="0" precision="FP32">
1036 <dim>-1</dim>
1037 <dim>1</dim>
1038 <dim>-1</dim>
1039 <dim>-1</dim>
1040 </port>
1041 </input>
1042 <output>
1043 <port id="1" precision="BOOL" names="70">
1044 <dim>-1</dim>
1045 <dim>1</dim>
1046 <dim>-1</dim>
1047 <dim>-1</dim>
1048 </port>
1049 </output>
1050 </layer>
1051 <layer id="75" name="__module.bert/aten::masked_fill/ConvertLike" type="Const" version="opset1">
1052 <data element_type="f32" shape="" offset="49452660" size="4" />
1053 <output>
1054 <port id="0" precision="FP32" />
1055 </output>
1056 </layer>
1057 <layer id="76" name="__module.bert/aten::masked_fill/Select" type="Select" version="opset1">
1058 <data auto_broadcast="numpy" />
1059 <input>
1060 <port id="0" precision="BOOL">
1061 <dim>-1</dim>
1062 <dim>1</dim>
1063 <dim>-1</dim>
1064 <dim>-1</dim>
1065 </port>
1066 <port id="1" precision="FP32" />
1067 <port id="2" precision="FP32">
1068 <dim>-1</dim>
1069 <dim>1</dim>
1070 <dim>-1</dim>
1071 <dim>-1</dim>
1072 </port>
1073 </input>
1074 <output>
1075 <port id="3" precision="FP32" names="71">
1076 <dim>-1</dim>
1077 <dim>1</dim>
1078 <dim>-1</dim>
1079 <dim>-1</dim>
1080 </port>
1081 </output>
1082 </layer>
1083 <layer id="77" name="__module.bert.encoder.layer.0.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1084 <data causal="false" />
1085 <input>
1086 <port id="0" precision="FP32">
1087 <dim>-1</dim>
1088 <dim>12</dim>
1089 <dim>-1</dim>
1090 <dim>32</dim>
1091 </port>
1092 <port id="1" precision="FP32">
1093 <dim>-1</dim>
1094 <dim>12</dim>
1095 <dim>-1</dim>
1096 <dim>32</dim>
1097 </port>
1098 <port id="2" precision="FP32">
1099 <dim>-1</dim>
1100 <dim>12</dim>
1101 <dim>-1</dim>
1102 <dim>32</dim>
1103 </port>
1104 <port id="3" precision="FP32">
1105 <dim>-1</dim>
1106 <dim>1</dim>
1107 <dim>-1</dim>
1108 <dim>-1</dim>
1109 </port>
1110 </input>
1111 <output>
1112 <port id="4" precision="FP32" names="117,attn_output.1">
1113 <dim>-1</dim>
1114 <dim>12</dim>
1115 <dim>-1</dim>
1116 <dim>32</dim>
1117 </port>
1118 </output>
1119 </layer>
1120 <layer id="78" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
1121 <data element_type="i32" shape="4" offset="49452664" size="16" />
1122 <output>
1123 <port id="0" precision="I32">
1124 <dim>4</dim>
1125 </port>
1126 </output>
1127 </layer>
1128 <layer id="79" name="__module.bert.encoder.layer.0.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
1129 <input>
1130 <port id="0" precision="FP32">
1131 <dim>-1</dim>
1132 <dim>12</dim>
1133 <dim>-1</dim>
1134 <dim>32</dim>
1135 </port>
1136 <port id="1" precision="I32">
1137 <dim>4</dim>
1138 </port>
1139 </input>
1140 <output>
1141 <port id="2" precision="FP32" names="118,attn_output.3">
1142 <dim>-1</dim>
1143 <dim>-1</dim>
1144 <dim>12</dim>
1145 <dim>32</dim>
1146 </port>
1147 </output>
1148 </layer>
1149 <layer id="80" name="Constant_305977" type="Const" version="opset1">
1150 <data element_type="i64" shape="3" offset="49452680" size="24" />
1151 <output>
1152 <port id="0" precision="I64">
1153 <dim>3</dim>
1154 </port>
1155 </output>
1156 </layer>
1157 <layer id="81" name="__module.bert.encoder.layer.0.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
1158 <data special_zero="true" />
1159 <input>
1160 <port id="0" precision="FP32">
1161 <dim>-1</dim>
1162 <dim>-1</dim>
1163 <dim>12</dim>
1164 <dim>32</dim>
1165 </port>
1166 <port id="1" precision="I64">
1167 <dim>3</dim>
1168 </port>
1169 </input>
1170 <output>
1171 <port id="2" precision="FP32" names="120">
1172 <dim>-1</dim>
1173 <dim>-1</dim>
1174 <dim>384</dim>
1175 </port>
1176 </output>
1177 </layer>
1178 <layer id="82" name="self.bert.encoder.layer.0.attention.output.dense.weight" type="Const" version="opset1">
1179 <data element_type="f32" shape="384, 384" offset="49452704" size="589824" />
1180 <output>
1181 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.attention.output.dense.weight">
1182 <dim>384</dim>
1183 <dim>384</dim>
1184 </port>
1185 </output>
1186 </layer>
1187 <layer id="83" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1188 <data transpose_a="false" transpose_b="true" />
1189 <input>
1190 <port id="0" precision="FP32">
1191 <dim>-1</dim>
1192 <dim>-1</dim>
1193 <dim>384</dim>
1194 </port>
1195 <port id="1" precision="FP32">
1196 <dim>384</dim>
1197 <dim>384</dim>
1198 </port>
1199 </input>
1200 <output>
1201 <port id="2" precision="FP32">
1202 <dim>-1</dim>
1203 <dim>-1</dim>
1204 <dim>384</dim>
1205 </port>
1206 </output>
1207 </layer>
1208 <layer id="84" name="Constant_305876" type="Const" version="opset1">
1209 <data element_type="f32" shape="1, 1, 384" offset="50042528" size="1536" />
1210 <output>
1211 <port id="0" precision="FP32">
1212 <dim>1</dim>
1213 <dim>1</dim>
1214 <dim>384</dim>
1215 </port>
1216 </output>
1217 </layer>
1218 <layer id="85" name="__module.bert.encoder.layer.0.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
1219 <data auto_broadcast="numpy" />
1220 <input>
1221 <port id="0" precision="FP32">
1222 <dim>-1</dim>
1223 <dim>-1</dim>
1224 <dim>384</dim>
1225 </port>
1226 <port id="1" precision="FP32">
1227 <dim>1</dim>
1228 <dim>1</dim>
1229 <dim>384</dim>
1230 </port>
1231 </input>
1232 <output>
1233 <port id="2" precision="FP32" names="125,input.3">
1234 <dim>-1</dim>
1235 <dim>-1</dim>
1236 <dim>384</dim>
1237 </port>
1238 </output>
1239 </layer>
1240 <layer id="86" name="__module.bert.encoder.layer.0.attention.output/aten::add/Add" type="Add" version="opset1">
1241 <data auto_broadcast="numpy" />
1242 <input>
1243 <port id="0" precision="FP32">
1244 <dim>-1</dim>
1245 <dim>-1</dim>
1246 <dim>384</dim>
1247 </port>
1248 <port id="1" precision="FP32">
1249 <dim>-1</dim>
1250 <dim>-1</dim>
1251 <dim>384</dim>
1252 </port>
1253 </input>
1254 <output>
1255 <port id="2" precision="FP32" names="127">
1256 <dim>-1</dim>
1257 <dim>-1</dim>
1258 <dim>384</dim>
1259 </port>
1260 </output>
1261 </layer>
1262 <layer id="87" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1263 <data element_type="i32" shape="1" offset="47675412" size="4" />
1264 <output>
1265 <port id="0" precision="I32">
1266 <dim>1</dim>
1267 </port>
1268 </output>
1269 </layer>
1270 <layer id="88" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1271 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1272 <input>
1273 <port id="0" precision="FP32">
1274 <dim>-1</dim>
1275 <dim>-1</dim>
1276 <dim>384</dim>
1277 </port>
1278 <port id="1" precision="I32">
1279 <dim>1</dim>
1280 </port>
1281 </input>
1282 <output>
1283 <port id="2" precision="FP32">
1284 <dim>-1</dim>
1285 <dim>-1</dim>
1286 <dim>384</dim>
1287 </port>
1288 </output>
1289 </layer>
1290 <layer id="89" name="Constant_305877" type="Const" version="opset1">
1291 <data element_type="f32" shape="1, 1, 384" offset="50044064" size="1536" />
1292 <output>
1293 <port id="0" precision="FP32">
1294 <dim>1</dim>
1295 <dim>1</dim>
1296 <dim>384</dim>
1297 </port>
1298 </output>
1299 </layer>
1300 <layer id="90" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1301 <data auto_broadcast="numpy" />
1302 <input>
1303 <port id="0" precision="FP32">
1304 <dim>-1</dim>
1305 <dim>-1</dim>
1306 <dim>384</dim>
1307 </port>
1308 <port id="1" precision="FP32">
1309 <dim>1</dim>
1310 <dim>1</dim>
1311 <dim>384</dim>
1312 </port>
1313 </input>
1314 <output>
1315 <port id="2" precision="FP32">
1316 <dim>-1</dim>
1317 <dim>-1</dim>
1318 <dim>384</dim>
1319 </port>
1320 </output>
1321 </layer>
1322 <layer id="91" name="Constant_305878" type="Const" version="opset1">
1323 <data element_type="f32" shape="1, 1, 384" offset="50045600" size="1536" />
1324 <output>
1325 <port id="0" precision="FP32">
1326 <dim>1</dim>
1327 <dim>1</dim>
1328 <dim>384</dim>
1329 </port>
1330 </output>
1331 </layer>
1332 <layer id="92" name="__module.bert.encoder.layer.0.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1333 <data auto_broadcast="numpy" />
1334 <input>
1335 <port id="0" precision="FP32">
1336 <dim>-1</dim>
1337 <dim>-1</dim>
1338 <dim>384</dim>
1339 </port>
1340 <port id="1" precision="FP32">
1341 <dim>1</dim>
1342 <dim>1</dim>
1343 <dim>384</dim>
1344 </port>
1345 </input>
1346 <output>
1347 <port id="2" precision="FP32" names="131,input_tensor.1">
1348 <dim>-1</dim>
1349 <dim>-1</dim>
1350 <dim>384</dim>
1351 </port>
1352 </output>
1353 </layer>
1354 <layer id="93" name="self.bert.encoder.layer.0.intermediate.dense.weight" type="Const" version="opset1">
1355 <data element_type="f32" shape="1536, 384" offset="50047136" size="2359296" />
1356 <output>
1357 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.intermediate.dense.weight">
1358 <dim>1536</dim>
1359 <dim>384</dim>
1360 </port>
1361 </output>
1362 </layer>
1363 <layer id="94" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1364 <data transpose_a="false" transpose_b="true" />
1365 <input>
1366 <port id="0" precision="FP32">
1367 <dim>-1</dim>
1368 <dim>-1</dim>
1369 <dim>384</dim>
1370 </port>
1371 <port id="1" precision="FP32">
1372 <dim>1536</dim>
1373 <dim>384</dim>
1374 </port>
1375 </input>
1376 <output>
1377 <port id="2" precision="FP32">
1378 <dim>-1</dim>
1379 <dim>-1</dim>
1380 <dim>1536</dim>
1381 </port>
1382 </output>
1383 </layer>
1384 <layer id="95" name="Constant_305879" type="Const" version="opset1">
1385 <data element_type="f32" shape="1, 1, 1536" offset="52406432" size="6144" />
1386 <output>
1387 <port id="0" precision="FP32">
1388 <dim>1</dim>
1389 <dim>1</dim>
1390 <dim>1536</dim>
1391 </port>
1392 </output>
1393 </layer>
1394 <layer id="96" name="__module.bert.encoder.layer.0.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
1395 <data auto_broadcast="numpy" />
1396 <input>
1397 <port id="0" precision="FP32">
1398 <dim>-1</dim>
1399 <dim>-1</dim>
1400 <dim>1536</dim>
1401 </port>
1402 <port id="1" precision="FP32">
1403 <dim>1</dim>
1404 <dim>1</dim>
1405 <dim>1536</dim>
1406 </port>
1407 </input>
1408 <output>
1409 <port id="2" precision="FP32" names="135">
1410 <dim>-1</dim>
1411 <dim>-1</dim>
1412 <dim>1536</dim>
1413 </port>
1414 </output>
1415 </layer>
1416 <layer id="97" name="__module.bert.encoder.layer.0.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
1417 <data approximation_mode="ERF" />
1418 <input>
1419 <port id="0" precision="FP32">
1420 <dim>-1</dim>
1421 <dim>-1</dim>
1422 <dim>1536</dim>
1423 </port>
1424 </input>
1425 <output>
1426 <port id="1" precision="FP32" names="136">
1427 <dim>-1</dim>
1428 <dim>-1</dim>
1429 <dim>1536</dim>
1430 </port>
1431 </output>
1432 </layer>
1433 <layer id="98" name="self.bert.encoder.layer.0.output.dense.weight" type="Const" version="opset1">
1434 <data element_type="f32" shape="384, 1536" offset="52412576" size="2359296" />
1435 <output>
1436 <port id="0" precision="FP32" names="self.bert.encoder.layer.0.output.dense.weight">
1437 <dim>384</dim>
1438 <dim>1536</dim>
1439 </port>
1440 </output>
1441 </layer>
1442 <layer id="99" name="__module.bert.encoder.layer.0.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
1443 <data transpose_a="false" transpose_b="true" />
1444 <input>
1445 <port id="0" precision="FP32">
1446 <dim>-1</dim>
1447 <dim>-1</dim>
1448 <dim>1536</dim>
1449 </port>
1450 <port id="1" precision="FP32">
1451 <dim>384</dim>
1452 <dim>1536</dim>
1453 </port>
1454 </input>
1455 <output>
1456 <port id="2" precision="FP32">
1457 <dim>-1</dim>
1458 <dim>-1</dim>
1459 <dim>384</dim>
1460 </port>
1461 </output>
1462 </layer>
1463 <layer id="100" name="Constant_305880" type="Const" version="opset1">
1464 <data element_type="f32" shape="1, 1, 384" offset="54771872" size="1536" />
1465 <output>
1466 <port id="0" precision="FP32">
1467 <dim>1</dim>
1468 <dim>1</dim>
1469 <dim>384</dim>
1470 </port>
1471 </output>
1472 </layer>
1473 <layer id="101" name="__module.bert.encoder.layer.0.output.dense/aten::linear/Add" type="Add" version="opset1">
1474 <data auto_broadcast="numpy" />
1475 <input>
1476 <port id="0" precision="FP32">
1477 <dim>-1</dim>
1478 <dim>-1</dim>
1479 <dim>384</dim>
1480 </port>
1481 <port id="1" precision="FP32">
1482 <dim>1</dim>
1483 <dim>1</dim>
1484 <dim>384</dim>
1485 </port>
1486 </input>
1487 <output>
1488 <port id="2" precision="FP32" names="141,input.5">
1489 <dim>-1</dim>
1490 <dim>-1</dim>
1491 <dim>384</dim>
1492 </port>
1493 </output>
1494 </layer>
1495 <layer id="102" name="__module.bert.encoder.layer.0.output/aten::add/Add" type="Add" version="opset1">
1496 <data auto_broadcast="numpy" />
1497 <input>
1498 <port id="0" precision="FP32">
1499 <dim>-1</dim>
1500 <dim>-1</dim>
1501 <dim>384</dim>
1502 </port>
1503 <port id="1" precision="FP32">
1504 <dim>-1</dim>
1505 <dim>-1</dim>
1506 <dim>384</dim>
1507 </port>
1508 </input>
1509 <output>
1510 <port id="2" precision="FP32" names="143">
1511 <dim>-1</dim>
1512 <dim>-1</dim>
1513 <dim>384</dim>
1514 </port>
1515 </output>
1516 </layer>
1517 <layer id="103" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
1518 <data element_type="i32" shape="1" offset="47675412" size="4" />
1519 <output>
1520 <port id="0" precision="I32">
1521 <dim>1</dim>
1522 </port>
1523 </output>
1524 </layer>
1525 <layer id="104" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
1526 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
1527 <input>
1528 <port id="0" precision="FP32">
1529 <dim>-1</dim>
1530 <dim>-1</dim>
1531 <dim>384</dim>
1532 </port>
1533 <port id="1" precision="I32">
1534 <dim>1</dim>
1535 </port>
1536 </input>
1537 <output>
1538 <port id="2" precision="FP32">
1539 <dim>-1</dim>
1540 <dim>-1</dim>
1541 <dim>384</dim>
1542 </port>
1543 </output>
1544 </layer>
1545 <layer id="105" name="Constant_305881" type="Const" version="opset1">
1546 <data element_type="f32" shape="1, 1, 384" offset="54773408" size="1536" />
1547 <output>
1548 <port id="0" precision="FP32">
1549 <dim>1</dim>
1550 <dim>1</dim>
1551 <dim>384</dim>
1552 </port>
1553 </output>
1554 </layer>
1555 <layer id="106" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
1556 <data auto_broadcast="numpy" />
1557 <input>
1558 <port id="0" precision="FP32">
1559 <dim>-1</dim>
1560 <dim>-1</dim>
1561 <dim>384</dim>
1562 </port>
1563 <port id="1" precision="FP32">
1564 <dim>1</dim>
1565 <dim>1</dim>
1566 <dim>384</dim>
1567 </port>
1568 </input>
1569 <output>
1570 <port id="2" precision="FP32">
1571 <dim>-1</dim>
1572 <dim>-1</dim>
1573 <dim>384</dim>
1574 </port>
1575 </output>
1576 </layer>
1577 <layer id="107" name="Constant_305882" type="Const" version="opset1">
1578 <data element_type="f32" shape="1, 1, 384" offset="54774944" size="1536" />
1579 <output>
1580 <port id="0" precision="FP32">
1581 <dim>1</dim>
1582 <dim>1</dim>
1583 <dim>384</dim>
1584 </port>
1585 </output>
1586 </layer>
1587 <layer id="108" name="__module.bert.encoder.layer.0.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
1588 <data auto_broadcast="numpy" />
1589 <input>
1590 <port id="0" precision="FP32">
1591 <dim>-1</dim>
1592 <dim>-1</dim>
1593 <dim>384</dim>
1594 </port>
1595 <port id="1" precision="FP32">
1596 <dim>1</dim>
1597 <dim>1</dim>
1598 <dim>384</dim>
1599 </port>
1600 </input>
1601 <output>
1602 <port id="2" precision="FP32" names="147,hidden_states.7">
1603 <dim>-1</dim>
1604 <dim>-1</dim>
1605 <dim>384</dim>
1606 </port>
1607 </output>
1608 </layer>
1609 <layer id="109" name="self.bert.encoder.layer.1.attention.self.query.weight" type="Const" version="opset1">
1610 <data element_type="f32" shape="384, 384" offset="54776480" size="589824" />
1611 <output>
1612 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.self.query.weight">
1613 <dim>384</dim>
1614 <dim>384</dim>
1615 </port>
1616 </output>
1617 </layer>
1618 <layer id="110" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
1619 <data transpose_a="false" transpose_b="true" />
1620 <input>
1621 <port id="0" precision="FP32">
1622 <dim>-1</dim>
1623 <dim>-1</dim>
1624 <dim>384</dim>
1625 </port>
1626 <port id="1" precision="FP32">
1627 <dim>384</dim>
1628 <dim>384</dim>
1629 </port>
1630 </input>
1631 <output>
1632 <port id="2" precision="FP32">
1633 <dim>-1</dim>
1634 <dim>-1</dim>
1635 <dim>384</dim>
1636 </port>
1637 </output>
1638 </layer>
1639 <layer id="111" name="Constant_305883" type="Const" version="opset1">
1640 <data element_type="f32" shape="1, 1, 384" offset="55366304" size="1536" />
1641 <output>
1642 <port id="0" precision="FP32">
1643 <dim>1</dim>
1644 <dim>1</dim>
1645 <dim>384</dim>
1646 </port>
1647 </output>
1648 </layer>
1649 <layer id="112" name="__module.bert.encoder.layer.1.attention.self.query/aten::linear/Add" type="Add" version="opset1">
1650 <data auto_broadcast="numpy" />
1651 <input>
1652 <port id="0" precision="FP32">
1653 <dim>-1</dim>
1654 <dim>-1</dim>
1655 <dim>384</dim>
1656 </port>
1657 <port id="1" precision="FP32">
1658 <dim>1</dim>
1659 <dim>1</dim>
1660 <dim>384</dim>
1661 </port>
1662 </input>
1663 <output>
1664 <port id="2" precision="FP32" names="160,x.13">
1665 <dim>-1</dim>
1666 <dim>-1</dim>
1667 <dim>384</dim>
1668 </port>
1669 </output>
1670 </layer>
1671 <layer id="113" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
1672 <data element_type="i64" shape="4" offset="48269848" size="32" />
1673 <output>
1674 <port id="0" precision="I64">
1675 <dim>4</dim>
1676 </port>
1677 </output>
1678 </layer>
1679 <layer id="114" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
1680 <data special_zero="true" />
1681 <input>
1682 <port id="0" precision="FP32">
1683 <dim>-1</dim>
1684 <dim>-1</dim>
1685 <dim>384</dim>
1686 </port>
1687 <port id="1" precision="I64">
1688 <dim>4</dim>
1689 </port>
1690 </input>
1691 <output>
1692 <port id="2" precision="FP32" names="164,x.15">
1693 <dim>-1</dim>
1694 <dim>-1</dim>
1695 <dim>12</dim>
1696 <dim>32</dim>
1697 </port>
1698 </output>
1699 </layer>
1700 <layer id="115" name="Constant_301289" type="Const" version="opset1">
1701 <data element_type="i64" shape="4" offset="48269880" size="32" />
1702 <output>
1703 <port id="0" precision="I64" names="165">
1704 <dim>4</dim>
1705 </port>
1706 </output>
1707 </layer>
1708 <layer id="116" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
1709 <input>
1710 <port id="0" precision="FP32">
1711 <dim>-1</dim>
1712 <dim>-1</dim>
1713 <dim>12</dim>
1714 <dim>32</dim>
1715 </port>
1716 <port id="1" precision="I64">
1717 <dim>4</dim>
1718 </port>
1719 </input>
1720 <output>
1721 <port id="2" precision="FP32" names="166">
1722 <dim>-1</dim>
1723 <dim>12</dim>
1724 <dim>-1</dim>
1725 <dim>32</dim>
1726 </port>
1727 </output>
1728 </layer>
1729 <layer id="117" name="self.bert.encoder.layer.1.attention.self.key.weight" type="Const" version="opset1">
1730 <data element_type="f32" shape="384, 384" offset="55367840" size="589824" />
1731 <output>
1732 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.self.key.weight">
1733 <dim>384</dim>
1734 <dim>384</dim>
1735 </port>
1736 </output>
1737 </layer>
1738 <layer id="118" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
1739 <data transpose_a="false" transpose_b="true" />
1740 <input>
1741 <port id="0" precision="FP32">
1742 <dim>-1</dim>
1743 <dim>-1</dim>
1744 <dim>384</dim>
1745 </port>
1746 <port id="1" precision="FP32">
1747 <dim>384</dim>
1748 <dim>384</dim>
1749 </port>
1750 </input>
1751 <output>
1752 <port id="2" precision="FP32">
1753 <dim>-1</dim>
1754 <dim>-1</dim>
1755 <dim>384</dim>
1756 </port>
1757 </output>
1758 </layer>
1759 <layer id="119" name="Constant_305884" type="Const" version="opset1">
1760 <data element_type="f32" shape="1, 1, 384" offset="55957664" size="1536" />
1761 <output>
1762 <port id="0" precision="FP32">
1763 <dim>1</dim>
1764 <dim>1</dim>
1765 <dim>384</dim>
1766 </port>
1767 </output>
1768 </layer>
1769 <layer id="120" name="__module.bert.encoder.layer.1.attention.self.key/aten::linear/Add" type="Add" version="opset1">
1770 <data auto_broadcast="numpy" />
1771 <input>
1772 <port id="0" precision="FP32">
1773 <dim>-1</dim>
1774 <dim>-1</dim>
1775 <dim>384</dim>
1776 </port>
1777 <port id="1" precision="FP32">
1778 <dim>1</dim>
1779 <dim>1</dim>
1780 <dim>384</dim>
1781 </port>
1782 </input>
1783 <output>
1784 <port id="2" precision="FP32" names="169,x.17">
1785 <dim>-1</dim>
1786 <dim>-1</dim>
1787 <dim>384</dim>
1788 </port>
1789 </output>
1790 </layer>
1791 <layer id="121" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
1792 <data element_type="i64" shape="4" offset="48269848" size="32" />
1793 <output>
1794 <port id="0" precision="I64">
1795 <dim>4</dim>
1796 </port>
1797 </output>
1798 </layer>
1799 <layer id="122" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
1800 <data special_zero="true" />
1801 <input>
1802 <port id="0" precision="FP32">
1803 <dim>-1</dim>
1804 <dim>-1</dim>
1805 <dim>384</dim>
1806 </port>
1807 <port id="1" precision="I64">
1808 <dim>4</dim>
1809 </port>
1810 </input>
1811 <output>
1812 <port id="2" precision="FP32" names="173,x.19">
1813 <dim>-1</dim>
1814 <dim>-1</dim>
1815 <dim>12</dim>
1816 <dim>32</dim>
1817 </port>
1818 </output>
1819 </layer>
1820 <layer id="123" name="Constant_301312" type="Const" version="opset1">
1821 <data element_type="i64" shape="4" offset="48269880" size="32" />
1822 <output>
1823 <port id="0" precision="I64" names="174">
1824 <dim>4</dim>
1825 </port>
1826 </output>
1827 </layer>
1828 <layer id="124" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
1829 <input>
1830 <port id="0" precision="FP32">
1831 <dim>-1</dim>
1832 <dim>-1</dim>
1833 <dim>12</dim>
1834 <dim>32</dim>
1835 </port>
1836 <port id="1" precision="I64">
1837 <dim>4</dim>
1838 </port>
1839 </input>
1840 <output>
1841 <port id="2" precision="FP32" names="175">
1842 <dim>-1</dim>
1843 <dim>12</dim>
1844 <dim>-1</dim>
1845 <dim>32</dim>
1846 </port>
1847 </output>
1848 </layer>
1849 <layer id="125" name="self.bert.encoder.layer.1.attention.self.value.weight" type="Const" version="opset1">
1850 <data element_type="f32" shape="384, 384" offset="55959200" size="589824" />
1851 <output>
1852 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.self.value.weight">
1853 <dim>384</dim>
1854 <dim>384</dim>
1855 </port>
1856 </output>
1857 </layer>
1858 <layer id="126" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
1859 <data transpose_a="false" transpose_b="true" />
1860 <input>
1861 <port id="0" precision="FP32">
1862 <dim>-1</dim>
1863 <dim>-1</dim>
1864 <dim>384</dim>
1865 </port>
1866 <port id="1" precision="FP32">
1867 <dim>384</dim>
1868 <dim>384</dim>
1869 </port>
1870 </input>
1871 <output>
1872 <port id="2" precision="FP32">
1873 <dim>-1</dim>
1874 <dim>-1</dim>
1875 <dim>384</dim>
1876 </port>
1877 </output>
1878 </layer>
1879 <layer id="127" name="Constant_305885" type="Const" version="opset1">
1880 <data element_type="f32" shape="1, 1, 384" offset="56549024" size="1536" />
1881 <output>
1882 <port id="0" precision="FP32">
1883 <dim>1</dim>
1884 <dim>1</dim>
1885 <dim>384</dim>
1886 </port>
1887 </output>
1888 </layer>
1889 <layer id="128" name="__module.bert.encoder.layer.1.attention.self.value/aten::linear/Add" type="Add" version="opset1">
1890 <data auto_broadcast="numpy" />
1891 <input>
1892 <port id="0" precision="FP32">
1893 <dim>-1</dim>
1894 <dim>-1</dim>
1895 <dim>384</dim>
1896 </port>
1897 <port id="1" precision="FP32">
1898 <dim>1</dim>
1899 <dim>1</dim>
1900 <dim>384</dim>
1901 </port>
1902 </input>
1903 <output>
1904 <port id="2" precision="FP32" names="178,x.21">
1905 <dim>-1</dim>
1906 <dim>-1</dim>
1907 <dim>384</dim>
1908 </port>
1909 </output>
1910 </layer>
1911 <layer id="129" name="__module.bert.encoder.layer.1.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
1912 <data element_type="i64" shape="4" offset="48269848" size="32" />
1913 <output>
1914 <port id="0" precision="I64">
1915 <dim>4</dim>
1916 </port>
1917 </output>
1918 </layer>
1919 <layer id="130" name="__module.bert.encoder.layer.1.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
1920 <data special_zero="true" />
1921 <input>
1922 <port id="0" precision="FP32">
1923 <dim>-1</dim>
1924 <dim>-1</dim>
1925 <dim>384</dim>
1926 </port>
1927 <port id="1" precision="I64">
1928 <dim>4</dim>
1929 </port>
1930 </input>
1931 <output>
1932 <port id="2" precision="FP32" names="182,x.23">
1933 <dim>-1</dim>
1934 <dim>-1</dim>
1935 <dim>12</dim>
1936 <dim>32</dim>
1937 </port>
1938 </output>
1939 </layer>
1940 <layer id="131" name="Constant_301335" type="Const" version="opset1">
1941 <data element_type="i64" shape="4" offset="48269880" size="32" />
1942 <output>
1943 <port id="0" precision="I64" names="183">
1944 <dim>4</dim>
1945 </port>
1946 </output>
1947 </layer>
1948 <layer id="132" name="__module.bert.encoder.layer.1.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
1949 <input>
1950 <port id="0" precision="FP32">
1951 <dim>-1</dim>
1952 <dim>-1</dim>
1953 <dim>12</dim>
1954 <dim>32</dim>
1955 </port>
1956 <port id="1" precision="I64">
1957 <dim>4</dim>
1958 </port>
1959 </input>
1960 <output>
1961 <port id="2" precision="FP32" names="184">
1962 <dim>-1</dim>
1963 <dim>12</dim>
1964 <dim>-1</dim>
1965 <dim>32</dim>
1966 </port>
1967 </output>
1968 </layer>
1969 <layer id="133" name="__module.bert.encoder.layer.1.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
1970 <data causal="false" />
1971 <input>
1972 <port id="0" precision="FP32">
1973 <dim>-1</dim>
1974 <dim>12</dim>
1975 <dim>-1</dim>
1976 <dim>32</dim>
1977 </port>
1978 <port id="1" precision="FP32">
1979 <dim>-1</dim>
1980 <dim>12</dim>
1981 <dim>-1</dim>
1982 <dim>32</dim>
1983 </port>
1984 <port id="2" precision="FP32">
1985 <dim>-1</dim>
1986 <dim>12</dim>
1987 <dim>-1</dim>
1988 <dim>32</dim>
1989 </port>
1990 <port id="3" precision="FP32">
1991 <dim>-1</dim>
1992 <dim>1</dim>
1993 <dim>-1</dim>
1994 <dim>-1</dim>
1995 </port>
1996 </input>
1997 <output>
1998 <port id="4" precision="FP32" names="185,attn_output.5">
1999 <dim>-1</dim>
2000 <dim>12</dim>
2001 <dim>-1</dim>
2002 <dim>32</dim>
2003 </port>
2004 </output>
2005 </layer>
2006 <layer id="134" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
2007 <data element_type="i32" shape="4" offset="49452664" size="16" />
2008 <output>
2009 <port id="0" precision="I32">
2010 <dim>4</dim>
2011 </port>
2012 </output>
2013 </layer>
2014 <layer id="135" name="__module.bert.encoder.layer.1.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
2015 <input>
2016 <port id="0" precision="FP32">
2017 <dim>-1</dim>
2018 <dim>12</dim>
2019 <dim>-1</dim>
2020 <dim>32</dim>
2021 </port>
2022 <port id="1" precision="I32">
2023 <dim>4</dim>
2024 </port>
2025 </input>
2026 <output>
2027 <port id="2" precision="FP32" names="186,attn_output.7">
2028 <dim>-1</dim>
2029 <dim>-1</dim>
2030 <dim>12</dim>
2031 <dim>32</dim>
2032 </port>
2033 </output>
2034 </layer>
2035 <layer id="136" name="Constant_305978" type="Const" version="opset1">
2036 <data element_type="i64" shape="3" offset="49452680" size="24" />
2037 <output>
2038 <port id="0" precision="I64">
2039 <dim>3</dim>
2040 </port>
2041 </output>
2042 </layer>
2043 <layer id="137" name="__module.bert.encoder.layer.1.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
2044 <data special_zero="true" />
2045 <input>
2046 <port id="0" precision="FP32">
2047 <dim>-1</dim>
2048 <dim>-1</dim>
2049 <dim>12</dim>
2050 <dim>32</dim>
2051 </port>
2052 <port id="1" precision="I64">
2053 <dim>3</dim>
2054 </port>
2055 </input>
2056 <output>
2057 <port id="2" precision="FP32" names="188">
2058 <dim>-1</dim>
2059 <dim>-1</dim>
2060 <dim>384</dim>
2061 </port>
2062 </output>
2063 </layer>
2064 <layer id="138" name="self.bert.encoder.layer.1.attention.output.dense.weight" type="Const" version="opset1">
2065 <data element_type="f32" shape="384, 384" offset="56550560" size="589824" />
2066 <output>
2067 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.attention.output.dense.weight">
2068 <dim>384</dim>
2069 <dim>384</dim>
2070 </port>
2071 </output>
2072 </layer>
2073 <layer id="139" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2074 <data transpose_a="false" transpose_b="true" />
2075 <input>
2076 <port id="0" precision="FP32">
2077 <dim>-1</dim>
2078 <dim>-1</dim>
2079 <dim>384</dim>
2080 </port>
2081 <port id="1" precision="FP32">
2082 <dim>384</dim>
2083 <dim>384</dim>
2084 </port>
2085 </input>
2086 <output>
2087 <port id="2" precision="FP32">
2088 <dim>-1</dim>
2089 <dim>-1</dim>
2090 <dim>384</dim>
2091 </port>
2092 </output>
2093 </layer>
2094 <layer id="140" name="Constant_305886" type="Const" version="opset1">
2095 <data element_type="f32" shape="1, 1, 384" offset="57140384" size="1536" />
2096 <output>
2097 <port id="0" precision="FP32">
2098 <dim>1</dim>
2099 <dim>1</dim>
2100 <dim>384</dim>
2101 </port>
2102 </output>
2103 </layer>
2104 <layer id="141" name="__module.bert.encoder.layer.1.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
2105 <data auto_broadcast="numpy" />
2106 <input>
2107 <port id="0" precision="FP32">
2108 <dim>-1</dim>
2109 <dim>-1</dim>
2110 <dim>384</dim>
2111 </port>
2112 <port id="1" precision="FP32">
2113 <dim>1</dim>
2114 <dim>1</dim>
2115 <dim>384</dim>
2116 </port>
2117 </input>
2118 <output>
2119 <port id="2" precision="FP32" names="193,input.7">
2120 <dim>-1</dim>
2121 <dim>-1</dim>
2122 <dim>384</dim>
2123 </port>
2124 </output>
2125 </layer>
2126 <layer id="142" name="__module.bert.encoder.layer.1.attention.output/aten::add/Add" type="Add" version="opset1">
2127 <data auto_broadcast="numpy" />
2128 <input>
2129 <port id="0" precision="FP32">
2130 <dim>-1</dim>
2131 <dim>-1</dim>
2132 <dim>384</dim>
2133 </port>
2134 <port id="1" precision="FP32">
2135 <dim>-1</dim>
2136 <dim>-1</dim>
2137 <dim>384</dim>
2138 </port>
2139 </input>
2140 <output>
2141 <port id="2" precision="FP32" names="195">
2142 <dim>-1</dim>
2143 <dim>-1</dim>
2144 <dim>384</dim>
2145 </port>
2146 </output>
2147 </layer>
2148 <layer id="143" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2149 <data element_type="i32" shape="1" offset="47675412" size="4" />
2150 <output>
2151 <port id="0" precision="I32">
2152 <dim>1</dim>
2153 </port>
2154 </output>
2155 </layer>
2156 <layer id="144" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2157 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2158 <input>
2159 <port id="0" precision="FP32">
2160 <dim>-1</dim>
2161 <dim>-1</dim>
2162 <dim>384</dim>
2163 </port>
2164 <port id="1" precision="I32">
2165 <dim>1</dim>
2166 </port>
2167 </input>
2168 <output>
2169 <port id="2" precision="FP32">
2170 <dim>-1</dim>
2171 <dim>-1</dim>
2172 <dim>384</dim>
2173 </port>
2174 </output>
2175 </layer>
2176 <layer id="145" name="Constant_305887" type="Const" version="opset1">
2177 <data element_type="f32" shape="1, 1, 384" offset="57141920" size="1536" />
2178 <output>
2179 <port id="0" precision="FP32">
2180 <dim>1</dim>
2181 <dim>1</dim>
2182 <dim>384</dim>
2183 </port>
2184 </output>
2185 </layer>
2186 <layer id="146" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2187 <data auto_broadcast="numpy" />
2188 <input>
2189 <port id="0" precision="FP32">
2190 <dim>-1</dim>
2191 <dim>-1</dim>
2192 <dim>384</dim>
2193 </port>
2194 <port id="1" precision="FP32">
2195 <dim>1</dim>
2196 <dim>1</dim>
2197 <dim>384</dim>
2198 </port>
2199 </input>
2200 <output>
2201 <port id="2" precision="FP32">
2202 <dim>-1</dim>
2203 <dim>-1</dim>
2204 <dim>384</dim>
2205 </port>
2206 </output>
2207 </layer>
2208 <layer id="147" name="Constant_305888" type="Const" version="opset1">
2209 <data element_type="f32" shape="1, 1, 384" offset="57143456" size="1536" />
2210 <output>
2211 <port id="0" precision="FP32">
2212 <dim>1</dim>
2213 <dim>1</dim>
2214 <dim>384</dim>
2215 </port>
2216 </output>
2217 </layer>
2218 <layer id="148" name="__module.bert.encoder.layer.1.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2219 <data auto_broadcast="numpy" />
2220 <input>
2221 <port id="0" precision="FP32">
2222 <dim>-1</dim>
2223 <dim>-1</dim>
2224 <dim>384</dim>
2225 </port>
2226 <port id="1" precision="FP32">
2227 <dim>1</dim>
2228 <dim>1</dim>
2229 <dim>384</dim>
2230 </port>
2231 </input>
2232 <output>
2233 <port id="2" precision="FP32" names="199,input_tensor.3">
2234 <dim>-1</dim>
2235 <dim>-1</dim>
2236 <dim>384</dim>
2237 </port>
2238 </output>
2239 </layer>
2240 <layer id="149" name="self.bert.encoder.layer.1.intermediate.dense.weight" type="Const" version="opset1">
2241 <data element_type="f32" shape="1536, 384" offset="57144992" size="2359296" />
2242 <output>
2243 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.intermediate.dense.weight">
2244 <dim>1536</dim>
2245 <dim>384</dim>
2246 </port>
2247 </output>
2248 </layer>
2249 <layer id="150" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2250 <data transpose_a="false" transpose_b="true" />
2251 <input>
2252 <port id="0" precision="FP32">
2253 <dim>-1</dim>
2254 <dim>-1</dim>
2255 <dim>384</dim>
2256 </port>
2257 <port id="1" precision="FP32">
2258 <dim>1536</dim>
2259 <dim>384</dim>
2260 </port>
2261 </input>
2262 <output>
2263 <port id="2" precision="FP32">
2264 <dim>-1</dim>
2265 <dim>-1</dim>
2266 <dim>1536</dim>
2267 </port>
2268 </output>
2269 </layer>
2270 <layer id="151" name="Constant_305889" type="Const" version="opset1">
2271 <data element_type="f32" shape="1, 1, 1536" offset="59504288" size="6144" />
2272 <output>
2273 <port id="0" precision="FP32">
2274 <dim>1</dim>
2275 <dim>1</dim>
2276 <dim>1536</dim>
2277 </port>
2278 </output>
2279 </layer>
2280 <layer id="152" name="__module.bert.encoder.layer.1.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
2281 <data auto_broadcast="numpy" />
2282 <input>
2283 <port id="0" precision="FP32">
2284 <dim>-1</dim>
2285 <dim>-1</dim>
2286 <dim>1536</dim>
2287 </port>
2288 <port id="1" precision="FP32">
2289 <dim>1</dim>
2290 <dim>1</dim>
2291 <dim>1536</dim>
2292 </port>
2293 </input>
2294 <output>
2295 <port id="2" precision="FP32" names="203">
2296 <dim>-1</dim>
2297 <dim>-1</dim>
2298 <dim>1536</dim>
2299 </port>
2300 </output>
2301 </layer>
2302 <layer id="153" name="__module.bert.encoder.layer.1.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
2303 <data approximation_mode="ERF" />
2304 <input>
2305 <port id="0" precision="FP32">
2306 <dim>-1</dim>
2307 <dim>-1</dim>
2308 <dim>1536</dim>
2309 </port>
2310 </input>
2311 <output>
2312 <port id="1" precision="FP32" names="204">
2313 <dim>-1</dim>
2314 <dim>-1</dim>
2315 <dim>1536</dim>
2316 </port>
2317 </output>
2318 </layer>
2319 <layer id="154" name="self.bert.encoder.layer.1.output.dense.weight" type="Const" version="opset1">
2320 <data element_type="f32" shape="384, 1536" offset="59510432" size="2359296" />
2321 <output>
2322 <port id="0" precision="FP32" names="self.bert.encoder.layer.1.output.dense.weight">
2323 <dim>384</dim>
2324 <dim>1536</dim>
2325 </port>
2326 </output>
2327 </layer>
2328 <layer id="155" name="__module.bert.encoder.layer.1.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2329 <data transpose_a="false" transpose_b="true" />
2330 <input>
2331 <port id="0" precision="FP32">
2332 <dim>-1</dim>
2333 <dim>-1</dim>
2334 <dim>1536</dim>
2335 </port>
2336 <port id="1" precision="FP32">
2337 <dim>384</dim>
2338 <dim>1536</dim>
2339 </port>
2340 </input>
2341 <output>
2342 <port id="2" precision="FP32">
2343 <dim>-1</dim>
2344 <dim>-1</dim>
2345 <dim>384</dim>
2346 </port>
2347 </output>
2348 </layer>
2349 <layer id="156" name="Constant_305890" type="Const" version="opset1">
2350 <data element_type="f32" shape="1, 1, 384" offset="61869728" size="1536" />
2351 <output>
2352 <port id="0" precision="FP32">
2353 <dim>1</dim>
2354 <dim>1</dim>
2355 <dim>384</dim>
2356 </port>
2357 </output>
2358 </layer>
2359 <layer id="157" name="__module.bert.encoder.layer.1.output.dense/aten::linear/Add" type="Add" version="opset1">
2360 <data auto_broadcast="numpy" />
2361 <input>
2362 <port id="0" precision="FP32">
2363 <dim>-1</dim>
2364 <dim>-1</dim>
2365 <dim>384</dim>
2366 </port>
2367 <port id="1" precision="FP32">
2368 <dim>1</dim>
2369 <dim>1</dim>
2370 <dim>384</dim>
2371 </port>
2372 </input>
2373 <output>
2374 <port id="2" precision="FP32" names="209,input.9">
2375 <dim>-1</dim>
2376 <dim>-1</dim>
2377 <dim>384</dim>
2378 </port>
2379 </output>
2380 </layer>
2381 <layer id="158" name="__module.bert.encoder.layer.1.output/aten::add/Add" type="Add" version="opset1">
2382 <data auto_broadcast="numpy" />
2383 <input>
2384 <port id="0" precision="FP32">
2385 <dim>-1</dim>
2386 <dim>-1</dim>
2387 <dim>384</dim>
2388 </port>
2389 <port id="1" precision="FP32">
2390 <dim>-1</dim>
2391 <dim>-1</dim>
2392 <dim>384</dim>
2393 </port>
2394 </input>
2395 <output>
2396 <port id="2" precision="FP32" names="211">
2397 <dim>-1</dim>
2398 <dim>-1</dim>
2399 <dim>384</dim>
2400 </port>
2401 </output>
2402 </layer>
2403 <layer id="159" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
2404 <data element_type="i32" shape="1" offset="47675412" size="4" />
2405 <output>
2406 <port id="0" precision="I32">
2407 <dim>1</dim>
2408 </port>
2409 </output>
2410 </layer>
2411 <layer id="160" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
2412 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
2413 <input>
2414 <port id="0" precision="FP32">
2415 <dim>-1</dim>
2416 <dim>-1</dim>
2417 <dim>384</dim>
2418 </port>
2419 <port id="1" precision="I32">
2420 <dim>1</dim>
2421 </port>
2422 </input>
2423 <output>
2424 <port id="2" precision="FP32">
2425 <dim>-1</dim>
2426 <dim>-1</dim>
2427 <dim>384</dim>
2428 </port>
2429 </output>
2430 </layer>
2431 <layer id="161" name="Constant_305891" type="Const" version="opset1">
2432 <data element_type="f32" shape="1, 1, 384" offset="61871264" size="1536" />
2433 <output>
2434 <port id="0" precision="FP32">
2435 <dim>1</dim>
2436 <dim>1</dim>
2437 <dim>384</dim>
2438 </port>
2439 </output>
2440 </layer>
2441 <layer id="162" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
2442 <data auto_broadcast="numpy" />
2443 <input>
2444 <port id="0" precision="FP32">
2445 <dim>-1</dim>
2446 <dim>-1</dim>
2447 <dim>384</dim>
2448 </port>
2449 <port id="1" precision="FP32">
2450 <dim>1</dim>
2451 <dim>1</dim>
2452 <dim>384</dim>
2453 </port>
2454 </input>
2455 <output>
2456 <port id="2" precision="FP32">
2457 <dim>-1</dim>
2458 <dim>-1</dim>
2459 <dim>384</dim>
2460 </port>
2461 </output>
2462 </layer>
2463 <layer id="163" name="Constant_305892" type="Const" version="opset1">
2464 <data element_type="f32" shape="1, 1, 384" offset="61872800" size="1536" />
2465 <output>
2466 <port id="0" precision="FP32">
2467 <dim>1</dim>
2468 <dim>1</dim>
2469 <dim>384</dim>
2470 </port>
2471 </output>
2472 </layer>
2473 <layer id="164" name="__module.bert.encoder.layer.1.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
2474 <data auto_broadcast="numpy" />
2475 <input>
2476 <port id="0" precision="FP32">
2477 <dim>-1</dim>
2478 <dim>-1</dim>
2479 <dim>384</dim>
2480 </port>
2481 <port id="1" precision="FP32">
2482 <dim>1</dim>
2483 <dim>1</dim>
2484 <dim>384</dim>
2485 </port>
2486 </input>
2487 <output>
2488 <port id="2" precision="FP32" names="215,hidden_states.13">
2489 <dim>-1</dim>
2490 <dim>-1</dim>
2491 <dim>384</dim>
2492 </port>
2493 </output>
2494 </layer>
2495 <layer id="165" name="self.bert.encoder.layer.2.attention.self.query.weight" type="Const" version="opset1">
2496 <data element_type="f32" shape="384, 384" offset="61874336" size="589824" />
2497 <output>
2498 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.self.query.weight">
2499 <dim>384</dim>
2500 <dim>384</dim>
2501 </port>
2502 </output>
2503 </layer>
2504 <layer id="166" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
2505 <data transpose_a="false" transpose_b="true" />
2506 <input>
2507 <port id="0" precision="FP32">
2508 <dim>-1</dim>
2509 <dim>-1</dim>
2510 <dim>384</dim>
2511 </port>
2512 <port id="1" precision="FP32">
2513 <dim>384</dim>
2514 <dim>384</dim>
2515 </port>
2516 </input>
2517 <output>
2518 <port id="2" precision="FP32">
2519 <dim>-1</dim>
2520 <dim>-1</dim>
2521 <dim>384</dim>
2522 </port>
2523 </output>
2524 </layer>
2525 <layer id="167" name="Constant_305893" type="Const" version="opset1">
2526 <data element_type="f32" shape="1, 1, 384" offset="62464160" size="1536" />
2527 <output>
2528 <port id="0" precision="FP32">
2529 <dim>1</dim>
2530 <dim>1</dim>
2531 <dim>384</dim>
2532 </port>
2533 </output>
2534 </layer>
2535 <layer id="168" name="__module.bert.encoder.layer.2.attention.self.query/aten::linear/Add" type="Add" version="opset1">
2536 <data auto_broadcast="numpy" />
2537 <input>
2538 <port id="0" precision="FP32">
2539 <dim>-1</dim>
2540 <dim>-1</dim>
2541 <dim>384</dim>
2542 </port>
2543 <port id="1" precision="FP32">
2544 <dim>1</dim>
2545 <dim>1</dim>
2546 <dim>384</dim>
2547 </port>
2548 </input>
2549 <output>
2550 <port id="2" precision="FP32" names="228,x.25">
2551 <dim>-1</dim>
2552 <dim>-1</dim>
2553 <dim>384</dim>
2554 </port>
2555 </output>
2556 </layer>
2557 <layer id="169" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
2558 <data element_type="i64" shape="4" offset="48269848" size="32" />
2559 <output>
2560 <port id="0" precision="I64">
2561 <dim>4</dim>
2562 </port>
2563 </output>
2564 </layer>
2565 <layer id="170" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
2566 <data special_zero="true" />
2567 <input>
2568 <port id="0" precision="FP32">
2569 <dim>-1</dim>
2570 <dim>-1</dim>
2571 <dim>384</dim>
2572 </port>
2573 <port id="1" precision="I64">
2574 <dim>4</dim>
2575 </port>
2576 </input>
2577 <output>
2578 <port id="2" precision="FP32" names="232,x.27">
2579 <dim>-1</dim>
2580 <dim>-1</dim>
2581 <dim>12</dim>
2582 <dim>32</dim>
2583 </port>
2584 </output>
2585 </layer>
2586 <layer id="171" name="Constant_301512" type="Const" version="opset1">
2587 <data element_type="i64" shape="4" offset="48269880" size="32" />
2588 <output>
2589 <port id="0" precision="I64" names="233">
2590 <dim>4</dim>
2591 </port>
2592 </output>
2593 </layer>
2594 <layer id="172" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
2595 <input>
2596 <port id="0" precision="FP32">
2597 <dim>-1</dim>
2598 <dim>-1</dim>
2599 <dim>12</dim>
2600 <dim>32</dim>
2601 </port>
2602 <port id="1" precision="I64">
2603 <dim>4</dim>
2604 </port>
2605 </input>
2606 <output>
2607 <port id="2" precision="FP32" names="234">
2608 <dim>-1</dim>
2609 <dim>12</dim>
2610 <dim>-1</dim>
2611 <dim>32</dim>
2612 </port>
2613 </output>
2614 </layer>
2615 <layer id="173" name="self.bert.encoder.layer.2.attention.self.key.weight" type="Const" version="opset1">
2616 <data element_type="f32" shape="384, 384" offset="62465696" size="589824" />
2617 <output>
2618 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.self.key.weight">
2619 <dim>384</dim>
2620 <dim>384</dim>
2621 </port>
2622 </output>
2623 </layer>
2624 <layer id="174" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
2625 <data transpose_a="false" transpose_b="true" />
2626 <input>
2627 <port id="0" precision="FP32">
2628 <dim>-1</dim>
2629 <dim>-1</dim>
2630 <dim>384</dim>
2631 </port>
2632 <port id="1" precision="FP32">
2633 <dim>384</dim>
2634 <dim>384</dim>
2635 </port>
2636 </input>
2637 <output>
2638 <port id="2" precision="FP32">
2639 <dim>-1</dim>
2640 <dim>-1</dim>
2641 <dim>384</dim>
2642 </port>
2643 </output>
2644 </layer>
2645 <layer id="175" name="Constant_305894" type="Const" version="opset1">
2646 <data element_type="f32" shape="1, 1, 384" offset="63055520" size="1536" />
2647 <output>
2648 <port id="0" precision="FP32">
2649 <dim>1</dim>
2650 <dim>1</dim>
2651 <dim>384</dim>
2652 </port>
2653 </output>
2654 </layer>
2655 <layer id="176" name="__module.bert.encoder.layer.2.attention.self.key/aten::linear/Add" type="Add" version="opset1">
2656 <data auto_broadcast="numpy" />
2657 <input>
2658 <port id="0" precision="FP32">
2659 <dim>-1</dim>
2660 <dim>-1</dim>
2661 <dim>384</dim>
2662 </port>
2663 <port id="1" precision="FP32">
2664 <dim>1</dim>
2665 <dim>1</dim>
2666 <dim>384</dim>
2667 </port>
2668 </input>
2669 <output>
2670 <port id="2" precision="FP32" names="237,x.29">
2671 <dim>-1</dim>
2672 <dim>-1</dim>
2673 <dim>384</dim>
2674 </port>
2675 </output>
2676 </layer>
2677 <layer id="177" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
2678 <data element_type="i64" shape="4" offset="48269848" size="32" />
2679 <output>
2680 <port id="0" precision="I64">
2681 <dim>4</dim>
2682 </port>
2683 </output>
2684 </layer>
2685 <layer id="178" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
2686 <data special_zero="true" />
2687 <input>
2688 <port id="0" precision="FP32">
2689 <dim>-1</dim>
2690 <dim>-1</dim>
2691 <dim>384</dim>
2692 </port>
2693 <port id="1" precision="I64">
2694 <dim>4</dim>
2695 </port>
2696 </input>
2697 <output>
2698 <port id="2" precision="FP32" names="241,x.31">
2699 <dim>-1</dim>
2700 <dim>-1</dim>
2701 <dim>12</dim>
2702 <dim>32</dim>
2703 </port>
2704 </output>
2705 </layer>
2706 <layer id="179" name="Constant_301535" type="Const" version="opset1">
2707 <data element_type="i64" shape="4" offset="48269880" size="32" />
2708 <output>
2709 <port id="0" precision="I64" names="242">
2710 <dim>4</dim>
2711 </port>
2712 </output>
2713 </layer>
2714 <layer id="180" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
2715 <input>
2716 <port id="0" precision="FP32">
2717 <dim>-1</dim>
2718 <dim>-1</dim>
2719 <dim>12</dim>
2720 <dim>32</dim>
2721 </port>
2722 <port id="1" precision="I64">
2723 <dim>4</dim>
2724 </port>
2725 </input>
2726 <output>
2727 <port id="2" precision="FP32" names="243">
2728 <dim>-1</dim>
2729 <dim>12</dim>
2730 <dim>-1</dim>
2731 <dim>32</dim>
2732 </port>
2733 </output>
2734 </layer>
2735 <layer id="181" name="self.bert.encoder.layer.2.attention.self.value.weight" type="Const" version="opset1">
2736 <data element_type="f32" shape="384, 384" offset="63057056" size="589824" />
2737 <output>
2738 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.self.value.weight">
2739 <dim>384</dim>
2740 <dim>384</dim>
2741 </port>
2742 </output>
2743 </layer>
2744 <layer id="182" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
2745 <data transpose_a="false" transpose_b="true" />
2746 <input>
2747 <port id="0" precision="FP32">
2748 <dim>-1</dim>
2749 <dim>-1</dim>
2750 <dim>384</dim>
2751 </port>
2752 <port id="1" precision="FP32">
2753 <dim>384</dim>
2754 <dim>384</dim>
2755 </port>
2756 </input>
2757 <output>
2758 <port id="2" precision="FP32">
2759 <dim>-1</dim>
2760 <dim>-1</dim>
2761 <dim>384</dim>
2762 </port>
2763 </output>
2764 </layer>
2765 <layer id="183" name="Constant_305895" type="Const" version="opset1">
2766 <data element_type="f32" shape="1, 1, 384" offset="63646880" size="1536" />
2767 <output>
2768 <port id="0" precision="FP32">
2769 <dim>1</dim>
2770 <dim>1</dim>
2771 <dim>384</dim>
2772 </port>
2773 </output>
2774 </layer>
2775 <layer id="184" name="__module.bert.encoder.layer.2.attention.self.value/aten::linear/Add" type="Add" version="opset1">
2776 <data auto_broadcast="numpy" />
2777 <input>
2778 <port id="0" precision="FP32">
2779 <dim>-1</dim>
2780 <dim>-1</dim>
2781 <dim>384</dim>
2782 </port>
2783 <port id="1" precision="FP32">
2784 <dim>1</dim>
2785 <dim>1</dim>
2786 <dim>384</dim>
2787 </port>
2788 </input>
2789 <output>
2790 <port id="2" precision="FP32" names="246,x.33">
2791 <dim>-1</dim>
2792 <dim>-1</dim>
2793 <dim>384</dim>
2794 </port>
2795 </output>
2796 </layer>
2797 <layer id="185" name="__module.bert.encoder.layer.2.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
2798 <data element_type="i64" shape="4" offset="48269848" size="32" />
2799 <output>
2800 <port id="0" precision="I64">
2801 <dim>4</dim>
2802 </port>
2803 </output>
2804 </layer>
2805 <layer id="186" name="__module.bert.encoder.layer.2.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
2806 <data special_zero="true" />
2807 <input>
2808 <port id="0" precision="FP32">
2809 <dim>-1</dim>
2810 <dim>-1</dim>
2811 <dim>384</dim>
2812 </port>
2813 <port id="1" precision="I64">
2814 <dim>4</dim>
2815 </port>
2816 </input>
2817 <output>
2818 <port id="2" precision="FP32" names="250,x.35">
2819 <dim>-1</dim>
2820 <dim>-1</dim>
2821 <dim>12</dim>
2822 <dim>32</dim>
2823 </port>
2824 </output>
2825 </layer>
2826 <layer id="187" name="Constant_301558" type="Const" version="opset1">
2827 <data element_type="i64" shape="4" offset="48269880" size="32" />
2828 <output>
2829 <port id="0" precision="I64" names="251">
2830 <dim>4</dim>
2831 </port>
2832 </output>
2833 </layer>
2834 <layer id="188" name="__module.bert.encoder.layer.2.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
2835 <input>
2836 <port id="0" precision="FP32">
2837 <dim>-1</dim>
2838 <dim>-1</dim>
2839 <dim>12</dim>
2840 <dim>32</dim>
2841 </port>
2842 <port id="1" precision="I64">
2843 <dim>4</dim>
2844 </port>
2845 </input>
2846 <output>
2847 <port id="2" precision="FP32" names="252">
2848 <dim>-1</dim>
2849 <dim>12</dim>
2850 <dim>-1</dim>
2851 <dim>32</dim>
2852 </port>
2853 </output>
2854 </layer>
2855 <layer id="189" name="__module.bert.encoder.layer.2.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
2856 <data causal="false" />
2857 <input>
2858 <port id="0" precision="FP32">
2859 <dim>-1</dim>
2860 <dim>12</dim>
2861 <dim>-1</dim>
2862 <dim>32</dim>
2863 </port>
2864 <port id="1" precision="FP32">
2865 <dim>-1</dim>
2866 <dim>12</dim>
2867 <dim>-1</dim>
2868 <dim>32</dim>
2869 </port>
2870 <port id="2" precision="FP32">
2871 <dim>-1</dim>
2872 <dim>12</dim>
2873 <dim>-1</dim>
2874 <dim>32</dim>
2875 </port>
2876 <port id="3" precision="FP32">
2877 <dim>-1</dim>
2878 <dim>1</dim>
2879 <dim>-1</dim>
2880 <dim>-1</dim>
2881 </port>
2882 </input>
2883 <output>
2884 <port id="4" precision="FP32" names="253,attn_output.9">
2885 <dim>-1</dim>
2886 <dim>12</dim>
2887 <dim>-1</dim>
2888 <dim>32</dim>
2889 </port>
2890 </output>
2891 </layer>
2892 <layer id="190" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
2893 <data element_type="i32" shape="4" offset="49452664" size="16" />
2894 <output>
2895 <port id="0" precision="I32">
2896 <dim>4</dim>
2897 </port>
2898 </output>
2899 </layer>
2900 <layer id="191" name="__module.bert.encoder.layer.2.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
2901 <input>
2902 <port id="0" precision="FP32">
2903 <dim>-1</dim>
2904 <dim>12</dim>
2905 <dim>-1</dim>
2906 <dim>32</dim>
2907 </port>
2908 <port id="1" precision="I32">
2909 <dim>4</dim>
2910 </port>
2911 </input>
2912 <output>
2913 <port id="2" precision="FP32" names="254,attn_output.11">
2914 <dim>-1</dim>
2915 <dim>-1</dim>
2916 <dim>12</dim>
2917 <dim>32</dim>
2918 </port>
2919 </output>
2920 </layer>
2921 <layer id="192" name="Constant_305979" type="Const" version="opset1">
2922 <data element_type="i64" shape="3" offset="49452680" size="24" />
2923 <output>
2924 <port id="0" precision="I64">
2925 <dim>3</dim>
2926 </port>
2927 </output>
2928 </layer>
2929 <layer id="193" name="__module.bert.encoder.layer.2.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
2930 <data special_zero="true" />
2931 <input>
2932 <port id="0" precision="FP32">
2933 <dim>-1</dim>
2934 <dim>-1</dim>
2935 <dim>12</dim>
2936 <dim>32</dim>
2937 </port>
2938 <port id="1" precision="I64">
2939 <dim>3</dim>
2940 </port>
2941 </input>
2942 <output>
2943 <port id="2" precision="FP32" names="256">
2944 <dim>-1</dim>
2945 <dim>-1</dim>
2946 <dim>384</dim>
2947 </port>
2948 </output>
2949 </layer>
2950 <layer id="194" name="self.bert.encoder.layer.2.attention.output.dense.weight" type="Const" version="opset1">
2951 <data element_type="f32" shape="384, 384" offset="63648416" size="589824" />
2952 <output>
2953 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.attention.output.dense.weight">
2954 <dim>384</dim>
2955 <dim>384</dim>
2956 </port>
2957 </output>
2958 </layer>
2959 <layer id="195" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
2960 <data transpose_a="false" transpose_b="true" />
2961 <input>
2962 <port id="0" precision="FP32">
2963 <dim>-1</dim>
2964 <dim>-1</dim>
2965 <dim>384</dim>
2966 </port>
2967 <port id="1" precision="FP32">
2968 <dim>384</dim>
2969 <dim>384</dim>
2970 </port>
2971 </input>
2972 <output>
2973 <port id="2" precision="FP32">
2974 <dim>-1</dim>
2975 <dim>-1</dim>
2976 <dim>384</dim>
2977 </port>
2978 </output>
2979 </layer>
2980 <layer id="196" name="Constant_305896" type="Const" version="opset1">
2981 <data element_type="f32" shape="1, 1, 384" offset="64238240" size="1536" />
2982 <output>
2983 <port id="0" precision="FP32">
2984 <dim>1</dim>
2985 <dim>1</dim>
2986 <dim>384</dim>
2987 </port>
2988 </output>
2989 </layer>
2990 <layer id="197" name="__module.bert.encoder.layer.2.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
2991 <data auto_broadcast="numpy" />
2992 <input>
2993 <port id="0" precision="FP32">
2994 <dim>-1</dim>
2995 <dim>-1</dim>
2996 <dim>384</dim>
2997 </port>
2998 <port id="1" precision="FP32">
2999 <dim>1</dim>
3000 <dim>1</dim>
3001 <dim>384</dim>
3002 </port>
3003 </input>
3004 <output>
3005 <port id="2" precision="FP32" names="261,input.11">
3006 <dim>-1</dim>
3007 <dim>-1</dim>
3008 <dim>384</dim>
3009 </port>
3010 </output>
3011 </layer>
3012 <layer id="198" name="__module.bert.encoder.layer.2.attention.output/aten::add/Add" type="Add" version="opset1">
3013 <data auto_broadcast="numpy" />
3014 <input>
3015 <port id="0" precision="FP32">
3016 <dim>-1</dim>
3017 <dim>-1</dim>
3018 <dim>384</dim>
3019 </port>
3020 <port id="1" precision="FP32">
3021 <dim>-1</dim>
3022 <dim>-1</dim>
3023 <dim>384</dim>
3024 </port>
3025 </input>
3026 <output>
3027 <port id="2" precision="FP32" names="263">
3028 <dim>-1</dim>
3029 <dim>-1</dim>
3030 <dim>384</dim>
3031 </port>
3032 </output>
3033 </layer>
3034 <layer id="199" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3035 <data element_type="i32" shape="1" offset="47675412" size="4" />
3036 <output>
3037 <port id="0" precision="I32">
3038 <dim>1</dim>
3039 </port>
3040 </output>
3041 </layer>
3042 <layer id="200" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3043 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3044 <input>
3045 <port id="0" precision="FP32">
3046 <dim>-1</dim>
3047 <dim>-1</dim>
3048 <dim>384</dim>
3049 </port>
3050 <port id="1" precision="I32">
3051 <dim>1</dim>
3052 </port>
3053 </input>
3054 <output>
3055 <port id="2" precision="FP32">
3056 <dim>-1</dim>
3057 <dim>-1</dim>
3058 <dim>384</dim>
3059 </port>
3060 </output>
3061 </layer>
3062 <layer id="201" name="Constant_305897" type="Const" version="opset1">
3063 <data element_type="f32" shape="1, 1, 384" offset="64239776" size="1536" />
3064 <output>
3065 <port id="0" precision="FP32">
3066 <dim>1</dim>
3067 <dim>1</dim>
3068 <dim>384</dim>
3069 </port>
3070 </output>
3071 </layer>
3072 <layer id="202" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3073 <data auto_broadcast="numpy" />
3074 <input>
3075 <port id="0" precision="FP32">
3076 <dim>-1</dim>
3077 <dim>-1</dim>
3078 <dim>384</dim>
3079 </port>
3080 <port id="1" precision="FP32">
3081 <dim>1</dim>
3082 <dim>1</dim>
3083 <dim>384</dim>
3084 </port>
3085 </input>
3086 <output>
3087 <port id="2" precision="FP32">
3088 <dim>-1</dim>
3089 <dim>-1</dim>
3090 <dim>384</dim>
3091 </port>
3092 </output>
3093 </layer>
3094 <layer id="203" name="Constant_305898" type="Const" version="opset1">
3095 <data element_type="f32" shape="1, 1, 384" offset="64241312" size="1536" />
3096 <output>
3097 <port id="0" precision="FP32">
3098 <dim>1</dim>
3099 <dim>1</dim>
3100 <dim>384</dim>
3101 </port>
3102 </output>
3103 </layer>
3104 <layer id="204" name="__module.bert.encoder.layer.2.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3105 <data auto_broadcast="numpy" />
3106 <input>
3107 <port id="0" precision="FP32">
3108 <dim>-1</dim>
3109 <dim>-1</dim>
3110 <dim>384</dim>
3111 </port>
3112 <port id="1" precision="FP32">
3113 <dim>1</dim>
3114 <dim>1</dim>
3115 <dim>384</dim>
3116 </port>
3117 </input>
3118 <output>
3119 <port id="2" precision="FP32" names="267,input_tensor.5">
3120 <dim>-1</dim>
3121 <dim>-1</dim>
3122 <dim>384</dim>
3123 </port>
3124 </output>
3125 </layer>
3126 <layer id="205" name="self.bert.encoder.layer.2.intermediate.dense.weight" type="Const" version="opset1">
3127 <data element_type="f32" shape="1536, 384" offset="64242848" size="2359296" />
3128 <output>
3129 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.intermediate.dense.weight">
3130 <dim>1536</dim>
3131 <dim>384</dim>
3132 </port>
3133 </output>
3134 </layer>
3135 <layer id="206" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3136 <data transpose_a="false" transpose_b="true" />
3137 <input>
3138 <port id="0" precision="FP32">
3139 <dim>-1</dim>
3140 <dim>-1</dim>
3141 <dim>384</dim>
3142 </port>
3143 <port id="1" precision="FP32">
3144 <dim>1536</dim>
3145 <dim>384</dim>
3146 </port>
3147 </input>
3148 <output>
3149 <port id="2" precision="FP32">
3150 <dim>-1</dim>
3151 <dim>-1</dim>
3152 <dim>1536</dim>
3153 </port>
3154 </output>
3155 </layer>
3156 <layer id="207" name="Constant_305899" type="Const" version="opset1">
3157 <data element_type="f32" shape="1, 1, 1536" offset="66602144" size="6144" />
3158 <output>
3159 <port id="0" precision="FP32">
3160 <dim>1</dim>
3161 <dim>1</dim>
3162 <dim>1536</dim>
3163 </port>
3164 </output>
3165 </layer>
3166 <layer id="208" name="__module.bert.encoder.layer.2.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
3167 <data auto_broadcast="numpy" />
3168 <input>
3169 <port id="0" precision="FP32">
3170 <dim>-1</dim>
3171 <dim>-1</dim>
3172 <dim>1536</dim>
3173 </port>
3174 <port id="1" precision="FP32">
3175 <dim>1</dim>
3176 <dim>1</dim>
3177 <dim>1536</dim>
3178 </port>
3179 </input>
3180 <output>
3181 <port id="2" precision="FP32" names="271">
3182 <dim>-1</dim>
3183 <dim>-1</dim>
3184 <dim>1536</dim>
3185 </port>
3186 </output>
3187 </layer>
3188 <layer id="209" name="__module.bert.encoder.layer.2.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
3189 <data approximation_mode="ERF" />
3190 <input>
3191 <port id="0" precision="FP32">
3192 <dim>-1</dim>
3193 <dim>-1</dim>
3194 <dim>1536</dim>
3195 </port>
3196 </input>
3197 <output>
3198 <port id="1" precision="FP32" names="272">
3199 <dim>-1</dim>
3200 <dim>-1</dim>
3201 <dim>1536</dim>
3202 </port>
3203 </output>
3204 </layer>
3205 <layer id="210" name="self.bert.encoder.layer.2.output.dense.weight" type="Const" version="opset1">
3206 <data element_type="f32" shape="384, 1536" offset="66608288" size="2359296" />
3207 <output>
3208 <port id="0" precision="FP32" names="self.bert.encoder.layer.2.output.dense.weight">
3209 <dim>384</dim>
3210 <dim>1536</dim>
3211 </port>
3212 </output>
3213 </layer>
3214 <layer id="211" name="__module.bert.encoder.layer.2.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3215 <data transpose_a="false" transpose_b="true" />
3216 <input>
3217 <port id="0" precision="FP32">
3218 <dim>-1</dim>
3219 <dim>-1</dim>
3220 <dim>1536</dim>
3221 </port>
3222 <port id="1" precision="FP32">
3223 <dim>384</dim>
3224 <dim>1536</dim>
3225 </port>
3226 </input>
3227 <output>
3228 <port id="2" precision="FP32">
3229 <dim>-1</dim>
3230 <dim>-1</dim>
3231 <dim>384</dim>
3232 </port>
3233 </output>
3234 </layer>
3235 <layer id="212" name="Constant_305900" type="Const" version="opset1">
3236 <data element_type="f32" shape="1, 1, 384" offset="68967584" size="1536" />
3237 <output>
3238 <port id="0" precision="FP32">
3239 <dim>1</dim>
3240 <dim>1</dim>
3241 <dim>384</dim>
3242 </port>
3243 </output>
3244 </layer>
3245 <layer id="213" name="__module.bert.encoder.layer.2.output.dense/aten::linear/Add" type="Add" version="opset1">
3246 <data auto_broadcast="numpy" />
3247 <input>
3248 <port id="0" precision="FP32">
3249 <dim>-1</dim>
3250 <dim>-1</dim>
3251 <dim>384</dim>
3252 </port>
3253 <port id="1" precision="FP32">
3254 <dim>1</dim>
3255 <dim>1</dim>
3256 <dim>384</dim>
3257 </port>
3258 </input>
3259 <output>
3260 <port id="2" precision="FP32" names="277,input.13">
3261 <dim>-1</dim>
3262 <dim>-1</dim>
3263 <dim>384</dim>
3264 </port>
3265 </output>
3266 </layer>
3267 <layer id="214" name="__module.bert.encoder.layer.2.output/aten::add/Add" type="Add" version="opset1">
3268 <data auto_broadcast="numpy" />
3269 <input>
3270 <port id="0" precision="FP32">
3271 <dim>-1</dim>
3272 <dim>-1</dim>
3273 <dim>384</dim>
3274 </port>
3275 <port id="1" precision="FP32">
3276 <dim>-1</dim>
3277 <dim>-1</dim>
3278 <dim>384</dim>
3279 </port>
3280 </input>
3281 <output>
3282 <port id="2" precision="FP32" names="279">
3283 <dim>-1</dim>
3284 <dim>-1</dim>
3285 <dim>384</dim>
3286 </port>
3287 </output>
3288 </layer>
3289 <layer id="215" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3290 <data element_type="i32" shape="1" offset="47675412" size="4" />
3291 <output>
3292 <port id="0" precision="I32">
3293 <dim>1</dim>
3294 </port>
3295 </output>
3296 </layer>
3297 <layer id="216" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3298 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3299 <input>
3300 <port id="0" precision="FP32">
3301 <dim>-1</dim>
3302 <dim>-1</dim>
3303 <dim>384</dim>
3304 </port>
3305 <port id="1" precision="I32">
3306 <dim>1</dim>
3307 </port>
3308 </input>
3309 <output>
3310 <port id="2" precision="FP32">
3311 <dim>-1</dim>
3312 <dim>-1</dim>
3313 <dim>384</dim>
3314 </port>
3315 </output>
3316 </layer>
3317 <layer id="217" name="Constant_305901" type="Const" version="opset1">
3318 <data element_type="f32" shape="1, 1, 384" offset="68969120" size="1536" />
3319 <output>
3320 <port id="0" precision="FP32">
3321 <dim>1</dim>
3322 <dim>1</dim>
3323 <dim>384</dim>
3324 </port>
3325 </output>
3326 </layer>
3327 <layer id="218" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3328 <data auto_broadcast="numpy" />
3329 <input>
3330 <port id="0" precision="FP32">
3331 <dim>-1</dim>
3332 <dim>-1</dim>
3333 <dim>384</dim>
3334 </port>
3335 <port id="1" precision="FP32">
3336 <dim>1</dim>
3337 <dim>1</dim>
3338 <dim>384</dim>
3339 </port>
3340 </input>
3341 <output>
3342 <port id="2" precision="FP32">
3343 <dim>-1</dim>
3344 <dim>-1</dim>
3345 <dim>384</dim>
3346 </port>
3347 </output>
3348 </layer>
3349 <layer id="219" name="Constant_305902" type="Const" version="opset1">
3350 <data element_type="f32" shape="1, 1, 384" offset="68970656" size="1536" />
3351 <output>
3352 <port id="0" precision="FP32">
3353 <dim>1</dim>
3354 <dim>1</dim>
3355 <dim>384</dim>
3356 </port>
3357 </output>
3358 </layer>
3359 <layer id="220" name="__module.bert.encoder.layer.2.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3360 <data auto_broadcast="numpy" />
3361 <input>
3362 <port id="0" precision="FP32">
3363 <dim>-1</dim>
3364 <dim>-1</dim>
3365 <dim>384</dim>
3366 </port>
3367 <port id="1" precision="FP32">
3368 <dim>1</dim>
3369 <dim>1</dim>
3370 <dim>384</dim>
3371 </port>
3372 </input>
3373 <output>
3374 <port id="2" precision="FP32" names="283,hidden_states.19">
3375 <dim>-1</dim>
3376 <dim>-1</dim>
3377 <dim>384</dim>
3378 </port>
3379 </output>
3380 </layer>
3381 <layer id="221" name="self.bert.encoder.layer.3.attention.self.query.weight" type="Const" version="opset1">
3382 <data element_type="f32" shape="384, 384" offset="68972192" size="589824" />
3383 <output>
3384 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.self.query.weight">
3385 <dim>384</dim>
3386 <dim>384</dim>
3387 </port>
3388 </output>
3389 </layer>
3390 <layer id="222" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/MatMul" type="MatMul" version="opset1">
3391 <data transpose_a="false" transpose_b="true" />
3392 <input>
3393 <port id="0" precision="FP32">
3394 <dim>-1</dim>
3395 <dim>-1</dim>
3396 <dim>384</dim>
3397 </port>
3398 <port id="1" precision="FP32">
3399 <dim>384</dim>
3400 <dim>384</dim>
3401 </port>
3402 </input>
3403 <output>
3404 <port id="2" precision="FP32">
3405 <dim>-1</dim>
3406 <dim>-1</dim>
3407 <dim>384</dim>
3408 </port>
3409 </output>
3410 </layer>
3411 <layer id="223" name="Constant_305903" type="Const" version="opset1">
3412 <data element_type="f32" shape="1, 1, 384" offset="69562016" size="1536" />
3413 <output>
3414 <port id="0" precision="FP32">
3415 <dim>1</dim>
3416 <dim>1</dim>
3417 <dim>384</dim>
3418 </port>
3419 </output>
3420 </layer>
3421 <layer id="224" name="__module.bert.encoder.layer.3.attention.self.query/aten::linear/Add" type="Add" version="opset1">
3422 <data auto_broadcast="numpy" />
3423 <input>
3424 <port id="0" precision="FP32">
3425 <dim>-1</dim>
3426 <dim>-1</dim>
3427 <dim>384</dim>
3428 </port>
3429 <port id="1" precision="FP32">
3430 <dim>1</dim>
3431 <dim>1</dim>
3432 <dim>384</dim>
3433 </port>
3434 </input>
3435 <output>
3436 <port id="2" precision="FP32" names="296,x.37">
3437 <dim>-1</dim>
3438 <dim>-1</dim>
3439 <dim>384</dim>
3440 </port>
3441 </output>
3442 </layer>
3443 <layer id="225" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat" type="Const" version="opset1">
3444 <data element_type="i64" shape="4" offset="48269848" size="32" />
3445 <output>
3446 <port id="0" precision="I64">
3447 <dim>4</dim>
3448 </port>
3449 </output>
3450 </layer>
3451 <layer id="226" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape" type="Reshape" version="opset1">
3452 <data special_zero="true" />
3453 <input>
3454 <port id="0" precision="FP32">
3455 <dim>-1</dim>
3456 <dim>-1</dim>
3457 <dim>384</dim>
3458 </port>
3459 <port id="1" precision="I64">
3460 <dim>4</dim>
3461 </port>
3462 </input>
3463 <output>
3464 <port id="2" precision="FP32" names="300,x.39">
3465 <dim>-1</dim>
3466 <dim>-1</dim>
3467 <dim>12</dim>
3468 <dim>32</dim>
3469 </port>
3470 </output>
3471 </layer>
3472 <layer id="227" name="Constant_301735" type="Const" version="opset1">
3473 <data element_type="i64" shape="4" offset="48269880" size="32" />
3474 <output>
3475 <port id="0" precision="I64" names="301">
3476 <dim>4</dim>
3477 </port>
3478 </output>
3479 </layer>
3480 <layer id="228" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose" type="Transpose" version="opset1">
3481 <input>
3482 <port id="0" precision="FP32">
3483 <dim>-1</dim>
3484 <dim>-1</dim>
3485 <dim>12</dim>
3486 <dim>32</dim>
3487 </port>
3488 <port id="1" precision="I64">
3489 <dim>4</dim>
3490 </port>
3491 </input>
3492 <output>
3493 <port id="2" precision="FP32" names="302">
3494 <dim>-1</dim>
3495 <dim>12</dim>
3496 <dim>-1</dim>
3497 <dim>32</dim>
3498 </port>
3499 </output>
3500 </layer>
3501 <layer id="229" name="self.bert.encoder.layer.3.attention.self.key.weight" type="Const" version="opset1">
3502 <data element_type="f32" shape="384, 384" offset="69563552" size="589824" />
3503 <output>
3504 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.self.key.weight">
3505 <dim>384</dim>
3506 <dim>384</dim>
3507 </port>
3508 </output>
3509 </layer>
3510 <layer id="230" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/MatMul" type="MatMul" version="opset1">
3511 <data transpose_a="false" transpose_b="true" />
3512 <input>
3513 <port id="0" precision="FP32">
3514 <dim>-1</dim>
3515 <dim>-1</dim>
3516 <dim>384</dim>
3517 </port>
3518 <port id="1" precision="FP32">
3519 <dim>384</dim>
3520 <dim>384</dim>
3521 </port>
3522 </input>
3523 <output>
3524 <port id="2" precision="FP32">
3525 <dim>-1</dim>
3526 <dim>-1</dim>
3527 <dim>384</dim>
3528 </port>
3529 </output>
3530 </layer>
3531 <layer id="231" name="Constant_305904" type="Const" version="opset1">
3532 <data element_type="f32" shape="1, 1, 384" offset="70153376" size="1536" />
3533 <output>
3534 <port id="0" precision="FP32">
3535 <dim>1</dim>
3536 <dim>1</dim>
3537 <dim>384</dim>
3538 </port>
3539 </output>
3540 </layer>
3541 <layer id="232" name="__module.bert.encoder.layer.3.attention.self.key/aten::linear/Add" type="Add" version="opset1">
3542 <data auto_broadcast="numpy" />
3543 <input>
3544 <port id="0" precision="FP32">
3545 <dim>-1</dim>
3546 <dim>-1</dim>
3547 <dim>384</dim>
3548 </port>
3549 <port id="1" precision="FP32">
3550 <dim>1</dim>
3551 <dim>1</dim>
3552 <dim>384</dim>
3553 </port>
3554 </input>
3555 <output>
3556 <port id="2" precision="FP32" names="305,x.41">
3557 <dim>-1</dim>
3558 <dim>-1</dim>
3559 <dim>384</dim>
3560 </port>
3561 </output>
3562 </layer>
3563 <layer id="233" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_1" type="Const" version="opset1">
3564 <data element_type="i64" shape="4" offset="48269848" size="32" />
3565 <output>
3566 <port id="0" precision="I64">
3567 <dim>4</dim>
3568 </port>
3569 </output>
3570 </layer>
3571 <layer id="234" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_1" type="Reshape" version="opset1">
3572 <data special_zero="true" />
3573 <input>
3574 <port id="0" precision="FP32">
3575 <dim>-1</dim>
3576 <dim>-1</dim>
3577 <dim>384</dim>
3578 </port>
3579 <port id="1" precision="I64">
3580 <dim>4</dim>
3581 </port>
3582 </input>
3583 <output>
3584 <port id="2" precision="FP32" names="309,x.43">
3585 <dim>-1</dim>
3586 <dim>-1</dim>
3587 <dim>12</dim>
3588 <dim>32</dim>
3589 </port>
3590 </output>
3591 </layer>
3592 <layer id="235" name="Constant_301758" type="Const" version="opset1">
3593 <data element_type="i64" shape="4" offset="48269880" size="32" />
3594 <output>
3595 <port id="0" precision="I64" names="310">
3596 <dim>4</dim>
3597 </port>
3598 </output>
3599 </layer>
3600 <layer id="236" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_1" type="Transpose" version="opset1">
3601 <input>
3602 <port id="0" precision="FP32">
3603 <dim>-1</dim>
3604 <dim>-1</dim>
3605 <dim>12</dim>
3606 <dim>32</dim>
3607 </port>
3608 <port id="1" precision="I64">
3609 <dim>4</dim>
3610 </port>
3611 </input>
3612 <output>
3613 <port id="2" precision="FP32" names="311">
3614 <dim>-1</dim>
3615 <dim>12</dim>
3616 <dim>-1</dim>
3617 <dim>32</dim>
3618 </port>
3619 </output>
3620 </layer>
3621 <layer id="237" name="self.bert.encoder.layer.3.attention.self.value.weight" type="Const" version="opset1">
3622 <data element_type="f32" shape="384, 384" offset="70154912" size="589824" />
3623 <output>
3624 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.self.value.weight">
3625 <dim>384</dim>
3626 <dim>384</dim>
3627 </port>
3628 </output>
3629 </layer>
3630 <layer id="238" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/MatMul" type="MatMul" version="opset1">
3631 <data transpose_a="false" transpose_b="true" />
3632 <input>
3633 <port id="0" precision="FP32">
3634 <dim>-1</dim>
3635 <dim>-1</dim>
3636 <dim>384</dim>
3637 </port>
3638 <port id="1" precision="FP32">
3639 <dim>384</dim>
3640 <dim>384</dim>
3641 </port>
3642 </input>
3643 <output>
3644 <port id="2" precision="FP32">
3645 <dim>-1</dim>
3646 <dim>-1</dim>
3647 <dim>384</dim>
3648 </port>
3649 </output>
3650 </layer>
3651 <layer id="239" name="Constant_305905" type="Const" version="opset1">
3652 <data element_type="f32" shape="1, 1, 384" offset="70744736" size="1536" />
3653 <output>
3654 <port id="0" precision="FP32">
3655 <dim>1</dim>
3656 <dim>1</dim>
3657 <dim>384</dim>
3658 </port>
3659 </output>
3660 </layer>
3661 <layer id="240" name="__module.bert.encoder.layer.3.attention.self.value/aten::linear/Add" type="Add" version="opset1">
3662 <data auto_broadcast="numpy" />
3663 <input>
3664 <port id="0" precision="FP32">
3665 <dim>-1</dim>
3666 <dim>-1</dim>
3667 <dim>384</dim>
3668 </port>
3669 <port id="1" precision="FP32">
3670 <dim>1</dim>
3671 <dim>1</dim>
3672 <dim>384</dim>
3673 </port>
3674 </input>
3675 <output>
3676 <port id="2" precision="FP32" names="314,x.45">
3677 <dim>-1</dim>
3678 <dim>-1</dim>
3679 <dim>384</dim>
3680 </port>
3681 </output>
3682 </layer>
3683 <layer id="241" name="__module.bert.encoder.layer.3.attention.self/prim::ListConstruct/Concat_2" type="Const" version="opset1">
3684 <data element_type="i64" shape="4" offset="48269848" size="32" />
3685 <output>
3686 <port id="0" precision="I64">
3687 <dim>4</dim>
3688 </port>
3689 </output>
3690 </layer>
3691 <layer id="242" name="__module.bert.encoder.layer.3.attention.self/aten::view/Reshape_2" type="Reshape" version="opset1">
3692 <data special_zero="true" />
3693 <input>
3694 <port id="0" precision="FP32">
3695 <dim>-1</dim>
3696 <dim>-1</dim>
3697 <dim>384</dim>
3698 </port>
3699 <port id="1" precision="I64">
3700 <dim>4</dim>
3701 </port>
3702 </input>
3703 <output>
3704 <port id="2" precision="FP32" names="318,x">
3705 <dim>-1</dim>
3706 <dim>-1</dim>
3707 <dim>12</dim>
3708 <dim>32</dim>
3709 </port>
3710 </output>
3711 </layer>
3712 <layer id="243" name="Constant_301781" type="Const" version="opset1">
3713 <data element_type="i64" shape="4" offset="48269880" size="32" />
3714 <output>
3715 <port id="0" precision="I64" names="319">
3716 <dim>4</dim>
3717 </port>
3718 </output>
3719 </layer>
3720 <layer id="244" name="__module.bert.encoder.layer.3.attention.self/aten::permute/Transpose_2" type="Transpose" version="opset1">
3721 <input>
3722 <port id="0" precision="FP32">
3723 <dim>-1</dim>
3724 <dim>-1</dim>
3725 <dim>12</dim>
3726 <dim>32</dim>
3727 </port>
3728 <port id="1" precision="I64">
3729 <dim>4</dim>
3730 </port>
3731 </input>
3732 <output>
3733 <port id="2" precision="FP32" names="320">
3734 <dim>-1</dim>
3735 <dim>12</dim>
3736 <dim>-1</dim>
3737 <dim>32</dim>
3738 </port>
3739 </output>
3740 </layer>
3741 <layer id="245" name="__module.bert.encoder.layer.3.attention.self/aten::scaled_dot_product_attention/ScaledDotProductAttention" type="ScaledDotProductAttention" version="opset13">
3742 <data causal="false" />
3743 <input>
3744 <port id="0" precision="FP32">
3745 <dim>-1</dim>
3746 <dim>12</dim>
3747 <dim>-1</dim>
3748 <dim>32</dim>
3749 </port>
3750 <port id="1" precision="FP32">
3751 <dim>-1</dim>
3752 <dim>12</dim>
3753 <dim>-1</dim>
3754 <dim>32</dim>
3755 </port>
3756 <port id="2" precision="FP32">
3757 <dim>-1</dim>
3758 <dim>12</dim>
3759 <dim>-1</dim>
3760 <dim>32</dim>
3761 </port>
3762 <port id="3" precision="FP32">
3763 <dim>-1</dim>
3764 <dim>1</dim>
3765 <dim>-1</dim>
3766 <dim>-1</dim>
3767 </port>
3768 </input>
3769 <output>
3770 <port id="4" precision="FP32" names="321,attn_output.13">
3771 <dim>-1</dim>
3772 <dim>12</dim>
3773 <dim>-1</dim>
3774 <dim>32</dim>
3775 </port>
3776 </output>
3777 </layer>
3778 <layer id="246" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/ScatterElementsUpdate" type="Const" version="opset1">
3779 <data element_type="i32" shape="4" offset="49452664" size="16" />
3780 <output>
3781 <port id="0" precision="I32">
3782 <dim>4</dim>
3783 </port>
3784 </output>
3785 </layer>
3786 <layer id="247" name="__module.bert.encoder.layer.3.attention.self/aten::transpose/Transpose" type="Transpose" version="opset1">
3787 <input>
3788 <port id="0" precision="FP32">
3789 <dim>-1</dim>
3790 <dim>12</dim>
3791 <dim>-1</dim>
3792 <dim>32</dim>
3793 </port>
3794 <port id="1" precision="I32">
3795 <dim>4</dim>
3796 </port>
3797 </input>
3798 <output>
3799 <port id="2" precision="FP32" names="322,attn_output">
3800 <dim>-1</dim>
3801 <dim>-1</dim>
3802 <dim>12</dim>
3803 <dim>32</dim>
3804 </port>
3805 </output>
3806 </layer>
3807 <layer id="248" name="Constant_305980" type="Const" version="opset1">
3808 <data element_type="i64" shape="3" offset="49452680" size="24" />
3809 <output>
3810 <port id="0" precision="I64">
3811 <dim>3</dim>
3812 </port>
3813 </output>
3814 </layer>
3815 <layer id="249" name="__module.bert.encoder.layer.3.attention.self/aten::reshape/Reshape" type="Reshape" version="opset1">
3816 <data special_zero="true" />
3817 <input>
3818 <port id="0" precision="FP32">
3819 <dim>-1</dim>
3820 <dim>-1</dim>
3821 <dim>12</dim>
3822 <dim>32</dim>
3823 </port>
3824 <port id="1" precision="I64">
3825 <dim>3</dim>
3826 </port>
3827 </input>
3828 <output>
3829 <port id="2" precision="FP32" names="324">
3830 <dim>-1</dim>
3831 <dim>-1</dim>
3832 <dim>384</dim>
3833 </port>
3834 </output>
3835 </layer>
3836 <layer id="250" name="self.bert.encoder.layer.3.attention.output.dense.weight" type="Const" version="opset1">
3837 <data element_type="f32" shape="384, 384" offset="70746272" size="589824" />
3838 <output>
3839 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.attention.output.dense.weight">
3840 <dim>384</dim>
3841 <dim>384</dim>
3842 </port>
3843 </output>
3844 </layer>
3845 <layer id="251" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
3846 <data transpose_a="false" transpose_b="true" />
3847 <input>
3848 <port id="0" precision="FP32">
3849 <dim>-1</dim>
3850 <dim>-1</dim>
3851 <dim>384</dim>
3852 </port>
3853 <port id="1" precision="FP32">
3854 <dim>384</dim>
3855 <dim>384</dim>
3856 </port>
3857 </input>
3858 <output>
3859 <port id="2" precision="FP32">
3860 <dim>-1</dim>
3861 <dim>-1</dim>
3862 <dim>384</dim>
3863 </port>
3864 </output>
3865 </layer>
3866 <layer id="252" name="Constant_305906" type="Const" version="opset1">
3867 <data element_type="f32" shape="1, 1, 384" offset="71336096" size="1536" />
3868 <output>
3869 <port id="0" precision="FP32">
3870 <dim>1</dim>
3871 <dim>1</dim>
3872 <dim>384</dim>
3873 </port>
3874 </output>
3875 </layer>
3876 <layer id="253" name="__module.bert.encoder.layer.3.attention.output.dense/aten::linear/Add" type="Add" version="opset1">
3877 <data auto_broadcast="numpy" />
3878 <input>
3879 <port id="0" precision="FP32">
3880 <dim>-1</dim>
3881 <dim>-1</dim>
3882 <dim>384</dim>
3883 </port>
3884 <port id="1" precision="FP32">
3885 <dim>1</dim>
3886 <dim>1</dim>
3887 <dim>384</dim>
3888 </port>
3889 </input>
3890 <output>
3891 <port id="2" precision="FP32" names="329,input.15">
3892 <dim>-1</dim>
3893 <dim>-1</dim>
3894 <dim>384</dim>
3895 </port>
3896 </output>
3897 </layer>
3898 <layer id="254" name="__module.bert.encoder.layer.3.attention.output/aten::add/Add" type="Add" version="opset1">
3899 <data auto_broadcast="numpy" />
3900 <input>
3901 <port id="0" precision="FP32">
3902 <dim>-1</dim>
3903 <dim>-1</dim>
3904 <dim>384</dim>
3905 </port>
3906 <port id="1" precision="FP32">
3907 <dim>-1</dim>
3908 <dim>-1</dim>
3909 <dim>384</dim>
3910 </port>
3911 </input>
3912 <output>
3913 <port id="2" precision="FP32" names="331">
3914 <dim>-1</dim>
3915 <dim>-1</dim>
3916 <dim>384</dim>
3917 </port>
3918 </output>
3919 </layer>
3920 <layer id="255" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
3921 <data element_type="i32" shape="1" offset="47675412" size="4" />
3922 <output>
3923 <port id="0" precision="I32">
3924 <dim>1</dim>
3925 </port>
3926 </output>
3927 </layer>
3928 <layer id="256" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
3929 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
3930 <input>
3931 <port id="0" precision="FP32">
3932 <dim>-1</dim>
3933 <dim>-1</dim>
3934 <dim>384</dim>
3935 </port>
3936 <port id="1" precision="I32">
3937 <dim>1</dim>
3938 </port>
3939 </input>
3940 <output>
3941 <port id="2" precision="FP32">
3942 <dim>-1</dim>
3943 <dim>-1</dim>
3944 <dim>384</dim>
3945 </port>
3946 </output>
3947 </layer>
3948 <layer id="257" name="Constant_305907" type="Const" version="opset1">
3949 <data element_type="f32" shape="1, 1, 384" offset="71337632" size="1536" />
3950 <output>
3951 <port id="0" precision="FP32">
3952 <dim>1</dim>
3953 <dim>1</dim>
3954 <dim>384</dim>
3955 </port>
3956 </output>
3957 </layer>
3958 <layer id="258" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
3959 <data auto_broadcast="numpy" />
3960 <input>
3961 <port id="0" precision="FP32">
3962 <dim>-1</dim>
3963 <dim>-1</dim>
3964 <dim>384</dim>
3965 </port>
3966 <port id="1" precision="FP32">
3967 <dim>1</dim>
3968 <dim>1</dim>
3969 <dim>384</dim>
3970 </port>
3971 </input>
3972 <output>
3973 <port id="2" precision="FP32">
3974 <dim>-1</dim>
3975 <dim>-1</dim>
3976 <dim>384</dim>
3977 </port>
3978 </output>
3979 </layer>
3980 <layer id="259" name="Constant_305908" type="Const" version="opset1">
3981 <data element_type="f32" shape="1, 1, 384" offset="71339168" size="1536" />
3982 <output>
3983 <port id="0" precision="FP32">
3984 <dim>1</dim>
3985 <dim>1</dim>
3986 <dim>384</dim>
3987 </port>
3988 </output>
3989 </layer>
3990 <layer id="260" name="__module.bert.encoder.layer.3.attention.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
3991 <data auto_broadcast="numpy" />
3992 <input>
3993 <port id="0" precision="FP32">
3994 <dim>-1</dim>
3995 <dim>-1</dim>
3996 <dim>384</dim>
3997 </port>
3998 <port id="1" precision="FP32">
3999 <dim>1</dim>
4000 <dim>1</dim>
4001 <dim>384</dim>
4002 </port>
4003 </input>
4004 <output>
4005 <port id="2" precision="FP32" names="335,input_tensor">
4006 <dim>-1</dim>
4007 <dim>-1</dim>
4008 <dim>384</dim>
4009 </port>
4010 </output>
4011 </layer>
4012 <layer id="261" name="self.bert.encoder.layer.3.intermediate.dense.weight" type="Const" version="opset1">
4013 <data element_type="f32" shape="1536, 384" offset="71340704" size="2359296" />
4014 <output>
4015 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.intermediate.dense.weight">
4016 <dim>1536</dim>
4017 <dim>384</dim>
4018 </port>
4019 </output>
4020 </layer>
4021 <layer id="262" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4022 <data transpose_a="false" transpose_b="true" />
4023 <input>
4024 <port id="0" precision="FP32">
4025 <dim>-1</dim>
4026 <dim>-1</dim>
4027 <dim>384</dim>
4028 </port>
4029 <port id="1" precision="FP32">
4030 <dim>1536</dim>
4031 <dim>384</dim>
4032 </port>
4033 </input>
4034 <output>
4035 <port id="2" precision="FP32">
4036 <dim>-1</dim>
4037 <dim>-1</dim>
4038 <dim>1536</dim>
4039 </port>
4040 </output>
4041 </layer>
4042 <layer id="263" name="Constant_305909" type="Const" version="opset1">
4043 <data element_type="f32" shape="1, 1, 1536" offset="73700000" size="6144" />
4044 <output>
4045 <port id="0" precision="FP32">
4046 <dim>1</dim>
4047 <dim>1</dim>
4048 <dim>1536</dim>
4049 </port>
4050 </output>
4051 </layer>
4052 <layer id="264" name="__module.bert.encoder.layer.3.intermediate.dense/aten::linear/Add" type="Add" version="opset1">
4053 <data auto_broadcast="numpy" />
4054 <input>
4055 <port id="0" precision="FP32">
4056 <dim>-1</dim>
4057 <dim>-1</dim>
4058 <dim>1536</dim>
4059 </port>
4060 <port id="1" precision="FP32">
4061 <dim>1</dim>
4062 <dim>1</dim>
4063 <dim>1536</dim>
4064 </port>
4065 </input>
4066 <output>
4067 <port id="2" precision="FP32" names="339">
4068 <dim>-1</dim>
4069 <dim>-1</dim>
4070 <dim>1536</dim>
4071 </port>
4072 </output>
4073 </layer>
4074 <layer id="265" name="__module.bert.encoder.layer.3.intermediate.intermediate_act_fn/aten::gelu/Gelu" type="Gelu" version="opset7">
4075 <data approximation_mode="ERF" />
4076 <input>
4077 <port id="0" precision="FP32">
4078 <dim>-1</dim>
4079 <dim>-1</dim>
4080 <dim>1536</dim>
4081 </port>
4082 </input>
4083 <output>
4084 <port id="1" precision="FP32" names="340">
4085 <dim>-1</dim>
4086 <dim>-1</dim>
4087 <dim>1536</dim>
4088 </port>
4089 </output>
4090 </layer>
4091 <layer id="266" name="self.bert.encoder.layer.3.output.dense.weight" type="Const" version="opset1">
4092 <data element_type="f32" shape="384, 1536" offset="73706144" size="2359296" />
4093 <output>
4094 <port id="0" precision="FP32" names="self.bert.encoder.layer.3.output.dense.weight">
4095 <dim>384</dim>
4096 <dim>1536</dim>
4097 </port>
4098 </output>
4099 </layer>
4100 <layer id="267" name="__module.bert.encoder.layer.3.output.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4101 <data transpose_a="false" transpose_b="true" />
4102 <input>
4103 <port id="0" precision="FP32">
4104 <dim>-1</dim>
4105 <dim>-1</dim>
4106 <dim>1536</dim>
4107 </port>
4108 <port id="1" precision="FP32">
4109 <dim>384</dim>
4110 <dim>1536</dim>
4111 </port>
4112 </input>
4113 <output>
4114 <port id="2" precision="FP32">
4115 <dim>-1</dim>
4116 <dim>-1</dim>
4117 <dim>384</dim>
4118 </port>
4119 </output>
4120 </layer>
4121 <layer id="268" name="Constant_305910" type="Const" version="opset1">
4122 <data element_type="f32" shape="1, 1, 384" offset="76065440" size="1536" />
4123 <output>
4124 <port id="0" precision="FP32">
4125 <dim>1</dim>
4126 <dim>1</dim>
4127 <dim>384</dim>
4128 </port>
4129 </output>
4130 </layer>
4131 <layer id="269" name="__module.bert.encoder.layer.3.output.dense/aten::linear/Add" type="Add" version="opset1">
4132 <data auto_broadcast="numpy" />
4133 <input>
4134 <port id="0" precision="FP32">
4135 <dim>-1</dim>
4136 <dim>-1</dim>
4137 <dim>384</dim>
4138 </port>
4139 <port id="1" precision="FP32">
4140 <dim>1</dim>
4141 <dim>1</dim>
4142 <dim>384</dim>
4143 </port>
4144 </input>
4145 <output>
4146 <port id="2" precision="FP32" names="345,input.17">
4147 <dim>-1</dim>
4148 <dim>-1</dim>
4149 <dim>384</dim>
4150 </port>
4151 </output>
4152 </layer>
4153 <layer id="270" name="__module.bert.encoder.layer.3.output/aten::add/Add" type="Add" version="opset1">
4154 <data auto_broadcast="numpy" />
4155 <input>
4156 <port id="0" precision="FP32">
4157 <dim>-1</dim>
4158 <dim>-1</dim>
4159 <dim>384</dim>
4160 </port>
4161 <port id="1" precision="FP32">
4162 <dim>-1</dim>
4163 <dim>-1</dim>
4164 <dim>384</dim>
4165 </port>
4166 </input>
4167 <output>
4168 <port id="2" precision="FP32" names="347">
4169 <dim>-1</dim>
4170 <dim>-1</dim>
4171 <dim>384</dim>
4172 </port>
4173 </output>
4174 </layer>
4175 <layer id="271" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply" type="Const" version="opset1">
4176 <data element_type="i32" shape="1" offset="47675412" size="4" />
4177 <output>
4178 <port id="0" precision="I32">
4179 <dim>1</dim>
4180 </port>
4181 </output>
4182 </layer>
4183 <layer id="272" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/MVN" type="MVN" version="opset6">
4184 <data eps="9.999999960041972e-13" normalize_variance="true" eps_mode="INSIDE_SQRT" />
4185 <input>
4186 <port id="0" precision="FP32">
4187 <dim>-1</dim>
4188 <dim>-1</dim>
4189 <dim>384</dim>
4190 </port>
4191 <port id="1" precision="I32">
4192 <dim>1</dim>
4193 </port>
4194 </input>
4195 <output>
4196 <port id="2" precision="FP32">
4197 <dim>-1</dim>
4198 <dim>-1</dim>
4199 <dim>384</dim>
4200 </port>
4201 </output>
4202 </layer>
4203 <layer id="273" name="Constant_305911" type="Const" version="opset1">
4204 <data element_type="f32" shape="1, 1, 384" offset="76066976" size="1536" />
4205 <output>
4206 <port id="0" precision="FP32">
4207 <dim>1</dim>
4208 <dim>1</dim>
4209 <dim>384</dim>
4210 </port>
4211 </output>
4212 </layer>
4213 <layer id="274" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Multiply_1" type="Multiply" version="opset1">
4214 <data auto_broadcast="numpy" />
4215 <input>
4216 <port id="0" precision="FP32">
4217 <dim>-1</dim>
4218 <dim>-1</dim>
4219 <dim>384</dim>
4220 </port>
4221 <port id="1" precision="FP32">
4222 <dim>1</dim>
4223 <dim>1</dim>
4224 <dim>384</dim>
4225 </port>
4226 </input>
4227 <output>
4228 <port id="2" precision="FP32">
4229 <dim>-1</dim>
4230 <dim>-1</dim>
4231 <dim>384</dim>
4232 </port>
4233 </output>
4234 </layer>
4235 <layer id="275" name="Constant_305912" type="Const" version="opset1">
4236 <data element_type="f32" shape="1, 1, 384" offset="76068512" size="1536" />
4237 <output>
4238 <port id="0" precision="FP32">
4239 <dim>1</dim>
4240 <dim>1</dim>
4241 <dim>384</dim>
4242 </port>
4243 </output>
4244 </layer>
4245 <layer id="276" name="__module.bert.encoder.layer.3.output.LayerNorm/aten::layer_norm/Add" type="Add" version="opset1">
4246 <data auto_broadcast="numpy" />
4247 <input>
4248 <port id="0" precision="FP32">
4249 <dim>-1</dim>
4250 <dim>-1</dim>
4251 <dim>384</dim>
4252 </port>
4253 <port id="1" precision="FP32">
4254 <dim>1</dim>
4255 <dim>1</dim>
4256 <dim>384</dim>
4257 </port>
4258 </input>
4259 <output>
4260 <port id="2" precision="FP32" names="351,353,hidden_states">
4261 <dim>-1</dim>
4262 <dim>-1</dim>
4263 <dim>384</dim>
4264 </port>
4265 </output>
4266 </layer>
4267 <layer id="277" name="23" type="Const" version="opset1">
4268 <data element_type="i64" shape="" offset="47675396" size="8" />
4269 <output>
4270 <port id="0" precision="I64" names="23" />
4271 </output>
4272 </layer>
4273 <layer id="278" name="__module.bert.pooler/aten::select/Gather" type="Gather" version="opset8">
4274 <data batch_dims="0" />
4275 <input>
4276 <port id="0" precision="FP32">
4277 <dim>-1</dim>
4278 <dim>-1</dim>
4279 <dim>384</dim>
4280 </port>
4281 <port id="1" precision="I64" />
4282 <port id="2" precision="I64" />
4283 </input>
4284 <output>
4285 <port id="3" precision="FP32" names="354">
4286 <dim>-1</dim>
4287 <dim>384</dim>
4288 </port>
4289 </output>
4290 </layer>
4291 <layer id="279" name="self.bert.pooler.dense.weight" type="Const" version="opset1">
4292 <data element_type="f32" shape="384, 384" offset="76070048" size="589824" />
4293 <output>
4294 <port id="0" precision="FP32" names="self.bert.pooler.dense.weight">
4295 <dim>384</dim>
4296 <dim>384</dim>
4297 </port>
4298 </output>
4299 </layer>
4300 <layer id="280" name="__module.bert.pooler.dense/aten::linear/MatMul" type="MatMul" version="opset1">
4301 <data transpose_a="false" transpose_b="true" />
4302 <input>
4303 <port id="0" precision="FP32">
4304 <dim>-1</dim>
4305 <dim>384</dim>
4306 </port>
4307 <port id="1" precision="FP32">
4308 <dim>384</dim>
4309 <dim>384</dim>
4310 </port>
4311 </input>
4312 <output>
4313 <port id="2" precision="FP32">
4314 <dim>-1</dim>
4315 <dim>384</dim>
4316 </port>
4317 </output>
4318 </layer>
4319 <layer id="281" name="Constant_305913" type="Const" version="opset1">
4320 <data element_type="f32" shape="1, 384" offset="76659872" size="1536" />
4321 <output>
4322 <port id="0" precision="FP32">
4323 <dim>1</dim>
4324 <dim>384</dim>
4325 </port>
4326 </output>
4327 </layer>
4328 <layer id="282" name="__module.bert.pooler.dense/aten::linear/Add" type="Add" version="opset1">
4329 <data auto_broadcast="numpy" />
4330 <input>
4331 <port id="0" precision="FP32">
4332 <dim>-1</dim>
4333 <dim>384</dim>
4334 </port>
4335 <port id="1" precision="FP32">
4336 <dim>1</dim>
4337 <dim>384</dim>
4338 </port>
4339 </input>
4340 <output>
4341 <port id="2" precision="FP32" names="357">
4342 <dim>-1</dim>
4343 <dim>384</dim>
4344 </port>
4345 </output>
4346 </layer>
4347 <layer id="283" name="__module.bert.pooler.activation/aten::tanh/Tanh" type="Tanh" version="opset1">
4348 <input>
4349 <port id="0" precision="FP32">
4350 <dim>-1</dim>
4351 <dim>384</dim>
4352 </port>
4353 </input>
4354 <output>
4355 <port id="1" precision="FP32" names="358,input">
4356 <dim>-1</dim>
4357 <dim>384</dim>
4358 </port>
4359 </output>
4360 </layer>
4361 <layer id="284" name="self.classifier.weight" type="Const" version="opset1">
4362 <data element_type="f32" shape="1, 384" offset="76661408" size="1536" />
4363 <output>
4364 <port id="0" precision="FP32" names="self.classifier.weight">
4365 <dim>1</dim>
4366 <dim>384</dim>
4367 </port>
4368 </output>
4369 </layer>
4370 <layer id="285" name="__module.classifier/aten::linear/Add" type="MatMul" version="opset1">
4371 <data transpose_a="false" transpose_b="true" />
4372 <input>
4373 <port id="0" precision="FP32">
4374 <dim>-1</dim>
4375 <dim>384</dim>
4376 </port>
4377 <port id="1" precision="FP32">
4378 <dim>1</dim>
4379 <dim>384</dim>
4380 </port>
4381 </input>
4382 <output>
4383 <port id="2" precision="FP32" names="logits">
4384 <dim>-1</dim>
4385 <dim>1</dim>
4386 </port>
4387 </output>
4388 </layer>
4389 <layer id="286" name="Result_302809" type="Result" version="opset1">
4390 <input>
4391 <port id="0" precision="FP32">
4392 <dim>-1</dim>
4393 <dim>1</dim>
4394 </port>
4395 </input>
4396 </layer>
4397 </layers>
4398 <edges>
4399 <edge from-layer="0" from-port="0" to-layer="8" to-port="0" />
4400 <edge from-layer="1" from-port="0" to-layer="58" to-port="0" />
4401 <edge from-layer="2" from-port="0" to-layer="4" to-port="0" />
4402 <edge from-layer="2" from-port="0" to-layer="15" to-port="0" />
4403 <edge from-layer="3" from-port="0" to-layer="6" to-port="0" />
4404 <edge from-layer="4" from-port="1" to-layer="6" to-port="1" />
4405 <edge from-layer="5" from-port="0" to-layer="6" to-port="2" />
4406 <edge from-layer="6" from-port="3" to-layer="11" to-port="0" />
4407 <edge from-layer="7" from-port="0" to-layer="10" to-port="0" />
4408 <edge from-layer="8" from-port="1" to-layer="10" to-port="1" />
4409 <edge from-layer="9" from-port="0" to-layer="10" to-port="2" />
4410 <edge from-layer="10" from-port="3" to-layer="11" to-port="1" />
4411 <edge from-layer="11" from-port="2" to-layer="25" to-port="0" />
4412 <edge from-layer="12" from-port="0" to-layer="24" to-port="0" />
4413 <edge from-layer="13" from-port="0" to-layer="21" to-port="0" />
4414 <edge from-layer="14" from-port="0" to-layer="21" to-port="1" />
4415 <edge from-layer="15" from-port="1" to-layer="18" to-port="0" />
4416 <edge from-layer="15" from-port="1" to-layer="63" to-port="0" />
4417 <edge from-layer="15" from-port="1" to-layer="67" to-port="0" />
4418 <edge from-layer="16" from-port="0" to-layer="18" to-port="1" />
4419 <edge from-layer="17" from-port="0" to-layer="18" to-port="2" />
4420 <edge from-layer="18" from-port="3" to-layer="21" to-port="2" />
4421 <edge from-layer="19" from-port="0" to-layer="21" to-port="3" />
4422 <edge from-layer="20" from-port="0" to-layer="21" to-port="4" />
4423 <edge from-layer="21" from-port="5" to-layer="22" to-port="0" />
4424 <edge from-layer="22" from-port="1" to-layer="24" to-port="1" />
4425 <edge from-layer="23" from-port="0" to-layer="24" to-port="2" />
4426 <edge from-layer="24" from-port="3" to-layer="25" to-port="1" />
4427 <edge from-layer="25" from-port="2" to-layer="27" to-port="0" />
4428 <edge from-layer="26" from-port="0" to-layer="27" to-port="1" />
4429 <edge from-layer="27" from-port="2" to-layer="29" to-port="0" />
4430 <edge from-layer="28" from-port="0" to-layer="29" to-port="1" />
4431 <edge from-layer="29" from-port="2" to-layer="31" to-port="0" />
4432 <edge from-layer="30" from-port="0" to-layer="31" to-port="1" />
4433 <edge from-layer="31" from-port="2" to-layer="33" to-port="0" />
4434 <edge from-layer="31" from-port="2" to-layer="41" to-port="0" />
4435 <edge from-layer="31" from-port="2" to-layer="49" to-port="0" />
4436 <edge from-layer="31" from-port="2" to-layer="86" to-port="1" />
4437 <edge from-layer="32" from-port="0" to-layer="33" to-port="1" />
4438 <edge from-layer="33" from-port="2" to-layer="35" to-port="0" />
4439 <edge from-layer="34" from-port="0" to-layer="35" to-port="1" />
4440 <edge from-layer="35" from-port="2" to-layer="37" to-port="0" />
4441 <edge from-layer="36" from-port="0" to-layer="37" to-port="1" />
4442 <edge from-layer="37" from-port="2" to-layer="39" to-port="0" />
4443 <edge from-layer="38" from-port="0" to-layer="39" to-port="1" />
4444 <edge from-layer="39" from-port="2" to-layer="77" to-port="0" />
4445 <edge from-layer="40" from-port="0" to-layer="41" to-port="1" />
4446 <edge from-layer="41" from-port="2" to-layer="43" to-port="0" />
4447 <edge from-layer="42" from-port="0" to-layer="43" to-port="1" />
4448 <edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
4449 <edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
4450 <edge from-layer="45" from-port="2" to-layer="47" to-port="0" />
4451 <edge from-layer="46" from-port="0" to-layer="47" to-port="1" />
4452 <edge from-layer="47" from-port="2" to-layer="77" to-port="1" />
4453 <edge from-layer="48" from-port="0" to-layer="49" to-port="1" />
4454 <edge from-layer="49" from-port="2" to-layer="51" to-port="0" />
4455 <edge from-layer="50" from-port="0" to-layer="51" to-port="1" />
4456 <edge from-layer="51" from-port="2" to-layer="53" to-port="0" />
4457 <edge from-layer="52" from-port="0" to-layer="53" to-port="1" />
4458 <edge from-layer="53" from-port="2" to-layer="55" to-port="0" />
4459 <edge from-layer="54" from-port="0" to-layer="55" to-port="1" />
4460 <edge from-layer="55" from-port="2" to-layer="77" to-port="2" />
4461 <edge from-layer="56" from-port="0" to-layer="73" to-port="0" />
4462 <edge from-layer="57" from-port="0" to-layer="58" to-port="1" />
4463 <edge from-layer="57" from-port="0" to-layer="278" to-port="2" />
4464 <edge from-layer="58" from-port="2" to-layer="60" to-port="0" />
4465 <edge from-layer="59" from-port="0" to-layer="60" to-port="1" />
4466 <edge from-layer="60" from-port="2" to-layer="69" to-port="0" />
4467 <edge from-layer="61" from-port="0" to-layer="63" to-port="1" />
4468 <edge from-layer="62" from-port="0" to-layer="63" to-port="2" />
4469 <edge from-layer="63" from-port="3" to-layer="68" to-port="0" />
4470 <edge from-layer="64" from-port="0" to-layer="68" to-port="1" />
4471 <edge from-layer="65" from-port="0" to-layer="67" to-port="1" />
4472 <edge from-layer="66" from-port="0" to-layer="67" to-port="2" />
4473 <edge from-layer="67" from-port="3" to-layer="68" to-port="2" />
4474 <edge from-layer="68" from-port="3" to-layer="69" to-port="1" />
4475 <edge from-layer="69" from-port="2" to-layer="70" to-port="0" />
4476 <edge from-layer="70" from-port="1" to-layer="72" to-port="0" />
4477 <edge from-layer="71" from-port="0" to-layer="72" to-port="1" />
4478 <edge from-layer="72" from-port="2" to-layer="73" to-port="1" />
4479 <edge from-layer="73" from-port="2" to-layer="76" to-port="2" />
4480 <edge from-layer="73" from-port="2" to-layer="74" to-port="0" />
4481 <edge from-layer="74" from-port="1" to-layer="76" to-port="0" />
4482 <edge from-layer="75" from-port="0" to-layer="76" to-port="1" />
4483 <edge from-layer="76" from-port="3" to-layer="77" to-port="3" />
4484 <edge from-layer="76" from-port="3" to-layer="245" to-port="3" />
4485 <edge from-layer="76" from-port="3" to-layer="133" to-port="3" />
4486 <edge from-layer="76" from-port="3" to-layer="189" to-port="3" />
4487 <edge from-layer="77" from-port="4" to-layer="79" to-port="0" />
4488 <edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
4489 <edge from-layer="79" from-port="2" to-layer="81" to-port="0" />
4490 <edge from-layer="80" from-port="0" to-layer="81" to-port="1" />
4491 <edge from-layer="81" from-port="2" to-layer="83" to-port="0" />
4492 <edge from-layer="82" from-port="0" to-layer="83" to-port="1" />
4493 <edge from-layer="83" from-port="2" to-layer="85" to-port="0" />
4494 <edge from-layer="84" from-port="0" to-layer="85" to-port="1" />
4495 <edge from-layer="85" from-port="2" to-layer="86" to-port="0" />
4496 <edge from-layer="86" from-port="2" to-layer="88" to-port="0" />
4497 <edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
4498 <edge from-layer="88" from-port="2" to-layer="90" to-port="0" />
4499 <edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
4500 <edge from-layer="90" from-port="2" to-layer="92" to-port="0" />
4501 <edge from-layer="91" from-port="0" to-layer="92" to-port="1" />
4502 <edge from-layer="92" from-port="2" to-layer="94" to-port="0" />
4503 <edge from-layer="92" from-port="2" to-layer="102" to-port="1" />
4504 <edge from-layer="93" from-port="0" to-layer="94" to-port="1" />
4505 <edge from-layer="94" from-port="2" to-layer="96" to-port="0" />
4506 <edge from-layer="95" from-port="0" to-layer="96" to-port="1" />
4507 <edge from-layer="96" from-port="2" to-layer="97" to-port="0" />
4508 <edge from-layer="97" from-port="1" to-layer="99" to-port="0" />
4509 <edge from-layer="98" from-port="0" to-layer="99" to-port="1" />
4510 <edge from-layer="99" from-port="2" to-layer="101" to-port="0" />
4511 <edge from-layer="100" from-port="0" to-layer="101" to-port="1" />
4512 <edge from-layer="101" from-port="2" to-layer="102" to-port="0" />
4513 <edge from-layer="102" from-port="2" to-layer="104" to-port="0" />
4514 <edge from-layer="103" from-port="0" to-layer="104" to-port="1" />
4515 <edge from-layer="104" from-port="2" to-layer="106" to-port="0" />
4516 <edge from-layer="105" from-port="0" to-layer="106" to-port="1" />
4517 <edge from-layer="106" from-port="2" to-layer="108" to-port="0" />
4518 <edge from-layer="107" from-port="0" to-layer="108" to-port="1" />
4519 <edge from-layer="108" from-port="2" to-layer="142" to-port="1" />
4520 <edge from-layer="108" from-port="2" to-layer="126" to-port="0" />
4521 <edge from-layer="108" from-port="2" to-layer="118" to-port="0" />
4522 <edge from-layer="108" from-port="2" to-layer="110" to-port="0" />
4523 <edge from-layer="109" from-port="0" to-layer="110" to-port="1" />
4524 <edge from-layer="110" from-port="2" to-layer="112" to-port="0" />
4525 <edge from-layer="111" from-port="0" to-layer="112" to-port="1" />
4526 <edge from-layer="112" from-port="2" to-layer="114" to-port="0" />
4527 <edge from-layer="113" from-port="0" to-layer="114" to-port="1" />
4528 <edge from-layer="114" from-port="2" to-layer="116" to-port="0" />
4529 <edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
4530 <edge from-layer="116" from-port="2" to-layer="133" to-port="0" />
4531 <edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
4532 <edge from-layer="118" from-port="2" to-layer="120" to-port="0" />
4533 <edge from-layer="119" from-port="0" to-layer="120" to-port="1" />
4534 <edge from-layer="120" from-port="2" to-layer="122" to-port="0" />
4535 <edge from-layer="121" from-port="0" to-layer="122" to-port="1" />
4536 <edge from-layer="122" from-port="2" to-layer="124" to-port="0" />
4537 <edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
4538 <edge from-layer="124" from-port="2" to-layer="133" to-port="1" />
4539 <edge from-layer="125" from-port="0" to-layer="126" to-port="1" />
4540 <edge from-layer="126" from-port="2" to-layer="128" to-port="0" />
4541 <edge from-layer="127" from-port="0" to-layer="128" to-port="1" />
4542 <edge from-layer="128" from-port="2" to-layer="130" to-port="0" />
4543 <edge from-layer="129" from-port="0" to-layer="130" to-port="1" />
4544 <edge from-layer="130" from-port="2" to-layer="132" to-port="0" />
4545 <edge from-layer="131" from-port="0" to-layer="132" to-port="1" />
4546 <edge from-layer="132" from-port="2" to-layer="133" to-port="2" />
4547 <edge from-layer="133" from-port="4" to-layer="135" to-port="0" />
4548 <edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
4549 <edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
4550 <edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
4551 <edge from-layer="137" from-port="2" to-layer="139" to-port="0" />
4552 <edge from-layer="138" from-port="0" to-layer="139" to-port="1" />
4553 <edge from-layer="139" from-port="2" to-layer="141" to-port="0" />
4554 <edge from-layer="140" from-port="0" to-layer="141" to-port="1" />
4555 <edge from-layer="141" from-port="2" to-layer="142" to-port="0" />
4556 <edge from-layer="142" from-port="2" to-layer="144" to-port="0" />
4557 <edge from-layer="143" from-port="0" to-layer="144" to-port="1" />
4558 <edge from-layer="144" from-port="2" to-layer="146" to-port="0" />
4559 <edge from-layer="145" from-port="0" to-layer="146" to-port="1" />
4560 <edge from-layer="146" from-port="2" to-layer="148" to-port="0" />
4561 <edge from-layer="147" from-port="0" to-layer="148" to-port="1" />
4562 <edge from-layer="148" from-port="2" to-layer="150" to-port="0" />
4563 <edge from-layer="148" from-port="2" to-layer="158" to-port="1" />
4564 <edge from-layer="149" from-port="0" to-layer="150" to-port="1" />
4565 <edge from-layer="150" from-port="2" to-layer="152" to-port="0" />
4566 <edge from-layer="151" from-port="0" to-layer="152" to-port="1" />
4567 <edge from-layer="152" from-port="2" to-layer="153" to-port="0" />
4568 <edge from-layer="153" from-port="1" to-layer="155" to-port="0" />
4569 <edge from-layer="154" from-port="0" to-layer="155" to-port="1" />
4570 <edge from-layer="155" from-port="2" to-layer="157" to-port="0" />
4571 <edge from-layer="156" from-port="0" to-layer="157" to-port="1" />
4572 <edge from-layer="157" from-port="2" to-layer="158" to-port="0" />
4573 <edge from-layer="158" from-port="2" to-layer="160" to-port="0" />
4574 <edge from-layer="159" from-port="0" to-layer="160" to-port="1" />
4575 <edge from-layer="160" from-port="2" to-layer="162" to-port="0" />
4576 <edge from-layer="161" from-port="0" to-layer="162" to-port="1" />
4577 <edge from-layer="162" from-port="2" to-layer="164" to-port="0" />
4578 <edge from-layer="163" from-port="0" to-layer="164" to-port="1" />
4579 <edge from-layer="164" from-port="2" to-layer="166" to-port="0" />
4580 <edge from-layer="164" from-port="2" to-layer="174" to-port="0" />
4581 <edge from-layer="164" from-port="2" to-layer="182" to-port="0" />
4582 <edge from-layer="164" from-port="2" to-layer="198" to-port="1" />
4583 <edge from-layer="165" from-port="0" to-layer="166" to-port="1" />
4584 <edge from-layer="166" from-port="2" to-layer="168" to-port="0" />
4585 <edge from-layer="167" from-port="0" to-layer="168" to-port="1" />
4586 <edge from-layer="168" from-port="2" to-layer="170" to-port="0" />
4587 <edge from-layer="169" from-port="0" to-layer="170" to-port="1" />
4588 <edge from-layer="170" from-port="2" to-layer="172" to-port="0" />
4589 <edge from-layer="171" from-port="0" to-layer="172" to-port="1" />
4590 <edge from-layer="172" from-port="2" to-layer="189" to-port="0" />
4591 <edge from-layer="173" from-port="0" to-layer="174" to-port="1" />
4592 <edge from-layer="174" from-port="2" to-layer="176" to-port="0" />
4593 <edge from-layer="175" from-port="0" to-layer="176" to-port="1" />
4594 <edge from-layer="176" from-port="2" to-layer="178" to-port="0" />
4595 <edge from-layer="177" from-port="0" to-layer="178" to-port="1" />
4596 <edge from-layer="178" from-port="2" to-layer="180" to-port="0" />
4597 <edge from-layer="179" from-port="0" to-layer="180" to-port="1" />
4598 <edge from-layer="180" from-port="2" to-layer="189" to-port="1" />
4599 <edge from-layer="181" from-port="0" to-layer="182" to-port="1" />
4600 <edge from-layer="182" from-port="2" to-layer="184" to-port="0" />
4601 <edge from-layer="183" from-port="0" to-layer="184" to-port="1" />
4602 <edge from-layer="184" from-port="2" to-layer="186" to-port="0" />
4603 <edge from-layer="185" from-port="0" to-layer="186" to-port="1" />
4604 <edge from-layer="186" from-port="2" to-layer="188" to-port="0" />
4605 <edge from-layer="187" from-port="0" to-layer="188" to-port="1" />
4606 <edge from-layer="188" from-port="2" to-layer="189" to-port="2" />
4607 <edge from-layer="189" from-port="4" to-layer="191" to-port="0" />
4608 <edge from-layer="190" from-port="0" to-layer="191" to-port="1" />
4609 <edge from-layer="191" from-port="2" to-layer="193" to-port="0" />
4610 <edge from-layer="192" from-port="0" to-layer="193" to-port="1" />
4611 <edge from-layer="193" from-port="2" to-layer="195" to-port="0" />
4612 <edge from-layer="194" from-port="0" to-layer="195" to-port="1" />
4613 <edge from-layer="195" from-port="2" to-layer="197" to-port="0" />
4614 <edge from-layer="196" from-port="0" to-layer="197" to-port="1" />
4615 <edge from-layer="197" from-port="2" to-layer="198" to-port="0" />
4616 <edge from-layer="198" from-port="2" to-layer="200" to-port="0" />
4617 <edge from-layer="199" from-port="0" to-layer="200" to-port="1" />
4618 <edge from-layer="200" from-port="2" to-layer="202" to-port="0" />
4619 <edge from-layer="201" from-port="0" to-layer="202" to-port="1" />
4620 <edge from-layer="202" from-port="2" to-layer="204" to-port="0" />
4621 <edge from-layer="203" from-port="0" to-layer="204" to-port="1" />
4622 <edge from-layer="204" from-port="2" to-layer="206" to-port="0" />
4623 <edge from-layer="204" from-port="2" to-layer="214" to-port="1" />
4624 <edge from-layer="205" from-port="0" to-layer="206" to-port="1" />
4625 <edge from-layer="206" from-port="2" to-layer="208" to-port="0" />
4626 <edge from-layer="207" from-port="0" to-layer="208" to-port="1" />
4627 <edge from-layer="208" from-port="2" to-layer="209" to-port="0" />
4628 <edge from-layer="209" from-port="1" to-layer="211" to-port="0" />
4629 <edge from-layer="210" from-port="0" to-layer="211" to-port="1" />
4630 <edge from-layer="211" from-port="2" to-layer="213" to-port="0" />
4631 <edge from-layer="212" from-port="0" to-layer="213" to-port="1" />
4632 <edge from-layer="213" from-port="2" to-layer="214" to-port="0" />
4633 <edge from-layer="214" from-port="2" to-layer="216" to-port="0" />
4634 <edge from-layer="215" from-port="0" to-layer="216" to-port="1" />
4635 <edge from-layer="216" from-port="2" to-layer="218" to-port="0" />
4636 <edge from-layer="217" from-port="0" to-layer="218" to-port="1" />
4637 <edge from-layer="218" from-port="2" to-layer="220" to-port="0" />
4638 <edge from-layer="219" from-port="0" to-layer="220" to-port="1" />
4639 <edge from-layer="220" from-port="2" to-layer="222" to-port="0" />
4640 <edge from-layer="220" from-port="2" to-layer="230" to-port="0" />
4641 <edge from-layer="220" from-port="2" to-layer="238" to-port="0" />
4642 <edge from-layer="220" from-port="2" to-layer="254" to-port="1" />
4643 <edge from-layer="221" from-port="0" to-layer="222" to-port="1" />
4644 <edge from-layer="222" from-port="2" to-layer="224" to-port="0" />
4645 <edge from-layer="223" from-port="0" to-layer="224" to-port="1" />
4646 <edge from-layer="224" from-port="2" to-layer="226" to-port="0" />
4647 <edge from-layer="225" from-port="0" to-layer="226" to-port="1" />
4648 <edge from-layer="226" from-port="2" to-layer="228" to-port="0" />
4649 <edge from-layer="227" from-port="0" to-layer="228" to-port="1" />
4650 <edge from-layer="228" from-port="2" to-layer="245" to-port="0" />
4651 <edge from-layer="229" from-port="0" to-layer="230" to-port="1" />
4652 <edge from-layer="230" from-port="2" to-layer="232" to-port="0" />
4653 <edge from-layer="231" from-port="0" to-layer="232" to-port="1" />
4654 <edge from-layer="232" from-port="2" to-layer="234" to-port="0" />
4655 <edge from-layer="233" from-port="0" to-layer="234" to-port="1" />
4656 <edge from-layer="234" from-port="2" to-layer="236" to-port="0" />
4657 <edge from-layer="235" from-port="0" to-layer="236" to-port="1" />
4658 <edge from-layer="236" from-port="2" to-layer="245" to-port="1" />
4659 <edge from-layer="237" from-port="0" to-layer="238" to-port="1" />
4660 <edge from-layer="238" from-port="2" to-layer="240" to-port="0" />
4661 <edge from-layer="239" from-port="0" to-layer="240" to-port="1" />
4662 <edge from-layer="240" from-port="2" to-layer="242" to-port="0" />
4663 <edge from-layer="241" from-port="0" to-layer="242" to-port="1" />
4664 <edge from-layer="242" from-port="2" to-layer="244" to-port="0" />
4665 <edge from-layer="243" from-port="0" to-layer="244" to-port="1" />
4666 <edge from-layer="244" from-port="2" to-layer="245" to-port="2" />
4667 <edge from-layer="245" from-port="4" to-layer="247" to-port="0" />
4668 <edge from-layer="246" from-port="0" to-layer="247" to-port="1" />
4669 <edge from-layer="247" from-port="2" to-layer="249" to-port="0" />
4670 <edge from-layer="248" from-port="0" to-layer="249" to-port="1" />
4671 <edge from-layer="249" from-port="2" to-layer="251" to-port="0" />
4672 <edge from-layer="250" from-port="0" to-layer="251" to-port="1" />
4673 <edge from-layer="251" from-port="2" to-layer="253" to-port="0" />
4674 <edge from-layer="252" from-port="0" to-layer="253" to-port="1" />
4675 <edge from-layer="253" from-port="2" to-layer="254" to-port="0" />
4676 <edge from-layer="254" from-port="2" to-layer="256" to-port="0" />
4677 <edge from-layer="255" from-port="0" to-layer="256" to-port="1" />
4678 <edge from-layer="256" from-port="2" to-layer="258" to-port="0" />
4679 <edge from-layer="257" from-port="0" to-layer="258" to-port="1" />
4680 <edge from-layer="258" from-port="2" to-layer="260" to-port="0" />
4681 <edge from-layer="259" from-port="0" to-layer="260" to-port="1" />
4682 <edge from-layer="260" from-port="2" to-layer="262" to-port="0" />
4683 <edge from-layer="260" from-port="2" to-layer="270" to-port="1" />
4684 <edge from-layer="261" from-port="0" to-layer="262" to-port="1" />
4685 <edge from-layer="262" from-port="2" to-layer="264" to-port="0" />
4686 <edge from-layer="263" from-port="0" to-layer="264" to-port="1" />
4687 <edge from-layer="264" from-port="2" to-layer="265" to-port="0" />
4688 <edge from-layer="265" from-port="1" to-layer="267" to-port="0" />
4689 <edge from-layer="266" from-port="0" to-layer="267" to-port="1" />
4690 <edge from-layer="267" from-port="2" to-layer="269" to-port="0" />
4691 <edge from-layer="268" from-port="0" to-layer="269" to-port="1" />
4692 <edge from-layer="269" from-port="2" to-layer="270" to-port="0" />
4693 <edge from-layer="270" from-port="2" to-layer="272" to-port="0" />
4694 <edge from-layer="271" from-port="0" to-layer="272" to-port="1" />
4695 <edge from-layer="272" from-port="2" to-layer="274" to-port="0" />
4696 <edge from-layer="273" from-port="0" to-layer="274" to-port="1" />
4697 <edge from-layer="274" from-port="2" to-layer="276" to-port="0" />
4698 <edge from-layer="275" from-port="0" to-layer="276" to-port="1" />
4699 <edge from-layer="276" from-port="2" to-layer="278" to-port="0" />
4700 <edge from-layer="277" from-port="0" to-layer="278" to-port="1" />
4701 <edge from-layer="278" from-port="3" to-layer="280" to-port="0" />
4702 <edge from-layer="279" from-port="0" to-layer="280" to-port="1" />
4703 <edge from-layer="280" from-port="2" to-layer="282" to-port="0" />
4704 <edge from-layer="281" from-port="0" to-layer="282" to-port="1" />
4705 <edge from-layer="282" from-port="2" to-layer="283" to-port="0" />
4706 <edge from-layer="283" from-port="1" to-layer="285" to-port="0" />
4707 <edge from-layer="284" from-port="0" to-layer="285" to-port="1" />
4708 <edge from-layer="285" from-port="2" to-layer="286" to-port="0" />
4709 </edges>
4710 <rt_info>
4711 <Runtime_version value="2024.4.1-16618-643f23d1318-releases/2024/4" />
4712 <conversion_parameters>
4713 <framework value="pytorch" />
4714 <is_python_object value="True" />
4715 </conversion_parameters>
4716 <optimum>
4717 <optimum_intel_version value="1.20.1" />
4718 <optimum_version value="1.24.0" />
4719 <pytorch_version value="2.6.0+cu124" />
4720 <transformers_version value="4.52.0.dev0" />
4721 </optimum>
4722 </rt_info>
4723 </net>
4724