config.json
2.2 KB · 102 lines · json Raw
1 {
2 "architectures": [
3 "OmniVoice"
4 ],
5 "audio_codebook_weights": [
6 8,
7 8,
8 6,
9 6,
10 4,
11 4,
12 2,
13 2
14 ],
15 "audio_mask_id": 1024,
16 "audio_vocab_size": 1025,
17 "bos_token_id": null,
18 "dtype": "float32",
19 "eos_token_id": 151645,
20 "llm_config": {
21 "_name_or_path": "",
22 "architectures": [
23 "Qwen3ForCausalLM"
24 ],
25 "attention_bias": false,
26 "attention_dropout": 0.0,
27 "bos_token_id": 151643,
28 "chunk_size_feed_forward": 0,
29 "dtype": "float32",
30 "eos_token_id": 151645,
31 "head_dim": 128,
32 "hidden_act": "silu",
33 "hidden_size": 1024,
34 "id2label": {
35 "0": "LABEL_0",
36 "1": "LABEL_1"
37 },
38 "initializer_range": 0.02,
39 "intermediate_size": 3072,
40 "is_encoder_decoder": false,
41 "label2id": {
42 "LABEL_0": 0,
43 "LABEL_1": 1
44 },
45 "layer_types": [
46 "full_attention",
47 "full_attention",
48 "full_attention",
49 "full_attention",
50 "full_attention",
51 "full_attention",
52 "full_attention",
53 "full_attention",
54 "full_attention",
55 "full_attention",
56 "full_attention",
57 "full_attention",
58 "full_attention",
59 "full_attention",
60 "full_attention",
61 "full_attention",
62 "full_attention",
63 "full_attention",
64 "full_attention",
65 "full_attention",
66 "full_attention",
67 "full_attention",
68 "full_attention",
69 "full_attention",
70 "full_attention",
71 "full_attention",
72 "full_attention",
73 "full_attention"
74 ],
75 "max_position_embeddings": 40960,
76 "max_window_layers": 28,
77 "model_type": "qwen3",
78 "num_attention_heads": 16,
79 "num_hidden_layers": 28,
80 "num_key_value_heads": 8,
81 "output_attentions": false,
82 "output_hidden_states": false,
83 "pad_token_id": null,
84 "problem_type": null,
85 "return_dict": true,
86 "rms_norm_eps": 1e-06,
87 "rope_parameters": {
88 "rope_theta": 1000000,
89 "rope_type": "default"
90 },
91 "sliding_window": null,
92 "tie_word_embeddings": true,
93 "use_cache": true,
94 "use_sliding_window": false,
95 "vocab_size": 151676
96 },
97 "model_type": "omnivoice",
98 "num_audio_codebook": 8,
99 "pad_token_id": 151643,
100 "transformers_version": "5.3.0"
101 }
102