config.json
1.2 KB · 51 lines · json Raw
1 {
2 "_name_or_path": "openai/whisper-large-v3",
3 "activation_dropout": 0.0,
4 "activation_function": "gelu",
5 "apply_spec_augment": false,
6 "architectures": [
7 "WhisperForConditionalGeneration"
8 ],
9 "attention_dropout": 0.0,
10 "begin_suppress_tokens": [
11 220,
12 50257
13 ],
14 "bos_token_id": 50257,
15 "classifier_proj_size": 256,
16 "d_model": 1280,
17 "decoder_attention_heads": 20,
18 "decoder_ffn_dim": 5120,
19 "decoder_layerdrop": 0.0,
20 "decoder_layers": 32,
21 "decoder_start_token_id": 50258,
22 "dropout": 0.0,
23 "encoder_attention_heads": 20,
24 "encoder_ffn_dim": 5120,
25 "encoder_layerdrop": 0.0,
26 "encoder_layers": 32,
27 "eos_token_id": 50257,
28 "init_std": 0.02,
29 "is_encoder_decoder": true,
30 "mask_feature_length": 10,
31 "mask_feature_min_masks": 0,
32 "mask_feature_prob": 0.0,
33 "mask_time_length": 10,
34 "mask_time_min_masks": 2,
35 "mask_time_prob": 0.05,
36 "max_length": 448,
37 "max_source_positions": 1500,
38 "max_target_positions": 448,
39 "median_filter_width": 7,
40 "model_type": "whisper",
41 "num_hidden_layers": 32,
42 "num_mel_bins": 128,
43 "pad_token_id": 50256,
44 "scale_embedding": false,
45 "torch_dtype": "float16",
46 "transformers_version": "4.36.0.dev0",
47 "use_cache": true,
48 "use_weighted_layer_sum": false,
49 "vocab_size": 51866
50 }
51