todeschini commited on
Commit
c361654
1 Parent(s): f2b9d95

Model save

Browse files
README.md CHANGED
@@ -1,9 +1,11 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
- base_model: pierreguillou/whisper-medium-portuguese
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: whisper-finetuned
9
  results: []
@@ -14,7 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # whisper-finetuned
16
 
17
- This model is a fine-tuned version of [pierreguillou/whisper-medium-portuguese](https://huggingface.co/pierreguillou/whisper-medium-portuguese) on an unknown dataset.
 
 
 
18
 
19
  ## Model description
20
 
@@ -33,28 +38,42 @@ More information needed
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
- - learning_rate: 1e-05
37
- - train_batch_size: 16
38
- - eval_batch_size: 8
39
  - seed: 42
40
  - distributed_type: multi-GPU
41
  - num_devices: 4
42
- - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 256
44
- - total_eval_batch_size: 32
45
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
- - lr_scheduler_warmup_ratio: 0.15
48
- - num_epochs: 3
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
 
54
 
55
  ### Framework versions
56
 
57
  - Transformers 4.46.0.dev0
58
- - Pytorch 2.2.0+cu121
59
  - Datasets 3.0.2.dev0
60
  - Tokenizers 0.20.0
 
1
  ---
2
  library_name: transformers
3
+ license: mit
4
+ base_model: openai/whisper-large-v3-turbo
5
  tags:
6
  - generated_from_trainer
7
+ metrics:
8
+ - wer
9
  model-index:
10
  - name: whisper-finetuned
11
  results: []
 
16
 
17
  # whisper-finetuned
18
 
19
+ This model is a fine-tuned version of [openai/whisper-large-v3-turbo](https://huggingface.co/openai/whisper-large-v3-turbo) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.3011
22
+ - Wer: 30.6945
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 9e-06
42
+ - train_batch_size: 4
43
+ - eval_batch_size: 16
44
  - seed: 42
45
  - distributed_type: multi-GPU
46
  - num_devices: 4
47
+ - gradient_accumulation_steps: 8
48
+ - total_train_batch_size: 128
49
+ - total_eval_batch_size: 64
50
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
  - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 500
53
+ - training_steps: 6000
54
  - mixed_precision_training: Native AMP
55
 
56
  ### Training results
57
 
58
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
59
+ |:-------------:|:------:|:----:|:---------------:|:-------:|
60
+ | 0.3654 | 0.3680 | 500 | 0.3516 | 35.6868 |
61
+ | 0.3355 | 0.7360 | 1000 | 0.3341 | 35.0002 |
62
+ | 0.2826 | 1.1041 | 1500 | 0.3248 | 34.6635 |
63
+ | 0.2763 | 1.4721 | 2000 | 0.3171 | 33.6200 |
64
+ | 0.2715 | 1.8401 | 2500 | 0.3101 | 33.1267 |
65
+ | 0.2203 | 2.2081 | 3000 | 0.3071 | 31.3256 |
66
+ | 0.2202 | 2.5761 | 3500 | 0.3019 | 30.5031 |
67
+ | 0.2169 | 2.9442 | 4000 | 0.2975 | 30.7246 |
68
+ | 0.1765 | 3.3122 | 4500 | 0.3002 | 31.2968 |
69
+ | 0.1768 | 3.6802 | 5000 | 0.2985 | 30.5046 |
70
+ | 0.1594 | 4.0482 | 5500 | 0.3003 | 30.5781 |
71
+ | 0.1603 | 4.4162 | 6000 | 0.3011 | 30.6945 |
72
 
73
 
74
  ### Framework versions
75
 
76
  - Transformers 4.46.0.dev0
77
+ - Pytorch 2.1.0+cu118
78
  - Datasets 3.0.2.dev0
79
  - Tokenizers 0.20.0
events.out.tfevents.1728417412.9f9e9923fdea.9716.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c56b4050ef9fce8032cb00073b6246ebca28d8287da16f5d79d5de0df6430a35
3
- size 60399
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c3d420df69bf7c4fb8989274ab42298dec885bbf07efd62a9a02d48c673d98
3
+ size 60753
generation_config.json CHANGED
@@ -1,5 +1,30 @@
1
  {
2
- "_from_model_config": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "begin_suppress_tokens": [
4
  220,
5
  50257
@@ -7,8 +32,129 @@
7
  "bos_token_id": 50257,
8
  "decoder_start_token_id": 50258,
9
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "max_length": 448,
 
11
  "pad_token_id": 50257,
12
- "transformers_version": "4.46.0.dev0",
13
- "use_cache": false
 
 
 
 
 
 
14
  }
 
1
  {
2
+ "alignment_heads": [
3
+ [
4
+ 2,
5
+ 4
6
+ ],
7
+ [
8
+ 2,
9
+ 11
10
+ ],
11
+ [
12
+ 3,
13
+ 3
14
+ ],
15
+ [
16
+ 3,
17
+ 6
18
+ ],
19
+ [
20
+ 3,
21
+ 11
22
+ ],
23
+ [
24
+ 3,
25
+ 14
26
+ ]
27
+ ],
28
  "begin_suppress_tokens": [
29
  220,
30
  50257
 
32
  "bos_token_id": 50257,
33
  "decoder_start_token_id": 50258,
34
  "eos_token_id": 50257,
35
+ "forced_decoder_ids": [
36
+ [
37
+ 1,
38
+ null
39
+ ],
40
+ [
41
+ 2,
42
+ 50360
43
+ ]
44
+ ],
45
+ "is_multilingual": true,
46
+ "lang_to_id": {
47
+ "<|af|>": 50327,
48
+ "<|am|>": 50334,
49
+ "<|ar|>": 50272,
50
+ "<|as|>": 50350,
51
+ "<|az|>": 50304,
52
+ "<|ba|>": 50355,
53
+ "<|be|>": 50330,
54
+ "<|bg|>": 50292,
55
+ "<|bn|>": 50302,
56
+ "<|bo|>": 50347,
57
+ "<|br|>": 50309,
58
+ "<|bs|>": 50315,
59
+ "<|ca|>": 50270,
60
+ "<|cs|>": 50283,
61
+ "<|cy|>": 50297,
62
+ "<|da|>": 50285,
63
+ "<|de|>": 50261,
64
+ "<|el|>": 50281,
65
+ "<|en|>": 50259,
66
+ "<|es|>": 50262,
67
+ "<|et|>": 50307,
68
+ "<|eu|>": 50310,
69
+ "<|fa|>": 50300,
70
+ "<|fi|>": 50277,
71
+ "<|fo|>": 50338,
72
+ "<|fr|>": 50265,
73
+ "<|gl|>": 50319,
74
+ "<|gu|>": 50333,
75
+ "<|haw|>": 50352,
76
+ "<|ha|>": 50354,
77
+ "<|he|>": 50279,
78
+ "<|hi|>": 50276,
79
+ "<|hr|>": 50291,
80
+ "<|ht|>": 50339,
81
+ "<|hu|>": 50286,
82
+ "<|hy|>": 50312,
83
+ "<|id|>": 50275,
84
+ "<|is|>": 50311,
85
+ "<|it|>": 50274,
86
+ "<|ja|>": 50266,
87
+ "<|jw|>": 50356,
88
+ "<|ka|>": 50329,
89
+ "<|kk|>": 50316,
90
+ "<|km|>": 50323,
91
+ "<|kn|>": 50306,
92
+ "<|ko|>": 50264,
93
+ "<|la|>": 50294,
94
+ "<|lb|>": 50345,
95
+ "<|ln|>": 50353,
96
+ "<|lo|>": 50336,
97
+ "<|lt|>": 50293,
98
+ "<|lv|>": 50301,
99
+ "<|mg|>": 50349,
100
+ "<|mi|>": 50295,
101
+ "<|mk|>": 50308,
102
+ "<|ml|>": 50296,
103
+ "<|mn|>": 50314,
104
+ "<|mr|>": 50320,
105
+ "<|ms|>": 50282,
106
+ "<|mt|>": 50343,
107
+ "<|my|>": 50346,
108
+ "<|ne|>": 50313,
109
+ "<|nl|>": 50271,
110
+ "<|nn|>": 50342,
111
+ "<|no|>": 50288,
112
+ "<|oc|>": 50328,
113
+ "<|pa|>": 50321,
114
+ "<|pl|>": 50269,
115
+ "<|ps|>": 50340,
116
+ "<|pt|>": 50267,
117
+ "<|ro|>": 50284,
118
+ "<|ru|>": 50263,
119
+ "<|sa|>": 50344,
120
+ "<|sd|>": 50332,
121
+ "<|si|>": 50322,
122
+ "<|sk|>": 50298,
123
+ "<|sl|>": 50305,
124
+ "<|sn|>": 50324,
125
+ "<|so|>": 50326,
126
+ "<|sq|>": 50317,
127
+ "<|sr|>": 50303,
128
+ "<|su|>": 50357,
129
+ "<|sv|>": 50273,
130
+ "<|sw|>": 50318,
131
+ "<|ta|>": 50287,
132
+ "<|te|>": 50299,
133
+ "<|tg|>": 50331,
134
+ "<|th|>": 50289,
135
+ "<|tk|>": 50341,
136
+ "<|tl|>": 50348,
137
+ "<|tr|>": 50268,
138
+ "<|tt|>": 50351,
139
+ "<|uk|>": 50280,
140
+ "<|ur|>": 50290,
141
+ "<|uz|>": 50337,
142
+ "<|vi|>": 50278,
143
+ "<|yi|>": 50335,
144
+ "<|yo|>": 50325,
145
+ "<|yue|>": 50358,
146
+ "<|zh|>": 50260
147
+ },
148
+ "max_initial_timestamp_index": 50,
149
  "max_length": 448,
150
+ "no_timestamps_token_id": 50364,
151
  "pad_token_id": 50257,
152
+ "prev_sot_token_id": 50362,
153
+ "return_timestamps": false,
154
+ "suppress_tokens": [],
155
+ "task_to_id": {
156
+ "transcribe": 50360,
157
+ "translate": 50359
158
+ },
159
+ "transformers_version": "4.46.0.dev0"
160
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b11fa079f7d5ece9014bf0544d68a1c2c90f4b5eedc05d1711868f85ab017869
3
  size 3235581408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4e810beee26a08075203b9aa1a9d0efacd4dcbcf0e77687d692e84222930fc1
3
  size 3235581408