Marco-Cheung commited on
Commit
ac8befd
1 Parent(s): 165e96d

End of training

Browse files
Files changed (2) hide show
  1. README.md +11 -11
  2. generation_config.json +2 -2
README.md CHANGED
@@ -24,7 +24,7 @@ model-index:
24
  metrics:
25
  - name: Wer
26
  type: wer
27
- value: 114.19385194479297
28
  ---
29
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -34,9 +34,9 @@ should probably proofread and complete it, then remove this comment. -->
34
 
35
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 13 dataset.
36
  It achieves the following results on the evaluation set:
37
- - Loss: 0.2812
38
- - Wer Ortho: 116.9664
39
- - Wer: 114.1939
40
 
41
  ## Model description
42
 
@@ -57,19 +57,19 @@ More information needed
57
  The following hyperparameters were used during training:
58
  - learning_rate: 1e-05
59
  - train_batch_size: 16
60
- - eval_batch_size: 8
61
  - seed: 42
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: constant_with_warmup
64
- - lr_scheduler_warmup_steps: 50
65
- - training_steps: 1000
66
 
67
  ### Training results
68
 
69
- | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
70
- |:-------------:|:-----:|:----:|:---------------:|:---------:|:--------:|
71
- | 0.3111 | 0.57 | 500 | 0.3072 | 92.3269 | 105.6462 |
72
- | 0.1729 | 1.14 | 1000 | 0.2812 | 116.9664 | 114.1939 |
73
 
74
 
75
  ### Framework versions
 
24
  metrics:
25
  - name: Wer
26
  type: wer
27
+ value: 57.700752823086574
28
  ---
29
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
34
 
35
  This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 13 dataset.
36
  It achieves the following results on the evaluation set:
37
+ - Loss: 0.2487
38
+ - Wer Ortho: 57.8423
39
+ - Wer: 57.7008
40
 
41
  ## Model description
42
 
 
57
  The following hyperparameters were used during training:
58
  - learning_rate: 1e-05
59
  - train_batch_size: 16
60
+ - eval_batch_size: 16
61
  - seed: 42
62
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: constant_with_warmup
64
+ - lr_scheduler_warmup_steps: 10
65
+ - training_steps: 2000
66
 
67
  ### Training results
68
 
69
+ | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
70
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:|
71
+ | 0.1621 | 1.14 | 1000 | 0.2587 | 61.0824 | 65.0094 |
72
+ | 0.0767 | 2.28 | 2000 | 0.2487 | 57.8423 | 57.7008 |
73
 
74
 
75
  ### Framework versions
generation_config.json CHANGED
@@ -51,7 +51,7 @@
51
  "forced_decoder_ids": [
52
  [
53
  1,
54
- 50322
55
  ],
56
  [
57
  2,
@@ -164,7 +164,7 @@
164
  "<|yo|>": 50325,
165
  "<|zh|>": 50260
166
  },
167
- "language": "sinhalese",
168
  "max_initial_timestamp_index": 1,
169
  "max_length": 448,
170
  "no_timestamps_token_id": 50363,
 
51
  "forced_decoder_ids": [
52
  [
53
  1,
54
+ 50260
55
  ],
56
  [
57
  2,
 
164
  "<|yo|>": 50325,
165
  "<|zh|>": 50260
166
  },
167
+ "language": "chinese",
168
  "max_initial_timestamp_index": 1,
169
  "max_length": 448,
170
  "no_timestamps_token_id": 50363,