jonathanagustin commited on
Commit
889dcea
1 Parent(s): 9bc588b

Model save

Browse files
Files changed (2) hide show
  1. README.md +75 -40
  2. pytorch_model.bin +1 -1
README.md CHANGED
@@ -1,43 +1,78 @@
1
  ---
2
- {}
 
 
 
 
 
 
3
  ---
4
 
5
- ---
6
- language:
7
- - en
8
- license: mit
9
- tags:
10
- - question-answering
11
- - SQuAD
12
- - BERT
13
- datasets:
14
- - squad
15
- metrics:
16
- - f1
17
- - em
18
-
19
- model-index:
20
- - name: distilbert-finetuned-uncased-squad_v2
21
- results:
22
- - task:
23
- type: question-answering
24
- name: SQuAD Question Answering
25
- dataset:
26
- type: squad_v2
27
- name: SQuAD v2
28
- split: validation
29
- metrics:
30
- - type: f1
31
- value: 26.869992349988973
32
- name: F1 Score
33
- - type: em
34
- value: 23.347090036216628
35
- name: Exact Match
36
- verified: true
37
- ---
38
-
39
- # distilbert-finetuned-uncased
40
-
41
- This model is fine-tuned on SQuAD v2 for question answering tasks.
42
-
43
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ tags:
3
+ - generated_from_trainer
4
+ datasets:
5
+ - squad_v2
6
+ model-index:
7
+ - name: distilbert-finetuned-uncased-squad_v2
8
+ results: []
9
  ---
10
 
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # distilbert-finetuned-uncased-squad_v2
15
+
16
+ This model was trained from scratch on the squad_v2 dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 1.3332
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 2e-05
38
+ - train_batch_size: 64
39
+ - eval_batch_size: 64
40
+ - seed: 42
41
+ - gradient_accumulation_steps: 4
42
+ - total_train_batch_size: 256
43
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
+ - lr_scheduler_type: linear
45
+ - num_epochs: 4
46
+
47
+ ### Training results
48
+
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:-----:|:----:|:---------------:|
51
+ | 3.6437 | 0.39 | 100 | 2.1780 |
52
+ | 2.1596 | 0.78 | 200 | 1.6557 |
53
+ | 1.8138 | 1.18 | 300 | 1.5683 |
54
+ | 1.6987 | 1.57 | 400 | 1.5076 |
55
+ | 1.6586 | 1.96 | 500 | 1.5350 |
56
+ | 1.5957 | 1.18 | 600 | 1.4431 |
57
+ | 1.5825 | 1.37 | 700 | 1.4955 |
58
+ | 1.5523 | 1.57 | 800 | 1.4444 |
59
+ | 1.5346 | 1.76 | 900 | 1.3930 |
60
+ | 1.5098 | 1.96 | 1000 | 1.4285 |
61
+ | 1.4632 | 2.16 | 1100 | 1.3630 |
62
+ | 1.4468 | 2.35 | 1200 | 1.3710 |
63
+ | 1.4343 | 2.55 | 1300 | 1.3422 |
64
+ | 1.4225 | 2.75 | 1400 | 1.3971 |
65
+ | 1.408 | 2.94 | 1500 | 1.4355 |
66
+ | 1.3609 | 3.14 | 1600 | 1.3332 |
67
+ | 1.3398 | 3.33 | 1700 | 1.3792 |
68
+ | 1.3224 | 3.53 | 1800 | 1.4172 |
69
+ | 1.3152 | 3.73 | 1900 | 1.3956 |
70
+ | 1.3141 | 3.92 | 2000 | 1.3748 |
71
+
72
+
73
+ ### Framework versions
74
+
75
+ - Transformers 4.34.1
76
+ - Pytorch 2.1.0+cu118
77
+ - Datasets 2.14.5
78
+ - Tokenizers 0.14.1
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3014a9d75db768507f8291c1b3167335b3af67b40e635ab2c9708ff0fdceb4cf
3
  size 265493026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52b840f6d7f16d80e5a037be11ca007dc767ac82e97d5ddb9bf97093c0af6b01
3
  size 265493026