pabRomero commited on
Commit
1d789c6
1 Parent(s): 5720dc6

Training complete

Browse files
README.md CHANGED
@@ -1,72 +1,74 @@
1
- ---
2
- license: apache-2.0
3
- base_model: pabRomero/BERT-full-finetuned-ner-pablo
4
- tags:
5
- - generated_from_trainer
6
- metrics:
7
- - precision
8
- - recall
9
- - f1
10
- - accuracy
11
- model-index:
12
- - name: BERT-full-finetuned-ner-pablo
13
- results: []
14
- ---
15
-
16
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
- should probably proofread and complete it, then remove this comment. -->
18
-
19
- # BERT-full-finetuned-ner-pablo
20
-
21
- This model is a fine-tuned version of [pabRomero/BERT-full-finetuned-ner-pablo](https://huggingface.co/pabRomero/BERT-full-finetuned-ner-pablo) on the None dataset.
22
- It achieves the following results on the evaluation set:
23
- - Loss: 0.1343
24
- - Precision: 0.7987
25
- - Recall: 0.7855
26
- - F1: 0.7920
27
- - Accuracy: 0.9688
28
-
29
- ## Model description
30
-
31
- More information needed
32
-
33
- ## Intended uses & limitations
34
-
35
- More information needed
36
-
37
- ## Training and evaluation data
38
-
39
- More information needed
40
-
41
- ## Training procedure
42
-
43
- ### Training hyperparameters
44
-
45
- The following hyperparameters were used during training:
46
- - learning_rate: 0.0002
47
- - train_batch_size: 4
48
- - eval_batch_size: 4
49
- - seed: 42
50
- - gradient_accumulation_steps: 4
51
- - total_train_batch_size: 16
52
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
- - lr_scheduler_type: linear
54
- - lr_scheduler_warmup_ratio: 0.05
55
- - num_epochs: 3
56
- - mixed_precision_training: Native AMP
57
-
58
- ### Training results
59
-
60
- | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
61
- |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
62
- | 0.1038 | 0.9998 | 2608 | 0.1377 | 0.7579 | 0.7370 | 0.7473 | 0.9645 |
63
- | 0.0715 | 2.0 | 5217 | 0.1353 | 0.7700 | 0.7625 | 0.7663 | 0.9667 |
64
- | 0.0477 | 2.9994 | 7824 | 0.1343 | 0.7987 | 0.7855 | 0.7920 | 0.9688 |
65
-
66
-
67
- ### Framework versions
68
-
69
- - Transformers 4.44.0
70
- - Pytorch 2.4.0+cu124
71
- - Datasets 2.21.0
72
- - Tokenizers 0.19.1
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google-bert/bert-base-uncased
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - precision
8
+ - recall
9
+ - f1
10
+ - accuracy
11
+ model-index:
12
+ - name: BERT-full-finetuned-ner-pablo
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # BERT-full-finetuned-ner-pablo
20
+
21
+ This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on the None dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.1071
24
+ - Precision: 0.7993
25
+ - Recall: 0.7887
26
+ - F1: 0.7940
27
+ - Accuracy: 0.9768
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 0.0002
47
+ - train_batch_size: 16
48
+ - eval_batch_size: 16
49
+ - seed: 42
50
+ - gradient_accumulation_steps: 4
51
+ - total_train_batch_size: 64
52
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
+ - lr_scheduler_type: linear
54
+ - lr_scheduler_warmup_ratio: 0.05
55
+ - num_epochs: 5
56
+ - mixed_precision_training: Native AMP
57
+
58
+ ### Training results
59
+
60
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
61
+ |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
62
+ | No log | 0.9923 | 97 | 0.1037 | 0.7655 | 0.7399 | 0.7525 | 0.9725 |
63
+ | No log | 1.9949 | 195 | 0.0907 | 0.8123 | 0.7488 | 0.7792 | 0.9759 |
64
+ | No log | 2.9974 | 293 | 0.0922 | 0.7739 | 0.7872 | 0.7805 | 0.9758 |
65
+ | No log | 4.0 | 391 | 0.0986 | 0.7856 | 0.7895 | 0.7875 | 0.9760 |
66
+ | No log | 4.9616 | 485 | 0.1071 | 0.7993 | 0.7887 | 0.7940 | 0.9768 |
67
+
68
+
69
+ ### Framework versions
70
+
71
+ - Transformers 4.44.0
72
+ - Pytorch 2.4.0+cu121
73
+ - Datasets 2.21.0
74
+ - Tokenizers 0.19.1
runs/Aug20_15-46-03_669c7f3efb42/events.out.tfevents.1724168764.669c7f3efb42.2417.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08b9cde2efe86a6859ebcdd96887124c6ddd41a3b63c98d61945d874f8b5a4a5
3
- size 7623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ce55810aeb4a970510a12ec1056bcb958b399e74ca31e59f6794deb4aca1726
3
+ size 8449