ahmedmbutt commited on
Commit
45b40aa
1 Parent(s): d7f5c63

End of training

Browse files
README.md CHANGED
@@ -7,41 +7,7 @@ metrics:
7
  - rouge
8
  model-index:
9
  - name: PTS-Bart-Large-CNN
10
- results:
11
- - task:
12
- type: summarization
13
- name: Summarization
14
- dataset:
15
- name: PTS Dataset
16
- type: PTS-Dataset
17
- metrics:
18
- - name: Rouge1
19
- type: rouge
20
- value: 0.6591
21
- - name: Rouge2
22
- type: rouge
23
- value: 0.449
24
- - name: Rougel
25
- type: rouge
26
- value: 0.5635
27
- - name: Rougelsum
28
- type: rouge
29
- value: 0.5633
30
- datasets:
31
- - ahmedmbutt/PTS-Dataset
32
- language:
33
- - en
34
- library_name: transformers
35
- widget:
36
- - text: >-
37
- I have to say that I do miss talking to a good psychiatrist- however. I
38
- could sit and argue for ages with a psychiatrist who is intelligent and kind
39
- (quite hard to find- but they do exist). Especially now that I have a PhD in
40
- philosophy and have read everything that can be found on madness- including
41
- the notes they wrote about me when I was in the hospital. Nowadays-
42
- psychiatrists have a tendency to sign me off pretty quickly when I come onto
43
- their radar. They don’t wish to deal with me- I tire them out.
44
- pipeline_tag: summarization
45
  ---
46
 
47
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -49,14 +15,14 @@ should probably proofread and complete it, then remove this comment. -->
49
 
50
  # PTS-Bart-Large-CNN
51
 
52
- This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the PTS dataset.
53
  It achieves the following results on the evaluation set:
54
- - Loss: 1.1442
55
- - Rouge1: 0.6591
56
- - Rouge2: 0.449
57
- - Rougel: 0.5635
58
- - Rougelsum: 0.5633
59
- - Gen Len: 78.7977
60
 
61
  ## Model description
62
 
@@ -88,14 +54,14 @@ The following hyperparameters were used during training:
88
 
89
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
90
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
91
- | No log | 1.0 | 220 | 0.8235 | 0.6279 | 0.4019 | 0.5268 | 0.5267 | 82.8295 |
92
- | No log | 2.0 | 440 | 0.8053 | 0.6461 | 0.4278 | 0.5486 | 0.5484 | 78.6318 |
93
- | 0.7147 | 3.0 | 660 | 0.8889 | 0.6471 | 0.4324 | 0.5491 | 0.5488 | 79.4432 |
94
- | 0.7147 | 4.0 | 880 | 0.9679 | 0.6533 | 0.4391 | 0.5538 | 0.5534 | 80.2023 |
95
- | 0.2566 | 5.0 | 1100 | 0.9734 | 0.6563 | 0.4422 | 0.5574 | 0.5571 | 78.9727 |
96
- | 0.2566 | 6.0 | 1320 | 1.0504 | 0.6538 | 0.4436 | 0.559 | 0.5585 | 78.5682 |
97
- | 0.1136 | 7.0 | 1540 | 1.1172 | 0.6591 | 0.4474 | 0.5646 | 0.5647 | 78.6068 |
98
- | 0.1136 | 8.0 | 1760 | 1.1442 | 0.6591 | 0.449 | 0.5635 | 0.5633 | 78.7977 |
99
 
100
 
101
  ### Framework versions
@@ -103,4 +69,4 @@ The following hyperparameters were used during training:
103
  - Transformers 4.41.2
104
  - Pytorch 2.3.0+cu121
105
  - Datasets 2.20.0
106
- - Tokenizers 0.19.1
 
7
  - rouge
8
  model-index:
9
  - name: PTS-Bart-Large-CNN
10
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
15
 
16
  # PTS-Bart-Large-CNN
17
 
18
+ This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 1.1760
21
+ - Rouge1: 0.6551
22
+ - Rouge2: 0.4332
23
+ - Rougel: 0.5543
24
+ - Rougelsum: 0.5541
25
+ - Gen Len: 80.0886
26
 
27
  ## Model description
28
 
 
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
57
+ | No log | 1.0 | 220 | 0.8239 | 0.6263 | 0.3973 | 0.5238 | 0.5237 | 84.2023 |
58
+ | No log | 2.0 | 440 | 0.8201 | 0.6461 | 0.4184 | 0.5417 | 0.5416 | 81.1659 |
59
+ | 0.7121 | 3.0 | 660 | 0.8661 | 0.6479 | 0.4226 | 0.5448 | 0.5454 | 80.5409 |
60
+ | 0.7121 | 4.0 | 880 | 0.9784 | 0.6474 | 0.4242 | 0.5424 | 0.5425 | 82.2932 |
61
+ | 0.2619 | 5.0 | 1100 | 1.0645 | 0.655 | 0.4327 | 0.5517 | 0.5517 | 80.8386 |
62
+ | 0.2619 | 6.0 | 1320 | 1.1098 | 0.6548 | 0.4339 | 0.5542 | 0.5543 | 81.3545 |
63
+ | 0.1124 | 7.0 | 1540 | 1.1528 | 0.6528 | 0.4298 | 0.5511 | 0.551 | 80.5705 |
64
+ | 0.1124 | 8.0 | 1760 | 1.1760 | 0.6551 | 0.4332 | 0.5543 | 0.5541 | 80.0886 |
65
 
66
 
67
  ### Framework versions
 
69
  - Transformers 4.41.2
70
  - Pytorch 2.3.0+cu121
71
  - Datasets 2.20.0
72
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:152f698bc2f3c8d305b7ffb5146b9dc4ce244bd36740dc73a223ae948426e8f0
3
  size 1625422896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:decb05eb85026d23066f4d7c898a7942df3fc8ebe27585ae528b438f830947ee
3
  size 1625422896
runs/Jun24_16-23-47_a407000c0675/events.out.tfevents.1719246228.a407000c0675.257.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:184d286253d5a42450d492e3bcee9350aff1ee65edf819d733dc632386efce1d
3
- size 9727
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d2b5624a59c6350b968ea01ae584711ab985e9bd917b67f626fce0cd3da4af7
3
+ size 11131