fbellame commited on
Commit
7cec10d
1 Parent(s): cd1a067

End of training

Browse files
Files changed (1) hide show
  1. README.md +79 -72
README.md CHANGED
@@ -1,73 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
- ---
3
- language:
4
- - en
5
- library_name: transformers
6
- tags:
7
- - gpt
8
- - llm
9
- - large language model
10
- - trl
11
- inference: false
12
- thumbnail: https://h2o.ai/etc.clientlibs/h2o/clientlibs/clientlib-site/resources/images/favicon.ico
13
- ---
14
- # Model Card
15
- ## Summary
16
-
17
- This model was trained using [TRL SFTTrainer](https://huggingface.co/docs/trl/main/en/sft_trainer).
18
- - Base model: [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
19
-
20
- ## Model Description
21
- Fine tuning to produce json quizz stability:
22
-
23
- This quiz:
24
-
25
- ```
26
- Question: Who is considered the father of artificial intelligence?
27
-
28
- A: Alan Turing
29
- B: John McCarthy
30
- C: Elon Musk
31
- D: Stephen Hawking
32
- Correct Answer: B
33
- ```
34
-
35
- Should return this json quiz:
36
- ```json
37
- {
38
- "params": {
39
- "questions": [
40
- {
41
- "question": "Who is considered the father of artificial intelligence?",
42
- "A": "Alan Turing",
43
- "B": "John McCarthy",
44
- "C": "Elon Musk",
45
- "D": "Stephen Hawking",
46
- "reponse": "B"
47
- }
48
- ]
49
- }
50
- }
51
- ```
52
-
53
- ## How to Use
54
- Provide instructions on how to use the model.
55
-
56
- ## Training Data
57
- Describe the data you used to train the model.
58
-
59
- ## Training Procedure
60
- Describe the training process, including any important parameters.
61
-
62
- ## Evaluation Results
63
- Include evaluation results if available.
64
-
65
- ## Limitations and Bias
66
- Discuss any limitations and potential biases in your model.
67
-
68
- ## Ethical Considerations
69
- Include any ethical considerations here.
70
-
71
- ## Citation
72
- Provide a citation if applicable.
73
-
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - trl
6
+ - sft
7
+ - generated_from_trainer
8
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
9
+ model-index:
10
+ - name: mistral-7b-json-quizz-fine-tuned-trl
11
+ results: []
12
+ ---
13
 
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # mistral-7b-json-quizz-fine-tuned-trl
18
+
19
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.6757
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 0.0002
41
+ - train_batch_size: 1
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: cosine
46
+ - training_steps: 50
47
+ - mixed_precision_training: Native AMP
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss |
52
+ |:-------------:|:-----:|:----:|:---------------:|
53
+ | 0.1134 | 4.55 | 50 | 0.6757 |
54
+
55
+
56
+ ### Framework versions
57
+
58
+ - Transformers 4.37.0.dev0
59
+ - Pytorch 2.1.1+cu121
60
+ - Datasets 2.15.0
61
+ - Tokenizers 0.15.0
62
+ ## Training procedure
63
+
64
+
65
+ The following `bitsandbytes` quantization config was used during training:
66
+ - quant_method: bitsandbytes
67
+ - load_in_8bit: True
68
+ - load_in_4bit: False
69
+ - llm_int8_threshold: 6.0
70
+ - llm_int8_skip_modules: None
71
+ - llm_int8_enable_fp32_cpu_offload: False
72
+ - llm_int8_has_fp16_weight: False
73
+ - bnb_4bit_quant_type: fp4
74
+ - bnb_4bit_use_double_quant: False
75
+ - bnb_4bit_compute_dtype: float32
76
+
77
+ ### Framework versions
78
+
79
+
80
+ - PEFT 0.6.2