Training in progress, step 100
Browse files- README.md +5 -10
- adapter_model.safetensors +1 -1
README.md
CHANGED
@@ -2,7 +2,6 @@
|
|
2 |
license: apache-2.0
|
3 |
library_name: peft
|
4 |
tags:
|
5 |
-
- axolotl
|
6 |
- generated_from_trainer
|
7 |
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
8 |
model-index:
|
@@ -69,10 +68,10 @@ lora_target_modules:
|
|
69 |
- o_proj
|
70 |
|
71 |
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
|
77 |
|
78 |
gradient_accumulation_steps: 2
|
@@ -111,7 +110,7 @@ fsdp_config:
|
|
111 |
|
112 |
# mixtral-pb-20e
|
113 |
|
114 |
-
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on
|
115 |
|
116 |
## Model description
|
117 |
|
@@ -157,10 +156,6 @@ The following hyperparameters were used during training:
|
|
157 |
- lr_scheduler_warmup_steps: 10
|
158 |
- num_epochs: 20
|
159 |
|
160 |
-
### Training results
|
161 |
-
|
162 |
-
|
163 |
-
|
164 |
### Framework versions
|
165 |
|
166 |
- PEFT 0.7.0
|
|
|
2 |
license: apache-2.0
|
3 |
library_name: peft
|
4 |
tags:
|
|
|
5 |
- generated_from_trainer
|
6 |
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
7 |
model-index:
|
|
|
68 |
- o_proj
|
69 |
|
70 |
|
71 |
+
wandb_project: function-call
|
72 |
+
wandb_name: mixtral-instruct-lora--v1
|
73 |
+
wandb_log_model: end
|
74 |
+
hub_model_id: dyang415/mixtral-pb-20e
|
75 |
|
76 |
|
77 |
gradient_accumulation_steps: 2
|
|
|
110 |
|
111 |
# mixtral-pb-20e
|
112 |
|
113 |
+
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on an unknown dataset.
|
114 |
|
115 |
## Model description
|
116 |
|
|
|
156 |
- lr_scheduler_warmup_steps: 10
|
157 |
- num_epochs: 20
|
158 |
|
|
|
|
|
|
|
|
|
159 |
### Framework versions
|
160 |
|
161 |
- PEFT 0.7.0
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 27297032
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87ba511a3f83c87a3c172e1dd73a924b4cf84def31d05067854b0ce64ddcb551
|
3 |
size 27297032
|