Vezora commited on
Commit
4679b79
1 Parent(s): 8722432

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -50,7 +50,7 @@ Please note that Mistral-22b is still in a WIP. v0.3 has started training now, w
50
  - Thank you to [Tri Dao](https://twitter.com/tri_dao), for creating Flash Attention
51
  - Thank you to Microsoft, for the Lora paper, and the Slice-GPT paper.
52
  - Thank you to the Hugging Face team, for everything.❤️ We really do appreciate you guys and all your hard work and commitment to the open source community!❤️
53
- - Thank you to I used one of his DPO datasets converted to SFT, more info will be explained in paper.[Jon Durbin](https://x.com/jon_durbin?s=21)
54
 
55
 
56
  ## Future plans, train 4-5 more of these experimental models gather preliminary testing results, and then run evaluations on all the models I see have the best possibilities of excelling, then use the best one.
 
50
  - Thank you to [Tri Dao](https://twitter.com/tri_dao), for creating Flash Attention
51
  - Thank you to Microsoft, for the Lora paper, and the Slice-GPT paper.
52
  - Thank you to the Hugging Face team, for everything.❤️ We really do appreciate you guys and all your hard work and commitment to the open source community!❤️
53
+ - Thank you to [Jon Durbin](https://x.com/jon_durbin?s=21) I used one of his DPO datasets converted to SFT, more info will be explained in paper.
54
 
55
 
56
  ## Future plans, train 4-5 more of these experimental models gather preliminary testing results, and then run evaluations on all the models I see have the best possibilities of excelling, then use the best one.