BossRui commited on
Commit
725c85b
1 Parent(s): 2d0cec1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -98,7 +98,7 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
98
 
99
  ## Results
100
 
101
- ### Dense
102
  🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-0.5B" target="_blank">Apollo2-0.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-1.5B" target="_blank">Apollo2-1.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-2B" target="_blank">Apollo2-2B</a>
103
  🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-3.8B" target="_blank">Apollo2-3.8B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-7B" target="_blank">Apollo2-7B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-9B" target="_blank">Apollo2-9B</a>
104
 
@@ -109,7 +109,7 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
109
 
110
  </details>
111
 
112
- ### Post-MoE
113
  🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-MoE-0.5B" target="_blank">Apollo-MoE-0.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-MoE-1.5B" target="_blank">Apollo-MoE-1.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-MoE-7B" target="_blank">Apollo-MoE-7B</a>
114
 
115
  <details>
@@ -123,12 +123,12 @@ Covering 12 Major Languages including English, Chinese, French, Hindi, Spanish,
123
 
124
 
125
  ## Usage Format
126
- #### Apollo2
127
  - 0.5B, 1.5B, 7B: User:{query}\nAssistant:{response}<|endoftext|>
128
  - 2B, 9B: User:{query}\nAssistant:{response}\<eos\>
129
  - 3.8B: <|user|>\n{query}<|end|><|assisitant|>\n{response}<|end|>
130
 
131
- #### Apollo-MoE
132
  - 0.5B, 1.5B, 7B: User:{query}\nAssistant:{response}<|endoftext|>
133
 
134
  ## Dataset & Evaluation
 
98
 
99
  ## Results
100
 
101
+ #### Dense
102
  🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-0.5B" target="_blank">Apollo2-0.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-1.5B" target="_blank">Apollo2-1.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-2B" target="_blank">Apollo2-2B</a>
103
  🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-3.8B" target="_blank">Apollo2-3.8B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-7B" target="_blank">Apollo2-7B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo2-9B" target="_blank">Apollo2-9B</a>
104
 
 
109
 
110
  </details>
111
 
112
+ #### Post-MoE
113
  🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-MoE-0.5B" target="_blank">Apollo-MoE-0.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-MoE-1.5B" target="_blank">Apollo-MoE-1.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-MoE-7B" target="_blank">Apollo-MoE-7B</a>
114
 
115
  <details>
 
123
 
124
 
125
  ## Usage Format
126
+ ##### Apollo2
127
  - 0.5B, 1.5B, 7B: User:{query}\nAssistant:{response}<|endoftext|>
128
  - 2B, 9B: User:{query}\nAssistant:{response}\<eos\>
129
  - 3.8B: <|user|>\n{query}<|end|><|assisitant|>\n{response}<|end|>
130
 
131
+ ##### Apollo-MoE
132
  - 0.5B, 1.5B, 7B: User:{query}\nAssistant:{response}<|endoftext|>
133
 
134
  ## Dataset & Evaluation