GGUF
English
Inference Endpoints
conversational
ybelkada commited on
Commit
721af8d
1 Parent(s): 84468e1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -13
README.md CHANGED
@@ -131,19 +131,6 @@ Also, we evaluate our model on the benchmarks of the first leaderboard using `li
131
 
132
  Mostly, we took evaluation results from both leaderboards. For the models marked by *star* we evaluated the tasks internally, while for the models marked by two *stars* the results were taken from paper or model card.
133
 
134
- ## Throughput
135
-
136
- This model can achieve comparable throughput and performance compared to other transformer based models that use optimized kernels such as Flash Attention 2. Make sure to install the optimized Mamba kernels with the following commands:
137
-
138
- ```bash
139
- pip install "causal-conv1d>=1.4.0" mamba-ssm
140
- ```
141
-
142
- Refer to our [FalconMamba blogpost](https://huggingface.co/blog/falconmamba) for more details about performance evaluation.
143
-
144
-
145
- <br>
146
-
147
  # Technical Specifications
148
 
149
  ## Model Architecture and Objective
 
131
 
132
  Mostly, we took evaluation results from both leaderboards. For the models marked by *star* we evaluated the tasks internally, while for the models marked by two *stars* the results were taken from paper or model card.
133
 
 
 
 
 
 
 
 
 
 
 
 
 
 
134
  # Technical Specifications
135
 
136
  ## Model Architecture and Objective