Update README.md
Browse files
README.md
CHANGED
@@ -131,19 +131,6 @@ Also, we evaluate our model on the benchmarks of the first leaderboard using `li
|
|
131 |
|
132 |
Mostly, we took evaluation results from both leaderboards. For the models marked by *star* we evaluated the tasks internally, while for the models marked by two *stars* the results were taken from paper or model card.
|
133 |
|
134 |
-
## Throughput
|
135 |
-
|
136 |
-
This model can achieve comparable throughput and performance compared to other transformer based models that use optimized kernels such as Flash Attention 2. Make sure to install the optimized Mamba kernels with the following commands:
|
137 |
-
|
138 |
-
```bash
|
139 |
-
pip install "causal-conv1d>=1.4.0" mamba-ssm
|
140 |
-
```
|
141 |
-
|
142 |
-
Refer to our [FalconMamba blogpost](https://huggingface.co/blog/falconmamba) for more details about performance evaluation.
|
143 |
-
|
144 |
-
|
145 |
-
<br>
|
146 |
-
|
147 |
# Technical Specifications
|
148 |
|
149 |
## Model Architecture and Objective
|
|
|
131 |
|
132 |
Mostly, we took evaluation results from both leaderboards. For the models marked by *star* we evaluated the tasks internally, while for the models marked by two *stars* the results were taken from paper or model card.
|
133 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
134 |
# Technical Specifications
|
135 |
|
136 |
## Model Architecture and Objective
|