Update README.md
Browse files
README.md
CHANGED
@@ -30,7 +30,7 @@ pip install causal-conv1d>=1.2.0
|
|
30 |
pip install mamba-ssm
|
31 |
```
|
32 |
|
33 |
-
If any of these two is not installed, the "eager" implementation will be used(not recommended). Otherwise the more optimised `
|
34 |
|
35 |
### Generation
|
36 |
|
@@ -76,7 +76,7 @@ print(tokenizer.batch_decode(out))
|
|
76 |
|
77 |
### On-device Inference
|
78 |
|
79 |
-
Since Mambaoutai is only 1.6B parameters, it can run on a CPU
|
80 |
|
81 |
Here is an example of how to run it on llama.cpp:
|
82 |
|
|
|
30 |
pip install mamba-ssm
|
31 |
```
|
32 |
|
33 |
+
If any of these two is not installed, the "eager" implementation will be used(not recommended). Otherwise the more optimised `CUDA` kernels will be used.
|
34 |
|
35 |
### Generation
|
36 |
|
|
|
76 |
|
77 |
### On-device Inference
|
78 |
|
79 |
+
Since Mambaoutai is only 1.6B parameters, it can be run on a CPU with reasonable speed.
|
80 |
|
81 |
Here is an example of how to run it on llama.cpp:
|
82 |
|