File size: 615 Bytes
7daa177 af93e1b 4f68cb1 d343f8a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
---
license: apache-2.0
---
Mamba-Chat is the first chat language model based on a state-space model architecture, not a transformer.
The model is a fine-tune of Albert Gu's and Tri Dao's model [Mamba-2.8B](https://github.com/state-spaces/mamba) from their paper *Mamba: Linear-Time Sequence Modeling with Selective State Spaces*.
Check our our [Github repository](https://github.com/havenhq/mamba-chat/tree/main) for training and inference code.
The prompt format is the zephyr format:
```
<|user|> {user_message}
<|assistant|> {assistant_message}
<|user|> {user_message}
<|assistant|> {assistant_message}
``` |