File size: 615 Bytes
7daa177
af93e1b
4f68cb1
 
 
 
 
 
d343f8a
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
---
license: apache-2.0
---

Mamba-Chat is the first chat language model based on a state-space model architecture, not a transformer.

The model is a fine-tune of Albert Gu's and Tri Dao's model [Mamba-2.8B](https://github.com/state-spaces/mamba) from their paper *Mamba: Linear-Time Sequence Modeling with Selective State Spaces*.

Check our our [Github repository](https://github.com/havenhq/mamba-chat/tree/main) for training and inference code.

The prompt format is the zephyr format:

```
<|user|> {user_message}
<|assistant|> {assistant_message}
<|user|> {user_message}
<|assistant|> {assistant_message}
```