Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license: mit
|
|
5 |
## TwiBERT
|
6 |
## Model Description
|
7 |
TwiBERT is a pre-trained language model specifically designed for the Twi language, which is widely spoken in Ghana,
|
8 |
-
West Africa. This model
|
9 |
To optimize its performance, TwiBERT was trained using a combination of the Asanti Twi Bible and a dataset
|
10 |
sourced through crowdsourcing efforts.
|
11 |
|
@@ -25,6 +25,6 @@ The example code below illustrates how you can use the TwiBERT model on a downtr
|
|
25 |
|
26 |
```python
|
27 |
>>> from transformers import AutoTokenizer, AutoModelForTokenClassification
|
28 |
-
>>> model = AutoModelForTokenClassification.from_pretrained("sakrah/
|
29 |
-
>>> tokenizer = AutoTokenizer.from_pretrained("sakrah/
|
30 |
```
|
|
|
5 |
## TwiBERT
|
6 |
## Model Description
|
7 |
TwiBERT is a pre-trained language model specifically designed for the Twi language, which is widely spoken in Ghana,
|
8 |
+
West Africa. This model has 61 million parameters, 6 layers, 6 attention heads, 768 hidden units, and a feed-forward size of 3072.
|
9 |
To optimize its performance, TwiBERT was trained using a combination of the Asanti Twi Bible and a dataset
|
10 |
sourced through crowdsourcing efforts.
|
11 |
|
|
|
25 |
|
26 |
```python
|
27 |
>>> from transformers import AutoTokenizer, AutoModelForTokenClassification
|
28 |
+
>>> model = AutoModelForTokenClassification.from_pretrained("sakrah/TwiBERT")
|
29 |
+
>>> tokenizer = AutoTokenizer.from_pretrained("sakrah/TwiBERT")
|
30 |
```
|