Trained on the Falcon 7b
Browse files- README.md +2 -7
- adapter_config.json +1 -1
- adapter_model.safetensors +2 -2
README.md
CHANGED
@@ -1,11 +1,6 @@
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
-
|
4 |
-
base_model: bigscience/bloom-7b1
|
5 |
-
license: apache-2.0
|
6 |
-
tags:
|
7 |
-
- text-generation-inference
|
8 |
-
pipeline_tag: text-generation
|
9 |
---
|
10 |
|
11 |
# Model Card for Model ID
|
@@ -222,4 +217,4 @@ The following `bitsandbytes` quantization config was used during training:
|
|
222 |
### Framework versions
|
223 |
|
224 |
|
225 |
-
- PEFT 0.6.3.dev0
|
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
+
base_model: tiiuae/falcon-7b
|
|
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
6 |
# Model Card for Model ID
|
|
|
217 |
### Framework versions
|
218 |
|
219 |
|
220 |
+
- PEFT 0.6.3.dev0
|
adapter_config.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
-
"base_model_name_or_path": "
|
5 |
"bias": "none",
|
6 |
"fan_in_fan_out": false,
|
7 |
"inference_mode": true,
|
|
|
1 |
{
|
2 |
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "tiiuae/falcon-7b",
|
5 |
"bias": "none",
|
6 |
"fan_in_fan_out": false,
|
7 |
"inference_mode": true,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f24ca81da2ed2a1804ef875a0a761fd691cc80bfc53ddcaeeb786d30aa636b14
|
3 |
+
size 18883912
|