Upload folder using huggingface_hub
Browse files- README.md +3 -2
- adapter_config.json +1 -1
- adapter_model.safetensors +1 -1
README.md
CHANGED
@@ -9,7 +9,7 @@ tags:
|
|
9 |
- pytorch
|
10 |
- llama
|
11 |
- llama-2
|
12 |
-
base_model: DavidLanz/
|
13 |
model_name: Llama 3 8B Instruct
|
14 |
inference: false
|
15 |
model_creator: Meta Llama 3
|
@@ -71,7 +71,7 @@ bnb_config = BitsAndBytesConfig(
|
|
71 |
bnb_4bit_use_double_quant=use_nested_quant,
|
72 |
)
|
73 |
|
74 |
-
based_model_path = "DavidLanz/
|
75 |
adapter_path = "DavidLanz/llama3_8b_taiwan_stock_qlora"
|
76 |
|
77 |
base_model = AutoModelForCausalLM.from_pretrained(
|
@@ -128,4 +128,5 @@ print(outputs[0]["generated_text"][len(prompt):])
|
|
128 |
|
129 |
### Framework versions
|
130 |
|
|
|
131 |
- PEFT 0.10.0
|
|
|
9 |
- pytorch
|
10 |
- llama
|
11 |
- llama-2
|
12 |
+
base_model: DavidLanz/Llama3-tw-8B-Instruct
|
13 |
model_name: Llama 3 8B Instruct
|
14 |
inference: false
|
15 |
model_creator: Meta Llama 3
|
|
|
71 |
bnb_4bit_use_double_quant=use_nested_quant,
|
72 |
)
|
73 |
|
74 |
+
based_model_path = "DavidLanz/Llama3-tw-8B-Instruct"
|
75 |
adapter_path = "DavidLanz/llama3_8b_taiwan_stock_qlora"
|
76 |
|
77 |
base_model = AutoModelForCausalLM.from_pretrained(
|
|
|
128 |
|
129 |
### Framework versions
|
130 |
|
131 |
+
- PEFT 0.11.1
|
132 |
- PEFT 0.10.0
|
adapter_config.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
-
"base_model_name_or_path": "DavidLanz/
|
5 |
"bias": "none",
|
6 |
"fan_in_fan_out": false,
|
7 |
"inference_mode": true,
|
|
|
1 |
{
|
2 |
"alpha_pattern": {},
|
3 |
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "DavidLanz/Llama3-tw-8B-Instruct",
|
5 |
"bias": "none",
|
6 |
"fan_in_fan_out": false,
|
7 |
"inference_mode": true,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 109069176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7dc4d8a655546fa1623d4816c07adc10eea67b26de00078bc36b30db78508ce4
|
3 |
size 109069176
|