Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -79,9 +79,9 @@ login(token=os.environ["HF_ACCESS_TOKEN"]) #for read access!!!!
|
|
79 |
#Modelle und Tokenizer
|
80 |
|
81 |
#Alternativ mit beliebigen Modellen:
|
82 |
-
|
83 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = False (in load_tokenizer_and_model)
|
84 |
-
base_model = "EleutherAI/gpt-neo-1.3B" #load_8bit = False (in load_tokenizer_and_model)
|
85 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = True
|
86 |
|
87 |
#Tokenizer und Model laden
|
@@ -140,7 +140,7 @@ print ("training args")
|
|
140 |
batch_size = 2
|
141 |
|
142 |
training_args = TrainingArguments(
|
143 |
-
output_dir="alexkueck/li-tis-tuned-
|
144 |
overwrite_output_dir = 'True',
|
145 |
per_device_train_batch_size=batch_size, #batch_size = 2 for full training
|
146 |
per_device_eval_batch_size=batch_size,
|
@@ -221,8 +221,8 @@ print("Done Eval")
|
|
221 |
#Push to Hub
|
222 |
print("push to hub")
|
223 |
login(token=os.environ["HF_WRITE_TOKEN"])
|
224 |
-
|
225 |
-
tokenizer.push_to_hub("alexkueck/li-tis-tuned-
|
226 |
print("done")
|
227 |
|
228 |
|
|
|
79 |
#Modelle und Tokenizer
|
80 |
|
81 |
#Alternativ mit beliebigen Modellen:
|
82 |
+
base_model = "project-baize/baize-v2-7b" #load_8bit = False (in load_tokenizer_and_model)
|
83 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = False (in load_tokenizer_and_model)
|
84 |
+
#base_model = "EleutherAI/gpt-neo-1.3B" #load_8bit = False (in load_tokenizer_and_model)
|
85 |
#base_model = "TheBloke/airoboros-13B-HF" #load_8bit = True
|
86 |
|
87 |
#Tokenizer und Model laden
|
|
|
140 |
batch_size = 2
|
141 |
|
142 |
training_args = TrainingArguments(
|
143 |
+
output_dir="alexkueck/li-tis-tuned-2",
|
144 |
overwrite_output_dir = 'True',
|
145 |
per_device_train_batch_size=batch_size, #batch_size = 2 for full training
|
146 |
per_device_eval_batch_size=batch_size,
|
|
|
221 |
#Push to Hub
|
222 |
print("push to hub")
|
223 |
login(token=os.environ["HF_WRITE_TOKEN"])
|
224 |
+
trainer.push_to_hub("alexkueck/li-tis-tuned-2")
|
225 |
+
tokenizer.push_to_hub("alexkueck/li-tis-tuned-2")
|
226 |
print("done")
|
227 |
|
228 |
|