alexkueck commited on
Commit
a71b514
·
verified ·
1 Parent(s): fe4701a

Update utils.py

Browse files
Files changed (1) hide show
  1. utils.py +7 -2
utils.py CHANGED
@@ -19,7 +19,7 @@ import operator
19
  from typing import Annotated, Sequence, TypedDict
20
  import pprint
21
 
22
- from transformers import pipeline, AutoModelForSeq2SeqLM, AutoTokenizer, AutoModelForCausalLM
23
  from sentence_transformers import SentenceTransformer, util
24
  from typing import List, Dict
25
 
@@ -151,11 +151,16 @@ EMBEDDING_MODELL = "sentence-transformers/all-mpnet-base-v2"
151
 
152
  #Modell und Tokenizer, um die Summary über die relevanten Texte zu machen
153
  #mögliche Modelle: "HuggingFaceH4/zephyr-7b-alpha" #"t5-small" #"meta-llama/Meta-Llama-3-8B-Instruct" #"mistralai/Mistral-7B-Instruct-v0.3" #"microsoft/Phi-3-mini-4k-instruct" #"HuggingFaceH4/zephyr-7b-alpha"
 
 
 
 
 
154
  HF_MODELL = "microsoft/Phi-3-mini-4k-instruct"
155
  # Laden des Modells und Tokenizers
156
  modell_rag = AutoModelForCausalLM.from_pretrained(HF_MODELL)
157
  tokenizer_rag = AutoTokenizer.from_pretrained(HF_MODELL)
158
- """
159
  HF_MODELL = "t5-small"
160
  modell_rag = AutoModelForSeq2SeqLM.from_pretrained(HF_MODELL)
161
  tokenizer_rag = AutoTokenizer.from_pretrained(HF_MODELL)
 
19
  from typing import Annotated, Sequence, TypedDict
20
  import pprint
21
 
22
+ from transformers import pipeline, AutoModelForSeq2SeqLM, AutoTokenizer, AutoModelForCausalLM, GPTNeoForCausalLM, GPT2Tokenizer
23
  from sentence_transformers import SentenceTransformer, util
24
  from typing import List, Dict
25
 
 
151
 
152
  #Modell und Tokenizer, um die Summary über die relevanten Texte zu machen
153
  #mögliche Modelle: "HuggingFaceH4/zephyr-7b-alpha" #"t5-small" #"meta-llama/Meta-Llama-3-8B-Instruct" #"mistralai/Mistral-7B-Instruct-v0.3" #"microsoft/Phi-3-mini-4k-instruct" #"HuggingFaceH4/zephyr-7b-alpha"
154
+ HF_MODELL ="EleutherAI/gpt-neo-2.7B"
155
+ modell_rag = GPTNeoForCausalLM.from_pretrained(HF_MODELL)
156
+ tokenizer_rag = GPT2Tokenizer.from_pretrained(HF_MODELL)
157
+
158
+ """
159
  HF_MODELL = "microsoft/Phi-3-mini-4k-instruct"
160
  # Laden des Modells und Tokenizers
161
  modell_rag = AutoModelForCausalLM.from_pretrained(HF_MODELL)
162
  tokenizer_rag = AutoTokenizer.from_pretrained(HF_MODELL)
163
+
164
  HF_MODELL = "t5-small"
165
  modell_rag = AutoModelForSeq2SeqLM.from_pretrained(HF_MODELL)
166
  tokenizer_rag = AutoTokenizer.from_pretrained(HF_MODELL)