ruggsea commited on
Commit
01ef28b
·
verified ·
1 Parent(s): 9de54c4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -7,8 +7,8 @@ import spaces
7
  import torch
8
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
9
 
10
- MAX_MAX_NEW_TOKENS = 4096
11
- DEFAULT_MAX_NEW_TOKENS = 2048
12
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
13
 
14
  DESCRIPTION = """\
@@ -32,7 +32,7 @@ if not torch.cuda.is_available():
32
 
33
 
34
  if torch.cuda.is_available():
35
- model_id = "ruggsea/Llama3-stanford-encyclopedia-philosophy-QA"
36
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
37
  tokenizer = AutoTokenizer.from_pretrained(model_id)
38
  tokenizer.use_default_system_prompt = False
 
7
  import torch
8
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
9
 
10
+ MAX_MAX_NEW_TOKENS = 8000
11
+ DEFAULT_MAX_NEW_TOKENS = 4000
12
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
13
 
14
  DESCRIPTION = """\
 
32
 
33
 
34
  if torch.cuda.is_available():
35
+ model_id = "ruggsea/Llama3.1-Chat-stanford-encyclopedia-philosophy"
36
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
37
  tokenizer = AutoTokenizer.from_pretrained(model_id)
38
  tokenizer.use_default_system_prompt = False