John6666 commited on
Commit
22d064d
·
verified ·
1 Parent(s): 0d05e2e

Upload 2 files

Browse files
Files changed (2) hide show
  1. dc.py +2 -6
  2. llmdolphin.py +10 -13
dc.py CHANGED
@@ -538,10 +538,6 @@ class GuiSD:
538
  if save_generated_images:
539
  info_images += f"<br>{download_links}"
540
 
541
- ## BEGIN MOD
542
- img = save_images(img, metadata)
543
- ## END MOD
544
-
545
  info_state = "COMPLETE"
546
 
547
  yield info_state, img, info_images
@@ -710,7 +706,7 @@ def infer(prompt, negative_prompt, seed, randomize_seed, width, height, guidance
710
 
711
  progress(0, desc="Loading model...")
712
  for m in sd_gen.load_new_model(model_name, vae, TASK_MODEL_LIST[0]):
713
- progress(0.5, desc=m)
714
  progress(1, desc="Model loaded.")
715
  progress(0, desc="Starting Inference...")
716
  images = None
@@ -726,7 +722,7 @@ def infer(prompt, negative_prompt, seed, randomize_seed, width, height, guidance
726
  True, None, None, "plus_face", "original", 0.7, None, None, "base", "style", 0.7, 0.0,
727
  load_lora_cpu, verbose_info, gpu_duration
728
  ):
729
- progress(0.5, desc=info_state)
730
  images = stream_images
731
  progress(1, desc="Inference completed.")
732
  output_image = images[0][0] if images else None
 
538
  if save_generated_images:
539
  info_images += f"<br>{download_links}"
540
 
 
 
 
 
541
  info_state = "COMPLETE"
542
 
543
  yield info_state, img, info_images
 
706
 
707
  progress(0, desc="Loading model...")
708
  for m in sd_gen.load_new_model(model_name, vae, TASK_MODEL_LIST[0]):
709
+ print(m)
710
  progress(1, desc="Model loaded.")
711
  progress(0, desc="Starting Inference...")
712
  images = None
 
722
  True, None, None, "plus_face", "original", 0.7, None, None, "base", "style", 0.7, 0.0,
723
  load_lora_cpu, verbose_info, gpu_duration
724
  ):
725
+ print(info_state)
726
  images = stream_images
727
  progress(1, desc="Inference completed.")
728
  output_image = images[0][0] if images else None
llmdolphin.py CHANGED
@@ -1,5 +1,9 @@
1
  import spaces
2
  import gradio as gr
 
 
 
 
3
  from llama_cpp import Llama
4
  from llama_cpp_agent import LlamaCppAgent, MessagesFormatterType
5
  from llama_cpp_agent.providers import LlamaCppPythonProvider
@@ -7,7 +11,6 @@ from llama_cpp_agent.chat_history import BasicChatHistory
7
  from llama_cpp_agent.chat_history.messages import Roles
8
  from ja_to_danbooru.ja_to_danbooru import jatags_to_danbooru_tags
9
  import wrapt_timeout_decorator
10
- from pathlib import Path
11
  from llama_cpp_agent.messages_formatter import MessagesFormatter
12
  from formatter import mistral_v1_formatter, mistral_v2_formatter, mistral_v3_tekken_formatter
13
 
@@ -846,6 +849,7 @@ llm_languages = ["English", "Japanese", "Chinese", "Korean", "Spanish", "Portugu
846
  llm_models_tupled_list = []
847
  default_llm_model_filename = list(llm_models.keys())[0]
848
  override_llm_format = None
 
849
 
850
 
851
  def to_list(s):
@@ -858,7 +862,6 @@ def list_uniq(l):
858
 
859
  @wrapt_timeout_decorator.timeout(dec_timeout=3.5)
860
  def to_list_ja(s):
861
- import re
862
  s = re.sub(r'[、。]', ',', s)
863
  return [x.strip() for x in s.split(",") if not s == ""]
864
 
@@ -873,7 +876,6 @@ def is_japanese(s):
873
 
874
 
875
  def update_llm_model_tupled_list():
876
- from pathlib import Path
877
  global llm_models_tupled_list
878
  llm_models_tupled_list = []
879
  for k, v in llm_models.items():
@@ -890,7 +892,6 @@ def update_llm_model_tupled_list():
890
 
891
 
892
  def download_llm_models():
893
- from huggingface_hub import hf_hub_download
894
  global llm_models_tupled_list
895
  llm_models_tupled_list = []
896
  for k, v in llm_models.items():
@@ -904,7 +905,6 @@ def download_llm_models():
904
 
905
 
906
  def download_llm_model(filename):
907
- from huggingface_hub import hf_hub_download
908
  if not filename in llm_models.keys(): return default_llm_model_filename
909
  try:
910
  hf_hub_download(repo_id = llm_models[filename][0], filename = filename, local_dir = llm_models_dir)
@@ -965,8 +965,6 @@ def get_dolphin_model_format(filename):
965
 
966
 
967
  def add_dolphin_models(query, format_name):
968
- import re
969
- from huggingface_hub import HfApi
970
  global llm_models
971
  api = HfApi()
972
  add_models = {}
@@ -991,7 +989,6 @@ def add_dolphin_models(query, format_name):
991
  except Exception as e:
992
  print(e)
993
  return gr.update(visible=True)
994
- #print(add_models)
995
  llm_models = (llm_models | add_models).copy()
996
  update_llm_model_tupled_list()
997
  choices = get_dolphin_models()
@@ -1191,7 +1188,6 @@ Output should be enclosed in //GENBEGIN//:// and //://GENEND//. The text to be g
1191
 
1192
 
1193
  def get_dolphin_sysprompt():
1194
- import re
1195
  prompt = re.sub('<LANGUAGE>', dolphin_output_language, dolphin_system_prompt.get(dolphin_sysprompt_mode, ""))
1196
  return prompt
1197
 
@@ -1221,11 +1217,11 @@ def select_dolphin_language(lang: str):
1221
 
1222
  @wrapt_timeout_decorator.timeout(dec_timeout=5.0)
1223
  def get_raw_prompt(msg: str):
1224
- import re
1225
  m = re.findall(r'/GENBEGIN/(.+?)/GENEND/', msg, re.DOTALL)
1226
  return re.sub(r'[*/:_"#\n]', ' ', ", ".join(m)).lower() if m else ""
1227
 
1228
 
 
1229
  @spaces.GPU(duration=60)
1230
  def dolphin_respond(
1231
  message: str,
@@ -1239,7 +1235,6 @@ def dolphin_respond(
1239
  repeat_penalty: float = 1.1,
1240
  progress=gr.Progress(track_tqdm=True),
1241
  ):
1242
- from pathlib import Path
1243
  progress(0, desc="Processing...")
1244
 
1245
  if override_llm_format:
@@ -1320,6 +1315,7 @@ def dolphin_parse(
1320
  return ", ".join(prompts), gr.update(interactive=True), gr.update(interactive=True)
1321
 
1322
 
 
1323
  @spaces.GPU(duration=60)
1324
  def dolphin_respond_auto(
1325
  message: str,
@@ -1334,7 +1330,6 @@ def dolphin_respond_auto(
1334
  progress=gr.Progress(track_tqdm=True),
1335
  ):
1336
  #if not is_japanese(message): return [(None, None)]
1337
- from pathlib import Path
1338
  progress(0, desc="Processing...")
1339
 
1340
  if override_llm_format:
@@ -1391,7 +1386,7 @@ def dolphin_respond_auto(
1391
  )
1392
 
1393
  progress(0.5, desc="Processing...")
1394
-
1395
  outputs = ""
1396
  for output in stream:
1397
  outputs += output
@@ -1421,6 +1416,8 @@ def dolphin_parse_simple(
1421
  import cv2
1422
  cv2.setNumThreads(1)
1423
 
 
 
1424
  @spaces.GPU()
1425
  def respond_playground(
1426
  message,
 
1
  import spaces
2
  import gradio as gr
3
+ from pathlib import Path
4
+ import re
5
+ import torch
6
+ from huggingface_hub import hf_hub_download, HfApi
7
  from llama_cpp import Llama
8
  from llama_cpp_agent import LlamaCppAgent, MessagesFormatterType
9
  from llama_cpp_agent.providers import LlamaCppPythonProvider
 
11
  from llama_cpp_agent.chat_history.messages import Roles
12
  from ja_to_danbooru.ja_to_danbooru import jatags_to_danbooru_tags
13
  import wrapt_timeout_decorator
 
14
  from llama_cpp_agent.messages_formatter import MessagesFormatter
15
  from formatter import mistral_v1_formatter, mistral_v2_formatter, mistral_v3_tekken_formatter
16
 
 
849
  llm_models_tupled_list = []
850
  default_llm_model_filename = list(llm_models.keys())[0]
851
  override_llm_format = None
852
+ device = "cuda" if torch.cuda.is_available() else "cpu"
853
 
854
 
855
  def to_list(s):
 
862
 
863
  @wrapt_timeout_decorator.timeout(dec_timeout=3.5)
864
  def to_list_ja(s):
 
865
  s = re.sub(r'[、。]', ',', s)
866
  return [x.strip() for x in s.split(",") if not s == ""]
867
 
 
876
 
877
 
878
  def update_llm_model_tupled_list():
 
879
  global llm_models_tupled_list
880
  llm_models_tupled_list = []
881
  for k, v in llm_models.items():
 
892
 
893
 
894
  def download_llm_models():
 
895
  global llm_models_tupled_list
896
  llm_models_tupled_list = []
897
  for k, v in llm_models.items():
 
905
 
906
 
907
  def download_llm_model(filename):
 
908
  if not filename in llm_models.keys(): return default_llm_model_filename
909
  try:
910
  hf_hub_download(repo_id = llm_models[filename][0], filename = filename, local_dir = llm_models_dir)
 
965
 
966
 
967
  def add_dolphin_models(query, format_name):
 
 
968
  global llm_models
969
  api = HfApi()
970
  add_models = {}
 
989
  except Exception as e:
990
  print(e)
991
  return gr.update(visible=True)
 
992
  llm_models = (llm_models | add_models).copy()
993
  update_llm_model_tupled_list()
994
  choices = get_dolphin_models()
 
1188
 
1189
 
1190
  def get_dolphin_sysprompt():
 
1191
  prompt = re.sub('<LANGUAGE>', dolphin_output_language, dolphin_system_prompt.get(dolphin_sysprompt_mode, ""))
1192
  return prompt
1193
 
 
1217
 
1218
  @wrapt_timeout_decorator.timeout(dec_timeout=5.0)
1219
  def get_raw_prompt(msg: str):
 
1220
  m = re.findall(r'/GENBEGIN/(.+?)/GENEND/', msg, re.DOTALL)
1221
  return re.sub(r'[*/:_"#\n]', ' ', ", ".join(m)).lower() if m else ""
1222
 
1223
 
1224
+ @torch.inference_mode()
1225
  @spaces.GPU(duration=60)
1226
  def dolphin_respond(
1227
  message: str,
 
1235
  repeat_penalty: float = 1.1,
1236
  progress=gr.Progress(track_tqdm=True),
1237
  ):
 
1238
  progress(0, desc="Processing...")
1239
 
1240
  if override_llm_format:
 
1315
  return ", ".join(prompts), gr.update(interactive=True), gr.update(interactive=True)
1316
 
1317
 
1318
+ @torch.inference_mode()
1319
  @spaces.GPU(duration=60)
1320
  def dolphin_respond_auto(
1321
  message: str,
 
1330
  progress=gr.Progress(track_tqdm=True),
1331
  ):
1332
  #if not is_japanese(message): return [(None, None)]
 
1333
  progress(0, desc="Processing...")
1334
 
1335
  if override_llm_format:
 
1386
  )
1387
 
1388
  progress(0.5, desc="Processing...")
1389
+
1390
  outputs = ""
1391
  for output in stream:
1392
  outputs += output
 
1416
  import cv2
1417
  cv2.setNumThreads(1)
1418
 
1419
+
1420
+ @torch.inference_mode()
1421
  @spaces.GPU()
1422
  def respond_playground(
1423
  message,