Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -89,6 +89,20 @@ MODEL_CONTEXT_SIZES = {
|
|
89 |
"llama-3.1-70b-specdec": 131072,
|
90 |
"llama-3.2-1b-preview": 131072,
|
91 |
"llama-3.2-3b-preview": 131072,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
92 |
}
|
93 |
}
|
94 |
|
@@ -335,7 +349,9 @@ def build_prompts(snippets: List[str], prompt_instruction: str, custom_prompt: O
|
|
335 |
|
336 |
def send_to_model(prompt, model_selection, hf_model_choice, hf_custom_model, hf_api_key,
|
337 |
groq_model_choice, groq_api_key, openai_api_key, openai_model_choice):
|
338 |
-
|
|
|
|
|
339 |
|
340 |
if not prompt or not prompt.strip():
|
341 |
return "Error: No prompt provided", None
|
@@ -344,27 +360,69 @@ def send_to_model(prompt, model_selection, hf_model_choice, hf_custom_model, hf_
|
|
344 |
logging.info("sending to model preparation.")
|
345 |
|
346 |
# Basic input validation
|
347 |
-
|
|
|
348 |
return "Error: Invalid model selection", None
|
349 |
|
350 |
-
# Model-specific validation
|
351 |
if model_selection == "Groq API" and not groq_api_key:
|
352 |
return "Error: Groq API key required", None
|
353 |
elif model_selection == "OpenAI ChatGPT" and not openai_api_key:
|
354 |
return "Error: OpenAI API key required", None
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
355 |
|
356 |
def send_to_model_impl(prompt, model_selection, hf_model_choice, hf_custom_model, hf_api_key,
|
357 |
groq_model_choice, groq_api_key, openai_api_key, openai_model_choice):
|
358 |
"""Implementation of model sending with improved error handling."""
|
359 |
logging.info("send to model impl commencing...")
|
360 |
|
361 |
-
if model_selection == "Clipboard only":
|
362 |
-
return "Text copied to clipboard. Use paste for processing.", None
|
363 |
-
|
364 |
try:
|
|
|
|
|
|
|
365 |
if model_selection == "HuggingFace Inference":
|
|
|
366 |
model_id = hf_custom_model if hf_model_choice == "Custom Model" else model_registry.hf_models[hf_model_choice]
|
367 |
-
summary = send_to_hf_inference(prompt, model_id
|
|
|
|
|
|
|
368 |
|
369 |
elif model_selection == "Groq API":
|
370 |
summary = send_to_groq(prompt, groq_model_choice, groq_api_key)
|
@@ -372,12 +430,13 @@ def send_to_model_impl(prompt, model_selection, hf_model_choice, hf_custom_model
|
|
372 |
elif model_selection == "OpenAI ChatGPT":
|
373 |
summary = send_to_openai(prompt, openai_api_key, model=openai_model_choice)
|
374 |
|
|
|
|
|
|
|
375 |
else:
|
376 |
return "Error: Invalid model selection", None
|
377 |
|
378 |
# Validate response
|
379 |
-
logging.info("model told us:", summary)
|
380 |
-
|
381 |
if not summary or not isinstance(summary, str):
|
382 |
return "Error: Invalid response from model", None
|
383 |
|
@@ -391,12 +450,29 @@ def send_to_model_impl(prompt, model_selection, hf_model_choice, hf_custom_model
|
|
391 |
|
392 |
except Exception as e:
|
393 |
error_msg = str(e)
|
394 |
-
if not error_msg:
|
395 |
error_msg = "Unknown error occurred"
|
396 |
logging.error(f"Error in send_to_model_impl: {error_msg}")
|
397 |
return f"Error: {error_msg}", None
|
398 |
-
|
399 |
def send_to_hf_inference(prompt: str, model_name: str, api_key: str = None) -> str:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
400 |
"""Send prompt to HuggingFace Inference API with optional authentication."""
|
401 |
try:
|
402 |
# First try without authentication
|
@@ -451,17 +527,13 @@ def send_to_groq(prompt: str, model_name: str, api_key: str) -> str:
|
|
451 |
return response.choices[0].message.content
|
452 |
except Exception as e:
|
453 |
logging.error(f"Groq API error: {e}")
|
454 |
-
|
455 |
|
456 |
def send_to_openai(prompt: str, api_key: str, model: str = "gpt-3.5-turbo") -> str:
|
457 |
-
"""Send prompt to OpenAI API
|
458 |
try:
|
459 |
-
from openai import OpenAI
|
460 |
-
|
461 |
-
# Initialize the client with the API key
|
462 |
client = OpenAI(api_key=api_key)
|
463 |
-
|
464 |
-
# Create chat completion using new format
|
465 |
response = client.chat.completions.create(
|
466 |
model=model,
|
467 |
messages=[
|
@@ -473,17 +545,37 @@ def send_to_openai(prompt: str, api_key: str, model: str = "gpt-3.5-turbo") -> s
|
|
473 |
top_p=0.95
|
474 |
)
|
475 |
|
476 |
-
# Extract the response content
|
477 |
if response.choices and len(response.choices) > 0:
|
478 |
return response.choices[0].message.content
|
479 |
else:
|
480 |
-
|
481 |
|
482 |
except ImportError:
|
483 |
-
|
484 |
except Exception as e:
|
485 |
logging.error(f"OpenAI API error: {e}")
|
486 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
487 |
|
488 |
def copy_text_js(element_id: str) -> str:
|
489 |
return f"""function() {{
|
|
|
89 |
"llama-3.1-70b-specdec": 131072,
|
90 |
"llama-3.2-1b-preview": 131072,
|
91 |
"llama-3.2-3b-preview": 131072,
|
92 |
+
},
|
93 |
+
"Cohere API": {
|
94 |
+
"command-r-plus-08-2024": 131072, # 128k
|
95 |
+
"command-r-plus-04-2024": 131072,
|
96 |
+
"command-r-plus": 131072,
|
97 |
+
"command-r-08-2024": 131072,
|
98 |
+
"command-r-03-2024": 131072,
|
99 |
+
"command-r": 131072,
|
100 |
+
"command": 4096,
|
101 |
+
"command-nightly": 131072,
|
102 |
+
"command-light": 4096,
|
103 |
+
"command-light-nightly": 4096,
|
104 |
+
"c4ai-aya-expanse-8b": 8192,
|
105 |
+
"c4ai-aya-expanse-32b": 131072,
|
106 |
}
|
107 |
}
|
108 |
|
|
|
349 |
|
350 |
def send_to_model(prompt, model_selection, hf_model_choice, hf_custom_model, hf_api_key,
|
351 |
groq_model_choice, groq_api_key, openai_api_key, openai_model_choice):
|
352 |
+
"""Wrapper function for send_to_model_impl with comprehensive error handling."""
|
353 |
+
|
354 |
+
logging.info("send to model starting...")
|
355 |
|
356 |
if not prompt or not prompt.strip():
|
357 |
return "Error: No prompt provided", None
|
|
|
360 |
logging.info("sending to model preparation.")
|
361 |
|
362 |
# Basic input validation
|
363 |
+
valid_selections = ["Clipboard only", "HuggingFace Inference", "Groq API", "OpenAI ChatGPT", "Cohere API"]
|
364 |
+
if model_selection not in valid_selections:
|
365 |
return "Error: Invalid model selection", None
|
366 |
|
367 |
+
# Model-specific validation
|
368 |
if model_selection == "Groq API" and not groq_api_key:
|
369 |
return "Error: Groq API key required", None
|
370 |
elif model_selection == "OpenAI ChatGPT" and not openai_api_key:
|
371 |
return "Error: OpenAI API key required", None
|
372 |
+
|
373 |
+
# Call implementation with error handling
|
374 |
+
try:
|
375 |
+
logging.info("calling send_to_model_impl.")
|
376 |
+
summary, download_file = send_to_model_impl(
|
377 |
+
prompt=prompt.strip(),
|
378 |
+
model_selection=model_selection,
|
379 |
+
hf_model_choice=hf_model_choice,
|
380 |
+
hf_custom_model=hf_custom_model,
|
381 |
+
hf_api_key=hf_api_key,
|
382 |
+
groq_model_choice=groq_model_choice,
|
383 |
+
groq_api_key=groq_api_key,
|
384 |
+
openai_api_key=openai_api_key,
|
385 |
+
openai_model_choice=openai_model_choice
|
386 |
+
)
|
387 |
+
logging.info("summary received:", summary)
|
388 |
+
|
389 |
+
if summary is None or not isinstance(summary, str):
|
390 |
+
return "Error: No response from model", None
|
391 |
+
|
392 |
+
return summary, download_file
|
393 |
+
|
394 |
+
except Exception as impl_error:
|
395 |
+
error_msg = str(impl_error)
|
396 |
+
if not error_msg:
|
397 |
+
error_msg = "Unknown error occurred in model implementation"
|
398 |
+
logging.error(f"Model implementation error: {error_msg}")
|
399 |
+
return f"Error: {error_msg}", None
|
400 |
+
|
401 |
+
except Exception as e:
|
402 |
+
error_msg = str(e)
|
403 |
+
if not error_msg:
|
404 |
+
error_msg = "Unknown error occurred"
|
405 |
+
logging.error(f"Error in send_to_model: {error_msg}")
|
406 |
+
return f"Error: {error_msg}", None
|
407 |
+
finally:
|
408 |
+
logging.info("send to model completed.")
|
409 |
|
410 |
def send_to_model_impl(prompt, model_selection, hf_model_choice, hf_custom_model, hf_api_key,
|
411 |
groq_model_choice, groq_api_key, openai_api_key, openai_model_choice):
|
412 |
"""Implementation of model sending with improved error handling."""
|
413 |
logging.info("send to model impl commencing...")
|
414 |
|
|
|
|
|
|
|
415 |
try:
|
416 |
+
if model_selection == "Clipboard only":
|
417 |
+
return "Text copied to clipboard. Use paste for processing.", None
|
418 |
+
|
419 |
if model_selection == "HuggingFace Inference":
|
420 |
+
# First try without API key
|
421 |
model_id = hf_custom_model if hf_model_choice == "Custom Model" else model_registry.hf_models[hf_model_choice]
|
422 |
+
summary = send_to_hf_inference(prompt, model_id)
|
423 |
+
if summary.startswith("Error"):
|
424 |
+
if hf_api_key: # If first try failed and we have an API key, try with it
|
425 |
+
summary = send_to_hf_inference(prompt, model_id, hf_api_key)
|
426 |
|
427 |
elif model_selection == "Groq API":
|
428 |
summary = send_to_groq(prompt, groq_model_choice, groq_api_key)
|
|
|
430 |
elif model_selection == "OpenAI ChatGPT":
|
431 |
summary = send_to_openai(prompt, openai_api_key, model=openai_model_choice)
|
432 |
|
433 |
+
elif model_selection == "Cohere API":
|
434 |
+
summary = send_to_cohere(prompt)
|
435 |
+
|
436 |
else:
|
437 |
return "Error: Invalid model selection", None
|
438 |
|
439 |
# Validate response
|
|
|
|
|
440 |
if not summary or not isinstance(summary, str):
|
441 |
return "Error: Invalid response from model", None
|
442 |
|
|
|
450 |
|
451 |
except Exception as e:
|
452 |
error_msg = str(e)
|
453 |
+
if not error_msg:
|
454 |
error_msg = "Unknown error occurred"
|
455 |
logging.error(f"Error in send_to_model_impl: {error_msg}")
|
456 |
return f"Error: {error_msg}", None
|
457 |
+
|
458 |
def send_to_hf_inference(prompt: str, model_name: str, api_key: str = None) -> str:
|
459 |
+
"""Send prompt to HuggingFace Inference API with optional authentication."""
|
460 |
+
try:
|
461 |
+
client = InferenceClient(token=api_key) if api_key else InferenceClient()
|
462 |
+
response = client.text_generation(
|
463 |
+
prompt,
|
464 |
+
model=model_name,
|
465 |
+
max_new_tokens=500,
|
466 |
+
temperature=0.7,
|
467 |
+
top_p=0.95,
|
468 |
+
repetition_penalty=1.1
|
469 |
+
)
|
470 |
+
return str(response)
|
471 |
+
except Exception as e:
|
472 |
+
logging.error(f"HuggingFace inference error: {e}")
|
473 |
+
return f"Error with HuggingFace inference: {str(e)}" # Return error message instead of raising
|
474 |
+
|
475 |
+
def send_to_hf_inference_old(prompt: str, model_name: str, api_key: str = None) -> str:
|
476 |
"""Send prompt to HuggingFace Inference API with optional authentication."""
|
477 |
try:
|
478 |
# First try without authentication
|
|
|
527 |
return response.choices[0].message.content
|
528 |
except Exception as e:
|
529 |
logging.error(f"Groq API error: {e}")
|
530 |
+
raise # Re-raise to be handled by caller
|
531 |
|
532 |
def send_to_openai(prompt: str, api_key: str, model: str = "gpt-3.5-turbo") -> str:
|
533 |
+
"""Send prompt to OpenAI API."""
|
534 |
try:
|
535 |
+
from openai import OpenAI
|
|
|
|
|
536 |
client = OpenAI(api_key=api_key)
|
|
|
|
|
537 |
response = client.chat.completions.create(
|
538 |
model=model,
|
539 |
messages=[
|
|
|
545 |
top_p=0.95
|
546 |
)
|
547 |
|
|
|
548 |
if response.choices and len(response.choices) > 0:
|
549 |
return response.choices[0].message.content
|
550 |
else:
|
551 |
+
raise Exception("No response generated")
|
552 |
|
553 |
except ImportError:
|
554 |
+
raise Exception("Please install the latest version of openai package (pip install --upgrade openai)")
|
555 |
except Exception as e:
|
556 |
logging.error(f"OpenAI API error: {e}")
|
557 |
+
raise # Re-raise to be handled by caller
|
558 |
+
|
559 |
+
def send_to_cohere(prompt: str, api_key: str = None) -> str:
|
560 |
+
"""Send prompt to Cohere API with optional authentication."""
|
561 |
+
try:
|
562 |
+
import cohere
|
563 |
+
client = cohere.Client(api_key) if api_key else cohere.Client()
|
564 |
+
|
565 |
+
response = client.chat(
|
566 |
+
message=prompt,
|
567 |
+
temperature=0.7,
|
568 |
+
max_tokens=500,
|
569 |
+
)
|
570 |
+
|
571 |
+
if hasattr(response, 'text'):
|
572 |
+
return response.text
|
573 |
+
else:
|
574 |
+
return "Error: No response text from Cohere"
|
575 |
+
|
576 |
+
except Exception as e:
|
577 |
+
logging.error(f"Cohere API error: {e}")
|
578 |
+
return f"Error with Cohere API: {str(e)}" # Return error message instead of raising
|
579 |
|
580 |
def copy_text_js(element_id: str) -> str:
|
581 |
return f"""function() {{
|