Update app.py
Browse files
app.py
CHANGED
@@ -2748,46 +2748,7 @@ def chunk_transcript(transcript: str, chunk_duration: int, words_per_second) ->
|
|
2748 |
return chunks
|
2749 |
|
2750 |
|
2751 |
-
|
2752 |
-
words_per_second: int) -> str:
|
2753 |
-
if api_name not in summarizers: # See 'summarizers' dict in the main script
|
2754 |
-
return f"Unsupported API: {api_name}"
|
2755 |
-
|
2756 |
-
if not transcript:
|
2757 |
-
logging.error("Empty or None transcript provided to summarize_chunks")
|
2758 |
-
return "Error: Empty or None transcript provided"
|
2759 |
-
|
2760 |
-
text = extract_text_from_segments(transcript)
|
2761 |
-
chunks = chunk_transcript(text, chunk_duration, words_per_second)
|
2762 |
-
|
2763 |
-
custom_prompt = args.custom_prompt
|
2764 |
-
|
2765 |
-
summaries = []
|
2766 |
-
for chunk in chunks:
|
2767 |
-
if api_name == 'openai':
|
2768 |
-
# Ensure the correct model and prompt are passed
|
2769 |
-
summaries.append(summarize_with_openai(api_key, chunk, custom_prompt))
|
2770 |
-
elif api_name == 'anthropic':
|
2771 |
-
summaries.append(summarize_with_cohere(api_key, chunk, anthropic_model, custom_prompt))
|
2772 |
-
elif api_name == 'cohere':
|
2773 |
-
summaries.append(summarize_with_claude(api_key, chunk, cohere_model, custom_prompt))
|
2774 |
-
elif api_name == 'groq':
|
2775 |
-
summaries.append(summarize_with_groq(api_key, chunk, groq_model, custom_prompt))
|
2776 |
-
elif api_name == 'llama':
|
2777 |
-
summaries.append(summarize_with_llama(llama_api_IP, chunk, api_key, custom_prompt))
|
2778 |
-
elif api_name == 'kobold':
|
2779 |
-
summaries.append(summarize_with_kobold(kobold_api_IP, chunk, api_key, custom_prompt))
|
2780 |
-
elif api_name == 'ooba':
|
2781 |
-
summaries.append(summarize_with_oobabooga(ooba_api_IP, chunk, api_key, custom_prompt))
|
2782 |
-
elif api_name == 'tabbyapi':
|
2783 |
-
summaries.append(summarize_with_vllm(api_key, tabby_api_IP, chunk, summarize.llm_model, custom_prompt))
|
2784 |
-
elif api_name == 'local-llm':
|
2785 |
-
summaries.append(summarize_with_local_llm(chunk, custom_prompt))
|
2786 |
-
else:
|
2787 |
-
return f"Unsupported API: {api_name}"
|
2788 |
-
|
2789 |
-
return "\n\n".join(summaries)
|
2790 |
-
|
2791 |
# FIXME - WHole section needs to be re-written
|
2792 |
def summarize_with_detail_openai(text, detail, verbose=False):
|
2793 |
summary_with_detail_variable = rolling_summarize(text, detail=detail, verbose=True)
|
|
|
2748 |
return chunks
|
2749 |
|
2750 |
|
2751 |
+
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2752 |
# FIXME - WHole section needs to be re-written
|
2753 |
def summarize_with_detail_openai(text, detail, verbose=False):
|
2754 |
summary_with_detail_variable = rolling_summarize(text, detail=detail, verbose=True)
|