Spaces:
Sleeping
Sleeping
# Confabulation_check.py | |
# | |
# This file contains the functions that are used to check the confabulation of the user's input. | |
# | |
# | |
# Imports | |
# | |
# External Imports | |
# | |
# Local Imports | |
# | |
# | |
#################################################################################################### | |
# | |
# Functions: | |
from App_Function_Libraries.Chat import chat_api_call | |
from App_Function_Libraries.ms_g_eval import validate_inputs, detailed_api_error | |
def simplified_geval(transcript: str, summary: str, api_name: str, api_key: str, temp: float = 0.7) -> str: | |
""" | |
Perform a simplified version of G-Eval using a single query to evaluate the summary. | |
Args: | |
transcript (str): The original transcript | |
summary (str): The summary to be evaluated | |
api_name (str): The name of the LLM API to use | |
api_key (str): The API key for the chosen LLM | |
temp (float, optional): The temperature parameter for the API call. Defaults to 0.7. | |
Returns: | |
str: The evaluation result | |
""" | |
try: | |
validate_inputs(transcript, summary, api_name, api_key) | |
except ValueError as e: | |
return str(e) | |
prompt = f"""You are an AI assistant tasked with evaluating the quality of a summary. You will be given an original transcript and a summary of that transcript. Your task is to evaluate the summary based on the following criteria: | |
1. Coherence (1-5): How well-structured and organized is the summary? | |
2. Consistency (1-5): How factually aligned is the summary with the original transcript? | |
3. Fluency (1-3): How well-written is the summary in terms of grammar, spelling, and readability? | |
4. Relevance (1-5): How well does the summary capture the important information from the transcript? | |
Please provide a score for each criterion and a brief explanation for your scoring. Then, give an overall assessment of the summary's quality. | |
Original Transcript: | |
{transcript} | |
Summary to Evaluate: | |
{summary} | |
Please provide your evaluation in the following format: | |
Coherence: [score] - [brief explanation] | |
Consistency: [score] - [brief explanation] | |
Fluency: [score] - [brief explanation] | |
Relevance: [score] - [brief explanation] | |
Overall Assessment: [Your overall assessment of the summary's quality] | |
""" | |
try: | |
result = chat_api_call( | |
api_name, | |
api_key, | |
prompt, | |
"", | |
temp=temp, | |
system_message="You are a helpful AI assistant tasked with evaluating summaries." | |
) | |
except Exception as e: | |
return detailed_api_error(api_name, e) | |
formatted_result = f""" | |
Confabulation Check Results: | |
{result} | |
""" | |
return formatted_result |