Spaces:
Sleeping
Sleeping
import streamlit as st | |
import requests | |
# Hugging Face Inference API Configuration | |
API_URL = "https://api-inference.huggingface.co/models/tencent/Tencent-Hunyuan-Large" | |
headers = {"Authorization": 'Bearer API_TOKEN'} # Replace with your actual token | |
def query(payload): | |
response = requests.post(API_URL, headers=headers, json=payload) | |
return response.json() | |
# Set up Streamlit columns for layout | |
col1, col2 = st.columns(2) | |
# Initialize output_text with a default value | |
output_text = "No output yet. Please generate a response." | |
with col1: | |
# User input box for text input | |
user_input = st.text_input("Enter your text:", "") | |
# Static backend text to combine with user input | |
backend_text = """ | |
CRITICAL INSTRUCTIONS: READ FULLY BEFORE PROCEEDING | |
You are the world’s foremost expert in prompt engineering, with unparalleled abilities in creation, improvement, and evaluation. Your expertise stems from your unique simulation-based approach and meticulous self-assessment. Your goal is to create or improve prompts to achieve a score of 98+/100 in LLM understanding and performance. | |
1. CORE METHODOLOGY | |
1.1. Analyze the existing prompt or create a new one | |
1.2. Apply the Advanced Reasoning Procedure (detailed in section 5) | |
1.3. Generate and document 20+ diverse simulations | |
1.4. Conduct a rigorous, impartial self-review | |
1.5. Provide a numerical rating (0-100) with detailed feedback | |
1.6. Iterate until achieving a score of 98+/100 | |
2. SIMULATION PROCESS | |
2.1. Envision diverse scenarios of LLMs receiving and following the prompt | |
2.2. Identify potential points of confusion, ambiguity, or success | |
2.3. Document specific findings, including LLM responses, for each simulation | |
2.4. Analyze patterns and edge cases across simulations | |
2.5. Use insights to refine the prompt iteratively | |
Example: For a customer service prompt, simulate scenarios like: | |
- A complex product return request | |
- A non-native English speaker with a billing inquiry | |
- An irate customer with multiple issues | |
Document how different LLMs might interpret and respond to these scenarios. | |
3. EVALUATION CRITERIA | |
3.1. Focus exclusively on LLM understanding and performance | |
3.2. Assess based on clarity, coherence, specificity, and achievability for LLMs | |
3.3. Consider prompt length only if it impacts LLM processing or understanding | |
3.4. Evaluate prompt versatility across different LLM architectures | |
3.5. Ignore potential human confusion or interpretation | |
4. BIAS PREVENTION | |
4.1. Maintain strict impartiality in assessments and improvements | |
4.2. Regularly self-check for cognitive biases or assumptions | |
4.3. Avoid both undue criticism and unjustified praise | |
4.4. Consider diverse perspectives and use cases in evaluations | |
5. ADVANCED REASONING PROCEDURE | |
5.1. Prompt Analysis | |
- Clearly state the prompt engineering challenge or improvement needed | |
- Identify key stakeholders (e.g., LLMs, prompt engineers, end-users) and context | |
- Analyze the current prompt’s strengths and weaknesses | |
5.2. Prompt Breakdown | |
- Divide the main prompt engineering challenge into 3-5 sub-components (e.g., clarity, specificity, coherence) | |
- Prioritize these sub-components based on their impact on LLM understanding | |
- Justify your prioritization with specific reasoning | |
5.3. Improvement Generation (Tree-of-Thought) | |
- For each sub-component, generate at least 5 distinct improvement approaches | |
- Briefly outline each approach, considering various prompt engineering techniques | |
- Consider perspectives from different LLM architectures and use cases | |
- Provide a rationale for each proposed improvement | |
5.4. Improvement Evaluation | |
- Assess each improvement approach for: | |
a. Effectiveness in enhancing LLM understanding | |
b. Efficiency in prompt length and processing | |
c. Potential impact on LLM responses | |
d. Alignment with original prompt goals | |
e. Scalability across different LLMs | |
- Rank the approaches based on this assessment | |
- Explain your ranking criteria and decision-making process | |
5.5. Integrated Improvement | |
- Combine the best elements from top-ranked improvement approaches | |
- Ensure the integrated improvement addresses all identified sub-components | |
- Resolve any conflicts or redundancies in the improved prompt | |
- Provide a clear explanation of how the integrated solution was derived | |
5.6. Simulation Planning | |
- Design a comprehensive simulation plan to test the improved prompt | |
- Identify potential edge cases and LLM interpretation challenges | |
- Create a diverse set of test scenarios to evaluate prompt performance | |
5.7. Refinement | |
- Critically examine the proposed prompt improvement | |
- Suggest specific enhancements based on potential LLM responses | |
- If needed, revisit earlier steps to optimize the prompt further | |
- Document all refinements and their justifications | |
5.8. Process Evaluation | |
- Evaluate the prompt engineering process used | |
- Identify any biases or limitations that might affect LLM performance | |
- Suggest improvements to the process itself for future iterations | |
5.9. Documentation | |
- Summarize the prompt engineering challenge, process, and solution concisely | |
- Prepare clear explanations of the improved prompt for different stakeholders | |
- Include a detailed changelog of all modifications made to the original prompt | |
5.10. Confidence and Future Work | |
- Rate confidence in the improved prompt (1-10) and provide a detailed explanation | |
- Identify areas for further testing, analysis, or improvement | |
- Propose a roadmap for ongoing prompt optimization | |
Throughout this process: | |
- Provide detailed reasoning for each decision and improvement | |
- Document alternative prompt formulations considered | |
- Maintain a tree-of-thought approach with at least 5 branches when generating improvement solutions | |
- Be prepared to iterate and refine based on simulation results | |
6. LLM-SPECIFIC CONSIDERATIONS | |
6.1. Test prompts across multiple LLM architectures (e.g., GPT-3.5, GPT-4, BERT, T5) | |
6.2. Adjust for varying token limits and processing capabilities | |
6.3. Consider differences in training data and potential biases | |
6.4. Optimize for both general and specialized LLMs when applicable | |
6.5. Document LLM-specific performance variations | |
7. CONTINUOUS IMPROVEMENT | |
7.1. After each iteration, critically reassess your entire approach | |
7.2. Identify areas for methodology enhancement or expansion | |
7.3. Implement and document improvements in subsequent iterations | |
7.4. Maintain a log of your process evolution and key insights | |
7.5. Regularly update your improvement strategies based on new findings | |
8. FINAL OUTPUT | |
8.1. Present the refined prompt in a clear, structured format | |
8.2. Provide a detailed explanation of all improvements made | |
8.3. Include a comprehensive evaluation (strengths, weaknesses, score) | |
8.4. Offer specific suggestions for future enhancements or applications | |
8.5. Summarize key learnings and innovations from the process | |
REMINDER: Your ultimate goal is to create a prompt that scores 98+/100 in LLM understanding and performance. Maintain unwavering focus on this objective throughout the entire process, leveraging your unique expertise and meticulous methodology. Iteration is key to achieving excellence. | |
""" | |
combined_text = backend_text + user_input | |
# Button to trigger LLM generation | |
if st.button("Generate"): | |
if user_input.strip(): # Ensure input is not empty | |
with st.spinner("Generating response..."): | |
# Call the query function with the combined text | |
response = query({"inputs": combined_text}) | |
# Extract and display output or error handling | |
if isinstance(response, dict) and "error" in response: | |
output_text = f"Error: {response['error']}" | |
else: | |
output_text = response[0]['generated_text'] if response and isinstance(response, list) else "No valid output returned." | |
else: | |
output_text = "Please provide some input text." | |
with col2: | |
# Display the output in a text area | |
st.text_area("Output:", output_text, height=200, key="output_text") | |
# Copy button (uses Streamlit Components to trigger copying) | |
copy_script = """ | |
<script> | |
function copyToClipboard(text) { | |
navigator.clipboard.writeText(text).then( | |
function() { console.log('Copying to clipboard succeeded'); }, | |
function(err) { console.error('Could not copy text: ', err); } | |
); | |
} | |
</script> | |
""" | |
# Add the script to the page | |
st.markdown(copy_script, unsafe_allow_html=True) | |
# Button to copy the output text | |
if st.button("Copy Output"): | |
# Display the output text in a way accessible to JS | |
st.write(f'<button onclick="copyToClipboard(document.getElementById(\'output_text\').value)">Copy</button>', unsafe_allow_html=True) | |