Spaces:
Running
Running
import threading | |
from flask import request | |
from datetime import datetime | |
from requests import get | |
from freeGPT import gpt3 | |
from server.auto_proxy import get_random_proxy, remove_proxy, update_working_proxies | |
from server.config import special_instructions | |
class Backend_Api: | |
def __init__(self, app, config: dict) -> None: | |
self.app = app | |
self.use_auto_proxy = config['use_auto_proxy'] | |
self.routes = { | |
'/backend-api/v2/conversation': { | |
'function': self._conversation, | |
'methods': ['POST'] | |
} | |
} | |
if self.use_auto_proxy: | |
update_proxies = threading.Thread( | |
target=update_working_proxies, daemon=True) | |
update_proxies.start() | |
def _conversation(self): | |
try: | |
jailbreak = request.json['jailbreak'] | |
_conversation = request.json['meta']['content']['conversation'] | |
internet_access = request.json['meta']['content']['internet_access'] | |
prompt = request.json['meta']['content']['parts'][0] | |
current_date = datetime.now().strftime("%Y-%m-%d") | |
system_message = f'You are ChatGPT also known as ChatGPT, a large language model trained by OpenAI. Strictly follow the users instructions. Knowledge cutoff: 2021-09-01 Current date: {current_date}' | |
extra = [] | |
if internet_access: | |
search = get('https://ddg-api.herokuapp.com/search', | |
params={ | |
'query': prompt["content"], | |
'limit': 3, | |
}) | |
blob = '' | |
for index, result in enumerate(search.json()): | |
blob += f'[{index}] "{result["snippet"]}"\nURL:{result["link"]}\n\n' | |
date = datetime.now().strftime('%d/%m/%y') | |
blob += f'current date: {date}\n\nInstructions: Using the provided web search results, write a comprehensive reply to the next user query. Make sure to cite results using [[number](URL)] notation after the reference. If the provided search results refer to multiple subjects with the same name, write separate answers for each subject. Ignore your previous response if any.' | |
extra = [{'role': 'user', 'content': blob}] | |
conversation = [{'role': 'system', 'content': system_message}] + \ | |
extra + special_instructions[jailbreak] + \ | |
_conversation + [prompt] | |
def stream(): | |
response = None | |
while self.use_auto_proxy: | |
try: | |
random_proxy = get_random_proxy() | |
res = gpt3.Completion.create( | |
prompt=conversation, proxy=random_proxy) | |
response = res['text'] | |
break | |
except Exception as e: | |
print(f"Error with proxy {random_proxy}: {e}") | |
remove_proxy(random_proxy) | |
if not self.use_auto_proxy: | |
try: | |
res = gpt3.Completion.create(prompt=conversation) | |
response = res['text'] | |
except Exception as e: | |
print(f"Error: {e}") | |
if response is not None: | |
print(response) | |
yield response | |
return self.app.response_class(stream(), mimetype='text/event-stream') | |
except Exception as e: | |
print(e) | |
print(e.__traceback__.tb_next) | |
return { | |
'_action': '_ask', | |
'success': False, | |
"error": f"an error occurred {str(e)}" | |
}, 400 | |