|
""" |
|
Use this file to test how a model does with the system prompts, with no additional context. |
|
""" |
|
|
|
|
|
from openai import OpenAI |
|
import configparser |
|
from hundred_system_prompts import * |
|
|
|
|
|
|
|
config = configparser.ConfigParser() |
|
config.read('config.ini') |
|
client = OpenAI(api_key = config['openai']['api_key']) |
|
model_name = config['openai'].get('model', fallback='gpt-4') |
|
|
|
|
|
filename = f"{model_name}.txt" |
|
with open(filename, "w") as file: |
|
file.write("") |
|
|
|
|
|
for i, (prompt, probe, lambda_function) in enumerate(system_prompts): |
|
history = [ |
|
{"role": "system", "content": prompt}, |
|
{"role": "user", "content": probe} |
|
] |
|
|
|
completion = client.chat.completions.create( |
|
model = model_name, |
|
messages = history, |
|
temperature = 1, |
|
) |
|
response = completion.choices[0].message.content |
|
score = lambda_function(response) |
|
print(f"{i+1}: {score}") |
|
with open(filename, "a", encoding='utf-8') as file: |
|
stripped_prompt = prompt.replace('\n', '') |
|
stripped_response = response.replace('\n', '') |
|
file.write(f"{i+1}: {score}\n") |
|
file.write(f"\tprompt: {stripped_prompt}\n") |
|
file.write(f"\tresponse: {stripped_response}\n") |
|
|