Spaces:
Running
Running
Update metagpt/ext/spo/utils/llm_client.py
Browse files
metagpt/ext/spo/utils/llm_client.py
CHANGED
@@ -33,9 +33,9 @@ class SPO_LLM:
|
|
33 |
raise ValueError("'model' parameter is required")
|
34 |
|
35 |
try:
|
36 |
-
model_config = ModelsConfig.default().get(
|
37 |
if model_config is None:
|
38 |
-
raise ValueError(f"Model
|
39 |
|
40 |
config = model_config.model_copy()
|
41 |
|
@@ -83,25 +83,3 @@ def extract_content(xml_string: str, tag: str) -> Optional[str]:
|
|
83 |
return match.group(1).strip() if match else None
|
84 |
|
85 |
|
86 |
-
async def main():
|
87 |
-
# test LLM
|
88 |
-
SPO_LLM.initialize(
|
89 |
-
optimize_kwargs={"model": "gpt-4o", "temperature": 0.7},
|
90 |
-
evaluate_kwargs={"model": "gpt-4o-mini", "temperature": 0.3},
|
91 |
-
execute_kwargs={"model": "gpt-4o-mini", "temperature": 0.3},
|
92 |
-
)
|
93 |
-
|
94 |
-
llm = SPO_LLM.get_instance()
|
95 |
-
|
96 |
-
# test messages
|
97 |
-
hello_msg = [{"role": "user", "content": "hello"}]
|
98 |
-
response = await llm.responser(request_type=RequestType.EXECUTE, messages=hello_msg)
|
99 |
-
logger(f"AI: {response}")
|
100 |
-
response = await llm.responser(request_type=RequestType.OPTIMIZE, messages=hello_msg)
|
101 |
-
logger(f"AI: {response}")
|
102 |
-
response = await llm.responser(request_type=RequestType.EVALUATE, messages=hello_msg)
|
103 |
-
logger(f"AI: {response}")
|
104 |
-
|
105 |
-
|
106 |
-
if __name__ == "__main__":
|
107 |
-
asyncio.run(main())
|
|
|
33 |
raise ValueError("'model' parameter is required")
|
34 |
|
35 |
try:
|
36 |
+
model_config = ModelsConfig.default().get("gpt-4o-mini")
|
37 |
if model_config is None:
|
38 |
+
raise ValueError(f"Model gpt-4o-mini not found in configuration")
|
39 |
|
40 |
config = model_config.model_copy()
|
41 |
|
|
|
83 |
return match.group(1).strip() if match else None
|
84 |
|
85 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|