Spaces:
Running
Running
# # This tests the litelm proxy | |
# # it makes async Completion requests with streaming | |
# import openai | |
# openai.base_url = "http://0.0.0.0:8000" | |
# openai.api_key = "temp-key" | |
# print(openai.base_url) | |
# async def test_async_completion(): | |
# response = await ( | |
# model="gpt-3.5-turbo", | |
# prompt='this is a test request, write a short poem', | |
# ) | |
# print(response) | |
# print("test_streaming") | |
# response = await openai.chat.completions.create( | |
# model="gpt-3.5-turbo", | |
# prompt='this is a test request, write a short poem', | |
# stream=True | |
# ) | |
# print(response) | |
# async for chunk in response: | |
# print(chunk) | |
# import asyncio | |
# asyncio.run(test_async_completion()) | |