Spaces:
Sleeping
Sleeping
Create server.py
Browse files
server.py
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from pydantic import BaseModel,Json
|
2 |
+
from fastapi import FastAPI,Request,Response
|
3 |
+
from fastapi.middleware.cors import CORSMiddleware
|
4 |
+
import requests
|
5 |
+
|
6 |
+
|
7 |
+
app = FastAPI()
|
8 |
+
origins = ["*"]
|
9 |
+
|
10 |
+
app.add_middleware(
|
11 |
+
CORSMiddleware,
|
12 |
+
allow_origins=origins,
|
13 |
+
allow_credentials=True,
|
14 |
+
allow_methods=["*"],
|
15 |
+
allow_headers=["*"],
|
16 |
+
)
|
17 |
+
|
18 |
+
class Item(BaseModel):
|
19 |
+
transcript: Json
|
20 |
+
|
21 |
+
@app.get("/")
|
22 |
+
def base():
|
23 |
+
return "PROXY ONLINE"
|
24 |
+
|
25 |
+
prefix_prompt="""<s>[INST]Summarize the following transcript[/INST]\n"""
|
26 |
+
suffix_prompt="""\n"""
|
27 |
+
@app.post("/llm")
|
28 |
+
async def main(request: Request):
|
29 |
+
input_json = await request.json()
|
30 |
+
print(input_json)
|
31 |
+
final_prompt = prefix_prompt + str(input_json["prompt"]) + suffix_prompt
|
32 |
+
resp = requests.post(
|
33 |
+
"https://YOUR_MODEL_ID.api.baseten.co/production/predict",
|
34 |
+
headers={"Authorization": "Api-Key YOUR_API_KEY"},
|
35 |
+
json={'prompt': final_prompt ,'temperature': 0.001, 'max_new_tokens': 100, 'repetition_penalty':1.2},
|
36 |
+
)
|
37 |
+
llm_response = resp.json()
|
38 |
+
llm_response = llm_response.rsplit("[/INST]")[-1].split("</s>")[0];
|
39 |
+
print(llm_response)
|
40 |
+
return {"text":str(llm_response)}
|
41 |
+
|