iAkashPaul commited on
Commit
d67c495
·
verified ·
1 Parent(s): 6ed0ea6

Create server.py

Browse files
Files changed (1) hide show
  1. server.py +41 -0
server.py ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from pydantic import BaseModel,Json
2
+ from fastapi import FastAPI,Request,Response
3
+ from fastapi.middleware.cors import CORSMiddleware
4
+ import requests
5
+
6
+
7
+ app = FastAPI()
8
+ origins = ["*"]
9
+
10
+ app.add_middleware(
11
+ CORSMiddleware,
12
+ allow_origins=origins,
13
+ allow_credentials=True,
14
+ allow_methods=["*"],
15
+ allow_headers=["*"],
16
+ )
17
+
18
+ class Item(BaseModel):
19
+ transcript: Json
20
+
21
+ @app.get("/")
22
+ def base():
23
+ return "PROXY ONLINE"
24
+
25
+ prefix_prompt="""<s>[INST]Summarize the following transcript[/INST]\n"""
26
+ suffix_prompt="""\n"""
27
+ @app.post("/llm")
28
+ async def main(request: Request):
29
+ input_json = await request.json()
30
+ print(input_json)
31
+ final_prompt = prefix_prompt + str(input_json["prompt"]) + suffix_prompt
32
+ resp = requests.post(
33
+ "https://YOUR_MODEL_ID.api.baseten.co/production/predict",
34
+ headers={"Authorization": "Api-Key YOUR_API_KEY"},
35
+ json={'prompt': final_prompt ,'temperature': 0.001, 'max_new_tokens': 100, 'repetition_penalty':1.2},
36
+ )
37
+ llm_response = resp.json()
38
+ llm_response = llm_response.rsplit("[/INST]")[-1].split("</s>")[0];
39
+ print(llm_response)
40
+ return {"text":str(llm_response)}
41
+