medrag / medrag_multi_modal /assistant /medqa_assistant.py
geekyrakshit's picture
add: MedQAAssistant
49cde8e
raw
history blame
931 Bytes
from typing import Optional
import weave
from PIL import Image
from ..retrieval import SimilarityMetric
from .llm_client import LLMClient
class MedQAAssistant(weave.Model):
llm_client: LLMClient
retriever: weave.Model
top_k_chunks: int = 2
retrieval_similarity_metric: SimilarityMetric = SimilarityMetric.COSINE
@weave.op()
def predict(self, query: str, image: Optional[Image.Image] = None) -> str:
_image = image
retrieved_chunks = self.retriever.predict(
query, top_k=self.top_k_chunks, metric=self.retrieval_similarity_metric
)
retrieved_chunks = [chunk["text"] for chunk in retrieved_chunks]
system_prompt = """
You are a medical expert. You are given a query and a list of chunks from a medical document.
"""
return self.llm_client.predict(
system_prompt=system_prompt, user_prompt=retrieved_chunks
)