from inference import Inference import os model_path = os.getenv("MODEL_PATH", "saved_model/pytorch_model.bin") tokenizer_path = os.getenv("TOKENIZER_PATH", "saved_tokenizer") inference = Inference(model_path, tokenizer_path) def handler(event, context): prompt = event["data"]["prompt"] max_length = event["data"].get("max_length", 100) response = inference.predict(prompt, max_length) return {"response": response}