SpeakSmart / app.py
m-adil-ali's picture
Rename untitled.py to app.py
31f453f verified
import os
os.environ["GROQ_API_KEY"] = "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"
import os
import gradio as gr
import whisper
from gtts import gTTS
import io
from transformers import pipeline
from groq import Groq
# Initialize the Groq client
client = Groq(api_key=os.environ.get("GROQ_API_KEY"))
# Load the Whisper model
whisper_model = whisper.load_model("base") # You can choose other models like "small", "medium", "large"
# Initialize the grammar correction pipeline
corrector = pipeline("text2text-generation", model="pszemraj/flan-t5-large-grammar-synthesis")
def process_audio(file_path):
try:
# Load the audio file
audio = whisper.load_audio(file_path)
# Transcribe the audio using Whisper
result = whisper_model.transcribe(audio)
user_text = result["text"]
# Display the user input text
corrected_text = corrector(user_text)[0]['generated_text'].strip()
# Generate a response using Groq
chat_completion = client.chat.completions.create(
messages=[{"role": "user", "content": corrected_text}],
model="llama3-8b-8192", # Replace with the correct model if necessary
)
# Access the response using dot notation
response_message = chat_completion.choices[0].message.content.strip()
# Convert the response text to speech
tts = gTTS(response_message)
response_audio_io = io.BytesIO()
tts.write_to_fp(response_audio_io) # Save the audio to the BytesIO object
response_audio_io.seek(0)
# Save audio to a file to ensure it's generated correctly
with open("response.mp3", "wb") as audio_file:
audio_file.write(response_audio_io.getvalue())
# Return the original text, corrected text, and the path to the saved audio file
return user_text, corrected_text, "response.mp3"
except Exception as e:
return f"An error occurred: {e}", None, None
# Create a Gradio interface with a submit button
iface = gr.Interface(
fn=process_audio,
inputs=gr.Audio(type="filepath"), # Use type="filepath"
outputs=[
gr.Textbox(label="User voice input into text"), # Original user input text
gr.Textbox(label="Corrected version of user input"), # Corrected text
gr.Audio(label="Response Audio") # Response audio
],
live=False, # Ensure live mode is off to use a submit button
title="Audio Processing with Grammar Correction",
description="Upload an audio file, which will be transcribed, corrected for grammar, and then used to generate a response.",
allow_flagging="never"
)
iface.launch()
# except Exception as e:
# return f"An error occurred: {e}", None, None
# iface = gr.Interface(
# fn=process_audio,
# inputs=gr.Audio(type="filepath"), # Use type="filepath"
# outputs=[
# gr.Textbox(label="User voice input into text"), # Original user input text
# gr.Textbox(label="Corrected version of user input"), # Corrected text
# gr.Audio(label="Response Audio") # Response audio
# ],
# live=True
# )
# iface.launch()