import torch import gradio as gr from transformers import pipeline pipe = pipeline(task="automatic-speech-recognition", model="openai/whisper-small", device="cuda" if torch.cuda.is_available() else "cpu") def transcribe(audio): text = pipe(audio)["text"] return text interface = gr.Interface( fn=transcribe, inputs=gr.Audio(sources=["microphone", "upload"], type="filepath"), outputs="text", title="Whisper Small", description="Realtime demo for Speech recognition using a Whisper small model.", ) if __name__ == "__main__": interface.launch()