faster-whisper / whisper.py
ankush13r's picture
enable faster whisper
e000cc1
raw
history blame
571 Bytes
from faster_whisper import WhisperModel
import torch
device = "cuda" if torch.cuda.is_available() else "cpu"
torch_dtype = "float32"
MODEL_NAME = "Systran/faster-whisper-large-v3"
model = WhisperModel(MODEL_NAME, compute_type=torch_dtype)
def generate(audio_path):
#check audio lenght
segments, _ = model.transcribe(
audio_path,
# language="ca",
# chunk_length=30,
task="transcribe",
word_timestamps=False,
)
text = ""
for segment in segments:
text += " " + segment.text.strip()
return text