use the fastest whisper implementation on some hardware. Used by transcribee.
Currently, this package can dispatch to (in order of preference):
- mlx-whisper (On Apple Silicon)
- faster-whisper (Everything else)
from decent_whiser import available_models, transcribe
from decent_whisper.model import choose_model, download_model, is_model_downloaded
model_info = choose_model(
available_models(),
model_size="small",
)
if not model_info:
raise ValueError("No matching model found")
if not is_model_downloaded(model_info):
download_model(model_info)
iter, info = transcribe(
"audio.mp3",
model=model_info,
)
for segment in iter:
print("".join([segment.word for segment in segment]))