How can I simulate real-time streaming transcription using OpenAI API? #2307
Unanswered
Santoshchodipilli
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm working on a project where I want to convert speech to text in real-time using OpenAI's Whisper model. I see that Whisper's hosted API (whisper-1) currently only supports batch mode — sending a full audio file and receiving the full transcript.
I'm trying to achieve a streaming-like transcription experience, where I can start receiving partial transcriptions as audio is still being recorded or uploaded.
Is there a way to simulate streaming transcription using Whisper?
I'm using Python.
I considered chunking the audio into small parts and sending them sequentially.
Is that the best approach, or is there a better method?
Also, is there any public roadmap or timeline for when the official OpenAI Whisper API might support real-time streaming transcription?
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions