How can I simulate real-time streaming transcription using OpenAI API? #2307
Unanswered
Santoshchodipilli
asked this question in
Q&A
Replies: 1 comment 2 replies
-
You're on the right path of emulating streaming transcription with Whisper — that's the best workaround available at the moment, TheOpenAI's whisper-1 API is only capable of batch processing and not streaming. let me describe it for you -
Method: Chunked Streaming Simulation
If you're interested, I can help you set up a full real-time transcription. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm working on a project where I want to convert speech to text in real-time using OpenAI's Whisper model. I see that Whisper's hosted API (whisper-1) currently only supports batch mode — sending a full audio file and receiving the full transcript.
I'm trying to achieve a streaming-like transcription experience, where I can start receiving partial transcriptions as audio is still being recorded or uploaded.
Is there a way to simulate streaming transcription using Whisper?
I'm using Python.
I considered chunking the audio into small parts and sending them sequentially.
Is that the best approach, or is there a better method?
Also, is there any public roadmap or timeline for when the official OpenAI Whisper API might support real-time streaming transcription?
Thanks in advance!
Beta Was this translation helpful? Give feedback.
All reactions