Slower performance on Whisper model #269
-
Hi! I noticed a quite significant difference in performance, specifically Whisper Large model. Usually, a 1 hour audio will be done in 1-2 minutes, but now it's about 10-15 minutes. Thank you. |
Beta Was this translation helpful? Give feedback.
Replies: 5 comments 20 replies
-
Whisper is currently having some issues. You can keep an eye on our status page to see when it's resolved. |
Beta Was this translation helpful? Give feedback.
-
Hey @jjmaldonis, we too have been noticing that Whisper has been slowing down quite a bit in the past month. A month ago, it used to process 1-2 minute audio files in about 10 seconds, which was a bit slower than OpenAI's Whisper at 6-7 seconds, but it was manageable. Lately, for the same audio it may take between 40 seconds to 1.5 minutes, and sometimes even longer than 10 minutes. The response times have become unpredictable. Here are a couple of examples of request ids with extended wait times: d83d7966-437b-4198-9423-d2e2d6571c0f (more than 2 minutes) Do you know when this might be fixed? Can we expect Whisper will be more stable and fast in the near future? We're trying to figure out our game plan while we wait it to get fixed. Thanks! |
Beta Was this translation helpful? Give feedback.
-
guys, did you notice any improvement in speed? |
Beta Was this translation helpful? Give feedback.
-
Hey all, early last week we improved the underlying systems that run Whisper to support the growing demand from our users. You should have seen significant improvements in the number of errors received and the latency of API requests to Whisper. We will continue to improve the underlying infrastructure over the coming months, and may introduce additional rate limits if Whisper latency remains higher than desired. Many of our users rely on Deepgram to serve voice-to-text solutions to their own customer base. When Whisper latency is high (and request take longer than normal to complete), those customers can see a poorer user experience because the audio files they upload seem to "hang" for an unknown amount of time. If necessary, additional rate limits will improve the end customer's experience by providing a near-instantaneous message saying, "this API request cannot be fulfilled in a reasonable amount of time", which allows products that rely on Deepgram to handle this scenario in a way that meets their customers' expectations. If you are using or plan to use thousands of dollars of Whisper requests per month, please keep an eye out for additional rate limits; you may also want to reach out to our sales team to continue the conversation. |
Beta Was this translation helpful? Give feedback.
-
Whisper is ridiculously slow. ~45 seconds for a 1min audio file, where OpenAI processes in ~6 sec |
Beta Was this translation helpful? Give feedback.
Hey all, early last week we improved the underlying systems that run Whisper to support the growing demand from our users. You should have seen significant improvements in the number of errors received and the latency of API requests to Whisper.
We will continue to improve the underlying infrastructure over the coming months, and may introduce additional rate limits if Whisper latency remains higher than desired.
Many of our users rely on Deepgram to serve voice-to-text solutions to their own customer base. When Whisper latency is high (and request take longer than normal to complete), those customers can see a poorer user experience because the audio files they upload seem to "hang" for …