-
Notifications
You must be signed in to change notification settings - Fork 315
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
update cuda 11.2 to cuda 12.2 #1590
Conversation
Thx for the PR, but I got disappointed by CUDA 12... RTX 3050 GPU [diff vs CUDA 11]:
|
I'm all for keeping backwards compatibility with CUDA 11, if feasible. Not sure if Purfview was suggesting to not pursue CUDA 12 builds altogether...but if that's the case I'd definitely recommend a lot more testing. But in terms of backward compatibility, virtually always a good thing IMHO...coming from a non-professional and hobbyist developer so... (apologize in advance if backwards compatibility wasn't the correct phrase...) |
Thanks @Purfview for sharing information. |
Diff from the tests in a new environment [various optimizations for performance]:
EDIT: |
On my end (and I'm just one amateur developer among professionals), I know that my user base (however small) would appreciate CUDA 11 support for awhile longer at least. Not all computer setups support CUDA 12, python libraries, etc. so to have ctranslate2 only work with one version of CUDA at a time would be harsh. I've noticed PyTorch's policy is to generally advertise two major version support...Maybe that could be a policy of ctranslate2? And then ctranslate2 could have a repository of older builds, easy for user's to understand, showing which version of CUDA are supported up to which version of Ctranslate2...like PyTorch has an "old builds" page. Anyways, I'm excited! Just saw this...IMHO, a year is kinda long to not yet have CUDA 12 support without having to compile from source... |
Does anyone know if this is still being worked on? It was on the verge of being incorporating CUDA 12+ but it's been awhile. |
Get this merged asap! I see no regression on my end. |
Yes, please merge. I don't even think they're talking about removing support for CUDA 11.8, but just adding CUDA 12 support! |
@minhthuc2502 Is it possible to upload this to pypi.org now so that I can "pip install" the newer version that supports CUDA 12? |
Do you have an idea why I get a nice speedup with the small whisper model with Whisper config is the following: temperature: 0
beam_size: 1
condition_on_previous_text: false
vad_filter: false
Conclusion: Main problem is |
@ozancaglayan Use |
Yes I just noticed that but it's again weird that |
It's not weird at all. |
OK I updated the table, |
Try much longer tests, not the few seconds. Btw, are you saying that this inconsistency appeared with CUDA12? |
I'm now repeating the tests with CTranslate < 4 using CUDA11. Inconsistencies are there as well. I'm running each test 5 times on the same audio file of length 5 minutes so I think it's good enough. I'm now counting the length of the texts generated at each run and |
Ar you using a clear speech audio without noise and silence? |
Very clear speech with very silent blocks. But even if I apply VAD, |
I updated my previous table with final final results btw see #1590 (comment) |
I'm sure there is something wrong with your test than something else. |
Did you try to do similar benchmarking with and without |
No description provided.