-
Notifications
You must be signed in to change notification settings - Fork 309
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Using a partitioned A100 GPU via MIG with device_index and faster_index causing ctranslate2 error #1788
Comments
I can use device_index=[0,1], but this is not normal and it could be the issues with CUDA on your machine. Could you test multiple GPUs with other examples, apart from CTranslate2, to confirm? Also, try running |
Thanks for the reply and sorry for the delay. I couldn't get it to work at all. "Multiple GPUs" work in general, but to make it work with faster whisper I seem to have to only run one GPU per script rather than use device_index with multiple partitioned GPUs. I'm currently using two V100s instead. Did you manage to get it working with MIG or with multiple hardware GPUs? Thanks |
I tested on my server - 2 GPUs V100s with the option device_index=[0,1]. the model can be loaded on 2 GPUs. |
Ah ok, thanks but as I said in my original post, I've previously used it like this with multiple hardware GPUs so I knew that worked. It's when using it with MIG partitioning on an A100 (which allows up to 7 "virtual" GPUs) that I'm struggling. |
Ah sorry for the confusion. This case hasn't been tested before, so it's understandable that it doesn't work. Unfortunately, I don't have any immediate suggestions. |
No worries, I'm using V100s now instead. Hopefully this can be worked out at some point though as MIG is very useful (although obviously it can be worked around by running multiple instances of a script instead of using device_index). Thanks again. |
Hi,
I asked this a couple of days ago on the faster-whisper git but given it's a ctranslate2 error, I thought I better post it here too (apologies for the duplication but the original question is at SYSTRAN/faster-whisper#1018
I've previously used faster_whisper with device_index with multiple GPUs, but I'm currently using an A100 and I've partitioned it with MIG so it has 7 partitions (that may be too many but I need to test that first).
Using device_index doesn't seem to work. I'm getting a ctranslate2 error.
nvidia_smi is showing
I was hoping I'd be able to put device_index=[0,1,2..] but that doesn't work and neither does using 7,8,9...
ie
model = WhisperModel(model_size, device="cuda",device_index=[0,1],compute_type="bfloat16")
Is this possible?
The error is:
Many thanks
John
The text was updated successfully, but these errors were encountered: