Skip to content

cuda devices #2

@kullrich

Description

@kullrich

Dear Mesih,

I get an memory allocation error, this is because on our GPU server multiple devices are running and it seems that the code always only uses the first GPU:

torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 26.00 MiB (GPU 0; 10.75 GiB total capacity; 25.19 MiB already allocated; 11.62 MiB free; 42.00 MiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation.  See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF

Could you please provide a @click option to select the cuda devices for pytorch, e.g. cuda:1

Since on other devices the memory is not full and I would like to be able to select the device or even cuda:all

|   0  NVIDIA GeForce ...  On   | 00000000:1A:00.0 Off |                  N/A |
|  0%   22C    P8     1W / 250W |  10421MiB / 11264MiB |      0%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
|   1  NVIDIA GeForce ...  On   | 00000000:1B:00.0 Off |                  N/A |
|  0%   22C    P8    10W / 250W |    343MiB / 11264MiB |      0%      Default |
|                               |                      |                  N/A |

Thank you in anticipation

Best regards

Kristian

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions