Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The accelerator {'count': 1, 'kind': 'gpu', 'api': 'cuda', 'brand': 'nvidia'} could not be provided (Singularity container ) #1390

Open
aaannaw opened this issue May 17, 2024 · 3 comments

Comments

@aaannaw
Copy link

aaannaw commented May 17, 2024

Hello, professor
I used the GPU docker container to run cactus: singularity run --nv ./cactus_v2.8.2-gpu.sif cactus ./jobStore ./Evolve.Bathyergidae.txt ./Evolve.Bathyergidae.hal --gpu 1 --workDir /data/groups/lzu_public/home/u220220932211/wn-cactus/tmp --lastzMemory 10G --maxDisk 5T
However, I suffered from warning and error:

WARNING: The NVIDIA Driver was not detected.  GPU functionality will not be available.
   Use the NVIDIA Container Toolkit to start this container with GPU support; see
   https://docs.nvidia.com/datacenter/cloud-native/ .

raise InsufficientSystemResources(requirer, 'accelerators', self.accelerator_identities, details=[
toil.batchSystems.abstractBatchSystem.InsufficientSystemResources: The job 'run_lastz' kind-run_lastz/instance-tzo0sor0 v1 is requesting [{'count': 1, 'kind': 'gpu', 'api': 'cuda', 'brand': 'nvidia'}] accelerators, more than the maximum of [] accelerators that SingleMachineBatchSystem was configured with. The accelerator {'count': 1, 'kind': 'gpu', 'api': 'cuda', 'brand': 'nvidia'} could not be provided. Scale is set to 1.

The problem is similar with #900 , it looks like Toil doesn't believe I am on a machine with a GPU. I add the --nv parameters but it does not work. Could you give me any suggestions?
Thank you for much!

Best wishes!
Na Wan

@glennhickey
Copy link
Collaborator

What happens when you run

singularity run --nv ./cactus_v2.8.2-gpu.sif nvidia-smi

@aaannaw
Copy link
Author

aaannaw commented May 18, 2024

Hello, professor
Thanks for your response. When I run singularity run --nv ./cactus_v2.8.2-gpu.sif nvidia-smi, the eorror occurs:

==========
== CUDA ==
==========

CUDA Version 11.4.3

Container image Copyright (c) 2016-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.

This container image and its contents are governed by the NVIDIA Deep Learning Container License.
By pulling and using the container, you accept the terms and conditions of this license:
https://developer.nvidia.com/ngc/nvidia-deep-learning-container-license

A copy of this license is made available in this container at /NGC-DL-CONTAINER-LICENSE for your convenience.

WARNING: The NVIDIA Driver was not detected.  GPU functionality will not be available.
   Use the NVIDIA Container Toolkit to start this container with GPU support; see
   https://docs.nvidia.com/datacenter/cloud-native/ .

NVIDIA-SMI couldn't find libnvidia-ml.so library in your system. Please make sure that the NVIDIA Display Driver is properly installed and present in your system.
Please also try adding directory that contains libnvidia-ml.so to your system PATH.

However I only run nvidia-smi, it showed normal:

Sat May 18 11:30:34 2024       
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 525.60.13    Driver Version: 525.60.13    CUDA Version: 12.0     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|                               |                      |               MIG M. |
|===============================+======================+======================|
|   0  Tesla V100-PCIE...  Off  | 00000000:41:00.0 Off |                    0 |
| N/A   20C    P0    24W / 250W |      0MiB / 32768MiB |      0%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
|   1  Tesla V100-PCIE...  Off  | 00000000:45:00.0 Off |                    0 |
| N/A   21C    P0    25W / 250W |      0MiB / 32768MiB |      0%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
|   2  Tesla V100-PCIE...  Off  | 00000000:B5:00.0 Off |                    0 |
| N/A   20C    P0    25W / 250W |      0MiB / 32768MiB |      0%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
|   3  Tesla V100-PCIE...  Off  | 00000000:B9:00.0 Off |                    0 |
| N/A   20C    P0    23W / 250W |      0MiB / 32768MiB |      0%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
                                                                               
+-----------------------------------------------------------------------------+
| Processes:                                                                  |
|  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
|        ID   ID                                                   Usage      |
|=============================================================================|
|  No running processes found                                                 |
+-----------------------------------------------------------------------------+

I try to add the path of libnvidia-ml.so to the command"singularity run -B /usr/local/nvidia/lib --nv ./cactus_v2.8.2-gpu.sif cactus ./liutest2 ./Evolve.txt ./Evolve..hal --maxCore=20 --gpu 1 --workDir ./tmpdir --lastzMemory 10G --maxDisk 5T",it loos run normally.

I have another confusion about the parameter "--targettime" and we should how to set it. Could you give me any suggestions?
Thank you very much.
Best wishes!
Na Wan

@glennhickey
Copy link
Collaborator

--tarrgettime doesn't do anything unless you're using the aws autoscaler - just ignore it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants