Reputation: 2193
I am using slurm to access GPU resources. Is it possible to show GPU usage for a running slurm job? Just like using nvidia-smi
in a normal interactive shell.
Upvotes: 6
Views: 13059
Reputation: 89
I think most of the users, don't have the permission to compute-node, I mean ssh node-1
Here is the way in slurm,
squeue -u <your_username>
and you will get the jobid for that.
srun --jobid=123456 nvidia-smi
Upvotes: 5
Reputation: 1600
I suggest trying to launch your application manually in jupyter and access the terminal shell in jupyter.
Upvotes: 0
Reputation: 51
you can use ssh to login your job's node. Then use nvidia-smi. It works for me. For example, I use squeue check my job xxxxxx is current running at node x-x-x. Then I use ssh x-x-x to access to that node. After that, you can use nvidia-smi to check the usage of GPUs.
Upvotes: 5