slurm
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
slurm [2021/01/19 09:37] – kauffman | slurm [2025/01/13 12:45] (current) – [Where to begin] amcguire | ||
---|---|---|---|
Line 16: | Line 16: | ||
==== Discord ==== | ==== Discord ==== | ||
- | There is a dedicated text channel '' | + | There is a dedicated text channel '' |
===== Clusters ===== | ===== Clusters ===== | ||
+ | |||
+ | We have a couple different clusters. If you don't know where to start please use the '' | ||
* [[slurm: | * [[slurm: | ||
Line 25: | Line 27: | ||
==== Peanut Cluster ==== | ==== Peanut Cluster ==== | ||
- | Think of these machines as a dumping ground for discrete computing tasks that might be rude or disruptive to execute on the main (shared) shell servers (i.e., | + | Think of these machines as a dumping ground for discrete computing tasks that might be rude or disruptive to execute on the main (shared) shell servers (i.e., |
Additionally, | Additionally, | ||
Line 37: | Line 39: | ||
===== Where to begin ===== | ===== Where to begin ===== | ||
- | Slurm is a set of command line utilities that can be accessed via the command line from **most** any computer science system you can login to. Using our main shell servers (linux.cs.uchicago.edu) is expected to be our most common use case, so you should start there. | + | Slurm is a set of command line utilities that can be accessed via the command line from **most** any computer science system you can login to. Using our main shell servers ('' |
ssh user@linux.cs.uchicago.edu | ssh user@linux.cs.uchicago.edu | ||
- | If you want to use the AI Cluster you will need to login into: | + | If you want to use the AI Cluster you will need to have previously requested access by sending in a ticket. Afterwards, you may login into: |
ssh user@fe.ai.cs.uchicago.edu | ssh user@fe.ai.cs.uchicago.edu | ||
Line 345: | Line 347: | ||
Please make sure you specify $CUDA_HOME and if you want to take advantage of CUDNN libraries you will need to append / | Please make sure you specify $CUDA_HOME and if you want to take advantage of CUDNN libraries you will need to append / | ||
- | cuda_version=9.2 | + | cuda_version=11.1 |
export CUDA_HOME=/ | export CUDA_HOME=/ | ||
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH: | export LD_LIBRARY_PATH=$LD_LIBRARY_PATH: | ||
Line 361: | Line 363: | ||
The variable name is actually misleading; since it does NOT mean the amount of devices, but rather the physical device number assigned by the kernel (e.g. / | The variable name is actually misleading; since it does NOT mean the amount of devices, but rather the physical device number assigned by the kernel (e.g. / | ||
- | For example: If you requested multiple gpu's from Slurm (--gres=gpu: | + | For example: If you requested multiple gpu's from Slurm (--gres=gpu: |
+ | |||
+ | The numbering is relative and specific to you. For example: two users with one job which require two gpus each could be assigned non-sequential gpu numbers. However CUDA_VISIBLE_DEVICES will look like this for both users: 0,1 | ||
Line 410: | Line 415: | ||
STDOUT will look something like this: | STDOUT will look something like this: | ||
< | < | ||
- | cnetid@linux1:~$ cat $HOME/ | + | cnetid@focal0:~$ cat $HOME/ |
Device Number: 0 | Device Number: 0 | ||
Device name: Tesla M2090 | Device name: Tesla M2090 |
/var/lib/dokuwiki/data/attic/slurm.1611070630.txt.gz · Last modified: 2021/01/19 09:37 by kauffman