The two features we are interested in for GPU scheduling are cuda and bigcuda. Nodes with small GPUs (gt740) have a feature of cuda. Nodes with big GPUs (k20, titan) have a feature of bigcuda. This way, you can constrain your job to an approprately sized GPU.
|title||Run the job file SmallTraining.sh against any gpu available in the cluster|
sbatch --qos=work --gres=gpu SmallTraining.sh
|title||Run the job file LargeTraining.sh against large memory GPUs|
sbatch --qos=work --gres=gpu --constraint=bigcuda LargeTraining.sh
|Content by Label|
|cql||label in ("slurm","gpu","cuda") and type = "page" and space = "rc"|
|labels||GPU Cuda slurm|