AAA- Quick Start Guide

Status
Yes

Logging in:

To login to the clusters, the easiest way is with 'ssh'.  On Windows, open a PowerShell - on the Mac, open a terminal window (under Applications-Utilities).

ssh your-username@pod.cnsi.ucsb.edu   or the other clusters, eg. ssh your-username@braid2.cnsi.ucsb.edu

Also, note you will need to be on campus, or, if you're off campus, you'll need to use the VPN

 

Running jobs

We use Slurm on the clusters to submit and run jobs.  Basically you need to create a simple script file to tell the compute node what to do.  A simple example is here (assuming the code you want to execute is named 'a.out' and you want to load the Intel 19 compilers via a module)

 


#!/bin/bash 
#SBATCH -N 1 --cores=1 
cd $SLURM_SUBMIT_DIR 
module load intel/19 
./a.out 

there's quite a few Slurm variables you can set - see this expanded example


#!/bin/bash
#SBATCH -N 1 --cores=1
#SBATCH --time=01:00:00   #  Set the CPU time limit - here, as an hour
#SBATCH --job-name="testslurm"
#SBATCH --mail-type=begin    # Send me an email when the job starts
#SBATCH --mail-user=myemailaddress@ucsb.edu  # where to email!

### next few only useful on Braid2
#SBATCH --mem=1500G             #  How much RAM - if you need the 3TB node, instead of one (and using 'largemem' partition)

#SBATCH --constraint="Intel"    #  Can constrain to Intel or AMD types of CPUs.  Only use if necessary (braid2/largemem).

cd $SLURM_SUBMIT_DIR

module load intel/19


./a.out

A few other special Slurm settings:

For GPU jobs, you need to 'reserve' a GPU.

#SBATCH --gres=gpu:a40:4  (on braid2...on pod, just --gres=gpu:1 - they are all the same type of GPUs)

Note that you can request memory, say you want the higher memory 'batch' nodes (on braid2 - on Pod, all nodes have same, 192GB, of RAM)

#SBATCH --mem=400G