FurtherTopics/FurtherInteractive
From HPC
Contents
Node Resources
By default the interactive' command will give you an allocation to a single compute core on a node for 12 hours and a standard 4GB of RAM.
Increase CPU resources
To increase CPU resources use the flag -n<Number of cores>.
[username@login01 ~]$ interactive -n24 salloc: Granted job allocation 306849 Job ID 306849 connecting to c174, please wait...
Increase Memory
For more memory use the flag --mem<amount>G. If a job exceeds the requested about of memory, it will terminate with an error message.
[username@login01 ~]$ interactive --mem=24G salloc: Granted job allocation 306852 Job ID 306852 connecting to c068, please wait...
Node Reservations
This example is for a reservation of 327889 and the partition (queue) GPU, missing the partition name will default to the compute queue.
[username@login01 ~]$ interactive -pgpu --reservation=327889 salloc: Granted job allocation 306353 Job ID 306353 connecting to gpu04, please wait...
More Information
More information can be found by typing the following:
[username@login01 ~]$ interactive --help Parallel run options: -A, --account=name charge job to specified account --begin=time defer job until HH:MM MM/DD/YY --bell ring the terminal bell when the job is allocated --bb=<spec> burst buffer specifications --bbf=<file_name> burst buffer specification file -c, --cpus-per-task=ncpus number of cpus required per task --comment=name arbitrary comment --cpu-freq=min[-max[:gov]] requested cpu frequency (and governor) -d, --dependency=type:jobid defer job until condition on jobid is satisfied -D, --chdir=path change working directory --get-user-env used by Moab. See srun man page. --gid=group_id group ID to run job as (user root only) --gres=list required generic resources -H, --hold submit job in held state -I, --immediate[=secs] exit if resources not available in "secs" --jobid=id specify jobid to use -J, --job-name=jobname name of job -k, --no-kill do not kill job on node failure -K, --kill-command[=signal] signal to send terminating job -L, --licenses=names required license, comma separated -m, --distribution=type distribution method for processes to nodes (type = block|cyclic|arbitrary) --mail-type=type notify on state change: BEGIN, END, FAIL or ALL --mail-user=user who to send email notification for job state changes -n, --tasks=N number of processors required --nice[=value] decrease scheduling priority by value --no-bell do NOT ring the terminal bell --ntasks-per-node=n number of tasks to invoke on each node -N, --nodes=N number of nodes on which to run (N = min[-max]) -O, --overcommit overcommit resources --power=flags power management options --priority=value set the priority of the job to value --profile=value enable acct_gather_profile for detailed data value is all or none or any combination of energy, lustre, network or task -p, --partition=partition partition requested --qos=qos quality of service -Q, --quiet quiet mode (suppress informational messages) --reboot reboot compute nodes before starting job -s, --share share nodes with other jobs --sicp If specified, signifies job is to receive job id from the incluster reserve range. --signal=[B:]num[@time] send signal when time limit within time seconds --switches=max-switches{@max-time-to-wait} Optimum switches and max time to wait for optimum -S, --core-spec=cores count of reserved cores --thread-spec=threads count of reserved threads -t, --time=minutes time limit --time-min=minutes minimum time limit (if distinct) --uid=user_id user ID to run job as (user root only) -v, --verbose verbose mode (multiple -v's increase verbosity) --wckey=wckey wckey to run job under Constraint options: --contiguous demand a contiguous range of nodes -C, --constraint=list specify a list of constraints -F, --nodefile=filename request a specific list of hosts --mem=MB minimum amount of real memory --mincpus=n minimum number of logical processors (threads) per node --reservation=name allocate resources from named reservation --tmp=MB minimum amount of temporary disk -w, --nodelist=hosts... request a specific list of hosts -x, --exclude=hosts... exclude a specific list of hosts Consumable resources related options: --exclusive[=user] allocate nodes in exclusive mode when cpu consumable resource is enabled --mem-per-cpu=MB maximum amount of real memory per allocated cpu required by the job. --mem >= --mem-per-cpu if --mem is specified. Affinity/Multi-core options: (when the task/affinity plugin is enabled) -B --extra-node-info=S[:C[:T]] Expands to: --sockets-per-node=S number of sockets per node to allocate --cores-per-socket=C number of cores per socket to allocate --threads-per-core=T number of threads per core to allocate each field can be 'min' or wildcard '*' total cpus requested = (N x S x C x T) --ntasks-per-core=n number of tasks to invoke on each core --ntasks-per-socket=n number of tasks to invoke on each socket Help options: -h, --help show this help message -u, --usage display brief usage message Other options: -V, --version output version information and exit