Difference between revisions of "FurtherTopics/FurtherInteractive"
From HPC
m (→More Information) |
(→Increase CPU resources) |
||
| Line 3: | Line 3: | ||
===Increase CPU resources=== | ===Increase CPU resources=== | ||
| + | To increase CPU resources use the flag ''-n<Number of cores>''. | ||
<pre style="background-color: #000000; color: white; border: 2px solid black; font-family: monospace, sans-serif;"> | <pre style="background-color: #000000; color: white; border: 2px solid black; font-family: monospace, sans-serif;"> | ||
[username@login01 ~]$ interactive -n24 | [username@login01 ~]$ interactive -n24 | ||
Revision as of 09:33, 10 November 2022
Contents
Node Resources
By default the interactive' command will give you an allocation to a single compute core on a node for 12 hours and a standard 4GB of RAM.
Increase CPU resources
To increase CPU resources use the flag -n<Number of cores>.
[username@login01 ~]$ interactive -n24 salloc: Granted job allocation 306849 Job ID 306849 connecting to c174, please wait...
Increase Memory
If a job exceeds the requested about of memory, it will terminate with an error message.
[username@login01 ~]$ interactive --mem=24G salloc: Granted job allocation 306852 Job ID 306852 connecting to c068, please wait...
Node Reservations
This example is for a reservation of 327889 and the partition (queue) GPU, missing the partition name will default to the compute queue.
[username@login01 ~]$ interactive -pgpu --reservation=327889 salloc: Granted job allocation 306353 Job ID 306353 connecting to gpu04, please wait...
More Information
More information can be found by typing the following:
[username@login01 ~]$ interactive --help
Parallel run options:
-A, --account=name charge job to specified account
--begin=time defer job until HH:MM MM/DD/YY
--bell ring the terminal bell when the job is allocated
--bb=<spec> burst buffer specifications
--bbf=<file_name> burst buffer specification file
-c, --cpus-per-task=ncpus number of cpus required per task
--comment=name arbitrary comment
--cpu-freq=min[-max[:gov]] requested cpu frequency (and governor)
-d, --dependency=type:jobid defer job until condition on jobid is satisfied
-D, --chdir=path change working directory
--get-user-env used by Moab. See srun man page.
--gid=group_id group ID to run job as (user root only)
--gres=list required generic resources
-H, --hold submit job in held state
-I, --immediate[=secs] exit if resources not available in "secs"
--jobid=id specify jobid to use
-J, --job-name=jobname name of job
-k, --no-kill do not kill job on node failure
-K, --kill-command[=signal] signal to send terminating job
-L, --licenses=names required license, comma separated
-m, --distribution=type distribution method for processes to nodes
(type = block|cyclic|arbitrary)
--mail-type=type notify on state change: BEGIN, END, FAIL or ALL
--mail-user=user who to send email notification for job state
changes
-n, --tasks=N number of processors required
--nice[=value] decrease scheduling priority by value
--no-bell do NOT ring the terminal bell
--ntasks-per-node=n number of tasks to invoke on each node
-N, --nodes=N number of nodes on which to run (N = min[-max])
-O, --overcommit overcommit resources
--power=flags power management options
--priority=value set the priority of the job to value
--profile=value enable acct_gather_profile for detailed data
value is all or none or any combination of
energy, lustre, network or task
-p, --partition=partition partition requested
--qos=qos quality of service
-Q, --quiet quiet mode (suppress informational messages)
--reboot reboot compute nodes before starting job
-s, --share share nodes with other jobs
--sicp If specified, signifies job is to receive
job id from the incluster reserve range.
--signal=[B:]num[@time] send signal when time limit within time seconds
--switches=max-switches{@max-time-to-wait}
Optimum switches and max time to wait for optimum
-S, --core-spec=cores count of reserved cores
--thread-spec=threads count of reserved threads
-t, --time=minutes time limit
--time-min=minutes minimum time limit (if distinct)
--uid=user_id user ID to run job as (user root only)
-v, --verbose verbose mode (multiple -v's increase verbosity)
--wckey=wckey wckey to run job under
Constraint options:
--contiguous demand a contiguous range of nodes
-C, --constraint=list specify a list of constraints
-F, --nodefile=filename request a specific list of hosts
--mem=MB minimum amount of real memory
--mincpus=n minimum number of logical processors (threads)
per node
--reservation=name allocate resources from named reservation
--tmp=MB minimum amount of temporary disk
-w, --nodelist=hosts... request a specific list of hosts
-x, --exclude=hosts... exclude a specific list of hosts
Consumable resources related options:
--exclusive[=user] allocate nodes in exclusive mode when
cpu consumable resource is enabled
--mem-per-cpu=MB maximum amount of real memory per allocated
cpu required by the job.
--mem >= --mem-per-cpu if --mem is specified.
Affinity/Multi-core options: (when the task/affinity plugin is enabled)
-B --extra-node-info=S[:C[:T]] Expands to:
--sockets-per-node=S number of sockets per node to allocate
--cores-per-socket=C number of cores per socket to allocate
--threads-per-core=T number of threads per core to allocate
each field can be 'min' or wildcard '*'
total cpus requested = (N x S x C x T)
--ntasks-per-core=n number of tasks to invoke on each core
--ntasks-per-socket=n number of tasks to invoke on each socket
Help options:
-h, --help show this help message
-u, --usage display brief usage message
Other options:
-V, --version output version information and exit