An SSH client is required to connect to all HPC systems. For windows users SmarTTY, MobaXterm and PuTTY are good free clients but there are many others. Mac OS X users can use the builtin Terminal app.
Your account will have been enabled only for the HPC system appropriate for your project(s).
Access to HPC clusters is via jumpbox.tpac.org.au. Once connected to jumpbox please read the “Message of the Day” (MOTD) for cluster status and instructions for how to connect the login node of each cluster.
Notes:
- The TPAC HPC clusters operate in a similar manner to most other HPC clusters in that they use a login node from where jobs are submitted to a scheduler to be then run in the cluster. Jobs found running on the login nodes will be terminated.
- Jobs cannot be submitted from jumpbox only from the login node of the relevant cluster.
Note: Rosalind users should refer to documentation provided separately regarding storage.
For kunanyi home (/u) and scratch (/scratch) directory file systems are provided by a ceph cluster using cephfs. In front of the cluster is a NFS server. Directories are auto mounted so your scratch directory will not be visible until accessed the first time i.e. doing “ls /scratch” may not show your directory.
Quotas are applied to both home and scratch directories. The default quota is 8TiB but can be increased given sufficient reason although we would ask that you do some house keeping before making a request.
Your current usage and quota is printed as part of the message of the day (MOTD) on logging into the kunanyi login node. You can also display this information by running “my-quota-usage”. The data provided is updated approximately every 15 mins.
Transferring data into your HPC home directory
For small amounts of data < 100GB you can use ‘scp’ like tools e.g. WinSCP. Also SmarTTY and MobaXterm both have inbuilt tools to transfer files as well as to make ssh connections. The following linux command will transfer data to your home directory.
scp file.tar.gz {username}@jumpbox.tpac.org.au:~/
Please consider others when transferring data and ensure you are transferring compressed archives (zip, gzip etc.)
For large amounts of data > 100GB please contact us via the UTas Service Catalog.
PBS at TPAC
The Portable Batch Systems (PBS) used at TPAC varies between compute systems.
PBS Pro User Guide
PBS Pro Reference Guide
The following is intended to be generic information for more detailed information refer to the “man” pages or the user documentation above.
For kunanyi example qsub scripts are available in /share/apps/pbs_script_examples
Quick Syntax guide
qstat
|
Standard queue status command supplied by PBS. See man qstat for details of options.
Some common uses are:
|
||||||||||||||||||||||||||
qdel jobid | Delete your unwanted jobs from the queues. The jobid is returned by qsub at jobsubmission time, and is also displayed in the nqstat output. |
||||||||||||||||||||||||||
qsub |
Submit jobs to the queues. The simplest use of the qsub command is typifiedby the following example (note that there is a carriage-return after ./a.out ):
or simply
where
You may need to enter data to the program and may be used to doing this interactively There are two ways of doing this in batch jobs. If, for example, the program requires the numbers 1000 then 50 to be entered when
then run the program as
or the data can be included in the batch job script as follows:
Notice that the PBS directives are all at the start of the script, that there are qsub options of note:
|
||||||||||||||||||||||||||
qps jobid | show the processes of a running job | ||||||||||||||||||||||||||
qls jobid | list the files in a job’s jobfs directory | ||||||||||||||||||||||||||
qcat jobid | show a running job’s stdout, stderr or script | ||||||||||||||||||||||||||
qcp jobid | copy a file from a running job’s jobfs directory |
The man pages for these commands on the system detail the various options you will probably need to use.
On each of the compute systems you can run “module avail” which will provide a list of installed software and their versions. See “module help” for information about how to load modules and other ways to search for modules.
Additional software and or versions can be requested via the UTas Service Catalog.
Requests to TPAC helpdesk can be submitted online via Service Now.
How to get online help
To find a command or library routine that performs a required function, try searching by keyword e.g.
man -k keyword
or
apropos keyword
Use “man command_name” to find details on how to use a unix command or library e.g.
man cat
man ls
If no man page is found, try “info command_name” e.g.
info module
Manipulating files and directories
- ls
List contents of current directory - cd
Change directory - rm
Remove file of directory - mkdir
Make a new directory
Use the “man” command for more information on the above.
A few notes on Unix directory names.
A Unix file full path name is constructed of the directory and subdirectory names separated by slashes “/”. ie. /u/jsmith/work/file1. When you first login you will be in your “home” directory at TPAC this is usually /u/username. In most shells this can also be referenced as ~username.
For example if your username is asmith then “cd ~asmith/work” will take you to the “work” directory in your home directory.
All Unix full path names start with “/”, (There are no Drive/Volume names as in Windows). Hence any filename starting with “/” is a full pathname.
A filename containing one or more slashes “/” will refer to a subdirectory of the “current working directory”. The current working directory may also be referenced as dot “.” i.e. ./subdirectory/file.
The parent of the “current working directory” may be referenced as dot-dot “..”. For example if you have two directories in your home directory work1 and work2 and you cd to work1 you can then change to work2 by typing the command “cd ../work2”.
If the application you are running on a node produces output on STDOUT it is important to ensure that this output is captured to a file in your home directory. If it isn’t redirected it will be captured to a file within the node’s local storage which has limited space. If the local storage file system fills up it may cause the job that is running to terminate early or produce inconsistent data.
It is recommended that you use the -e and -o options when running qsub. To ensure these are not forgotten it would be best to create a shell script to start your job as follows:
qsub_job.sh:
#!/bin/bash
#PBS -e errors.txt
#PBS -o output.txt
cmds
Alternatively the two output streams can be joined together. The following sends the error stream to the output stream:
qsub_job.sh:
#!/bin/bash
#PBS -j oe
#PBS -o output.txt
cmds
It is also possible on kunanyi to see the output and error streams while the job is still running using jobtail.sh and jobcat.sh.
Use “qstat -f” to provide information about the current status of the job.
In your job startup script you should redirect STDOUT to a file in your home directory. This may give you information about what the job is doing depending on your application.
#PBS -l mem=600mb
Because of the unique architecture of eddy cpus and memory are linked. If you specify both it is possible to end up with a job that can’t be run and will remain queued until deleted. When submitting jobs on Eddy specify memory or cpus but not both.
On kunanyi each node has 128GB of RAM and 28 CPUs if you are using whole nodes then there is no need to specify memory. If you are using a portion of a node then specifying memory accurately will allow the schedule to find a suitable node more easily to run your jobs
When specifying an array request to qsub please be aware of the following. The version of PBS running on kunanyi has replaced the “-t” option with “-J”. This change has not been reflected in the qsub man pages which still refer to “-t” but qsub will complain with invalid option ‘-t’ if used.
StarCCM, Matlab and Ansys all provide an X-Windows GUI. To use these applications interactively you will need to run an application on your desktop that provides an X-Server. e.g. SmarTTY, MobaXterm, Mac OS (with XQuartz) or any linux desktop.
With MacOS start a Terminal from within the XQuartz application.
If using SmarTTY or MobaXterm connect to jumpbox.tpac.org.au X-Server is enabled by default if not please refer to the help for the respective application. If using linux or Mac OS X terminal run:
ssh -Y jumpbox.tpac.org.au
The -Y option enables forwarding of X sessions from the remote server.
On jumpbox connect to the kunanyi login node using:
ssh -Y kunanyi
No applications other than compiling software should be run on the login node. If they are all your login sessions will be terminated. You must submit a job to the cluster first using qsub. An example session is as follows:
[kunanyi-01]~% qsub -I -X -l select=1:ncpus=28
qsub: waiting for job 14237.kunanyi-ohpc.tpac.org.au to start
qsub: job 14237.kunanyi-ohpc.tpac.org.au ready
localhost:50.0
[n209]~%
The above will allocate all 28 cpus on one node to your job. You can also use “select=2:ncpu=28” to request 28 cpus on 2 nodes for a total of 56 cpus etc. Run “man qsub” on the login node for more information about the qsub options. After running the qsub command you will now be in a shell running on a compute node. The prompt will change to include “nXXX” where XXX is in the range 002 to 256. You can now load the environment module for the application you wish to run e.g.
module load starccm+ module load matlab module load ansys
The above will load the most recent version that is available on the cluster. To see which versions are available run:
module avail starccm+ module avail matlab module avail ansys
Then use the module load command with the full name including the version as shown in the modul avail listing.
Maintaining a High Performance Computing environment is a time consuming and expensive task. Funding requests require evidence that there is a return on those funds. One of the best ways to show this is when research that benefited from these facilities acknowledges their use in publications.
Please consider including a statement similar to the following in your publications:
We would like to acknowledge the use of the high performance computing facilities provided by Digital Research Services, IT Services at the University of Tasmania
Also please let us know via helpdesk@tpac.org.au when and where your work is published.
Recent Comments