First things first is to connect to WUSM-secure wifi or ethernet.
While working off campus, you need to use a VPN. Download and install the Cisco AnyConnect app.
Connecting address is msvpn.wusm.wustl.edu. Use your wustl key and password for authentication.
HTCF can be logged onto from your local computer on a mac or linux by using the pre-installed terminal application.
- To open a terminal window. You can either search for "terminal" in the search bar or can navigate to the application by opening a finder window and then navigating to Applications > Utilities > Terminal.
- Type the ssh command in the terminal, using your assigned username.
ssh <your_wustl_id>@login.htcf.wustl.edu
- When prompted, enter
<your_wustl_password>. This will be the same password you use with your wustl key.
You are now in a HTCF login shell. Do NOT execute any heavy computation in the login shell. Instead, there are two ways to execute your computation jobs listed below.
Use one of these two methods for executing computation:
- Interactive session: Interactive sessions are useful for quick tasks, troubleshooting, or developing code. To start an interactive job on a single node for a job requiring 8G of memory, use the following command. You may change the memory quota for heavy computation.
srun --mem=8G --cpus-per-task=1 -J interactive -p interactive --pty /bin/bash -l
- SBATCH job: Submit a SBATCH job to the SLURM scheduler. Once your job exits the queue it will run in the background. This is useful when your task needs many resources (e.g. multiple threads, time). More information is provided in the documentation.
To submit an SBATCH job requiring a single node and 8G of memory that will take less than 6 hours, you must first create a script following this template:
#!/bin/bash
#SBATCH --job-name=my_job ## name job to make identification in the queue easier
#SBATCH --nodes=1 ## the number of nodes needed to run the job
#SBATCH --mem=8gb ## the total memory required to complete the job. suffixes specify units (e.g., G = gigabyte). if memory exceeds this the job will be cancelled.
#SBATCH --time=06:00:00 ## the max amount of time your job will run for. if it exceeds this time it will be cancelled.
#SBATCH --output=my_job.out ## name of file containing standard output streams
#SBATCH --error=my_job.err ## name of file containing standard error streams
date ## this will print the time the job started in my_job.out
hostname ## this will print the nodes the job is running on
[insert your code]
date ## this will print the time the job finished
If the script created above is called "my_script.sbatch", then submit your job script as follows:
sbatch my_script.sbatch
After submission, check your the job progress.
squeue -u $USER
(Adapted from HTCF documentation by Emma Johnson and Yiming Kang)