Source repo: ciml-summer-institute-2024 | Branch:
main| Last synced: 2026-04-24 10:27:17.425 UTC
CIML Summer Institute: Accounts, Login, Environments, Running Jobs, Logging into Expanse User Portal
Session: 1.2_accounts_login_environments_running_jobs_expanse_portal
Date: Tuesday, June 18, 2024
Presented by: Mary Thomas ( mpthomas @ ucsd.edu )
Title: Expanse Webinar: Accessing and Running Jobs on Expanse
Reading and Presentations:
-
Lecture material:
-
Presentation Slides: Access, accounts, modules & Envs
https://github.com/ciml-org/ciml-summer-institute-2024/blob/main/1.2_accounts_login_environment_expanse_portal/CIML-SI24_Jun_18_2024-PrepDay_accts_login_envs_jobs.pdf -
Source Code/Examples:
https://github.com/sdsc-hpc-training-org/hpctr-examples
-
-
Exercises:
-
Exercise 1: Log onto Expanse Use your training account and password to log on, find out what project you are on, and list out your module environment.
quantum:ciml accountname$ ssh -Y train111@login.expanse.sdsc.edu
(train111@login.expanse.sdsc.edu) Password:
Welcome to Bright release 9.0
Based on Rocky Linux 8
ID: #000002
--------------------------------------------------------------------------------
WELCOME TO
_______ __ ____ ___ _ _______ ______
/ ____/ |/ // __ \/ | / | / / ___// ____/
/ __/ | // /_/ / /| | / |/ /\__ \/ __/
/ /___ / |/ ____/ ___ |/ /| /___/ / /___
/_____//_/|_/_/ /_/ |_/_/ |_//____/_____/
--------------------------------------------------------------------------------
Use the following commands to adjust your environment:
'module avail' - show available modules
'module add `module`' - adds a module to your environment for this session
'module initadd `module`' - configure module to be loaded at every login
-------------------------------------------------------------------------------
Last login: Mon Jun 17 19:18:38 2024 from 75.80.45.222
connect /private/tmp/com.apple.launchd.HbagVgBfXZ/org.xquartz:0: Connection refused
[train111@login02 ~]$ -
Exercise 2: Clone the HPC Training GitHub Repo:
[train111@login02 ~]$ git clone https://github.com/sdsc-hpc-training-org/hpctr-examples.git
Cloning into 'hpctr-examples'...
remote: Enumerating objects: 485, done.
remote: Counting objects: 100% (485/485), done.
remote: Compressing objects: 100% (318/318), done.
remote: Total 485 (delta 212), reused 421 (delta 158), pack-reused 0
Receiving objects: 100% (485/485), 27.65 MiB | 27.87 MiB/s, done.
Resolving deltas: 100% (212/212), done.
Updating files: 100% (304/304), done.
[train111@login02 ~]$ ll hpctr-examples/
total 214
drwxr-xr-x 15 train111 gue998 19 Jun 17 22:35 .
drwxr-x--- 10 train111 gue998 20 Jun 17 22:35 ..
drwxr-xr-x 9 train111 gue998 10 Jun 17 22:35 basic_par
drwxr-xr-x 2 train111 gue998 6 Jun 17 22:35 calc-pi
drwxr-xr-x 2 train111 gue998 5 Jun 17 22:35 calc-prime
drwxr-xr-x 6 train111 gue998 7 Jun 17 22:35 cuda
drwxr-xr-x 2 train111 gue998 5 Jun 17 22:35 env_info
-rw-r--r-- 1 train111 gue998 5772 Jun 17 22:35 file-tree.txt
drwxr-xr-x 8 train111 gue998 13 Jun 17 22:35 .git
[snip]
drwxr-xr-x 2 train111 gue998 21 Jun 17 22:35 mpi
drwxr-xr-x 2 train111 gue998 3 Jun 17 22:35 netcdf
drwxr-xr-x 2 train111 gue998 17 Jun 17 22:35 openacc
drwxr-xr-x 2 train111 gue998 8 Jun 17 22:35 openmp
-rw-r--r-- 1 train111 gue998 5772 Jun 17 22:35 README.md
[train111@login02 ~]$ -
Exercise 3: create an interactive CPU node The following example will request one regular compute node, 4 cores, in the debug partition for 30 minutes.
[train111@exp-9-55 ~]$ hostname
login02
[train111@login02 ~]$ srun --partition=debug --pty --account=gue998 --nodes=1 --ntasks-per-node=4 --mem=8G
-t 00:30:00 --wait=0 --export=ALL /bin/bash
srun: job 31426273 queued and waiting for resources
srun: job 31426273 has been allocated resources
[train111@exp-9-55 ~]$ hostname
exp-9-55
[train111@exp-9-55 ~]$ exit
[train111@login02 ~]$ -
Exercise 4: create an interactive GPU node The following example will request a GPU node, 10 cores, 1 GPU and 96G in the debug partition for 30 minutes. To ensure the GPU environment is properly loaded, please be sure run both the module purge and module restore commands.
[train111@login02 ~]$ hostname
login02
login02$ srun --partition=gpu-debug --pty --account=gue998 --ntasks-per-node=10
--nodes=1 --mem=96G --gpus=1 -t 00:30:00 --wait=0 --export=ALL /bin/bash -
Exercise 5: compile the MPI Hello World code.
cd hpctr-examples/mpi
[train111@login02 mpi]$ module purge
[train111@login02 mpi]$ module load slurm
[train111@login02 mpi]$ module load cpu/0.15.4
[train111@login02 mpi]$ module load gcc/10.2.0
[train111@login02 mpi]$ module load openmpi/4.0.4
[train111@login02 mpi]$ mpif90 -o hello_mpi hello_mpi.f90
[train111@login02 mpi]$
[train111@login02 mpi]$ mpicc -o mpi_hello mpi_hello.c
[train111@login02 mpi]$ ll mpi_hello
-rwxr-xr-x 1 train111 gue998 18120 Jun 17 23:22 mpi_hello -
Exercise 6: Log onto the Expanse User Portal: Use your training account and password
https://portal.expanse.sdsc.edu/training
-