User Tools

Site Tools


howto:submissions
This version is outdated by a newer approved version.DiffThis version (2018/08/24 09:08) is a draft.
Approvals: 0/1

This is an old revision of the document!


Submit a cluster job from the GUI

Most SPHIRE jobs require heavy computing and are most suited for a cluster environment, although all SPHIRE pipeline commands now support a workstation (or a single node system) since Beta Patch 1 release (sphire_beta_20170602).

Since every cluster environment is a little bit different and everybody has a particular preference for job schedulers we cannot provide support for all the possible cluster configurations. Nevertheless, SPHIRE's GUI can create a submission script for you, given that you provide simple template script. Currently, these are the variables than can be parsed by the GUI:

Variable Description
XXX_SXMPI_NPROC_XXX Defines the total number of cores to be used.
XXX_SXMPI_JOB_NAME_XXX Defines the name of the job. This can be used to name the stdout and stderr and will define the name of the submission script created.
XXX_SXCMD_LINE_XXX This will be replaced by the actual command to run SPHIRE.

For instance, if you want to submit a 2D clustering job (ISAC) from the GUI, you should configure the MPI section of like below: SPHIRE GUI

This will create a file called qsub_sxisac_test.sh with the correct setup and submit it to the cluster.


Example Scripts

Please contact your system administrator in order to create an appropriate template in case your cluster uses a different job scheduler.

Example scripts can be downloaded here: QSUB EXAMPLES.

Son of Grid Engine (SGE)

Below you can find one example of the kind of submission templates that we would use in our local cluster.

#!/bin/bash
set echo on

#$ -pe mpi_fillup XXX_SXMPI_NPROC_XXX
#$ -N XXX_SXMPI_JOB_NAME_XXX
#$ -cwd

# Sets up the environment for SPHIRE. 
source /work/software/Sphire/EMAN2/eman2.bashrc
MPIRUN=$(which mpirun)

#Creates a file with the nodes used for the calculation. It can be given to mpirun, but 
#with the current setup it is not necessary.
set hosts = `cat $PE_HOSTFILE | cut -f 1 -d . | sort | fmt -w 30 | sed 's/ /,/g'`
cat $PE_HOSTFILE | awk '{print $1}'  > hostfile.$JOB_ID

${MPIRUN} -np XXX_SXMPI_NPROC_XXX XXX_SXCMD_LINE_XXX 

If you use this script and define the variable XXX_SXMPI_JOB_NAME_XXX as above (sxisac_test), then the stderr and stdout will be written to files called sxisac_test.e and sxisac_test.o respectively.

PBS/Torque

Below you can find one example of the kind of submission templates that we would be used for PBS/Torque queueing system.

#!/bin/csh
#PBS -N XXX_SXMPI_JOB_NAME_XXX
#PBS -l nodes=8:ppn=23

module load sphire

cd $PBS_O_WORKDIR
MPIRUN=$(which mpirun)

${MPIRUN} -np XXX_SXMPI_NPROC_XXX --hostfile $PBS_NODEFILE XXX_SXCMD_LINE_XXX

In this case, the user would provide 184 CPU in the GUI.

8 * 23 = 184

Slurm

Below you can find one example of the kind of submission templates that would be used for Slurm queueing system.

#!/bin/zsh
#SBATCH --partition all
#SBATCH --time 7-00:00
#SBATCH --ntasks XXX_SXMPI_NPROC_XXX
#SBATCH --output XXX_SXMPI_JOB_NAME_XXX.log
#SBATCH --job-name XXX_SXMPI_JOB_NAME_XXX

module load sphire
mpirun --map-by node -n XXX_SXMPI_NPROC_XXX XXX_SXCMD_LINE_XXX
howto/submissions.1535094495.txt.gz ยท Last modified: 2018/08/24 09:08 by mstabrin