Translating slurm and lsf-hpc jobids, 7 using the bhist command (long output), Example 9-7 – HP XC System 3.x Software User Manual

Page 83

Advertising
background image

Example 9-7

Using the bhist Command (Long Output)

$ bhist -l 24

Job <24>, User <lsfadmin>, Project <default>,

Interactive pseudo-terminal shell mode,

Extsched <SLURM[nodes=4]>, Command </bin/bash>

date and time stamp: Submitted from host <n2>,

to Queue <normal>, CWD <$HOME>,

4 Processors Requested, Requested Resources <type=any>;

date and time stamp: Dispatched to 4 Hosts/Processors

<4*lsfhost.localdomain>;

date and time stamp: slurm_id=22;ncpus=8;slurm_alloc=n[5-8];

date and time stamp: Starting (Pid 4785);

Summary of time in seconds spent in various states by

date and time stamp

PEND PSUSP RUN USUSP SSUSP UNKWN TOTAL

11 0 124 0 0 0 135

Translating SLURM and LSF-HPC JOBIDs

LSF-HPC and SLURM are independent resource management components of the HP XC system. They maintain
their own job identifiers (JOBIDs). It may be useful to be able to determine which the SLURM_JOBID
environment variable matches an LSF JOBID, and vice versa.

When a job is submitted to LSF-HPC, it is given an LSF JOBID, as in this example:

$ bsub -o %J.out -n 8 sleep 300

Job <99> is submitted to default queue <normal>

The following is the sequence of events when a SLURM JOBID is assigned:.

No SLURM_JOBID exists while the job is PENDing in LSF-HPC.

After LSF-HPC determines that the resources are available in SLURM for this job, LSF-HPC requests an
allocation in SLURM.

After the SLURM allocation is established, there is a corresponding SLURM JOBID for the LSF JOBID.

Use the bjobs command to view the SLURM JOBID:

$ bjobs -l 99 | grep slurm

date and time stamp: slurm_id=123;ncpus=8;slurm_alloc=n[13-16];

The SLURM JOBID is 123 for the LSF JOBID 99.

You can also find the allocation information in the output of the bhist command:

$ bhist -l 99 | grep slurm

date and time stamp: slurm_id=123;ncpus=8;slurm_alloc=n[13-16];

When LSF-HPC creates an allocation in SLURM, it constructs a name for the allocation by combining the LSF
cluster name with the LSF-HPC JOBID. You can see this name with the scontrol and sacct commands
while the job is running:

$ scontrol show job | grep Name

Name=hptclsf@99

$ sacct -j 123

Jobstep Jobname Partition Ncpus Status Error

---------- ------------------ ---------- ------- ---------- -----

123 hptclsf@99 lsf 8 RUNNING 0

123.0 hptclsf@99 lsf 0 RUNNING 0 

In these examples, the jobname is hptclsf@99; the LSF job ID is 99.

Note that the scontrol show job command keeps jobs briefly after they finish, then it purges itself; this
is similar with the bjobs command. The sacct command continues to provide job information after the
job has finished; this is similar to bhist command:

Using LSF-HPC

83

Advertising