Description
Details
- Slurm Version: 23.02.4
- Python Version: 3.6.8
- Cython Version: 3.0.0
- PySlurm Branch: main
- Linux Distribution: RHEL 8.8
Issue
Code that processes the memory of jobs was crashing due to a None being included.
I identified that this job (scontrol output) was causing pyslurm to give a value of "None" for job.memory.
JobId=5909485 JobName=interactive
UserId=xvy5180(5712646) GroupId=xvy5180(5712646) MCS_label=N/A
Priority=300000 Nice=0 Account=mxs2589_e_gpu QOS=normal
JobState=RUNNING Reason=None Dependency=(null)
Requeue=0 Restarts=0 BatchFlag=0 Reboot=0 ExitCode=0:0
RunTime=05:12:33 TimeLimit=10:00:00 TimeMin=N/A
SubmitTime=2023-09-18T10:57:13 EligibleTime=2023-09-18T10:57:13
AccrueTime=2023-09-18T10:57:13
StartTime=2023-09-18T10:57:31 EndTime=2023-09-18T20:57:32 Deadline=N/A
PreemptEligibleTime=2023-09-18T10:57:31 PreemptTime=None
SuspendTime=None SecsPreSuspend=0 LastSchedEval=2023-09-18T10:57:31 Scheduler=Main
Partition=sla-prio AllocNode:Sid=submit02:526250
ReqNodeList=(null) ExcNodeList=(null)
NodeList=p-gc-3003
BatchHost=p-gc-3003
NumNodes=1 NumCPUs=1 NumTasks=1 CPUs/Task=1 ReqB:S:C:T=0:0::
ReqTRES=cpu=1,mem=79488M,node=1,billing=1,gres/gpu=1
AllocTRES=cpu=1,mem=79488M,node=1,billing=1,gres/gpu=1
Socks/Node=* NtasksPerN:B:S:C=0:0:: CoreSpec=*
MinCPUsNode=1 MinMemoryNode=0 MinTmpDiskNode=0
Features=gc DelayBoot=00:00:00
OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null)
Command=/bin/sh
Power=
MemPerTres=gres:gpu:79488
TresPerJob=gres:gpu:1