Skip to content

Instantly share code, notes, and snippets.

@stevenjswanson
Last active August 21, 2023 05:49
Show Gist options
  • Save stevenjswanson/fab9ce8510804919fafe36804fd417f6 to your computer and use it in GitHub Desktop.
Save stevenjswanson/fab9ce8510804919fafe36804fd417f6 to your computer and use it in GitHub Desktop.
slurm-stack_c1.0.8ebgoxao0lxa@142dev | ---> Starting the MUNGE Authentication service (munged) ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | * Starting MUNGE munged
slurm-stack_c1.0.8ebgoxao0lxa@142dev | ...done.
slurm-stack_c1.0.8ebgoxao0lxa@142dev | ---> Waiting for slurmctld to become active before starting slurmd...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is not available. Sleeping ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is not available. Sleeping ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is not available. Sleeping ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is not available. Sleeping ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is not available. Sleeping ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is not available. Sleeping ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | -- slurmctld is now active ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | ---> Starting the Slurm Node Daemon (slurmd) ...
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: Log file re-opened
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: hwloc_topology_init
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: hwloc_topology_load
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: hwloc_topology_export_xml
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: CPUs:16 Boards:1 Sockets:1 CoresPerSocket:8 ThreadsPerCore:2
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: error: Node configuration differs from hardware: CPUs=1:16(hw) Boards=1:1(hw) SocketsPerBoard=1:1(hw) CoresPerSocket=1:8(hw) ThreadsPerCore=1:2(hw)
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: hwloc_topology_init
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: xcpuinfo_hwloc_topo_load: xml file (/var/spool/slurmd/hwloc_topo_whole.xml) found
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: CPUs:16 Boards:1 Sockets:1 CoresPerSocket:8 ThreadsPerCore:2
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: topology/none: init: topology NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: route/default: init: route default plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Gathering cpu frequency information for 16 cpus
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: Resource spec: No specialized cores configured by default on this node
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: Resource spec: Reserved system memory limit not configured for this node
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: task/none: init: task NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: auth/munge: init: Munge authentication plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: spank: opening plugin stack /etc/slurm/plugstack.conf
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: /etc/slurm/plugstack.conf: 1: include "/etc/slurm/plugstack.conf.d/*.conf"
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: cred/munge: init: Munge credential signature plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: slurmd version 21.08.5 started
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: jobacct_gather/linux: init: Job accounting gather LINUX plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: job_container/none: init: job_container none plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: switch/none: init: switch NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: switch Cray/Aries plugin loaded.
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: slurmd started on Mon, 21 Aug 2023 05:46:03 +0000
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: CPUs=1 Boards=1 Sockets=1 Cores=1 Threads=1 Memory=31952 TmpDisk=448557 Uptime=2970117 CPUSpecList=(null) FeaturesAvail=(null) FeaturesActive=(null)
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: acct_gather_energy/none: init: AcctGatherEnergy NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: acct_gather_Profile/none: init: AcctGatherProfile NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: acct_gather_interconnect/none: init: AcctGatherInterconnect NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: acct_gather_filesystem/none: init: AcctGatherFilesystem NONE plugin loaded
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No acct_gather.conf file (/etc/slurm/acct_gather.conf)
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: _handle_node_reg_resp: slurmctld sent back 8 TRES.
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Start processing RPC: REQUEST_LAUNCH_TASKS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Processing RPC: REQUEST_LAUNCH_TASKS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: launch task StepId=1.0 request from UID:0 GID:0 HOST:10.0.39.3 PORT:36176
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: Checking credential with 420 bytes of sig data
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: _insert_job_state: we already have a job state for job 1. No big deal, just an FYI.
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Finish processing RPC: REQUEST_LAUNCH_TASKS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Start processing RPC: REQUEST_TERMINATE_JOB
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Processing RPC: REQUEST_TERMINATE_JOB
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: _rpc_terminate_job: uid = 990 JobId=1
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: credential for job 1 revoked
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No steps in jobid 1 to send signal 18
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No steps in jobid 1 to send signal 15
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: set revoke expiration for jobid 1 to 1692596920 UTS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Finish processing RPC: REQUEST_TERMINATE_JOB
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Start processing RPC: REQUEST_LAUNCH_TASKS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Processing RPC: REQUEST_LAUNCH_TASKS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: launch task StepId=3.0 request from UID:0 GID:0 HOST:10.0.39.4 PORT:56740
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: Checking credential with 420 bytes of sig data
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: _insert_job_state: we already have a job state for job 3. No big deal, just an FYI.
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Finish processing RPC: REQUEST_LAUNCH_TASKS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Start processing RPC: REQUEST_KILL_TIMELIMIT
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Processing RPC: REQUEST_KILL_TIMELIMIT
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No steps in jobid 3 to send signal 996
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No steps in jobid 3 to send signal 15
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: Job 3: timeout: sent SIGTERM to 0 active steps
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: _rpc_terminate_job: uid = 990 JobId=3
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: credential for job 3 revoked
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No steps in jobid 3 to send signal 18
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: No steps in jobid 3 to send signal 15
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: set revoke expiration for jobid 3 to 1692597002 UTS
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug: JobId=3: sent epilog complete msg: rc = 0
slurm-stack_c1.0.8ebgoxao0lxa@142dev | slurmd: debug2: Finish processing RPC: REQUEST_KILL_TIMELIMIT
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment