Summary: | exact TRES allocations/affinities are not stored anywhere | ||
---|---|---|---|
Product: | Slurm | Reporter: | Sergey Meirovich <sergey_meirovich> |
Component: | Accounting | Assignee: | Unassigned Developer <dev-unassigned> |
Status: | RESOLVED DUPLICATE | QA Contact: | |
Severity: | 5 - Enhancement | ||
Priority: | --- | ||
Version: | 19.05.x | ||
Hardware: | Linux | ||
OS: | Linux | ||
Site: | AMAT | Alineos Sites: | --- |
Atos/Eviden Sites: | --- | Confidential Site: | --- |
Coreweave sites: | --- | Cray Sites: | --- |
DS9 clusters: | --- | HPCnow Sites: | --- |
HPE Sites: | --- | IBM Sites: | --- |
NOAA SIte: | --- | NoveTech Sites: | --- |
Nvidia HWinf-CS Sites: | --- | OCF Sites: | --- |
Recursion Pharma Sites: | --- | SFW Sites: | --- |
SNIC sites: | --- | Tzag Elita Sites: | --- |
Linux Distro: | --- | Machine Name: | |
CLE Version: | Version Fixed: | ||
Target Release: | --- | DevPrio: | --- |
Emory-Cloud Sites: | --- |
Description
Sergey Meirovich
2019-01-25 15:30:49 MST
Hi Sergey Meirovich, This is an interesting idea but at this time there are no current plans to tackle these changes. Hello Jason, Thanks for you answer. Could you please look into second part of my question? "... So even if SchedMD is not planning to look into this in a foreseeable future it would be at least interesting to understand where: slurm accounting db?/elastic search completion plugin?/somewhere else it seems logical to SchedMD to store information about exact CPU/GPU assignments." ? Hi Sergey Meirovich, > "... So even if SchedMD is not planning to look into this in a foreseeable future it would be at least interesting to understand where: slurm accounting db?/elastic search completion plugin?/somewhere else it seems logical to SchedMD to store information about exact CPU/GPU assignments." Slurm does not currently document task placement / gpu placement in the accounting database. It does give an overview of what was used. e.g. jason@nh-grey:~/slurm/master$ sacct -j 280 -o JobID,AllocGRES,AllocCPUS JobID AllocGRES AllocCPUS ------------ ------------ ---------- 280 gpu:0 2 280.batch gpu:0 2 280.extern gpu:0 2 Note that you can run "scontrol show job -d <job_id>", and query some more information in that output.. ... Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=* Nodes=m1 CPU_IDs=0-1 Mem=0 GRES=gpu(IDX:0) ... An epilogctld, "EpilogSlurmctld, may be able to capture this job's comment after the job is done. https://slurm.schedmd.com/prolog_epilog.html Rolling this into one ticket. *** This ticket has been marked as a duplicate of ticket 2047 *** |