| Summary: | Handle squeue filtering for --nodelist within the slurmctld to minimize data transfers | ||
|---|---|---|---|
| Product: | Slurm | Reporter: | CSC sysadmins <csc-slurm-tickets> |
| Component: | slurmctld | Assignee: | Unassigned Developer <dev-unassigned> |
| Status: | OPEN --- | QA Contact: | |
| Severity: | 5 - Enhancement | ||
| Priority: | --- | ||
| Version: | 17.02.10 | ||
| Hardware: | Linux | ||
| OS: | Linux | ||
| Site: | CSC - IT Center for Science | Alineos Sites: | --- |
| Atos/Eviden Sites: | --- | Confidential Site: | --- |
| Coreweave sites: | --- | Cray Sites: | --- |
| DS9 clusters: | --- | HPCnow Sites: | --- |
| HPE Sites: | --- | IBM Sites: | --- |
| NOAA SIte: | --- | OCF Sites: | --- |
| Recursion Pharma Sites: | --- | SFW Sites: | --- |
| SNIC sites: | --- | Linux Distro: | --- |
| Machine Name: | CLE Version: | ||
| Version Fixed: | Target Release: | --- | |
| DevPrio: | --- | Emory-Cloud Sites: | --- |
|
Description
CSC sysadmins
2018-04-19 09:50:58 MDT
Hi Tommi - I'd recommend reworking your epilog to avoid that call. We've discussed moving some of this filtering into the slurmctld - but there's an inherent tradeoff in having it work out a lot of this filtering instead of the squeue command itself. If you'd like us to look into that on 18.08+ I can reclassify this as an enhancement, but at the moment the RPCs for 17.02 and 17.11 cannot be changed to accommodate that. I will note that we highly recommend use of proctrack/cgroup and task/cgroup alongside pam_slurm_adopt to avoid the need for this type of node cleanup in an Epilog script. - Tim Hi, I'd guess that on the road towards exascale systems this may need some attention. I was surprised that squeue will transfer also job out/err/script paths and one particular user likes hyper long paths and long argument lists + lot of jobs which caused this problem for us (paths looks more like an ascii art). I'd still like to get a simple tool which tells if node is occupied or not (enable power saving, clean up shm/ipcs which are not in control of cgroup? etc.) Node slurmd does not have that information (count of running jobs overall and per user) ? -Tommi (In reply to Tommi Tervo from comment #2) > Hi, > > I'd guess that on the road towards exascale systems this may need some > attention. > > I was surprised that squeue will transfer also job out/err/script paths and > one particular user likes hyper long paths and long argument lists + lot of > jobs which caused this problem for us (paths looks more like an ascii art). > > I'd still like to get a simple tool which tells if node is occupied or not > (enable power saving, clean up shm/ipcs which are not in control of cgroup? > etc.) > > Node slurmd does not have that information (count of running jobs overall > and per user) ? Not directly. If you're on the node, you can enumerate that (and fetch some additional details) from the slurmstepd processes over their unix sockets. I'm retagging this as an enhancement to look at handling node-filtering on the slurmctld side, although I can't promise when/if we'll tackle this. - Tim |