SaNSA - the Supercomputer and Node State Architecture Neil Agarwal∗y, Hugh Greenberg∗, Sean Blanchard∗, and Nathan DeBardeleben∗ ∗ Ultrascale Systems Research Center Los Alamos National Laboratory High Performance Computing Los Alamos, New Mexico, 87545 Email: fhng, seanb,
[email protected] y University of California, Berkeley Berkeley, California, 94720 Email:
[email protected] Abstract—In this work we present SaNSA, the Supercomputer vendor, today there exists relatively poor fine-grained informa- and Node State Architecture, a software infrastructure for histor- tion about the state of individual nodes of a cluster. Operators ical analysis and anomaly detection. SaNSA consumes data from and technicians of data centers have dashboards which actively multiple sources including system logs, the resource manager, scheduler, and job logs. Furthermore, additional context such provide real time feedback about coarse-grained information as scheduled maintenance events or dedicated application run such as whether a node is up or down. times for specific science teams can be overlaid. We discuss how The Slurm Workload Manager, which runs on most of the this contextual information allows for more nuanced analysis. U.S. Department of Energy’s clusters, has one view of the SaNSA allows the user to apply arbitrary attributes, for instance, state of the nodes in the system. Slurm uses this to know positional information where nodes are located in a data center. We show how using this information we identify anomalous whether nodes are available for scheduling, jobs are running, behavior of one rack of a 1,500 node cluster. We explain the and when they complete. However, there are other views such design of SaNSA and then test it on four open compute clusters as the node’s own view as expressed via system logs it reports at LANL.