The job-level data is typically processed daily, with the various scripts run as cron jobs. The workflow is described below:
Before setting up the batch processing workflow, we highly recommend manually running the various scripts in debug mode following the instructions on the Inital Ingest Setup page.
Enable the following script to run everyday via a cron job. It should be executed after the Open XDMoD daily update process is expected to finish.
This script calls indexarchives.py and summarize_jobs.py in turn while providing a locking mechanisms so that processes do not conflict with each other.
The job-level summaries are ingested into Open XDMoD with the following command:
An example cron configuration file is shown below to illustrate a typical setup. The actual content of the cron file will depend on whether or not the summarization software is installed on a different host than the Open XDMoD package. The amount of time taken to run the various scripts depends on many factors mainly the number and size of HPC jobs and the IO bandwidth of the host that runs the scripts.
# Shred and ingest accounting data 0 1 * * * root /usr/bin/xdmod-slurm-helper -q -r resource-name && /usr/bin/xdmod-ingestor -q # Create job level summaries 0 2 * * * root /usr/bin/supremm_update # Ingest job level sumamries into XDMoD and run aggregation 0 4 * * * root /usr/bin/aggregate_supremm.sh