Slurmprolog
and epilog
Starting with AWS ParallelCluster version 3.6.0, the Slurm configuration that's deployed with AWS ParallelCluster
includes the Prolog
and Epilog
configuration parameters:
# PROLOG AND EPILOG Prolog=/opt/slurm/etc/scripts/prolog.d/* Epilog=/opt/slurm/etc/scripts/epilog.d/* SchedulerParameters=nohold_on_prolog_fail BatchStartTimeout=180
For more information, see the Prolog and Epilog Guide
AWS ParallelCluster includes the following prolog and epilog scripts:
-
(in the90_plcuster_health_check_manager
Prolog
folder) -
(in the90_pcluster_noop
Epilog
folder)
Note
Both the Prolog
and Epilog
folder must contain at least one file.
You can use your own custom prolog
or epilog
scripts by adding them to
the corresponding Prolog
and Epilog
folders.
Warning
Slurm runs every script in the folders, in reverse alphabetical order.
The run time duration of the prolog
and epilog
scripts impact the time needed to run a job.
Update the BatchStartTimeout
configuration setting when running multiple or long running
prolog
scripts. The default is 3 minutes.
If you are using custom prolog
and epilog
scripts, locate the scripts in the respective
Prolog
and Epilog
folders.
We recommend that you keep the 90_plcuster_health_check_manager
script that runs before every custom script.
For more information, see Slurm configuration customization.