Rechner-Cluster
Mehr Informationen zu dem Service finden Sie in unserem Dokumentationsportal.
Batch System Controller Unavailable
We are experiencing issues with Slurm's batch controller which causes all Slurm commands (sbatch, squeue etc.) to time out. Our team is investigating the issue and working on a quick solution. Until then, job submission will not be possible. Running jobs might run as usual but exit with failures if connections to the controller fails. Results might still be valid.
Maintenance of login23-g-1 due to GPU errors
Due to repeated issues with the GPUs on the GPU dialog node login23-g-1, the node will be under shorthand maintenance on 2025-12-10 and cannot be used until further notice.
A technician will perform an on-site hardware diagnosis and component modifications which require the node to be shut down.
Please consider using an interactive GPU job or the Jupyterhub with the interactive node n23i0001 for short GPU computations.
Slurm downtime
Monday morning we will have to restart our Slurm database and this will cause Slurm commands to be unavailable for the planned downtime.
Jobs will not be able to be submitted.
Slurm emergency downtime
Due to unforeseen circumstances we are forced to fully stop and restart the Slurm batch system infrastructure.
During this short downtime the submission of new jobs wont be possible, Slurm commands will not be available and other Slurm related tasks will not be possible.
Already running jobs should be able to continue and finish without issue.