The Maxwell-Cluster is a resource dedicated to parallel and multi-threaded application, which can use at least some of the specific characteristics. In addition to serving as a medium scale High-Performance-Cluster, Maxwell incorporates resources for Photon Science data analysis, resources of CFEL, CSSB, Petra4, the European XFEL...
The Maxwell-Cluster is composed of a core partition (maxwell) and group specific partitions. All compute nodes are however available for everyone!
The Maxwell-Cluster is primarily intended for parallel computation making best use of the multi-core architectures, the infiniband low-latency network, fast storage and available memory. The cluster is hence not suited for single-core computations or embarrassingly parallel jobs like Monte-Carlo productions. Use BIRD, Grid or your groups workgroup server (WGS) for this kind of tasks.
The entire cluster is managed by SLURM scheduler (with some notable exceptions). The SLURM scheduler essentially works on a "who comes first" basis. The group specific partitions however have slightly different rules: though everyone can run jobs on group specific nodes, members of the group will have a higher priority and will compete non-group jobs off the partition. See Groups and Partitions on Maxwell for details.
- To get started, please have a look at the Getting Started page!
- The Maxwell Hardware page provides a list of currently available nodes & configurations.
- The Maxwell Partitions page provides a quick overview of the nodes, capacities, features and limits of the individual partitions.
Read the documentation! It should cover at least the essentials. If you come across incorrect or outdated information: please let us know!
Acknowledgments and References
If you find the resource useful for your work, we would greatly appreciate to learn about publications, which have been substantially benefiting from the Maxwell-Cluster. Drop us a mail at maxwell.service@desy.de. Acknowledgement of the maxwell-resource would also be greatly appreciated. It'll help to foster the cluster, for example: "This research was supported in part through the Maxwell computational resources operated at Deutsches Elektronen-Synchrotron DESY, Hamburg, Germany"
Contact
For any questions, problems, suggestions please contact our ticket system: maxwell.service@desy.de
All Announcements will be sent via maxwell-user@desy.de.
We strongly recommend that all maxwell-users subscribe to the mailing list. Only user of the maxwell partition, are automatically added.
User help user
We offer an forum where maxwell user can help other maxwell user. It's by no means a replacement for the regular support channels and helpdesk.
Search the compute space
Blog Posts
-
Blog: max-display3 (aka max-display004,5) as Centos_8 testbed
created by
Nov 10, 2020 18:00
-
Blog: Changes to Maxwell partitioning
created by
Oct 15, 2020 09:55
-
Blog: Software update: octave 5.2.0
created by
Aug 19, 2020 12:46
-
Blog: temporary urgent changes to maxwell partition
created by
Aug 19, 2020 12:45
-
Blog: Maxwell storage outage 10th August 2020
created by
Aug 11, 2020 09:31
-
Blog: Software Updates 05/08/2020
created by
Aug 05, 2020 09:22
-
Blog: Julia on Maxwell
created by
Jun 24, 2020 15:32
-
Blog: OpenMPI problems
created by
Jun 02, 2020 12:36
-
Blog: mmlsquota temporarily disabled
created by
May 13, 2020 10:34
-
Blog: Anaconda3 updates
created by
May 07, 2020 12:53
-
Blog: RESOLVED: Major outage in GPFS
created by
Mar 05, 2020 09:31
-
Blog: docker removed from workgroup and login server
created by
Jan 30, 2020 13:18
-
Blog: Software Updates
created by
Nov 20, 2019 11:07
-
Blog: RESOLVED: Infiniband trouble on Maxwell - no login possible
created by
Oct 08, 2019 14:05
-
Blog: New Jupyterhub extensions
created by
Sep 23, 2019 20:34
- No labels