Recent Changes on Hazel Cluster
- login.hpc.ncsu.edu now goest to RHEL 9.2 login nodes
- login01.hpc.ncsu.edu remains available with CentOS 7.9
- RHEL 9.2 HPC login node image available via VCL
- New GPU nodes available from gpu LSF queue
- Nvidia L40 and H100 GPUs (from NSF CC* grant award)
- All GPU nodes are now running Red Hat Enterprise Linux 9.2 and are available from 'gpu' queue
- Update from CentOS 7.9 to Red Hat Enterprise Linux 9.2 well underway with most compute nodes now updated
Note Regarding Message Passing Interconnects
With the RHEL 9.2 update a number of older Flex node InfiniBand interfaces
were no longer supported. All Flex nodes are now using 10Gbps Ethernet for message passing. These nodes do not have
connections between chassis and are primarily accessible from the
single_chassis queue. These Flex nodes have an LSF resource of e10g
assigned to them.
Newer nodes with InfiniBand interfaces have an LSF resource 'ib' assigned.
Work that Remains In-Progress
- Update videos to reflect new cluster name and default shell
Significant Changes between Hazel and Henry2
- Running CentOS 7.9 (currently being updated to RHEL 9.2) - Henry2 had been running 7.5
- Running LSF 10.1.0.13 (has been updated to 10.1.0.14) - Henry2 had been running 10.1.0.10 [this is actually a significant difference as IBM includes significant feature changes in the minor release levels]
- Default shell is bash - Henry2 had been tcsh
- Began operation with significantly fewer cores than Henry2 (which had about 10,000 cores)
- Operation began Jan 9 with about 3000 cores
- Added about 3800 cores from Henry2 nodes (what had been racks 2b and 2l)
- Added about 1600 cores with new partner nodes
- Added about 2000 Flex Chassis node cores from Henry2
- Some additional Flex Chassis nodes may be added - however rack and cooling capacity is being consumed largely by new partner nodes