InfraGRID – UVT HPC Center – InfraGRID Cluster
InfraGRID Cluster consists of 50 compute nodes powered by IBM BladeCenter-H technology. The entire solution is built up from 4 blade center chassis, each with 14 HS21 blade server. Each blade server has a dual quad core Intel Xeon E5504 CPU (clocked at 2.00Ghz) and 10GB RAM memory. The connectivity is delivered by: (a) Infiniband for interconnect and storage, (b) fiber channel for dedicated storage and (c) Gigabit for service networking. This cluster was initially setup in 2009. In 2011 the InfraGRID Cluster received an update by adding a new IBM BladeCenter-H chassis powered with 7 dual CPU/GPU HS22 blade servers. The CPU is powered by Intel XEON technology with a clock speed at 3.46Ghz and 32GB RAM memory. The GPU cards are NVidia Tesla M2070Q (448 GPU cores and 6GB GDDR5 RAM memory). All hardware details are available in the table below.
The cluster is managed by two service nodes, one dedicated to user access (also called head node) and one dedicated exclusively for service actions and cluster management. The storage is shared using GPFS file system over Infiniband using two dedicated NSD nodes. The service management is conducted using IBM BladeCenter the advanced management module (AMM) that is built in into both blade center chassis and blade servers (IMM). This feature allows remote administration and monitoring of any of the installed hardware. Infragrid Cluster is cooled with air. The displacement of the cooling units is in-row with a well delimited hot-cold area. Currently three APC InRow cooling units are installed and working in a cluster configuration to obtain cooling unit high-availability behavior.
Administrative data
Name | InfraGRID Cluster |
Short Description | UVT HPC Center – InfraGRID Cluster |
Owner | Universitatea de Vest din Timisoara (UVT) |
Country | Romania |
Dedication to VI-SEEM
CPU (percent) | 10% |
Storage (percent) | 0% |
Accelerators (percent) | 20% |
CPU (core-hours per year) | 798,912 |
Storage in TB | 0 |
Accelerators (hours per year) | 5,494,272 |
Integration
System operational since | Sep 2009 |
Available to the project from | PM04 |
Expected date to be phased | - |
Interfaces | SSH |
Computational Power
Number of servers | 57 | Interconnect type | QDR InfiniBand | Peak performance (Tflops) | 2.11 | |
Server specification | IBM BladeCenter HS21 | Interconnect latency | 2.5 μs | Real performance (Tflops) | 3.5 | |
CPU per server | 2 | Interconnect bandwidth | 40 Gbps | Operating system | Red Hat Enterprise Linux | |
RAM per server | 10 GB | Local filesystem type | GPFS | Version | 6 | |
Total number of CPU-cores | 400 | Total storage (TB) | 50 TB | Batch system/scheduler | LoadLeveler | |
Max number of parallel processes | 800 | Peak performance CPU (Tflops) | 3.5 | Accelerators type | Nvidia Tesla M2070Q | |
Number of cores | 448 | Servers equipped with accelerators | 7 | Peak performance accelerators (Tflops) | 3.5 | |
Accelerators per server | 1 | |||||
Development tools | IIntel Compilers (C/C++, FORTRAN), GNU Compilers, OpenMPI, CUDA, gdb, pgdbg | |||||
Libraries | Intel MKL, HDF5, FFTW, NetCDF, GSL, LAPACK, Boost, BLAS | |||||
Applications | Misc. |