Dell Broadcom NetXtreme Family of Adapters Broadcom NetXtreme 57XX User Guide - Page 158

High-Performance Computing Cluster

Page 158 highlights

NOTE: Microsoft Network Load Balancing is not supported with Microsoft Cluster Software. High-Performance Computing Cluster Gigabit Ethernet is typically used for the following three purposes in high-performance computing cluster (HPCC) applications: 1. Inter-Process Communications (IPC): For applications that do not require low-latency high-bandwidth interconnects (such as Myrinet, InfiniBand), Gigabit Ethernet can be used for communication between the compute nodes. 2. I/O: Ethernet can be used for file sharing and serving the data to the compute nodes. This can be done simply using an NFS server or using parallel file systems such as PVFS. 3. Management & Administration: Ethernet is used for out-of-band (ERA) and in-band (OMSA) management of the nodes in the cluster. It can also be used for job scheduling and monitoring. In our current HPCC offerings, only one of the on-board adapters is used. If Myrinet or IB is present, this adapter serves I/O and administration purposes; otherwise, it is also responsible for IPC. In case of an adapter failure, the administrator can use the Felix package to easily configure adapter 2. Adapter teaming on the host side is neither tested nor supported in HPCC.

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • 31
  • 32
  • 33
  • 34
  • 35
  • 36
  • 37
  • 38
  • 39
  • 40
  • 41
  • 42
  • 43
  • 44
  • 45
  • 46
  • 47
  • 48
  • 49
  • 50
  • 51
  • 52
  • 53
  • 54
  • 55
  • 56
  • 57
  • 58
  • 59
  • 60
  • 61
  • 62
  • 63
  • 64
  • 65
  • 66
  • 67
  • 68
  • 69
  • 70
  • 71
  • 72
  • 73
  • 74
  • 75
  • 76
  • 77
  • 78
  • 79
  • 80
  • 81
  • 82
  • 83
  • 84
  • 85
  • 86
  • 87
  • 88
  • 89
  • 90
  • 91
  • 92
  • 93
  • 94
  • 95
  • 96
  • 97
  • 98
  • 99
  • 100
  • 101
  • 102
  • 103
  • 104
  • 105
  • 106
  • 107
  • 108
  • 109
  • 110
  • 111
  • 112
  • 113
  • 114
  • 115
  • 116
  • 117
  • 118
  • 119
  • 120
  • 121
  • 122
  • 123
  • 124
  • 125
  • 126
  • 127
  • 128
  • 129
  • 130
  • 131
  • 132
  • 133
  • 134
  • 135
  • 136
  • 137
  • 138
  • 139
  • 140
  • 141
  • 142
  • 143
  • 144
  • 145
  • 146
  • 147
  • 148
  • 149
  • 150
  • 151
  • 152
  • 153
  • 154
  • 155
  • 156
  • 157
  • 158
  • 159
  • 160
  • 161
  • 162
  • 163
  • 164
  • 165
  • 166
  • 167
  • 168
  • 169
  • 170
  • 171
  • 172
  • 173
  • 174
  • 175
  • 176
  • 177
  • 178

NOTE: Microsoft Network Load Balancing is not supported with Microsoft Cluster Software.
High-Performance Computing Cluster
Gigabit Ethernet is typically used for the following three purposes in high-performance computing cluster (HPCC) applications:
1. Inter-Process Communications (IPC): For applications that do not require low-latency high-bandwidth interconnects
(such as Myrinet, InfiniBand), Gigabit Ethernet can be used for communication between the compute nodes.
2. I/O: Ethernet can be used for file sharing and serving the data to the compute nodes. This can be done simply using an
NFS server or using parallel file systems such as PVFS.
3. Management & Administration: Ethernet is used for out-of-band (ERA) and in-band (OMSA) management of the nodes
in the cluster. It can also be used for job scheduling and monitoring.
In our current HPCC offerings, only one of the on-board adapters is used. If Myrinet or IB is present, this adapter serves I/O
and administration purposes; otherwise, it is also responsible for IPC. In case of an adapter failure, the administrator can use
the Felix package to easily configure adapter 2. Adapter teaming on the host side is neither tested nor supported in HPCC.