Dell PowerEdge C4140 Deep Learning Performance Comparison - Scale-up vs. Scale - Page 34

PowerEdge C4140-K-V100-16GB and V100-32GB: SXM2 Multi Node

Page 34 highlights

Deep Learning Performance: Scale-up vs Scale-out 7.2.2 PowerEdge C4140-K-V100-16GB and V100-32GB: SXM2 Multi Node Figure 27: Training with PowerEdge C4140-V100-16&32GB-SXM2 in multi-node PowerEdge C4140-V100-16GB-SXM2 and PowerEdge C4140-V100-32GB-SXM2 with 4 GPUs each were configured in multi-node to run the TensorFlow in distributed mode, extract the throughput performance, and determine its scaling efficiency. The GPUs scale very well within a node to 97% and 90% across the nodes. The ideal performance is computed by multiplying the single-GPU throughput by the number of GPUs in the system. See Figure 28 Architectures & Technologies Dell EMC | Infrastructure Solutions Group 33

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • 31
  • 32
  • 33
  • 34
  • 35
  • 36
  • 37
  • 38
  • 39
  • 40
  • 41
  • 42
  • 43
  • 44
  • 45
  • 46
  • 47
  • 48
  • 49
  • 50
  • 51
  • 52
  • 53

Deep Learning Performance: Scale-up vs Scale-out
Architectures & Technologies
Dell
EMC
| Infrastructure Solutions Group
33
7.2.2
PowerEdge C4140-K-V100-16GB and V100-32GB: SXM2 Multi Node
Figure 27: Training with PowerEdge C4140-V100-16&32GB-SXM2 in multi-node
PowerEdge C4140-V100-16GB-SXM2 and PowerEdge C4140-V100-32GB-SXM2 with 4 GPUs each
were configured in multi-node to run the TensorFlow in distributed mode, extract the throughput
performance, and determine its scaling efficiency. The GPUs scale very well within a node to 97%
and 90% across the nodes. The ideal performance is computed by multiplying the single-GPU
throughput by the number of GPUs in the system. See
Figure 28