BOXX GX8 Rackmount Server Overview
The BOXX GX8 Rackmount Series supports the Intel® Xeon® processor E5-2600 v4 product families and up to 10 PCIe Gen 3.0 x16 compatible devices. GX8 features multi-device peering on a single PCIe root complex, such as the NVIDIA® Tesla® V100 GPU cards, powered by NVIDIA Volta, which is driving the AI revolution and enabling HPC breakthroughs, so you get only the best for your accelerated applications. This multi-device peering on a single PCIe root complex makes it a perfect solution for GPU accelerated applications and libraries like those used for deep learning, data analytics and molecular dynamics such as PyTorch, MXNet, Caffe2, and TensorFlow.
The GX8 Series rackmount servers are different from other GPU supporting hardware implementations. Most of the hardware configurations available today only provide maximum performance between specific pairs of GPUs; and since GPUs are paired up, jobs requiring communication between arbitrary GPUs experience a performance impact. Additionally, there can be significant performance impacts with trying to scale more than four GPUs on multi-socket systems. These have been persistent problems for customers who are pushing the limits of GPUs with large, complex data sets and calculations, or where data must be streamed between GPUs. BOXX has been able to overcome these issues, and achieve near linear performance scaling with its design.
By utilizing the SR3615 PCIe 96-lane switch expander, the BOXX GX8 supports up to eight NVIDIA Tesla V100 GPU cards and provides room for additional InfiniBand® or NVMe storage devices while enabling increased bandwidth and lower latencies between PCIe Gen3 devices than are possible in traditional systems. By enabling up to 8 discrete GPU accelerators to communicate directly with each other on the PCI bus, free of the need for host CPU intervention, they can create a "micro-cluster", sharing a single memory address space.
NEURAL NETWORK TRAINING: DONE.
Maximize PCIe Bandwidth
BOXX is a strong believer in utilizing a technology to its fullest potential whenever possible and GPUs and GPU Accelerators are no exception. If the GPU has a PCIe Gen3 x16 link, then it should use it when communicating with other GPUs — any other GPUs. Our switch expander technology allows us to scale and peer multiple PCIe x16 Gen3 cards on a single root hub ensuring that the maximum PCIe bandwidth is available utilized for inter-card communication.
Minimize Intercard Latency and Obtain Consistent Performance Between GPUs
Our switch expander allows GPUs to communicate as if they are all on the same bus... because they are. Gone are the days of needing a bounce-buffer in host memory, or leaving GPU DMA engines unused because they couldn't address other devices in the system. This reduces intercard latency while helping to maintain a consistent performance level between GPUs.
Supports the Largest Number of GPU Offerings
We work closely with our technology partners to ensure you're given the broadest offerings for your application. The BOXX GX8 Series supports both professional and consumer cards from the leading manufacturers including ground-breaking GPU accelerators, such as the NVIDIA Tesla® V100 accelerators designed specifically for deep learning deployments.
What’s the Difference Between Artificial Intelligence, Machine Learning, and Deep Learning?
Artificial intelligence is the future. Artificial intelligence is science fiction. Artificial intelligence is already part of our everyday lives. All those statements are true, it just depends on what flavor of AI you are referring to. In this multi-part series, long-time tech journalist Michael Copeland explains the fundamentals of deep learning.
MORE THAN JUST HARDWARE.
WE KNOW YOUR SOFTWARE, TOO!
BOXX supports multiple configurations of its products and prefers to work closely with its customers and partners to determine the best fit for their company's needs. We work hard to listen and understand and can tailor any of our products to your specific requirements. If your’re looking to accelerate training and inference of deep neural networks using applications like TensorFlow, Caffe, Torch 7, Theano, Neon, and AMBER one of our performance specialists can guide you to the appropriate solution and configuration. Click below to connect with us.