It is that time of the year at Mellanox, where we proudly present some of the coolest things our team has worked on! This time it is going to be at the Open Compute Project (OCP) Summit which will be held in the heart of Silicon Valley – San Jose Convention Center on March 11-12, 2015. It is impressive to see how hyper-scale architecture has been revolutionized in just 4 years.
What started as a small project from the basement of Facebook office in Palo Alto has come alive in the form of some cutting edge innovation in racks, server, networking and storage. Some of these innovations from Mellanox will take the center stage during the OCP summit that will accelerate the advancement of data center components, mainly server and networking. Key highlights during the OCP events are:
ConnectX-4 and Multi-Host: Back in November, Mellanox announced the industry’s first 100GbE interconnect adapter pushing the innovation in the networking arena in HPC, Cloud, Web2.0, storage and enterprise applications. With a throughput of 100 Gb/s, bidirectional throughput of 195 Gb/s, application latency of 610 nanoseconds and message rate of 149.5 million messages per second, ConnectX-4 InfiniBand adapters provide the means to increase data center return on investment while reducing IT costs.
Pushing the frontiers of science and technology will require extreme-scale computing with machines that are 500-to-1,000 times more capable than today’s supercomputers. As researchers continuously refine their models, the demand for more parallel computation and advanced networking capabilities is paramount.
As a result of the ubiquitous data explosion and the ascendance of big data, today’s systems need to move enormous amounts of data and perform more sophisticated analysis; the interconnect truly becomes the critical element of enabling the use of data.
One of the biggest catchphrases in modern science is Human Genome–the DNA coding that largely pre-determines who we are and many of our medical outcomes. By mapping and analyzing the structure of the human genetic code, scientists and doctors have already started to identify the causes of many diseases and to pinpoint effective treatments based on the specific genetic sequence of a given patient. With the advanced data that such analysis provides, doctors can offer more targeted strategies for potentially terminal patients at times when no other clinically relevant treatment options exist.
The University of Edinburgh’s entry into the ISC 2014 Student Cluster Competition, EPCC, has been awarded first place in the LINPACK test. The EPCC team harnessed Boston’s HPC cluster to smash the 10Tflop mark for the first time – shattering the previous record of 9.27Tflops set by students at ASC14 earlier this month. The team recorded a score of 10.14Tflops producing 3.38 Tflops/kW which would achieve a rank of #4 in the Green500, a list of the most energy efficient supercomputers in the world.
This achievement was made possible thanks to the provisioning of a high performance, liquid cooled GPU cluster by Boston. The system consisted on four 1U Supermicro servers, each comprising of two Intel® Xeon™ ‘Ivy Bridge’ processors and two NVIDIA® K40 Tesla GPUs, and Mellanox FDR 56Gb/s InfiniBand adapters, switches and cables.
As data continues to grow exponentially storing today’s data volumes in an efficient way is a challenge. Many traditional storage solutions neither scale-out nor make it feasible from Capex and Opex perspective, to deploy Peta-Byte or Exa-Byte data stores.
In this newly published whitepaper, we summarize the installation and performance benchmarks of a Ceph storage solution. Ceph is a massively scalable, open source, software-defined storage solution, which uniquely provides object, block and file system services with a single, unified Ceph storage cluster. The testing emphasizes the careful network architecture design necessary to handle users’ data throughput and transaction requirements.
Mellanox recently announced a collaboration with IBM to produce a tightly integrated server and storage solutions that incorporate our end-to-end FDR 56Gb/s InfiniBand and 10/40 Gigabit Ethernet interconnect solutions with IBM POWER CPUs. By combining IBM POWER CPUs with the world’s highest-performance interconnect solution will drive data at optimal rates, maximizing performance and efficiency for all types of applications and workloads, as well as enable dynamic storage solutions to allow multiple applications to efficiently share data repositories.
Advances in high-performance applications are enabling analysts, researchers, scientists and engineers to run more complex and detailed simulations and analyses in a bid to gather game-changing insights and deliver new products to market. This is placing greater demand on existing IT infrastructures, driving a need for instant access to resources – compute, storage, and network.
Companies are looking for faster and more efficient ways to drive business value from their applications and data. The combination of IBM processor technologies and Mellanox high-speed interconnect solutions can provide clients with an advanced and efficient foundation to achieve their goals.
Shout out to anyone who happens to attend the GPU Technology Conference 2014! This conference is touted as the world’s biggest and most important GPU developer conference. Follow all the social conversation around the event using the hashtag #GTC2014. The conference will be held next week, March 24-27,2014 at the San Jose McEnery Convention Center in San Jose, CA.
This is the fourth year I am attending this event and I will be hanging out at the “Ask the Expert Table” at the GTC. Feel free to swing by and chat about any of your burning questions you may have on GPUDirect RDMA with Mellanox InfiniBand!
New advances in Big Data applications are enabling analysts, researchers, scientists and engineers to run more complex and detailed simulations and analyses than ever before. These applications deliver game-changing insights, bring new products to market and place greater demand on existing IT infrastructures.
This ever-growing demand drives the need for instant access to resources – compute, storage, and network. Users are seeking cutting-edge technologies and tools to help them better capture, understand and leverage increasing volumes of data as well as build infrastructures that are energy-efficient and can easily scale as their business grow.
The HPC Advisory Council published a best practices paper showing record application performance for LS-DYNA® Automotive Crash Simulation, one of the automotive industry’s most computational and network intensive applications for automotive design and safety. The paper can be downloaded here: HPC Advisory Council : LS-Dyna Performance Benchmark and Profiling.
The LS-DYNA benchmarks were tested on a Dell™ PowerEdge R720 based-cluster comprised of 32 nodes and with networking provided by Mellanox Connect-IB™ 56Gb/s InfiniBand adapters and switch. The results demonstrate that the combined solution delivers world-leading performance versus any given system at these sizes, or versus larger core count system based on Ethernet or proprietary interconnect solution based supercomputers.
The TopCrunch project is used to track the aggregate performance trends of high performance computer systems and engineering software. Rather than using a synthetic benchmark, actual engineering software applications are used with real datasets and run on high performance computer systems.
Author: Scot Schultz is a HPC technology specialist with broad knowledge in operating systems, high speed interconnects and processor technologies. Prior to joining Mellanox, he spent the past 17 years at AMD in various engineering and leadership roles, most recently in strategic HPC technology ecosystem enablement. Scot was also instrumental with the growth and development of the Open Fabrics Alliance as co-chair of the board of directors. Follow him on Twitter: @ScotSchultz.
Windows Azure continues to be the leader in High-Performance Computing Cloud services. Delivering a HPC solution built on top of Windows Server technology and Microsoft HPC Pack, Windows Azure offers the performance and scalability of a world-class supercomputing center to everyone, on demand, in the cloud.
Customers can now run compute-intensive workloads such as parallel Message Passing Interface (MPI) applications with HPC Pack in Windows Azure. By choosing compute intensive instances such as A8 and A9 for the cloud compute resources, customers can deploy these compute resources on demand in Windows Azure in a “burst to the cloud” configuration, and take advantage of InfiniBand interconnect technology with low-latency and high-throughput, including Remote Direct Memory Access (RDMA) technology for maximum efficiency. The new high performance A8 and A9 compute instances also provide customers with ample memory and the latest CPU technology.
Author: Eli Karpilovski manages the Cloud Market Development at Mellanox Technologies. In addition, Mr. Karpilovski serves as the Cloud Advisory Council Chairman. Mr. Karpilovski served as product manager for the HCA Software division at Mellanox Technologies. Mr. Karpilovski holds a Bachelor of Science in Engineering from the Holon Institute of Technology and a Master of Business Administration from The Open University of Israel. Follow him on Twitter.