All posts by Scot Schultz

About Scot Schultz

Scot Schultz is a HPC technology specialist with broad knowledge in operating systems, high speed interconnects and processor technologies. Joining the Mellanox team in March 2013 as Director of HPC and Technical Computing, Schultz is a 25-year veteran of the computing industry. Prior to joining Mellanox, he spent the past 17 years at AMD in various engineering and leadership roles, most recently in strategic HPC technology ecosystem enablement. Scot was also instrumental with the growth and development of the Open Fabrics Alliance as co-chair of the board of directors. Scot currently maintains his role as Director of Educational Outreach, founding member of the HPC Advisory Council and of various other industry organizations. Follow him on Twitter: @ScotSchultz

Stanford Researcher is the Medical Worlds’ Braveheart

The average human heart beats about 4,800 times per hour or 115,200 times per day. That means if you live to be 80 years old, your heart would have beaten approximately 3,363,840,000 times.

A normal human heart has four chambers: two upper chambers called atria, which receive blood into the heart, and two lower chambers called ventricles, which pump blood out of the heart. A single ventricle defect, also called single defect anomaly or single ventricle physiology, refers to a variety of cardiac defects where only one of the heart’s two ventricles functions properly. A single ventricle defect is a congenital (meaning it is present at birth) condition and is among the most complex of heart defects. This condition occurs in five of every 100,000 live births as the heart develops during the first eight weeks of the mother’s pregnancy. Current surgical correction, which has been practiced over the past five decades, shows poor outcome with mortality rates as high as 30 percent. This is complicated by the fact that the condition requires multiple, complex surgeries and while survival rates continue to improve, the condition still carries some of the highest morbidity and mortality rates related to congenital heart surgery.

That’s where Stanford Researcher and Assistant Professor at Cornell, Dr. Mahdi Esmaily, comes in. Dr. Esmaily, or Dr. Braveheart, as some of us like to call him, has been conducting research that is nothing short of ground breaking and lifesaving. Dr. Esmaily, Postdoctoral scholar at the Center for Turbulence Research – Stanford University, has applied his engineering expertise to a long standing problem in the medical field. He has conducted computational simulations of blood flow to evaluate an innovative new surgical procedure. The outcome is that infant cardiac surgeons may be able to use a radical new surgery by adopting a frankly engineering concept. This is accomplished by constructing an anatomy that is inspired by an ejector pump, a device that is typically found in industrial power plants. In this operation, the high-energy flow from the systemic circulation is injected into the low-energy flow from the upper body, assisting its drainage into the pulmonary arteries. This is counter to the current surgery which is known to cause inadequate pulmonary blood flow in already stressed patients ‒ particularly true in tiny infants.

The High Performance Computing Center (HPCC) at Stanford University was founded to provide high performance computing resources and services that enable computationally-intensive research within the School of Engineering and to support the efforts of scientists performing sponsored research. The High Performance Computing Center leverages Mellanox InfiniBand in their High Performance Computing research to enable larger simulations, analyses and faster computation times than are possible using computers available to individual researchers.

Dr. Esmaily used high performance computing resources to simulate realistic 3-dimensional models of neonatal circulations. These simulations predicted that his proposed surgery provides significantly higher oxygen delivery at a lower single-ventricular workload in comparison to the conventional operation, meaning that part of the heart will not have to work as hard. The conclusion is that while there is more research to be done, the use of an ejector pump, augmented by Dr. Esmaily’s technique, holds the promise of giving surgeons a brand new, less risky option, for newborns undergoing this kind of surgery and with it, hope for the families of infants born with heart defects.

Supporting Resources:

Climate Change and Global Health: What NCI is Doing to Help

Climate change is poised to become what the prestigious British medical publication The Lancet has called “the biggest global health threat of the 21st century.” The health risks associated with climate change are numerous and alarming. Just to name a few: increases in heat-related illnesses and death; extreme weather-related injuries and mortality; aggravated chronic illnesses; spread of infectious diseases; increases in asthma and respiratory allergies, and upsurge in chronic respiratory disorders; rising malnutrition and child development complications; increases in stress-related and mental health disorders…the list goes on and on. In addition, there are tangible impacts related to both population displacement and migration; as well as climate-triggered instability (famine) and subsequent conflict. The healthcare sector is just beginning to understand that climate change will have major impacts on health care costs, services and delivery. The World Health Organization has estimated some of impact of climate change to be:

  • Between 2030 and 2050, climate change is expected to cause approximately 250,000 additional deaths per year, from malnutrition, malaria, diarrhea and heat stress.
  • The direct damage costs to health (i.e. excluding costs in health-determining sectors such as agriculture and water and sanitation), is estimated to be between US$ 2-4 billion/year by 2030.
  • Areas with weak health infrastructure – mostly in developing countries – will be the least able to cope without assistance to prepare and respond.
  • Reducing emissions of greenhouse gases through better transport, food and energy-use choices can result in improved health, particularly through reduced air pollution.

This is why climate change and tracking the climate has such far-reaching implications. Last quarter, I wrote about the growing impact of global warming on The Great Barrier Reef and coral bleaching and the ground-breaking research being done at the Australia’s National Computational Institute. Now, I’d like to highlight the important work they are doing in climate modeling.

ACCESS is Australia’s national earth system science and climate modelling suite, a complex coupled-system model that comprises atmosphere, ocean, sea-ice, land, and other components — derived from the best of the UK, USA, France, and Australia to provide national weather and climate prediction capability. The model’s complexity comes from the vast span of time scales over which it has to work: from hours for extreme weather events (storms and bushfires), days for general weather prediction, months for seasonal prediction, decades for planning for environmental change, through to centuries and millennia for long-term climate change.

ACCESS is developed through a collaboration of the Bureau of Meteorology, CSIRO, the academic community through the ARC Centre of Excellence in Climate System Science, and NCI.

A flooded road with depth indications in Queensland, Australia.

NCI’s role is mission critical, not only as the collaborative, integrated development platform, but also through its unique expertise in optimizing the performance of key model components. Performance improvements of 30-40 percent and much higher code scalability (up to 20-fold improvements with some codes now exploiting up to 20,000 cores), are enabling greater operational efficiency and productivity with faster time to results, more accurate simulations that enable new scientific outcomes and insight, and much heightened prediction capability.

In real terms, the outcomes are wide-ranging and include multi-billion dollar benefits for agriculture through more accurate seasonal prediction, the reduction of severe economic losses and the mitigation of dangers to public safety and health from extreme weather events.

As part of this, NCI is using Mellanox’s interconnect solutions allow for faster inter-node connectivity and access to storage, providing Australian researchers and scientific research organizations with critical on-demand access to NCI’s high-performance cloud. This cloud facilitates scientific workloads with a deployment that combines the Mellanox CloudX solution with OpenStack software to support high performance workloads on a scalable and easy to manage cloud platform. CloudX simplifies and automates the orchestration of cloud platforms and reduces deployment time from days to hours. The NCI deployment is based on Mellanox 40/56 Gb/s Virtual Protocol Interconnect adapters and switches supporting both InfiniBand and Ethernet. NCI also has Mellanox’s 100Gbit/s EDR InfiniBand interconnect for its new Lenovo NextScale supercomputer. This powerful combination of storage and compute power enable NCI to deliver extremely complex simulations and more accurate predictions, all with the aim of improving the human condition.

 Supporting Resources:

 

How Artificial Intelligence is Revolutionizing Personalized Medicine

Imagine becoming gravely ill and yet being able to receive an accurate diagnosis with a recommended treatment plan in just 10 minutes. Sound like the future? This is actually happening now with the help of Artificial Intelligence (AI).  The University of Tokyo recently reported that Watson, IBM’s cognitive supercomputer, correctly diagnosed a rare form of leukemia in a 60-year-old woman. Doctors originally thought the woman had acute myeloid leukemia, but after examining 20 million cancer research papers in 10 minutes, Watson was able to correctly determine the actual disease and recommend a personalized treatment plan.  AI – and its related applications, Machine Learning (ML) and Deep Learning (DL) – are changing healthcare as we know it. The advancements made in AI will revolutionize research and, ultimately, personalized medicine.

The Historical Challenge with Data

Big Data has been a buzz word for several years now. Hospitals, like enterprises, have been drowning in big data. From the moment doctors begin keeping patient records, they – and now hospitals – have been amassing large quantities of complex data within patient medical records; including handwritten notes, X-ray results, blood samples, vital signs, DNA sequences, and more. Historically, this data has been disparate and existed in hard copies only, making it nearly impossible to analyze in aggregate. Now with AI, analytic tools and other technological advancements, there is a way to actually organize, analyze and cross reference the data, enabling hospitals, doctors and researchers to finally put that data to use. With medical devices alone, the influx of data is staggering, necessitating that hospitals – and the medical industry in general – rethink the way they collect, store and analyze data.

AI Revolutionizes Personalized Medicine  

While there is generally a solution to any problem, often times it isn’t that we can’t see the solution, it’s that we can’t correctly identify the problem. AI is able to learn from each piece of data it is given and rapidly re-evaluate its analysis as more data and more is received. This enables doctors and researchers to better identify problems and, subsequently, the potential solutions to those problems. A door to a world of possibilities has now been opened, and with it, the potential to find cures for the thus-far incurable diseases, perhaps even within our lifetime.

AI is not limited to traditional data on a spread sheet. It can interpret and aggregate imaging, text, handwritten notes, test results, sensor data, and even demographic and geo-spatial data. AI will be able to cross reference data, find commonalities and draw insights that were previously impossible due to data silos or the sheer amount of time it would take for a human to crunch the numbers. It can also consider seemingly unrelated or outside factors that doctors and researchers may not immediate see as relevant. For example, environmental factor, such as elevation, humidity and proximity to certain dense mineral deposits, factories or agriculture. This ability to rapidly analyze data, and potential correlations, creates a more comprehensive and holistic view into a patient’s health.

AI in Action Today

  • The National Cancer Institute has partnered with NVIDIA to develop an AI framework – powered by Mellanox InfiniBand adapters and aimed at supercharging cancer research. The framework, CANDLE (Cancer Distributed Learning Environment), will leverage AI to extract and study millions of patient records with the goal of understanding how cancer spreads and reoccurs. This is an example of AI being able to pour through large amounts of genomic data in a quick manner so doctors can draw conclusions.
  • A recent study published in Neurobiology of Aging found that AI could help detect signs of Alzheimer’s in patient brain scans before doctors. AI is currently being used to study scans of healthy brains and brains affected by Alzheimer’s to learn and identify the telling systems of the disease.
  • Medecision, a leading global healthcare information technology provider, is also employing AI to sweep through large amounts of data, identifying variables and predicting avoidable hospitalizations in diabetes patients. The medical community has only just begun to scratch the surface of what can be achieved with AI.

Preparing for Impact

AI is impacting nearly every aspect of the healthcare industry from patient care such as the examples described herein to hospital security and pharmaceutical drug development (stay tuned for a future post on how AI may just be the solution to rising drug prices). Mellanox is committed to the cause and is helping to accelerate many of the world’s leading AI, ML and DL systems with solutions like RDMA, GPUDirect RDMA, SHARP and intelligent interconnects that are able to handle the highest rates of real-time data and mitigate network congestion. We have only just begun to scratch the surface of AI’s potential and Mellanox believes that AI has the potential to improve our quality of life, find cures for life’s most threating illnesses and provide a deeper understanding of our own evolution.

The Heat is On

It’s on everyone’s bucket list; to experience the Great Barrier Reef. In fact, Traveler Magazine puts in on the Ultimate Travel Bucket List. However, before you book that dream vacation to this World Heritage site, know that the reef is fragile and scientific research says it is getting more fragile with each passing day. So much so that scientists are working frantically to preserve it. In fact, the work being done at Australia’s National Computational Infrastructure is nothing less than a virtual wake up call for as all of us about to the impact global warming is having on the reef.

NCI is a national provider of high-performance research computing and data services. NCI systems currently support around 4,000 researchers working on more than 500 projects. These researchers come from more than 34 Australian universities, national science agencies and medical research institutes, with an increasing industry user base. As Australia’s national research computing service, the organization provides world-class, high-end services to Australia’s researchers, the primary objectives of which are to raise the ambition, impact, and outcomes of Australian research through access to advance computational and data-intensive methods, support, and high-performance infrastructure.

Back in 2013, NCI selected Mellanox’s interconnect to support Australia’s national research computing infrastructure which provides world-class, high-end services to Australia’s researchers. Mellanox’s interconnect solutions allow for faster inter-node connectivity and access to storage, providing researchers and scientific research organizations with critical on-demand access to NCI’s high-performance cloud. The system is designed around Mellanox CloudX™ , a reference architecture that provides the most efficient, highest performing scalable clouds based on Mellanox’s superior interconnect.

Then, in 2016, NCI chose Mellanox’s 100Gbit/s EDR InfiniBand interconnect for its newest Lenovo® NextScale supercomputer. The new system added a whopping 40 percent performance increase in NCI computational capacity.

Some research can take years — even decades — to come to fruition but at NCI, the results are already in.  Earlier in 2017, it was reported that more than two-thirds of the coral in Australia’s Great Barrier Reef was experiencing enormous amounts of bleaching.  Similar bleaching events occurred in 2016, and 2017 has already devastated a 1,500 km (900 miles) stretch of the reef. Before 2016, there had only been two bleaching events along the Great Barrier Reef in the past two decades, reflecting an alarming trend. That’s when researchers came to NCI to get their work done.

Leveraging the computing resources at NCI, researchers looked at four key extreme Australian events; the Angry Summer 2012/13; the Coral Sea marine heatwave of 2016; the severe rain event in NE Australia in 2010; and the 2006 drought in southeast Australia. They modeled how often these events would occur under each scenario. Results showed that keeping global temperatures from rising less than 1.5°C would have a clear benefit for Australia in terms of reducing extreme weather events and the costs associated with such incidences. Further, scientists at NCI reported that if the global average surface temperatures increase just 1.5°C above pre-industrial conditions, then a repeat of the coral bleaching that severely damaged the Great Barrier Reef earlier this year would become more than twice as likely. Scientific modeling results from NCI also revealed that if the world heats up by just 2°C more than the pre-industrial world, it nearly triples the odds of another mass bleaching event.

These findings from University of Melbourne scientists at the ARC Centre of Excellence for Climate System Science, are the result of research using NCI’s Mellanox-enabled HPC facilities to look at how Australian extremes in heat, drought, precipitation and ocean warming will change in a world 1.5°C and 2°C warmer than pre-industrial conditions.

As the world continues to vigorously debate the impact of global warming, NCI is helping to bring the facts to the table — not in months or weeks, but now, today.  For a World Heritage site in peril, these findings have long-term implications that require action now.

 

Supporting Resources:

 

Mellanox with IBM Research: Deep Learning, THIS is how you do it!

Mellanox’s role in record setting deep learning performance with IBM PowerAI

This year has already proven to be a game-changer in the next generation performance for artificial intelligence, as the race continues to solve the challenges of scalable deep learning.  Most popular deep learning frameworks today can scale to multiple GPUs in a server, but its difficult with multiple servers with GPUs.

This is challenge in particular is where Mellanox has been the clear leader as the only interconnect solution able to deliver the needed performance and offload capabilities to unlock the power of scalable AI.

IBM Research just announced their amazing achievement in unprecedented performance and close to ideal scaling with new distributed deep learning software which achieved record communication overhead and 95% scaling efficiency on the Caffe deep learning framework with Mellanox InfiniBand and over 256 NVIDIA GPUs in 64 IBM Power systems. 

With the IBM DDL (Distributed Deep Learning) library, it took just 7 hours to train ImageNet-22K using ResNet-101. From 16 days down to 7 hours changes the workflow of data scientists.

 

That’s a 58x speedup!

 

You can read more at the IBM blogs : https://www.ibm.com/blogs/research/2017/08/distributed-deep-learning/ and https://www.ibm.com/blogs/systems/scaling-tensorflow-and-caffe-to-256-gpus/

And download the whitepaper here : https://arxiv.org/abs/1708.02188

A technical preview of this IBM Research Distributed Deep Learning code is available today in IBM PowerAI 4.0 distribution for TensorFlow and Caffe.

 

EPYC-Curious? A New Datacenter Recipe for a Refreshing and Satisfying Datacenter Featuring Mellanox’s Key Ingredients

Hungry for something new and innovative? At the SIGGRAPH 2017 event, AMD showcased their latest EPYC™ processors and 80 Radeon™ Instinct MI25 accelerators featuring Mellanox’s EDR 100Gb/s InfiniBand Smart Interconnect. This distinctive new blend of ingredients ushers in a new era in tackling today’s most complex workloads. Unveiling the first systems from Inventec, the single rack of Inventec P47 systems can unleash 1 PetaFLOPS of compute power at full 32-bit precision, well suited to address the most challenging deep learning, HPC and enterprise-class applications. This is another advancement in the way we think of datacenter deployments, because this enables more cores, threads, compute units, IO lanes and memory channels in use at one time than in any other system in history.

These latest Inventec P47 systems include Samsung Electronics HBM2 memory which is used across the Vega architecture Radeon Instinct MI25 accelerators. Samsung also provided high-performance NVMe SSD storage and high-speed DDR4 memory. For connecting the Inventec P47 systems, AMD and Mellanox Technologies teamed up to ensure the full potential of this state-of-the-art compute-powerhouse. Because high-performance compute and artificial intelligence rely on increasingly complex computations, it’s causing data bandwidth and high speed storage requirements to spiral upward. The most savvy users in the world already deploy Mellanox end-to-end, knowing there is no other interconnect solution that can even compete at the high-end of High Performance Computing (HPC) environments and Enterprise Data Centers (EDC) they want to maximize their performance-per-TCO-dollar. They will need every last bit of bandwidth delivered to this new class of compute and I/O performance with Mellanox’s InfiniBand high-speed, and ultra-low latency smart HCA’s and switches.

Welcome again, AMD to the race in providing the key ingredients to satisfy the most data-hungry datacenters with capabilities that are ready for the challenges of today and the workloads of tomorrow!

 

 

 

Mellanox Educates on Caffe, Chainer, and TensorFlow

Why Mellanox is Participating in the OpenPOWER Developer Congress

Mellanox is not only a founding member of the OpenPOWER Foundation, but also a founding member of its Machine Learning Work Group.  AI/cognitive computing will improve our quality of life, drive emerging markets, and surely play a leading role in global economics. But to achieve real scalable performance with AI, the ability to leverage cutting-edge interconnect capabilities is paramount. Typical vanilla networking just doesn’t scale, so it’s important that developers are aware of the additional performance that can be achieved by understanding the critical role of the network.

Because Deep Learning applications are well-suited to exploit the POWER architecture, it is also extremely important to have an advanced network that unlocks the scalable performance of deep learning systems, and that is where the Mellanox interconnect comes in. The benefits of RDMA, ultra-low latency, and In-Network Computing deliver an optimal environment for data-ingest at the critical performance levels required by POWER-based systems.

Mellanox is committed to working with the industry’s thought leaders to drive technologies in the most open way. Its core audience has always been end users — understanding their challenges and working with them to deliver real solutions. Today, more than ever, the developers, data-centric architects, and data scientists are the new generation of end users that drive the data center. They are defining the requirements of the data center, establishing its performance metrics, and delivering the fastest time to solution by exploiting the capabilities of the OpenPOWER architecture.  Mellanox believes that participating in the OpenPOWER Developer Congress gives us an opportunity to educate developers on its state-of-art-networking and also demonstrates its commitment to innovation with open development and open standards.

 

What is Mellanox Bringing to the Developer Congress?

Mellanox will provide on-site expertise to discuss the capabilities of Mellanox Interconnect Solutions. Dror Goldenberg, VP of Software Architecture at Mellanox, will be present to further dive into areas of machine learning acceleration and the frameworks that already take advantage of Mellanox capabilities, such as Caffe, Chainer, TensorFlow, and others.

Mellanox is the interconnect leader in AI/cognitive computing data centers, and already accelerates machine learning frameworks to achieve from 2x to 18x speedup for image recognition, NLP, voice recognition, and more. The company’s goal is to assist developers with their applications to achieve maximum scalability on POWER-based systems.

For more information check out: https://openpowerfoundation.org/openpower-developer-congress/

Mellanox: The World’s Most Advanced Interconnect, Pushing the Boundaries of Science and Medicine

Each year, Earth Day—April 22—marks the anniversary of the birth of the modern environmental movement in 1970.  Earth Day has transformed into its current status as the largest modern-day observance in the world, celebrated by more than a billion people every year, and a day of action that changes human behavior and provokes policy changes.

Earth Day is about creating change and bringing awareness to societies, that every individual has a keen responsibility to help create a greener, more sustainable future for today and for the future.

I thought to include an example in this special Earth Day blog; how Mellanox, and our high speed intelligent networking fabric, is very much a part of the efforts to advance the understanding of science; as science serves all of us and is a vital role in our daily lives.  In the most recent example, we demonstrate how Mellanox is a contributing technology in helping scientists to use state-of-the-art capabilities to assemble the 1.2 billion letter genome of the Culex quinquefasciatus mosquito, in particular carries the West Nile virus.

A team from Rice University, Baylor College of Medicine, Texas Children’s Hospital and the Broad Institute of MIT and Harvard has developed a new way to sequence genomes that can generate a human genome from scratch for roughly the cost of an MRI exam.  The new geonome will help scientists to better combat the spread of the West Nile virus. You can watch a short video clip on this technology.

“Taking advantage of IBM POWER8 and Mellanox InfiniBand interconnect, we are now able to change the way we assemble a genome,” said Olga Dudchenko, a postdoctoral fellow at The Center for Genome Architecture at Baylor College of Medicine. “And while we originally created Voltron to sequence the human genome, the method can be applied to a dizzying array of species. This gives us an opportunity to explore mosquitoes, which carry diseases that impact many people around the globe.”


“Taking advantage of IBM POWER8 and Mellanox InfiniBand interconnect, we are now able to change the way we assemble a genome,” said Olga Dudchenko, a postdoctoral fellow at The Center for Genome Architecture at Baylor College of Medicine. “And while we originally created Voltron to sequence the human genome, the method can be applied to a dizzying array of species. This gives us an opportunity to explore mosquitoes, which carry diseases that impact many people around the globe.”


A new high performance computing (HPC) system dubbed, “VOLTRON,” which is based on network capabilities from Mellanox, and the latest IBM Power Systems platform, provides the scalable HPC capabilities necessary to accommodate a broad spectrum of data-enabled research activities. Baylor College of Medicine joins leading supercomputing agencies globally – the Department of Energy’s Oak Ridge and Lawrence Livermore National Labs and the U.K. government’s Science and Technology Facilities Council’s Hartree Centre – that have selected these technologies for cutting-edge HPC research.

We couldn’t be more proud to participate in the efforts of Earth Day, all year round!  Even if our role is somewhat less known in larger scheme of things, it is as equally important.  At Mellanox, we continue to improve upon the most advanced high speed interconnect solutions for both InfiniBand and Ethernet, and our technology helps to push the boundaries of science and data to achieve a better understanding of who we are while improving the quality of life for generations to come.

Mellanox and Zettar Crush World Record LOSF Performance Using ESnet OSCARS Test Circuit

In the wake of SC16, Mellanox has just broken the record for Lots of Small Files (LOSF) performance using the ESnet OSCARS test circuit measuring 70Gb/s. Preliminary results show a ten-fold performance improvement (40+Gbps) compared to the best results DOE researchers have reported thus far (4Gbps), even with TLS encryption enabled, for LOSF[i], despite the bandwidth cap and QoS limitations of the ESnet OSCARS test circuit.

At SC16, Mellanox and Zettar demonstrated real-time transfers, round-trip from SLAC to Atlanta, Georgia, and back to SLAC over a 5000-mile ESNet OSCARS loop. The two companies also exhibited real-time data transfers using two 100Gbps LAN links that show line rate performance of moving data from memory-to-memory and file-to-file between clusters. The configuration leveraged Mellanox 100Gb/s InfiniBand connectivity on the storage backend as well as Mellanox 100Gb/s Ethernet connectivity at the front end. The motivation of such endeavors is due to the fact that the next generation Linear Coherent Light Source experiment (LCLS-II) at SLAC is anticipated to achieve an event rate of 1000 times that of today’s LCLS. The majority of the data analysis will be performed at the NERSC supercomputer center at Lawrence Berkley National Laboratory, where it is essential to have a solution that is capable of supporting this distributed, data-intensive project.

Mellanox was delighted and honored to participate with this important technology demonstration that leveraged a complete state-of-the-art 100Gb/s InfiniBand and Ethernet connectivity solution. By showcasing that the foundational interconnect requirements of the LCLS-II project, we now have hard evidence that co-design and open standards is on the proper trajectory needed to drive future generation requirements for both science and data.

“Providing a scale-out data transfer solution consisting of matching software and a transfer system design will be paramount for the amount of data generated by projects such as LCLS-II,” said Dr. Chin Fang, CEO and Founder, Zettar, Inc. “Harnessing the latest capabilities of RDMA, 100Gb/s InfiniBand and Ethernet with Zettar’s scale-out data transfer solution, we can achieve the performance needed to satisfy the demands of the future data centers for the most data-intensive research such as LCLS-II, even with the formidable challenges found in LOSF transfers.”

“The rates to transfer the data to NERSC is expected to reach several hundred Gb/s soon after the project turns on in 2020 and exceed a terabyte per second by 2025,” said Dr. Les Cottrell, SLAC National Accelerator Laboratory. “This demonstration will bring to light the growing need we are experiencing for data transfer and High Performance Computing (HPC) for analysis.”

ESnet provides the high-bandwidth, reliable connections that link scientists at national laboratories, universities and other research institutions, enabling them to collaborate on some of the world’s most important scientific challenges including energy, climate science, and the origins of the universe. Funded by the DOE Office of Science, ESnet is managed and operated by the Scientific Networking Division at Lawrence Berkeley National Laboratory. As a nationwide infrastructure and DOE User Facility, ESnet provides scientists with access to unique DOE research facilities and extensive computing resources.

Zettar Inc. delivers a scale-out data transfer software and architected a data transfer cluster design that proved the feasibility of using compact and energy-efficient high-density servers for high-performance big-data transfers.   The design leverages the industry leading Mellanox 100Gb/s ConnectX-4, SwitchIB-2 and the Mellanox SN2410 Spectrum-based 48-port 25GbE + 8-port 100GbE Open Ethernet Platform switches.

Supporting Resources:

[i] See http://datasys.cs.iit.edu/publications/2013_CCGrid13-100Gbps.pdf, Figures 6 and 7

Mellanox Technologies Honored with Six HPCwire Readers’ and Editors’ Choice Awards at the Supercomputing Conference

At SC 16, Mellanox’s commitment to innovation and dedication to cultivating the HPC community has led to multiple honors, including outstanding leadership in HPC.  The Company was honored with six HPCwire Readers’ and Editors’ Choice Awards. The six awards span a variety of categories and acknowledge the company’s achievements in delivering high-performance interconnect technology that enables the highest performance, most efficient compute and storage platforms. Mellanox Vice President of Marketing, Gilad Shainer, also received an award for outstanding leadership in HPC for his individual contributions, including his eight years of service as Chairman of the HPC Advisory council and his role in the development of Co-Design architecture, to the community over the course of his career.

We were further honored to be recognized for these awards by Tom Tabor of Tabor communications and publisher of HPCwire: “HPCwire is honored to recognize Mellanox’s technological achievements and applaud Mellanox’s commitment to the HPC community,” said Tom Tabor, CEO of Tabor communications, publisher of HPCwire. “Mellanox has long been a thought leader in the HPC space, recognizing the important role it plays in research, science and global progress, and the company continues to push the innovation envelope with each passing year. This is not the first year – and I’m sure it will not be the last – that our readers and editors have recognized Mellanox’s incredible contributions to the advancement of HPC.”

Mellanox was honored with the following Reader’s Choice Awards:

  • Best Interconnect Product Or Technology: Mellanox EDR InfiniBand
  • Top Five New Products Or Technologies To Watch: Mellanox ConnectX-5
  • Outstanding Leadership in HPC: Gilad Shainer, VP of Marketing, Mellanox

Mellanox was honored with the following Editor’s Choice Awards:

  • Best HPC Collaboration Between Government & Industry:PNNL’s Center for Advanced Technology Evaluation (CENATE), a program for early evaluation of technologies, is currently assessing products from Micron Technology, Mellanox Technologies, Penguin Computing, NVIDIA, IBM, and Data Vortex.
  • Best Use Of HPC In The Cloud: The San Diego Supercomputer Center and Mellanox collaboration, which made a historical and major advancement by recording the gravitational waves from two black holes colliding millions of light-years away.
  • Best HPC Interconnect Product Or Technology: Mellanox EDR InfiniBand

The show is in full swing and Mellanox is in the thick of the action. Tonight, we host a party with some seriously funny talent, a.k.a., “Fluffy” Gabriel Iglesias. Follow us and don’t miss a minute of the action on: Twitter, Facebook, Google+, LinkedIn, and YouTube

Visit Mellanox Technologies at SC16 (November 14-17, 2016)

Visit Mellanox Technologies at SC16 (booth #2631) to learn more on the new 200G HDR InfiniBand solutions and to see the full suite of Mellanox’s end-to-end high-performance InfiniBand and Ethernet solutions.

For more information on Mellanox’s booth and speaking activities at SC16, please visit: http://www.mellanox.com/sc16/.

As you can see, we have a lot going on at Supercomputing, so stay tuned in, more blogs and news to come.