Virtual Modular Switch (VMS): A Network Evolution Story – Part 2

Distributed elements, in any sector, have their basic benefits and drawbacks compared to a single large tool.  It is similar to the preference of using small aircraft over a jumbo 747 for carrying passengers between proximate airfields or to using a bus vs. multiple private cars to move a football team around.

 

In networking, the analysis between a Virtual Modular Switch (VMS) and a Modular switch is cost and performance driven. A network facilitator will prefer a solution that gets the job done at the lowest cost. Such an analysis will produce different results based on the cluster’s size. If the number of network ports required for the solution can be fitted into a single chassis based device, this means that the use of the chassis, although equipped with redundant peripheral elements such as fans and power units, is presenting a single point of failure in the network. In order to solve this, a second chassis is introduced for sharing the load and provide connectivity in case of chassis failure.

 

From a financial point of view, assuming you had a chassis of 1000 ports in full use, you need to deploy a solution of 2000 ports for high availability purposes which means a 100% price increase. Using 2/3 of the ports in the chassis will translate to 200% increase on top of the real required investment and more such examples are easy to find. Other problem with the chassis is that it comes in very few form factors so if your solution requires 501 ports while the chassis of choice supports 500, you need to add another and pay the double cost.

 

Alternatively, breaking the solution into multiple devices in a VMS gives both improved granularity in terms of port count and high availability in terms of impact from failure. In loose terms, if the VMS consists of 20 switches, the failure of a switch translates to 5% loss of network capacity. Regardless from how powerful and complicated the chassis is, this is a classic case where the strength of many tops the strength of one.

 

Ran Almog VMS Part 2

 

 

Continue reading

Mellanox Video Recap of Supercomputing Conference 2013 – Denver

We want thank everyone for joining us at SC13 in Denver, Colorado last month. We hope you had a chance to become more familiar with our end-to-end interconnect solutions for HPC.

 

Check out the  videos of the presentations given during the Mellanox Evening Event, held on November 20, 2013 in the Sheraton Denver Downtown Hotel. The event was keynoted by Eyal Waldman, President and CEO of Mellanox Technologies:

Continue reading

Symantec’s Clustered File Storage over InfiniBand

Last week (on December 9th, 2013), Symantec announced the GA of their clustered file storage (CFS). The new solution enables customers to access mission critical data and applications 400% faster than traditional Storage Area Networks (SANs) at 60% of the cost.

Faster is cheaper! Sounds like magic! How they are doing it?

Try to understand the “magic”:  It is important to understand the advantages that using SSD with high performance interconnect enable in the modern scale-out (or clustered) storage systems.   Up to now, SAN-based storage has typically been used to increase performance and provide data availability for multiple applications and clustered systems. However, with the recent high-performance applications demand, SAN vendors are trying to add SSD into the storage array itself to provide higher bandwidth and lower latency response.

Since SSDs offer an incredibly high number of IOPS and bandwidth, it is important to use the right interconnect technology and to avoid bottlenecks associated with access to storage. Old fabric, like Fibre Channel (FC) cannot cope with faster pipe demands, as 8Gb/s (or even 16Gb/s) bandwidth performance is not good enough to satisfy the applications requirements.  While 40Gb/s Ethernet may look like an alternative, InfiniBand (IB) currently supports up to 56Gb/s, with a roadmap to 100Gb/s in next year.

Continue reading

CloudNFV Proof-of-Concept Approved by ETSI ISG

Mellanox is a CloudNFV integration partner providing ConnectX-3 and ConnectX-3 PRO 10/40GbE NIC on Dell Servers

The CloudNFV team will be starting PoC execution in mid-January, reporting on our results beginning of February, and contributing four major documents to the ISG’s process through the first half of 2014.” said Tom Nolle, President of CIMI Corporation, Chief Architect of CloudNFV in his recent related blog.and enabling active high performance data center.  Telefonica and Sprint have agreed to sponsor

tomnolle_new

the CloudNFV PoC.

We’re already planning additional PoCs, some focusing on specific areas and developed by our members and some advancing the boundaries of NFV into the public and private cloud and into the world of pan-provider services and global telecommunications.

Mellanox server and storage interconnect enable telecom data plane virtual network functions with near bare metal server performance in OpenStack Cloud environment through integration to NFV Orchestration and SDN platforms.

Read more:   The CloudNFV Proof-of-Concept Was Approved by the ETSI ISG!

eran bello
Author: As a Director of Business Development at Mellanox, Eran Bello handles the business, solutions and product development and strategy for the growing Telecom and Security markets. Prior to joining Mellanox, Eran was Director of Sales and Business Development at Anobit Technologies where he was responsible for the development of the ecosystem for Anobit new Enterprise SSD business as well as portfolio introduction and business engagements with key Server OEMs, Storage Solution providers and mega datacenters. Earlier on Eran was VP of Marketing and Sales for North and Central America at Runcom Technologies, the first company to deliver Mobile WiMAX/4G End to End solution and was a member of the WiMAX/4G Forum.

The Train Has Left the Station, Open Ethernet is Happening

Authored by:   Amit Katz – Sr. Director, Product Management

Customers are tired of paying huge sums of money for Ethernet switches for no good reason. At some point, OpenFlow seemed like the way to change the networking world, but various factors such as overlay networks, changing market interests, and other unforeseen developments, it is hard to view OpenFlow today as a game-changer. While it remains a very important technology and provides a valuable mean of implementing certain functionalities, it has not created a revolution in the networking industry.

 

The real revolution that is occurring today is based on a combination of the momentum gained by the Open Compute Platform and the increasing number of switch software and hardware suppliers. Initiatives to open the switch, such as Mellanox’s Open Ethernet that was announced earlier this year, have placed us on the right path to bringing networking to where servers are today: affordable, open, and software-defined.

 

But is this revolution all about saving on cost? Not at all – cost is important but flexibility, openness, and the freedom to choose are equally important. One of the key elements in enabling vendor selection elasticity is Open Network Install Environment (ONIE), which decouples the switch hardware from its software, enabling vendors to provide something very similar to what we see in the server world: hardware without an Operating System. That means customer can buy a server with many ports and install their choice of OS on top of it. In the event that the customer wants to change the OS, the lion’s share of the investment (the hardware piece) is protected.

Continue reading

Mellanox Congratulates Yarden Gerbi

Mellanox congratulates Yarden Gerbi for winning the Gold medal in the recent Israeli Judo competition.  Mellanox will sponsor Gerbi throughout her training toward the 2016 Rio Olympic games.  Yarden Gerbi is the 2013 Judo World Champion in the under 63kg (139 lbs.) category and ranked first worldwide.  Mellanox will sponsor her as she attempts to qualify for and compete in the Olympic Games in Rio de Janeiro, Brazil.

 

Photo Credit:  Oron Kochman
Photo Credit: Oron Kochman

Continue reading

Mellanox FDR 56Gb/s InfiniBand Helps Lead SC’13 Student Cluster Competition Teams to Victory

Mellanox’s end-to-end FDR 56Gb/s InfiniBand solutions helped lead The University of Texas at Austin to victory at the SC Student Cluster Competition’s Standard Track during SC’13. Utilizing Mellanox’s FDR InfiniBand solutions, The University of Texas at Austin achieved superior application run-time and sustained performance within a 26-amp of 120-volt power limit, allowing them to complete workloads faster while achieving top benchmark performance. Special recognition was also provided to China’s National University of Defense Technology (NUDT), which through the use of Mellanox’s FDR 56Gb/s InfiniBand, won the award for highest LINPACK performance.

 

Held as part of HPC Interconnections, the SC Student Cluster Competition is designed to introduce the next generation of students to the high-performance computing community. In this real-time, non-stop, 48-hour challenge, teams of undergraduate students assembled a small cluster on the SC13 exhibit floor and raced to demonstrate the greatest sustained performance across a series of applications. The winning team was determined based on a combined score for workload completed, benchmark performance, conference attendance, and interviews.

Continue reading

Mellanox at SuperComputing Conference 2013 – Denver, CO

 

Attending the SC13 conference in Denver next week?

gfx_02086.jpg

Yes? Be sure to stop by the Mellanox booth at booth #2722 and check out the latest products, technology demonstrations, and FDR InfiniBand performance with Connect-IB!   We have a long list of theater presentations with our partners at the Mellanox booth. We will have giveaways at every presentation and a lucky attendee will go home with a new Apple iPad3 Mini at the end of each day!

Don’t forget to sign up for Mellanox Special Evening Event During SC13 on Wednesday night.  Register here:  http://www.mellanox.com/sc13/event.php

Location
Sheraton Denver Downtown Hotel
Plaza Ballroom
1550 Court Place
Denver, Colorado 80202
Phone: (303) 893-3333
  Map It  

Time:
Wednesday, November 20th
7:00PM – 10:00PM

Also download the Print ‘n Fly guide to SC13 in Denver from insideHPC!

print'nfly cover

Finally, come to hear from our experts in the SC13 sessions:

 

Speaker: Gilad Shainer, VP Marketing; Richard Graham, Sr. Solutions Architect

Title: “OpenSHMEM BoF”

Date: Wednesday, November 20, 2013

Time: 5:30PM – 7:00PM

Room: 201/203

 

Speaker: Richard Graham, Sr. Solutions Architect

Title: “Technical Paper Session Chair: Inter-Node Communication

Date: Thursday, November 21, 2013

Time: 10:30AM – 12:00PM

Room: 405/406/407

 

Speaker: Richard Graham, Sr. Solutions Architect

Title: “MPI Forum BoF”

Date: Thursday, November 21, 2013

Time: 12:15PM-1:15PM

Room: 705/707/709/711

P.S.  Stop by the Mellanox booth [2272]  to see our Jelly bean jar.  Comment on this post with your guess, and you could win a $50 Amazon Gift Card!  Winner will be announced at the end of the conference.  Follow all of our activities on our social channels including Twitter, Facebook and our Community!

Guess How Many?

 See you in Denver!

 

 

pak Author: Pak Lui is the Applications Performance Manager at Mellanox Technologies, responsible for managing the application performance, application characterization, profiling and testing. His main focus is to optimize HPC applications on products, explore new technologies and solutions and their effect on real workloads. Pak has been working in the HPC industry for over 12 years. Prior to joining Mellanox Technologies, Pak worked as a Cluster Engineer, responsible for building and testing HPC cluster configurations from different OEMs and ISVs. Pak holds a B.Sc. in Computer Systems Engineering and a M.Sc. in Computer Science from Boston University in the United States.

Mellanox-Based Clouds: A Key Ingredient for Your Start-up Success

Mellanox’s Ethernet and InfiniBand interconnects enable and enhance world-leading cloud infrastructures around the globe. Utilizing Mellanox’s fast server and storage interconnect solutions, these cloud vendors maximized their cloud efficiency and reduced their cost-per-application.

 

Mellanox is now working with a variety of incubators, accelerators, co-working spaces and venture capitalists to introduce these cloud vendors that are based on Mellanox interconnect cloud solution to new evolving startup companies.   These new companies can enjoy best performance with the added benefit of reduced cost, as they advance application development.  In this post, we will discuss the advantages of using Mellanox based clouds.

 

RDMA (Remote Direct Memory Access) is a critical element in building the most scalable and cost-effective cloud environments and to achieve the highest return-on-investment.  For example, Microsoft Azure’s InfiniBand based cloud, as listed on the world’s top performance capable systems (TOP500), demonstrated 33% lower application cost compared to other clouds on the same list.

 

Mellanox’s InfiniBand and RoCE (RDMA over Converged Ethernet) cloud solutions deliver world-leading Ethernet based interconnect density, compute and storage.  Mellanox’s Virtual Protocol Interconnect (VPI) technology incorporates both InfiniBand and Ethernet into the same solution to provide interconnect flexibility for cloud providers.

  • Higher Performance
    • 56Gb/s per port with RDMA
    • 2us for VM to VM connectivity
    • 3.5x faster VM migration
    • 6x faster storage access
  • Cost Effective Storage
    • Higher storage density with RDMA
    • Utilization of existing disk bays
  • Higher Infrastructure Efficiency
    • Support more VMs per server
    • Offload hypervisor CPU
    • Unlimited scalability
    • I/O consolidation (one wire)

 

Accelerating Cloud Performance
Accelerating Cloud Performance

 


 

Don’t waste resources worried about bringing up dedicated cloud infrastructures. Instead, keep your developers focused on developing applications that are strategic to your business. By choosing a RDMA-based cloud from one of our partners, you can be rest assured that you will have the most efficient, scalable, and cost-effective cloud platform available.

 

Learn more, click here:  Mellanox Based Clouds

eli karpilovski
Author: Eli Karpilovski manages the Cloud Market Development at Mellanox Technologies. In addition, Mr. Karpilovski serves as the Cloud Advisory Council Chairman. Mr. Karpilovski served as product manager for the HCA Software division at Mellanox Technologies. Mr. Karpilovski holds a Bachelor of Science in Engineering from the Holon Institute of Technology and a Master of Business Administration from The Open University of Israel.

Mellanox joins the CloudNFV Initiative

Today, we are pleased to announce that Mellanox has joined the CloudNFV initiative as an Integration Partner to contribute to the success of the effort and eco-system.  CloudNFV is a collaboration between its memberCloudNFV companies with a focus to prove the value of NFV.  The organization currently include the following member companies:  6WIND, CIMI, Dell, EnterpriseWeb, Overture, Qosmos, MetaSwitch and Mellanox.

“I am excited to welcome Mellanox into CloudNFV as an Integration Partner.  NFV and the cloud demand an efficient data center network and storage structure, and Mellanox is a global leader in both areas with high-availability and high-performance fabric connectivity that’s a perfect match to NFV implementations.  We’re already working to integrate Mellanox into our lab at Dell’s facilities in California, and they’ll be a key element in our public demonstration of high-performance NFV-based UCC services in December.”  Tom Nolle, President of CIMI Corporation, Chief Architect of CloudNFV

Cloud Network Functions Virtualization (NFV) is an ISG activity within ETSI, dedicated to creating an architecture to host network features and functions on general-purpose servers instead of on purpose-built network appliances or devices.  CloudNFV is a platform to test the integration of Cloud computing, SDN, NFV for Carrier Telecom Cloud.   Read more about this Initiative here.

“We are excited to join and collaborate with the CloudNFV team and contribute to this important initiative. We are integrating Mellanox’s ConnectX-3 PRO 10/40/56Gbps InfiniBand and Ethernet high performance, low latency NIC with the CloudNFV platform to enable data plane network functions running as virtual machines with near bare metal performance in an OpenStack environment through Mellanox Software Define Networking solutions” Eran Bello, Director of Business Development, Mellanox.

To find more information about Mellanox Telecom NFV solutions and CloudNV (http://www.cloudnfv.com) or to schedule a meeting during SDN & OpenFlow World Congress in Bad Homburg, Frankfurt, 15-18 October 2013 please contact Eran Bello at eranb@mellanox.com

 

eran bello
Author: As a Director of Business Development at Mellanox, Eran Bello handles the business, solutions and product development and strategy for the growing Telecom and Security markets. Prior to joining Mellanox, Eran was Director of Sales and Business Development at Anobit Technologies where he was responsible for the development of the ecosystem for Anobit new Enterprise SSD business as well as portfolio introduction and business engagements with key Server OEMs, Storage Solution providers and mega datacenters. Earlier on Eran was VP of Marketing and Sales for North and Central America at Runcom Technologies, the first company to deliver Mobile WiMAX/4G End to End solution and was a member of the WiMAX/4G Forum.