Welcome!

@CloudExpo Authors: LeanTaaS Blog, Pat Romanski, Liz McMillan, Kevin Jackson, Elizabeth White

Related Topics: @CloudExpo, Microservices Expo

@CloudExpo: Article

Hyperscale Computing Driving Small-Scale Designs

Are mid-scale offerings soon to be obsolete?

Multi-million user social networks, cloud hosting, Internet search and Big Data problems such as meteorology, complex physics and business informatics, all share one basic need - they each require incredibly large, complex and varied computer platforms. However, a common requirement across these systems is to "optimize the unit cost of computing." At this degree of hyperscale computing, the network, system, software, facility, and maintenance all add up to 10s or 100s of millions of dollars per project, and optimizations of a single element or the coordination of multiple elements can save the business millions. A good example of this holistic approach is Facebook's OpenCompute project, which saved the company 38% in efficiency and costs 24% less in build expense.

Similar to the automobile industry, where the racing technology from Indy, F1, and NASCAR end up in passenger vehicles, the hyperscale compute innovations we're seeing in juggernauts like Facebook will end up as line-item part numbers from vendors that are available to everyone. The timing couldn't be better, as solid state hard drives are becoming affordable and most enterprises are ramping up private cloud initiatives within their firms.

In a hyperscale design, premium computing constructs (like those seen in blade systems) are normally abandoned, favoring stripped down commodity designs that do the job at a fraction of the price. Because of the size of the deployment, rewriting an application to take advantage of the commodity compute fabric, or moving a task that was done in purpose-built hardware into custom software (e.g., disaster fault recovery), becomes cost-effective. Essentially, the decreased investment in hardware funds the software investments with ease. So what design elements are being abandoned in favor of hyperscale computing?

An example of the complex monolithic system that is being abandoned

  • Premium storage array networks with expensive optical connectivity and recovery features are being replaced with a mix of locally attached and network-attached storage, eliminating the heavy burden on the storage network
  • Dedicated compute, manage and storage networks are being replaced, favoring virtual LANs that reduce cabling and network costs
  • High cost per port network switching is being replaced, favoring commodity network components
  • High cost per socket blade systems are being replaced, favoring commodity compute components
  • Devices for monitoring and management are being replaced, favoring software tools and thoughtfully architected applications
  • Hot-swappable devices for high availability are being replaced, favoring streamlined hardware configuration
  • Redundant power supplies are being eliminated

The best visualization for this kind of unit cost of computing design is the Google Platform from 1998 that integrated individual parts without the purchase of machine cases.

Previously, creating the best optimized hyperscale compute fabric meant that a full staff of hardware/network/applications/systems/facilities engineers was needed to drive out the costs. Today, there are firms that are using hyperscale designs to create private cloud solutions affordable for small to medium-sized business markets or for business units in large firms. Companies working in this space aim to create the highest performance per IOP private cloud solution, delivering highly scalable infrastructure solutions.

Ideally, this architecture comes in the form of a single unit that uses converged networking, a mix of local and network-attached storage, and management software included in a small form factor. There are a handful of innovative vendors offering these solutions today. Customers adopting this type of solution enjoy an extremely low-cost commitment as a minimally configured system is capable of running a base level of virtual machines in a private and dedicated system with the potential to scale as needed. Hyperscale designs also work well in large-scale deployments, where 100,000s of virtual machines are being run.

The mid-scale cloud market, comprised of 10,000s of virtual machines, is also an interesting space. Currently, mid-market integrated private cloud offerings require large upfront costs and ongoing operational costs for dedicated staff to manage and maintain the complicated compute, storage, and networking, in addition to the expensive per socket and port hardware. Buyers in this space should certainly be asking vendors the cost per VM and the cost per terabyte of storage before they purchase, as well as determining the skills that are required to maintain an infrastructure of that kind. At this point, the mid-scale solutions look obsolete, as evolving hyperscale formats require lower cost commitments, and deliver high price performance coupled with compute, network and storage cooperation.

When discussing application considerations, hyperscale architecture is a natural platform for applications designed to leverage its key features - horizontal scalability (for high throughput and increased performance), and redundancy (for high availability and fault tolerance). Earlier hyperscale architectures, as mentioned earlier, took a different approach toward performance and reliability. Data access performance and high availability relied on premium storage array networks with expensive optical connectivity and recovery features. Compute performance relied on a high cost per socket blade system and high cost per port network switches.

The service orientation and "assumed failure" approach to cloud applications puts the burden of performance and reliability assurance on the application architecture. By constructing applications as a collection of loosely coupled services, greater performance can be achieved by distributing and replicating services horizontally across commodity compute, network, and storage components. High availability can also be achieved in a similar fashion by replicating application services across the hyperscale environment and introducing a failover mechanism to mirrored services upon service failure detection.

It's important to note several additional benefits achieved by this synergy between the hyperscale architecture and applications designed to leverage it. From a performance standpoint, system monitoring software can easily be configured to detect business policy-driven performance thresholds and automatically scale or contract services based on such policies. A similar strategy can be established for high availability policies. Should the number of redundant backup services fall below a certain threshold, additional backup services can be launched before any danger of service disruption is reached. Without going into exhaustive detail, it's clear that another hyperscale benefit is the ease in which applications and platform components can be patched and replaced without service disruption. Finally, the same mechanism by which patches are applied and platforms are replaced makes it easy to test and launch new features in line with the company's business strategy.

In conclusion, organizations across a wide variety of markets require robust servers with high density performance at an affordable entry price for all levels of businesses. A hyperscale architecture combined with well-designed applications provides enterprises with a powerful tool to operate an agile business, staying ahead of the competition and exploiting new business opportunities to its advantage.

More Stories By Lee Thompson

Lee Thompson is passionate about using cutting-edge technology to automate businesses, and was one of the key architects of E*TRADE FINANCIAL, using technology to price financial services products affordable for everyone. Lee currently brings his broad experience to Morphlabs as Chief Technology Officer, and to dev2ops.org, where he is a contributor.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@CloudExpo Stories
Digital transformation is about embracing digital technologies into a company's culture to better connect with its customers, automate processes, create better tools, enter new markets, etc. Such a transformation requires continuous orchestration across teams and an environment based on open collaboration and daily experiments. In his session at 21st Cloud Expo, Alex Casalboni, Technical (Cloud) Evangelist at Cloud Academy, explored and discussed the most urgent unsolved challenges to achieve f...
In his Opening Keynote at 21st Cloud Expo, John Considine, General Manager of IBM Cloud Infrastructure, led attendees through the exciting evolution of the cloud. He looked at this major disruption from the perspective of technology, business models, and what this means for enterprises of all sizes. John Considine is General Manager of Cloud Infrastructure Services at IBM. In that role he is responsible for leading IBM’s public cloud infrastructure including strategy, development, and offering m...
There is a huge demand for responsive, real-time mobile and web experiences, but current architectural patterns do not easily accommodate applications that respond to events in real time. Common solutions using message queues or HTTP long-polling quickly lead to resiliency, scalability and development velocity challenges. In his session at 21st Cloud Expo, Ryland Degnan, a Senior Software Engineer on the Netflix Edge Platform team, will discuss how by leveraging a reactive stream-based protocol,...
"I focus on what we are calling CAST Highlight, which is our SaaS application portfolio analysis tool. It is an extremely lightweight tool that can integrate with pretty much any build process right now," explained Andrew Siegmund, Application Migration Specialist for CAST, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that Evatronix will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Evatronix SA offers comprehensive solutions in the design and implementation of electronic systems, in CAD / CAM deployment, and also is a designer and manufacturer of advanced 3D scanners for professional applications.
SYS-CON Events announced today that Synametrics Technologies will exhibit at SYS-CON's 22nd International Cloud Expo®, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. Synametrics Technologies is a privately held company based in Plainsboro, New Jersey that has been providing solutions for the developer community since 1997. Based on the success of its initial product offerings such as WinSQL, Xeams, SynaMan and Syncrify, Synametrics continues to create and hone inn...
As many know, the first generation of Cloud Management Platform (CMP) solutions were designed for managing virtual infrastructure (IaaS) and traditional applications. But that's no longer enough to satisfy evolving and complex business requirements. In his session at 21st Cloud Expo, Scott Davis, Embotics CTO, explored how next-generation CMPs ensure organizations can manage cloud-native and microservice-based application architectures, while also facilitating agile DevOps methodology. He expla...
To get the most out of their data, successful companies are not focusing on queries and data lakes, they are actively integrating analytics into their operations with a data-first application development approach. Real-time adjustments to improve revenues, reduce costs, or mitigate risk rely on applications that minimize latency on a variety of data sources. In his session at @BigDataExpo, Jack Norris, Senior Vice President, Data and Applications at MapR Technologies, reviewed best practices to ...
"Evatronix provides design services to companies that need to integrate the IoT technology in their products but they don't necessarily have the expertise, knowledge and design team to do so," explained Adam Morawiec, VP of Business Development at Evatronix, in this SYS-CON.tv interview at @ThingsExpo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
DevOps promotes continuous improvement through a culture of collaboration. But in real terms, how do you: Integrate activities across diverse teams and services? Make objective decisions with system-wide visibility? Use feedback loops to enable learning and improvement? With technology insights and real-world examples, in his general session at @DevOpsSummit, at 21st Cloud Expo, Andi Mann, Chief Technology Advocate at Splunk, explored how leading organizations use data-driven DevOps to close th...
Digital Transformation (DX) is not a "one-size-fits all" strategy. Each organization needs to develop its own unique, long-term DX plan. It must do so by realizing that we now live in a data-driven age, and that technologies such as Cloud Computing, Big Data, the IoT, Cognitive Computing, and Blockchain are only tools. In her general session at 21st Cloud Expo, Rebecca Wanta explained how the strategy must focus on DX and include a commitment from top management to create great IT jobs, monitor ...
"WineSOFT is a software company making proxy server software, which is widely used in the telecommunication industry or the content delivery networks or e-commerce," explained Jonathan Ahn, COO of WineSOFT, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Recently, WebRTC has a lot of eyes from market. The use cases of WebRTC are expanding - video chat, online education, online health care etc. Not only for human-to-human communication, but also IoT use cases such as machine to human use cases can be seen recently. One of the typical use-case is remote camera monitoring. With WebRTC, people can have interoperability and flexibility for deploying monitoring service. However, the benefit of WebRTC for IoT is not only its convenience and interopera...
In a recent survey, Sumo Logic surveyed 1,500 customers who employ cloud services such as Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP). According to the survey, a quarter of the respondents have already deployed Docker containers and nearly as many (23 percent) are employing the AWS Lambda serverless computing framework. It’s clear: serverless is here to stay. The adoption does come with some needed changes, within both application development and operations. Tha...
Sanjeev Sharma Joins June 5-7, 2018 @DevOpsSummit at @Cloud Expo New York Faculty. Sanjeev Sharma is an internationally known DevOps and Cloud Transformation thought leader, technology executive, and author. Sanjeev's industry experience includes tenures as CTO, Technical Sales leader, and Cloud Architect leader. As an IBM Distinguished Engineer, Sanjeev is recognized at the highest levels of IBM's core of technical leaders.
Nordstrom is transforming the way that they do business and the cloud is the key to enabling speed and hyper personalized customer experiences. In his session at 21st Cloud Expo, Ken Schow, VP of Engineering at Nordstrom, discussed some of the key learnings and common pitfalls of large enterprises moving to the cloud. This includes strategies around choosing a cloud provider(s), architecture, and lessons learned. In addition, he covered some of the best practices for structured team migration an...
The dynamic nature of the cloud means that change is a constant when it comes to modern cloud-based infrastructure. Delivering modern applications to end users, therefore, is a constantly shifting challenge. Delivery automation helps IT Ops teams ensure that apps are providing an optimal end user experience over hybrid-cloud and multi-cloud environments, no matter what the current state of the infrastructure is. To employ a delivery automation strategy that reflects your business rules, making r...
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the bene...
The past few years have brought a sea change in the way applications are architected, developed, and consumed—increasing both the complexity of testing and the business impact of software failures. How can software testing professionals keep pace with modern application delivery, given the trends that impact both architectures (cloud, microservices, and APIs) and processes (DevOps, agile, and continuous delivery)? This is where continuous testing comes in. D
Most technology leaders, contemporary and from the hardware era, are reshaping their businesses to do software. They hope to capture value from emerging technologies such as IoT, SDN, and AI. Ultimately, irrespective of the vertical, it is about deriving value from independent software applications participating in an ecosystem as one comprehensive solution. In his session at @ThingsExpo, Kausik Sridhar, founder and CTO of Pulzze Systems, discussed how given the magnitude of today's application ...