Welcome!

Cloud Expo Authors: Elizabeth White, Lisa Pope, Pat Romanski, Yeshim Deniz, Liz McMillan

Related Topics: Cloud Expo, Virtualization

Cloud Expo: Article

Five Capacity Management Challenges for Private Clouds

Capacity management enables cloud operators to maximize compute cycles delivered to a customer at the lowest possible cost

Organizations that are seeking to deploy cloud-based business models for their infrastructure face unique capacity management challenges. This article will review these challenges to enable cloud providers, either public or private, to avoid the pitfalls of improper capacity management.  While the article discusses both types of clouds, the needs of private cloud providers will be especially highlighted due to the unique challenges they face with this business model.

IT Becomes a Business within a Business
For years, we have been hearing "IT must behave more like a business." A hypervisor's ability to deliver utility computing moves this vision closer to reality. Using hypervisors to deploy an infrastructure cloud fundamentally changes the relationship between IT and their customers. Application portability combined with competing cloud offerings from companies like Amazon change the dynamic between application owners and corporate IT. If end users can't explicitly move their applications between cloud providers, they can at least compare prices and service levels between providers. Right or wrong, Amazon S3 becomes a measuring stick in price, performance and service for IT organizations.

To deliver an Amazon S3 experience requires significant retooling in IT processes. Capacity management is one of the areas requiring retooling. Virtualization alone causes changes to capacity management (for more information, see http://www.vkernel.com/solutions/capacity-planning). But virtualization deployed as an infrastructure cloud adds nuances to the capacity management problem. Capacity management for cloud providers is unique for the following five reasons:

  • Capacity monitoring in addition to planning
  • Chargeback is mandatory
  • Efficiency drives return on assets
  • Tenant reporting requirements are unique
  • Optimization is a value add

Variable Demand Drives Criticality of Capacity Monitoring
Cloud deployments of virtualization technology introduce many operational changes for IT administrators. The first is a change in the amount of control IT has over the loads deployed on their hardware. With cloud deployments, either public or private, end users deploy applications using self-service portals as they see fit, load them as they desire and consume resources at whatever pace they need.

Hence, unlike the careful P2V sizing process undertaken for the first wave of virtualization where applications are sized, scheduled and deployed in a methodical manner, clouds have no careful sizing or timing that the cloud operator is aware of. Applications of unknown sizes appear, consume resources, and may just as quickly disappear. Without adequate capacity, these applications will fail to perform to customer expectations. Without real-time capacity monitoring, application deployment can dramatically impact other applications.

Hence capacity planning, a well-thought-out process for making sure there is sufficient capacity en masse, must be supplemented with capacity monitoring. Capacity monitoring is a real-time process that takes raw performance and utilization data and transforms it into actionable information concerning system-level capacity requirements. Without capacity monitoring, system administrators are left to interpret real-time utilization metrics from individual virtual machines. VKernel's research has shown that properly monitoring capacity in real time involves collecting over 20 metrics per VM at least 10 times per hour, and keeping this information for at least 30 days. A 100 VM environment would require about 17 million data points to accurately monitor capacity in the environment. This capacity monitoring, however, is a must-have to augment standard capacity planning and prevent performance issues from impacting the cloud.

Chargeback Matters
For any cloud where resources can be deployed in a self-service fashion, charging back for resources becomes a necessity. Without a method to chargeback or show back, self-service clouds would quickly find themselves at capacity since resources are essentially free.

But chargeback is a tricky area. For commercial cloud and private cloud operators, charging back for allocated resources is fairly straightforward. But, since the private chargeback operator is simply shifting costs insides the company and not impacting the bottom line, the motivations for chargeback are different. The public cloud operator is indifferent to allocated resources and utilized resources. If the public cloud operator is charging for an allocated resource pool, they make money. In fact, the higher the ratio between allocated and utilized, the more over allocation of resources is possible and the higher the profit margins. For private cloud operators, however, the goal is to actually lower the costs for the company. Hence, the private cloud operator wants the allocated resource usage to be very close to actual usage to drive resource efficiency. Highlighting the difference between actual resource usage and allocated resource usage shows internal business units the amount of corporate resources they are wasting. This motivation can then be used to right size environments and reduce overall IT spend.

While chargeback is important, cloud operators need to be mindful of what they charge. For public operators, there are competitive pressures. For private operators, charging provides a way to directly compare internal IT costs vs. external costs such as Amazon.

Is a simple $/CPU comparison between internal clouds and Amazon a fair comparison? Does Amazon contain the same level of compliance? Of control? Is the company comfortable with data outside the company firewall?

More important, chargeback for private cloud operators is primarily a means to minimize the difference between allocated and utilized resources to drive up efficiencies and VM densities. Chargeback or showback becomes a control mechanism as opposed to an actual financial transfer mechanism. Hence the rate of chargeback is not as important as the difference between allocated and actual usage.

Even here, the challenges for private cloud operators are greater. Let's assume a private cloud operator hosts 200 internal customers. Assume each of these internal customers is wasting 50% of their resource allocation. On an individual basis, the absolute value of the wasted resources may be insignificant. But across all 200 customers, the magnitude of the IT spend could be quite large. The greater good theory for IT would require that IT actually reduce resource usage for all the internal customers to claim additional savings for the company despite what the internal customers want. Private cloud operators must operate not only for their internal customers' needs, but also for the company's needs.

Setting rates for chargeback is the final tricky area for cloud operators. For a public cloud operator, the rate needs to be competitive, provide some profit margin and match customer's value. Easy enough. But what about private cloud operators? Once again, being a private operator makes things difficult. What are the rates for chargeback for a private cloud operator? Is the goal to set rates to make an internal profit when 50% of the VMs slots are filled? 75%? But if the internal cost center is making a profit, is that the right thing to do from a budgeting standpoint? Is the goal cost reclamation or efficiency?

The net of this is that for cloud operators, chargeback is critical. For private cloud operators, chargeback's purpose needs to be clearly defined to align IT not only with their customer's goals, but also the broader corporate goals.

Capacity Planning Impacts Revenue and Cost
Customers expecting to use a cloud service have high expectations with regards to time to deploy a service. For public clouds, this expectation will be a nearly instant deployment after the service request. Private cloud operators may not have quite an instantaneous expectation for their customers. Either way, the "acceptable" wait time of several weeks to deploy a new server is gone. Immediate is the word, not days.

To enable this immediate capability, sufficient capacity must be on hand to deploy new virtual machines based on both steady state and unexpected increases in demands. To meet this accelerated time duration, a high degree of capacity planning must take place to predict future capacity needs ahead of demand and allow for the slower process of procuring and installing physical servers, networks and storage.

It's easy to meet rapid deployment expectations by over procuring hardware. The danger here is that over procurement impacts cash flows and profitability for a cloud. Having large amounts of depreciating assets sitting around is not a sound business strategy. If these assets are plugged in and configured, the added power costs worsen an already bad situation. Under procuring hardware is just as bad since cloud operators will be unable to meet customer needs should their systems not have available capacity. Hence the goal is to have a solid understanding of consumption and then apply a safety factor to allow for unexpected demand.

Understanding capacity needs across the entire IT infrastructure is important. But it's just as important to understand where there are available VM slots for the best placement of VMs from a performance and a capacity standpoint. Utilizing available VM slot reporting ensures the performance of the running VMs and increases the VM density per host, which is critical to achieving a high return on assets.

Capacity planning is critical to cloud operators to generate a high return on assets while also meeting customer demand for near instantaneous deployment requests.

Tenant Reporting
With a standard virtualized environment, the IT organization may report on environment status to a few senior IT leaders. Not so with an infrastructure cloud. For public clouds and private clouds, there is a greater expectation of visibility into the environment. Reporting for cloud tenants could involve availability, resource allocation, resource utilization, current charges, and pricing plan. The amount of information revealed to a customer could depend on the business philosophy or type of cloud. For a private cloud, revealing differences between allocated resources and utilized resources and the savings a customer could achieve by reducing their resources allocation makes sense for a company trying to save money. For a public cloud provider, however, suggesting ways to reduce resource allocations may not be in the provider's interest.

Beyond questions around what type of information to provide is the manner in which information is provided. For public cloud operators, online portals are most likely the reporting distribution mechanism of choice. For private clouds, however, information needs to flow seamlessly into the enterprises existing reporting infrastructure. This could involve connections with SharePoint, with IT service management frameworks, internal portals, or simple email distribution of reports on an ongoing basis.

Optimization Is a Value Add
Many times, when virtual machines are first deployed, they are over allocated CPU, memory and storage. In a cloud deployment, this over allocation does not cost the cloud operator revenue. In fact, just the opposite occurs. The cloud operator deploys the requested resources, but despite not being used, the cloud operator still collects the revenue. As discussed earlier, the cloud operator can decide whether to reveal this to the tenant or not.

While the cloud operator may not care about wasted resources to an application, the end customer does as does the firm's CFO. Decreasing requested CPU, memory and storage reduces monthly tenant costs. Hence, optimization becomes an optional value-add service for the provider to offer tenants to reduce spending. This would be the equivalent of your cell phone company contacting you to suggest a lower monthly plan. While it lowers the monthly revenue of the cloud operator, it dramatically improves customer loyalty.

Conclusion
Because of the increased need for chargeback, monitoring, capacity planning, and reporting, capacity management takes on added urgency for cloud operators. Capacity management performed correctly enables cloud operators to maximize compute cycles delivered to a customer at the lowest possible cost and, thus, capacity management is a key building block for any cloud implementation.

More Stories By Bryan Semple

Bryan Semple is Chief Marketing Officer at VKernel. A 15+ year high-tech veteran, he has spent the last 8 years working in server and storage companies focused on virtualization technologies. He comes to VKernel from NetApp where he was the general manager of the storage virtualization business unit. Under his leadership, the group experienced record growth, expanded engineering operations to India, and built global awareness for NetApp’s industry leading storage virtualization solutions.

Prior to NetApp, Bryan was VP of Marketing at Onaro where he established the company as a leader in storage management software and built the marketing processes that supported the company’s profitability and successful acquisition by NetApp in 2008. Before Onaro, he was the VP of Product Marketing and Strategy at server blade virtualization pioneer Egenera. At Egenera, he worked with early adopters of infrastructure and server virtualization technologies in the financial services industry as the company scaled from one to several hundred customers.

Bryan holds a BS in Systems Engineering from the US Naval Academy and an MBA from Stanford University.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@CloudExpo Stories
Leysin American School is an exclusive, private boarding school located in Leysin, Switzerland. Leysin selected an OpenStack-powered, private cloud as a service to manage multiple applications and provide development environments for students across the institution. Seeking to meet rigid data sovereignty and data integrity requirements while offering flexible, on-demand cloud resources to users, Leysin identified OpenStack as the clear choice to round out the school's cloud strategy. Additional...
Explosive growth in connected devices. Enormous amounts of data for collection and analysis. Critical use of data for split-second decision making and actionable information. All three are factors in making the Internet of Things a reality. Yet, any one factor would have an IT organization pondering its infrastructure strategy. How should your organization enhance its IT framework to enable an Internet of Things implementation? In his session at Internet of @ThingsExpo, James Kirkland, Chief Ar...
Technology is enabling a new approach to collecting and using data. This approach, commonly referred to as the "Internet of Things" (IoT), enables businesses to use real-time data from all sorts of things including machines, devices and sensors to make better decisions, improve customer service, and lower the risk in the creation of new revenue opportunities. In his General Session at Internet of @ThingsExpo, Dave Wagstaff, Vice President and Chief Architect at BSQUARE Corporation, discuss the ...
The security devil is always in the details of the attack: the ones you've endured, the ones you prepare yourself to fend off, and the ones that, you fear, will catch you completely unaware and defenseless. The Internet of Things (IoT) is nothing if not an endless proliferation of details. It's the vision of a world in which continuous Internet connectivity and addressability is embedded into a growing range of human artifacts, into the natural world, and even into our smartphones, appliances, a...
The major cloud platforms defy a simple, side-by-side analysis. Each of the major IaaS public-cloud platforms offers their own unique strengths and functionality. Options for on-site private cloud are diverse as well, and must be designed and deployed while taking existing legacy architecture and infrastructure into account. Then the reality is that most enterprises are embarking on a hybrid cloud strategy and programs. In this Power Panel at 15th Cloud Expo (http://www.CloudComputingExpo.com...
"BSQUARE is in the business of selling software solutions for smart connected devices. It's obvious that IoT has moved from being a technology to being a fundamental part of business, and in the last 18 months people have said let's figure out how to do it and let's put some focus on it, " explained Dave Wagstaff, VP & Chief Architect, at BSQUARE Corporation, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
The 4th International DevOps Summit, co-located with16th International Cloud Expo – being held June 9-11, 2015, at the Javits Center in New York City, NY – announces that its Call for Papers is now open. Born out of proven success in agile development, cloud computing, and process automation, DevOps is a macro trend you cannot afford to miss. From showcase success stories from early adopters and web-scale businesses, DevOps is expanding to organizations of all sizes, including the world's large...
Verizon Enterprise Solutions is simplifying the cloud-purchasing experience for its clients, with the launch of Verizon Cloud Marketplace, a key foundational component of the company's robust ecosystem of enterprise-class technologies. The online storefront will initially feature pre-built cloud-based services from AppDynamics, Hitachi Data Systems, Juniper Networks, PfSense and Tervela. Available globally to enterprises using Verizon Cloud, Verizon Cloud Marketplace provides a one-stop shop fo...
Software-driven innovation is becoming a primary approach to how businesses create and deliver new value to customers. A survey of 400 business and IT executives by the IBM Institute for Business Value showed businesses that are more effective at software delivery are also more profitable than their peers nearly 70 percent of the time (1). DevOps provides a way for businesses to remain competitive, applying lean and agile principles to software development to speed the delivery of software that ...

ARMONK, N.Y., Nov. 20, 2014 /PRNewswire/ --  IBM (NYSE: IBM) today announced that it is bringing a greater level of control, security and flexibility to cloud-based application development and delivery with a single-tenant version of Bluemix, IBM's

"Our premise is Docker is not enough. That's not a bad thing - we actually love Docker. At ActiveState all our products are based on open source technology and Docker is an up-and-coming piece of open source technology," explained Bart Copeland, President & CEO of ActiveState Software, in this SYS-CON.tv interview at DevOps Summit at Cloud Expo®, held Nov 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
DevOps Summit 2015 New York, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete...
Infor has announced a new feature Infor CloudSuite™ Aerospace & Defense (A&D) to aid compliance with International Traffic in Arms Regulations (ITAR). The ITAR function will serve as a complementary function for new or existing Infor CloudSuite A&D customers, to facilitate compliance for Infor customers that are creating a US defense article or performing a US defense service and wish to benefit from cloud-services. The ITAR regulation serves to manage handling and access requirements for dat...
What do a firewall and a fortress have in common? They are no longer strong enough to protect the valuables housed inside. Like the walls of an old fortress, the cracks in the firewall are allowing the bad guys to slip in - unannounced and unnoticed. By the time these thieves get in, the damage is already done and the network is already compromised. Intellectual property is easily slipped out the back door leaving no trace of forced entry. If we want to reign in on these cybercriminals, it's hig...
Bit6 today issued a challenge to the technology community implementing Web Real Time Communication (WebRTC). To leap beyond WebRTC’s significant limitations and fully leverage its underlying value to accelerate innovation, application developers need to consider the entire communications ecosystem.
The 3rd International @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The Internet of Things (IoT) is the most profound change in personal and enterprise IT since the creation of the Worldwide Web more than 20 years ago. All major researchers estimate there will be tens of billions devices - computers, smartphones, tablets, and sensors - connected to th...
The Internet of Things is not new. Historically, smart businesses have used its basic concept of leveraging data to drive better decision making and have capitalized on those insights to realize additional revenue opportunities. So, what has changed to make the Internet of Things one of the hottest topics in tech? In his session at @ThingsExpo, Chris Gray, Director, Embedded and Internet of Things, discussed the underlying factors that are driving the economics of intelligent systems. Discover ...
Cochlear Limited, the global leader in implantable hearing solutions, has selected AppZero for easy migration of its Microsoft Windows Server 2003 applications, a major priority for IT organizations before Microsoft ends support on July 14, 2015. AppZero software enables server application migration from old operating systems to new platforms or clouds and has been proven to be ten times faster, more reliable and efficient than alternative approaches.
Fundamentally, SDN is still mostly about network plumbing. While plumbing may be useful to tinker with, what you can do with your plumbing is far more intriguing. A rigid interpretation of SDN confines it to Layers 2 and 3, and that's reasonable. But SDN opens opportunities for novel constructions in Layers 4 to 7 that solve real operational problems in data centers. "Data center," in fact, might become anachronistic - data is everywhere, constantly on the move, seemingly always overflowing. Net...
SYS-CON Media announced today the IBM, which offers the world’s deepest portfolio of technologies and expertise that are transforming the future of work, has launched ad campaigns on SYS-CON’s numerous online magazines such as Cloud Computing Journal, DevOps Journal, Virtualization Journal, and IoT Journal. IBM's campaigns focus on application testing, improving application development processes, common challenges in testing composite applications, continuous testing as part of the DevOps lifec...