Click here to close now.




















Welcome!

@CloudExpo Authors: Elizabeth White, Dana Gardner, Tom Lounibos, Liz McMillan, Pat Romanski

Related Topics: @CloudExpo, Microservices Expo, Containers Expo Blog

@CloudExpo: Article

Challenges and Best Practices for Load Testing with the Cloud

Part 2: The key is to think of cloud testing as the delivery mechanism

In Part 1 I described how the cloud is revolutionizing load testing and the advantages it provides to ensure that your web applications perform well in production. We also looked at what capabilities you should seek out when selecting a load testing solution.

In Part 2, I will offer the limitations of a test strategy that relies solely on cloud-based testing, highlighting the need for a complementary internal load testing solution. I will also discuss several best practices for load testing in the cloud. Understanding how to apply the right tools and practices to make the most of the cloud is fundamental to cloud-based testing and vital to ultimately going live with total peace of mind.

The advantages to load testing with the cloud are clear, but internal testing still has its place in the overall test plan, particularly when testing from outside the firewall is not feasible. Internal testing also helps you to isolate effects that are due to your own application or infrastructure from those that are outside your firewall and potentially beyond your control.

The advantages to load testing with the cloud are clear, but internal testing still has its place in the overall test plan, particularly when testing from outside the firewall is not feasible. Internal testing also helps you to isolate effects that are due to your own application or infrastructure from those that are outside your firewall and potentially beyond your control.

Isolating Root Cause
When load testing uncovers a problem, the next step is identifying which layer in the delivery chain is causing the problem. You can use monitors to check performance metrics such has hits, average response time per request, and average bandwidth for each layer or piece of equipment in the chain. You can also employ application performance management (APM) solutions to identify bottlenecks in the code. These techniques work reasonably well when there is a single source for the performance slow-down.

When there are multiple problems, both inside and outside the firewall, it can be difficult to sort out the root causes because the symptoms are often commingled, making them difficult to isolate and resolve individually. For this reason it's important to have a cloud load testing solution that you can also apply within the firewall. You can then separate the problems that exist within the firewall from those caused by layers outside it. The ability to test a subset of the delivery chain in this way makes it much easier to find the root causes of performance problems.

Reproducing Tests
Often, you need to precisely measure the effect of changes made to the application code or settings. For example, you may need to determine the performance improvement that results from resolving a specific defect or evaluate performance for a range of cache sizes and other settings.

With cloud load testing, such precise measurements are difficult because of variations in Internet traffic and bandwidth availability at the data center level. Such variations can make it almost impossible to duplicate conditions from day to day or even within the same day.

As with isolating root causes, this situation also highlights the need for internal testing, in which you can better manage the conditions of the test, stabilize the testing environment, and take precise measurements to get more reliable performance metrics for comparison.

Conducting Small Scale Tests
Not all load testing requires hundreds of load generators. Even applications that anticipate many thousands of concurrent users are initially tested with a small population. These smaller scale tests require only a few machines may be easier and less expensive to conduct internally if the machines have already been purchased and are available for use. These tests can be carried out within the firewall to conduct tests that don't require a heavy load or the full delivery chain. To keep cloud expenses down, use available real machines when they can provide the information you need, and employ load testing with the cloud when necessary for larger scale, more realistic tests.

Testing Inside the Firewall
Of course, some testing use cases preclude the use of the cloud. If you're developing an enterprise web application that was not designed to be accessed from the Internet, then it's not a good idea to expose it outside the firewall solely for the purpose of load testing with the cloud. Likewise, if your pre-production environment is not set up to be accessed from the Internet, you'll want to have an internal testing solution that can be used within the firewall. Ideally, you want to use the same load testing solution for both internal testing and testing with the cloud, so that the scripts you developed for internal pre-production testing can be reused in production cloud-based testing. Using different tools for internal and cloud testing would not only require a rewrite of the scripts, it would also increase licensing and training costs.

Best Practices
The following best practices can help you maximize the advantages - and minimize the challenges - of load testing with the cloud.

Employ a Two-Stage Process
A two-stage process for load testing enables engineers to employ internal and cloud testing in the situations for which they are most effective and appropriate. In the first stage of the process you conduct internal tests with a medium load to quickly identify and resolve preliminary performance issues. Then increase the load incrementally with one or more load generators in the test lab. After the performance has been validated internally, proceed to the second stage, cloud-based load testing, for large scale tests that validate the entire delivery chain of the application.

This hybrid approach addresses the key challenges facing organizations that attempt to rely on testing from the cloud only:

  • It enables teams to isolate problems. The source of any performance issue identified in the first stage is clearly within the firewall (because no other systems are involved in the test). It's easier to pinpoint and fix internal problems when they are not being compounded by other issues that originate outside the firewall.
  • It enables earlier testing. With the two-stage process, you don't have to wait for the application to be deployed and accessible from the Internet to test it. You can test internally earlier in the application lifecycle when defects are easier and less expensive to fix.
  • It enables reproducible tests. With internal testing you have much more control over the environment, so you can precisely measure the effect of code or configuration changes on application performance.
  • It provides a better understanding of each major area of the delivery chain. You can compare the results of the same test scenario run internally and from the cloud to get a clearer picture of how the application server and network infrastructure contribute to overall response times.
  • It lowers costs. Cloud testing is based on a pay-per-use model. When you can test internally on hardware you already have, you can reduce the amount of testing that you need to perform from the cloud and cut costs.

Use Different Cloud Providers
There are several advantages to using multiple cloud providers. First, it helps you test from more geographical regions, which provides more realistic results that capture the effects of various third-party servers and content delivery networks. Second, it's more scalable. For exceptionally large scale tests, you can engage multiple providers simultaneously to bypass limitations that a single provider may place on bandwidth or the number of machines in use. Third, it enables you to detect potential network issues at the cloud provider level. If test results from virtually all providers show acceptable performance, but you're seeing significantly worse performance from machines on a particular provider, then you can safely conclude that there is a temporary problem with only that provider, not your application. Load testing solutions that are locked into a single provider limit the test engineer's ability to conduct realistic, reliable, large-scale tests.

Secure Your Data
In internal pre-production testing, the data used is often fake - not actual customer or user information. Further, you can be reasonably assured that any real data used is safe because testing is being conducted within the firewall. This is not the case when testing from the cloud on production data. You'll need to take steps to ensure that any accounts, scenarios, detailed error messages, and other sensitive data involved in your tests are secured.

Encrypt the communication between your controller and load generators. This helps secure data sent to the load generators during the test (including account information) as well as the data that is retrieved (including error messages). If possible, use SSL to secure the communication between the browser and the tested server. Last, ensure your load generators are secured with their own firewalls to protect them from outside threats.

Tune Load Generators
To ensure that your load generator machines in the cloud are capable of generating large loads, you must properly tune the system to support the creation of a high number of sockets and threads per process. In addition, allocate an appropriate heap size for Java-based load generators. The default settings for a typical machine allow all programs to share its resources fairly. In the case of load generators, the machine is dedicated to a single task, so you can improve performance by allocating a significantly larger share of the available resources to the load generation tasks.

Monitor Your Servers
Once you've identified a performance bottleneck, you need information to track down its root cause. This information should be gathered during the test by monitoring each component of the infrastructure including application servers and database servers. Specifically, you want to monitor both the system - including the operating system, disks, and network - and the server software - including connection pools, threads, cache hits, and indexes.

Linking all the information gathered during the tests with the tests themselves is much easier when the monitoring is integrated with your load testing tool. This enables you to correlate the response times and errors generated by load testing with the monitored data to track down the cause of problems quickly. A cloud testing solution that has no ability to monitor activity inside the firewall cannot integrate and correlate the tests it initiates from the outside with what is happening on the inside. With such a setup, test engineers will not have all the information they need to quickly identify the sources of performance problems.

Summing It Up
Even with all its potential benefits, cloud testing cannot meet all the challenges facing performance test engineers. In practice, cloud testing is most effective when combined with internal load testing in a two-stage process that makes use of multiple cloud providers and both internal and external infrastructure. You can be optimistic about using the cloud, but don't get caught up in all the hype. The key is to think of cloud testing as the delivery mechanism that is just one (albeit an important one) part of an overall load and performance test strategy.

More Stories By Steve Weisfeldt

Steve Weisfeldt is a Senior Performance Engineer at Neotys, a provider of load testing software for Web applications. Previously, he has worked as the President of Engine 1 Consulting, a services firm specializing in all facets of test automation. Prior to his involvement at Engine 1 Consulting, he was a Senior Systems Engineer at Aternity. Prior to that, Steve spent seven years at automated testing vendor Segue Software (acquired by Borland). While spending most of his time at Segue delivering professional services and training, he was also involved in pre-sales and product marketing efforts.

Being in the load and performance testing space since 1999, Steve has been involved in load and performance testing projects of all sizes, in industries that span the retail, financial services, insurance and manufacturing sectors. His expertise lies in enabling organizations to optimize their ability to develop, test and launch high-quality applications efficiently, on-time and on-budget. Steve graduated from the University of Massachusetts-Lowell with a BS in Electrical Engineering and an MS in Computer Engineering.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@CloudExpo Stories
SYS-CON Events announced today that VividCortex, the monitoring solution for the modern data system, will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. The database is the heart of most applications, but it’s also the part that’s hardest to scale, monitor, and optimize even as it’s growing 50% year over year. VividCortex is the first unified suite of database monitoring tools specifically desi...
In their session at 17th Cloud Expo, Hal Schwartz, CEO of Secure Infrastructure & Services (SIAS), and Chuck Paolillo, CTO of Secure Infrastructure & Services (SIAS), provide a study of cloud adoption trends and the power and flexibility of IBM Power and Pureflex cloud solutions. In his role as CEO of Secure Infrastructure & Services (SIAS), Hal Schwartz provides leadership and direction for the company.
Container technology is sending shock waves through the world of cloud computing. Heralded as the 'next big thing,' containers provide software owners a consistent way to package their software and dependencies while infrastructure operators benefit from a standard way to deploy and run them. Containers present new challenges for tracking usage due to their dynamic nature. They can also be deployed to bare metal, virtual machines and various cloud platforms. How do software owners track the usag...
As organizations shift towards IT-as-a-service models, the need for managing and protecting data residing across physical, virtual, and now cloud environments grows with it. CommVault can ensure protection and E-Discovery of your data – whether in a private cloud, a Service Provider delivered public cloud, or a hybrid cloud environment – across the heterogeneous enterprise. In his session at 17th Cloud Expo, Randy De Meno, Chief Technologist - Windows Products and Microsoft Partnerships at Com...
Mobile, social, Big Data, and cloud have fundamentally changed the way we live. “Anytime, anywhere” access to data and information is no longer a luxury; it’s a requirement, in both our personal and professional lives. For IT organizations, this means pressure has never been greater to deliver meaningful services to the business and customers.
SYS-CON Events announced today that MobiDev, a software development company, will exhibit at the 17th International Cloud Expo®, which will take place November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. MobiDev is a software development company with representative offices in Atlanta (US), Sheffield (UK) and Würzburg (Germany); and development centers in Ukraine. Since 2009 it has grown from a small group of passionate engineers and business managers to a full-scale mobi...
There are many considerations when moving applications from on-premise to cloud. It is critical to understand the benefits and also challenges of this migration. A successful migration will result in lower Total Cost of Ownership, yet offer the same or higher level of robustness. In his session at 15th Cloud Expo, Michael Meiner, an Engineering Director at Oracle, Corporation, analyzed a range of cloud offerings (IaaS, PaaS, SaaS) and discussed the benefits/challenges of migrating to each offe...
"We've just seen a huge influx of new partners coming into our ecosystem, and partners building unique offerings on top of our API set," explained Seth Bostock, Chief Executive Officer at IndependenceIT, in this SYS-CON.tv interview at 16th Cloud Expo, held June 9-11, 2015, at the Javits Center in New York City.
Digital Transformation is the ultimate goal of cloud computing and related initiatives. The phrase is certainly not a precise one, and as subject to hand-waving and distortion as any high-falutin' terminology in the world of information technology. Yet it is an excellent choice of words to describe what enterprise IT—and by extension, organizations in general—should be working to achieve. Digital Transformation means: handling all the data types being found and created in the organizat...
SYS-CON Events announced today that HPM Networks will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. For 20 years, HPM Networks has been integrating technology solutions that solve complex business challenges. HPM Networks has designed solutions for both SMB and enterprise customers throughout the San Francisco Bay Area.
Chuck Piluso presented a study of cloud adoption trends and the power and flexibility of IBM Power and Pureflex cloud solutions. Prior to Secure Infrastructure and Services, Mr. Piluso founded North American Telecommunication Corporation, a facilities-based Competitive Local Exchange Carrier licensed by the Public Service Commission in 10 states, serving as the company's chairman and president from 1997 to 2000. Between 1990 and 1997, Mr. Piluso served as chairman & founder of International Te...
The Software Defined Data Center (SDDC), which enables organizations to seamlessly run in a hybrid cloud model (public + private cloud), is here to stay. IDC estimates that the software-defined networking market will be valued at $3.7 billion by 2016. Security is a key component and benefit of the SDDC, and offers an opportunity to build security 'from the ground up' and weave it into the environment from day one. In his session at 16th Cloud Expo, Reuven Harrison, CTO and Co-Founder of Tufin,...
With SaaS use rampant across organizations, how can IT departments track company data and maintain security? More and more departments are commissioning their own solutions and bypassing IT. A cloud environment is amorphous and powerful, allowing you to set up solutions for all of your user needs: document sharing and collaboration, mobile access, e-mail, even industry-specific applications. In his session at 16th Cloud Expo, Shawn Mills, President and a founder of Green House Data, discussed h...
One of the hottest areas in cloud right now is DRaaS and related offerings. In his session at 16th Cloud Expo, Dale Levesque, Disaster Recovery Product Manager with Windstream's Cloud and Data Center Marketing team, will discuss the benefits of the cloud model, which far outweigh the traditional approach, and how enterprises need to ensure that their needs are properly being met.
The Internet of Everything (IoE) brings together people, process, data and things to make networked connections more relevant and valuable than ever before – transforming information into knowledge and knowledge into wisdom. IoE creates new capabilities, richer experiences, and unprecedented opportunities to improve business and government operations, decision making and mission support capabilities.
Explosive growth in connected devices. Enormous amounts of data for collection and analysis. Critical use of data for split-second decision making and actionable information. All three are factors in making the Internet of Things a reality. Yet, any one factor would have an IT organization pondering its infrastructure strategy. How should your organization enhance its IT framework to enable an Internet of Things implementation? In his session at @ThingsExpo, James Kirkland, Red Hat's Chief Arch...
Malicious agents are moving faster than the speed of business. Even more worrisome, most companies are relying on legacy approaches to security that are no longer capable of meeting current threats. In the modern cloud, threat diversity is rapidly expanding, necessitating more sophisticated security protocols than those used in the past or in desktop environments. Yet companies are falling for cloud security myths that were truths at one time but have evolved out of existence.
Public Cloud IaaS started its life in the developer and startup communities and has grown rapidly to a $20B+ industry, but it still pales in comparison to how much is spent worldwide on IT: $3.6 trillion. In fact, there are 8.6 million data centers worldwide, the reality is many small and medium sized business have server closets and colocation footprints filled with servers and storage gear. While on-premise environment virtualization may have peaked at 75%, the Public Cloud has lagged in adop...
The time is ripe for high speed resilient software defined storage solutions with unlimited scalability. ISS has been working with the leading open source projects and developed a commercial high performance solution that is able to grow forever without performance limitations. In his session at Cloud Expo, Alex Gorbachev, President of Intelligent Systems Services Inc., shared foundation principles of Ceph architecture, as well as the design to deliver this storage to traditional SAN storage co...
MuleSoft has announced the findings of its 2015 Connectivity Benchmark Report on the adoption and business impact of APIs. The findings suggest traditional businesses are quickly evolving into "composable enterprises" built out of hundreds of connected software services, applications and devices. Most are embracing the Internet of Things (IoT) and microservices technologies like Docker. A majority are integrating wearables, like smart watches, and more than half plan to generate revenue with ...