Welcome!

@CloudExpo Authors: Bob Gourley, Elizabeth White, Liz McMillan, Pat Romanski, Sematext Blog

Related Topics: @CloudExpo, Microservices Expo, Containers Expo Blog

@CloudExpo: Article

Challenges and Best Practices for Load Testing with the Cloud

Part 2: The key is to think of cloud testing as the delivery mechanism

In Part 1 I described how the cloud is revolutionizing load testing and the advantages it provides to ensure that your web applications perform well in production. We also looked at what capabilities you should seek out when selecting a load testing solution.

In Part 2, I will offer the limitations of a test strategy that relies solely on cloud-based testing, highlighting the need for a complementary internal load testing solution. I will also discuss several best practices for load testing in the cloud. Understanding how to apply the right tools and practices to make the most of the cloud is fundamental to cloud-based testing and vital to ultimately going live with total peace of mind.

The advantages to load testing with the cloud are clear, but internal testing still has its place in the overall test plan, particularly when testing from outside the firewall is not feasible. Internal testing also helps you to isolate effects that are due to your own application or infrastructure from those that are outside your firewall and potentially beyond your control.

The advantages to load testing with the cloud are clear, but internal testing still has its place in the overall test plan, particularly when testing from outside the firewall is not feasible. Internal testing also helps you to isolate effects that are due to your own application or infrastructure from those that are outside your firewall and potentially beyond your control.

Isolating Root Cause
When load testing uncovers a problem, the next step is identifying which layer in the delivery chain is causing the problem. You can use monitors to check performance metrics such has hits, average response time per request, and average bandwidth for each layer or piece of equipment in the chain. You can also employ application performance management (APM) solutions to identify bottlenecks in the code. These techniques work reasonably well when there is a single source for the performance slow-down.

When there are multiple problems, both inside and outside the firewall, it can be difficult to sort out the root causes because the symptoms are often commingled, making them difficult to isolate and resolve individually. For this reason it's important to have a cloud load testing solution that you can also apply within the firewall. You can then separate the problems that exist within the firewall from those caused by layers outside it. The ability to test a subset of the delivery chain in this way makes it much easier to find the root causes of performance problems.

Reproducing Tests
Often, you need to precisely measure the effect of changes made to the application code or settings. For example, you may need to determine the performance improvement that results from resolving a specific defect or evaluate performance for a range of cache sizes and other settings.

With cloud load testing, such precise measurements are difficult because of variations in Internet traffic and bandwidth availability at the data center level. Such variations can make it almost impossible to duplicate conditions from day to day or even within the same day.

As with isolating root causes, this situation also highlights the need for internal testing, in which you can better manage the conditions of the test, stabilize the testing environment, and take precise measurements to get more reliable performance metrics for comparison.

Conducting Small Scale Tests
Not all load testing requires hundreds of load generators. Even applications that anticipate many thousands of concurrent users are initially tested with a small population. These smaller scale tests require only a few machines may be easier and less expensive to conduct internally if the machines have already been purchased and are available for use. These tests can be carried out within the firewall to conduct tests that don't require a heavy load or the full delivery chain. To keep cloud expenses down, use available real machines when they can provide the information you need, and employ load testing with the cloud when necessary for larger scale, more realistic tests.

Testing Inside the Firewall
Of course, some testing use cases preclude the use of the cloud. If you're developing an enterprise web application that was not designed to be accessed from the Internet, then it's not a good idea to expose it outside the firewall solely for the purpose of load testing with the cloud. Likewise, if your pre-production environment is not set up to be accessed from the Internet, you'll want to have an internal testing solution that can be used within the firewall. Ideally, you want to use the same load testing solution for both internal testing and testing with the cloud, so that the scripts you developed for internal pre-production testing can be reused in production cloud-based testing. Using different tools for internal and cloud testing would not only require a rewrite of the scripts, it would also increase licensing and training costs.

Best Practices
The following best practices can help you maximize the advantages - and minimize the challenges - of load testing with the cloud.

Employ a Two-Stage Process
A two-stage process for load testing enables engineers to employ internal and cloud testing in the situations for which they are most effective and appropriate. In the first stage of the process you conduct internal tests with a medium load to quickly identify and resolve preliminary performance issues. Then increase the load incrementally with one or more load generators in the test lab. After the performance has been validated internally, proceed to the second stage, cloud-based load testing, for large scale tests that validate the entire delivery chain of the application.

This hybrid approach addresses the key challenges facing organizations that attempt to rely on testing from the cloud only:

  • It enables teams to isolate problems. The source of any performance issue identified in the first stage is clearly within the firewall (because no other systems are involved in the test). It's easier to pinpoint and fix internal problems when they are not being compounded by other issues that originate outside the firewall.
  • It enables earlier testing. With the two-stage process, you don't have to wait for the application to be deployed and accessible from the Internet to test it. You can test internally earlier in the application lifecycle when defects are easier and less expensive to fix.
  • It enables reproducible tests. With internal testing you have much more control over the environment, so you can precisely measure the effect of code or configuration changes on application performance.
  • It provides a better understanding of each major area of the delivery chain. You can compare the results of the same test scenario run internally and from the cloud to get a clearer picture of how the application server and network infrastructure contribute to overall response times.
  • It lowers costs. Cloud testing is based on a pay-per-use model. When you can test internally on hardware you already have, you can reduce the amount of testing that you need to perform from the cloud and cut costs.

Use Different Cloud Providers
There are several advantages to using multiple cloud providers. First, it helps you test from more geographical regions, which provides more realistic results that capture the effects of various third-party servers and content delivery networks. Second, it's more scalable. For exceptionally large scale tests, you can engage multiple providers simultaneously to bypass limitations that a single provider may place on bandwidth or the number of machines in use. Third, it enables you to detect potential network issues at the cloud provider level. If test results from virtually all providers show acceptable performance, but you're seeing significantly worse performance from machines on a particular provider, then you can safely conclude that there is a temporary problem with only that provider, not your application. Load testing solutions that are locked into a single provider limit the test engineer's ability to conduct realistic, reliable, large-scale tests.

Secure Your Data
In internal pre-production testing, the data used is often fake - not actual customer or user information. Further, you can be reasonably assured that any real data used is safe because testing is being conducted within the firewall. This is not the case when testing from the cloud on production data. You'll need to take steps to ensure that any accounts, scenarios, detailed error messages, and other sensitive data involved in your tests are secured.

Encrypt the communication between your controller and load generators. This helps secure data sent to the load generators during the test (including account information) as well as the data that is retrieved (including error messages). If possible, use SSL to secure the communication between the browser and the tested server. Last, ensure your load generators are secured with their own firewalls to protect them from outside threats.

Tune Load Generators
To ensure that your load generator machines in the cloud are capable of generating large loads, you must properly tune the system to support the creation of a high number of sockets and threads per process. In addition, allocate an appropriate heap size for Java-based load generators. The default settings for a typical machine allow all programs to share its resources fairly. In the case of load generators, the machine is dedicated to a single task, so you can improve performance by allocating a significantly larger share of the available resources to the load generation tasks.

Monitor Your Servers
Once you've identified a performance bottleneck, you need information to track down its root cause. This information should be gathered during the test by monitoring each component of the infrastructure including application servers and database servers. Specifically, you want to monitor both the system - including the operating system, disks, and network - and the server software - including connection pools, threads, cache hits, and indexes.

Linking all the information gathered during the tests with the tests themselves is much easier when the monitoring is integrated with your load testing tool. This enables you to correlate the response times and errors generated by load testing with the monitored data to track down the cause of problems quickly. A cloud testing solution that has no ability to monitor activity inside the firewall cannot integrate and correlate the tests it initiates from the outside with what is happening on the inside. With such a setup, test engineers will not have all the information they need to quickly identify the sources of performance problems.

Summing It Up
Even with all its potential benefits, cloud testing cannot meet all the challenges facing performance test engineers. In practice, cloud testing is most effective when combined with internal load testing in a two-stage process that makes use of multiple cloud providers and both internal and external infrastructure. You can be optimistic about using the cloud, but don't get caught up in all the hype. The key is to think of cloud testing as the delivery mechanism that is just one (albeit an important one) part of an overall load and performance test strategy.

More Stories By Steve Weisfeldt

Steve Weisfeldt is a Senior Performance Engineer at Neotys, a provider of load testing software for Web applications. Previously, he has worked as the President of Engine 1 Consulting, a services firm specializing in all facets of test automation. Prior to his involvement at Engine 1 Consulting, he was a Senior Systems Engineer at Aternity. Prior to that, Steve spent seven years at automated testing vendor Segue Software (acquired by Borland). While spending most of his time at Segue delivering professional services and training, he was also involved in pre-sales and product marketing efforts.

Being in the load and performance testing space since 1999, Steve has been involved in load and performance testing projects of all sizes, in industries that span the retail, financial services, insurance and manufacturing sectors. His expertise lies in enabling organizations to optimize their ability to develop, test and launch high-quality applications efficiently, on-time and on-budget. Steve graduated from the University of Massachusetts-Lowell with a BS in Electrical Engineering and an MS in Computer Engineering.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@CloudExpo Stories
"Once customers get a year into their IoT deployments, they start to realize that they may have been shortsighted in the ways they built out their deployment and the key thing I see a lot of people looking at is - how can I take equipment data, pull it back in an IoT solution and show it in a dashboard," stated Dave McCarthy, Director of Products at Bsquare Corporation, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
In IT, we sometimes coin terms for things before we know exactly what they are and how they’ll be used. The resulting terms may capture a common set of aspirations and goals – as “cloud” did broadly for on-demand, self-service, and flexible computing. But such a term can also lump together diverse and even competing practices, technologies, and priorities to the point where important distinctions are glossed over and lost.
What happens when the different parts of a vehicle become smarter than the vehicle itself? As we move toward the era of smart everything, hundreds of entities in a vehicle that communicate with each other, the vehicle and external systems create a need for identity orchestration so that all entities work as a conglomerate. Much like an orchestra without a conductor, without the ability to secure, control, and connect the link between a vehicle’s head unit, devices, and systems and to manage the ...
All clouds are not equal. To succeed in a DevOps context, organizations should plan to develop/deploy apps across a choice of on-premise and public clouds simultaneously depending on the business needs. This is where the concept of the Lean Cloud comes in - resting on the idea that you often need to relocate your app modules over their life cycles for both innovation and operational efficiency in the cloud. In his session at @DevOpsSummit at19th Cloud Expo, Valentin (Val) Bercovici, CTO of Soli...
Complete Internet of Things (IoT) embedded device security is not just about the device but involves the entire product’s identity, data and control integrity, and services traversing the cloud. A device can no longer be looked at as an island; it is a part of a system. In fact, given the cross-domain interactions enabled by IoT it could be a part of many systems. Also, depending on where the device is deployed, for example, in the office building versus a factory floor or oil field, security ha...
Amazon has gradually rolled out parts of its IoT offerings in the last year, but these are just the tip of the iceberg. In addition to optimizing their back-end AWS offerings, Amazon is laying the ground work to be a major force in IoT – especially in the connected home and office. Amazon is extending its reach by building on its dominant Cloud IoT platform, its Dash Button strategy, recently announced Replenishment Services, the Echo/Alexa voice recognition control platform, the 6-7 strategic...
"Qosmos has launched L7Viewer, a network traffic analysis tool, so it analyzes all the traffic between the virtual machine and the data center and the virtual machine and the external world," stated Sebastien Synold, Product Line Manager at Qosmos, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Everyone knows that truly innovative companies learn as they go along, pushing boundaries in response to market changes and demands. What's more of a mystery is how to balance innovation on a fresh platform built from scratch with the legacy tech stack, product suite and customers that continue to serve as the business' foundation. In his General Session at 19th Cloud Expo, Michael Chambliss, Head of Engineering at ReadyTalk, discussed why and how ReadyTalk diverted from healthy revenue and mor...
Without a clear strategy for cost control and an architecture designed with cloud services in mind, costs and operational performance can quickly get out of control. To avoid multiple architectural redesigns requires extensive thought and planning. Boundary (now part of BMC) launched a new public-facing multi-tenant high resolution monitoring service on Amazon AWS two years ago, facing challenges and learning best practices in the early days of the new service. In his session at 19th Cloud Exp...
As data explodes in quantity, importance and from new sources, the need for managing and protecting data residing across physical, virtual, and cloud environments grow with it. Managing data includes protecting it, indexing and classifying it for true, long-term management, compliance and E-Discovery. Commvault can ensure this with a single pane of glass solution – whether in a private cloud, a Service Provider delivered public cloud or a hybrid cloud environment – across the heterogeneous enter...
You have great SaaS business app ideas. You want to turn your idea quickly into a functional and engaging proof of concept. You need to be able to modify it to meet customers' needs, and you need to deliver a complete and secure SaaS application. How could you achieve all the above and yet avoid unforeseen IT requirements that add unnecessary cost and complexity? You also want your app to be responsive in any device at any time. In his session at 19th Cloud Expo, Mark Allen, General Manager of...
Financial Technology has become a topic of intense interest throughout the cloud developer and enterprise IT communities. Accordingly, attendees at the upcoming 20th Cloud Expo at the Javits Center in New York, June 6-8, 2017, will find fresh new content in a new track called FinTech.
@DevOpsSummit taking place June 6-8, 2017 at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @DevOpsSummit at Cloud Expo New York Call for Papers is now open.
"We are a leader in the market space called network visibility solutions - it enables monitoring tools and Big Data analysis to access the data and be able to see the performance," explained Shay Morag, VP of Sales and Marketing at Niagara Networks, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
The 20th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held June 6-8, 2017, at the Javits Center in New York City, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Containers, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal ...
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor - all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
"We are a modern development application platform and we have a suite of products that allow you to application release automation, we do version control, and we do application life cycle management," explained Flint Brenton, CEO of CollabNet, in this SYS-CON.tv interview at DevOps at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Regulatory requirements exist to promote the controlled sharing of information, while protecting the privacy and/or security of the information. Regulations for each type of information have their own set of rules, policies, and guidelines. Cloud Service Providers (CSP) are faced with increasing demand for services at decreasing prices. Demonstrating and maintaining compliance with regulations is a nontrivial task and doing so against numerous sets of regulatory requirements can be daunting task...
Bert Loomis was a visionary. This general session will highlight how Bert Loomis and people like him inspire us to build great things with small inventions. In their general session at 19th Cloud Expo, Harold Hannon, Architect at IBM Bluemix, and Michael O'Neill, Strategic Business Development at Nvidia, discussed the accelerating pace of AI development and how IBM Cloud and NVIDIA are partnering to bring AI capabilities to "every day," on-demand. They also reviewed two "free infrastructure" pr...
"This is specifically designed to accommodate some of the needs for high availability and failover in a network managed system for the major Korean corporations," stated Thomas Masters, Managing Director at InfranicsUSA, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.