Welcome!

@CloudExpo Authors: James Carlini, Yeshim Deniz, Xenia von Wedel, Elizabeth White, Nate Vickery

Related Topics: Java IoT, Linux Containers, Containers Expo Blog, @DevOpsSummit

Java IoT: Blog Feed Post

7 Spices of #ContinuousDelivery Pipeline | @DevOpsSummit #CD #DevOps

A Continuous Delivery pipeline as part of an Agile transformation is like spices in a meal

The Seven Spices of a Continuous Delivery Pipeline
By Andreas Prins

A Continuous Delivery pipeline as part of an Agile transformation is like spices in a meal. Without them, the food is bland and worthless. On the other hand, the right blend of spices will leave you craving more, stimulating your senses and energizing you. But as any good cook will tell you, it can be a bit difficult to find the exact right blend of spices for a specific dish.

Salt and pepper are usually basic requirements, but knowing how to give your creation a boost by adding more complex spices, like turmeric, star anise, ginger or coriander, is a little trickier. It requires selecting the spices with care to make the dish tasteful—and that’s exactly like choosing tools for your CD pipeline and building your pipeline up. In short, creating a Continuous Delivery pipeline is not like using a standard set of spices you store in your kitchen drawer. You must carefully choose your tools per the goals of your team.

Want to become the master chef of your Continuous Delivery pipeline? Here are 7 tips:

  1. Avoid creating monoliths
  2. Strike a balance between fixed and flexible components
  3. Treat your CD Pipeline as a value stream, not a bunch of tech tools
  4. Use the MVP approach to build and extend your pipeline
  5. Embrace a model that allows you to easily experiment
  6. Limit the number of homegrown solutions you build—don’t miss out on all the great tools already on the market
  7. Setup your CD pipeline like it’s your most critical piece of software

Avoid creating monoliths
Monolithic applications can be a useful part of backend operations. But as you increase your customer-facing applications and interactions, monoliths become difficult to handle, impeding and restricting agility. Keep in mind that most of the changes you’ll experience over the next few months and years will probably affect the monoliths. Tools grow and go, new frameworks force you to adapt, and compliance is no longer a department, but more like oxygen—it’s everywhere and crucial to surviving. You need to accept that, in increasingly agile environments, monoliths are a thing of the past, and should be avoided as you create your CD pipeline.

Strike a balance between fixed and flexible components
Every IT team in every organization whether, insurance, government or banking, faces mandatory requirements. My advice is to set up a pipeline with a dual focus.

First, create fixed processes for addressing elements of the pipeline that are mandatory for getting software into production. Examples include version control, 4-eye principle, peer review, secure code review and user access management. Think of this as the salt and pepper of your delivery process—it’s the foundation of a tasty dish.

Next, make all non-mandatory processes flexible by using a modular approach. Allow teams to choose from a set of OP tools that can be easily attached to the pipeline. Not every team, for example, requires the same performance test tool. Depending on the technology, type, moment and maturity of testing, you could give let them select from, say, a set of five tools. This gives them freedom of choice while allowing you to maintain control and reduce maintenance.

Treat your CD Pipeline as a value stream, not a bunch of tech tools
CD pipelines require care and feeding to keep them running at peak efficiency. You can’t just pull together a bunch of tools and watch your pipeline magically transform. Your pipeline is better seen as a value stream that allows you to visualize your release process, understanding its throughput times, identifying bottlenecks and so on, so you can continually optimize your delivery cycle. The shorter a release is from “Merge to master” to “Deployment in production,” the faster feedback will flow throughout your organization, and the better your ability to respond quickly to last-minute demands.

Use the MVP approach to building and extending the pipeline
Building a CD pipeline is hard work. Don’t expect to create it overnight or to onboard teams in the blink of an eye. Start small and grow with the maturity of the team. It’s like learning to cook great meals. You might start by buying some pre-packaged seasonings from the supermarket and adding water. The more knowledgeable you become, the better you get at picking your own spices, which gives you the confidence to start experimenting. Why? Because you come to understand the subtleties of flavor and the effect of certain combinations.

Embrace a model that allows for experimentation
Speaking of experimenting, if you want to become an Agile organization, your CD pipeline should be flexible enough that you can try new things. Let’s be honest, tools in this field are like Roman emperors: they rise, shine and fall, so you must be able to experiment without breaking things. For example, if you balance your pipeline between fixed and flexible components as suggested above, you could try using multiple Docker containers without destabilizing the pipeline.

Limit the number of homegrown solutions you build
With all the great tools on the market, there’s no need to start building your own CD tools. You can create robust pipelines that fit into the above principles using off-the-shelf tools.

Setup your CD Pipeline like it’s your most critical software
From time to time I still hear people say that, when faced with urgent business demands, they skip over the pipeline and place software directly into production. The argument is that the pipeline is simply too slow to meet the demand. Here’s what I suggest:

  • Make your pipeline stable enough to work in every critical situation
  • Ensure that the pipeline extends from end-to-end so you can optimize all activities towards getting working software into production
  • Make the pipeline fast enough so you’re not tempted to do everything manually

Achieving availability, integrity, reliability, and speed is definitely hard work. But if you structure your pipeline as suggested with both fixed and flexible parts, it will be a lot easier to start small and grow into a highly efficient Continuous Delivery pipeline.

About the author
Andreas Prins is facilitator and manager of several DevOps teams. He loves to think and write about topics like transforming organizations, coaching teams and speeding up the delivery process. You can read his personal blog at IdeeTotIT.nl (Dutch).

The post The 7 Spices of a Continuous Delivery Pipeline appeared first on XebiaLabs.

More Stories By XebiaLabs Blog

XebiaLabs is the technology leader for automation software for DevOps and Continuous Delivery. It focuses on helping companies accelerate the delivery of new software in the most efficient manner. Its products are simple to use, quick to implement, and provide robust enterprise technology.

@CloudExpo Stories
In his keynote at 18th Cloud Expo, Andrew Keys, Co-Founder of ConsenSys Enterprise, provided an overview of the evolution of the Internet and the Database and the future of their combination – the Blockchain. Andrew Keys is Co-Founder of ConsenSys Enterprise. He comes to ConsenSys Enterprise with capital markets, technology and entrepreneurial experience. Previously, he worked for UBS investment bank in equities analysis. Later, he was responsible for the creation and distribution of life settle...
The cloud era has reached the stage where it is no longer a question of whether a company should migrate, but when. Enterprises have embraced the outsourcing of where their various applications are stored and who manages them, saving significant investment along the way. Plus, the cloud has become a defining competitive edge. Companies that fail to successfully adapt risk failure. The media, of course, continues to extol the virtues of the cloud, including how easy it is to get there. Migrating...
Blockchain. A day doesn’t seem to go by without seeing articles and discussions about the technology. According to PwC executive Seamus Cushley, approximately $1.4B has been invested in blockchain just last year. In Gartner’s recent hype cycle for emerging technologies, blockchain is approaching the peak. It is considered by Gartner as one of the ‘Key platform-enabling technologies to track.’ While there is a lot of ‘hype vs reality’ discussions going on, there is no arguing that blockchain is b...
You know you need the cloud, but you’re hesitant to simply dump everything at Amazon since you know that not all workloads are suitable for cloud. You know that you want the kind of ease of use and scalability that you get with public cloud, but your applications are architected in a way that makes the public cloud a non-starter. You’re looking at private cloud solutions based on hyperconverged infrastructure, but you’re concerned with the limits inherent in those technologies.
Coca-Cola’s Google powered digital signage system lays the groundwork for a more valuable connection between Coke and its customers. Digital signs pair software with high-resolution displays so that a message can be changed instantly based on what the operator wants to communicate or sell. In their Day 3 Keynote at 21st Cloud Expo, Greg Chambers, Global Group Director, Digital Innovation, Coca-Cola, and Vidya Nagarajan, a Senior Product Manager at Google, discussed how from store operations and ...
In his session at 21st Cloud Expo, Raju Shreewastava, founder of Big Data Trunk, provided a fun and simple way to introduce Machine Leaning to anyone and everyone. He solved a machine learning problem and demonstrated an easy way to be able to do machine learning without even coding. Raju Shreewastava is the founder of Big Data Trunk (www.BigDataTrunk.com), a Big Data Training and consulting firm with offices in the United States. He previously led the data warehouse/business intelligence and B...
"IBM is really all in on blockchain. We take a look at sort of the history of blockchain ledger technologies. It started out with bitcoin, Ethereum, and IBM evaluated these particular blockchain technologies and found they were anonymous and permissionless and that many companies were looking for permissioned blockchain," stated René Bostic, Technical VP of the IBM Cloud Unit in North America, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Conventi...
A strange thing is happening along the way to the Internet of Things, namely far too many devices to work with and manage. It has become clear that we'll need much higher efficiency user experiences that can allow us to more easily and scalably work with the thousands of devices that will soon be in each of our lives. Enter the conversational interface revolution, combining bots we can literally talk with, gesture to, and even direct with our thoughts, with embedded artificial intelligence, whic...
The need for greater agility and scalability necessitated the digital transformation in the form of following equation: monolithic to microservices to serverless architecture (FaaS). To keep up with the cut-throat competition, the organisations need to update their technology stack to make software development their differentiating factor. Thus microservices architecture emerged as a potential method to provide development teams with greater flexibility and other advantages, such as the abili...
Product connectivity goes hand and hand these days with increased use of personal data. New IoT devices are becoming more personalized than ever before. In his session at 22nd Cloud Expo | DXWorld Expo, Nicolas Fierro, CEO of MIMIR Blockchain Solutions, will discuss how in order to protect your data and privacy, IoT applications need to embrace Blockchain technology for a new level of product security never before seen - or needed.
Leading companies, from the Global Fortune 500 to the smallest companies, are adopting hybrid cloud as the path to business advantage. Hybrid cloud depends on cloud services and on-premises infrastructure working in unison. Successful implementations require new levels of data mobility, enabled by an automated and seamless flow across on-premises and cloud resources. In his general session at 21st Cloud Expo, Greg Tevis, an IBM Storage Software Technical Strategist and Customer Solution Architec...
The use of containers by developers -- and now increasingly IT operators -- has grown from infatuation to deep and abiding love. But as with any long-term affair, the honeymoon soon leads to needing to live well together ... and maybe even getting some relationship help along the way. And so it goes with container orchestration and automation solutions, which are rapidly emerging as the means to maintain the bliss between rapid container adoption and broad container use among multiple cloud host...
Blockchain is a shared, secure record of exchange that establishes trust, accountability and transparency across business networks. Supported by the Linux Foundation's open source, open-standards based Hyperledger Project, Blockchain has the potential to improve regulatory compliance, reduce cost as well as advance trade. Are you curious about how Blockchain is built for business? In her session at 21st Cloud Expo, René Bostic, Technical VP of the IBM Cloud Unit in North America, discussed the b...
Imagine if you will, a retail floor so densely packed with sensors that they can pick up the movements of insects scurrying across a store aisle. Or a component of a piece of factory equipment so well-instrumented that its digital twin provides resolution down to the micrometer.
While some developers care passionately about how data centers and clouds are architected, for most, it is only the end result that matters. To the majority of companies, technology exists to solve a business problem, and only delivers value when it is solving that problem. 2017 brings the mainstream adoption of containers for production workloads. In his session at 21st Cloud Expo, Ben McCormack, VP of Operations at Evernote, discussed how data centers of the future will be managed, how the p...
As DevOps methodologies expand their reach across the enterprise, organizations face the daunting challenge of adapting related cloud strategies to ensure optimal alignment, from managing complexity to ensuring proper governance. How can culture, automation, legacy apps and even budget be reexamined to enable this ongoing shift within the modern software factory? In her Day 2 Keynote at @DevOpsSummit at 21st Cloud Expo, Aruna Ravichandran, VP, DevOps Solutions Marketing, CA Technologies, was jo...
"Since we launched LinuxONE we learned a lot from our customers. More than anything what they responded to were some very unique security capabilities that we have," explained Mark Figley, Director of LinuxONE Offerings at IBM, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
Nordstrom is transforming the way that they do business and the cloud is the key to enabling speed and hyper personalized customer experiences. In his session at 21st Cloud Expo, Ken Schow, VP of Engineering at Nordstrom, discussed some of the key learnings and common pitfalls of large enterprises moving to the cloud. This includes strategies around choosing a cloud provider(s), architecture, and lessons learned. In addition, he covered some of the best practices for structured team migration an...
Smart cities have the potential to change our lives at so many levels for citizens: less pollution, reduced parking obstacles, better health, education and more energy savings. Real-time data streaming and the Internet of Things (IoT) possess the power to turn this vision into a reality. However, most organizations today are building their data infrastructure to focus solely on addressing immediate business needs vs. a platform capable of quickly adapting emerging technologies to address future ...