Welcome!

@CloudExpo Authors: Elizabeth White, Craig Lowell, Pat Romanski, Carmen Gonzalez, ManageEngine IT Matters

Blog Feed Post

Why Workflow and BPM Suck

The verities and the balderdash the impact of the cloud

I originally wrote this paper back in 2005 as a bit of a rant against the positioning of Workflow and BPM. I was reminded of it the other day and took another look only to discover that things still haven’t changed that much. So I’ve decided to revamp it a bit to encompass cloudy type things and what the impact of social media etc has had in the ensuing years. So for your amusement or edification here’s a revised version.

Many of us that were involved in the field of Workflow Automation and Business Process Management (BPM) a few years ago (and some still are I’m sure) argued long and hard about where the two technologies overlapped, where they were different, which mathematical models should be used, which standards were applicable to which part of the technology stack and all that associated puff.

Well these arguments and discussions are well and truly over more or less forgotten; the demarcation lines were well defined and drawn; the road ahead became clear.

The fact that Business Process Management has its roots in Workflow technology is well known – many of today’s leading products are, in fact, evolutions of the original forms processing packages. So there is no longer a need to debate, what is now, a moot point.

But what has happened is that BPM also changed. Rather than being an extension of workflow concepts BPM was seen as systems-to-systems technology exclusively used in the deployment of concepts such as SOA solutions. I’m over simplifying things I know, but it seemed as though BPM was destined to become an IT Technology solution as opposed to the business process solution it was meant to be. Somewhere along the way, one of the key elements in a business process – a person – dropped off the agenda. The fact that the majority of business processes (some 85% according to some very old Forrester research) involve carbon based resources was overlooked – think BPEL for a moment – doesn’t the attempt to develop that particular standard tell you something about the general direction of BPM? But be warned, even today, many vendors will tell you that their BPM products support Human interaction, but what they are talking about will be simple work item handling and form filling – this is a long way from the collaboration and interaction management we will talk about below.

The problem stems from the fact that most Workflow products were flawed and as a result, the problem in the gene pool rippled through to the evolved BPM species. So what was wrong with workflow? It’s quite simple when you think about it; most workflow products assumed that work moved from one resource to another. One user entered the loan details, another approved it. But business doesn’t work like that.

This flawed thinking is probably the main reason why workflow was never quite the success most pundits thought it would be; the solutions were just not flexible enough, since the majority of processes are unsuited to this way of working. Paradoxically, it is the exact reason why BPM is so suited to the world of SOA and systems to systems processes. A rigid approach to systems processes is essential, where people are concerned; the name of the game is flexibility.

Why do we need the flexibility?

Let’s take a simple analogy so that the concept is more easily understood.

Supposing you were playing golf; using the BPM approach would be like hitting a hole in one every time you tee off. Impressive – 18 shots, and a round finished in 25 minutes.

But as we all know, the reality is somewhat different (well my golf is different) – there’s a lot that happens between teeing off and finishing a hole. Ideally about four shots (think nodes in a process) – but you have to deal with the unexpected even though you know the unexpected is very likely; sand traps, water hazards, lost balls, free drops, collaboration with fellow players, unexpected consultation with the referee – and so it goes on. Then there are 17 more holes to do – the result is an intricate and complex process with 18 targets but about 72 operations.

As mentioned earlier, we have to deal with the unexpected. This is not just about using a set of tools to deal with every anticipated business outcome or rule; we are talking about the management of true interaction that takes place between individuals and groups which cannot be predicted or encapsulated beforehand. This is because Business Processes exist at 2 levels – the predictable (the systems) and the un-predictable (the people).

The predictable aspects of the process are easily and well catered for by BPMS solutions – which is why the term Business Process Management is a misnomer since the perceived technology only addresses the integration aspects – with the close coupling with SOA (SOA needs BPM, the converse is not true) there still iis an argument for renaming BPM to Services Process Management (SPM).

Proposals such as BPEL4People didn’t fix the problem either, all that managed to achieve is replicating the shortcomings of Workflow. Anyone who has tried to put together a business case for buying SOA/BPM will know the entire proposition will be a non-starter.

Understanding the business processes exist at 2 levels (the Silicon and the Carbon) takes us a long way towards understanding how we solve this problem. The key point is to recognize that the unpredictable actions of the carbon components are not ad-hoc processes, nor are they exception handling (ask anyone with a six sigma background about exceptions and you’ll understand very quickly what I mean). This is all about the unstructured interactions between people – in particular knowledge workers.  These unstructured and unpredictable interactions can, and do, take place all the time – and it’s only going to get worse! The advent of social networking, SaaS etc. etc.,  are already having, and will continue to have, a profound effect on the way we manage and do business.

Process based technology that understands the needs of people and supports the inherent “spontaneity” of the human mind is the next logical step, and we might be tempted to name this potential paradigm shift “A Business Operations Platform”. [1]

But what makes a BOP different from what’s gone before?

One of the key innovations (and there are many) is the collaborative nature of the platform. At last there is an environment that allows, encourages even, the business world and the technology world to align. Given that the business process is where these two worlds collide then the BOP becomes the place where the two worlds can achieve the most in terms of collaborative development and common understanding. Eliminating decades of misunderstanding. The Business Operations Platform does six main jobs.

It:

  1. Puts existing and new application software under the direct control of business managers.
  2. Facilitates communication between business and IT.
  3. Makes it easier for the business to improve existing processes and create new ones.
  4. Enables the automation of processes across the entire organization, and beyond it.
  5. Gives managers real-time information on the performance of processes.
  6. Allows organizations to take full advantage of new computing services.

Unlike early BPM offerings that were stitched together from fragments of technologies past, a BOP must be built on a standards-based and modern architecture.. With a service oriented architecture (SOA) and full BPM capabilities companies can create a complete business operations environment that can drive innovation, efficiency and agility for their enterprise. It must be Cloud enabled and capable of being deployed as BPMAAS as. It is the BOP that sets “enterprise cloud computing” apart from “consumer cloud Computing”

So why does workflow suck? It sucks because it made the fatal assumption that a business process was simply modelled as “a to b to c” – but business, as we all know, doesn’t quite work like that. BPM succeeds because of the heritage these products is in the workflow world – but BPM sucks as well because it ignores the requirement to include people.

Jon Pyke


[1] Since I wrote this paper Gartner coined the term “Intelligent BPM” but that begs the question as to what went before “Stupid BPM” ? So I’ll use BOP if that is OK with you the reader.

The post Why Workflow and BPM Suck appeared first on Cloud Computing Best Practices.

Read the original blog entry...

More Stories By Cloud Best Practices Network

The Cloud Best Practices Network is an expert community of leading Cloud pioneers. Follow our best practice blogs at http://CloudBestPractices.net

@CloudExpo Stories
Businesses and business units of all sizes can benefit from cloud computing, but many don't want the cost, performance and security concerns of public cloud nor the complexity of building their own private clouds. Today, some cloud vendors are using artificial intelligence (AI) to simplify cloud deployment and management. In his session at 20th Cloud Expo, Ajay Gulati, Co-founder and CEO of ZeroStack, will discuss how AI can simplify cloud operations. He will cover the following topics: why clou...
The Internet of Things (IoT) promises to simplify and streamline our lives by automating routine tasks that distract us from our goals. This promise is based on the ubiquitous deployment of smart, connected devices that link everything from industrial control systems to automobiles to refrigerators. Unfortunately, comparatively few of the devices currently deployed have been developed with an eye toward security, and as the DDoS attacks of late October 2016 have demonstrated, this oversight can ...
Without a clear strategy for cost control and an architecture designed with cloud services in mind, costs and operational performance can quickly get out of control. To avoid multiple architectural redesigns requires extensive thought and planning. Boundary (now part of BMC) launched a new public-facing multi-tenant high resolution monitoring service on Amazon AWS two years ago, facing challenges and learning best practices in the early days of the new service. In his session at 19th Cloud Exp...
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
Internet-of-Things discussions can end up either going down the consumer gadget rabbit hole or focused on the sort of data logging that industrial manufacturers have been doing forever. However, in fact, companies today are already using IoT data both to optimize their operational technology and to improve the experience of customer interactions in novel ways. In his session at @ThingsExpo, Gordon Haff, Red Hat Technology Evangelist, will share examples from a wide range of industries – includin...
"We build IoT infrastructure products - when you have to integrate different devices, different systems and cloud you have to build an application to do that but we eliminate the need to build an application. Our products can integrate any device, any system, any cloud regardless of protocol," explained Peter Jung, Chief Product Officer at Pulzze Systems, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo 2016 in New York. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be! Internet of @ThingsExpo, taking place June 6-8, 2017, at the Javits Center in New York City, New York, is co-located with 20th Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry p...
@DevOpsSummit at Cloud taking place June 6-8, 2017, at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long developm...
"We are an all-flash array storage provider but our focus has been on VM-aware storage specifically for virtualized applications," stated Dhiraj Sehgal of Tintri in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Between 2005 and 2020, data volumes will grow by a factor of 300 – enough data to stack CDs from the earth to the moon 162 times. This has come to be known as the ‘big data’ phenomenon. Unfortunately, traditional approaches to handling, storing and analyzing data aren’t adequate at this scale: they’re too costly, slow and physically cumbersome to keep up. Fortunately, in response a new breed of technology has emerged that is cheaper, faster and more scalable. Yet, in meeting these new needs they...
"We're a cybersecurity firm that specializes in engineering security solutions both at the software and hardware level. Security cannot be an after-the-fact afterthought, which is what it's become," stated Richard Blech, Chief Executive Officer at Secure Channels, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
When it comes to cloud computing, the ability to turn massive amounts of compute cores on and off on demand sounds attractive to IT staff, who need to manage peaks and valleys in user activity. With cloud bursting, the majority of the data can stay on premises while tapping into compute from public cloud providers, reducing risk and minimizing need to move large files. In his session at 18th Cloud Expo, Scott Jeschonek, Director of Product Management at Avere Systems, discussed the IT and busin...
According to Forrester Research, every business will become either a digital predator or digital prey by 2020. To avoid demise, organizations must rapidly create new sources of value in their end-to-end customer experiences. True digital predators also must break down information and process silos and extend digital transformation initiatives to empower employees with the digital resources needed to win, serve, and retain customers.
The IoT is changing the way enterprises conduct business. In his session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, discussed how businesses can gain an edge over competitors by empowering consumers to take control through IoT. He cited examples such as a Washington, D.C.-based sports club that leveraged IoT and the cloud to develop a comprehensive booking system. He also highlighted how IoT can revitalize and restore outdated business models, making them profitable ...
SaaS companies can greatly expand revenue potential by pushing beyond their own borders. The challenge is how to do this without degrading service quality. In his session at 18th Cloud Expo, Adam Rogers, Managing Director at Anexia, discussed how IaaS providers with a global presence and both virtual and dedicated infrastructure can help companies expand their service footprint with low “go-to-market” costs.
Get deep visibility into the performance of your databases and expert advice for performance optimization and tuning. You can't get application performance without database performance. Give everyone on the team a comprehensive view of how every aspect of the system affects performance across SQL database operations, host server and OS, virtualization resources and storage I/O. Quickly find bottlenecks and troubleshoot complex problems.
"Once customers get a year into their IoT deployments, they start to realize that they may have been shortsighted in the ways they built out their deployment and the key thing I see a lot of people looking at is - how can I take equipment data, pull it back in an IoT solution and show it in a dashboard," stated Dave McCarthy, Director of Products at Bsquare Corporation, in this SYS-CON.tv interview at @ThingsExpo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
"We are the public cloud providers. We are currently providing 50% of the resources they need for doing e-commerce business in China and we are hosting about 60% of mobile gaming in China," explained Yi Zheng, CPO and VP of Engineering at CDS Global Cloud, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Predictive analytics tools monitor, report, and troubleshoot in order to make proactive decisions about the health, performance, and utilization of storage. Most enterprises combine cloud and on-premise storage, resulting in blended environments of physical, virtual, cloud, and other platforms, which justifies more sophisticated storage analytics. In his session at 18th Cloud Expo, Peter McCallum, Vice President of Datacenter Solutions at FalconStor, discussed using predictive analytics to mon...
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform and how we integrate our thinking to solve complicated problems. In his session at 19th Cloud Expo, Craig Sproule, CEO of Metavine, demonstrated how to move beyond today's coding paradigm and sh...