Welcome!

@CloudExpo Authors: Pat Romanski, Yeshim Deniz, Liz McMillan, Jason Bloomberg, Elizabeth White

Related Topics: @CloudExpo, @BigDataExpo, @ThingsExpo

@CloudExpo: Blog Post

The #IoT and #Analytics | @ThingsExpo #BigData #BI #AI #DX #MachineLearning

The Internet of Things promises to change everything by enabling “smart” environments and smart products

The Internet of Things (IoT) and Analytics at The Edge

The Internet of Things (IoT) promises to change everything by enabling “smart” environments (homes, cities, hospitals, schools, stores, etc.) and smart products (cars, trucks, airplanes, trains, wind turbines, lawnmowers, etc.). I recently wrote about the importance of moving beyond “connected” to “smart” in a blog titled “Internet of Things: Connected Does Not Equal Smart”. The article discusses the importance of moving beyond just collecting the data, to transitioning to leveraging this new wealth of IoT data to improve the decisions that these smart environments and products need to make: to help these environments and products to self-monitor, self-diagnose and eventually, self-direct.

But one of the key concepts in enabling this transition from connected to smart is the ability to perform “analytics at the edge.” Shawn Rogers, Chief Research Officer at Dell Statistica, had the following quote in an article in Information Management titled “Will the Citizen Data Scientist Inherit the World?”:

“Organizations are fast coming to the realization that IoT implementations are only going to become more vast and more pervasive, and that as that happens, the traditional analytic model of pulling all data in to a centralized source such as a data warehouse or analytic sandbox is going to make less and less sense.

So, most of the conversations I’m having around IoT analytics today revolve around looking at how companies can flip that model on its head and figure out ways to push the analytics out to the edge. If you can run analytics at the edge, you not only can eliminate the time, bandwidth and expense required to transport the data, but you make it possible to take immediate action in response to the insight. You speed up and simplify the analytic process in a way that’s never been done before.”

So I asked Shawn and his boss John Thompson, General Manager of Advanced Analytics at Dell, to help me understand what exactly they mean by “analytics at the edge.” It really boils down to these questions:

  • Are we really developing analytics at the edge?
  • If not, then what sorts of analytics are we performing at the edge?
  • Where are the analytic models actually being built?
  • And finally, what the heck does “at the edge” really mean?
  • So let’s actually start with that last question: What does “at the edge” really mean?

Question #1: What Is “At The Edge”?
“At the edge” refers to the multitude of devices or sensors that are scattered across any network or embedded throughout a product (car, jet engine, CT Scan) that is generating data about the operations and performance of that specific device or sensor.

For example, the current Airbus A350 model has close to 6,000 sensors and generates 2.5 Tb of data per day, while an even newer model – expected to be available in 2020 – will capture more than triple that amount! It is becoming more and more common for everyday common products to have hundreds if not thousands of embedded sensors that are generating readings every couple of seconds on the operations and performance of that particular product (see Figure 1).

Figure 1: Sensors at the Edge

But collecting these huge and real-time volumes of data doesn’t do anything to directly create business advantage. It is what you do with that data that drives the business value, which brings us to…

Question #2: Are We Really Developing Analytics “At The Edge”?
Are we really “performing analytics” (collecting the data, storing the data, preparing the data, running analytic algorithms, validating the analytic goodness of fit and then acting on the results) at the edges, or are we just “executing the analytic models” at the edges? It’s one thing to “execute the analytic models” (e.g., scores, rules, recommendations) at the edges, but something entirely different to actually “perform analytics” at the edges.

Per Shawn and John, “We can deliver analytic models to any end point. We can execute the analytic models in any environment – large or small. We can execute all the steps in performing analytics in a wide range of environments, but there are limits at the edge. The limits are on the robustness of the environment (i.e. cannot deliver an executable to an environment that does not have the memory or processing power to store it or execute it. We cannot change the laws of physics…;-).)”

Question #3: What Sorts Of Analytics Are We Performing At The Edge?
In our airplane example with 6,000 sensors on the plane generating over 2.5 Tb of data per day, how are we performing the analytics at the end?

Per John and Shawn, if the jet engine has a place to house a Java Virtual Machine (JVM) and an analytic model (i.e., lightweight rules based model), then we can execute the model on the engine itself. If the model streams the data to a network, we can execute the analytic model on a gateway, or intermediate server (see Figure 2).

Figure 2: Executing Analytic Models at The Edge

Think of the network as having concentric rings. Each ring can have many servers. Each server can do either – either executing an analytic model or building the analytic models. Now think of many network networks with concentric rings that interlock at various intersections. Analytics can be at any or all levels including at the core, in a data center or in the cloud.

Per Shawn, “By working in tandem with Dell Boomi, we’ve given users the ability to deploy JVM’s with the analytic models on any edge device or gateway anywhere on the network or device. This edge scoring capability enables organizations to address nearly any IoT analytics use case by executing the analytic models at the edge of the network where data is being created.”

Question #4: Where Are The Analytic Models Actually Being Built?
Okay, so we “execute” the pre-built modes at the edge, but we actually build (test, refine, test, refine) the analytic models by bringing the detailed sensor data back to a central data and analytics environment (a.k.a. the Data Lake). Figure 3, courtesy of Joel Dodd of Pivotal, shows the data flow and the supporting analytics execution.

Figure 3: “At the Edge” Analytic Model Execution

Final point, even if you are doing all the sensor/IoT analysis at the edges, you are likely still going to want to bring the raw IoT data back into the data lake for more extensive analysis in order to house the detailed IoT history. For example, we have major economic cycles every 4 to 7 years. You might want to quantify the impact of these economic changes on your network demand and performance. That would eventually require 8 to 14 years of data. And that’s why you are going to want a data lake as the foundation of the transition from a “connected” IoT world to a “smart” IoT world.

The post The Internet of Things (IoT) and Analytics at The Edge appeared first on InFocus.

Read the original blog entry...

More Stories By William Schmarzo

Bill Schmarzo, author of “Big Data: Understanding How Data Powers Big Business”, is responsible for setting the strategy and defining the Big Data service line offerings and capabilities for the EMC Global Services organization. As part of Bill’s CTO charter, he is responsible for working with organizations to help them identify where and how to start their big data journeys. He’s written several white papers, avid blogger and is a frequent speaker on the use of Big Data and advanced analytics to power organization’s key business initiatives. He also teaches the “Big Data MBA” at the University of San Francisco School of Management.

Bill has nearly three decades of experience in data warehousing, BI and analytics. Bill authored EMC’s Vision Workshop methodology that links an organization’s strategic business initiatives with their supporting data and analytic requirements, and co-authored with Ralph Kimball a series of articles on analytic applications. Bill has served on The Data Warehouse Institute’s faculty as the head of the analytic applications curriculum.

Previously, Bill was the Vice President of Advertiser Analytics at Yahoo and the Vice President of Analytic Applications at Business Objects.

@CloudExpo Stories
With major technology companies and startups seriously embracing Cloud strategies, now is the perfect time to attend 21st Cloud Expo October 31 - November 2, 2017, at the Santa Clara Convention Center, CA, and June 12-14, 2018, at the Javits Center in New York City, NY, and learn what is going on, contribute to the discussions, and ensure that your enterprise is on the right path to Digital Transformation.
SYS-CON Events announced today that Cloud Academy named "Bronze Sponsor" of 21st International Cloud Expo which will take place October 31 - November 2, 2017 at the Santa Clara Convention Center in Santa Clara, CA. Cloud Academy is the industry’s most innovative, vendor-neutral cloud technology training platform. Cloud Academy provides continuous learning solutions for individuals and enterprise teams for Amazon Web Services, Microsoft Azure, Google Cloud Platform, and the most popular cloud com...
With major technology companies and startups seriously embracing Cloud strategies, now is the perfect time to attend 21st Cloud Expo October 31 - November 2, 2017, at the Santa Clara Convention Center, CA, and June 12-14, 2018, at the Javits Center in New York City, NY, and learn what is going on, contribute to the discussions, and ensure that your enterprise is on the right path to Digital Transformation.
SYS-CON Events announced today that CA Technologies has been named "Platinum Sponsor" of SYS-CON's 21st International Cloud Expo®, which will take place October 31-November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. CA Technologies helps customers succeed in a future where every business - from apparel to energy - is being rewritten by software. From planning to development to management to security, CA creates software that fuels transformation for companies in the applic...
Multiple data types are pouring into IoT deployments. Data is coming in small packages as well as enormous files and data streams of many sizes. Widespread use of mobile devices adds to the total. In this power panel at @ThingsExpo, moderated by Conference Chair Roger Strukhoff, panelists looked at the tools and environments that are being put to use in IoT deployments, as well as the team skills a modern enterprise IT shop needs to keep things running, get a handle on all this data, and deliver...
In his session at @ThingsExpo, Eric Lachapelle, CEO of the Professional Evaluation and Certification Board (PECB), provided an overview of various initiatives to certify the security of connected devices and future trends in ensuring public trust of IoT. Eric Lachapelle is the Chief Executive Officer of the Professional Evaluation and Certification Board (PECB), an international certification body. His role is to help companies and individuals to achieve professional, accredited and worldwide re...
Amazon started as an online bookseller 20 years ago. Since then, it has evolved into a technology juggernaut that has disrupted multiple markets and industries and touches many aspects of our lives. It is a relentless technology and business model innovator driving disruption throughout numerous ecosystems. Amazon’s AWS revenues alone are approaching $16B a year making it one of the largest IT companies in the world. With dominant offerings in Cloud, IoT, eCommerce, Big Data, AI, Digital Assista...
New competitors, disruptive technologies, and growing expectations are pushing every business to both adopt and deliver new digital services. This ‘Digital Transformation’ demands rapid delivery and continuous iteration of new competitive services via multiple channels, which in turn demands new service delivery techniques – including DevOps. In this power panel at @DevOpsSummit 20th Cloud Expo, moderated by DevOps Conference Co-Chair Andi Mann, panelists examined how DevOps helps to meet the de...
Both SaaS vendors and SaaS buyers are going “all-in” to hyperscale IaaS platforms such as AWS, which is disrupting the SaaS value proposition. Why should the enterprise SaaS consumer pay for the SaaS service if their data is resident in adjacent AWS S3 buckets? If both SaaS sellers and buyers are using the same cloud tools, automation and pay-per-transaction model offered by IaaS platforms, then why not host the “shrink-wrapped” software in the customers’ cloud? Further, serverless computing, cl...
You know you need the cloud, but you’re hesitant to simply dump everything at Amazon since you know that not all workloads are suitable for cloud. You know that you want the kind of ease of use and scalability that you get with public cloud, but your applications are architected in a way that makes the public cloud a non-starter. You’re looking at private cloud solutions based on hyperconverged infrastructure, but you’re concerned with the limits inherent in those technologies.
The taxi industry never saw Uber coming. Startups are a threat to incumbents like never before, and a major enabler for startups is that they are instantly “cloud ready.” If innovation moves at the pace of IT, then your company is in trouble. Why? Because your data center will not keep up with frenetic pace AWS, Microsoft and Google are rolling out new capabilities. In his session at 20th Cloud Expo, Don Browning, VP of Cloud Architecture at Turner, posited that disruption is inevitable for comp...
No hype cycles or predictions of zillions of things here. IoT is big. You get it. You know your business and have great ideas for a business transformation strategy. What comes next? Time to make it happen. In his session at @ThingsExpo, Jay Mason, Associate Partner at M&S Consulting, presented a step-by-step plan to develop your technology implementation strategy. He discussed the evaluation of communication standards and IoT messaging protocols, data analytics considerations, edge-to-cloud tec...
"When we talk about cloud without compromise what we're talking about is that when people think about 'I need the flexibility of the cloud' - it's the ability to create applications and run them in a cloud environment that's far more flexible,” explained Matthew Finnie, CTO of Interoute, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
Wooed by the promise of faster innovation, lower TCO, and greater agility, businesses of every shape and size have embraced the cloud at every layer of the IT stack – from apps to file sharing to infrastructure. The typical organization currently uses more than a dozen sanctioned cloud apps and will shift more than half of all workloads to the cloud by 2018. Such cloud investments have delivered measurable benefits. But they’ve also resulted in some unintended side-effects: complexity and risk. ...
It is ironic, but perhaps not unexpected, that many organizations who want the benefits of using an Agile approach to deliver software use a waterfall approach to adopting Agile practices: they form plans, they set milestones, and they measure progress by how many teams they have engaged. Old habits die hard, but like most waterfall software projects, most waterfall-style Agile adoption efforts fail to produce the results desired. The problem is that to get the results they want, they have to ch...
In 2014, Amazon announced a new form of compute called Lambda. We didn't know it at the time, but this represented a fundamental shift in what we expect from cloud computing. Now, all of the major cloud computing vendors want to take part in this disruptive technology. In his session at 20th Cloud Expo, Doug Vanderweide, an instructor at Linux Academy, discussed why major players like AWS, Microsoft Azure, IBM Bluemix, and Google Cloud Platform are all trying to sidestep VMs and containers wit...
The Internet giants are fully embracing AI. All the services they offer to their customers are aimed at drawing a map of the world with the data they get. The AIs from these companies are used to build disruptive approaches that cannot be used by established enterprises, which are threatened by these disruptions. However, most leaders underestimate the effect this will have on their businesses. In his session at 21st Cloud Expo, Rene Buest, Director Market Research & Technology Evangelism at Ara...
"We are a monitoring company. We work with Salesforce, BBC, and quite a few other big logos. We basically provide monitoring for them, structure for their cloud services and we fit into the DevOps world" explained David Gildeh, Co-founder and CEO of Outlyer, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
When growing capacity and power in the data center, the architectural trade-offs between server scale-up vs. scale-out continue to be debated. Both approaches are valid: scale-out adds multiple, smaller servers running in a distributed computing model, while scale-up adds fewer, more powerful servers that are capable of running larger workloads. It’s worth noting that there are additional, unique advantages that scale-up architectures offer. One big advantage is large memory and compute capacity...
SYS-CON Events announced today that IBM has been named “Diamond Sponsor” of SYS-CON's 21st Cloud Expo, which will take place on October 31 through November 2nd 2017 at the Santa Clara Convention Center in Santa Clara, California.