Welcome!

@CloudExpo Authors: Liz McMillan, Pat Romanski, Elizabeth White, Nate Vickery, Gopala Krishna Behara

Related Topics: @CloudExpo, Artificial Intelligence, @DXWorldExpo

@CloudExpo: Blog Post

The Human Body and Data Center Automation | @CloudExpo #AI #ML #DataCenter

The nervous system has two components: the central nervous system and the peripheral nervous system

The Human Body and Data Center Automation - Part 2

Disclaimer: I am an IT guy and my knowledge on human body is limited to my daughter's high school biology class book and information obtained from search engines. So, excuse me if any of the information below is not represented accurately !!

Human body is the most complex machine ever created. With a complex network of interconnected organs, millions of cells and the most advanced processor, human body is the most automated system in this planet. In this article, we will draw comparisons between working of a human body to that of a data center. We will draw parallels between human body automation to data center automation and explain different levels of automation we need to drive in data centers. This article is divided into four parts covering each of body main functions and drawing parallels on automation. This is the second article in the human body series. Please click here for the link for first article

The nervous system
The nervous system is a complex collection of nerves and specialized cells known as neurons that transmit signals between different parts of the body. It is through the nervous system that we communicate with the outside world and, at the same time, many mechanisms inside our body are controlled. The nervous system takes in information through our senses, processes the information and triggers reactions, such as making your muscles move or causing you to feel pain. The closest comparison to nervous system to what is in our data center is the network. Much like the network connects everything together in the data center, nervous system essentially is the body's network system.

The nervous system has two components: the central nervous system and the peripheral nervous system. The central nervous system is made up of the brain, spinal cord and nerves. The peripheral nervous system consists of sensory neurons, ganglia (clusters of neurons) and nerves that connect to one another and to the central nervous system. Imagine this as core network (central) and data center network (peripheral). However, what is so fascinating about our nervous system is the way it works. Let's take a deep dive inside our body and learn how can we make our networks more efficient.

Image Source: Livescience.com

The nervous system has two main subdivisions: the somatic, or voluntary, component; and the autonomic, or involuntary, component. The autonomic nervous system regulates certain body processes, such as blood pressure and the rate of breathing, that work without conscious effort. It is constantly active, regulating things such as breathing, heart beat and metabolic processes. It does this by receiving signals from the brain and passing them on to the body. It can also send signals in the other direction - from the body to the brain - providing your brain with information about how full your bladder is or how quickly your heart is beating.

Now can you think what system in our data center comes close to autonomic nervous system? It's our monitoring system. The function of monitoring system in data center is to monitor health of various components(hardware/software) in our data center and alert us when the thresholds are breached or an error has occurred. In most of the modern data centers today there is some tool which does this job. Alerts and error logs are collected at all layers and once an error has occurred or particular KPIs crosses certain threshold, an event is generated and humans are notified to take action. However, where the human body defeats any modern monitoring system is ability to take autonomic actions based on situation. Let's imagine you on a treadmill and running. When you are running and as the heart rates goes up, the brain is not just sending out alerts to you indicating your heart rate is going up but it also taking appropriate actions to ensure the body continues to function. The first action is to breakdown glycogen, a form of glucose to give you extra dose of energy.

The second action is to draw more blood towards your muscles which are under stress and away from non-needed functions like digestion (unless you are eating while excising). Since the body needs more oxygen for your muscles it signals your lungs to intake more oxygen and hence your breathing rate goes up. As the body burns more glucose and your body heats up, your brain sends signals to your sweat glands to release moisture to keep the body cool and hence maintains temperature inside the body. All these actions without you telling your body what to do to keep you healthy. Only if the thresholds crosses beyond certain rate and your body is not able to fix you, it will signal us to take action like resting or slowing down. This is exactly how our monitoring system should work. However, what happens in most of the enterprise is a sorry state of affairs.

Let's consider a very common issue most of the enterprise faces - a performance issue. Consider that the mission critical business application running on your server is experiencing performance issues and the users are complaining. In a typical organization, an application user will do first level analysis and based on his/her analysis he will open an incident ticket with command center. Everyone from systems engineer, storage engineer, network engineer and specialized performance engineers are paged to figure out what's happening. Hours are spent to detect where in the fabric there is contention which is leading to performance issue. Once the issue is detected another few hours are spent to finalize the action plan and finally the fix is put in place. Sounds familiar.

Now imagine we can learn from our human body and can design our data center in such a way that the system should automatically detect something is going wrong in the fabric and find out where in the fabric there is issue. Once the issue is detected it identifies appropriate fix and implements the fix. If the system detects performance issue is because of underlying CPU constraint on one of the VMs, the system should either scale up CPU capacity on the VM or automatically horizontally scale the application by adding another VM or container. If the issue was detected at network level, system should be in a position to move entire VLAN to another healthy leaf switch. If the issue happened at DC level, the system should automatically fail over all the impacted applications to another DC. While some of the modern cloud native applications works in similar fashion, the same level of maturity is not seen in traditional applications.

The somatic system consists of nerves that connect the brain and spinal cord with muscles and sensory receptors in the skin. The voluntary nervous system (somatic nervous system) controls all the things that we are aware of and can consciously influence, such as moving our arms, legs and other parts of the body. The nerves (like the network cables in our DC), starts at the brain and central cord and branches out to every part of our body. Neurons (intelligent code) send signals to other cells though thin fibers called axons, which causes chemical known as neuro transmitter to be released at junctions called synapses. A synapse gives command to the cell and the entire communication takes a fraction of a second. Such is the speed of transmission in our human body that our fastest router in the world cannot come close to this.

Let's take an example. Imagine someone tapping you lightly on your shoulder and your immediate reaction is to turn around and see who is doing that. The sensory neurons (cells) on our shoulder transmit the signals to your brain via the nerves at such a fast pace that you immediately react. Now imagine someone tapping at your shoulder and it takes few seconds to a minute for your body to react to the signal J . The way our body reacts to various form of sense (touch, smell, taste, etc.) and the fact we don't have to manage every action indicates how advance is our body's automation system. The body sensor systems are like the sensors in our data center. The role of sensors is to collect the data and send it for further processing. While we have lot of maturity to collect data what we lack is how fast can we analyze the data to take appropriate action. This is where our Brain checkmates even the fastest of all computers including IBM Watson. Our brain is a combination of Big Data system, e.g., Hadoop, the intelligence of IBM Watson and fastest super compute in the world all combined into one. Let's look at our brain.

Image Source: diseasespictures.com

Brain - Our Brain is the intelligence of our body. It controls all actions in our body. It acts as both CPU and memory for our body and without brains you are almost like walking zombie who has no control of his or her actions. Inside the data center, CPU/Memory inside our servers combined with the software which runs on top of these acts like brain. However, we are still far to match the amount of computing capacity our brain has and more important the learning and intuitive capabilities our brain has.

The fastest supercomputer in the world is China's Sunway Taihulight  and has a maximum processing speed of 93 petaFLOPS. A petaFLOP is a quadrillion (one thousand trillion) floating point calculations per second. Still this does not come close to processing speed of the human brain. It is postulated that the human brain operates at 1 exaFLOP, which is equivalent to a billion billion calculations per second. While the hardware which is, the muscular structure can be compared with chip set we have in our computers, it's the software which makes the difference.

Our brain controls the nervous system, the muscular system and other vital parts of our body. It also has tremendous learning capability. When a child is born, our brain is almost empty but is in a learning mode. It quickly learns how to interact with the outside world and starts to read data from the sense organs. This is how we react to taste, touch and sound. As we grow we learn how to talk, write and communicate. We learn how to walk, run and jump! In the software world, we call it as AI - Artificial intelligence. Humans have been trying for centuries on how to develop brain like self-learning capabilities and while we now have self-driving cars which is pushing the envelope, we are still far away from truly matching our brain's power.

Now the brain cannot work out of isolation. It gets all the data it needs from our sensory organs - eyes, nose, tongue. We interact with the world with the help of our sensory organs. Eyes gives us visual data , nose gives us smell related data while the tongue allows us to taste. Do you know that our tongue alone has millions of sensors which allows to distinguish various tastes from sour to sweet and from hot to cold! Imagine the status of all restaurants if we did not have these sensors. On the other hand, our nose has sensors to not only detect various smell but also acts as a self-defense organ.

In our data centers, we need similar capabilities. Inside the compute, we have sensors which tells us that the filesystem is getting full or inside the router we can tell if we have packet drops happening. We name them as alerts and in a given day millions of alerts are generated by all the systems running in an enterprise. The difference here is what do we do the alerts? If every alert is dependent on human to take manual intervention it will be as good as our tongue telling you that the coffee you are drinking is very hot and waiting for an action to be decided by you on whether you should stop drinking. Your tongue sends alert to your brain, your brain processes information and decides it's too dangerous and immediately takes action to control your actions. Now you may still drink it cause harm but the body takes immediate action to prevent the harm.

Similarly, with the alerts coming from our systems, we need to develop systems which can take immediate actions (self-heal) and not wait for human intervention all the time. If the filesystem is full, take immediate action to detect and fix what is causing it to be full. If the security intrusion detection system has detected malicious emails, block the emails immediately. If network port is dropping packets, isolate the port and move traffic to alternate port. The more autonomic actions we can take, the better we will be in managing our data center.

To summarize here's what we need in our data center: L-C-C-A

  • Lightning fast network of intelligent sensors across the data center stack
  • Central monitoring system which can monitor alerts/error logs at all layers from application down to the server
  • Co-relation engine which can correlate various alerts and error logs and pin points where in the data center there is issue
  • Artificial intelligence (AI) capable run-book automation engine which can trigger autonomous action (self-heal) based on the issue identified and implements the fix

In our next article on human body and data center automation we will focus on our circulator system which is responsible for flow of blood, oxygen and nutrients in our body and we will learn how our data center should learn from same. Until next time.

More Stories By Ashish Nanjiani

Ashish Nanjiani is a Senior IT Manager within Cisco IT managing Cisco worldwide IT data centers as an operations manager. With 20 years of IT experience, he is an expert in data center operations and automation. He has spoken in many inter-company events on data center automation and helps IT professionals digitize their IT operations. He is also an entrepreneur and has been successfully running a website business for 10+ years.

Ashish holds a Bachelor of Science degree in Electrical and Electronics and a Masters in Business Administration. He is a certified PMP, Scrum master. He is married and has two lovely daughters. He enjoys playing with technology during his free time. [email protected]

@CloudExpo Stories
In his Opening Keynote at 21st Cloud Expo, John Considine, General Manager of IBM Cloud Infrastructure, led attendees through the exciting evolution of the cloud. He looked at this major disruption from the perspective of technology, business models, and what this means for enterprises of all sizes. John Considine is General Manager of Cloud Infrastructure Services at IBM. In that role he is responsible for leading IBM’s public cloud infrastructure including strategy, development, and offering m...
Digital transformation is about embracing digital technologies into a company's culture to better connect with its customers, automate processes, create better tools, enter new markets, etc. Such a transformation requires continuous orchestration across teams and an environment based on open collaboration and daily experiments. In his session at 21st Cloud Expo, Alex Casalboni, Technical (Cloud) Evangelist at Cloud Academy, explored and discussed the most urgent unsolved challenges to achieve f...
In his session at 21st Cloud Expo, Raju Shreewastava, founder of Big Data Trunk, provided a fun and simple way to introduce Machine Leaning to anyone and everyone. He solved a machine learning problem and demonstrated an easy way to be able to do machine learning without even coding. Raju Shreewastava is the founder of Big Data Trunk (www.BigDataTrunk.com), a Big Data Training and consulting firm with offices in the United States. He previously led the data warehouse/business intelligence and B...
The past few years have brought a sea change in the way applications are architected, developed, and consumed—increasing both the complexity of testing and the business impact of software failures. How can software testing professionals keep pace with modern application delivery, given the trends that impact both architectures (cloud, microservices, and APIs) and processes (DevOps, agile, and continuous delivery)? This is where continuous testing comes in. D
Mobile device usage has increased exponentially during the past several years, as consumers rely on handhelds for everything from news and weather to banking and purchases. What can we expect in the next few years? The way in which we interact with our devices will fundamentally change, as businesses leverage Artificial Intelligence. We already see this taking shape as businesses leverage AI for cost savings and customer responsiveness. This trend will continue, as AI is used for more sophistica...
The “Digital Era” is forcing us to engage with new methods to build, operate and maintain applications. This transformation also implies an evolution to more and more intelligent applications to better engage with the customers, while creating significant market differentiators. In both cases, the cloud has become a key enabler to embrace this digital revolution. So, moving to the cloud is no longer the question; the new questions are HOW and WHEN. To make this equation even more complex, most ...
Blockchain is a shared, secure record of exchange that establishes trust, accountability and transparency across business networks. Supported by the Linux Foundation's open source, open-standards based Hyperledger Project, Blockchain has the potential to improve regulatory compliance, reduce cost as well as advance trade. Are you curious about how Blockchain is built for business? In her session at 21st Cloud Expo, René Bostic, Technical VP of the IBM Cloud Unit in North America, discussed the b...
SYS-CON Events announced today that Synametrics Technologies will exhibit at SYS-CON's 22nd International Cloud Expo®, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. Synametrics Technologies is a privately held company based in Plainsboro, New Jersey that has been providing solutions for the developer community since 1997. Based on the success of its initial product offerings such as WinSQL, Xeams, SynaMan and Syncrify, Synametrics continues to create and hone in...
With tough new regulations coming to Europe on data privacy in May 2018, Calligo will explain why in reality the effect is global and transforms how you consider critical data. EU GDPR fundamentally rewrites the rules for cloud, Big Data and IoT. In his session at 21st Cloud Expo, Adam Ryan, Vice President and General Manager EMEA at Calligo, examined the regulations and provided insight on how it affects technology, challenges the established rules and will usher in new levels of diligence arou...
As you move to the cloud, your network should be efficient, secure, and easy to manage. An enterprise adopting a hybrid or public cloud needs systems and tools that provide: Agility: ability to deliver applications and services faster, even in complex hybrid environments Easier manageability: enable reliable connectivity with complete oversight as the data center network evolves Greater efficiency: eliminate wasted effort while reducing errors and optimize asset utilization Security: imple...
Nordstrom is transforming the way that they do business and the cloud is the key to enabling speed and hyper personalized customer experiences. In his session at 21st Cloud Expo, Ken Schow, VP of Engineering at Nordstrom, discussed some of the key learnings and common pitfalls of large enterprises moving to the cloud. This includes strategies around choosing a cloud provider(s), architecture, and lessons learned. In addition, he covered some of the best practices for structured team migration an...
Companies are harnessing data in ways we once associated with science fiction. Analysts have access to a plethora of visualization and reporting tools, but considering the vast amount of data businesses collect and limitations of CPUs, end users are forced to design their structures and systems with limitations. Until now. As the cloud toolkit to analyze data has evolved, GPUs have stepped in to massively parallel SQL, visualization and machine learning.
The 22nd International Cloud Expo | 1st DXWorld Expo has announced that its Call for Papers is open. Cloud Expo | DXWorld Expo, to be held June 5-7, 2018, at the Javits Center in New York, NY, brings together Cloud Computing, Digital Transformation, Big Data, Internet of Things, DevOps, Machine Learning and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding busin...
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the bene...
Kubernetes is an open source system for automating deployment, scaling, and management of containerized applications. Kubernetes was originally built by Google, leveraging years of experience with managing container workloads, and is now a Cloud Native Compute Foundation (CNCF) project. Kubernetes has been widely adopted by the community, supported on all major public and private cloud providers, and is gaining rapid adoption in enterprises. However, Kubernetes may seem intimidating and complex ...
In his session at 21st Cloud Expo, Michael Burley, a Senior Business Development Executive in IT Services at NetApp, described how NetApp designed a three-year program of work to migrate 25PB of a major telco's enterprise data to a new STaaS platform, and then secured a long-term contract to manage and operate the platform. This significant program blended the best of NetApp’s solutions and services capabilities to enable this telco’s successful adoption of private cloud storage and launching ...
In his general session at 21st Cloud Expo, Greg Dumas, Calligo’s Vice President and G.M. of US operations, discussed the new Global Data Protection Regulation and how Calligo can help business stay compliant in digitally globalized world. Greg Dumas is Calligo's Vice President and G.M. of US operations. Calligo is an established service provider that provides an innovative platform for trusted cloud solutions. Calligo’s customers are typically most concerned about GDPR compliance, application p...
In a recent survey, Sumo Logic surveyed 1,500 customers who employ cloud services such as Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP). According to the survey, a quarter of the respondents have already deployed Docker containers and nearly as many (23 percent) are employing the AWS Lambda serverless computing framework. It’s clear: serverless is here to stay. The adoption does come with some needed changes, within both application development and operations. Tha...
You know you need the cloud, but you’re hesitant to simply dump everything at Amazon since you know that not all workloads are suitable for cloud. You know that you want the kind of ease of use and scalability that you get with public cloud, but your applications are architected in a way that makes the public cloud a non-starter. You’re looking at private cloud solutions based on hyperconverged infrastructure, but you’re concerned with the limits inherent in those technologies.
Smart cities have the potential to change our lives at so many levels for citizens: less pollution, reduced parking obstacles, better health, education and more energy savings. Real-time data streaming and the Internet of Things (IoT) possess the power to turn this vision into a reality. However, most organizations today are building their data infrastructure to focus solely on addressing immediate business needs vs. a platform capable of quickly adapting emerging technologies to address future ...