Data Center Workload Measurement and Analysis

You might also like

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 24

Data Center Workload Measurement

and Analysis

1
Sensitivity: LNT Construction Internal Use
What is a Data center ?
•A large group of networked computer servers typically used by
organizations for the remote storage, processing, or distribution of
large amounts of data.

•It doesn’t house only servers but also contains backup power
supplies, communication connections, air conditioning, fire
supplies etc.

•“A data center is a factory that transforms and stores bits”

Sensitivity: LNT Construction Internal Use


A few glimpses of Data Center of a few organizations
… - Richardson,TX
Rackspace Facebook – Lulea, Sweden

Google- Douglas County, Georgia Amazon – Virginia, outside Washington D.C

3
Sensitivity: LNT Construction Internal Use
Data Center workload
• Amount of processing that the computer has been given to do at a
given time.
• Workload — in the form of web requests, data analysis, multimedia rendering,
or other applications – is placed in the data center

5
Sensitivity: LNT Construction Internal Use
Classification of workloads based on time
criticality

Critical Workloads Non-critical Workloads

“Cannot tolerate even a few minutes can tolerate a wide range of outage times
of downtime”

Sensitivity: LNT Construction Internal Use


Ways to improve data
protection
• Prevent downtime by reducing resource contention :
Managers accommodates drastically changing demands on workloads by allowing easy creation of
additional workloads without changing or customizing its applications.

• Replicate workloads into cloud to create asymmetric “Hot back-ups”:


Clone the complete workload stack. Import into public/private cloud

• Using dissimilar infrastructure for off-premises redundancies:


Workloads are replicated off-site to different cloud providers.

• Concept of “Failures or Failback”  reserved only for critical workloads:


Automating the switching of users or processes from production to recovery instances

Sensitivity: LNT Construction Internal Use


Characterizing Data Analysis workloads in Data Centers
• Data Analysis is important improving future performance of data center
• Data center workloads services workload (web search, media streaming)
data analysis workload ( business intelligence,
machine
• We concentrate on internet services learninghere
workload )
• Data analysis workloads are diverse in speedup performance and micro-architectural characteristics.
Therefore, there is a need to analyze many applications

• 3 important application domains are in internet services are : 1) search engine 2) social networks 3)
electronic commerce

7
Sensitivity: LNT Construction Internal Use
Workload requirements :
1) most important application domain
2)data is distributed, data can not be processed on single node
3)consider recently used data

8
Sensitivity: LNT Construction Internal Use
Methodologies

9
Sensitivity: LNT Construction Internal Use
Workflow
Phases

Extract Partition Aggregate


• Look for raw data • Divides stream into • Combines/reduces
• Generates stream buckets
of data

10
Sensitivity: LNT Construction Internal Use
Patterns comprising traffic in Data
Center

Work- Scatter gather


seeks- pattern
bandwidth

11
Sensitivity: LNT Construction Internal Use
Monitoring Data Center
Workload
• For coordinated monitoring and control of data centers, the most
commonly approaches are based on Monitor, Analyze ,Plan and
Execute (MAPE ) control loops.

Overview
12
Sensitivity: LNT Construction Internal Use
Modern Data Center
Operation
• Workload in the form of web requests, data analysis, etc is placed in the
data center.

• An instrumentation infrastructure logs sensor readings.

• The results are fed into a policy engine that creates a plan to utilize
resources.

• External interfaces or Actuators implement the plan.

13
Sensitivity: LNT Construction Internal Use
Workload Monitoring using
Splice
• Splice aggregates sensor and performance data in a relational
database.
• It also gathers data from many sources through different interfaces
with different formats.
• Splice uses change of value filter that retains only those values that
differ significantly from the previously logged values.
• It reduces minimal loss of information.

14
Sensitivity: LNT Construction Internal Use
Analysi
s
• Data analysis is done by two main classes- attribute behavior and
correlation.

• Attribute behavior describes the value of the observed readings and


how those values change over time.

• Data correlation methods determine the strength of the correlations


among the attributes affecting each other.

15
Sensitivity: LNT Construction Internal Use
Virtualization in Data Centers
• Virtualization is a combination of software and hardware features that creates
virtual CPUs (vCPU) or virtual systems-on-chip (vSoC).
• Virtualization provides the required level of isolation and partitioning of
resources.
• Each VM is protected from interference from another VM.

Sensitivity: LNT Construction Internal Use


Why Virtualization
• Reduced power consumption and building space, providing high
availability for critical applications and streamlining application
deployment and migration.
• To support multiple operating systems and consolidation of services on a
single server by defining multiple VMs.
• Multiple VMs can run on a single server, the advantage is of reduced server
inventory and better server utilization.

Sensitivity: LNT Construction Internal Use


Benefits Of Virtualization

Sensitivity: LNT Construction Internal Use


Multi Core Processing
• A multi-core processor is a single computing component with two or more
independent actual processing units (called "cores"), which are the units that
read and execute program instructions.

Sensitivity: LNT Construction Internal Use


Data center
Reliability
Network Reliability

Analyzing the
Characterizing most failure Estimating the impact of
effectiveness of network
prone network elements failures
redundancy

Sensitivity: LNT Construction Internal Use


Key Observations
• Data center networks are reliable
• Low-cost, commodity switches are highly reliable
• Load balancers experience a high number of software faults
• Failures potentially cause loss of a large number of small packets.
• Network redundancy helps, but it is not entirely effective

Sensitivity: LNT Construction Internal Use


Reasons to change from traditional
Significant changes in computing power, network bandwidth, and
network file system usage
• Network file system workloads
• No CIFS protocol studies
• Limited file system workloads

Sensitivity: LNT Construction Internal Use


• Sequentiality Analysis:

Sequential Access

Entire Partial

Sensitivity: LNT Construction Internal Use


Load
Balancing
Load Balancing

Load Spreading: Load Balancing:


requests spread evenly over a place load balancers in front of the
pool of servers actual servers

Sensitivity: LNT Construction Internal Use

You might also like