Professional Documents
Culture Documents
Ch4 OpenStack New (Autosaved)
Ch4 OpenStack New (Autosaved)
• Federated Identity
– Logging into multiple open stack nodes through single
user ID
– Special request from europian organization for nuclear
research(CERN)
Features of OpenStack
• Trove
– manage database resources like MySQL for manipulating
users and schema
– Manipulation is done through Trove APIs
– Original term use is Project Red Dwarf
– MongoDB, , Cassandra, etc
04/21/20 55
Google App Engine
• Features
– Easy to build.
– Easy to maintain.
– Easy to scale as the traffic and storage needs grow.
04/21/20 56
Google App Engine
• GAE makes it easy to build and deploy an application that
runs reliably even under heavy load and with large
amounts of data.
• Features:
– Persistent storage with queries, sorting, and transactions.
– Automatic scaling and load balancing.
– Asynchronous task queues for performing work outside the
scope of a request.
– Scheduled tasks for triggering events at specified times or
regular intervals.
– Integration with other Google cloud services and APIs.
04/21/20 57
Google App Engine: Is it free?
• Yes
• free for upto 1 GB of storage and enough CPU
• bandwidth to support 5 million page views a
month
• 10 Applications per Google account
04/21/20 58
GAE: Programming
languages support
• Google App Engine supports apps written in a
variety of programming languages.
• Java
• Python
• PHP
• Go
04/21/20 59
GAE: Programming languages
support for java
• Java:
• App Engine runs JAVA apps on a JAVA 7 virtual machine
(currently supports JAVA 6 as well).
• Uses JAVA Servlet standard for web applications:
• WAR (Web Applications ARchive) directory structure.
• Servlet classes
• Java Server Pages (JSP)
• Static and data files
• Deployment descriptor (web.xml)
• Other configuration files
• Getting started :
– https://developers.google.com/appengine/docs/java/getti
04/21/20
ngstarted/ 60
GAE: Programming languages
support for python
Python:
• Uses WSGI (Web Server Gateway Interface) standard.
• Python applications can be written using:
• Webapp2 framework
• Django framework
• Any python code that uses the CGI (Common Gateway
Interface) standard.
•Getting started :
– https://developers.google.com/appengine/docs/python/g
ettingstartedpython27/
04/21/20 61
GAE: Programming languages
support for PHP
• PHP (Experimental support):
• Local development servers are available to anyone for
developing and testing local applications.
• Only whitelisted applications can be deployed on Google App
Engine. (https://gaeforphp.appspot.com/).
• Getting started:
https://developers.google.com/appengine/docs/php/
04/21/20 62
Google App Engine: Programming
languages support for java
• Google’s Go:
• Go is an Google’s open source programming environment.
• Tightly coupled with Google App Engine.
• Applications can be written using App Engine’s Go SDK.
• Getting started:
https://developers.google.com/appengine/docs/go/overview
04/21/20 63
When to use GAE
04/21/20 66
Google File System
(GFS)
GFS
• Google File System is a scalable distributed file system
for large distributed data-intensive applications.
• It provides fault tolerance while running on
inexpensive commodity hardware (Commodity
hardware is a term for affordable devices that are
generally compatible with other such devices.
• In a process called commodity computing or
commodity cluster computing, these devices are often
networked to provide more processing power when
those who own them cannot afford to purchase more
elaborate supercomputers, or want to maximize
savings in IT design.),
04/21/20 68
GFS
• and it delivers high aggregate performance to a large
number of clients.
• GFS shares many of the same goals as previous
distributed file systems such as performance,
scalability, reliability, and availability.
GFS has designed with certain assumptions that also
provide opportunities for developers researchers
• Automatic recovery from component failure on
routine basis
• Efficient storage support for large size files as huge
amount of data to be processed is stored in these file
Storage support for small size files without requiring
04/21/20 69
any optimization for them
• With the workloads mainly consisting of 2
large streaming reads and small random reads
the system should be performance conscious
(aware of and responding to one's) so that
small reads are made steady rather than going
back and forth
• Semantics are defined well
• Atomicity is maintained least overhead due to
synchronization
• Provision for sustained bandwidth is given
priority than latency(the delay before a
transfer of data begins following an instruction
for its transfer.)
• Google takes the assumption listed above into
consideration and support its cloud platform,
Google App Engine through GFS
04/21/20 72
3.2. Masterservers.
Chunk
servers
3. Chunk servers.
GFS Architecture
• GFS is clusters of computers. A cluster is
simply a network of computers. Each cluster
might contain hundreds or even thousands of
machines. In each GFS clusters there are three
main entities:
04/21/20 73
http://programming-project.blogspot.com/2014/04/general-architecture-of-
google-file.html
1. Clients
2. Master servers
3. Chunk servers
• Client can be other computers or computer
applications and make a file request.
• Requests can range from retrieving and
manipulating existing files to creating new
files on the system.
• Clients can be thought as customers of the
GFS
• Master Server is the coordinator for the cluster. Its task
include:-
• Maintaining an operation log, that keeps track of the
activities of the cluster.
• The operation log helps keep service interruptions to a
minimum if the master server crashes, a replacement
server that has monitored the operation log can take its
place.
04/21/20 81
GFS Architecture
• A GFS cluster consists of a single master and multiple
chunkservers and is accessed by multiple clients.
• Each of these is typically a commodity Linux machine
running a user-level server process.
• Files are divided into fixed-size chunks. Each chunk is
identified by an immutable and globally unique 64 bit chunk
handle assigned by the master at the time of chunk
creation.
• Chunkservers store chunks on local disks as Linux files and
read or write chunk data specified by a chunk handle and
byte range.
04/21/20 82
GFS Architecture
• For reliability, each chunk is replicated on multiple
chunk servers. By default, GFS store three replicas,
though users can designate different replication
levels for different regions of the file namespace.
• The master maintains all file system metadata. This
includes the namespace, access control information,
the mapping from files to chunks, and the current
locations of chunks.
04/21/20 83
GFS Architecture
• Cluster Computing
Single Master
Multiple Chunk Servers
– Stores 64 MB file chunks
Multiple clients
04/21/20 84
GFS cluster
04/21/20 85
GFS cluster
• Single Master:
– Minimal Master Load.
– Fixed chunk Size.
– The master also predicatively provide chunk locations immediately
following those requested by unique id.
• Chunk Size :
– 64 MB size.
– Read and write operations on same chunk.
– Reduces network overhead and size of metadata in the master
• Operation Log:
– Keeps track of activities.
– It stores on multiple remote locations .
04/21/20 86
GFS cluster
• Types of Metadata:
– File and chunk namespaces
– Mapping from files to chunks
– Location of each chunks replicas
• In-memory data structures:
– Master operations are fast.
– Periodic scanning entire state is easy and efficient
• Chunk Locations:
– Master polls chunk server for the information.
– Client request data from chunk server.
04/21/20 87
GFS cluster
• Atomic Record Appends:
– GFS offers Record Append .
– Clients on different machines append to the same file concurrently.
– The data is written at least once as an atomic unit.
• Snapshot:
– It creates quick copy of files or a directory .
– Master revokes lease for that file
– Duplicate metadata
– On first write to a chunk after the snapshot operation
– All chunk servers create new chunk
– Data can be copied locally
04/21/20 88
GFS : Master Operation
Namespace Management and Locking:
– GFS maps full pathname to Metadata in a table.
– Each master operation acquires a set of locks.
– Locking scheme allows concurrent mutations in same
directory.
– Locks are acquired in a consistent total order to prevent
deadlock.
Replica Placement:
– Maximizes reliability, availability and network bandwidth
utilization.
– Spread chunk replicas across racks
04/21/20 89
GFS : Master Operation
Create:
o Equalize disk utilization.
o Limit the number of creation on chunk server.
o Spread replicas across racks.
Re-replication:
o Re-replication of chunk happens on priority.
Rebalancing:
o Move replica for better disk space and load balancing.
o Remove replicas on chunk servers with below average free space.
Data Integrity:
o Check sum every 64 MB block in each chunk.
04/21/20 90
GFS : Master Operation
Garbage Collection:
o Makes system Simpler and more reliable.
o Master logs the deletion, renames the file to a hidden name.
Stale Replica detection:
o Chunk version number identifies the stale replicas.
o Client or chunk server verifies the version number.
High availability:
o Fast recovery.
o Chunk replication.
o Shadow Masters.
04/21/20 91
GFS Interaction
04/21/20 92
GFS Interaction
04/21/20 93
GFS Interaction
1. The client asks the master which chunkserver holds the current
lease for the chunk and the locations of the other replicas.
2. The master replies with the identity of the primary and the
locations of the other (secondary) replicas.
3. The client pushes the data to all the replicas. A client can do so
in any order.
4. Once all the replicas have acknowledged receiving the data, the
client sends a write request to the primary. The request
identifies the data pushed earlier to all of the replicas. The
primary assigns consecutive serial numbers to all the mutations
it receives, possibly from multiple clients, which provides the
necessary serialization.
04/21/20 94
GFS Interaction
5. The primary forwards the write request to all secondary
replicas. Each secondary replica applies mutations in the
same serial number order assigned by the primary.
6. The secondaries all reply to the primary indicating that they
have completed the operation.
7. The primary replies to the client. Any errors encountered at
any of the replicas are reported to the client. In case of
errors, the write may have succeeded at the primary and an
arbitrary subset of the secondary replicas. (If it had failed at
the primary, it would not have been assigned a serial
number and forwarded.)
04/21/20 95
GFS: Issues and solution
• More Infrastructure Requirements
– Disciplined programming approach
• Corruption of data
– Checksum for detection
• Single reader – writer problem
– Replace mmap() by pread() which requires an
extra copy of entire data
04/21/20 96
BigTable as Google’s NoSQL System:
A Distributed Storage System
Types of NoSQL databases-
There are 4 basic types of NoSQL databases:
•Key-Value Store – It has a Big Hash Table of keys
& values {Example- Riak, Amazon S3 (Dynamo)}
•Document-based Store- It stores documents
made up of tagged elements. {Example- CouchDB}
•Column-based Store- Each storage block contains
data from only one column, {Example- HBase,
Cassandra}
•Graph-based-A network database that uses
edges and nodes to represent and store data.
{Example- Neo4J}
Big Table
• Cloud Bigtable is a sparsely populated table that can scale
to billions of rows and thousands of columns, enabling you
to store terabytes or even petabytes of data.
• A single value in each row is indexed; this value is known
as the row key.
• Cloud Bigtable is ideal for storing very large amounts of
single-keyed data with very low latency.
• It supports high read and write throughput at low latency,
and it is an ideal data source for MapReduce operations
04/21/20 99
• You can use Cloud Bigtable to store and query
all of the following types of data:
Time-series data, such as CPU and memory
usage over time for multiple servers.
Marketing data, such as purchase histories and
customer preferences.
Financial data, such as transaction histories,
stock prices, and currency exchange rates.
Internet of Things data, such as usage reports
from energy meters and home appliances.
Graph data, such as information about how
users are connected to one another.
Cloud Bigtable storage model
• Cloud Bigtable stores data in massively scalable
tables, each of which is a sorted key/value map.
• The table is composed of rows, each of which
typically describes a single entity, and columns,
which contain individual values for each row.
• Each row is indexed by a single row key, and
columns that are related to one another are
typically grouped together into a column family.
• Each column is identified by a combination of the
column family and a column qualifier, which is a
unique name within the column family.
• Each row/column intersection can contain
multiple cells at different timestamps,
providing a record of how the stored data has
been altered over time.
• Cloud Bigtable tables are sparse; if a cell does
not contain any data, it does not take up any
space.
• For example, suppose you're building a social
network for United States presidents—let's
call it Prezzy. Each president can follow posts
from other presidents. The following
illustration shows a Cloud Bigtable table that
tracks who each president is following on
Prezzy:
• The table contains one column family, the follows
family. This family contains multiple column
qualifiers.
• Column qualifiers are used as data. This design
choice takes advantage of the sparseness of Cloud
Bigtable tables, and the fact that new column
qualifiers can be added on the fly.
• The username is used as the row key. Assuming
usernames are evenly spread across the alphabet,
data access will be reasonably uniform across the
entire table.
Cloud Bigtable architecture
• all client requests go through a front-end server
before they are sent to a Cloud Bigtable node.
(In the original Bigtable whitepaper, these nodes
are called "tablet servers.")
• The nodes are organized into a Cloud Bigtable
cluster, which belongs to a Cloud Bigtable
instance, a container for the cluster
• Each node in the cluster handles a subset of
the requests to the cluster. By adding nodes
to a cluster, you can increase the number of
simultaneous requests that the cluster can
handle, as well as the maximum throughput
for the entire cluster.
• If you enable replication by adding a second
cluster, you can also send different types of
traffic to different clusters
• you can fail over to one cluster if the other
cluster becomes unavailable.
• A Cloud Bigtable table is sharded (partitions for
fast access)into blocks of contiguous rows,
called tablets, to help balance the workload of
queries. (Tablets are similar to HBase regions.)
Tablets are stored on Colossus, Google's file
system, in SSTable format.
04/21/20 115
Big Table
• Lots of (semi-)structured data at Google
– URLs:
• Contents, crawl metadata, links, anchors, pagerank, …
– Per-user data:
• User preference settings, recent queries/search results, …
– Geographic locations:
• Physical entities (shops, restaurants, etc.), roads, satellite image
data, user annotations, …
• Scale is large
– Billions of URLs, many versions/page (~20K/version)
– Hundreds of millions of users, thousands or q/sec
– 100TB+ of satellite image data
04/21/20 116
BigTable
Goals
•Want asynchronous processes to be continuously
updating different pieces of data
– Want access to most current data at any time
•Need to support:
– Very high read/write rates (millions of IOPS)
– Efficient scans over all or interesting subsets of data
– Efficient joins of large one-to-one and one-to-many datasets
•Often want to examine data changes over time
– E.g. Contents of a web page over multiple crawls
04/21/20 117
BigTable
Features
•Distributed multi-level map
•Fault-tolerant, persistent
•Scalable
– Thousands of servers
– Terabytes of in-memory data
– Petabyte of disk-based data
– Millions of reads/writes per second, efficient scans
•Self-managing
– Servers can be added/removed dynamically
– Servers adjust to load imbalance
04/21/20 118
BigTable: Building Blocks
1. Google File System (GFS): Raw storage
– stores persistent data (SSTable file format for storage )
04/21/20 119
BigTable: Basic Data Model
04/21/20 120
Basic Data Model
• A BigTable is a sparse, distributed, persistent, multi-
dimensional, sorted map
(row, column, timestamp) -> cell contents
04/21/20 121
Basic Data Model: Rows
04/21/20 123
Basic Data Model : Timestamps
04/21/20 124
Example
04/21/20 125
04/21/20 126
Example 1
04/21/20 127
Example 2
04/21/20 128
Example 3
04/21/20 129
Bigtable Example
04/21/20 130
BigTable: Components
04/21/20 131
BigTable: Components
• Client
• One master server
– Responsible for
• Assigning tablets to tablet servers
• Detecting addition and expiration of tablet servers
• Balancing tablet-server load
• Garbage collection
• Many tablet servers
– Tablet servers handle read and write requests to its table
– Splits tablets that have grown too large
04/21/20 132
BigTable: Tablets
• Large tables broken into tablets at row boundaries
– Tablet holds contiguous range of rows
• Clients can often choose row keys to achieve locality
– Aim for ~100MB to 200MB of data per tablet
• Serving machine responsible for ~100 tablets
– Fast recovery:
• 100 machines each pick up 1 tablet for failed machine
– Fine-grained load balancing:
• Migrate tablets away from overloaded machine
• Master makes load-balancing decisions
04/21/20 133
04/21/20 134
BigTable: Tablet Assignment
04/21/20 135
Big Table API
• Metadata operations
– Create/delete tables, column families, change metadata
• Writes (atomic)
– Set(): write cells in a row
– DeleteCells(): delete cells in a row
– DeleteRow(): delete all cells in a row
• Reads
– Scanner: read arbitrary cells in a bigtable
• Each row read is atomic
• Can restrict returned rows to a particular range
• Can ask for just data from 1 row, all rows, etc.
• Can ask for all columns, just certain column families, or specific columns
04/21/20 136
Benefits of BigTable
• Elastic scaling
• Bigger Data Handling Capability
• Maintaining NoSQL Servers is Cheaper
• Lesser Server Cost
• No Schema or Fixed Data model
• Integrated Caching Facility
04/21/20 137
Chubby
(Google Distributed Lock Service)
Chubby : Distributed Lock service
• provide reliable storage to the loosely coupled distributed
system
• Synchronize access to shared resources
• Goals
– High availability
– Reliability
• Anti-goals:
– High performance
– Throughput
– Storage capacity
04/21/20 139
Chubby
• Presents a simple distributed file system
• Clients can open/close/read/write files
– Reads and writes are whole-file
– Also supports advisory reader/writer locks
– Clients can register for notification of file update
04/21/20 140
Chubby : Architecture
04/21/20 141
Chubby : Architecture
• Chubby cell has set of servers (or replicas)
• All client requests are directed to master
– updates propagated to replicas
– master periodically polls for failed replicas
• Chubby cell is usually 5 replicas
– 3 must be alive for cell to be viable
• Periodically elected master from 5 replicas
• How do replicas in Chubby agree on their own master,
official lock values?
– PAXOS (distributed) algorithm
04/21/20 142
Chubby : Architecture
• Master election is simple: all replicas try to acquire a
write lock on designated file. The one who gets the
lock is the master.
– Master can then write its address to file; other replicas can
read this file to discover the chosen master name.
04/21/20 145
Design - Sequencer for lock
• The Status of locks after they acquired can described
using descriptor strings called as sequencer
– introduce sequence numbers into interactions
that use locks
– lock holder requests a sequencer, pass it to file
server to validate
04/21/20 146
Design - Events
• Client subscribes to events when creating handle
• Event types
– file contents modified
– child node added / removed / modified
– Chubby master failed over
– handle / lock have become invalid
– lock acquired / conflicting lock request (rarely used)
• Events are delivered when the action that corresponds to it is
completed
• Chubby is implemented using following API
– Creation of handle using Open() method
– Destruction of handles using Close() method
04/21/20 147
Mobile Cloud Computing
Outline
• Introduction
• Definition
• Architecture
• Benefits
• Challenges
MCC
• MCC is the combination of cloud computing, mobile
computing and wireless networks to bring rich
computational resources to mobile users, network
operators, as well as cloud computing providers.
• NIST defines MCC as “a model for enabling
convenient, on-demand network access to a shared
pool of configurable computing resources (e.g.
network, server, storage, applications and services)
that can be rapidly provisioned and released with
minimal management effort or service provider
interaction.”
Architecture
• Cloud as a large scale distributed systems is based
on various servers that are connected to data
centers.
• Layered Architecture for cloud
• Data Center
• IaaS
• PaaS
• SaaS
Architecture
Benefits
• Extended Lifetime of the battery
• Improved data storage capacity and processing
power
• Improved reliability
• Dynamic provisioning
• Scalability
• Multi-tenancy
• Ease of Integration
Challenges
• Challenges at Mobile End
– Network latency and limited bandwidth
– Service availability
– Heterogeneity of platform, devices and service providers
• Challenges at cloud End
– Computing offload
– Security
– Enhancing the efficiency of Data Access
– Context Aware Mobile cloud Services
Summary
• Definition
• Architecture
• Benefits
• Challenges
AAA Administration for clouds
Outline
• AAA model
• SSO for Clouds
• Authentication management in cloud
• Authorization management in clouds
• Accounting for Resource utilization in cloud
AAA Model
• AAA (a sequence of events when a user logs in)
• Authentication- Security server first checks if the
login name and password are legitimate then user
is authenticated
• Authorization - The user is given access to
modules of application or sets of data that he can
use or view
• Accounting -The server keeps a log or account of
all the resources utilized and the user activities
1. Authentication
04/21/20 179
AWS
• AWS is a secure cloud services platform, offering
compute power, database storage, content delivery and
other functionality to help businesses scale and grow.
• solutions to build sophisticated applications with
increased flexibility, scalability and reliability.
• The AWS Cloud provides a broad set of infrastructure
services, such as computing power, storage options,
networking and databases, delivered as a utility: on-
demand, available in seconds, with pay-as-you-go
pricing.
Advantages
• Trade capital expense for variable expense
• Benefit from massive economies of scale
• Stop guessing capacity
• Increase speed and agility
• Stop spending money on running and
maintaining data centers
• Go global in minutes
AWS
• Amazon operates at least 30 data centers in its global
network, with another 10 to 15 on the drawing board.
• Amazon doesn’t disclose the full scope of its
infrastructure, but third-party estimates peg its U.S.
data center network at about 600 megawatts of IT
capacity.(<100,000 servers per data center)
• Figuring out the upper end of the range is more
difficult, but could range as high as 5.6 million,
according to calculations by Timothy Prickett Morgan at
the Platform.
AWS data centers
Cloud Solutions
• Solutions- Mobile Services, Websites, Backup
and Recovery
• Hundreds of thousands of customers have joined
the Amazon Web Services (AWS) community and
use AWS solutions to build their businesses.
• The AWS cloud computing platform provides the
flexibility to build your application, regardless of
your use case or industry.
• You can save time, money, and let AWS manage
your infrastructure, without compromising
scalability, security, or dependability.
Cloud Products & Services
• Amazon Web Services (AWS) offers a broad
set of global compute, storage, database,
analytics, application, and deployment
services that help organizations move faster,
lower IT costs, and scale applications.
• Eg EC2, S3, RDS, Aurora
Amazon EC2
• Amazon Elastic Compute Cloud (Amazon EC2)
is a web service that provides secure, resizable
compute capacity in the cloud. It is designed
to make web-scale cloud computing easier for
developers.
Amazon S3
• Amazon Simple Storage Service (Amazon S3) is
object storage with a simple web service
interface to store and retrieve any amount of
data from anywhere on the web.
Amazon RDS
• Amazon Relational Database Service makes it
easy to set up, operate, and scale a relational
database in the cloud.
AWS Lambda
• AWS Lambda lets you run code without
provisioning or managing servers. You pay
only for the compute time you consume -
there is no charge when your code is not
running.
Amazon QuickSight
• Amazon QuickSight is a fast, cloud-powered
business analytics service that makes it easy
to build visualizations, perform ad-hoc
analysis, and quickly get business insights
from your data.
AWS IoT
• AWS IoT is a managed cloud platform that lets
connected devices easily and securely interact
with cloud applications and other devices.
Amazon Aurora
• Amazon Aurora is a MySQL and PostgreSQL-
compatible relational database engine that
combines the speed and availability of high-end
commercial databases with the simplicity and
cost-effectiveness of open source databases.
• Amazon Aurora provides up to five times better
performance than MySQL with the security,
availability, and reliability of a commercial
database at one tenth the cost.
Amazon Aurora
Amazon EC2
Amazon EC2
• Amazon Elastic Compute Cloud (Amazon EC2) is a web
service that provides secure, resizable compute capacity in
the cloud. It is designed to make web-scale cloud
computing easier for developers.
• Benefits
– Elastic Web-Scale Computing
– Completely Controlled
– Flexible Cloud Hosting Services
– Integrated
– Reliable
– Secure
– Inexpensive
– Easy to start
Amazon EC2 Pricing
• The pricing below includes the cost to run private and public
AMIs on the specified operating system (“Windows Usage”
prices apply to Windows Server 2003 R2, 2008, 2008 R2,
2012, 2012 R2, and 2016). Amazon also provides you with
additional instances for Amazon EC2 running Microsoft
Windows with SQL Server, Amazon EC2 running SUSE Linux
Enterprise Server, Amazon EC2 running Red Hat Enterprise
Linux and Amazon EC2 running IBM that are priced differently.
Amazon EC2 Spot Instances Pricing
• Spot instances provide you with access to unused Amazon EC2 capacity at steep
discounts relative to On-Demand prices. The Spot price fluctuates based on the
supply and demand of available unused EC2 capacity.
• When you request Spot instances, you specify the maximum Spot price you are
willing to pay. Your Spot instance is launched when the Spot price is lower than
the price you specified, and will continue to run until you choose to terminate it or
the Spot price exceeds the maximum price you specified.
• With Spot instances, you will never be charged more than the maximum price you
specified. While your instance runs, you are charged the Spot price that is in effect
for that period. If the Spot price exceeds your specified price, your instance will
receive a two-minute notification before it is terminated, and you will not be
charged for the partial hour that your instance has run.
• If you include a duration requirement with your Spot instances request, your
instance will continue to run until you choose to terminate it, or until the specified
duration has ended; your instance will not be terminated due to changes in the
Spot price.
Amazon EC2 Reserved Instances
Pricing
• Reserved Instances provide you with a
significant discount (up to 75%) compared to
On-Demand instance pricing. In addition,
when Reserved Instances are assigned to a
specific Availability Zone, they provide a
capacity reservation, giving you additional
confidence in your ability to launch instances
when you need them.
Amazon EC2 Dedicated Hosts Pricing
• The price for a Dedicated Host varies by instance family, region, and
payment option. Regardless of the quantity or the size of instances that
you choose to launch on a particular Dedicated Host you pay hourly for
each active Dedicated Host, and you are not billed for instance usage.
• When you allocate a Dedicated Host host for use, you must choose an
instance type configuration for the host. This selection will define the
number of sockets and physical cores per host, the type of instance you
can run on the host, and the number of instances that you can run on
each host.
• After you have allocated a Dedicated Host, you will pay On-Demand
unless you have a Dedicated Host Reservation. A Dedicated Host
Reservation provides you with a discount of up to 70% compared to On-
Demand pricing.
AWS Marketplace(CSB)
• AWS Marketplace provides a new sales
channel for ISVs and Consulting Partners to
sell their solutions to AWS customers. They
make it easy for customers to find, buy,
deploy and manage software solutions,
including SaaS, in a matter of minutes.
Amazon S3
Amazon S3
• Amazon Simple Storage Service is object storage with a simple
web service interface to store and retrieve any amount of data
from anywhere on the web. It is designed to deliver
99.999999999% durability, and scale past trillions of objects
worldwide.
• Customers use S3 as primary storage for cloud-native
applications; as a bulk repository, or "data lake," for analytics;
as a target for backup & recovery and disaster recovery; and
with server less computing.
• It's simple to move large volumes of data into or out of
Amazon S3 with Amazon's cloud data migration options.
• Once data is stored in S3, it can be automatically tiered into
lower cost, longer-term cloud storage classes like S3 Standard
- Infrequent Access and Amazon Glacier for archiving.
Amazon S3
Benefits
•Simple
•Durable
•Scalable
•Secure
•Available
•Low cost
•Simple Data Transfer
•Integrated
•Easy to manage
Amazon S3 Use cases
• Backup & Archiving
• Disaster Recovery
•Highly Scalable
•Secure
•Inexpensive
Summary
• Google App Engine
• Google File System
• Big Table
• Chubby
• Amazon EC2
• Amazon S3
• Amazon EBS
• Amazon Simple DB
04/21/20 229
Thank you !
04/21/20 230
04/21/20 231
Amazon EBS
• Amazon Elastic Block Store (Amazon EBS)
provides persistent block storage volumes for
use with Amazon EC2 instances in the AWS
Cloud. Each Amazon EBS volume is
automatically replicated within its Availability
Zone to protect you from component failure,
offering high availability and durability.
EC2
• AWS offers cloud programming support to its cloud
environment EC2 through various systems including
Simple Storage Service (S3 ), Elastic Block Store (EBS),
Simple DB
• Benefits
– Up and down scalability of web services
– Complete control over computing resources
– Flexibility of services for cloud hosting
– Support for other Amazon web services
– High reliability and security
– Cost efficiency
04/21/20 233
Amazon S3
• Amazon Simple Storage Service (Amazon S3), provides
developers and IT teams with secure, durable, highly-
scalable object storage.
• Amazon S3 is easy to use, with a simple web services
interface to store and retrieve any amount of data from
anywhere on the web.
• With Amazon S3, you pay only for the storage you actually
use. There is no minimum fee and no setup cost.
• Amazon S3 provides cost-effective object storage for cloud
applications, content distribution, backup and archiving,
disaster recovery, and big data analytics.
04/21/20 234
Amazon S3
• Design requirements
– Durable- durability of 99.999999999% of objects.
– Low Cost
– Available- 99.99% availability of objects
– Secure- data transfer over SSL and automatic encryption of your data
once it is uploaded.
– Scalable
– Send Event Notifications
– High Performance
– Integrated
– Easy to use
04/21/20 235
Amazon EBS
• Amazon Elastic Block Store (Amazon EBS) provides
persistent block level storage volumes for use with Amazon
EC2 instances in the AWS Cloud.
• Each Amazon EBS volume is automatically replicated within
its Availability Zone to protect you from component failure,
offering high availability and durability.
• Amazon EBS volumes offer the consistent and low-latency
performance needed to run your workloads.
• With Amazon EBS, you can scale your usage up or down
within minutes – all while paying a low price for only what
you provision.
04/21/20 236
Amazon EBS
• Amazon EBS top patterns
– Storing key-value – data stored on basis of primary key
access resulted into creation of S3
– Storing Data in simple and Structured form
– Storing Data in Blocks
• Benefits
– Reliable, secure , Consistent storage system
– High performance and low-latency performance
– Quickly scale up, easily scale down
– Backup, restore, innovate
04/21/20 237
Amazon EBS
• EBS is utilized in various ways to render benefits
– EBS usage as database
– EBS usage in applications developed for
enterprises
– EBS usage as NoSQL
– EBS usage in development and testing
environments
– EBS usage in continuing businesses
– EBS usage in file workloads
04/21/20 238
Amazon SimpleDB
• Amazon SimpleDB is a fast, scalable system that would
provide fully managed database services
• It is a highly available and flexible non-relational data
store that offloads the work of database administration.
• Developers simply store and query data items via web
services requests and Amazon SimpleDB does the rest.
• Amazon SimpleDB is optimized to provide high
availability and flexibility, with little or no administrative
burden.
04/21/20 239
Amazon SimpleDB
• Amazon SimpleDB creates and manages multiple
geographically distributed replicas of your data
automatically to enable high availability and data durability.
• The service charges only for the resources actually
consumed in storing your data and serving requests. You
can change your data model on the fly, and data is
automatically indexed for you.
• With Amazon SimpleDB, you can focus on application
development without worrying about infrastructure
provisioning, high availability, software maintenance,
schema and index management, or performance tuning.
04/21/20 240
Amazon SimpleDB
Characteristics
– Scalable- Scaling the storage automatically, Provisioning of throughput, Fully
Distributed Architecture without sharing
– Easy Administration
– Flexible
– Fast, Predictable Performance
– Built-in Fault Tolerance
– Schema-less
– Strong consistency, Atomic Counters
– Cost Effective
– Secure
– Integrated Monitoring
04/21/20 241
Q2) Describe Platform as a Service
OR
OR