Professional Documents
Culture Documents
Lecture 1 Ho PDF
Lecture 1 Ho PDF
Distributed Database
Management Systems
Dist ributed DBMS 2
Outline
B Introduction
B Distributed DBMS Architecture
B Distributed Database Design
B Distributed Query Processing
B Distributed Concurrency Control
B Distributed Reliability Protocols
Page 2
Dist ributed DBMS 3
Outline
O Introduction
=What is a distributed DBMS
=Problems
=Current state-of-affairs
O Distributed DBMS Architecture
O Distributed Database Design
O Distributed Query Processing
O Distributed Concurrency Control
O Distributed Reliability Protocols
Dist ributed DBMS 4
Motivation
Database
Technology
Computer
Networks
i ntegration distribut ion
i ntegration
integrati on central ization
Distributed
Database
Systems
Page 3
Dist ributed DBMS 5
What is a Distributed Database
System?
A distributed database (DDB) is a collection of multiple,
logically interrelated databases distributed over a computer
network.
A distributed database management system (DDBMS) is
the software that manages the DDB and provides an
access mechanism that makes this distribution transparent
to the users.
Distributed database system (DDBS) = DDB + DDBMS
Dist ributed DBMS 6
B A timesharing computer system
B A loosely or tightly coupled multiprocessor system
B A database system which resides at one of the
nodes of a network of computers - this is a
centralized database on a network node
What is not a DDBS?
Page 4
Dist ributed DBMS 7
Centralized DBMS on a Network
Communication
Network
Site 5
Site 1
Site 2
Site 3
Site 4
Dist ributed DBMS 8
Distributed DBMS Environment
Communication
Network
Site 5
Site 1
Site 2
Site 3
Site 4
Page 5
Dist ributed DBMS 9
Implicit Assumptions
B Data stored at a number of sites each site
logically consists of a single processor.
B Processors at different sites are interconnected
by a computer network no multiprocessors
=parallel database systems
B Distributed database is a database, not a
collection of files data logically related as
exhibited in the users access patterns
=relational data model
B D-DBMS is a full-fledged DBMS
=not remote file system, not a TP system
Dist ributed DBMS 10
Distributed DBMS Promises
0Transparent management of distributed,
fragmented, and replicated data
OImproved reliability/availability through distributed
transactions
OImproved performance
OEasier and more economical system expansion
Page 6
Dist ributed DBMS 11
Transparency
B Transparency is the separation of the higher level
semantics of a system from the lower level
implementation issues.
B Fundamental issue is to provide
data independence
in the distributed environment
=Network (distribution) transparency
=Replication transparency
=Fragmentation transparency
+ horizontal fragmentation: selection
+ vertical fragmentation: projection
+ hybrid
Dist ributed DBMS 12
Example
TITLE SAL
PAY
Elect. Eng. 40000
Syst. Anal. 34000
Mech. Eng. 27000
Programmer 24000
PROJ
PNO PNAME BUDGET
ENO ENAME TITLE
E1 J. Doe Elect. Eng.
E2 M. Smi th Syst. Anal.
E3 A. Lee Mech. Eng.
E4 J. Mi l ler Programmer
E5 B. Casey Syst. Anal.
E6 L. Chu Elect. Eng.
E7 R. Davis Mech. Eng.
E8 J. Jones Syst. Anal.
EMP
ENO PNO RESP
E1 P1 Manager 12
DUR
E2 P1 Analyst 24
E2 P2 Analyst 6
E3 P3 Consul tant 10
E3 P4 Engineer 48
E4 P2 Programmer 18
E5 P2 Manager 24
E6 P4 Manager 48
E7 P3 Engineer 36
E8 P3 Manager 40
ASG
P1 I nst rumentat ion 150000
P3 CAD/CAM 250000
P2 Database Develop. 135000
P4 Maintenance 310000
E7 P5 Engineer 23
Page 7
Dist ributed DBMS 13
Transparent Access
SELECT ENAME,SAL
FROM EMP,ASG,PAY
WHERE DUR > 12
AND EMP.ENO = ASG.ENO
AND PAY.TITLE = EMP.TITLE
Paris projects
Paris employees
Paris assignments
Boston employees
Montreal projects
Paris projects
New York projects
with budget > 200000
Montreal employees
Montreal assignments
Boston
Communication
Network
Mont real
Paris
New
York
Boston projects
Boston employees
Boston assignments
Boston projects
New York employees
New York projects
New York assignments
Tokyo
Dist ributed DBMS 14
Distributed Database
Distributed Database
User View
Page 8
Dist ributed DBMS 15
Distributed DBMS - Reality
Communication
Subsystem
User
Query
DBMS
Software
DBMS
Software
User
Application
DBMS
Software
User
Application
User
Query
DBMS
Software
User
Query
DBMS
Software
Dist ributed DBMS 16
Potentially Improved
Performance
B Proximity of data to its points of use
=Requires some support for fragmentation and replication
B Parallelism in execution
=Inter-query parallelism
=Intra-query parallelism
Page 9
Dist ributed DBMS 17
Parallelism Requirements
B Have as much of the data required by each
application at the site where the application
executes
=Full replication
B How about updates?
=Updates to replicated data requires implementation of
distributed concurrency control and commit protocols
Dist ributed DBMS 18
System Expansion
B Issue is database scaling
B Emergence of microprocessor and workstation
technologies
=Demise of Grosh's law
=Client-server model of computing
B Data communication cost vs telecommunication
cost
Page 10
Dist ributed DBMS 19
Distributed DBMS Issues
BDistributed Database Design
=how to distribute the database
=replicated & non-replicated database distribution
=a related problem in directory management
B Query Processing
=convert user transactions to data manipulation instructions
=optimization problem
=min{cost = data transmission + local processing}
=general formulation is NP-hard
Dist ributed DBMS 20
Distributed DBMS Issues
B Concurrency Control
=synchronization of concurrent accesses
=consistency and isolation of transactions' effects
=deadlock management
B Reliability
=how to make the system resilient to failures
=atomicity and durability
Page 11
Dist ributed DBMS 21
Di rectory
Management
Relationship Between Issues
Reliabi l i ty
Deadlock
Management
Query
Processing
Concurrency
Cont rol
Dist ribut ion
Design
Dist ributed DBMS 22
Outline
B Introduction
B Distributed DBMS Architecture
O Distributed Database Design
=Fragmentation
=Data Placement
O Distributed Query Processing
O Distributed Concurrency Control
O Distributed Reliability Protocols
Page 12
Dist ributed DBMS 23
Design Problem
B In the general setting :
=Making decisions about the placement of data and programs
across the sites of a computer network as well as possibly
designing the network itself.
B In Distributed DBMS, the placement of
applications entails
=placement of the distributed DBMS software; and
=placement of the applications that run on the database
Dist ributed DBMS 24
Distribution Design
B Top-down
=mostly in designing systems from scratch
=mostly in homogeneous systems
B Bottom-up
=when the databases already exist at a number of sites
Page 13
Dist ributed DBMS 25
Top-Down Design
User Input
View Integration
User Input
Requirements
Analysis
Objectives
Conceptual
Design
View Design
Access
Information
ESs GCS
Distribution
Design
Physical
Design
LCSs
LISs
Dist ributed DBMS 26
Distribution Design Issues
0Why fragment at all?
OHow to fragment?
OHow much to fragment?
OHow to test correctness?
OHow to allocate?
OInformation requirements?
Page 14
Dist ributed DBMS 27
Fragmentation
B Can't we just distribute relations?
B What is a reasonable unit of distribution?
=relation
+ views are subsets of relations locality
+ extra communication
=fragments of relations (sub-relations)
+ concurrent execution of a number of transactions that
access different portions of a relation
+ views that cannot be defined on a single fragment will
require extra processing
+ semantic data control (especially integrity enforcement)
more difficult
Dist ributed DBMS 28
PROJ
1
: projects with budgets less
than $200,000
PROJ
2
: projects with budgets
greater than or equal to
$200,000
PROJ
1
PNO PNAME BUDGET
LOC
P3 CAD/CAM 250000 New York
P4 Maintenance 310000 Paris
P5 CAD/CAM 500000 Boston
PNO PNAME
LOC
P1 Instrumentation 150000 Montreal
P2 Database Develop. 135000 New York
BUDGET
PROJ
2
Fragmentation Alternatives
Horizontal
New York
New York
PROJ
PNO PNAME BUDGET LOC
P1 Instrumentation 150000 Montreal
P3 CAD/CAM 250000
P2 Database Develop. 135000
P4 Maintenance 310000 Paris
P5 CAD/CAM 500000 Boston
New York
New York
Page 15
Dist ributed DBMS 29
Fragmentation Alternatives
Vertical
PROJ
1
: information about
project budgets
PROJ
2
: information about
project names and
locations
PNO BUDGET
P1 150000
P3 250000
P2 135000
P4 310000
P5 500000
PNO PNAME LOC
P1 Instrumentation Montreal
P3 CAD/CAM New York
P2 Database Develop. New York
P4 Maintenance Paris
P5 CAD/CAM Boston
PROJ
1
PROJ
2
New York
New York
PROJ
PNO PNAME BUDGET LOC
P1 Instrumentation 150000 Montreal
P3 CAD/CAM 250000
P2 Database Develop. 135000
P4 Maintenance 310000 Paris
P5 CAD/CAM 500000 Boston
New York
New York
Dist ributed DBMS 30
Degree of Fragmentation
Finding the sui table level of part i t ioning
wi thi n this range
tuples
or
att ributes
relations
fini te number of al ternati ves
Page 16
Dist ributed DBMS 31
B Completeness
= Decomposition of relation R into fragments R
1
, R
2
, ..., R
n
is complete
iff each data item in R can also be found in some R
i
B Reconstruction
= If relation R is decomposed into fragments R
1
, R
2
, ..., R
n
, then there
should exist some relational operator such that
R =
1in
R
i
B Disjointness
= If relation R is decomposed into fragments R
1
, R
2
, ..., R
n
, and data
item d
i
is in R
j
, then d
i
should not be in any other fragment R
k
(k j ).
Correctness of Fragmentation
Dist ributed DBMS 32
Allocation Alternatives
B Non-replicated
= partitioned : each fragment resides at only one site
B Replicated
= fully replicated : each fragment at each site
= partially replicated : each fragment at some of the sites
B Rule of thumb:
If replication is advantageous,
otherwise replication may cause problems
read - only queries
update queries
1
Page 17
Dist ributed DBMS 33
B Horizontal Fragmentation (HF)
=Primary Horizontal Fragmentation (PHF)
=Derived Horizontal Fragmentation (DHF)
B Vertical Fragmentation (VF)
B Hybrid Fragmentation (HF)
Fragmentation
Dist ributed DBMS 34
Definition :
R
j
=
F
j
(R ), 1 j w
where F
j
is a selection formula.
Therefore,
A horizontal fragment R
i
of relation R consists of all
the tuples of R which satisfy a predicate p
i
.
all si tes
Page 25
Dist ributed DBMS 49
B Query Processing Cost
Processing component
access cost + integrity enforcement cost + concurrency control cost
=Access cost
=Integrity enforcement and concurrency control costs
+ Can be similarly calculated
Allocation Model
(no. of update accesses+ no. of read accesses)
all fragments
all si tes
x
i j
local processing cost at a site
Dist ributed DBMS 50
B Query Processing Cost
Transmission component
cost of processing updates + cost of processing retrievals
=Cost of updates
=Retrieval Cost
Allocation Model
update message cost +
all fragments
all si tes
acknowledgment cost
all fragments
all si tes
( min
all si tes
all fragments
ENAME
(
DUR>37EMP.ENO=ASG.ENO
(EMP ASG))
Strategy 2
ENAME
(EMP
ENO
(
DUR>37
(ASG)))
Strategy 2 avoids Cartesian product, so is better
Selecting Alternatives
Dist ributed DBMS 58
What is the Problem?
Si te 1 Si te 2 Si te 3 Si te 4 Si te 5
EMP
1
=
ENOE3
(EMP) EMP
2
=
ENO>E3
(EMP) ASG
2
=
ENO>E3
(ASG) ASG
1
=
ENOE3
(ASG) Result
Si te 5
Si te 1 Si te 2 Si te 3 Si te 4
ASG
1
EMP
1
EMP
2
ASG
2
result
2
=(EMP
1
EMP
2
)
ENO
DUR>37
(ASG
1
ASG
1
)
Si te 4
result = EMP
1
EMP
2
Si te 3
Si te 1 Si te 2
EMP
2
=EMP
2 ENO
ASG
2
EMP
1
=EMP
1 ENO
ASG
1
ASG
1
=
DUR>37
(ASG
1
) ASG
2
=
DUR>37
(ASG
2
)
Si te 5
ASG
2
ASG
1
EMP
1
EMP
2
Page 30
Dist ributed DBMS 59
B Assume:
= size(EMP) = 400, size(ASG) = 1000
= tuple access cost = 1 unit; tuple transfer cost = 10 units
B Strategy 1
0 produce ASG': (10+10)tuple access cost 20
O transfer ASG' to the sites of EMP: (10+10)tuple transfer cost 200
O produce EMP': (10+10) tuple access cost2 40
O transfer EMP' to result site: (10+10) tuple transfer cost 200
Total cost 460
B Strategy 2
0 transfer EMP to site 5:400tuple transfer cost 4,000
O transfer ASG to site 5 :1000tuple transfer cost 10,000
O produce ASG':1000tuple access cost 1,000
O join EMP and ASG':40020tuple access cost 8,000
Total cost 23,000
Cost of Alternatives
Dist ributed DBMS 60
Minimize a cost function
I/O cost + CPU cost + communication cost
These might have different weights in different
distributed environments
Wide area networks
= communication cost will dominate
+ low bandwidth
+ low speed
+ high protocol overhead
= most algorithms ignore all other cost components
Local area networks
= communication cost not that dominant
= total cost function should be considered
Can also maximize throughput
Query Optimization Objectives
Page 31
Dist ributed DBMS 61
Query Optimization Issues
Types of Optimizers
B Exhaustive search
=cost-based
=optimal
=combinatorial complexity in the number of relations
B Heuristics
=not optimal
=regroup common sub-expressions
=perform selection, projection first
=replace a join by a series of semijoins
=reorder operations to reduce intermediate relation size
=optimize individual operations
Dist ributed DBMS 62
Query Optimization Issues
Optimization Granularity
B Single query at a time
=cannot use common intermediate results
B Multiple queries at a time
=efficient if many similar queries
=decision space is much larger
Page 32
Dist ributed DBMS 63
Query Optimization Issues
Optimization Timing
B Static
=compilation optimize prior to the execution
=difficult to estimate the size of the intermediate results
error propagation
=can amortize over many executions
=R*
B Dynamic
=run time optimization
=exact information on the intermediate relation sizes
=have to reoptimize for multiple executions
=Distributed INGRES
B Hybrid
=compile using a static algorithm
=if the error in estimate sizes > threshold, reoptimize at run
time
=MERMAID
Dist ributed DBMS 64
Query Optimization Issues
Statistics
B Relation
=cardinality
=size of a tuple
=fraction of tuples participating in a join with another relation
B Attribute
=cardinality of domain
=actual number of distinct values
B Common assumptions
=independence between different attribute values
=uniform distribution of attribute values within their domain
Page 33
Dist ributed DBMS 65
Query Optimization
Issues Decision Sites
B Centralized
=single site determines the best schedule
=simple
=need knowledge about the entire distributed database
B Distributed
=cooperation among sites to determine the schedule
=need only local information
=cost of cooperation
B Hybrid
=one site determines the global schedule
=each site optimizes the local subqueries
Dist ributed DBMS 66
Query Optimization Issues
Network Topology
B Wide area networks (WAN) point-to-point
=characteristics
+ low bandwidth
+ low speed
+ high protocol overhead
=communication cost will dominate; ignore all other cost factors
=global schedule to minimize communication cost
=local schedules according to centralized query optimization
B Local area networks (LAN)
=communication cost not that dominant
=total cost function should be considered
=broadcasting can be exploited (joins)
=special algorithms exist for star networks
Page 34
Dist ributed DBMS 67
Distributed Query Processing
Methodology
Calculus Query on Dist ributed
Relations
CONTROL
SITE
LOCAL
SITES
Query
Decomposition
Query
Decomposition
Data
Localization
Data
Localization
Algebraic Query on Dist ributed
Relations
Global
Optimization
Global
Optimization
Fragment Query
Local
Optimization
Local
Optimization
Opti mized Fragment Query
wi th Communicati on Operations
Opti mized Local
Queries
GLOBAL
SCHEMA
GLOBAL
SCHEMA
FRAGMENT
SCHEMA
FRAGMENT
SCHEMA
STATS ON
FRAGMENTS
STATS ON
FRAGMENTS
LOCAL
SCHEMAS
LOCAL
SCHEMAS
Dist ributed DBMS 68
Step 1 Query Decomposition
Input : Calculus query on global relations
B Normalization
= manipulate query quantifiers and qualification
B Analysis
= detect and reject incorrect queries
= possible for only a subset of relational calculus
B Simplification
= eliminate redundant predicates
B Restructuring
= calculus query algebraic query
= more than one translation is possible
= use transformation rules
Page 35
Dist ributed DBMS 69
B Lexical and syntactic analysis
= check validity (similar to compilers)
= check for attributes and relations
= type checking on the qualification
B Put into normal form
= Conjunctive normal form
(p
11
p
12
p
1n
) (p
m1
p
m2
p
mn
)
= Disjunctive normal form
(p
11
p
12
p
1n
) (p
m1
p
m2
p
mn
)
= OR's mapped into union
= AND's mapped into join or selection
Normalization
Dist ributed DBMS 70
B Refute incorrect queries
B Type incorrect
= If any of its attribute or relation names are not defined
in the global schema
= If operations are applied to attributes of the wrong type
B Semantically incorrect
= Components do not contribute in any way to the
generation of the result
= Only a subset of relational calculus queries can be
tested for correctness
= Those that do not contain disjunction and negation
= To detect
+ connection graph (query graph)
+ join graph
Analysis
Page 36
Dist ributed DBMS 71
B Why simplify?
= Remember the example
B How? Use transformation rules
= elimination of redundancy
+ idempotency rules
p
1
( p
1
) false
p
1
(p
1
p
2
) p
1
p
1
false p
1
= application of transitivity
= use of integrity rules
Simplification
Dist ributed DBMS 72
SELECT TITLE
FROM EMP
WHERE EMP.ENAME = J. Doe
OR (NOT(EMP.TITLE = Programmer)
AND (EMP.TITLE = Programmer
OR EMP.TITLE = Elect. Eng.)
AND NOT(EMP.TITLE = Elect. Eng.))
SELECT TITLE
FROM EMP
WHERE EMP.ENAME = J. Doe
Simplification Example
Page 37
Dist ributed DBMS 73
B Convert relational calculus to
relational algebra
B Make use of query trees
B Example
Find the names of employees other than J.
Doe who worked on the CAD/CAM project
for either 1 or 2 years.
SELECT ENAME
FROM EMP, ASG, PROJ
WHERE EMP.ENO = ASG.ENO
AND ASG.PNO = PROJ.PNO
AND ENAME J. Doe
AND PNAME = CAD/CAM
AND (DUR = 12 OR DUR = 24)
Restructuring
ENAME
DUR=12 OR DUR=24
PNAME=CAD/CAM
ENAMEJ. DOE
PROJ ASG EMP
Project
Select
Join
PNO
ENO
Dist ributed DBMS 74
B Commutativity of binary operations
= R S S R
= R S S R
= R S S R
B Associativity of binary operations
= ( R S ) T R (S T)
= ( R S ) T R (S T )
B Idempotence of unary operations
=
A
(
A
(R))
A
(R)
=
p
1
(A
1
)
(
p
2
(A
2
)
(R)) =
p
1
(A
1
)
p
2
(A
2
)
(R)
where R[A] and A' A, A" A and A' A"
B Commuting selection with projection
Restructuring Transformation
Rules
Page 38
Dist ributed DBMS 75
B Commuting selection with binary operations
=
p(A)
(R S) (
p(A)
(R)) S
=
p(A
i
)
(R
(A
j
,B
k
)
S) (
p(A
i
)
(R))
(A
j
,B
k
)
S
=
p(A
i
)
(R T)
p(A
i
)
(R)
p(A
i
)
(T)
where A
i
belongs to R and T
B Commuting projection with binary operations
=
C
(R S)
A
(R)
B
(S)
=
C
(R
(A
j
,B
k
)
S)
A
(R)
(A
j
,B
k
)
B
(S)
=
C
(R S)
C
(R)
C
(S)
where R[A] and S[B]; C = A' B' where A' A, B' B
Restructuring
Transformation Rules
Dist ributed DBMS 76
Example
Recall the previous example:
Find the names of employees other than J.
Doe who worked on the CAD/CAM project
for either one or two years.
SELECT ENAME
FROM PROJ, ASG, EMP
WHERE ASG.ENO=EMP.ENO
AND ASG.PNO=PROJ.PNO
AND ENAMEJ. Doe
AND PROJ.PNAME=CAD/CAM
AND (DUR=12 OR DUR=24)
ENAME
DUR=12 OR DUR=24
PNAME=CAD/CAM
ENAMEJ. DOE
PROJ ASG EMP
Project
Select
Join
PNO
ENO
Page 39
Dist ributed DBMS 77
Equivalent Query
ENAME
PNAME=CAD/CAM
(DUR=12 DUR=24)
ENAMEJ. DOE
ENO
Dist ributed DBMS 78
EMP
ENAME
PNO,ENAME
PNAME = "CAD/CAM"
PNO
PNO,ENO
PNO,ENAME
Restructuring
PNO
ENO
Page 40
Dist ributed DBMS 79
Step 2 Data Localization
Input: Algebraic query on distributed relations
B Determine which fragments are involved
B Localization program
=substitute for each global query its materialization program
=optimize
Dist ributed DBMS 80
Example
Assume
= EMP is fragmented into EMP
1
, EMP
2
,
EMP
3
as follows:
+ EMP
1
=
ENOE3
(EMP)
+ EMP
2
=
E3<ENOE6
(EMP)
+ EMP
3
=
ENOE6
(EMP)
= ASG fragmented into ASG
1
and ASG
2
as follows:
+ ASG
1
=
ENOE3
(ASG)
+ ASG
2
=
ENO>E3
(ASG)
Replace EMP by (EMP
1
EMP
2
EMP
3
)
and ASG by (ASG
1
ASG
2
) in any
query
ENAME
DUR=12 OR DUR=24
PNAME=CAD/CAM
ENAMEJ. DOE
PROJ
EMP
1
EMP
2
EMP
3
ASG
1
ASG
2
PNO
ENO
Page 41
Dist ributed DBMS 81
Provides Parallellism
EMP
3
ASG
1
EMP
2
ASG
2
EMP
1
ASG
1
EMP
3
ASG
2
ENO ENO ENO ENO
Dist ributed DBMS 82
Eliminates Unnecessary Work
EMP
2
ASG
2
EMP
1
ASG
1
EMP
3
ASG
2
ENO ENO ENO
Page 42
Dist ributed DBMS 83
B Reduction with selection
= Relation R and F
R
={R
1
, R
2
, , R
w
} where R
j
=
p
j
(R)
p
i
(R
j
)= if x in R: (p
i
(x) p
j
(x))
= Example
SELECT *
FROM EMP
WHERE ENO=E5
Reduction for PHF
ENO=E5
EMP
1
EMP
2
EMP
3
EMP
2
ENO=E5
Dist ributed DBMS 84
B Reduction with join
= Possible if fragmentation is done on join attribute
= Distribute join over union
(R
1
R
2
) S (R
1
S) (R
2
S)
= Given R
i
=
p
i
(R) and R
j
=
p
j
(R)
R
i
R
j
= if x in R
i
, y in R
j
: (p
i
(x) p
j
(y))
Reduction for PHF
Page 43
Dist ributed DBMS 85
B Reduction with join - Example
=Assume EMP is fragmented as before and
ASG
1
:
ENO "E3"
(ASG)
ASG
2
:
ENO > "E3"
(ASG)
=Consider the query
SELECT*
FROM EMP, ASG
WHERE EMP.ENO=ASG.ENO
Reduction for PHF
EMP
1
EMP
2
EMP
3
ASG
1
ASG
2
ENO
Dist ributed DBMS 86
B Reduction with join - Example
=Distribute join over unions
=Apply the reduction rule
Reduction for PHF
EMP
1
ASG
1
EMP
2
ASG
2
EMP
3
ASG
2
ENO ENO ENO
Page 44
Dist ributed DBMS 87
B Find useless (not empty) intermediate relations
Relation R defined over attributes A = {A
1
, ..., A
n
} vertically fragmented
as R
i
=
A'
(R) where A' A:
D,K
(R
i
) is useless if the set of projection attributes D is not in A'
Example: EMP
1
=
ENO,ENAME
(EMP); EMP
2
=
ENO,TITLE
(EMP)
SELECT ENAME
FROM EMP
Reduction for VF
ENAME
EMP
1
EMP
1
EMP
2
ENAME
ENO
Dist ributed DBMS 88
Step 3 Global Query
Optimization
Input: Fragment query
B Find the best (not necessarily optimal) global
schedule
=Minimize a cost function
=Distributed join processing
+ Bushy vs. linear trees
+ Which relation to ship where?
+ Ship-whole vs ship-as-needed
=Decide on the use of semijoins
+ Semijoin saves on communication at the expense of
more local processing.
=Join methods
+ nested loop vs ordered joins (merge join or hash join)
Page 45
Dist ributed DBMS 89
Cost-Based Optimization
B Solution space
=The set of equivalent algebra expressions (query trees).
B Cost function (in terms of time)
=I/O cost + CPU cost + communication cost
=These might have different weights in different distributed
environments (LAN vs WAN).
=Can also maximize throughput
B Search algorithm
=How do we move inside the solution space?
=Exhaustive search, heuristic algorithms (iterative
improvement, simulated annealing, genetic,)
Dist ributed DBMS 90
Query Optimization Process
Search Space
Generation
Search
Strategy
Equivalent QEP
Input Query
Transformation
Rules
Cost Model
Best QEP
Page 46
Dist ributed DBMS 91
Search Space
B Search space characterized by
alternative execution plans
B Focus on join trees
B For N relations, there are O(N!)
equivalent join trees that can be
obtained by applying
commutativity and associativity
rules
SELECT ENAME,RESP
FROM EMP, ASG, PROJ
WHERE EMP.ENO=ASG.ENO
AND ASG.PNO=PROJ.PNO
PROJ
ASG EMP
PROJ
ASG
EMP
PROJ
ASG
EMP
ENO
ENO
PNO
PNO
ENO,PNO
Dist ributed DBMS 92
Search Space
B Restrict by means of heuristics
=Perform unary operations before binary operations
=
B Restrict the shape of the join tree
=Consider only linear trees, ignore bushy ones
R
2
R
1
R
3
R
4
Linear Join Tree
R
2
R
1
R
4
R
3
Bushy Join Tree
Page 47
Dist ributed DBMS 93
Search Strategy
B How to move in the search space.
B Deterministic
=Start from base relations and build plans by adding one
relation at each step
=Dynamic programming: breadth-first
=Greedy: depth-first
B Randomized
=Search for optimalities around a particular starting point
=Trade optimization time for execution time
=Better when > 5-6 relations
=Simulated annealing
=Iterative improvement
Dist ributed DBMS 94
Search Strategies
B Deterministic
B Randomized
R
2
R
1
R
3
R
4
R
2
R
1
R
2
R
1
R
3
R
2
R
1
R
3
R
3
R
1
R
2
Page 48
Dist ributed DBMS 95
B Total Time (or Total Cost)
= Reduce each cost (in terms of time) component individually
= Do as little of each cost component as possible
= Optimizes the utilization of the resources
Increases system throughput
B Response Time
= Do as many things as possible in parallel
= May increase total time because of increased total activity
Cost Functions
Dist ributed DBMS 96
Summation of all cost factors
Total cost = CPU cost + I/O cost + communication cost
CPU cost = unit instruction cost no.of instructions
I/O cost = unit disk I/O cost no. of disk I/Os
communication cost = message initiation + transmission
Total Cost
Page 49
Dist ributed DBMS 97
B Wide area network
=message initiation and transmission costs high
=local processing cost is low (fast mainframes or
minicomputers)
=ratio of communication to I/O costs = 20:1
B Local area networks
=communication and local processing costs are more or less
equal
=ratio = 1:1.6
Total Cost Factors
Dist ributed DBMS 98
Elapsed time between the initiation and the completion of a
query
Response time = CPU time + I/O time + communication time
CPU time = unit instruction time no. of sequential instructions
I/O time = unit I/O time no. of sequential I/Os
communication time = unit msg initiation time no. of
sequential msg + unit transmission time no. of
sequential bytes
Response Time
Page 50
Dist ributed DBMS 99
Assume that only the communication cost is considered
Total time = 2 message initialization time + unit transmission
time (x+y)
Response time = max {time to send x from 1 to 3, time to send y
from 2 to 3}
time to send x from 1 to 3 = message initialization time + unit
transmission time x
time to send y from 2 to 3 = message initialization time + unit
transmission time y
Example
Site 1
Site 2
x units
y units
Site 3
Dist ributed DBMS 100
B Primary cost factor: size of intermediate relations
B Make them precise more costly to maintain
= For each relation R[A
1
, A
2
, , A
n
] fragmented as R
1
, , R
r
+ length of each attribute: length(Ai)
+ the number of distinct values for each attribute in each fragment:
card(
A
i
R
j
)
+ maximum and minimum values in the domain of each attribute:
min(A
i
), max(A
i
)
+ the cardinalities of each domain: card(dom[A
i
])
+ the cardinalities of each fragment: card(R
j
)
= Selectivity factor of each operation for relations
+ For joins
Optimization Statistics
SF (R,S) =
card(R S)
card(R) card(S)
Page 51
Dist ributed DBMS 101
Selection
size(R) = card(R) length(R)
card(
F
(R)) = SF
(F) card(R)
where
Intermediate Relation
Sizes
S F
(A = value) =
card(
A
(R))
1
S F
(A > value) =
max(A) min(A)
max(A) value
S F
(A < value) =
max(A) min(A)
value max(A)
SF
(p(A
i
) p(A
j
)) = SF
(p(A
i
)) SF
(p(A
j
))
SF
(p(A
i
) p(A
j
)) = SF
(p(A
i
)) + SF
(p(A
j
)) (SF
(p(A
i
)) SF
(p(A
j
)))
SF
(A value) = SF