Query Processing and Query Optmization

Post on 23-Jan-2016

118 views 0 download

description

CS157B Lecture 19. Query Processing and Query Optmization. Prof. Sin-Min Lee Department of Computer Science San Jose State University. Transaction Management. Until now, our concept of database has been one in which programs accessing the database are run one - PowerPoint PPT Presentation

Transcript of Query Processing and Query Optmization

Query Processing and Query Optmization

Prof. Sin-Min Lee

Department of Computer Science

San Jose State University

Transaction ManagementUntil now, our concept of database has been one in

which programs accessing the database are run one

at a time (serially). Often this is indeed the case.

However, there are also numerous applications in

which more than one program, or different

executions of the same program, run simultaneously

(concurrently). This is where TRANSACTION

MANAGEMENT comes in handy

Example #1An airline reservation system, where at one time,

several agents may be selling tickets, and therefore,

changing lists of passengers and counts of available

seats. The problem here is that if we are not careful

when we allow two or more processes to access the

database, we could sell the same seat twice. Here 2

processes that read and change the value of the same

object must not be allowed to run concurrently,

because they might interact in undesirable ways.

SAFETY

Example #2

Statistical database, such as census data, where

many people may be querying the database at

once. Here, as long as no one is changing the

data, we do not really care in what order the

processes read data. We can let the operating

system schedule simultaneous read requests.

Here we want to allow maximum concurrent

operations, so time can be saved

SPEED

Important Terms

Serializable schedule -- is a linear

arrangement of the database calls from several transactions with the property: the final database state obtained by executing the calls in schedule order is the same as the obtained by running the transactions in some unspecified serial order

Important Terms (cont…)

Lock -- is an access privilege on a database

object, which the DBMS grants to a particular transaction. To allow the privileged transaction to complete its work without undue interference, the lock restricts the access of competing transactions to the database object.

Important Terms (cont…)

Two-phase locking protocol -- is a discipline

that transactions can use to ensure serializability. The first phase is characterized by the monotonic acquisition of new locks or the strengthening of existing locks. The second phase involves the monotonic downgrade or release of existing locks. This lock is the most important technique for

managing concurrency.

Important Terms (cont…)

Strict two-phase locking protocol -- each transaction holds all its locks until it commits or rolls back. When the commit or rollback is secure, it releases all the locks. In other words, the shrinking phase occurs all at once, immediately after the transaction terminates.

Important Terms (cont…)Timestamping -- is another method for

securing concurrency in the face of conflicting transactions.

Timestamp -- is a time of origin of each

transaction given by transaction manager portion of DBMS

Items -- units of data to which access is

controlled

Important Terms (cont…)

Transaction -- is simply a single execution of a program. This program may be a simple query expressed in one of the query languages or an elaborate host language program with embedded calls to a query language. Several independent executions of the same program may be in progress simultaneously; each is a different transaction

Main Idea

To a large extent, transaction management can be seen as an attempt to make complex operations appear atomic. That is, they either occur in their entirety or do not occur at all, and if they occur, nothing else apparently went on during the time of their occurrence. The normal approach to ensuring atomicity of transactions is ‘serialization’, which forces transactions to run concurrently in a way that makes it appear that they ran one-at-a-time (serially).

Main Idea

In reality, transaction are sequences of more

elementary steps, such as reading or writing of

single items from the database, and

performing simple arithmetic steps in the

workspace. When concurrency control is

provided, other primitive steps are also

needed, steps which set and release locks,

commit (complete) transactions, and others

Items

To manage concurrency, the database must be partitioned into items, which are the units of data to which access is controlled. The nature and size of items are for the system designer to choose. In the relational model of data, for example, we could choose large items, like relations, or small items like individual tuples or even components of tuples.

Locks

The most common way in which access to items is controlled is by “locks.” Lock manager is the part of a DBMS that records, for each item I, whether one or more transactions are reading or writing any part of I. If so, the manager will forbid another transaction from gaining access to I, provided the type of access (read or write) could cause a conflict, such as the duplicate selling of an airline seat.

Locks

As it is typical for only a small subset of the

items to have locks on them at any one time,

the lock manager can store the current locks in

a lock table which consists of records

(<item>,<lock type>,<transaction> The

meaning of record (I,L,T) is that transaction T

has a lock of type L on item I.

Example of locks

Lets consider two transaction T1 and T2. Each accesses an item A, which we assume has an integer value, and adds one to A.

Read A;A:=A+1;Write A;-----------------------------------------------------------T1: Read A A:=A+1 Write AT2: Read A A:=A+1 Write A-----------------------------------------------------------

Example of locks (cont…)

The most common solution to this problem is to provide a lock on A. Before reading A, a transaction T must lock A, which prevents another transaction from accessing A until T is finished with A. Furthermore, the need for T to set a lock on A prevents T from accessing A if some other transaction is already using A. T must wait until the other transaction unlocks A, which it should do only after finishing with A.

Introduction

Goals:

• get the right answer to a query

• get it fast

Possible query representations (after parsing)

• relational algebra (not usually)

• query graph (usually)

Strategies for making relational algebra queries “better”

• Push down selects – eg.: r1(r1no, a, b, c, d) r2(r2no, x, y, z, r1no)

σ r1.a = 7 (r1 JOIN r2) (σ r1.a = 7 r1) JOIN r2

• push down projects– Πr1.a, r1.b(r1 JOIN r2) Πr1.a, r1.b(

(Πr1no,a,b (r1)) JOIN r2)

Strategies for making relational algebra queries “better”(cntd)

• Eliminate products– σr1.r1no=r2.r1no (r1 r2) r1 JOIN r1.r1no=r2.r1no r2

• replace– σp1 and p2 (e1) by σp1 (σp2 (e1))

Strategies for making relational algebra queries “better”(cntd)

• maybe there is an index to support p2 solution

eg,:

σ (sal > 10000) and (name = “elmo”) emp

do name = “elmo” first with index

test results for sal > 10000

Strategies for making relational algebra queries “better”(cntd)

• Join relations in increasing order of size to keep intermediate results small ( saves time)

| e1| = 1

| e2| = 200

| e3| = 100000

eg.: e1 JOIN (e2 JOIN e3) (e1 JOIN e2) JOIN e3

Query Processing Cost Estimation

• Catalog Statistics– r = a relation– nr = |r| = # tuples in r– sr = size of record r in bytes– V(A,r) = # unique attribute A values in r

Given these statistics,• |r s| = nrns

• size of a tuple of r s = sr + ss

Query Processing Cost Estimation

If ‘a’ is a constant, then

|σ r.A = a| = 1 * nr

V(A,r)

Note: histograms can be better for selectivity estimation than V(A,r) statistic.

Selection Predicate Selectivity Estimates

sel(r.A=a) 1 / V(A,r)

sel(r.A<a) (a-min(A,r))/(max(A,r)-min(A,r))

sel(a1 <r.A< a2) (a2 - a1) / (max(A,r) - min(A,r))

sel(p(R.a)) 1/3 guess!!

Join Predicate Selectivity

• What is | r1 (R1) JOIN r2 (R2)| ?

– If R1 R2 is a key of r1, then at most 1 tuple of r1 joins to each r2 tuple. Hence,

| r1 JOIN r2| | r2 |

– if R1 R2 is a key of neither r1 nor r2, then,

• assume R1 R2 = {A}

• a tuple t r1 produces approximately n r2 / V(A, r2) tuples when joined to r2.

• Using the above, | r1 JOIN r2 | (n r1n r2 )/V(A, r2)

Join Predicate Selectivity(cntd)

• If we had used t r2 instead of r1 to compute the above estimate, we would have gotten

| r1 JOIN r2 | (n r1 n r2 )/V(A, r1)

If V(A,r1) V(A,r2), there are likely to be some dangling tuples that don’t participate in the join. Hence, the smaller of the two (join size) estimates is probably better.

Maintaining Statistics

• Doing it after every update is too much overhead

• Hence, most systems gather statistics periodically during slack time ( often this is initiated manually).

1-table selection subquery cost

• Without index– must use a sequential scan

– C I/O = cost of an I/O 11 msec

– C CPU = cost of CPU to process 1 tuple .1msec

cost(σ p (r)) = nr C CPU + pages ( r ) * C I/O

1-table selection subquery cost(cntd)

with clustered index• B+ tree on r.A• p= ( r.A = a)

cost (σ p (r)) = (1/ V(A,r)) * nr C CPU +

(height (r.A B+tree) +(1/V(A,r))*pages ( r )) C I/O

1-table selection subquery cost(cntd)

With unclustered index

• secondary B+ tree on r.A

• p = (r.A = a)

cost (σ p (r)) = (1/ V(A,r)) * nr C CPU +

(height (r.A B+tree) +(1/V(A,r)) * nr) C I/O

Exercise

• What is the cost of solving

σ a1 < r.A < a2 (r)

using a clustered B+ tree on R?

Join StrategiesNested Loop Join ( Simple Iteration) NLJto solve r1 JOINp r2

for each t1 r1 do begin

for t2 r2 do begin

if p(t1 , t2) = true then

output(t1 , t2) as part of result endend

NOTE: outer scan is over r1

inner scan is over r2

(this could have been reversed)

QUIZ• Should outer scan be over biggest or smallest table?• Hint: Consider buffer pool size.

• Variation of nested loop join: block oriented iteration

• Scan inner table once per outer table block instead of outer table tuple.

Sort Merge Join

To do r1 JOIN r2

r1.a = r2 . b

• sort r1 on a

• sort r2 on b

• merge– scan r1 sequentially

– iterate over only needed portion of r2 for each r1 tuple

Sort Merge Join(cntd)

r1 x a r2 b y

a 1 1 p

b 1 2 q

c 2 2 r

d 2 3 s

e 3

cost = sort cost + constant * | r1 JOIN r2|

Use of an Index

• In nested loop join, if there is an index on JOIN attribute of inner table, you can use it.

• In sort-merge join, if there is a B+tree index (preferably clustered) on the join attribute(s) you can avoid sorting that table (read it in sorted order).

Conclusion

• Query processing• Query processing operations

– sequential scan

– index scan

– join• nested loop

• sort merge

• Query processing operation costs• Next time: Hash join