1. Trang chủ
  2. » Công Nghệ Thông Tin

Tài liệu Advances in Database Technology- P15 docx

50 396 0

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Tiêu đề Efficient Similarity Search for Hierarchical Data in Large Databases
Tác giả K. Kailing, et al.
Thể loại tài liệu
Định dạng
Số trang 50
Dung lượng 1,01 MB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

For any two trees and the of the leaf distance histograms is a lower bound of the degree-2 edit distance of and Proof.. Theorem 1 and 2 also allow us to use leaf distance histograms as a

Trang 1

682 K Kailing et al.

Fig 5. Folding techniques for histograms: The technique of Papadopoulos and Manolopoulos

(top) and the modulo folding technique (bottom).

Theorem 2 For any two trees and the of the leaf distance histograms

is a lower bound of the degree-2 edit distance of and

Proof. Analogously to the proof of theorem 1

Theorem 1 and 2 also allow us to use leaf distance histograms as a filter for the

weighted edit and weighted degree-2 edit distance This statement is justified by the

fol-lowing considerations As shown above, the of two leaf distance histograms

gives a lower bound for the insert and delete operations that are necessary to transform

the two corresponding trees into each other This fact also holds for weighted relabeling

operations, as weights do not have any influence on the necessary structural

modifica-tions But even when insert/delete operations are weighted, our filter can be used as long

as their exists a smallest possible weight for an insert or delete operation In this

case, the term is a lower bound for the weighted edit and

degree-2 edit distance between the trees and Since we assume metric properties as

well as the symmetry of insertions and deletions for the distance, the triangle

inequal-ity guarantees the existence of such a minimum weight Otherwise, any relabeling of a

node would be performed cheaper by a deletion and a corresponding insertion operation

Moreover, structural differences of objects would be reflected only weakly if structural

changes are not weighted properly

Histogram folding. Another property of leaf distance histograms is that their size

is unbounded as long as the height of the trees in the database is also unbounded

This problem arises for several feature vector types, including the degree histograms

presented in section 4.2 Papadopoulos and Manolopoulos [10] address this problem by

folding the histograms into vectors with fixed dimension This is done in a piecewise

grouping process For example, when a 5-dimensional feature vector is desired, the

first one fifth of the histogram bins is summed up and the result is used as the first

component of the feature vector This is done analogously for the rest of the histogram

bins The above approach could also be used for leaf distance histograms, but it has

the disadvantage that the maximal height of all trees in the database has to be known

in advance For dynamic data sets, this precondition cannot be fulfilled Therefore, we

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 2

Efficient Similarity Search for Hierarchical Data in Large Databases 683

propose a different technique that yields fixed-size n-dimensional histograms by adding

up the values of certain entries in the leaf distance histogram Instead of summing up

adjacent bins in the histogram, we add up those with the same index modulo n, as

depicted in figure 5 This way, histograms of distinct length can be compared, and there

is no bound for the length of the original histograms

Definition 7 (folded histogram) A folded histogram of a histogram for a

given parameter is a vector of size where the value of any bin is

the sum of all bins in with mod i.e.

The following theorem justifies to use folded histograms in a multi-step query

pro-cessing architecture

Theorem 3 For any two histograms and and any parameter the

of the folded histograms of and is a lower bound for the

of and

Proof. Let be the length of and If necessary, and

are extended with bins containing 0 until and Then the following

holds:

4.2 Filtering Based on Degree of Nodes

The degrees of the nodes are another structural property of trees which can be used

as a filter for the edit distances Again, a simple filter can be obtained by using the

Trang 3

684 K Kailing et al.

maximal degree of all nodes in a tree denoted by as a single feature

The difference between the maximal degrees of two trees is an obvious lower bound for

the edit distance as well as for the degree-2 edit distance As before, this single-valued

filter is very coarse and using a degree histogram clearly increases the selectivity

Definition 8 (degree histogram) The degree histogram of a tree is a vector of

length where the value of any bin is the number of

nodes that share the degree i.e.

Theorem 4 For any two trees and the of the degree histograms

divided by three is a lower bound of the edit distance of and

Proof. Given two arbitrary trees and let us consider an edit sequence

that transforms into We proceed by induction over the length of the

and of both are equal to zero For let us assume that the lower-bounding

property already holds for and i.e

When extending the sequence by to S, the right hand side of the

inequality is increased by The situation on the left hand side is as follows

The edit step may be a relabeling, an insert or a delete operation Obviously, for a

relabeling, the degree histogram does not change, i.e

The insertion of a single node affects the histogram and the of the

histograms in the following way:

1

2

3

The inserted node causes an increase in the bin of degree That may change

the by at most one

The degree of parent node may change In the worst case this affects two bins

The bin of former degree is decreased by one while the bin of its new degree

is increased by one Therefore, the may additionally be changed by at

most two

No other nodes are affected

From the above three points it follows that the of the two histograms

and changes by at most three Therefore, the following holds:

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 4

Efficient Similarity Search for Hierarchical Data in Large Databases 685

As the above considerations also hold for the degree-2 edit distance, theorem 4 holds

analogously for this similarity measure

4.3 Filtering Based on Node Labels

Apart from the structure of the trees, the content features, expressed through node labels,

have an impact on the similarity of attributed trees The node labels can be used to define

a filter function To be useful in our filter-refinement architecture, this filter method has

to deliver a lower bound for the edit cost when transforming two trees into each other

The difference between the distribution of the values within a tree and the distribution

of the values in another tree can be used to develop a lower-bounding filter To ensure

efficient evaluation of the filter, the distribution of those values has to be approximated

for the filter step

One way to approximate the distribution of values is to use histograms In this case,

an histogram is derived by dividing the range of the node label into bins

Then, each bin is assigned the number of nodes in the tree whose value is in the range

of the bin To estimate the edit distance or the degree-2 edit distance between two trees,

half of the of their corresponding label histograms is appropriate A single

insert or delete operation changes exactly one bin of such a label histogram, a single

relabeling operation can influence at most two histogram bins If a node is assigned to

a new bin after relabeling, the entry in the old bin is decreased by one and the entry in

the new bin is increased by one (cf figure 6) Otherwise, a relabeling does not change

the histogram This method also works for weighted variants of the edit distance and the

degree-2 edit distance as long as there is a minimal weight for a relabeling operation In

this case, the calculated filter value has to be multiplied by this minimal weight in order

to gain a lower-bounding filter

This histogram approach applies to discrete label distributions very well However,

for continuous label spaces, the use of a continuous weight function which may become

arbitrarily small, can be reasonable In this case, a discrete histogram approach can not

be used An example for such a weight function is the Euclidean distance in the color

space, assuming trees where the node labels are colors Here, the cost for changing a

color value is proportional to the Euclidean distance between the original and the target

color As this distance can be infinitely small, it is impossible to estimate the relabeling

cost based on a label histogram as in the above cases

Fig 6. A single relabeling operation may result in a label histogram distance of two.

Trang 5

686 K Kailing et al.

Fig 7. Filtering for continuous weight functions.

More formally, when using the term ‘continuous weight function’ we mean that the

cost for changing a node label from value to value is proportional to

Let be the maximal possible difference between two attribute values Then

has to be normalized to [0,1] by dividing it through assuming thatthe maximal cost for a single insertion, deletion or relabeling is one To develop a filter

method for attributes with such a weight function, we exploit the following property of

the edit distance measure The cost-minimal edit sequence between two trees removes

the difference between the distributions of attribute values of those two trees It does not

matter whether this is achieved through relabelings, insertions or deletions

For our filter function we define the following feature value for a tree

Here is the attribute value of the node in and is the size of tree The absolute

difference between two such feature values is an obvious lower bound for the difference

between the distribution of attribute values of the corresponding trees Consequently,

we use

as a filter function for continuous label spaces, see figure 7 for an illustration Once

more, the above considerations also hold for the degree-2 edit distance

To simplify the presentation we assumed that a node label consists of just one single

attribute But usually a node will carry several different attributes If possible, the attribute

with the highest selectivity can be chosen for filtering In practice, there is often no such

single attribute In this case, filters for different attributes can be combined with the

technique described in the following section

4.4 Combining Filter Methods

All of the above filters use a single feature of an attributed tree to approximate the edit

distance or degree-2 edit distance As the filters are not equally selective in each situation,

we propose a method to combine several of the presented filters

A very flexible way of combining different filters is to follow the inverted list

ap-proach, i.e to apply the different filters independently from each other and then intersect

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 6

Efficient Similarity Search for Hierarchical Data in Large Databases 687

the resulting candidate sets With this approach, separate index structures for the

differ-ent filters have to be maintained and for each query, a time-consuming intersection step

is necessary To avoid those disadvantages, we concatenate the different filter histograms

and filter values for each object and use a combined distance function as a similarity

function

Definition 9 (Combined distance function) Let be a set of distance functions

for trees Then, the combined distance function is defined to be the maximum of the

component functions:

Theorem 5 For every set of lower-bounding distance functions

i.e for all trees and the combined distance function is

a lower bound of the edit distance function

Proof. For all trees and the following equivalences hold:

The final inequality represents the precondition

Justified by theorem 5, we apply each separate filter function to its corresponding

component of the combined histogram The combined distance function is derived from

the results of this step

For our tests, we implemented a filter and refinement architecture according to the

optimal multi-step k-nearest-neighbor search approach as proposed in [16] Naturally,

the positive effects which we show in the following experiments for k-nn-queries also

hold for range queries and for all data mining algorithms based on range queries or

k-nn-queries (e.g clustering, k-nn-classification) As similarity measure for trees, we

implemented the degree-2 edit distance algorithm as presented in [14] The filter

his-tograms were organized in an X-tree [17] All algorithms were implemented in Java 1.4

and the experiments were run on a workstation with a Xeon 1,7 GHz processor and 2

GB main memory under Linux

To show the efficiency of our approach, we chose two different applications, an

image database and a database of websites which are described in the following

Trang 7

688 K Kailing et al.

Fig 8. Structural and content-based information of a picture represented as a tree.

5.1 Image Databases

As one example of tree structured objects we chose images, because for images, both,

content-based as well as structural information are important Figure 8 gives an idea of

the two aspects which are present in a picture

The images we used for our experiments were taken from three real-world databases:

a set of 705 black and white pictographs, a set of 8,536 commercially available color

images and a set of 43,000 color TV-Images We extracted trees from those images in

a two-step process First, the images were divided into segments of similar color by a

segmentation algorithm In the second step, a tree was created from those segments by

iteratively applying a region-growing algorithm which merges neighboring segments if

their colors are similar This is done until all segments are merged into a single node As

a result, we obtain a set of labeled unordered trees where each node label describes the

color, size and horizontal as well as vertical extension of the associated segment Table

1 shows some statistical information about the trees we generated

For the first experiments, we used label histograms as described in section 4.3 To

derive a discrete label distribution, we reduced the number of different attribute values

to 16 different color values for each color channel and 4 different values each for size

and extensions We used a relabeling function with a minimal weight of 0.5 Later on

we also show some experiments where we did not reduce the different attribute values

and used a continuous weight function for relabeling

Comparison of our filter types. For our first experiment we used 10,000 TV-images We

created 10-dimensional height and degree histograms and combined them as described in

section 4.4 We also built a 24-dimensional combined label histogram which considered

the color, size and extensions of all node labels (6 attributes with histograms of size 4)

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 8

Efficient Similarity Search for Hierarchical Data in Large Databases 689

Fig 9. Runtime and number of candidates for k-nn-queries on 10,000 color TV-images.

Finally, the combination of this combined label histogram and a 4-dimensional height

histogram was taken as another filter criterion Let us note, that the creation of the filter

X-trees took between 25 sec for the height histogram and 62 sec for the combined

height-label histogram

We ran 70 k-nearest-neighbor queries (k = 1, 10, 100) for each of our filters Figure

9 shows the selectivity of our filters, measured in the average number of candidates

with respect to the size of the data set The figures show that filtering based solely on

structural (height or degree histogram) or content-based features (label histogram) is not

as effective as their combination Figure 9 also shows that for this data the degree filter

is less selective than the height filter The method which combines the filtering based on

the height of the nodes and on content features is most effective Figure 5.1 additionally

depicts the average runtime of our filters compared to the sequential scan As one can see,

we reduced the runtime by a factor of up to 5 Furthermore, the comparison of the two

diagrams in figure 9 shows that the runtime is dominated by the number of candidates,

whereas the additional overhead due to the filtering is negligible

Influence of histogram size. In a next step we tested to what extent the size of the

histogram influences the size of the candidate set and the corresponding runtime The

results for nearest neighbor queries on 10,000 color TV-images are shown in figure 10

With increasing dimension, the number of candidates as well as the runtime decrease The

comparison of the two diagrams in figure 10 shows that the runtime is again dominated

Fig 10. Influence of dimensionality of histograms.

Trang 9

690 K Kailing et al.

Fig 11. Scalability versus size of data set.

by the number of candidates, while the additional overhead due to higher dimensional

histograms is negligible

Scalability of filters versus size of data set. For this experiment we united all three

image data sets and chose three subsets of size 10,000, 25,000 and 50,000 On these

subsets we performed several representative 5-nn queries Figure 11 shows that the

selectivity of our structural filters does not depend on the size of the data set

Comparison of different filters for a continuous weight function. As mentioned

above, we also tested our filters when using a continuous weight function for relabeling

For this experiment, we used the same 10,000 color images as in 5.1 Figure 12 shows

the results averaged over 200 k-nn queries In this case, both the height histogram and

the label filter are very selective Unfortunately, the combination of both does not further

enhance the runtime While there is a slight decrease in the number of candidates, this

is used up by the additional overhead of evaluating two different filter criteria

Comparison with a metric tree. In [18] other efficient access methods for similarity

search in metric spaces are presented In order to support dynamic datasets, we use

the X-tree that can be updated at any time Therefore, we chose to compare our filter

methods to the M-tree which analogously is a dynamic index structure for metric spaces

Fig 12. Runtime and number of candidates when using a continuous weight function.

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 10

Efficient Similarity Search for Hierarchical Data in Large Databases 691

Fig 13.Runtime and number of distance computations of filter methods compared to the M-Tree.

We implemented the M-tree as described in [ 19] by using the best split policy mentioned

there

The creation of an M-tree for 1,000 tree objects already took more than one day, due

to the split policy that has quadratic time-complexity The time for the creation of the

filter vectors, on the other hand, was in the range of a few seconds As can be seen in

figure 13, the M-tree outperformed the sequential scan for small result sizes However,

all of our filtering techniques significantly outperform the sequential scan and the M-tree

index for all result set sizes This observation is mainly due to the fact that the filtering

techniques reduce the number of necessary distance calculations far more than the

M-tree index This behavior results in speed-up factors between 2.5 and 6.2 compared to

the M-tree index and even higher factors compared to a simple sequential scan This

way, our multi-step query processing architecture is a significant improvement over the

standard indexing approach

5.2 Web Site Graphs

As demonstrated in [20], the degree-2 edit distance is well suitable for approximate

website matching In website management it can be used for searching similar websites

In [21] web site mining is described as a new way to spot competitors, customers and

suppliers in the world wide web

By choosing the main page as the root, one can represent a website as a rooted,

labeled, unordered tree Each node in the tree represents a webpage of the site and is

labeled with the URL of that page All referenced pages are children of that node and

the borders of the website where chosen carefully See figure 14 for an illustration

Fig 14. Part of a website tree.

Trang 11

692 K Kailing et al.

Fig 15. Average runtime and number of candidates for 5-nn queries.

For our experiment, we used a compressed form of the 207 web sites described in

[21], resulting in trees that have 67 nodes on the average We ran 5-nn-queries on this

data The results are shown in figure 15 We notice that even if the degree filter produces

a lot more candidates than the height filter, it results in a better run time This is due to

the fact that it filters out those trees, where the computation of the degree-2 edit distance

is especially time-consuming Using the combination of both histograms, the runtime is

reduced by a factor of 4

In this paper, we presented a new approach for efficient similarity search in large

databases of tree structures Based on the degree-2 edit distance as similarity measure,

we developed a multi-step query architecture for similarity search in tree structures For

structural as well as for content-based features of unordered attributed trees, we

sug-gested several filter methods These filter methods significantly reduce the number of

complex edit distance calculations necessary for a similarity search The main idea

be-hind our filter methods is to approximate the distribution of structural and content-based

features within a tree by means of feature histograms Furthermore, we proposed a new

technique for folding histograms and a new way to combine different filter methods in

order to improve the filter selectivity We performed extensive experiments on two sets

of real data from the domains of image similarity and website mining Our experiments

showed that filtering significantly accelerates the complex task of similarity search for

tree-structured objects Moreover, it turned out that no single feature of a tree is sufficient

for effective filtering, but only the combination of structural and content-based filters

yields good results

In our future work, we will explore how different weights for edit operations influence

the selectivity of our filter methods Additionally, we intend to investigate other structural

features of trees for their appropriateness in the filter step In a recent publication [5], an

edit distance for XML-documents has been proposed An interesting question is, how

our architecture and filters can be applied to the problem of similarity search in large

databases of XML-documents

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 12

Efficient Similarity Search for Hierarchical Data in Large Databases 693

Jiang, T., Wang, L., Zhang, K.: Alignment of trees - an alternative to tree edit Proc Int Conf.

on Combinatorial Pattern Matching (CPM), LNCS 807 (1994) 75–86

Selkow, S.: The tree-to-tree editing problem Information Processing Letters 6 (1977) 576–

584

Zhang, K.: A constrained editing distance between unordered labeled trees Algorithmica 15

(1996) 205–222

Wang, J.T.L., Zhang, K., Chang, G., Shasha, D.: Finding approximate pattersn in undirected

acyclic graphs Pattern Recognition 35 (2002) 473–483

Nierman, A., Jagadish, H.V.: Evaluating structural similarity in XML documents In: Proc.

5th Int Workshop on the Web and Databases (WebDB 2002), Madison, Wisconsin, USA.

(2002) 61–66

Sebastian, T.B., Klein, P.N., Kimia, B.B.: Recognition of shapes by editing shock graphs.

In: Proc 8th Int Conf on Computer Vision (ICCV’01), Vancouver, BC, Canada Volume 1.

(2001) 755–762

Bunke, H., Shearer, K.: A graph distance metric based on the maximal common subgraph.

Pattern Recognition Letters 19 (1998) 255–259

Chartrand, G., Kubicki, G., Schultz, M.: Graph similarity and distance in graphs Aequationes

Mathematicae 55 (1998) 129–145

Kubicka, E., Kubicki, G., Vakalis, I.: Using graph distance in object recognition In: Proc.

ACM Computer Science Conference (1990) 43–48

Papadopoulos, A., Manolopoulos,Y.: Structure-based similarity search with graph histograms.

In: Proc DEXA/IWOSS Int Workshop on Similarity Search (1999) 174–178

Levenshtein, V.: Binary codes capable of correcting deletions, insertions and reversals Soviet

Physics-Doklady 10 (1966) 707–710

Wagner, R.A., Fisher, M.J.: The string-to-string correction problem Journal of the ACM 21

(1974) 168–173

Zhang, K., Statman, R., Shasha, D.: On the editing distance between unordered labeled trees.

Information Processing Letters 42 (1992) 133–139

Zhang, K., Wang, J., Shasha, D.: On the editing distance between undirected acyclic graphs.

International Journal of Foundations of Computer Science 7 (1996) 43–57

Agrawal, R., Faloutsos, C., Swami, A.N.: Efficient similarity search in sequence databases.

In: Proc 4th Int Conf of Foundations of Data Organization and Algorithms (FODO) (1993)

69–84

Seidl, T., Kriegel, H.P.: Optimal multi-step k-nearest neighbor search In Haas, L.M., Tiwary,

A., eds.: Proc ACM SIGMOD Int Conf on Managment of Data, ACM Press (1998) 154–165

Berchtold, S., Keim, D., Kriegel, H.P.: The X-tree: An index structure for high-dimensional

data In: 22nd Conference on Very Large Databases, Bombay, India (1996) 28–39

Chavez, E., Navarro, G., Baeza-Yates, R., Marroquin, J.: Searching in metric spaces ACM

Computing Surveys 33 (2001) 273–321

Ciaccia, P., Patella, M., Zezula, P.: M-tree: An efficient access method for similarity search in

metric spaces In: VLDB’97, Proc 23rd Int Conf on Very Large Databases, August 25-29,

1997, Athens, Greece (1997) 426–435

Wang, J., Zhang, K., Jeong, K., Shasha, D.: A system for approximate tree matching IEEE

Transactions on Knowledge and Data Engineering 6 (1994) 559–571

Ester, M., Kriegel, H.P., Schubert, M.: Web site mining: A new way to spot competitors,

customers and suppliers in the world wide web In: Proc 8th Int Conf on Knowledge

Discovery in Databases (SIGKDD’02), Edmonton, Alberta, Canada (2002) 249–258

Trang 13

QuaSAQ: An Approach to Enabling End-to-End

QoS for Multimedia Databases

Yi-Cheng Tu, Sunil Prabhakar, Ahmed K Elmagarmid, and Radu Sion

Purdue University, West Lafayette IN 47907, USA

Abstract. The paper discusses the design and prototype tion of a QoS-aware multimedia database system Recent research in multimedia databases has devoted little attention to the aspect of the integration of QoS support at the user level Our proposed architec- ture to enable end-to-end QoS control, the QoS-Aware Query Processor (QuaSAQ), satisfies user specified quality requirements The users need not be aware of detailed low-level QoS parameters, but rather specifies high-level, qualitative attributes In addition to an overview of key re- search issues in the design of QoS-aware databases, this paper presents our proposed solutions, and system implementation details An impor- tant issue relates to the enumeration and evaluation of alternative plans for servicing QoS-enhanced queries This step follows the conventional query execution which results in the identification of objects of interest

implementa-to the user We propose a novel cost model for media delivery that

ex-plicitly takes the resource utilization of the plan and the current system contention level into account Experiments run on the QuaSAQ proto- type show significantly improved QoS and system throughput.

As compared to traditional applications, multimedia applications have special

requirements with respect to search and playback with satisfactory quality The

problem of searching multimedia data has received significant attention from

researchers with the resulting development of content-based retrieval for

multi-media databases The problem of efficient delivery and playback of such data

(especially video data), on the other hand, has not received the same level of

attention From the point of view of multimedia DBMS design, one has to be

concerned about not only the correctness but also the quality of the query results.

The set of quality parameters that describes the temporal/spatial constraints of

media-related applications is called Quality of Service (QoS) [1] Guaranteeing

QoS for the user requires an end-to-end solution – all the way from the retrieval

of data at the source to the playback of the data at the user

In spite of the fact that research in multimedia databases has covered many

key issues such as data models, system architectures, query languages, algorithms

for effective data organization and retrieval [2], little effort has been devoted to

the aspect of the integration of QoS support In the context of general

multime-dia system, research on QoS has concentrated on system and network support

E Bertino et al (Eds.): EDBT 2004, LNCS 2992, pp 694–711, 2004.

© Springer-Verlag Berlin Heidelberg 2004

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 14

QuaSAQ: An Approach to Enabling End-to-End QoS 695

with little concern for QoS control on the higher (user, application) levels

High-level QoS support is essential in any multimedia systems because the satisfaction

of human users is the primary concern in defining QoS [3] Simply deploying a

multimedia DBMS on top of a QoS-provisioning system will not provide

end-to-end QoS Moreover, such a solution is unable to exploit the application level

flexibility such as the user’s acceptable range of quality For example, for a

physi-cian diagnosing a patient, the jitter-free playback of very high frame rate and

resolution video of the patient’s test data is critical; whereas a nurse accessing

the same data for organization purposes may not require the same high quality

Such information is only available at the user or application levels

We envision users such as medical professionals accessing these databases via

a simple user interface In addition to specifying the multimedia items of interest

(directly or via content-based similarity to other items), the user specifies a set

of desired quality parameter bounds The quality bounds could be specified

explicitly or automatically generated based upon the user’s profile The user

should not need to be aware of detailed system QoS parameters but rather

specifies high-level qualitative attributes (e.g “high resolution”, or “CD quality

audio”) Thus a QoS-enabled database will search for multimedia objects that

satisfy the content component of the query and at the same time can be delivered

to the user with the desired level of quality

In this paper we discuss the design and prototype implementation of our

QoS-aware multimedia DBMS We describe the major challenges to enabling

end-to-end QoS, and present our proposed solutions to these problems To the

best of our knowledge, this is the first prototype system that achieves

end-to-end QoS for multimedia databases We present experimental results from our

prototype that establish the feasibility and advantages of such a system Our

implementation builds upon the VDBMS prototype multimedia database system

developed by our group at Purdue University [4] Among other enhancements,

QuaSAQ extends VDBMS to build a distributed QoS-aware multimedia DBMS

with multiple copies of storage/streaming manager

To address the structure of a QoS-provisioning networked multimedia

sys-tem, four levels of QoS have been proposed: user QoS, application QoS, system

QoS, and network QoS [1,5] We consider a series of QoS parameters in our

research as shown in Table 1 QoS guarantees for individual requests and the

overall system performance are in most cases two conflicting goals since the

en-tire QoS problem is caused by scarcity of resources Most current research on

QoS fail to address the optimization of system performance In this paper, we

highlight the key elements of our proposed approach to supporting end-to-end

QoS and achieving high performance in a multimedia database environment

The approach is motivated by query processing and optimization techniques in

conventional distributed databases

The key idea of our approach is to augment the query evaluation and

opti-mization modules of a distributed database management system (D-DBMS) to

directly take QoS into account To incorporate QoS control into the database,

user-level QoS parameters are translated into application QoS and become an

Trang 15

696 Y.-C Tu et al.

augmented component of the query For each raw media object, a number

of copies with different application QoS parameters are generated offline by

transcoding and these copies are replicated on the distributed servers Based on

the information of data replication and runtime QoS adaptation options (e.g

frame dropping), the query processor generates various plans for each query

and evaluates them according to a predefined cost model The query

evalua-tion/optimization module also takes care of resource reservation to satisfy

low-level QoS For this part, we propose the design of a unified API and

implemen-tation module that enables negotiation and control of the underlying system

and network QoS APIs, thereby providing a single entry-point to a multitude of

QoS layers (system and network) The major contributions of this paper are: 1)

We propose a query processing architecture for multimedia databases for

han-dling queries enhanced with QoS parameters; 2) We propose a cost model that

evaluates QoS-aware queries by their resource utilization with consideration of

current system status; 3) We implement the proposed query processor within a

multimedia DBMS and evaluate our design via experiments run on this

proto-type

The paper is organized as follows: Section 2 deals with the main issues

en-countered in the process of designing and implementing the system Section 3

presents the actual architecture of the Quality of Service Aware Query

Proces-sor (QuaSAQ) We also discuss details pertaining to the design of individual

components in the architecture The prototype implementation of QuaSAQ is

detailed in Section 4 Section 5 presents the evaluation of the proposed QuaSAQ

architecture In Section 6, we compare our work with relevant research efforts

Section 7 concludes the paper

Building a distributed multimedia DBMS requires a careful design of many

com-plex modules as well as effective interactions between these components This

becomes further complicated if the system is to support non-trivial aspects such

as QoS In order to extend the D-DBMS approach to address end-to-end QoS,

several important requirements have to be met These include:

1 Smart QoS-aware data replication algorithms have to be developed

Indi-vidual multimedia objects need to be replicated on various nodes of the

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 16

QuaSAQ: An Approach to Enabling End-to-End QoS 697

2

3

4

5.

database Each replica may satisfy different application QoS in order to

closely meet the requirements of user inputs The total number and choice

of QoS of pre-stored media replicas should reflect the access pattern of media

content Therefore, dynamic online replication and migration has to be

per-formed to make the system converge to the current status of user requests

Another concern in replication is the storage space

Mapping of QoS parameters between different layers has to be achieved First

of all, user-level qualitative QoS inputs (e.g DVD-quality video) need to be

translated into application QoS (e.g spatial resolution) since the underlying

query processor only understands the latter One critical point here is that

the mapping from user QoS to application QoS highly depends on the user’s

personal preference Resource consumption of query plans is essential for

cost estimation and query optimization in QoS-aware multimedia databases

This requires mapping application QoS in our QoS-enhanced queries to QoS

parameters on the system and network level

A model for the search space of possible execution plans The search space

is of a very different structure from that of a traditional D-DBMS In the

latter, the primary data model for search space comprises a query tree The

query optimizer then explores the space using strategies such as dynamic

programming and randomized search to find the “best” plan according to a

cost model [6] In our system, various components such as encryption,

encod-ing, and filtering must be individually considered in addition to the choice of

database server and physical media object Depending on the system status,

any of the above components can be the dominant factor in terms of cost

A cost estimation model is needed to evaluate the generated QoS-aware

plans Unlike the static cost estimates in traditional D-DBMS, it is critical

that the costs under current system status (e.g based upon current load

on a link) be factored into the choice of an acceptable plan Furthermore,

the cost model in our query processor should also consider optimization

criteria other than the total time1, which is normally the only metric used in

D-DBMS A very important optimization goal in multimedia applications is

system throughput Resource consumption of each query has to be estimated

and controlled for the system to achieve maximum throughput and yet QoS

constraints of individual requests are not violated

Once an acceptable quality plan has been chosen, the playback of the media

objects in accordance with the required quality has to be achieved

Gener-ally, QoS control in multimedia systems are achieved in two ways: resource

reservation and adaptation [1] Both strategies require deployment of a

QoS-aware resource management module, which is featured with admission

con-trol and reservation mechanisms There may also be need for renegotiation

(adaptation) of the QoS constraints due to user actions during playback

Our research addresses all above challenges In the next section, we present a

framework for QoS provisioning in a distributed multimedia database

environ-ment with the focus on our solutions to items 3 and 4 listed above For items

1

Sometimes response time is also used, as in distributed INGRES.

Trang 17

698 Y.-C Tu et al.

Fig 1. QuaSAQ architecture

2 and 5, we concentrate on the implementation and evaluation of known

ap-proaches within the context of multimedia databases Item1 will be covered in a

follow-up paper

Figure 1 describes in detail the proposed architecture of our QoS-aware

dis-tributed multimedia DBMS, which we call Quality-of-Service Aware Query

Pro-cessor (QuaSAQ) In this section, we present detailed descriptions of the various

components of QuaSAQ

3.1 Offline Components

The offline components of QuaSAQ provide a basis for the database

adminis-trators to accomplish QoS-related database maintenance Two major activities,

offline replication and QoS sampling, are performed for each media object

in-serted into the database As a result of those, relevant information such as the

quality, location and resource consumption pattern of each replica of the

newly-inserted object is fed into the Distributed Metadata Engine as metadata Please

refer to [7] for more details of replication and QoS mapping in QuaSAQ

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 18

QuaSAQ: An Approach to Enabling End-to-End QoS 699

3.2 QoP Browser

The QoP Browser is the user interface to the underlying storage, processing

and retrieval system It enables certain QoP parameter control, generation of

QoS-aware queries, and execution of the resulting presentation plans The main

entities of the QoP Browser include: The User Profile contains high-level QoP

parameter mappings to lower level QoS parameter settings as well as various user

related statistics acquired over time, enabling better renegotiation decisions in

case of resource failure The Query Producer takes as input some user actions

(requests with QoP inputs) and the current settings from the user profile and

generates a query As compared to those of traditional DBMS, the queries

gener-ated in QuaSAQ are enhanced with QoS requirements We call them QoS-aware

queries The Plan Executor is in charge of actually running the chosen plan.

It basically performs actual presentation, synchronization as well as runtime

maintenance of underlying QoS parameters

Quality of Presentation. From a user’s perspective, QoS translates into the

more qualitative notion of Quality of Presentation (QoP) The user is not

ex-pected to understand low level quality parameters such as frame rates or packet

loss rate Instead, the user specifies high-level qualitative parameters to the best

of his/her understanding of QoS Some key QoP parameters that are often

con-sidered in multimedia systems include: spatial resolution, temporal resolution or

period, color depth, reliability, and audio quality Before being integrated into a

database query, the QoP inputs are translated into application QoS based on the

information stored in the User Profile For example, a user input of “VCD-like

spatial resolution” can be interpreted as a resolution range of 320×240 – 352×288

pixels The application QoS parameters are quantitative and we achieve some

flexibility by allowing one QoP mapped to a range of QoS values QoS

require-ments are allowed to be modified during media playback and a renegotiation is

expected Another scenario for renegotiation is when the user-specified QoP is

rejected by the admission control module due to low resource availability Under

such circumstances, a number of admittable alternative plans will be presented

as a “second chance” for the query to be serviced

One important weakness of these qualitative formulations of QoP is their

lack of flexibility (i.e failure to capture differences between users) For example,

when renegotiation has to be performed, one user may prefer reduction in the

temporal resolution while another user may prefer a reduction in the spatial

resolution We remedy this by introducing a per-user weighting of the quality

parameters as part of the User Profile

3.3 Distributed Metadata Engine

In a multimedia DBMS, operations such as content-based searching depend

heav-ily, if not exclusively, on the metadata of the media objects [2] As mentioned

in Section 2, video objects are stored in several locations, each copy with

dif-ferent representation characteristics This requires more items in the metadata

collection Specifically, we require at least the following types of metadata for a

QoS-aware DBMS:

Trang 19

700 Y.-C Tu et al.

Content Metadata: describe the content of objects to enable multimedia

query, search, and retrieval In our system, a number of visual and

seman-tic descriptors such as shot detection, frame extraction, segmentation, and

camera motion are extracted

Quality Metadata: describe the quality characteristics (in the form of

appli-cation level QoS) of physical media objects For our QoS-aware DBMS, the

following parameters are kept as metadata for each video object: resolution,

color depth, frame rate, and file format

Distribution Metadata: describe the physical locations (i.e paths, servers,

proxies, etc.) of the media objects It records the OIDs of objects and the

mapping between media content and media file

QoS profile: describe the resource consumption in the delivery of individual

media objects The data in QoS profiles is obtained via static QoS

map-ping performed by the QoS sampler The QoS profiles are the basis for cost

estimation of QoS-aware query execution plans

We distribute the metadata in various locations enabling ease of use and

migra-tion Caching is used to accelerate non-local metadata accesses

3.4 Quality Manager

The Quality Manager is the focal point of the entire system It is heavily

inte-grated with the Composite QoS in order to enable reservation and

negoti-ation It has the following main components:

Plan Generator. The Plan Generator is in charge of generating plans that

enable the execution of the query from the Query Producer The Content

Meta-data is used to identify logical objects that satisfy the content component of

the query (e.g videos with images of George Bush or Sunsets) A given logical

object may be replicated at multiple sites and further with different formats For

example, a given video may be stored in different resolutions and color depth at

two different sites The plan generator determines which of the alternatives can

be used to satisfy the request and also the necessary steps needed to present it

to the user

The final execution of QoS-aware query plans can be viewed as a series of

server activities that may include retrieval, decoding, transcoding between

dif-ferent formats and/or qualities, and encryption Therefore, the search space of

alternative QoS-aware plans consists of all possible combinations of media

repos-itories, target objects, and server activities mentioned above We can model the

search space as a universe of disjoint sets Each set represents a target media

object or a server activity whose possible choices serve as elements in the set

Suppose we have such sets then an execution plan is an ordered

set satisfying the following conditions:

Trang 20

QuaSAQ: An Approach to Enabling End-to-End QoS 701The semantics of the above conditions are: (1) The total number of compo-

nents in a plan cannot exceed the number of possible server activities; (2) All

components in a plan come from some disjoint set; and (3) No two components

in a plan come from the same set The size of the search space is huge even

with the above restrictions Suppose each set of server activity has elements,

the number of possible plans is Fortunately, there are also some other

system-specific rules that further reduce the number of alternative plans One

salient rule is related to the order of server activities For example, the first

server activity should always be the retrieval of a media object from a certain

site, all other activities such as transcoding, encryption have to follow the the

media retrieval in a plan If the order of all server activity sets are fixed, the size

of search space decreases to

Fig 2 Illustrative plan generation in QuaSAQ

Runtime QoS Evaluation and Plan Drop. The Plan Generator described

above does not check generated plans for any QoS constraints We can perform

those verifications by applying a set of static and dynamic rules First of all,

decisions can be made instantly based on QoS inputs in the query For example,

we cannot retrieve a video with resolution lower than that required by the user

Similarly, it makes no sense to transcode from low resolution to high resolution

Therefore, QoS constraints help further reduce the size of search space by

de-creasing the appropriate set size In practice, can be regarded as a constant

Some of the plans can be immediately dropped by the Plan Generator if their

costs are intolerably high This requires QuaSAQ to be aware of some obvious

performance pitfalls For example, encryption should always follow the frame

dropping since it is a waste of CPU cycles to encrypt the data in frames that

will be dropped Once a suitable plan has been discovered, the Plan Generator

computes its resource requirements (in the form of a resource vector) and feeds

it to the next component down the processing pipe-line

Trang 21

702 Y.-C Tu et al.

Illustrative examples of plans. The path in solid lines shown in Figure 2

rep-resents a query plan with the following details: 1 retrieve physical copy number

1 of the requested media from the disk of server B; 2 transfer the media to

server A; 3 transcode to MPEG1 format with certain target QoS; 4 drop all

the B frames during delivery; 5 encrypt the media data using algorithm 1 The

dotted line corresponds to a simpler plan: retrieve the same object and transcode

with the same target QoS, no frame dropping or encryption is needed An even

simpler plan would be a single node in set A1, meaning the object is sent without

further processing

Runtime Cost Evaluator. The Runtime Cost Evaluator is the main

compo-nent that computes (at runtime) estimated costs for generated plans It sorts

the plans in ascending cost order and passes them to the Plan Executor in the

QoP Browser The first plan in this order that satisfies the QoS requirements is

used to service the query In a traditional D-DBMS, the cost of a query is

gener-ally expressed as the sum of time spent on CPU, I/O and data transferring In

QuaSAQ, the total time for executing any query plans is exactly the same since

the streaming time for a media object is fixed As a result, processing time is no

longer a valid metric for cost estimation of the QoS-aware query plans

We propose a cost model that focuses on the resource consumption of

alter-native query execution plans Multimedia delivery is generally resource intensive,

especially on the network bandwidth Thus, to improve system throughput is an

important design goal of media systems Intuitively, the execution plan we may

choose should be one that consumes as few resources as possible and yet meets

all the QoS requirements Our cost model is designed to capture the ‘amount’

of resources used in each plan Furthermore, the cost model id also valid for

other global optimization goals such as minimal waste of resources, maximized

user satisfaction, and fairness Our ultimate goal is to build a configurable query

optimizer whose optimization goal can be configured according to user (DBA)

inputs We then evaluate plans by their cost efficiency that can be denoted as:

where C is the cost function, the resource vector of the plan being evaluated,

and G the gain of servicing the query following the plan of interest An optimal

plan is the one with the highest cost efficiency The generation of the G value

of a plan depends on the optimization goal used For instance, a utility function

can be used when our goal is to maximize the satisfiability of user perception of

media streams [8] A detailed discussion of the configurable cost model mentioned

above is beyond the scope of this paper Instead, we present a simple cost model

that aims to maximize system throughput

Lowest Resource Bucket (LRB) model. Suppose there are types of

re-sources to be considered in QuaSAQ, we denote the total amount of resource

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 22

QuaSAQ: An Approach to Enabling End-to-End QoS 703

as In our algorithm, we build a virtual resource bucket for each individual

resource All values are standardized into the height of the buckets

There-fore, the height of all buckets is 1 (or 100%) The buckets are filled when the

relevant resources are being used and drained when the resources are released

Therefore, the height of the filled part of any bucket is the percentage of

re-source that is being used For example, the filled part of bucket R2 in Figure

3d has height 42, which means 42% of is currently in use The cost evaluation

is done as follows: for any plan we first transform the items in resource

vector into standardized heights related to the corresponding bucket (denoted

as we then fill the buckets accordingly using the transformed

re-source vector and record the largest height among all the buckets The query

that leads to the smallest such maximum bucket height wins In Figure 3, the

cost of three plans (a, b, c) are marked by dotted lines Putting them all

to-gether, we found the filled height of plan 2 is the lowest and plans 2 is chosen

for execution Formally, the cost function of the LRB model can be expressed as

where is the current usage of resource The input is the resource vector of

the plan being evaluated

Fig 3. Cost evaluation by the Lowest Resource Bucket model

The reasoning of the above algorithm is easy to understand: the goal is to

make the filling rate of all the buckets distribute evenly Since no queries can

be served if we have an overflowing bucket, we should prevent any single bucket

from growing faster than the others This algorithm is not guaranteed to be

optimal, it works fairly well, as shown by our experiments (Section 5.2)

Trang 23

704 Y.-C Tu et al.

3.5 QoS APIs

The Composite QoS API hides implementation and access details of

underly-ing APIs (i.e system and network) and offers control to upper layers (e.g Plan

Generator ) at the same time The major functionality provided by the

Com-posite QoS API is QoS-related resource management, which is generally

accom-plished in the following aspects: 1 Admission control, which determines whether

a query/plan can be accepted under current system status; 2.Resource

reserva-tion, an important strategy toward QoS control by guaranteeing resources needed

during the lifetime of media delivery jobs; 3.Renegotiation that are mainly

per-formed under two scenarios mentioned in Section 3.2

Transport API. It is basically composed of the underlying packetization and

synchronization mechanisms of continuous media, similar to those found in

gen-eral media servers The Transport API has to honor the full reservation of

re-sources This is done through interactions with the Composite QoS API The

interface to some of the other server activities such as encryption, transcoding,

and filtering are also integrated into the Transport API

We implement a prototype of QuaSAQ on top of the Video Database

Man-agement System (VDBMS) developed at Purdue University [4] The QuaSAQ

development is done using C++ under the Solaris 2.6 environment Figure 4

shows the architecture of VDBMS enhanced with the QuaSAQ prototype

Fig 4. System architecture for QoS-aware multimedia DBMS prototypePlease purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Trang 24

QuaSAQ: An Approach to Enabling End-to-End QoS 705

QuaSAQ and VDBMS. Developed from the object-relational database

en-gine PREDATOR[9] with Shore[10] as the underlying storage manager (SM),

VDBMS is a multimedia DBMS that supports full-featured video operations (e.g

content-based searching, streaming) and complex queries Most of the VDBMS

development was done by adding features to PREDATOR We extended the

cur-rent release of VDBMS, which runs only on a single node, to a distributed version

by realizing communication and data transferring functionalities among

differ-ent sites As shown in Figure 4, QuaSAQ augmdiffer-ents VDBMS and sits between

Shore and PREDATOR in the query processing path In our QuaSAQ-enhanced

database, queries on videos are processed in two steps: 1 searching and

identifi-cation of video objects done by the original VDBMS; 2 QoS-constrained delivery

of the video by QuaSAQ [3] In VDBMS, the query processor returns an object

ID (OID), by which Shore retrieves the video from disk With QuaSAQ, these

OIDs refer to the video content (represented by logical OID) rather than the

entity in storage (physical OID) since multiple copies of the same video exist

In QuaSAQ, the mapping between logical OIDs and physical OIDs are stored

as part of the metadata (Section 3.3) Upon receiving the logical OID of the

video of interest from PREDATOR, the Quality Manager of QuaSAQ annotates

a series of plans for QoS-guaranteed delivery and chooses one to execute It

communicates with either QuaSAQ modules in remote sites or local Shore

com-ponent (depending on the plan it chooses) to initiate the video transmission

Note the sender of the video data is not necessarily the site at which the query

was received and processed

QuaSAQ Components. Most of the QuaSAQ components are developed by

modifying and augmenting relevant modules in VDBMS (e.g client program,

SQL parser, query optimizer) Replicas for all videos in the database are

gener-ated using a commercial video transcoding/encoding software VideoMach2 The

choice of quality parameters is determined in a way that the bitrate of the

re-sulting video replicas fit the bandwidth of typical network connections such as

T1, DSL, and modems [7] To obtain an online video transcoder, we modified the

source code of the popular Linux video processing tool named transcode3 and

integrated it into the Transport API of our QuaSAQ prototype The major part

of the Transport API is developed on the basis of a open-source media

stream-ing program4 It decodes the layering information of MPEG stream files and

leverages the synchronization functionality of the Real Time Protocol (RTP)

We also implement various frame dropping strategies for MPEG1 videos as part

of the Transport API We build the Composite QoS APIs using a QoS-aware

middleware named GARA [11] as substrate GARA contains separate managers

for individual resources (e.g CPU, network bandwidth and storage bandwidth)

The CPU manager in GARA is based on the application-level CPU scheduler

DSRT [12] developed in the context of the QualMan project [13] QoS-aware

network protocols are generally the solution to network resource management,

Trang 25

706 Y.-C Tu et al.

which requires participation of both end-systems and routers In GARA, the

DiffSrv mechanism provided by the Internet Protocol (IP) is used

We evaluated the performance of QuaSAQ in comparison with the original

VDBMS system The experiments are focused on the QoS in video delivery as

well as system throughput An important metric in measuring QoS of networked

video streaming tasks is the inter-frame delay, which is defined as the interval

between the processing time of two consecutive frames in a video stream [12,

14] Ideally, the inter-frame delay should be the reciprocal of the frame rate of

the video For the system throughput, we simply use the number of concurrent

streaming sessions and the reject rate of queries

Experimental setup. The experiments are performed on a small distributed

system containing three servers and a number of client machines The servers

are all Intel machines (one Pentium 4 2.4GHz CPU and 1GB memory)

run-ning Solaris 2.6 The servers are located at three different 100Mbps Ethernets

in the domain of purdue.edu Each server has a total streaming bandwidth of

3200KBps The clients are deployed on machines that are generally 2-3 hops

away from the servers Due to lack of router support of the DiffSrv mechanism,

only admission control is performed in network management A reasonable

as-sumption here is that the bottlenecking link is always the outband link of the

severs and those links are dedicated for our experiments Instead of user

in-puts from a GUI-based client program [4], the queries for the experiments are

from a traffic generator Our experimental video database contains 15 videos in

MPEG-1 format with playback time ranging from 30 seconds to 18 minutes For

each video, three to four copies with different quality are generated and fully

replicated on three servers so that each server has all copies

5.1 Improvement of QoS by QuaSAQ

Figure 5 shows the inter-frame delay of a representative streaming session for

a video with frame rate of 23.97 fps The data is collected on the server side,

e.g the processing time is when the video frame is first handled Only end-point

system resources should be considered in analyzing server-side results The left

two graphs of Figure 5 represent the result of the original VDBMS while the

right two graphs show those with QuaSAQ We compare the performance of

both systems by their response to various contention levels On the first row,

streaming is done without competition from other programs (low contention)

while the number of concurrent video streams are high (high contention) for

experiments on the second row

Under low contention, both systems (Fig 5a and 5b) demonstrated timely

processing of almost all the frames, as shown by their relatively low variance of

inter-frame delay (Table 2) Note that some variance are inevitable in dealing

with Variable Bitrate (VBR) media streams such as MPEG video because the

Please purchase PDF Split-Merge on www.verypdf.com to remove this watermark.

Ngày đăng: 24/12/2013, 02:18

TỪ KHÓA LIÊN QUAN