International Journal of Application or Innovation in Engineering & Management... Web Site: www.ijaiem.org Email: , Volume 2, Issue 12, December 2013

advertisement
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
Analysis of Classical Web-Centric Query
Optimization Techniques
Atul Garg
Associate Professor, MMICT & BM, MMU, Mullana, Ambala-133 207
ABSTRACT
The growth of Internet and web is changing the life style of current generation. In its result, heterogeneous data is available on
the web and to retrieve integrated data is a challenge for the researchers and developers. Query optimization techniques are used
for information retrieval. In this paper, web centric classical queries processing and optimization techniques, classification of the
different present techniques and a comprehensive framework is discussed.
Keywords: Mediator, Disco, Garlic, Hermes, Tsimmis, Infomaster, HiQIQ etc.
1. INTRODUCTION
Nowadays, Internet is playing an important role in the overall growth of the current generation. These days, people are
engaging in more and more online activities ranging from reading news, banking, banking to conducting business,
advertisement, online shopping, education, research etc. This has been particularly enabled by the technical advances
achieved in computer science. The various fields of computer science like high speed networking, communication
protocols, mark-up languages, graphical interfaces, Java technologies, and communication middleware are growing.
Earlier (in evolution period of web) the text-based web pages were provided to the users. But, as the web become more
powerful and more in demand the pressure from the user’s side is increased. The pressure was in the form of type of data,
Internet speed and Information Retrieval methods. As a result, these days, a variety of data which can be multimedia-rich
is available on the web. So, as the different type of data is available on the Internet in its result different type of
information. To retrieve different type of information the web developers are facing various challenges. Few of them are
discussed below:
 Heterogeneous Data: The data on the web is not only text but it is a combination of text, images and multimedia.
To retrieve this type of data the user sends only one query or request.
 Security Issues: The people on the Internet are doing various secure transactions like online shopping, Internet
banking etc. To retrieve or use the secure transaction the users are using the same Internet and tools.
 Performance Issues: The good performance is the key of any web site. If it is poor in performance then the users
select another site.
 Remote Servers: With the global growth of Internet the data is stored in near or far severs. So, to retrieve the data
from remote server it increases the response time and security issues.
 Decontrol: The traditional information retrieval systems were in control due to only one server and that was also in
control and in security. But now, in web-centric information retrieval system there are various servers and to
search the information the request information being moving from one server to another. So, the web-centric
queries are decontrolled after the submitting the query.
 Lager number of clients: With the revolution of electronic devices, the web users are growing day by day and
become a challenge for the web developers. Because, the developer don’t know in advance with type of device is
going to be use to retrieve the information.
Query optimization is a function of many relational database management systems in which multiple query plans for
satisfying a query are examined and a good query plan is identified [11]. It takes a central stage in data integration
systems on the web. The features of web make the task of enabling efficient querying even more difficult. The mediator
approach [8] is adopted by most of the described systems for data integration. The related systems are those that match
information requests from consumers, individuals or applications, to information providers [23].
Further, the Web’s volatility and highly dynamic nature are a challenge when the expectation is that queries always
return specific results. Also, all information sources are not expected to provide the same query capabilities. So, the
working of query processor becomes more difficult and it needs to make sure that the generated query execution plan is
feasible with respect to these limitations. In this paper, web centric classical queries processing and optimization
techniques are discussed. A classification of the different present techniques and a comprehensive framework is also
discussed. Mediator approach [5] is mostly used for data integration. The classical systems match the request from the
user or users to the host. The classical query optimization techniques are classified based on optimization techniques. The
Volume 2, Issue 12, December 2013
Page 101
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
numerous techniques like cost-based optimization techniques, adaptive query optimization, quality-based optimization,
and query optimization are discussed.
2. BACKGROUND
In this section, the pre work done by the renowned researchers is discussed. In the conventional query optimization
system a query for example in MySQL, there could be many execution plans to retrieve the results. The different methods
respond the same results. Though, the different plans are different in their performance, response time and their
parameters (resources). Different researchers are doing work on query optimization focusing on the performance, fast
accessibility and security of the web-centric quires.
Unel et al., [7] focused on query optimization techniques for multimedia databases. Further, query processing strategies
and query optimization techniques are different in multimedia and traditional databases. Researchers proposed query
optimization approaches for processing spatio-temporal queries in video database systems and implemented on a specific
system called BilVideo [7 and 19]. They applied reordering algorithms named internal node reordering and leaf node
reordering on query execution tree. Rao [29] presented sampling techniques based on selecting points from the parameter
space at random and optimizing them using a conventional optimizer. The set of plans returned was a subset of the
parametric optimal set of plans.
Ouzzani and Bouguettaya [23] discussed the difficulties of query optimization for unstructured information on the web.
The authors pointed out that querying abundant information scattered on the web raises several difficulties in achieving
efficiency. The authors surveying their research on fundamental problems to efficiently process queries over web data
integration systems were evaluated and in fact four categories of query optimization techniques i.e. cost-based, qualitybased, capability-based, and adaptive-based, in the context of web-based data integration were highlighted.
Rossi et al., [9] focused on personalized application. The researchers proposed Object-Oriented Hypermedia Design
Method (OOHDM) [9] to attract scientists to design-oriented discussion of personalization for specifying and designing
personalized behaviours in web applications. According to Liu and Ling [22] queries for the internal hierarchical
structure of web documents is more important. The researchers [22] proposed conceptual model based Web Query
Language (WebQL) [22]. WebQL is a rule-based language for querying the HTML documents for web documents.
Further, WebQL provides a simple and very powerful way to query both the structure and contents of the HTML
documents and to restructure the results. Cardellini et al., [32] considered different categories of web applications, and
evaluated how static, dynamic and secure requests affect performance and quality of service of heterogeneous information
on the web. Authors proposed a distributed web architecture that is able to guarantee the assessed service level agreement
for all client requests.
Srivastava et al., have proposed the overall goal of a general-purpose Web Service Management System (WSMS) [31],
enabling clients to query a collection of web services in a transparent and integrated approach. The authors focused on
new query optimization issues that arose in a WSMS. Their execution model consists of pipelined query processing over
web services and they derived the “bottleneck” cost metric to characterize the cost of a pipelined plan. For this cost metric,
the new algorithms not only to decide the optimal arrangements of web services in a pipelined plan respecting precedence
constraints but also to decide the optimal chunk size to use when sending data to each web service were devised. These
algorithms formed the basis of a WSMS query optimizer.
Lam and Ozsu [30] stated that high volatility of web, make querying of data difficult. A system called WebQA that
provides a declarative query-based approach to web data retrieval that uses question-answering technology in extracting
information from web sites that are retrieved by search engines was introduced. Hu and Bagga [14] focused on the idea
that the web documents are increasingly accessed from wireless devices with small screens and low bandwidths. Web
documents contain a large number of images and in fact, to identify the function of each image according to the device is
an important issue in web content repurposing. The researchers proposed to explore image-recognition techniques to
interpret images more effectively in the icon/logo category.
Cali and Calvanese [3] argued that the problem of optimizing conjunctive query answering in relational database. They
showed an effective optimization technique for pruning unnecessary accesses from query plans; their approach is
applicable in the case of full conjunctive queries that are different from literature. Meinecke et al., [13] introduced Web
Composition Architecture Model (WAM) as an overall modeling approach tailored to aspects of highly distributed
systems with federation as a related issue. Meinecke [13] and his co-researcher applied the technology used for the
functional parts of the architecture and provided a web service for querying and changing the model.
Volume 2, Issue 12, December 2013
Page 102
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
Gupta [4] et al., presented a system that matches user queries with web services operations. The research employed a selflearning mechanism and a set of algorithms and optimizations to match user queries with corresponding operations in
web services. Their system uses lexical analysis, domain-independent matching techniques, domain-specific ontologies
and a set of specialized algorithms and optimizations to match simple free-form queries. They proved that the proposed
approach is better than traditional approach in terms of effectiveness and performance.
3. ANALYSIS AND COMPARISON
In order to make use of vast amount of data, proficient techniques to access web document information based on its
content were needed to be developed. As a result, the role of Information Retrieval (IR) systems became most important.
On the whole, the most significant and difficult operations in information retrieval was to generate queries that can
succinctly identify relevant documents and reject irrelevant documents [25]. Therefore, in order to address the
prominent i.e. moving from data-centric approach to web-centric approach, more known classical web centric query
optimization techniques were discussed and are presented. A classification of the different existing techniques and a
comparison among them is highlighted in Figure 1.
Figure 1: Classification of Query Optimization
The extended query optimization techniques was already analyzed and compared in [1]. Now in classical query
optimization techniques a review of cost-based optimization techniques, adaptive query optimization, quality-based
optimization, and query optimization in presence of sources with limited capabilities is being presented.
Wiederhold [6] presented an architecture named as mediators. It was based on for some basic issues like maintenance,
problem modelling, learning and knowledge attainment. Mediators represent a schema for multiple heterogeneous and
independent information sources. Further, in this schema an interface is used for location, model and transparency. A
wrapper is used for each source to show its participation in the mediator system. Wiederhold [6] apply this schema over
a specific application domain. The working of mediator is explained in Figure 2 [6 & 23].
Figure 2: Mediator Based Architecture
3.1 Cost Based Optimization Techniques
Cost-based optimization techniques help to find or select the best efficient execution plan from the available network
paths with the help of available schemas. The basic objective this technique or approach is best throughput with
minimal resources. Information on web is in heterogeneous form. Now, to retrieve the information the basic challenge is
to optimize the query and the response information. As discussed above the mediator keeps all the information with
wrappers, so the estimated or provided information can be fetched from wrapper. Cost based optimization techniques
are further divided in different types. Few of them are discussed below:
 Disco (Distributed Information Search Component): Tomasic et al., [2], define the user of web as end users
(Focus on data), application developers (focus on presentation), Database Administrators (focus on wrapping) and
Database Implementers (focus on performance). Naacke [10], Disco is a cost based optimization technique. It combines
a generic cost model with specific cost information exported by wrappers The Disco approach deals with the unavailable
data. The Disco mediator contains its own internal database (see Figure 3) [2, 10 & 23]. Disco returns a biased answer.
In the internal database on Disco it contains interfaces, views, data information, registration details etc. By [23], there are
two basic algorithms are used named as cost formula integration and cost estimation. First algorithm is used for root to
leaf, formulas are applied on nodes and in these rules are used to connect with the mediator cost model and 2nd algorithm
Volume 2, Issue 12, December 2013
Page 103
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
is applied for its back traversal which is based on plans and operator’s cost is computed. In this approach, from the root to
the leaves and back method is used to calculate the cost.
Figure 3: Disco Architecture
 Garlic Architectures: Garlic [18, 20 and 21] approach is used for heterogeneous multimedia (text, videos,
images etc.) information. An incorporated view of heterogeneous information sources shown in Figure 4 is presented by
this approach.
Figure 4: Garlic Architecture
The query processor in the garlic architecture generates plans to apply accurate query on the heterogeneous repositories of
information. This approach decomposes the quires and repositories in to pieces to distribute the load of repositories. In
the Garlic approach wrappers plays an important role and with the help of wrappers this technique follows and improve
the cost estimation techniques. In this framework, cost model, cost formulas, and statistics techniques are used for cost
estimation.
 Ariadne: Ambite and Knoblock [15] proposed a new approach PbR (Planning by Rewriting) based on mediator
approach. The planning for a query is most important. For example it is worth to plan more and more if the total cost is
very small. With the flexibility of PbR approach the planning is applied on different scale projects differently and in its
continuation it reduces the cost. Secondly, in PbR an initial plan is followed by rewriting rules. The rewriting rules are
iteratively applied until a satisfactory result or a resource limit is reached.
 Hermes (Heterogeneous Reasoning and Mediator System): Subrahmanian et al.,[33] proposed a mediator
language for consistency to access data sources and reasoning system. Further they discuss main aspects of mediator
approach. Domain integration is used to add new data or reasoning system to maintain the whole system. Whereas, in
Semantic Integration the procedure used to determine conflicts, to find mutual pool information, and define new
operations. In Hermes, query optimization is based on cost-based model, because it uses caching technique [28]. All the
calls which are locally available are not sent to the data servers. With this technique, the execution time is improved
and further the cost of possible execution plans is estimated.
3.2 Quality-Based Optimization Techniques
In Web-centric query optimization the quality of the information is more important. In these cases, the cost is estimated
with different parameters. These optimization techniques are as much important as others and few of them are as
follows:
Volume 2, Issue 12, December 2013
Page 104
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
 ObjectGlobe: Braumandl et al., [26] developed a distributed query processing system based on quality of service
(QoS) and query evaluation plan (QeP). The authors [26] proposed new services named as cycle and function providers
for the users and the information providers. The developers Braumandl et al., [26], compare their system with the garlic
systems. In the garlic system no new functionality can be transport to the data, but this technique is more dynamic and
it is possible. The researchers [26 & 23] propose the four major steps i) lookup service is queried to find relevant
information sources, cycle providers, and query operators that might be useful to execute the query. Some more
information regarding cost is also collected during lookup ii) an optimal plan is built to fulfil the quality constraints of
the user, based on information gathered from lookup iii) Plug in this step the plan is spread to the external query
operators and the cycle providers iv) in the last step the plan is executed with an iterator model. Further, the quality
constraints are applied on each and every phase. The query plan is dynamically adapted or the query is aborted, based
on that QoS concept, the goal of ObjectGlobe is to increase the ratio of successful queries.
 HiQIQ (High Quality Information Querying) Naumann and Lesser [5] incorporated Information Quality (IQ)
in the query plan. The approach is a combination of source centric and query-centric approaches used in mediators.
Different criteria are applied in Information Quality (IQ) with query planning. The first criteria, which are Query
Correspondence Assertions (QCA) specific criteria, include response time, availability, price, accuracy, relevancy and
consistency. The second, query-specific criteria check the quality of the answer received corresponding to each query.
The Information Quality is applied based on above criteria to improve the quality.
3.3 Adaptive query optimization
Adaptive or dynamic query optimization is most important in heterogeneous web-centric query optimization technique.
These techniques remove the unpredictable events during query execution. Adaptive Query Optimization techniques are
active at run time to change the execution plan, repeat the optimization query with some specific operators to increase
the flexibility. The adaptive query, dynamic optimization finds its applications in various projects and approaches.
 Telegraphy: Hellerstein [16] et al., proposed dataflow architecture based on rivers and eddies named as
telegraphy query engine. Telegraphy is architected for fine-grained adaptivity and interactive user control. To improve
both the extent and the effects is the main goal of telegraphy. In this approach a metadata is used to store the
information at end of each query and this information is used for the next queries, whereas to enhance the set of effects
from rivers and eddies is more complicated. The working of eddy [27] is as in the following steps i) the eddy set order of
tuples to operators, ii) the operators executes an independent thread, iii)and back it return tuples to the eddy. The eddy
sends a tuple to the output only when it has been handled by all the operators.
 Interleaving scheduling and optimization: Bouganim [17] proposed a new dynamic scheduling approach to
deal with limited memory. The researchers basically focus on the performance problems due to random and wanted data
release rates in data integration systems. This approach monitors the arrival rates at the information sources and
memory availability. Further, the execution plan is revised as there is any significant change. The above statements
prove that planning and execution phases are interleaved. In their query execution plan they proposed two types of
edges names as blocking edge and pipelineable edge respectively. In the first edge the data is completely produced
before consumed, it means the blocking input have to wait for the whole data before start. Whereas, on another edge the
consumer can start with a single tuple and no need to wait.
Further, Query engine divides the query execution into several components: dynamic query optimizer, dynamic query
scheduler and processor, and communication manager [17]. The dynamic query optimizer implements dynamic reoptimization techniques. Each planning phase of the dynamic query optimizer can modify the query execution plan
(QEP). Dynamic query scheduler takes the input from QEP and schedule the planning. The dynamic query processor
implements the execution component of the system and processes concurrently. The Communication Manager (CM)
receives data from the wrappers and makes it available to the dynamic query processor.
3.4 Optimizing queries over sources with limited capabilities
The web-centric queries optimization technique deals with different sources and queries. In the web queries the
sources follow the specific requirements and limited constraints. For example, users can send only one query at a time.
He/she can’t check the whole database with a single query. There are various reasons behind to send a single query at a
time. The main reasons are: i) security ii) performance and iii) specific business. To overcome on these problems there
are many query optimization techniques named as Tsimmis, Infomaster, Information Manifold, Annotated query plans
and source sequence plan with limited capabilities. In analysis, Tsimmis and Infomaster are important so these are
discussed in detail.
 Tsimmis: Molina et al., [12], proposed a prototype to integrate volatile information sources for mediation. The
basic principles of Tsimmis (Figure 5 and Figure 6) [12 & 23] are: i) Object Exchange Model (OEM) is a lightweight due
Volume 2, Issue 12, December 2013
Page 105
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
to less typing objects, it convey the information between components, ii) Mediator specification Language (MSL) is logic
based which targets OEM, iii) the extension of MSL are wrappers or translators which are specific with Wrapper
Specification Language (WSL) to permit for the description of source contents and querying capabilities iv) Wrapper- and
mediator-generators. Tsimmis approach focuses on feasible query plans with limited capabilities of available sources.
Figure 5: Tsimmis Prototype
Figure 6: Tsimmis Query Processor
 Infomaster: By Duschka & Genesereth [24], Infomaster gives an illusion of centralized volatile information
system. It an information integration system for homogeneous data. For mediation, Infomaster [Figure 7] is sourcecentric, with different type of relations: i) for users to prepare queries interface-relations are available ii) for data
information source-relations are available and iii) to generate a schema for the system and to describe the above relation a
world-relation is used.
Figure 7: Infomaster Architecture
By [23 and 24], the working of query processor in Infomaster follow the five basic steps i) Reduction: in this step the
system rewrite each atom of query in terms of base relations ii) Abduction: base relations based expression is produced
and with a set of definitions. The set of all consistent and minimal conjunctions and retrievable atoms is produced iii)
Conjunctive minimization: this step eliminates all the redundant conjunct iv) Disjunctive minimization: in this step
Infomaster drops all disjuncts that can be shown in the union of the remaining disjuncts v) Grouping: the atoms are
grouped within each conjunction for presentation.
Table 1 lists the comparison of the optimization techniques presented so for. It is evident that none of technique can be
called as the best and in fact all approaches are application specific rather than being generic.
Table 1 : Comparative Analyses of Various Classical Query Optimization
Volume 2, Issue 12, December 2013
Page 106
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
4. CONCLUSIONS
In addition to difficulties encountered in optimizing queries various query optimization techniques are explored. To
optimize the query and to control query execution is a difficult after submitted the query on the web. Further, with these
problems the information source may show different behaviour from its expectations. So, it may reduce the expected
results. It shows that, the traditional optimization techniques that are based on static information may not be appropriate
for heterogeneous information. Further, expectations from web are increasing day by day and as its result the dynamic
nature of web is also changing. Web centric query optimization could not work in the boundaries of classical query
optimization techniques.
REFERENCES:
[1] A. Tomasic, L. Rashid, and P. Valduriez, “Scaling heterogeneous database and design of DISCO,” in Proceedings of
the 16th International Conference on Distributing Computing Systems (ICDCS), Hong Kong, May 1996.
[2] Andrea Cali and Diego Calvanese, “Optimising Query Answering in the Presence of Access Limitations”, in
proceedings of the 17th International Conference on Database and Expert Systems Applications (DEXA'06), 0-76952641-1/06, IEEE Computer Society-2006.
[3] Atul Garg and Dimple Juneja “A Comparison and Analysis of various extended Techniques of Query Optimization”,
in International Journal of Advancements in Technology, Vol. 3(3), July 2012, ISSN : 0976-4860 (Online).
[4] Chaitali Gupta, Rajdeep Bhowmik, Michael R. Head, Madhusudhan Govindaraju, and Weiyi Meng, “Improving
Performance of Web Services Query Matchmaking with Automated Knowledge Acquisition” 2007 IEEE/WIC/ACM
International Conference on Web Intelligence, 0-7695-3026-5/07, IEEE Computer Society-2007.
[5] Felix Naumann, Ulf Lesser, “Quality–driven integration of heterogeneous information systems,” in Proceedings of
the 25th International Conference on Very Large Data Bases (VLDB), Edinburgh, UK, Sept. 1999.
[6] G. Wiederhold, “Mediators in the architecture of future information systems,” IEEE Computer, vol. 25, no. 3,1992.
[7] G. Unel, M.E. Donderler, O. Ulusoy and U. Gudukbay, “An efficient query optimization strategy for spatio-temporal
queriesin video databases”, The Journal of Systems and Software 73 (2004) 113–131, www.elsevier.com/locate/jss, 5
December 2003.
[8] G. Wiederhold, “Mediators in the architecture of future information systems”, IEEE Computer, vol. 25, no. 3, 1992.
[9] Gustavo Rossi, Daniel Schwabe and Robson Guimaraes, “Designing Personalized Web Applications”, WWW10,
Hong Kong, China, Copyright 2001 ACM 1-58113-000-0/00/0000, May 1-4, 2000.
[10] H. Naacke, G. Gardarin, and A. Tomasic, “Leveraging mediator cost models with heterogeneous data sources,” in
Proceedings of the 14th International Conference on Data Engineering, Orlando, Florida, Feb. 1998.
[11] http://en.wikipedia.org/wiki/Query_optimization
[12] H. Garcia-Molina, Y. Papakonstantinou, D. Quass, A. Rajaraman, Y. Sagiv, J.D. Ullman, V. Vassalos, and J.
Widom, “The TSIMMIS approach to mediation: Data models and languages,” Journal of Intelligent Information
Systems, vol. 8, no. 2, 1997.
[13] J. Meinecke, M. Gaedke, F. Majer, and A. Brändle, “Capturing the Essentials of Federated Systems”, in proceedings
of the Fourteenth International World Wide Web Conference (WWW), Pages 895-896, Edinburgh, UK, 23-26 May,
2006.
[14] Jianying Hu and Amit Bagga, “Categorizing Images in Web Documents”, 1070-986X/04, IEEE Computer Society,
January-March, 2004.
[15] Jous Luis Ambite and Craig A. Knoblock, “Flexible and scalable query planning in distributed and heterogeneous
environments,” in proceedings of the Fourth International Conference on Artificial Intelligence Planning Systems,
Pitsburg, USA, June 1998.
[16] J.M. Hellerstein, M.J. Franklin, S. Chnadrasekaran, A. Deshpande, K. Hildrum, S. Madden, V. Ramana, and M.A.
Shah, “Adaptive query processing: Technology in evolution,” IEEE Data Engineering Bulletin, vol. 23, no. 2, 2000.
[17] L. Bouganim, F. Fabret, C. Mohan, and P. Valduriez, “Dynamic query scheduling in data integration systems,”in
Proceedings of the 16th International Conference on Data Engineering, San Diego, CA, USA, Feb.–March 2000.
[18] M. Carey, L. Haas, et al., “Towards Heterogeneous Multimedia Information Systems: The Garlic Approach”, RIDEDOM ‘95, Taiwan 3/95. (5th International Workshop on Research Issues in Data Engineering: Distributed Object
Management.)
[19] M. E. Donderler, O. Ulusoy and U. Gudukbay, “A rule-based approach to represent spatio-temporal relations in video
data”, In: Yakhno, T. (Ed.), Advances in Information Systems (ADVIS’00), Lecture Notes in Computer Science,
1909, Springer, Berlin, pp. 409–418, 2000.
[20] M.T. Roth and P. Schwarz, “Don’t scrap it, wrap it! A wrapper architecture for legacy data sources”, in Proceeedings
of the 23rd International Conference on Very Large Data Bases (VLDB), Athens, Greece, Aug. 1997.
[21] M.T. Roth, M Arya, L Haas, M Carey, W Cody, R Fagin, P Schwarz, J Thomas, E Wimmers and P. Schwarz, “The
Garlic Project”, pp 557, ACM, 1997.
[22] Mengchi Liu and Tok Wang Ling, “A Rule-based Query Language for HTML”, 0-7695-0996-7/01, IEEE-2001.
Volume 2, Issue 12, December 2013
Page 107
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 2, Issue 12, December 2013
ISSN 2319 - 4847
[23] Mourad Ouzzani and Athman Bouguettaya, “Query Processing and Optimization on the Web”, Distributed and
Parallel Databases, Kluwer Academic Publishers. Manufactured in The Netherlands, 15, 187–218, 2004.
[24] O.M. Duschka and M.R. Genesereth, “Query planning in infomaster,” in Proceedings of the Twelfth Annual ACM
Symposium on Applied Computing, SAC ’97, San Jose, CA, USA, Feb. 1997.
[25] R. H. Arpaci-Dusseau, E. Anderson, N. Treuhaft, D.E. Culler, J.M. Hellerstein, D. Patterson, and K. Yelick, “Cluster
I/O with River: Making the fast case common,” in proceedings of the Sixth Workshop on I/O in Parallel and
Distributed Systems. ACM Press, May 1999.
[26] Reinhard Braumandl, Markus Keidl, Alfons Kemper, Donald Kossmann, Stefan Seltzsam, Konrad Stocker,
“ObjectGlobe: Open Distributed Query Processing Services on the Internet,” in the Bulletin of the Technical
Committee on March 2001 Vol. 24 No. 1, IEEE Computer Society.
[27] Ron Avnur and Joseph M. Hellerstein, “Eddies: Continuously adaptive query processing,” in Proceedings of the
ACM SIGMOD International Conference on Management of Data, Dallas, Texas, USA, May 2000.
[28] S. Adali, K.S. Candan, Y. Papakonstantinou, and V.S. Subrahmanian, “Query caching and optimization in
distributed mediator systems,” in Proceeedings of ACM SIGMOD International Conference on Management of Data,
Montreal, Canada, June 1996.
[29] S. V. U. M. Rao., “Parametric query optimization: A non-geometric approach”, Technical report, Indian Institute of
Technology, Kanpur, Available at http://www.cse.iitk.ac.in/research/mtech1997, Feb 1999.
[30] Sunny K.S. Lam and M. Tamer Ozsu, “Querying Web Data – The WebQA Approach”, in the proceedings of the 3rd
International Conference on Web Information Systems Engineering (WISE.02), 0-7695-1766-8/02, IEEE Computer
Society, 2007.
[31] Utkarsh Srivastava, Kamesh Munagala, Jennifer Widom and Rajeev Motwani, “Query Optimization over Web
Services”, VLDB ‘06, Seoul, Korea, ACM 1595933859/06/09, 2006.
[32] Valeria Cardellini, Emiliano Casalicchio and Michele Colajanni, “A Performance Study of Distributed Architectures
for the Quality of Web Services”, in proceedings of the 34th Hawaii International Conference on System Sciences –
2001, 0-7695-0981-9/01, IEEE, 2001.
[33] V.S. Subrahmanian, Sibel Adali, Anne Brink, James J. Lu, Adil Rajput, Timothy J. Rogers, Robert Ross, Charles
Ward, “A Heterogeneous Reasoning and Mediator System”, http://www.cs.umd.edu/projects/hermes/overview/paper/.
Atul Garg received degree of Master of Computer Applications from Kurukshetra University, Kurukshetra in
2004 and completed his Ph. D degree from Maharishi Markandeshwar University, Mullana (Ambala) in 2013.
Currently, he is working as an Associate Professor at M. M. I. C. T. & B. M., Maharishi Markandeshwar
University, Mullana (Ambala), Haryana. His area of interest is web and Query Optimizations.
Volume 2, Issue 12, December 2013
Page 108
Download