WO2016012903A1 - A system for querying heterogeneous data sources and a method thereof - Google Patents

A system for querying heterogeneous data sources and a method thereof Download PDF

Info

Publication number
WO2016012903A1
WO2016012903A1 PCT/IB2015/055343 IB2015055343W WO2016012903A1 WO 2016012903 A1 WO2016012903 A1 WO 2016012903A1 IB 2015055343 W IB2015055343 W IB 2015055343W WO 2016012903 A1 WO2016012903 A1 WO 2016012903A1
Authority
WO
WIPO (PCT)
Prior art keywords
query
data sources
data
data source
processing time
Prior art date
Application number
PCT/IB2015/055343
Other languages
French (fr)
Inventor
Keeratpal SINGH
Ettikan Kandasamy Karuppiah
Fazli MAT NOR
Original Assignee
Mimos Berhad
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mimos Berhad filed Critical Mimos Berhad
Publication of WO2016012903A1 publication Critical patent/WO2016012903A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2455Query execution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2458Special types of queries, e.g. statistical queries, fuzzy queries or distributed queries
    • G06F16/2471Distributed queries

Definitions

  • the present invention relates to the field of data warehousing and data mining, particularly to a system and method for processing queries for heterogeneous data sources incorporated within a data warehouse.
  • choosing big data processing technology for querying small data sets, or choosing RDBMS to perform real time queries on big data results in inefficient use of resources which leads to low output performance, inefficient use of resource time and latency in observing simulating parameter changes, latency of obtaining real time Bl (Business Intelligence) reports and latency of generating analytics result.
  • Usage of different data sources through various databases (or file system) of different data sizes for the purpose of real time analytics and simulation leads to longer response time to achieve query results due to inefficient selection of resource models.
  • the expression 'heterogeneous data sources' used hereinafter in the specification relates to a plurality of data sets or file systems having different sizes or volumes.
  • the data source may represent 'Big Data' clusters, small / medium relational databases or the like.
  • the expression 'query' used hereinafter in the specification relates to a structured and syntactical coding language used for retrieving information from a data source / file system.
  • the expression 'route' used hereinafter in the specification relates to an interface which connects to a set of resources including a data source such as a database; hardware including Random Access Memory (RAM), processor, and graphics processor and software such as operating system / file system.
  • a data source such as a database
  • hardware including Random Access Memory (RAM), processor, and graphics processor and software such as operating system / file system.
  • class of service used hereinafter in the specification relates to an indicator which determines processing time based on the type of application/ query. It also determines priority for processing a query and a time within which a response for that query must be obtained.
  • class of service may be categorized as Class 1 , Class 2, Class 3 and Class 4, wherein Class 1 gets highest priority and represents applications including simulation, prescriptive analytics, Bl real time reports and web response which must be processed within 15 seconds; Class 2 gets second priority and represents application such as generation of Big Data static reports which must be generated within 60 seconds; Class 3 gets third priority and represents batch processing applications which must be processed within a time range between one minute and a hour; and Class 4 gets lowest priority and represents applications which take more than an hour.
  • machine readable media used hereinafter in the specification refers to RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor.
  • the expression 'computer program product' is defined as a manufactured product embodied in a machine readable medium as defined herein above.
  • the system uses machine learning and classification rules for predicting a best data source to execute a query in order to be able to minimize the response time for meeting real time response constraints.
  • the system comprises a plurality of data sources, wherein each data source has a discrete volume, discrete hardware resources and tools for facilitating processing of queries.
  • the system also comprises a central orchestrator unit to select at least one machine learning module for processing of queries; a predictor module to receive at least one query for retrieving information from at least one of the data sources; a selector module having a multiple connector interface for connecting with each of the data sources; and an output module cooperating with the data sources.
  • the predictor module predicts discrete processing time for executing the query for available routes in the system based on the machine learning module selected by the central orchestrator unit. Subsequently, the selector module compares the predicted processing time for each route generated by the predictor module with the class of service of the query. The selector module selects an appropriate data source for execution of the query via the interface based on the comparison of the predicted processing time for each route and class of service. The query is then processed by the selected data source within the processing time predicted for that route and the query results are displayed on the output module.
  • FIGURE 1 discloses a schematic of the system for querying heterogeneous data sources in accordance with the present invention
  • FIGURE 2 is a flowchart showing the steps for querying heterogeneous data sources in accordance with the present invention
  • FIGURE 3 is a flowchart showing the steps performed by a query editor in accordance with this invention.
  • FIGURE 4 is a flowchart showing the steps performed by a central orchestrator unit for selecting a machine learning module for improving accuracy of selecting a best route for queries in accordance with this invention.
  • the present invention proposes a system and a method for querying heterogeneous data sources.
  • the proposed system performs predictive planning via machine learning to predict and select a best proposed route for executing a query within a minimum response time bounded by benchmarks (predicted processing time).
  • FIGURE 1 of the accompanying drawings discloses a system (100) for querying heterogeneous data sources (102 1 ...102 n ), collectively represented by reference numeral (102), and FIGURE 2 discloses a method for querying heterogeneous data sources (102).
  • the system (100) comprises the following components: a plurality of data sources (102); a Central Orchestrator Unit (COU) (106); a query editor (104); a predictor module (108); a selector module (1 10); and an output module (1 12).
  • the data sources (102) include one or more tables that hold requisite information for performing simulation, analytics or similar complex analysis.
  • Each of the data sources (102) have a discrete volume, discrete hardware resources and tools for facilitating processing of queries.
  • the data source (102 ⁇ may be a Relational Database Management System (RDMS) holding small and medium sized data and data source (102 2 ) may be a Big Data cluster.
  • RDMS Relational Database Management System
  • the query editor (104) accepts queries for retrieving information from the data sources (102), step (2000) of FIGURE 2.
  • FIGURE 3 discloses the steps performed by a query editor (104) in accordance with this invention.
  • the query editor (104) preemptively corrects and completes a query while it is being typed using information from a memory module (1 16).
  • the memory module (1 16) is communicably linked to the data sources (102) for holding metadata of the data sources (102) and any previously executed queries.
  • the query editor (104) uses the metadata information stored in the memory module (1 16) to check syntax of a query to determine if typed column names exist within a table name chosen for a data source (102), and also suggests column names or strings while the query is being typed (3002).
  • the query editor (104) checks the query for different syntax as the query may be executed on any one of the available data sources (102). For instance, the query editor (104) checks the syntax for query languages/tools such as Postgres or Hive (Hadoop) or SQL parser of GPU. Further, in case any error is identified in the query, then the query editor (104) highlights appropriate portions of the query with a distinctive color (such as red color) and/or an error message to indicate occurrence of the error (3002). When the query is error-free, the query editor (104) invokes the COU (106) for execution of the query (3004). Typically, the query editor (104) interface includes an execute command button. Once the execute command button is pressed, the query editor (104) invokes the COU (106) for execution of the query.
  • the query editor (104) interface includes an execute command button. Once the execute command button is pressed, the query editor (104) invokes the COU (106) for execution of the query.
  • the COU (106) receives the request from the query editor (104) and facilitates in selection of a best route for a query by selecting at least one machine learning module.
  • the COU (106) firstly identifies a data source (102) which includes the table name appearing in the query. Once the data source (102) has been identified, the COU (106) checks the volume of that data source (102) to determine if the volume is small, general (medium) or big. The COU (106) directly routes the query to the identified data source (102) if the volume of the data source is classified as small. The query is then executed at the data source (102) and results are displayed on the output module (1 12). If the volume is not small and the table name is appearing in all the available data sources then COU (106) forwards the query to the predictor module (108).
  • the predictor module (108) uses the machine learning module selected by COU (106) to predict processing time for executing a query for each of the available data sources (102), step (2004) of FIGURE 2.
  • the predictor module (108) calculates the processing time required by each of the data sources (102 ⁇ , (102 2 ) to (102 n ) for executing the received query.
  • the machine learning modules are a combination of at least one regression model and one classification model to facilitate in carrying out the prediction for each route. These machine learning modules include at least one module from regression analysis, decision trees and neural networks. While performing prediction, the predictor module (108) also takes into account loading capabilities of the data sources (102) and current load and utilization of available hardware resources.
  • the predictor module (108) forwards the processing time predicted for each of the available data sources / routes to the selector module (1 10).
  • the selector module (1 10) receives the 'class of service' classification information from the COU (106).
  • the selector module (1 10) then designates a 'class of service' to the query based on the classification rules and further compares the predicted processing time for each route with the 'class of service' of the query, step (2006) of FIGURE 2.
  • the selector module (1 10) based on result of the comparison selects a route / data source (102) having the lowest processing time for executing the query, step (2008) and (2010) of FIGURE 2.
  • the selector module (1 10) has a multiple connector interface (1 18) for connecting to each of the data sources (102).
  • the selector module (1 10) via the interface (1 18) sends the query to the selected data source (102) for execution.
  • the execution of the query is carried out using tools and techniques well known in the art.
  • the COU (106) facilitates in final selection of a data source (102) in the event that the selector module (1 10) is unable to find a suitable data source (102)/ route for execution of the query.
  • the COU (106) simultaneously sends the query to all available data sources based on the highest class of service of the query requested. The query is then simultaneously executed at the available data sources (102). And results obtained from a data source (102) that has executed the query first are displayed on the output module (1 12).
  • the selected data source (102) then executes the query within the predicted processing time and returns the results to the output module (1 12) as per a required format, step (2012) of FIGURE 2.
  • COU (106) performs additional functions to ensure accuracy of the system (100) and efficient use of resources. The functions are explained hereinafter.
  • the COU (106) includes a scheduler (not shown in the figures) for triggering the functions performed by it, which include: maintaining and classifying the data sources; monitoring utilization of the hardware resources; initiating a self-learning process based on new training data; and testing accuracy of the machine learning module.
  • the scheduler carries out the aforementioned activities during off-peak hours.
  • the self-learning module (1 14) is a database which stores the processing details to enable COU (106) to initiate a self-learning process and test accuracy of the machine learning modules.
  • COU (106) COU
  • execution details of the query along with the predicted processing time and class of service are appended to a training table (not shown in the figures) in the self-learning module (1 14) as new training data.
  • FIGURE 4 which discloses the steps followed by the COU (106) in initiating self-learning rules to re-train and correct conflicting data by selecting various models of machine learning for the purpose of improving accuracy of selecting the best route when queries are generated.
  • the COU (106) fetches the new training data from the self-learning module (1 14) for initiating the self-learning process.
  • the COU (106) selects a machine learning module to evaluate accuracy of processing the failed queries using the new machine learning module (4000).
  • the COU (106) then re-evaluates accuracy of processing the failed queries using a different machine learning module (4002).
  • the accuracy of machine learning modules from step (4000) and (4002) are then compared (4006) to select a machine learning module.
  • the machine learning modules with Area of Receiver Operator Characteristic (ROC) close to 1 are chosen.
  • the selected machine learning module is thereafter used by the predictor module (108). Testing accuracy of the machine learning module
  • the COU (106) regularly checks the system for any failed queries. If there is no improvement in accuracy, the previously existing machine learning module is selected for use by the predictor module (108). If after several days, accuracy further reduces, then the COU (106) randomly chooses one of the previously used machine learning modules having accuracy (>90%) and ROC closer to 1 . If accuracy still degrades ( ⁇ 80%) after a few days for instance, 3 days after using different models, the entry of new training data are compared for "near" conflicting attributes with almost similar attributes in the training table which produce different selection of a data source / route.
  • the instances are run in real time by the COU (106) to verify and take one instance that matches the entry in the self-learning module (1 14) for training data pertaining to regression machine learning modules. For instance, if a first regression model takes lesser time than a second regression model then the first regression model is selected by the COU (106) for prediction. Thereafter, the records in the training table which have "near" conflicting attributes and use the second regression model are eliminated. In this manner the machine learning module is self-corrected. Still further, the COU (106) checks Linear Regression prediction for processing time for daily queries on each data source (102) during execution of queries and compares it with real query processing time. Error difference for each prediction is difference of time predicted with real time.
  • the selected route's training data is updated with new self-learning queries for each data source (102).
  • the COU (106) performs the aforementioned functions for testing the accuracy of the machine learning modules to ensure that an optimum route for executing the query is selected.
  • the COU (106) performs daily checks on all tables of the data sources (102).
  • the COU (106) classifies the tables based on their volume into big data, general data and small data categories. For instance, if tables exceed 100GB, then they are classified as big data and tables below 500MB as classified as small or general data.
  • the COU (106) balances the volume of the data sources (102). If data is not small, then the scheduler initiates import of data from the data sources on incremental basis such as every 10 minutes to a file.
  • the COU (106) continuously monitors real time server load for all the data sources and their hardware resources.
  • the COU (106) generates emails and reports for system administrators if it detects any one of the following conditions: (i) server loading issues during processing of a query, (ii) if queries appear to be taking longer time for execution than predicted, (iii) data sources are unable to meet the class of service requirements or (iv) if accuracy issues appear in routing the queries.
  • the present invention thereby achieves the objectives of efficient utilization of the available resources, usage of an optimum data source to execute the query and achieve results/ output of the query within near real time constraints.

Abstract

A system (100) and a method for querying heterogeneous data sources have been disclosed. The system (100) based on machine learning and classification rules, predicts a best route for querying at least one data source from the heterogeneous data sources. The system (100) minimizes inefficient use of hardware resources and obtains near real time response from the data source selected through the predicted route for executing the query.

Description

A SYSTEM FOR QUERYING HETEROGENEOUS DATA SOURCES AND A METHOD
THEREOF
FIELD OF THE INVENTION
The present invention relates to the field of data warehousing and data mining, particularly to a system and method for processing queries for heterogeneous data sources incorporated within a data warehouse.
BACKGROUND
In today's fierce market, enterprises are constantly relying on analytical and strategic decision making for staying competitive. These complex decisions rely on mining large volumes and varied sources of data for providing necessary analytics. Although, the data comes in various forms and volumes it has to be processed to provide analytical results within real-time response constraints.
These sources of data have volume ranging from big data, medium sized data and small data. Traditional relational databases such as MySQL do not have capability to cope up with increasing demands of analytics especially for providing near real time query response for big data. Likewise, when the data is small, using big data tools such as Hadoop, MapReduce, parallel distributed solutions and algorithms written for general-purpose computing on graphics processing units are not fruitful, and lead to inefficient usage of resources since big data clusters generate overheads in processing time to a query. When data is small or manageable, Relational Database Management System (RDBMS) still process data efficiently fast.
Hence, choosing big data processing technology for querying small data sets, or choosing RDBMS to perform real time queries on big data results in inefficient use of resources which leads to low output performance, inefficient use of resource time and latency in observing simulating parameter changes, latency of obtaining real time Bl (Business Intelligence) reports and latency of generating analytics result. Usage of different data sources through various databases (or file system) of different data sizes for the purpose of real time analytics and simulation, leads to longer response time to achieve query results due to inefficient selection of resource models.
There is therefore felt a need for a system and method for querying heterogeneous data sources, that is data sources having different volumes, resources and forms and obtaining analytics and/or results which meet near real time response constraints.
DEFINITIONS
The terms used throughout this specification are defined as follows, unless otherwise limited in specific instances.
The expression 'heterogeneous data sources' used hereinafter in the specification relates to a plurality of data sets or file systems having different sizes or volumes. For example, the data source may represent 'Big Data' clusters, small / medium relational databases or the like. The expression 'query' used hereinafter in the specification relates to a structured and syntactical coding language used for retrieving information from a data source / file system.
The expression 'route' used hereinafter in the specification relates to an interface which connects to a set of resources including a data source such as a database; hardware including Random Access Memory (RAM), processor, and graphics processor and software such as operating system / file system.
The expression 'class of service' used hereinafter in the specification relates to an indicator which determines processing time based on the type of application/ query. It also determines priority for processing a query and a time within which a response for that query must be obtained. For instance, class of service may be categorized as Class 1 , Class 2, Class 3 and Class 4, wherein Class 1 gets highest priority and represents applications including simulation, prescriptive analytics, Bl real time reports and web response which must be processed within 15 seconds; Class 2 gets second priority and represents application such as generation of Big Data static reports which must be generated within 60 seconds; Class 3 gets third priority and represents batch processing applications which must be processed within a time range between one minute and a hour; and Class 4 gets lowest priority and represents applications which take more than an hour.
The expression 'machine readable media' used hereinafter in the specification refers to RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor.
The expression 'computer program product' is defined as a manufactured product embodied in a machine readable medium as defined herein above.
These definitions are in addition to those expressed in the art.
SUMMARY
Inefficient selection of resources for querying data sources of varied sizes, for the purpose of real time analytics and simulation led to longer response time to achieve query results. This drawback of the prior art led the present invention to propose a system for querying heterogeneous data sources, wherein a response to the query for obtaining analytics and/or simulation results meets near real time response constraints.
The proposed system uses machine learning and classification rules for predicting a best data source to execute a query in order to be able to minimize the response time for meeting real time response constraints. In accordance with this invention, the system comprises a plurality of data sources, wherein each data source has a discrete volume, discrete hardware resources and tools for facilitating processing of queries. The system also comprises a central orchestrator unit to select at least one machine learning module for processing of queries; a predictor module to receive at least one query for retrieving information from at least one of the data sources; a selector module having a multiple connector interface for connecting with each of the data sources; and an output module cooperating with the data sources.
According to this invention, the predictor module predicts discrete processing time for executing the query for available routes in the system based on the machine learning module selected by the central orchestrator unit. Subsequently, the selector module compares the predicted processing time for each route generated by the predictor module with the class of service of the query. The selector module selects an appropriate data source for execution of the query via the interface based on the comparison of the predicted processing time for each route and class of service. The query is then processed by the selected data source within the processing time predicted for that route and the query results are displayed on the output module.
Additional aspects, features and advantages of the invention will become apparent to those skilled in the art upon consideration of the following detailed description of preferred embodiments of the invention in conjunction with the drawings listed below.
BRIEF DESCRIPTION OF THE ACCOMPANYING DRAWINGS
A system and method for querying heterogeneous data sources, in accordance with the present invention will now be described with the help of the accompanying drawings, in which:
FIGURE 1 discloses a schematic of the system for querying heterogeneous data sources in accordance with the present invention;
FIGURE 2 is a flowchart showing the steps for querying heterogeneous data sources in accordance with the present invention;
FIGURE 3 is a flowchart showing the steps performed by a query editor in accordance with this invention; and FIGURE 4 is a flowchart showing the steps performed by a central orchestrator unit for selecting a machine learning module for improving accuracy of selecting a best route for queries in accordance with this invention.
DETAILED DESCRIPTION
The system for querying heterogeneous data sources, of the present disclosure will now be described with reference to the embodiment shown in the accompanying drawings. The embodiment does not limit the scope and ambit of the disclosure. The description relates purely to the exemplary embodiment and its suggested applications.
The embodiment herein and the various features and advantageous details thereof are explained with reference to the non-limiting embodiment in the following description. Descriptions of well-known components and processing techniques are omitted so as to not unnecessarily obscure the embodiments herein. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiment herein may be practiced and to further enable those of skill in the art to practice the embodiment herein. Accordingly, the description should not be construed as limiting the scope of the embodiment herein.
The description herein after, of the specific embodiment will so fully reveal the general nature of the embodiments herein that others can, by applying current knowledge, readily modify and/or adapt for various applications such specific embodiment without departing from the generic concept, and, therefore, such adaptations and modifications should and are intended to be comprehended within the meaning and range of equivalents of the disclosed embodiments. It is to be understood that the phraseology or terminology employed herein is for the purpose of description and not of limitation.
Additional aspects, features and advantages of the invention will become apparent to those skilled in the art upon consideration of the following detailed description of preferred embodiments of the invention in conjunction with the drawings listed below The present invention proposes a system and a method for querying heterogeneous data sources. The proposed system performs predictive planning via machine learning to predict and select a best proposed route for executing a query within a minimum response time bounded by benchmarks (predicted processing time).
The following detailed description of the preferred embodiments will now be described in accordance with the attached drawings, either individually or in combination.
FIGURE 1 of the accompanying drawings discloses a system (100) for querying heterogeneous data sources (1021...102n), collectively represented by reference numeral (102), and FIGURE 2 discloses a method for querying heterogeneous data sources (102).
Referring to FIGURE 1 and FIGURE 2, the system (100) comprises the following components: a plurality of data sources (102); a Central Orchestrator Unit (COU) (106); a query editor (104); a predictor module (108); a selector module (1 10); and an output module (1 12). Each of these components and their operation will be described in detail hereinafter. The data sources (102) include one or more tables that hold requisite information for performing simulation, analytics or similar complex analysis. Each of the data sources (102) have a discrete volume, discrete hardware resources and tools for facilitating processing of queries. For instance, the data source (102^ may be a Relational Database Management System (RDMS) holding small and medium sized data and data source (1022) may be a Big Data cluster.
The query editor (104) accepts queries for retrieving information from the data sources (102), step (2000) of FIGURE 2.
FIGURE 3 discloses the steps performed by a query editor (104) in accordance with this invention. Apart from accepting queries (3000), the query editor (104) preemptively corrects and completes a query while it is being typed using information from a memory module (1 16). The memory module (1 16) is communicably linked to the data sources (102) for holding metadata of the data sources (102) and any previously executed queries. The query editor (104) uses the metadata information stored in the memory module (1 16) to check syntax of a query to determine if typed column names exist within a table name chosen for a data source (102), and also suggests column names or strings while the query is being typed (3002).
In addition, the query editor (104) checks the query for different syntax as the query may be executed on any one of the available data sources (102). For instance, the query editor (104) checks the syntax for query languages/tools such as Postgres or Hive (Hadoop) or SQL parser of GPU. Further, in case any error is identified in the query, then the query editor (104) highlights appropriate portions of the query with a distinctive color (such as red color) and/or an error message to indicate occurrence of the error (3002). When the query is error-free, the query editor (104) invokes the COU (106) for execution of the query (3004). Typically, the query editor (104) interface includes an execute command button. Once the execute command button is pressed, the query editor (104) invokes the COU (106) for execution of the query.
The COU (106) receives the request from the query editor (104) and facilitates in selection of a best route for a query by selecting at least one machine learning module.
The COU (106) firstly identifies a data source (102) which includes the table name appearing in the query. Once the data source (102) has been identified, the COU (106) checks the volume of that data source (102) to determine if the volume is small, general (medium) or big. The COU (106) directly routes the query to the identified data source (102) if the volume of the data source is classified as small. The query is then executed at the data source (102) and results are displayed on the output module (1 12). If the volume is not small and the table name is appearing in all the available data sources then COU (106) forwards the query to the predictor module (108). The predictor module (108) uses the machine learning module selected by COU (106) to predict processing time for executing a query for each of the available data sources (102), step (2004) of FIGURE 2. The predictor module (108) calculates the processing time required by each of the data sources (102^, (1022) to (102n) for executing the received query. The machine learning modules are a combination of at least one regression model and one classification model to facilitate in carrying out the prediction for each route. These machine learning modules include at least one module from regression analysis, decision trees and neural networks. While performing prediction, the predictor module (108) also takes into account loading capabilities of the data sources (102) and current load and utilization of available hardware resources.
The predictor module (108) forwards the processing time predicted for each of the available data sources / routes to the selector module (1 10). The selector module (1 10) receives the 'class of service' classification information from the COU (106). The selector module (1 10) then designates a 'class of service' to the query based on the classification rules and further compares the predicted processing time for each route with the 'class of service' of the query, step (2006) of FIGURE 2. The selector module (1 10) based on result of the comparison selects a route / data source (102) having the lowest processing time for executing the query, step (2008) and (2010) of FIGURE 2.
The selector module (1 10) has a multiple connector interface (1 18) for connecting to each of the data sources (102). The selector module (1 10) via the interface (1 18) sends the query to the selected data source (102) for execution. The execution of the query is carried out using tools and techniques well known in the art.
In accordance with another embodiment of the present invention, the COU (106) facilitates in final selection of a data source (102) in the event that the selector module (1 10) is unable to find a suitable data source (102)/ route for execution of the query. To address this situation, the COU (106) simultaneously sends the query to all available data sources based on the highest class of service of the query requested. The query is then simultaneously executed at the available data sources (102). And results obtained from a data source (102) that has executed the query first are displayed on the output module (1 12).
The selected data source (102) then executes the query within the predicted processing time and returns the results to the output module (1 12) as per a required format, step (2012) of FIGURE 2.
In accordance with an aspect of the present invention, COU (106) performs additional functions to ensure accuracy of the system (100) and efficient use of resources. The functions are explained hereinafter.
The COU (106) includes a scheduler (not shown in the figures) for triggering the functions performed by it, which include: maintaining and classifying the data sources; monitoring utilization of the hardware resources; initiating a self-learning process based on new training data; and testing accuracy of the machine learning module. The scheduler carries out the aforementioned activities during off-peak hours.
I. Initiating A Self-Learning Process Based On New Training Data At each stage of processing and routing of queries, feedback and/or execution details are saved in a self-learning module (1 14). The self-learning module (1 14) is a database which stores the processing details to enable COU (106) to initiate a self-learning process and test accuracy of the machine learning modules. In the event that any query failed to execute within the predicted processing time then execution details of the query along with the predicted processing time and class of service are appended to a training table (not shown in the figures) in the self-learning module (1 14) as new training data. Referring to FIGURE 4 which discloses the steps followed by the COU (106) in initiating self-learning rules to re-train and correct conflicting data by selecting various models of machine learning for the purpose of improving accuracy of selecting the best route when queries are generated. The COU (106) fetches the new training data from the self-learning module (1 14) for initiating the self-learning process. The COU (106) selects a machine learning module to evaluate accuracy of processing the failed queries using the new machine learning module (4000). The COU (106) then re-evaluates accuracy of processing the failed queries using a different machine learning module (4002). The accuracy of machine learning modules from step (4000) and (4002) are then compared (4006) to select a machine learning module. The machine learning modules with Area of Receiver Operator Characteristic (ROC) close to 1 are chosen. The selected machine learning module is thereafter used by the predictor module (108). Testing accuracy of the machine learning module
On selection of a new machine learning module, the COU (106) regularly checks the system for any failed queries. If there is no improvement in accuracy, the previously existing machine learning module is selected for use by the predictor module (108). If after several days, accuracy further reduces, then the COU (106) randomly chooses one of the previously used machine learning modules having accuracy (>90%) and ROC closer to 1 . If accuracy still degrades (<80%) after a few days for instance, 3 days after using different models, the entry of new training data are compared for "near" conflicting attributes with almost similar attributes in the training table which produce different selection of a data source / route. Then the instances are run in real time by the COU (106) to verify and take one instance that matches the entry in the self-learning module (1 14) for training data pertaining to regression machine learning modules. For instance, if a first regression model takes lesser time than a second regression model then the first regression model is selected by the COU (106) for prediction. Thereafter, the records in the training table which have "near" conflicting attributes and use the second regression model are eliminated. In this manner the machine learning module is self-corrected. Still further, the COU (106) checks Linear Regression prediction for processing time for daily queries on each data source (102) during execution of queries and compares it with real query processing time. Error difference for each prediction is difference of time predicted with real time. If error difference is above 10%, then the selected route's training data is updated with new self-learning queries for each data source (102). The COU (106) performs the aforementioned functions for testing the accuracy of the machine learning modules to ensure that an optimum route for executing the query is selected.
III. Maintaining and classifying the data sources
The COU (106) performs daily checks on all tables of the data sources (102). The COU (106) classifies the tables based on their volume into big data, general data and small data categories. For instance, if tables exceed 100GB, then they are classified as big data and tables below 500MB as classified as small or general data.
Furthermore, the COU (106) balances the volume of the data sources (102). If data is not small, then the scheduler initiates import of data from the data sources on incremental basis such as every 10 minutes to a file.
IV. Monitoring utilization of the hardware resources The COU (106) continuously monitors real time server load for all the data sources and their hardware resources.
In addition, the COU (106) generates emails and reports for system administrators if it detects any one of the following conditions: (i) server loading issues during processing of a query, (ii) if queries appear to be taking longer time for execution than predicted, (iii) data sources are unable to meet the class of service requirements or (iv) if accuracy issues appear in routing the queries.
The technical / economical advancements offered by one or more aspects of the system (100) for querying heterogeneous data sources (102), include the realization of:
• routing of a query to an optimum data source (102) having the required resources; and
• execution of the query within a time bound by benchmarks determined by a machine learning module selected by the COU (106). The present invention thereby achieves the objectives of efficient utilization of the available resources, usage of an optimum data source to execute the query and achieve results/ output of the query within near real time constraints.
The terminology used herein is for the purpose of describing particular example embodiments only and is not intended to be limiting. As used herein, the singular forms "a", "an" and "the" may be intended to include the plural forms as well, unless the context clearly indicates otherwise. The terms "comprises," "comprising," "including," and "having," are inclusive and therefore specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. The method steps, processes, and operations described herein are not to be construed as necessarily requiring their performance in the particular order discussed or illustrated, unless specifically identified as an order of performance. It is also to be understood that additional or alternative steps may be employed.
The use of the expression "at least" or "at least one" suggests the use of one or more elements, as the use may be in one of the embodiments to achieve one or more of the desired objects or results.

Claims

CLAIMS:
1 . A system (100) for querying heterogeneous data sources (102), comprising:
i. a plurality of data sources (102), each data source having a discrete volume, discrete hardware resources and tools for facilitating processing of queries;
ii. a central orchestrator unit (106) for monitoring the data sources (102) and the hardware resources and to select at least one machine learning module for processing of queries;
iii. a predictor module (108) to receive at least one query for retrieving information from at least one of said data sources (102); and
iv. a selector module (1 10) having a multiple connector interface (1 18) for connecting with each of said data sources (102), wherein said predictor module (108) predicts processing time for executing a query using each of the data sources (102) based on said at least one selected machine learning module, said selector module (1 10) compares the predicted processing time for each data source with a class of service of the query to route the query to a data source (102) via the interface (1 18) for execution of the query within the predicted processing time for that data source (102).
2. A system (100) as claimed in claim 1 , wherein said central orchestrator unit (106) includes a scheduler for (i) maintaining and classifying the data sources; (ii) monitoring utilization of the hardware resources; (iii) initiating a self-learning process based on new training data; and (iv) testing accuracy of the machine learning module.
3. A system (100) as claimed in claim 1 , wherein said machine learning module is selected from the group consisting of regression analysis, decision trees and neural networks.
4. A system (100) as claimed in claims 1 , wherein said selector module (1 10) routes the query to a data source (102) having a lowest processing time based on the comparison of predicted processing time for each data source with a class of service of the query.
A system (100) as claimed in claim 1 , the system (100) includes a query editor (104) to receive at least one query for retrieving information from at least one of said data sources (102).
A system (100) as claimed in claim 5, wherein said query editor (104) co-operates with a memory module (1 16) to preemptively auto-complete and auto-correct the query.
A system (100) as claimed in claim 5, wherein said query editor (104) co-operates with the central orchestrator unit (106) to initiate routing of the query to at least one data source in the event said query editor (104) receives instructions to execute the query.
A system (100) as claimed in claim 6, wherein said memory module (1 16) is communicably linked to the data sources (102) for holding metadata of the data sources (102) and any previously executed queries.
A method for querying heterogeneous data sources, wherein each data source (102) having a discrete volume, discrete hardware resources and tools for facilitating processing of queries, the method comprising the following steps:
i. receiving at least one query for retrieving information from the data sources (102);
ii. selecting at least one machine learning module for routing a query to an appropriate data source (102);
iii. predicting processing time for executing the query for each of the data sources (102) based on the machine learning module;
iv. comparing the processing time for each of the data sources (102) with a class of service of the query;
v. selecting a data source (102) having a lowest processing time based on result of comparison obtained from step (iv); and vi. executing the query at the selected data source (102) within the predicted processing time for the selected data source (102).
10. A method as claimed in claim 9, wherein the step of receiving at least one query for retrieving information from said data sources (102) includes:
i. receiving the query via a query editor (104) interface;
ii. checking the query while it is being typed in the query editor (104) (a) to determine if typed column names exist within a table name chosen for an data source (102), and (b) to suggest column names or strings while the query is being typed;
iii. checking the query language syntax and detecting and highlighting appropriate portions of the query in the event an error is detected in (a) or (b) of step (ii); and
iv. invoking a central orchestrator unit (106) on receiving instructions for executing the query.
1 1. A method as claimed in claim 9, wherein the step of executing the query at the selected data source (102) within the predicted processing time further includes adding execution details of the query along with the predicted processing time and class of service in a self-learning module (1 14) as a training data if the query failed to execute within the predicted processing time.
12. A method as claimed in claim 9, wherein the method further includes the steps of: i. scheduling checks on all tables of the data sources (102);
ii. classifying the tables based on their volume into big data, general data and small data categories;
iii. balancing the volume of the data sources (102); and
iv. monitor real-time load on the hardware resources associated with the data sources (102).
13. A method as claimed in claim 10, wherein prior to the step of invoking a central orchestrator unit (106) the method includes determining a data source which includes the table name appearing in the query; checking the volume of that data source; routing the query to that data source and skipping steps (iii) to (v) of Claim 9 if the volume of the data source is classified as small.
14. A method as claimed in claims 1 1 , wherein prior to the step of selecting at least one machine learning module for routing a query, the method further includes the steps of:
i. using the training data in the self-learning module (1 14) to evaluate accuracy of processing failed queries using a new machine learning module; ii. re-evaluating accuracy of processing the failed queries using another machine learning module; and
iii. comparing accuracy of machine learning model selected in step (i) and step (ii) to select a machine learning module for routing a query.
15. A computer program product for querying heterogeneous data sources (102), wherein each data source (102) having a discrete volume, discrete hardware resources and tools for facilitating processing of queries, said computer program product having instructions operable to cause one or more modules to perform the following steps:
i. receiving at least one query for retrieving information from the data sources
(102);
ii. selecting at least one machine learning module for routing a query to an appropriate data source (102);
iii. predicting processing time for executing the query for each of the data sources (102) based on the machine learning module;
iv. comparing the processing time for each of the data sources (102) with a class of service of the query;
v. selecting a data source (102) having a lowest processing time based on result of comparison obtained from step (iv); and
vi. executing the query at the selected data source (102) within the predicted processing time for the selected data source (102).
16. A computer program product for querying heterogeneous data sources (102), wherein each data source (102) having a discrete volume, discrete hardware resources and tools for facilitating processing of queries, said computer program product being tangibly implemented on a machine readable media and comprising: i. a query editor (104) to receive at least one query for retrieving information from the data sources (102);
ii. a central orchestrator unit (106) for monitoring the data sources (102) and the hardware resources and to select at least one machine learning module for processing of queries;
iii. a predictor module (108) to receive at least one query from the query editor (104);
iv. a selector module (1 10) having a multiple connector interface for connecting with each of said data sources (102); and
v. an output module (1 12) cooperating with the data sources (102), wherein said predictor module (108) predicts processing time for executing a query using each of the data sources (102) based on said at least one selected machine learning module, said selector module (1 10) compares the predicted processing time for each data source with a class of service of the query to route the query to a data source (102) via the interface (1 18) for execution of said query within the predicted processing time for that data source (102) and display of the query results on the output module (1 12).
PCT/IB2015/055343 2014-07-23 2015-07-15 A system for querying heterogeneous data sources and a method thereof WO2016012903A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
MYPI2014002164 2014-07-23
MYPI2014002164A MY186962A (en) 2014-07-23 2014-07-23 A system for querying heterogeneous data sources and a method thereof

Publications (1)

Publication Number Publication Date
WO2016012903A1 true WO2016012903A1 (en) 2016-01-28

Family

ID=53724407

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2015/055343 WO2016012903A1 (en) 2014-07-23 2015-07-15 A system for querying heterogeneous data sources and a method thereof

Country Status (2)

Country Link
MY (1) MY186962A (en)
WO (1) WO2016012903A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101828522B1 (en) 2016-02-16 2018-02-12 전북대학교산학협력단 System of Parallel Distributed Processing System for Heterogeneous Data Processing

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060195416A1 (en) * 2005-02-28 2006-08-31 Ewen Stephan E Method and system for providing a learning optimizer for federated database systems
WO2009097438A2 (en) * 2008-01-29 2009-08-06 Hewlett-Packard Development Company, L. P. Query deployment plan for a distributed shared stream processing system
WO2012112980A2 (en) * 2011-02-18 2012-08-23 Microsoft Corporation Dynamic distributed query execution over heterogeneous sources

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060195416A1 (en) * 2005-02-28 2006-08-31 Ewen Stephan E Method and system for providing a learning optimizer for federated database systems
WO2009097438A2 (en) * 2008-01-29 2009-08-06 Hewlett-Packard Development Company, L. P. Query deployment plan for a distributed shared stream processing system
WO2012112980A2 (en) * 2011-02-18 2012-08-23 Microsoft Corporation Dynamic distributed query execution over heterogeneous sources

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101828522B1 (en) 2016-02-16 2018-02-12 전북대학교산학협력단 System of Parallel Distributed Processing System for Heterogeneous Data Processing

Also Published As

Publication number Publication date
MY186962A (en) 2021-08-26

Similar Documents

Publication Publication Date Title
US20200183796A1 (en) Recovery strategy for a stream processing system
US11086688B2 (en) Managing resource allocation in a stream processing framework
US10983895B2 (en) System and method for data application performance management
US9965330B2 (en) Maintaining throughput of a stream processing framework while increasing processing load
US10943009B2 (en) System and method to infer investigation steps for security alerts using crowd sourcing
Yang et al. A system architecture for manufacturing process analysis based on big data and process mining techniques
US20210035026A1 (en) Diagnosing &amp; triaging performance issues in large-scale services
US20170024433A1 (en) Query plan post optimization analysis and reoptimization
US20200013070A1 (en) Identifying clusters for service management operations
US10740336B2 (en) Computerized methods and systems for grouping data using data streams
US9141908B1 (en) Dynamic information assembly for designated purpose based on suitability reasoning over metadata
US10311364B2 (en) Predictive intelligence for service and support
US11693371B2 (en) Potential replacement algorithm selection based on algorithm execution context information
US20210303532A1 (en) Streamlined transaction and dimension data collection
US11567936B1 (en) Platform agnostic query acceleration
US20210304073A1 (en) Method and system for developing a machine learning model
US20220121981A1 (en) Cognitive Error Recommendation Based on Log Data
US11775867B1 (en) System and methods for evaluating machine learning models
US10896176B1 (en) Machine learning based query optimization for federated databases
CN111582488A (en) Event deduction method and device
US10250715B2 (en) Dynamic adjustment of boxcarring of action requests from component-driven cloud applications
WO2016012903A1 (en) A system for querying heterogeneous data sources and a method thereof
US10250716B2 (en) Priority-driven boxcarring of action requests from component-driven cloud applications
US20190065987A1 (en) Capturing knowledge coverage of machine learning models
US11809865B2 (en) Method and system for evidence servicing

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15742108

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15742108

Country of ref document: EP

Kind code of ref document: A1