WO1995021407A2 - Improved method and apparatus for data access in multiprocessor digital data processing systems - Google Patents

Improved method and apparatus for data access in multiprocessor digital data processing systems Download PDF

Info

Publication number
WO1995021407A2
WO1995021407A2 PCT/US1995/001356 US9501356W WO9521407A2 WO 1995021407 A2 WO1995021407 A2 WO 1995021407A2 US 9501356 W US9501356 W US 9501356W WO 9521407 A2 WO9521407 A2 WO 9521407A2
Authority
WO
WIPO (PCT)
Prior art keywords
signals
signal
subquery
representative
database table
Prior art date
Application number
PCT/US1995/001356
Other languages
English (en)
French (fr)
Other versions
WO1995021407A3 (en
Inventor
David Reiner
Jeffrey M. Miller
David C. Wheat
Original Assignee
Sun Microsystems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Microsystems, Inc. filed Critical Sun Microsystems, Inc.
Priority to JP7520740A priority Critical patent/JPH09511347A/ja
Priority to EP95911601A priority patent/EP0753176A4/en
Publication of WO1995021407A2 publication Critical patent/WO1995021407A2/en
Publication of WO1995021407A3 publication Critical patent/WO1995021407A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2453Query optimisation
    • G06F16/24532Query optimisation of parallel queries
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/964Database arrangement
    • Y10S707/966Distributed
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99932Access augmentation or optimizing
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99931Database or file accessing
    • Y10S707/99933Query processing, i.e. searching
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99941Database schema or data structure
    • Y10S707/99943Generating database or data structure, e.g. via user interface
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users

Definitions

  • This invention relates to digital data processing and. more particularly, to methods and apparatus for database management systems on multiprocessor digital data processing systems.
  • One or more indexes on large tables are generally provided to facilitate the most common data accesses, e.g., look-ups based on employee name.
  • corresponding rows in two or more tables are identified by matching data values in one or more columns.
  • the department name corresponding to a given employee may be identified by matching his or her department number to a row in a department data table that gives department numbers and department names.
  • This is in contrast to hierarchical, network, and other DBMS's that use pointers instead of data values to indicate corresponding rows when tables are combined, or "joined.”
  • Relational DBMS's typically permit the operator to access information in the database via a query. This is a command that specifies which data fields (columns) are to be retrieved from a database table and which records (rows) those fields are to be selected from.
  • a query for the names of all employees in department 10 might be fashioned as follows:
  • a particular relational data table need not be stored in a single computer file but, rather, can be partitioned among many files. This makes such tables particularly suited for use on multiprocessor computer systems, i.e., computer systems having multiple processors and multiple disk drives (or other storage devices) of the type disclosed in U.S. Patent 5,055,999.
  • prior art DBMS's have not proven capable of taking full advantage of the power of such multiprocessing systems and, particularly, their power to simultaneously process data (in parallel) from multiple partitions on multiple storage devices with multiple central processing units.
  • an object of the invention is to provide improved methods and apparatus for database management and, particularly, improved methods and apparatus for data base management capable of operating on multiprocessor systems.
  • a further object of the invention is to provide improved systems for database management capable of effectively accessing a relational database contained in multiple tables and multiple partitions.
  • a still further object is to provide improved methods and apparatus for storing and retrieving data for access by a DBMS.
  • DBMS database management system
  • the improvement is characterized by a parallel interface that intercepts selected queries prior to substantive processing by the standard interface.
  • the standard interface is often called the "server" interface; it is accessed by clients that are the source of queries.
  • a decomposition element within the parallel interface generates multiple subqueries from the intercepted query. Those subqueries, each representing a request for access to data stored in a respective partition of the table, are applied in parallel to the standard interface in lieu of the intercepted query.
  • Responses by the DBMS to the subqueries are reassembled to generate a final response representing the response the DBMS would have generated to the intercepted query signal itself.
  • Such reassembly can include interleaving the data contained in the responses (e.g., to create a single sorted list) or applying an aggregate function (e.g., sum or average) to that data.
  • the decomposition element generates the subqueries to be substantially identical to the intercepted signal but including an "intersecting predicate" (i.e., additional query conditions) that evaluates true for all data records in respective partitions of said database table and ⁇ /-?e for all others.
  • This can be, for example, a logically AND'ed condition that evaluates true for records in the respective partition.
  • a subquery for the first partition could be generated as follows (where rowid has three parts, the last of which indicates the partition number):
  • the invention contemplates a further improvement to a digital data processing system of the type described above, wherein the DBMS responds to selected queries for accessing data records joined from one or more of database tables, and wherein the DBMS includes an optimizer for determining an optimal strategy for applying such queries to the tables.
  • the improvement of this aspect is characterized by an element for identifying, from output of the optimizer, a driving table whose partitions will be targeted by subqueries generated in responding to an intercepted query.
  • the improvement is further characterized by generating the subqueries to include, in addition to the predicate list of the intercepted query, an intersecting predicate for all data records in respective partitions of the driving database table.
  • tables referenced in the query other than the driving table need not be identically partitioned to the driving table, nor co-located with its partitions on storage devices.
  • Tables may be accessed through either full- table scans or indexed scans, i.e., whether the DBMS searches all blocks of the relevant partition or only those indicated by a relevant index.
  • the invention provides an improvement to a digital data processing system of the type described, wherein the DBMS's standard interface is invoked by a procedure or function call.
  • the improvement is characterized by functionality for invoking the parallel interface in lieu of the client-side portion of the standard interface in response to such a procedure/function call. .And, by responding to a query for generating plural subqueries in the form of further procedures/functions to the standard server interface.
  • the parallel interface can form part of an object code library for linking with a computer program including procedures/function calls for invoking the DBMS.
  • the invention contemplates an improvement to a digital data processing system as described above, wherein the standard interface normally responds to insert/select queries by placing requested data from the database table means in a further database table (i.e., as opposed to merely printing the requested data or otherwise outputting it in text form or merely returning the data to the requesting program).
  • the improvement of this aspect is characterized by generating the plural subqueries so as to cause the DBMS to place the data requested from each respective partition in the designated database table.
  • a digital data processing system as described above can include functionality for executing multiple threads, or "lightweight processes," each for applying a respective subquery signal to the DBMS's interface element. Those threads can be executed in parallel on multiple central processing units, and can be serviced by multiple server processes within the DBMS that also execute in parallel.
  • a digital data processing system of the type having a storage element (e.g., a disk drive or other random-access media 1 for storing and retrieving data records, as well as a DBMS having (i) a hashing element to effect storage of data records in "hash bucket' ' regions in the storage element, where each record is stored in a root hash bucket region corresponding to a hash function of a selected value of the data record or, alternatively, to effect storage of data records in an overflow hash bucket region associated with that root hash bucket region; and (2) an indexing element to index each stored data record for direct access in accord with a respective value of that data record.
  • a storage element e.g., a disk drive or other random-access media 1 for storing and retrieving data records
  • DBMS having (i) a hashing element to effect storage of data records in "hash bucket' ' regions in the storage element, where each record is stored in a root hash bucket region corresponding to a hash function of a selected
  • the improvement is characterized by a scatter cluster retrieval element that responds to a request for accessing a data record previously stored via the hashing element, by invoking the indexing element to retrieve that record in accord with the index value thereof, where stored records have previously been indexed by the indexing element with respect to the same fields (columns) used by the hashing element.
  • the hashing element stores the data records in hash bucket regions that are sized so as to create at least one overflow hash bucket region per root bucket region, and such that overflow bucket regions for a given root bucket region are distributed roughly evenly across different storage partitions.
  • Another aspect of the invention provides a digital data processing system of the type described above, in which plural subcursor buffers are associated with each subquery signal for storing results generated by the DBMS's standard interface means in response to that subquery signal.
  • a root buffer stores a then- current result, while a fetching element simultaneously assembles a final result signal based upon those results currently stored in selected subcursor buffers.
  • results are taken from each of those buffers, they are emptied.
  • a subquery is applied to the standard interface asynchronously with respect to demand for that buffer's contents in assembling the final result.
  • the root buffer stores then-current results in a temporary table to be queried later by an aggregate query generated by the decomposition element.
  • the invention provides a method for digital data processing paralleling the operation of the digital data processing system described above; i.e., "transparent" to the DBMS client other than by improved performance.
  • Figure 1 depicts a preferred multiprocessing system used to practice the invention.
  • FIG. 1 illustrates in greater detail processing cells and their interconnection within the processing system of Figure 1.
  • Figure 3 A depicts a standard arrangement of processes and software modules utilized in digital data processor 10 without query decomposition and data access according to the invention.
  • Figure 3B depicts a preferred arrangement of threads, processes and software modules utilized in digital data processor 10 for query decomposition and data access according to the invention.
  • Figure 4 shows the operation of assembler 74B on results generated by the DBMS 76 and threads 78A, 78B, 78C in response to the subquery signals.
  • Figure 5 depicts a preferred mechanism, referred to as “scatter clustering,” for storing and retrieving data from database 72.
  • Figures UM 9-1 and UM 9-2 are used in connection with the discussion of the operation and use of a preferred query decomposition system according to the invention.
  • Figures 26 - 1 through 26 - 3 are used in connection with the discussion of design provided in Database Note #26.
  • Figures 61 - 1 through 61 - 3 are used in connection with the discussion of query decomposition for applications running on client workstations in Database Note #61.
  • Figures 32 - 1 through 32 - 3 are used in connection with the discussion of the framework of rules for automating query decomposition in Database Note #32.
  • Figures 36 - 1 through 36 - 7 are used in connection with the discussion of parallel cursor building blocks in Database Note #36.
  • Figures 37 - 1 and 37 - 2 are used in connection with the discussion of parse tree requirements for query decomposition in Database Note #37.
  • Figures 41 - 1 and 41 - 2 are used in connection with the discussion of query decomposition control structures in Database Notes #41.
  • Figures 42 - 1 through 42 - 3 are used in connection with the discussion of upper tree parallelism in parallel cursors in Database Note #42.
  • FIG. 1 depicts a preferred multiprocessing system used to practice the invention.
  • the illustrated system 10 includes three information transfer levels: leve O, level: 1, and level:2.
  • Each information transfer level includes one or more level segments, characterized by a bus element and a plurality of interface elements.
  • levehO of the illustrated system 10 includes six segments, designated 12A, 12B, 12C, 12D, 12E and 12F, respectively.
  • level: 1 includes segments 14A and 14B
  • level:2 includes segment 16.
  • Each segment of levehO i.e., segments 12A, 12B, ... 12F, comprise a plurality of processing cells.
  • segment 12A includes cells 18 A, 18B and 18C;
  • segment 12B includes cells 18D, 18E and 18F; and so forth.
  • Each of those cells include a central processing unit and a memory element, interconnected along an intracellular processor bus (not shown).
  • the memory element contained in each cells stores all control and data signals used by its associated central processing unit.
  • Certain cells of the processing system 10 are connected to secondary storage devices.
  • cell 18C is coupled with disk drive 19A
  • cell 18D is coupled with disk drive 19B
  • cell 180 is coupled with disk drive 19C.
  • the disk drives 19A - 19C are of conventional design and can be selected from any of several commercially available devices. It will be appreciated that secondary storage devices other than disk drives, e.g., tape drives, can also be used to store information.
  • FIG. 2 illustrates in greater detail processing cells and their interconnection within the processing system of Figure 1.
  • plural central processing units 40 A, 40B and 40C are coupled, respectively, to associated memory elements 42A, 42B and 42C. Communications between the processing and memory units of each pair are carried along buses 44A, 44B and 44C, as shown.
  • Network 46 representing the aforementioned level segments and routing cells, transfers information packets (passed to the network 46 over buses 48A, 48B and 48C) between the illustrated processing cells 42A - 42C.
  • the central processing units 40A, 40B and 40C each include an access request element, labeled 50A, 50B and 50C, respectively.
  • These access request elements generate requests for access to data stored in the memory elements 42 A, 42B and 42C.
  • access requests signals generated by elements 50A, 50B and 50C is the ownership-request, representing a request for exclusive, modification access to a datum stored in the memory elements.
  • access request elements 50A, 50B and 50C comprise a subset of an instruction set implemented on CPU's 40A, 40B and 40C. This instruction subset is described below.
  • the central processing units 40A, 40B, 40C operate under control of an operating system 51 , portions 51 A, 51 B and 51 C of which are resident on respective ones of the central processing units.
  • the operating system 51 provides an interface between applications programs executing on the central processing units and the system 10 facilities, and includes a virtual memory management system for managing data accesses and allocations.
  • a preferred operating system for controlling central processing units 40A, 40B and 40C is a UNIX-like operating system and, more preferably, OSF/1, modified in accord with the teachings herein.
  • the memory elements 40A, 40B and 40C include cache control units 52A, 52B and 52C. respectively. Each of these cache control units interfaces a data storage area 54A. 54B and 54C via a corresponding directory element 56A, 56B and 56C, as shown. Stores 54A, 54B and 54C are utilized by the illustrated system to provide physical storage space for data and instruction signals needed by their respective central processing units.
  • Figure 3 A depicts a standard arrangement of processes and software modules utilized in digital data processor 10 without query decomposition and data access according to the invention.
  • Figure 3B depicts a preferred arrangement of processes and software modules utilized in digital data processor 10 for query decomposition and data access according to the invention.
  • An. initiating process 70 generates a query for accessing data stored in relational database 72 having data partitions 72A, 72B, 72C.
  • the query is generated in a conventional format otherwise intended for a conventional DBMS 76.
  • that conventional format is SQL
  • that conventional DBMS is the ORACLE7 Database Management System (hereinafter, "ORACLE” or "ORACLE Version 7”) of Oracle Corporation.
  • ORACLE ORACLE7 Database Management System
  • DBMS such as ORACLE Version 7
  • ORACLE Version 7 DBMS
  • Element 74A (responsible for decomposing the query) routes queries not susceptible to decomposition to DBMS 76, but for a decomposable query it generates a set of subqueries, each of which is based on the initial query but which is directed to data in one or more respective of the partitions 72A, 72B, 72C of database 72. Then element 74A initiates and invokes threads 78A, 78B, 78C, which initiate execution of the subqueries.
  • PUPI parallel user program interface
  • the subqueries corresponding to threads 78A, 78B, 78C are routed to the user program interface ("UPI" or "standard interface") of DBMS 76 (in lieu of the intercepted query), as shown in the drawing.
  • UPI user program interface
  • Multiple subqueries are preferably applied to the UPI of DBMS 76 in parallel with one another, thus capitalizing on the database partitions and on the multiprocessing nature of the preferred digital data processing system 10.
  • Each thread routes its subquery to a separate server process in DBMS 76.
  • the DBMS 76 responds in the conventional manner to each subquery by generating appropriate requests (e.g., a disk read) for access to the database 73 and, particularly, for access to respective partitions of that database (unless the data requested is already in memory).
  • Data retrieved from the database 72 in response to each subquery is processed in the normal manner by DBMS 76 and is routed to processes 76A, 76D and 76G.
  • Those responses are routed to parallel interface assembly section 74B which assembles a response like that which would have been generated by the DBMS 76 had the intercepted response been applied directly to it.
  • the assembled response produced by assembly section 74B is generally returned to the initiating process 70 more quickly than that which would have been generated by the DBMS 76 had the intercepted query been applied directly to it. This is a consequence of decomposition of the intercepted query and its parallel application to the UPI of DBMS 76. It is also a consequence of the architecture of the underlying multiprocessor, which permits multiple server processes to run simultaneously. Though it will be appreciated that, even when running on a uniprocessor, the concurrent execution of multiple subqueries could speed access where there is overlapping I/O and CPU processing.
  • the decomposer 74A generates subqueries based on the conventional- format query intercepted from the initiating process. For simple, single-table queries, the decomposer 74A generates corresponding subqueries by duplicating the query and appending a predicate for matching records in the corresponding table partition.
  • rowid has three parts, the last of which indicates the partition number.
  • Other subqueries would be of similar form, with changes to the partition numbers referenced in the rowid predicates.
  • the decomposer 74A For queries joining two or more tables, the decomposer 74A generates corresponding subqueries by duplicating the query and appending a predicate for matching records in the corresponding table partition of the driving table, which is selected by the decomposer 74A based on the access strategy chosen by the query optimizer portion 76B of the DBMS 76.
  • information from the optimizer 76B including possible tables to be chosen as the driving table, can be obtained from data files generated by the DBMS 76 in connection with the query, and accessed by use of the "EXPL.A-IN" command.
  • Figure 4 shows the operation of assembler 74B on results generated by the UPI of DBMS 76 and threads 78A, 78B, 78C in response to the subquery signals. More particularly, the drawing shows that for intercepted queries that call for aggregate data functions, element 74C performs a like or related data function of the results of the subqueries. Thus, for example, if the intercepted query seeks a minimum data value from the database table ⁇ and, likewise, the subqueries seek the same minimum value from their respective partitions — then element 74C generates a final result signal representing the minimum among those reported to the assembler 74B by the DBMS 76 and threads 78A, 78B, 78C.
  • element 74C generates an average table value through a weighted average of the reported subquery results.
  • the decomposer 74A generates subqueries requesting related functions of the data, e.g., the sum, count and sum of the squares of the data.
  • Such aggregate processing is preferably applied to, for example, intercepted queries requesting (i) a minimum or maximum of an item in the records (ii) an average of selected items, (iii) a standard deviation and variance of selected items, and (iv) a sum and a count of selected items.
  • element 74D For intercepted queries that call for non-aggregate data functions, element 74D generates a final result signal by interleaving the results of the subqueries.
  • element 74D generates a final result signal by interleaving (in the specified sort order) the items presented in the results reported to the assembler 74B by the DBMS 76 and threads 78A, 78B, 78C.
  • Other non-aggregate queries involving, for example, (i) a distinct value of an entire result row, (ii) a nested selection of items, and/or (iii) a correlated selection of items are processed accordingly.
  • the decomposer 74A For queries involving grouping operations, the decomposer 74A generates corresponding subqueries by duplicating the query, along with the grouping clause in its predicate list. For each group, data retrieved by the DBMS in response to those subqueries is placed in a temporary table. For that group, the assembly section 74B generates and passes to the DBMS a "group by" combining query to be applied to the temporary table. The results of those queries are returned to the initiating process 70 in lieu of the response that would have been generated by the DBMS 76 had the intercepted query been applied directly to it.
  • the decomposer 74A and assembly section 74B operate in the manner describe above, except, that the "having" clause is not included in the subqueries. That clause is, however, incorporated into the combining queries that are executed on the temporary table.
  • Figure 5 depicts a preferred mechanism, referred to as “scatter clustering” or “small bucket hashing,” for storing and retrieving data from database 72.
  • the mechanism combines cluster-storage and index-access techniques to disperse and retrieve data records from storage media 80A, 80B, 80C (e.g., disk drives) upon which database 72 is contained.
  • Data records are stored using the DBMS's 76 cluster-storing capabilities, based on a conventional hash function of its key value (as generated by element 76B), and using a smaller-than-normal bucket size chosen to insure that at least one overflow hash bucket will be created for each root bucket. More preferably, the bucket size is chosen to insure that hash buckets are spread over storage devices to maximize the potential for parallel access.
  • Each stored record is simultaneously indexed for direct access in accord with the same key value(s) used by the hash function.
  • the DBMS 76 responds to requests to store data records by invoking the hashing element 76B to store those data records in accord with a hash on their key values.
  • the DBMS 76 also populates index 76C by invoking DBMS's 76 corresponding indexing functionality.
  • the decomposer 74A When accessing data records, the decomposer 74A generates subqueries specifying that requested data records are to be accessed via the index element 76c, not the hashing element 76b.
  • the database 72 is organized to achieve the best mix of I/O parallelism and hit ratio.
  • I/O parallelism the more threads 78A, 78B, 78C can be used, in parallel, to initiate data retrievals.
  • hit ratio the greater the number of relevant records each thread 78A, 78B, 78C gets with each retrieval.
  • the aforementioned scatter clustering technique achieves a good mix of I/O parallelism and hit ratio. It does this by storing the data records using the DBMS's 76 hash- based storage techniques with abnormally small bucket size, thereby distributing small bucket-size clusters of related information around the disk, and by retrieving the data using the DBMS's indexing mechanism.
  • the invention contemplates operating on database tables with any plurality of partitions. .And, that the invention contemplates using any plurality of subqueries (and corresponding threads) to execute retrievals against those partitions. Moreover, it will be appreciated that the invention does not require that the number of partitions and subqueries be identical. Preferably, the number of subqueries (and threads) is an integral divisor, greater than one. of the number of partitions. Thus, for example, three subqueries can be beneficially run against six partitions.
  • ORACLE database management system (which is commercially available from Oracle Corporation and can be adapted for operation with a number of computer systems, including the Kendall Square Research Corporation multiprocessors). Each of the sections which follow is identified by a “Database Note Number " (or DBN #). Those identifications are used to cross-reference the sections, typically, in lieu of their titles. The inventors are alternatively referred to as “we, " “ “I, " “KSR, " and other like terms.
  • KSR QD The KSR Query Decomposer
  • Fig. UM 9-1 shows the basic steps of query decomposition:
  • KSR QD intercepts it and generates subqueries against existing data partitions on disks.
  • KSR QD is compatible with the ORACLE architecture (transparent to your applications), while at the same time leveraging ORACLE'S existing optimization strategies.
  • Fig. UM 9-2 is a conceptual view of the ORACLE program interface.
  • the UPI User Program Interface
  • the UPI is the common point of access to the ORACLE kernel for all applications.
  • the KSR QD implementation sits between the UPI external interface and the UPI program library, transforming a serial interface into a parallel one.
  • KSR QD Decisions made automatically by KSR QD include the following:
  • KSR QD will not notify you whether your query was decomposed, because everything is meant to happen transparently—unless you want it to be otherwise. If you want more details, you should issue an EXPLAIN PLAN command to see the actual execution plan for your query. If it is being decomposed, a row of the execution plan will be labeled KSR PARALLEL EXECUTION. For further information, see Section 11.4.
  • End users do not need to be aware of KSR Query Decomposer activities.
  • the database administrator must set up a default environment that enhances the performance of queries normally issued in your application environment.
  • the DBA's most impo ⁇ ant step is to distribute data in a way to take advantage of the parallelism provided by KSR QD.
  • KSR QD parallelizes a query by dividing it into subqueries, each of which uses a rowid range predicate on the driving table to specify one or more files to which that query's reads will be restricted.
  • the approach depends on partitioning tables across files on multiple disk drives, so the files can be read in parallel.
  • the DBA partitions large database tables over multiple disks to maximize parallel reads from disk. There may be tens or even hundreds of partitions for a given table. This is described in Chapter 10.
  • the only change in the subqueries is the addition of the rowid range predicates.
  • the first subquery will read only the blocks of the EMP table which are in file 1 : the second, file 2; the third, file 3. This is an example of decomposing a full-table scan.
  • the overall query needs to read all blocks of the table, and you gain near-linear speedup by reading the separate files across which the table is partitioned in parallel. The total number of reads is not changed, but they occur in parallel.
  • Query decomposition also can work with queries that use an index. Suppose you have the following query:
  • the only change in the subqueries is the addition of rowid range predicates.
  • the subqueries read the index blocks concurrently and process them in parallel. Index blocks are cached, so the index blocks should be read in from disk only once.
  • a subquery finds an index entry for DEPTNO 5, however, it will examine the rowid stored in that index entry to see whether it falls within the range for that subquery. Only if it does will that subquery read the data page containing the row with that DEPTNO value and rowid.
  • ORACLE The component of ORACLE whose goal is to choose the most efficient way to execute an SQL statement.
  • a table must be partitioned to be used as a partitioning table by KSR QD.
  • the view definition must not contain a GROUP BY clause, an aggregate function, or any construct not listed above.
  • KSR QD will not alert you as to whether decomposition has taken place. You will receive the results of your query without any interrupting messages. KSR QD always runs transparently to your operation.
  • KSR QD supports the following products in a local client/server environment:
  • KSR QD is a separate software component that interacts seamlessly with ORACLE for KSR. It is developed and distributed by Kendall Square for use on KSR/Series computers. KSR QD is not expected to impact general database administration procedures.
  • ORACLE installation includes KSR QD components and KSR QD files in the correct directories.
  • Section 10.1 "General KSR QD Start-up Information" - This section provides information for performing a post-installation assessment to make certain all KSR QD files are in place. It provides a procedure to establish the views and tables necessary for KSR QD operation.
  • libora.qd.a This is a customized libora.a file containing all the KSR QD code. There are KSR QD modules and modified UPI modules in this copy of the library. It is located in $ORACLE_HOME/lib. sqlplus.qd This is SQL*Plus built with KSR QD linked in. It is located in $ORACLE_HOME/bin.
  • KSR QD Product Overview describes how KSR QD is related to the ORACLE UPI.
  • the programs comprising the UPI normally are found in the libora.a library.
  • the libora.qd.a file in your distribution is the modified ORACLE libora.a file containing KSR QD code.
  • the sqlplus.qd executable you received was linked with this modified libora.qd.a file. . Any new applications automatically will include KSR QD when they are built with libora.qd.a.
  • KSR_.ALL_TABLE_FILES This view must exist in a database for query decomposition to be possible. It permits KSR QD to get a count and list of the files into which a given table is partitioned. It is owned by SYS and must have SELECT privilege granted to public (or. at the DBA's discretion, only to those users permitted to use KSR QD). It is created, with appropriate grants, by the SQL script
  • KSR_PLAN_TABLE This is the default plan table for the temporary execution plans generated by KSR QD when it calls EXPL.AIN PLAN. This plan table has the same format as a standard ORACLE plan table, is owned by SYS. must have SELECT.
  • Standard ORACLE includes a script called utlxplan.sql, which any user can employ to create a private plan table (named PLAN_TABLE by default). For a given user. KSR QD will use PLAN_TABLE if it exists: otherwise, it will use KSR_PLAN_TABLE. If KSR_PLAN_TABLE does not exist in a given database, only users with a private plan table explicitly named PLAN_TABLE can use KSR QD in that database.
  • KSR_DISABLE_QD This is a public synonym for the table DUAL: which is transparently created when the script $ORACLE_HOME/rdbms/admin/ksr_disable_gd.sgl is run. It is transparently dropped when you subsequently run the script
  • KSR QD is enabled for that database.
  • KSR QD creates a temporary table when it decomposes a query containing aggregate functions. It uses this table to gather partial aggregate results from the parallel subqueries.
  • a KSR QD temporary table has a name beginning with QDIT (for Query Decomposer Intermediate Table) with a unique 9-digit suffix. It is owned by the user executing the query and is created in that user's TEMPORARY tablespace (which defaults to SYSTEM).
  • a user must have the ability to create tables in his or her TEMPORARY tablespace to apply query decomposition to queries containing aggregate functions.
  • the ALTER USER statement can be used to assign a TEMPORARY tablespace for a particular user. The user
  • RULE 26 can be enabled to create tables in that tablespace by either granting the RESOURCE role to the user (which enables creating tables in any tablespace) or using the QUOTA clause of the ALTER USER statement to grant the user a storage quota in a specified tablespace. If a quota is used, it must be sufficient to permit creating a table with default storage attributes. The minimum necessary quota varies depending on the database configuration (e.g., in a database with DB_BLOCK_SIZE of 8192. the minimum quota required to create KSR QD intermediate tables is 40 KB).
  • Each file of the tablespace constitutes a KSR QD "partition" for each table.
  • ORACLE extents are automatically allocated in a balanced manner across the tablespace as data is loaded.
  • a database may contain multiple partitioned tablespaces. each of which may have a different number of partitions.
  • a table created in a partitioned tablespace potentially has the same number of partitions as the tablespace.
  • KSR QD will consider as partitions only those files which contain at least one extent of the table in question. For example, if a tablespace has 20 files, but a particular table in the tablespace has extents in only 10 of those files, KSR QD considers that table to have 10 partitions, not 20. This means KSR QD generates at most 10 parallel subqueries for a query with this table as its partitioning table.
  • the maximum effective degree of partitioning of a tablespace is limited by the number of disk drives, because KSR QD does not benefit significantly from using more than one partition on the same disk, for a given tablespace.
  • the number of processors does not impose a hard limit on the degree of partitioning, because multiple KSR QD threads may run on a given processor in a time-sharing fashion.
  • a gradual decline of added speedup per additional degree of partitioning will occur as the degree of partitioning exceeds the number of processors. It is recommended the degree of partitioning not exceed the number of processors for the initial run.
  • the degree of partitioning of a tablespace determines the maximum potential degree of parallelism KSR QD can use for queries whose driving table is in that tablespace.
  • the actual maximum degree of parallelism is equal to the number of files in the tablespace that contain at least one extent of the table.
  • KSR QD can be made to use a smaller degree of parallelism than the maximum by using the KSR_MAXPARTITIONS environment variable or query directive, as explained in Section 11.1, "KSR QD Environment Variables.”
  • KSR QD divides its work according to the partitions of one table, the "driving" table of the join (see Section 11.4, "Kendall Square Extensions to EXPLAIN PLAN,” to find out how to determine which is the partitioning table for a given query).
  • a given KSR QD parallel subquery is responsible for fetching rows from a particular partition of the partitioning table, and for each of those rows, finding the matching rows from the other tables. If more than one of the tables being joined is partitioned across the same disks, one parallel subquery looking for matching rows from a non-partitioning table may contend for the same disk with another parallel subquery looking for rows matching its own partition of the partitioning table. Using a cluster solves this problem, because the rows of non-partitioning tables usually are in the same block (and thus the same partition) as the corresponding rows of the partitioning table to which they join.
  • a useful rule of thumb when creating tablespaces to be used for query decomposition is to define all files of your tablespace to be the same size. This will provide even data distribution across all the files.
  • KSR QD Some useful rules of thumb when creating tables to be used by KSR QD include the following:
  • KSR QD can effectively decompose queries using an index to retrieve rows from the driving table (although speedup will not tend to be as dramatic as for full-table scans), but the reading of the index itself is not decomposed (i.e.. each parallel subquery reads the same index blocks). Placing the index in a separate tablespace avoids disk contention between index reads by one parallel subquery and table reads by another parallel subquery.
  • KSR QD is most effective when target data is evenly distributed among all the files of a tablespace. This is so because KSR QD divides the work of a query so each parallel subquery covers the same number of table partitions as nearly as possible. For example, if a given table has 20 partitions (i.e.. it has at least one extent in each of 20 files) and the user has specified a maximum degree of decomposition of 10 (using the KSR_MAXPARTITIONS directive, described in Section 11.2). each parallel subquery will be responsible for retrieving data from two partitions. If no maximum degree of decomposition has been specified, each parallel subquery will retrieve data from one partition. If each of the 20 partitions contains roughly the same amount of data, each parallel subquery has roughly the same amount of work to do. Therefore, when they are executed in parallel, all the parallel subqueries will complete at about the same time. The execution time for the overall query is always slightly
  • a given file contains no extents for a given table, that file is not considered a partition of that table by KSR QD. If, for example, a table in a 20-file tablespace has equal amounts of data in each of 10 files and has no extents in the remaining 10 files, a query on that table can be decomposed into 10 parallel subqueries. If the same data were instead distributed among all 20 files, a higher degree of parallelism, and thus a higher degree of speedup, would be possible.
  • data blocks of the table are read only if they contain at least one row which satisfies the predicates applied to the index (e.g.. if an index on DEPTNO is used, only those data blocks are read which contain at least one row with a DEPTNO value in the range specified in the query's WHERE clause). Therefore, for a given indexed query, the skew of distribution of data blocks containing rows falling within the specified ranges on the indexed columns is important. This type of skew is more difficult to avoid for two reasons: First, one must predict the most frequently queried value ranges to determine the ideal data distribution.
  • Scatter clustering can be quite valuable.
  • the goal of scatter clustering is to create a hashed cluster with a large number of overflow blocks, each of which contains records with the same key value.
  • KSR QD can ameliorate the problem by splitting up the load in parallel but if the index does not provide speedup relative to full-table scan without query decomposition, it will not provide that speedup with query decomposition either.
  • each index entry consists of a key value/rowid pair, which points directly to the block containing the row in question. It also can be used for range predicates as well as direct match predicates.
  • ORACLE optimizer When presented with a query with an exact-match predicate on the hash-key columns, the ORACLE optimizer will choose hashed access rather than using the index on those same columns, because under normal circumstances, hashed access would unquestionably be faster.
  • KSR QD notices (in the execution plan) ORACLE has chosen hashed access and there is a regular index which has all the columns of the hash key as its leading columns. it generates an INDEX optimizer hint in the parallel subqueries. coercing the ORACLE optimizer to use the regular index rather than hashing. Since the parallel subqueries have rowid range predicates, this regular indexed query can be decomposed like any other. Because the data is clustered on the same column values with blocks for each cluster-key value well distributed among the files of the partitioned table, many fewer blocks need to be read than if this were not a hashed table.
  • HASHEDJTABLE a partitioned table hashed on the column HASHKEY_COLUMN. and there also is an index called REGULAR_INDEX on the same column.
  • the regular index optionally may contain additional trailing columns beyond those which match columns of the hash key. This means it can be used to further restrict the rows read. according to additional predicates in the querv. This could be particularly useful to give added flexibility, because a hash key must be chosen by a DBA before a table is created. Once the hashed table is populated, it requires a complete reorganization to add additional hash-key columns. It is much easier, however, to add columns to an index (or replace it with a different index) without affecting the data itself. So. if additional, frequently used selection criteria are identified after a hash table exists, these columns can be added to the regular index.
  • KSR QD must choose one of these indexes arbitrarily for ORACLE to use. In this event, however, the user optionally may choose the index by placing the INDEX optimizer hint in the original query. KSR QD always leaves any hints from the original query in the parallel subqueries to provide the user this extra degree of customized control over optimization when needed.
  • KSR QD determines whether you are running KSR QD. If your SQL*Plus has KSR QD linked in, then when you run SQL*Plus, you will see an indication KSR QD is running (e.g., KSR QD version 1.0).
  • KSR_NOPARTITION environment variable or directive If you are running KSR QD, try the KSR_NOPARTITION environment variable or directive. This disables KSR QD as explained in Section 11.1.
  • KSR QD intermediate tables are dropped automatically upon completion of query execution, but in exceptional cases (e.g.. if an application is canceled during query execution), they may not be dropped.
  • a utility called qdcleanup is provided for easily removing KSR QD temporary tables.
  • ORACLE To run qdcleanup. ORACLE must be started and the database must be open.
  • 0 KSR QD is intended to enhance the performance of complex decision-support queries executing over large databases. It works in coordination with the ORACLE optimizer to automatically provide what it believes to be the optimal strategy for running your query, based on what it knows about your application and the structure of your data.
  • KSR QD works on your behalf without you even being aware that it is there, except for the performance enhancements you see in the execution of your queries.
  • KSR QD can not be 5 expected to deduce. To further optimize your queries, that information may be made available to KSR QD. This is the purpose of the KSR QD user interface.
  • This control is provided in a granularity allowing for individual customization:
  • KSR QD control variable being used as an environment variable. See Table 11-1 for a list of KSR QD control variables.
  • Variables can be unset by issuing the unsetenv command with a null value.
  • KSRQD_ARRAY_S If set to an unsigned integer, indicates the array size for array fetch
  • KSR QD array fetch minimizes the number of client-server messages by fetching multiple rows per message in arrays.
  • KRSQD _ARRAY_SIZE must be set to a non-zero value. The default value is zero and means no array fetch. A value of one also means no array fetch, although the array fetch mechanism would be employed.
  • KRSQD_ARRAY If set to an unsigned integer, indicates the array fetch threshold. A JTHRESHOLD given parallel subquery starts using array fetch only after this many rows have been fetched. The default value is 0. meaning start right away (i.e., before the first row is fetched). If KSRQD_ARRAY_SIZE is zero, there will be no array fetching.
  • KSRQD_DISPLAY If set to any value, causes KSR QD timing information to be JTIME computed and displayed. By default, this variable is not set. See Section 11.6 for more information.
  • KSRQD_DISPLAY JTIME is also _FILE set. indicates the file where timing information is to be written. By default, this variable is not set. and all information is displayed on standard output.
  • KSRQD If set to a positive integer, determines the maximum degree of _MAXPARTITION parallelism for queries (even if the partitioning table has more
  • the actual degree of parallelism for a query will be the minimum of this value (if set), and the number of partitions in the partitioning table of the query.
  • the default value is the actual number of partitions. This parameter must be greater than zero.
  • KSRQD If set to any value, disables KSR QD within the user environment. _NOPARTITION It cannot be overridden by a KSRQD_MAXPARTITIONS query directive. Unsetting or removing this variable re-enables KSR QD within the local shell. By default, this parameter is not set. 11.2 KSR QD Directives
  • a directive comment is any comment which immediately follows the SELECT keyword (with or without intervening white space), and has a plus sign (+) immediately after the open-comment delimiter (/ * or — ).
  • a directive comment may contain zero or more KSR QD directives separated by blanks.
  • KSR QD will use a maximum of 10-way parallelism, and it will use array fetching to fetch 100 rows at a time, per parallel subquery.
  • the full (emp) ORACLE optimizer hint will be treated as a comment by KSR QD but passed along to the ORACLE optimizer in the parallel subqueries, forcing ORACLE to use a full-table scan on the emp table.
  • KSR QD can be enabled/disabled at three levels:
  • KSR QD • Per query (via KSR QD directives) Disabling KSR QD at a higher level prevents enabling it at a lower level (i.e. if KSR QD is disabled in a database, it cannot be enabled within user environments for queries on that database: if it is disabled in a user environment, it cannot be enabled for any query run in that environment). At each level, the default is for KSR QD to be enabled, unless it is disabled at a higher level.
  • Enabling/disabling KSR QD is largely transparent except insofar as it affects performance. The same queries will return the same results. Queries containing explicit KSR QD directives can be run in an environment where KSR QD is disabled: the KSR QD directives will simply be ignored, and the query will not be decomposed.
  • KSR QD is disabled will not show parallel execution as one of the steps for any query.
  • KSR QD When KSR QD is used on a query containing aggregate functions or a GROUP BY clause, there is an implicit commit when the cursor for the query is opened, and another implicit commit when it is closed (caused by KSR QD creating and dropping the temporary table used to collect aggregate results from the parallel subqueries). Users are advised not to use KSR QD on queries in transactions having uncommitted updates unless the implicit commits generated by KSR QD are perfectly acceptable.
  • KSR QD is enabled by default in a newly created database, provided the scripts described in Section 10.1 have been run. Any user with DBA privilege may disable query decomposition for that database by executing the following SQL script:
  • a DBA may re-enable KSR QD for that database as follows:
  • KSR QD is enabled by default in a user environment, unless the user accesses a database for which KSR QD is disabled.
  • KSR QD may be disabled in a user environment by setting the environment variable KSRQD_NOPARTITION to any value.
  • KSR QD may be re-enabled by unsetting that environment variable.
  • KSR QD is enabled by default for a query, unless it is disabled at the database or environment level. It may be disabled by specifying the KSRQD_NOPARTITION directive within the query. KSRQD_NOPARTITION overrides any other KSR QD directives specified in the same query (so a user may add it and remove it without making any other editing changes to the query). When KSR QD is disabled for a query, the query will not be decomposed. When KSR QD is enabled, it may be decomposed at the discretion of KSR QD. based on whether analysis of the query indicates decomposition is likely to be effective.
  • EXPLAIN PLAN When query decomposition is enabled and EXPLAIN PLAN is invoked for an SQL query, if the query in question would be decomposed, EXPLAIN PLAN produces an execution plan that includes a row providing information about how KSR QD is used for this query. The other rows of the plan show the optimization strategy ORACLE has chosen for executing the parallel subqueries.
  • OBJECT DWNER Indicate the owner of the partitioning table.
  • OBJECT_NAME Indicate the name of the partitioning table.
  • SEARCH_COLUMNS Indicate the degree of partitioning (i.e.. the number of parallel subqueries).
  • PARENT ID Indicate the ID of the logical parent of this row. Set to null for the special KSR QD row (PARENT D is always null for the row whose ID is 1). 11.4.1 EXPLAIN PLAN Examples
  • the first EXPLAIN PLAN statement is for a simple query (no ORDER BY or GROUP BY clauses, joins, or aggregates).
  • the emp table has 20 partitions.
  • the second EXPLAIN PLAN statement is for a query requesting DISTINCT values.
  • KSR QD uses a MERGE combining function in this case.
  • EXPLAIN PLAN SET STATEMENTJD 'query2' FOR SELECT DISTINCT LNAME FROM EMP: SELECT OPERATION. OPTIONS. OBJECT_NAME. ID. PARENT JD.
  • the third EXPLAIN PLAN statement is for a query joining two tables and has aggregation and grouping.
  • KSR QD chooses emp as the partitioning table because it is the driving table in ORACLE'S plan for the join.
  • ORACLE uses a nested-loops join and uses the unique key pk_dept to retrieve dept in the parallel subqueries.
  • ORACLE hints are allowed in SELECT, DELETE, or UPDATE statements, but since DELETE and UPDATE statements are never decomposed, KSR QD directives only have meaning after a SELECT statement.
  • ORACLE also permits non-hint comments to be interspersed with hints. From ORACLE'S perspective, the KSR QD directives are comments, while from KSR QD's perspective. ORACLE hints are comments.
  • KSR QD directives and ORACLE hints are semantically independent.
  • the presence of ORACLE hints in a query does not affect KSR QD's decision whether to decompose the query, except insofar as the hints yield an optimizer plan appropriate for decomposition.
  • the presence of a KSR QD directive in a query has no affect on ORACLE'S optimization. Note, however, that the presence of any ORACLE hint other than NOCOST implies use of the cost-based optimizer, even in cases where statistics are not available for any table in the query. Without the hints, the heuristic optimizer will be used.
  • KSR QD parser does not need to understand ORACLE hints, and it ignores them along with anything else within a comment that is not a KSR QD directive.
  • KSR QD passes the input query to the ORACLE EXPLAIN PLAN facility, the latter will take any hints into account in producing the execution plan.
  • KSR QD does not need to know whether the plan was influenced by hints, only what the plan is.
  • any hints are replicated in the subqueries. This ensures ORACLE chooses the same strategy for executing the subqueries as when EXPLAINING the input query (KSR QD relies on the assumption this will be the case).
  • KSR QD always partitions on the driving table of a join, provided that table is retrieved in a manner facilitating partitioning (indexed and full-table scans permit partitioning).
  • ORACLE hints provide a way to control the choice of driving table and the manner of retrieving that table, and thereby provide a way to control the choice of the partitioning table.
  • ORACLE optimizer 's default strategy for a query may not always be the most efficient strategy to use in conjunction with KSR QD.
  • ORACLE hints provide a means for users to second-guess the ORACLE optimizer, based on their ability to take KSR QD into account and coerce a different choice of driving table.
  • the timing utility can be enabled at the session level with an environment variable.
  • the timing utility is accessed through SQL*Plus or a KSR QD linked Pro*C program.
  • KSRQD JSPLAY JTIME controls display of KSR QD's timing information. If it is set. KSR QD measures and displays timing statistics for various steps in the execution of a query.
  • KRSQD_DISPLAY_FILE Setting the environment variable KRSQD_DISPLAY_FILE to a filename causes the timing statistics to be placed in that file. If this is not set. the information is sent to standard output.
  • the timing utility provides information about five rocessing phases, as shown in Table 1 1-2.
  • Table 11-2 The timing utility provides information about five rocessing phases, as shown in Table 1 1-2. Table 11-2
  • Preparing Query Full Preparation time Time to analyze and prepare a query (including parsing the query, performing semantic analysis, checking whether the query would benefit from decomposition, and generating data structures needed for later KSR QD processing)
  • Timing statistics are provided only for significant steps in the KSR QD process.
  • KSR Query Decomposer Version 0.1.4.10
  • Array processing can improve performance by reducing the number of calls from an application to ORACLE. Array processing allows an application to fetch multiple rows with only a single call to the ORACLE kernel. This is an important performance technique you can use in your applications.
  • An array size of 1,000 reduces the number of calls to 10. This performance gain is relatively small compared to the gain from increasing the array size from 1 to 100. Increasing the array size to 1,000 also increases the amount of memory needed to hold the array.
  • ORACLE application tools can take advantage of array processing. These tools include:
  • SQL*Plus for example, uses array processing automatically to return many rows from the database at once. It allows you to control the number of rows returned at a time through the SQL*Plus variable ARRAYSIZE (see “SOL *Plus User's Guide and Reference”). You can improve the performance of your queries by setting ARRAYSIZE appropriately.
  • the trade ⁇ off is SQL*Plus needs a larger buffer to store the rows of your query if it is fetching them using a bigger array size.
  • Array processing can be used at the query level, the subquery level (by KSR QD), or both. Both query and subquery array processing give you an analogous way to improve your query performance, but they operate independently.
  • KSR QD Users of KSR QD require the ability to create tables in their TEMPORARY tablespace to use KSR QD on queries using an intermediate table (all queries containing aggregate functions).
  • TEMPORARY tablespace is SYSTEM, but this can be altered by the TEMPORARY TABLESPACE clause of the ALTER USER statement.
  • the TEMPORARY tablespace is used by ORACLE to create temporary segments (e.g., for sorts), and this does not require any special privilege on the user's part.
  • KSR QD intermediate tables while temporary from KSR QD's stand ⁇ point, are ordinary tables as far as ORACLE is concerned. The user must have the privilege to create tables.
  • a user can be enabled to create tables in a given tablespace either by granting the RESOURCE role to the user, or by using the QUOTA clause of the ALTER USER state ⁇ ment to grant the user a storage quota in a specified tablespace. If a quota is used, it must be sufficient to permit creating a table with default storage attributes. The quota also must be sufficient to hold all the rows of the intermediate table, which
  • Queries using an intermediate table may run out of space for the intermediate table.
  • the KSR QD intermediate table stays quite small, because at any given time it contains at most one row for each parallel subquery (i.e., number of rows ⁇ degree of partitioning).
  • the intermediate table can potentially grow much larger: queries which contain both a GROUP BY clause and an ORDER BY clause. For such queries, the maximum number of rows in
  • the intermediate table equals the degree of partitioning times the number of groups in the query result. It is very unlikely this will exhaust the space in the tem ⁇ porary tablespace, because it is less temporary space than ORACLE needs for each parallel subquery (which already will be freed up before the stage of query execu ⁇ tion at which the KSR QD intermediate table is populated). It is possible a user's personal quota will be exhausted, however, since ORACLE'S temporary segments for sorts are not governed by the quota, but the KSR QD intermediate table is gov-
  • KSR QD decides whether to decompose a query based on analyzing the query and ORACLE'S explain plan. If KSR QD decides not to decompose, it silently lets ORACLE execute the query, as if KSR QD were not present.
  • KSR QD requires the driving table of a query have at least KSRQD_ INPARTITIONS partitions and be retrieved by either indexed scan or full-table scan (not hash access), to decompose a query.
  • KSR QD There is a performance overhead for KSR QD.
  • KSR QD Cost to check whether KSR QD is enabled. This cost is payed for all queries (whether or not KSR QD is enabled) when running a client application (e.g., SQL*Plus) with KSR QD linked in.
  • client application e.g., SQL*Plus
  • KSR QD intermediate tables may not get dropped in some cases, if errors occur during query execution.
  • a KSR QD intermediate table is created when a user executes a query containing aggregate functions, using KSR QD. It is owned by that user, -.. and has a unique name starting with QDIT (for example, QDIT169538333).
  • KSR QD is supposed to drop the intermediate table at the end of a query, or if any error occurs during query execution, but in some cases (particularly if the appli ⁇ cation executing the query crashes), this does not occur.
  • the presence of these tables is relatively benign, but eventually they may exhaust a user's storage quo ⁇ tas, so it is best to drop them. They may be dropped using the DROP TABLE state ⁇ ment or running the qdcleanup program (see Section 10.3.2).
  • Described below is a "front-end" to the ORACLE database management sytem that can parallelize a reasonable class of decision support queries without requiring major changes to the DBMS itself.
  • query decomposition in which parallel subqueries are submitted to the DBMS, matching the physical data declustering already permitted through table "striping" in ORACLE.
  • query decomposition is applicable to a very significant class of decision support queries, has excellent potential for performance gain for this class, and will be achievable with reasonable engineering effort at KSR.
  • this is an approach that can eventually benefit all users of ORACLE on parallel and shared-memory multiprocessor machines.
  • Section 2 (of this database note) describes our query decomposition approach in more detail, including a simple example.
  • Section 3 discusses the critical problems that need to be solved to implement this approach.
  • Section 4 analyzes the applicability of query decomposition with respect to a number of sample queries.
  • ORACLE permits the DBA to specify table "striping" in the CREATE TABLESPACE command.
  • a large table may be broken up into a number of files, spread across multiple disks. This is mainly viewed as an OLTP-oriented technique, aimed at optimizing random access to tables. Depending on how the file extents are populated, there may be some degree of data skew in terms of tuple distributions. However, striping is effectively a physical partitioning that we believe is adequate to support query decomposition.
  • Query decomposition is done by making a number of copies of the original query, and then appending additional predicates to each subquery to make it match one of the existing partitions of one of the tables in the query. These subqueries are then executed in parallel. Finally, a combining query (or function) over the subquery results produces the result of the original query. Most commonly, this is the union over the subquery results.
  • the degree of parallelism is limited by the number of physical partitions of the partitioning table, but not by the inherent parallelism in the query, as is the case for inter-operator parallelism.
  • it should be possible to leverage our initial work by basing query decomposition on hash-partitioned data, or by decomposing queries according to other criteria than matching data partitions.
  • in_interval(t.FILEID,i) is true for tuples in the ith group of files for table t.
  • the predicate translates into the appropriate conditions on FILEIDs (i.e., on SUBSTRft. ROWID, 15,4)), as was shown in the example in section 2.
  • index(t.x) means there exists an index on the J attribute of table t.
  • a nested loops join, with a as the outer table and b as the inner will be written NU(a,b).
  • a merge join of a and b will be written MJ(a,b).
  • Queries Ql through Q12 are against tables a, b, and c.
  • tables a, b, and c By starting with simple, abstract queries and adding increasingly complex conditions, we hope to better characterize the applicability of the query decomposition approach. Given our decision-support orientation, we have considered just read-only queries, and not data manipulation statements that do updates, deletions, or modifications.
  • index(a.x) Assume index(a.x). According to ORACLE, the index will be used to apply the predicate on a.x and the predicates on FILEID. This effectively parallelizes the subqueries. If there is no index, then the query can be treated as was Ql, with the a.x predicate being checked against all rows scanned by each subquery.
  • index (a. z) and index (b.z) If index (a. z) and index (b.z), then one of a and b will be chosen by the optimizer as the outer table, and should also be used as the partitioning table. By default, the optimizer will pick the smaller table as the outer one. However, if the smaller table has very few partitions, it is preferable to direct the optimizer to choose the larger table as the outer one, and to use it as the partitioning table as well. In either case, the subqueries can be effectively parallelized.
  • ORACLE will generate MJ(a,b), and will sort both a and b before performing the join. While the query can still be decomposed into subqueries, say Q3/a/i, the problem is that each subquery will sort the entire b table. The likely result is relatively little performance speedup. Note that a parallel hash join operator would help in this case, if it were available.
  • NU(a,b) will be generated.
  • the index on a.x will be used to apply the predicate and to get FILEIDs; this is straightforward and effective.
  • NU(a,b) will also be generated.
  • index(a.x) and index(a.z) and indexfb.z) are generated if index(a.x) and index(a.z) and indexfb.z), with the two indexes on a being intersected before a tuples are retrieved.
  • Q4 reduces to the Q3 case. In other words, there is no problem unless not index(a.x) and not index(a.z) and not index ⁇ .z). In that case, MJ(a,b) will be generated, and the subqueries cannot be effectively parallelized.
  • index(a.x) and index ⁇ .y) and index(a.z) and index ⁇ .z) then nested loop joins are possible with either a or b as the outer table.
  • the choice will be made based on the selectivity of the two single-table predicates D the more selective predicate will be applied to the outer table. If NU(a,b) is generated, then Q5/a/i is appropriate; if it is NU(b,a), then Q5/b/i is the preferred decomposition into subqueries. Either way, the subqueries can be effectively parallelized.
  • index(a.x) If only one of the indexes supporting single-table predicates is present, say index(a.x), then Q5 reduces to the Q4 case. If neither is present, then Q5 reduces to the Q3 case.
  • the preferred join order of tables is: first, the largest unindexed table, if one exists; followed by all indexed tables, in order of decreasing predicate selectivity (including both join predicates and single-table predicates); followed by all remaining unindexed tables, if any.
  • the subqueries can be effectively parallelized. Since ORACLE currently does a sort on a.x for each subquery in order to weed out duplicates, the subquery results are assumed to be sorted on this field. Combining the subquery results then requires just one more level of duplicate elimination.
  • the keyword DISTINCT can also appear inside of an aggregate function (e.g., A VG (DISTINCT a.y)). This construct cannot be effectively parallelized; it is impossible to combine subquery results in a meaningful way.
  • a VG DISTINCT a.y
  • partitioning table also referred to as "driving table” elsewhere
  • join order the set of decomposable queries (assuming that the underlying tables are all partitioned). We expect these rules to be refined over time.
  • a first implementation may use the first table in the optimizer's EXPL.AIN plan as the partitioning table.
  • the preferred join order of tables is: first, the largest unindexed table, if one exists; followed by all indexed tables, in order of decreasing predicate selectivity (including both join predicates and single-table predicates); followed by all remaining unindexed tables, if any. This supports access plans that consist of one or more nested loops joins, followed by zero or more merge joins.
  • Non-flattenable nested subqueries can be effectively parallelized, if they do not contain any other problematic constructs.
  • Decomposition will be done when the answers to (a), (b), and (c) are yes.
  • the user will always retain the ability to disable decomposition if desired. We intend to automate the answers to all of these questions.
  • An application programmer can override any of the automatic decomposition decisions by using directives in the SELECT statement, in the form of embedded comments.
  • the exact form of these directives are not described in this database note, but will adhere to the style used in ORACLE. For purposes of this database note, we will make some rational guesses about what they might look like.
  • Query decomposition can be used with Pro*COBOL, Pro*C, SQL*Plus, OCI, SQL*Report, and possibly SQL*ReportWriter when it gets rewritten to use UPI in ORACLE version 7.0. (It might also work with the precompilers for other languages, but we will make no special effort to insure that.) We would like to support QD for PL/SQL, but have not yet determined how much additional work would be needed, if any.
  • the parallel execution of queries via QD can be selectively enabled and disabled without changing any application code.
  • a parallel application can be written and initially tested in serial mode. After it is working correctly, parallelization can be turned on with some kind of switch.
  • ORACLE application processes queries by iteratively performing fetches on a cursor, which steps through a virtual table of result rows. This result table does not necessarily exist as a complete entity at any point in time. It is frequently constructed on the fly, so that the result
  • One of our design goals is to modularize query decomposition to allow that code to be maintained separately from the rest of the ORACLE code. This follows Oracle's policies on port-specific modifications and will simplify the appropriate sharing of maintenance between KSR and Oracle.
  • the UPI (User Program Interface) is the common point of access to the ORACLE kernel for all applications.
  • a parallel UPI library (PUPI, pronounced "puppy") will be developed that intercepts each call to UPI (for performing operations like connect, parse, fetch, etc.) and generates multiple calls to UPI, which generally will be executed in parallel (see Figure 26 - 1).
  • KSR ORACLE should be runable without the PUPI.
  • PUPI will pass the original query on to UPI to have it parsed and verify that the syntax is correct. After that, the query will be scanned to parse the parallel directives, if any. By default, we will decompose any queries where it is correct and effective to do so, as long as decomposition has been enabled. The user can override the decision to decompose or the choice of partitioning table. Once the partitioning table has been determined, the PUPI will look up the table name in ORACLE'S catalog to find out the number of files comprising it and the list of file Jd's. The number of files determines the number of subqueries and, therefore, the number of additional connections to ORACLE that are needed.
  • Rows will be fetched asynchronously from the subcursors and returned to the application as needed.
  • the rows returned from the subcursors may need to be combined or ordered in some way before the root cursor's fetch can be satisfied. See the Parallel Cursors section below for more details.
  • the PUPI will consist of a set of functions that have the same external interface as their UPI counterparts, but will call the appropriate UPI functions multiple times. Not all the UPI functions will be duplicated in the PUPI, since not all of them can be or need to be parallelized. We need a way to easily switch between serial and parallel query processing. At different times, the same application may call either UPI or PUPI functions without (by our own requirements) changing any code. (See Figure 26 - 3. The three functions shown in each library parse a query, execute it, and fetch the results. There are many more functions that need to be implemented.) The "Application" in this figure can be assumed to include SQLLIB and OCI, i.e., everything above the UPI level.
  • the UPI maintains a hstdef " (host definition) structure for every connection that exists. We will allocate a hstdef for each additional connection we need (one for each subquery). The proper hstdef for each connection must be referenced when performing any actions related to the subqueries.
  • the extra connections can't be made until after the original query has been parsed and the number of subqueries has been determined. At that time, we will also have access to the hstdef that was set up on the first connection, which may contain information we need in order to make additional connections to the same database. (We need to have access to the connect string (user, password, host, etc.), or its equivalent. Without that, we have no way of knowing where the original connection was made.) We may also need access to the
  • ALL_DATA_FILES doesn't yet exist, but could be created as a duplicate of
  • DBA_DATA_FILES with the additional condition that the tablespace_name must exist in .ALL_T.ABLES.
  • a public synonym could be created for DBA_DATA_FILES, with public select access. It depends on how concerned users are about letting everyone see what database files exist on the system.
  • a set of combining functions will be developed to produce a single result row for the application from all of the subquery rows available for consideration. Only the most recent row from each subquery needs to be considered.
  • the specific method used for merging or ordering the subquery results is completely dependent on the nature of the query. The existence of aggregate functions, ORDER BY, or GROUP BY clauses are the main factors to consider. Sometimes multiple combining functions need to be applied to the same query. For example, the query
  • each subquery will have returned a single row containing the aggregate result for its partition. Combine all of these rows into a single row, using the appropriate aggregate function(s).
  • the combining function cannot simply take one row from each subquery and combine them. It needs to select and combine rows where the group values match each other. For the first root cursor fetch, all the DEPTNO 10's will be combined; the next fetch will combine the 20's, etc. Since GROUP BY implies ascending ordering before the aggregate function was applied, we can select the lowest available group value and all of its duplicates.
  • the need to combine multiple rows implies that the query has at least one aggregate. Combining can be viewed as collapsing several rows into one. All the eligible subquery rows are identical in the non-aggregate columns. These columns can simply be copied into the result row.
  • the aggregate columns can be combined by calling the appropriate combining function, passing the column number and pointers to the relevant rows. Note that averages need some special handling - the corresponding COUNT column also needs to be identified and taken into account by the combining function.
  • Airay fetches will need some special consideration.
  • the combining functions may have to be called iteratively until the array is full.
  • the user requested decomposition and the query cannot be decomposed correctly .
  • the user requested decomposition and the query can be correctly decomposed, but not effectively. It may even run slower.
  • Data skew or partition skew in the partitioning table with respect to the query.
  • data skew here to mean any distribution of data that causes result rows to be fetched from the subcursors in something other than round-robin fashion. For example, sorted output may appear in clumps so that several rows in succession from the same subcursor are returned to the root cursor. During such periods of time, little, if any, parallel fetching will occur. This phenomenon may appear and disappear many times during the course of a single query. Increasing the number of fetch buffers per subquery will help to minimize the effects of this type of data skew.
  • Partition skew is defined as a distribution of data that results in unequal-sized partitions. During the latter part of query execution, and possibly even during the entire query, some partitions will have no more rows to fetch. This will reduce the degree of parallelism for the remainder of the query.
  • the database partitions may actually be equal in size, but the effective partition size for any given query might be reduced by predicates in the query.
  • ORACLE or OS limits on the number of processes, threads, connections, etc.
  • query decomposition is designed to work in conjunction with other parallel processing techniques, such as parallel relational operators and pipelining.
  • parallel relational operators and pipelining.
  • Our Query Decomposition parallelizes a query by dividing it into subqueries, each of which use a rowid range predicate to specify one or more files to which that query's reads will be restricted.
  • the approach depends on partioning tables across files on multiple disk drives, so that the files can be read in parallel. So, for a trivial example, if the table EMP is partitioned across 3 files with ORACLE fileid's 1, 2, and 3, then the query SELECT * FROM EMP can be decomposed into three subqueries:
  • the first query will only read blocks of the EMP table which are in file 1, the second will only read blocks from file 2, and the third from file 3.
  • ORACLE has been modified to restrict reads during full table scans, based on rowid range predicates, as a necessary prerequisite to implementing this approach.
  • Query Decomposition can also work with queries that use an index.
  • Each of these subqueries must redundantly read the same index blocks, to find index entries for DEPTNO 5, but hopefully the index blocks will be cached by the first subquery which gets to each one, so they are only read once.
  • a subquery finds an index entry for DEPTNO 5, however, it will examine the rowid stored in that index entry, to see whether it fall within the range for that subquery. Only if it does will that subquery read the data page containing the row with that DEPTNO value and rowid. Speedup is not as close to linear as with full table scans, because only the table reads are partitioned. Logically, the total reads are increased due to redundant reading of the index, but the redundant reading happens in parallel, and hopefully caching will eliminate most actual redundant I/O.
  • Query Decomposition as described above speeds up query execution by parallelizing the reads involved in a query, but not by reducing their total number. While this improves individual query response time, it does not improve system throughput (and may even reduce throughput, due to the added overhead of additional threads and processes, and of redundant index reads).
  • ORACLE'S clusters and hashed clusters are approaches to speeding up query execution by greatly reducing the number of reads needed to accomplish certain queries.
  • "Regular" (i.e. non-hashed) clusters reduce the reads needed for commonly-executed joins by clustering together the rows of several related tables based on common join column values, further reducing the number of blocks needed to read a related set of rows by storing each cluster key value only once for all rows of all tables sharing that key value.
  • This kind of cluster still has an associated index on the cluster key, but the index entries simply point the to root block for the cluster key value, rather than having separate rowid entries for individual rows.
  • Hashed clusters reduce reads for queries which seek rows of an individual table that exactly match a given key value. Rows with key values that hash to the same hash key value are clustered together, and no index is needed to navigate directly to the root block for a given hash key value.
  • Query Decomposition has more general applicability: as long as a DBA decides in advance to partition a given table across multiple disks, Query Decomposition can be used on that table for any query that uses either a full table scan or any regular index, rather than being restricted to queries with predicates on certain predetermined columns.
  • Query Decomposition and clustering cannot be used in conjunction to optimize access to the same table in the same query. This is so because accessing a table through a cluster key, whether hashed or otherwise, does not use either a full table scan or a regular indexed scan. Instead, it uses the cluster index (for regular clusters) or hashing to find the root block for the cluster key value. Then, if all rows for the specified cluster key value are in that one block, that's all that has to be read, so there's no opportunity for parallel partitioning. Otherwise, all of the chained blocks for that cluster key value must be read in sequence, whether they are in the same or different files.
  • the index entry for a particular key value just points to the first block of the overflow chain, so there's no opportunity to examine rowid's and decide whether they fall in a specified range, to decide whether to read a data block.
  • an index has a fairly small number of distinct values, relative to the number of rows in a table, and if rows with a given index value can be scattered anywhere in the table, without regard to their key value on that index, then even after using the index, a much larger volume of data may have to be read from the table than the volume represented by rows with the desired key values, because only a small fraction of each block read consists of the desired rows. In the worst cases, all blocks of the table must be read, so that performance is worse than if the index isn't used at all (because of the extra reads of the index, and because of the higher proportion of random to sequential I/O's).
  • QD can ameliorate the problem by splitting up the load in parallel, but it remains the case that if the index doesn't provide speedup relative to full table scan without QD, then it won't provide speedup relative to full table scan with QD.
  • each index entry consists of a key value /rowid pair, which points directly to the block containing the row in question. Also because it is a regular index, it can be used for range predicates as well as direct match predicates.
  • the ORACLE optimizer When presented with a query that has an exact-match predicate on the hash key columns, the ORACLE optimizer will choose hashed access rather than using the index on those same columns, because under normal circumstances, hashed access would unquestionably be faster.
  • the Query Decomposer notices (in the EXPLAIN plan) that ORACLE has chosen hashed access, and that there is a regular index which has all of the columns of the hash key as its leading columns, it can generate an INDEX optimizer hint in the parallel subqueries, coercing the ORACLE optimizer to use the regular index rather than hashing. Since the parallel subqueries have rowid range predicates, this regular indexed query can be decomposed like any other. But because the data is clustered on the same column values, with blocks for each cluster key value well-distributed among the files of the partitioned table, many fewer blocks need to be read than if this were not a hashed table.
  • the regular index may optionally contain additional trailing columns, beyond those which match columns of the hash key. This means it can be used to further restrict the rows read, according to additional predicates in the query. This could be particularly useful to give added flexibility, because a hash key must be decided upon by a DBA before a table is created, and once the hashed table is populated, it would require a complete reorg to add additional hash key columns. It is much easier, however, to add columns to an index (or replace it with a different index) without affecting the data itself. So if additional frequently- used selection criteria are identified after a hash table already exists, these columns could be added to the regular index.
  • the Query Decomposer must choose one of these indexes arbitrarily, as the one it will tell ORACLE to use, because it is not equipped to perform the function of a full-fledged query optimizer, to analyze the predicates in the query and decide which index would be best to use. In this event, however, the user may optionally choose the index by placing the INDEX optimizer hint in the original query.
  • the Query Decomposer always leaves any hints from the original query in the parallel subqueries, to provide the user this extra degree of customized control over optimization when needed in this or other situations.
  • ORACLE applications use a client-server architecture in which all database access is performed on behalf of an application program by a separate server or "shadow" process. While this architecture is used even when the client application and the server are running on the same machine, ORACLE'S SQL*Net network software supports the seamless connection of remote clients and servers running on heterogeneous platforms. This permits the KSR1 to play the role of database server for a network of workstations, a configuration which is becoming increasingly prevalent, and may be preferred or even required by some potential KSR customers.
  • Section 1 explains why remote workstations cannot be supported by the current QD architecture; Sections 3 and 4 present alternate architectures to solve the problem; and Section 5 draws conclusions about which architecture is likely to be preferable, and how much effort will be required to implement it.
  • PUPI Parallel User Program Interface
  • This set of routines emulates the calling sequence and behavior of the UPI routines, but is also capable of decomposing a query into parallel subqueries, creating and managing the threads in which those parallel subqueries are executed, and combining the results to emulate the result of the original query. For each parallel subquery, a separate thread is created, and a connection is made from within that thread to a separate ORACLE server.
  • PUPI routine is called for a task which does not require parallelism, behavior is the same as for an ordinary UPI routine, and the call is serviced by the server from the original user connection (which we may designate the primary server to distinguish it from the servers used for parallel subqueries).
  • This architecture is shown in Figure 61 - 1.
  • This architecture takes advantage of ORACLE'S separation of client and server processes, even for local connections, to manage parallelism inside the client process, thereby requiring minimal change to the server. Unfortunately, this only works when the client is executing on the KSR1. To support a remote client, the architecture must be changed so that parallelism can be managed on the server side of the remote client/server boundary.
  • An advantage of this solution is that it introduces no new processes or connections, other than those specifically needed for executing parallel subqueries.
  • a client program makes sends a message to the server which does not require parallel processing, that call is simply passed on into the kernel, without requiring an additional message.
  • the ORACLE server is playing a dual role, both as a standard ORACLE server, and as a QD server.
  • ORACLE kernel From the standpoint of detailed design and implementation, changes of this nature to the ORACLE kernel present much room for unpredictable difficulties and side effects.
  • Prior experience indicates that it can be very difficult to emulate client behavior inside a server, since the two sides of a client/server interface, if not specifically implemented to allow for this, may contain variables with corresponding names and purposes, but which are used in subtly different ways.
  • the current implementation of QD assumes its residence in the client; ORACLE functions are called which have similar but different counterparts on the server side.
  • the QD server would incorporate routines from the outermost, message handling layers of the ORACLE kernel (in particular, modules of the SQL*Net and Two Task Common, or TTC, layers), but its dispatcher would call PUPI routines, rather than OPI or POPI routines, to service requests.
  • This architecture is shown in Figure 61 - 3 below.
  • a key advantage of this approach is that, while it incorporates some peripheral kernel routines, it does not constitute modification of the ORACLE kernel itself.
  • QD code is completely segregated from the kernal. There are likely to be fewer dangers of side effects, and much less danger of unintentional security violations (the latter danger is not entirely eliminated, because emulating an ORACLE server from the client's perspective may still require access to the ORACLE SGA, but in a better-isolated and more easily-controlled context).
  • Some degree of message translation may be necessary to relay incoming messages, intended to be processed by OPI calls, to UPI or PUPI calls which will pass them along to an ORACLE server. Furthermore, while the majority of UPI calls do not require PUPI counterparts in the current implementation, because they are not directly related to retrieving query results (e.g. calls for managing transactions, for connecting to ORACLE, or for modifying data), a QD server would need to be able to relay
  • ORACLE server More detailed study of the ORACLE code will be required to determine the amount of effort involved, and whether it outweighs the advantages of leaving QD in the PUPI layer. It could turn out that this approach is not as different from the approach of relocating QD inside the OPI layer as it would superficially appear to be.
  • Another disadvantage of this approach is that requests for database operations which do not require parallelization must make an extra message hop to get from the client application to the ORACLE server which will service them. Since the QD code decides whether a given UPI call requires parallelization, if the QD code is in the QD server rather than in the application program, then the application program can't "know" whether to send a given request to the QD server or the ORACLE server, so it must always choose one or the other. We can provide mechanisms to let the DBA or application user decide globally or per application whether to enable QD for remote queries, so that applications with little or no need for QD can avoid the extra overhead of the intermediate QD server.
  • a hybrid approach could place inside the application program those portions of QD logic which determine whether to decompose a query, while managing the parallelism in a QD server. This approach, however, would require substantially more effort to implement, since it would involve a re-partitioning of QD functionality among processes.
  • the QD server approach appears preferable to the approach of locating QD in the ORACLE server, but not dramatically so.
  • the QD server approach avoids modifying the ORACLE kernel, but this is somewhat offset by the added architectural complexity and possible complications in packaging and code integration. Maintaining the same QD/ORACLE interface for remote and local clients is certainly preferable conceptually, but may be offset by difficulties in relocating some kernel routines in a separate server, and in relaying messages to UPI routines which were intended for OPI routines.
  • the QD server approach introduces extra performance overhead for non- parallelized ORACLE calls; this can be limited at the cost of slight extra administrative complexity, and might be reduced further by optional hybrid approaches, at the cost of greater development effort.
  • a reasonably conservative initial estimate of development cost would be one person-month to implement the basic QD server functionality, with an additional two to three weeks to resolve peripheral issues of administration, configuration, and packaging.
  • the initial phase of development would involve a detailed examination of the relevant ORACLE code, which would facilitate making a final decision between the alternate approaches, and producing a more reliable development cost estimate and task breakdown.
  • This paper provides a conceptual framework for automating the process of query decomposition proposed in Database Notes #21 and #26.
  • This framework can be viewed as a general structure within which to answer the question "What do we know, and when do we know it?", during the stages of transformation from an original input query to a decomposed query ready for parallel execution.
  • this paper provides a breakdown of the categories of rules involved in query decomposition, their input information and goals, and the categories of generated queries associated with them.
  • OAT model whose name is an acronym for three forms through which a collection of information passes during a transformation: the original form (O-form), the analyzed form (A-form), and the transformed form (T-form).
  • the process of query decomposition consists of producing, for a given input query, the collection of parallel subqueries, combining queries, combining function control structures, and other control structures needed to retrieve data in parallel and combine it to emulate the result table of the original query.
  • This can be viewed conceptually as a transformation of the original query (which we will designate as the O-form of the query) to that collection of objects which comprise the decomposed query (which we will designate the T-form of the query).
  • To automate this process we must specify a collection of rules whose starting point is the O-form of a query, and whose ultimate goal is the T-form. This highest-level goal path is shown in Figure 32 - 1.
  • An SQL query submitted to the system does not contain within itself all of the information needed to decompose it. Strategic information such as index usage, table cardinalities, predicate selectivity, and join order and method must be obtained from the query optimizer to make decisions about decomposition strategy, such as choice of a partitioning table. Semantic information about tables, columns, clauses and expressions in the query must be gathered from the data dictionary to determine the details of combining functions and queries (for example, what kind of comparisons to perform for a merge sort,
  • the A-form includes the original query definition and any needed cross-references between that definition and the other collected information, so that no information is lost in the transition from O-form to A-form.
  • the A-form provides a valuable "fire wall" between the gathering/analyzing rules and the transformation rules. It prevents radical differences in the gathering/analyzing approach from having any effect on the transformation approach (for example, the difference between parsing the input query and then querying the data dictionary to bind semantic information to the parsed query, or obtaining a parse tree with already-bound semantic information from the query optimizer, and translating that to our standardized A- form). It also permits us to expand our repertoire of parallelization techniques relatively independently of the gathering/analyzing rules.
  • query decomposition process Much of the query decomposition process, both in the gathering/analyzing and transformation phases, is accomplished through the generation and execution of queries.
  • query is used in the broad sense to include DDL commands such as CREATE and DROP, para-DML commands such as EXPLAIN, and logical equivalents to these and other DML commands which do not necessarily involve explicit generation or processing of SQL.
  • Query generation is used to mean applying rules to define a query and prepare it for execution.
  • Query execution is used to mean retrieving information through the query.
  • Queries can be broken down into five categories: probing
  • gathering/analyzing rules can be divided into two classes: gathering rules which govern the generation and execution of probing queries, and analyzing rules which analyze and restructure the gathered information to produce the A-form of the query.
  • Probing queries also fall into two groups: those which gather information on query optimizer strategy and associated cardinality and selectivity estimates; and those which gather semantic information about objects referenced in the query from the data dictionary. (This may be an over-simplification in some cases. For example, queries about file partitioning have more to do with retrieval strategy than semantics, but formally they may have more in common with data dictionary queries than with optimizer queries, if the file partition information is accessed through a data dictionary view.)
  • Optimizer strategy information can be obtained by invoking EXPLAIN to produce an access plan for the query, and then generating and executing appropriate queries against the plan table to obtain information about join order, join methods (nested loop vs. merge), and index usage. (If a later release of EXPLAIN also provides cardinality and selectivity estimates, these will be gathered as well.)
  • Semantic information can be obtained by asking queries against data dictionary views, and by using DESCRIBE SELECT to generate a SQLDA structure describing the output columns (select list items) of the original input query, or of transformations of that query. In some instances alternate strategies for obtaining information are possible.
  • Additional data dictionary queries beyond those which gather basic semantic information, may be needed in some cases to establish cross-references between the semantically-augmented parse tree and the query optimizer plan. These could be needed, for example, to determine which index name in the optimizer plan corresponds to which table name in the query definition, or to match table synonyms used in the query definition to actual table names.
  • Set-up queries are generated during the transformation phase of query decomposition, and, as the name implies, they are executed during an initial set-up phase of query execution. They fall into two general groups: DDL set-up queries to create temporary tables or indexes; and DML set-up queries, which could be used in multi-stage execution strategies to populate temporary tables with intermediate results. Potentially, a DML set-up query could itself be decomposed and executed in parallel.
  • Temporary tables may be created at set-up time, and populated during main query execution, to gather rows from parallel subqueries for final aggregation or testing of a HAVING clause by a combimng query.
  • Creating temporary indexes, and populating intermediate sorted tables during set ⁇ up, are also steps of alternative approaches to merge joins which avoid redundant sorting of the non-driving table in the join by each parallel subquery, either by pre-sorting or by pre- indexing the non-driving table. If presorting is used, only those rows which satisfy single- table predicates are inserted in a temporary table, which is indexed on the join columns, and the temporary table replaces the original table in the FROM clauses of the parallel subqueries. If pre-indexing is used, the entire table must be indexed on the join columns. Either way, the resulting table can now be used as the inner table in a nested loops join.
  • Clean-up queries are generated at the same time set-up queries are generated, and are executed when the overall parallel cursor is closed.
  • Output rows from parallel subqueries provide the input rows to the combining functions and queries discussed below.
  • the combining functions or queries dynamically merge the output streams of the parallel subqueries, so that the parallel subqueries do not have to be executed to completion before executing the combining functions or queries.
  • a combination of combining functions and queries is used to merge the output streams of parallel subqueries, producing a single output stream identical except possibly for ordering to that which would have been produced by directly executing the O-form of the query.
  • a single combimng function is used to produce the logical "union all" of the separate parallel streams.
  • More complex cases can involve multiple functions or queries working together to perform merging of sorted streams, merging of groups, aggregation, and expression evalution (e.g. testing of HAVING clauses), as well as the set operations UNION, INTERSECT, and MINUS. The means by which multiple combining functions and queries can coordinate their efforts are discussed in detail in DBN #36.
  • Combining functions are generic and predefined (e.g. one predefined grouping function, one predefined merging function, etc.), but their roles in executing a particular decomposed query are governed by control structures which are generated during the transformation phase of query decomposition. The interconnection of these structures governs the way in which the different combining functions and queries coordinate their work.
  • a control structure When a combining query is called for, a control structure will be generated as for a combining function, but in addition, the query itself must be generated. This is done by starting from the A-form of the query, and applying transformations analogous to, but different from, those used to generate parallel subqueries. These can include the following:
  • a goal of the transformation phase of query decomposition is the generation of control structures to glue together and coordinate the overall parallel cursor, and to keep track of housekeeping details such as memory buffers and DBMS connections.
  • this means that the several types of queries produced by transformation rules are not separate and independent goals, but rather coordinated pieces which together constitute the embodiment of a parallel execution strategy, which is the T- form of a query.
  • probing queries differ from the other four in that they are created during the gathering/analyzing phase of query decomposition, rather than during the transformation phase. They also differ in that while their generation is a goal of some of the gathering rules, they are used as a tool by other gathering rules, and the output of their execution serves as input to the analyzing rules, and so, indirectly, to the transformation phase of query decomposition.
  • the remaining categories of queries can all be considered end products of query decomposition, and collectively (together with parallel cursor control structures) they constitute the T-form of a query.
  • Figure 32 - 3 summarizes the query decomposition process. Solid arrows in the diagram represent the application of rules, and point towards the goals of those rules. .Arrows with dashed lines indicate query execution, and point from the query being executed to the query which depends on the output of that execution. Note that while there is
  • 97 - is a sequence of execution dependencies between the four types of queries belonging to the T-form, the rules which generate them can conceptually be applied in parallel.
  • Prolog provides an ideal tool for the definition, prototyping, and "proof-of-concept" testing of the rules of query decomposition. Rules can be specified clearly, concisely, and non-procedurally in Prolog, which can greatly facilitate testing of complex combinations of rules. Prolog also supports syntax for concise specification of grammar, which would facilitate developing a basic SQL parser to drive the rule testing. Once the set of rules has been verified in Prolog, it can be hard-coded in C for optimal efficiency of the actual implementation. As rules change or new rules are added to the system in subsequent releases, the Prolog prototype will provide a flexible tool for testing them together with the existing rules before adding them to the C implementation. The present document provides a framework within which to define and test specific rules in the Prolog prototype.
  • pnodes referred to as “building blocks” or “bb's” elsewhere. These can be arranged into a doubly- linked tree called a pnode tree. Each pnode has one pointer to its parent, and zero or more pointers to its children, depending on its node type (some node types have a variable number of pointers to children). Other attributes of all pnodes include:
  • Node ID Uniquely identifies this pnode within a particular pnode tree
  • Each node type has its own executor function
  • a variant portion will contain attributes particular to each node type, sometimes including additional state attributes.
  • Each node type also has a specialized executor function, but all executor functions take the same two parameters: a request code indicating the type of operation to perform, and an array of pointers to buffers which is used to locate data.
  • pnodes are specialized row sources.
  • Pnode trees are parent-driven.
  • a parent "pulls" rows from its children, which passively respond to parent requests.
  • a parent pulls a child by calling the child's executor function, passing it a request code to distinguish the specific nature of the request. Since all executor functions are of the same type, and since the generic portion of the pnode contains a pointer to its function, a parent can call a child's function without knowing the child's node type, or what specific function to call.
  • Request codes might include:
  • RESET_CACHE Reset to beginning of cached group of rows, return first
  • NEW_CACHE Start a new cached group of rows, return first
  • CLEANUP Perform any necessary cleanup, e.g. close cursors
  • a second (perhaps overlapping) series of reply codes is returned to the parent by the child, as the return value of its executor function.
  • These might include:
  • a third (again perhaps overlapping) series of state codes will be maintained by a pnode's execution function as values of its state field, to let the pnode remember its context from one pull to the next.
  • State codes might include:
PCT/US1995/001356 1994-01-31 1995-01-31 Improved method and apparatus for data access in multiprocessor digital data processing systems WO1995021407A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP7520740A JPH09511347A (ja) 1994-01-31 1995-01-31 マルチプロセッサディジタルデータプロセッシングシステムにおけるデータアクセス方法および装置
EP95911601A EP0753176A4 (en) 1994-01-31 1995-01-31 METHOD AND APPARATUS FOR DATA ACCESS IN A DIGITAL COMPUTER SYSTEM WITH SEVERAL PROCESSORS

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/189,497 US5742806A (en) 1994-01-31 1994-01-31 Apparatus and method for decomposing database queries for database management system including multiprocessor digital data processing system
US08/189,497 1994-01-31

Publications (2)

Publication Number Publication Date
WO1995021407A2 true WO1995021407A2 (en) 1995-08-10
WO1995021407A3 WO1995021407A3 (en) 1995-09-14

Family

ID=22697584

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1995/001356 WO1995021407A2 (en) 1994-01-31 1995-01-31 Improved method and apparatus for data access in multiprocessor digital data processing systems

Country Status (5)

Country Link
US (3) US5742806A (US06816854-20041109-P00003.png)
EP (1) EP0753176A4 (US06816854-20041109-P00003.png)
JP (1) JPH09511347A (US06816854-20041109-P00003.png)
CA (1) CA2180252A1 (US06816854-20041109-P00003.png)
WO (1) WO1995021407A2 (US06816854-20041109-P00003.png)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013156773A1 (en) * 2012-04-17 2013-10-24 Dataline Software Ltd Methods of querying a relational database
US11663179B2 (en) * 2020-12-21 2023-05-30 International Business Machines Corporation Data simulation for regression analysis
US11797520B2 (en) * 2019-11-29 2023-10-24 Oracle International Corporation ROWID elimination rewrite

Families Citing this family (521)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2297180B (en) 1993-09-27 1998-05-20 Oracle Corp Method and apparatus for parallel processing in a database system
US5742806A (en) 1994-01-31 1998-04-21 Sun Microsystems, Inc. Apparatus and method for decomposing database queries for database management system including multiprocessor digital data processing system
JPH10501086A (ja) * 1995-11-02 1998-01-27 インターナシヨナル・ビジネス・マシーンズ・コーポレーシヨン 記憶プレーン編成及びそれに基づく記憶システム
JP3510042B2 (ja) 1996-04-26 2004-03-22 株式会社日立製作所 データベース管理方法及びシステム
JP3747525B2 (ja) * 1996-08-28 2006-02-22 株式会社日立製作所 並列データベースシステム検索方法
US5978799A (en) * 1997-01-30 1999-11-02 Hirsch; G. Scott Search engine including query database, user profile database, information templates and email facility
US5903891A (en) * 1997-02-25 1999-05-11 Hewlett-Packard Company Hierarchial information processes that share intermediate data and formulate contract data
US5860069A (en) * 1997-04-11 1999-01-12 Bmc Software, Inc. Method of efficient collection of SQL performance measures
US5963932A (en) * 1997-04-29 1999-10-05 Oracle Corporation Method and apparatus for transforming queries
US6115705A (en) * 1997-05-19 2000-09-05 Microsoft Corporation Relational database system and method for query processing using early aggregation
JP3779431B2 (ja) * 1997-06-13 2006-05-31 富士通株式会社 リレーショナルデータベース管理装置,中間リンクテーブル自動作成処理方法およびプログラム記憶媒体
US6092062A (en) * 1997-06-30 2000-07-18 International Business Machines Corporation Relational database query optimization to perform query evaluation plan, pruning based on the partition properties
US5864840A (en) * 1997-06-30 1999-01-26 International Business Machines Corporation Evaluation of existential and universal subquery in a relational database management system for increased efficiency
US5893086A (en) * 1997-07-11 1999-04-06 International Business Machines Corporation Parallel file system and method with extensible hashing
US6006220A (en) * 1997-09-30 1999-12-21 International Business Machines Corporation Determining the optimal access path for a query at execution time using an actual value for each variable in a query for estimating a filter factor
US5937411A (en) * 1997-11-21 1999-08-10 International Business Machines Corporation Method and apparatus for creating storage for java archive manifest file
US6076087A (en) * 1997-11-26 2000-06-13 At&T Corp Query evaluation on distributed semi-structured data
US6125370A (en) * 1998-04-01 2000-09-26 International Business Machines Corporation Repartitioning data
US6701516B1 (en) 1998-05-21 2004-03-02 Qifang Li P++ software
US6625593B1 (en) * 1998-06-29 2003-09-23 International Business Machines Corporation Parallel query optimization strategies for replicated and partitioned tables
US7197570B2 (en) * 1998-07-22 2007-03-27 Appstream Inc. System and method to send predicted application streamlets to a client device
US20010044850A1 (en) 1998-07-22 2001-11-22 Uri Raz Method and apparatus for determining the order of streaming modules
US6606617B1 (en) 1998-09-24 2003-08-12 International Business Machines Corporation Optimized technique for prefetching LOB table space pages
US6343293B1 (en) 1998-09-24 2002-01-29 International Business Machines Corporation Storing the uncompressed data length in a LOB map to speed substring access within a LOB value
US6366902B1 (en) 1998-09-24 2002-04-02 International Business Machines Corp. Using an epoch number to optimize access with rowid columns and direct row access
US6363389B1 (en) 1998-09-24 2002-03-26 International Business Machines Corporation Technique for creating a unique quasi-random row identifier
US6470359B1 (en) 1998-09-24 2002-10-22 International Business Machines Corporation Fast technique for recovering an index on an auxiliary table
US6694340B1 (en) 1998-09-24 2004-02-17 International Business Machines Corporation Technique for determining the age of the oldest reading transaction with a database object
US6343286B1 (en) 1998-09-24 2002-01-29 International Business Machines Corporation Efficient technique to defer large object access with intermediate results
US6144970A (en) * 1998-09-24 2000-11-07 International Business Machines Corporation Technique for inplace reorganization of a LOB table space
US6360218B1 (en) * 1998-10-26 2002-03-19 Microsoft Corporation Compact record format for low-overhead databases
US6718320B1 (en) 1998-11-02 2004-04-06 International Business Machines Corporation Schema mapping system and method
US6546381B1 (en) 1998-11-02 2003-04-08 International Business Machines Corporation Query optimization system and method
US6886012B1 (en) 1998-11-18 2005-04-26 International Business Machines Corporation Providing traditional update semantics when updates change the location of data records
US6629132B1 (en) * 1998-12-23 2003-09-30 Novell, Inc. Predicate indexing of data stored in a computer with application to indexing cached data
US7047232B1 (en) * 1999-01-13 2006-05-16 Ab Initio Software Corporation Parallelizing applications of script-driven tools
US6408299B1 (en) * 1999-01-28 2002-06-18 International Business Machines Corporation Type convertor registry
DE19910537A1 (de) * 1999-03-09 2000-09-14 Siemens Ag Automatisierungssystem mit Automatisierungsobjekten mit Verzeichnisstruktur und Verfahren zur Verwaltung von Automatisierungsobjekten in einer Verzeichnisstruktur
US6836768B1 (en) * 1999-04-27 2004-12-28 Surfnotes Method and apparatus for improved information representation
EP1049030A1 (en) * 1999-04-28 2000-11-02 SER Systeme AG Produkte und Anwendungen der Datenverarbeitung Classification method and apparatus
NO992269D0 (no) * 1999-05-10 1999-05-10 Fast Search & Transfer Asa S°kemotor med todimensjonalt skalerbart, parallell arkitektur
US6938147B1 (en) 1999-05-11 2005-08-30 Sun Microsystems, Inc. Processor with multiple-thread, vertically-threaded pipeline
US6542991B1 (en) 1999-05-11 2003-04-01 Sun Microsystems, Inc. Multiple-thread processor with single-thread interface shared among threads
US6351808B1 (en) 1999-05-11 2002-02-26 Sun Microsystems, Inc. Vertically and horizontally threaded processor with multidimensional storage for storing thread data
US6507862B1 (en) 1999-05-11 2003-01-14 Sun Microsystems, Inc. Switching method in a multi-threaded processor
US6341347B1 (en) 1999-05-11 2002-01-22 Sun Microsystems, Inc. Thread switch logic in a multiple-thread processor
US6738755B1 (en) * 1999-05-19 2004-05-18 International Business Machines Corporation Query optimization method for incrementally estimating the cardinality of a derived relation when statistically correlated predicates are applied
US6421663B1 (en) 1999-06-14 2002-07-16 International Business Machines Corporation Optimization of joined table expressions by extended access path selection
US6397204B1 (en) * 1999-06-25 2002-05-28 International Business Machines Corporation Method, system, and program for determining the join ordering of tables in a join query
US6446063B1 (en) * 1999-06-25 2002-09-03 International Business Machines Corporation Method, system, and program for performing a join operation on a multi column table and satellite tables
US6374235B1 (en) * 1999-06-25 2002-04-16 International Business Machines Corporation Method, system, and program for a join operation on a multi-column table and satellite tables including duplicate values
CA2279028C (en) * 1999-07-29 2002-09-10 Ibm Canada Limited-Ibm Canada Limitee Dropped database table recovery
US6408292B1 (en) * 1999-08-04 2002-06-18 Hyperroll, Israel, Ltd. Method of and system for managing multi-dimensional databases using modular-arithmetic based address data mapping processes on integer-encoded business dimensions
US6385604B1 (en) * 1999-08-04 2002-05-07 Hyperroll, Israel Limited Relational database management system having integrated non-relational multi-dimensional data store of aggregated data elements
JP3634681B2 (ja) * 1999-08-17 2005-03-30 日本電信電話株式会社 検索要求並列処理方法,及びその方法の実現に用いられるプログラム記録媒体
JP2001084257A (ja) * 1999-09-13 2001-03-30 Hitachi Ltd 問合せ処理方法及びシステム
US6466933B1 (en) 1999-09-21 2002-10-15 International Business Machines Corporation Delayed delivery of query results or other data from a federated server to a federated client until such information is needed
US6792416B2 (en) 1999-09-21 2004-09-14 International Business Machines Corporation Managing results of federated searches across heterogeneous datastores with a federated result set cursor object
US6748389B1 (en) 1999-09-21 2004-06-08 International Business Machines Corporation Method, system, and program for inverting columns in a database table
US7113939B2 (en) * 1999-09-21 2006-09-26 International Business Machines Corporation Architecture to enable search gateways as part of federated search
US6920443B1 (en) 1999-09-21 2005-07-19 International Business Machines, Corporation Method, system, program, and data structure for transforming database tables
US6965888B1 (en) 1999-09-21 2005-11-15 International Business Machines Corporation Method, system, program, and data structure for cleaning a database table using a look-up table
US6604095B1 (en) 1999-09-21 2003-08-05 International Business Machines Corporation Method, system, program, and data structure for pivoting columns in a database table
US7120638B1 (en) 1999-09-21 2006-10-10 International Business Machines Corporation Method, system, program, and data structure for cleaning a database table
US7197491B1 (en) 1999-09-21 2007-03-27 International Business Machines Corporation Architecture and implementation of a dynamic RMI server configuration hierarchy to support federated search and update across heterogeneous datastores
US6370541B1 (en) 1999-09-21 2002-04-09 International Business Machines Corporation Design and implementation of a client/server framework for federated multi-search and update across heterogeneous datastores
US6598058B2 (en) * 1999-09-22 2003-07-22 International Business Machines Corporation Method and apparatus for cross-node sharing of cached dynamic SQL in a multiple relational database management system environment
US6353820B1 (en) * 1999-09-29 2002-03-05 Bull Hn Information Systems Inc. Method and system for using dynamically generated code to perform index record retrieval in certain circumstances in a relational database manager
US6353819B1 (en) * 1999-09-29 2002-03-05 Bull Hn Information Systems Inc. Method and system for using dynamically generated code to perform record management layer functions in a relational database manager
US6493710B1 (en) * 1999-10-04 2002-12-10 Oracle Corporation Method and apparatus for reducing costs associated with manipulating data
US6438538B1 (en) * 1999-10-07 2002-08-20 International Business Machines Corporation Data replication in data warehousing scenarios
US7487439B1 (en) * 1999-10-19 2009-02-03 International Business Machines Corporation Method and apparatus for converting between data sets and XML documents
JP3608993B2 (ja) * 1999-11-10 2005-01-12 富士通株式会社 コンパイラ装置及びコンパイラプログラムを記録した記録媒体
JP4206586B2 (ja) * 1999-11-12 2009-01-14 株式会社日立製作所 データベース管理方法および装置並びにデータベース管理プログラムを記録した記憶媒体
DE19957594B4 (de) * 1999-11-30 2004-08-26 OCé PRINTING SYSTEMS GMBH Verfahren zum Synchronisieren von threads eines Computerprogramms
US7069264B2 (en) * 1999-12-08 2006-06-27 Ncr Corp. Stratified sampling of data in a database system
US6457029B1 (en) * 1999-12-22 2002-09-24 International Business Machines Corporation Computer method and system for same document lookup with different keywords from a single view
US6546403B1 (en) 2000-01-19 2003-04-08 International Business Machines Corporation Mechanism to resubmit queries in a parallel database system
ES2208164T3 (es) * 2000-02-23 2004-06-16 Ser Solutions, Inc Metodo y aparato para procesar documentos electronicos.
US20020029207A1 (en) * 2000-02-28 2002-03-07 Hyperroll, Inc. Data aggregation server for managing a multi-dimensional database and database management system having data aggregation server integrated therein
US6957209B1 (en) * 2000-02-29 2005-10-18 Unisys Corporation Sizing servers for database management systems via user defined workloads
GB2368666B (en) * 2000-03-29 2004-10-06 Jarg Corp Knowledge extraction system and method
US6473763B1 (en) 2000-03-31 2002-10-29 International Business Machines Corporation System, method and computer program for filtering multi-action rule set
US6567803B1 (en) * 2000-05-31 2003-05-20 Ncr Corporation Simultaneous computation of multiple moving aggregates in a relational database management system
US6505189B1 (en) * 2000-06-15 2003-01-07 Ncr Corporation Aggregate join index for relational databases
US6505188B1 (en) * 2000-06-15 2003-01-07 Ncr Corporation Virtual join index for relational databases
US7246111B1 (en) 2000-06-30 2007-07-17 Ncr Corporation Capturing database system information
US7185000B1 (en) * 2000-06-30 2007-02-27 Ncr Corp. Method and apparatus for presenting query plans
US6778534B1 (en) 2000-06-30 2004-08-17 E. Z. Chip Technologies Ltd. High-performance network processor
US6738756B1 (en) * 2000-06-30 2004-05-18 Ncr Corporation Analysis method and apparatus for a parallel system
US9177828B2 (en) 2011-02-10 2015-11-03 Micron Technology, Inc. External gettering method and device
US9252955B2 (en) * 2000-08-18 2016-02-02 United States Postal Service Apparatus and methods for the secure transfer of electronic data
EP1182577A1 (en) * 2000-08-18 2002-02-27 SER Systeme AG Produkte und Anwendungen der Datenverarbeitung Associative memory
US7302582B2 (en) 2000-08-21 2007-11-27 United States Postal Service Delivery point validation system
US6598041B1 (en) 2000-09-07 2003-07-22 International Business Machines Corporation Method, system, and program for processing modifications to data in tables in a database system
US7565394B1 (en) 2000-09-07 2009-07-21 ReportEdge, LLC Distributed report processing system and methods
US6757894B2 (en) * 2000-09-26 2004-06-29 Appstream, Inc. Preprocessed applications suitable for network streaming applications and method for producing same
US20020087717A1 (en) * 2000-09-26 2002-07-04 Itzik Artzi Network streaming of multi-application program code
US6748377B1 (en) * 2000-10-18 2004-06-08 International Business Machines Corporation Facilitating query pushdown in a multi-tiered database environment
US7080101B1 (en) * 2000-12-01 2006-07-18 Ncr Corp. Method and apparatus for partitioning data for storage in a database
US6713688B2 (en) * 2000-12-27 2004-03-30 Matsushita Electric Industrial Co., Ltd. Circuit board and its manufacture method
DE10104831A1 (de) * 2001-02-01 2002-08-08 Sap Ag Datenstruktur für Informationssysteme
US6721847B2 (en) * 2001-02-20 2004-04-13 Networks Associates Technology, Inc. Cache hints for computer file access
IL141599A0 (en) * 2001-02-22 2002-03-10 Infocyclone Inc Information retrieval system
US20020129145A1 (en) * 2001-03-06 2002-09-12 Accelerate Software Inc. Method and system for real-time querying, retrieval and integration of data from database over a computer network
US7634756B2 (en) 2001-03-22 2009-12-15 Robert Dean Bjornson Method and apparatus for dataflow creation and execution
US7614036B2 (en) * 2001-03-22 2009-11-03 Robert D Bjornson Method and system for dataflow creation and execution
US6691109B2 (en) * 2001-03-22 2004-02-10 Turbo Worx, Inc. Method and apparatus for high-performance sequence comparison
US7203678B1 (en) * 2001-03-27 2007-04-10 Bea Systems, Inc. Reconfigurable query generation system for web browsers
US6856996B2 (en) 2001-03-30 2005-02-15 International Business Machines Corporation Method, system, and program for accessing rows in one or more tables satisfying a search criteria
US9183317B1 (en) * 2001-06-20 2015-11-10 Microstrategy Incorporated System and method for exporting report results from a reporting system
US6792420B2 (en) * 2001-06-29 2004-09-14 International Business Machines Corporation Method, system, and program for optimizing the processing of queries involving set operators
ES2375403T3 (es) 2001-08-27 2012-02-29 BDGB Enterprise Software Sàrl Un método para la indexación automática de documentos.
US7213025B2 (en) * 2001-10-16 2007-05-01 Ncr Corporation Partitioned database system
AUPR894801A0 (en) * 2001-11-20 2001-12-13 Unisearch Limited A system and method for searching data sources
WO2003048972A2 (en) * 2001-12-06 2003-06-12 Schlumberger Systemes Method to query an embebbed database
US6907422B1 (en) * 2001-12-18 2005-06-14 Siebel Systems, Inc. Method and system for access and display of data from large data sets
CA2365433A1 (en) * 2001-12-19 2003-06-19 Alcatel Canada Inc. System and method for multiple-threaded access to a database
US7421436B2 (en) * 2001-12-21 2008-09-02 International Business Machines Corporation Decentralized many-to-many relationship management in an object persistence management system
CA2366196A1 (en) * 2001-12-21 2003-06-21 Ibm Canada Limited-Ibm Canada Limitee Unique identification of sql cursor occurrences in repetitive, nested environment
KR100850255B1 (ko) * 2001-12-28 2008-08-19 인터내셔널 비지네스 머신즈 코포레이션 실시간 데이터 웨어하우징
US7080093B2 (en) * 2002-01-14 2006-07-18 Sun Microsystems, Inc. System and method for database design
CA2372092C (en) * 2002-02-15 2010-04-06 Cognos Incorporated A queuing model for a plurality of servers
US20030158842A1 (en) * 2002-02-21 2003-08-21 Eliezer Levy Adaptive acceleration of retrieval queries
US8244702B2 (en) * 2002-02-26 2012-08-14 International Business Machines Corporation Modification of a data repository based on an abstract data representation
US6996558B2 (en) 2002-02-26 2006-02-07 International Business Machines Corporation Application portability and extensibility through database schema and query abstraction
US7398263B2 (en) * 2002-02-26 2008-07-08 International Business Machines Corporation Sequenced modification of multiple entities based on an abstract data representation
US7664731B2 (en) * 2002-03-21 2010-02-16 United States Postal Service Method and system for storing and retrieving data using hash-accessed multiple data stores
US7587408B2 (en) * 2002-03-21 2009-09-08 United States Postal Service Method and system for storing and retrieving data using hash-accessed multiple data stores
US7213011B1 (en) * 2002-04-08 2007-05-01 Oracle International Corporation Efficient processing of multi-column and function-based in-list predicates
US6954748B2 (en) * 2002-04-25 2005-10-11 International Business Machines Corporation Remote data access and integration of distributed data sources through data schema and query abstraction
US7010525B2 (en) * 2002-04-25 2006-03-07 International Business Machines Corporation Method and system for ensuring system awareness with data base connection on demand
US7010521B2 (en) * 2002-05-13 2006-03-07 Netezza Corporation Optimized database appliance
US6910032B2 (en) * 2002-06-07 2005-06-21 International Business Machines Corporation Parallel database query processing for non-uniform data sources via buffered access
US6915291B2 (en) 2002-06-07 2005-07-05 International Business Machines Corporation Object-oriented query execution data structure
US7089230B2 (en) * 2002-06-07 2006-08-08 International Business Machines Corporation Method for efficient processing of multi-state attributes
US6999958B2 (en) * 2002-06-07 2006-02-14 International Business Machines Corporation Runtime query optimization for dynamically selecting from multiple plans in a query based upon runtime-evaluated performance criterion
US7181460B2 (en) * 2002-06-18 2007-02-20 International Business Machines Corporation User-defined aggregate functions in database systems without native support
US7406469B1 (en) * 2002-06-20 2008-07-29 Oracle International Corporation Linear instance mapping for query rewrite
US6598044B1 (en) * 2002-06-25 2003-07-22 Microsoft Corporation Method for choosing optimal query execution plan for multiple defined equivalent query expressions
US6990483B2 (en) * 2002-07-08 2006-01-24 International Business Machines Corporation Method, system and program product for automatically retrieving documents
US20040019587A1 (en) * 2002-07-25 2004-01-29 You-Chin Fuh Method and device for processing a query in a database management system
US7159119B2 (en) * 2002-09-06 2007-01-02 United States Postal Service Method and system for efficiently retrieving secured data by securely pre-processing provided access information
US7047230B2 (en) * 2002-09-09 2006-05-16 Lucent Technologies Inc. Distinct sampling system and a method of distinct sampling for optimizing distinct value query estimates
WO2004027649A1 (en) * 2002-09-18 2004-04-01 Netezza Corporation Asymmetric streaming record data processor method and apparatus
US7096217B2 (en) 2002-10-31 2006-08-22 International Business Machines Corporation Global query correlation attributes
WO2004045123A1 (en) 2002-11-06 2004-05-27 International Business Machines Corporation Confidential data sharing and anonymous entity resolution
US7016845B2 (en) * 2002-11-08 2006-03-21 Oracle International Corporation Method and apparatus for providing speech recognition resolution on an application server
US7240059B2 (en) * 2002-11-14 2007-07-03 Seisint, Inc. System and method for configuring a parallel-processing database system
US7089356B1 (en) * 2002-11-21 2006-08-08 Oracle International Corporation Dynamic and scalable parallel processing of sequence operations
JP4161693B2 (ja) * 2002-11-25 2008-10-08 松下電器産業株式会社 マルチキャリア送信装置およびマルチキャリア受信装置ならびにマルチキャリア通信装置
US7293011B1 (en) * 2002-11-27 2007-11-06 Oracle International Corporation TQ distribution that increases parallism by distributing one slave to a particular data block
US7487140B2 (en) 2002-12-17 2009-02-03 International Business Machines Corporation Method for executing a query having multiple distinct key columns
US7146360B2 (en) * 2002-12-18 2006-12-05 International Business Machines Corporation Method and system for improving response time for database query execution
US8620937B2 (en) * 2002-12-27 2013-12-31 International Business Machines Corporation Real time data warehousing
US7158996B2 (en) 2003-01-27 2007-01-02 International Business Machines Corporation Method, system, and program for managing database operations with respect to a database table
US7657540B1 (en) 2003-02-04 2010-02-02 Seisint, Inc. Method and system for linking and delinking data records
US7054877B2 (en) * 2003-03-31 2006-05-30 International Business Machines Corporation Dealing with composite data through data model entities
CA2744925C (en) * 2003-04-08 2014-06-03 Grant L. Hutchison Method and system for executing a database query
US7464073B2 (en) * 2003-04-10 2008-12-09 International Business Machines Corporation Application of queries against incomplete schemas
US7392239B2 (en) * 2003-04-14 2008-06-24 International Business Machines Corporation System and method for querying XML streams
US7716187B2 (en) * 2003-05-21 2010-05-11 Microsoft Corporation System and method for transparent storage reorganization
CA2429910A1 (en) * 2003-05-27 2004-11-27 Cognos Incorporated System and method of query transformation
EP1482418A1 (en) * 2003-05-28 2004-12-01 Sap Ag A data processing method and system
US8112458B1 (en) 2003-06-17 2012-02-07 AudienceScience Inc. User segmentation user interface
US7966333B1 (en) 2003-06-17 2011-06-21 AudienceScience Inc. User segment population techniques
US7406714B1 (en) 2003-07-01 2008-07-29 Symantec Corporation Computer code intrusion detection system based on acceptable retrievals
US7568229B1 (en) 2003-07-01 2009-07-28 Symantec Corporation Real-time training for a computer code intrusion detection system
DE60330955D1 (de) 2003-07-08 2010-03-04 Sap Ag Verfahren und Computersystem zur Abfrageverarbeitung
US8490096B2 (en) * 2003-07-11 2013-07-16 Ca, Inc. Event processor for job scheduling and management
US7085757B2 (en) * 2003-07-11 2006-08-01 International Business Machines Corporation Abstract data linking and joining interface
US7337295B2 (en) * 2003-07-24 2008-02-26 Sap Aktiengesellschaft Memory management frame handler
US7310719B2 (en) 2003-07-24 2007-12-18 Sap Aktiengesellschaft Memory management tile optimization
KR100922141B1 (ko) * 2003-09-15 2009-10-19 아브 이니티오 소프트웨어 엘엘시 데이터 프로파일링 방법 및 시스템
US7899843B2 (en) * 2003-09-19 2011-03-01 International Business Machines Corporation Expanding the scope of an annotation to an entity level
US7835953B2 (en) * 2003-09-29 2010-11-16 International Business Machines Corporation Method and structure for monitoring moving objects
US7313554B2 (en) * 2003-09-29 2007-12-25 International Business Machines Corporation System and method for indexing queries, rules and subscriptions
US8972380B2 (en) * 2003-09-29 2015-03-03 International Business Machines Corporaton System and method for monitoring events against continual range queries
US7299126B2 (en) * 2003-11-03 2007-11-20 International Business Machines Corporation System and method for evaluating moving queries over moving objects
US7124142B2 (en) * 2003-11-10 2006-10-17 Conversive, Inc. Method and system for responding to requests relating to complex data maintained in a structured form
US7900133B2 (en) 2003-12-09 2011-03-01 International Business Machines Corporation Annotation structure type determination
US8321420B1 (en) * 2003-12-10 2012-11-27 Teradata Us, Inc. Partition elimination on indexed row IDs
US6944633B1 (en) * 2003-12-10 2005-09-13 Ncr Corporation Performing a join in a partitioned database system
US20050131893A1 (en) * 2003-12-15 2005-06-16 Sap Aktiengesellschaft Database early parallelism method and system
US7685095B2 (en) * 2003-12-16 2010-03-23 Oracle International Corporation Executing a parallel single cursor model
US7958160B2 (en) * 2003-12-16 2011-06-07 Oracle International Corporation Executing filter subqueries using a parallel single cursor model
US7451133B2 (en) * 2003-12-16 2008-11-11 Oracle International Corporation Executing nested subqueries of parallel table functions in the parallel single cursor model
US7340452B2 (en) * 2003-12-16 2008-03-04 Oracle International Corporation Parallel single cursor model on multiple-server configurations
US8086645B2 (en) * 2003-12-16 2011-12-27 Oracle International Corporation Compilation and processing a parallel single cursor model
US7376638B2 (en) * 2003-12-24 2008-05-20 International Business Machines Corporation System and method for addressing inefficient query processing
US20050160101A1 (en) * 2004-01-16 2005-07-21 International Business Machines Corporation Method and apparatus using dynamic SQL for item create, retrieve, update delete operations in a content management application
US8037102B2 (en) 2004-02-09 2011-10-11 Robert T. and Virginia T. Jenkins Manipulating sets of hierarchical data
US8782024B2 (en) * 2004-02-12 2014-07-15 International Business Machines Corporation Managing the sharing of logical resources among separate partitions of a logically partitioned computer system
US7254574B2 (en) * 2004-03-08 2007-08-07 Microsoft Corporation Structured indexes on results of function applications over data
US7406477B2 (en) * 2004-03-12 2008-07-29 Sybase, Inc. Database system with methodology for automated determination and selection of optimal indexes
US8266177B1 (en) 2004-03-16 2012-09-11 Symantec Corporation Empirical database access adjustment
US20050210023A1 (en) * 2004-03-18 2005-09-22 Renato Barrera Query optimizer using implied predicates
US7464405B2 (en) * 2004-03-25 2008-12-09 International Business Machines Corporation Method for preventing loading and execution of rogue operating systems in a logical partitioned data processing system
US7890497B2 (en) * 2004-04-14 2011-02-15 Oracle International Corporation Using estimated cost to schedule an order for refreshing a set of materialized views (MVS)
US7363303B2 (en) * 2004-05-28 2008-04-22 International Business Machines Corporation Multi-column multi-data type internationalized sort extension method for web applications
US9646107B2 (en) 2004-05-28 2017-05-09 Robert T. and Virginia T. Jenkins as Trustee of the Jenkins Family Trust Method and/or system for simplifying tree expressions such as for query reduction
US7356525B2 (en) * 2004-05-28 2008-04-08 International Business Machines Corporation Multi-column multi-data type internationalized sort extension method for web applications
US7370030B2 (en) * 2004-06-17 2008-05-06 International Business Machines Corporation Method to provide management of query output
US7590620B1 (en) * 2004-06-18 2009-09-15 Google Inc. System and method for analyzing data records
US7702627B2 (en) * 2004-06-22 2010-04-20 Oracle International Corporation Efficient interaction among cost-based transformations
US20050283471A1 (en) * 2004-06-22 2005-12-22 Oracle International Corporation Multi-tier query processing
ATE450011T1 (de) * 2004-06-23 2009-12-15 Sap Ag System und verfahren zur datenverarbeitung
US7814042B2 (en) * 2004-08-17 2010-10-12 Oracle International Corporation Selecting candidate queries
US7472186B2 (en) * 2004-09-09 2008-12-30 International Business Machines Corporation Method for using SNMP as an RPC mechanism for exporting the data structures of a remote library
US7574424B2 (en) * 2004-10-13 2009-08-11 Sybase, Inc. Database system with methodology for parallel schedule generation in a query optimizer
US20060100992A1 (en) * 2004-10-21 2006-05-11 International Business Machines Corporation Apparatus and method for data ordering for derived columns in a database system
US7627591B2 (en) 2004-10-29 2009-12-01 Skyler Technology, Inc. Method and/or system for manipulating tree expressions
US7801923B2 (en) 2004-10-29 2010-09-21 Robert T. and Virginia T. Jenkins as Trustees of the Jenkins Family Trust Method and/or system for tagging trees
US20060095406A1 (en) * 2004-10-29 2006-05-04 International Business Machines Corporation Displaying explain data for a SQL query of a database
US7346625B2 (en) * 2004-11-05 2008-03-18 International Business Machines Corporation Methods and apparatus for performing structural joins for answering containment queries
US20060101045A1 (en) * 2004-11-05 2006-05-11 International Business Machines Corporation Methods and apparatus for interval query indexing
US7752181B2 (en) * 2004-11-08 2010-07-06 Oracle International Corporation System and method for performing a data uniqueness check in a sorted data set
US7610272B2 (en) * 2004-11-29 2009-10-27 Sap Ag Materialized samples for a business warehouse query
US7630995B2 (en) 2004-11-30 2009-12-08 Skyler Technology, Inc. Method and/or system for transmitting and/or receiving data
US7636727B2 (en) 2004-12-06 2009-12-22 Skyler Technology, Inc. Enumeration of trees from finite number of nodes
US7801925B2 (en) * 2004-12-22 2010-09-21 United States Postal Service System and method for electronically processing address information
US8316059B1 (en) 2004-12-30 2012-11-20 Robert T. and Virginia T. Jenkins Enumeration of rooted partial subtrees
US7640237B2 (en) * 2005-01-11 2009-12-29 International Business Machines Corporation System and method for database query with on demand database query reduction
US8122012B2 (en) 2005-01-14 2012-02-21 International Business Machines Corporation Abstract record timeline rendering/display
US7499917B2 (en) * 2005-01-28 2009-03-03 International Business Machines Corporation Processing cross-table non-Boolean term conditions in database queries
US8615530B1 (en) 2005-01-31 2013-12-24 Robert T. and Virginia T. Jenkins as Trustees for the Jenkins Family Trust Method and/or system for tree transformation
US20080022136A1 (en) * 2005-02-18 2008-01-24 Protegrity Corporation Encryption load balancing and distributed policy enforcement
US20070174271A1 (en) * 2005-02-18 2007-07-26 Ulf Mattsson Database system with second preprocessor and method for accessing a database
US7681177B2 (en) 2005-02-28 2010-03-16 Skyler Technology, Inc. Method and/or system for transforming between trees and strings
US7444331B1 (en) 2005-03-02 2008-10-28 Symantec Corporation Detecting code injection attacks against databases
US7814044B2 (en) * 2005-03-22 2010-10-12 Sap Ag Data access service queries
US8126870B2 (en) * 2005-03-28 2012-02-28 Sybase, Inc. System and methodology for parallel query optimization using semantic-based partitioning
US7676467B1 (en) 2005-04-14 2010-03-09 AudienceScience Inc. User segment population techniques
US7809752B1 (en) 2005-04-14 2010-10-05 AudienceScience Inc. Representing user behavior information
US7610265B2 (en) 2005-04-29 2009-10-27 Sap Ag Data query verification
US7899821B1 (en) 2005-04-29 2011-03-01 Karl Schiffmann Manipulation and/or analysis of hierarchical data
US8046374B1 (en) 2005-05-06 2011-10-25 Symantec Corporation Automatic training of a database intrusion detection system
WO2006124910A2 (en) 2005-05-17 2006-11-23 United States Postal Service System and method for automated management of an address database
US7558796B1 (en) 2005-05-19 2009-07-07 Symantec Corporation Determining origins of queries for a database intrusion detection system
US7774361B1 (en) 2005-07-08 2010-08-10 Symantec Corporation Effective aggregation and presentation of database intrusion incidents
US7690037B1 (en) 2005-07-13 2010-03-30 Symantec Corporation Filtering training data for machine learning
US7984039B2 (en) * 2005-07-14 2011-07-19 International Business Machines Corporation Merging of results in distributed information retrieval
JP4611830B2 (ja) * 2005-07-22 2011-01-12 優 喜連川 データベース管理システム及び方法
US7475056B2 (en) * 2005-08-11 2009-01-06 Oracle International Corporation Query processing in a parallel single cursor model on multi-instance configurations, using hints
JP3962417B2 (ja) * 2005-08-31 2007-08-22 インターナショナル・ビジネス・マシーンズ・コーポレーション データベースに対するアクセスを制御するシステム、およびその方法
US20070067337A1 (en) * 2005-09-22 2007-03-22 Morris John M Method of managing retrieval of data objects from a storage device
US7814091B2 (en) * 2005-09-27 2010-10-12 Oracle International Corporation Multi-tiered query processing techniques for minus and intersect operators
US7617189B2 (en) * 2005-09-27 2009-11-10 Oracle International Corporation Parallel query processing techniques for minus and intersect operators
US7512700B2 (en) * 2005-09-30 2009-03-31 International Business Machines Corporation Real-time mining and reduction of streamed data
US7877379B2 (en) * 2005-09-30 2011-01-25 Oracle International Corporation Delaying evaluation of expensive expressions in a query
US7444332B2 (en) * 2005-11-10 2008-10-28 International Business Machines Corporation Strict validation of inference rule based on abstraction environment
US20070112827A1 (en) * 2005-11-10 2007-05-17 International Business Machines Corporation Abstract rule sets
US7440945B2 (en) * 2005-11-10 2008-10-21 International Business Machines Corporation Dynamic discovery of abstract rule set required inputs
US7415454B2 (en) * 2005-11-18 2008-08-19 Microsoft Corporation Allocation performance by query optimization
JP4200456B2 (ja) * 2005-12-28 2008-12-24 ブラザー工業株式会社 周辺装置、プログラム、制御方法
US20070150585A1 (en) * 2005-12-28 2007-06-28 Microsoft Corporation Multi-dimensional aggregation on event streams
US20070156734A1 (en) * 2005-12-30 2007-07-05 Stefan Dipper Handling ambiguous joins
US20070179982A1 (en) * 2006-01-31 2007-08-02 International Business Machines Corporation Temporal summarization of a data subset
US20070192215A1 (en) * 2006-02-10 2007-08-16 Taylor Thomas B Computer-implemented registration for providing inventory fulfillment services to merchants
US7809713B2 (en) * 2006-03-15 2010-10-05 Oracle International Corporation Efficient search space analysis for join factorization
US7945562B2 (en) * 2006-03-15 2011-05-17 Oracle International Corporation Join predicate push-down optimizations
US7676450B2 (en) * 2006-03-15 2010-03-09 Oracle International Corporation Null aware anti-join
US7644062B2 (en) * 2006-03-15 2010-01-05 Oracle International Corporation Join factorization of union/union all queries
US8151360B1 (en) 2006-03-20 2012-04-03 Netapp, Inc. System and method for administering security in a logical namespace of a storage system environment
US9118697B1 (en) 2006-03-20 2015-08-25 Netapp, Inc. System and method for integrating namespace management and storage management in a storage system environment
US20070226176A1 (en) * 2006-03-23 2007-09-27 International Business Machines Corporation Apparatus and method for optimizing a query to a partitioned database table using a virtual maintained temporary index that spans multiple database partitions
US7797312B2 (en) * 2006-04-06 2010-09-14 International Business Machines Corporation Database query processing method and system
US7680787B2 (en) * 2006-04-06 2010-03-16 International Business Machines Corporation Database query generation method and system
US7698351B1 (en) 2006-04-28 2010-04-13 Netapp, Inc. GUI architecture for namespace and storage management
US8635247B1 (en) * 2006-04-28 2014-01-21 Netapp, Inc. Namespace and storage management application infrastructure for use in management of resources in a storage system environment
US8018471B2 (en) * 2006-05-15 2011-09-13 Microsoft Corporation Visual component/clause merging
US8423569B2 (en) 2006-08-09 2013-04-16 International Business Machines Corporation Decomposed query conditions
US7792819B2 (en) * 2006-08-31 2010-09-07 International Business Machines Corporation Priority reduction for fast partitions during query execution
US8799266B2 (en) * 2006-08-31 2014-08-05 International Business Machines Corporation Method and system for managing operation of a user-defined function on a partitioned database
US7831620B2 (en) * 2006-08-31 2010-11-09 International Business Machines Corporation Managing execution of a query against a partitioned database
US7962442B2 (en) * 2006-08-31 2011-06-14 International Business Machines Corporation Managing execution of a query against selected data partitions of a partitioned database
US20080071755A1 (en) * 2006-08-31 2008-03-20 Barsness Eric L Re-allocation of resources for query execution in partitions
JP2008112934A (ja) * 2006-10-31 2008-05-15 Oki Electric Ind Co Ltd 半導体記憶装置及びその製造方法
US20100312787A1 (en) * 2006-11-04 2010-12-09 Nokia Corporation Control of Search Scope
US8204831B2 (en) 2006-11-13 2012-06-19 International Business Machines Corporation Post-anonymous fuzzy comparisons without the use of pre-anonymization variants
US8019771B2 (en) * 2006-11-30 2011-09-13 International Business Machines Corporation Method for dynamically finding relations between database tables
US8150870B1 (en) * 2006-12-22 2012-04-03 Amazon Technologies, Inc. Scalable partitioning in a multilayered data service framework
US7672925B2 (en) * 2006-12-28 2010-03-02 Sybase, Inc. Accelerating queries using temporary enumeration representation
US7941414B2 (en) * 2007-03-28 2011-05-10 Microsoft Corporation Executing non-blocking parallel scans
US8065319B2 (en) * 2007-04-01 2011-11-22 Nec Laboratories America, Inc. Runtime semantic query optimization for event stream processing
US8140557B2 (en) 2007-05-15 2012-03-20 International Business Machines Corporation Ontological translation of abstract rules
US9037493B2 (en) * 2007-05-17 2015-05-19 Oracle International Corporation Security using EPCIS data and a virtual private database
US7853480B2 (en) * 2007-05-21 2010-12-14 Amazon Technologies, Inc. System and method for providing export services to merchants
US20100031321A1 (en) 2007-06-11 2010-02-04 Protegrity Corporation Method and system for preventing impersonation of computer system user
US8074219B2 (en) * 2007-06-27 2011-12-06 Microsoft Corporation Order preservation in data parallel operations
US8201171B2 (en) * 2007-06-27 2012-06-12 Microsoft Corporation Adjacent data parallel and streaming operator fusion
US8832073B2 (en) * 2007-06-29 2014-09-09 Alcatel Lucent Method and apparatus for efficient aggregate computation over data streams
US7676461B2 (en) 2007-07-18 2010-03-09 Microsoft Corporation Implementation of stream algebra over class instances
US8805799B2 (en) 2007-08-07 2014-08-12 International Business Machines Corporation Dynamic partial uncompression of a database table
US7747585B2 (en) * 2007-08-07 2010-06-29 International Business Machines Corporation Parallel uncompression of a partially compressed database table determines a count of uncompression tasks that satisfies the query
US20090083238A1 (en) * 2007-09-21 2009-03-26 Microsoft Corporation Stop-and-restart style execution for long running decision support queries
US20090112932A1 (en) * 2007-10-26 2009-04-30 Microsoft Corporation Visualizing key performance indicators for model-based applications
US8438152B2 (en) * 2007-10-29 2013-05-07 Oracle International Corporation Techniques for bushy tree execution plans for snowstorm schema
US8943057B2 (en) * 2007-12-12 2015-01-27 Oracle America, Inc. Method and system for distributed bulk matching and loading
US9292567B2 (en) 2007-12-12 2016-03-22 Oracle International Corporation Bulk matching with update
US8051091B2 (en) * 2008-02-19 2011-11-01 Sap Ag Parallelizing data manipulation by data set abstraction
US8181155B2 (en) * 2008-02-29 2012-05-15 Microsoft Corporation Unified expression and location framework
US8108401B2 (en) * 2008-03-28 2012-01-31 International Business Machines Corporation Applying various hash methods used in conjunction with a query with a group by clause
US20090248631A1 (en) * 2008-03-31 2009-10-01 International Business Machines Corporation System and Method for Balancing Workload of a Database Based Application by Partitioning Database Queries
US8266168B2 (en) 2008-04-24 2012-09-11 Lexisnexis Risk & Information Analytics Group Inc. Database systems and methods for linking records and entity representations with sufficiently high confidence
US8386508B2 (en) * 2008-04-28 2013-02-26 Infosys Technologies Limited System and method for parallel query evaluation
US8290930B2 (en) * 2008-06-02 2012-10-16 Microsoft Corporation Query result generation based on query category and data source category
US8290917B2 (en) * 2008-06-02 2012-10-16 Microsoft Corporation Reordering of data elements in a data parallel system
US8364751B2 (en) 2008-06-25 2013-01-29 Microsoft Corporation Automated client/server operation partitioning
CA2748625C (en) 2008-07-02 2015-10-27 Lexisnexis Risk Solutions Fl Inc. Entity representation identification based on a search query using field match templates
US8195645B2 (en) * 2008-07-23 2012-06-05 International Business Machines Corporation Optimized bulk computations in data warehouse environments
US8140522B2 (en) * 2008-08-12 2012-03-20 International Business Machines Corporation Method, apparatus, and computer program product for adaptive query parallelism partitioning with look-ahead probing and feedback
US8150889B1 (en) * 2008-08-28 2012-04-03 Amazon Technologies, Inc. Parallel processing framework
US8984390B2 (en) 2008-09-15 2015-03-17 Palantir Technologies, Inc. One-click sharing for screenshots and related documents
US20100088325A1 (en) 2008-10-07 2010-04-08 Microsoft Corporation Streaming Queries
US20100115246A1 (en) * 2008-10-31 2010-05-06 Yahoo! Inc. System and method of data partitioning for parallel processing of dynamically generated application data
US8682998B2 (en) * 2008-10-31 2014-03-25 Software Ag Method and server cluster for map reducing flow services and large documents
US9244793B1 (en) 2008-11-04 2016-01-26 Teradata Us, Inc. Using target database system statistics in emulation
US20100114929A1 (en) * 2008-11-06 2010-05-06 Yahoo! Inc. Diverse query recommendations using clustering-based methodology
US8943040B1 (en) * 2008-11-25 2015-01-27 Teradota Us, Inc. Using a table to track whether a dynamic partitioning condition evaluates to true
US8046394B1 (en) 2008-11-25 2011-10-25 Teradata Us, Inc. Dynamic partitioning for an ordered analytic function
US8412734B2 (en) * 2008-12-30 2013-04-02 International Business Machines Corporation Unifying hetrogenous data
US9519712B2 (en) * 2009-01-06 2016-12-13 At&T Intellectual Property I, L.P. Systems and methods to evaluate search qualities
US8161076B1 (en) * 2009-04-02 2012-04-17 Netapp, Inc. Generation and use of a data structure for distributing responsibilities among multiple resources in a network storage system
US8266172B2 (en) * 2009-04-03 2012-09-11 Microsoft Corporation Data parallel query analysis
US9152883B2 (en) * 2009-11-02 2015-10-06 Harry Urbschat System and method for increasing the accuracy of optical character recognition (OCR)
US8321357B2 (en) * 2009-09-30 2012-11-27 Lapir Gennady Method and system for extraction
US9213756B2 (en) * 2009-11-02 2015-12-15 Harry Urbschat System and method of using dynamic variance networks
US9158833B2 (en) * 2009-11-02 2015-10-13 Harry Urbschat System and method for obtaining document information
US8832074B2 (en) * 2009-10-21 2014-09-09 Teradata Us, Inc. System, method and computer-readable medium for dynamic skew avoidance for generic queries
US9158816B2 (en) 2009-10-21 2015-10-13 Microsoft Technology Licensing, Llc Event processing with XML query based on reusable XML query template
US8566341B2 (en) * 2009-11-12 2013-10-22 Oracle International Corporation Continuous aggregation on a data grid
US8688683B2 (en) * 2009-11-30 2014-04-01 Business Objects Software Ltd. Query plan reformulation
CA2686993A1 (en) * 2009-12-03 2011-06-03 Ibm Canada Limited - Ibm Canada Limitee Semantic verification of multidimensional data sources
US9411859B2 (en) 2009-12-14 2016-08-09 Lexisnexis Risk Solutions Fl Inc External linking based on hierarchical level weightings
US8392920B2 (en) * 2010-01-27 2013-03-05 Microsoft Corporation Parallel query engine with dynamic number of workers
CN102141995B (zh) * 2010-01-29 2013-06-12 国际商业机器公司 简化并行计算系统中的传输的系统与方法
CN102844754B (zh) * 2010-03-30 2015-11-25 惠普发展公司,有限责任合伙企业 用于对并行数据库管理系统的外连接的系统和方法
US10838957B2 (en) 2010-06-17 2020-11-17 Microsoft Technology Licensing, Llc Slicing relational queries using spool operators
US8332388B2 (en) * 2010-06-18 2012-12-11 Microsoft Corporation Transformation rule profiling for a query optimizer
US20110314001A1 (en) * 2010-06-18 2011-12-22 Microsoft Corporation Performing query expansion based upon statistical analysis of structured data
US8868484B2 (en) * 2010-07-08 2014-10-21 Oracle International Corporation Efficiently updating rows in a data warehouse
US20120011144A1 (en) 2010-07-12 2012-01-12 Frederik Transier Aggregation in parallel computation environments with shared memory
US9189505B2 (en) 2010-08-09 2015-11-17 Lexisnexis Risk Data Management, Inc. System of and method for entity representation splitting without the need for human interaction
US8903805B2 (en) * 2010-08-20 2014-12-02 Oracle International Corporation Method and system for performing query optimization using a hybrid execution plan
US8645338B2 (en) 2010-10-28 2014-02-04 International Business Machines Corporation Active memory expansion and RDBMS meta data and tooling
US8442988B2 (en) 2010-11-04 2013-05-14 International Business Machines Corporation Adaptive cell-specific dictionaries for frequency-partitioned multi-dimensional data
WO2012074529A1 (en) * 2010-12-03 2012-06-07 Hewlett-Packard Development Company, L.P. Systems and methods for performing a nested join operation
US9355145B2 (en) 2011-01-25 2016-05-31 Hewlett Packard Enterprise Development Lp User defined function classification in analytical data processing systems
US8538954B2 (en) 2011-01-25 2013-09-17 Hewlett-Packard Development Company, L.P. Aggregate function partitions for distributed processing
US8856151B2 (en) 2011-01-25 2014-10-07 Hewlett-Packard Development Company, L.P. Output field mapping of user defined functions in databases
CN103348598B (zh) 2011-01-28 2017-07-14 起元科技有限公司 生成数据模式信息
US9020953B1 (en) * 2011-03-08 2015-04-28 Pmc-Sierra Us, Inc. Search table for data networking matching
US8392408B1 (en) * 2011-05-04 2013-03-05 Google Inc. Coordinating successive search queries using a query cursor
US20120310975A1 (en) * 2011-05-31 2012-12-06 Nokia Corporation Method and apparatus for dynamic partitioning of data into data silos
US9092482B2 (en) 2013-03-14 2015-07-28 Palantir Technologies, Inc. Fair scheduling for mixed-query loads
US8732574B2 (en) 2011-08-25 2014-05-20 Palantir Technologies, Inc. System and method for parameterizing documents for automatic workflow generation
US8504542B2 (en) 2011-09-02 2013-08-06 Palantir Technologies, Inc. Multi-row transactions
US8380701B1 (en) * 2011-09-02 2013-02-19 International Business Machines Corporation Using a partially built index in a computer database system
US8959110B2 (en) 2011-09-18 2015-02-17 Microsoft Technology Licensing, Llc Dynamic query for external data connections
US10331664B2 (en) * 2011-09-23 2019-06-25 Hartford Fire Insurance Company System and method of insurance database optimization using social networking
US8694486B2 (en) 2011-09-27 2014-04-08 International Business Machines Corporation Deadline-driven parallel execution of queries
US8433702B1 (en) * 2011-09-28 2013-04-30 Palantir Technologies, Inc. Horizon histogram optimizations
US8560494B1 (en) 2011-09-30 2013-10-15 Palantir Technologies, Inc. Visual data importer
US8826284B1 (en) 2011-10-31 2014-09-02 Google Inc. Scalable task scheduling
US9110933B1 (en) * 2011-11-04 2015-08-18 Google Inc. Processing data triggers in an untrusted environment based on information stored in a trusted environment
US8862588B1 (en) 2011-11-30 2014-10-14 Google Inc. Generating an empirically-determined schema for a schemaless database
US9148329B1 (en) 2011-11-30 2015-09-29 Google Inc. Resource constraints for request processing
US8938444B2 (en) * 2011-12-29 2015-01-20 Teradata Us, Inc. Techniques for external application-directed data partitioning in data exporting from a database management system
US8938480B2 (en) * 2011-12-29 2015-01-20 Teradata Us, Inc. Techniques for fast loading of data from an external distributed file system to a database management system
US20130173662A1 (en) * 2012-01-03 2013-07-04 International Business Machines Corporation Dependency based prioritization of sub-queries and placeholder resolution
US9317470B1 (en) 2012-01-13 2016-04-19 Google Inc. Method and system for incremental cache lookup and insertion
US8898796B2 (en) 2012-02-14 2014-11-25 International Business Machines Corporation Managing network data
US9378526B2 (en) 2012-03-02 2016-06-28 Palantir Technologies, Inc. System and method for accessing data objects via remote references
US9235607B1 (en) 2012-03-29 2016-01-12 Google Inc. Specifying a predetermined degree of inconsistency for test data
US8762407B2 (en) * 2012-04-17 2014-06-24 Renmin University Of China Concurrent OLAP-oriented database query processing method
US10417227B2 (en) * 2012-04-27 2019-09-17 Hitachi, Ltd. Database management system, computer, and database management method
US9342553B1 (en) * 2012-05-13 2016-05-17 Google Inc. Identifying distinct combinations of values for entities based on information in an index
US9798768B2 (en) 2012-09-10 2017-10-24 Palantir Technologies, Inc. Search around visual queries
US8635373B1 (en) * 2012-09-22 2014-01-21 Nest Labs, Inc. Subscription-Notification mechanisms for synchronization of distributed states
US9471370B2 (en) 2012-10-22 2016-10-18 Palantir Technologies, Inc. System and method for stack-based batch evaluation of program instructions
CA2887670C (en) 2012-10-22 2021-05-25 Ab Initio Technology Llc Profiling data with location information
US9348677B2 (en) 2012-10-22 2016-05-24 Palantir Technologies Inc. System and method for batch evaluation programs
US8892599B2 (en) * 2012-10-24 2014-11-18 Marklogic Corporation Apparatus and method for securing preliminary information about database fragments for utilization in mapreduce processing
US9152672B2 (en) * 2012-12-17 2015-10-06 General Electric Company Method for storage, querying, and analysis of time series data
US9152671B2 (en) * 2012-12-17 2015-10-06 General Electric Company System for storage, querying, and analysis of time series data
WO2014109109A1 (ja) * 2013-01-11 2014-07-17 日本電気株式会社 インデックスキー生成装置及びインデックスキー生成方法並びに検索方法
US9892026B2 (en) 2013-02-01 2018-02-13 Ab Initio Technology Llc Data records selection
US9367463B2 (en) 2013-03-14 2016-06-14 Palantir Technologies, Inc. System and method utilizing a shared cache to provide zero copy memory mapped database
US9665621B1 (en) * 2013-03-14 2017-05-30 EMC IP Holding Company LLC Accelerated query execution within a storage array
US8909656B2 (en) 2013-03-15 2014-12-09 Palantir Technologies Inc. Filter chains with associated multipath views for exploring large data sets
US8868486B2 (en) 2013-03-15 2014-10-21 Palantir Technologies Inc. Time-sensitive cube
US8903717B2 (en) 2013-03-15 2014-12-02 Palantir Technologies Inc. Method and system for generating a parser and parsing complex data
US9740369B2 (en) 2013-03-15 2017-08-22 Palantir Technologies Inc. Systems and methods for providing a tagging interface for external content
US9218379B2 (en) * 2013-03-15 2015-12-22 Informatica Llc Method, apparatus, and computer-readable medium for efficiently performing operations on distinct data values
US8930897B2 (en) 2013-03-15 2015-01-06 Palantir Technologies Inc. Data integration tool
US8855999B1 (en) 2013-03-15 2014-10-07 Palantir Technologies Inc. Method and system for generating a parser and parsing complex data
US9898167B2 (en) 2013-03-15 2018-02-20 Palantir Technologies Inc. Systems and methods for providing a tagging interface for external content
US9329899B2 (en) 2013-06-24 2016-05-03 Sap Se Parallel execution of parsed query based on a concurrency level corresponding to an average number of available worker threads
US9105000B1 (en) 2013-12-10 2015-08-11 Palantir Technologies Inc. Aggregating data from a plurality of data sources
IL229907A (en) * 2013-12-10 2015-02-26 David Almer Mobile device with enhanced security
US11487732B2 (en) 2014-01-16 2022-11-01 Ab Initio Technology Llc Database key identification
US10002254B2 (en) * 2014-01-20 2018-06-19 Prevoty, Inc. Systems and methods for SQL type evaluation to detect evaluation flaws
US11100218B2 (en) 2014-01-20 2021-08-24 Prevoty, Inc. Systems and methods for improving accuracy in recognizing and neutralizing injection attacks in computer services
US10025936B2 (en) * 2014-01-20 2018-07-17 Prevoty, Inc. Systems and methods for SQL value evaluation to detect evaluation flaws
US10635669B1 (en) 2014-01-27 2020-04-28 Microstrategy Incorporated Data engine integration and data refinement
US10255320B1 (en) 2014-01-27 2019-04-09 Microstrategy Incorporated Search integration
US11921715B2 (en) 2014-01-27 2024-03-05 Microstrategy Incorporated Search integration
US11386085B2 (en) 2014-01-27 2022-07-12 Microstrategy Incorporated Deriving metrics from queries
US9952894B1 (en) * 2014-01-27 2018-04-24 Microstrategy Incorporated Parallel query processing
US9953074B2 (en) 2014-01-31 2018-04-24 Sap Se Safe synchronization of parallel data operator trees
US9870390B2 (en) 2014-02-18 2018-01-16 Oracle International Corporation Selecting from OR-expansion states of a query
CN103870340B (zh) * 2014-03-06 2017-11-07 华为技术有限公司 流计算系统中的数据处理方法、控制节点及流计算系统
US10114674B2 (en) * 2014-03-06 2018-10-30 International Business Machines Corporation Sorting database collections for parallel processing
KR102361153B1 (ko) 2014-03-07 2022-02-09 아브 이니티오 테크놀로지 엘엘시 데이터 유형에 관련된 데이터 프로파일링 동작 관리
WO2015137919A1 (en) * 2014-03-10 2015-09-17 Hewlett-Packard Development Company, L.P. Parallelizing sql user defined transformation functions
US8924429B1 (en) 2014-03-18 2014-12-30 Palantir Technologies Inc. Determining and extracting changed data from a data source
US9785669B2 (en) 2014-05-21 2017-10-10 International Business Machines Corporation Revising policy statements using hyperlinks
CN104021161B (zh) * 2014-05-27 2018-06-15 华为技术有限公司 一种聚簇存储方法及装置
KR101679011B1 (ko) * 2014-06-26 2016-11-24 주식회사 알티베이스 데이터베이스에서 데이터 이동을 처리하는 방법 및 장치
US20160026923A1 (en) 2014-07-22 2016-01-28 Palantir Technologies Inc. System and method for determining a propensity of entity to take a specified action
US10229208B2 (en) * 2014-07-28 2019-03-12 Facebook, Inc. Optimization of query execution
US10740331B2 (en) * 2014-08-07 2020-08-11 Coupang Corp. Query execution apparatus, method, and system for processing data, query containing a composite primitive
US9710532B2 (en) * 2014-08-21 2017-07-18 Codership Oy Method for avoiding conflicts in database cluster
US10585887B2 (en) 2015-03-30 2020-03-10 Oracle International Corporation Multi-system query execution plan
US9348880B1 (en) 2015-04-01 2016-05-24 Palantir Technologies, Inc. Federated search of multiple sources with conflict resolution
US20160342646A1 (en) * 2015-05-20 2016-11-24 International Business Machines Corporation Database query cursor management
US10503706B2 (en) * 2015-06-01 2019-12-10 Sap Se Deferred data definition statements
US10642833B2 (en) 2015-08-11 2020-05-05 Sybase, Inc. Accelerating database queries using composite union enumeration
US10467228B2 (en) 2015-08-11 2019-11-05 Sybase, Inc. Accelerating database queries using equivalence union enumeration
US9514205B1 (en) 2015-09-04 2016-12-06 Palantir Technologies Inc. Systems and methods for importing data from electronic data files
US10558339B1 (en) 2015-09-11 2020-02-11 Palantir Technologies Inc. System and method for analyzing electronic communications and a collaborative electronic communications user interface
US9772934B2 (en) 2015-09-14 2017-09-26 Palantir Technologies Inc. Pluggable fault detection tests for data pipelines
US10073871B2 (en) 2015-11-09 2018-09-11 International Business Machines Corporation Database entity analysis
US10095738B1 (en) * 2015-12-22 2018-10-09 Amazon Technologies, Inc. Dynamic assignment of logical partitions according to query predicate evaluations
US9652510B1 (en) 2015-12-29 2017-05-16 Palantir Technologies Inc. Systems and user interfaces for data analysis including artificial intelligence algorithms for generating optimized packages of data items
EP3188038B1 (en) * 2015-12-31 2020-11-04 Dassault Systèmes Evaluation of a training set
US10554516B1 (en) 2016-06-09 2020-02-04 Palantir Technologies Inc. System to collect and visualize software usage metrics
US9678850B1 (en) 2016-06-10 2017-06-13 Palantir Technologies Inc. Data pipeline monitoring
US10997175B2 (en) * 2016-06-24 2021-05-04 Teradata Us, Inc. Method for predicate evaluation in relational database systems
US10133782B2 (en) 2016-08-01 2018-11-20 Palantir Technologies Inc. Techniques for data extraction
US10621314B2 (en) 2016-08-01 2020-04-14 Palantir Technologies Inc. Secure deployment of a software package
US11256762B1 (en) 2016-08-04 2022-02-22 Palantir Technologies Inc. System and method for efficiently determining and displaying optimal packages of data items
US10552531B2 (en) 2016-08-11 2020-02-04 Palantir Technologies Inc. Collaborative spreadsheet data validation and integration
US10373078B1 (en) 2016-08-15 2019-08-06 Palantir Technologies Inc. Vector generation for distributed data sets
EP3282374A1 (en) 2016-08-17 2018-02-14 Palantir Technologies Inc. User interface data sample transformer
US10650086B1 (en) 2016-09-27 2020-05-12 Palantir Technologies Inc. Systems, methods, and framework for associating supporting data in word processing
US20180095996A1 (en) 2016-10-03 2018-04-05 Ocient Llc Database system utilizing forced memory aligned access
US10152306B2 (en) 2016-11-07 2018-12-11 Palantir Technologies Inc. Framework for developing and deploying applications
US11061876B2 (en) * 2016-11-15 2021-07-13 Sap Se Fast aggregation on compressed data
US10261763B2 (en) 2016-12-13 2019-04-16 Palantir Technologies Inc. Extensible data transformation authoring and validation system
WO2018112056A1 (en) 2016-12-14 2018-06-21 Ocient Llc Efficient database management system utilizing silo and manifest
EP3555756A4 (en) 2016-12-14 2020-09-23 Ocient, Inc. SYSTEM AND METHOD OF USING A DETAILED LEADER IN A DATABASE MANAGEMENT SYSTEM
US11157951B1 (en) 2016-12-16 2021-10-26 Palantir Technologies Inc. System and method for determining and displaying an optimal assignment of data items
US10528599B1 (en) 2016-12-16 2020-01-07 Amazon Technologies, Inc. Tiered data processing for distributed data
US11074261B1 (en) 2016-12-16 2021-07-27 Amazon Technologies, Inc. Format independent processing for distributed data
US10509844B1 (en) 2017-01-19 2019-12-17 Palantir Technologies Inc. Network graph parser
US10180934B2 (en) 2017-03-02 2019-01-15 Palantir Technologies Inc. Automatic translation of spreadsheets into scripts
US10572576B1 (en) 2017-04-06 2020-02-25 Palantir Technologies Inc. Systems and methods for facilitating data object extraction from unstructured documents
US10846318B1 (en) 2017-04-18 2020-11-24 Microstrategy Incorporated Natural language visualizations
US10824604B1 (en) 2017-05-17 2020-11-03 Palantir Technologies Inc. Systems and methods for data entry
US10747765B2 (en) 2017-05-30 2020-08-18 Ocient Inc. System and method for optimizing large database management systems with multiple optimizers
US10534595B1 (en) 2017-06-30 2020-01-14 Palantir Technologies Inc. Techniques for configuring and validating a data pipeline deployment
US10210240B2 (en) * 2017-06-30 2019-02-19 Capital One Services, Llc Systems and methods for code parsing and lineage detection
US10204119B1 (en) 2017-07-20 2019-02-12 Palantir Technologies, Inc. Inferring a dataset schema from input files
US10754820B2 (en) 2017-08-14 2020-08-25 Palantir Technologies Inc. Customizable pipeline for integrating data
US11016936B1 (en) 2017-09-05 2021-05-25 Palantir Technologies Inc. Validating data for integration
US11182125B2 (en) 2017-09-07 2021-11-23 Ocient Inc. Computing device sort function
US11379525B1 (en) 2017-11-22 2022-07-05 Palantir Technologies Inc. Continuous builds of derived datasets in response to other dataset updates
US10783162B1 (en) 2017-12-07 2020-09-22 Palantir Technologies Inc. Workflow assistant
US10552524B1 (en) 2017-12-07 2020-02-04 Palantir Technolgies Inc. Systems and methods for in-line document tagging and object based data synchronization
US10360252B1 (en) 2017-12-08 2019-07-23 Palantir Technologies Inc. Detection and enrichment of missing data or metadata for large data sets
US11176116B2 (en) 2017-12-13 2021-11-16 Palantir Technologies Inc. Systems and methods for annotating datasets
US10853352B1 (en) 2017-12-21 2020-12-01 Palantir Technologies Inc. Structured data collection, presentation, validation and workflow management
GB201800595D0 (en) 2018-01-15 2018-02-28 Palantir Technologies Inc Management of software bugs in a data processing system
US10599762B1 (en) 2018-01-16 2020-03-24 Palantir Technologies Inc. Systems and methods for creating a dynamic electronic form
US11068540B2 (en) 2018-01-25 2021-07-20 Ab Initio Technology Llc Techniques for integrating validation results in data profiling and related systems and methods
US11138230B2 (en) * 2018-03-26 2021-10-05 Mcafee, Llc Methods, apparatus, and systems to aggregate partitioned computer database data
US10685031B2 (en) * 2018-03-27 2020-06-16 New Relic, Inc. Dynamic hash partitioning for large-scale database management systems
US10885021B1 (en) 2018-05-02 2021-01-05 Palantir Technologies Inc. Interactive interpreter and graphical user interface
US11263263B2 (en) 2018-05-30 2022-03-01 Palantir Technologies Inc. Data propagation and mapping system
US11061542B1 (en) 2018-06-01 2021-07-13 Palantir Technologies Inc. Systems and methods for determining and displaying optimal associations of data items
US10795909B1 (en) 2018-06-14 2020-10-06 Palantir Technologies Inc. Minimized and collapsed resource dependency path
US11709835B2 (en) 2018-10-15 2023-07-25 Ocient Holdings LLC Re-ordered processing of read requests
US11256696B2 (en) 2018-10-15 2022-02-22 Ocient Holdings LLC Data set compression within a database system
US11249916B2 (en) 2018-10-15 2022-02-15 Ocient Holdings LLC Single producer single consumer buffering in database systems
US11880368B2 (en) 2018-10-15 2024-01-23 Ocient Holdings LLC Compressing data sets for storage in a database system
US11886436B2 (en) 2018-10-15 2024-01-30 Ocient Inc. Segmenting a partition of a data set based on a data storage coding scheme
US10341172B1 (en) * 2018-10-22 2019-07-02 CSG Media, LLC System and method for configuring networks
CN109508335B (zh) * 2018-12-03 2022-10-28 中国电波传播研究所(中国电子科技集团公司第二十二研究所) 一种海量地杂波数据分类存储方法
US11195050B2 (en) 2019-02-05 2021-12-07 Microstrategy Incorporated Machine learning to generate and evaluate visualizations
US11157565B2 (en) * 2019-04-04 2021-10-26 Sap Se Parallel partition-wise insert sub-select
US11093500B2 (en) 2019-10-28 2021-08-17 Ocient Holdings LLC Enforcement of minimum query cost rules required for access to a database system
US11106679B2 (en) 2019-10-30 2021-08-31 Ocient Holdings LLC Enforcement of sets of query rules for access to data supplied by a plurality of data providers
US11614970B2 (en) 2019-12-06 2023-03-28 Microstrategy Incorporated High-throughput parallel data transmission
US11609911B2 (en) * 2019-12-19 2023-03-21 Ocient Holdings LLC Selecting a normalized form for conversion of a query expression
US11567965B2 (en) 2020-01-23 2023-01-31 Microstrategy Incorporated Enhanced preparation and integration of data sets
US11853364B2 (en) 2020-01-31 2023-12-26 Ocient Holdings LLC Level-based queries in a database system and methods for use therewith
US11061910B1 (en) 2020-01-31 2021-07-13 Ocient Holdings LLC Servicing concurrent queries via virtual segment recovery
US11599463B2 (en) 2020-03-25 2023-03-07 Ocient Holdings LLC Servicing queries during data ingress
US11238041B2 (en) 2020-03-25 2022-02-01 Ocient Holdings LLC Facilitating query executions via dynamic data block routing
US11580102B2 (en) 2020-04-02 2023-02-14 Ocient Holdings LLC Implementing linear algebra functions via decentralized execution of query operator flows
US11256694B2 (en) * 2020-04-27 2022-02-22 Hewlett Packard Enterprise Development Lp Tolerance level-based tuning of query processing
US11294916B2 (en) 2020-05-20 2022-04-05 Ocient Holdings LLC Facilitating query executions via multiple modes of resultant correctness
US11321314B2 (en) 2020-05-22 2022-05-03 International Business Machines Corporation Query content-based data generation
US11586608B1 (en) 2020-06-23 2023-02-21 Amazon Technologies, Inc. Handling requests to access separately stored items in a non-relational database
US11775529B2 (en) 2020-07-06 2023-10-03 Ocient Holdings LLC Recursive functionality in relational database systems
US11880716B2 (en) 2020-08-05 2024-01-23 Ocient Holdings LLC Parallelized segment generation via key-based subdivision in database systems
US11755589B2 (en) 2020-08-05 2023-09-12 Ocient Holdings LLC Delaying segment generation in database systems
US11321288B2 (en) 2020-08-05 2022-05-03 Ocient Holdings LLC Record deduplication in database systems
US11704314B2 (en) 2020-08-06 2023-07-18 International Business Machines Corporation Multiplexing data operation
US11822532B2 (en) 2020-10-14 2023-11-21 Ocient Holdings LLC Per-segment secondary indexing in database systems
US11507578B2 (en) 2020-10-19 2022-11-22 Ocient Holdings LLC Delaying exceptions in query execution
US11675757B2 (en) 2020-10-29 2023-06-13 Ocient Holdings LLC Maintaining row durability data in database systems
US11297123B1 (en) 2020-12-11 2022-04-05 Ocient Holdings LLC Fault-tolerant data stream processing
US11314743B1 (en) 2020-12-29 2022-04-26 Ocient Holdings LLC Storing records via multiple field-based storage mechanisms
US11645273B2 (en) 2021-05-28 2023-05-09 Ocient Holdings LLC Query execution utilizing probabilistic indexing
US20220405263A1 (en) * 2021-06-21 2022-12-22 International Business Machines Corporation Increasing Index Availability in Databases
US20230024553A1 (en) * 2021-07-20 2023-01-26 Oracle International Corporation Subsumption of views and subqueries
US11803544B2 (en) 2021-10-06 2023-10-31 Ocient Holdings LLC Missing data-based indexing in database systems
CN116204136B (zh) * 2023-05-04 2023-08-15 山东浪潮科学研究院有限公司 一种数据存储、查询方法、装置、设备及存储介质

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4860201A (en) * 1986-09-02 1989-08-22 The Trustees Of Columbia University In The City Of New York Binary tree parallel processor
US4870568A (en) * 1986-06-25 1989-09-26 Thinking Machines Corporation Method for searching a database system including parallel processors
US5121494A (en) * 1989-10-05 1992-06-09 Ibm Corporation Joining two database relations on a common field in a parallel relational database field

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4876643A (en) * 1987-06-24 1989-10-24 Kabushiki Kaisha Toshiba Parallel searching system having a master processor for controlling plural slave processors for independently processing respective search requests
GB8719572D0 (en) * 1987-08-19 1987-09-23 Krebs M S Sigscan text retrieval system
US5060143A (en) * 1988-08-10 1991-10-22 Bell Communications Research, Inc. System for string searching including parallel comparison of candidate data block-by-block
US5469354A (en) * 1989-06-14 1995-11-21 Hitachi, Ltd. Document data processing method and apparatus for document retrieval
US5146540A (en) * 1990-02-22 1992-09-08 International Business Machines Corp. Processor allocation method and apparatus for multiprocessor execution of a constraint satisfaction search
US5210870A (en) * 1990-03-27 1993-05-11 International Business Machines Database sort and merge apparatus with multiple memory arrays having alternating access
US5325525A (en) 1991-04-04 1994-06-28 Hewlett-Packard Company Method of automatically controlling the allocation of resources of a parallel processor computer system by calculating a minimum execution time of a task and scheduling subtasks against resources to execute the task in the minimum time
US5367675A (en) 1991-12-13 1994-11-22 International Business Machines Corporation Computer automated system and method for optimizing the processing of a query in a relational database system by merging subqueries with the query
US5379420A (en) * 1991-12-26 1995-01-03 Trw Inc. High-speed data searching apparatus and method capable of operation in retrospective and dissemination modes
US5423037A (en) * 1992-03-17 1995-06-06 Teleserve Transaction Technology As Continuously available database server having multiple groups of nodes, each group maintaining a database copy with fragments stored on multiple nodes
US5446885A (en) 1992-05-15 1995-08-29 International Business Machines Corporation Event driven management information system with rule-based applications structure stored in a relational database
US5548770A (en) * 1993-02-25 1996-08-20 Data Parallel Systems, Inc. Method and apparatus for improving retrieval of data from a database
US5765146A (en) * 1993-11-04 1998-06-09 International Business Machines Corporation Method of performing a parallel relational database query in a multiprocessor environment
US5437032A (en) * 1993-11-04 1995-07-25 International Business Machines Corporation Task scheduler for a miltiprocessor system
US5495606A (en) * 1993-11-04 1996-02-27 International Business Machines Corporation System for parallel processing of complex read-only database queries using master and slave central processor complexes
US5590319A (en) * 1993-12-15 1996-12-31 Information Builders, Inc. Query processor for parallel processing in homogenous and heterogenous databases
US5742806A (en) * 1994-01-31 1998-04-21 Sun Microsystems, Inc. Apparatus and method for decomposing database queries for database management system including multiprocessor digital data processing system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4870568A (en) * 1986-06-25 1989-09-26 Thinking Machines Corporation Method for searching a database system including parallel processors
US4860201A (en) * 1986-09-02 1989-08-22 The Trustees Of Columbia University In The City Of New York Binary tree parallel processor
US5121494A (en) * 1989-10-05 1992-06-09 Ibm Corporation Joining two database relations on a common field in a parallel relational database field

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP0753176A1 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013156773A1 (en) * 2012-04-17 2013-10-24 Dataline Software Ltd Methods of querying a relational database
US9672268B2 (en) 2012-04-17 2017-06-06 Dataline Software, Ltd. Methods of querying a relational database
EP3654199A1 (en) * 2012-04-17 2020-05-20 Dataline Software Ltd Methods of querying a relational database
US11797520B2 (en) * 2019-11-29 2023-10-24 Oracle International Corporation ROWID elimination rewrite
US11663179B2 (en) * 2020-12-21 2023-05-30 International Business Machines Corporation Data simulation for regression analysis

Also Published As

Publication number Publication date
US5742806A (en) 1998-04-21
WO1995021407A3 (en) 1995-09-14
US6816854B2 (en) 2004-11-09
EP0753176A4 (en) 1998-06-17
CA2180252A1 (en) 1995-08-10
EP0753176A1 (en) 1997-01-15
US6289334B1 (en) 2001-09-11
JPH09511347A (ja) 1997-11-11
US20020032676A1 (en) 2002-03-14

Similar Documents

Publication Publication Date Title
US5742806A (en) Apparatus and method for decomposing database queries for database management system including multiprocessor digital data processing system
US5778354A (en) Database management system with improved indexed accessing
US7171399B2 (en) Method for efficient query execution using dynamic queries in database environments
US4769772A (en) Automated query optimization method using both global and parallel local optimizations for materialization access planning for distributed databases
EP1107135B1 (en) Parallel optimized triggers in parallel processing database systems
DeWitt et al. Parallel database systems: The future of database processing or a passing fad?
US6625593B1 (en) Parallel query optimization strategies for replicated and partitioned tables
US6618719B1 (en) Database system with methodology for reusing cost-based optimization decisions
US5325525A (en) Method of automatically controlling the allocation of resources of a parallel processor computer system by calculating a minimum execution time of a task and scheduling subtasks against resources to execute the task in the minimum time
US8126870B2 (en) System and methodology for parallel query optimization using semantic-based partitioning
US5615361A (en) Exploitation of uniqueness properties using a 1-tuple condition for the optimization of SQL queries
US5379419A (en) Methods and apparatus for accesssing non-relational data files using relational queries
US6105033A (en) Method and apparatus for detecting and removing obsolete cache entries for enhancing cache system operation
US6826562B1 (en) Method of simplifying and optimizing scalar subqueries and derived tables that return exactly or at most one tuple
US20050097103A1 (en) Performing sequence analysis as a multipart plan storing intermediate results as a relation
EP1649344A2 (en) Sql code generation for heterogeneous environment
Mehta et al. Batch scheduling in parallel database systems
Kranas et al. Parallel query processing in a polystore
US7099864B2 (en) System and method for slow materialization sorting of partially ordered inputs in a database system
Manegold et al. A multi-query optimizer for Monet
Smith et al. Polar: An architecture for a parallel ODMG compliant object database
von Bültzingsloewen et al. Design and implementation of KARDAMOM—A set-oriented data flow database machine
Salza et al. Performance Modeling of parallel database systems
von Bultzingsloewen et al. Kardamom—a dataflow database machine for real-time applications
Luo et al. Data language requirements of database machines

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): CA JP

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

AK Designated states

Kind code of ref document: A3

Designated state(s): CA JP

AL Designated countries for regional patents

Kind code of ref document: A3

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2180252

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 1995911601

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1995911601

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1995911601

Country of ref document: EP