US8909615B2 - System and method of managing capacity of search index partitions - Google Patents

System and method of managing capacity of search index partitions Download PDF

Info

Publication number
US8909615B2
US8909615B2 US13/595,031 US201213595031A US8909615B2 US 8909615 B2 US8909615 B2 US 8909615B2 US 201213595031 A US201213595031 A US 201213595031A US 8909615 B2 US8909615 B2 US 8909615B2
Authority
US
United States
Prior art keywords
partition
threshold
capacity utilization
search
mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/595,031
Other versions
US20140181071A1 (en
Inventor
Patrick Thomas Sidney Pidduck
Aleksandra Czarlinska
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Open Text SA
Original Assignee
Open Text SA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Open Text SA filed Critical Open Text SA
Priority to US13/595,031 priority Critical patent/US8909615B2/en
Assigned to Open Text S.A. reassignment Open Text S.A. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CZARLINSKA, ALEKSANDRA, PIDDUCK, PATRICK THOMAS SIDNEY
Publication of US20140181071A1 publication Critical patent/US20140181071A1/en
Priority to US14/539,542 priority patent/US9836541B2/en
Application granted granted Critical
Publication of US8909615B2 publication Critical patent/US8909615B2/en
Assigned to OPEN TEXT SA ULC reassignment OPEN TEXT SA ULC CERTIFICATE OF AMALGAMATION Assignors: IP OT SUB ULC
Assigned to IP OT SUB ULC reassignment IP OT SUB ULC CERTIFICATE OF CONTINUANCE Assignors: OT IP SUB, LLC
Assigned to OT IP SUB, LLC reassignment OT IP SUB, LLC IP BUSINESS SALE AGREEMENT Assignors: Open Text S.A.
Assigned to BARCLAYS BANK PLC reassignment BARCLAYS BANK PLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OPEN TEXT SA ULC
Assigned to BARCLAYS BANK PLC reassignment BARCLAYS BANK PLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OPEN TEXT SA ULC
Assigned to BARCLAYS BANK PLC reassignment BARCLAYS BANK PLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OPEN TEXT SA ULC
Assigned to THE BANK OF NEW YORK MELLON reassignment THE BANK OF NEW YORK MELLON SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OPEN TEXT SA ULC
Assigned to OPEN TEXT SA ULC reassignment OPEN TEXT SA ULC RELEASE OF SECURITY INTEREST IN PATENTS (REEL/FRAME 063559/0628) Assignors: BARCLAYS BANK PLC
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/951Indexing; Web crawling techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/316Indexing structures
    • G06F16/328Management therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/16File or folder operations, e.g. details of user interfaces specifically adapted to file systems
    • G06F16/162Delete operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/23Updating
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/27Replication, distribution or synchronisation of data between databases or within a distributed database system; Distributed database system architectures therefor
    • G06F16/278Data partitioning, e.g. horizontal or vertical partitioning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9538Presentation of query results
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/16Threshold monitoring

Definitions

  • This disclosure relates generally to the field of computerized search systems, and more particularly to systems, methods and products for managing capacity of search index partitions in a computerized search system.
  • Document management systems often employ a search engine to allow for fast retrieval of information.
  • a search engine can search the metadata and text of documents in a search index to determine which documents match search criteria without having to parse the document itself.
  • Each partition has a capacity limit, based on resources such as available memory, disk space or other capacity constraints.
  • partitions are filled with new data from indexing operations, they approach a point at which they are eventually deemed to be full.
  • the administrators need to monitor the sizes of the partitions, and make configuration changes as the partitions increase or decrease in size. This creates a system management burden, and can even result in a partition becoming inoperable if configuration changes are not made in a timely manner.
  • Another solution is for external automated applications to regularly check the status of the partitions, and notify the administrators that action should be taken based upon configuration rules.
  • the disadvantages here are the need for external programs to be created to monitor the partitions, and this still leaves room for errors if the administrator cannot react to the notifications in a timely manner.
  • search system may also be possible for an external system to monitor the status of partitions and make configuration changes automatically. This places a burden on external technology to anticipate the internal behavior of search partitions.
  • This disclosure relates generally to the field of computerized search systems. More particularly, embodiments disclosed herein provide for a system, method and computer program product that can manage search index partitions in a computerized search system in an automated, efficient, and intelligent manner.
  • a computerized search system may comprise a search engine.
  • This search engine may maintain a search index for objects (e.g., documents or other objects in a repository).
  • This search index may be partitioned, each having a set capacity.
  • the search engine may allow objects to be added to the partition and also allow existing objects in the partition to be updated or deleted.
  • An update operation may, for example, be to replace or modify the object.
  • the search engine may change from an Add mode to an Update mode.
  • the search engine may refuse indexing requests to add objects to the partition, but still allow existing objects in the partition to be updated or deleted.
  • the search engine may change from the Update mode to a Rebalancing mode.
  • the search engine may refuse indexing requests to add objects to the partition, allow existing objects in the partition to be deleted, and move objects, for which indexing requests have been received to update them, to other partition(s) that are not yet full.
  • a partition manager may transform an object that is subject to an update into an indexing request. The indexing request thus generated is then sent to another partition in the search index for processing.
  • the partition manager may automatically and proactively move existing objects out of the partition that is in the Rebalancing mode without waiting for incoming indexing requests to modify or delete existing objects.
  • the partition manager may be configured to perform such a moving operation until the capacity utilization of the partition drops below a third threshold and back into the Update mode.
  • One embodiment provides a computer program product that comprises a non-transitory computer readable medium having a set of computer instructions stored on it.
  • the instructions are executable by a computer processor to perform a method essentially as described above.
  • Another embodiment provides a computerized search system comprising at least one non-transitory computer readable medium storing instructions translatable by at least one processor to perform a method essentially as described above.
  • Embodiments disclosed herein can provide many advantages. For example, by having a range of capacity utilization reserved for updates to a partition, an authorized user can estimate the space that will be needed for updates and therefore eliminate or reduce the amount of rebalancing that will be required when the partition is full. Further, having an Update mode with a Rebalancing mode also provide advantages. With these modes, the rebalancing process can be optimized and the partition manager can make better decisions about allocating indexing requests among partitions.
  • FIG. 1 depicts one embodiment of a computing environment in which one embodiment of a search system can be implemented.
  • FIG. 2 depicts a functional block diagram illustrating one embodiment of an object analyzer.
  • FIG. 3 depicts a functional block diagram illustrating one embodiment of a search engine.
  • FIG. 4 depicts a diagrammatic representation of one embodiment of regions or fields for a portion of an index.
  • FIG. 5 depicts a functional block diagram illustrating one embodiment of an indexing engine.
  • FIG. 6 depicts a diagrammatic representation of one embodiment of an index with multiple partitions connected to and being managed by a partition manager.
  • FIG. 7 depicts a diagrammatic representation of one embodiment of a partition to which multiple thresholds are assigned.
  • FIG. 8 depicts a flow diagram illustrating one embodiment of a method for managing an index partition.
  • Document management systems often employ a search engine to allow for fast retrieval of information.
  • a search engine can search the metadata and text of documents in a search index to determine which documents match search criteria without having to parse the document itself.
  • Embodiments of a search engine described herein provide systems and methods for managing partitions to prevent partitions from becoming unmanageable. It may be helpful to first discuss an example search system.
  • FIG. 1 depicts a block diagram illustrating one embodiment of a computing environment 100 with object search system 101 .
  • Computing environment 100 includes an object repository 105 storing objects 107 of interest (documents, images, emails or other objects that may be searched).
  • Object repository 105 may comprise a file server or database system or other storage mechanism remotely or locally accessible by search system 101 .
  • search system 101 comprises a server having a central processing unit 112 connected to a memory 114 and storage unit 118 via a bus.
  • Central processing unit 112 may represent a single processor, multiple processors, a processor(s) with multiple processing cores and the like.
  • Storage unit 118 may include a non-transitory storage medium such as hard-disk drives, flash memory devices, optical media and the like.
  • Search system 101 may be connected to a data communications network (not shown).
  • Storage unit 118 stores computer executable instructions 119 and index 124 .
  • Computer executable instructions 119 can represent multiple programs and operating system code.
  • instructions 119 are executable to provide an object analyzer 120 and search engine 122 .
  • Object analyzer 120 and search engine 122 may be portions of the same program or may be separate programs. According to one embodiment, for example, object analyzer 120 is a component of a document management system while search engine 122 is a separate program that interfaces with the document management system.
  • object analyzer 120 and search engine 122 can be implemented on different computing systems and can, themselves, be distributed.
  • Index 124 includes metadata used to identify objects in response to a search query and may also include text used to identify objects.
  • Index 124 can include a single index containing metadata and text, separate metadata and text indices or other arrangements of information. While shown as a single index, index 124 may include multiple indices. Further, as will be described further below, index 124 may be partitioned, with different objects being represented in each partition.
  • Client computer system 130 may include components similar to those of the server of search system 101 , such as CPU 138 , memory 136 , and storage 140 . Additionally, client computer system 130 may include executable instructions 132 to provide a user interface 134 that allows a user to enter a search query. The user interface may be provided through a web browser, file system interface or other program.
  • the search system 101 of FIG. 1 is merely an example of a computing system and embodiments of a search system that may be implemented using other computing systems (e.g., desktop computers, laptops, mobile computing devices or other computing devices with adequate processing and memory) including multiple computers acting together to provide a search system (e.g., a cluster of servers or other computing devices connected by a network acting together to provide the search system).
  • client computer 130 may include any suitable desktop computer, laptop, mobile device, server or other computing system.
  • object analyzer 120 analyzes objects in object repository 105 to determine information to be indexed in index 124 .
  • Object analyzer 120 can send indexing instructions to search engine 122 to direct search engine 122 to add/modify/or delete metadata or text in index 124 , as discussed below.
  • search engine 122 can search the information in index 124 to identify objects responsive to the search query and return a list or other representation of those objects to client computer 130 .
  • FIG. 2 depicts a diagrammatic representation of one embodiment of an object analyzer 120 for analyzing an object 107 .
  • Object analyzer 120 can comprise various modules to process an object 107 .
  • Reading source data module 154 can open the object 107 .
  • Format identification module 156 examines the object to determine what type of file or data the object 107 comprises.
  • Archive expansion module 158 unzips files or otherwise decompresses files if the object 107 is a compressed file.
  • Decryption module 160 decrypts all or part of the data in the object 107 .
  • Text extraction module 162 applies rules to text in the object 107 to extract text for index 124 .
  • Language detection module 164 examines the text to determine the language in which the text is written.
  • Classification module 166 applies rules based upon text and metadata to classify content.
  • Encoding module 168 can convert text to a supported encoding.
  • Randomness detection module 170 can analyze data to be indexed to reject random information.
  • Object analyzer 120 may include modules that can derive metadata for object 107 .
  • a document management system may provide a limited amount of metadata with the object 107 .
  • Object analyzer 120 can derive other metadata from the metadata provided, text or other properties of the object 107 .
  • a filter or piece or code that extracts the text from a PowerPoint presentation might also create metadata about the presentation.
  • the metadata which is not provided by the document management system and which is derived by object analyzer 120 may include the number of slides in the presentation, the title of the file, the name of the presentation author, or the size of paper the presentation was designed to print on.
  • Metadata might include statistical analysis to generate a list of keyword or key concepts in the document; determining the subject person or company of the text; sentiment analysis—is the tone of the text positive or negative; or language identification—in what language is the text written.
  • Further examples of metadata may either be provided by the document management system (or other application) or derived by the analyzer may be the date the object was created, the size of the object in bytes, the name of the object, a description of the object or the like.
  • Object analyzer 120 may include any number of other modules to analyze an object and extract text 174 and metadata 172 to be indexed.
  • Object analyzer 120 may be a portion of a larger program, such as a document management program, may be a separate program or may be implemented according any suitable programming architecture.
  • the process of determining metadata 172 and text 174 to be indexed may be carried out by any number of different programs on a computer system or distributed across computer systems.
  • Metadata 172 and text 174 thus processed by object analyzer 120 may be provided to a search engine.
  • An example search engine will now be described with reference to FIG. 3 .
  • FIG. 3 depicts a diagrammatic representation of logical blocks for one embodiment of a search engine 122 .
  • Search engine 122 may provide an indexing interface 200 that receives indexing requests (e.g., from object analyzer 120 ) or other source.
  • a distributor module 210 may distribute the indexing requests to indexing engine 220 that act on an indexing request to update index 124 .
  • Search engine 122 may also include a search interface 230 to receive queries (e.g., from a document management system or other source).
  • Search interface 230 may send queries to search modules 240 .
  • Federator 245 gathers the results from all search modules together, and generates a response to the query received through search interface 230 .
  • Search modules 240 are responsible for performing searches on an index partition, and performing tasks such as computing relevance score, sorting results, and retrieving metadata regions to return in a query.
  • Search interface 230 may be configured to receive a search query from a user, and search index 124 for objects that meet the criteria set forth in the search query.
  • Query language may also be configured to permit sorting results of a search.
  • Various rules may be used to determine the sort order.
  • a user constructed the search query. It should be noted, however, that the user could be any system that issues queries to the search system, and may include other computer programs searching on behalf of other users, creating reports or running automatic processes. Additionally, as described above, there can be many different types of metadata in the search index. Thus, the search queries are not restricted to “text” based search terms.
  • search term represents a technical concept or interpretation.
  • a search term in the context of this disclosure can be a word, a string, or any combination of the following:
  • search term can also refer to any term that is used in a query and that has been modified or generated by any commonly used techniques.
  • Search engine 122 may include any number of other modules or configurations to update and search an index.
  • search modules 240 and indexing engines 220 may be a single module.
  • Search engine 122 may be a portion of a larger program, such as a document management program, may be a separate program or may be implemented according to any suitable programming architecture.
  • the processes of search engine 122 may be distributed across multiple computer systems.
  • index 124 is illustrated as a single index, index 124 may comprise a set of smaller indexes. For example, a separate index can be used by each indexing engine.
  • FIG. 4 depicts a diagrammatic representation of one embodiment of regions or fields for a portion of index 124 .
  • Index 124 includes a list of some or all objects 107 in repository 105 ( FIG. 1 ), each identified by a unique identifier 301 (also referred to as object ID).
  • Index 124 further includes a set of metadata regions 300 (also referred to as metadata fields).
  • a metadata field 300 may include more than one entry for an object.
  • the metadata fields can each have associated values in value storage locations within storage unit 118 . In other embodiments, the values may be discarded.
  • the index may include a list of dictionary terms contained in the metadata values of the object and pointers to where the metadata values corresponding to the field are stored.
  • Index 124 may also include other regions for an object, such as a text region 302 . Text region 302 may, example, include a list of terms in the text of an object.
  • Index 124 may include some or all of the content of an object.
  • index 124 may be partitioned.
  • index partitioning in one embodiment, the index of objects in repository 105 may be split into multiple indexes such that some objects are listed in one index partition, while other objects are listed in the other index partitions.
  • a ‘partition’ comprises a portion or fragment of index 124 and is associated with indexing engine 220 and search module 240 . Note it is possible to copy a partition and associate a different index engine and search engine with this partition. Index partitioning may also reduce resource usage and search time. Furthermore, separate indexes may be maintained for metadata and text and/or different metadata regions or fields.
  • Index 124 can be stored according to any suitable storage scheme. Example storage schemes may include “Memory Storage,” “Disk Storage” and “Retrieval Storage”:
  • Memory Storage in this storage scheme, all the elements of the index are kept in memory. This provides the fastest operation when search results must be retrieved, since the memory storage mode minimizes disk activity. Conversely, memory storage consumes the most memory in partitions. Text regions which are frequently searched and retrieved for display should typically be held in memory.
  • Disk Storage in this storage scheme, the dictionary and index are kept in memory, but the value storage is located on disk within a Checkpoint file. Keyword searches are still fast, but search queries which need to examine the original data, such as phrase searches, are generally slower. Retrieving values from disk for display is also slower. For regions which are not commonly searched and displayed, disk storage is a good choice. Disk storage is also suitable as a storage mode for systems utilizing solid state hardware.
  • Retrieval Storage in this storage scheme, storage is optimized for text metadata regions which need to be retrieved and displayed, but do not need to be searchable.
  • text values may be stored on disk within the Checkpoint file, and there is no dictionary or index at all.
  • This storage scheme can be used, for example, for regions such as Hot Phrases and Summaries.
  • FIG. 5 depicts a diagrammatic representation of one embodiment of an indexing engine 220 to maintain a partition of index 124 .
  • indexing engine 220 can include an indexing controller 305 , a metadata update component 310 , and a text update component 315 .
  • index 124 is maintained as a separate metadata index 312 , which contains metadata for objects in repository 105
  • text index 314 which contains content text from objects in repository 105 , with a known relationship between the text and metadata components for each object in the index.
  • Indexing controller 305 receives indexing requests (e.g., from a distributor, another application or other source).
  • An indexing request received at the indexing controller 305 may include an instruction to add an object, delete an object, modify an object or replace an object in index 124 .
  • Such an indexing request may also include the information to be added or changed, such as the full text content to be indexed and the associated metadata for the object.
  • An indexing request may also contain derived metadata.
  • the text (derived text or full text content) of an indexing request may be a text file. It could be data exported from a database or other information system. Commonly, the text is the human-readable information within a document composed on a computer. In this scenario, a file such as a Microsoft Word document would be analyzed by a filtering step to extract the text, which can be stripped of unnecessary information such as fonts, styles, or page layout information.
  • the metadata portion of an indexing request may specifically be provided by an application providing the indexing request. This might be data such as an identifier for the object, the date or time it was first added to the system, or the identity of the user who manages the object.
  • a portion of the metadata can be derived metadata.
  • Derived metadata can include metadata inferred from the text content.
  • the filter or code that extracts the text from a PowerPoint presentation might also create metadata about the presentation.
  • the generated metadata may include the number of slides in the presentation, the title of the file, the name of the presentation author stored in the PowerPoint file, or the size of paper the presentation was designed to print on.
  • More complex examples of derived metadata might include statistical analysis to generate a list of keyword or key concepts in the document, determining the subject person or company of the text, sentiment analysis (the positive or negative tone of the text), or identification of the language in which the text is written.
  • Derived metadata may also include data inferred from processing an object. For example, in processing a PowerPoint presentation, derived metadata may include a timestamp of the time the PowerPoint was processed or the location where the PowerPoint presentation was processed.
  • An indexing engine can receive an indexing request 510 from an application, distributor or other source.
  • Indexing request 510 specifies an operation to be taken on index 124 for an object and any metadata or text for that action.
  • an application that generates an indexing request may be a corporate document management system, a web site with a search capability such as an online store, or a desktop search program for email.
  • an indexing request can take the form of an indexing object that includes a unique identification for an object, an operation, the metadata or text regions affected and the metadata and/or text for the index.
  • indexing operations may include adding, replacing, modifying and deleting information in the index, or combinations thereof. The following provides some exemplary operations that may be included in indexing requests.
  • AddOrReplace this operation can be used to create new objects in the index. According to one embodiment, if the object does not exist, it will be created, but if an entry with the same object identification exists, then it will be completely replaced with the new data, equivalent to a delete and add.
  • AddOrReplace this function may distinguish between content and metadata. If an object already exists, and metadata only is provided, the existing full text content is retained.
  • AddOrModify this operation will update an existing object, or create a new object if it does not already exist.
  • modifying an existing object only the provided content and metadata is updated. Any metadata regions that already exist which are not specified in the AddOrModify command will be left intact.
  • this operation will remove an object from the index, including both the metadata and the content.
  • Indexing controller 305 is a component which interprets the indexing request to determine how it should be processed. Indexing controller 305 can identify whether a text indexing command exists, and, if so, send the command with the necessary parameters to the text update component 315 . Indexing controller 305 can likewise determine if any metadata indexing operations are required, and if so, send the command with necessary parameters to the metadata update component 310 .
  • Text update component 315 is responsible for processing requests to index full text content. This may include tasks such as maintaining a dictionary of search terms, maintaining the internal search data structures, and updating the storage representation of the text portion of the search index in memory or on disk as appropriate. Text update component 315 may support instructions such as Add an Object, Replace an Object, or Delete an Object.
  • Metadata update component 310 is responsible for processing requests to index metadata 312 associated with an object in index 124 . This may include building and maintaining dictionaries of search terms, maintaining internal search data structures, and updating the representation of the metadata portion of the search index in memory or on disk as appropriate. Metadata update component 310 may support instructions such as Add an Object, Replace an Object, or Delete an Object.
  • Indexing engine 220 may include any number of other modules to update and search an index.
  • Indexing engine 220 may be a portion of a larger program, such as a document management program, may be a separate program or may be implemented according any suitable programming architecture.
  • the processes of indexing engine 220 may be distributed across multiple computer systems.
  • FIG. 6 depicts a diagrammatic representation of one embodiment of system 600 for managing partitions.
  • index 124 is divided into “n” partitions 400 , with each partition including a metadata index and a text index.
  • each partition can have its own indexing engine 220 and search module 240 .
  • Partition manager 505 is a component of a search system that accepts indexing requests, and determines which partition should service an indexing request, and provides the indexing request to the appropriate indexing engine 220 .
  • partition manager 505 can be a logical function of a search engine in the search system which, in turn, can be part of a document management system.
  • partition manager 505 can be a logical function of distributor 210 shown in FIG. 3 .
  • An indexing engine (e.g., indexing engine 220 shown in FIG. 3 ) for a partition performs the actual indexing operations of adding, deleting or modifying data in the partition.
  • partition manager 505 may be able to federate search queries to multiple search engines 240 associated with multiple partitions and combine the results. In one embodiment, this function of partition manager 505 may be incorporated in a federator (e.g., federator 245 shown in FIG. 3 ).
  • a search system can be configured so that each partition can have a set capacity (number of objects, memory, disk space or other measure of capacity). Additionally, thresholds can be set for each partition. The amount of information in the partition relative to the thresholds can be used to determine which operations are performed on the partition. An example of this methodology is described below with reference to FIGS. 7-8 .
  • FIG. 7 depicts one embodiment of partition 400 to which three thresholds are assigned: first threshold 405 , second threshold 410 and third threshold 415 .
  • the behavior of the search system can change depending upon the amount of information in a partition relative to various thresholds. These thresholds may be the same or different from partition to partition.
  • Each partition 400 in index 124 can have its own set of thresholds.
  • the capacity of the partition is described below using percentages.
  • the capacity utilization of a partition may be measured using one or a combination of values, such as memory, disk space, the number of objects contained within a partition, or other suitable measures.
  • the partition In the empty state, the partition is capable of accepting new objects. That is, the indexing engine is able to add new objects to the partition. The partition therefore is operating in an “Add” mode, which is represented in FIG. 7 as range 420 .
  • the indexing engine is also capable of servicing indexing requests to delete or modify objects already within the partition.
  • the capacity utilization of partition 400 may increase or decrease as operations are performed to add objects to the partition, delete objects from the partition, or modify objects in the partition.
  • the partition can continue to accept new objects while in the “Add” mode until the capacity utilization reaches first threshold 405 . In the example of FIG. 8 , this is referred to as “Add” mode ( 810 ).
  • indexing engine 220 will change its mode of operation from “Add” to “Update” ( 812 ). In the Update mode, the indexing engine for the partition will refuse to accept indexing requests to add new objects. Specifically, indexing engine 220 will communicate with partition manager 505 to convey this status and partition manager 505 will direct any indexing requests to add objects to other partitions.
  • the indexing engine can continue to accept requests to delete existing objects, which will reduce the capacity utilization.
  • the indexing engine will also accept requests to change or modify existing objects, which may increase or decrease the capacity utilization depending upon the nature of the change. If the capacity utilization of the partition drops below the Update Threshold, the partition will change its behavior back to the Add mode ( 814 ), and can begin accepting new objects again ( 810 ).
  • the amount of information in the partition decreases from capacity utilization range 425 to capacity utilization range 420 (below first threshold 405 )
  • indexing engine 220 will change its mode of operation from “Update” to “Add” and distributor 210 (where partition manager 505 resides in this example) will again direct indexing requests to add objects to partition 400 .
  • the partition will remain in the Rebalancing mode until the capacity utilization drops below the Exit Rebalancing Threshold.
  • indexing engine 220 will remain in the Rebalancing mode until the capacity utilization drops below third threshold 415 .
  • the fact that the capacity utilization may drop below second threshold 410 but not third threshold 415 will not cause indexing engine 220 to change from the Rebalancing mode to the Update mode.
  • indexing engine 220 will change from the Rebalancing mode to the Update mode.
  • third threshold 415 is not needed and second threshold 410 can server as a single trigger point by which indexing engine 220 enters or exits the Rebalancing mode. In one embodiment, having third threshold 415 being lower than second threshold 410 may provide better efficiency, as described below.
  • partition manager 505 may extract the object to be modified and transform it to an indexing request for another partition. In this case, partition manager 505 can select which target partition should be used and forward the indexing request that partition manager 505 had prepared to an appropriate indexing engine 220 . Partition manager 505 may then cause the object to be deleted from the partition that is in the Rebalancing mode. In this way, the capacity utilization can be reduced any time an object is deleted from the partition, or an object is modified since modification causes the object to be moved from the partition that is in the Rebalancing mode to another partition that is not in the Rebalancing mode.
  • partition manager 505 can wait for delete and modify operations to reduce the capacity utilization while a partition is in the Rebalancing mode.
  • An alternative implementation would allow a more aggressive Rebalancing operation in which objects are moved proactively to other partitions to reduce the capacity utilization below the Rebalancing Threshold, instead of waiting for indexing requests to delete and/or modify specific objects.
  • the system will switch the mode of operation from the Rebalancing mode to the Update mode for the partition ( 818 ).
  • Having an Update mode with a Rebalancing mode also provide advantages. Without these modes, objects are moved each time an upper limit is reached. This can result in thrashing of the index. For example, an object is added and that pushes the size past the upper limit. The next modification moves an object to another partition, allowing another add operation to occur, with the cycle repeating.
  • the rebalancing process can be optimized and the partition manager can make better decisions about allocating indexing requests among partitions.
  • thresholds 405 , 410 and 415 are adjustable. This ability to configure the thresholds allows system users to optimize the behavior of the capacity management system for their situation. For example, if the search system is part of a long-term archival product, then changes to objects are infrequent, and the estimated capacity reserved for updates can be small. If the search system is part of a customer management system where customer information is changed frequently, then reserving more capacity for updates to objects is appropriate. Moreover, the number of thresholds and level of capacity of utilization of each threshold can vary between partitions.
  • partitions may be assigned additional thresholds. For example, while the same threshold is used to enter and exit the update mode of operation in FIG. 7 , the partition can be assigned a different enter update threshold and exit update threshold.
  • Partition manager 505 can create new partitions as needed.
  • partition manager 505 can create a new partition when all the other partitions are in an update or rebalancing mode.
  • thresholds 405 , 410 and 415 can be percentage thresholds. For example, a threshold of 85% capacity utilization may be used for the Update threshold. In this case, a majority of available memory is dedicated to building and updating the index. When the capacity utilization meets or exceeds the 85% threshold, the index is deemed to be full. The system enters into the Update mode and new objects are added to other partitions that are not full. Capacity utilization range 425 , which is defined by the Update Threshold and the Enter Rebalancing Threshold, essentially represents space reserved for updates and changes to existing objects after the index partition is deemed “full”. Some considerations for adjusting this threshold setting may include:
  • a threshold of 100% capacity utilization may be used to mark an entry point for the Rebalancing mode. Note that there can be memory reserved beyond the 100% capacity utilization threshold for program use and other temporary data needs. As an example, there can be 300 MB reserved working space in the memory that is not part of the capacity utilization for the partition.
  • the 100% capacity utilization threshold can serve as an Enter Rebalancing Threshold and a 97% capacity utilization threshold can serve as an Exit Rebalancing Threshold.
  • the index partition enters the Rebalancing mode when the Enter Rebalancing Threshold is met or exceeded. As described above, further updates in the Rebalancing mode will cause objects to be moved to other partitions. When the size of the partition in the Rebalancing mode drops below the Exit Rebalancing Threshold, the partition will again accept updates to existing object as the partition re-enters into the Update mode.
  • embodiments of a search system can maintain a search index of metadata and text for objects in a repository, repositories or distributed across a network.
  • the search index can be divided into partitions with a partition assigned a first capacity utilization threshold and a second capacity utilization threshold. If the capacity utilization of the partition is below the first threshold, the search system can add, update and delete information in the partition. If the capacity utilization of the partition is above the first threshold, the search system can update and delete information in the partition, but cannot add information for new objects to the partition. If the capacity utilization of the partition is above the second threshold, the search system can enter a rebalancing mode in which it seeks to rebalance capacity utilization between partitions. Other alternative embodiments are also possible.
  • the invention disclosed here describes a system where the partition maintenance is not visible or otherwise transparent to the end user. Specifically, a user query issued during any stage of partition maintenance is answered using a consistent and complete view of all objects in the repository regardless of the current partition in which a given object resides.
  • the intelligent partition management described above can therefore occur ‘online’ or ‘live’.
  • ‘online’ or ‘live’ means that there is no interruption in the search service from the end user's perspective.
  • Embodiments discussed herein can be implemented in a computer communicatively coupled to a network (for example, the Internet), another computer, or in a standalone computer.
  • a suitable computer can include a central processing unit (“CPU”), at least one read-only memory (“ROM”), at least one random access memory (“RAM”), at least one hard drive (“HD”), and one or more input/output (“I/O”) device(s).
  • the I/O devices can include a keyboard, monitor, printer, electronic pointing device (for example, mouse, trackball, stylus, touch pad, etc.), or the like.
  • ROM, RAM, and HD are computer memories for storing computer-executable instructions executable by the CPU or capable of being compiled or interpreted to be executable by the CPU. Suitable computer-executable instructions may reside on a computer readable medium (e.g., ROM, RAM, and/or HD), hardware circuitry or the like, or any combination thereof.
  • a computer readable medium is not limited to ROM, RAM, and HD and can include any type of data storage medium that can be read by a processor.
  • a computer-readable medium may refer to a data cartridge, a data backup magnetic tape, a floppy diskette, a flash memory drive, an optical data storage drive, a CD-ROM, ROM, RAM, HD, or the like.
  • the processes described herein may be implemented in suitable computer-executable instructions that may reside on a computer readable medium (for example, a disk, CD-ROM, a memory, etc.).
  • a computer readable medium for example, a disk, CD-ROM, a memory, etc.
  • the computer-executable instructions may be stored as software code components on a direct access storage device array, magnetic tape, floppy diskette, optical storage device, or other appropriate computer-readable medium or storage device.
  • Any suitable programming language can be used to implement the routines, methods or programs of embodiments of the invention described herein, including C, C++, Java, JavaScript, HTML, or any other programming or scripting code, etc.
  • Other software/hardware/network architectures may be used.
  • the functions of the disclosed embodiments may be implemented on one computer or shared/distributed among two or more computers in or across a network. Communications between computers implementing embodiments can be accomplished using any electronic, optical, radio frequency signals, or other suitable methods and tools of communication in compliance with known network protocols.
  • Any particular routine can execute on a single computer processing device or multiple computer processing devices, a single computer processor or multiple computer processors. Data may be stored in a single storage medium or distributed through multiple storage mediums, and may reside in a single database or multiple databases (or other data storage techniques).
  • steps, operations, or computations may be presented in a specific order, this order may be changed in different embodiments. In some embodiments, to the extent multiple steps are shown as sequential in this specification, some combination of such steps in alternative embodiments may be performed at the same time.
  • the sequence of operations described herein can be interrupted, suspended, or otherwise controlled by another process, such as an operating system, kernel, etc.
  • the routines can operate in an operating system environment or as stand-alone routines. Functions, routines, methods, steps and operations described herein can be performed in hardware, software, firmware or any combination thereof.
  • Embodiments described herein can be implemented in the form of control logic in software or hardware or a combination of both.
  • the control logic may be stored in an information storage medium, such as a computer-readable medium, as a plurality of instructions adapted to direct an information processing device to perform a set of steps disclosed in the various embodiments.
  • an information storage medium such as a computer-readable medium
  • a person of ordinary skill in the art will appreciate other ways and/or methods to implement the invention.
  • a “computer-readable medium” may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, system or device.
  • the computer readable medium can be, by way of example only but not by limitation, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, system, device, propagation medium, or computer memory.
  • Such computer-readable medium shall generally be machine readable and include software programming or code that can be human readable (e.g., source code) or machine readable (e.g., object code).
  • non-transitory computer-readable media can include random access memories, read-only memories, hard drives, data cartridges, magnetic tapes, floppy diskettes, flash memory drives, optical data storage devices, compact-disc read-only memories, and other appropriate computer memories and data storage devices.
  • some or all of the software components may reside on a single server computer or on any combination of separate server computers.
  • a computer program product implementing an embodiment disclosed herein may comprise one or more non-transitory computer readable media storing computer instructions translatable by one or more processors in a computing environment.
  • a “processor” includes any, hardware system, mechanism or component that processes data, signals or other information.
  • a processor can include a system with a general-purpose central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor can perform its functions in “real-time,” “offline,” in a “batch mode,” etc. Portions of processing can be performed at different times and at different locations, by different (or the same) processing systems.
  • the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having,” or any other variation thereof, are intended to cover a non-exclusive inclusion.
  • a process, product, article, or apparatus that comprises a list of elements is not necessarily limited only those elements but may include other elements not expressly listed or inherent to such process, product, article, or apparatus.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Computing Systems (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Software Systems (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

A search system can maintain a search index of metadata and text for objects in a repository, repositories or distributed across a network. The search index can be divided into partitions with a partition assigned a first capacity utilization threshold and a second capacity utilization threshold. If the capacity utilization of the partition is below the first threshold, the system can add, update and delete information in the partition. If the capacity utilization of the partition is above the first threshold, the system can update and delete information in the partition, but cannot add information for new objects to the partition. If the capacity utilization of the partition is above the second threshold, the system can enter a rebalancing mode in which it seeks to rebalance capacity utilization between partitions. The behavior of the system can change depending upon the size of a partition relative to its configurable thresholds.

Description

CROSS-REFERENCE TO RELATED APPLICATION(S)
This is a conversion of and claims a benefit of priority from U.S. Provisional Application No. 61/528,967, filed Aug. 30, 2011, entitled “SYSTEM AND METHOD OF MANAGING CAPACITY OF SEARCH INDEX PARTITIONS,” which is incorporated by reference as if set forth herein in its entirety.
TECHNICAL FIELD
This disclosure relates generally to the field of computerized search systems, and more particularly to systems, methods and products for managing capacity of search index partitions in a computerized search system.
BACKGROUND OF THE RELATED ART
Document management systems often employ a search engine to allow for fast retrieval of information. A search engine can search the metadata and text of documents in a search index to determine which documents match search criteria without having to parse the document itself.
As the volume of information committed to a search system increases, a need arises to have multiple search systems sharing responsibility for managing the search index. The index needs to be split into smaller components, called partitions. Each partition has a capacity limit, based on resources such as available memory, disk space or other capacity constraints.
As partitions are filled with new data from indexing operations, they approach a point at which they are eventually deemed to be full. In a traditional system, the administrators need to monitor the sizes of the partitions, and make configuration changes as the partitions increase or decrease in size. This creates a system management burden, and can even result in a partition becoming inoperable if configuration changes are not made in a timely manner.
One solution is for the administrator of the system to check the conditions of the partitions on a regular basis. This introduces the prospect of user error, and is problematic if the administrators are not available.
Another solution is for external automated applications to regularly check the status of the partitions, and notify the administrators that action should be taken based upon configuration rules. The disadvantages here are the need for external programs to be created to monitor the partitions, and this still leaves room for errors if the administrator cannot react to the notifications in a timely manner.
If the search system provides suitable integration points, it may also be possible for an external system to monitor the status of partitions and make configuration changes automatically. This places a burden on external technology to anticipate the internal behavior of search partitions.
None of these solutions, however, provide for an understanding when a partition is too full and moving appropriate data from a full partition to one with available space.
There are implementations that exist today which are capable of moving data to other partitions. However, these solutions move data inefficiently, have only one mode of operation, and only move data once extreme limits have been exceeded. Consequently, there is always room for innovations and improvements.
SUMMARY OF THE DISCLOSURE
This disclosure relates generally to the field of computerized search systems. More particularly, embodiments disclosed herein provide for a system, method and computer program product that can manage search index partitions in a computerized search system in an automated, efficient, and intelligent manner.
A computerized search system may comprise a search engine. This search engine may maintain a search index for objects (e.g., documents or other objects in a repository). This search index may be partitioned, each having a set capacity. In one embodiment, when a capacity utilization of a partition in the search index is under a first threshold, the search engine may allow objects to be added to the partition and also allow existing objects in the partition to be updated or deleted. An update operation may, for example, be to replace or modify the object.
When the capacity utilization of the partition meets or exceeds the first threshold, the search engine may change from an Add mode to an Update mode. In the Update mode, the search engine may refuse indexing requests to add objects to the partition, but still allow existing objects in the partition to be updated or deleted.
When the capacity utilization of the partition meets or exceeds a second threshold, the search engine may change from the Update mode to a Rebalancing mode. In the Rebalancing mode, the search engine may refuse indexing requests to add objects to the partition, allow existing objects in the partition to be deleted, and move objects, for which indexing requests have been received to update them, to other partition(s) that are not yet full. To do so, a partition manager may transform an object that is subject to an update into an indexing request. The indexing request thus generated is then sent to another partition in the search index for processing. Optionally, the partition manager may automatically and proactively move existing objects out of the partition that is in the Rebalancing mode without waiting for incoming indexing requests to modify or delete existing objects. The partition manager may be configured to perform such a moving operation until the capacity utilization of the partition drops below a third threshold and back into the Update mode. These thresholds are configurable and may differ from partition to partition.
One embodiment provides a computer program product that comprises a non-transitory computer readable medium having a set of computer instructions stored on it. The instructions are executable by a computer processor to perform a method essentially as described above.
Another embodiment provides a computerized search system comprising at least one non-transitory computer readable medium storing instructions translatable by at least one processor to perform a method essentially as described above.
Embodiments disclosed herein can provide many advantages. For example, by having a range of capacity utilization reserved for updates to a partition, an authorized user can estimate the space that will be needed for updates and therefore eliminate or reduce the amount of rebalancing that will be required when the partition is full. Further, having an Update mode with a Rebalancing mode also provide advantages. With these modes, the rebalancing process can be optimized and the partition manager can make better decisions about allocating indexing requests among partitions.
These, and other, aspects of the invention will be better appreciated and understood when considered in conjunction with the following description and the accompanying drawings. The following description, while indicating various embodiments of the invention and numerous specific details thereof, is given by way of illustration and not of limitation. Many substitutions, modifications, additions or rearrangements may be made within the scope of the invention, and the invention includes all such substitutions, modifications, additions or rearrangements.
BRIEF DESCRIPTION OF THE DRAWINGS
The drawings accompanying and forming part of this specification are included to depict certain aspects of the invention. A clearer impression of the invention, and of the components and operation of systems provided with the invention, will become more readily apparent by referring to the exemplary, and therefore non-limiting, embodiments illustrated in the drawings, wherein identical reference numerals designate the same components. Note that the features illustrated in the drawings are not necessarily drawn to scale.
FIG. 1 depicts one embodiment of a computing environment in which one embodiment of a search system can be implemented.
FIG. 2 depicts a functional block diagram illustrating one embodiment of an object analyzer.
FIG. 3 depicts a functional block diagram illustrating one embodiment of a search engine.
FIG. 4 depicts a diagrammatic representation of one embodiment of regions or fields for a portion of an index.
FIG. 5 depicts a functional block diagram illustrating one embodiment of an indexing engine.
FIG. 6 depicts a diagrammatic representation of one embodiment of an index with multiple partitions connected to and being managed by a partition manager.
FIG. 7 depicts a diagrammatic representation of one embodiment of a partition to which multiple thresholds are assigned.
FIG. 8 depicts a flow diagram illustrating one embodiment of a method for managing an index partition.
DETAILED DESCRIPTION
The disclosure and various features and advantageous details thereof are explained more fully with reference to the exemplary, and therefore non-limiting, embodiments illustrated in the accompanying drawings and detailed in the following description. Descriptions of known starting materials and processes may be omitted so as not to unnecessarily obscure the disclosure in detail. It should be understood, however, that the detailed description and the specific examples, while indicating the preferred embodiments, are given by way of illustration only and not by way of limitation. Various substitutions, modifications, additions and/or rearrangements within the spirit and/or scope of the underlying inventive concept will become apparent to those skilled in the art from this disclosure.
Document management systems often employ a search engine to allow for fast retrieval of information. A search engine can search the metadata and text of documents in a search index to determine which documents match search criteria without having to parse the document itself. Embodiments of a search engine described herein provide systems and methods for managing partitions to prevent partitions from becoming unmanageable. It may be helpful to first discuss an example search system.
FIG. 1 depicts a block diagram illustrating one embodiment of a computing environment 100 with object search system 101. Computing environment 100 includes an object repository 105 storing objects 107 of interest (documents, images, emails or other objects that may be searched). Object repository 105 may comprise a file server or database system or other storage mechanism remotely or locally accessible by search system 101.
In the embodiment of FIG. 1, search system 101 comprises a server having a central processing unit 112 connected to a memory 114 and storage unit 118 via a bus. Central processing unit 112 may represent a single processor, multiple processors, a processor(s) with multiple processing cores and the like. Storage unit 118 may include a non-transitory storage medium such as hard-disk drives, flash memory devices, optical media and the like. Search system 101 may be connected to a data communications network (not shown).
Storage unit 118 stores computer executable instructions 119 and index 124. Computer executable instructions 119 can represent multiple programs and operating system code. In one embodiment, instructions 119 are executable to provide an object analyzer 120 and search engine 122. Object analyzer 120 and search engine 122 may be portions of the same program or may be separate programs. According to one embodiment, for example, object analyzer 120 is a component of a document management system while search engine 122 is a separate program that interfaces with the document management system. Furthermore, object analyzer 120 and search engine 122 can be implemented on different computing systems and can, themselves, be distributed.
Index 124 includes metadata used to identify objects in response to a search query and may also include text used to identify objects. Index 124 can include a single index containing metadata and text, separate metadata and text indices or other arrangements of information. While shown as a single index, index 124 may include multiple indices. Further, as will be described further below, index 124 may be partitioned, with different objects being represented in each partition.
Client computer system 130 may include components similar to those of the server of search system 101, such as CPU 138, memory 136, and storage 140. Additionally, client computer system 130 may include executable instructions 132 to provide a user interface 134 that allows a user to enter a search query. The user interface may be provided through a web browser, file system interface or other program.
The search system 101 of FIG. 1 is merely an example of a computing system and embodiments of a search system that may be implemented using other computing systems (e.g., desktop computers, laptops, mobile computing devices or other computing devices with adequate processing and memory) including multiple computers acting together to provide a search system (e.g., a cluster of servers or other computing devices connected by a network acting together to provide the search system). Similarly, client computer 130 may include any suitable desktop computer, laptop, mobile device, server or other computing system.
In operation, object analyzer 120 analyzes objects in object repository 105 to determine information to be indexed in index 124. Object analyzer 120 can send indexing instructions to search engine 122 to direct search engine 122 to add/modify/or delete metadata or text in index 124, as discussed below. When a search query is received, search engine 122 can search the information in index 124 to identify objects responsive to the search query and return a list or other representation of those objects to client computer 130.
FIG. 2 depicts a diagrammatic representation of one embodiment of an object analyzer 120 for analyzing an object 107. Object analyzer 120 can comprise various modules to process an object 107. Reading source data module 154 can open the object 107. Format identification module 156 examines the object to determine what type of file or data the object 107 comprises. Archive expansion module 158 unzips files or otherwise decompresses files if the object 107 is a compressed file. Decryption module 160 decrypts all or part of the data in the object 107. Text extraction module 162 applies rules to text in the object 107 to extract text for index 124. Language detection module 164 examines the text to determine the language in which the text is written. Classification module 166 applies rules based upon text and metadata to classify content. Encoding module 168 can convert text to a supported encoding. Randomness detection module 170 can analyze data to be indexed to reject random information.
Object analyzer 120 may include modules that can derive metadata for object 107. For example, a document management system may provide a limited amount of metadata with the object 107. Object analyzer 120 can derive other metadata from the metadata provided, text or other properties of the object 107. As a specific example, a filter or piece or code that extracts the text from a PowerPoint presentation might also create metadata about the presentation. In this example, the metadata which is not provided by the document management system and which is derived by object analyzer 120 may include the number of slides in the presentation, the title of the file, the name of the presentation author, or the size of paper the presentation was designed to print on. More complex examples of derived metadata might include statistical analysis to generate a list of keyword or key concepts in the document; determining the subject person or company of the text; sentiment analysis—is the tone of the text positive or negative; or language identification—in what language is the text written. Further examples of metadata that may either be provided by the document management system (or other application) or derived by the analyzer may be the date the object was created, the size of the object in bytes, the name of the object, a description of the object or the like.
The embodiment of FIG. 2 is provided by way of example. Object analyzer 120 may include any number of other modules to analyze an object and extract text 174 and metadata 172 to be indexed. Object analyzer 120 may be a portion of a larger program, such as a document management program, may be a separate program or may be implemented according any suitable programming architecture. In one embodiment, the process of determining metadata 172 and text 174 to be indexed may be carried out by any number of different programs on a computer system or distributed across computer systems.
Metadata 172 and text 174 thus processed by object analyzer 120 may be provided to a search engine. An example search engine will now be described with reference to FIG. 3.
FIG. 3 depicts a diagrammatic representation of logical blocks for one embodiment of a search engine 122. Search engine 122 may provide an indexing interface 200 that receives indexing requests (e.g., from object analyzer 120) or other source. A distributor module 210 may distribute the indexing requests to indexing engine 220 that act on an indexing request to update index 124. Search engine 122 may also include a search interface 230 to receive queries (e.g., from a document management system or other source). Search interface 230 may send queries to search modules 240. Federator 245 gathers the results from all search modules together, and generates a response to the query received through search interface 230. Search modules 240 are responsible for performing searches on an index partition, and performing tasks such as computing relevance score, sorting results, and retrieving metadata regions to return in a query.
Search interface 230 may be configured to receive a search query from a user, and search index 124 for objects that meet the criteria set forth in the search query. Query language may also be configured to permit sorting results of a search. Various rules may be used to determine the sort order. In this example, a user constructed the search query. It should be noted, however, that the user could be any system that issues queries to the search system, and may include other computer programs searching on behalf of other users, creating reports or running automatic processes. Additionally, as described above, there can be many different types of metadata in the search index. Thus, the search queries are not restricted to “text” based search terms.
In the context of this disclosure, the phrase “search term” represents a technical concept or interpretation. For example, a search term in the context of this disclosure can be a word, a string, or any combination of the following:
    • phrases
    • numbers
    • strings
    • logical operations (e.g., AND, OR, NOT, etc.)
    • ordering or operations (e.g., using parenthesis)
    • relationships (e.g., greater than, less than, not equal to, etc.)
    • similarities based on thesaurus, stemming, sounds-like, etc.
    • wildcards and pattern matching
To this end, a search term can also refer to any term that is used in a query and that has been modified or generated by any commonly used techniques.
For context, a search term could be a word “John” or a more complex expression like: (>“bob” or !=(“123” or a*)) and (sounds-like “smith” or thesaurus “doctor” or “medical doctor” or stem “medical”).
The embodiment of FIG. 3 is provided by way of example. Search engine 122 may include any number of other modules or configurations to update and search an index. For example, search modules 240 and indexing engines 220 may be a single module. Search engine 122 may be a portion of a larger program, such as a document management program, may be a separate program or may be implemented according to any suitable programming architecture. In one embodiment, the processes of search engine 122 may be distributed across multiple computer systems. Furthermore, while in FIG. 3, index 124 is illustrated as a single index, index 124 may comprise a set of smaller indexes. For example, a separate index can be used by each indexing engine.
FIG. 4 depicts a diagrammatic representation of one embodiment of regions or fields for a portion of index 124. Index 124 includes a list of some or all objects 107 in repository 105 (FIG. 1), each identified by a unique identifier 301 (also referred to as object ID). Index 124 further includes a set of metadata regions 300 (also referred to as metadata fields). A metadata field 300 may include more than one entry for an object. The metadata fields can each have associated values in value storage locations within storage unit 118. In other embodiments, the values may be discarded. The index may include a list of dictionary terms contained in the metadata values of the object and pointers to where the metadata values corresponding to the field are stored. Index 124 may also include other regions for an object, such as a text region 302. Text region 302 may, example, include a list of terms in the text of an object. Index 124 may include some or all of the content of an object.
While shown as a single index, index 124 may be partitioned. In index partitioning, in one embodiment, the index of objects in repository 105 may be split into multiple indexes such that some objects are listed in one index partition, while other objects are listed in the other index partitions. As described below with reference to FIGS. 5 and 6, a ‘partition’ comprises a portion or fragment of index 124 and is associated with indexing engine 220 and search module 240. Note it is possible to copy a partition and associate a different index engine and search engine with this partition. Index partitioning may also reduce resource usage and search time. Furthermore, separate indexes may be maintained for metadata and text and/or different metadata regions or fields. Index 124 can be stored according to any suitable storage scheme. Example storage schemes may include “Memory Storage,” “Disk Storage” and “Retrieval Storage”:
Memory Storage: in this storage scheme, all the elements of the index are kept in memory. This provides the fastest operation when search results must be retrieved, since the memory storage mode minimizes disk activity. Conversely, memory storage consumes the most memory in partitions. Text regions which are frequently searched and retrieved for display should typically be held in memory.
Disk Storage: in this storage scheme, the dictionary and index are kept in memory, but the value storage is located on disk within a Checkpoint file. Keyword searches are still fast, but search queries which need to examine the original data, such as phrase searches, are generally slower. Retrieving values from disk for display is also slower. For regions which are not commonly searched and displayed, disk storage is a good choice. Disk storage is also suitable as a storage mode for systems utilizing solid state hardware.
Retrieval Storage: in this storage scheme, storage is optimized for text metadata regions which need to be retrieved and displayed, but do not need to be searchable. As an example, text values may be stored on disk within the Checkpoint file, and there is no dictionary or index at all. This storage scheme can be used, for example, for regions such as Hot Phrases and Summaries.
FIG. 5 depicts a diagrammatic representation of one embodiment of an indexing engine 220 to maintain a partition of index 124. In this embodiment, indexing engine 220 can include an indexing controller 305, a metadata update component 310, and a text update component 315. In this embodiment, index 124 is maintained as a separate metadata index 312, which contains metadata for objects in repository 105, and text index 314, which contains content text from objects in repository 105, with a known relationship between the text and metadata components for each object in the index.
Indexing controller 305 receives indexing requests (e.g., from a distributor, another application or other source). An indexing request received at the indexing controller 305 may include an instruction to add an object, delete an object, modify an object or replace an object in index 124. Such an indexing request may also include the information to be added or changed, such as the full text content to be indexed and the associated metadata for the object. An indexing request may also contain derived metadata.
The text (derived text or full text content) of an indexing request may be a text file. It could be data exported from a database or other information system. Commonly, the text is the human-readable information within a document composed on a computer. In this scenario, a file such as a Microsoft Word document would be analyzed by a filtering step to extract the text, which can be stripped of unnecessary information such as fonts, styles, or page layout information.
The metadata portion of an indexing request may specifically be provided by an application providing the indexing request. This might be data such as an identifier for the object, the date or time it was first added to the system, or the identity of the user who manages the object.
A portion of the metadata can be derived metadata. Derived metadata can include metadata inferred from the text content. For example, the filter or code that extracts the text from a PowerPoint presentation might also create metadata about the presentation. In this example, the generated metadata may include the number of slides in the presentation, the title of the file, the name of the presentation author stored in the PowerPoint file, or the size of paper the presentation was designed to print on. More complex examples of derived metadata might include statistical analysis to generate a list of keyword or key concepts in the document, determining the subject person or company of the text, sentiment analysis (the positive or negative tone of the text), or identification of the language in which the text is written. Derived metadata may also include data inferred from processing an object. For example, in processing a PowerPoint presentation, derived metadata may include a timestamp of the time the PowerPoint was processed or the location where the PowerPoint presentation was processed.
An indexing engine can receive an indexing request 510 from an application, distributor or other source. Indexing request 510 specifies an operation to be taken on index 124 for an object and any metadata or text for that action. For context, an application that generates an indexing request may be a corporate document management system, a web site with a search capability such as an online store, or a desktop search program for email.
According to one embodiment, for example, an indexing request can take the form of an indexing object that includes a unique identification for an object, an operation, the metadata or text regions affected and the metadata and/or text for the index. By way of example, but not limitation, indexing operations may include adding, replacing, modifying and deleting information in the index, or combinations thereof. The following provides some exemplary operations that may be included in indexing requests.
AddOrReplace: this operation can be used to create new objects in the index. According to one embodiment, if the object does not exist, it will be created, but if an entry with the same object identification exists, then it will be completely replaced with the new data, equivalent to a delete and add.
AddOrReplace: this function may distinguish between content and metadata. If an object already exists, and metadata only is provided, the existing full text content is retained.
AddOrModify: this operation will update an existing object, or create a new object if it does not already exist. When modifying an existing object, only the provided content and metadata is updated. Any metadata regions that already exist which are not specified in the AddOrModify command will be left intact.
Delete: this operation will remove an object from the index, including both the metadata and the content.
Indexing controller 305, according to one embodiment, is a component which interprets the indexing request to determine how it should be processed. Indexing controller 305 can identify whether a text indexing command exists, and, if so, send the command with the necessary parameters to the text update component 315. Indexing controller 305 can likewise determine if any metadata indexing operations are required, and if so, send the command with necessary parameters to the metadata update component 310.
Text update component 315 is responsible for processing requests to index full text content. This may include tasks such as maintaining a dictionary of search terms, maintaining the internal search data structures, and updating the storage representation of the text portion of the search index in memory or on disk as appropriate. Text update component 315 may support instructions such as Add an Object, Replace an Object, or Delete an Object.
Metadata update component 310 is responsible for processing requests to index metadata 312 associated with an object in index 124. This may include building and maintaining dictionaries of search terms, maintaining internal search data structures, and updating the representation of the metadata portion of the search index in memory or on disk as appropriate. Metadata update component 310 may support instructions such as Add an Object, Replace an Object, or Delete an Object.
The embodiment of FIG. 5 is provided by way of example. Indexing engine 220 may include any number of other modules to update and search an index. Indexing engine 220 may be a portion of a larger program, such as a document management program, may be a separate program or may be implemented according any suitable programming architecture. In one embodiment, the processes of indexing engine 220 may be distributed across multiple computer systems.
As discussed above, an index may be partitioned. FIG. 6 depicts a diagrammatic representation of one embodiment of system 600 for managing partitions. In the embodiment of FIG. 6, index 124 is divided into “n” partitions 400, with each partition including a metadata index and a text index. As illustrated in FIG. 5, in one embodiment, each partition can have its own indexing engine 220 and search module 240.
A partition manager can be configured to manage these partitions. Partition manager 505 is a component of a search system that accepts indexing requests, and determines which partition should service an indexing request, and provides the indexing request to the appropriate indexing engine 220. In one embodiment, partition manager 505 can be a logical function of a search engine in the search system which, in turn, can be part of a document management system. In one embodiment, partition manager 505 can be a logical function of distributor 210 shown in FIG. 3. An indexing engine (e.g., indexing engine 220 shown in FIG. 3) for a partition performs the actual indexing operations of adding, deleting or modifying data in the partition. Likewise, partition manager 505 may be able to federate search queries to multiple search engines 240 associated with multiple partitions and combine the results. In one embodiment, this function of partition manager 505 may be incorporated in a federator (e.g., federator 245 shown in FIG. 3).
A search system can be configured so that each partition can have a set capacity (number of objects, memory, disk space or other measure of capacity). Additionally, thresholds can be set for each partition. The amount of information in the partition relative to the thresholds can be used to determine which operations are performed on the partition. An example of this methodology is described below with reference to FIGS. 7-8. FIG. 7 depicts one embodiment of partition 400 to which three thresholds are assigned: first threshold 405, second threshold 410 and third threshold 415. As flow 800 of FIG. 8 illustrates, the behavior of the search system can change depending upon the amount of information in a partition relative to various thresholds. These thresholds may be the same or different from partition to partition. Each partition 400 in index 124 can have its own set of thresholds.
Assume there are several partitions, and the initial state is that each is empty. For the purpose of discussion and not of limitation, the capacity of the partition is described below using percentages. In practice, the capacity utilization of a partition may be measured using one or a combination of values, such as memory, disk space, the number of objects contained within a partition, or other suitable measures.
In the empty state, the partition is capable of accepting new objects. That is, the indexing engine is able to add new objects to the partition. The partition therefore is operating in an “Add” mode, which is represented in FIG. 7 as range 420. The indexing engine is also capable of servicing indexing requests to delete or modify objects already within the partition. Thus, the capacity utilization of partition 400 may increase or decrease as operations are performed to add objects to the partition, delete objects from the partition, or modify objects in the partition. The partition can continue to accept new objects while in the “Add” mode until the capacity utilization reaches first threshold 405. In the example of FIG. 8, this is referred to as “Add” mode (810).
Once the Update Threshold is exceeded and the partition is in capacity utilization range 425, indexing engine 220 will change its mode of operation from “Add” to “Update” (812). In the Update mode, the indexing engine for the partition will refuse to accept indexing requests to add new objects. Specifically, indexing engine 220 will communicate with partition manager 505 to convey this status and partition manager 505 will direct any indexing requests to add objects to other partitions.
While in the Update mode the indexing engine can continue to accept requests to delete existing objects, which will reduce the capacity utilization. The indexing engine will also accept requests to change or modify existing objects, which may increase or decrease the capacity utilization depending upon the nature of the change. If the capacity utilization of the partition drops below the Update Threshold, the partition will change its behavior back to the Add mode (814), and can begin accepting new objects again (810). Referring to FIG. 7, when the amount of information in the partition decreases from capacity utilization range 425 to capacity utilization range 420 (below first threshold 405), may cause indexing engine 220 as well as distributor 210 to change behavior. Specifically, indexing engine 220 will change its mode of operation from “Update” to “Add” and distributor 210 (where partition manager 505 resides in this example) will again direct indexing requests to add objects to partition 400.
While the capacity utilization of the partition is in the Update mode, as indexing requests to modify objects are processed, the capacity utilization may increase. Eventually, the capacity utilization may reach an Enter Rebalancing Threshold, which is referred to as second threshold 410 in FIG. 7. At this point, the partition is now over the safe capacity for adding new objects or modifying existing objects. The partition enters a “Rebalancing” mode of operation (816) and into capacity utilization range 430. As shown in FIG. 7, capacity utilization range 430 is defined by third threshold 415, which is normally lower than second threshold 410. Third threshold 415 may be referred to as an Exit Rebalancing Threshold. Once in the Rebalancing mode, the partition will remain in the Rebalancing mode until the capacity utilization drops below the Exit Rebalancing Threshold. Thus, once partition 400 enters into capacity utilization range 430, indexing engine 220 will remain in the Rebalancing mode until the capacity utilization drops below third threshold 415. In this embodiment, the fact that the capacity utilization may drop below second threshold 410 but not third threshold 415 will not cause indexing engine 220 to change from the Rebalancing mode to the Update mode. In another embodiment, when the capacity utilization drops below second threshold 410, indexing engine 220 will change from the Rebalancing mode to the Update mode. In this embodiment, third threshold 415 is not needed and second threshold 410 can server as a single trigger point by which indexing engine 220 enters or exits the Rebalancing mode. In one embodiment, having third threshold 415 being lower than second threshold 410 may provide better efficiency, as described below.
While in the Rebalancing mode, the partition will not accept new objects to index. In one embodiment, if an indexing request arrives to modify an existing object in the partition, partition manager 505 may extract the object to be modified and transform it to an indexing request for another partition. In this case, partition manager 505 can select which target partition should be used and forward the indexing request that partition manager 505 had prepared to an appropriate indexing engine 220. Partition manager 505 may then cause the object to be deleted from the partition that is in the Rebalancing mode. In this way, the capacity utilization can be reduced any time an object is deleted from the partition, or an object is modified since modification causes the object to be moved from the partition that is in the Rebalancing mode to another partition that is not in the Rebalancing mode.
In one embodiment, partition manager 505 can wait for delete and modify operations to reduce the capacity utilization while a partition is in the Rebalancing mode. An alternative implementation would allow a more aggressive Rebalancing operation in which objects are moved proactively to other partitions to reduce the capacity utilization below the Rebalancing Threshold, instead of waiting for indexing requests to delete and/or modify specific objects. When the capacity utilization drops below the Exit Rebalancing Threshold, the system will switch the mode of operation from the Rebalancing mode to the Update mode for the partition (818).
In any application that incorporates a search system such as a document management system, having the ability to reserve a capacity utilization range for updates is a material advantage. Without an Update mode, the partitions will add information until they are full. Once full, the system must perform rebalancing every time an object must be modified in a way that increases the capacity utilization. This can impose significant performance degradation on the indexing system, since moving objects to other partitions may be computationally expensive. By having a range of capacity utilization reserved for updates, an authorized user such as a system operator or administrator can estimate the space that will be needed for updates and therefore eliminate or reduce the amount of rebalancing that will be required.
Having an Update mode with a Rebalancing mode also provide advantages. Without these modes, objects are moved each time an upper limit is reached. This can result in thrashing of the index. For example, an object is added and that pushes the size past the upper limit. The next modification moves an object to another partition, allowing another add operation to occur, with the cycle repeating. By introducing hysteresis with the Rebalancing mode, the rebalancing process can be optimized and the partition manager can make better decisions about allocating indexing requests among partitions.
According to one embodiment, thresholds 405, 410 and 415 are adjustable. This ability to configure the thresholds allows system users to optimize the behavior of the capacity management system for their situation. For example, if the search system is part of a long-term archival product, then changes to objects are infrequent, and the estimated capacity reserved for updates can be small. If the search system is part of a customer management system where customer information is changed frequently, then reserving more capacity for updates to objects is appropriate. Moreover, the number of thresholds and level of capacity of utilization of each threshold can vary between partitions.
Furthermore, partitions may be assigned additional thresholds. For example, while the same threshold is used to enter and exit the update mode of operation in FIG. 7, the partition can be assigned a different enter update threshold and exit update threshold.
Partition manager 505 can create new partitions as needed. By way of example, but not of limitation, partition manager 505 can create a new partition when all the other partitions are in an update or rebalancing mode.
In one embodiment, thresholds 405, 410 and 415 can be percentage thresholds. For example, a threshold of 85% capacity utilization may be used for the Update threshold. In this case, a majority of available memory is dedicated to building and updating the index. When the capacity utilization meets or exceeds the 85% threshold, the index is deemed to be full. The system enters into the Update mode and new objects are added to other partitions that are not full. Capacity utilization range 425, which is defined by the Update Threshold and the Enter Rebalancing Threshold, essentially represents space reserved for updates and changes to existing objects after the index partition is deemed “full”. Some considerations for adjusting this threshold setting may include:
    • If the system has applications or custom modules known to add significant new metadata to existing objects, more space should be allowed for updates.
    • Archival systems which rarely modify metadata can reduce the space reserved for updates.
A threshold of 100% capacity utilization may be used to mark an entry point for the Rebalancing mode. Note that there can be memory reserved beyond the 100% capacity utilization threshold for program use and other temporary data needs. As an example, there can be 300 MB reserved working space in the memory that is not part of the capacity utilization for the partition. In this case, the 100% capacity utilization threshold can serve as an Enter Rebalancing Threshold and a 97% capacity utilization threshold can serve as an Exit Rebalancing Threshold. The index partition enters the Rebalancing mode when the Enter Rebalancing Threshold is met or exceeded. As described above, further updates in the Rebalancing mode will cause objects to be moved to other partitions. When the size of the partition in the Rebalancing mode drops below the Exit Rebalancing Threshold, the partition will again accept updates to existing object as the partition re-enters into the Update mode.
Users or an automated system can periodically review the percent full status of partitions, and adjust the partition percent full thresholds based upon actual usage patterns. Again, the capacity utilization of a partition can be measured in various ways and is not limited by the use of percentage.
Thus, embodiments of a search system can maintain a search index of metadata and text for objects in a repository, repositories or distributed across a network. The search index can be divided into partitions with a partition assigned a first capacity utilization threshold and a second capacity utilization threshold. If the capacity utilization of the partition is below the first threshold, the search system can add, update and delete information in the partition. If the capacity utilization of the partition is above the first threshold, the search system can update and delete information in the partition, but cannot add information for new objects to the partition. If the capacity utilization of the partition is above the second threshold, the search system can enter a rebalancing mode in which it seeks to rebalance capacity utilization between partitions. Other alternative embodiments are also possible.
The invention disclosed here describes a system where the partition maintenance is not visible or otherwise transparent to the end user. Specifically, a user query issued during any stage of partition maintenance is answered using a consistent and complete view of all objects in the repository regardless of the current partition in which a given object resides. The intelligent partition management described above can therefore occur ‘online’ or ‘live’. In this context, ‘online’ or ‘live’ means that there is no interruption in the search service from the end user's perspective. This is an advantage over ‘offline’ systems where the search service(s) must be temporarily disabled (taken offline) to allow maintenance of the partitions. In such an ‘offline’ system, a user may not be able to issue queries during maintenance as such queries would produce incomplete search results.
Although the invention has been described with respect to specific embodiments thereof, these embodiments are merely illustrative, and not restrictive of the invention. The description herein of illustrated embodiments of the invention, including the description in the Abstract and Summary, is not intended to be exhaustive or to limit the invention to the precise forms disclosed herein (and in particular, the inclusion of any particular embodiment, feature or function within the Abstract or Summary is not intended to limit the scope of the invention to such embodiment, feature or function). Rather, the description is intended to describe illustrative embodiments, features and functions in order to provide a person of ordinary skill in the art context to understand the invention without limiting the invention to any particularly described embodiment, feature or function, including any such embodiment feature or function described in the Abstract or Summary. While specific embodiments of, and examples for, the invention are described herein for illustrative purposes only, various equivalent modifications are possible within the spirit and scope of the invention, as those skilled in the relevant art will recognize and appreciate. As indicated, these modifications may be made to the invention in light of the foregoing description of illustrated embodiments of the invention and are to be included within the spirit and scope of the invention. Thus, while the invention has been described herein with reference to particular embodiments thereof, a latitude of modification, various changes and substitutions are intended in the foregoing disclosures, and it will be appreciated that in some instances some features of embodiments of the invention will be employed without a corresponding use of other features without departing from the scope and spirit of the invention as set forth. Therefore, many modifications may be made to adapt a particular situation or material to the essential scope and spirit of the invention.
Reference throughout this specification to “one embodiment”, “an embodiment”, or “a specific embodiment” or similar terminology means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment and may not necessarily be present in all embodiments. Thus, respective appearances of the phrases “in one embodiment”, “in an embodiment”, or “in a specific embodiment” or similar terminology in various places throughout this specification are not necessarily referring to the same embodiment. Furthermore, the particular features, structures, or characteristics of any particular embodiment may be combined in any suitable manner with one or more other embodiments. It is to be understood that other variations and modifications of the embodiments described and illustrated herein are possible in light of the teachings herein and are to be considered as part of the spirit and scope of the invention.
In the description herein, numerous specific details are provided, such as examples of components and/or methods, to provide a thorough understanding of embodiments of the invention. One skilled in the relevant art will recognize, however, that an embodiment may be able to be practiced without one or more of the specific details, or with other apparatus, systems, assemblies, methods, components, materials, parts, and/or the like. In other instances, well-known structures, components, systems, materials, or operations are not specifically shown or described in detail to avoid obscuring aspects of embodiments of the invention. While the invention may be illustrated by using a particular embodiment, this is not and does not limit the invention to any particular embodiment and a person of ordinary skill in the art will recognize that additional embodiments are readily understandable and are a part of this invention.
Embodiments discussed herein can be implemented in a computer communicatively coupled to a network (for example, the Internet), another computer, or in a standalone computer. As is known to those skilled in the art, a suitable computer can include a central processing unit (“CPU”), at least one read-only memory (“ROM”), at least one random access memory (“RAM”), at least one hard drive (“HD”), and one or more input/output (“I/O”) device(s). The I/O devices can include a keyboard, monitor, printer, electronic pointing device (for example, mouse, trackball, stylus, touch pad, etc.), or the like.
ROM, RAM, and HD are computer memories for storing computer-executable instructions executable by the CPU or capable of being compiled or interpreted to be executable by the CPU. Suitable computer-executable instructions may reside on a computer readable medium (e.g., ROM, RAM, and/or HD), hardware circuitry or the like, or any combination thereof. Within this disclosure, the term “computer readable medium” is not limited to ROM, RAM, and HD and can include any type of data storage medium that can be read by a processor. For example, a computer-readable medium may refer to a data cartridge, a data backup magnetic tape, a floppy diskette, a flash memory drive, an optical data storage drive, a CD-ROM, ROM, RAM, HD, or the like. The processes described herein may be implemented in suitable computer-executable instructions that may reside on a computer readable medium (for example, a disk, CD-ROM, a memory, etc.). Alternatively, the computer-executable instructions may be stored as software code components on a direct access storage device array, magnetic tape, floppy diskette, optical storage device, or other appropriate computer-readable medium or storage device.
Any suitable programming language can be used to implement the routines, methods or programs of embodiments of the invention described herein, including C, C++, Java, JavaScript, HTML, or any other programming or scripting code, etc. Other software/hardware/network architectures may be used. For example, the functions of the disclosed embodiments may be implemented on one computer or shared/distributed among two or more computers in or across a network. Communications between computers implementing embodiments can be accomplished using any electronic, optical, radio frequency signals, or other suitable methods and tools of communication in compliance with known network protocols.
Different programming techniques can be employed such as procedural or object oriented. Any particular routine can execute on a single computer processing device or multiple computer processing devices, a single computer processor or multiple computer processors. Data may be stored in a single storage medium or distributed through multiple storage mediums, and may reside in a single database or multiple databases (or other data storage techniques). Although the steps, operations, or computations may be presented in a specific order, this order may be changed in different embodiments. In some embodiments, to the extent multiple steps are shown as sequential in this specification, some combination of such steps in alternative embodiments may be performed at the same time. The sequence of operations described herein can be interrupted, suspended, or otherwise controlled by another process, such as an operating system, kernel, etc. The routines can operate in an operating system environment or as stand-alone routines. Functions, routines, methods, steps and operations described herein can be performed in hardware, software, firmware or any combination thereof.
Embodiments described herein can be implemented in the form of control logic in software or hardware or a combination of both. The control logic may be stored in an information storage medium, such as a computer-readable medium, as a plurality of instructions adapted to direct an information processing device to perform a set of steps disclosed in the various embodiments. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will appreciate other ways and/or methods to implement the invention.
It is also within the spirit and scope of the invention to implement in software programming or code an of the steps, operations, methods, routines or portions thereof described herein, where such software programming or code can be stored in a computer-readable medium and can be operated on by a processor to permit a computer to perform any of the steps, operations, methods, routines or portions thereof described herein. The invention may be implemented by using software programming or code in one or more general purpose digital computers, by using application specific integrated circuits, programmable logic devices, field programmable gate arrays, optical, chemical, biological, quantum or nanoengineered systems, components and mechanisms may be used. In general, the functions of the invention can be achieved by any means as is known in the art. For example, distributed, or networked systems, components and circuits can be used. In another example, communication or transfer (or otherwise moving from one place to another) of data may be wired, wireless, or by any other means.
A “computer-readable medium” may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, system or device. The computer readable medium can be, by way of example only but not by limitation, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, system, device, propagation medium, or computer memory. Such computer-readable medium shall generally be machine readable and include software programming or code that can be human readable (e.g., source code) or machine readable (e.g., object code). Examples of non-transitory computer-readable media can include random access memories, read-only memories, hard drives, data cartridges, magnetic tapes, floppy diskettes, flash memory drives, optical data storage devices, compact-disc read-only memories, and other appropriate computer memories and data storage devices. In an illustrative embodiment, some or all of the software components may reside on a single server computer or on any combination of separate server computers. As one skilled in the art can appreciate, a computer program product implementing an embodiment disclosed herein may comprise one or more non-transitory computer readable media storing computer instructions translatable by one or more processors in a computing environment.
A “processor” includes any, hardware system, mechanism or component that processes data, signals or other information. A processor can include a system with a general-purpose central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor can perform its functions in “real-time,” “offline,” in a “batch mode,” etc. Portions of processing can be performed at different times and at different locations, by different (or the same) processing systems.
It will also be appreciated that one or more of the elements depicted in the drawings/figures can also be implemented in a more separated or integrated manner, or even removed or rendered as inoperable in certain cases, as is useful in accordance with a particular application. Additionally, any signal arrows in the drawings/figures should be considered only as exemplary, and not limiting, unless otherwise specifically noted.
As used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having,” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, product, article, or apparatus that comprises a list of elements is not necessarily limited only those elements but may include other elements not expressly listed or inherent to such process, product, article, or apparatus.
Furthermore, the term “or” as used herein is generally intended to mean “and/or” unless otherwise indicated. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present). As used herein, including the claims that follow, a term preceded by “a” or “an” (and “the” when antecedent basis is “a” or “an”) includes both singular and plural of such term, unless clearly indicated within the claim otherwise (i.e., that the reference “a” or “an” clearly indicates only the singular or only the plural). Also, as used in the description herein and throughout the claims that follow, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise. The scope of the present disclosure should be determined by the following claims and their legal equivalents.

Claims (20)

What is claimed is:
1. A computerized search system, comprising:
a search engine communicatively connected to a client device over a physical connection;
a processor; and
a storage unit coupled to the processor and storing a search index and instructions, wherein the instructions when translated by the processor cause the search engine to perform:
when a capacity utilization of a partition in the search index is under a first threshold:
allowing any object to be added to the partition; and
allowing any existing object in the partition to be updated or deleted;
when the capacity utilization of the partition is above the first threshold:
disallowing any object to be added to the partition; and
allowing any existing object in the partition to be updated or deleted; and
when the capacity utilization of the partition is above a second threshold:
disallowing any object to be added to the partition;
allowing any existing object in the partition to be deleted; and
moving any existing object in the partition that is subject to an update to another partition in the search index along with an indexing request.
2. The system of claim 1, wherein the indexing request is generated by a partition manager based on the update associated with the existing object.
3. The system of claim 2, wherein the moving is automatically and proactively performed by the partition manager when the capacity utilization of the partition is above the second threshold.
4. The system of claim 2, wherein the moving is performed by the partition manager in response to another indexing request received at the partition manager.
5. The system of claim 1, wherein the instructions when translated by the processor further cause the search engine to perform:
changing from an Add mode to an Update mode when the capacity utilization of the partition meets or exceeds the first threshold; and
changing from the Update mode to a Rebalancing mode when the capacity utilization of the partition meets or exceeds the second threshold.
6. The system of claim 1, wherein the instructions when translated by the processor further cause the search engine to perform:
changing from an Update mode to an Add mode when the capacity utilization of the partition drops below the first threshold.
7. The system of claim 1, wherein the instructions when translated by the processor further cause the search engine to perform:
changing from a Rebalancing mode to an Update mode when the capacity utilization of the partition drops below a third threshold.
8. A computer program product comprising at least one non-transitory computer readable medium storing instructions translatable by a processor, wherein the instructions when translated by the processor cause a search engine to perform:
when a capacity utilization of a partition in a search index is under a first threshold:
allowing any object to be added to the partition; and
allowing any existing object in the partition to be updated or deleted;
when the capacity utilization of the partition is above the first threshold:
disallowing any object to be added to the partition; and
allowing any existing object in the partition to be updated or deleted; and
when the capacity utilization of the partition is above a second threshold:
disallowing any object to be added to the partition;
allowing any existing object in the partition to be deleted; and
moving any existing object in the partition that is subject to an update to another partition in the search index along with an indexing request, the search index being connected to the search engine, the search engine being communicatively connected to a client device over a physical connection.
9. The computer program product of claim 8, wherein the indexing request is generated by a partition manager based on the update associated with the existing object.
10. The computer program product of claim 9, wherein the moving is automatically and proactively performed by the partition manager when the capacity utilization of the partition is above the second threshold.
11. The computer program product of claim 9, wherein the moving is performed by the partition manager in response to another indexing request received at the partition manager.
12. The computer program product of claim 8, wherein the instructions when translated by the processor further cause the search engine to perform:
changing from an Add mode to an Update mode when the capacity utilization of the partition meets or exceeds the first threshold; and
changing from the Update mode to a Rebalancing mode when the capacity utilization of the partition meets or exceeds the second threshold.
13. The computer program product of claim 8, wherein the instructions when translated by the processor further cause the search engine to perform:
changing from an Update mode to an Add mode when the capacity utilization of the partition drops below the first threshold.
14. The computer program product of claim 8, wherein the instructions when translated by the processor further cause the search engine to perform:
changing from a Rebalancing mode to an Update mode when the capacity utilization of the partition drops below a third threshold.
15. A method for managing partitions in a computerized search system, the method performed by a search engine embodied on at least one non-transitory computer readable medium including instructions translable by a processor, the method comprising:
when a capacity utilization of a partition in a search index is under a first threshold:
allowing any object to be added to the partition; and
allowing any existing object in the partition to be updated or deleted;
when the capacity utilization of the partition is above the first threshold:
disallowing any object to be added to the partition; and
allowing any existing object in the partition to be updated or deleted; and
when the capacity utilization of the partition is above a second threshold:
disallowing any object to be added to the partition;
allowing any existing object in the partition to be deleted; and
moving any existing object in the partition that is subject to an update to another partition in the search index along with an indexing request, the search index being connected to the search engine, the search engine being communicatively connected to a client device over a physical connection.
16. The method of claim 15, wherein the indexing request is generated by a partition manager based on the update associated with the existing object.
17. The method of claim 16, wherein the moving is automatically and proactively performed by the partition manager when the capacity utilization of the partition is above the second threshold.
18. The method of claim 16, wherein the moving is performed by the partition manager in response to another indexing request received at the partition manager.
19. The method of claim 15, further comprising:
changing from an Add mode to an Update mode when the capacity utilization of the partition meets or exceeds the first threshold; and
changing from the Update mode to a Rebalancing mode when the capacity utilization of the partition meets or exceeds the second threshold.
20. The method of claim 15, further comprising:
changing from a Rebalancing mode to an Update mode when the capacity utilization of the partition drops below a third threshold.
US13/595,031 2011-08-30 2012-08-27 System and method of managing capacity of search index partitions Active 2033-05-20 US8909615B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/595,031 US8909615B2 (en) 2011-08-30 2012-08-27 System and method of managing capacity of search index partitions
US14/539,542 US9836541B2 (en) 2011-08-30 2014-11-12 System and method of managing capacity of search index partitions

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201161528967P 2011-08-30 2011-08-30
US13/595,031 US8909615B2 (en) 2011-08-30 2012-08-27 System and method of managing capacity of search index partitions

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/539,542 Continuation US9836541B2 (en) 2011-08-30 2014-11-12 System and method of managing capacity of search index partitions

Publications (2)

Publication Number Publication Date
US20140181071A1 US20140181071A1 (en) 2014-06-26
US8909615B2 true US8909615B2 (en) 2014-12-09

Family

ID=50975883

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/595,031 Active 2033-05-20 US8909615B2 (en) 2011-08-30 2012-08-27 System and method of managing capacity of search index partitions
US14/539,542 Active US9836541B2 (en) 2011-08-30 2014-11-12 System and method of managing capacity of search index partitions

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/539,542 Active US9836541B2 (en) 2011-08-30 2014-11-12 System and method of managing capacity of search index partitions

Country Status (1)

Country Link
US (2) US8909615B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150074080A1 (en) * 2011-08-30 2015-03-12 Open Text SA System and method of managing capacity of search index partitions

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9165406B1 (en) 2012-09-21 2015-10-20 A9.Com, Inc. Providing overlays based on text in a live camera view
US9047326B2 (en) 2012-10-12 2015-06-02 A9.Com, Inc. Index configuration for searchable data in network
US9507750B2 (en) * 2012-10-12 2016-11-29 A9.Com, Inc. Dynamic search partitioning
US9299123B2 (en) * 2013-01-22 2016-03-29 Qualcomm Incorporated Indexed streamout buffers for graphics processing
US9087055B2 (en) * 2013-01-28 2015-07-21 International Business Machines Corporation Segmenting documents within a full text index
US9613096B2 (en) * 2014-03-04 2017-04-04 International Business Machines Corporation Dynamic result set caching with a database accelerator
CN104965826B (en) * 2014-04-18 2019-04-16 腾讯科技(深圳)有限公司 Search method and retrieval device based on browser
ES2813903T3 (en) 2014-04-30 2021-03-25 Huawei Tech Co Ltd Device and search method
US9842428B2 (en) * 2014-06-27 2017-12-12 Samsung Electronics Co., Ltd. Dynamically optimized deferred rendering pipeline
CN104598550B (en) * 2014-12-31 2018-09-25 北京奇艺世纪科技有限公司 A kind of update method and device of Internet video index
US11062129B2 (en) * 2015-12-30 2021-07-13 Veritas Technologies Llc Systems and methods for enabling search services to highlight documents
CN107203557A (en) * 2016-03-17 2017-09-26 伊姆西公司 The method and device of object to be searched for handling
US10263877B1 (en) 2016-08-12 2019-04-16 Pinterest, Inc. Generating collections of sets based on user provided annotations
US10936613B2 (en) 2017-05-03 2021-03-02 Servicenow, Inc. Table-per-partition
EP3649564A4 (en) * 2017-07-06 2020-12-16 Open Text SA ULC System and method of managing indexing for search index partitions
US11182437B2 (en) * 2017-10-26 2021-11-23 International Business Machines Corporation Hybrid processing of disjunctive and conjunctive conditions of a search query for a similarity search
CN111143113B (en) 2018-11-02 2023-07-07 伊姆西Ip控股有限责任公司 Method, electronic device and computer program product for copying metadata
RU2733482C2 (en) 2018-11-16 2020-10-01 Общество С Ограниченной Ответственностью "Яндекс" Method and system for updating search index database
CN110096478B (en) * 2019-05-09 2021-06-29 中国联合网络通信集团有限公司 Document index generation method and device
CN114328558B (en) * 2021-12-30 2022-12-27 北京达佳互联信息技术有限公司 List updating method, apparatus, device and storage medium
US11947822B2 (en) 2022-03-31 2024-04-02 Microsoft Technology Licensing, Llc. Maintaining a record data structure using page metadata of a bookkeeping page
CN115809268B (en) * 2022-11-23 2024-01-16 深圳计算科学研究院 Adaptive query method and device based on fragment index

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050165750A1 (en) * 2004-01-20 2005-07-28 Microsoft Corporation Infrequent word index for document indexes
US7174346B1 (en) * 2003-07-31 2007-02-06 Google, Inc. System and method for searching an extended database
US7293016B1 (en) * 2004-01-22 2007-11-06 Microsoft Corporation Index partitioning based on document relevance for document indexes
US20090157666A1 (en) * 2007-12-14 2009-06-18 Fast Search & Transfer As Method for improving search engine efficiency
US20100306238A1 (en) * 2009-05-29 2010-12-02 International Business Machines, Corporation Parallel segmented index supporting incremental document and term indexing

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5537588A (en) * 1994-05-11 1996-07-16 International Business Machines Corporation Partitioned log-structured file system and methods for operating the same
US6125370A (en) * 1998-04-01 2000-09-26 International Business Machines Corporation Repartitioning data
US6269373B1 (en) * 1999-02-26 2001-07-31 International Business Machines Corporation Method and system for persisting beans as container-managed fields
US7136883B2 (en) * 2001-09-08 2006-11-14 Siemens Medial Solutions Health Services Corporation System for managing object storage and retrieval in partitioned storage media
US7146389B2 (en) * 2002-08-30 2006-12-05 Hitachi, Ltd. Method for rebalancing free disk space among network storages virtualized into a single file system view
US7809762B1 (en) * 2003-07-11 2010-10-05 Bmc Software, Inc. Outage-less database change operation
US7644087B2 (en) * 2005-02-24 2010-01-05 Xeround Systems Ltd. Method and apparatus for data management
US8364677B2 (en) * 2008-08-01 2013-01-29 International Business Machines Corporation Method and apparatus for generating partitioning keys for a range-partitioned database
US8732139B2 (en) * 2008-12-18 2014-05-20 Sap Ag Method and system for dynamically partitioning very large database indices on write-once tables
US8533181B2 (en) * 2009-04-29 2013-09-10 Oracle International Corporation Partition pruning via query rewrite
US9477671B2 (en) * 2009-05-27 2016-10-25 Oracle International Corporation System and method for implementing effective date constraints in a role hierarchy
US8930332B2 (en) * 2010-03-12 2015-01-06 Salesforce.Com, Inc. Method and system for partitioning search indexes
WO2012072879A1 (en) * 2010-11-30 2012-06-07 Nokia Corporation Method and apparatus for updating a partitioned index
US8909615B2 (en) 2011-08-30 2014-12-09 Open Text S.A. System and method of managing capacity of search index partitions
US8898204B1 (en) * 2011-10-21 2014-11-25 Applied Micro Circuits Corporation System and method for controlling updates of a data structure

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7174346B1 (en) * 2003-07-31 2007-02-06 Google, Inc. System and method for searching an extended database
US20050165750A1 (en) * 2004-01-20 2005-07-28 Microsoft Corporation Infrequent word index for document indexes
US7293016B1 (en) * 2004-01-22 2007-11-06 Microsoft Corporation Index partitioning based on document relevance for document indexes
US20090157666A1 (en) * 2007-12-14 2009-06-18 Fast Search & Transfer As Method for improving search engine efficiency
US20100306238A1 (en) * 2009-05-29 2010-12-02 International Business Machines, Corporation Parallel segmented index supporting incremental document and term indexing

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150074080A1 (en) * 2011-08-30 2015-03-12 Open Text SA System and method of managing capacity of search index partitions
US9836541B2 (en) * 2011-08-30 2017-12-05 Open Text Sa Ulc System and method of managing capacity of search index partitions

Also Published As

Publication number Publication date
US20150074080A1 (en) 2015-03-12
US20140181071A1 (en) 2014-06-26
US9836541B2 (en) 2017-12-05

Similar Documents

Publication Publication Date Title
US9836541B2 (en) System and method of managing capacity of search index partitions
US11748323B2 (en) System and method of search indexes using key-value attributes to searchable metadata
US11853334B2 (en) Systems and methods for generating and using aggregated search indices and non-aggregated value storage
US11669576B2 (en) System, method and computer program product for protecting derived metadata when updating records within a search engine
US9367581B2 (en) System and method of quality assessment of a search index
US11449564B2 (en) System and method for searching based on text blocks and associated search operators
US11977581B2 (en) System and method for searching chains of regions and associated search operators
US20230109772A1 (en) System and method for value based region searching and associated search operators
US20200218705A1 (en) System and method of managing indexing for search index partitions
JPWO2012081165A1 (en) Database management apparatus and database management method

Legal Events

Date Code Title Description
AS Assignment

Owner name: OPEN TEXT S.A., LUXEMBOURG

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PIDDUCK, PATRICK THOMAS SIDNEY;CZARLINSKA, ALEKSANDRA;SIGNING DATES FROM 20120821 TO 20120824;REEL/FRAME:028849/0360

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: OPEN TEXT SA ULC, CANADA

Free format text: CERTIFICATE OF AMALGAMATION;ASSIGNOR:IP OT SUB ULC;REEL/FRAME:039872/0662

Effective date: 20160708

Owner name: OT IP SUB, LLC, DELAWARE

Free format text: IP BUSINESS SALE AGREEMENT;ASSIGNOR:OPEN TEXT S.A.;REEL/FRAME:039872/0605

Effective date: 20160701

Owner name: IP OT SUB ULC, CANADA

Free format text: CERTIFICATE OF CONTINUANCE;ASSIGNOR:OT IP SUB, LLC;REEL/FRAME:039986/0689

Effective date: 20160702

FEPP Fee payment procedure

Free format text: SURCHARGE FOR LATE PAYMENT, LARGE ENTITY (ORIGINAL EVENT CODE: M1554)

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

AS Assignment

Owner name: BARCLAYS BANK PLC, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:OPEN TEXT SA ULC;REEL/FRAME:063559/0758

Effective date: 20230501

Owner name: BARCLAYS BANK PLC, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:OPEN TEXT SA ULC;REEL/FRAME:063559/0647

Effective date: 20230430

Owner name: BARCLAYS BANK PLC, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:OPEN TEXT SA ULC;REEL/FRAME:063559/0628

Effective date: 20230428

AS Assignment

Owner name: THE BANK OF NEW YORK MELLON, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:OPEN TEXT SA ULC;REEL/FRAME:064761/0256

Effective date: 20230430

AS Assignment

Owner name: OPEN TEXT SA ULC, CANADA

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS (REEL/FRAME 063559/0628);ASSIGNOR:BARCLAYS BANK PLC;REEL/FRAME:067807/0053

Effective date: 20240621