WO2000033216A1 - A natural knowledge acquisition method - Google Patents

A natural knowledge acquisition method Download PDF

Info

Publication number
WO2000033216A1
WO2000033216A1 PCT/US1999/028226 US9928226W WO0033216A1 WO 2000033216 A1 WO2000033216 A1 WO 2000033216A1 US 9928226 W US9928226 W US 9928226W WO 0033216 A1 WO0033216 A1 WO 0033216A1
Authority
WO
WIPO (PCT)
Prior art keywords
text information
information
database
semantic
text
Prior art date
Application number
PCT/US1999/028226
Other languages
French (fr)
Inventor
James D. Pustejovsky
John H. Clippinger
Robert Ingria
Original Assignee
Lexeme Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lexeme Corporation filed Critical Lexeme Corporation
Priority to AU19263/00A priority Critical patent/AU1926300A/en
Priority to EP99962917A priority patent/EP1151401A4/en
Publication of WO2000033216A1 publication Critical patent/WO2000033216A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • G06F40/35Discourse or dialogue representation

Definitions

  • This invention generally relates to the field of information management. More particularly, the present invention provides a technique including a method for extraction and automatic classification of document content for any machine-readable text.
  • LR Information retrieval
  • the indexing technique includes full-text indexing, in which content words in a document are used as keywords.
  • Full text searching had been one of the most promising of recent LR approaches.
  • full text searching has many limitations. For example, full text searching lacks precision and often retrieves literally thousands of "hits" or related documents, which then require further refinement and filtering. Additionally, full text searching has limited recall characteristics. Accordingly, full text searching has much room for improvement.
  • Domain knowledge techniques can enhance an effectiveness of a full-text searching system.
  • Domain knowledge techniques often provide related terms that can be used to refine the full-text searching process. That is, domain knowledge often can broaden, narrow, or refocus a query at retrieval time. Likewise, domain knowledge may be applied at indexing time to do word sense disambiguation or simple content analysis. Unfortunately, for many domains, such knowledge, even in the form of a thesaurus, is either generally not available, or is often incomplete with respect to the vocabulary of the texts indexed.
  • the method and system described in Dahlgren employs a natural language understanding system to provide a "concept annotation" of text for subsequent retrieval. Furthermore, when the system is used to query a database, it matches on pointers to the text provided by the annotation rather than an answer to the query.
  • the results of the queries are "hits” rather than “answers”; that is, a hit is the entire text that matches the indexing criteria, while an answer on the other hand is the actual utterance (or portion of the text) that satisfied a user query. For example, if the query were "Who are the officers of Microsoft, Inc?", a hit-based system would return ali the documents that contain this information anywhere within them, whereas an answer- based system would return the actual value of the answer, namely the officers.
  • the present invention provides a method using a combination of syntactic and semantic information objects.
  • the present invention provides a natural language database forming method. The method includes providing text information comprising a plurality of related words. A step of tagging each word in the text information also is included. The method forms an object that has syntactic information and semantic information from each word in the text information. The object is placed or mapped into a object oriented relational database.
  • the present invention provides a natural language knowledge acquisition method.
  • the method includes providing text information (e.g., electronic form) including a plurality of related words.
  • the method tags each word in the text information.
  • the method also forms an object comprising syntactic information and semantic information from each word in the text information.
  • the object is placed into a relational, object-oriented, or mixed relational/object oriented database.
  • the methods of providing, tagging, forming, and placing are repeated to populate the database.
  • a user can access the information in the database.
  • the user forms a query, which is entered, processed by the system, and selects an object based upon entity relationships to achieve a unique output, which can actually be an answer to the query.
  • the present invention provides a method for recognizing lexical objects within text and typing these lexical objects into semantic categories. These semantic representations can then be utilized in a variety of ways, including, persisting them in various forms (such as relational, object, or mixed object/relational databases), text summarization, keyword extraction, semantic indexing. Moreover, this method of deriving semantic representations from lexical objects in input text can be used both for extracting information and for querying an already existing database of knowledge (including those created by this engine). Thus both database population and database retrieval of said objects may be performed. Numerous advantages are achieved by way of the present invention. In one embodiment, the present invention provides a relational database that can be queried using a natural language approach.
  • the present invention provides methods using a combination of data coupled with logic.
  • the invention can also provide knowledge extraction in other embodiments.
  • the invention provides object creation, and provides conversational access to the database in other embodiments. Accordingly, the present invention can provide an acquisition technique that actually provides answers to queries (rather than hits), which can be singular.
  • queries queries
  • the present invention can provide an acquisition technique that actually provides answers to queries (rather than hits), which can be singular.
  • one or more of these advantages can be present.
  • FIG. 1 is a simplified diagram of an information acquisition method according to an embodiment of the present invention.
  • Fig. 2 is a simplified diagram of an information acquisition method according to an alternative embodiment of the present invention.
  • the present invention provides a method using a combination of syntactic and semantic information objects.
  • the present invention provides a modular, object-oriented, and collaborative approach to semantic typing, interpreting, and extraction, of knowledge objects from text sources into databases.
  • the invention provides a highly general object- oriented method for using lexically-based knowledge to identify and extract semantic objects in text and to represent them in a database.
  • the invention offers savings in the time, effort, and costs for constructing, populating, and updating a wide variety of databases.
  • a method according the present invention is briefly outlined below: 1. Providing text information sources; 2. Tokenizing;
  • Step 7 can use either a relational, object-oriented, or mixed relational/object database. If a relational database is used, the step 7 of translation into the relational model is required (step 6). Any relational database can be used in Step 7. As merely an example, the database can be made by a company called Oracle of Redwood City, California. Alternatively, other companies such as Informix, Sybase, and others also manufacture database designs that can incorporate the present invention. Similarly, any object-oriented or mixed relational/object database can be used. Details of the above steps are briefly described according to Fig. 1, for example.
  • Fig. 1 is a simplified diagram 100 of an information acquisition method according to an embodiment of the present invention.
  • the method 100 begins at start, step 101, which includes steps of providing text information (step 103), tokenizing (step 105), tagging (step 107), stemming (step 109), interpreting (step 111), and extracting objects (step 125). Further details of each of these steps are provided below.
  • the method provides (step 103) information such as text information from a variety of sources, including all digital sources.
  • the sources include, among others, newspapers, magazines, research, web sites, product information, internets, ,intranets, and spoken language inputs.
  • the text source are generally in electronic form, which can be read, organized, and categorized by way of a computer.
  • the text source can be in any suitable electronic form such as ASCII, HTML, XML, LaTeX, word processing applications, and presentation slides (e.g., Microsoft Power PointTM).
  • the method tokenizes the text information.
  • the text may be "tokenized,” for example, split up into textual elements separated by a delimiter, such as a "white space” or "blank” character. Tokenization normalizes the input text into a form that is usable by subsequent steps " of the method.
  • the tokenizer separates punctuation (e.g., periods, apostrophes, quotes, etc.) from words.
  • punctuation e.g., periods, apostrophes, quotes, etc.
  • the present invention provides a step of tagging (step 107) the tokenized text information.
  • a part-of-speech (herein "POS") tagger can be used.
  • a goal of the part-of-speech tagger is to assign grammatical category labels to each tokenized element in the text produced from the tokenizer. For example, a tagger will convert the input below:
  • the present method performs a step 109 of stemming.
  • Stemming is yet another stage in normalization for further processing.
  • all stems are orthographically lower case.
  • the stem will be the dictionary look up form of the token (e.g., 'man' for 'men', 'run' for 'ran').
  • Stemming can use dictionary lookup in the case of known inflected words.
  • He/PRP is/NBZ baseoYNB ⁇ in/I ⁇ Panama ⁇ P as/T ⁇ chief/ ⁇ ofI ⁇ U.S./ ⁇ P forces/ ⁇ S in/T ⁇ Latin/ ⁇ P America/ ⁇ P and/CC ould/MD replace/NB retiring/NBG U.S./ ⁇ P Army ⁇ P Gen/ ⁇ P ./. George ⁇ P Joulwan/ ⁇ P as/T ⁇ Supreme ⁇ P Allied/ ⁇ P Commander/ ⁇ P of I ⁇ ⁇ ATO ⁇ P in/T ⁇ Europe/ ⁇ P (/(
  • the method performs interpreting (step 111), using an interpreting module or the like.
  • a specific embodiment of the interpreting includes three sub- modules: accessing of a lexicon and type system 113, parsing 115, and identification of semantic types 117, e.g., qualia roles.
  • the first sub-module 113 accessing of a lexicon and type system 113, uses two knowledge bases.
  • the first is a lexicon (resource A), indexed by stem within a particular part of speech. For example, 'base' as a noun and 'base' as a verb will have two separate entries.
  • Each lexical entry contains a type property which is a name of a semantic type in the type system (resource B).
  • Each lexical entry contains appropriate syntactic information. This information is combined to create the appropriate type of syntactic constituent (e.g., a noun for a stem with a noun tag), with the appropriate semantic representation.
  • the parser sub-module 115 takes the output of the accessing of a lexicon and type system sub-module 113, and composes these into larger syntactic and semantic structures that make up the sentences of text in natural language.
  • This sub- module 115 uses an engine embodying an all paths parser (Younger, D., "Recognition and Parsing of Context-free Languages in Time n 3 ", Information and Control 10:189-208,
  • syntactic- semantic composition is accomplished by means of grammar rules, 120, which specify both how syntactic elements are to be combined, and also how their semantic interpretations are to be composed.
  • grammar rules can contain constraints which specify the conditions under which a rule can apply. If the constraint fails, the rule is not even considered, which improves the performance of the interpreter, since useless search paths are not pursued.
  • This rule states that a VP (Verb Phrase) can be constructed from a VP followed by an NP (Noun Phrase).
  • ⁇ Transitive ⁇ represents a constraint. This rule may only be fired if the VP is transitive: i.e., if the semantics of the VP allows for a direct object and this direct object position has not yet been filled.
  • [DirectObject] represents a role, i.e., names of pieces of code that contain (1) constraints on the other dependent(s) of the rule; and (2) specify how the semantics of the dependents of the rule are combined to create the semantics of the dominating constituent (i.e. the left hand side). This is the semantic composition that takes place if the rule succeeds.
  • the semantics of the NP is checked for compatibility with the semantic type specified for the direct object in the semantics of the VP. Moreover, if the type of the NP is compatible with the type of the direct object but is less specific, the type of the NP will be changed to the more specific type. It is in this way that the system acquires new knowledge, by using its existing knowledge bases to learn or further specify the meanings of words it has not previously encountered.
  • the semantics of the NP is then bound to the direct object position of the semantics of the existing VP, and this new semantic representation is made the semantics of the newly constructed VP.
  • the present interpreter 111 uses a single Interpreter object.
  • the interpreter 111 can manipulate a plurality of types of data structures such as those noted below:
  • WH elements i.e., words like 'who' and 'what' that appear 'dislocated' from the 'logical' syntactic positions in which they receive their semantic interpretation).
  • Edges interact to create new Edge on the basis of two other classes of objects: GrammarRule objects, which represent information about how new constituents can be built out of existing constituents, and Constituent objects, which represent traditional grammatical elements such as nouns, verbs, and sentences.
  • the interpreting process can be defined as follows: the Interpreter is sent the output of the Stemmer, which is an Array of underspecified objects of type Constituent. Each of these objects includes the following pieces of information:
  • token the unit of the input string (i.e. orthographic word or punctuation) found by the Tokenizer; b).
  • tag the part of speech tag assigned by the Tagger; c).
  • stem the dictionary lookup form of the token added by the Stemmer; d).
  • offset the numerical position of the current token in the input text computed by the Stemmer.
  • the Interpreter object processes each of the underspecified Constituent objects in its input in sequence and tells it to transform itself into a fully specified object of the appropriate syntactic category.
  • the Interpreter causes an associated Edge object to be created, which stores its paired constituent in an instance variable.
  • the Interpreter consults the GrammarRule class object (step 120) to find out what grammar rules involve the new constituent.
  • the Interpreter causes a new Edge to be built, which stores information about the rule that sanctioned its creation and the constituent that corresponds to the portion of the rule already found in the input. To illustrate this process, let us look at what happens when the Interpreter object receives the output of the Stemmer for the input text 'green apples', given the existence of a GrammarRule object that constructs a NounPhrase out of an Adjective and a Noun.
  • the Stemmer's output has two Constituents, the first with the token 'green' and tag 'JJ'; the second with the token 'apples' and tag *NNS'.
  • the Interpreter object processes these constituents in left to right order in a specific embodiment.
  • the Interpreter therefore, causes a new Edge to be built, which records that it has found the newly created Adjective object, and that it is trying to form a NounPhrase on the basis of the rule just given.
  • the Interpreter looks for any existing Edges (to the right of the Adjective), that the newly created Edge can interact with, but at this point there are none.
  • the Interpreter causes a new associated Edge object to be built. It also looks for any (active) Edges immediately preceding 'apples'. It finds one, the active Edge built off of the Adjective object 'green', which is looking for an immediately following Noun.
  • the Interpreter passes the new Edge to this existing Edge, which checks to see if it can use the newly formed Edge. It can, and therefore adds the new Noun to its collection of found constituent and marks itself as complete, i.e. as having found a NounPhrase.
  • the Interpreter will also, as in the case of the preceding Adjective object, look for any new rules that can use the newly found Noun object and, if any such rules are found, will cause the corresponding Edges to be built, and will look for any Edges to the right of these new Edges that they can consume. These activities are irrelevant in the current example, however, so we omit the details.
  • the present method using selected grammar rules (step 120) for a interpreter (step 111).
  • the rules are noted below and are defined according to the following representations.
  • the elements between curly brackets ⁇ are constraints, i.e. names of pieces of code that the left edge of a rule (the left-most element on the right hand side) should satisfy before the rule will be considered.
  • the elements between square brackets [] are roles, e.g., [DirectObject];
  • ComplementVP rules (3): ComplementVP > WhPhrase ComplementVP [WhQuestion]
  • ComplementVP COMP ComplementVP ⁇ InfinitivalComp ⁇ [Complementizer]
  • ComplementVP TO VP [Infinitive]
  • VBar rules (6): VBar > V
  • VBar VBar ⁇ PossibleAuxiliary ⁇ [Auxiliary]
  • VBar V NEG VBar ⁇ PossibleAuxiliary ⁇ [Auxiliary] [Negation]
  • VBar Modal VBar [Auxiliary]
  • VBar Modal NEG VBar [Auxiliary] [Negation]
  • VBar AdvP VBar [AdverbialModifier]
  • VP VP NP ⁇ Copula ⁇ [PredicateNominal]
  • VP VP AdjP ⁇ Copula ⁇ [Predicate Adjective]
  • VP VP AdjP ⁇ TakesAdjectiveComplement ⁇ [AdjectiveComplement]
  • VP > VP PP ⁇ TakesPPComplement ⁇ [PPComplement]
  • VP > VP Complements ⁇ TakesClause ⁇ [ClausalComplement]
  • VP > VP NP ⁇ TakesClause ⁇ [QuestionOnClausalComplement]
  • NBar N NBar ⁇ PossiblePreName ⁇ [PreNameModifier]
  • NBar V NBar ⁇ PossibleVerbalModifier ⁇ [VerbalModifier]
  • NBar AdjBar NBar [AdjectiveModifier]
  • CoreNP Title NBar [TitleModifier]
  • CoreNP DeterminerGroup NBar [NPSpec]
  • NP NP Complements ⁇ TakesClause ⁇ [ClausalComplement]
  • NP NP NumAppositive [NumAppositiveModifier]
  • AdjBar NBar Punctuation AdjBar
  • AdjP AdjBar
  • the present invention uses a combination of syntactic and semantic composition.
  • syntactic and semantic composition As noted, the previous portions of the specification described how the interactions of objects of the classes Interpreter, Edge, GrammarRule, and Constituent produce syntactic interpret structures. This description was simplified in one respect, however: GrammarRules, and the Edges that are associated with them, do not merely check to determine whether the constituent associated with a candidate Edge matches the syntactic category specified in a rule: they also check for the eligibility of the candidate in terms of more fine-grained syntactic and/or semantic information.
  • Edges create not only new syntactic constituents from the information contained in grammar rules, they also compose the semantics of the syntactic dependents of that constituent, to form a new semantic object that is the associated meaning representation of the newly constructed constituent.
  • the finer grained syntactic and semantic well-formedness conditions can be expressed in the form of Roles,e.g., [DirectObject], which are pieces of code associated with GrammarRule objects.
  • Roles e.g., [DirectObject]
  • the semantics associated with the candidate NounPhrase is compatible with the semantic type requirement imposed on the Subject argument by the VerbPhrase semantics (e.g. the semantics of the verb 'sell' requires that its Subject be either a Person or an Organization).
  • Roles can be represented in rules, for example, as simple symbolic names. These names are often associated with the actual code that is used to perform the type checking and semantic composition by lookup in a table. This allows the code for the same named Role to be used in multiple rules. For example, the Subject role will appear in main clause, complement clauses, relative clauses, declarative sentences, questions, etc.
  • Semantic representations can be constructed during the course of interpreting.
  • the semantic representations associated with phrases and clauses are created by the interpreting process, by means of the composition of the semantics of dependent constituents; e.g., a sentence gets its semantics by the composition of the semantics of its subject NounPhrase and main predicate VerbPhrase; a transitive VerbPhrase gets its semantics from the composition of the semantics of its head Verb and its NounPhrase direct object, etc.
  • PreTerminals i.e. the constituents corresponding to the actual words in the sentence, such as Noun, Verb, Adjective get their semantics either by lexical lookup or by default.
  • a new, fully specified PreTerminal object such as a Noun
  • a LexicalEntry class object of the appropriate type (e.g. NounEntry for Noun, VerbEntry for Verb, etc.) to determine if its stem has an associated lexical entry in that category. If it does, the PreTerminal object uses the semantic information in the LexicalEntry to create its associated semantic representation object. If it does not, the PreTerminal creates a default semantic representation appropriate for its syntactic category; e.g.
  • the Interpreting module (step 111) also uses a sub-module for identification of semantic types associated with words and phrases (step 117).
  • the present invention uses this sub-module 117 that contains the core conceptual knowledge for the system.
  • the sub-module 117 provides a structure for a semantic type system 113 (resource B), which underlies the processes of lexical inference.
  • the type system 113 is structured along multiple dimensions, where each dimension corresponds to a different aspect of word meaning. As a result each dimension involves a different way of understanding a given entity in the domain and thus corresponds to a different set of questions (i.e. queries) concerning that entity.
  • each conceptual type in the type system is a data structure that incorporates the set of inferences that are available as well as the relations between that type and other entities.
  • the present invention provides types as identifiers for other types.
  • types are not merely used for structuring information for instances of each particular type, but they also play a crucial role in identifying other types in the text.
  • the structure of the typing information drives part of the knowledge acquisition process.
  • the information that is associated with a noun denoting a professional role i.e. "doctor”
  • the noun ""maker” becomes a fairly reliable identifier of entities that are typed as "products”, as in "software maker”.
  • the present invention also provides a qualia structure as a basis for acquiring knowledge.
  • the qualia structure of a lexical item can be identified in terms of patterns occurring in the text.
  • Each qualia represents a well defined component of meaning which is talked about in texts as well as in people's conversations.
  • an object which occurs with the predicate "manufacture” — as its direct object — is understood as an artifactual entity which is brought about through a manufacturing process. This information is exploited to acquire information concerning the AGENTIVE role of a given entity.
  • This strategy is extensible at different levels of specificity, in a way that allows the engine to acquire information concerning the relationship between Microsoft and Windows 98.
  • Patterns that identify information concerning the TELIC of an entity can be useful for simultaneously acquiring the same information about another entity from a different perspective. For example, while we are assigning Microsoft as the entity that specifies the AGENTIVE role of one of its products, we are also building that information as part of Microsoft's TELIC. Qualia Structure can also be used as a basis for querying and reasoning over a database. Given the close relation between syntactic patterns and the semantics they convey relative to a given entity, the querying exploits the same principles used in the knowledge acquisition. The process, however, is taken further. Once we have available the information concerning the AGENTIVE role of a set of different software products, for instance, then it is possible to ask about competitors.
  • the S&P500 stock index rose 36.46 points.
  • Tokenized and tagged The/DT S&P500/NNP stock NN index/NN rose/VBD 36.46/CD points/NNS ./.
  • Lexicon system (Resource AV)
  • a specific embodiment of the present method uses a step of extracting objects, step 125.
  • the present method creates a semantic representation of objects, during interpreting, including syntactic-semantic composition, and qualia induction, that serve as an interface to a relational database model. Since these semantic representation objects are a combination of data and procedures, similar to objects in an object-oriented systems, they implement their own procedures (i.e., methods) to translate themselves from their object representation into SQL statements that map to a relational database. While there are several classes of semantic representation objects (LexLF) that are used during the course of interpreting and semantic composition, there are two classes that are relevant to interactions with the relational database: EntityLexLF — which represents the semantics of entities (i.e. person, places, things, concepts)
  • Both of these classes implement methods for two modes of database interaction: insertion and retrieval.
  • the LexLF transforms its semantics into a SQL INSERT statement to add information to the database. Since a LexLF contains various pieces of information that must be present in the database, the LexLF negotiates with the database to discover whether supporting pieces of information exist already or not. For example, FunctionLexLF contains information about the predicate type of the associated predicate and about the various entities that fill its argument positions. Each of these elements must exist in the appropriate table of the database; when one of these elements already exists, the LexLF merely looks up the primary key for it; when it does not, the LexLF must first cause the element to be inserted in the database, to preserve relational integrity.
  • FunctionLexLF contains information about the predicate type of the associated predicate and about the various entities that fill its argument positions. Each of these elements must exist in the appropriate table of the database; when one of these elements already exists, the LexLF merely looks up the primary key for it; when it does not, the LexLF must first cause the element to be inserted in the database, to preserve relational integrity.
  • EntityLexLF for '36.46 points' produces the following: insert into Entities(CanonicalName) values('point') insert into Types(EntityLD, DocumentID, Offset, Type) values(5231,405,380,'Measure')
  • step 125 Once the objects have been extracted, step 125, they can be mapped onto a relational database, which is illustrated by a simplified method diagram of Fig. 2. This diagram is merely an example and should not limit the scope of the claims herein. One of ordinary skill in the art would recognize other variations, modifications, and alternatives.
  • the present method 200 begins at step 201, which can derive from step 125 in the above
  • the method maps (step 203) objects in a relational manner, which can be defined by an entity relation diagram, for example.
  • the objects are inserted (step 205) into an objected oriented relational database 211. Once the objects have been inserted they are now ready for use by a user.
  • the user can query (step 207) the database 211.
  • the present process stops at step 209, but is not limited. As merely an example, which uses the above database population technique, it would be possible to find an answer to the following question:
  • semantics of the interrogative pronoun 'What' is interpreted in its 'logical' position, i.e. as the direct object of the main verb 'do'.
  • semantic representation of 'What' includes a
  • QuantifierLexLF that has #Wh as the value of its #quantifier. This indicates that this is the logical argument that is being asked about in this query. Semantic representations for content queries of this type are processed for database lookup in the following manner.
  • the S&P500 stock index rose 36.46 points.
  • this information is passed to the user interface in the format:
  • ⁇ ⁇ ⁇ > which contains the source of the response text, a URL that points to the complete source document (the location of the host machine is not hard-coded, but is determined dynamically to point to the host machine that is running the system), and the actual response text.
  • This is presented in a more user-friendly format: only the name of the source and the response text are displayed, with the source name being a hyperlink that points to the full source text, to allow the user to examine the entire text, if desired.
  • the database lookup is done in two stages: an initial lookup of the EntityLD of the subject and then a lookup from the Relations table, using this EntityLD, rather than as a single select statement that does a database join over these two tables.
  • the reason for the decomposition of the lookup process in this manner is to allow for a more responsive interaction with the end user. Ifa simple join query fails to retrieve any data, the source of the failure is unknown. It could be due to, for example, one of several reasons:
  • the system can explore the possibility of presenting the user with a response that is more informative than a simple 'No Answer'. It can provide the user with an exact answer if one is available, and with the closest appropriate answer where an exact match is impossible. 'No Answer' responses are only given when the system cannot find anything that is relevant to the user's question.
  • retrieval e.g., query
  • the LexLF transforms its semantics into a SQL SELECT statement. This procedure is the mirror image of insertion in many ways: elements from multiple tables should be queried to produce the desired answer.
  • the LexLF may engage in iterative interactions with the database, as it falls back and widens the parameters of its retrieval space to find information relevant to the user's request, but which may not literally match the query. For example, if the user asks 'Did Brand X buy The Highpriced Spread?', the database may not contain information about the buy relation between Brand X and The Highpriced Spread, but it may contain other relations involving them.
  • the LexLF will engage in a series of transactions with the database to find something at least loosely matching the user's search parameters before giving up and returning no information. There are two sorts of transactions possible with the database, when there is no direct answer to the query: a. If there are other relations, in which the object being queried about appears, they will be presented, when related to the original query relation. b. If there are more general relations or object type descriptions than those present in the query, these will be returned.
  • the software functionality can be further combined or even separated.
  • the hardware functionality can be further combined, or even separated.
  • the software functionality can be implemented in terms of hardware or a combination of hardware and software.
  • the hardware functionality can be implemented in software or a combination of hardware and software. Any number of different combinations can occur depending upon the application.

Abstract

A natural language database forming method. The method includes providing text information (103) comprising a plurality of related words. A step of tagging (107) each word in the text information is also included. The method forms an object (125) that has syntactic information and semantic information from each word in the text information. The object is placed or mapped into an object oriented relational database (127).

Description

A NATURAL KNOWLEDGE ACQUISITION METHOD
CROSS-REFERENCES TO RELATED APPLICATIONS
This application claims priority from the following provisional patent application, the disclosure of which is herein incorporated by reference for all purposes:
U.S. Provisional Patent Application No. 60/110,190 in the names of James D. Pustejovsky, et al. titled,"Natural Knowledge Acquisition Method, System, and Code," filed November 30, 1998.
The following one commonly-owned co-pending provisional application is being filed concurrently and is hereby incorporated by reference in its entirety for all purposes:
U.S. Provisional Patent Application Serial No., , in the name of James D. Pustejovsky, titled, "A Method of Using a Natural Knowledge Acquisition System," (Attorney Docket Number 019497-000140)
BACKGROUND OF THE INVENTION This invention generally relates to the field of information management. More particularly, the present invention provides a technique including a method for extraction and automatic classification of document content for any machine-readable text.
The expansion of the Internet has proliferated "on-line" textual information. Such on-line textual information includes newspapers, magazines, WebPages, email, advertisements, commercial publications, and the like in electronic form. By way of the Internet, millions if not billions of pieces of information can be accessed using simple "browser" programs. Information retrieval (herein "LR") engines such as those made by companies such as Yahoo! allow a user to access such information using an indexing technique. The indexing technique includes full-text indexing, in which content words in a document are used as keywords. Full text searching had been one of the most promising of recent LR approaches. Unfortunately, full text searching has many limitations. For example, full text searching lacks precision and often retrieves literally thousands of "hits" or related documents, which then require further refinement and filtering. Additionally, full text searching has limited recall characteristics. Accordingly, full text searching has much room for improvement.
Techniques such as the use of "domain knowledge" can enhance an effectiveness of a full-text searching system. Domain knowledge techniques often provide related terms that can be used to refine the full-text searching process. That is, domain knowledge often can broaden, narrow, or refocus a query at retrieval time. Likewise, domain knowledge may be applied at indexing time to do word sense disambiguation or simple content analysis. Unfortunately, for many domains, such knowledge, even in the form of a thesaurus, is either generally not available, or is often incomplete with respect to the vocabulary of the texts indexed.
There have been attempts to use natural language understanding in some applications. As merely an example, U.S. Patent No. 5,794,050 in the names of Dahlgren et al. (herein Dahlgren.) utilized a conventional rule based system for providing searches on text information. Dahlgren, et al. use a naive semantic lexicon to "reason" about word senses. This simple semantic lexicon brings some "common sense" world knowledge to many stages of the natural language understanding process. Unfortunately, the design of such a semantic lexicon follows fairly standard taxonomic knowledge representation techniques, and hence the reasoning process making use of this taxonomy is generally incomplete. That is, it may provide a first level method for performing a relatively simple search, but often lacks a general ability to conduct a detailed retrieval to provide a comprehensive answer to a query. Fundamentally, the method and system described in Dahlgren, employs a natural language understanding system to provide a "concept annotation" of text for subsequent retrieval. Furthermore, when the system is used to query a database, it matches on pointers to the text provided by the annotation rather than an answer to the query.
Although some of the above techniques are fairly sophisticated compared to the information retrieval search engines so ubiquitous on the internet (e.g., Inktomi or Alta Vista), the results of the queries are "hits" rather than "answers"; that is, a hit is the entire text that matches the indexing criteria, while an answer on the other hand is the actual utterance (or portion of the text) that satisfied a user query. For example, if the query were "Who are the officers of Microsoft, Inc?", a hit-based system would return ali the documents that contain this information anywhere within them, whereas an answer- based system would return the actual value of the answer, namely the officers.
From the above, it is seen that a technique for improved information retrieval is highly desirable.
SUMMARY OF THE INVENTION According to the present invention, a technique including a method for acquiring information is provided. In a specific embodiment, the present invention provides a method using a combination of syntactic and semantic information objects. In a specific embodiment, the present invention provides a natural language database forming method. The method includes providing text information comprising a plurality of related words. A step of tagging each word in the text information also is included. The method forms an object that has syntactic information and semantic information from each word in the text information. The object is placed or mapped into a object oriented relational database.
In an alternative embodiment, the present invention provides a natural language knowledge acquisition method. The method includes providing text information (e.g., electronic form) including a plurality of related words. The method tags each word in the text information. The method also forms an object comprising syntactic information and semantic information from each word in the text information. The object is placed into a relational, object-oriented, or mixed relational/object oriented database. The methods of providing, tagging, forming, and placing are repeated to populate the database. Next, a user can access the information in the database. Here, the user forms a query, which is entered, processed by the system, and selects an object based upon entity relationships to achieve a unique output, which can actually be an answer to the query.
In another specific embodiment, the present invention provides a method for recognizing lexical objects within text and typing these lexical objects into semantic categories. These semantic representations can then be utilized in a variety of ways, including, persisting them in various forms (such as relational, object, or mixed object/relational databases), text summarization, keyword extraction, semantic indexing. Moreover, this method of deriving semantic representations from lexical objects in input text can be used both for extracting information and for querying an already existing database of knowledge (including those created by this engine). Thus both database population and database retrieval of said objects may be performed. Numerous advantages are achieved by way of the present invention. In one embodiment, the present invention provides a relational database that can be queried using a natural language approach. In other aspects, the present invention provides methods using a combination of data coupled with logic. The invention can also provide knowledge extraction in other embodiments. The invention provides object creation, and provides conversational access to the database in other embodiments. Accordingly, the present invention can provide an acquisition technique that actually provides answers to queries (rather than hits), which can be singular. Depending upon the embodiment, one or more of these advantages can be present. These and other advantages, however, are described throughout the present specification and more particularly below. These and other embodiments of the present invention are described in more detail in conjunction with the text below and attached Figs.
BRIEF DESCRIPTION OF THE DRAWINGS Fig. 1 is a simplified diagram of an information acquisition method according to an embodiment of the present invention; and
Fig. 2 is a simplified diagram of an information acquisition method according to an alternative embodiment of the present invention
DESCRIPTION OF THE SPECIFIC EMBODIMENTS According to the present invention, a technique including a method for acquiring information is provided. In a specific embodiment, the present invention provides a method using a combination of syntactic and semantic information objects. In one or more aspects, the present invention provides a modular, object-oriented, and collaborative approach to semantic typing, interpreting, and extraction, of knowledge objects from text sources into databases. The invention provides a highly general object- oriented method for using lexically-based knowledge to identify and extract semantic objects in text and to represent them in a database. The invention offers savings in the time, effort, and costs for constructing, populating, and updating a wide variety of databases.
A method according the present invention is briefly outlined below: 1. Providing text information sources; 2. Tokenizing;
3. Performing part-of-speech tagging;
4. Stemming tagged items;
5. Interpreting including type composition and type induction, semantic-syntactic composition, and instantiation of qualia. 6. Translating the resulting expression into a relational model;
7. Inserting this into a relational database;
8. Performing other steps, as desired.
The above sequence of steps generally provides a method for natural language input into a relational database. The present sequence of steps using a combination of syntactic and semantic information using an object-oriented approach. Step 7 can use either a relational, object-oriented, or mixed relational/object database. If a relational database is used, the step 7 of translation into the relational model is required (step 6). Any relational database can be used in Step 7. As merely an example, the database can be made by a company called Oracle of Redwood City, California. Alternatively, other companies such as Informix, Sybase, and others also manufacture database designs that can incorporate the present invention. Similarly, any object-oriented or mixed relational/object database can be used. Details of the above steps are briefly described according to Fig. 1, for example.
Fig. 1 is a simplified diagram 100 of an information acquisition method according to an embodiment of the present invention. This diagram is merely an example and should not limit the scope of the claims herein. One of ordinary skill in the art would recognize other variations, modifications, and alternatives. The method 100 begins at start, step 101, which includes steps of providing text information (step 103), tokenizing (step 105), tagging (step 107), stemming (step 109), interpreting (step 111), and extracting objects (step 125). Further details of each of these steps are provided below. The method provides (step 103) information such as text information from a variety of sources, including all digital sources. The sources include, among others, newspapers, magazines, research, web sites, product information, internets, ,intranets, and spoken language inputs. The text source are generally in electronic form, which can be read, organized, and categorized by way of a computer. The text source can be in any suitable electronic form such as ASCII, HTML, XML, LaTeX, word processing applications, and presentation slides (e.g., Microsoft Power Point™). In the next step 105, the method tokenizes the text information. The text may be "tokenized," for example, split up into textual elements separated by a delimiter, such as a "white space" or "blank" character. Tokenization normalizes the input text into a form that is usable by subsequent steps" of the method. In a specific embodiment, the tokenizer separates punctuation (e.g., periods, apostrophes, quotes, etc.) from words. As merely an example, the following phrase:
"Thomas E. Wheeler, CTIA's President" is converted into:
" Thomas_E._Wheeler_,_CTIA_'s_President_" where the underscore character "_" is used to represent a blank space for readability purposes only. Other examples of tokenization are expansions of contractions:
he'd => he d I'll ==> I_'ll
Other examples are shown below:
.... business day. ==> ..._business_day_. .... reason to be cheerful: ==> ..._reason_to_be_cheerful_:
"They are in trouble." =:> "_They_are_in_trouble_. " where again the underscore character "_" is used to represent a blank space for readability purposes only.
It is, however, desirable that abbreviations and initials be preserved and not separately tokenized, i.e., the period is not separated from them. For example, Mr., M.D., Mrs., Esq., and so on. It is also important not to split names that have false punctuation in them; e.g., index.html, http://www.company.com. In a specific embodiment, the present invention provides a step of tagging (step 107) the tokenized text information. In a specific embodiment, a part-of-speech (herein "POS") tagger can be used. A goal of the part-of-speech tagger is to assign grammatical category labels to each tokenized element in the text produced from the tokenizer. For example, a tagger will convert the input below:
The_new_company_. into
The/DT_new/JJ_company/NN_./. where the underscore character "_" is again used to represent a blank space for readability purposes only and where the "/tag" labels, e.g., DT, /JJ, /NN, refer to a standard set of POS tags used in the computational linguistics community. An example of such a POS tagging system is described in Brill (Brill, "AEric, "A simple rule-based part-of-speech ".tagger". In Third Conference on Applied Natural Language Processing, pages 152-155, Trento, Italy, 1992, which is herein incorporated by reference). An example of such a POS tagging system is described in Brill (Brill, Eric, Third Conference on Applied Natural Language Processing, pages 152-155, Trento, Italy, 1992).
The present method performs a step 109 of stemming. Stemming is yet another stage in normalization for further processing. For example, all stems are orthographically lower case. In addition, for example, in the case of inflected categories, such as a plural noun or a past tense verb, the stem will be the dictionary look up form of the token (e.g., 'man' for 'men', 'run' for 'ran'). Stemming can use dictionary lookup in the case of known inflected words. If the particular token does not occur in the dictionary, then it can be passed on to a stripped down version of the Porter Stemmer (Porter, M.F., "An Algorithm for Suffix Stripping," Program 14 (3), July 1980, pp. 130- 137, which is herein incorporated by reference ), which strips off affixes in certain orthographic contexts.
As merely an example, which should not limit the scope of the present invention herein, an illustration of steps of tokenizing, tagging, and stemming a text taken from a newspaper corpus is shown below. Original Text: Clinton Picks General to Command NATO WASHINGTON (Reuter) - President Clinton has chosen U.S. Army Gen. Wesley Clark to become commander of all allied NATO forces and American troops in Europe, a senior Pentagon official said Monday. Clark , 52, speaks Russian and was a member of the American team that helped broker the 1995
Dayton peace accords on Bosnia . He is based in Panama as chief of U.S. forces in Latin America and would replace retiring U.S. Army Gen. George Joulwan as Supreme Allied Commander of NATO in Europe (SACEUR) based in Mons, Belgium. Examples of tokenizing, tagging and stemming of the above text:
Clinton/NNP Picks/NBZ General/ΝΝP to/TO CommanαVΝΝP ΝATO/ΝΝP WASHIΝGTOΝ/ΝΝP (/( Reuter/ΝΝP )/SYM -/: President/ ΝP Clinton/ΝΝP has NBZ chosen/NBΝ U.S./ΝΝP Army ΝΝP GenVΝΝP Wesley/ΝΝP Clark/ΝΝP to/TO become/NB commander ΝΝ of/IΝ all/DT allied/NBΝ ΝATO ΝΝP forces/ΝΝS and/CC American/JJ troops/ΝΝS in/TΝ Europe ΝΝP ,/, a/DT senior/JJ Pentagon/ΝΝP official ΝΝ saiαWBD Monday/ΝΝP ./. Clark/ΝΝP ,/, 52/CD ,/, speaks/NBZ Russian/ΝΝP and/CC was/VBD a DT member/ΝΝ ofIΝ the/DT American/JJ team/ΝΝ that WDT helped/NBD broker ΝΝ the/DT 1995/CD Dayton/ΝΝP peace/ΝΝ accords/ΝΝS on IΝ Bosnia/ΝΝP ./. He/PRP is/NBZ baseoYNBΝ in/IΝ Panama ΝΝP as/TΝ chief/ΝΝ ofIΝ U.S./ΝΝP forces/ΝΝS in/TΝ Latin/ΝΝP America/ΝΝP and/CC ould/MD replace/NB retiring/NBG U.S./ΝΝP Army ΝΝP Gen/ΝΝP ./. George ΝΝP Joulwan/ΝΝP as/TΝ Supreme ΝΝP Allied/ΝΝP Commander/ΝΝP of IΝ ΝATO ΝΝP in/TΝ Europe/ΝΝP (/(
SACEUR/ΝΝP )/SYM based/VBΝ in/TΝ Mons/ΝΝS ,/, Belgium ΝΝP ./.
Next, the method performs interpreting (step 111), using an interpreting module or the like. A specific embodiment of the interpreting, includes three sub- modules: accessing of a lexicon and type system 113, parsing 115, and identification of semantic types 117, e.g., qualia roles.
The first sub-module 113, accessing of a lexicon and type system 113, uses two knowledge bases. The first is a lexicon (resource A), indexed by stem within a particular part of speech. For example, 'base' as a noun and 'base' as a verb will have two separate entries. Each lexical entry contains a type property which is a name of a semantic type in the type system (resource B). Each lexical entry contains appropriate syntactic information. This information is combined to create the appropriate type of syntactic constituent (e.g., a noun for a stem with a noun tag), with the appropriate semantic representation.
Next, the parser sub-module 115 takes the output of the accessing of a lexicon and type system sub-module 113, and composes these into larger syntactic and semantic structures that make up the sentences of text in natural language. This sub- module 115 uses an engine embodying an all paths parser (Younger, D., "Recognition and Parsing of Context-free Languages in Time n3", Information and Control 10:189-208,
1967, and Graham, Harrison, and Ruzzo, "An Improved Context-free Recongizer", ACM Transactions on Programming Languages and Systems 2:415-462, 1980, both of which are herein incorporated by reference), that uses a chart to hold information about the syntactic constituents found. Unlike a chart parser implemented in a procedural language, however, the present parser does not contain a single "controlling" module that acts in isolation to build interpreted structures out of passive data elements. Rather, interpreting can be accomplished only by collaboration among active objects of different classes. As part of the Interpreting Step 111, syntactic- semantic composition is accomplished by means of grammar rules, 120, which specify both how syntactic elements are to be combined, and also how their semantic interpretations are to be composed. Moreover, grammar rules can contain constraints which specify the conditions under which a rule can apply. If the constraint fails, the rule is not even considered, which improves the performance of the interpreter, since useless search paths are not pursued.
An example of such a rule is
VP => VP NP {Transitive} [DirectObject]
This rule states that a VP (Verb Phrase) can be constructed from a VP followed by an NP (Noun Phrase). {Transitive} represents a constraint. This rule may only be fired if the VP is transitive: i.e., if the semantics of the VP allows for a direct object and this direct object position has not yet been filled. [DirectObject] represents a role, i.e., names of pieces of code that contain (1) constraints on the other dependent(s) of the rule; and (2) specify how the semantics of the dependents of the rule are combined to create the semantics of the dominating constituent (i.e. the left hand side). This is the semantic composition that takes place if the rule succeeds. In this case, the semantics of the NP is checked for compatibility with the semantic type specified for the direct object in the semantics of the VP. Moreover, if the type of the NP is compatible with the type of the direct object but is less specific, the type of the NP will be changed to the more specific type. It is in this way that the system acquires new knowledge, by using its existing knowledge bases to learn or further specify the meanings of words it has not previously encountered.
The semantics of the NP is then bound to the direct object position of the semantics of the existing VP, and this new semantic representation is made the semantics of the newly constructed VP.
In a specific embodiment, the present interpreter 111 uses a single Interpreter object. The interpreter 111 can manipulate a plurality of types of data structures such as those noted below:
1. The elements that represent completed syntactic elements (i.e., those that have been found);
2. The elements that represent incomplete syntactic elements (i.e. those that are in the process of being constructed); and
3. WH elements (i.e., words like 'who' and 'what' that appear 'dislocated' from the 'logical' syntactic positions in which they receive their semantic interpretation).
These types of syntactic elements are represented by objects of the class Edge. It is the interactions among the Edges that the Interpreter object "facilitates" in the construction of the larger syntactic structure: As active and inactive Edges are inserted into the chart representation that can be maintained by Interpreter, the Interpreter passes information about their presence to other Edges already in the chart that may be able to interact with them to create new Edges (i.e. to form new syntactic constituents). Similarly, the Interpreter passes information about Edges already in the chart (i.e. syntactic constituents already found) to the new Edge that are added. The Interpreter also passes information about the existence of Edges representing WH elements to active Edges that may make use of them.
Edges interact to create new Edge on the basis of two other classes of objects: GrammarRule objects, which represent information about how new constituents can be built out of existing constituents, and Constituent objects, which represent traditional grammatical elements such as nouns, verbs, and sentences.
In a specific embodiment, the interpreting process can be defined as follows: the Interpreter is sent the output of the Stemmer, which is an Array of underspecified objects of type Constituent. Each of these objects includes the following pieces of information:
a), token: the unit of the input string (i.e. orthographic word or punctuation) found by the Tokenizer; b). tag: the part of speech tag assigned by the Tagger; c). stem: the dictionary lookup form of the token added by the Stemmer; d). offset: the numerical position of the current token in the input text computed by the Stemmer.
The Interpreter object processes each of the underspecified Constituent objects in its input in sequence and tells it to transform itself into a fully specified object of the appropriate syntactic category. For example, a Constituent with token 'men' and tag 'NNS' can be transformed into an object of class Noun with the features (instance variables) 'proper = false' and 'number = plural'; a constituent with token 'sang' and tag 'VBD' is transformed into an object of class Verb with the feature 'tense = past'; etc.
As each input Constituent is processed, the Interpreter causes an associated Edge object to be created, which stores its paired constituent in an instance variable. In addition, the Interpreter consults the GrammarRule class object (step 120) to find out what grammar rules involve the new constituent. For each such grammar rule found, the Interpreter causes a new Edge to be built, which stores information about the rule that sanctioned its creation and the constituent that corresponds to the portion of the rule already found in the input. To illustrate this process, let us look at what happens when the Interpreter object receives the output of the Stemmer for the input text 'green apples', given the existence of a GrammarRule object that constructs a NounPhrase out of an Adjective and a Noun. The Stemmer's output has two Constituents, the first with the token 'green' and tag 'JJ'; the second with the token 'apples' and tag *NNS'. The Interpreter object processes these constituents in left to right order in a specific embodiment.
First it tells the Constituent for 'apple' to transform itself into a completely specified syntactic object: this produces an object of class Adjective with the feature 'degree = positive'. The Interpreter then causes an (inactive) Edge associated with this constituent to be built. It then checks to see if there are any existing Edges to the left of the Adjective that can use it, but since this is the first element in the sentence, there are none. Consulting with the GrammarRule class object, together they find the rule NounPhrase => Adjective Noun, which can make use of the Adjective object just created. The Interpreter, therefore, causes a new Edge to be built, which records that it has found the newly created Adjective object, and that it is trying to form a NounPhrase on the basis of the rule just given. The Interpreter then looks for any existing Edges (to the right of the Adjective), that the newly created Edge can interact with, but at this point there are none.
The Interpreter then tells the next Constituent object ('apples') to transform itself: this produces an object of class Noun with the features 'proper = false' and 'number = plural'. As before, the Interpreter causes a new associated Edge object to be built. It also looks for any (active) Edges immediately preceding 'apples'. It finds one, the active Edge built off of the Adjective object 'green', which is looking for an immediately following Noun. The Interpreter passes the new Edge to this existing Edge, which checks to see if it can use the newly formed Edge. It can, and therefore adds the new Noun to its collection of found constituent and marks itself as complete, i.e. as having found a NounPhrase. The Interpreter will also, as in the case of the preceding Adjective object, look for any new rules that can use the newly found Noun object and, if any such rules are found, will cause the corresponding Edges to be built, and will look for any Edges to the right of these new Edges that they can consume. These activities are irrelevant in the current example, however, so we omit the details. In a specific embodiment, the present method using selected grammar rules (step 120) for a interpreter (step 111). The rules are noted below and are defined according to the following representations. The elements between curly brackets {} are constraints, i.e. names of pieces of code that the left edge of a rule (the left-most element on the right hand side) should satisfy before the rule will be considered. The elements between square brackets [] are roles, e.g., [DirectObject];
Utterance rules (8):
Utterance => RootS Utterance => Interj ection
Utterance => NP
Utterance => VP {Imperative} [ImperativeHead]
Utterance -> RootS EndPunct [SemanticHead]
Utterance => Interjection EndPunct Utterance => NP EndPunct [NPUtterance]
Utterance => VP EndPunct {Imperative} [ImperativeHead]
RootS rules (8 ):
RootS => WhPhrase RootS [WhQuestion] RootS => NP VP [Subject]
RootS => V RootS {PossibleAuxiliary} [Subject Auxlnversion]
RootS => V NEG RootS {PossibleAuxiliary} [Subject Auxlnversion]
[Negation]
RootS => V NP AdjP {Copula} [InvertedSubject] [InvertedPredicateAdjective]
RootS => Modal RootS [Subject Auxlnversion]
RootS => Modal NEG RootS [Subject Auxlnversion] [Negation]
RootS => AdvP RootS [AdverbialModifier] Complements rules (3 V.
Complements => WhPhrase Complements [WhQuestion] Complements => NP VP [Subject] Complements => COMP Complements [Complementizer]
RelS rules (TV.
RelS => WhRelPhrase Complements [RelativePronoun]
ComplementVP rules (3): ComplementVP => WhPhrase ComplementVP [WhQuestion]
ComplementVP => COMP ComplementVP {InfinitivalComp} [Complementizer] ComplementVP => TO VP [Infinitive]
WhPhrase rules (3):
WhPhrase => NP {WhElement} WhPhrase => AdjP {WhElement} WhPhrase => PP {WhElement}
WhRelPhrase rules (2):
WhRelPhrase => NP {RelElement} WhRelPhrase => PP {RelElement}
VBar rules (6): VBar => V
VBar => V VBar {PossibleAuxiliary} [Auxiliary]
VBar => V NEG VBar {PossibleAuxiliary} [Auxiliary] [Negation]
VBar => Modal VBar [Auxiliary]
VBar => Modal NEG VBar [Auxiliary] [Negation] VBar => AdvP VBar [AdverbialModifier] VP rules (TOV. VP => VBar
VP => VP NP {Copula} [PredicateNominal] VP => VP AdjP {Copula} [Predicate Adjective] VP => VP PP {Copula} [PredicatePP]
VP => VP NP {Transitive} [DirectObject]
VP => VP AdjP {TakesAdjectiveComplement} [AdjectiveComplement] VP => VP PP {TakesPPComplement} [PPComplement] VP => VP Complements {TakesClause} [ClausalComplement] VP => VP NP {TakesClause} [QuestionOnClausalComplement]
VP => VP AdvP [AdverbialModifier]
NBar rules (6):
NBar => N NBar => N NBar {PossibleNounModifier} [NounModifier]
NBar => N NBar {PossiblePreName} [PreNameModifier]
NBar => N Conj N {ProperNoun} [AmpersandConjuntion]
[ProperNameWithAmpersand]
NBar => V NBar {PossibleVerbalModifier} [VerbalModifier] NBar => AdjBar NBar [AdjectiveModifier]
CoreNP rules (5): CoreNP => NBar
CoreNP => NBar NBar {TitleNoun} [TitleModifier] CoreNP => NBar Identifier {PossiblePreName} [IdentifierModifier]
CoreNP => Title NBar [TitleModifier] CoreNP => DeterminerGroup NBar [NPSpec]
DeterminerGroup rules (4): DeterminerGroup => Num
DeterminerGroup => Determiner DeterminerGroup => NP POS [PossessiveHead] DeterminerGroup => PreDet Determiner [Predeterminer]
NP rules (6):
NP => Pronoun
NP => CoreNP
NP => NP PP {TakesPPComplement} [PPComplement]
NP => NP Complements {TakesClause} [ClausalComplement]
NP => NP NPAppositive [AppositiveModifier]
NP => NP NumAppositive [NumAppositiveModifier]
NPAppositive rules 0):
NPAppositive => Punctuation NP Punctuation {AppositivePunctuation}
[AppositiveSemantics] [ClosingAppositivePunctuation]
NumAppositive rules (IV.
NumAppositive => Punctuation Num Punctuation
{AppositivePunctuation}
[AppositiveSemantics] [ClosingAppositivePunctuation]
AdiBar rules (3V.
AdjBar => Adj
AdjBar => NBar AdjBar [NounModifierTo Adjective]
AdjBar => NBar Punctuation AdjBar
[NounLocationModifierToAdjective] [LocationPunctuation]
AdiP rules (2):
AdjP => AdjBar
AdjP => AdjP PP {TakesPPComplement} [PPComplement]
AdvP rules ("4V. AdvP => Adv AdvP => DayOfWeek AdvP => Prep DayOfWeek AdvP => ReportingAdvP Pronoun
PP rules (IV
PP => Prep NP [PPObject]
DatePhrase rules (IV DatePhrase => MonthDay
MonthDav rules (4V MonthDay => Num MONTH MonthDay => MONTH Num MonthDay => MONTH ORD MonthDay => ORD MONTH
ReportingAdvP rules (2V
ReportingAdvP => V {SayingVerb}
ReportingAdvP => V AdvP {SayingVerb}
In specific embodiment, the present invention uses a combination of syntactic and semantic composition. As noted, the previous portions of the specification described how the interactions of objects of the classes Interpreter, Edge, GrammarRule, and Constituent produce syntactic interpret structures. This description was simplified in one respect, however: GrammarRules, and the Edges that are associated with them, do not merely check to determine whether the constituent associated with a candidate Edge matches the syntactic category specified in a rule: they also check for the eligibility of the candidate in terms of more fine-grained syntactic and/or semantic information. Moreover, Edges create not only new syntactic constituents from the information contained in grammar rules, they also compose the semantics of the syntactic dependents of that constituent, to form a new semantic object that is the associated meaning representation of the newly constructed constituent. The finer grained syntactic and semantic well-formedness conditions can be expressed in the form of Roles,e.g., [DirectObject], which are pieces of code associated with GrammarRule objects. For example, the rule that constructs a Sentence out of a NounPhrase and a VerbPhrase, has an associated Role named Subject that allows the Sentence to be constructed only if:
(1) the semantics associated with the VerbPhrase has not yet filled its Subject argument position; and
(2) the semantics associated with the candidate NounPhrase is compatible with the semantic type requirement imposed on the Subject argument by the VerbPhrase semantics (e.g. the semantics of the verb 'sell' requires that its Subject be either a Person or an Organization).
If these conditions are met, the semantics of the subject NounPhrase is bound to the
Subject position of the semantics of the VerbPhrase, and the new semantic object formed by binding this argument position is passed to the Sentence that is created as its semantics.
Roles can be represented in rules, for example, as simple symbolic names. These names are often associated with the actual code that is used to perform the type checking and semantic composition by lookup in a table. This allows the code for the same named Role to be used in multiple rules. For example, the Subject role will appear in main clause, complement clauses, relative clauses, declarative sentences, questions, etc.
Semantic representations can be constructed during the course of interpreting. The semantic representations associated with phrases and clauses are created by the interpreting process, by means of the composition of the semantics of dependent constituents; e.g., a sentence gets its semantics by the composition of the semantics of its subject NounPhrase and main predicate VerbPhrase; a transitive VerbPhrase gets its semantics from the composition of the semantics of its head Verb and its NounPhrase direct object, etc. PreTerminals (i.e. the constituents corresponding to the actual words in the sentence, such as Noun, Verb, Adjective) get their semantics either by lexical lookup or by default. When a new, fully specified PreTerminal object, such as a Noun, is formed from an underspecified Constituent, the first thing it does is to consult with a LexicalEntry class object of the appropriate type (e.g. NounEntry for Noun, VerbEntry for Verb, etc.) to determine if its stem has an associated lexical entry in that category. If it does, the PreTerminal object uses the semantic information in the LexicalEntry to create its associated semantic representation object. If it does not, the PreTerminal creates a default semantic representation appropriate for its syntactic category; e.g. a Verb creates a default semantic representation whose semantic type is Event, the least specific type of activity; a Noun creates a default semantic representation whose semantic type is TopType, the least specific semantic type for persons, places, things, or concepts, etc. In a specific embodiment, the Interpreting module (step 111) also uses a sub-module for identification of semantic types associated with words and phrases (step 117). The present invention uses this sub-module 117 that contains the core conceptual knowledge for the system. The sub-module 117 provides a structure for a semantic type system 113 (resource B), which underlies the processes of lexical inference. The type system 113 is structured along multiple dimensions, where each dimension corresponds to a different aspect of word meaning. As a result each dimension involves a different way of understanding a given entity in the domain and thus corresponds to a different set of questions (i.e. queries) concerning that entity.
These different aspects of word meaning are expressed by means of qualia structure, namely "modes of understanding" of an entity. This is described in J.
Pustejovsky, "The Generative Lexicon", MIT Press, 1995, which is herein incorporated by reference. A structured conceptual type involving qualia roles may be defined relative to the following four qualia roles:
formal: the kind of entity constitutive: the mode of individuation of that entity telic: the purpose or function of the entity agentive: how the entity comes into being
Qualia roles can provide building blocks for structuring a concept, such that the types in our type system differ in terms of their internal complexity. Thus, concepts are not organized in terms of a taxonomic ISA link uniquely. EachRather, each conceptual type in the type system is a data structure that incorporates the set of inferences that are available as well as the relations between that type and other entities.
In a specific embodiment, the present invention provides types as identifiers for other types. Here, types are not merely used for structuring information for instances of each particular type, but they also play a crucial role in identifying other types in the text. In other words, the structure of the typing information drives part of the knowledge acquisition process. For example, the information that is associated with a noun denoting a professional role, i.e. "doctor", permits identification of the type of the associated institution, namely a hospital or a health clinic. Similarly, given the semantics of certain head nouns, it is possible to type their modifiers in compound nominal constructions. For example the noun ""maker"" becomes a fairly reliable identifier of entities that are typed as "products", as in "software maker".
The present invention also provides a qualia structure as a basis for acquiring knowledge. The qualia structure of a lexical item can be identified in terms of patterns occurring in the text. Each qualia represents a well defined component of meaning which is talked about in texts as well as in people's conversations. For instance, an object which occurs with the predicate "manufacture" — as its direct object — is understood as an artifactual entity which is brought about through a manufacturing process. This information is exploited to acquire information concerning the AGENTIVE role of a given entity. This strategy is extensible at different levels of specificity, in a way that allows the engine to acquire information concerning the relationship between Microsoft and Windows 98. Similarly, there are grammatical constructs such as "used for", "used in", "good for" which provide reliable identifiers for the TELIC of an entity. There are also patterns that indicate the CONSTITUTIVE information: for companies, expressions involving headquarters, location, and address, all provide the basis for specifying the constitutive aspect; for products, constructions such as "made of or "made from" indicate the specific components of an entity.
Patterns that identify information concerning the TELIC of an entity, can be useful for simultaneously acquiring the same information about another entity from a different perspective. For example, while we are assigning Microsoft as the entity that specifies the AGENTIVE role of one of its products, we are also building that information as part of Microsoft's TELIC. Qualia Structure can also be used as a basis for querying and reasoning over a database. Given the close relation between syntactic patterns and the semantics they convey relative to a given entity, the querying exploits the same principles used in the knowledge acquisition. The process, however, is taken further. Once we have available the information concerning the AGENTIVE role of a set of different software products, for instance, then it is possible to ask about competitors. To achieve this it is a matter of finding the list of companies that appear in the AGENTIVE role of products with the same type. Similarly, it is possible to query a set of products or entities that are fulfilling a given function that which is specified in the TELIC role. According to one example the entire process of populating a database from natural language input in text can be provided, as shown below: Input document = 0000077400.txt (a Reuters text)
Input sentence (appears at offset position 380 in text):
The S&P500 stock index rose 36.46 points.
Tokenized and tagged: The/DT S&P500/NNP stock NN index/NN rose/VBD 36.46/CD points/NNS ./.
Examples of the resources accessed from the lexicon and type system sub- module 113 are shown below:
Lexicon system (Resource AV
[VerbEntry stem: 'rise' type: 'financial rise activity'; subjectRole: #theme; objectRole: #measure
]
[NounEntry stem: 'point' type: 'Measure'
]
Type system (Resource BV [GLEventType name: 'financial rise activity' formal: #([[rise activity]]) argumentStructure: theme: [[Abstract Object]] externalArgument: [[Measure]] ]
[GLType //comment: Qualia Role name: 'Measure' formal: #([[Abstract Object]]) ]
The resulting semantic root node for this text is as follows: :
[UtteranceLexLF type: [[Opinion]] illocutionaryForce: #Assertion content: [FunctionLexLF type: [[rise activity]] predicateStem: 'rise' complements: (#Subject -> [EntityLexLF type: [[Abstract Object.Company]] value: 'S&P500 stock index' quantification: [QuantifierLexLF type: [[Abstract Object]] value: 'The']]
#DirectObject -> [EntityLexLF type: [[Measure]] value: 'points' quantification: [CountLexLF type: [[Number]] value: 36.46]])]]
Next, a specific embodiment of the present method uses a step of extracting objects, step 125. The present method creates a semantic representation of objects, during interpreting, including syntactic-semantic composition, and qualia induction, that serve as an interface to a relational database model. Since these semantic representation objects are a combination of data and procedures, similar to objects in an object-oriented systems, they implement their own procedures (i.e., methods) to translate themselves from their object representation into SQL statements that map to a relational database. While there are several classes of semantic representation objects (LexLF) that are used during the course of interpreting and semantic composition, there are two classes that are relevant to interactions with the relational database: EntityLexLF — which represents the semantics of entities (i.e. person, places, things, concepts)
FunctionLexLF — which represents the semantics of relations among entities
Both of these classes implement methods for two modes of database interaction: insertion and retrieval.
In insertion mode, the LexLF transforms its semantics into a SQL INSERT statement to add information to the database. Since a LexLF contains various pieces of information that must be present in the database, the LexLF negotiates with the database to discover whether supporting pieces of information exist already or not. For example, FunctionLexLF contains information about the predicate type of the associated predicate and about the various entities that fill its argument positions. Each of these elements must exist in the appropriate table of the database; when one of these elements already exists, the LexLF merely looks up the primary key for it; when it does not, the LexLF must first cause the element to be inserted in the database, to preserve relational integrity.
The entities are inserted, then the relation is built on them.
The EntityLexLF for 'S&P500 stock index' produces the following SQL statements:
Insert the entity: insert into Entities(CanonicalName) values('S&P500 stock index')
Then insert the type information: (autonumbering in database at time of insertion gives 'S&P500 stock index1 an EntityID of 5230) insert into Types(EntityLD, DocumentID, Offset, Type) values(5230,405,380,'Abstract Object.Company')
The EntityLexLF for '36.46 points' produces the following: insert into Entities(CanonicalName) values('point') insert into Types(EntityLD, DocumentID, Offset, Type) values(5231,405,380,'Measure')
The FunctionLexLF for the whole utterance produces the following:
Insert the predicate: insert into Predicates(PredicateName, PredicateType ) values('rise-rise activity','rise activity')
Then the relation proper:
(the Entityld for unfilled arguments is '0' by default) insert into Relations(PredicateLD,DocumentLD,Offset,Subject, Object,
ClausalArgjExtraArgl, ExtraArg2,ExtraArg3) values(23,405,380,5230,5231,0,0,0,0) Insert the cardinality information from the Direct Object update Relations set ObjectCardinality = '36.46' where Relation!!) = 776
Once the objects have been extracted, step 125, they can be mapped onto a relational database, which is illustrated by a simplified method diagram of Fig. 2. This diagram is merely an example and should not limit the scope of the claims herein. One of ordinary skill in the art would recognize other variations, modifications, and alternatives.
The present method 200 begins at step 201, which can derive from step 125 in the above
Fig. The method maps (step 203) objects in a relational manner, which can be defined by an entity relation diagram, for example. The objects are inserted (step 205) into an objected oriented relational database 211. Once the objects have been inserted they are now ready for use by a user.
In a specific embodiment, the user can query (step 207) the database 211.
The present process stops at step 209, but is not limited. As merely an example, which uses the above database population technique, it would be possible to find an answer to the following question:
What did the S&P stock index do? As in the previous example, this utterance would go through the stages of at least tagging and tokenization:
What/WP did/VBD the/DT S&P500/NNP stock/NN index/NN do/VB ?/.
and would produce a semantic representation of the following form:
[UtteranceLexLF type: [[Question]] illocutionaryForce: #WhQuestion content: [FunctionLexLF type: [[QueryDo]] predicateStem: 'do' complements: (#Subject -> [EntityLexLF type: [[Abstract Object]] value: 'S&P500 stock index' quantification: [QuantifierLexLF type: [[Abstract Object]] value: 'The']] #DirectObject -> [EntityLexLF type: [[Entity]] value: 'What' quantification: [QuantifierLexLF type: [[Entity]] value: 'what' quantifier: #Wh]])]]
There are several features of this semantic form. First, the semantics of the interrogative pronoun 'What' is interpreted in its 'logical' position, i.e. as the direct object of the main verb 'do'. Second, the semantic representation of 'What' includes a
QuantifierLexLF that has #Wh as the value of its #quantifier. This indicates that this is the logical argument that is being asked about in this query. Semantic representations for content queries of this type are processed for database lookup in the following manner.
First, the Entity LD of the subject is retrieved:
select EntityLD from Entities where CanonicalName = 'S&P500 stock index1
This will retrieve the EntityLD 5230, which is then used to construct a select statement on the Relations table:
select * from Relations where Subject = 5230
This will retrieve the row:
(776,23,405,380,5230,null,5231,'36.46',0,0,null,0,null,0,null,0)
Finally, for presentation to the user, the system will use this information to retrieve the sentence:
The S&P500 stock index rose 36.46 points.
i.e. the sentence at offset position 380, in the document with DocumentID 405, whose filename is '0000077400'. In the current implementation, this information is passed to the user interface in the format:
<DISPLAY-FULL-OBJECT "" { "Reuters"
"http://199.103.231.59/demo- code/source.pl display=0000077400,380#380"
"The S&P500 stock index rose 36.46 points." } { } > which contains the source of the response text, a URL that points to the complete source document (the location of the host machine is not hard-coded, but is determined dynamically to point to the host machine that is running the system), and the actual response text. This is presented in a more user-friendly format: only the name of the source and the response text are displayed, with the source name being a hyperlink that points to the full source text, to allow the user to examine the entire text, if desired.
Note that the database lookup is done in two stages: an initial lookup of the EntityLD of the subject and then a lookup from the Relations table, using this EntityLD, rather than as a single select statement that does a database join over these two tables. The reason for the decomposition of the lookup process in this manner is to allow for a more responsive interaction with the end user. Ifa simple join query fails to retrieve any data, the source of the failure is unknown. It could be due to, for example, one of several reasons:
1. one or more of the entities asked about are unknown to the system;
2. the entities are known, and appear as arguments to one or more relations, but not as arguments to the relation the user specifically asked about;
3. the types of the entities are known, but there is no relation that connects them;
4. etc.
By decomposing the lookup process into separate stages, the system can explore the possibility of presenting the user with a response that is more informative than a simple 'No Answer'. It can provide the user with an exact answer if one is available, and with the closest appropriate answer where an exact match is impossible. 'No Answer' responses are only given when the system cannot find anything that is relevant to the user's question. Using the above example in retrieval (e.g., query) mode, the LexLF transforms its semantics into a SQL SELECT statement. This procedure is the mirror image of insertion in many ways: elements from multiple tables should be queried to produce the desired answer. Moreover, to provide robustness in the system's replies to the user, the LexLF may engage in iterative interactions with the database, as it falls back and widens the parameters of its retrieval space to find information relevant to the user's request, but which may not literally match the query. For example, if the user asks 'Did Brand X buy The Highpriced Spread?', the database may not contain information about the buy relation between Brand X and The Highpriced Spread, but it may contain other relations involving them. The LexLF will engage in a series of transactions with the database to find something at least loosely matching the user's search parameters before giving up and returning no information. There are two sorts of transactions possible with the database, when there is no direct answer to the query: a. If there are other relations, in which the object being queried about appears, they will be presented, when related to the original query relation. b. If there are more general relations or object type descriptions than those present in the query, these will be returned.
Although the above functionality has generally been described in terms of specific hardware and software, it would be recognized that the invention has a much broader range of applicability. For example, the software functionality can be further combined or even separated. Similarly, the hardware functionality can be further combined, or even separated. The software functionality can be implemented in terms of hardware or a combination of hardware and software. Similarly, the hardware functionality can be implemented in software or a combination of hardware and software. Any number of different combinations can occur depending upon the application.
Many modifications and variations of the present invention are possible in light of the above teachings. Therefore, it is to be understood that within the scope of the appended claims, the invention may be practiced otherwise than as specifically described.

Claims

WHAT IS CLAIMED IS:
1. A natural language database forming method, said method comprising: providing text information comprising a plurality of related words; tagging each word in said text information; forming an object comprising syntactic information and semantic information from each word in said text information; and placing into a object oriented relational database said object.
2. The method of claim 1 wherein said tag can be selected from a verb, a noun, an adjective, an adverb, a numeral, a conjunction, a determiner, and a preposition.
3. The method of claim 1 wherein said text information is derived from publications, e-mail, newspapers, news feeds, and wires.
4. The method of claim 1 wherein said relational object oriented database is a mixed objected oriented database.
5. The method of claim 1 further comprising: forming a query; and selecting an object based upon an entity relationships to achieve a unique output.
6. The method of claim 5 wherein said output comprises text information.
7. The method of claim 5 wherein said unique output comprises an answer.
8. The method of claim 6 wherein said text information supports the answer.
9. The method of claim 5 wherein is said text information comprises a plurality of headings.
10. The method of claim 1 wherein said text information is provided in electronic form.
11 . A natural language knowledge acquisition method, said method comprising: providing text information comprising a plurality of related words, said text information being in electronic form; tagging each word in said text information; forming an object comprising syntactic information and semantic information from each word in said text information; placing into a object oriented relational database said object; repeating said providing, tagging, forming, and placing to populate said relational object oriented database; forming a query; and selecting an object based upon an entity relationships to achieve a unique output.
12. The method of claim 11 wherein said tag can be selected from a . verb, a noun, an adjective, an adverb, a numeral, a connection, a determiner, and a preposition.
13. The method of claim 11 wherein said text information is derived from publications, e-mail, newspapers, news feeds, and wires.
14. The method of claim 11 wherein said relational object oriented database is a mixed objected oriented database.
15. The method of claim 11 wherein said output comprises text information.
16. The method of claim 11 wherein said unique output comprises an answer.
17. The method of claim 11 wherein said text information supports the answer.
18. The method of claim 11 wherein is said text information comprises a plurality of headings.
PCT/US1999/028226 1998-11-30 1999-11-29 A natural knowledge acquisition method WO2000033216A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
AU19263/00A AU1926300A (en) 1998-11-30 1999-11-29 A natural knowledge acquisition method
EP99962917A EP1151401A4 (en) 1998-11-30 1999-11-29 A natural knowledge acquisition method

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US11019098P 1998-11-30 1998-11-30
US60/110,190 1998-11-30
US16334599P 1999-11-03 1999-11-03
US43363099A 1999-11-03 1999-11-03
US60/163,345 1999-11-03
US09/433,630 1999-11-03

Publications (1)

Publication Number Publication Date
WO2000033216A1 true WO2000033216A1 (en) 2000-06-08

Family

ID=27380808

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1999/028226 WO2000033216A1 (en) 1998-11-30 1999-11-29 A natural knowledge acquisition method

Country Status (3)

Country Link
EP (1) EP1151401A4 (en)
AU (1) AU1926300A (en)
WO (1) WO2000033216A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002010960A2 (en) * 2000-07-31 2002-02-07 Eliyon Technologies Corporation Computer method and apparatus for extracting data from web pages

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11017172B2 (en) 2018-08-01 2021-05-25 Signals Analytics Ltd. Proposition identification in natural language and usage thereof for search and retrieval

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4688195A (en) * 1983-01-28 1987-08-18 Texas Instruments Incorporated Natural-language interface generating system
US4829423A (en) * 1983-01-28 1989-05-09 Texas Instruments Incorporated Menu-based natural language understanding system
US5555367A (en) * 1994-09-30 1996-09-10 General Electric Company Method and system for generating computer programs for queries formed by manipulating object-oriented diagrams
US5584024A (en) * 1994-03-24 1996-12-10 Software Ag Interactive database query system and method for prohibiting the selection of semantically incorrect query parameters
US5909678A (en) * 1996-09-13 1999-06-01 International Business Machines Corporation Computer systems, method and program for constructing statements by dragging and dropping iconic representations of subcomponent statements onto a phrase template

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
NL8900587A (en) * 1989-03-10 1990-10-01 Bso Buro Voor Systeemontwikkel METHOD FOR DETERMINING THE SEMANTIC RELATION OF LEXICAL COMPONENTS IN A TEXT
US5794050A (en) * 1995-01-04 1998-08-11 Intelligent Text Processing, Inc. Natural language understanding system
US5963940A (en) * 1995-08-16 1999-10-05 Syracuse University Natural language information retrieval system and method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4688195A (en) * 1983-01-28 1987-08-18 Texas Instruments Incorporated Natural-language interface generating system
US4829423A (en) * 1983-01-28 1989-05-09 Texas Instruments Incorporated Menu-based natural language understanding system
US5584024A (en) * 1994-03-24 1996-12-10 Software Ag Interactive database query system and method for prohibiting the selection of semantically incorrect query parameters
US5555367A (en) * 1994-09-30 1996-09-10 General Electric Company Method and system for generating computer programs for queries formed by manipulating object-oriented diagrams
US5909678A (en) * 1996-09-13 1999-06-01 International Business Machines Corporation Computer systems, method and program for constructing statements by dragging and dropping iconic representations of subcomponent statements onto a phrase template

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP1151401A4 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002010960A2 (en) * 2000-07-31 2002-02-07 Eliyon Technologies Corporation Computer method and apparatus for extracting data from web pages
WO2002010960A3 (en) * 2000-07-31 2003-07-17 Eliyon Technologies Corp Computer method and apparatus for extracting data from web pages
US6618717B1 (en) 2000-07-31 2003-09-09 Eliyon Technologies Corporation Computer method and apparatus for determining content owner of a website
US6778986B1 (en) 2000-07-31 2004-08-17 Eliyon Technologies Corporation Computer method and apparatus for determining site type of a web site
US6983282B2 (en) 2000-07-31 2006-01-03 Zoom Information, Inc. Computer method and apparatus for collecting people and organization information from Web sites
US7065483B2 (en) 2000-07-31 2006-06-20 Zoom Information, Inc. Computer method and apparatus for extracting data from web pages
US7356761B2 (en) 2000-07-31 2008-04-08 Zoom Information, Inc. Computer method and apparatus for determining content types of web pages

Also Published As

Publication number Publication date
EP1151401A1 (en) 2001-11-07
EP1151401A4 (en) 2002-03-06
AU1926300A (en) 2000-06-19

Similar Documents

Publication Publication Date Title
Plaza et al. A semantic graph-based approach to biomedical summarisation
Moens Automatic indexing and abstracting of document texts
US8265925B2 (en) Method and apparatus for textual exploration discovery
US8977953B1 (en) Customizing information by combining pair of annotations from at least two different documents
US7587389B2 (en) Question answering system, data search method, and computer program
US6957213B1 (en) Method of utilizing implicit references to answer a query
US6584470B2 (en) Multi-layered semiotic mechanism for answering natural language questions using document retrieval combined with information extraction
US6076088A (en) Information extraction system and method using concept relation concept (CRC) triples
US6732098B1 (en) Relational text index creation and searching
US6738765B1 (en) Relational text index creation and searching
US6732097B1 (en) Relational text index creation and searching
US6728707B1 (en) Relational text index creation and searching
US7171349B1 (en) Relational text index creation and searching
US20120303358A1 (en) Semantic textual analysis
US20070179776A1 (en) Linguistic user interface
JPH1125129A (en) Method for retrieving information stored in text data base
US20020046019A1 (en) Method and system for acquiring and maintaining natural language information
Dittenbach et al. A natural language query interface for tourism information
Torres-Parejo et al. MTCIR: A multi-term tag cloud information retrieval system
Kangavari et al. Information retrieval: Improving question answering systems by query reformulation and answer validation
Litkowski Question Answering Using XML-Tagged Documents.
Litkowski Summarization experiments in DUC 2004
Vilares et al. Extraction of complex index terms in non-English IR: A shallow parsing based approach
EP1151401A1 (en) A natural knowledge acquisition method
Mihalcea et al. Automatic Acquisition of Sense Tagged Corpora.

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 1999962917

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWP Wipo information: published in national office

Ref document number: 1999962917

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1999962917

Country of ref document: EP