WO2001098941A1 - Method of sentence analysis - Google Patents

Method of sentence analysis Download PDF

Info

Publication number
WO2001098941A1
WO2001098941A1 PCT/AU2001/000731 AU0100731W WO0198941A1 WO 2001098941 A1 WO2001098941 A1 WO 2001098941A1 AU 0100731 W AU0100731 W AU 0100731W WO 0198941 A1 WO0198941 A1 WO 0198941A1
Authority
WO
WIPO (PCT)
Prior art keywords
buffer
relational
syntactic
sentence
retrieval
Prior art date
Application number
PCT/AU2001/000731
Other languages
French (fr)
Inventor
Simon Dennis
Original Assignee
The University Of Queensland
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by The University Of Queensland filed Critical The University Of Queensland
Priority to AU2001265700A priority Critical patent/AU2001265700A1/en
Publication of WO2001098941A1 publication Critical patent/WO2001098941A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates

Definitions

  • the invention relates to the general field of computer-based information retrieval. In particular it relates to the extraction of information from databases of textual material. Specifically, the invention provides a method of sentence analysis to extract relational information, and of accessing the relational information to answer questions.
  • Sentence interpretation may then be construed as a memory retrieval problem involving three critical memory systems: syntactic; relational; and lexical.
  • a syntactic trace is the set of syntagmatic associations within a sentence.
  • a relational trace is the set of paradigmatic associations activated by a sentence.
  • a lexical trace is the set of paradigmatic associates of a word across sentences.
  • the sentence to be interpreted can be used as a probe to a long-term memory containing sentences to which the system has been exposed and the retrieved information can form constraints on working-memory resolution.
  • the representation and extraction of information from structured databases is a well-understood problem with well-developed and highly successful solutions. However, as general information retrieval mechanisms, structured databases have a number of inherent limitations.
  • database methodologies both for constructing information stores and for querying these stores requires substantial expertise, limiting the accessibility of these techniques to expert users.
  • database methodologies typically require the information technology professional to create a predetermined organization for the information domain of interest. Altering this organization once the store has been constructed can be difficult.
  • the organization chosen when establishing the store can place limitations on the ways in which that information can be accessed, leading to inflexibility as retrieval needs evolve.
  • Open text requires only - language skills to construct, which allows for more general participation in the development of information bases.
  • open text is difficult to query.
  • Current information retrieval methods typically require the user to supply a set of keywords that are used to specify the general information domain. Documents related to that domain are returned and it is up to the user to trawl through these documents looking for the specific information they require. This can be a time consuming and error fraught process.
  • This approach is used almost universally by common search engines to extract information from databases accessible by the World Wide Web and much of the known prior art is primarily suitable for supporting this style of retrieval operation.
  • the invention resides in a method of sentence analysis including the steps of: segmenting the sentence to be analysed; initialising a number of buffers for storing word vectors; lexical retrieval; syntactic retrieval and resolution; and relational retrieval and resolution.
  • the step of segmenting the sentence involves segmenting the sentence into a sequence of individual words.
  • the sentence is segmented into a sequence of tokens which include root words, prefixes or suffixes, and punctuation.
  • the step of initializing buffers includes initializing an input buffer and a processing buffer with the sequence of individual words or tokens representing the sentence. Initially each word is stored as the representation of itself. A syntactic buffer and a relational buffer are also established.
  • the step of lexical retrieval includes retrieving a vector of paradigmatic associations for each word from a lexical memory storage means and replacing each word in the processing buffer with the corresponding vector.
  • syntactic retrieval includes the steps of constructing a syntactic probe in the syntactic buffer from the contents of the processing buffer and using the probe to retrieve syntactic traces from a syntactic memory means, to update the syntactic probe.
  • the updated probe is a weighted average of the retrieved syntactic traces.
  • Syntactic resolution includes the step of modifying the contents of the processing buffer to reflect the constraints in the syntactic buffer. An optimization technique is suitably applied.
  • Suitable optimization techniques include gradient descent, conjugate gradient, simplex, and alignment.
  • relational retrieval includes the steps of constructing a relational probe in the relational buffer from the contents of the processing buffer and the input buffer. The probe is used to retrieve relational traces from a relational memory means to update the relational probe. The updated probe is a weighted average of the retrieved relational traces.
  • Relational resolution includes the step of modifying the contents of the processing buffer to reflect the constraints in the relational buffer. An optimization technique is suitably applied.
  • Suitable optimization techniques include gradient descent, conjugate gradient and simplex.
  • the invention resides in a method of retrieving the answer to a natural language question, the method including the steps of: a) segmenting the question to be answered; b) initializing an input buffer and a processing buffer with the segmented question; c) performing lexical retrieval; d) performing syntactic retrieval and resolution; and e) performing relational retrieval and resolution; said steps c) to e) being performed one or more times until the processing buffer contains the answer to the question.
  • steps c) to e) may be performed in any order and as many times as necessary to analyze the sentence, answer the question, or retrieve the desired text.
  • the invention resides in a method of establishing an ensemble of facts from a text document including the steps of: a) organising the text document into a collection of sentences; b) segmenting each sentence; c) initialising a number of buffers for storing word vectors, including an input buffer, a processing buffer, a syntactic buffer and a relational buffer; d) lexical retrieval; e) syntactic retrieval and resolution; f) relational retrieval and resolution; g) g) performing steps b) to f) until the relational buffer contains a fact in the form of a matrix; h) repeating step g) for all sentences in the document to compile the ensemble of facts.
  • the method may further include the step of performing steps a) to h) for a first document to produce a first ensemble of facts, performing steps a) to h) for a second document to produce a second ensemble of facts, and comparing the first and second ensemble of facts to establish a similarity measure between the first and second document.
  • the ensemble of facts may be compared by determining a dot product for each fact of the first document with each fact of the second document, and calculating an average of the maximum dot products.
  • a cosine of the angle between each fact of the first document with each fact of the second document may be calculated, and an average of the maximum cosines determined.
  • FIG 1 shows the architecture of a sentence analysis engine
  • FIG 2 shows a flowchart of a document comparison process.
  • FIG 1 is shown the architecture of a sentence analysis engine 1 for working the method.
  • the engine consists of syntactic long-term memory system 2 coding syntactic traces (between slot, per sentence), relational long-term memory system 3 coding relational traces (within slot, per sentence) and lexical long-term memory system 4 coding lexical traces (within slot, across sentences).
  • An input sequence buffer 5 contains the sentence to be processed and a processing sequence buffer 6 contains the current interpretation of each of the words in the input buffer.
  • Each slot in the processing buffer is a vector each component of which represents a possible token (word, punctuation symbol or affix) and will contain the set of possible paradigmatic associates of the input word.
  • the results of retrieval from the syntactic long term memory are stored in a syntactic buffer 7.
  • a similar relational buffer 8 stores the result of retrieval from the relational long term memory.
  • Sentence analysis follows the same basic process for building a knowledge base of sentence structures or for using the knowledge base to answer questions. Processing of a sentence involves the following steps: 1. Text Preprocessing: Firstly, the text representing the sentence or question must be segmented. Words and punctuation marks are separated into tokens. Performance of the model can be improved by also segmenting the affixes from each root word (e.g. ing, ly, un). Note that unlike other information retrieval techniques the affixes are retained as tokens because they aid syntactic interpretation. 2.
  • Initialize Input and Processing Buffers The sequence of tokens derived during step one is copied into the input and processing buffers, so that the initial interpretation of each word is the word itself.
  • Lexical Retrieval The vector of paradigmatic associates of each of the input words is retrieved from Lexical Memory and added to the corresponding processing buffer slot.
  • Syntactic Probe Construction Using the current contents of the processing buffer a syntactic probe is formed (see below for details) and placed in the Syntactic Buffer.
  • Syntactic Retrieval The probe in the Syntactic Buffer is used to retrieve syntactic traces from syntactic memory. The probe is updated from the retrieved traces.
  • Syntactic Resolution An optimization technique is used to alter the processing buffer representations to reflect the constraints stored in the Syntactic Buffer.
  • Relational Probe Construction Using the current contents of the input and processing buffers a relational probe is formed (see below for details) and placed in the Relational Buffer.
  • Relational Retrieval The probe in the Relational Buffer is used to retrieve relational traces from relational memory. The probe is updated from the retrieved traces.
  • Relational Resolution Again, an optimization technique is used to alter the processing buffer representations to reflect the constraints stored in the Relational Buffer. Having outlined the main process it remains to elucidate how each of the operations such as memory retrieval, lexical, syntactic, and relational trace construction and syntactic and relational resolution occur.
  • the memory retrieval mechanism is based on the Minerva II model of episodic memory [Hintzman, D.L. (1984); Minerva 2: A Simulation Model of Human Memory; Behavior Research Methods, Instruments and Computers 6 (2) 96-101].
  • Minerva II memories are stored as vectors. Retrieval involves comparing a retrieval probe, also represented as a vector, against each of the traces in memory using a dot product operation.
  • a new probe is constructed by multiplying each trace by its similarity raised to a power ( ⁇ ) and summing:
  • is an odd number (usually 3 or 5) so that the multiplier retains the sign of the dot product.
  • the retrieved vector thus formed can then be used to initiate further retrieval. To avoid probes that grow or shrink rapidly the trace multipliers are normalized.
  • a syntactic trace is the set of syntagmatic associates within a sentence.
  • Each sentence is represented by a word-by-word matrix.
  • a 1 in a given row and column indicates that the row word is followed by the column word.
  • Such a representation has important advantages in a domain such as sentence processing where embedded structure is the norm rather than the exception. For instance, if memory contains the sentence:
  • the girl is loved by John and Mary who was sick is loved by John despite the fact that the component words do not align. Using the syntagmatic representation of the sentence ensures that there will be overlap. "The girl is loved by John” is represented by:
  • the input buffer 5 contains the sentence to be processed as an array of locally encoded vectors. For instance, the sentence "Mary is loved by John” is coded as:
  • the syntactic buffer 7 contains a set of constraints on the representations in the processing buffer 6. To apply these constraints we need to convert the order based sentence matrix into a sequence-based buffer. To do this, the method applies an optimization technique.
  • a suitable optimization technique is the gradient descent procedure.
  • An alternative approach is an alignment algorithm described in "Speech discrimination by dynamic programming" by Vinysyuk, T. K. in Cynemetics 4(1): 52-57 (1968). The following example uses the gradient descent procedure.
  • Sij is the matrix of syntagmatic associations retrieved from memory then we need to minimize the cost function:
  • C represents the balance between the degree to which the cost function is minimised and the processing time. A smaller C requires longer processing time whereas a larger C represents less accuracy.
  • Relational traces constructed in this way can then be stored in relational memory to be later retrieved from partial relational cues.
  • b ⁇ is the processing buffer vector in the z th location.
  • the lexical trace for word W* is defined as:
  • b t is the state of the buffer following syntactic resolution and lexical retrieval involves adding the trace of the corresponding input word to each processing buffer location:
  • lexical, syntactic and relational memory systems are constructed and the sentence-processing algorithm is applied with the question as input. Additional slots beyond the end of the question are included in the processing buffer to hold the answer.
  • the lexical long-term memory system 4 and syntactic long-term memory system 2 contain general linguistic knowledge and hence can be pre- trained on a large corpus (containing some answered questions). The specific information required to answer the question will be contained in the relational long-term memory system 3. Consequently, it must be trained on the text expected to contain the answer.
  • the current invention differs from prior art question-answer systems in that it produces a representation of the deep structure of a sentence but does not require an elaborate set of parsing or semantic rules. Instead, it extracts the lexical, syntactic and relational information required to interpret a passage automatically from large corpora of electronic text. As a consequence, it is able to incorporate a more complete knowledge base than is possible when system designers must code these components. Furthermore, the automated nature of knowledge acquisition makes it feasible to track and respond to changes in language use.
  • the inventors consider that the primary application of the method is in information retrieval, particularly extraction of information from a database in response to a natural language question. However, other applications of the method can include document filtering, document categorisation and document comparison. One particular application of document comparison is essay marking.
  • FIG 2 A master document is divided into individual sentences and each sentence is analysed to create a relational representation in the form of a matrix.
  • the ensemble of matrices is a representation of the facts that are expressed in the master document.
  • the same process is applied to a target document to produce a second ensemble of facts.
  • the ensembles are compared to determine a similarity measure that indicates the degree to which the target document contains the same facts as the master document.
  • the first step is to divide the document into a collection of sentences. This may be done, for example, by detecting punctuation. Other algorithms known in the field could also be used depending upon the specific situation.
  • Each sentence is segmented into words or tokens and analysed in the manner described in detail above.
  • a document can be divided into a collection of sentences before further analysis, or preferably, each sentence is analysed progressively before the next sentence is identified. For each sentence the process produces a matrix that defines a fact. If there are further sentences the process iterates until the ensemble of matrices (facts) is collated. Once the process has completed a comparison is made between the ensemble of matrices for the target document and the ensemble of matrices of the master document. The outcome of the comparison is a similarity measure that provides an indication of the degree of overlap of facts in the target document compared to the master document.
  • each fact in the master document is compared against each fact in the target document using a similarity metric such as the dot product (other alternatives include the cosine of the angle between the matrices and the Euclidean metric).
  • the fact from the target document that has the maximal dot product with each fact from the master document is the one most likely to contain the same information.
  • This dot product is then a measure of the likelihood that the target document contains the same fact as the master document. By taking the average of these maximal dot products an overall measure of the amount of factual content from the master document that appears in the target document is produced.
  • the similarity measure is the average of the maximum cosines.
  • Other similarity measures such as Minkowski distance (of which the Euclidean distance is one case) may also be used.

Abstract

To analyse a sentence, the sentence is first segmented into a sequence of tokens and each token is placed in a processing buffer. The sentence is then subjected to lexical retrieval whereby each token is replaced by a vector of paradigmatic associations for each word from a lexical long term memory. The segmented sentence is also copied into a syntactic buffer and into a relational buffer. A syntactic probe is constructed and used to retrieve syntactic traces from a syntactic long term memory and to iteratively up-date the syntactic probe. Similarly a relational probe is constructed and used to retrieve relational traces from a relational long term memory.

Description

METHOD OF SENTENCE ANALYSIS
The invention relates to the general field of computer-based information retrieval. In particular it relates to the extraction of information from databases of textual material. Specifically, the invention provides a method of sentence analysis to extract relational information, and of accessing the relational information to answer questions.
BACKGROUND TO THE INVENTION
It has been hypothesized that humans learn language by acquiring the syntagmatic (within sentence) and paradigmatic (between sentence) associations among words. For instance, the fact that the word "green" is often followed by the word "grass" will lead to the formation of a syntagmatic association between "green" and "grass". Similarly, the fact that the word "green" often fills the same slot in a sentence as the word "red" (e.g. the balloon was green, the balloon was red) will lead to the formation of a paradigmatic association between "green" and "red".
Sentence interpretation may then be construed as a memory retrieval problem involving three critical memory systems: syntactic; relational; and lexical. A syntactic trace is the set of syntagmatic associations within a sentence. A relational trace is the set of paradigmatic associations activated by a sentence. A lexical trace is the set of paradigmatic associates of a word across sentences. The sentence to be interpreted can be used as a probe to a long-term memory containing sentences to which the system has been exposed and the retrieved information can form constraints on working-memory resolution. The representation and extraction of information from structured databases is a well-understood problem with well-developed and highly successful solutions. However, as general information retrieval mechanisms, structured databases have a number of inherent limitations. The application of database techniques both for constructing information stores and for querying these stores requires substantial expertise, limiting the accessibility of these techniques to expert users. In addition, database methodologies typically require the information technology professional to create a predetermined organization for the information domain of interest. Altering this organization once the store has been constructed can be difficult. In addition, the organization chosen when establishing the store can place limitations on the ways in which that information can be accessed, leading to inflexibility as retrieval needs evolve.
The majority of electronically stored information, particularly that available on line, comes in the form of open text. Open text requires only - language skills to construct, which allows for more general participation in the development of information bases. However, open text is difficult to query. Current information retrieval methods typically require the user to supply a set of keywords that are used to specify the general information domain. Documents related to that domain are returned and it is up to the user to trawl through these documents looking for the specific information they require. This can be a time consuming and error fraught process. This approach, however, is used almost universally by common search engines to extract information from databases accessible by the World Wide Web and much of the known prior art is primarily suitable for supporting this style of retrieval operation. For instance, reference may be had to United States patent number 5056021 in the name of Ausborn. The patent describes a method of abstracting concepts from natural language using a word by word categorization process. Many of these retrieval systems utilize co-occurrence information (syntagmatic information) to build semantic representations of words and documents. For instance, US patent number 4839853 assigned to Bell Communications Research Inc. describes a system for extracting latent semantic structure from corpora to support the representation and retrieval of textual objects. Similarly, United States patent number 5406480 and US patent number 5099425, both assigned to Matsushita Electric Industrial Co., Ltd., describe methods for extracting semantic representations from co-occurrence information. A solution to the difficulties outlined above is to allow users to supply their queries as natural language questions. The retrieval mechanism must then take each question and return just that information required to answer it. Current approaches to question-answer systems fall into one of two categories. Either they use symbolic techniques to parse each sentence rigorously extracting the content in accordance with syntactic and semantic grammars, and are consequently easily derailed by language examples that were not anticipated by the designers. Or they act like more general information retrieval engines using a set of heuristics to isolate in the document base a specific piece of text (word, phrase or sentence) that might contain the content of interest. In these systems, there is no attempt to code the deep structure of each sentence. United States patent number 5331554, assigned to Ricoh Corporation, describes the application of such a method to the question-answering task using a tree based pattern recognition approach.
Thus the known prior art either relies on a designer generated abstract grammar that makes it susceptible to unexpected input, or is limited to the extraction of simple information that is evident in the surface form of the text, thus missing relational structure that can be expressed in many surface forms. A more complete method is required that can construct lexical, syntactic and relational memories, which may then be used for extracting answers to natural language questions.
DISCLOSURE OF THE INVENTION
In one form, although it need not be the only or indeed the broadest form, the invention resides in a method of sentence analysis including the steps of: segmenting the sentence to be analysed; initialising a number of buffers for storing word vectors; lexical retrieval; syntactic retrieval and resolution; and relational retrieval and resolution.
The step of segmenting the sentence involves segmenting the sentence into a sequence of individual words. Suitably the sentence is segmented into a sequence of tokens which include root words, prefixes or suffixes, and punctuation. Suitably the step of initializing buffers includes initializing an input buffer and a processing buffer with the sequence of individual words or tokens representing the sentence. Initially each word is stored as the representation of itself. A syntactic buffer and a relational buffer are also established.
The step of lexical retrieval includes retrieving a vector of paradigmatic associations for each word from a lexical memory storage means and replacing each word in the processing buffer with the corresponding vector. Suitably, syntactic retrieval includes the steps of constructing a syntactic probe in the syntactic buffer from the contents of the processing buffer and using the probe to retrieve syntactic traces from a syntactic memory means, to update the syntactic probe. The updated probe is a weighted average of the retrieved syntactic traces. Syntactic resolution includes the step of modifying the contents of the processing buffer to reflect the constraints in the syntactic buffer. An optimization technique is suitably applied.
Suitable optimization techniques include gradient descent, conjugate gradient, simplex, and alignment. Suitably, relational retrieval includes the steps of constructing a relational probe in the relational buffer from the contents of the processing buffer and the input buffer. The probe is used to retrieve relational traces from a relational memory means to update the relational probe. The updated probe is a weighted average of the retrieved relational traces. Relational resolution includes the step of modifying the contents of the processing buffer to reflect the constraints in the relational buffer. An optimization technique is suitably applied.
Suitable optimization techniques include gradient descent, conjugate gradient and simplex. In a further form, the invention resides in a method of retrieving the answer to a natural language question, the method including the steps of: a) segmenting the question to be answered; b) initializing an input buffer and a processing buffer with the segmented question; c) performing lexical retrieval; d) performing syntactic retrieval and resolution; and e) performing relational retrieval and resolution; said steps c) to e) being performed one or more times until the processing buffer contains the answer to the question.
It will be appreciated that the steps c) to e) may be performed in any order and as many times as necessary to analyze the sentence, answer the question, or retrieve the desired text.
In a yet further form the invention resides in a method of establishing an ensemble of facts from a text document including the steps of: a) organising the text document into a collection of sentences; b) segmenting each sentence; c) initialising a number of buffers for storing word vectors, including an input buffer, a processing buffer, a syntactic buffer and a relational buffer; d) lexical retrieval; e) syntactic retrieval and resolution; f) relational retrieval and resolution; g) g) performing steps b) to f) until the relational buffer contains a fact in the form of a matrix; h) repeating step g) for all sentences in the document to compile the ensemble of facts. The method may further include the step of performing steps a) to h) for a first document to produce a first ensemble of facts, performing steps a) to h) for a second document to produce a second ensemble of facts, and comparing the first and second ensemble of facts to establish a similarity measure between the first and second document. The ensemble of facts may be compared by determining a dot product for each fact of the first document with each fact of the second document, and calculating an average of the maximum dot products. Alternatively, a cosine of the angle between each fact of the first document with each fact of the second document may be calculated, and an average of the maximum cosines determined.
BRIEF DETAILS OF THE DRAWINGS
To assist in understanding the invention, preferred embodiments will now be described with reference to the following figures in which:
FIG 1 shows the architecture of a sentence analysis engine; and
FIG 2 shows a flowchart of a document comparison process.
DETAILED DESCRIPTION OF THE DRAWINGS
In the drawings, like reference numerals refer to like parts. In FIG 1 is shown the architecture of a sentence analysis engine 1 for working the method. The engine consists of syntactic long-term memory system 2 coding syntactic traces (between slot, per sentence), relational long-term memory system 3 coding relational traces (within slot, per sentence) and lexical long-term memory system 4 coding lexical traces (within slot, across sentences).
An input sequence buffer 5 contains the sentence to be processed and a processing sequence buffer 6 contains the current interpretation of each of the words in the input buffer. Each slot in the processing buffer is a vector each component of which represents a possible token (word, punctuation symbol or affix) and will contain the set of possible paradigmatic associates of the input word.
The results of retrieval from the syntactic long term memory are stored in a syntactic buffer 7. A similar relational buffer 8 stores the result of retrieval from the relational long term memory. Sentence analysis follows the same basic process for building a knowledge base of sentence structures or for using the knowledge base to answer questions. Processing of a sentence involves the following steps: 1. Text Preprocessing: Firstly, the text representing the sentence or question must be segmented. Words and punctuation marks are separated into tokens. Performance of the model can be improved by also segmenting the affixes from each root word (e.g. ing, ly, un). Note that unlike other information retrieval techniques the affixes are retained as tokens because they aid syntactic interpretation. 2. Initialize Input and Processing Buffers: The sequence of tokens derived during step one is copied into the input and processing buffers, so that the initial interpretation of each word is the word itself.
3. Lexical Retrieval: The vector of paradigmatic associates of each of the input words is retrieved from Lexical Memory and added to the corresponding processing buffer slot.
4. Syntactic Probe Construction: Using the current contents of the processing buffer a syntactic probe is formed (see below for details) and placed in the Syntactic Buffer.
5. Syntactic Retrieval: The probe in the Syntactic Buffer is used to retrieve syntactic traces from syntactic memory. The probe is updated from the retrieved traces.
6. Syntactic Resolution: An optimization technique is used to alter the processing buffer representations to reflect the constraints stored in the Syntactic Buffer. 7. Relational Probe Construction: Using the current contents of the input and processing buffers a relational probe is formed (see below for details) and placed in the Relational Buffer.
8. Relational Retrieval: The probe in the Relational Buffer is used to retrieve relational traces from relational memory. The probe is updated from the retrieved traces.
9. Relational Resolution: Again, an optimization technique is used to alter the processing buffer representations to reflect the constraints stored in the Relational Buffer. Having outlined the main process it remains to elucidate how each of the operations such as memory retrieval, lexical, syntactic, and relational trace construction and syntactic and relational resolution occur.
The memory retrieval mechanism is based on the Minerva II model of episodic memory [Hintzman, D.L. (1984); Minerva 2: A Simulation Model of Human Memory; Behavior Research Methods, Instruments and Computers 6 (2) 96-101]. In Minerva II, memories are stored as vectors. Retrieval involves comparing a retrieval probe, also represented as a vector, against each of the traces in memory using a dot product operation.
st = p(t)J,
Where Sj is the similarity of the t'th trace, p is the probe vector and 7} is the tth trace.
A new probe is constructed by multiplying each trace by its similarity raised to a power (α) and summing:
Figure imgf000010_0001
Typically, α is an odd number (usually 3 or 5) so that the multiplier retains the sign of the dot product. The retrieved vector thus formed can then be used to initiate further retrieval. To avoid probes that grow or shrink rapidly the trace multipliers are normalized.
Figure imgf000010_0002
The method disclosed herein, assumes that a syntactic trace is the set of syntagmatic associates within a sentence. Each sentence is represented by a word-by-word matrix. A 1 in a given row and column indicates that the row word is followed by the column word.
For instance, the sentence "Mary is loved by John" would be coded with the matrix:
Figure imgf000011_0001
Such a representation has important advantages in a domain such as sentence processing where embedded structure is the norm rather than the exception. For instance, if memory contains the sentence:
Mary is loved by John Then this trace should be well matched by the probes:
The girl is loved by John and Mary who was sick is loved by John despite the fact that the component words do not align. Using the syntagmatic representation of the sentence ensures that there will be overlap. "The girl is loved by John" is represented by:
Figure imgf000011_0002
The underlined components overlap with the original sentence representation.
Similarly, "Mary who was sick is loved by John" is represented by:
Figure imgf000012_0003
Suppose that the input buffer 5 contains the sentence to be processed as an array of locally encoded vectors. For instance, the sentence "Mary is loved by John" is coded as:
Figure imgf000012_0004
Then the syntactic trace can be expressed as:
M forward -
Figure imgf000012_0001
where n is the number of words in the sentence. We could also code the backward associations:
A 4βdbrørf =
Figure imgf000012_0002
Note, however, that Mforward =Mb T ackward so there is no need to calculate this matrix separately.
After syntactic retrieval the syntactic buffer 7 contains a set of constraints on the representations in the processing buffer 6. To apply these constraints we need to convert the order based sentence matrix into a sequence-based buffer. To do this, the method applies an optimization technique. A suitable optimization technique is the gradient descent procedure. An alternative approach is an alignment algorithm described in "Speech discrimination by dynamic programming" by Vinysyuk, T. K. in Cynemetics 4(1): 52-57 (1968). The following example uses the gradient descent procedure.
If Sij is the matrix of syntagmatic associations retrieved from memory then we need to minimize the cost function:
where
Figure imgf000013_0001
and bj is the processing buffer vector in the i location. To update the buffer vectors the following equation is used:
t k = ε\ J (S-E) + ∑b,(S-E)7
1=1 k-\ l=k+\ n
where ε is a rate of change parameter and bk are constrained to be positive. Resolution continues until: max(Δb/t ) < C
where C is a parameter of the process.
C represents the balance between the degree to which the cost function is minimised and the processing time. A smaller C requires longer processing time whereas a larger C represents less accuracy.
By coding the order relations we now have a mechanism by which appropriate surface forms, independent of embedded structure, can be matched. To be able to answer questions, however, we need a method for extracting a representation of the relational content of the sentence that is independent of surface structure. The sentences:
Mary is loved by John
Ellen is loved by Bert will match quite well using the order relations (is-loved, loved-by, is-by) even though they code quite different relational content, while: Mary is loved by John John loves Mary do not match at all despite the fact that these sentences code the same relational content. In this case, what we would like is a way of associating fillers to their roles that is independent of the surface structure and that does not require us to specify in advance what the roles are. To see how this is achieved suppose the engine is trained on the following sentences:
Ann is loved by Bert
Josie is loved by Steve
Ellen is loved by Dave
Bert loves Ann Steve loves Josie
Dave loves Ellen
Who does Bert love? Ann
Who does Steve love? Josie
Who does Dave love? Ellen
In each of these constructions there is a lover and a lovee.
Furthermore, in all three constructions the same people fill the lover and lovee roles. Bert, Steve and Dave are the lovers and Ann, Josie and Ellen are the lovees. The role is represented as the distributed pattern of those words that fill the same surface slots in different constructions
(paradigmatic associates). That is, we can call the distributed pattern where Bert, Steve and Dave are active the lover role vector and the distributed pattern in which Ann, Josie and Ellen are active the lovee role vector. To interpret the sentence "Mary is loved by John" we would code the order relations (as above) and retrieve from syntactic long-term memory. We would then find that Ann, Josie and Ellen have filled the "Mary" slot in the similar surface structures previously, and Bert, Steve and Dave have filled the "John" slot. As a consequence, the following mappings would be formed:
{Ann, Josie, Ellen} -> Mary {Bert, Steve, Dave} -> John
Furthermore, the same would be true independent of the surface form. "John loves Mary" produces the same bindings. The method takes advantage of the fact that questions (including their answers) are simply an alternative syntactic form of the relational content in a corresponding sentence. To ask the question "Who does John love?", the words would be instantiated in the Input and Processing Buffers. Then syntactic retrieval and resolution would be executed. As a consequence, Bert, Steve and Dave would be activated in the "John" slot and Ann, Josie and Ellen would be activated in the slot following the question mark. The mapping {Bert, Steve, Dave} -> John can then be used to retrieve the trace above. To fulfill the relational constraints imposed by this trace requires that there be a {Ann, Josie, Ellen} -> Mary mapping generated by the Input and Processing Buffers. This constraint can be resolved by increasing the activation of Mary in the final slot. In the process the question is answered.
In the preceding paragraphs, the relational traces were described as a set of mappings. Within the engine these sets of mappings are expressed as a sum of outer products with the following equation:
R = ∑bi(ii -biγ ι=l I - hi is used to ensure that only the components that are not well predicted (i.e. the implicit variables in the utterance) are coded in the relational representation, thus making it insensitive to surface form. (Note an alternative form of this equation which has the same effect as
R = fjbiIi - diαg(fjbiIi)).
1=1 ι=l For example, suppose the sentences listed above are stored in syntactic memory and we wish to find the relational representation for "Mary is loved by John". The input buffer will contain:
Figure imgf000015_0001
If a syntactic probe is generated, used to retrieve from syntactic memory and then used to resolve the retrieved syntactic matrix in the processing buffer B we get:
Figure imgf000016_0003
where Bi is a distributed vector representing the lovee role and B5 is a distributed vector representing the lover role. Then R =
Figure imgf000016_0004
Relational traces constructed in this way can then be stored in relational memory to be later retrieved from partial relational cues.
Again, we use, for example, gradient descent with an objective function equal to the sum of squared differences between the current and retrieved relational representations to resolve the Processing buffer. If Ry is the matrix of paradigmatic associations retrieved from memory then we need to minimize:
Figure imgf000016_0001
where
Figure imgf000016_0002
and bι is the processing buffer vector in the zth location.
To update buffer vectors, then, we use:
Abk = ε bk (R - B)
Where £ is a rate of change parameter and bk are constrained to be positive. Again, resolution continues until: max(ΔbA) < C where C is a parameter of the process.
While the syntactic and relational memory mechanisms are sufficient to allow the extraction of relational representations of a sentence, a great many sentences would be required to ensure that each word had been seen within the many contexts in which it could appear. Performance can be enhanced by adding the distributed pattern of paradigmatic associates of a word to each slot prior to syntactic probe construction. Now, although a given word may not have appeared in a certain position within a sentence, a similar word contained in the input words lexical trace can precipitate syntactic retrieval.
The lexical trace for word W* is defined as:
Figure imgf000017_0001
a wntences
Figure imgf000017_0002
where bt is the state of the buffer following syntactic resolution and lexical retrieval involves adding the trace of the corresponding input word to each processing buffer location:
b, = b, + L, 3 l, = W,
To answer a question, lexical, syntactic and relational memory systems are constructed and the sentence-processing algorithm is applied with the question as input. Additional slots beyond the end of the question are included in the processing buffer to hold the answer. Note that the lexical long-term memory system 4 and syntactic long-term memory system 2 contain general linguistic knowledge and hence can be pre- trained on a large corpus (containing some answered questions). The specific information required to answer the question will be contained in the relational long-term memory system 3. Consequently, it must be trained on the text expected to contain the answer.
The current invention differs from prior art question-answer systems in that it produces a representation of the deep structure of a sentence but does not require an elaborate set of parsing or semantic rules. Instead, it extracts the lexical, syntactic and relational information required to interpret a passage automatically from large corpora of electronic text. As a consequence, it is able to incorporate a more complete knowledge base than is possible when system designers must code these components. Furthermore, the automated nature of knowledge acquisition makes it feasible to track and respond to changes in language use. The inventors consider that the primary application of the method is in information retrieval, particularly extraction of information from a database in response to a natural language question. However, other applications of the method can include document filtering, document categorisation and document comparison. One particular application of document comparison is essay marking. The use of the invention in this application is shown in FIG 2. A master document is divided into individual sentences and each sentence is analysed to create a relational representation in the form of a matrix. The ensemble of matrices is a representation of the facts that are expressed in the master document. The same process is applied to a target document to produce a second ensemble of facts. The ensembles are compared to determine a similarity measure that indicates the degree to which the target document contains the same facts as the master document. Referring to FIG 2, the first step is to divide the document into a collection of sentences. This may be done, for example, by detecting punctuation. Other algorithms known in the field could also be used depending upon the specific situation. Each sentence is segmented into words or tokens and analysed in the manner described in detail above. It will be appreciated that a document can be divided into a collection of sentences before further analysis, or preferably, each sentence is analysed progressively before the next sentence is identified. For each sentence the process produces a matrix that defines a fact. If there are further sentences the process iterates until the ensemble of matrices (facts) is collated. Once the process has completed a comparison is made between the ensemble of matrices for the target document and the ensemble of matrices of the master document. The outcome of the comparison is a similarity measure that provides an indication of the degree of overlap of facts in the target document compared to the master document.
To compare the documents, each fact in the master document is compared against each fact in the target document using a similarity metric such as the dot product (other alternatives include the cosine of the angle between the matrices and the Euclidean metric). The fact from the target document that has the maximal dot product with each fact from the master document is the one most likely to contain the same information. This dot product is then a measure of the likelihood that the target document contains the same fact as the master document. By taking the average of these maximal dot products an overall measure of the amount of factual content from the master document that appears in the target document is produced.
If the cosine of the angle between the matrices is used instead of the dot product, the similarity measure is the average of the maximum cosines. Other similarity measures, such as Minkowski distance (of which the Euclidean distance is one case) may also be used.
It will be appreciated that the master document must be processed following the flowchart of FIG 2 before a comparison can be made with a target document.
Throughout the specification the aim has been to describe the preferred embodiments of the invention without limiting the invention to any one embodiment or specific collection of features.

Claims

1. A method of sentence analysis including the steps of: segmenting the sentence to be analysed; initialising a number of buffers for storing word vectors; lexical retrieval; syntactic retrieval and resolution; and relational retrieval and resolution.
2. The method of claim 1 wherein the step of segmenting the sentence includes segmenting the sentence into a sequence of individual words or a sequence of tokens.
3. The method of claim 2 wherein the sentence is segmented into tokens which include root words, prefixes or suffixes, and punctuation.
4. The method of claim 1 wherein the step of initializing buffers includes initializing an input buffer and a processing buffer with the sequence representing the sentence.
5. The method of claim 4 wherein the step of initializing buffers further includes establishing a syntactic buffer and a relational buffer.
6. The method of claim 1 wherein the step of lexical retrieval includes the steps of: retrieving a vector of paradigmatic associations for each word from a lexical long-term memory; and replacing each word in the processing buffer with the corresponding vector.
7. The method of claim 1 wherein the step of syntactic retrieval includes the steps of: constructing a syntactic probe in a syntactic buffer from the contents of a processing buffer; and using the probe to retrieve syntactic traces from a syntactic long-term memory, to update the syntactic probe.
8. The method of claim 7 wherein the step of syntactic resolution includes the step of modifying the contents of the processing buffer to reflect the constraints in the syntactic buffer.
9. The method of claim 1 wherein the step of relational retrieval includes the steps of: constructing a relational probe in a relational buffer from the contents of a processing buffer and an input buffer; and using the relational probe to retrieve relational traces from a relational long term memory to update the relational probe.
10. The method of claim 9 wherein the step of relational resolution includes the step of modifying the contents of the processing buffer to reflect the constraints in the relational buffer.
11. A method of retrieving an answer to a natural language question, the method including the steps of: a) segmenting the question to be answered; b) initializing an input buffer and a processing buffer with the segmented question; c) performing lexical retrieval; d) performing syntactic retrieval and resolution; and e) performing relational retrieval and resolution; said steps c) to e) being performed one or more times until the processing buffer contains the answer to the question.
12. The method of claim 11 further including the step of establishing a syntactic buffer and a relational buffer.
13. The method of claim 11 wherein step c) includes the steps of: retrieving a vector of paradigmatic associations for each word from a lexical long-term memory; and replacing each word in the processing buffer with the corresponding vector.
14. The method of claim 11 wherein step d) includes the steps of: constructing a syntactic probe in a syntactic buffer from the contents of a processing buffer; using the probe to retrieve syntactic traces from a syntactic long-term memory, to update the syntactic probe; and modifying the contents of the processing buffer to reflect the constraints in the syntactic buffer.
15. The method of claim 11 wherein step e) includes the steps of: constructing a relational probe in a relational buffer from the contents of a processing buffer and an input buffer; using the relational probe to retrieve relational traces from a relational long term memory to update the relational probe; and modifying the contents of the processing buffer to reflect the constraints in the relational buffer.
16. A method of establishing an ensemble of facts from a text document including the steps of: a) organising the text document into a collection of sentences; b) segmenting each sentence; c) initialising a number of buffers for storing word matrices, including an input buffer, a processing buffer, a syntactic buffer and a relational buffer; d) lexical retrieval; e) syntactic retrieval and resolution; f) relational retrieval and resolution; g) performing steps b) to f) for each sentence until the relational buffer contains a fact in the form of a matrix; h) repeating step g) for all sentences in the document to compile the ensemble of facts.
17. The method of claim 16 wherein the input buffer and the processing buffer store word vectors.
18. The method of claim 16 further including the step of performing steps a) to h) for a first document to produce a first ensemble of facts, performing steps a) to h) for a second document to produce a second ensemble of facts, and comparing the first and second ensemble of facts to establish a similarity measure between the first and second document.
19. The method of claim 18 wherein the comparison is performed by determining a dot product for each fact of the first document with each fact of the second document, and calculating an average of the maximum dot products.
20. The method of claim 18 wherein the comparison is performed by determining a cosine of the angle between each fact of the first document with each fact of the second document, and calculating an average of the maximum cosines.
PCT/AU2001/000731 2000-06-20 2001-06-20 Method of sentence analysis WO2001098941A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2001265700A AU2001265700A1 (en) 2000-06-20 2001-06-20 Method of sentence analysis

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
AUPQ8253 2000-06-20
AUPQ8253A AUPQ825300A0 (en) 2000-06-20 2000-06-20 Method of sentence analysis

Publications (1)

Publication Number Publication Date
WO2001098941A1 true WO2001098941A1 (en) 2001-12-27

Family

ID=3822320

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/AU2001/000731 WO2001098941A1 (en) 2000-06-20 2001-06-20 Method of sentence analysis

Country Status (2)

Country Link
AU (1) AUPQ825300A0 (en)
WO (1) WO2001098941A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100057710A1 (en) * 2008-08-28 2010-03-04 Yahoo! Inc Generation of search result abstracts
CN111221976A (en) * 2019-11-14 2020-06-02 北京京航计算通讯研究所 Knowledge graph construction method based on bert algorithm model
CN112464654A (en) * 2020-11-27 2021-03-09 科技日报社 Keyword generation method and device, electronic equipment and computer readable medium
CN113378568A (en) * 2020-03-09 2021-09-10 株式会社理光 Relationship extraction method, device, system and computer readable storage medium
WO2023282781A1 (en) * 2021-07-06 2023-01-12 Публичное Акционерное Общество "Сбербанк России" Method and device for providing a service to a client by means of a virtual assistant

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10254880A (en) * 1997-03-10 1998-09-25 Atr Onsei Honyaku Tsushin Kenkyusho:Kk Language structure analyzer of example control type
US5841895A (en) * 1996-10-25 1998-11-24 Pricewaterhousecoopers, Llp Method for learning local syntactic relationships for use in example-based information-extraction-pattern learning
US5878385A (en) * 1996-09-16 1999-03-02 Ergo Linguistic Technologies Method and apparatus for universal parsing of language
WO2000011576A1 (en) * 1998-08-24 2000-03-02 Virtual Research Associates, Inc. Natural language sentence parser
EP1020803A1 (en) * 1997-03-04 2000-07-19 Ltd. Sintokogio Language analysis system and method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5878385A (en) * 1996-09-16 1999-03-02 Ergo Linguistic Technologies Method and apparatus for universal parsing of language
US5841895A (en) * 1996-10-25 1998-11-24 Pricewaterhousecoopers, Llp Method for learning local syntactic relationships for use in example-based information-extraction-pattern learning
EP1020803A1 (en) * 1997-03-04 2000-07-19 Ltd. Sintokogio Language analysis system and method
JPH10254880A (en) * 1997-03-10 1998-09-25 Atr Onsei Honyaku Tsushin Kenkyusho:Kk Language structure analyzer of example control type
WO2000011576A1 (en) * 1998-08-24 2000-03-02 Virtual Research Associates, Inc. Natural language sentence parser

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
PATENT ABSTRACTS OF JAPAN *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100057710A1 (en) * 2008-08-28 2010-03-04 Yahoo! Inc Generation of search result abstracts
US8984398B2 (en) * 2008-08-28 2015-03-17 Yahoo! Inc. Generation of search result abstracts
CN111221976A (en) * 2019-11-14 2020-06-02 北京京航计算通讯研究所 Knowledge graph construction method based on bert algorithm model
CN113378568A (en) * 2020-03-09 2021-09-10 株式会社理光 Relationship extraction method, device, system and computer readable storage medium
CN112464654A (en) * 2020-11-27 2021-03-09 科技日报社 Keyword generation method and device, electronic equipment and computer readable medium
WO2023282781A1 (en) * 2021-07-06 2023-01-12 Публичное Акционерное Общество "Сбербанк России" Method and device for providing a service to a client by means of a virtual assistant

Also Published As

Publication number Publication date
AUPQ825300A0 (en) 2000-07-13

Similar Documents

Publication Publication Date Title
Brill et al. An overview of empirical natural language processing
US7035789B2 (en) Supervised automatic text generation based on word classes for language modeling
US7286978B2 (en) Creating a language model for a language processing system
US7546235B2 (en) Unsupervised learning of paraphrase/translation alternations and selective application thereof
US20030144832A1 (en) Machine translation system
Dennis A memory‐based theory of verbal cognition
Schwartz et al. Language understanding using hidden understanding models
Simske et al. Functional Applications of Text Analytics Systems
Alkahtani Building and verifying parallel corpora between Arabic and English
WO2001098941A1 (en) Method of sentence analysis
Wu et al. A survey on statistical approaches to natural language processing
US20230140480A1 (en) Utterance generation apparatus, utterance generation method, and program
Lee Natural Language Processing: A Textbook with Python Implementation
Lane et al. Interactive word completion for morphologically complex languages
CN113011141A (en) Buddha note model training method, Buddha note generation method and related equipment
Dandapat Part-of-Speech tagging for Bengali
Chen et al. A POST parser-based learner model for template-based ICALL for Japanese-English writing skills
Wen Text mining using HMM and PMM
KR102496958B1 (en) Book story data base generating method for reading evaluation
Ekpenyong et al. Agent-based framework for intelligent natural language interface
Kamm Active Learning for acoustic speech recognition modeling
McMahon Statistical language processing based on self-organising word classification
Reeve Integrating hidden markov models into semantic web annotation platforms
Henrich et al. LISGrammarChecker: Language Independent Statistical Grammar Checking
Ahmed Detection of foreign words and names in written text

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP