CN106649259B - A method of learning dependence between extracting blocks of knowledge automatically from courseware text - Google Patents
A method of learning dependence between extracting blocks of knowledge automatically from courseware text Download PDFInfo
- Publication number
- CN106649259B CN106649259B CN201610874480.3A CN201610874480A CN106649259B CN 106649259 B CN106649259 B CN 106649259B CN 201610874480 A CN201610874480 A CN 201610874480A CN 106649259 B CN106649259 B CN 106649259B
- Authority
- CN
- China
- Prior art keywords
- knowledge
- blocks
- term
- formula
- dependence
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/284—Lexical analysis, e.g. tokenisation or collocates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/36—Creation of semantic tools, e.g. ontology or thesauri
- G06F16/374—Thesaurus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/12—Use of codes for handling textual entities
- G06F40/151—Transformation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/205—Parsing
- G06F40/216—Parsing using statistical methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/237—Lexical tools
- G06F40/247—Thesauruses; Synonyms
Abstract
The invention discloses it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, the text in courseware is corresponded to by handling blocks of knowledge, obtain candidate terms set, then the synonymous term in candidate terms set is handled, and each term is calculated to the criticality of blocks of knowledge, construct optimal model, by solving the study dependence extraction model optimized, courseware text can be automatically analyzed, it identifies the term in text and calculates term to the criticality of blocks of knowledge, and the model that study dependence is excavated is obtained by optimizing the relationship between term, locality of the process independent of study dependence, it can be used to excavate the study dependence that theme is associated between farther away blocks of knowledge, more complete knowledge navigation service is provided for learner.
Description
Technical field
The present invention relates to the methods of study dependence, and in particular to it is a kind of extract blocks of knowledge automatically from courseware text between
Learn the method for dependence.
Background technique
With the fast development of human sciences' technology, human knowledge total amount shows explosive growth.According to joint state religion
The statistics of section's text tissue, the nearly 30 years knowledge accumulated of the mankind account for the 90% of knowledge total amount since the dawn of human civilization, and the multiplication of knowledge
Period is still constantly shortening, and has reduced to 5-7 at present.The rapid growth of knowledge total amount is that effective acquisition of knowledge and expression are brought
Serious challenge.It is that user feedback goes out relevant documentation that traditional solution, which is by search engine,.This mode cannot be directly anti-
The interested knowledge of user is presented, the very big energy of user effort is needed to be screened from a large amount of relevant documentations.Knowledge mapping technology
Using RDF triple indicate semantic network, it is intended to realize search engine from " machinery is enumerated " to " network collection is known " develop, for
Family provides semantization, correlation information retrieval, alleviates the above problem to a certain extent.But knowledge mapping is not configured to face
Cognitive learning to theme can not embody the cognition relationship between each theme, and study is easy to cause to get lost problem.Knowledge Map foundation
The characteristics of human cognitive learns forms the efficient expression of one kind and knows by the relational organization between knowledge and knowledge at the form of figure
The mode of institutional framework between knowledge and knowledge provides effective method to alleviate the study problem of getting lost.
Study dependence describes the relationship to interdepend in cognitive process between blocks of knowledge.Determine two knowledge
Whether unit has relationship, is the Xiang Jiben in Knowledge Map building but very important work.Currently, the knowledge of high quality
Figure, it is still necessary to which domain expert marks the study dependence between blocks of knowledge according to domain knowledge, and building process is relatively more slow
Slowly.Therefore, the effective study dependence mining algorithm of design will greatly improve Knowledge Map building speed, reduce manpower and disappear
Consumption, helps to push the research and application of the navigation learning based on Knowledge Map.
For the method for learning dependence excavation between blocks of knowledge, Patent No. ZL201110312882.1, title
For a kind of blocks of knowledge incidence relation method for digging of text-oriented, the method for proposition includes the following steps: that (1) textual association is dug
Pick: clustering text collection, finds the text pair with similar topic, and the asymmetry being distributed using central term,
Excavate the linear correlation relationship between text;(2) candidate blocks of knowledge pair is generated: using the locality of blocks of knowledge incidence relation,
Generate candidate blocks of knowledge pair;(3) feature selecting and blocks of knowledge incidence relation excavate: the term word of knowledge based unit pair
Frequently, distance and semantic type feature, using SVM classifier by candidate blocks of knowledge to progress two-value classification, Extracting Knowledge list
Incidence relation between member.This method can greatly reduce candidate blocks of knowledge number, under the premise of guaranteeing precision, be effectively reduced
The time complexity of relation excavation.Since it makes use of the locality of study dependence, the above method is difficult to extract distance
Study dependence between farther away blocks of knowledge.
Summary of the invention
In order to solve the problems in the prior art, the present invention proposes to learn between one kind extracts blocks of knowledge from courseware text automatically
The method for practising dependence, can automatically analyze courseware text, identify the term in text and calculate term pair
The criticality of blocks of knowledge, and the model that study dependence is excavated, process are obtained by optimizing the relationship between term
Independent of the locality of study dependence, it can be used to excavate the study that theme is associated between farther away blocks of knowledge and rely on
Relationship provides more complete knowledge navigation service for learner.
In order to achieve the goal above, the technical scheme adopted by the invention is as follows: the following steps are included:
1) candidate terms based on mutual information generate: courseware document being converted into text formatting first, and is carried out at participle
Reason;Then the tightness degree that adjacent words combine is measured using mutual information, and processing is merged to compact vocabulary, from
And obtain candidate terms set;
2) synonymous term based on wikipedia about subtracts: crawling the corresponding wikipedia page of term, utilizes wikipedia
Redirection mark and multilingual link in the page, are about subtracted processing to synonymous term;
3) term criticality is measured: being calculated the corresponding TF-IDF parameter value of each term first, is then utilized knowledge list
First name feature and format character are weighted processing to TF-IDF parameter value, measure each term to the pass of blocks of knowledge with this
Stroke degree;
4) it optimal model building and solution: establishes between blocks of knowledge and learns to determine between dependence and term relationship
Amount indicates, converts optimization problem for model solution problem, constructs the objective function of optimization, and calculate using gradient decline
Method carries out model solution, and completion learns dependence between extracting blocks of knowledge automatically from courseware text.
The step 1) the following steps are included:
1.1) blocks of knowledge is corresponded to the text in courseware using poi kit to extract, and is segmented, removes and stop
Word processing;
1.2) after setting participle, original character string c is divided into two words a and b, by character string c being total in corpus
Existing frequency is denoted as f (c), co-occurrence probabilities of the character string c in corpus is denoted as p (c), then according to maximum likelihood estimate, in number
According to measure it is sufficiently large in the case where, it is believed that p (c) be equal to f (c), each word is considered as event, for character string c=ab, mutual information
Formula are as follows:The internal combustion tightness degree that character string is measured using mutual information, is obtained
Candidate terms set.
The step 2) the following steps are included:
2.1) thesaurus expands: based on thesaurus, utilizing redirection mark in the wikipedia page and more
Language link, expands thesaurus;
2.2) synonymous term about subtracts: using the thesaurus expanded by wikipedia, to same in candidate terms set
Adopted term is about subtracted processing.
About subtract the mode of processing in the step 2.2) are as follows: for containing the term A of synonymous lexical item, find and have with term A
There are identical meanings and the highest term B of the frequency of occurrences, replaces term A with term B in candidate terms set.
The step 3) the following steps are included:
3.1) to each of candidate terms set CT' term, it is calculated to each knowledge list by TF-IDF index
The basic criticality of member, TF-IDF index calculation formula are as follows:Formula
In: fijIndicate term i in document djIn word frequency;dfiIndicate the document word frequency of term i;N indicates total number of documents;niIndicate document
The middle number of files for term i occur;
3.2) weighting of knowledge based unit title: by investigating whether term appears in blocks of knowledge title to original
TF-IDF parameter be weighted, weighted formula are as follows: Namei,j=wname×bi,j, in formula: wnameIndicate that blocks of knowledge title adds
Weigh weight;bi,jIndicate whether term i appears in the title of blocks of knowledge j;
3.3) based on the weighting of format character: by the font size of term position, to the criticality of term into
Row weighting processing, weighted formula are as follows:In formula: wfontIndicate font size weighting
Weight;K indicates that blocks of knowledge j corresponds to all different fonts sizes in courseware;fi,kIndicate whether term i is gone out with font size k
It is existing;rankkAfter indicating all font size backward sequences, the ranking value of font size k;
3.4) aggregative weighted is carried out to original TF-IDF parameter by blocks of knowledge title and courseware font, obtains term
Criticality, the formula of weighting are as follows: scorei,j=wi,j×(1+Namei,j+Fonti,j), in formula: scorei,jIndicate i pairs of term
The criticality of blocks of knowledge j.
The step 4) the following steps are included:
4.1) objective function constructs: for blocks of knowledge i and blocks of knowledge j, measuring to exist between them by following formula and learn
A possibility that practising dependence:In formula: xiIt is the key that by all terms to blocks of knowledge i journey
The vector constituted is spent, each element represents corresponding term to the criticality of blocks of knowledge i in vector;A matrix representative model
Parameter;
To blocks of knowledge i, if set omegai=(i, j) | yij=1, j=1,2 ..., n } it is all to be deposited with blocks of knowledge i
In the blocks of knowledge of study dependence and the node pair of blocks of knowledge i composition, setFor it is all with blocks of knowledge i there is no the blocks of knowledge of study dependence with know
Know the node pair of unit i composition, enablesOptimization is defined as follows to ask
Topic:
In formula: X is a matrix, the i-th row in matrix
ByIt constitutes;(1-v)+Represent hinge loss function;||A||FRepresent this black norm of not Luo Beini of matrix A;
4.2) it model solution: to optimization problem, is solved using accelerating gradient decline:
It enablesThen former objective function is write as:Formula to A derivation,
Obtain gradient:
In formula:ei、ej、ekAll it is
Unit vector;
4.3) study dependence is excavated: the most optimized parameter A matrix of model is obtained by step 4.2), for any two
A blocks of knowledge is judged between them by optimal model with the presence or absence of study dependence.
Compared with prior art, the present invention corresponds to the text in courseware by handling blocks of knowledge, obtains candidate terms collection
It closes, then handles the synonymous term in candidate terms set, and calculate each term to the criticality of blocks of knowledge, construct
Optimal model can divide automatically courseware text by solving the study dependence extraction model optimized
Analysis identifies the term in text and calculates term to the criticality of blocks of knowledge, and passes through and optimize between term
Relationship obtains the model that study dependence is excavated, and locality of the process independent of study dependence can be used to dig
Pick theme is associated with the study dependence between farther away blocks of knowledge, provides more complete knowledge navigation clothes for learner
Business.The present invention can be corresponded to the text in courseware using blocks of knowledge and extract study dependence between blocks of knowledge automatically,
The cost for reducing artificial constructed Knowledge Map, helps to push the research of the navigation learning based on Knowledge Map and answers
With.
Detailed description of the invention
Fig. 1 is method flow frame diagram of the invention;
Fig. 2 is the graphical schematic diagram of formula (6);
Fig. 3 is that " Java language " learns dependence excavation partial data exemplary diagram.
Specific embodiment
Below with reference to specific embodiment and Figure of description the present invention will be further explained explanation.
Referring to Fig. 1, the present invention specifically includes the following steps:
1) candidate terms based on mutual information generate, and mainly include 2 steps:
1.1) blocks of knowledge is corresponded to the text in PPT courseware using poi kit to extract, and is segmented, is gone
Except stop words processing;
1.2) after setting participle, word string c originally is divided into two words a and b.By co-occurrence of the character string c in corpus
Frequency is denoted as f (c), and co-occurrence probabilities of the character string c in corpus are denoted as p (c), then, then according to maximum likelihood estimate,
In the case that data volume is sufficiently large, it is believed that p (c) is equal to f (c), if each word is considered as event, for character string c=
Ab, mutual information formula are as follows:
The internal combustion tightness degree of character string is measured using mutual information, to achieve the purpose that extract candidate terms.
Candidate terms generating algorithm based on mutual information is shown in specific step is as follows:
Input: word segmentation result set Tokens={ word1,word2,...,wordn, vocabulary number n in Tokens, word
Frequency statistical information TF, threshold value w
Algorithm flow:
Output: candidate terms set CT
2) synonymous term based on wikipedia about subtracts, and mainly includes 2 steps:
2.1) thesaurus expands: based on " Chinese thesaurus ", utilizing the redirection mark in the wikipedia page
With multilingual link, synonymicon is expanded, it is as follows that thesaurus expands specific algorithm:
Input: candidate terms set CT={ term1,term2,...,termn, vocabulary number n, thesaurus D=in CT
{(term1,...,termi), wherein (term1,...,termi) it is the i terms with identical meanings
Algorithm flow:
Output: by the thesaurus D expanded
2.2) synonymous term based on thesaurus about subtracts: right herein using the thesaurus expanded by wikipedia
Synonymous term in candidate terms set is uniformly processed, and the basic mode of processing is, for containing the art of synonymous lexical item
Language A, finding has identical meanings and the highest term B of the frequency of occurrences with term A, replaces in candidate terms concentration term B
Term A, respective algorithms are as follows:
Input: candidate terms set CT={ term1,term2,...,termn, vocabulary number n, thesaurus D=in CT
{(term1,...,termi), wherein (term1,...,termi) it is the i terms with identical meanings, word frequency statistics information
TF
Algorithm flow:
Output: CT'
3) term criticality is measured, and mainly includes 4 steps:
3.1) to each of candidate terms set CT' term, it is calculated to each knowledge list by TF-IDF index
The basic criticality of member, TF-IDF index calculation formula are as follows:
In formula: fijIndicate term i in document djIn word frequency;dfiIndicate the document word frequency of term i;N indicates that document is total
Number;niIndicate the number of files of term i occur in document;
3.2) weighting of knowledge based unit title: appearing in the term in blocks of knowledge title is likely to be the knowledge
The Key Term of unit, can be by investigating whether term appears in blocks of knowledge title to original TF-IDF parameter progress
Weighting, weighted formula are as follows:
Namei,j=wname×bi,j (3)
In formula: wnameIndicate that blocks of knowledge title weights weight;bi,jIndicate whether term i appears in the name of blocks of knowledge j
In title;
3.3) weighting based on PPT format character: in the classification of PPT is shown, level is higher, and font is generally bigger,
Expressed content is more important, therefore, can be weighted by the font size of term position to the criticality of term
Processing, weighted formula are as follows:
In formula: wfontIndicate that font size weights weight;It is big that k indicates that blocks of knowledge j corresponds to all different fonts in courseware
It is small;fi,kIndicate term i whether with font size k appearance;rankkAfter indicating all font size backward sequences, font size k
Ranking value;
3.4) formula of aggregative weighted is carried out to original TF-IDF parameter by blocks of knowledge title and courseware font are as follows:
scorei,j=wi,j×(1+Namei,j+Fonti,j) (5)
In formula: scorei ,jIndicate term i to the criticality of blocks of knowledge j;
4) optimal model building and solution, mainly include 3 steps:
4.1) objective function constructs: firstly, for blocks of knowledge i and blocks of knowledge j, being measured by following formula and is deposited between them
A possibility that learning dependence:
In formula: xiIt is the vector being made of criticality of all terms to blocks of knowledge i, each element generation in vector
Criticality of the table corresponding term to blocks of knowledge i;The parameter of A matrix representative model, as shown in Fig. 2, formula (6) can be regarded as
The sum of weight from paths all blocks of knowledge i to blocks of knowledge j;
To blocks of knowledge i, if set omegai=(i, j) | yij=1, j=1,2 ..., n }, i.e., the set be it is all with know
Know unit i and there is the blocks of knowledge of study dependence and the node pair of blocks of knowledge i composition, setI.e. the set is all knowledge that study dependence is not present with blocks of knowledge i
The node pair of unit and blocks of knowledge i composition, model parameter A reasonable for one, it is desirable to each pair of knowledge in set omega i
The score value of formula corresponding to unit will be greater than setIn each pair of blocks of knowledge score value, enableIt is defined as follows optimization problem:
In formula: X is a matrix, in matrix the i-th row byIt constitutes;(1-v)+Represent hinge loss function;||A||FGeneration
This black norm of the not Luo Beini of table matrix A;
4.2) for optimization problem shown in formula (7), accelerating gradient decline model solution: can be used
(AcceleratedGradient Descent) is solved:
It enablesThen former objective function can be write as:
Formula (8) obtains gradient to A derivation:
In formula:ei、ej、ekAll it is unit vector, circular is as follows:
Input: X, T, λ, η, maximum number of iterations N
Algorithm flow:
Output: A
4.3) study dependence is excavated: by step 4.2), the most optimized parameter A matrix of available model, at this point,
For any two blocks of knowledge, can judge to whether there is study dependence between them by the optimal model,
If Fig. 3 illustrates the few examples of " Java language " course learning dependence Result, solid line represents known learn in figure
Dependence is practised, dotted line represents the blocks of knowledge pair for needing to judge, the number on dotted line represents the blocks of knowledge to study
A possibility that dependence.
Claims (4)
1. it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, which is characterized in that including following
Step:
1) candidate terms based on mutual information generate: courseware document being converted into text formatting first, and carries out word segmentation processing;So
The tightness degree that adjacent words combine is measured using mutual information afterwards, and processing is merged to compact vocabulary, thus
To candidate terms set;
2) synonymous term based on wikipedia about subtracts: crawling the corresponding wikipedia page of term, utilizes the wikipedia page
In redirection mark and multilingual link, about subtracted processing to synonymous term;
3) term criticality is measured: being calculated the corresponding TF-IDF parameter value of each term first, is then utilized blocks of knowledge name
Claim feature and format character to be weighted processing to TF-IDF parameter value, each term is measured to the crucial journey of blocks of knowledge with this
Degree;
4) optimal model building and solution: the quantitative table learnt between dependence and term relationship is established between blocks of knowledge
Show, convert optimization problem for model solution problem, construct the objective function of optimization, and using gradient descent algorithm into
Row model solution, completion learn dependence between extracting blocks of knowledge automatically from courseware text;
The step 3) the following steps are included:
3.1) to each of candidate terms set CT' term, it is calculated to each blocks of knowledge by TF-IDF index
Basic criticality, TF-IDF index calculation formula are as follows:In formula:
fijIndicate term i in document djIn word frequency;dfiIndicate the document word frequency of term i;N indicates total number of documents;niIt indicates in document
There is the number of files of term i;
3.2) weighting of knowledge based unit title: by investigating whether term appears in blocks of knowledge title to original
TF-IDF parameter is weighted, weighted formula are as follows: Namei,j=wname×bi,j, in formula: wnameIndicate the weighting of blocks of knowledge title
Weight;bi,jIndicate whether term i appears in the title of blocks of knowledge j;
3.3) based on the weighting of format character: by the font size of term position, adding to the criticality of term
Power processing, weighted formula are as follows:In formula: wfontIndicate that font size weights weight;
K indicates that blocks of knowledge j corresponds to all different fonts sizes in courseware;fi,kIndicate term i whether with font size k appearance;
rankkAfter indicating all font size backward sequences, the ranking value of font size k;
3.4) aggregative weighted is carried out to original TF-IDF parameter by blocks of knowledge title and courseware font, obtains term key
Degree, the formula of weighting are as follows: scorei,j=wi,j×(1+Namei,j+Fonti,j), in formula: scorei,jIndicate term i to knowledge
The criticality of unit j;
The step 4) the following steps are included:
4.1) objective function constructs: for blocks of knowledge i and blocks of knowledge j, measured between them by following formula exist study according to
A possibility that relationship of relying:In formula: xiIt is the criticality structure by all terms to blocks of knowledge i
At vector, each element represents corresponding term to the criticality of blocks of knowledge i in vector;The ginseng of A matrix representative model
Number;
To blocks of knowledge i, if set omegai=(i, j) | yij=1, j=1,2 ..., n } it is all and blocks of knowledge i presence
Practise the blocks of knowledge of dependence and the node pair of blocks of knowledge i composition, setFor
The node pair of all blocks of knowledge and blocks of knowledge i composition that study dependence is not present with blocks of knowledge i, enablesIt is defined as follows optimization problem:
In formula: X is a matrix, in matrix the i-th row by
It constitutes;(1-v)+Represent hinge loss function;||A||FRepresent this black norm of not Luo Beini of matrix A;
4.2) it model solution: to optimization problem, is solved using accelerating gradient decline:
It enables,Then former objective function is write as:Formula obtains A derivation
To gradient:
In formula:ei、ej、ekIt is all unit
Vector;
4.3) study dependence is excavated: obtaining the most optimized parameter A matrix of model by step 4.2), any two are known
Know unit, is judged between them by optimal model with the presence or absence of study dependence.
2. it is according to claim 1 it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method,
It is characterized in that, the step 1) the following steps are included:
1.1) blocks of knowledge is corresponded to the text in courseware using poi kit to extract, and is segmented, removes stop words
Processing;
1.2) after setting participle, original character string c is divided into two words a and b, by co-occurrence frequency of the character string c in corpus
Rate is denoted as f (c), co-occurrence probabilities of the character string c in corpus is denoted as p (c), then according to maximum likelihood estimate, in data volume
In the case where sufficiently large, it is believed that p (c) is equal to f (c), each word is considered as event, for character string c=ab, mutual information formula
Are as follows:The internal combustion tightness degree that character string is measured using mutual information, obtains candidate
Term set.
3. it is according to claim 1 it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method,
It is characterized in that, the step 2) the following steps are included:
2.1) thesaurus expands: based on thesaurus, utilizing redirection mark in the wikipedia page and multilingual
Link, expands thesaurus;
2.2) synonymous term about subtracts: using the thesaurus expanded by wikipedia, to the synonymous art in candidate terms set
Language is about subtracted processing.
4. it is according to claim 3 it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method,
It is characterized in that, about subtracting the mode of processing in the step 2.2) are as follows: for containing the term A of synonymous lexical item, find and term A
With identical meanings and the highest term B of the frequency of occurrences, term A is replaced with term B in candidate terms set.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610874480.3A CN106649259B (en) | 2016-09-30 | 2016-09-30 | A method of learning dependence between extracting blocks of knowledge automatically from courseware text |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610874480.3A CN106649259B (en) | 2016-09-30 | 2016-09-30 | A method of learning dependence between extracting blocks of knowledge automatically from courseware text |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106649259A CN106649259A (en) | 2017-05-10 |
CN106649259B true CN106649259B (en) | 2019-05-24 |
Family
ID=58854709
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610874480.3A Active CN106649259B (en) | 2016-09-30 | 2016-09-30 | A method of learning dependence between extracting blocks of knowledge automatically from courseware text |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106649259B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107545791B (en) * | 2017-08-29 | 2020-03-06 | 广州思涵信息科技有限公司 | System and method for automatically generating classroom teaching knowledge map by courseware |
CN108021682A (en) * | 2017-12-11 | 2018-05-11 | 西安交通大学 | Open information extracts a kind of Entity Semantics method based on wikipedia under background |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102436480A (en) * | 2011-10-15 | 2012-05-02 | 西安交通大学 | Incidence relation excavation method for text-oriented knowledge unit |
CN104484454A (en) * | 2014-12-27 | 2015-04-01 | 西安交通大学 | Knowledge map oriented network learning behavior and efficiency analysis method |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160041720A1 (en) * | 2014-08-06 | 2016-02-11 | Kaybus, Inc. | Knowledge automation system user interface |
-
2016
- 2016-09-30 CN CN201610874480.3A patent/CN106649259B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102436480A (en) * | 2011-10-15 | 2012-05-02 | 西安交通大学 | Incidence relation excavation method for text-oriented knowledge unit |
CN104484454A (en) * | 2014-12-27 | 2015-04-01 | 西安交通大学 | Knowledge map oriented network learning behavior and efficiency analysis method |
Non-Patent Citations (2)
Title |
---|
Mining preorder relation between knowledge units from text;J Liu 等;《SAC "10 Proceedings of the 2010 ACM Symposium on Applied Computing》;20100326;第1047-1053页 * |
一种从术语定义句中自动抽取知识单元的方法;宋培彦 等;《情报杂志》;20140418;第33卷(第4期);第139-143 * |
Also Published As
Publication number | Publication date |
---|---|
CN106649259A (en) | 2017-05-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104391942B (en) | Short essay eigen extended method based on semantic collection of illustrative plates | |
CN107766324B (en) | Text consistency analysis method based on deep neural network | |
CN110633409B (en) | Automobile news event extraction method integrating rules and deep learning | |
JP5904559B2 (en) | Scenario generation device and computer program therefor | |
CN108255813B (en) | Text matching method based on word frequency-inverse document and CRF | |
CN113239186B (en) | Graph convolution network relation extraction method based on multi-dependency relation representation mechanism | |
CN106649275A (en) | Relation extraction method based on part-of-speech information and convolutional neural network | |
CN103544242A (en) | Microblog-oriented emotion entity searching system | |
CN105512209A (en) | Biomedicine event trigger word identification method based on characteristic automatic learning | |
CN111143672B (en) | Knowledge graph-based professional speciality scholars recommendation method | |
WO2015093540A1 (en) | Phrase pair gathering device and computer program therefor | |
CN103399901A (en) | Keyword extraction method | |
CN111144119B (en) | Entity identification method for improving knowledge migration | |
CN107798624A (en) | A kind of technical label in software Ask-Answer Community recommends method | |
CN109918649B (en) | Suicide risk identification method based on microblog text | |
CN105975475A (en) | Chinese phrase string-based fine-grained thematic information extraction method | |
CN115860006B (en) | Aspect-level emotion prediction method and device based on semantic syntax | |
Sadr et al. | Unified topic-based semantic models: A study in computing the semantic relatedness of geographic terms | |
CN111651983A (en) | Causal event extraction method based on self-training and noise model | |
CN107526721A (en) | A kind of disambiguation method and device to electric business product review vocabulary | |
CN106649259B (en) | A method of learning dependence between extracting blocks of knowledge automatically from courseware text | |
Lin et al. | Sensitive information detection based on convolution neural network and bi-directional LSTM | |
CN110245234A (en) | A kind of multi-source data sample correlating method based on ontology and semantic similarity | |
Tianxiong et al. | Identifying chinese event factuality with convolutional neural networks | |
CN111008285B (en) | Author disambiguation method based on thesis key attribute network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |