CN106649259B - A method of learning dependence between extracting blocks of knowledge automatically from courseware text - Google Patents

A method of learning dependence between extracting blocks of knowledge automatically from courseware text Download PDF

Info

Publication number
CN106649259B
CN106649259B CN201610874480.3A CN201610874480A CN106649259B CN 106649259 B CN106649259 B CN 106649259B CN 201610874480 A CN201610874480 A CN 201610874480A CN 106649259 B CN106649259 B CN 106649259B
Authority
CN
China
Prior art keywords
knowledge
blocks
term
formula
dependence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610874480.3A
Other languages
Chinese (zh)
Other versions
CN106649259A (en
Inventor
魏笔凡
王晨晨
刘均
郑庆华
曾宏伟
姚思雨
吴蓓
石磊
郭朝彤
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN201610874480.3A priority Critical patent/CN106649259B/en
Publication of CN106649259A publication Critical patent/CN106649259A/en
Application granted granted Critical
Publication of CN106649259B publication Critical patent/CN106649259B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/374Thesaurus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/151Transformation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/216Parsing using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/237Lexical tools
    • G06F40/247Thesauruses; Synonyms

Abstract

The invention discloses it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, the text in courseware is corresponded to by handling blocks of knowledge, obtain candidate terms set, then the synonymous term in candidate terms set is handled, and each term is calculated to the criticality of blocks of knowledge, construct optimal model, by solving the study dependence extraction model optimized, courseware text can be automatically analyzed, it identifies the term in text and calculates term to the criticality of blocks of knowledge, and the model that study dependence is excavated is obtained by optimizing the relationship between term, locality of the process independent of study dependence, it can be used to excavate the study dependence that theme is associated between farther away blocks of knowledge, more complete knowledge navigation service is provided for learner.

Description

A method of learning dependence between extracting blocks of knowledge automatically from courseware text
Technical field
The present invention relates to the methods of study dependence, and in particular to it is a kind of extract blocks of knowledge automatically from courseware text between Learn the method for dependence.
Background technique
With the fast development of human sciences' technology, human knowledge total amount shows explosive growth.According to joint state religion The statistics of section's text tissue, the nearly 30 years knowledge accumulated of the mankind account for the 90% of knowledge total amount since the dawn of human civilization, and the multiplication of knowledge Period is still constantly shortening, and has reduced to 5-7 at present.The rapid growth of knowledge total amount is that effective acquisition of knowledge and expression are brought Serious challenge.It is that user feedback goes out relevant documentation that traditional solution, which is by search engine,.This mode cannot be directly anti- The interested knowledge of user is presented, the very big energy of user effort is needed to be screened from a large amount of relevant documentations.Knowledge mapping technology Using RDF triple indicate semantic network, it is intended to realize search engine from " machinery is enumerated " to " network collection is known " develop, for Family provides semantization, correlation information retrieval, alleviates the above problem to a certain extent.But knowledge mapping is not configured to face Cognitive learning to theme can not embody the cognition relationship between each theme, and study is easy to cause to get lost problem.Knowledge Map foundation The characteristics of human cognitive learns forms the efficient expression of one kind and knows by the relational organization between knowledge and knowledge at the form of figure The mode of institutional framework between knowledge and knowledge provides effective method to alleviate the study problem of getting lost.
Study dependence describes the relationship to interdepend in cognitive process between blocks of knowledge.Determine two knowledge Whether unit has relationship, is the Xiang Jiben in Knowledge Map building but very important work.Currently, the knowledge of high quality Figure, it is still necessary to which domain expert marks the study dependence between blocks of knowledge according to domain knowledge, and building process is relatively more slow Slowly.Therefore, the effective study dependence mining algorithm of design will greatly improve Knowledge Map building speed, reduce manpower and disappear Consumption, helps to push the research and application of the navigation learning based on Knowledge Map.
For the method for learning dependence excavation between blocks of knowledge, Patent No. ZL201110312882.1, title For a kind of blocks of knowledge incidence relation method for digging of text-oriented, the method for proposition includes the following steps: that (1) textual association is dug Pick: clustering text collection, finds the text pair with similar topic, and the asymmetry being distributed using central term, Excavate the linear correlation relationship between text;(2) candidate blocks of knowledge pair is generated: using the locality of blocks of knowledge incidence relation, Generate candidate blocks of knowledge pair;(3) feature selecting and blocks of knowledge incidence relation excavate: the term word of knowledge based unit pair Frequently, distance and semantic type feature, using SVM classifier by candidate blocks of knowledge to progress two-value classification, Extracting Knowledge list Incidence relation between member.This method can greatly reduce candidate blocks of knowledge number, under the premise of guaranteeing precision, be effectively reduced The time complexity of relation excavation.Since it makes use of the locality of study dependence, the above method is difficult to extract distance Study dependence between farther away blocks of knowledge.
Summary of the invention
In order to solve the problems in the prior art, the present invention proposes to learn between one kind extracts blocks of knowledge from courseware text automatically The method for practising dependence, can automatically analyze courseware text, identify the term in text and calculate term pair The criticality of blocks of knowledge, and the model that study dependence is excavated, process are obtained by optimizing the relationship between term Independent of the locality of study dependence, it can be used to excavate the study that theme is associated between farther away blocks of knowledge and rely on Relationship provides more complete knowledge navigation service for learner.
In order to achieve the goal above, the technical scheme adopted by the invention is as follows: the following steps are included:
1) candidate terms based on mutual information generate: courseware document being converted into text formatting first, and is carried out at participle Reason;Then the tightness degree that adjacent words combine is measured using mutual information, and processing is merged to compact vocabulary, from And obtain candidate terms set;
2) synonymous term based on wikipedia about subtracts: crawling the corresponding wikipedia page of term, utilizes wikipedia Redirection mark and multilingual link in the page, are about subtracted processing to synonymous term;
3) term criticality is measured: being calculated the corresponding TF-IDF parameter value of each term first, is then utilized knowledge list First name feature and format character are weighted processing to TF-IDF parameter value, measure each term to the pass of blocks of knowledge with this Stroke degree;
4) it optimal model building and solution: establishes between blocks of knowledge and learns to determine between dependence and term relationship Amount indicates, converts optimization problem for model solution problem, constructs the objective function of optimization, and calculate using gradient decline Method carries out model solution, and completion learns dependence between extracting blocks of knowledge automatically from courseware text.
The step 1) the following steps are included:
1.1) blocks of knowledge is corresponded to the text in courseware using poi kit to extract, and is segmented, removes and stop Word processing;
1.2) after setting participle, original character string c is divided into two words a and b, by character string c being total in corpus Existing frequency is denoted as f (c), co-occurrence probabilities of the character string c in corpus is denoted as p (c), then according to maximum likelihood estimate, in number According to measure it is sufficiently large in the case where, it is believed that p (c) be equal to f (c), each word is considered as event, for character string c=ab, mutual information Formula are as follows:The internal combustion tightness degree that character string is measured using mutual information, is obtained Candidate terms set.
The step 2) the following steps are included:
2.1) thesaurus expands: based on thesaurus, utilizing redirection mark in the wikipedia page and more Language link, expands thesaurus;
2.2) synonymous term about subtracts: using the thesaurus expanded by wikipedia, to same in candidate terms set Adopted term is about subtracted processing.
About subtract the mode of processing in the step 2.2) are as follows: for containing the term A of synonymous lexical item, find and have with term A There are identical meanings and the highest term B of the frequency of occurrences, replaces term A with term B in candidate terms set.
The step 3) the following steps are included:
3.1) to each of candidate terms set CT' term, it is calculated to each knowledge list by TF-IDF index The basic criticality of member, TF-IDF index calculation formula are as follows:Formula In: fijIndicate term i in document djIn word frequency;dfiIndicate the document word frequency of term i;N indicates total number of documents;niIndicate document The middle number of files for term i occur;
3.2) weighting of knowledge based unit title: by investigating whether term appears in blocks of knowledge title to original TF-IDF parameter be weighted, weighted formula are as follows: Namei,j=wname×bi,j, in formula: wnameIndicate that blocks of knowledge title adds Weigh weight;bi,jIndicate whether term i appears in the title of blocks of knowledge j;
3.3) based on the weighting of format character: by the font size of term position, to the criticality of term into Row weighting processing, weighted formula are as follows:In formula: wfontIndicate font size weighting Weight;K indicates that blocks of knowledge j corresponds to all different fonts sizes in courseware;fi,kIndicate whether term i is gone out with font size k It is existing;rankkAfter indicating all font size backward sequences, the ranking value of font size k;
3.4) aggregative weighted is carried out to original TF-IDF parameter by blocks of knowledge title and courseware font, obtains term Criticality, the formula of weighting are as follows: scorei,j=wi,j×(1+Namei,j+Fonti,j), in formula: scorei,jIndicate i pairs of term The criticality of blocks of knowledge j.
The step 4) the following steps are included:
4.1) objective function constructs: for blocks of knowledge i and blocks of knowledge j, measuring to exist between them by following formula and learn A possibility that practising dependence:In formula: xiIt is the key that by all terms to blocks of knowledge i journey The vector constituted is spent, each element represents corresponding term to the criticality of blocks of knowledge i in vector;A matrix representative model Parameter;
To blocks of knowledge i, if set omegai=(i, j) | yij=1, j=1,2 ..., n } it is all to be deposited with blocks of knowledge i In the blocks of knowledge of study dependence and the node pair of blocks of knowledge i composition, setFor it is all with blocks of knowledge i there is no the blocks of knowledge of study dependence with know Know the node pair of unit i composition, enablesOptimization is defined as follows to ask Topic:
In formula: X is a matrix, the i-th row in matrix ByIt constitutes;(1-v)+Represent hinge loss function;||A||FRepresent this black norm of not Luo Beini of matrix A;
4.2) it model solution: to optimization problem, is solved using accelerating gradient decline:
It enablesThen former objective function is write as:Formula to A derivation, Obtain gradient:
In formula:ei、ej、ekAll it is Unit vector;
4.3) study dependence is excavated: the most optimized parameter A matrix of model is obtained by step 4.2), for any two A blocks of knowledge is judged between them by optimal model with the presence or absence of study dependence.
Compared with prior art, the present invention corresponds to the text in courseware by handling blocks of knowledge, obtains candidate terms collection It closes, then handles the synonymous term in candidate terms set, and calculate each term to the criticality of blocks of knowledge, construct Optimal model can divide automatically courseware text by solving the study dependence extraction model optimized Analysis identifies the term in text and calculates term to the criticality of blocks of knowledge, and passes through and optimize between term Relationship obtains the model that study dependence is excavated, and locality of the process independent of study dependence can be used to dig Pick theme is associated with the study dependence between farther away blocks of knowledge, provides more complete knowledge navigation clothes for learner Business.The present invention can be corresponded to the text in courseware using blocks of knowledge and extract study dependence between blocks of knowledge automatically, The cost for reducing artificial constructed Knowledge Map, helps to push the research of the navigation learning based on Knowledge Map and answers With.
Detailed description of the invention
Fig. 1 is method flow frame diagram of the invention;
Fig. 2 is the graphical schematic diagram of formula (6);
Fig. 3 is that " Java language " learns dependence excavation partial data exemplary diagram.
Specific embodiment
Below with reference to specific embodiment and Figure of description the present invention will be further explained explanation.
Referring to Fig. 1, the present invention specifically includes the following steps:
1) candidate terms based on mutual information generate, and mainly include 2 steps:
1.1) blocks of knowledge is corresponded to the text in PPT courseware using poi kit to extract, and is segmented, is gone Except stop words processing;
1.2) after setting participle, word string c originally is divided into two words a and b.By co-occurrence of the character string c in corpus Frequency is denoted as f (c), and co-occurrence probabilities of the character string c in corpus are denoted as p (c), then, then according to maximum likelihood estimate, In the case that data volume is sufficiently large, it is believed that p (c) is equal to f (c), if each word is considered as event, for character string c= Ab, mutual information formula are as follows:
The internal combustion tightness degree of character string is measured using mutual information, to achieve the purpose that extract candidate terms. Candidate terms generating algorithm based on mutual information is shown in specific step is as follows:
Input: word segmentation result set Tokens={ word1,word2,...,wordn, vocabulary number n in Tokens, word Frequency statistical information TF, threshold value w
Algorithm flow:
Output: candidate terms set CT
2) synonymous term based on wikipedia about subtracts, and mainly includes 2 steps:
2.1) thesaurus expands: based on " Chinese thesaurus ", utilizing the redirection mark in the wikipedia page With multilingual link, synonymicon is expanded, it is as follows that thesaurus expands specific algorithm:
Input: candidate terms set CT={ term1,term2,...,termn, vocabulary number n, thesaurus D=in CT {(term1,...,termi), wherein (term1,...,termi) it is the i terms with identical meanings
Algorithm flow:
Output: by the thesaurus D expanded
2.2) synonymous term based on thesaurus about subtracts: right herein using the thesaurus expanded by wikipedia Synonymous term in candidate terms set is uniformly processed, and the basic mode of processing is, for containing the art of synonymous lexical item Language A, finding has identical meanings and the highest term B of the frequency of occurrences with term A, replaces in candidate terms concentration term B Term A, respective algorithms are as follows:
Input: candidate terms set CT={ term1,term2,...,termn, vocabulary number n, thesaurus D=in CT {(term1,...,termi), wherein (term1,...,termi) it is the i terms with identical meanings, word frequency statistics information TF
Algorithm flow:
Output: CT'
3) term criticality is measured, and mainly includes 4 steps:
3.1) to each of candidate terms set CT' term, it is calculated to each knowledge list by TF-IDF index The basic criticality of member, TF-IDF index calculation formula are as follows:
In formula: fijIndicate term i in document djIn word frequency;dfiIndicate the document word frequency of term i;N indicates that document is total Number;niIndicate the number of files of term i occur in document;
3.2) weighting of knowledge based unit title: appearing in the term in blocks of knowledge title is likely to be the knowledge The Key Term of unit, can be by investigating whether term appears in blocks of knowledge title to original TF-IDF parameter progress Weighting, weighted formula are as follows:
Namei,j=wname×bi,j (3)
In formula: wnameIndicate that blocks of knowledge title weights weight;bi,jIndicate whether term i appears in the name of blocks of knowledge j In title;
3.3) weighting based on PPT format character: in the classification of PPT is shown, level is higher, and font is generally bigger, Expressed content is more important, therefore, can be weighted by the font size of term position to the criticality of term Processing, weighted formula are as follows:
In formula: wfontIndicate that font size weights weight;It is big that k indicates that blocks of knowledge j corresponds to all different fonts in courseware It is small;fi,kIndicate term i whether with font size k appearance;rankkAfter indicating all font size backward sequences, font size k Ranking value;
3.4) formula of aggregative weighted is carried out to original TF-IDF parameter by blocks of knowledge title and courseware font are as follows:
scorei,j=wi,j×(1+Namei,j+Fonti,j) (5)
In formula: scorei ,jIndicate term i to the criticality of blocks of knowledge j;
4) optimal model building and solution, mainly include 3 steps:
4.1) objective function constructs: firstly, for blocks of knowledge i and blocks of knowledge j, being measured by following formula and is deposited between them A possibility that learning dependence:
In formula: xiIt is the vector being made of criticality of all terms to blocks of knowledge i, each element generation in vector Criticality of the table corresponding term to blocks of knowledge i;The parameter of A matrix representative model, as shown in Fig. 2, formula (6) can be regarded as The sum of weight from paths all blocks of knowledge i to blocks of knowledge j;
To blocks of knowledge i, if set omegai=(i, j) | yij=1, j=1,2 ..., n }, i.e., the set be it is all with know Know unit i and there is the blocks of knowledge of study dependence and the node pair of blocks of knowledge i composition, setI.e. the set is all knowledge that study dependence is not present with blocks of knowledge i The node pair of unit and blocks of knowledge i composition, model parameter A reasonable for one, it is desirable to each pair of knowledge in set omega i The score value of formula corresponding to unit will be greater than setIn each pair of blocks of knowledge score value, enableIt is defined as follows optimization problem:
In formula: X is a matrix, in matrix the i-th row byIt constitutes;(1-v)+Represent hinge loss function;||A||FGeneration This black norm of the not Luo Beini of table matrix A;
4.2) for optimization problem shown in formula (7), accelerating gradient decline model solution: can be used (AcceleratedGradient Descent) is solved:
It enablesThen former objective function can be write as:
Formula (8) obtains gradient to A derivation:
In formula:ei、ej、ekAll it is unit vector, circular is as follows:
Input: X, T, λ, η, maximum number of iterations N
Algorithm flow:
Output: A
4.3) study dependence is excavated: by step 4.2), the most optimized parameter A matrix of available model, at this point, For any two blocks of knowledge, can judge to whether there is study dependence between them by the optimal model, If Fig. 3 illustrates the few examples of " Java language " course learning dependence Result, solid line represents known learn in figure Dependence is practised, dotted line represents the blocks of knowledge pair for needing to judge, the number on dotted line represents the blocks of knowledge to study A possibility that dependence.

Claims (4)

1. it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, which is characterized in that including following Step:
1) candidate terms based on mutual information generate: courseware document being converted into text formatting first, and carries out word segmentation processing;So The tightness degree that adjacent words combine is measured using mutual information afterwards, and processing is merged to compact vocabulary, thus To candidate terms set;
2) synonymous term based on wikipedia about subtracts: crawling the corresponding wikipedia page of term, utilizes the wikipedia page In redirection mark and multilingual link, about subtracted processing to synonymous term;
3) term criticality is measured: being calculated the corresponding TF-IDF parameter value of each term first, is then utilized blocks of knowledge name Claim feature and format character to be weighted processing to TF-IDF parameter value, each term is measured to the crucial journey of blocks of knowledge with this Degree;
4) optimal model building and solution: the quantitative table learnt between dependence and term relationship is established between blocks of knowledge Show, convert optimization problem for model solution problem, construct the objective function of optimization, and using gradient descent algorithm into Row model solution, completion learn dependence between extracting blocks of knowledge automatically from courseware text;
The step 3) the following steps are included:
3.1) to each of candidate terms set CT' term, it is calculated to each blocks of knowledge by TF-IDF index Basic criticality, TF-IDF index calculation formula are as follows:In formula: fijIndicate term i in document djIn word frequency;dfiIndicate the document word frequency of term i;N indicates total number of documents;niIt indicates in document There is the number of files of term i;
3.2) weighting of knowledge based unit title: by investigating whether term appears in blocks of knowledge title to original TF-IDF parameter is weighted, weighted formula are as follows: Namei,j=wname×bi,j, in formula: wnameIndicate the weighting of blocks of knowledge title Weight;bi,jIndicate whether term i appears in the title of blocks of knowledge j;
3.3) based on the weighting of format character: by the font size of term position, adding to the criticality of term Power processing, weighted formula are as follows:In formula: wfontIndicate that font size weights weight; K indicates that blocks of knowledge j corresponds to all different fonts sizes in courseware;fi,kIndicate term i whether with font size k appearance; rankkAfter indicating all font size backward sequences, the ranking value of font size k;
3.4) aggregative weighted is carried out to original TF-IDF parameter by blocks of knowledge title and courseware font, obtains term key Degree, the formula of weighting are as follows: scorei,j=wi,j×(1+Namei,j+Fonti,j), in formula: scorei,jIndicate term i to knowledge The criticality of unit j;
The step 4) the following steps are included:
4.1) objective function constructs: for blocks of knowledge i and blocks of knowledge j, measured between them by following formula exist study according to A possibility that relationship of relying:In formula: xiIt is the criticality structure by all terms to blocks of knowledge i At vector, each element represents corresponding term to the criticality of blocks of knowledge i in vector;The ginseng of A matrix representative model Number;
To blocks of knowledge i, if set omegai=(i, j) | yij=1, j=1,2 ..., n } it is all and blocks of knowledge i presence Practise the blocks of knowledge of dependence and the node pair of blocks of knowledge i composition, setFor The node pair of all blocks of knowledge and blocks of knowledge i composition that study dependence is not present with blocks of knowledge i, enablesIt is defined as follows optimization problem:
In formula: X is a matrix, in matrix the i-th row by It constitutes;(1-v)+Represent hinge loss function;||A||FRepresent this black norm of not Luo Beini of matrix A;
4.2) it model solution: to optimization problem, is solved using accelerating gradient decline:
It enables,Then former objective function is write as:Formula obtains A derivation To gradient:
In formula:ei、ej、ekIt is all unit Vector;
4.3) study dependence is excavated: obtaining the most optimized parameter A matrix of model by step 4.2), any two are known Know unit, is judged between them by optimal model with the presence or absence of study dependence.
2. it is according to claim 1 it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, It is characterized in that, the step 1) the following steps are included:
1.1) blocks of knowledge is corresponded to the text in courseware using poi kit to extract, and is segmented, removes stop words Processing;
1.2) after setting participle, original character string c is divided into two words a and b, by co-occurrence frequency of the character string c in corpus Rate is denoted as f (c), co-occurrence probabilities of the character string c in corpus is denoted as p (c), then according to maximum likelihood estimate, in data volume In the case where sufficiently large, it is believed that p (c) is equal to f (c), each word is considered as event, for character string c=ab, mutual information formula Are as follows:The internal combustion tightness degree that character string is measured using mutual information, obtains candidate Term set.
3. it is according to claim 1 it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, It is characterized in that, the step 2) the following steps are included:
2.1) thesaurus expands: based on thesaurus, utilizing redirection mark in the wikipedia page and multilingual Link, expands thesaurus;
2.2) synonymous term about subtracts: using the thesaurus expanded by wikipedia, to the synonymous art in candidate terms set Language is about subtracted processing.
4. it is according to claim 3 it is a kind of from courseware text extract blocks of knowledge automatically between learn dependence method, It is characterized in that, about subtracting the mode of processing in the step 2.2) are as follows: for containing the term A of synonymous lexical item, find and term A With identical meanings and the highest term B of the frequency of occurrences, term A is replaced with term B in candidate terms set.
CN201610874480.3A 2016-09-30 2016-09-30 A method of learning dependence between extracting blocks of knowledge automatically from courseware text Active CN106649259B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610874480.3A CN106649259B (en) 2016-09-30 2016-09-30 A method of learning dependence between extracting blocks of knowledge automatically from courseware text

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610874480.3A CN106649259B (en) 2016-09-30 2016-09-30 A method of learning dependence between extracting blocks of knowledge automatically from courseware text

Publications (2)

Publication Number Publication Date
CN106649259A CN106649259A (en) 2017-05-10
CN106649259B true CN106649259B (en) 2019-05-24

Family

ID=58854709

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610874480.3A Active CN106649259B (en) 2016-09-30 2016-09-30 A method of learning dependence between extracting blocks of knowledge automatically from courseware text

Country Status (1)

Country Link
CN (1) CN106649259B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107545791B (en) * 2017-08-29 2020-03-06 广州思涵信息科技有限公司 System and method for automatically generating classroom teaching knowledge map by courseware
CN108021682A (en) * 2017-12-11 2018-05-11 西安交通大学 Open information extracts a kind of Entity Semantics method based on wikipedia under background

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102436480A (en) * 2011-10-15 2012-05-02 西安交通大学 Incidence relation excavation method for text-oriented knowledge unit
CN104484454A (en) * 2014-12-27 2015-04-01 西安交通大学 Knowledge map oriented network learning behavior and efficiency analysis method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160041720A1 (en) * 2014-08-06 2016-02-11 Kaybus, Inc. Knowledge automation system user interface

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102436480A (en) * 2011-10-15 2012-05-02 西安交通大学 Incidence relation excavation method for text-oriented knowledge unit
CN104484454A (en) * 2014-12-27 2015-04-01 西安交通大学 Knowledge map oriented network learning behavior and efficiency analysis method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Mining preorder relation between knowledge units from text;J Liu 等;《SAC "10 Proceedings of the 2010 ACM Symposium on Applied Computing》;20100326;第1047-1053页 *
一种从术语定义句中自动抽取知识单元的方法;宋培彦 等;《情报杂志》;20140418;第33卷(第4期);第139-143 *

Also Published As

Publication number Publication date
CN106649259A (en) 2017-05-10

Similar Documents

Publication Publication Date Title
CN104391942B (en) Short essay eigen extended method based on semantic collection of illustrative plates
CN107766324B (en) Text consistency analysis method based on deep neural network
CN110633409B (en) Automobile news event extraction method integrating rules and deep learning
JP5904559B2 (en) Scenario generation device and computer program therefor
CN108255813B (en) Text matching method based on word frequency-inverse document and CRF
CN113239186B (en) Graph convolution network relation extraction method based on multi-dependency relation representation mechanism
CN106649275A (en) Relation extraction method based on part-of-speech information and convolutional neural network
CN103544242A (en) Microblog-oriented emotion entity searching system
CN105512209A (en) Biomedicine event trigger word identification method based on characteristic automatic learning
CN111143672B (en) Knowledge graph-based professional speciality scholars recommendation method
WO2015093540A1 (en) Phrase pair gathering device and computer program therefor
CN103399901A (en) Keyword extraction method
CN111144119B (en) Entity identification method for improving knowledge migration
CN107798624A (en) A kind of technical label in software Ask-Answer Community recommends method
CN109918649B (en) Suicide risk identification method based on microblog text
CN105975475A (en) Chinese phrase string-based fine-grained thematic information extraction method
CN115860006B (en) Aspect-level emotion prediction method and device based on semantic syntax
Sadr et al. Unified topic-based semantic models: A study in computing the semantic relatedness of geographic terms
CN111651983A (en) Causal event extraction method based on self-training and noise model
CN107526721A (en) A kind of disambiguation method and device to electric business product review vocabulary
CN106649259B (en) A method of learning dependence between extracting blocks of knowledge automatically from courseware text
Lin et al. Sensitive information detection based on convolution neural network and bi-directional LSTM
CN110245234A (en) A kind of multi-source data sample correlating method based on ontology and semantic similarity
Tianxiong et al. Identifying chinese event factuality with convolutional neural networks
CN111008285B (en) Author disambiguation method based on thesis key attribute network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant