CN117332768A - Data processing system for acquiring text generation template - Google Patents

Data processing system for acquiring text generation template Download PDF

Info

Publication number
CN117332768A
CN117332768A CN202311308051.6A CN202311308051A CN117332768A CN 117332768 A CN117332768 A CN 117332768A CN 202311308051 A CN202311308051 A CN 202311308051A CN 117332768 A CN117332768 A CN 117332768A
Authority
CN
China
Prior art keywords
model
text
model type
candidate
list
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311308051.6A
Other languages
Chinese (zh)
Other versions
CN117332768B (en
Inventor
石江枫
于伟
靳雯
王全修
赵洲洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Rizhao Ruian Information Technology Co ltd
Beijing Rich Information Technology Co ltd
Original Assignee
Rizhao Ruian Information Technology Co ltd
Beijing Rich Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Rizhao Ruian Information Technology Co ltd, Beijing Rich Information Technology Co ltd filed Critical Rizhao Ruian Information Technology Co ltd
Priority to CN202311308051.6A priority Critical patent/CN117332768B/en
Publication of CN117332768A publication Critical patent/CN117332768A/en
Application granted granted Critical
Publication of CN117332768B publication Critical patent/CN117332768B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/186Templates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/338Presentation of query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a data processing system for acquiring a text generation template, which comprises the following steps: the method comprises the following steps of key text corresponding to a target user, a preset model type tree, a processor and a memory storing a computer program, wherein when the computer program is executed by the processor, the following steps are realized: obtaining a candidate template text; obtaining a target NLP model according to the candidate template text and a plurality of first preset sample texts; acquiring a middle confidence coefficient list according to the key text and the target NLP model; acquiring a middle text generation template according to the middle confidence coefficient list, the number of characters of all keywords corresponding to the key text and the number of characters in the key text; adjusting the intermediate text generation template to obtain a target text generation template; according to the invention, the text generation template does not need to be manually manufactured or selected, so that the accuracy of acquiring the text generation template is improved, the thinking of a user can be widened, and the working efficiency of the user is improved.

Description

Data processing system for acquiring text generation template
Technical Field
The invention relates to the technical field of text processing, in particular to a data processing system for acquiring a text generation template.
Background
With the wide application of the internet, when a user needs to generate a report, the report can be directly generated through the text generation template without manually writing the report, so that it is very necessary to obtain an accurate and proper text generation template.
However, the above method also has the following technical problems:
the existing templates are limited in online template making or self-use, uncontrollable, a user needs to make or select a template through a large amount of thinking, and when the thinking of the user is not divergent enough, the made template is not accurate enough, and the application range of the template is limited, so that the text generation template obtained through the method is not accurate enough.
Disclosure of Invention
Aiming at the technical problems, the invention adopts the following technical scheme:
a data processing system for retrieving a text generation template, comprising: key text T corresponding to target user 0 The method comprises the following steps of:
s1, according to T 0 And A, obtain T 0 Corresponding candidate template text list t= { T 1 ,T 2 ,……,T b ,……,T c },T b The candidate template text for the b-th candidate template, b=1, 2, … …, c, c is the number of candidate templates, and the candidate template text is a text for describing the applicable range of the candidate template.
S2, training an NLP model according to the T and a plurality of first preset sample texts to obtain a target NLP model, wherein the result of the output of the NLP model is that the first preset sample texts are respectively matched with the T 1 ,T 2 ,……,T b ,……,T c Confidence between them.
S3, T is taken as 0 Input into target NLP model to obtain T 0 Corresponding intermediate confidence list T 1 ={T 1 1 ,T 1 2 ,……,T 1 b ,……,T 1 c },T 1 b Is T 0 And T b An intermediate confidence level between.
S4, when AT/BT is more than or equal to YT, T is processed according to a first processing method 0 T and T 1 Processing to obtain an intermediate text generation template, wherein AT is T 0 The sum of the number of characters of all corresponding keywords, BT is T 0 The first processing method in S4 includes the following steps:
s41, acquiring a first keyword list GJ= { GJ 1 ,GJ 2 ,……,GJ (ai) ,……,GJ (am) },GJ (ai) For the ai first keyword, ai=1, 2, … …, am, am is the number of first keywords, and the first keyword is T 0 Corresponding keywords.
S42, acquiring a second keyword list GJ corresponding to the T 0 ={GJ 0 1 ,GJ 0 2 ,……,GJ 0 b ,……,GJ 0 c },GJ 0 b ={GJ 0 b1 ,GJ 0 b2 ,……,GJ 0 b(aj) ,……,GJ 0 b(an) },GJ 0 b(aj) Is T b Corresponding toIs a second keyword list GJ of (2) 0 b Aj=1, 2, … …, an, an is the number of second keywords in the second keyword list, and the second keywords are keywords in the candidate template text.
S43, according to GJ and GJ 0 Acquiring a list similarity list GJ corresponding to the GJ 1 ={GJ 1 1 ,GJ 1 2 ,……,GJ 1 b ,……,GJ 1 c },GJ 1 b Is GJ and GJ 0 b List similarity between, wherein GJ 1 b Meets the following conditions:
GJ 1 b =Σ am ai=1an aj=1 XS (ai) (aj) /an)/am,XS (ai) (aj) for GJ (ai) With GJ 0 b(aj) Word similarity between.
S44 according to T 1 And GJ 1 Acquiring a first priority list YX= { YX corresponding to T 1 ,YX 2 ,……,YX b ,……,YX c },YX b Is T b Corresponding first priority, YX b Meets the following conditions:
YX b =T 1 b +GJ 1 b
s45, when there is only one maximum value in YX, determining max (YX 1 ,YX 2 ,……,YX b ,……,YX c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
S46, when there are a plurality of maximum values in YX, according to T 0 And T b An intermediate text generation template is determined.
S5, when AT/BT is less than YT, obtaining an intermediate text generation template according to a second processing method.
S6, the target user adjusts the intermediate text generation template to obtain a target text generation template.
The invention has at least the following beneficial effects:
the invention provides a data processing system for acquiring a text generation template, which comprises the following steps: the method comprises the following steps of key text corresponding to a target user, a preset model type tree, a processor and a memory storing a computer program, wherein when the computer program is executed by the processor, the following steps are realized: obtaining a candidate template text according to the key text and a preset model type tree; obtaining a target NLP model according to the candidate template text and a plurality of first preset sample texts; inputting the key text into a target NLP model, acquiring an intermediate confidence list corresponding to the key text, acquiring an intermediate text generation template according to a first processing method when the ratio of the number of characters of all keywords corresponding to the key text to the number of characters in the key text is not smaller than a preset character number ratio threshold, otherwise, acquiring the intermediate text generation template according to a second processing method; the target user adjusts the intermediate text generation template to obtain a target text generation template; according to the method, the candidate template text list can be obtained through the key text input by the user, the intermediate text generation template is determined according to the key text, the key words corresponding to the key text and the key words in the candidate template text, the intermediate text generation template is adjusted, and the target text generation template is generated.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings required for the description of the embodiments will be briefly described below, and it is apparent that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flowchart of a data processing system executing a computer program for obtaining a text generation template according to an embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to fall within the scope of the invention.
An embodiment of the present invention provides a data processing system for acquiring a text generation template, including: key text T corresponding to target user 0 The system comprises a preset model type tree A, a processor and a memory for storing a computer program, wherein the A comprises a plurality of preset model type nodes, each preset model type node corresponds to a plurality of appointed text generation templates, and when the computer program is executed by the processor, the following steps are realized as shown in fig. 1:
s1, according to T 0 And A, obtain T 0 Corresponding candidate template text list t= { T 1 ,T 2 ,……,T b ,……,T c },T b The candidate template text for the b-th candidate template, b=1, 2, … …, c, c is the number of candidate templates, and the candidate template text is a text for describing the applicable range of the candidate template.
Specifically, the target user is a user who needs to generate a template using text.
Specifically, the candidate template text corresponding to the candidate template is stored in the system.
Specifically, S1 includes the steps of obtaining a candidate template:
s11, T is taken 0 Inputting into a preset keyword extraction model to obtain T 0 The corresponding keyword information list includes a plurality of keyword information, wherein, the person skilled in the art knows that the preset keyword extraction model is a model trained in advance by the person skilled in the art according to actual requirements, and the details are not repeated here.
Specifically, the keyword information includes: keywords, keyword types.
S12, acquiring a candidate word type list B= { B according to the keyword information list 1 ,B 2 ,……,B e ,……,B f },B e For the e candidate word type, e=1, 2, … …, f, f is the number of candidate word types, and the candidate word type is the keyword type obtained by de-duplicating the keyword types in all the keyword information.
S13, acquiring a third model type tree according to A and B, wherein A= { A 1 ,A 2 ,……,A i ,……,A m },A i ={A i1 ,A i2 ,……,A ij ,……,A in(i) },A ij For the j-th preset model type node in the i-th layer of the preset model type tree, i=1, 2, … …, m, m is the number of layers of the preset model type tree, j=1, 2, … …, n (i), and n (i) is the number of preset model type nodes in the i-th layer of the preset model type tree.
Specifically, S13 includes the steps of obtaining a third model type tree:
s131, let all A ij NULL to obtain a first specified model type tree corresponding to A 1 ={A 1 1 ,A 1 2 ,……,A 1 i ,……,A 1 m },A 1 i ={A 1 i1 ,A 1 i2 ,……,A 1 ij ,……,A 1 in(i) },A 1 ij Is A ij The corresponding designated model type node, where the person skilled in the art knows that the preset model type tree is a tree preset by the person skilled in the art according to the actual requirement, and will not be described herein again.
S132, obtaining A ij Corresponding preset model type A 0 ij The preset model type is the model type presented by the preset model type node.
S133, obtaining A 0 ij And B is connected with e Type similarity AB between 0e ij The type similarity is the similarity between the preset model type and the candidate word type, wherein, the model type is known to those skilled in the art and can be understood as a label, and any method for obtaining the similarity between two labels in the prior art belongs to the protection scope of the present invention, and is not described herein in detail, for example: cosine similarity, edit distance.
Specifically, the greater the type similarity, the more similar the preset model type and the candidate word type.
S134, when B e Any one of corresponding AB 0e ij ≥A 2 When max (AB 0e 11 ,AB 0e 12 ,……,AB 0e 1j ,……,AB 0e 1n(i) ,AB 0e 21 ,AB 0e 22 ,……,AB 0e 2j ,……,AB 0e 2n(i) ,……,AB 0e i1 ,AB 0e i2 ,……,AB 0e ij ,……,AB 0e in(i) ,……,AB 0e m1 ,AB 0e m2 ,……,AB 0e mj ,……,AB 0e mn(i) ) Corresponding A 0 ij Corresponding A ij Corresponding A 1 ij Replaced by B e To obtain a second specified model tree, wherein max () is a maximum value obtaining function, A 2 And presetting a similarity threshold.
Specifically, A 2 The value range of (5) is [0.8,1 ]]Wherein, the person skilled in the art knows that the person skilled in the art can realize the invention according to the actual requirements [0.8,1 ]]Specific values of the preset similarity threshold are set, and are not described herein.
S135, when B e All corresponding AB 0e ij <A 2 At the time, obtain A ij And B is connected with e Similarity priority between YX e ij Wherein YX e ij Meets the following conditions:
YX e ij =FJ e ij +ZJ e ij ,FJ e ij is A ij The preset model type corresponding to the father node of (B) e Type similarity between ZJ e ij Is A ij The type of the preset model corresponding to the child node of (B) e Type similarity between, when A ij FJ without parent node e ij 0, when A ij ZJ without child node e ij Is 0.
S136, max (YX e 11 ,YX e 12 ,……,YX e 1j ,……,YX e 1n(i) ,YX e 21 ,YX e 22 ,……,YX e 2j ,……,YX e 2n(i) ,……,YX e i1 ,YX e i2 ,……,YX e ij ,……,YX e in(i) ,……,YX e m1 ,YX e m2 ,……,YX e mj ,……,YX e mn(i) ) Corresponding A ij Corresponding A 1 ij Replaced by B e To obtain a third specified model tree.
S137, when the j-th second designated model node of the i-th layer of the second designated model tree is NULL and the j-th third designated model node of the i-th layer of the third designated model tree is not NULL, replacing the j-th second designated model node of the i-th layer of the second designated model tree with the j-th third designated model node of the i-th layer of the third designated model tree to obtain a third model type tree.
And when the similarity of all types corresponding to the candidate word type is smaller than the preset similarity threshold, explaining that the candidate word type is quite similar to the preset model type, determining the position of the candidate word type in the third model type tree, generating a third model type tree, determining the position of the candidate word type in the third model type tree according to the preset model type, determining the candidate word type, and generating a candidate text, wherein the candidate word type is not similar to the preset model type, the candidate word type cannot be determined in the third model type tree, the similarity priority corresponding to the candidate word type is acquired, the similarity priority is compared, the position of the candidate word type in the third model type tree is determined, the position of the candidate word type in the third model type tree is used for determining, the second model type tree is fused with the third model type tree, the node in the third model type tree can be accurately determined, the third model type tree is generated, the position of the candidate word type in the third model type tree is not determined according to the preset similarity threshold, the candidate word type is not corresponding to the first model type tree, the candidate text is not used as the candidate text, and the candidate text is prevented from being generated.
S14, determining the level of the nearest public ancestor ZX of all non-NULL nodes in the third model type tree and the position of the nearest public ancestor ZX in the level, wherein a person skilled in the art knows that any mode for acquiring the nearest public ancestor in the prior art belongs to the protection scope of the invention and is not repeated herein; it is understood that determining ZX is the third model type node of the third model type tree at what level.
S15, acquiring a first candidate model type node from the A, wherein the first candidate model type node is a level in the A and a preset model type node with the same position in the level as the level of ZX in a third model type tree and the position in the level; it can be understood that: if ZX is the j-th third model type node of the i-th layer of the third model type tree, the first candidate model type node is the j-th first preset model type node of the i-th layer of the preset model type nodes
S16, acquiring a second candidate model type node list according to the first candidate model type node, wherein the second candidate model type node list comprises a plurality of second candidate model type nodes, and the second candidate model type nodes are preset model type nodes which have the same father node as the first candidate model type nodes.
Specifically, the text generation template is designated as the template stored by the system before the current point in time.
S17, obtaining a candidate template, wherein the candidate template is any appointed text generation template corresponding to any second candidate model type node.
And according to the third model type tree and the preset model type tree, determining a first candidate model type node from the preset model type tree, further determining a second candidate model type node, and taking a specified text generation template corresponding to the second candidate model type node as a candidate template.
S2, training an NLP model according to the T and a plurality of first preset sample texts to obtain a target NLP model, wherein the result of the output of the NLP model is that the first preset sample texts are respectively matched with the T 1 ,T 2 ,……,T b ,……,T c The confidence level between the two, wherein, the person skilled in the art knows that the first preset sample text is a text preset by the person skilled in the art according to the actual requirement, and any method for training the model in the prior art belongs to the protection scope of the present invention, and is not described in detail herein, for example: unsupervised training and supervised training.
Specifically, the first preset sample text and T respectively 1 ,T 2 ,……,T b ,……,T c The sum of the confidence levels between them is 1.
S3, T is taken as 0 Input into target NLP model to obtain T 0 Corresponding intermediate confidence list T 1 ={T 1 1 ,T 1 2 ,……,T 1 b ,……,T 1 c },T 1 b Is T 0 And T b An intermediate confidence level between.
Specifically, Σ c b=1 T 1 b =1。
S4, when AT/BT is more than or equal to YT, T is processed according to a first processing method 0 T and T 1 Processing to obtain an intermediate text generation template, wherein AT is T 0 The sum of the number of characters of all corresponding keywords, BT is T 0 The number of the medium characters, YT, is a preset character number proportion threshold value.
Specifically, the value range of YT is [0.6,1], where those skilled in the art know that those skilled in the art set the preset character number proportion threshold in [0.6,1] according to the actual requirement, and the details are not repeated here.
Specifically, the first processing method in S4 includes the following steps:
s41, acquiring a first keyword list GJ= { GJ 1 ,GJ 2 ,……,GJ (ai) ,……,GJ (am) },GJ (ai) For the ai first keyword, ai=1, 2, … …, am, am is the number of first keywords, and the first keyword is T 0 Corresponding keywords.
S42, acquiring a second keyword list GJ corresponding to the T 0 ={GJ 0 1 ,GJ 0 2 ,……,GJ 0 b ,……,GJ 0 c },GJ 0 b ={GJ 0 b1 ,GJ 0 b2 ,……,GJ 0 b(aj) ,……,GJ 0 b(an) },GJ 0 b(aj) Is T b Corresponding second keyword list GJ 0 b The aj=1, 2, … …, an, an is the number of second keywords in the second keyword list, and the second keywords are keywords in the candidate template text, where, as known by those skilled in the art, the manner of obtaining the keywords in the candidate template text is the same as the manner of obtaining the keywords corresponding to the keyword text, and will not be described herein again.
S43, according to GJ and GJ 0 Acquiring a list similarity list GJ corresponding to the GJ 1 ={GJ 1 1 ,GJ 1 2 ,……,GJ 1 b ,……,GJ 1 c },GJ 1 b Is GJ and GJ 0 b List similarity between, wherein GJ 1 b Meets the following conditions:
GJ 1 b =Σ am ai=1an aj=1 XS (ai) (aj) /an)/am,XS (ai) (aj) for GJ (ai) With GJ 0 b(aj) The term similarity between the two terms, wherein, the person skilled in the art knows that any mode of obtaining the term similarity between the two terms in the prior art belongs to the protection scope of the present invention, and is not repeated here.
Specifically, the larger the value of the term similarity, the more similar the terms are.
Further, the larger the value of the list similarity, the more similar the first keyword list and the second keyword list are.
S44 according to T 1 And GJ 1 Acquiring a first priority list YX= { YX corresponding to T 1 ,YX 2 ,……,YX b ,……,YX c },YX b Is T b Corresponding first priority, YX b Meets the following conditions:
YX b =T 1 b +GJ 1 b
s45, when there is only one maximum value in YX, determining max (YX 1 ,YX 2 ,……,YX b ,……,YX c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
S46, when there are a plurality of maximum values in YX, according to T 0 And T b An intermediate text generation template is determined.
When the ratio of the number of characters of all keywords corresponding to the key text to the number of characters of the key text is not smaller than the preset character number proportion threshold, the meaning carried by the text can be expressed by the keywords, the similarity between all keywords corresponding to the key text and all keywords corresponding to the candidate template text, namely the list similarity, is obtained, the sum of the confidence coefficient and the list similarity is used as the first priority corresponding to the candidate template text, the candidate template corresponding to the candidate template text with the largest first priority is selected as the intermediate text generation template, the accuracy of obtaining the intermediate text generation template is improved, when a plurality of maximum first priority values exist, the intermediate text generation template cannot be accurately obtained according to the first priority, the key text and the candidate template text need to be analyzed again, and the intermediate text generation template is determined.
Specifically, S46 includes the steps of:
s461, T 0 Inputting into a preset vector extraction model to obtain T 0 Corresponding first text vector list U 0 ={U 0 1 ,U 0 2 ,……,U 0 v ,……,U 0 w },U 0 v Is T 0 The corresponding v-th first text vector, v=1, 2, … …, w, w is the number of first text vectors, wherein the preset vector extraction model is a neural network model capable of extracting text vectors, and those skilled in the art know that any neural network model capable of obtaining text vectors in the prior art belongs to the protection scope of the present invention, and is not repeated herein.
S462, will T b Inputting into a preset vector extraction model to obtain T b Corresponding second text vector list U b ={U b1 ,U b2 ,……,U bv ,……,U bw },U bv Is T b A corresponding v-th second text vector.
S463 according to U 0 And U b Obtaining T 0 Respectively sum to T 1 ,T 2 ,……,T b ,……,T c Text similarity L between 0 1 ,L 0 2 ,……,L 0 b ,……,L 0 c Wherein L is 0 b Is T 0 And T b Text similarity between, and L 0 b Meets the following conditions:
s464 according to T 1 b And L 0 b Acquiring a second priority list YX corresponding to T 0 ={YX 0 1 ,YX 0 2 ,……,YX 0 b ,……,YX 0 c },YX 0 b Is T b Corresponding second priority, YX 0 b Meets the following conditions:
YX 0 b =T 1 b +L 0 b
s465, YX 0 When there is only one maximum value, max (YX 0 1 ,YX 0 2 ,……,YX 0 b ,……,YX 0 c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
S466, when YX 0 When there are a plurality of maximum values in (c), max (T 1 1 +GJ 1 1 +L 0 1 ,T 1 2 +GJ 1 2 +L 0 2 ,……,T 1 b +GJ 1 b +L 0 b ,……,T 1 c +GJ 1 c +L 0 c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
When the plurality of maximum first priority values exist, the text similarity between the key text and the candidate template text is obtained, the sum of the confidence coefficient and the text similarity is taken as the second priority corresponding to the candidate template text, the candidate template corresponding to the candidate template text with the maximum second priority is selected as the intermediate text generation template, the accuracy of obtaining the intermediate text generation template is improved, when the plurality of maximum second priority values exist, the intermediate text generation template cannot be obtained accurately according to the second priority, and at the moment, the candidate template corresponding to the candidate template text with the maximum sum of the confidence coefficient, the list similarity and the text similarity is required to be selected as the intermediate text generation template, and the accuracy of obtaining the intermediate text generation template is improved.
S5, when AT/BT is less than YT, obtaining an intermediate text generation template according to a second processing method.
Specifically, the second processing method in S5 includes the following steps:
s51, when YX 0 When there is only one maximum value, max (YX 0 1 ,YX 0 2 ,……,YX 0 b ,……,YX 0 c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
S53, when YX 0 When there are a plurality of maximum values in (c), max (T 1 1 +GJ 1 1 +L 0 1 ,T 1 2 +GJ 1 2 +L 0 2 ,……,T 1 b +GJ 1 b +L 0 b ,……,T 1 c +GJ 1 c +L 0 c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
When the ratio of the number of characters of all keywords corresponding to the key text to the number of characters of the key text is smaller than the preset character number proportion threshold, it is stated that the keywords may not express the meaning carried by the text, and the intermediate text generation template cannot be acquired mainly by the keywords, and at this time, the candidate template corresponding to the candidate template text with the largest second priority is selected as the intermediate text generation template, which is favorable for improving the accuracy of acquiring the intermediate text generation template, and when a plurality of the largest second priority values exist, the intermediate text generation template cannot be acquired accurately according to the second priority, and at this time, the candidate template corresponding to the candidate template text with the largest sum of confidence, list similarity and text similarity needs to be selected as the intermediate text generation template, which is favorable for improving the accuracy of acquiring the intermediate text generation template.
S6, adjusting the intermediate text generation template by the target user to obtain the target text generation template, wherein the method for adjusting the intermediate text generation template by the target user is known to the person skilled in the art and is set by the person skilled in the art according to actual requirements, and details are omitted here, for example: and adding the data statistical model in the intermediate text generating template, and deleting the data statistical model in the intermediate text generating template.
The invention provides a data processing system for acquiring a text generation template, which comprises the following steps: the method comprises the following steps of key text corresponding to a target user, a preset model type tree, a processor and a memory storing a computer program, wherein when the computer program is executed by the processor, the following steps are realized: obtaining a candidate template text according to the key text and a preset model type tree; obtaining a target NLP model according to the candidate template text and a plurality of first preset sample texts; inputting the key text into a target NLP model, acquiring an intermediate confidence list corresponding to the key text, acquiring an intermediate text generation template according to a first processing method when the ratio of the number of characters of all keywords corresponding to the key text to the number of characters in the key text is not smaller than a preset character number ratio threshold, otherwise, acquiring the intermediate text generation template according to a second processing method; the target user adjusts the intermediate text generation template to obtain a target text generation template; according to the method, the candidate template text list can be obtained through the key text input by the user, the intermediate text generation template is determined according to the key text, the key words corresponding to the key text and the key words in the candidate template text, the intermediate text generation template is adjusted, and the target text generation template is generated.
The invention also provides a data statistics system based on the text generation template, which takes the target text generation template as an initial text generation template and comprises the following steps after S6:
s100, acquiring a first model type list according to an initial text generation template, wherein the first model type list comprises a plurality of first model types, the first model types are model types of initial data statistical models in the initial text generation template, wherein the initial data statistical models are known to a person skilled in the art, and are models which are trained in advance by the person skilled in the art and can be used for counting data in a data set and outputting result data, and the details are not repeated.
Specifically, the initial text generation template comprises a plurality of initial data statistical models.
Specifically, the different initial data statistical models output different result data.
Further, the form of the result data output by the initial data statistical model is one or more of a data table form, a text form, a digital form, a data set form and the like.
Specifically, S100 includes the following steps:
s101, acquiring an initial data statistical model name list from an initial text generation template, wherein the initial data statistical model name list comprises a plurality of initial data statistical model names in the initial text generation template, and the initial data statistical model names are names of initial data statistical models.
S103, acquiring a second model type list corresponding to the initial data statistical model name list according to the initial data statistical model name list, wherein the second model type list comprises a plurality of second model types, and the second model types are model types of initial data statistical models corresponding to the initial data statistical model names.
Specifically, model types of the initial data statistical model are stored in the system.
S105, performing de-duplication processing on the second model type list to obtain a first model type list, wherein any de-duplication method known to those skilled in the art belongs to the protection scope of the present invention, and is not described herein.
According to the method, the second model type list is obtained through the initial data statistical model names in the initial text generation template, the duplicate removal processing is carried out on the second model type list, the first model type list is obtained, the first model type tree is further generated, the target statistical data set is obtained from the root to the leaf according to the sequence from the key data statistical model corresponding to the root node of the first model type tree, the repeated processing of the initial data statistical model of the same model type can be avoided, the repeated calculation and the resource waste are avoided, and the operation efficiency of the system is improved.
S200, according to the A and the first model type list, a first model type tree list C= { C corresponding to the first model type list is obtained 1 ,C 2 ,……,C r ,……,C s },C r ={C r1 ,C r2 ,……,C rg ,……,C rh },C rg ={C 1 rg ,C 2 rg ,……,C x rg ,……,C p rg },C x rg For the r first model type tree C r The x first model type node of the g layer, r=1, 2, … …, s, s the number of first model type trees, g=1, 2, … …, h, h is the number of layers of the first model type tree, x=1, 2, … …, p, p is the number of first model type nodes of one layer of the first model type tree. For example: if the node of the first layer of the preset model type tree, namely the root node, is of a text type, and the node of the child node of the root node, namely the node of the second layer, is: the nodes of the third layer are as follows: the child nodes of the plain text type are: a traditional Chinese character type and a simplified Chinese character type; the child nodes of the plain digital text type are: decimal type, integer type, fractional type; the child nodes of the pure English text type are: capital letter types, lowercase letter types; the child nodes of the mixed text type are: chinese-English mixing typeText and number combination types; if the first model type is: words, english, mixed, traditional, simplified, lower case letters, numbers and Chinese, then 3 first model type trees can be obtained, and the root nodes are respectively: text, english, mix; the child nodes of the characters are: complex and simplified; the subnodes in English are: lower case letters; the mixed child nodes are: numbers and chinese.
Specifically, S200 includes the following steps:
s201, acquiring a second model type tree according to the A and the first model type list, wherein the second model type tree comprises a plurality of second model type nodes, and a person skilled in the art knows that the mode of acquiring the second model type tree according to the preset model type tree and the first model type list is the same as the mode of acquiring a third model type tree according to the preset model type tree list and the candidate word type list in S131-S137, and the description is omitted herein; the second model type tree may be understood as a third model type tree obtained in S131-S137 when the candidate word type in S131-S137 is replaced with the first model type.
S203, deleting the second model type node which is NULL in the second model type tree to acquire C.
According to the type similarity between the preset model type and the first model type, the second model type tree is determined, all empty nodes in the second model tree are deleted, the first model type tree can be accurately acquired, the target statistical data set is acquired from the root to the leaf according to the sequence from the key data statistical model corresponding to the root node of the first model type tree, repeated calculation can be avoided, resource waste is avoided, and the running efficiency of the system is improved.
S300, obtaining C x rg Corresponding key data statistical model list D x rg ={D x1 rg ,D x2 rg ,……,D xy rg ,……,D xq rg },D xy rg Is C x rg The corresponding y-th key data statistical model, y=1, 2, … …, q, q is the first modelThe number of the key data statistical models corresponding to the type nodes is the number of the key data statistical models, and the key data statistical models are initial data statistical models with the same model type as that presented by the first model type node.
Specifically, each first model type node corresponds to q key data statistical models, q varies with x, r, g, for example: if the first model type node is a traditional Chinese, if the model types of the traditional Chinese text name statistical model, the traditional Chinese text quantity statistical model and the traditional Chinese text publication time statistical model are all traditional Chinese, the quantity of the key data statistical models corresponding to the first model type node is 3 and the key data statistical model is: a traditional text name statistical model, a traditional text quantity statistical model and a traditional text publishing time statistical model; if the first model type node is a simple body, if the model types of the simple text quantity statistical model and the simple text publication time statistical model are both simple bodies, the quantity of the key data statistical models corresponding to the first model type node is 2, and the key data statistical models are: the simplified text quantity statistical model and the simplified text publishing time statistical model.
S400, when g=1, input the initial dataset to D xy rg In order to obtain D xy rg Corresponding target statistics set G xy rg Wherein the target statistics set comprises a plurality of pieces of target statistics.
In particular, the initial data set includes all data for data statistics and the initial data set is stored in the system.
Further, the initial data set includes a plurality of pieces of initial data.
S500, when g is not equal to 1, C is r(g-1) As C in x rg C of parent node of (C) x r(g-1) All corresponding D xy r(g-1) Corresponding G xy r(g-1) Target statistics as C x rg Corresponding intermediate statistics to obtain C x rg Corresponding intermediate statistical data set H x rg And S600 is performed.
In particular, the intermediate statistics set comprises several pieces of intermediate statistics.
S600, H x rg Input to D xy rg In order to obtain D xy rg Corresponding target statistics set G xy rg
And when the first model type node corresponding to the key data statistical model is not the root node, acquiring the target statistical data set from the target statistical data set corresponding to all the key data statistical models corresponding to the father node of the corresponding first model type node, avoiding repeated calculation, being beneficial to avoiding resource waste and improving the operation efficiency of the system.
While certain specific embodiments of the invention have been described in detail by way of example, it will be appreciated by those skilled in the art that the above examples are for illustration only and are not intended to limit the scope of the invention. Those skilled in the art will also appreciate that many modifications may be made to the embodiments without departing from the scope and spirit of the invention.

Claims (8)

1. A data processing system for obtaining a text generation template, the system comprising: key text T corresponding to target user 0 The method comprises the following steps of:
s1, according to T 0 And A, obtain T 0 Corresponding candidate template text list t= { T 1 ,T 2 ,……,T b ,……,T c },T b B=1, 2, … …, c, c is the number of candidate templates, and the candidate template text is a text for describing the application range of the candidate template;
s2, training an NLP model according to the T and a plurality of first preset sample texts to obtain a target NLP model, wherein the result of the output of the NLP model is that the first preset sample texts are respectively matched with the T 1 ,T 2 ,……,T b ,……,T c Confidence between;
s3, T is taken as 0 Input into target NLP model to obtain T 0 Corresponding intermediate confidence list T 1 ={T 1 1 ,T 1 2 ,……,T 1 b ,……,T 1 c },T 1 b Is T 0 And T b An intermediate confidence level between;
s4, when AT/BT is more than or equal to YT, T is processed according to a first processing method 0 T and T 1 Processing to obtain an intermediate text generation template, wherein AT is T 0 The sum of the number of characters of all corresponding keywords, BT is T 0 The first processing method in S4 includes the following steps:
s41, acquiring a first keyword list GJ= { GJ 1 ,GJ 2 ,……,GJ (ai) ,……,GJ (am) },GJ (ai) For the ai first keyword, ai=1, 2, … …, am, am is the number of first keywords, and the first keyword is T 0 Corresponding keywords;
s42, acquiring a second keyword list GJ corresponding to the T 0 ={GJ 0 1 ,GJ 0 2 ,……,GJ 0 b ,……,GJ 0 c },GJ 0 b ={GJ 0 b1 ,GJ 0 b2 ,……,GJ 0 b(aj) ,……,GJ 0 b(an) },GJ 0 b(aj) Is T b Corresponding second keyword list GJ 0 b The aj second keywords in the list of the second keywords, aj=1, 2, … …, an, an are the number of the second keywords in the list of the second keywords, and the second keywords are keywords in the candidate template text;
s43, according to GJ and GJ 0 Acquiring a list similarity list GJ corresponding to the GJ 1 ={GJ 1 1 ,GJ 1 2 ,……,GJ 1 b ,……,GJ 1 c },GJ 1 b Is GJ and GJ 0 b List similarity between, wherein GJ 1 b Meets the following conditions:
GJ 1 b =Σ am ai=1an aj=1 XS (ai) (aj) /an)/am,XS (ai) (aj) for GJ (ai) With GJ 0 b(aj) Word similarity between;
s44 according to T 1 And GJ 1 Acquiring a first priority list YX= { YX corresponding to T 1 ,YX 2 ,……,YX b ,……,YX c },YX b Is T b Corresponding first priority, YX b Meets the following conditions:
YX b =T 1 b +GJ 1 b
s45, when there is only one maximum value in YX, determining max (YX 1 ,YX 2 ,……,YX b ,……,YX c ) Corresponding T b Generating templates for the intermediate texts by using the corresponding candidate templates;
s46, when there are a plurality of maximum values in YX, according to T 0 And T b Determining an intermediate text generation template;
s5, when AT/BT is less than YT, acquiring an intermediate text generation template according to a second processing method;
s6, the target user adjusts the intermediate text generation template to obtain a target text generation template.
2. The data processing system for obtaining a text generation template according to claim 1, wherein S1 comprises the steps of obtaining a candidate template:
s11, T is taken 0 Inputting into a preset keyword extraction model to obtain T 0 The corresponding keyword information list comprises a plurality of keyword information, wherein the keyword information comprises: keywords, keyword types;
s12, acquiring a candidate word type list B= { B according to the keyword information list 1 ,B 2 ,……,B e ,……,B f },B e For the e candidate word type, e=1, 2, … …, f, f is the number of candidate word types, and the candidate word type is the keyword type obtained by de-duplicating the keyword types in all the keyword information;
s13, acquiring a third model type tree according to A and B, wherein A= { A 1 ,A 2 ,……,A i ,……,A m },A i ={A i1 ,A i2 ,……,A ij ,……,A in(i) },A ij For the j-th preset model type node in the i-th layer of the preset model type tree, i=1, 2, … …, m, m are the number of layers of the preset model type tree, j=1, 2, … …, n (i), n (i) are the number of preset model type nodes in the i-th layer of the preset model type tree, and each preset model type node corresponds to a plurality of appointed text generation templates;
s14, determining the level of the nearest common ancestor ZX of all non-NULL nodes in the third model type tree and the positions of the nearest common ancestor ZX in the level;
s15, acquiring a first candidate model type node from the A, wherein the first candidate model type node is a level in the A and a preset model type node with the same position in the level as the level of ZX in a third model type tree and the position in the level;
s16, acquiring a second candidate model type node list according to the first candidate model type node, wherein the second candidate model type node list comprises a plurality of second candidate model type nodes, and the second candidate model type nodes are preset model type nodes which have the same father node as the first candidate model type nodes;
s17, obtaining a candidate template, wherein the candidate template is any appointed text generation template corresponding to any second candidate model type node.
3. The data processing system for retrieving a text generation template of claim 2, wherein the candidate template text corresponding to the candidate template is stored in the system.
4. The data processing system for obtaining a text generation template according to claim 2, wherein S13 comprises the steps of obtaining a third model type tree:
s131, let all A ij NULL to obtain a first specified model type tree corresponding to A 1 ={A 1 1 ,A 1 2 ,……,A 1 i ,……,A 1 m },A 1 i ={A 1 i1 ,A 1 i2 ,……,A 1 ij ,……,A 1 in(i) },A 1 ij Is A ij Corresponding designated model type nodes;
s132, obtaining A ij Corresponding preset model type A 0 ij The preset model type is the model type presented by the preset model type node;
s133, obtaining A 0 ij And B is connected with e Type similarity AB between 0e ij The type similarity is the similarity between the preset model type and the candidate word type;
s134, when B e Any one of corresponding AB 0e ij ≥A 2 When max (AB 0e 11 ,AB 0e 12 ,……,AB 0e 1j ,……,AB 0e 1n(i) ,AB 0e 21 ,AB 0e 22 ,……,AB 0e 2j ,……,AB 0e 2n(i) ,……,AB 0e i1 ,AB 0e i2 ,……,AB 0e ij ,……,AB 0e in(i) ,……,AB 0e m1 ,AB 0e m2 ,……,AB 0e mj ,……,AB 0e mn(i) ) Corresponding A 0 ij Corresponding A ij Corresponding A 1 ij Replaced by B e To obtain a second specified model tree, wherein max () is a maximum value obtaining function, A 2 A similarity threshold is preset;
s135, when B e All corresponding AB 0e ij <A 2 At the time, obtain A ij And B is connected with e Similarity priority between YX e ij Wherein YX e ij Meets the following conditions:
YX e ij =FJ e ij +ZJ e ij ,FJ e ij is A ij The preset model type corresponding to the father node of (B) e Type similarity between ZJ e ij Is A ij The type of the preset model corresponding to the child node of (B) e Type similarity between, when A ij FJ without parent node e ij 0, when A ij ZJ without child node e ij Is 0;
s136, max (YX e 11 ,YX e 12 ,……,YX e 1j ,……,YX e 1n(i) ,YX e 21 ,YX e 22 ,……,YX e 2j ,……,YX e 2n(i) ,……,YX e i1 ,YX e i2 ,……,YX e ij ,……,YX e in(i) ,……,YX e m1 ,YX e m2 ,……,YX e mj ,……,YX e mn(i) ) Corresponding A ij Corresponding A 1 ij Replaced by B e To obtain a third specified model tree;
s137, when the j-th second designated model node of the i-th layer of the second designated model tree is NULL and the j-th third designated model node of the i-th layer of the third designated model tree is not NULL, replacing the j-th second designated model node of the i-th layer of the second designated model tree with the j-th third designated model node of the i-th layer of the third designated model tree to obtain a third model type tree.
5. The data processing system for obtaining a text generation template according to claim 1, wherein S46 comprises the steps of:
s461, T 0 Inputting into a preset vector extraction model to obtain T 0 Corresponding first text vector list U 0 ={U 0 1 ,U 0 2 ,……,U 0 v ,……,U 0 w },U 0 v Is T 0 The corresponding v first text vector, v=1, 2, … …, w, w being the number of first text vectors;
s462, will T b Inputting into a preset vector extraction model to obtain T b Corresponding second text vector list U b ={U b1 ,U b2 ,……,U bv ,……,U bw },U bv Is T b A corresponding v-th second text vector;
s463 according to U 0 And U b Obtaining T 0 Respectively sum to T 1 ,T 2 ,……,T b ,……,T c Text similarity L between 0 1 ,L 0 2 ,……,L 0 b ,……,L 0 c Wherein L is 0 b Is T 0 And T b Text similarity between, and L 0 b Meets the following conditions:
s464 root and rootAccording to T 1 b And L 0 b Acquiring a second priority list YX corresponding to T 0 ={YX 0 1 ,YX 0 2 ,……,YX 0 b ,……,YX 0 c },YX 0 b Is T b Corresponding second priority, YX 0 b Meets the following conditions:
YX 0 b =T 1 b +L 0 b
s465, YX 0 When there is only one maximum value, max (YX 0 1 ,YX 0 2 ,……,YX 0 b ,……,YX 0 c ) Corresponding T b Generating templates for the intermediate texts by using the corresponding candidate templates;
s466, when YX 0 When there are a plurality of maximum values in (c), max (T 1 1 +GJ 1 1 +L 0 1 ,T 1 2 +GJ 1 2 +L 0 2 ,……,T 1 b +GJ 1 b +L 0 b ,……,T 1 c +GJ 1 c +L 0 c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
6. The data processing system for obtaining a text generation template according to claim 5, wherein the second processing method in S5 comprises the steps of:
s51, when YX 0 When there is only one maximum value, max (YX 0 1 ,YX 0 2 ,……,YX 0 b ,……,YX 0 c ) Corresponding T b Generating templates for the intermediate texts by using the corresponding candidate templates;
s53, when YX 0 When there are a plurality of maximum values in (c), max (T 1 1 +GJ 1 1 +L 0 1 ,T 1 2 +GJ 1 2 +L 0 2 ,……,T 1 b +GJ 1 b +L 0 b ,……,T 1 c +GJ 1 c +L 0 c ) Corresponding T b The corresponding candidate templates are templates for the intermediate text generation.
7. The data processing system for retrieving a text generation template as recited in claim 4, wherein a 2 The value range of (5) is [0.8,1 ]]。
8. The data processing system for obtaining a text generation template according to claim 1, wherein the target text generation template is used as an initial text generation template and after S6 comprises the steps of obtaining a target statistical data set:
s100, acquiring a first model type list according to an initial text generation template, wherein the first model type list comprises a plurality of first model types, and the first model types are model types of initial data statistical models in the initial text generation template;
s200, according to the A and the first model type list, a first model type tree list C= { C corresponding to the first model type list is obtained 1 ,C 2 ,……,C r ,……,C s },C r ={C r1 ,C r2 ,……,C rg ,……,C rh },C rg ={C 1 rg ,C 2 rg ,……,C x rg ,……,C p rg },C x rg For the r first model type tree C r The x first model type node of the g layer, r=1, 2, … …, s, s the number of first model type trees, g=1, 2, … …, h, h is the number of layers of the first model type tree, x=1, 2, … …, p, p is the number of first model type nodes of one layer of the first model type tree;
s300, obtaining C x rg Corresponding key data statistical model list D x rg ={D x1 rg ,D x2 rg ,……,D xy rg ,……,D xq rg },D xy rg Is C x rg The corresponding y-th key data statistical model is that y=1, 2, … …, q and q are the number of the key data statistical models corresponding to the first model type node, and the key data statistical model is an initial data statistical model with the same model type as that presented by the first model type node;
s400, when g=1, input the initial dataset to D xy rg In order to obtain D xy rg Corresponding target statistics set G xy rg Wherein the initial data set comprises all data for data statistics and the initial data set is stored in the system, and the target statistics set comprises a plurality of pieces of target statistics;
s500, when g is not equal to 1, C is r(g-1) As C in x rg C of parent node of (C) x r(g-1) All corresponding D xy r(g-1) Corresponding G xy r(g-1) Target statistics as C x rg Corresponding intermediate statistics to obtain C x rg Corresponding intermediate statistical data set H x rg And performs S600;
s600, H x rg Input to D xy rg In order to obtain D xy rg Corresponding target statistics set G xy rg
CN202311308051.6A 2023-10-10 2023-10-10 Data processing system for acquiring text generation template Active CN117332768B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311308051.6A CN117332768B (en) 2023-10-10 2023-10-10 Data processing system for acquiring text generation template

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311308051.6A CN117332768B (en) 2023-10-10 2023-10-10 Data processing system for acquiring text generation template

Publications (2)

Publication Number Publication Date
CN117332768A true CN117332768A (en) 2024-01-02
CN117332768B CN117332768B (en) 2024-03-08

Family

ID=89275136

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311308051.6A Active CN117332768B (en) 2023-10-10 2023-10-10 Data processing system for acquiring text generation template

Country Status (1)

Country Link
CN (1) CN117332768B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160124933A1 (en) * 2014-10-30 2016-05-05 International Business Machines Corporation Generation apparatus, generation method, and program
CN116561388A (en) * 2023-05-09 2023-08-08 每日互动股份有限公司 Data processing system for acquiring labels

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160124933A1 (en) * 2014-10-30 2016-05-05 International Business Machines Corporation Generation apparatus, generation method, and program
CN116561388A (en) * 2023-05-09 2023-08-08 每日互动股份有限公司 Data processing system for acquiring labels

Also Published As

Publication number Publication date
CN117332768B (en) 2024-03-08

Similar Documents

Publication Publication Date Title
US11636264B2 (en) Stylistic text rewriting for a target author
CN111209738B (en) Multi-task named entity recognition method combining text classification
WO2020063092A1 (en) Knowledge graph processing method and apparatus
CN112328742B (en) Training method and device based on artificial intelligence, computer equipment and storage medium
CN108491511B (en) Data mining method and device based on graph data and model training method and device
CN106168954B (en) A kind of negative information mode Method of Fuzzy Matching based on editing distance
CN103870001B (en) A kind of method and electronic device for generating candidates of input method
CN111340661B (en) Automatic application problem solving method based on graph neural network
WO2022048363A1 (en) Website classification method and apparatus, computer device, and storage medium
WO2020147409A1 (en) Text classification method and apparatus, computer device, and storage medium
CN113627797B (en) Method, device, computer equipment and storage medium for generating staff member portrait
WO2021159803A1 (en) Text summary generation method and apparatus, and computer device and readable storage medium
CN113051914A (en) Enterprise hidden label extraction method and device based on multi-feature dynamic portrait
CN112434131A (en) Text error detection method and device based on artificial intelligence, and computer equipment
CN113821635A (en) Text abstract generation method and system for financial field
US20180018321A1 (en) Avoiding sentiment model overfitting in a machine language model
CN112101042A (en) Text emotion recognition method and device, terminal device and storage medium
CN114372475A (en) Network public opinion emotion analysis method and system based on RoBERTA model
CN112418320A (en) Enterprise association relation identification method and device and storage medium
CN117312641A (en) Method, device, equipment and storage medium for intelligently acquiring information
Lee et al. Lossless online Bayesian bagging
CN117711600A (en) LLM model-based electronic medical record question-answering system
CN114373554A (en) Drug interaction relation extraction method using drug knowledge and syntactic dependency relation
CN113837307A (en) Data similarity calculation method and device, readable medium and electronic equipment
CN113988048A (en) Emotional cause pair extraction method based on multi-wheel machine reading understanding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant