CN113392196A - Topic retrieval method and system based on multi-mode cross comparison - Google Patents
Topic retrieval method and system based on multi-mode cross comparison Download PDFInfo
- Publication number
- CN113392196A CN113392196A CN202110622823.8A CN202110622823A CN113392196A CN 113392196 A CN113392196 A CN 113392196A CN 202110622823 A CN202110622823 A CN 202110622823A CN 113392196 A CN113392196 A CN 113392196A
- Authority
- CN
- China
- Prior art keywords
- topic
- similarity
- question
- representation
- text
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 238000004364 calculation method Methods 0.000 claims abstract description 27
- 238000007405 data analysis Methods 0.000 claims abstract description 11
- 239000013598 vector Substances 0.000 claims description 18
- 238000007781 pre-processing Methods 0.000 abstract description 3
- 238000012545 processing Methods 0.000 abstract description 2
- 244000025254 Cannabis sativa Species 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000013473 artificial intelligence Methods 0.000 description 3
- 239000011248 coating agent Substances 0.000 description 3
- 238000000576 coating method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 238000012549 training Methods 0.000 description 3
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000004140 cleaning Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 241000282414 Homo sapiens Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007619 statistical method Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation or dialogue systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Human Computer Interaction (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computational Linguistics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention provides a topic retrieval system and method based on multi-mode cross comparison, wherein the system comprises: the system comprises a question data analysis module, a question similarity calculation module and a result output module; the system comprises a question data analysis module, a question information preprocessing module and a question data processing module, wherein the question data analysis module is used for receiving question information input by a user, and carrying out preprocessing and structured arrangement; the question similarity calculation module is used for alternately calculating the similarity between the questions input by the user and the text representation and the picture representation of the questions in the question bank and calculating the comprehensive similarity of the questions and the picture representation in the question bank in a weighting manner; and the result output module is used for returning relevant information such as questions, answers and the like in the question bank with the comprehensive similarity larger than the preset subject threshold to the user. By the system, the retrieval result of each subject in the subject library can be more accurate.
Description
Technical Field
The invention relates to the technical field of computers, in particular to a topic retrieval method and system based on multi-mode cross comparison.
Background
In recent years, with the development of internet and artificial intelligence technology, topic question-answering systems are greatly developed, and great help is provided for personalized teaching. It is becoming more and more important how to quickly and accurately retrieve the same or similar questions as the user input questions and then give answers in the question bank according to the user's questions.
Currently, the implementation of the topic retrieval system is generally implemented by comparing the text similarity between topics, a user transmits a text for describing topic information to the topic retrieval system, and the topic retrieval system compares the text similarity between the topic text input by the user and the topic text in the topic library, and then selects the topic with the highest similarity as a retrieval result to return to the user. If the topic information input by the user is in the form of a picture, comparing the similarity of the topics by comparing the similarity of the pictures.
The current text similarity calculation methods mainly fall into two categories. The text similarity calculation method based on characters and the text similarity calculation method based on vector space are respectively provided.
The text similarity calculation method based on characters, such as the traditional methods of editing distance, hamming distance, Jaccard, LCS and the like, evaluates the text similarity by directly comparing the same characters and their sequence relationship between two text strings.
Vector space-based text similarity calculation methods such as TF-IDF and BM25, and calculation of similarity between texts by cosine similarity after text is subjected to vector representation and direct comparison of similarity between texts by a neural network.
With the development of multimedia, users describe the title information in a text-only manner, and more situations are currently available in which the title information is described by text and combined with pictures.
Currently, the mainstream topic retrieval service and system in the market only support a retrieval mode facing text topics or picture topics, such as "easy search topic" (address: https:// www.xuesai.cn/souti /) ape search topic, and the like, as shown in fig. 1 and fig. 2.
The prior art is mainly concerned with the following problems, which are illustrated and described below.
(1) Only text content is input, and users often have situations where the words are not known or cannot be expressed clearly for subjects in the form of mathematics, physics, or speech-through-the-picture.
(2) Only picture content is input, and although correct answers can be returned, the answers may not meet the requirements of the user, for example, as shown in fig. 2, the user may not be familiar with the triangle rule of vector addition in the answers, and thus, even if the answers are seen, it may not be clear how to do the questions. Thus, if a user is allowed to input his or her needs, the user can be better assisted with a text.
(3) There is also a question-and-answer system that lets a user input a topic text portion and a topic picture portion at the same time, but it recognizes the topic picture portion as a text representation by a picture text recognition technology, and then directly splices the topic text portion and the recognized picture text content and compares the text similarity between topics. Due to the defects of the picture text recognition technology, the problem content recognition is wrong, for example, the same problem is different due to different light rays, angles and the like, and the finally recognized results are different, so that the original same problem is judged to be dissimilar due to the wrong picture text recognition.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a topic retrieval method and a system based on multi-mode cross comparison, which are characterized in that (1) input topics of a user are structurally arranged through a topic data analysis module, and (2) the similarity between the input topics of the user and candidate topics is calculated one by one through a topic similarity calculation module, and the information of the result is returned to the user.
In order to achieve the above object, the present invention is achieved by the following technical solutions.
According to an aspect of the present invention, a topic retrieval system based on multi-modal cross-comparison is provided, which includes: the system comprises a question data analysis module, a question similarity calculation module and a structure output module; wherein,
the question data analysis module is used for receiving the question information input by the user and carrying out pretreatment;
the question similarity calculation module is used for calculating the similarity between the questions input by the user and the questions in the question bank;
and the result output module is used for returning the questions in the question bank with the similarity larger than a preset subject threshold value to the user.
Further, in the topic similarity calculation module, the method includes:
a. splicing the text after the title is cleaned and the text after the title content is cleaned to be used as text representation, and using the title picture text identification content as picture representation;
b. the text representation and the picture representation of the user input questions are represented by T1 and P1, the text representation and the picture representation of the questions in the question bank are represented by T2 and P2, the similarity between T1 and T2, between T1 and P2, between P1 and T2 and between P1 and P2 are calculated and are respectively represented by S1, S2, S3 and S4;
c. and calculating the comprehensive similarity s.
Further, the similarity S1 is calculated by using a Jaccard method; calculating similarities S2, S3, and S4 by cosine similarity; preferably, the topic picture text recognition is converted into a vector representation through a BERT model, the topic picture is converted into a vector through a LeNet convolutional network model, and then the two vectors are spliced to be used as a vectorized representation of the picture representation.
Further, the calculation formula of the comprehensive similarity s is as follows:
w is the subject weight.
Further, the subject weight is:
subject of the scientific discipline | Weight (w1, w2, w3, w4) | Subject of the scientific discipline | Weight (w1, w2, w3, w4) |
Chinese language | 5.5,2,2,0.5 | Physics of physics | 5,1,1,3 |
History of | 5,2,2,1 | Mathematics, and | 4,2,2,2 |
|
5,2,2,1 | Chemistry | 5,1,1,3 |
|
4,2,2,2 | |
4,2,2,2 |
Further, in the result output module, if the comprehensive similarity is greater than the subject threshold corresponding to the subject input by the user, the problem in the subject library is taken as a candidate problem.
Further, the subject threshold is:
subject of the scientific discipline | Threshold value | Subject of the scientific discipline | Threshold value |
Chinese language | 0.8 | Physics of physics | 0.7 |
History of | 0.8 | Mathematics, and | 0.5 |
geography | 0.7 | Chemistry | 0.5 |
Politics | 0.6 | Biological organisms | 0.5 |
According to another aspect of the present invention, a topic retrieval method based on multi-modal cross-comparison is provided, which includes:
step 1, receiving question information input by a user;
and 3, returning the questions in the question bank with the similarity larger than the subject threshold value to the user as candidate questions.
Further, the step 2 comprises:
a. acquiring text representation and picture representation of title information;
b. representing the text representation and the picture representation of the title by T1 and P1, representing the text representation and the picture representation of the title in the title library by T2 and P2, and then cross-comparing the similarity of four parts of content S1, S2, S3 and S4; wherein, S1 is calculated by adopting a Jaccard method; a cosine similarity calculation method is adopted when S2, S3 and S4 are calculated;
c. and calculating the comprehensive similarity s.
Further, in the step 3, the questions in the question bank with the comprehensive similarity greater than the subject threshold corresponding to the user input question are taken as candidate questions.
Compared with the prior art, the invention has the beneficial effects that:
(1) according to the invention, for topic retrieval, a cross comparison method based on multiple modes is adopted, and compared with the method of directly comparing the similarity of topic texts and comparing the similarity of topic pictures, the accuracy of topic comparison is greatly improved.
(2) When the comprehensive similarity calculation of the topics is carried out on the four parts of similarity of the cross comparison, different weight parameter combinations are set based on experience according to different subjects, and the accuracy of topic comparison is improved for the topic calculation of different subjects.
(3) According to the invention, the comprehensive similarity result of the topic comparison is compared with the preset threshold values corresponding to different subjects for judging the topic similarity, so that the accuracy of the topic comparison is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a schematic interface diagram of a prior art question-answering system;
FIG. 2 is a schematic diagram of another interface of a prior art question-answering system;
FIG. 3 is a schematic diagram of a topic retrieval system according to one embodiment of the present invention;
FIG. 4 is a schematic flow chart of topic data parsing according to one embodiment of the invention;
FIG. 5 is a schematic diagram of a cross-comparison in accordance with one embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, a technical solution in an embodiment of the present invention will be described in detail and completely with reference to the accompanying drawings in the embodiment of the present invention, and it is obvious that the described embodiment is a part of embodiments of the present invention, but not all embodiments. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention.
The invention provides a question searching system based on multi-mode cross comparison, which can be applied to a question answering system, receives questions input by a user, searches in a question bank, and returns the same or similar questions or answers to the user so as to help the user to complete a learning task. As shown in fig. 3, the system includes the following modules:
(1) a topic data analysis module: receiving subject information input by a user, wherein the subject information comprises subject information of subjects, subject titles, subject descriptions, subject pictures and subject type information. The topic data analysis module can be arranged on a front-end webpage or a mobile-end APP.
After receiving the topic data, the topic data analysis module firstly divides the topic data into topic text data and topic picture data, and respectively preprocesses the topic text data and the topic picture data. The topic data parsing module is shown in FIG. 4.
The preprocessing of the subject text data comprises text cleaning and word segmentation. The text cleaning work comprises unified coding processing, illegal character removal, emoji symbol removal, character symbols, HTML labels and other symbols, and invalid character removal.
The invalid characters are characters which are irrelevant to subject contents and are sorted after analyzing, counting and finishing subject data in a subject database, and the invalid characters mainly comprise the following characters:
thank you, solve online, ask, teacher's help, read, ask for help, do not, solve, each location, help, hard, you, trouble, explain, help me to say, give a message, want to help me to solve, question how to do, ask how to answer, help me, check, give a message, not understand, not clear, explain, help me to see, baituo, answer, hope teacher's help to see, i's go, do not know, how to write, how to explain, how to ask, how to list, how to do, detail, not give, do, not understand, not read, help analyze, ask, give a question, give a message to me, give a message, how to give a message, do not especially, not know, how to do, do not think, do, not think, give a message, thought, not know, point, give a message to do, do not know, help, do not particularly, do not understand, do not know, do not like, give a message, or a message, give a message, or a message, instruction, wish details, model, volume, middle exam, end of term, notice, version.
The topic picture data is a problem picture, picture text identification needs to be carried out on the problem picture, topic information in the picture is identified as text information through the existing picture identification service, and then the data are subjected to structured arrangement.
The topic text and the topic picture data are combined into topic data after being preprocessed, and the structure of the topic data is shown as follows, wherein picture information stored in the topic data is a picture address.
{
"query _ id": subject data unique identifier,
"query _ title _ original": title header original text information,
"query content original": title content original text information,
"query _ type": title type,
"query _ create _ time" title creation time,
"query _ subject" subject information,
"query _ pic" title picture information,
"query _ pic _ content": title picture text identification content,
"query _ title _ clean": the subject title cleared text,
"query _ content _ clean": title content clean-up text
}
Example (c):
{
"question_id":00063274-1008-4525-9b7d-9f297f72bde1,
"query _ title _ original" please help the teacher to see how do it about the title of "bad house name? ,
"query _ content _ original" analyzes the material in the question, asking for "mountain is not high, there is a immortal name. "what is the idea the author wants to express? ,
"query _ type" is a question and answer question,
"question_create_time":2019-06-27 20:58:28,
"query _ subject" in the language,
"question_pic":https://cs.101.com/v0.1/download/actions/direct?dentryId=9e563a62-a96a-49e2-a238-a810b550b001&serviceName=fep,
"query _ pic _ content": crude Chamber Ming ": mountain is not high, so it is named after immortal. If water is not deep, Longshiling can be used. The gaussian is a poor room, but wu De Xin. The green tongue coating and the green grass color. There are julians in the conversation, and there are no white dices in the past.
"query title clean" in relation to "hovel inscription",
"query _ content _ clean" analyzes the material in the question, asking for "mountain is not high, there is a immortal name. "what the author wants to express is
}
(2) And the question similarity calculation module is used for comparing the question data with the question data in the question bank one by one and returning the same and similar question information to the user. As shown in fig. 5, the specific comparison method is as follows:
a. and splicing the text after the topic titles are cleaned and the text after the topic contents are cleaned to be used as text representation, and using the topic picture text identification contents and the topic pictures as picture representation together.
b. The text representation and the picture representation of title 1 are represented by T1 and P1, the text representation and the picture representation of title 2 are represented by T2 and P2, and then the similarities of four parts of content, namely (T1, T2), (T1, P2), (P1, T2), (P1, P2), are cross-compared and represented by S1, S2, S3 and S4 respectively.
When the similarity S1 is calculated, namely the similarity between the text representations of the comparison topics, the Jaccard method is adopted for calculation, the text representation to be compared is subjected to word segmentation, stop words and punctuation marks are removed, the similarity is calculated by comparing the intersection and union of the words contained in the text, and a specific calculation formula is shown as follows.
Wherein J (A, B) is epsilon [0, 1 ].
When calculating the similarity S2 and S3, that is, the similarity between the text representation and the picture representation, vectorizing the text representation by a BERT model, and then vectorizing the picture representation, specifically: the text recognition of the title picture is also converted into vector representation through a BERT model, the title picture is converted into vectors through a LeNet convolution network model, then the two vectors are spliced to be used as vectorization representation of picture representation, and then the similarity between the vectors is calculated through cosine similarity. The cosine similarity calculation formula is shown below.
When calculating the similarity S4, that is, comparing the similarity represented by the pictures, the topic pictures are simultaneously vectorized and then the cosine similarity is calculated.
After the similarity result of the cross comparison is obtained, different weights Wi are given to different similarity comparison values according to the prior experience after the prior statistical analysis, and the comprehensive similarity of the two questions to be compared is represented as follows:
for each subject, the corresponding weights are set as follows (subject weights can be set empirically and can also be obtained by training of neural networks):
table 1 weight setting table
Subject of the scientific discipline | Weight (w1, w2, w3, w4) | Subject of the scientific discipline | Weight (w1, w2, w3, w4) |
Chinese language | 5.5,2,2,0.5 | Physics of physics | 5,1,1,3 |
History of | 5,2,2,1 | Mathematics, and | 4,2,2,2 |
|
5,2,2,1 | Chemistry | 5,1,1,3 |
|
4,2,2,2 | |
4,2,2,2 |
The method has the advantages that the comparison accuracy can be improved, compared with the original method, the method for cross comparison separately compares the text content of the theme with the text content of the picture, and the theme content is analyzed in a more detailed mode during comparison.
As exemplified by the following examples:
topic 1:
topic content: the question was asked about "mountain is not high, but has a name of immortal rule" to analyze the material in the question. "what is the idea the author wants to express?
Topic picture text content: "Houchan Ming" (Lou Ming-Shi): mountain is not high, so it is named after immortal. If water is not deep, Longshiling can be used. The gaussian is a poor room, but wu De Xin. The green tongue coating and the green grass color. There are julians in the conversation, and there are no white dices in the past.
Topic 2:
topic content: ask "the green upper part of the tongue mark and the green color of grass. What is it?
Topic picture text content: "Houchan Ming" (Lou Ming-Shi): mountain is not high, so it is named after immortal. If water is not deep, Longshiling can be used. The gaussian is a poor room, but wu De Xin. The green tongue coating and the green grass color.
If the topic text content and the topic picture content are directly spliced, then the similarity between the topic content of topic 1 and the topic content of topic 2 is calculated to be 0.68 by using Jaccard.
After the cross-comparison, the similarity between the text content of topic 1 and the text content of topic 2 is 0.05, the similarity between the picture content of topic 1 and the text content of topic 2 is 0.14, the similarity between the text content of topic 1 and the picture content of topic 2 is 0.19, the similarity between the picture content of topic 1 and the picture content of topic 2 is 0.95, and the similarity between topic 1 and topic 2 is (5.5 + 0.05+2 + 0.14+2 + 0.19+0.5 + 0.95)/4 is 0.3 by comprehensive calculation based on the preset weight information (5.5, 2, 2, 0.5) of the linguistic subjects, which indicates that the similarity between the two is small or not similar. In fact, the two items can be judged to be dissimilar by human beings, so that the result 0.3 of the cross comparison can reflect the similarity of the items more than the result 0.68 of the previous method.
(3) And the result output module is used for comparing the calculated comprehensive similarity with a preset subject threshold, wherein different comparison thresholds are set for different subjects, as shown in table 2. And if the comprehensive similarity is greater than the subject threshold corresponding to the subject input by the user, considering that the problem in the subject library is similar to the problem input by the user and taking the problem as a candidate problem. Thresholds may also be obtained based on the data sets and artificial intelligence network training. After all questions in the question bank are compared, the candidate questions are ranked according to the magnitude of the comprehensive similarity, and the first five questions are selected and returned to the user; if the number of the candidate topics is less than five, all the candidate topics meeting the conditions are returned to the user; if not, prompting the user that the similar problem is not found, and issuing the problem input by the user as a new problem.
TABLE 2 threshold setting table
Subject of the scientific discipline | Threshold value | Subject of the scientific discipline | Threshold value |
Chinese language | 0.8 | Physics of physics | 0.7 |
History of | 0.8 | Mathematics, and | 0.5 |
geography | 0.7 | Chemistry | 0.5 |
Politics | 0.6 | Biological organisms | 0.5 |
According to another aspect of the present invention, a topic retrieval method based on multi-modal cross-comparison is provided, which includes:
step 1, receiving question information input by a user;
and 3, returning the questions in the question bank with the similarity larger than the subject threshold value to the user as candidate questions.
In step 1, after receiving the topic data, the topic data is divided into topic text data and topic picture data, and the topic text data and the topic picture data are preprocessed respectively. The specific process is described in a topic parsing data module.
In step 2, a, acquiring text representation and picture representation of user input title information; b. representing the text representation and the picture representation of the title by T1 and P1, representing the text representation and the picture representation of the title in the title library by T2 and P2, and then cross-comparing the similarity of four parts of content S1, S2, S3 and S4; wherein, S1 is calculated by adopting a Jaccard method; when calculating S2 and S3, firstly, vectorizing representation is carried out on text representation through a BERT model, then vectorizing representation is carried out on the subject picture, specifically, text identification of the subject picture is converted into vector representation through the BERT model, the subject picture is converted into vectors through a LeNet convolution network model, then the two vectors are spliced to be used as vectorized representation of picture representation, and then the similarity between the vectors is calculated through cosine similarity; when the similarity is calculated S4, calculating the cosine similarity of the vectorization expression of the topic picture and the topic pictures in the topic library; c. and calculating the comprehensive similarity s. The specific calculation method is as described above.
In step 3, the calculated similarity is compared with a preset subject threshold, wherein different comparison thresholds are set for different subjects, as shown in table 2. And if the similarity is greater than the subject threshold corresponding to the subject input by the user, considering that the problem in the subject library is similar to the problem input by the user and taking the problem as a candidate problem. Thresholds may also be obtained based on the data sets and artificial intelligence network training. After all questions in the question bank are compared, the candidate questions are sorted according to the similarity, and the first five questions are selected and returned to the user; if the number of the candidate topics is less than five, all the candidate topics meeting the conditions are returned to the user; if not, prompting the user that the similar problem is not found, and issuing the problem input by the user as a new problem.
The above examples are only for illustrating the technical solutions of the present invention, and not for limiting the same; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions.
Claims (10)
1. A topic retrieval system based on multimodal cross-comparisons, the system comprising: the system comprises a question data analysis module, a question similarity calculation module and a structure output module; wherein,
the question data analysis module is used for receiving the question information input by the user and carrying out pretreatment;
the question similarity calculation module is used for calculating the similarity between the questions input by the user and the questions in the question bank;
and the result output module is used for returning the questions in the question bank with the similarity larger than a preset subject threshold value to the user.
2. The topic retrieval system according to claim 1, wherein the topic similarity calculation module comprises:
a. splicing the text after the title is cleaned and the text after the title content is cleaned to be used as text representation, and using the title picture text identification content and the title picture as picture representation;
b. the text representation and the picture representation of the user input questions are represented by T1 and P1, the text representation and the picture representation of the questions in the question bank are represented by T2 and P2, the similarity between T1 and T2, between T1 and P2, between P1 and T2 and between P1 and P2 are calculated and are respectively represented by S1, S2, S3 and S4;
c. and calculating the comprehensive similarity s.
3. The title retrieval system of claim 2, wherein the similarity S1 is calculated by a Jaccard method; calculating similarities S2, S3, and S4 by cosine similarity; preferably, the topic picture text recognition is converted into a vector representation through a BERT model, the topic picture is converted into a vector through a LeNet convolutional network model, and then the two vectors are spliced to be used as a vectorized representation of the picture representation.
6. the topic retrieval system of claim 1, wherein in the result output module, if the integrated similarity is greater than a topic threshold corresponding to a topic input by a user, the question in the topic library is taken as a candidate question.
7. The topic retrieval system of claim 6, wherein the topic threshold is:
8. A topic retrieval method based on multi-modal cross comparison comprises the following steps:
step 1, receiving question information input by a user;
step 2, calculating the similarity between the received questions and the questions in the question bank;
and 3, returning the questions in the question bank with the similarity larger than the subject threshold value to the user as candidate questions.
9. The title retrieval method according to claim 1, wherein the step 2 comprises:
a. acquiring text representation and picture representation of title information;
b. representing the text representation and the picture representation of the title by T1 and P1, representing the text representation and the picture representation of the title in the title library by T2 and P2, and then cross-comparing the similarity of four parts of content S1, S2, S3 and S4; wherein, S1 is calculated by adopting a Jaccard method; a cosine similarity calculation method is adopted when S2, S3 and S4 are calculated;
c. and calculating the comprehensive similarity s.
10. The topic retrieval method according to claim 9, wherein in the step 3, topics in the topic database having a comprehensive similarity greater than a topic threshold corresponding to a user-input topic are used as candidate questions.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110622823.8A CN113392196B (en) | 2021-06-04 | 2021-06-04 | Question retrieval method and system based on multi-mode cross comparison |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110622823.8A CN113392196B (en) | 2021-06-04 | 2021-06-04 | Question retrieval method and system based on multi-mode cross comparison |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113392196A true CN113392196A (en) | 2021-09-14 |
CN113392196B CN113392196B (en) | 2023-04-21 |
Family
ID=77618168
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110622823.8A Active CN113392196B (en) | 2021-06-04 | 2021-06-04 | Question retrieval method and system based on multi-mode cross comparison |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113392196B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113920335A (en) * | 2021-09-28 | 2022-01-11 | 苏州冷王网络科技有限公司 | Deep learning-based image and text embedded drug label identification method |
CN114048354A (en) * | 2022-01-10 | 2022-02-15 | 广州启辰电子科技有限公司 | Test question retrieval method, device and medium based on multi-element characterization and metric learning |
Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104794169A (en) * | 2015-03-30 | 2015-07-22 | 明博教育科技有限公司 | Subject term extraction method and system based on sequence labeling model |
CN105760507A (en) * | 2016-02-23 | 2016-07-13 | 复旦大学 | Cross-modal subject correlation modeling method based on deep learning |
CN106886601A (en) * | 2017-03-02 | 2017-06-23 | 大连理工大学 | A kind of Cross-modality searching algorithm based on the study of subspace vehicle mixing |
CN107145571A (en) * | 2017-05-05 | 2017-09-08 | 广东艾檬电子科技有限公司 | A kind of searching method and device |
CN107562812A (en) * | 2017-08-11 | 2018-01-09 | 北京大学 | A kind of cross-module state similarity-based learning method based on the modeling of modality-specific semantic space |
CN108269110A (en) * | 2016-12-30 | 2018-07-10 | 华为技术有限公司 | Item recommendation method, system and user equipment based on community's question and answer |
CN109213853A (en) * | 2018-08-16 | 2019-01-15 | 昆明理工大学 | A kind of Chinese community's question and answer cross-module state search method based on CCA algorithm |
CN109558498A (en) * | 2018-11-07 | 2019-04-02 | 南京邮电大学 | Multi-modal hash method based on deep learning |
CN110059217A (en) * | 2019-04-29 | 2019-07-26 | 广西师范大学 | A kind of image text cross-media retrieval method of two-level network |
CN110196930A (en) * | 2019-05-22 | 2019-09-03 | 山东大学 | A kind of multi-modal customer service automatic reply method and system |
CN110287951A (en) * | 2019-06-21 | 2019-09-27 | 北京百度网讯科技有限公司 | A kind of method and device of Text region |
CN111753190A (en) * | 2020-05-29 | 2020-10-09 | 中山大学 | Meta learning-based unsupervised cross-modal Hash retrieval method |
CN112035669A (en) * | 2020-09-09 | 2020-12-04 | 中国科学技术大学 | Social media multi-modal rumor detection method based on propagation heterogeneous graph modeling |
CN112287130A (en) * | 2019-07-23 | 2021-01-29 | 小船出海教育科技(北京)有限公司 | Searching method, device and equipment for graphic questions |
US20210057058A1 (en) * | 2019-08-23 | 2021-02-25 | Alibaba Group Holding Limited | Data processing method, apparatus, and device |
US20210150151A1 (en) * | 2019-01-08 | 2021-05-20 | Institute Of Automation, Chinese Academy Of Sciences | Autonomous evolution intelligent dialogue method, system, and device based on a game with a physical environment |
CN112905827A (en) * | 2021-02-08 | 2021-06-04 | 中国科学技术大学 | Cross-modal image-text matching method and device and computer readable storage medium |
-
2021
- 2021-06-04 CN CN202110622823.8A patent/CN113392196B/en active Active
Patent Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104794169A (en) * | 2015-03-30 | 2015-07-22 | 明博教育科技有限公司 | Subject term extraction method and system based on sequence labeling model |
CN105760507A (en) * | 2016-02-23 | 2016-07-13 | 复旦大学 | Cross-modal subject correlation modeling method based on deep learning |
CN108269110A (en) * | 2016-12-30 | 2018-07-10 | 华为技术有限公司 | Item recommendation method, system and user equipment based on community's question and answer |
CN106886601A (en) * | 2017-03-02 | 2017-06-23 | 大连理工大学 | A kind of Cross-modality searching algorithm based on the study of subspace vehicle mixing |
CN107145571A (en) * | 2017-05-05 | 2017-09-08 | 广东艾檬电子科技有限公司 | A kind of searching method and device |
CN107562812A (en) * | 2017-08-11 | 2018-01-09 | 北京大学 | A kind of cross-module state similarity-based learning method based on the modeling of modality-specific semantic space |
CN109213853A (en) * | 2018-08-16 | 2019-01-15 | 昆明理工大学 | A kind of Chinese community's question and answer cross-module state search method based on CCA algorithm |
CN109558498A (en) * | 2018-11-07 | 2019-04-02 | 南京邮电大学 | Multi-modal hash method based on deep learning |
US20210150151A1 (en) * | 2019-01-08 | 2021-05-20 | Institute Of Automation, Chinese Academy Of Sciences | Autonomous evolution intelligent dialogue method, system, and device based on a game with a physical environment |
CN110059217A (en) * | 2019-04-29 | 2019-07-26 | 广西师范大学 | A kind of image text cross-media retrieval method of two-level network |
CN110196930A (en) * | 2019-05-22 | 2019-09-03 | 山东大学 | A kind of multi-modal customer service automatic reply method and system |
CN110287951A (en) * | 2019-06-21 | 2019-09-27 | 北京百度网讯科技有限公司 | A kind of method and device of Text region |
CN112287130A (en) * | 2019-07-23 | 2021-01-29 | 小船出海教育科技(北京)有限公司 | Searching method, device and equipment for graphic questions |
US20210057058A1 (en) * | 2019-08-23 | 2021-02-25 | Alibaba Group Holding Limited | Data processing method, apparatus, and device |
CN111753190A (en) * | 2020-05-29 | 2020-10-09 | 中山大学 | Meta learning-based unsupervised cross-modal Hash retrieval method |
CN112035669A (en) * | 2020-09-09 | 2020-12-04 | 中国科学技术大学 | Social media multi-modal rumor detection method based on propagation heterogeneous graph modeling |
CN112905827A (en) * | 2021-02-08 | 2021-06-04 | 中国科学技术大学 | Cross-modal image-text matching method and device and computer readable storage medium |
Non-Patent Citations (1)
Title |
---|
刘泳文: "基于图像识别的搜题系统的研究与实现" * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113920335A (en) * | 2021-09-28 | 2022-01-11 | 苏州冷王网络科技有限公司 | Deep learning-based image and text embedded drug label identification method |
CN114048354A (en) * | 2022-01-10 | 2022-02-15 | 广州启辰电子科技有限公司 | Test question retrieval method, device and medium based on multi-element characterization and metric learning |
Also Published As
Publication number | Publication date |
---|---|
CN113392196B (en) | 2023-04-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110298033B (en) | Keyword corpus labeling training extraction system | |
CN116166782A (en) | Intelligent question-answering method based on deep learning | |
CN115269857A (en) | Knowledge graph construction method and device based on document relation extraction | |
CN114416942A (en) | Automatic question-answering method based on deep learning | |
CN110119510B (en) | Relationship extraction method and device based on transfer dependency relationship and structure auxiliary word | |
CN111339269A (en) | Knowledge graph question-answer training and application service system with automatically generated template | |
CN114036281B (en) | Knowledge graph-based citrus control question-answering module construction method and question-answering system | |
CN113821605B (en) | Event extraction method | |
CN115564393A (en) | Recruitment requirement similarity-based job recommendation method | |
CN112328800A (en) | System and method for automatically generating programming specification question answers | |
CN113157885A (en) | Efficient intelligent question-answering system for knowledge in artificial intelligence field | |
CN113392196A (en) | Topic retrieval method and system based on multi-mode cross comparison | |
CN113505589B (en) | MOOC learner cognitive behavior recognition method based on BERT model | |
CN112349294B (en) | Voice processing method and device, computer readable medium and electronic equipment | |
CN112541337A (en) | Document template automatic generation method and system based on recurrent neural network language model | |
CN112364132A (en) | Similarity calculation model and system based on dependency syntax and method for building system | |
CN111782793A (en) | Intelligent customer service processing method, system and equipment | |
CN112579666A (en) | Intelligent question-answering system and method and related equipment | |
CN117609419A (en) | Domain retrieval method based on meta learning and knowledge enhancement | |
CN106897274B (en) | Cross-language comment replying method | |
CN111666374A (en) | Method for integrating additional knowledge information into deep language model | |
CN111460118A (en) | Artificial intelligence conflict semantic recognition method and device | |
CN112711666B (en) | Futures label extraction method and device | |
CN116522165B (en) | Public opinion text matching system and method based on twin structure | |
CN115774996B (en) | Intelligent interview topdressing problem generation method and device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |