CN111666882B - Method for extracting answers of handwriting test questions - Google Patents

Method for extracting answers of handwriting test questions Download PDF

Info

Publication number
CN111666882B
CN111666882B CN202010510103.8A CN202010510103A CN111666882B CN 111666882 B CN111666882 B CN 111666882B CN 202010510103 A CN202010510103 A CN 202010510103A CN 111666882 B CN111666882 B CN 111666882B
Authority
CN
China
Prior art keywords
graph
matching
picture
point
points
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010510103.8A
Other languages
Chinese (zh)
Other versions
CN111666882A (en
Inventor
王家奎
王旺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Veilytech Co ltd
Original Assignee
Wuhan Veilytech Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Veilytech Co ltd filed Critical Wuhan Veilytech Co ltd
Priority to CN202010510103.8A priority Critical patent/CN111666882B/en
Publication of CN111666882A publication Critical patent/CN111666882A/en
Application granted granted Critical
Publication of CN111666882B publication Critical patent/CN111666882B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/30Writer recognition; Reading and verifying signatures
    • G06V40/33Writer recognition; Reading and verifying signatures based only on signature image, e.g. static signature recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Abstract

The invention discloses a method for extracting answers of handwriting test questions, which comprises the steps that one picture is a source test question picture A, the other picture is an answer picture B to be extracted, and photographing and uploading are carried out; the method also comprises the following steps: s1: performing corresponding preprocessing, such as size scaling, on the graphs A and B, and extracting feature points; s2: performing primary matching on the characteristic points of the graphs A and B by using a coarse matching method, calculating a homography matrix from the graph B to the graph A, and correspondingly mapping the graph B into a graph C, wherein the sizes of the graph C and the graph A are consistent, and the angles are close; the invention has scientific and reasonable structure, safe and convenient use, improves the on-line examination question review function, reduces the problems of lower angle and flatness of the uploaded pictures of students, facilitates the extraction of answers, reduces the inconvenience brought by the extremely high precision requirement in the past, and is suitable for popularization and use.

Description

Method for extracting answers of handwriting test questions
Technical Field
The invention relates to the technical field of picture processing, in particular to a method for extracting answers of handwriting test questions.
Background
The examination is a strict knowledge level identification method, the learning ability and other abilities of students can be checked through the examination, in order to ensure the fairness and fairness of the results, the examination room is required to have strong discipline constraint, and is specially provided with a main examination, an invigilation and other examination supervising processes, any cheating actions are absolutely forbidden, otherwise legal and criminal responsibilities are about to be born, the examination is that a group of people with different educational resources can complete the same answer sheet within a certain time, however, the meaning of the examination is not limited to the same, and the examination can also be carried out in an omnibearing way on a person for one target, so that the examination is really that people with different social positions can have the opportunity to change themselves, and a test sheet can be realized, the examination questions of the test sheet are compared, and correct answers are adjusted;
with the continuous development of online education, the acceptance degree of online teaching is gradually increased, and with the increase of the number of students, the workload of teachers is increased, the online examination question review function can effectively lighten the work of teachers, more energy is put in teaching work and knowledge grasping conditions of study students, and the current online examination question review function is often insufficient in accuracy for acquiring examination questions uploaded by the students and extracting answer parts, or has higher requirements on the angle and flatness of uploading pictures by the students, so that the online examination question review function is inconvenient to use.
Disclosure of Invention
The invention provides a handwriting test question answer extraction method, which can effectively solve the problems that the acceptance degree of network teaching is gradually increased along with the continuous development of online education in the background technology, the workload of teachers is increased along with the increase of the number of students, the work of teachers can be effectively lightened by an online test question review function, more energy is put on teaching work and the knowledge grasping condition of study students, the accuracy of the part for acquiring test questions uploaded by the students and extracting answers is often insufficient, or the requirements on the angle and the flatness of the pictures uploaded by the students are higher, and the use is inconvenient.
In order to achieve the above purpose, the present invention provides the following technical solutions: a method for extracting answers of handwriting test questions comprises taking one picture as a source test question picture A and the other picture as an answer picture B to be extracted, and taking a picture for uploading;
the method also comprises the following steps:
s1: performing corresponding preprocessing, such as size scaling, on the graphs A and B, and extracting feature points;
s2: performing primary matching on the characteristic points of the graphs A and B by using a coarse matching method, calculating a homography matrix from the graph B to the graph A, and correspondingly mapping the graph B into a graph C, wherein the sizes of the graph C and the graph A are consistent, and the angles are close;
s3: extracting characteristic points from the graphs A and C by using a characteristic point matching method based on a graph neural network, and performing characteristic point matching to obtain characteristic point matching pairs corresponding to the graphs A and C;
s4: the feature point matching pairs calculate the positions of the labeling frames in the graph A in the graph C by using a corresponding algorithm;
s5: and extracting answer information in the graph C.
According to the technical scheme, the source test question picture A is marked in advance by marking the answer position to be extracted.
According to the above technical solution, in step S1, the image preprocessing portion, the size scaling is beneficial to the overall running speed, but the accuracy is reduced, and the sift technique is used for feature point extraction.
According to the above technical solution, in step S2, the matching technique may use KNN technique or FLANN technique, and the calculation of the homography matrix may be performed through a corresponding interface provided by OpenCV;
the purpose of this step is to rotate the graph B to an angle close to the graph a through this matching, and can remove the possible redundant content of the graph B in the shooting process, and reduce the error in the next matching, but due to the possible distortion, deformation and other conditions of the graph B, the homography matrix cannot be used to make the graph B perfectly coincide with the graph a, so that the position coordinates marked by the graph a cannot be directly adopted.
According to the above technical solution, in the step S3, the neural network model of the graph may perform feature extraction on the input image by using a CNN network, perform matching through the GNN network, and finally return to the matched feature point pair; the higher the accuracy of the feature point matching technique used in this step, the better.
According to the above technical solution, in the step S4, since the coordinates of the marked frame in the graph a are not necessarily feature points, the approximate position of the marked frame in the graph C is calculated by using an algorithm;
algorithm design: the label box of FIG. A has four vertices, respectively (x 1, y 1), (x 2, y 2), (x 3, y 3), and (x 4, y 4), followed by two steps;
the first step: inputting (x 1, y 1), calculating Euclidean distance between the point and all the characteristic points, taking a plurality of points closest to the point, calculating an average position (x, y) by using the closest points, and then calculating the relative distance between the point and the input point (w=x1-x, h=y1-y);
and a second step of: in the graph C, finding the matching point of the nearest characteristic point of the vertex of the graph A, similarly calculating the average position (x ', y'), and calculating the approximate position (w+x ', h+y') of the input point in the graph C by using the relative position calculated in the first step;
and sequentially inputting the remaining three points to obtain the region mapped by the labeling frame of the graph A in the graph C.
According to the above technical solution, in step S5, a corresponding area is obtained in step S4, and the content of the area is extracted, i.e. the answer portion in the test question.
Compared with the prior art, the invention has the beneficial effects that: the invention has scientific and reasonable structure, safe and convenient use, improves the on-line examination question review function, reduces the problems of lower angle and flatness of the uploaded pictures of students, facilitates the extraction of answers, reduces the inconvenience caused by the extremely high precision requirement in the past, can also achieve the rapid and accurate extraction of the handwriting examination papers in the online examination papers, reduces the degree of the past examination papers, and is not suitable for the application, but the extraction of the method greatly facilitates the answer extraction of the examination papers, is convenient, quick and reliable, and is suitable for popularization and use.
Drawings
The accompanying drawings are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate the invention and together with the embodiments of the invention, serve to explain the invention.
In the drawings:
FIG. 1 is a schematic flow diagram of an example of the present invention;
FIG. 2 is a schematic diagram of a blank test question (panel A);
FIG. 3 is a schematic diagram of a blank test question labeling frame;
FIG. 4 is a schematic diagram of a handwriting test question (panel B);
FIG. 5 is a graph (panel C) of a schematic diagram of a handwriting after homography matrix transformation;
FIG. 6 is a block diagram of the neural network module of FIG. 6;
FIG. 7 is a block diagram of a portion of the GNN module of the neural network of FIG. 7;
FIG. 8 is a feature point matching pair wiring diagram;
FIG. 9 is a schematic diagram of the labeling frame of FIG. A drawn directly on FIG. C;
fig. 10 is a schematic diagram of the final answer area drawn on the graph C.
Detailed Description
The preferred embodiments of the present invention will be described below with reference to the accompanying drawings, it being understood that the preferred embodiments described herein are for illustration and explanation of the present invention only, and are not intended to limit the present invention.
Examples: the invention provides a technical scheme, a method for extracting answers of handwriting test questions, which comprises the following steps:
s1, respectively extracting features of two pictures by using a sift method provided by OpenCV;
s2, performing feature matching on feature points of the two pictures by using a KNN method, calculating a homography matrix H by matching the output feature points with a findhomography method of input OpenCV, and obtaining a picture C on a picture B according to the matrix H;
s3, inputting the graph A and the graph C into a SuperGLUE graph neural network model to obtain feature point matching pairs;
s4, calculating the position of the marked frame in the graph A by utilizing the feature point matching pair;
s5, extracting the handwriting answer in the graph C.
According to the above technical solution, in the step S1, the two pictures are respectively a prepared blank test question picture (picture a) and are marked;
referring to fig. 2 and 3, a same set of test question pictures (picture B) with answers filled in;
see fig. 4; the feature point extraction method used is the sift method provided by OpenCV.
According to the above technical solution, in step S2, a KNN matching method is used to obtain a feature point matching pair of two pictures, and then a homography method provided by OpenCV is used to calculate a homography matrix H for affine transformation from the graph B to the graph a, and the graph B is transformed correspondingly to obtain the graph C, see fig. 5.
According to the above technical solution, step S3 of the above step, the neural network model structure of the graph is shown in fig. 6 and 7, which includes two parts, respectively as follows:
extracting features of the input picture through a CNN network, wherein the specific structure is shown in fig. 6;
the extracted features are matched through the GNN network, and the specific structure is shown in fig. 6;
in the GNN network, the Attention module is executed for 9 times, and the model structure is shown in fig. 7;
the matching results are connected with each other as shown in fig. 8.
According to the technical scheme, the step S4 is divided into the following steps after more accurate characteristic point pairs are obtained in the previous step:
circularly acquiring a labeling frame in the graph A;
acquiring four vertexes of each labeling frame, and sequentially inputting each vertex into an algorithm;
inputting points (Px, py), traversing Euclidean distance between each characteristic point of the graph A and the point, and finally obtaining 5 points closest to the point through screening;
acquiring 5 corresponding characteristic points of the 5 points on the graph C;
average positions (xA, yA) and (xC, yC) of 5 points on the map a and the map C were calculated, respectively, that is:
xA=(x1+x2+x3+x4+x5)/5;
yA=(y1+y2+y3+y4+y5)/5;
xC=(x’1+x’2+x’3+x’4+x’5)/5;
yC=(y’1+y’2+y’3+y’4+y’5)/5;
calculating the relative distance (w, h) of the input point (Px, py) from (xA, yA) on the graph a, namely:
w=Px-xA;
h=Py-yA;
the relative distance is obtained according to the previous step, and the approximate position (Px ', py') of the map C is calculated by using the relative distance and the calculated average position (xC, yC) on the map C, namely:
Px’=w+xC;
Py’=h+yC;
sequentially inputting, and finally obtaining the approximate positions of the four vertexes of each labeling frame to form a quadrangle, and framing the position of the area where the answer on the graph C is located;
finally, the positions of all the label frame mappings are obtained, see fig. 10.
According to the technical scheme, in the step 5, the mapping positions obtained in the previous step are summarized, and the answer of the handwriting test question can be obtained by screenshot on the graph C.
Compared with the prior art, the invention has the beneficial effects that: the invention has scientific and reasonable structure, safe and convenient use, improves the on-line examination question review function, reduces the problems of lower angle and flatness of the uploaded pictures of students, facilitates the extraction of answers, reduces the inconvenience caused by the extremely high precision requirement in the past, can also achieve the rapid and accurate extraction of the handwriting examination papers in the online examination papers, reduces the degree of the past examination papers, and is not suitable for the application, but the extraction of the method greatly facilitates the answer extraction of the examination papers, is convenient, quick and reliable, and is suitable for popularization and use.
Finally, it should be noted that: the foregoing is merely a preferred example of the present invention, and the present invention is not limited thereto, but it is to be understood that modifications and equivalents of some of the technical features described in the foregoing embodiments may be made by those skilled in the art, although the present invention has been described in detail with reference to the foregoing embodiments. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (6)

1. A method for extracting answers of handwriting test questions is characterized in that: the method comprises the steps that one picture is a source test question picture A, the other picture is an answer picture B to be extracted, and photographing and uploading are carried out;
the method also comprises the following steps:
s1: performing corresponding preprocessing, size scaling and feature point extraction on the graphs A and B;
s2: performing primary matching on the characteristic points of the graphs A and B by using a coarse matching method, calculating a homography matrix from the graph B to the graph A, and correspondingly mapping the graph B into a graph C, wherein the sizes of the graph C and the graph A are consistent, and the angles are close;
s3: extracting characteristic points from the graphs A and C by using a characteristic point matching method based on a graph neural network, and performing characteristic point matching to obtain characteristic point matching pairs corresponding to the graphs A and C;
s4: the feature point matching pairs calculate the positions of the labeling frames in the graph A in the graph C by using a corresponding algorithm;
s5: extracting answer information in the graph C;
in the step S4, since the coordinates of the marked frame in the graph a are not necessarily feature points, the approximate position of the marked frame in the graph C is calculated by an algorithm;
algorithm design: the label box of FIG. A has four vertices, respectively (x 1, y 1), (x 2, y 2), (x 3, y 3), and (x 4, y 4), followed by two steps;
the first step: inputting (x 1, y 1), calculating Euclidean distance between the point and all the characteristic points, taking a plurality of points closest to the point, calculating an average position (x, y) by using the closest points, and then calculating the relative distance between the point and the input point (w=x1-x, h=y1-y);
and a second step of: in the graph C, finding the matching point of the nearest characteristic point of the vertex of the graph A, similarly calculating the average position (x ', y'), and calculating the approximate position (w+x ', h+y') of the input point in the graph C by using the relative position calculated in the first step;
and sequentially inputting the remaining three points to obtain the region mapped by the labeling frame of the graph A in the graph C.
2. The method for extracting answers to test questions of handwriting according to claim 1, wherein the source test question picture a is labeled in advance by labeling processing, and the answer position to be extracted is labeled.
3. The method according to claim 1, wherein in the step S1, the image preprocessing portion is scaled to a size that is favorable for the overall operation speed, but the accuracy is reduced, and the feature point extraction is performed by using a sift technique.
4. The method according to claim 1, wherein in step S2, the matching technique uses KNN technique or FLANN technique, and the calculation of the homography matrix is performed through a corresponding interface provided by OpenCV.
5. The method according to claim 1, wherein the neural network model of the graph in step S3 performs feature extraction on the input image by using a CNN network, performs matching by using the GNN network, and finally returns the matched feature point pairs.
6. The method according to claim 1, wherein in step S5, a corresponding area is obtained in step S4, and the content of the area is extracted to obtain an answer portion in the test question.
CN202010510103.8A 2020-06-08 2020-06-08 Method for extracting answers of handwriting test questions Active CN111666882B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010510103.8A CN111666882B (en) 2020-06-08 2020-06-08 Method for extracting answers of handwriting test questions

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010510103.8A CN111666882B (en) 2020-06-08 2020-06-08 Method for extracting answers of handwriting test questions

Publications (2)

Publication Number Publication Date
CN111666882A CN111666882A (en) 2020-09-15
CN111666882B true CN111666882B (en) 2023-04-25

Family

ID=72386942

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010510103.8A Active CN111666882B (en) 2020-06-08 2020-06-08 Method for extracting answers of handwriting test questions

Country Status (1)

Country Link
CN (1) CN111666882B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112419374B (en) * 2020-11-11 2022-12-27 北京航空航天大学 Unmanned aerial vehicle positioning method based on image registration
CN114565750A (en) * 2022-02-22 2022-05-31 杭州布谷蓝途科技有限公司 Method and system for processing paper test questions

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103473774A (en) * 2013-09-09 2013-12-25 长安大学 Vehicle locating method based on matching of road surface image characteristics
CN105184327A (en) * 2015-10-30 2015-12-23 上海海事大学 Vertex trisection strategy-based remote sensing image feature point matching method
CN105488512A (en) * 2015-11-27 2016-04-13 南京理工大学 Sift feature matching and shape context based test paper inspection method
CN108228703A (en) * 2017-10-31 2018-06-29 北京市商汤科技开发有限公司 Image answering method, device, system and storage medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200160177A1 (en) * 2018-11-16 2020-05-21 Royal Bank Of Canada System and method for a convolutional neural network for multi-label classification with partial annotations
CN109583429B (en) * 2018-12-26 2021-04-30 杭州大拿科技股份有限公司 Method and device for correcting application questions in test paper in batches
CN109801317A (en) * 2018-12-29 2019-05-24 天津大学 The image matching method of feature extraction is carried out based on convolutional neural networks
CN110110714A (en) * 2019-04-28 2019-08-09 重庆学析优科技有限公司 Method and system are corrected automatically on a kind of line of papery operation
CN110992217B (en) * 2019-11-15 2020-12-08 广东工业大学 Method and device for expressing and searching multi-view features of design patent
CN111079741A (en) * 2019-12-02 2020-04-28 腾讯科技(深圳)有限公司 Image frame position detection method and device, electronic equipment and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103473774A (en) * 2013-09-09 2013-12-25 长安大学 Vehicle locating method based on matching of road surface image characteristics
CN105184327A (en) * 2015-10-30 2015-12-23 上海海事大学 Vertex trisection strategy-based remote sensing image feature point matching method
CN105488512A (en) * 2015-11-27 2016-04-13 南京理工大学 Sift feature matching and shape context based test paper inspection method
CN108228703A (en) * 2017-10-31 2018-06-29 北京市商汤科技开发有限公司 Image answering method, device, system and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
许明星 ; 李重 ; .基于图像的人体特征点提取与尺寸测量.计算机系统应用.2018,(06),全文. *

Also Published As

Publication number Publication date
CN111666882A (en) 2020-09-15

Similar Documents

Publication Publication Date Title
CN109993160B (en) Image correction and text and position identification method and system
CN111160352B (en) Workpiece metal surface character recognition method and system based on image segmentation
CN111666882B (en) Method for extracting answers of handwriting test questions
CN111783622A (en) Method, device and equipment for recognizing facial expressions and computer-readable storage medium
CN109242856A (en) A kind of papery operation electronization reads and makes comments method and device
CN110689000A (en) Vehicle license plate identification method based on vehicle license plate sample in complex environment
CN111401156A (en) Image identification method based on Gabor convolution neural network
CN110443235B (en) Intelligent paper test paper total score identification method and system
Shao et al. InMAS: Deep learning for designing intelligent making system
CN111738177B (en) Student classroom behavior identification method based on attitude information extraction
CN111274863A (en) Text prediction method based on text peak probability density
CN116912872A (en) Drawing identification method, device, equipment and readable storage medium
CN115984968A (en) Student time-space action recognition method and device, terminal equipment and medium
CN115810197A (en) Multi-mode electric power form recognition method and device
CN112884642B (en) Real-time facial aging simulation method based on face recognition technology
CN114155540A (en) Character recognition method, device and equipment based on deep learning and storage medium
CN112184835B (en) Deformation recovery method and system for piano keyboard image
CN114120443A (en) Classroom teaching gesture recognition method and system based on 3D human body posture estimation
CN113435441A (en) Bi-LSTM mechanism-based four-fundamental operation formula image intelligent batch modification method
CN114550176A (en) Examination paper correcting method based on deep learning
CN111914683A (en) Handwriting score input system based on bionic image enhancement algorithm and FPGA hardware acceleration
CN113160291A (en) Change detection method based on image registration
CN116259050B (en) Method, device, equipment and detection method for positioning and identifying label characters of filling barrel
CN116739949B (en) Blastomere edge enhancement processing method of embryo image
CN111414844B (en) Container number identification method based on convolutional neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant