CN113362032A - Verification and approval method based on artificial intelligence image recognition - Google Patents

Verification and approval method based on artificial intelligence image recognition Download PDF

Info

Publication number
CN113362032A
CN113362032A CN202110638429.3A CN202110638429A CN113362032A CN 113362032 A CN113362032 A CN 113362032A CN 202110638429 A CN202110638429 A CN 202110638429A CN 113362032 A CN113362032 A CN 113362032A
Authority
CN
China
Prior art keywords
picture
image
essential
data
conv1
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110638429.3A
Other languages
Chinese (zh)
Inventor
王进一
周玄红
李茜茹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guizhou Kaifa Future Computer Technology Co ltd
Original Assignee
Guizhou Kaifa Future Computer Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guizhou Kaifa Future Computer Technology Co ltd filed Critical Guizhou Kaifa Future Computer Technology Co ltd
Priority to CN202110638429.3A priority Critical patent/CN113362032A/en
Publication of CN113362032A publication Critical patent/CN113362032A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/103Workflow collaboration or project management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Strategic Management (AREA)
  • Human Resources & Organizations (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Molecular Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a checking and approving method based on artificial intelligence image recognition, which comprises the following steps: 1) the user transacts business through a manual window or a network window and submits the result of photographing: uploading the essential picture and the essential field to an approval system; the approval system classifies the essential picture and judges whether the category of the essential picture meets the current business process; 2) after the picture classification of the essential element meets the acceptance requirement, after the acceptance authority user clicks and accepts, intelligently checking the essential element field, judging whether the essential element information is correct or not and whether the essential element information meets the auditing standard or not, prompting whether the reference opinion is approved or not, and judging whether the case passes or not by an auditor; if the case passes the verification, the case handling is completed. In conclusion, different templates of the same requirement are provided with different requirement checking points through different templates; the model is configured to complete perfect butt joint of the model and the service; the purpose of intelligent verification is achieved.

Description

Verification and approval method based on artificial intelligence image recognition
Technical Field
The invention belongs to the technical field of verification and approval methods, and particularly relates to the technical field of verification and approval methods based on artificial intelligence image recognition.
Background
The existing transaction flow is that when a client applies for transaction, information (pictures) of a plurality of specific requirements needs to be uploaded in a matched manner, for example: an identity card. The business handling personnel take pictures of specific essential components, upload the essential components, submit the essential components, and then carry out handling and checking work by other personnel, manually group the essential components, and carry out comparison, verification and checking on the filled information and the information in the business system. Due to the fact that the number of human factors such as configuration factors of business structure personnel, personnel operation habits, auditing standards and the like is too large. The specific service handling efficiency cannot be fixed, and the handling accuracy cannot be guaranteed; the uploaded key element (picture) is not directly related to the information to be compared and the service information, and the key element and the service information are repeatedly checked, compared and checked; the sorting of the requirements also needs to be done manually. The method not only increases the workload of review, but also increases the possibility of errors in service handling, and simultaneously prolongs the service handling time.
Disclosure of Invention
The invention provides a checking and approving method based on artificial intelligence image recognition, aiming at solving the defects of the problems.
The invention is realized by adopting the following technical scheme.
The invention discloses a verification and approval method based on artificial intelligence image recognition, which comprises the following steps:
1) the user transacts business through a manual window or a network window and submits the result of photographing: uploading the essential picture and the essential field to an approval system;
the approval system classifies the essential picture and judges whether the category of the essential picture meets the current business process;
if the condition that the essential picture is absent exists, the verification result of the approval system is failed;
2) after the picture classification of the essential element meets the acceptance requirement, after the acceptance authority user clicks and accepts, intelligently checking the essential element field, judging whether the essential element information is correct or not and whether the essential element information meets the auditing standard or not, prompting whether the reference opinion is approved or not, and judging whether the case passes or not by an auditor; if the case passes the verification, the case handling is completed.
The classification and intelligent verification method is a paper image classification method based on deep learning image target detection, and specifically comprises the following steps: 1) acquisition and labeling of training image data: the image acquired by the terminal image acquisition equipment is uploaded to an image processing server, the preliminary operation is to manually screen and classify and label the image, and the label file names correspond to the image file names one by one. The label format of the label is txt text file. The content of the label is generated by an open source tool LabelImg label.
The classification and intelligent verification method comprises the following steps: 2) performing K-means clustering on the initially screened original image data: replacing the initial anchor frame of YOLO-v3 with the initial anchor frame of the clustered anchor;
the modified K-means clustering algorithm applicable to the anchor frame comprises the following steps:
i. data represent: presetting the width and height sizes of 9 clusters:
S0={cw1,ch1,cw2,ch2,......cwn,chn},n=9,
wherein cwi,chiRespectively, the initial width and height, S0Is the initial width-height set, the subscript represents the number of current clusters, and n represents the total number. The following is a broad-high representation of all data sets, D being the set of data, wi,hiIs the width and height in the data, n represents the total:
D={w1,h1,w2,h2,......wn,hn},n=Size of Dataset.
distance calculation: and calculating the Euclidean distance from each data in the data set to any cluster by taking the cluster width and height S as a center:
d=(dwj,dhj),
wherein, i, j respectively represent the serial number of the cluster and the serial numbers of all data:
Figure BDA0003106762210000031
Figure BDA0003106762210000032
i={1,2,3......,9},j={1,2,3......,Size of Dataset}.
update clusters: after the distance from each data to each cluster is calculated, all data are sequentially classified into the closest cluster, and then the new cluster width and height are calculated according to each cluster, wherein i represents a cluster class, and j represents data in different clusters:
Figure BDA0003106762210000033
Figure BDA0003106762210000034
i={1,2,3......,9},n=Size of Each Dataset.
recombining into a new set of clusters (the letter meaning in the formula is consistent with the previous meaning)
S1={cw1,ch1,cw2,ch2,......cwn,chn},n=9.
Repeating steps ii and iii until cluster SnApproaching stabilization, this stabilized cluster was applied to the Yolo-v3 anchor frame.
The classification and intelligent verification method comprises the following steps: 3) improvement of the Yolo-v3 network model: the network model structure adopted by the image classification is a Yolo-v3 target detection model;
the core backbone network-residual structure of the target detection model structure is updated, and an HS-Block module is used for replacing 3X3 convolution operation, and the method comprises the following steps:
HS-Block Module:
Input:x(Tensor)
v1,v2,v3,v4,v5=split(x),
v21,v22=split(conv3×3(v2)),
v31,v32=split(conv3×3(concat(conv3×3(v3),v22))),
v41,v42=split(conv3×3(concat(conv3×3(v4),v32))),
v51=conv3×3(concat(conv3×3(v5),v42)).
Output:y(Tensor)
y=concat(v1,v21,v31,v41,v51).
in the above formula, conv1 × 1 and conv3 × 3 represent convolution, split represents a feature map divided by layer average, and v representsiRepresenting the profile, concat represents the per-layer connection profile.
In the hyper-parameter setting, the picture input (608x608) with the maximum Yolo-v3 at present is taken as the input of the network model;
in the aspect of preprocessing the pictures, data augmentation is carried out on all the pictures, and the specific operation is as follows: and rotating the left and right sides by 90 degrees respectively, randomly adjusting the exposure rate, and splicing, zooming and finely adjusting the angle of the picture.
The classification and intelligent verification method comprises the following steps: 3) performing image enhancement processing in a training process:
before the beginning of the Yolo-v3 model, a pre-trained repetitive channel and space attention model is firstly connected to the vector matrix of the input picture, and after the attention model is used for amplifying the more obvious region of the picture in a centralized way, the more obvious region of the picture is placed into a Yolo-v3 network, so that the detection effect of the network model on the more obvious region of the picture can be increased:
channel Fa(x) Space Fb(x) The attention module is as follows:
Input:x(Tensor)
maxpool,avgpool=split(x)
Fa(x)=conv1×1(relu(bn(conv1×1(concat(maxpool,avgpool))))),
Fb1(x)=conv1×1(relu(bn(conv1×1(x)))),
Fb2(x)=conv1×1(relu(bn(conv3×1(conv1×3(x))))),
Fb(x)=Fb1(x)*Fb2(x).
Output:y(Tensor)
y=(Softmax(Fa(x))*x)*Softmax(Fb(x))+x.
in the above formula, conv1 × 1, conv1 × 3, conv1 × 3 represent convolution operations of different sizes, and Fa、FbRepresenting the profile output, Softmax is the mapping function.
The classification and intelligent verification method comprises the following steps: 4) improvements in sustainable training: the preliminarily trained model is directly applied to the image acquisition and screening work, and in the process, the filtering threshold value is reduced, and the acquisition efficiency of the image data set is increased;
roughly screening the original picture by using the model parameters of the first training, continuously optimizing and updating the model parameters, screening the original picture by using the new model parameters, and continuously updating and finely adjusting the hyper-parameters of the Yolo-v3 model by manpower in the process; and continuously circulating the process until the accuracy of the image classifier meets the service requirement.
The invention has the beneficial effects that different requirements and key information needing to be checked, such as the name and the certificate number of the identity card, are configured through the template. Different business types are matched with different model information, and different essential information which needs to be checked in business is intelligently checked. Through the screening of the keyword information, the essential elements are effectively classified, and the essential elements of the same type are distinguished and identified. And correcting and cutting the pictures which are photographed irregularly through an image verification algorithm. And (4) carrying out image recognition, and checking and comparing the key field information on the key element with the information in the system to judge whether the key field information passes the system. Unidentified pictures are uniformly placed. Finally, a suggestion whether the audit is passed is given. The efficiency of the auditors is improved, and the service handling time is reduced. In conclusion, different templates of the same requirement are provided with different requirement checking points through different templates; and the perfect butt joint of the model and the service is completed by configuring the model. The purpose of intelligent verification is achieved.
The invention is further explained below with reference to the drawings and the detailed description.
Drawings
FIG. 1 is a general flow diagram of the present invention.
FIG. 2 is a flowchart of the image classifier operation of the present invention.
FIG. 3 is a comparison of the original residual Block of the present invention with an alternative 3 × 3 HS-Block Block.
FIG. 4 is a flow chart of the channel attention operation of the present invention.
FIG. 5 is a flow chart of the spatial attention work of the present invention.
FIG. 6 is a flow chart of the operation of the present invention for the combination of channel attention and spatial attention.
Detailed Description
A verification and approval method based on artificial intelligence image recognition is disclosed, the general flow is shown in FIG. 1, and the following general description is made on the system:
1) and when the user transacts the business in the manual window/the network window, the user finishes shooting and uploading the business and submits the business. The system firstly classifies the essential picture and judges whether the essential picture category accords with the current business process. And if the condition picture is absent, the verification result is referred to and suggested as failed.
2) After the group of the essential picture meets the acceptance requirement and the acceptance authority user clicks and accepts, intelligently checking essential fields, judging whether essential information is correct or not and whether the essential information meets the auditing standard or not, prompting whether the reference opinion is approved or not, and judging whether the case is approved or not by an auditor; and completing case handling through verification.
Specifically, the image classifier is used for classifying and checking; the paper image classification method based on deep learning image target detection comprises the following steps:
1) acquisition and labeling of training image data: the image acquired by the terminal image acquisition equipment is uploaded to an image processing server, the preliminary operation is to manually screen and classify and label the image, the label format of the label is txt text file, the label file names correspond to the image file names one by one, and the label content is generated by the label of an open source tool LabelImg.
2) Performing K-means clustering on the initially screened original image data: the clustered initial anchor frame is used for replacing the YOLO-v3 initial anchor frame, and the method can remarkably improve the regression accuracy of the small target detection result frame.
The modified K-means clustering algorithm applicable to the anchor frame comprises the following steps: (Note: all the formula characters in the following are consistent with the preceding text)
i. Data represent: presetting the width and height sizes of 9 clusters:
S0={cw1,ch1,cw2,ch2,......cwn,chn},n=9,
wherein cwi,chiRespectively, the initial width and height, S0Is the initial width-height set, the subscript represents the number of current clusters, and n represents the total number. The following is a broad-high representation of all data sets, D being the set of data, wi,hiIs the width and height in the data, n represents the total:
D={w1,h1,w2,h2,......wn,hn},n=Size of Dataset.
distance calculation: calculating Euclidean distance from each data to any cluster in data set by taking cluster width and height S as center
d=(dwj,dhj),
Wherein, i, j respectively represent the serial number of the cluster and the serial numbers of all data:
Figure BDA0003106762210000071
Figure BDA0003106762210000072
i={1,2,3......,9},j={1,2,3......,Size of Dataset}.
update clusters: after the distance from each data to each cluster is calculated, all data are sequentially classified into the closest cluster, and then the new cluster width and height are calculated according to each cluster, wherein i represents a cluster class, and j represents data in different clusters:
Figure BDA0003106762210000073
Figure BDA0003106762210000074
i={1,2,3......,9},n=Size of Each Dataset.
recombined into a new set of clusters
S1={cw1,ch1,cw2,ch2,......cwn,chn},n=9.
Repeating steps ii and iii until cluster SnApproaching stabilization, this stabilized cluster was applied to the Yolo-v3 anchor frame.
3) Improvement of the Yolo-v3 network model: the network model structure adopted by the image classification is a Yolo-v3 target detection model (the image classification task is performed by a target detection method, so that the problem that the classification accuracy of the original image classification on paper text pictures is not high is effectively solved, and experiments show that in the task, the local feature detection and identification performance of the target detection is superior to the global feature identification and classification of the image classification). Wherein, the core backbone network-residual error structure of the target detection model structure is updated, the original residual error module is compared with the HS-Block module replacing the 3x3 module as follows, and the attached figure 3 is matched:
original residual module:
Input:x(Tensor)
Output:y(Tensor)
y=add(conv1×1(conv3×3(conv1×1(x))),x).
HS-Block Module:
Input:x(Tensor)
v1,v2,v3,v4,v5=split(x),
v21,v22=split(conv3×3(v2)),
v31,v32=split(conv3×3(concat(conv3×3(v3),v22))),
v41,v42=split(conv3×3(concat(conv3×3(v4),v32))),
v51=conv3×3(concat(conv3×3(v5),v42)).
Output:y(Tensor)
y=concat(v1,v21,v31,v41,v51).
in the above formula, conv1 × 1 and conv3 × 3 represent convolution, split represents a layer-wise average segmentation feature map, add represents bitwise addition, and v isiRepresenting the profile, concat represents the per-layer connection profile.
As shown above (HS-Block structure referred to from paper HS-ResNet: Hierarchical-Split Block on probabilistic Neural Network), the original residual Block kernel 3x3 convolution is replaced by a HS-Block Hierarchical convolution Block.
The original network and this improved network are trained via the same picture dataset.
Experimental data show that the loss of the original network is stabilized to be about 6.0, while the loss of the improved network is stabilized to be about 4.5, and the loss is improved to a certain extent.
At the hyper-parameter setting, the picture input (608x608) with the largest Yolo-v3 at present is taken as the input of the network model.
Because the mobile terminal runs on the server instead of the terminal mobile device, the requirement on the speed is not very high, so that more resources can be biased to the precision between the balance detection precision and the speed, and a great improvement space is left for the identification precision.
On the preprocessing of the picture, because different terminal image acquisition devices all differ the paper image direction, the illumination intensity, the angle in the image acquisition process, so can do data augmentation to all pictures on the preprocessing, and the concrete operation is: the left and the right are respectively rotated by 90 degrees, the exposure rate is randomly adjusted, and the picture is spliced, zoomed and finely adjusted by an angle (small-angle rotation). Thereby enhancing the generalization ability of the model.
And further image enhancement processing is carried out in the training process.
Before the beginning of the Yolo-v3 model, a pre-trained repetitive channel and space attention model is connected to the input picture vector matrix, and after the attention model is used for amplifying the more obvious regions of the image in a concentrated manner, the more obvious regions of the image are placed into the Yolo-v3 network, so that the detection effect of the network model on the more obvious regions of the image can be improved.
Channel Fa(x) Space Fb(x) The attention module is as follows, and is provided with figures 4, 5 and 6:
Input:x(Tensor)
maxpool,avgpool=split(x)
Fa(x)=conv1×1(relu(bn(conv1×1(concat(maxpool,avgpool))))),
Fb1(x)=conv1×1(relu(bn(conv1×1(x)))),
Fb2(x)=conv1×1(relu(bn(conv3×1(conv1×3(x))))),
Fb(x)=Fb1(x)*Fb2(x).
Output:y(Tensor)
y=(Softmax(Fa(x))*x)*Softmax(Fb(x))+x.
in the above formula, conv1 × 1, conv1 × 3, conv1 × 3 represent convolution operations of different sizes, and Fa、FbRepresenting the profile output, Softmax is the mapping function.
4) Improvements in sustainable training: because the workload of manual image data screening is too large, the model trained primarily is directly applied to the image acquisition and screening work, and in the process, the filtering threshold value needs to be reduced, so that the acquisition efficiency of the image data set can be increased to a great extent.
The specific image classifier workflow is shown in fig. 2.
It can be seen from the flowchart that the initial picture is roughly screened by using the model parameters of the first training, then the model parameters are continuously optimized and updated, the initial picture is screened by using the new model parameters, and the hyper-parameters of the Yolo-v3 model are continuously updated and fine-tuned manually in the process. And continuously circulating the process until the accuracy of the image classifier meets the service requirement.
The invention has the following remarkable innovation points:
1) and replacing the initial anchor frame of YOLO-v3 with the initial anchor frame after clustering, and setting the size of the anchor frame according to different data sets.
2) The adopted network model structure is a YOLO-v3 target detection model, and an image classification task is performed by a target detection method. The 3x3 convolution of the core backbone network-residual Block of the model structure can be replaced by either HS-Block or Ghost network structures.
3) In the preprocessing, data enhancement is performed on all pictures, and the specific operations are as follows: the left and the right are respectively rotated by 90 degrees, the random adjustment of the exposure rate is carried out, the picture is spliced, zoomed and slightly rotated (small-angle rotation), and 608x608x3 is adopted as model input in the super-parameter setting.
4) Before the Yolo-v3 model starts, a mechanism of attention (spatial and channel attention) is engaged to the input picture vector matrix.
And directly applying the preliminarily trained model to the acquisition work of the picture, and reducing the filtering threshold value of the classifier from 0.6 in the formal production environment to 0.3.
The above description is only a part of specific embodiments of the present invention (since the embodiments of the present invention are not exhaustive, the scope of the present invention is defined by the description scope of the present invention and other technical points), and the detailed contents or common general knowledge known in the schemes are not described herein too much. It should be noted that the above-mentioned embodiments do not limit the present invention in any way, and all technical solutions obtained by means of equivalent substitution or equivalent transformation for those skilled in the art are within the protection scope of the present invention. The scope of the claims of the present application shall be determined by the contents of the claims, and the description of the embodiments and the like in the specification shall be used to explain the contents of the claims.

Claims (8)

1. A verification and approval method based on artificial intelligence image recognition is characterized by comprising the following steps:
1) the user transacts business through a manual window or a network window and submits the result of photographing: uploading the essential picture and the essential field to an approval system;
the approval system classifies the essential picture and judges whether the category of the essential picture meets the current business process;
if the condition that the essential picture is absent exists, the verification result of the approval system is failed;
2) after the picture classification of the essential element meets the acceptance requirement, after the acceptance authority user clicks and accepts, intelligently checking the essential element field, judging whether the essential element information is correct or not and whether the essential element information meets the auditing standard or not, prompting whether the reference opinion is approved or not, and judging whether the case passes or not by an auditor; if the case passes the verification, the case handling is completed.
2. The verification and approval method according to claim 1, wherein the classification and intelligent verification method is a paper image classification method based on deep learning image target detection, and specifically comprises the following steps: 1) acquisition and labeling of training image data: the image acquired by the terminal image acquisition equipment is uploaded to an image processing server, the preliminary operation is to manually screen and classify and label the image, and the label file names correspond to the image file names one by one.
3. The validation and approval method of claim 2 wherein said label is in the format of a txt text file.
4. The validation approval method of claim 2 wherein the contents of the label are generated by the open source tool LabelImg label.
5. The validation and approval method of claim 2, wherein said classification, intelligent validation method comprises the steps of: 2) performing K-means clustering on the initially screened original image data: replacing the initial anchor frame of YOLO-v3 with the initial anchor frame of the clustered anchor;
the modified K-means clustering algorithm applicable to the anchor frame comprises the following steps:
i. data represent: presetting the width and height sizes of 9 clusters:
S0={cw1,ch1,cw2,ch2,......cwn,chn},n=9,
wherein cwi,chiRespectively, the initial width and height, S0Is a set with initial width and height, subscript represents the current clustering times, and n represents the total number; the following is a broad-high representation of all data sets, D being the set of data, wi,hiIs the width and height in the data, n represents the total:
D=[w1,h1,w2,h2,......wn,hn},n=Size of Dataset.
distance calculation: and calculating the Euclidean distance from each data in the data set to any cluster by taking the cluster width and height S as a center:
d=(dwj,dhj),
wherein, i, j respectively represent the serial number of the cluster and the serial numbers of all data:
Figure FDA0003106762200000021
Figure FDA0003106762200000022
i={1,2,3......,9},j={1,2,3......,Size of Dataset}.
update clusters: after the distance from each data to each cluster is calculated, all data are sequentially classified into the closest cluster, and then the new cluster width and height are calculated according to each cluster, wherein i represents a cluster class, and j represents data in different clusters:
Figure FDA0003106762200000023
Figure FDA0003106762200000024
i={1,2,3......,9},n=Size of Each Dataset.
recombining into a new set of clusters (the letter meaning in the formula is consistent with the previous meaning)
S1={cw1,ch1,cw2,ch2,......cwn,chn},n=9.
Repeating steps ii and iii until cluster SnApproaching stabilization, this stabilized cluster was applied to the Yolo-v3 anchor frame.
6. The validation and approval method of claim 5, wherein said classification, intelligent validation method comprises the steps of: 3) improvement of the Yolo-v3 network model: the network model structure adopted by the image classification is a Yolo-v3 target detection model;
the 3X3 convolution operation of replacing the residual structure by an HS-Block module is as follows:
HS-Block Module:
Input:x(Tensor)
v1,v2,v3,v4,v5=split(x),
v21,v22=split(conv3×3(v2)),
v31,v32=split(conv3×3(concat(conv3×3(v3),v22))),
v41,v42=split(conv3×3(concat(conv3×3(v4),v32))),
v51=conv3×3(concat(conv3×3(v5),v42)).
Output:y(Tensor)
y=concat(v1,v21,v31,v41,v51).
in the above formula, conv1 × 1 and conv3 × 3 represent convolution, split represents a feature map divided by layer average, and v representsiRepresenting a feature map, concat representing a per-layer connection feature map;
in the hyper-parameter setting, the picture input (608x608) with the maximum Yolo-v3 at present is taken as the input of the network model;
in the aspect of preprocessing the pictures, data augmentation is carried out on all the pictures, and the specific operation is as follows: and rotating the left and right sides by 90 degrees respectively, randomly adjusting the exposure rate, and splicing, zooming and finely adjusting the angle of the picture.
7. The validation and approval method of claim 6, wherein said classification, intelligent validation method comprises the steps of: 3) performing image enhancement processing in a training process:
before the beginning of the Yolo-v3 model, a pre-trained repetitive channel and space attention model is firstly connected to the vector matrix of the input picture, and after the attention model is used for amplifying the more obvious region of the picture in a centralized way, the more obvious region of the picture is placed into a Yolo-v3 network, so that the detection effect of the network model on the more obvious region of the picture can be increased:
channel Fa(x) Space Fb(x) The attention module is as follows:
Input:x(Tensor)
maxpool,avgpool=split(x)
Fa(x)=conv1×1(relu(bn(conv1×1(concat(maxpool,avgpool))))),
Fb1(x)=conv1×1(relu(bn(conv1×1(x)))),
Fb2(x)=conv1×1(relu(bn(conv3×1(conv1×3(x))))),
Fb(x)=Fb1(x)*Fb2(x).
Output:y(Tensor)
y=(Softmax(Fa(x))*x)*Softmax(Fb(x))+x.
in the above formula, conv1 × 1, conv1 × 3, conv1 × 3 represent convolution operations of different sizes, and Fa、FbRepresenting the profile output, Softmax is the mapping function.
8. The validation and approval method of claim 6 or 7 wherein said classification, intelligent validation method comprises the steps of: 4) improvements in sustainable training: the preliminarily trained model is directly applied to the image acquisition and screening work, and in the process, the filtering threshold value is reduced, and the acquisition efficiency of the image data set is increased; roughly screening the original picture by using the model parameters of the first training, continuously optimizing and updating the model parameters, screening the original picture by using the new model parameters, and continuously updating and finely adjusting the hyper-parameters of the Yolo-v3 model by manpower in the process; and continuously circulating the process until the accuracy of the image classifier meets the service requirement.
CN202110638429.3A 2021-06-08 2021-06-08 Verification and approval method based on artificial intelligence image recognition Pending CN113362032A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110638429.3A CN113362032A (en) 2021-06-08 2021-06-08 Verification and approval method based on artificial intelligence image recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110638429.3A CN113362032A (en) 2021-06-08 2021-06-08 Verification and approval method based on artificial intelligence image recognition

Publications (1)

Publication Number Publication Date
CN113362032A true CN113362032A (en) 2021-09-07

Family

ID=77533270

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110638429.3A Pending CN113362032A (en) 2021-06-08 2021-06-08 Verification and approval method based on artificial intelligence image recognition

Country Status (1)

Country Link
CN (1) CN113362032A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024016575A1 (en) * 2022-07-22 2024-01-25 重庆文理学院 Cbam mechanism-based residual network medical image auxiliary detection method

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106022807A (en) * 2016-06-02 2016-10-12 国网上海市电力公司 On-site business handling processing method based on on-line off-line service platform
CN109544096A (en) * 2018-10-18 2019-03-29 国网福建省电力有限公司 Low pressure industry based on artificial intelligence expands business handling method and device on line
CN111353445A (en) * 2020-03-05 2020-06-30 三构科技(武汉)有限公司 Patient assistance intelligent auditing system based on deep learning
CN111401371A (en) * 2020-06-03 2020-07-10 中邮消费金融有限公司 Text detection and identification method and system and computer equipment
CN111612751A (en) * 2020-05-13 2020-09-01 河北工业大学 Lithium battery defect detection method based on Tiny-yolov3 network embedded with grouping attention module
CN112069868A (en) * 2020-06-28 2020-12-11 南京信息工程大学 Unmanned aerial vehicle real-time vehicle detection method based on convolutional neural network
CN112434672A (en) * 2020-12-18 2021-03-02 天津大学 Offshore human body target detection method based on improved YOLOv3
CN112529090A (en) * 2020-12-18 2021-03-19 天津大学 Small target detection method based on improved YOLOv3
CN112686862A (en) * 2020-12-30 2021-04-20 浙江托普云农科技股份有限公司 Pest identification and counting method, system and device and readable storage medium
CN112766427A (en) * 2021-04-06 2021-05-07 浙江口碑网络技术有限公司 Training method, device and equipment of image recognition model

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106022807A (en) * 2016-06-02 2016-10-12 国网上海市电力公司 On-site business handling processing method based on on-line off-line service platform
CN109544096A (en) * 2018-10-18 2019-03-29 国网福建省电力有限公司 Low pressure industry based on artificial intelligence expands business handling method and device on line
CN111353445A (en) * 2020-03-05 2020-06-30 三构科技(武汉)有限公司 Patient assistance intelligent auditing system based on deep learning
CN111612751A (en) * 2020-05-13 2020-09-01 河北工业大学 Lithium battery defect detection method based on Tiny-yolov3 network embedded with grouping attention module
CN111401371A (en) * 2020-06-03 2020-07-10 中邮消费金融有限公司 Text detection and identification method and system and computer equipment
CN112069868A (en) * 2020-06-28 2020-12-11 南京信息工程大学 Unmanned aerial vehicle real-time vehicle detection method based on convolutional neural network
CN112434672A (en) * 2020-12-18 2021-03-02 天津大学 Offshore human body target detection method based on improved YOLOv3
CN112529090A (en) * 2020-12-18 2021-03-19 天津大学 Small target detection method based on improved YOLOv3
CN112686862A (en) * 2020-12-30 2021-04-20 浙江托普云农科技股份有限公司 Pest identification and counting method, system and device and readable storage medium
CN112766427A (en) * 2021-04-06 2021-05-07 浙江口碑网络技术有限公司 Training method, device and equipment of image recognition model

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
PENGCHENG YUAN等: "HS-ResNet: Hierarchical-Split Block on Convolutional Neural Network", 《ARXIV:2010.07621V1[CS.CV]》 *
潘喆琼等: "基于图像识别的业务智能化审核技术研究", 《科学技术创新》 *
郝文宁等: "《数据分析与数据挖掘实验指导书》", 31 March 2016, 国防工业出版社 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024016575A1 (en) * 2022-07-22 2024-01-25 重庆文理学院 Cbam mechanism-based residual network medical image auxiliary detection method

Similar Documents

Publication Publication Date Title
US11816165B2 (en) Identification of fields in documents with neural networks without templates
CN108776772B (en) Cross-time building change detection modeling method, detection device, method and storage medium
WO2018166116A1 (en) Car damage recognition method, electronic apparatus and computer-readable storage medium
US7650035B2 (en) Optical character recognition based on shape clustering and multiple optical character recognition processes
KR20190021187A (en) Vehicle license plate classification methods, systems, electronic devices and media based on deep running
CN111652232B (en) Bill identification method and device, electronic equipment and computer readable storage medium
CN105894046A (en) Convolutional neural network training and image processing method and system and computer equipment
US20080063277A1 (en) High resolution replication of document based on shape clustering
US20120020564A1 (en) Shape clustering in post optical character recognition processing
CN109934255B (en) Model fusion method suitable for classification and identification of delivered objects of beverage bottle recycling machine
CN111814785B (en) Invoice recognition method, training method of relevant model, relevant equipment and device
CN107067228A (en) A kind of hand-held authentication intelligent checks system and its checking method
CN112052845A (en) Image recognition method, device, equipment and storage medium
WO2023029353A1 (en) Service data processing method and apparatus based on multi-modal hybrid model
CN111444792B (en) Bill identification method, electronic equipment, storage medium and device
US11023720B1 (en) Document parsing using multistage machine learning
CN108364037A (en) Method, system and the equipment of Handwritten Chinese Character Recognition
US11455713B2 (en) Automatic document processing
US11620842B2 (en) Automated data extraction and document generation
CN117112782A (en) Method for extracting bid announcement information
CN113362032A (en) Verification and approval method based on artificial intelligence image recognition
CN110796210A (en) Method and device for identifying label information
RU2744769C1 (en) Method for image processing using adaptive technologies based on neural networks and computer vision
CN114140649A (en) Bill classification method, bill classification device, electronic apparatus, and storage medium
CN112508000A (en) Method and equipment for generating OCR image recognition model training data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210907

RJ01 Rejection of invention patent application after publication