CN116403203A - Label generation method, system, electronic equipment and storage medium - Google Patents

Label generation method, system, electronic equipment and storage medium Download PDF

Info

Publication number
CN116403203A
CN116403203A CN202310660056.9A CN202310660056A CN116403203A CN 116403203 A CN116403203 A CN 116403203A CN 202310660056 A CN202310660056 A CN 202310660056A CN 116403203 A CN116403203 A CN 116403203A
Authority
CN
China
Prior art keywords
tag
template
label
filled
article
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310660056.9A
Other languages
Chinese (zh)
Other versions
CN116403203B (en
Inventor
柳雄
胡新平
李�杰
杜智洁
朱道标
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Jingchen Wisdom Logo Technology Co ltd
Original Assignee
Wuhan Jingchen Wisdom Logo Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Jingchen Wisdom Logo Technology Co ltd filed Critical Wuhan Jingchen Wisdom Logo Technology Co ltd
Priority to CN202310660056.9A priority Critical patent/CN116403203B/en
Publication of CN116403203A publication Critical patent/CN116403203A/en
Application granted granted Critical
Publication of CN116403203B publication Critical patent/CN116403203B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/63Scene text, e.g. street names
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/19007Matching; Proximity measures
    • G06V30/19093Proximity measures, i.e. similarity or distance measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/191Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
    • G06V30/1918Fusion techniques, i.e. combining data from various sources, e.g. sensor fusion
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a label generation method, a label generation system, electronic equipment and a storage medium. The method comprises the following steps: acquiring an article tag image and a plurality of predefined tag templates, and searching a tag template with the highest matching degree for the article tag image to be used as a tag template to be filled; inputting the article tag image into a first deep learning model based on deep learning, wherein the first deep learning model is used for outputting fields and attribute values in the article tag image by adopting semantic entity identification and entity relation extraction technology; inputting the label template to be filled into the first deep learning model, and outputting the field of the label template to be filled; and matching and filling the fields and the attribute values in the article tag image with the fields of the tag template to be filled, and generating a tag based on the tag template to be filled. The method and the device can improve the user experience of label generation.

Description

Label generation method, system, electronic equipment and storage medium
Technical Field
The invention belongs to the technical field of computers, and in particular relates to a label generation method, a label generation system, electronic equipment and a storage medium.
Background
The label refers to the literal, graphic and symbol on the product and all the explanatory matters. The label is widely applied to the scenes of super-retail business, industrial production, express delivery, clothing, office management and the like, such as price labels, product description labels, shelf labels, bar code labels, clothing labels, document labels, file storage labels, various articles, stationery labels, express delivery face sheets and the like. The user can edit the label by using label printing software matched with the labeler, and the label is sent to the labeler for printing after the editing is finished.
In order to facilitate user operation and improve user experience, the existing label printing software proposes to automatically generate an electronic label on the label printing software by scanning paper labels on articles such as clothing, goods and the like. However, by visiting a large number of commercial users, there are problems at present, for example, after scanning is completed, each content block in the generated electronic tag is easy to mix in one paragraph, and cannot be directly used, and further editing and beautifying operations are required for the users.
Disclosure of Invention
Aiming at the defects or improvement demands of the prior art, the invention provides a label generation method, a label generation system, electronic equipment and a storage medium, which can promote user experience of label generation.
To achieve the above object, according to one aspect of the present invention, there is provided a tag generation method comprising the steps of:
acquiring an article tag image and a plurality of predefined tag templates, and searching a tag template with the highest matching degree for the article tag image to be used as a tag template to be filled;
inputting the article tag image to a first deep learning model based on deep learning, wherein the first deep learning model is used for outputting fields and attribute values in the article tag image;
inputting the label template to be filled into the first deep learning model, and outputting the field of the label template to be filled;
and matching and filling the fields and the attribute values in the article tag image with the fields of the tag template to be filled, and generating a tag based on the tag template to be filled.
Further, searching a label template with highest matching degree for the article label image comprises the following steps:
identifying texts contained in the article tag images, inputting the texts contained in the article tag images into a second deep learning model, and obtaining embedded representations of the article tag images;
identifying texts contained in each label template, and respectively inputting the texts contained in each label template into the second deep learning model to obtain embedded representations of each label template;
and calculating the similarity between the embedded representation of the article tag image and the embedded representation of each tag template, and searching the tag template with the highest similarity for the article tag image.
Further, the calculating the similarity between the embedded representation of the item label image and the embedded representation of each label template includes the steps of:
clustering the embedded representations of all the tag templates, and classifying the embedded representations of all the tag templates into multiple classes;
and determining a class matched with the embedded representation of the article tag image, and calculating the similarity between the embedded representation of the article tag image and the embedded representation of each tag template in the class.
Further, the method further comprises the steps of:
identifying layout information and pictures contained in the article tag image, carrying out multi-modal fusion on texts contained in the article tag image and the layout information and pictures contained in the article tag image, and then inputting the texts into the second deep learning model;
and identifying the layout information and the pictures contained in each label template, carrying out multi-modal fusion on the text contained in each label template and the layout information and the pictures contained in each label template, and then inputting the text into the second deep learning model.
Further, the matching and filling steps include:
the fields in the article label image are marked as Ai, i is more than or equal to 1 and less than or equal to N, N is the total number of the fields in the article label image, the fields in the label template to be filled are marked as Bj, j is more than or equal to 1 and less than or equal to M, and M is the total number of the fields in the label template to be filled;
for each field Ai, searching a field Bj with the highest matching degree with the field Ai, and taking the attribute value of the field Ai as the attribute value of the field with the highest matching degree.
Further, if the field with the highest matching degree is different from the field Ai, replacing the field with the highest matching degree with the field Ai in the template to be filled.
Further, the generating the label based on the label template to be filled comprises the steps of:
and after the label template to be filled is filled, adjusting the character layout style of the attribute value based on the size of the attribute value filling frame in the label template to be filled.
According to a second aspect of the present invention, there is also provided a tag generation system comprising:
the matching module is used for acquiring an article tag image and a plurality of predefined tag templates, searching the tag template with the highest matching degree for the article tag image, and taking the tag template with the highest matching degree as a tag template to be filled;
the identification module is used for inputting the article tag image into a first deep learning model based on deep learning, wherein the first deep learning model is used for outputting fields and attribute values in the article tag image, inputting the tag template to be filled into the first deep learning model and outputting the fields of the tag template to be filled;
and the filling module is used for matching and filling the fields and the attribute values in the article label image with the fields of the label template to be filled, and generating a label based on the label template to be filled.
According to a third aspect of the present invention there is also provided an electronic device comprising at least one processing unit and at least one storage unit, wherein the storage unit stores a computer program which, when executed by the processing unit, causes the processing unit to perform the steps of any of the methods described above.
According to a fourth aspect of the present invention there is also provided a storage medium storing a computer program executable by an access authentication device, the computer program, when run on the access authentication device, causing the access authentication device to perform the steps of any of the methods described above.
In general, compared with the prior art, the technical scheme of the invention can select the optimal label template after scanning the object label, and identify and fill the fields and the attribute values based on the optimal label template, so that the risk of overlapping a plurality of contents in the generated label can be greatly reduced, the user is not required to carry out subsequent further editing and beautifying, and the user experience is greatly improved.
Drawings
FIG. 1 is a flow chart of a label generation method according to an embodiment of the present invention;
FIG. 2 is a schematic illustration of an image of an item tag according to an embodiment of the present invention;
fig. 3 is a schematic diagram of a tag generation system of an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention. In addition, the technical features of the embodiments of the present invention described below may be combined with each other as long as they do not collide with each other.
In the description of the embodiments of the present application, the terms "first," "second," and the like are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include at least one such feature. The meaning of "a plurality of" means at least two, e.g., two, three, etc., unless explicitly defined otherwise.
The terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, apparatus, article, or device that comprises a list of steps or modules is not necessarily limited to those steps or modules that are expressly listed or inherent to such process, method, article, or device.
The naming or numbering of the steps in the embodiments of the present invention does not mean that the steps in the method flow must be executed according to the time/logic sequence indicated by the naming or numbering, and the named or numbered flow steps may change the execution order according to the technical purpose to be achieved, so long as the same or similar technical effects can be achieved.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the invention. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those of skill in the art will explicitly and implicitly appreciate that the embodiments described herein may be combined with other embodiments.
As shown in fig. 1, an embodiment of the present invention provides a tag generation method, which includes steps S101 to S104.
S101, acquiring an article label image and a plurality of predefined label templates, and searching the label template with the highest matching degree for the article label image to be used as a label template to be filled.
The item label may be scanned by a mobile device camera to obtain an item label image, such as a scanned price label, a product description label, a shelf label, a barcode label, a clothing label.
A plurality of different label templates are predefined, and the fields and styles of each label template are different. For example, according to the characteristics of industry labels, label templates of different industries may be preset, fields of label templates of clothing industries include names, manufacturers, materials, washing modes, and the like, and fields of label templates of food industries may include names, manufacturers, validity periods, and the like. Attribute values may also be predefined for the fields. For the same industry, tag templates with different layout can be designed.
Further, a picture may also be provided on the label template.
Further, searching a label template with highest matching degree for the label image of the article comprises the following steps: identifying texts contained in the object tag images, inputting the texts contained in the object tag images into a second deep learning model, and obtaining embedded representations of the object tag images; identifying texts contained in each label template, and respectively inputting the texts contained in each label template into a second deep learning model to obtain embedded representations of each label template; and calculating the similarity between the embedded representation of the object tag image and the embedded representation of each tag template, and searching the tag template with the highest similarity for the object tag image.
The second deep learning model may convert the input information into an embedded representation that is back-end processable. The second deep learning model may be any existing model for semantic analysis, for example, an ERNIE 3.0 model may be used, and the ERNIE 3.0 model may learn knowledge of various aspects of language, such as morphology, syntax, and semantic information, by setting a pre-training task.
After the embedded representation of the object tag image and the embedded representation of each tag template are obtained, the similarity of the embedded representation of the object tag image and each tag template can be calculated one by one, and the tag template with the highest similarity is used as the tag template to be filled.
Further, the method further comprises the steps of:
identifying layout information and pictures contained in the article tag image, carrying out multi-modal fusion on texts contained in the article tag image and the layout information and pictures contained in the article tag image, and then inputting the texts into a second deep learning model;
and identifying the layout information and the pictures contained in each label template, carrying out multi-modal fusion on the text contained in each label template and the layout information and the pictures contained in each label template, and then inputting the text into the second deep learning model.
In this embodiment, the characteristics of the text, the layout information and the picture after the multimodal fusion can be used as the input of the second deep learning model. Compared with a method based on text only, the method can simultaneously utilize text information, layout information and pictures, so that matching precision can be improved. The layout information comprises the field and the position information of the picture; fields, relative position information between pictures, etc.
Further, calculating a similarity of the embedded representation of the item label image and the embedded representation of each label template, comprising the steps of:
clustering the embedded representations of all the tag templates, and classifying the embedded representations of all the tag templates into multiple classes;
determining the class to which the embedded representation of the item label image belongs, and calculating the similarity between the embedded representation of the item label image and the embedded representation of each label template in the class.
The clustering can adopt various existing clustering algorithms, the clustering algorithm is adopted to divide the embedded representations of all the label templates into several types, the embedded representations of different types are greatly different, and then the embedded representation of the label image of the article is determined to belong to which of the several types. In calculating the similarity, it is not necessary to calculate the similarity between the embedded representation of the item label image and the embedded representations of all the label templates, but only the similarity between the embedded representation of the item label image and the embedded representation of the label templates in the class to which it belongs. Thus, the data volume for calculating the similarity can be greatly reduced, and the efficiency for calculating the similarity is improved.
S102, inputting the object tag image into a first deep learning model based on deep learning, wherein the first deep learning model is used for outputting fields and attribute values in the object tag image.
The first deep learning model may employ a BERT (Bidirectional Encoder Representation from Transformers) model that outputs field and attribute values in the item-tag image using semantic entity identification and entity relationship extraction techniques.
Semantic entity identification, each detected text may be classified, such as into name, price, etc. Entity relationship extraction can identify relationships between text, take fields as keys, and find corresponding values for each key.
For example, as shown in fig. 2, it is recognized that the article tag image includes "article", "95 gasoline", "unit", "liter", and "article" and "unit" as fields, and "95 gasoline" as an attribute value of "article" and "liter" as an attribute value of "unit".
S103, inputting the label template to be filled into the first deep learning model, and outputting the field of the label template to be filled.
And similarly, inputting the label template with the highest matching degree screened in the step S101 into the first deep learning model. The fields of the tag template may have corresponding attribute values, or may be empty. If the fields of the label template have corresponding attribute values, the attribute values corresponding to each field can also be identified.
S104, matching and filling the fields and the attribute values in the object label image with the fields of the label template to be filled, and generating a label based on the label template to be filled.
Further, matching and filling are performed, comprising the steps of:
the fields in the article label image are marked as Ai, i is more than or equal to 1 and less than or equal to N, N is the total number of fields in the article label image, the fields in the label template to be filled are marked as Bj, j is more than or equal to 1 and less than or equal to M, and M is the total number of fields in the label template to be filled.
For each field Ai, searching a field Bj with the highest matching degree with the field Ai, and taking the attribute value of the field Ai as the attribute value of the field with the highest matching degree.
The field with the highest matching degree with A1 is determined, the field with the highest matching degree with A2 is determined, the field with the highest matching degree with A3 is determined, and the field with the highest matching degree with A2 is determined, and the field with the highest matching degree with A3 is determined, and the field with the highest matching degree with A2 is determined. And taking the attribute value of A1 as the attribute value of B2, directly filling the attribute value of A1 into B2 if the attribute value of B2 is null, and directly replacing the attribute value of A1 with the attribute value of B2 if the attribute value of B2 is not null. For example, if the label template to be filled contains a field "name", the article label image also contains a field "name", and the attribute value is "95 gasoline", then "95 gasoline" is taken as the attribute value of the field "name" in the label template to be filled.
It should be noted here that for the same meaning, the descriptions in the item label image and the label template may be different, e.g. the field in the item label image is "good", while in the label template to be filled the field that best matches the field "good" is "name", both expressed in the same meaning. By adopting the method provided by the embodiment of the invention, the expansion of the paraphrasing and the synonyms can be supported, and the matching probability is increased.
Further, if the field with the highest matching degree is different from the field Ai, the field with the highest matching degree is replaced by the field Ai in the template to be filled.
In order to further enhance the user experience, the expression habit of the user is respected, so that the field 'name' of the label template to be filled is replaced by 'commodity', and the description of 'commodity' and 'name' is adopted in the finally generated label. But the layout etc. information is generated based on the label template to be populated. For example, the size and position of the field "commodity" are generated according to the preconfigured size and position information of the label template to be filled, and the characters are replaced by the "commodity".
Further, generating a label based on the label template to be filled comprises the steps of:
after the label template to be filled is filled, the character layout style of the attribute value is adjusted based on the size of the attribute value filling frame in the label template to be filled. For example, if the character of the attribute value is too large, the character needs to be scaled down to be displayed entirely within the filled-in frame. Therefore, related information such as related text fonts, sizes, line feed widths, styles and the like can be predefined, and only the image library related interface is required to be called for adjusting the predefined format.
As shown in fig. 3, according to a second aspect of the present invention, there is also provided a tag generation system including:
the matching module 201 is configured to obtain an article tag image and a predefined plurality of tag templates, search a tag template with the highest matching degree for the article tag image, and use the tag template as a tag template to be filled;
the identification module 202 is configured to input an article tag image to a first deep learning model based on deep learning, where the first deep learning model is configured to output fields and attribute values in the article tag image, and input a tag template to be filled to the first deep learning model, and output the fields and attribute values of the tag template to be filled;
and the filling module 203 is configured to match and fill the fields and the attribute values in the article tag image with the fields and the attribute values of the tag template to be filled, and generate a tag based on the tag template to be filled.
The principle and effect of the label generation system are the same as those of the label generation, and are not repeated here.
The embodiment of the invention also provides electronic equipment, which comprises at least one processing unit and at least one storage unit, wherein the storage unit stores a computer program, and the computer program, when executed by the processing unit, causes the processing unit to execute the steps of any one of the methods.
The electronic device may be a portable electronic device such as a mobile phone, a tablet computer, a wearable device, and the like.
The embodiment of the invention also provides a storage medium storing a computer program executable by an access authentication device, which when run on the access authentication device causes the access authentication device to perform the steps of any of the methods described above.
The storage medium may include, but is not limited to, any type of disk including floppy disks, optical disks, DVDs, CD-ROMs, micro-drives, and magneto-optical disks, ROM, RAM, EPROM, EEPROM, DRAM, VRAM, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.
It will be readily appreciated by those skilled in the art that the foregoing description is merely a preferred embodiment of the invention and is not intended to limit the invention, but any modifications, equivalents, improvements or alternatives falling within the spirit and principles of the invention are intended to be included within the scope of the invention.

Claims (10)

1. A label generation method, characterized by comprising the steps of:
acquiring an article tag image and a plurality of predefined tag templates, and searching a tag template with the highest matching degree for the article tag image to be used as a tag template to be filled;
inputting the article tag image to a first deep learning model based on deep learning, wherein the first deep learning model is used for outputting fields and attribute values in the article tag image;
inputting the label template to be filled into the first deep learning model, and outputting the field of the label template to be filled;
and matching and filling the fields and the attribute values in the article tag image with the fields of the tag template to be filled, and generating a tag based on the tag template to be filled.
2. The tag generation method of claim 1, wherein searching for the tag template having the highest matching degree for the article tag image comprises the steps of:
identifying texts contained in the article tag images, inputting the texts contained in the article tag images into a second deep learning model, and obtaining embedded representations of the article tag images;
identifying texts contained in each label template, and respectively inputting the texts contained in each label template into the second deep learning model to obtain embedded representations of each label template;
and calculating the similarity between the embedded representation of the article tag image and the embedded representation of each tag template, and searching the tag template with the highest similarity for the article tag image.
3. The label generation method of claim 2, wherein said calculating the similarity of the embedded representation of the item label image and the embedded representation of each label template comprises the steps of:
clustering the embedded representations of all the tag templates, and classifying the embedded representations of all the tag templates into multiple classes;
and determining the class to which the embedded representation of the article tag image belongs, and calculating the similarity between the embedded representation of the article tag image and the embedded representation of each tag template in the class.
4. The tag generation method of claim 2, further comprising the step of:
identifying layout information and pictures contained in the article tag image, performing multi-mode fusion on texts contained in the article tag image and the layout information and pictures contained in the article tag image, and then inputting the multi-mode fusion into the second deep learning model;
and identifying the layout information and the pictures contained in each label template, carrying out multi-mode fusion on the text contained in each label template and the layout information and the pictures contained in each label template, and inputting the text and the layout information and the pictures into the second deep learning model.
5. The tag generation method of claim 1, wherein the matching and filling comprises the steps of:
the fields in the article label image are marked as Ai, i is more than or equal to 1 and less than or equal to N, N is the total number of the fields in the article label image, the fields in the label template to be filled are marked as Bj, j is more than or equal to 1 and less than or equal to M, and M is the total number of the fields in the label template to be filled;
for each field Ai, searching a field Bj with the highest matching degree with the field Ai, and taking the attribute value of the field Ai as the attribute value of the field with the highest matching degree.
6. The tag generation method according to claim 5, wherein if the field with the highest matching degree is different from the field Ai, the field with the highest matching degree is replaced with the field Ai in the template to be filled.
7. The tag generation method of claim 1, wherein the generating the tag based on the tag template to be filled comprises the steps of:
and after the label template to be filled is filled, adjusting the character layout style of the attribute value based on the size of the attribute value filling frame in the label template to be filled.
8. A label producing system, comprising:
the matching module is used for acquiring an article tag image and a plurality of predefined tag templates, searching the tag template with the highest matching degree for the article tag image, and taking the tag template with the highest matching degree as a tag template to be filled;
the identification module is used for inputting the article tag image into a first deep learning model based on deep learning, wherein the first deep learning model is used for outputting fields and attribute values in the article tag image, inputting the tag template to be filled into the first deep learning model and outputting the fields of the tag template to be filled;
and the filling module is used for matching and filling the fields and the attribute values in the article label image with the fields of the label template to be filled, and generating a label based on the label template to be filled.
9. An electronic device comprising at least one processing unit and at least one storage unit, wherein the storage unit stores a computer program that, when executed by the processing unit, causes the processing unit to perform the steps of the method of any of claims 1-7.
10. A storage medium storing a computer program executable by an access authentication device, the computer program, when run on the access authentication device, causing the access authentication device to perform the steps of the method of any one of claims 1 to 7.
CN202310660056.9A 2023-06-06 2023-06-06 Label generation method, system, electronic equipment and storage medium Active CN116403203B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310660056.9A CN116403203B (en) 2023-06-06 2023-06-06 Label generation method, system, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310660056.9A CN116403203B (en) 2023-06-06 2023-06-06 Label generation method, system, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN116403203A true CN116403203A (en) 2023-07-07
CN116403203B CN116403203B (en) 2023-08-29

Family

ID=87016429

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310660056.9A Active CN116403203B (en) 2023-06-06 2023-06-06 Label generation method, system, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN116403203B (en)

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140281871A1 (en) * 2013-03-15 2014-09-18 Meditory Llc Method for mapping form fields from an image containing text
CN108052408A (en) * 2017-12-14 2018-05-18 杭州富春云科技有限公司 Alarm work order based on warning information content quickly submits method and system
CN108229299A (en) * 2017-10-31 2018-06-29 北京市商汤科技开发有限公司 The recognition methods of certificate and device, electronic equipment, computer storage media
CN108388831A (en) * 2018-01-10 2018-08-10 链家网(北京)科技有限公司 A kind of identification of spare part and finish message method and device
US20200050845A1 (en) * 2018-08-13 2020-02-13 International Business Machines Corporation Method and system for extracting information from an image of a filled form document
CN112800848A (en) * 2020-12-31 2021-05-14 中电金信软件有限公司 Structured extraction method, device and equipment of information after bill identification
CN113065557A (en) * 2021-04-16 2021-07-02 潍坊工程职业学院 Image matching method based on character extraction
CN113920335A (en) * 2021-09-28 2022-01-11 苏州冷王网络科技有限公司 Deep learning-based image and text embedded drug label identification method
US20220027611A1 (en) * 2021-03-03 2022-01-27 Beijing Baidu Netcom Science Technology Co., Ltd. Image classification method, electronic device and storage medium
US20220044058A1 (en) * 2020-08-07 2022-02-10 Salesforce.Com, Inc. Template-Based Key-Value Extraction for Inferring OCR Key Values Within Form Images
CN114117171A (en) * 2021-11-16 2022-03-01 华中师范大学 Intelligent project file collecting method and system based on energized thinking
US20220092276A1 (en) * 2020-09-22 2022-03-24 Samsung Electronics Co., Ltd. Multimodal translation method, apparatus, electronic device and computer-readable storage medium
CN114821623A (en) * 2022-05-09 2022-07-29 深圳市商汤科技有限公司 Document processing method and device, electronic equipment and storage medium
CN115527220A (en) * 2022-10-11 2022-12-27 太保科技有限公司 Document image classification method and device
US20220415012A1 (en) * 2021-06-24 2022-12-29 Walmart Apollo, Llc Image processing based methods and apparatus for planogram compliance
CN115810197A (en) * 2022-12-19 2023-03-17 北京中电普华信息技术有限公司 Multi-mode electric power form recognition method and device

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140281871A1 (en) * 2013-03-15 2014-09-18 Meditory Llc Method for mapping form fields from an image containing text
CN108229299A (en) * 2017-10-31 2018-06-29 北京市商汤科技开发有限公司 The recognition methods of certificate and device, electronic equipment, computer storage media
CN108052408A (en) * 2017-12-14 2018-05-18 杭州富春云科技有限公司 Alarm work order based on warning information content quickly submits method and system
CN108388831A (en) * 2018-01-10 2018-08-10 链家网(北京)科技有限公司 A kind of identification of spare part and finish message method and device
US20200050845A1 (en) * 2018-08-13 2020-02-13 International Business Machines Corporation Method and system for extracting information from an image of a filled form document
US20220044058A1 (en) * 2020-08-07 2022-02-10 Salesforce.Com, Inc. Template-Based Key-Value Extraction for Inferring OCR Key Values Within Form Images
US20220092276A1 (en) * 2020-09-22 2022-03-24 Samsung Electronics Co., Ltd. Multimodal translation method, apparatus, electronic device and computer-readable storage medium
CN112800848A (en) * 2020-12-31 2021-05-14 中电金信软件有限公司 Structured extraction method, device and equipment of information after bill identification
US20220027611A1 (en) * 2021-03-03 2022-01-27 Beijing Baidu Netcom Science Technology Co., Ltd. Image classification method, electronic device and storage medium
CN113065557A (en) * 2021-04-16 2021-07-02 潍坊工程职业学院 Image matching method based on character extraction
US20220415012A1 (en) * 2021-06-24 2022-12-29 Walmart Apollo, Llc Image processing based methods and apparatus for planogram compliance
CN113920335A (en) * 2021-09-28 2022-01-11 苏州冷王网络科技有限公司 Deep learning-based image and text embedded drug label identification method
CN114117171A (en) * 2021-11-16 2022-03-01 华中师范大学 Intelligent project file collecting method and system based on energized thinking
CN114821623A (en) * 2022-05-09 2022-07-29 深圳市商汤科技有限公司 Document processing method and device, electronic equipment and storage medium
CN115527220A (en) * 2022-10-11 2022-12-27 太保科技有限公司 Document image classification method and device
CN115810197A (en) * 2022-12-19 2023-03-17 北京中电普华信息技术有限公司 Multi-mode electric power form recognition method and device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
PRANJAL DHAKAL ET AL.: "One-Shot Template Matching for Automatic Document Data Capture", 《2019 ARTIFICIAL INTELLIGENCE FOR TRANSFORMING BUSINESS AND SOCIETY (AITB)》, pages 1 - 6 *
曾紫薇: "基于OCR技术的发票验证小程序设计与实现", 《中国优秀硕博学位论文全文数据库 信息科技辑》, pages 138 - 603 *

Also Published As

Publication number Publication date
CN116403203B (en) 2023-08-29

Similar Documents

Publication Publication Date Title
US11715313B2 (en) Apparatus and methods for extracting data from lineless table using delaunay triangulation and excess edge removal
US10726198B2 (en) Method, device, and system, for identifying data elements in data structures
CN108391446B (en) Automatic extraction of training corpus for data classifier based on machine learning algorithm
CN109783635A (en) Use machine learning and fuzzy matching AUTOMATIC ZONING classifying documents and identification metadata
CN112800848A (en) Structured extraction method, device and equipment of information after bill identification
US20210209289A1 (en) Method and apparatus for generating customized content based on user intent
US10699112B1 (en) Identification of key segments in document images
US20230206670A1 (en) Semantic representation of text in document
CN107679208A (en) A kind of searching method of picture, terminal device and storage medium
CN110874534B (en) Data processing method and data processing device
Ha et al. Recognition of OCR invoice metadata block types
CN110147545A (en) The structuring output method and system of text, storage medium and computer equipment
Romero et al. Modern vs diplomatic transcripts for historical handwritten text recognition
CN111753535A (en) Method and device for generating patent application text
CN110738050A (en) Text recombination method, device and medium based on word segmentation and named entity recognition
CN114022891A (en) Method, device and equipment for extracting key information of scanned text and storage medium
CN111274799B (en) Article title generation method, device and storage medium
CN116403203B (en) Label generation method, system, electronic equipment and storage medium
CN114090630A (en) Commodity data integration method based on distributed micro-service cluster
JP5510221B2 (en) Information processing apparatus and information processing program
US20230308731A1 (en) Method for providing service of producing multimedia conversion content by using image resource matching, and apparatus thereof
US11315351B2 (en) Information processing device, information processing method, and information processing program
CN116798417B (en) Voice intention recognition method, device, electronic equipment and storage medium
CN110837740B (en) Comment aspect opinion level mining method based on dictionary improvement LDA model
EP4336379A1 (en) Tracking concepts within content in content management systems and adaptive learning systems

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
PE01 Entry into force of the registration of the contract for pledge of patent right
PE01 Entry into force of the registration of the contract for pledge of patent right

Denomination of invention: A label generation method, system, electronic device, and storage medium

Granted publication date: 20230829

Pledgee: Bank of China Limited Wuhan Donghu New Technology Development Zone Branch

Pledgor: WUHAN JINGCHEN WISDOM LOGO TECHNOLOGY Co.,Ltd.

Registration number: Y2024980008335