US11144799B2 - Image classification method, computer device and medium - Google Patents
Image classification method, computer device and medium Download PDFInfo
- Publication number
- US11144799B2 US11144799B2 US16/556,697 US201916556697A US11144799B2 US 11144799 B2 US11144799 B2 US 11144799B2 US 201916556697 A US201916556697 A US 201916556697A US 11144799 B2 US11144799 B2 US 11144799B2
- Authority
- US
- United States
- Prior art keywords
- classified
- image
- level
- features
- contained
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G06K9/726—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8851—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G06K9/6268—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/40—Image enhancement or restoration using histogram techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
- G06T7/0008—Industrial image inspection checking presence/absence
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/174—Segmentation; Edge detection involving the use of two or more images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/191—Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06V30/19173—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/26—Techniques for post-processing, e.g. correcting the recognition result
- G06V30/262—Techniques for post-processing, e.g. correcting the recognition result using context analysis, e.g. lexical, syntactic or semantic context
- G06V30/274—Syntactic or semantic context, e.g. balancing
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8851—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
- G01N2021/8854—Grading and classifying of flaws
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
Definitions
- the present disclosure relates to the field of image processing technology, and more particularly, to an image classification method, a computer device, and a medium.
- an image classification method comprising:
- acquiring a middle-level semantic feature of an image to be classified through a visual dictionary comprises:
- acquiring a target to be classified contained in the image to be classified comprises: acquiring the target to be classified contained in the image to be classified through image segmentation.
- the image segmentation is implemented using a minimum circumscribed rectangle algorithm.
- the low-level feature is a histogram of oriented gradient feature.
- acquiring the middle-level semantic feature of the image to be classified according to the low-level feature through the visual dictionary comprises: acquiring, according to the low-level feature, a visual word which has the closest Euclidean distance to the histogram of oriented gradient feature of the target to be classified contained in the image to be classified as the middle-level semantic feature of the image to be classified.
- the method before acquiring the target to be classified contained in the image to be classified, the method further comprises: performing image enhancement on the image to be classified.
- performing image enhancement comprises: performing graying, wavelet de-noising, otsu threshold segmentation, binary expansion, median filtering, and binary corrosion in sequence.
- the method before acquiring a middle-level semantic feature of an image to be classified through a visual dictionary, the method further comprises:
- the method before classifying the image to be classified according to the middle-level semantic feature of the image to be classified using a classification model based on middle-level semantic features, the method further comprises:
- acquiring the targets to be classified contained in the plurality of training images comprises: acquiring the targets to be classified contained in the plurality of training images through image segmentation.
- the image segmentation is implemented using a minimum circumscribed rectangle algorithm.
- the low-level features are histogram of oriented gradient features.
- constructing the visual dictionary according to the low-level features of the targets to be classified contained in the plurality of training images comprises: clustering histogram of oriented gradient features of the targets to be classified contained in the plurality of training images using a K-means algorithm to obtain visual words, and constructing the visual dictionary according to the visual words.
- acquiring middle-level semantic features of the plurality of training images according to the low-level features of the targets to be classified contained in the plurality of training images through the visual dictionary comprises: acquiring, according to the low-level features, visual words which have the closest Euclidean distances to the histogram of oriented gradient features of the targets to be classified contained in the plurality of training images as the middle-level semantic features of the plurality of training images.
- the method before acquiring the targets to be classified contained in the plurality of training images, the method further comprises: performing image enhancement on the plurality of training images.
- performing image enhancement comprises: performing graying, wavelet de-noising, otsu threshold segmentation, binary expansion, median filtering, and binary corrosion in sequence.
- a computer device comprising a memory, a processor, and a computer program stored on the memory and operative on the processor, wherein the program, when executed by the processor, implements the image classification method according to the first aspect of the present disclosure.
- a computer readable storage medium having stored thereon a computer program, wherein the program, when executed by a processor, implements the image classification method according to the first aspect of the present disclosure.
- FIG. 1 illustrates a flowchart of an image classification method according to an embodiment of the present disclosure.
- FIG. 2 illustrates a flowchart of performing image enhancement.
- FIG. 3 illustrates a schematic diagram of constructing a visual dictionary according to low-level features of targets to be classified contained in a plurality of training images.
- FIG. 4 illustrates a flowchart of an image classification method for classifying display screen defects according to an embodiment of the present disclosure.
- FIG. 5 illustrates a schematic diagram of an image classification system according to an embodiment of the present disclosure.
- FIG. 6 illustrates a schematic structural diagram of a computer system for performing image classification according to an embodiment of the present disclosure.
- Classification of display screen defects is taken as an example.
- display screen defects which are generally classified into spot defects, line defects and mura defects.
- the spot defects comprise dark spots, bright spots etc.
- the line defects comprise film defects, scratches etc.
- the mura defects comprise glass breakage, liquid leakage etc.
- camera parameters, a shooting mode, and a shooting environment of a camera change while the camera collects a display image of a display screen under test
- low-level features of the same type of defects in the image such as geometric shapes, textures, local description etc., may also change accordingly.
- an embodiment of the present disclosure provides an image classification method.
- step 101 a middle-level semantic feature of an image to be classified is acquired through a visual dictionary.
- step 102 the image to be classified is classified according to the middle-level semantic feature of the image to be classified using a classification model based on middle-level semantic features.
- the middle-level semantic feature of the image to be classified is acquired and the image is classified based on the middle-level semantic feature of the image to be classified, which reduces a semantic gap between a low-level feature and a high-level feature, and has advantages such as high accuracy, high robustness, high efficiency etc. Furthermore computer storage space and a large amount of calculation can be reduced.
- acquiring a middle-level semantic feature of an image to be classified through a visual dictionary further comprises:
- An image has hierarchical features.
- Low-level features are basic features of the image, and represent objective features of the image. During extraction, the low-level features are extracted based on an image level without any prior knowledge.
- Low-level features of images which are widely used currently comprise geometric shapes, textures, local invariant features, transform domain features, etc.
- Middle-level features are middle-level semantic features obtained by analyzing a statistical distribution of the low-level features, and comprise a bag of visual words and a semantic topic.
- High-level features are more abstract semantics of the image, which correspond to behavioral semantics, emotional semantics, and scene semantics etc. The scene semantics describe content of the image, the behavioral semantics describe motion information in the image, and the emotional semantics describe emotions, such as happiness. sadness etc. which are transferred by the image to humans.
- the middle-level semantic feature of the image to be classified is formed by analyzing a statistical distribution of the low-level feature of the target to be classified contained in the image to be classified, i.e., establishing a relationship with semantics by analyzing a statistical distribution of the low-level feature, which reduces a semantic gap between a low-level feature and a high-level feature, and has advantages such as high accuracy, high robustness, high efficiency etc. Furthermore computer storage space and a large amount of calculation can be reduced.
- the method before acquiring the target to be classified contained in the image to be classified, the method further comprises: performing image enhancement on the image to be classified.
- the accuracy of the image classification may further be improved.
- performing image enhancement on the image to be classified further comprises: performing graying, wavelet de-noising, otsu threshold segmentation, binary expansion, median filtering, and binary corrosion in sequence on the image to be classified.
- the graying may greatly reduce the amount of calculation for subsequent image processing based on the distribution characteristics of brightness and chromaticity of the image.
- the wavelet de-noising may highlight high-frequency information such as an edge, a structure etc. of the target to be classified contained in the image, and improve a contrast between the target to be classified and a background, thereby highlighting the target to be classified contained in the image.
- the otsu threshold segmentation is used to binarize the image.
- the median filtering may perform edge smoothing on the target contained in the binarized image.
- the binary expansion and the binary corrosion may remove holes and isolated false points from the binarized image.
- acquiring a target to be classified contained in the image to be classified further comprises: acquiring the target to be classified contained in the image to be classified through image segmentation.
- acquiring the target to be classified contained in the image to be classified through image segmentation further comprises: performing image segmentation on the image to be classified based on a minimum circumscribed rectangle algorithm, to obtain the target to be classified contained in the image to be classified.
- a minimum circumscribed rectangle of the target contained in the image is obtained using a target rotation method. Specifically, the target contained in the image is rotated at equal intervals in a range of 90°, an area of a circumscribed rectangle parallel to a coordinate axis direction is calculated each time the target is rotated, a circumscribed rectangle with the smallest area is obtained as a minimum circumscribed rectangle of the target to be classified contained in the image, and then a region of the minimum circumscribed rectangle is segmented to obtain the target to be classified contained in the image.
- the low-level feature of the target to be classified contained in the image to be classified is a Histogram of Oriented Gradient (HOG) feature.
- the histogram of oriented gradient feature is composed of histograms of oriented gradient of local regions of the image. Gradients in an image or oriented densities of edges in the image may accurately represent shape attributes of local regions of the target to be classified. Therefore, the histogram of oriented gradient feature may be used as the low-level feature to ensure the accuracy of the image classification.
- a flow of extracting an HOG feature is as follows.
- a magnitude value and an orientation of the gradient at the pixel point (x, y) are as follows respectively:
- a gradient histogram is established, and specifically, the image is firstly divided into a plurality of cells, for example, cells each composed of 2*2 pixels, an orientation value of the gradient histogram is set to 0-180°, and a bin is obtained every 20°, wherein a weight of oriented gradient within the gradient histogram is determined by a magnitude value of the gradient.
- acquiring the middle-level semantic feature of the image to be classified according to the low-level feature of the target to be classified contained in the image to be classified through the visual dictionary further comprises: acquiring a visual word which has the closest Euclidean distance to the histogram of oriented gradient feature of the target to be classified contained in the image to be classified as the middle-level semantic feature of the image to be classified.
- the method before acquiring a middle-level semantic feature of an image to be classified through a visual dictionary, the method further comprises:
- the method before acquiring the targets to be classified contained in the plurality of training images, the method further comprises: performing image enhancement on the plurality of training images.
- performing image enhancement on the plurality of training images further comprises: performing graying, wavelet de-noising, otsu threshold segmentation, binary expansion, median filtering, and binary corrosion in sequence on the plurality of training images.
- the method before classifying the image to be classified according to the middle-level semantic feature of the image to be classified using a classification model based on middle-level semantic features, the method further comprises: acquiring middle-level semantic features of the plurality of training images according to the low-level features of the targets to be classified contained in the plurality of training images through the visual dictionary; and
- acquiring the targets to be classified contained in the plurality of training images further comprises: acquiring the targets to be classified contained in the plurality of training images through image segmentation.
- acquiring the targets to be classified contained in the plurality of training images through image segmentation further comprises: performing image segmentation on the plurality of training images based on a minimum circumscribed rectangle algorithm, to obtain the targets to be classified contained in the plurality of training images.
- the low-level features of the targets to be classified contained in the plurality of training images are histogram of oriented gradient features.
- constructing the visual dictionary according to the low-level features of the targets to be classified contained in the plurality of training images further comprises: clustering histogram of oriented gradient features of the targets to be classified contained in the plurality of training images using a K-means algorithm to obtain visual words, and constructing the visual dictionary according to the visual words.
- acquiring middle-level semantic features of the plurality of training images according to the low-level features of the targets to be classified contained in the plurality of training images through the visual dictionary comprises: acquiring visual words which have the closest Euclidean distances to the histogram of oriented gradient features of the targets to be classified contained in the plurality of training images as the middle-level semantic features of the plurality of training images.
- a Bag of Words (BoW) model is such a middle-level feature representation method.
- the BoW model treats the image as a document composed of visual words, so that the BoW model and related theory in the field of text classification are applied to the understanding of the image, without analyzing and interpreting specific composition of targets in an object, and instead, a plurality of sample images are applied as training samples, low-level features of the sample images are quantized into visual words, and content of an unknown target is expressed using a distribution histogram of the visual words of the image.
- the histogram of oriented gradient features of the targets to be classified contained in the plurality of training images are clustered using a K-means algorithm to obtain visual words, and a flow of constructing a visual dictionary according to the visual words is as follows.
- Visual words for example, t 1 , t 2 , t 3 , t 4 as shown in FIG. 3 , are formed from the histogram of oriented gradient features of the targets to be classified contained in the plurality of training images using the K-means algorithm.
- t 1 , t 2 , t 3 , t 4 are centroids of clusters, and all the visual words form a visual dictionary.
- the K-means algorithm is an unsupervised machine learning algorithm. When clustering is performed using the algorithm, N objects are divided into k classes based on a criterion that there is a high similarity within the classes and a low similarity between the classes.
- a specific flow is as follows.
- a specific flow of acquiring a visual word which has the closest Euclidean distance to the histogram of oriented gradient feature of the target to be classified contained in the image to be classified as the middle-level semantic feature of the image to be classified is as follows.
- V i [ v 1 +v 2 + . . . +v k ], v k ⁇ R k ⁇ 1
- the histogram of the visual words is the middle-level semantic feature of the image, and is applied to sample images. Firstly, low-level features of targets to be classified contained in the images are quantized into visual words, and then content of the unknown targets is expressed by a distribution histogram of the visual words of the images. Since overall statistical information of the images is applied to the middle-level semantic feature without analyzing specific composition of the targets contained in the images, there are high accuracy and high robustness in modeling the features of the targets of the images.
- visual words having the closest Euclidean distances to the histogram of oriented gradient features of the targets to be classified contained in the plurality of training images may be acquired as middle-level semantic features of the plurality of training images in a specific flow similar to that described above, and will not be described in detail here.
- Image enhancement is performed on the training samples and the test sample respectively.
- the middle-level semantic features of the training samples are trained using a machine learning algorithm to obtain a classification model, and the test sample is classified according to the middle-level semantic feature of the test sample using the classification model.
- Feature extraction is a key step in the classification of display screen defects.
- Low-level features such as geometry, textures, shapes, local descriptors etc. used in the existing display screen defect classification method based on image classification are extracted based on information at an image level. Since the feature extraction has not been sufficiently generalized and abstracted, it is far from a concept level, display screen defect classification based on low-level features is not highly accurate, and it is difficult to accurately classify an image having a defect outside a training set.
- the specific flow of the present example comprises two phases, which are a training phase and a classification phase respectively.
- image enhancement is performed on the training samples
- a visual dictionary is constructed according to the low-level features of the display screen defect targets of the training samples, and middle-level semantic features of the training samples are acquired through the visual dictionary;
- the middle-level semantic features of the training samples are trained using a machine learning algorithm to obtain a classification model.
- At least one second display image of at least one second display screen to be classified under test is collected, and the second display image is used as a test sample;
- a middle-level semantic feature of the test sample is acquired through the visual dictionary
- FIG. 5 another embodiment of the present disclosure provides an image classification system which performs the image classification method described above, comprising:
- a middle-level semantic feature acquisition module configured to acquire a middle-level semantic feature of a target to be classified in an image to be classified through a visual dictionary
- a classification module configured to classify the image to be classified according to the middle-level semantic feature of the image to be classified using a classification model based on middle-level semantic features.
- a computer system suitable for implementing the image classification system according to the present embodiment, comprising a Central Processing Unit (CPU) which may perform various appropriate actions and processes based on a program stored in a Read Only Memory (ROM) or a program loaded from a storage portion into a Random Access Memory (RAM).
- ROM Read Only Memory
- RAM Random Access Memory
- Various programs and data required for operations of a computer system are also stored in the RAM.
- the CPU, the ROM, and the RAM are connected to each other through a bus.
- An Input/Output (I/O) interface is also connected to the bus.
- the following components are connected to the I/O interface: an input part comprising a keyboard, a mouse, etc.; an output part comprising a Liquid Crystal Display (LCD), etc., and a speaker; a storage part comprising a hard disk etc.; and a communication part comprising a network interface card such as a LAN card, a modem, etc.
- the communication part performs communication processing via a network such as the Internet.
- the driver is also connected to the I/O interface as needed.
- a removable medium such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory etc., is mounted on the driver as needed, so that a computer program read therefrom is installed into the storage part as needed.
- the processes described above in the flowcharts may be implemented as a computer software program.
- the present embodiment comprises a computer program product comprising a computer program included tangibly on a computer readable medium, the computer program comprising program codes for performing the methods illustrated in the above flowcharts.
- the computer program may be downloaded and installed from the network via a communication portion, and/or installed from a removable medium.
- each block of the flowcharts or diagrams may represent a module, a program segment, or a portion of codes, which comprises one or more executable instructions for implementing specified logic functions.
- the functions illustrated in the blocks may also occur in a different order than that illustrated in the accompanying drawings. For example, two successively represented blocks may in fact be executed substantially in parallel, or they may sometimes be executed in a reverse order, depending upon functionality involved.
- each block of the diagrams and/or flowcharts, and combinations of blocks in the diagrams and/or flowcharts may be implemented by a dedicated hardware-based system which performs specified functions or operations, or may be implemented by combination of dedicated hardware and computer instructions.
- Modules described in the present embodiment may be implemented by software, or may be implemented by hardware, and the described modules may also be disposed in a processor.
- a processor comprising a middle-level semantic feature acquisition module and a classification module.
- names of these modules do not in any way constitute a limitation on the modules themselves.
- the middle-level semantic feature acquisition module may also be described as “visual dictionary module”.
- the present embodiment further provides a non-volatile computer storage medium, which may be a non-volatile computer storage medium included in the above apparatus in the above embodiments; or may exist separately but is not assembled into a non-volatile computer storage medium in a terminal.
- the above non-volatile computer storage medium stores one or more programs which, when executed by a device, cause the device to:
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- Pathology (AREA)
- Immunology (AREA)
- General Health & Medical Sciences (AREA)
- Biochemistry (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
I(x,y)=I(x,y)gamma
where I(x, y) is the minimum circumscribed rectangle of the target to be classified contained in the image which is obtained in the above example, and gamma=½.
G x(x,y)=I(x+1,y)−I(x−1,y) and
G y(x,y)=I(x,y+1)−I(x,y−1)
where e is a constant term with a small value.
R k ={i|∥x i −u k ∥≤∥x i −u i ∥,k≠l},k=1,2, . . . ,K
U={u 1 ,u 2 , . . . ,u k },u k ∈R d.
where X is a low-level feature space of each image, U is a visual dictionary obtained by K-means clustering, V represents a correspondence relationship between low-level features and visual words, a conditional constraint ∥vi∥0=1 indicates that there may only be one non-zero value in a vector vi, a conditional constraint ∥vi∥1=1 indicates that a cumulative sum of absolute values of numbers in the vector vi is 1, and thereby ∥vi∥0=∥vi∥1=1 defines that there is only one “1” in the vector vi, that is, a low-level feature may be quantized into a visual word having the closest Euclidean distance thereto.
V i=[v 1 +v 2 + . . . +v k],v k ∈R k×1
Claims (18)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910175377.3 | 2019-03-08 | ||
| CN201910175377.3A CN109858570A (en) | 2019-03-08 | 2019-03-08 | Image classification method and system, computer equipment and medium |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200285917A1 US20200285917A1 (en) | 2020-09-10 |
| US11144799B2 true US11144799B2 (en) | 2021-10-12 |
Family
ID=66900257
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/556,697 Active 2039-11-08 US11144799B2 (en) | 2019-03-08 | 2019-08-30 | Image classification method, computer device and medium |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US11144799B2 (en) |
| CN (1) | CN109858570A (en) |
Families Citing this family (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110414538B (en) * | 2019-07-24 | 2022-05-27 | 京东方科技集团股份有限公司 | Defect classification method, defect classification training method and device thereof |
| US12056880B2 (en) * | 2020-08-03 | 2024-08-06 | Korea Advanced Institute Of Science And Technology | Method of classifying lesion of chest x-ray radiograph based on data normalization and local patch and apparatus thereof |
| CN112183559B (en) * | 2020-10-27 | 2022-01-11 | 深圳市威富视界有限公司 | Image recognition model training method, image recognition method and device |
| CN112347899B (en) * | 2020-11-03 | 2023-09-19 | 广州杰赛科技股份有限公司 | A moving target image extraction method, device, equipment and storage medium |
| CN112668567A (en) * | 2020-12-25 | 2021-04-16 | 深圳太极云软技术有限公司 | Image clipping algorithm based on deep learning |
| CN115205571B (en) * | 2021-04-13 | 2026-01-02 | 武汉Tcl集团工业研究院有限公司 | An image classification method, apparatus, terminal device, and storage medium |
| CN115619800A (en) * | 2021-07-13 | 2023-01-17 | 电子科技大学 | Segmentation method and device for COVID-19 CT images based on adaptive threshold segmentation |
| CN116863198A (en) * | 2023-05-26 | 2023-10-10 | 中国银行股份有限公司 | Image classification method, apparatus, computer device and storage medium |
| CN116721099B (en) * | 2023-08-09 | 2023-11-21 | 山东奥洛瑞医疗科技有限公司 | Image segmentation method of liver CT image based on clustering |
| CN116994109B (en) * | 2023-08-10 | 2025-09-16 | 商汤集团有限公司 | Image processing method, device, computer equipment and storage medium |
Citations (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN102208038A (en) | 2011-06-27 | 2011-10-05 | 清华大学 | Image classification method based on visual dictionary |
| US20120075638A1 (en) * | 2010-08-02 | 2012-03-29 | Case Western Reserve University | Segmentation and quantification for intravascular optical coherence tomography images |
| CN102509110A (en) | 2011-10-24 | 2012-06-20 | 中国科学院自动化研究所 | Method for classifying images by performing pairwise-constraint-based online dictionary reweighting |
| US20120290577A1 (en) | 2011-05-13 | 2012-11-15 | Microsoft Corporation | Identifying visual contextual synonyms |
| US20130080426A1 (en) * | 2011-09-26 | 2013-03-28 | Xue-wen Chen | System and methods of integrating visual features and textual features for image searching |
| CN104850859A (en) | 2015-05-25 | 2015-08-19 | 电子科技大学 | Multi-scale analysis based image feature bag constructing method |
| US20160132754A1 (en) * | 2012-05-25 | 2016-05-12 | The Johns Hopkins University | Integrated real-time tracking system for normal and anomaly tracking and the methods therefor |
| CN107515905A (en) | 2017-08-02 | 2017-12-26 | 北京邮电大学 | An Interactive Image Search and Fusion Method Based on Sketch |
| US20190331301A1 (en) * | 2016-12-30 | 2019-10-31 | Du Yuchuan | Method for leakage detection of underground pipeline corridor based on dynamic infrared thermal image processing |
-
2019
- 2019-03-08 CN CN201910175377.3A patent/CN109858570A/en active Pending
- 2019-08-30 US US16/556,697 patent/US11144799B2/en active Active
Patent Citations (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20120075638A1 (en) * | 2010-08-02 | 2012-03-29 | Case Western Reserve University | Segmentation and quantification for intravascular optical coherence tomography images |
| US20120290577A1 (en) | 2011-05-13 | 2012-11-15 | Microsoft Corporation | Identifying visual contextual synonyms |
| CN102208038A (en) | 2011-06-27 | 2011-10-05 | 清华大学 | Image classification method based on visual dictionary |
| US20130080426A1 (en) * | 2011-09-26 | 2013-03-28 | Xue-wen Chen | System and methods of integrating visual features and textual features for image searching |
| CN102509110A (en) | 2011-10-24 | 2012-06-20 | 中国科学院自动化研究所 | Method for classifying images by performing pairwise-constraint-based online dictionary reweighting |
| US20160132754A1 (en) * | 2012-05-25 | 2016-05-12 | The Johns Hopkins University | Integrated real-time tracking system for normal and anomaly tracking and the methods therefor |
| CN104850859A (en) | 2015-05-25 | 2015-08-19 | 电子科技大学 | Multi-scale analysis based image feature bag constructing method |
| US20190331301A1 (en) * | 2016-12-30 | 2019-10-31 | Du Yuchuan | Method for leakage detection of underground pipeline corridor based on dynamic infrared thermal image processing |
| CN107515905A (en) | 2017-08-02 | 2017-12-26 | 北京邮电大学 | An Interactive Image Search and Fusion Method Based on Sketch |
Non-Patent Citations (2)
| Title |
|---|
| First Chinese Office Action dated Jul. 23, 2020, for corresponding Chinese Application No. 201910175377.3. |
| Liu Yuhuan, "Product Defect Detection Based on Computer Vision", Mar. 2018. |
Also Published As
| Publication number | Publication date |
|---|---|
| US20200285917A1 (en) | 2020-09-10 |
| CN109858570A (en) | 2019-06-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11144799B2 (en) | Image classification method, computer device and medium | |
| US11334982B2 (en) | Method for defect classification, method for training defect classifier, and apparatus thereof | |
| CN110334706B (en) | Image target identification method and device | |
| Al-Shemarry et al. | Ensemble of adaboost cascades of 3L-LBPs classifiers for license plates detection with low quality images | |
| Chen et al. | Text detection and recognition in images and video frames | |
| Jia et al. | Region-based license plate detection | |
| Pan et al. | A robust system to detect and localize texts in natural scene images | |
| EP2701098B1 (en) | Region refocusing for data-driven object localization | |
| CN111126115B (en) | Violent sorting behavior identification method and device | |
| WO2019232853A1 (en) | Chinese model training method, chinese image recognition method, device, apparatus and medium | |
| Molina-Moreno et al. | Efficient scale-adaptive license plate detection system | |
| CN111507344A (en) | Method and device for recognizing text from images | |
| Liu et al. | A novel SVM network using HOG feature for prohibition traffic sign recognition | |
| Shi et al. | License plate localization in complex environments based on improved GrabCut algorithm | |
| Ismail et al. | Statistical Binarization Techniques for Document Image Analysis. | |
| CN104834891A (en) | Method and system for filtering Chinese character image type spam | |
| CN113344047A (en) | Platen state identification method based on improved K-means algorithm | |
| Gaceb et al. | Adaptative smart-binarization method: For images of business documents | |
| Gaddour et al. | A new method for arabic text detection in natural scene image based on the color homogeneity | |
| Qin et al. | Video scene text frames categorization for text detection and recognition | |
| Huang et al. | Text extraction in natural scenes using region-based method | |
| Tung et al. | Efficient uneven-lighting image binarization by support vector machines | |
| Pack et al. | Augmentation-based pseudo-ground truth generation for deep learning in historical document segmentation for greater levels of archival description and access | |
| Kaur et al. | Text Extraction from Natural Scene using PCA. | |
| Kosala et al. | Robust License Plate Detection in Complex Scene using MSER-Dominant Vertical Sobel. |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: BEIJING BOE OPTOELECTRONICS TECHNOLOGY CO., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PENG, XIANGJUN;WANG, YUNQI;ZHAO, CHENXI;AND OTHERS;SIGNING DATES FROM 20190603 TO 20190718;REEL/FRAME:050237/0281 Owner name: BOE TECHNOLOGY GROUP CO., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PENG, XIANGJUN;WANG, YUNQI;ZHAO, CHENXI;AND OTHERS;SIGNING DATES FROM 20190603 TO 20190718;REEL/FRAME:050237/0281 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |