CN111931782A - Semantic segmentation method, system, medium, and apparatus - Google Patents

Semantic segmentation method, system, medium, and apparatus Download PDF

Info

Publication number
CN111931782A
CN111931782A CN202010808133.7A CN202010808133A CN111931782A CN 111931782 A CN111931782 A CN 111931782A CN 202010808133 A CN202010808133 A CN 202010808133A CN 111931782 A CN111931782 A CN 111931782A
Authority
CN
China
Prior art keywords
semantic segmentation
semantic
picture
edge
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010808133.7A
Other languages
Chinese (zh)
Other versions
CN111931782B (en
Inventor
舒睿俊
陈铭弘
李嘉茂
张晓林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Institute of Microsystem and Information Technology of CAS
Original Assignee
Shanghai Institute of Microsystem and Information Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Institute of Microsystem and Information Technology of CAS filed Critical Shanghai Institute of Microsystem and Information Technology of CAS
Priority to CN202010808133.7A priority Critical patent/CN111931782B/en
Publication of CN111931782A publication Critical patent/CN111931782A/en
Application granted granted Critical
Publication of CN111931782B publication Critical patent/CN111931782B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/18Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Computational Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Mathematical Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Mathematical Physics (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Operations Research (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a semantic segmentation method, a semantic segmentation system, a semantic segmentation medium and a semantic segmentation device, wherein the method comprises the following steps: acquiring a picture needing semantic segmentation, and recording the size of the picture as H multiplied by W multiplied by M, wherein H represents the height of the picture, W represents the width of the picture, and M represents the number of channels; performing edge processing on the picture to obtain an edge connected graph S; performing semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi(ii) a According to the edge connected graph S, a four-neighbor domain connected set F with the pixel of 0 is generated, each connected domain k in the F is traversed, and a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain k is found out respectivelyminDetermining connectivityFinal semantic labels for all pixels in the domain k
Figure DDA0002629910860000011
Where (h, w) e.k. The invention discloses a semantic segmentation method, a semantic segmentation system, a semantic segmentation medium and a semantic segmentation device, which are used for improving the precision and the robustness of semantic segmentation in strange scenes.

Description

Semantic segmentation method, system, medium, and apparatus
Technical Field
The present invention relates to the field of image segmentation technologies, and in particular, to a semantic segmentation method, system, medium, and apparatus.
Background
In the semantic recognition system, the method combined with deep learning is the most widely applied at present, and the method with supervised learning is the most widely applied. In practical application, a method of supervised training of a certain data set with semantic tags is often needed, and the goal is to learn a stable model with good performance in all aspects. Due to the limited expression capability of the trained model, the obtained model can be a preferred model.
Ensemble learning is intended to obtain a better and more comprehensive model by combining multiple learners (models), and the core idea is that even if one learner gets a wrong prediction, the other learners can correct the error back, as shown in fig. 1 a. Each individual learner is generated from training data using an existing learning algorithm, such as the C4.5 decision tree algorithm, the BP neural network algorithm, and the like. The effect of reducing variance, deviation or improving prediction can be achieved by combining a plurality of machine learning technologies with the same form or different forms into a meta-algorithm of a prediction model.
In combination, ensemble learning can be divided into two types. The first is a sequence integration approach, in which the underlying learners participating in training are generated in order (e.g., AdaBoost). The principle of the sequence method is to utilize the dependency between the basic learners. By assigning higher weights to the samples that were incorrectly labeled in the previous training, the overall prediction effect can be improved. The second is a parallel integration approach, where the underlying learners participating in the training are generated in parallel (e.g., Random Forest). The principle of the parallel method is that the independence between the basic learners is utilized, and errors can be obviously reduced through averaging.
Ensemble learning can be divided into two categories in terms of binding strategies. The first is an averaging method, mainly aimed at outputting h in numerical formi(x) Regression for e.RAnd (5) class tasks. The averaging method is classified into a simple averaging method and a weighted averaging method. The simple averaging method formula is as follows:
Figure BDA0002629910840000011
the weighted average method is as follows:
Figure BDA0002629910840000012
where w isiPresentation individual learning machine hiWeight of (2), usually requiring wi≥0,
Figure BDA0002629910840000013
The weights of the weighted average method are generally learned from training data, and training samples in real tasks are usually insufficient or noisy, which makes the learned weights not completely reliable. Thus, both experiments and applications show that weighted average scores are not necessarily superior to naive simple averaging. In general, weighted averaging is preferred when the performance of individual learners varies greatly, while simple averaging is preferred when the performance of individual learners is similar.
Another combination strategy is voting, mainly aiming at classification tasks. For the classification task, learner hiWill tag the set c from the category1,c2,...,cNOne marker is predicted, and the most common binding strategy is to use voting. For ease of discussion, we will refer to hiThe prediction output at sample X is represented as an N-dimensional vector, where
Figure BDA0002629910840000021
Is hiIn category label cjAn output of (c). The voting method is classified into three kinds, an absolute majority voting method, a relative majority voting method, and a weighted voting method. The absolute majority voting formula is as follows:
Figure BDA0002629910840000022
the relative majority voting formula is as follows:
Figure BDA0002629910840000023
the weighted voting formula is as follows:
Figure BDA0002629910840000024
in the existing scene perception and understanding algorithm, an edge graph is used as auxiliary information to improve results, wherein the work is less, part of the work is as the same as the Classification with edge of d.marmanis on the isps 2018, the Classification with edge of the d.marmanis is inputted into an edge detection network together with an input picture and a DEM signal before the classic edge detection network HED is added into a full volume network (FCN) -based semantic segmentation network, then the edge likelihood graph outputted by the edge detection network and the two signals are inputted into the semantic segmentation network together, and finally the whole network is supervised and trained by a semantic segmentation label, as shown in fig. 1 b. In the training phase, the HED network and the FCN network need to be trained respectively aiming at respective tasks and then spliced together for training.
In the existing edge auxiliary semantic segmentation method, only simple concatee (namely, two pictures are aligned and stacked) operation is carried out on an edge graph and an input picture, and edge information is integrated by a method of independently learning an edge likelihood graph by a network in a training process, so that the method is not clear in an auxiliary means; the method of marmanis is a method corresponding to a specific data set, and does not consider the effect of semantic segmentation when a network is placed in a real scene. The pictures collected in the real scene and the pictures in the data set during training have data distribution difference, and are also the data set of the automatic driving scene, and the semantic segmentation network trained on the Cityscapes data set cannot be well represented on the ADE20k data set, not to mention the representation capability in the real scene.
Therefore, it is desirable to solve the problem of how to obtain a model without preference and how to improve the precision and robustness of semantic segmentation in an unfamiliar scene.
Disclosure of Invention
In view of the above drawbacks of the prior art, an object of the present invention is to provide a semantic segmentation method, system, medium, and apparatus, which are used to solve the problems of how to obtain a model without preference and how to improve the precision and robustness of semantic segmentation in an unfamiliar scene in the prior art.
To achieve the above and other related objects, the present invention provides a semantic segmentation method, comprising the steps of: acquiring a picture needing semantic segmentation, and recording the size of the picture as H multiplied by W multiplied by M, wherein H represents the height of the picture, W represents the width of the picture, and M represents the number of channels; performing edge processing on the picture to obtain an edge connected graph S; performing semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi(ii) a According to the edge connected graph S, a four-neighbor domain connected set F with the pixel of 0 is generated, each connected domain k in the F is traversed to find out a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain kminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000031
Where (h, w) e.k.
In an embodiment of the present invention, the performing the edge processing on the picture to obtain the edge connected graph S includes: performing edge detection on the picture through an edge detection network to generate an initial edge detection image E, wherein the size of the initial edge detection image E is H multiplied by W; thinning the initial edge detection graph E to obtain a thinned edge graph; carrying out binarization processing on the refined edge map to obtain an initial binarization edge map T; and carrying out large connected block full connected processing on the initial binarization edge image T to obtain a closed edge image S.
In the hairIn an embodiment of the invention, the semantic segmentation is performed on the picture based on a semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map PiThe method comprises the following steps: carrying out scene classification on the picture based on a scene recognition network to acquire the scene type of the picture; calling n semantic segmentation models which accord with scene attributes from a semantic segmentation model library based on the scene types; traversing each semantic segmentation model, and recording the current semantic segmentation model as i, i belongs to [1, n ]]Performing semantic segmentation on the picture to obtain an initial semantic segmentation result of the model i, wherein the initial semantic segmentation result comprises the following steps: semantic segmentation probability map PiDimension H × W × CiIn which C isiThe number of semantic categories which can be output by the semantic segmentation model; semantic segmentation label graph GiAnd the size is H multiplied by W.
In an embodiment of the present invention, the semantic segmentation model i with the minimum average information entropy is found out by traversing each connected domain k in FminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000032
Wherein (h, w) ∈ k includes: counting the semantic label voting condition of the semantic model i at the position of the pixel position set on the pixel position set in the connected domain k, and selecting the label with the most votes as the semantic label result of the semantic model i at the pixel position in the connected domain k; calculating the average information entropy l of each semantic segmentation model i on the set of pixel positions in the connected domain ki
Figure BDA0002629910840000033
Figure BDA0002629910840000034
Wherein (h, w) ∈ k, CiRepresenting the total number of semantic tags; finding the average information entropy liMinimum semantic segmentation model imin(ii) a Determining final semantic tags for all pixels within connected domain k
Figure BDA0002629910840000035
Where (h, w) e.k.
In order to achieve the above object, the present invention further provides a semantic segmentation system, including: the system comprises an acquisition module, an edge processing module, a semantic segmentation module and a label module; the acquisition module is used for acquiring a picture needing semantic segmentation, and recording the size of the picture as H multiplied by W multiplied by M, wherein H represents the height of the picture, W represents the width of the picture, and M represents the number of channels; the edge processing module is used for carrying out edge processing on the picture to obtain an edge connected graph S; the semantic segmentation module is used for performing semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi(ii) a The label module is used for generating a four-adjacent domain connected set F with a pixel of 0 according to the edge connected graph S, traversing each connected domain k in the F and respectively finding out a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain kminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000041
Where (h, w) e.k.
In an embodiment of the present invention, the performing the edge processing on the picture to obtain the edge connected graph S includes: performing edge detection on the picture through an edge detection network to generate an initial edge detection image E, wherein the size of the initial edge detection image E is H multiplied by W; thinning the initial edge detection graph E to obtain a thinned edge graph; carrying out binarization processing on the refined edge map to obtain an initial binarization edge map T; and carrying out large connected block full connected processing on the initial binarization edge image T to obtain a closed edge image S.
In an embodiment of the present invention, the semantic segmentation is performed on the picture based on a semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map PiThe method comprises the following steps: carrying out scene classification on the picture based on a scene recognition network to acquire the scene type of the picture; calling n semantic segmentation models which accord with scene attributes from a semantic segmentation model library based on the scene types; go throughEach semantic segmentation model records the current semantic segmentation model as i, i belongs to [1, n ]]Performing semantic segmentation on the picture to obtain an initial semantic segmentation result of the model i, wherein the initial semantic segmentation result comprises the following steps: semantic segmentation probability map PiDimension H × W × CiIn which C isiThe number of semantic categories which can be output by the semantic segmentation model; semantic segmentation label graph GiAnd the size is H multiplied by W.
In an embodiment of the present invention, the semantic segmentation model i with the minimum average information entropy is found out by traversing each connected domain k in FminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000042
Wherein (h, w) ∈ k includes: counting the semantic label voting condition of the semantic model i at the position of the pixel position set on the pixel position set in the connected domain k, and selecting the label with the most votes as the semantic label result of the semantic model i at the pixel position in the connected domain k; calculating the average information entropy l of each semantic segmentation model i on the set of pixel positions in the connected domain ki
Figure BDA0002629910840000043
Figure BDA0002629910840000044
Wherein (h, w) ∈ k, CiRepresenting the total number of semantic tags; finding the average information entropy liMinimum semantic segmentation model imin(ii) a Determining final semantic tags for all pixels within connected domain k
Figure BDA0002629910840000045
Where (h, w) e.k.
To achieve the above object, the present invention further provides a computer-readable storage medium having a computer program stored thereon, which when executed by a processor implements any of the above semantic segmentation methods.
In order to achieve the above object, the present invention further provides a semantic segmentation apparatus, including: a processor and a memory; the memory is used for storing a computer program; the processor is connected with the memory and is used for executing the computer program stored in the memory so as to enable the semantic segmentation device to execute any one of the semantic segmentation methods.
As described above, the semantic segmentation method, system, medium, and apparatus of the present invention have the following advantageous effects: the method is used for improving the precision and robustness of semantic segmentation in an unfamiliar scene.
Drawings
FIG. 1a is a schematic diagram of an ensemble learning basic system according to an embodiment of the semantic segmentation method of the present invention;
FIG. 1b is a flow chart of a Classification with an edge network according to an embodiment of the semantic segmentation method of the present invention;
FIG. 1c is a flow chart of a semantic segmentation method according to an embodiment of the invention;
FIG. 1d is a schematic diagram of an eight neighborhood region in an embodiment of the semantic segmentation method of the present invention;
FIG. 2 is a schematic diagram illustrating a semantic segmentation system according to an embodiment of the present invention;
FIG. 3 is a schematic diagram illustrating a semantic segmentation apparatus according to an embodiment of the present invention.
Description of the element reference numerals
21 acquisition module
22 edge processing module
23 semantic segmentation module
24 label module
31 processor
32 memory
Detailed Description
The embodiments of the present invention are described below with reference to specific embodiments, and other advantages and effects of the present invention will be easily understood by those skilled in the art from the disclosure of the present specification. The invention is capable of other and different embodiments and of being practiced or of being carried out in various ways, and its several details are capable of modification in various respects, all without departing from the spirit and scope of the present invention. It is to be noted that the features in the following embodiments and examples may be combined with each other without conflict.
It should be noted that the drawings provided in the following embodiments are only for illustrating the basic idea of the present invention, so that the components related to the present invention are only shown in the drawings rather than drawn according to the number, shape and size of the components in actual implementation, the type, quantity and proportion of the components in actual implementation can be changed freely, and the layout of the components can be more complicated.
The semantic segmentation method, the semantic segmentation system, the semantic segmentation medium and the semantic segmentation device are used for improving the precision and the robustness of semantic segmentation in strange scenes.
As shown in fig. 1c, in an embodiment, the semantic segmentation method of the present invention includes the following steps:
and step S11, acquiring a picture needing semantic segmentation, and recording the size of the picture as H multiplied by W multiplied by M, wherein H represents the height of the image, W represents the width of the image, and M represents the number of channels.
Specifically, the method further includes loading of various modules of the system, configuration of system parameters and the like.
Specifically, a picture requiring semantic segmentation is acquired from a perception system, and the picture size is recorded as H multiplied by W multiplied by 3, wherein H represents the image height (unit: pixel), W represents the image width (unit: pixel), and 3 represents 3 channels of red, green and blue.
And step S12, performing edge processing on the picture to obtain an edge connected graph S.
Specifically, the performing edge processing on the picture to obtain an edge connected graph S includes:
and carrying out edge detection on the picture through an edge detection network to generate an initial edge detection image E, wherein the size of the initial edge detection image E is H multiplied by W. Specifically, edge detection is performed on an input picture by using an existing edge detection network, and an initial edge detection graph E is generated, wherein the size of E is H × W (the height and the width of the input picture are consistent), an edge detection value is stored in each pixel position in E, the value is a real number, the value range is 0-255, and the larger the value is, the higher the probability that the pixel is an object edge point is.
And thinning the initial edge detection image E to obtain a thinned edge image. Specifically, the edge is represented by using gradient change, and since the area with large gradient change is usually wide, a normal arctan (y/x) is determined by using gradients in the directions of an x axis and a y axis on an image coordinate system, then whether the gradient value of the current pixel is a peak value (or a local maximum value) is judged in the normal direction, if so, the edge detection value (gradient value) of the current pixel is retained, if not, the edge detection value of the current pixel is suppressed, the edge detection value of the current pixel is set to be 0, and finally, a refined edge line is obtained, wherein the width is 1-2 pixels, and a refined edge image is obtained.
And carrying out binarization processing on the refined edge map to obtain an initial binarization edge map T. Specifically, all pixel points in the refined edge graph are traversed, if the edge detection value of the pixel point is greater than 0, the edge detection value is uniformly set to be a fixed value (for example, 255), and the pixel point is called as a valued pixel point; otherwise, the edge detection value of the pixel point keeps 0 unchanged. And after traversing is finished, obtaining a binary edge map T.
And carrying out large connected block full connected processing on the initial binarization edge image T to obtain a closed edge image S. Specifically, firstly, a connected domain set C of valued pixels in T is detected. Here, the connection refers to the concept of eight neighborhood connections in the image processing field, as shown in fig. 1d, it is assumed that a pixel point 4 has a value, and if one of eight points around the pixel point 4 has a value, the point is said to be connected with the pixel point 4, and the connected pixel points form a line.
Secondly, traversing each connected domain in the connected domain set C, recording the current connected domain as t, and processing the connected domain t as follows:
a. according to the eight-neighborhood connected method, traversing each pixel point in the connected domain t, and judging whether the number of valued pixel points in the eight-neighborhood is more than 2: if yes, the pixel point is a line segment inner point; otherwise, the pixel is an endpoint in the connected domain. Generating the endpoint set S in the connected domain after the traversal is finishedt
b. Traverse StIn the method, each endpoint is marked as j, valued pixel points in the neighborhood radius r of the endpoint j are searched, and all valued pixel points which belong to the same connected domain with the endpoint j form a pixel point set I in the connected domainj(ii) a All valued pixels which do not belong to the same connected domain with the end point j form a connected domain outside pixel point set Oj. If set OjIf not, then for OjAnd (4) sorting the inner pixel points according to the distance from the inner pixel point to the end point j, connecting the closest point with the end point j, and updating the binary edge map T. If set IjIf not, all pixels in the neighborhood radius r of the endpoint j are taken to form a graph, and connected domain detection is carried out on the graph to obtain a connected domain set L of the graphjAnd to IjThe inner pixel points are sorted according to the distance from the end point j, and sequentially judge I according to the sequence from far to nearjWhether the inner pixel point and the end point j belong to LjIf so, connecting the two points, updating the binary edge map T and jumping out of the loop.
After this step is completed, an edge connected graph S with closed edges can be obtained. The closed concept, still taking fig. 1d as an example, wherein if 0, 1, 2, 3, 5, 6, 7, 8 has a value, it means that each point has at least two neighboring points in the eight-neighboring domain having values, so the line passing through the above pixels is a closed line. Similarly, lines passing through 1, 3, 5, 7 are also closed lines. Firstly, prior, the edge detection network focuses more on the lower-level semantic information in the image, and the semantic segmentation network focuses more on the higher-level semantic information in the image. Edge detection is used to separate each object in the picture using lines, while semantic segmentation is used to determine why each pixel location in the picture should classify an object. Edge detection performs better in detail on the edge portion of an object and another object, and therefore the result of edge detection is used as a dividing line between the object and the object. These dividing lines divide the entire picture into several connected domains. However, after the edge detection network performs edge detection on a picture, the generated edge map is a probability value, which means a probability value of an edge at the pixel position (which may be understood as generating a map of H × W, where each position has a value ranging from 0 to 255, and the larger the value, the more likely the edge is). Most areas throughout the graph are of values greater than 0, except for the size of the values. If it is desired to separate the object from the object, it is most desirable that the edge map be a single pixel wide line. Therefore, the non-maximum suppression (NMS) operation is carried out on the edge map obtained by the edge detection, and a refined edge map is obtained. However, the lines in the edge graph are broken, so a method is proposed to close the lines, so that each object or each block in the object can form a connected domain. The accuracy and robustness of the segmentation can be further improved. This is because the characteristic improvement system that can be guided by edge detection has an effect of improving the segmentation accuracy of the edge portion of the object in the semantic segmentation result; the semantic fusion can pick the most expressive semantic segmentation model i from the k connected region given by the edge detection to mark the k region with a semantic label, so that the semantic segmentation model i with better expressive ability in each region k can be picked for different scenes, and the robustness of the whole system is improved.
Step S13, performing semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi
Specifically, the pre-trained semantic segmentation model i is trained, and the number of the semantic segmentation models i is i ∈ [1, n ].
Specifically, the semantic segmentation is carried out on the picture based on a semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map PiThe method comprises the following steps:
and carrying out scene classification on the picture based on a scene recognition network to acquire the scene type of the picture. Specifically, the pictures are subjected to scene classification by using an existing scene recognition network, and scene types (such as outdoor, indoor, outdoor and the like) are generated.
And calling the semantic segmentation models which accord with the scene attributes from a semantic segmentation model library based on the scene types, wherein the number of the semantic segmentation models is n. Specifically, according to the identified scene type, the semantic segmentation models conforming to the scene type are called from a semantic segmentation model library, and n semantic segmentation models conforming to the scene type are assumed.
Traversing each semantic segmentation model, recording the current semantic segmentation model as i, i belongs to [1, n ], performing semantic segmentation on the picture, and obtaining an initial semantic segmentation result of the model i, wherein the initial semantic segmentation result comprises the following steps:
semantic segmentation probability map PiDimension H × W × CiIn which C isiThe number of semantic categories which can be output by the semantic segmentation model;
Figure BDA0002629910840000081
represents PiPosition (d) is the value of (h, w, c);
Figure BDA0002629910840000082
probability value representing that the pixel (h, w) position of the input picture belongs to semantic type C, C ∈ [1, C]。
Semantic segmentation label graph GiThe size is H multiplied by W;
Figure BDA0002629910840000083
represents GiThe value of position (h, w) in (c),
Figure BDA0002629910840000084
a number indicating that the pixel (h, w) position of the input picture belongs to semantic type c. Each semantic segmentation model i carries out semantic segmentation once, so that the best-performing semantic segmentation model is selected for each region k, and the result of the semantic segmentation model is used as a fused label. GiAnd PiIs required for later calculation of information entropy and voting.
Step S14, according to the edge connected graph S, generating a four-adjacent domain connected set F with a pixel of 0, traversing each connected domain k in the F, and respectively finding out a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain kminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000085
Where (h, w) e.k. Such as P (h, w, c), c e (1, c)max). This totals cmaxAnd the value represents the probability value that the pixel point belongs to each class label c respectively. The quality of a model's expressive power is considered to be the deterministic size (confidence) for the target decision. For example, an expert may consider a thing to be not a good expert if the analysis of the thing is ambiguous, but consider a good expert if an expert considers that a thing must be something. The same is true in this probability P, for which c probability values an entropy is calculated, which low represents that the model determines that this point (h, w) is of a certain class, and vice versa.
Specifically, according to the edge connected graph S, generating the four-neighborhood connected set F with a pixel of 0 includes: the four neighborhood connected concept here is similar to eight neighborhood connected. As shown in fig. 1d, the four-neighborhood region of pixel 4 is pixels 1, 3, 5, 7. When the pixel 4 has a non-zero value, and a non-zero value exists somewhere on the positions 1, 3, 5 and 7, the positions are said to form four-neighborhood communication with the pixel 4.
Specifically, each connected domain k in the traversal F finds out a semantic segmentation model i with the minimum average information entropyminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000091
Wherein (h, w) ∈ k includes: traversing each connected domain k in F, and dividing the corresponding category c of the network i in the connected domain k due to different semantemesiThere may be differences, so the present invention trusts by calculating the confidence of different semantic probability maps and selecting the one with the highest confidence.
And counting the semantic label voting condition of the semantic model i at the position of the pixel position set on the pixel position set in the connected domain k, and selecting the label with the most votes as the semantic label result of the semantic model i at the pixel position in the connected domain k. And on a pixel position set in the connected domain k, counting the semantic label voting condition of each semantic model i at the positions according to the concept of a voting method in ensemble learning, and selecting the label with the most votes as a semantic label result of the semantic model at the pixel position in the connected domain k.
Calculating the average information entropy l of each semantic segmentation model i on the set of pixel positions in the connected domain ki
Figure BDA0002629910840000092
Figure BDA0002629910840000093
Wherein (h, w) ∈ k, CiRepresenting the total number of semantic tags. The connected component k is a concept defined from the length and width dimensions of the picture. For example, a picture in which a connected component k divides a basketball in the picture, all the pixels representing the basketball are all the elements in the connected component. That is, (h, w) is a set of pixels included in the connected domain belonging to k.
Finding the average information entropy liThe smallest semantic segmentation model is denoted as imin
Determine the final semantic label of all pixels in the connected component k as
Figure BDA0002629910840000094
Where (h, w) e.k.
Figure BDA0002629910840000095
Meaning semantically segmented label graph GiI is iminG of (A) timei
Figure BDA0002629910840000096
To represent
Figure BDA0002629910840000097
The value of position (h, w) in (f) represents the semantic segmentation label of the pixel (h, w).
Specifically, each connected domain in F is traversed, and a final semantic segmentation label graph R is generated.
As shown in fig. 2, in an embodiment, the semantic segmentation system of the present invention includes: obtainA fetching module 21, an edge processing module 22, a semantic segmentation module 23 and a labeling module 24. The obtaining module 21 is configured to obtain a picture to be subjected to semantic segmentation, and record that the size of the picture is H × W × M, where H represents an image height, W represents an image width, and M represents a channel number; the edge processing module 22 is configured to perform edge processing on the picture to obtain an edge connected graph S; the semantic segmentation module 23 is configured to perform semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi(ii) a The labeling module 24 is configured to generate a connected set F of four neighboring domains with a pixel of 0 according to the edge connected graph S, traverse each connected domain k in F, and find a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain k respectivelyminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000098
Where (h, w) e.k.
In an embodiment of the present invention, the performing the edge processing on the picture to obtain the edge connected graph S includes: performing edge detection on the picture through an edge detection network to generate an initial edge detection image E, wherein the size of the initial edge detection image E is H multiplied by W; thinning the initial edge detection graph E to obtain a thinned edge graph; carrying out binarization processing on the refined edge map to obtain an initial binarization edge map T; and carrying out large connected block full connected processing on the initial binarization edge image T to obtain a closed edge image S.
In an embodiment of the present invention, the semantic segmentation is performed on the picture based on a semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map PiThe method comprises the following steps: carrying out scene classification on the picture based on a scene recognition network to acquire the scene type of the picture; calling n semantic segmentation models which accord with scene attributes from a semantic segmentation model library based on the scene types; traversing each semantic segmentation model, and recording the current semantic segmentation model as i, i belongs to [1, n ]]Performing semantic segmentation on the picture to obtain the initial value of the model iThe starting semantic segmentation result comprises: semantic segmentation probability map PiDimension H × W × CiIn which C isiThe number of semantic categories which can be output by the semantic segmentation model; semantic segmentation label graph GiAnd the size is H multiplied by W.
In an embodiment of the present invention, the semantic segmentation model i with the minimum average information entropy is found out by traversing each connected domain k in FminDetermining the final semantic label of all pixels in the connected component k
Figure BDA0002629910840000101
Wherein (h, w) ∈ k includes: counting the semantic label voting condition of the semantic model i at the position of the pixel position set on the pixel position set in the connected domain k, and selecting the label with the most votes as the semantic label result of the semantic model i at the pixel position in the connected domain k; calculating the average information entropy l of each semantic segmentation model i on the set of pixel positions in the connected domain ki
Figure BDA0002629910840000102
Figure BDA0002629910840000103
Wherein (h, w) ∈ k, CiRepresenting the total number of semantic tags; finding the average information entropy liMinimum semantic segmentation model imin(ii) a Determining final semantic tags for all pixels within connected domain k
Figure BDA0002629910840000104
Where (h, w) e.k.
It should be noted that the structures and principles of the obtaining module 21, the edge processing module 22, the semantic segmentation module 23, and the label module 24 correspond to the steps in the semantic segmentation method one to one, and therefore are not described herein again.
It should be noted that the division of the modules of the above system is only a logical division, and the actual implementation may be wholly or partially integrated into one physical entity, or may be physically separated. And these modules can be realized in the form of software called by processing element; or may be implemented entirely in hardware; and part of the modules can be realized in the form of calling software by the processing element, and part of the modules can be realized in the form of hardware. For example, the x module may be a processing element that is set up separately, or may be implemented by being integrated in a chip of the apparatus, or may be stored in a memory of the apparatus in the form of program code, and the function of the x module may be called and executed by a processing element of the apparatus. Other modules are implemented similarly. In addition, all or part of the modules can be integrated together or can be independently realized. The processing element described herein may be an integrated circuit having signal processing capabilities. In implementation, each step of the above method or each module above may be implemented by an integrated logic circuit of hardware in a processor element or an instruction in the form of software.
For example, the above modules may be one or more integrated circuits configured to implement the above methods, such as: one or more Specific Integrated circuits (ASICs), or one or more Microprocessors (MPUs), or one or more Field Programmable Gate Arrays (FPGAs), etc. For another example, when one of the above modules is implemented in the form of a Processing element scheduler code, the Processing element may be a general-purpose processor, such as a Central Processing Unit (CPU) or other processor capable of calling program code. For another example, these modules may be integrated together and implemented in the form of a system-on-a-chip (SOC).
In an embodiment of the present invention, the present invention further includes a computer-readable storage medium, on which a computer program is stored, where the computer program is executed by a processor to implement any of the above semantic segmentation methods.
Those of ordinary skill in the art will understand that: all or part of the steps for implementing the above method embodiments may be performed by hardware associated with a computer program. The aforementioned computer program may be stored in a computer readable storage medium. When executed, the program performs steps comprising the method embodiments described above; and the aforementioned storage medium includes: various media that can store program codes, such as ROM, RAM, magnetic or optical disks.
As shown in fig. 3, in an embodiment, the semantic segmentation apparatus of the present invention includes: a processor 31 and a memory 32; the memory 32 is for storing a computer program; the processor 31 is connected to the memory 32, and is configured to execute the computer program stored in the memory 32, so as to enable the semantic segmentation apparatus to execute any one of the semantic segmentation methods.
Specifically, the memory 32 includes: various media that can store program codes, such as ROM, RAM, magnetic disk, U-disk, memory card, or optical disk.
Preferably, the Processor 31 may be a general-purpose Processor, including a Central Processing Unit (CPU), a Network Processor (NP), and the like; the Integrated Circuit may also be a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic device, or discrete hardware components.
In summary, the semantic segmentation method, system, medium and apparatus of the present invention are used to improve the precision and robustness of semantic segmentation in strange scenes. Therefore, the invention effectively overcomes various defects in the prior art and has high industrial utilization value.
The foregoing embodiments are merely illustrative of the principles and utilities of the present invention and are not intended to limit the invention. Any person skilled in the art can modify or change the above-mentioned embodiments without departing from the spirit and scope of the present invention. Accordingly, it is intended that all equivalent modifications or changes which can be made by those skilled in the art without departing from the spirit and technical spirit of the present invention be covered by the claims of the present invention.

Claims (10)

1. A semantic segmentation method is characterized by comprising the following steps:
acquiring a picture needing semantic segmentation, and recording the size of the picture as H multiplied by W multiplied by M, wherein H represents the height of the picture, W represents the width of the picture, and M represents the number of channels;
performing edge processing on the picture to obtain an edge connected graph S;
performing semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi
According to the edge connected graph S, a four-neighbor domain connected set F with the pixel of 0 is generated, each connected domain k in the F is traversed, and a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain k is found out respectivelyminDetermining the final semantic label of all pixels in the connected component k
Figure FDA0002629910830000011
Where (h, w) e.k.
2. The semantic segmentation method according to claim 1, wherein the edge processing the picture to obtain an edge connected graph S comprises:
performing edge detection on the picture through an edge detection network to generate an initial edge detection image E, wherein the size of the initial edge detection image E is H multiplied by W;
thinning the initial edge detection graph E to obtain a thinned edge graph;
carrying out binarization processing on the refined edge map to obtain an initial binarization edge map T;
and carrying out large connected block full connected processing on the initial binarization edge image T to obtain a closed edge image S.
3. The semantic segmentation method according to claim 1, wherein the semantic segmentation is performed on the picture based on a semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map PiThe method comprises the following steps:
carrying out scene classification on the picture based on a scene recognition network to acquire the scene type of the picture;
calling n semantic segmentation models which accord with scene attributes from a semantic segmentation model library based on the scene types;
traversing each semantic segmentation model, recording the current semantic segmentation model as i, i belongs to [1, n ], performing semantic segmentation on the picture, and obtaining an initial semantic segmentation result of the model i, wherein the initial semantic segmentation result comprises the following steps:
semantic segmentation probability map PiDimension H × W × CiIn which C isiThe number of semantic categories which can be output by the semantic segmentation model i;
semantic segmentation label graph GiAnd the size is H multiplied by W.
4. The semantic segmentation method according to claim 1, wherein each connected component k in the traversal F finds the semantic segmentation model i with the smallest average entropyminDetermining the final semantic label of all pixels in the connected component k
Figure FDA0002629910830000021
Wherein (h, w) ∈ k includes:
counting the semantic label voting condition of the semantic model i at the position of the pixel position set on the pixel position set in the connected domain k, and selecting the label with the most votes as the semantic label result of the semantic model i at the pixel position in the connected domain k;
calculating the average information entropy l of each semantic segmentation model i on the set of pixel positions in the connected domain ki
Figure FDA0002629910830000022
Figure FDA0002629910830000023
Wherein (h, w) ∈ k, CiRepresenting the total number of semantic tags;
finding the average information entropy liMinimum sizeSemantic segmentation model imin
Determining final semantic tags for all pixels within connected domain k
Figure FDA0002629910830000024
Where (h, w) e.k.
5. A semantic segmentation system, comprising: the system comprises an acquisition module, an edge processing module, a semantic segmentation module and a label module;
the acquisition module is used for acquiring a picture needing semantic segmentation, and recording the size of the picture as H multiplied by W multiplied by M, wherein H represents the height of the picture, W represents the width of the picture, and M represents the number of channels;
the edge processing module is used for carrying out edge processing on the picture to obtain an edge connected graph S;
the semantic segmentation module is used for performing semantic segmentation on the picture based on a pre-trained semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map Pi
The label module is used for generating a four-adjacent domain connected set F with a pixel of 0 according to the edge connected graph S, traversing each connected domain k in the F and respectively finding out a semantic segmentation model i with the minimum average information entropy corresponding to each connected domain kminDetermining the final semantic label of all pixels in the connected component k
Figure FDA0002629910830000025
Where (h, w) e.k.
6. The semantic segmentation system according to claim 5, wherein the edge processing the picture to obtain an edge connected graph S comprises:
performing edge detection on the picture through an edge detection network to generate an initial edge detection image E, wherein the size of the initial edge detection image E is H multiplied by W;
thinning the initial edge detection graph E to obtain a thinned edge graph;
carrying out binarization processing on the refined edge map to obtain an initial binarization edge map T;
and carrying out large connected block full connected processing on the initial binarization edge image T to obtain a closed edge image S.
7. The semantic segmentation system according to claim 5, wherein the semantic segmentation is performed on the picture based on a semantic segmentation model i; generating a semantic tag graph GiAnd semantic probability map PiThe method comprises the following steps:
carrying out scene classification on the picture based on a scene recognition network to acquire the scene type of the picture;
calling n semantic segmentation models which accord with scene attributes from a semantic segmentation model library based on the scene types;
traversing each semantic segmentation model, recording the current semantic segmentation model as i, i belongs to [1, n ], performing semantic segmentation on the picture, and obtaining an initial semantic segmentation result of the model i, wherein the initial semantic segmentation result comprises the following steps:
semantic segmentation probability map PiDimension H × W × CiIn which C isiThe number of semantic categories which can be output by the semantic segmentation model i;
semantic segmentation label graph GiAnd the size is H multiplied by W.
8. The semantic segmentation system according to claim 5, wherein each connected component k in the traversal F finds the semantic segmentation model i with the smallest average entropyminDetermining the final semantic label of all pixels in the connected component k
Figure FDA0002629910830000031
Wherein (h, w) ∈ k includes:
counting the semantic label voting condition of the semantic model i at the position of the pixel position set on the pixel position set in the connected domain k, and selecting the label with the most votes as the semantic label result of the semantic model i at the pixel position in the connected domain k;
calculating the average information entropy l of each semantic segmentation model i on the set of pixel positions in the connected domain ki
Figure FDA0002629910830000032
Figure FDA0002629910830000033
Wherein (h, w) ∈ k, CiRepresenting the total number of semantic tags;
finding the average information entropy liMinimum semantic segmentation model imin
Determining final semantic tags for all pixels within connected domain k
Figure FDA0002629910830000034
Where (h, w) e.k.
9. A computer-readable storage medium, on which a computer program is stored, the computer program being executable by a processor to implement the semantic segmentation method according to any one of claims 1 to 4.
10. A semantic segmentation apparatus, comprising: a processor and a memory;
the memory is used for storing a computer program;
the processor is connected to the memory and is configured to execute the computer program stored in the memory to cause the semantic segmentation apparatus to perform the semantic segmentation method according to any one of claims 1 to 4.
CN202010808133.7A 2020-08-12 2020-08-12 Semantic segmentation method, system, medium and device Active CN111931782B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010808133.7A CN111931782B (en) 2020-08-12 2020-08-12 Semantic segmentation method, system, medium and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010808133.7A CN111931782B (en) 2020-08-12 2020-08-12 Semantic segmentation method, system, medium and device

Publications (2)

Publication Number Publication Date
CN111931782A true CN111931782A (en) 2020-11-13
CN111931782B CN111931782B (en) 2024-03-01

Family

ID=73310815

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010808133.7A Active CN111931782B (en) 2020-08-12 2020-08-12 Semantic segmentation method, system, medium and device

Country Status (1)

Country Link
CN (1) CN111931782B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112633148A (en) * 2020-12-22 2021-04-09 杭州景联文科技有限公司 Method and system for detecting authenticity of signature fingerprint
CN112700450A (en) * 2021-01-15 2021-04-23 北京睿芯高通量科技有限公司 Image segmentation method and system based on ensemble learning
CN114049569A (en) * 2022-01-13 2022-02-15 自然资源部第三地理信息制图院 Deep learning model performance evaluation method and system

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107424159A (en) * 2017-07-28 2017-12-01 西安电子科技大学 Image, semantic dividing method based on super-pixel edge and full convolutional network
CN109886238A (en) * 2019-03-01 2019-06-14 湖北无垠智探科技发展有限公司 Unmanned plane Image Change Detection algorithm based on semantic segmentation
CN109919159A (en) * 2019-01-22 2019-06-21 西安电子科技大学 A kind of semantic segmentation optimization method and device for edge image
CN110322445A (en) * 2019-06-12 2019-10-11 浙江大学 A kind of semantic segmentation method based on maximization prediction and impairment correlations function between label
CN110443818A (en) * 2019-07-02 2019-11-12 中国科学院计算技术研究所 A kind of Weakly supervised semantic segmentation method and system based on scribble
CN110570427A (en) * 2019-07-19 2019-12-13 武汉珈和科技有限公司 Remote sensing image semantic segmentation method and device fusing edge detection
US20200134375A1 (en) * 2017-08-01 2020-04-30 Beijing Sensetime Technology Development Co., Ltd. Semantic segmentation model training methods and apparatuses, electronic devices, and storage media
CN111127499A (en) * 2019-12-20 2020-05-08 北京工业大学 Security inspection image cutter detection segmentation method based on semantic contour information
CN111161314A (en) * 2019-12-17 2020-05-15 中国科学院上海微系统与信息技术研究所 Target object position area determining method and device, electronic equipment and storage medium
CN111192265A (en) * 2019-12-25 2020-05-22 中国科学院上海微系统与信息技术研究所 Point cloud based semantic instance determination method and device, electronic equipment and storage medium
US10672129B1 (en) * 2019-03-22 2020-06-02 Lunit Inc. Method for semantic segmentation and apparatus thereof
CN111242027A (en) * 2020-01-13 2020-06-05 北京工业大学 Unsupervised learning scene feature rapid extraction method fusing semantic information
CN111340047A (en) * 2020-02-28 2020-06-26 江苏实达迪美数据处理有限公司 Image semantic segmentation method and system based on multi-scale feature and foreground and background contrast
WO2020156303A1 (en) * 2019-01-30 2020-08-06 广州市百果园信息技术有限公司 Method and apparatus for training semantic segmentation network, image processing method and apparatus based on semantic segmentation network, and device and storage medium

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107424159A (en) * 2017-07-28 2017-12-01 西安电子科技大学 Image, semantic dividing method based on super-pixel edge and full convolutional network
US20200134375A1 (en) * 2017-08-01 2020-04-30 Beijing Sensetime Technology Development Co., Ltd. Semantic segmentation model training methods and apparatuses, electronic devices, and storage media
CN109919159A (en) * 2019-01-22 2019-06-21 西安电子科技大学 A kind of semantic segmentation optimization method and device for edge image
WO2020156303A1 (en) * 2019-01-30 2020-08-06 广州市百果园信息技术有限公司 Method and apparatus for training semantic segmentation network, image processing method and apparatus based on semantic segmentation network, and device and storage medium
CN109886238A (en) * 2019-03-01 2019-06-14 湖北无垠智探科技发展有限公司 Unmanned plane Image Change Detection algorithm based on semantic segmentation
US10672129B1 (en) * 2019-03-22 2020-06-02 Lunit Inc. Method for semantic segmentation and apparatus thereof
CN110322445A (en) * 2019-06-12 2019-10-11 浙江大学 A kind of semantic segmentation method based on maximization prediction and impairment correlations function between label
CN110443818A (en) * 2019-07-02 2019-11-12 中国科学院计算技术研究所 A kind of Weakly supervised semantic segmentation method and system based on scribble
CN110570427A (en) * 2019-07-19 2019-12-13 武汉珈和科技有限公司 Remote sensing image semantic segmentation method and device fusing edge detection
CN111161314A (en) * 2019-12-17 2020-05-15 中国科学院上海微系统与信息技术研究所 Target object position area determining method and device, electronic equipment and storage medium
CN111127499A (en) * 2019-12-20 2020-05-08 北京工业大学 Security inspection image cutter detection segmentation method based on semantic contour information
CN111192265A (en) * 2019-12-25 2020-05-22 中国科学院上海微系统与信息技术研究所 Point cloud based semantic instance determination method and device, electronic equipment and storage medium
CN111242027A (en) * 2020-01-13 2020-06-05 北京工业大学 Unsupervised learning scene feature rapid extraction method fusing semantic information
CN111340047A (en) * 2020-02-28 2020-06-26 江苏实达迪美数据处理有限公司 Image semantic segmentation method and system based on multi-scale feature and foreground and background contrast

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
D.MARMANIS等: "Classification with an edge:Improving semantic image segmentation with boundary detection", ARXIV, pages 31 *
JINGANG TAN等: "HCFS3D: Hierarchical coupled feature selection network for 3D semantic and instance segmentation", IMAGE AND VISION COMPUTING, pages 1 - 12 *
RUIJUN SHU等: "CSF: Close d-mask-guide d semantic fusion method for semantic perception of unknown scenes", PATTERN RECOGNITION LETTERS, pages 101 - 107 *
舒睿俊 等: "基于SDR的智能电网系统PBCH加扰实现与优化", 电子设计工程, vol. 26, no. 6, pages 81 - 85 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112633148A (en) * 2020-12-22 2021-04-09 杭州景联文科技有限公司 Method and system for detecting authenticity of signature fingerprint
CN112633148B (en) * 2020-12-22 2022-08-09 杭州景联文科技有限公司 Method and system for detecting authenticity of signature fingerprint
CN112700450A (en) * 2021-01-15 2021-04-23 北京睿芯高通量科技有限公司 Image segmentation method and system based on ensemble learning
CN114049569A (en) * 2022-01-13 2022-02-15 自然资源部第三地理信息制图院 Deep learning model performance evaluation method and system
CN114049569B (en) * 2022-01-13 2022-03-18 自然资源部第三地理信息制图院 Deep learning model performance evaluation method and system

Also Published As

Publication number Publication date
CN111931782B (en) 2024-03-01

Similar Documents

Publication Publication Date Title
WO2019100723A1 (en) Method and device for training multi-label classification model
CN111931782A (en) Semantic segmentation method, system, medium, and apparatus
US9519660B2 (en) Information processing apparatus, clustering method, and recording medium storing clustering program
CN108920580A (en) Image matching method, device, storage medium and terminal
US20060221090A1 (en) Image processing apparatus, method, and program
CN103927510B (en) Pattern recognition device and image-recognizing method
US8036468B2 (en) Invariant visual scene and object recognition
CN106202329A (en) Sample data process, data identification method and device, computer equipment
CN111310800B (en) Image classification model generation method, device, computer equipment and storage medium
JP6997369B2 (en) Programs, ranging methods, and ranging devices
KR20170026222A (en) Method and device for classifying an object of an image and corresponding computer program product and computer-readable medium
CN111738036B (en) Image processing method, device, equipment and storage medium
CN110852327A (en) Image processing method, image processing device, electronic equipment and storage medium
CN115131363A (en) Positioning method and device based on semantic information and terminal equipment
CN112241736B (en) Text detection method and device
JP2021051589A5 (en)
CN111914809A (en) Target object positioning method, image processing method, device and computer equipment
CN116958809A (en) Remote sensing small sample target detection method for feature library migration
CN112785595A (en) Target attribute detection, neural network training and intelligent driving method and device
CN111062385A (en) Network model construction method and system for image text information detection
CN113837236B (en) Method and device for identifying target object in image, terminal equipment and storage medium
CN112750124B (en) Model generation method, image segmentation method, model generation device, image segmentation device, electronic equipment and storage medium
CN111435537B (en) Model training method and device and pose optimization method and device based on mosaic
CN113033612A (en) Image classification method and device
CN116071625B (en) Training method of deep learning model, target detection method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant