CN113487624B - Human body instance segmentation method, terminal equipment and storage medium - Google Patents
Human body instance segmentation method, terminal equipment and storage medium Download PDFInfo
- Publication number
- CN113487624B CN113487624B CN202110707258.5A CN202110707258A CN113487624B CN 113487624 B CN113487624 B CN 113487624B CN 202110707258 A CN202110707258 A CN 202110707258A CN 113487624 B CN113487624 B CN 113487624B
- Authority
- CN
- China
- Prior art keywords
- output
- steps
- method comprises
- carrying
- treatment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/11—Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
Abstract
The invention relates to a human body instance segmentation method, a terminal device and a storage medium, wherein the method comprises the following steps: s1: calculating the distance between pixels through different distance calculation algorithms based on the color information of the color picture, and converting all the distance calculation results into picture features through a convolution layer after connecting all the distance calculation results in series; s2: dividing a picture into a plurality of blocks, and obtaining a feature representation vector E of the picture according to the block feature and the pixel position of each block; s3: constructing a converter encoder/decoder network, and converting the feature representation vector E into an optimized feature F through the converter encoder/decoder network d The method comprises the steps of carrying out a first treatment on the surface of the S4: will optimize feature F d And F is obtained after connection with the characteristic expression vector E c The method comprises the steps of carrying out a first treatment on the surface of the S5: by means of attentiveness mechanism to F c After the important features in the model are aggregated, the aggregation result is input into a convolutional neural network to obtain a human body example segmentation result. The method fully utilizes the advantages of the transducer to optimize the human body instance segmentation result, and conventional operations such as clustering, candidate frame and non-maximum value suppression operation are not needed.
Description
Technical Field
The present invention relates to the field of computer vision, and in particular, to a human body instance segmentation method, a terminal device, and a storage medium.
Background
Current example segmentation techniques fall into two main categories: 1) Candidate target frame-free algorithms (Proposal-free methods) whose core idea is to learn a distinguishable feature for each point and then cluster the points of the same instance together by a clustering method such as mean-shift. The problem with this type of approach is that the examples that are eventually clustered together are less targeted. Furthermore, the time cost of such process post-treatment steps is typically high; 2) Candidate target frame based algorithms (Proposal-based methods) typically generate a large number of candidate frames prior to the second stage of optimization. But such an approach is difficult to work with if the deviation of the candidate boxes is relatively large. At the same time, operations similar to non-maximum suppression are needed to select dense candidate frames, which has the disadvantage of being extremely computationally intensive, and when there is a highly overlapping object, the non-maximum suppression considers one of the candidate frames as a repeated candidate region and then deletes it. This means that it is difficult to deal with cases where objects overlap highly based on an instance of a candidate box.
Disclosure of Invention
In order to solve the problems, the invention provides a human body instance segmentation method, a terminal device and a storage medium, which overcome the defects of a traditional instance segmentation algorithm aiming at a non-candidate frame algorithm and a candidate frame-based algorithm, and do not need clustering, candidate frames and non-maximum suppression operation.
The specific scheme is as follows:
a human body instance segmentation method, comprising the steps of:
s1: when a color picture containing a human body is received, calculating the distance between pixels through different distance calculation algorithms based on the color information of the picture, and converting all the distance calculation results into picture features through a convolution layer after connecting all the distance calculation results in series;
s2: dividing a picture into a plurality of blocks, acquiring block characteristics corresponding to each block according to pixel positions and picture characteristics corresponding to each block, and acquiring a characteristic representation vector E of the picture according to the block characteristics and the pixel positions corresponding to each block;
s3: constructing a converter encoder/decoder network, and converting the feature representation vector E into an optimized feature F through the converter encoder/decoder network d ;
S4: will optimize feature F d And F is obtained after connection with the characteristic expression vector E c ;
S5: by means of attentiveness mechanism to F c After the important features in the model are aggregated, the aggregation result is input into a convolutional neural network to obtain a human body example segmentation result.
Further, the distance calculation algorithm includes a Euclidean distance, a Chebyshev distance, and a cosine distance.
Further, the feature expression vector e= [ f 1 +p 1 ,f 2 +p 2 ,...,f N +p N ]Wherein f N Block characteristics representing the nth block, p N The pixel position of the nth block is represented, and N represents the number of blocks.
Further, the implementation process of step S3 is as follows:
s301: inputting the feature expression vector E into a transducer coding block, and outputting x by a multi-head self-attention mechanism in the transducer coding block 1 Then, the output of the multi-head self-attention mechanism is connected with the characteristic expression vector E through residual connection and then x is output 2 The method comprises the steps of carrying out a first treatment on the surface of the For x 2 Normalized and output x 3 The method comprises the steps of carrying out a first treatment on the surface of the Will x 3 Input fully connected network layer post output with activation function x 4 The method comprises the steps of carrying out a first treatment on the surface of the Will x 4 Input of fully connected network layer post output x without activation function 5 The method comprises the steps of carrying out a first treatment on the surface of the Setting x 5 An output encoding a block for a transducer;
s302: stacking a plurality of transducer encoding blocks to form a transducer encoder so that an encoding information matrix M of the output of the transducer encoder is consistent with the dimension of the feature representation vector E;
s303: taking the coding information matrix M output by the transducer encoder as the input of the transducer decoder; the transform decoding block contains a two-pass multi-headed self-attention mechanism;
the output of the first self-attention mechanism is y 1 The method comprises the steps of carrying out a first treatment on the surface of the The output M of the transducer encoder and the output y of the first self-attention mechanism are connected by a residual error 1 Post-connection output y 2 The method comprises the steps of carrying out a first treatment on the surface of the For y 2 Normalized and output y 3 ;
Second self-attentionThe input of the first decoding layer in the mechanism is characteristic representation vector E, and the corresponding output is y 4 The method comprises the steps of carrying out a first treatment on the surface of the Output y of the first self-attention mechanism via residual connection 1 Output y from the second self-attention mechanism 4 Post-connection output y 5 The method comprises the steps of carrying out a first treatment on the surface of the For y 5 Normalized and output y 6 The method comprises the steps of carrying out a first treatment on the surface of the Will y 6 Input fully connected network layer post output y with activation function 7 The method comprises the steps of carrying out a first treatment on the surface of the Will y 7 Input fully connected network layer post output y without activation function 8 The method comprises the steps of carrying out a first treatment on the surface of the Setting y 8 An output for the transform decoding block;
s304: stacking multiple transducer decoding blocks to form a transducer decoder, the output of which is the optimized feature F d 。
Further, in step S4, the feature F is optimized by a Concat operation d Connected to the feature representation vector E.
The human body instance segmentation terminal equipment comprises a processor, a memory and a computer program stored in the memory and capable of running on the processor, wherein the processor realizes the steps of the method according to the embodiment of the invention when executing the computer program.
A computer readable storage medium storing a computer program which, when executed by a processor, implements the steps of the method described above for embodiments of the present invention.
According to the method, the device and the system, the advantages of the Transformer are fully utilized to optimize the human body example segmentation result, the model with better interpretability is generated through the multi-head self-attention mechanism, and conventional operations such as clustering, candidate frame and non-maximum value inhibition operation are not needed.
Drawings
Fig. 1 is a flowchart of a first embodiment of the present invention.
Detailed Description
For further illustration of the various embodiments, the invention is provided with the accompanying drawings. The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate embodiments and together with the description, serve to explain the principles of the embodiments. With reference to these matters, one of ordinary skill in the art will understand other possible embodiments and advantages of the present invention.
The invention will now be further described with reference to the drawings and detailed description.
Embodiment one:
the embodiment of the invention provides a human body example segmentation method, as shown in fig. 1, which is a flow chart of the human body example segmentation method according to the embodiment of the invention, wherein the method comprises the following steps:
s1: when a color picture containing a human body is received, calculating the distance between pixels through different distance calculation algorithms based on the color information of the picture, and converting all the distance calculation results into picture features through a convolution layer after connecting all the distance calculation results in series.
In this embodiment the color picture is denoted as xR 3xHxW Wherein 3 represents R, G, B three elements, H represents the height of the picture, and W represents the width of the picture.
The distance calculation algorithm comprises a Euclidean distance, a Chebyshev distance and a cosine distance, and is used for calculating the distance between two pixels, and the calculation formulas are respectively as follows:
euclidean distance
Chebyshev distance d 2 =max(|R 1 -R 2 |,|G 1 -G 2 |,|B 1 -B 2 |);
Cosine distance
Wherein R is 1 、G 1 、B 1 Respectively representing chromaticity values of red, green and blue channels of the first pixel point, R 2 、G 2 、B 2 And respectively representing the chromaticity values of the red, green and blue channels of the second pixel point.
In other embodiments, other distance calculation algorithms may be selected as desired, and are not limited herein.
It should be noted that, all the distance calculation results are connected in series, that is, the pixel points are connected in series with the distances calculated by all the distance calculation algorithms of each pixel point except the pixel points, and assuming that n pixel points are total, a× (n-1) distances are connected in series for each pixel point, and a represents the number of the distance calculation algorithms, which is 3 in this embodiment.
The image characteristics converted by the convolution layer are fixed dimensions, namely d dimensions, and the value of d can be set by a person skilled in the art, and the method is not limited herein.
S2: dividing a picture into a plurality of blocks (patch), acquiring the characteristics corresponding to each block according to the pixel position and the picture characteristics corresponding to each block, and acquiring the characteristic representation vector E of the picture according to the characteristics and the pixel position corresponding to each block.
In this embodiment, the number of the set blocks is N, and the calculation formula of N is:where s represents the area of the block.
Feature representation vector e= [ f 1 +p 1 ,f 2 +p 2 ,...,f N +p N ]Wherein f N Block characteristics representing the nth block, p N Representing the pixel location of the nth block.
By combining picture convolution with color coding, block features and pixel locations are learned so that the constructed feature representation vector can be better adapted to the input of a transform encoder.
S3: constructing a converter encoder/decoder network, and converting the feature representation vector E into an optimized feature F through the converter encoder/decoder network d 。
The implementation procedure of step S3 in this embodiment is as follows:
s301: inputting the feature expression vector E into a transducer coding block, and outputting x by a multi-head self-attention mechanism in the transducer coding block 1 Then, the output of the multi-head self-attention mechanism is connected with the characteristic expression vector E through residual connection and then x is output 2 The method comprises the steps of carrying out a first treatment on the surface of the For x 2 Normalized and output x 3 The method comprises the steps of carrying out a first treatment on the surface of the Will x 3 Input fully connected network layer post output with activation function x 4 The method comprises the steps of carrying out a first treatment on the surface of the Will x 4 Input of fully connected network layer post output x without activation function 5 The method comprises the steps of carrying out a first treatment on the surface of the Setting x 5 An output encoding a block for a transducer;
s302: stacking a plurality of transducer encoding blocks to form a transducer encoder so that an encoding information matrix M of the output of the transducer encoder is consistent with the dimension of the feature representation vector E;
s303: taking the coding information matrix M output by the transducer encoder as the input of the transducer decoder; the transform decoding block contains a two-pass multi-headed self-attention mechanism;
the output of the first self-attention mechanism is y 1 The method comprises the steps of carrying out a first treatment on the surface of the The output M of the transducer encoder and the output y of the first self-attention mechanism are connected by a residual error 1 Post-connection output y 2 The method comprises the steps of carrying out a first treatment on the surface of the For y 2 Normalized and output y 3 ;
The input of the first layer decoding layer in the second self-attention mechanism is the feature representation vector E, and the corresponding output is y 4 The method comprises the steps of carrying out a first treatment on the surface of the Output y of the first self-attention mechanism via residual connection 1 Output y from the second self-attention mechanism 4 Post-connection output y 5 The method comprises the steps of carrying out a first treatment on the surface of the For y 5 Normalized and output y 6 The method comprises the steps of carrying out a first treatment on the surface of the Will y 6 Input fully connected network layer post output y with activation function 7 The method comprises the steps of carrying out a first treatment on the surface of the Will y 7 Input fully connected network layer post output y without activation function 8 The method comprises the steps of carrying out a first treatment on the surface of the Setting y 8 An output for the transform decoding block;
s304: stacking multiple transducer decoding blocks to form a transducer decoder, the output of which is the optimized feature F d 。
In the above steps, the normalization process is used to accelerate network convergence.
S4: will optimize feature F d And F is obtained after connection with the characteristic expression vector E c 。
In this embodiment feature F will be optimized by the Concat operation d Connected to the feature representation vector E.
S5: by means of attentiveness mechanism to F c After polymerization of the important features, the polymerization result F A Inputting the human body example segmentation result into a convolutional neural network.
The discrimination of feature maps can be enhanced by a mechanism of attention.
The embodiment of the invention is an effective improvement on the traditional candidate frame-based algorithm and the candidate frame-free algorithm, fully utilizes the advantages of a transducer to optimize the human body example segmentation result, generates a model with more interpretability through a multi-head self-attention mechanism, and does not need conventional operations such as clustering, candidate frame and non-maximum value inhibition operation and the like.
Embodiment two:
the invention also provides a human body instance segmentation terminal device, which comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the steps in the method embodiment of the first embodiment of the invention are realized when the processor executes the computer program.
Further, as an executable scheme, the human body instance segmentation terminal device may be a computing device such as a desktop computer, a notebook computer, a palm computer, a cloud server, and the like. The human body instance segmentation terminal device may include, but is not limited to, a processor, a memory. It will be understood by those skilled in the art that the above-described constituent structure of the human body example division terminal device is merely an example of the human body example division terminal device, and does not constitute limitation of the human body example division terminal device, and may include more or less components than those described above, or may combine some components, or different components, for example, the human body example division terminal device may further include an input/output device, a network access device, a bus, etc., which is not limited in the embodiment of the present invention.
Further, as an implementation, the processor may be a central processing unit (Central Processing Unit, CPU), other general purpose processor, digital signal processor (Digital Signal Processor, DSP), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), field programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic device, discrete gate or transistor logic device, discrete hardware components, etc. The general processor may be a microprocessor or the processor may be any conventional processor or the like, which is a control center of the human body instance division terminal device, and connects the respective parts of the whole human body instance division terminal device using various interfaces and lines.
The memory may be used to store the computer program and/or the module, and the processor may implement various functions of the human body instance segmentation terminal device by running or executing the computer program and/or the module stored in the memory and invoking data stored in the memory. The memory may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, at least one application program required for a function; the storage data area may store data created according to the use of the cellular phone, etc. In addition, the memory may include high-speed random access memory, and may also include non-volatile memory, such as a hard disk, memory, plug-in hard disk, smart Media Card (SMC), secure Digital (SD) Card, flash Card (Flash Card), at least one disk storage device, flash memory device, or other volatile solid-state storage device.
The present invention also provides a computer readable storage medium storing a computer program which when executed by a processor implements the steps of the above-described method of an embodiment of the present invention.
The modules/units integrated by the human body instance segmentation terminal device may be stored in a computer readable storage medium if implemented in the form of software functional units and sold or used as a separate product. Based on such understanding, the present invention may implement all or part of the flow of the method of the above embodiment, or may be implemented by a computer program to instruct related hardware, where the computer program may be stored in a computer readable storage medium, and when the computer program is executed by a processor, the computer program may implement the steps of each of the method embodiments described above. Wherein the computer program comprises computer program code which may be in source code form, object code form, executable file or some intermediate form etc. The computer readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a U disk, a removable hard disk, a magnetic disk, an optical disk, a computer Memory, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a software distribution medium, and so forth.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.
Claims (6)
1. A human body instance segmentation method, characterized by comprising the steps of:
s1: when a color picture containing a human body is received, calculating the distance between pixels through different distance calculation algorithms based on the color information of the picture, and converting all the distance calculation results into picture features through a convolution layer after connecting all the distance calculation results in series;
s2: dividing a picture into a plurality of blocks, acquiring block characteristics corresponding to each block according to pixel positions and picture characteristics corresponding to each block, and acquiring a characteristic representation vector E of the picture according to the block characteristics and the pixel positions corresponding to each block;
s3: constructing a converter encoder/decoder network, and converting the feature representation vector E into an optimized feature F through the converter encoder/decoder network d The method comprises the steps of carrying out a first treatment on the surface of the The implementation process of step S3 is as follows:
s301: inputting the feature expression vector E into a transducer coding block, and outputting x by a multi-head self-attention mechanism in the transducer coding block 1 Then, the output of the multi-head self-attention mechanism is connected with the characteristic expression vector E through residual connection and then outputx 2 The method comprises the steps of carrying out a first treatment on the surface of the For x 2 Normalized and output x 3 The method comprises the steps of carrying out a first treatment on the surface of the Will x 3 Input fully connected network layer post output with activation function x 4 The method comprises the steps of carrying out a first treatment on the surface of the Will x 4 Input of fully connected network layer post output x without activation function 5 The method comprises the steps of carrying out a first treatment on the surface of the Setting x 5 An output encoding a block for a transducer;
s302: stacking a plurality of transducer encoding blocks to form a transducer encoder so that an encoding information matrix M of the output of the transducer encoder is consistent with the dimension of the feature representation vector E;
s303: taking the coding information matrix M output by the transducer encoder as the input of the transducer decoder; the transform decoding block contains a two-pass multi-headed self-attention mechanism;
the output of the first self-attention mechanism is y 1 The method comprises the steps of carrying out a first treatment on the surface of the The output M of the transducer encoder and the output y of the first self-attention mechanism are connected by a residual error 1 Post-connection output y 2 The method comprises the steps of carrying out a first treatment on the surface of the For y 2 Normalized and output y 3 ;
The input of the first layer decoding layer in the second self-attention mechanism is the feature representation vector E, and the corresponding output is y 4 The method comprises the steps of carrying out a first treatment on the surface of the Output y of the first self-attention mechanism via residual connection 1 Output y from the second self-attention mechanism 4 Post-connection output y 5 The method comprises the steps of carrying out a first treatment on the surface of the For y 5 Normalized and output y 6 The method comprises the steps of carrying out a first treatment on the surface of the Will y 6 Input fully connected network layer post output y with activation function 7 The method comprises the steps of carrying out a first treatment on the surface of the Will y 7 Input fully connected network layer post output y without activation function 8 The method comprises the steps of carrying out a first treatment on the surface of the Setting y 8 An output for the transform decoding block;
s304: stacking multiple transducer decoding blocks to form a transducer decoder, the output of which is the optimized feature F d ;
S4: will optimize feature F d And F is obtained after connection with the characteristic expression vector E c ;
S5: by means of attentiveness mechanism to F c After the important features in the model are aggregated, the aggregation result is input into a convolutional neural network to obtain a human body example segmentation result.
2. The human body instance segmentation method according to claim 1, characterized in that: the distance calculation algorithm includes a Euclidean distance, a Chebyshev distance, and a cosine distance.
3. The human body instance segmentation method according to claim 1, characterized in that: feature representation vector e= [ f 1 +p 1 ,f 2 +p 2 ,...,f N +p N ]Wherein f N Block characteristics representing the nth block, p N The pixel position of the nth block is represented, and N represents the number of blocks.
4. The human body instance segmentation method according to claim 1, characterized in that: optimizing feature F by Concat operation in step S4 d Connected to the feature representation vector E.
5. A human body instance segmentation terminal device, characterized in that: comprising a processor, a memory and a computer program stored in the memory and running on the processor, which processor, when executing the computer program, carries out the steps of the method according to any one of claims 1 to 4.
6. A computer-readable storage medium storing a computer program, characterized in that: the computer program, when executed by a processor, implements the steps of the method according to any one of claims 1 to 4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110707258.5A CN113487624B (en) | 2021-06-24 | 2021-06-24 | Human body instance segmentation method, terminal equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110707258.5A CN113487624B (en) | 2021-06-24 | 2021-06-24 | Human body instance segmentation method, terminal equipment and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113487624A CN113487624A (en) | 2021-10-08 |
CN113487624B true CN113487624B (en) | 2023-10-13 |
Family
ID=77936230
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110707258.5A Active CN113487624B (en) | 2021-06-24 | 2021-06-24 | Human body instance segmentation method, terminal equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113487624B (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112184780A (en) * | 2020-10-13 | 2021-01-05 | 武汉斌果科技有限公司 | Moving object instance segmentation method |
CN112862830A (en) * | 2021-01-28 | 2021-05-28 | 陕西师范大学 | Multi-modal image segmentation method, system, terminal and readable storage medium |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10839259B2 (en) * | 2017-10-27 | 2020-11-17 | Google Llc | Attention-based image generation neural networks |
-
2021
- 2021-06-24 CN CN202110707258.5A patent/CN113487624B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112184780A (en) * | 2020-10-13 | 2021-01-05 | 武汉斌果科技有限公司 | Moving object instance segmentation method |
CN112862830A (en) * | 2021-01-28 | 2021-05-28 | 陕西师范大学 | Multi-modal image segmentation method, system, terminal and readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN113487624A (en) | 2021-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP2894113B2 (en) | Image clustering device | |
CN109410123B (en) | Deep learning-based mosaic removing method and device and electronic equipment | |
US20220335583A1 (en) | Image processing method, apparatus, and system | |
CN110717953B (en) | Coloring method and system for black-and-white pictures based on CNN-LSTM (computer-aided three-dimensional network-link) combination model | |
CN109948441B (en) | Model training method, image processing method, device, electronic equipment and computer readable storage medium | |
WO2022166258A1 (en) | Behavior recognition method and apparatus, terminal device, and computer-readable storage medium | |
CN112801918A (en) | Training method of image enhancement model, image enhancement method and electronic equipment | |
US20220084165A1 (en) | System and method for single-modal or multi-modal style transfer and system for random stylization using the same | |
CN116091823A (en) | Single-feature anchor-frame-free target detection method based on fast grouping residual error module | |
CN116863194A (en) | Foot ulcer image classification method, system, equipment and medium | |
CN111179276B (en) | Image processing method and device | |
CN111784699A (en) | Method and device for carrying out target segmentation on three-dimensional point cloud data and terminal equipment | |
CN109712075B (en) | Method and device for identifying original image of digital image data | |
CN112614110B (en) | Method and device for evaluating image quality and terminal equipment | |
CN113487624B (en) | Human body instance segmentation method, terminal equipment and storage medium | |
CN110717405B (en) | Face feature point positioning method, device, medium and electronic equipment | |
CN111681165A (en) | Image processing method, image processing device, computer equipment and computer readable storage medium | |
CN111814884A (en) | Target detection network model upgrading method based on deformable convolution | |
CN111932557B (en) | Image semantic segmentation method and device based on ensemble learning and probability map model | |
CN110647898B (en) | Image processing method, image processing device, electronic equipment and computer storage medium | |
CN115809959A (en) | Image processing method and device | |
CN115700785A (en) | Image feature extraction method and device, electronic equipment and storage medium | |
CN112669240A (en) | High-definition image restoration method and device, electronic equipment and storage medium | |
CN111597375B (en) | Picture retrieval method based on similar picture group representative feature vector and related equipment | |
CN113256743B (en) | Image processing method and device, electronic equipment and readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |