CN110766645A - Target person reproduction graph generation method based on person identification and segmentation - Google Patents

Target person reproduction graph generation method based on person identification and segmentation Download PDF

Info

Publication number
CN110766645A
CN110766645A CN201911017510.9A CN201911017510A CN110766645A CN 110766645 A CN110766645 A CN 110766645A CN 201911017510 A CN201911017510 A CN 201911017510A CN 110766645 A CN110766645 A CN 110766645A
Authority
CN
China
Prior art keywords
picture
target person
face
image
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911017510.9A
Other languages
Chinese (zh)
Other versions
CN110766645B (en
Inventor
姜光
来滇之
史梦真
马全盟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Electronic Science and Technology
Original Assignee
Xian University of Electronic Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Electronic Science and Technology filed Critical Xian University of Electronic Science and Technology
Priority to CN201911017510.9A priority Critical patent/CN110766645B/en
Publication of CN110766645A publication Critical patent/CN110766645A/en
Application granted granted Critical
Publication of CN110766645B publication Critical patent/CN110766645B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Abstract

The invention discloses a target character reproduction graph generation method based on character recognition and segmentation, which comprises the following implementation scheme: in a plurality of input pictures collected in the same scene, one picture is arbitrarily selected as a reference picture for generating a reproduction picture of a target person, the target person in each picture is determined by a face recognition and detection algorithm, non-target persons are removed in the reference picture by an example segmentation algorithm, the background of the reference picture is supplemented by the backgrounds of other input pictures except the reference picture, and the regional images of the target persons in the other pictures are used for replacing the regional images of the corresponding positions in the reference picture, so that a reproduction picture which has complete background information and consists of the target persons in all the input pictures is obtained. The method has the advantages that the target person is accurately selected by using a face recognition and detection algorithm, and the background information of the picture is supplemented by using an example segmentation algorithm.

Description

Target person reproduction graph generation method based on person identification and segmentation
Technical Field
The invention belongs to the technical field of image processing, and further relates to a person target recurrence map generation method based on person identification and segmentation in the technical field of computer vision. The method can be used for removing redundant characters from a plurality of pictures collected at the dense crowd and generating a character target reproduction picture from the plurality of pictures from which the redundant characters are removed.
Background
In a natural picture shot in a complex scene, due to numerous people and a disordered background, a target person cannot be highlighted, and the background is seriously shielded, accurate target person and enough background information cannot be acquired for image processing such as person reproduction and contrast. For example, when a person wants to take a picture containing only the target person at a crowded place such as an airport, a tourist attraction, a shopping center, a physical fitness or a sports training field, the person needs to find the angle in a time-consuming and labor-consuming manner, and background information is lost when the person selects the place. Removing non-target persons may cause the image to contain more background information, highlighting the target person. In addition, the target characters at different positions in multiple pictures in the same scene can appear in the same picture, so that the interest and the enjoyment of the picture can be improved, and the picture can contain more information, such as a sentence represented by multiple sign language actions in the same picture.
In the patent document "a street photograph target person extraction method" (application No. 201711135299.1 application date: 2017.11.15 publication No. 109145911a) applied by the university of petroleum in china, a method of extracting a target person from a street photograph is disclosed. The method first generates a candidate Region using Region pro-potential Network (RPN). And then, extracting the overall features of the image by using the convolutional layer of fast rcnn to obtain a feature map of each candidate region in the image, and performing pixel correction on each candidate region by using RoIAlign. And after the feature map of each candidate region is obtained, predicting each candidate region to obtain the category and the region bounding box of the candidate region. And predicting the category of each pixel point in each candidate region by using a designed FCN frame for each candidate region to finally obtain an image instance segmentation result. And extracting the individual image of the target person by using the mask matrix obtained by the segmentation result and manual interaction. The method has the disadvantages that the target person is not judged, the image of the target person needs to be manually extracted, the operation is complex, and the wrong target person can be selected.
Patent document "a person image processing method and apparatus" (application No. 201510235866.5 application date: 2015.5.11 publication No. 104794462B) filed by changdu outlook digital technology limited discloses a person image processing method. The method comprises the steps of obtaining a face area in a target image according to a preset face recognition algorithm; under the condition that the number of the obtained face areas is not less than two, the obtained face areas are divided into a foreground face area and a background face area according to a preset classification algorithm and/or according to selection operation of a user; and processing the background face region obtained by the distinguishing according to a preset first image processing algorithm, so that the visual effect of the background face region obtained by the distinguishing is poor. The method has the defects that the position and pixel information of the target person are lost due to the fact that the target person and the background are blurred while the non-target person face area is subjected to blurring processing.
Disclosure of Invention
The present invention is directed to provide a method for generating a target person recurrence map based on person identification and segmentation, which is used to solve the problems of too many persons in a picture and a disordered picture background.
The technical idea for realizing the purpose of the invention is as follows: and removing non-target characters from the input picture, and integrating all target characters in one picture.
The implementation steps of the invention comprise the following steps:
step 1, inputting pictures:
inputting at least two pictures collected in the same scene, wherein each picture at least comprises a target person to be determined;
step 2, selecting a reference graph:
arbitrarily selecting an input picture as a reference picture for generating a target character reproduction picture;
step 3, establishing a face data set:
carrying out face detection on each input picture and a reference picture by using a face detection algorithm, carrying out face correction, and forming a face data set by all corrected face pictures;
step 4, determining a target person:
inputting the pictures in the face data set into a trained face recognition network, outputting the face characteristic vector of each figure, and taking the figure corresponding to the characteristic vector with the largest number as the determined target figure in each picture;
step 5, determining the position of the non-target person in the reference image:
obtaining a non-target person region image except the target person and position information of the non-target person in the reference image by using an example segmentation algorithm;
step 6, finding the positions of the other pictures after the target person is determined, wherein the positions of the other pictures are the same as the positions of the non-target persons in the reference picture; judging whether the positions in the other pictures have complete backgrounds or not, and if not, executing the step 7; otherwise, executing step 8;
step 7, inputting a picture meeting the condition in the step 1 and then executing the step 3;
step 8, replacing the non-target character region image in the corresponding position region in the reference image by the background region in the complete background picture to obtain an updated reference image;
step 9, determining the position of the target person:
determining the target person region image and the position information of the target person in the other pictures by adopting the same example segmentation algorithm as the step 5;
step 10, obtaining a reproduction diagram:
finding the position which is the same as the position of the target person in the rest pictures in the updated reference picture; and replacing the corresponding position area image in the reference image with the target person area image in the other images to obtain a reproduction image consisting of the target persons in all the input images.
Compared with the prior art, the invention has the following advantages:
1, the method determines the target person region image and the position information of the target person in the picture by using the example segmentation algorithm, thereby overcoming the problems that the operation of manually selecting the target person is complicated and wrong target persons can be selected in the prior art, and having the advantage that the target person can be accurately selected to generate a correct target person reproduction picture.
2, because the background area existing in the complete background picture is used for replacing the image of the non-target person area in the corresponding position area in the reference picture, the problem that the position and pixel information of the target person is lost due to the fact that the target person and the background are blurred while the fuzzy processing is carried out on the face area of the non-target person in the prior art is solved, and the method has the advantage that the background information can be completely reserved in the generated recurrent picture.
Drawings
FIG. 1 is a flow chart of the present invention.
Detailed Description
The specific steps of the implementation of the present invention are further described below with reference to fig. 1.
Step 1, inputting a picture.
At least two pictures collected in the same scene are input, and each picture at least comprises a target person to be determined. In the embodiment of the invention, 8 pictures collected at the same place of a school are input, and each picture comprises a target person and non-target persons with different numbers of people.
And 2, selecting a reference graph.
One input picture is arbitrarily selected from the 8 pictures input in the embodiment of the present invention as a reference picture for generating a reproduction picture of a target person.
And step 3, establishing a face data set.
And performing face detection on the reference image and the remaining 7 images in the embodiment of the invention by adopting a face detection algorithm, performing face rectification, and forming a face data set by all the rectified face images.
Firstly, each input picture and a reference picture are sequentially input into a trained face region suggestion network, and each face frame in each picture is output.
The trained face area suggestion network is a full-connection network comprising a full-connection layer and a face key point locator. The fully connected layer of the network can mark each face frame from each input picture.
In the embodiment of the invention, the trained face region suggestion Network adopts a region suggestion Network in a Multi-target cascade convolution Network Multi-task Cascaded generalized Convolitional Network.
Secondly, performing regression processing on each face frame by using a trained convolutional neural network, outputting coordinates of the upper left corner and the lower right corner of the regressed face frame, and outputting five feature points corresponding to five sense organs of each face by using a face key point positioner;
the regression processing of the invention adopts a trained convolutional neural network which comprises a full connection layer boundary box regression-Boxregresseo. The method comprises the steps that a full connection layer boundary frame of the network regresses Bounding-Box regression to manually select a real face frame Ground route for each face in each input picture, whether the Intersection ratio of a face frame calibrated in the first step and the manually selected real face frame is larger than an interaction over Unit is judged, if the Intersection ratio is larger than a threshold value, coordinates of the upper left corner and the lower right corner of the calibrated face frame are output, otherwise, fine adjustment is carried out on the calibrated face frame until the Intersection ratio is larger than the threshold value, and the coordinates of the upper left corner and the lower right corner of the regressed face frame are output.
The output feature points of the invention adopt a face key point positioner, which can extract five feature points corresponding to five sense organs for each face from an input picture containing the face and output the pixel coordinates of each feature point.
In the embodiment of the invention, the trained boundary box regression and face key point positioner adopts a refined Network Refine Network in a Multi-target cascade convolution Network Multi-task Cascade connected conditional Network.
And thirdly, determining each face frame after regression processing by using the coordinates of the upper left corner and the lower right corner of the face frame, cutting each face frame from each input picture and a reference picture, and aligning the face by using five feature points corresponding to each face after regression.
In the embodiment of the invention, a picture without pixel information is preset, the picture comprises five characteristic point positions corresponding to five sense organs of the front face image, and the face picture cut out in the third step is mapped to the pixel position corresponding to the preset picture through affine transformation to obtain the corrected face picture.
And 4, determining the target person.
And inputting the pictures in the face data set into a trained face recognition network, outputting the face feature vector of each person, and taking the person corresponding to the feature vector with the maximum number as the determined target person in each picture.
The invention outputs the facial feature vector of each person through the trained full-connection layer convolutional neural network, the full-connection layer of the network converts each face picture in the input face data set into the feature vector, and the person corresponding to the feature vector with the largest number is used as the determined target person in each picture.
In the embodiment of the invention, the face recognition network adopts a deep neural network ArcFace.
And 5, determining the position of the non-target person in the reference image.
The following example segmentation algorithm is employed to obtain the image of the non-target person region other than the target person and the position information of the non-target person in the reference map.
Firstly, establishing an image segmentation model formed by sequentially connecting a convolutional neural network, an interested region recommendation network and a segmentation network;
the convolutional neural network adopts a residual error network with 50 layers;
the network structure of the region of interest is a tree structure and comprises a trunk and two branches, the size of the trunk convolutional layer convolution kernel is set to be 3 multiplied by 3, and the size of the branch convolutional layer convolution kernel is set to be 1 multiplied by 1;
the segmentation network structure is composed of 6 convolutional layers and 2 fully-connected layers, the size of a convolutional kernel of each convolutional layer is set to be 3 x 3, and the size of a convolutional kernel of each fully-connected layer is set to be 7 x 7;
secondly, training the image segmentation model by using an image set containing characters until the characters can be segmented pixel by pixel to obtain a trained image segmentation model; the reference map is input to the trained image segmentation model, and the non-target person region image and the position information of the non-person are output.
Step 6, finding the positions of the other pictures after the target person is determined, wherein the positions of the other pictures are the same as the positions of the non-target persons in the reference picture; judging whether the positions in the other pictures have complete backgrounds or not, and if not, executing the step 7; otherwise, step 8 is performed.
And 7, inputting a picture meeting the condition in the step 1 and then executing the step 3.
And 8, replacing the non-target character region image in the corresponding position region in the reference image by using the background region in the complete background picture to obtain an updated reference image.
And 9, determining the position of the target person.
Determining the target person region image and the position information of the target person in the other pictures by adopting the following example segmentation algorithm;
firstly, establishing an image segmentation model formed by sequentially connecting a convolutional neural network, an interested region recommendation network and a segmentation network;
the convolutional neural network adopts a residual error network with 50 layers;
the network structure of the region of interest is a tree structure and comprises a trunk and two branches, the size of the trunk convolutional layer convolution kernel is set to be 3 multiplied by 3, and the size of the branch convolutional layer convolution kernel is set to be 1 multiplied by 1;
the segmentation network structure is composed of 6 convolutional layers and 2 fully-connected layers, the size of a convolutional kernel of each convolutional layer is set to be 3 x 3, and the size of a convolutional kernel of each fully-connected layer is set to be 7 x 7;
secondly, training the image segmentation model by using an image set containing characters until the characters can be segmented pixel by pixel to obtain a trained image segmentation model; and inputting the other pictures except the reference picture into the trained image segmentation model, and outputting the target person region image and the position information of the person.
And step 10, obtaining a reproduction diagram.
Finding the position which is the same as the position of the target person in the rest pictures in the updated reference picture; and replacing the corresponding position area image in the reference image with the target person area image in the other images to obtain a reproduction image consisting of the target persons in all the input images.
In the embodiment of the invention, the target person in 7 pictures except the reference picture is used for replacing the image of the corresponding position area in the reference picture, and a reproduction picture of the target person appearing 8 times at different positions is obtained.

Claims (4)

1. A target person reproduction image generation method based on person recognition and segmentation is characterized in that a target person is judged according to facial features of the person, non-target persons are removed through an example segmentation algorithm, and a target person reproduction image is generated, wherein the method comprises the following steps:
step 1, inputting pictures:
inputting at least two pictures collected in the same scene, wherein each picture at least comprises a target person to be determined;
step 2, selecting a reference graph:
arbitrarily selecting an input picture as a reference picture for generating a target character reproduction picture;
step 3, establishing a face data set:
carrying out face detection on each input picture and a reference picture by using a face detection algorithm, carrying out face correction, and forming a face data set by all corrected face pictures;
step 4, determining a target person:
inputting the pictures in the face data set into a trained face recognition network, outputting the face characteristic vector of each figure, and taking the figure corresponding to the characteristic vector with the largest number as the determined target figure in each picture;
step 5, determining the position of the non-target person in the reference image:
obtaining a non-target person region image except the target person and position information of the non-target person in the reference image by using an example segmentation algorithm;
step 6, finding the positions of the other pictures after the target person is determined, wherein the positions of the other pictures are the same as the positions of the non-target persons in the reference picture; judging whether the positions in the other pictures have complete backgrounds or not, and if not, executing the step 7; otherwise, executing step 8;
step 7, inputting a picture meeting the condition in the step 1 and then executing the step 3;
step 8, replacing the non-target character region image in the corresponding position region in the reference image by the background region in the complete background picture to obtain an updated reference image;
step 9, determining the position of the target person:
determining the target person region image and the position information of the target person in the other pictures by adopting the same example segmentation algorithm as the step 5;
step 10, obtaining a reproduction diagram:
finding the position which is the same as the position of the target person in the rest pictures in the updated reference picture; and replacing the corresponding position area image in the reference image with the target person area image in the other images to obtain a reproduction image consisting of the target persons in all the input images.
2. The method of claim 1, wherein the target person recurrence map generation method comprises: the steps of the face detection algorithm in step 3 are as follows:
firstly, sequentially inputting each input picture and a reference picture into a trained face region suggestion network, and outputting each face frame in each picture;
secondly, performing regression processing on each face frame by using a trained convolutional neural network, outputting coordinates of the upper left corner and the lower right corner of the regressed face frame, and outputting five feature points corresponding to five sense organs of each face by using a face key point positioner;
and thirdly, determining each face frame after regression processing by using the coordinates of the upper left corner and the lower right corner of the face frame, cutting each face frame from each input picture and a reference picture, and aligning the face by using five feature points corresponding to each face after regression.
3. The method of claim 1, wherein the target person recurrence map generation method comprises: and 4, adopting a deep neural network ArcFace as the face recognition network in the step 4.
4. The method of claim 1, wherein the target person recurrence map generation method comprises: the example segmentation algorithm in the steps 5 and 8 comprises the following steps:
firstly, establishing an image segmentation model formed by sequentially connecting a convolutional neural network, an interested region recommendation network and a segmentation network;
the convolutional neural network adopts a residual error network with 50 layers;
the network structure of the region of interest is a tree structure and comprises a trunk and two branches, the size of the trunk convolutional layer convolution kernel is set to be 3 multiplied by 3, and the size of the branch convolutional layer convolution kernel is set to be 1 multiplied by 1;
the segmentation network structure is composed of 6 convolutional layers and 2 fully-connected layers, the size of a convolutional kernel of each convolutional layer is set to be 3 x 3, and the size of a convolutional kernel of each fully-connected layer is set to be 7 x 7;
secondly, training the image segmentation model by using an image set containing characters until the characters can be segmented pixel by pixel to obtain a trained image segmentation model; a picture is input to the trained image segmentation model, and a person region image and position information of a person, the person being a non-target person in step 5, a target person in step 8, the picture being a reference picture in step 5, and the remaining pictures excluding the reference picture in step 8, are output.
CN201911017510.9A 2019-10-24 2019-10-24 Target person recurrence map generation method based on person identification and segmentation Active CN110766645B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911017510.9A CN110766645B (en) 2019-10-24 2019-10-24 Target person recurrence map generation method based on person identification and segmentation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911017510.9A CN110766645B (en) 2019-10-24 2019-10-24 Target person recurrence map generation method based on person identification and segmentation

Publications (2)

Publication Number Publication Date
CN110766645A true CN110766645A (en) 2020-02-07
CN110766645B CN110766645B (en) 2023-03-10

Family

ID=69333391

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911017510.9A Active CN110766645B (en) 2019-10-24 2019-10-24 Target person recurrence map generation method based on person identification and segmentation

Country Status (1)

Country Link
CN (1) CN110766645B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111626091A (en) * 2020-03-09 2020-09-04 咪咕文化科技有限公司 Face image annotation method and device and computer readable storage medium
CN112446820A (en) * 2020-10-31 2021-03-05 浙江工业大学 Method for removing irrelevant portrait of scenic spot photo
WO2023087215A1 (en) * 2021-11-18 2023-05-25 Citrix Systems, Inc. Online meeting non-participant detection and remediation

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108932536A (en) * 2018-07-18 2018-12-04 电子科技大学 Human face posture method for reconstructing based on deep neural network
CN109961006A (en) * 2019-01-30 2019-07-02 东华大学 A kind of low pixel multiple target Face datection and crucial independent positioning method and alignment schemes
WO2019128508A1 (en) * 2017-12-28 2019-07-04 Oppo广东移动通信有限公司 Method and apparatus for processing image, storage medium, and electronic device
CN109993089A (en) * 2019-03-22 2019-07-09 浙江工商大学 A kind of video object removal and background recovery method based on deep learning

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019128508A1 (en) * 2017-12-28 2019-07-04 Oppo广东移动通信有限公司 Method and apparatus for processing image, storage medium, and electronic device
CN108932536A (en) * 2018-07-18 2018-12-04 电子科技大学 Human face posture method for reconstructing based on deep neural network
CN109961006A (en) * 2019-01-30 2019-07-02 东华大学 A kind of low pixel multiple target Face datection and crucial independent positioning method and alignment schemes
CN109993089A (en) * 2019-03-22 2019-07-09 浙江工商大学 A kind of video object removal and background recovery method based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张守东等: "基于多特征融合的显著性目标检测算法", 《计算机科学与探索》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111626091A (en) * 2020-03-09 2020-09-04 咪咕文化科技有限公司 Face image annotation method and device and computer readable storage medium
CN111626091B (en) * 2020-03-09 2023-09-22 咪咕文化科技有限公司 Face image labeling method and device and computer readable storage medium
CN112446820A (en) * 2020-10-31 2021-03-05 浙江工业大学 Method for removing irrelevant portrait of scenic spot photo
WO2023087215A1 (en) * 2021-11-18 2023-05-25 Citrix Systems, Inc. Online meeting non-participant detection and remediation

Also Published As

Publication number Publication date
CN110766645B (en) 2023-03-10

Similar Documents

Publication Publication Date Title
CN109299274B (en) Natural scene text detection method based on full convolution neural network
Ding et al. Semantic segmentation of large-size VHR remote sensing images using a two-stage multiscale training architecture
CN110766645B (en) Target person recurrence map generation method based on person identification and segmentation
CN110335277A (en) Image processing method, device, computer readable storage medium and computer equipment
US11854244B2 (en) Labeling techniques for a modified panoptic labeling neural network
CN108229397A (en) Method for text detection in image based on Faster R-CNN
CN108549893A (en) A kind of end-to-end recognition methods of the scene text of arbitrary shape
CN112261477B (en) Video processing method and device, training method and storage medium
CN107180226A (en) A kind of dynamic gesture identification method based on combination neural net
CN109872278B (en) Image cloud layer removing method based on U-shaped network and generation countermeasure network
CN108388882A (en) Based on the gesture identification method that the overall situation-part is multi-modal RGB-D
CN110766020A (en) System and method for detecting and identifying multi-language natural scene text
CN110765833A (en) Crowd density estimation method based on deep learning
CN110263768A (en) A kind of face identification method based on depth residual error network
CN106897681A (en) A kind of remote sensing images comparative analysis method and system
CN112232199A (en) Wearing mask detection method based on deep learning
CN110413816A (en) Colored sketches picture search
CN108921850B (en) Image local feature extraction method based on image segmentation technology
CN110956681B (en) Portrait background automatic replacement method combining convolution network and neighborhood similarity
CN109919157A (en) A kind of vision positioning method and device
CN110909724B (en) Thumbnail generation method of multi-target image
CN106845513A (en) Staff detector and method based on condition random forest
CN110533026A (en) The competing image digitization of electricity based on computer vision and icon information acquisition methods
CN110427819A (en) The method and relevant device of PPT frame in a kind of identification image
CN114120389A (en) Network training and video frame processing method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant