CN111666842A - Shadow detection method based on double-current-cavity convolution neural network - Google Patents

Shadow detection method based on double-current-cavity convolution neural network Download PDF

Info

Publication number
CN111666842A
CN111666842A CN202010449023.6A CN202010449023A CN111666842A CN 111666842 A CN111666842 A CN 111666842A CN 202010449023 A CN202010449023 A CN 202010449023A CN 111666842 A CN111666842 A CN 111666842A
Authority
CN
China
Prior art keywords
shadow
layer
pooling
channel
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010449023.6A
Other languages
Chinese (zh)
Other versions
CN111666842B (en
Inventor
李大威
王思凡
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Donghua University
Original Assignee
Donghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Donghua University filed Critical Donghua University
Priority to CN202010449023.6A priority Critical patent/CN111666842B/en
Publication of CN111666842A publication Critical patent/CN111666842A/en
Application granted granted Critical
Publication of CN111666842B publication Critical patent/CN111666842B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a shadow detection method based on a double-current cavity convolution neuron network, which comprises the following steps: inputting the image with shadow into the network in a RGB three-channel mode; respectively extracting image features by using a pooling channel and a residual channel; performing fusion of global and local features on the feature map through a multilevel cavity pooling module; the pooling channel utilizes a decoder form to up-sample the feature map to be the same as the size of the input image, the residual channel continuously keeps low-dimensional features, and the two channels are subjected to feature fusion after the up-sampling is carried out to be the same as the size of the input image; training a network by using a cross entropy loss function to obtain a group of weights with the lowest loss values; the weights are used to detect shadows in the test image, and the argmax function is used to generate a shadow binary map. The invention has higher shadow detection accuracy and better shadow edge keeping effect. The method can be used for removing the wrongly detected figure and target shadow pixels after common algorithms such as target detection, change detection and the like.

Description

Shadow detection method based on double-current-cavity convolution neural network
Technical Field
The invention relates to the technical field of deep learning and image processing, in particular to a robust shadow detection method based on a semantic segmentation network for a single picture.
Background
Shadow detection, namely, marking shadow parts in a color image by taking pixels as units, wherein the task to be completed is to utilize a designed network structure to train, obtain a group of weights with highest accuracy to detect the shadow parts in a single picture, call the shadow parts as a foreground, mark the shadow parts as white, and call the rest parts as a background and mark the shadow parts as black. Shadows are ubiquitous in all natural scenes and can be created as long as objects block the path of light from a light source. However, in most cases, the shadow existing in the image is liable to bring interference and difficulty to the image processing, for example, when performing tasks such as foreground detection and segmentation, the shadow is often mistaken for the target because the shadow is the same as the detected target and has a significant difference from the background color, thereby greatly reducing the accuracy of detection. If the shadow can be detected before the machine vision task is carried out, the task accuracy rate can be greatly improved. Therefore, shadow detection has long been a key task in the field of machine vision.
At present, the research on the shadow detection method is developed from the angle of manually extracting features and the depth learning, but the two methods are different in nature. In the early traditional algorithm, some scholars analyze the structural features and the color features of shadows by using an optical angle and an image processing method, and need an algorithm designer to know a large amount of optical knowledge and image processing knowledge, such as analyzing a color histogram under an HSV channel and analyzing the influence of the intensity of ambient light and the transparency of an object on the shadows. At the present stage, some scholars use deep learning, namely, design a detector based on a convolutional neural network to detect shadows in images, and use sampling methods such as convolution, pooling and the like to extract a feature map of each image, so that the network learns the structural features of the shadows by self, thereby realizing end-to-end shadow detection.
Disclosure of Invention
The purpose of the invention is: the shadow area in the single image is accurately detected.
In order to achieve the above object, the technical solution of the present invention is to provide a shadow detection method based on a dual-flow void convolutional neural network, which is characterized by comprising the following steps:
step S1: inputting single pictures in the training set into a designed network in an RGB three-channel mode in sequence;
step S2: the image input network is firstly divided into two channel operations, namely a pooling channel and a residual channel, wherein: the pooling channel is subjected to down-sampling in the form of an encoder through a cavity convolution module, and high-dimensional features are gradually extracted; the residual channel is composed of a plurality of cross-flow residual modules, the characteristics are extracted by convolution through the cavity convolution module, and a layer of characteristic diagram and the characteristic information of the corresponding pooling channel are superposed to keep the low-dimensional characteristics
Step S3: sending feature maps obtained from the front four layers of the pooling channel into a multi-level cavity pooling module, pooling the feature maps obtained from the front three layers into the same size according to cavity convolutions with different expansion rates to obtain a first part three-layer feature map, performing global average pooling on the feature map of the fourth layer, performing bilinear interpolation on the feature map of the fourth layer to obtain a second part feature map, performing feature fusion on the feature map of the fourth layer to obtain the final output of a down-sampling part;
step S4: the method comprises the steps that a decoder is adopted to conduct upsampling which is completely symmetrical to a downsampling process on a multi-level cavity pooling module through a cavity convolution module, and finally an image is upsampled to the size same as that of an input image;
step S5: after the input layer, the hidden layer and the output layer of the network are determined, all the images and the labels in the data set are sent to the network for training according to the steps from S1 to S4, the labels are shadow binary images with the same size as the images and with shadow areas and non-shadow areas marked according to pixels, training algebra is determined according to the convergence trend of a training-obtained loss function, and the calculated result is obtainedThe process comprises two steps: the first step is to mark the logits value of the sample as x and convert the logits value of the sample into probability
Figure BDA0002506992490000021
Secondly, calculating a loss value by using a cross entropy formula with weight, z ×∑ y '× log (y), wherein y' is a label, y is a logits probability value calculated in the first step, and z is a self-defined weight;
step S6: the image needing to be detected is tested by utilizing the stored weight, after the parameter of the weight is determined, the detected image can output a shadow feature map and a non-shadow feature map after being input into a network, and then the two feature maps are converted into a detected shadow binary map through an argmax function.
Preferably, in step S2, the selected hole convolution module includes four layers: the first layer is a common convolution with a convolution kernel of 3 × 3; the second layer is a hole convolution with the expansion rate of 3 and the relative convolution kernel of 7 multiplied by 7; the third layer is the same as the second layer; the fourth layer is the same as the first layer.
Preferably, the loss function in step S5 is weighted cross-entropy loss.
Preferably, in step S6, after a single color image is input into a designed network, a shadow feature map and a non-shadow feature map are output, and stored in an array form, the argmax function compares the detected values of corresponding pixels in the two feature maps, if the value of the foreground of the part is large, the pixel is considered as a foreground, that is, a required shadow part, which is labeled 255, and is displayed as a white part in the shadow binary map, and the background is the same, that is, the non-shadow part, which is labeled 0, and is displayed as a black part in the shadow binary map, so as to obtain the detected shadow binary map by the above method.
The invention adopts a double-flow network structure, namely, residual flow is used for keeping the stability of low-level image features in learning, and pooling flow is used for keeping the extraction and fusion of the features from low level to deep level of the image. The network can have a good detection effect on large-scale shadows and broken shadows by introducing the hole convolution. The method can be used for removing the character and the target shadow pixels which are detected by mistake after algorithms such as common target detection, change detection and the like are carried out; the present invention can also be used alone as an image shadow area detector.
Due to the adoption of the technical scheme, compared with the prior art, the invention has the following advantages and positive effects:
1) the invention utilizes the cavity convolution module to extract the characteristics, and the cavity convolution is different from the common convolution, so that the cavity convolution can not only increase the receptive field, but also retain the pixel position information of the characteristics, thereby greatly improving the missing detection phenomenon caused by the shadow covering on different texture areas, such as: when the shadow of the vehicle covers the asphalt road with the white lane line, other detectors easily miss the shadow area on the white lane line, but our detector can detect all the shadow areas.
2) The multi-level hole pooling module method extracts the global features and fuses the global features and the local features, so that the network simultaneously extracts the local and global features of the shadow, and can better judge whether the dark color region is a shadow region or a region with a darker color, for example, other detectors can easily detect a person wearing black clothes and the shadow as the shadow together, and the detector can well distinguish the black clothes from the shadow.
3) In order to simplify the computational complexity of the network, the invention designs a cross-flow residual error module, which adds a residual error flow output characteristic diagram of the previous stage, an output characteristic diagram of the previous stage passing through a cavity convolution module and a corresponding pooling flow output characteristic diagram which is up-sampled to the same size, thereby keeping the low-dimensional characteristics and accelerating the training process.
4) And a large number of data sets are adopted for training, so that the robustness of the detector is enhanced, shadows on various textures can be accurately detected, and the detection method is fully adaptive to the change of a detection scene.
5) The detection speed is high, the detection time of a single 480 multiplied by 480 color image and a single image is only 0.12s, and the method is suitable for the shadow detection task in a video sequence.
Drawings
FIG. 1 is a flow chart of the present invention.
Fig. 2 is a network structure diagram of the inventive network, which is composed of three modules, namely a hole convolution module (ACM), a multi-level hole pooling module (MLAPM), and a cross-flow residual error module (CSRM), and the upper right corner is a legend.
Fig. 3 is an internal structure diagram of a hole convolution module, where rate is the expansion rate of a convolution kernel, that is, the size of a hole injected into one convolution kernel, one rate is 3, and a convolution kernel with a size of 3 × 3 performs hole convolution, and the size of its receptive field is the same as that of a 7 × 7 ordinary convolution kernel, in this module, we use ordinary convolution for the first and fourth layers, and use hole convolution with a rate of 3 for the second and third layers.
Fig. 4 is an internal structure of a multi-level cavity pooling module, which is composed of two parts, the first part is the fusion of the first three-layer network feature maps, the cavity convolution is performed by using expansion rates of different sizes, then three groups of feature maps are sampled to have the same size, the second part is simplified multi-scale pyramid pooling, global features are obtained by adopting global tie pooling, then up-sampling is performed to have the same size as the first part, and finally the features of the two parts are fused into one group.
Fig. 5 is a cross-flow residual module that fuses the information of the pooled channels with the residual channel information and preserves the features of the previous stage using the feature preservation method of the residual network to preserve the completeness of the local features.
FIG. 6 is a diagram showing the effect of shadow detection on SBU and ISTD data sets using the present invention, where input is the color image to be tested and ground route is the shadow image of the artificial mark.
Fig. 7 shows shadow detection on a "Bungalows" video sequence on a CDnet2012 data set, five frames are randomly selected for presentation, wherein a second behavior WeSamBe foreground detection algorithm can be seen to detect a foreground object and its shadow together, but the shadow belongs to false detection. The third row is the detection of the shadow area by our shadow detection algorithm based on the WeSamBe result, where the detected shadow area is shown in light gray.
Detailed Description
The invention will be further illustrated with reference to the following specific examples. It should be understood that these examples are for illustrative purposes only and are not intended to limit the scope of the present invention. Further, it should be understood that various changes or modifications of the present invention may be made by those skilled in the art after reading the teaching of the present invention, and such equivalents may fall within the scope of the present invention as defined in the appended claims.
As shown in fig. 1, the present embodiment discloses a shadow detection method based on a Double-flow void convolutional neural Network (Double-flow convolutional neural Network), which includes the following specific steps:
step S1: acquiring a single color image, and inputting the single color image into a designed network in an RGB (red, green and blue) form;
step S2: the image input network is first divided into two channel operations, a pooling channel and a residual channel. The pooling channel is downsampled in the form of an encoder through a cavity convolution module, and high-dimensional features such as semantics are gradually extracted. And the residual channel consists of a plurality of cross flow residual modules, namely, a cavity convolution module is used for convolution to extract features, and a layer of feature graph and feature information of a corresponding pooling channel are superposed to keep low-dimensional features. It is worth noting that the selected hole convolution module contains four layers, the first layer is a common convolution with a convolution kernel of 3 × 3; the second layer is a hole convolution with a dilation rate of 3 and a relative convolution kernel of 7 × 7, the third layer is the same as the second layer, and the fourth layer is the same as the first layer.
Step S3: and the second part is bilinear interpolation of the feature map of the fourth layer after global average pooling into the size after pooling with the feature map of the first part. And finally, performing feature fusion on the four layers of feature maps to obtain the final output of the down-sampling part.
Step S4: the output characteristic diagram of the multi-level cavity pooling module is up-sampled by a decoder through a cavity convolution module, the up-sampling is completely symmetrical to the down-sampling process, and finally the image is up-sampled to the size same as that of the input image.
Step S5: after an input layer, a hidden layer and an output layer of the network are determined, images and labels (with the same size as the images and binary images of shadow areas and non-shadow areas marked according to pixels) in a data set are all sent to the network for training according to the four steps, a training algebra can be determined according to the convergence trend of a loss function obtained by training, the loss function is weighted cross-error, and the calculation process is divided into two steps: the first step is to mark the sample's logits value as x, and first convert the sample's logits value to a probability
Figure BDA0002506992490000051
And secondly, calculating a loss value by using a cross entropy formula with weight, z ×∑ y '× log (y), wherein y' is a label, y is a logits probability value calculated in the first step, and z is self-defined weight.
Step S6: the stored weight is used for testing the picture to be detected, the parameters of the weight are input into the network to obtain two types of feature maps, and the two feature maps are converted into a detected shadow binary map through an argmax function.
The step S6 of converting the two feature maps into a shadow binary map specifically includes the following steps:
step 6.1, the detected image is sent into a network and then stored in an array form, and the pixels are divided into two types by our network: one is foreground (shaded) and one is background (unshaded);
step 6.2, coding by using an argmax function, wherein the argmax function compares values of corresponding positions in the two characteristic graphs, if the value of the foreground of the part is large, the pixel is regarded as the foreground, namely the needed shadow part is marked as 255 and is displayed as white; the same background, i.e., the non-shaded portion, marked as 0, is displayed in black, and the detected shaded binary image is obtained by the above method.

Claims (4)

1. A shadow detection method based on a double-current cavity convolution neural network is characterized by comprising the following steps:
step S1: inputting single pictures in the training set into a designed network in an RGB three-channel mode in sequence;
step S2: the image input network is firstly divided into two channel operations, namely a pooling channel and a residual channel, wherein: the pooling channel is subjected to down-sampling in the form of an encoder through a cavity convolution module, and high-dimensional features are gradually extracted; the residual channel is composed of a plurality of cross-flow residual modules, the characteristics are extracted by convolution through the cavity convolution module, and a layer of characteristic diagram and the characteristic information of the corresponding pooling channel are superposed to keep the low-dimensional characteristics
Step S3: sending feature maps obtained from the front four layers of the pooling channel into a multi-level cavity pooling module, pooling the feature maps obtained from the front three layers into the same size according to cavity convolutions with different expansion rates to obtain a first part three-layer feature map, performing global average pooling on the feature map of the fourth layer, performing bilinear interpolation on the feature map of the fourth layer to obtain a second part feature map, performing feature fusion on the feature map of the fourth layer to obtain the final output of a down-sampling part;
step S4: the method comprises the steps that a decoder is adopted to conduct upsampling which is completely symmetrical to a downsampling process on a multi-level cavity pooling module through a cavity convolution module, and finally an image is upsampled to the size same as that of an input image;
step S5: after an input layer, a hidden layer and an output layer of the network are determined, images and labels in a data set are all sent to the network for training according to the steps from S1 to S4, the labels are shadow binary images with the same size as the images and mark shadow areas and non-shadow areas according to pixels, a training algebra is determined according to the convergence trend of a training gain loss function, and the calculation process is divided into two steps: the first step is to mark the logits value of the sample as x and convert the logits value of the sample into probability
Figure FDA0002506992480000011
The second step calculates the loss value using the cross entropy formula with weight-z ×∑ y '× log (y), where y' is a label and y is a labelThe logits probability value calculated for the first step, z being the self-defined weight;
step S6: the image needing to be detected is tested by utilizing the stored weight, after the parameter of the weight is determined, the detected image can output a shadow feature map and a non-shadow feature map after being input into a network, and then the two feature maps are converted into a detected shadow binary map through an argmax function.
2. The method for detecting the shadow based on the dual-flow hole convolution neural network of claim 1, wherein in the step S2, the selected hole convolution module includes four layers: the first layer is a common convolution with a convolution kernel of 3 × 3; the second layer is a hole convolution with the expansion rate of 3 and the relative convolution kernel of 7 multiplied by 7; the third layer is the same as the second layer; the fourth layer is the same as the first layer.
3. The method for detecting the shadow based on the dual-flow hole convolution neural network as claimed in claim 1, wherein the loss function in step S5 is weighted cross-entropy loss.
4. The method according to claim 1, wherein in step S6, after the single color image is input into the designed network, a shadow feature map and a non-shadow feature map are output and stored in an array form, an argmax function compares the detected values of corresponding pixels in the two feature maps, if the foreground value of the portion is large, the pixel is considered as a foreground, i.e., a required shadow portion, which is marked as 255, and displayed as a white portion in the shadow binary map, and the background is the same, i.e., the non-shadow portion, which is marked as 0, and displayed as a black portion in the shadow binary map, so as to obtain the detected shadow binary map.
CN202010449023.6A 2020-05-25 2020-05-25 Shadow detection method based on double-current-cavity convolution neural network Active CN111666842B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010449023.6A CN111666842B (en) 2020-05-25 2020-05-25 Shadow detection method based on double-current-cavity convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010449023.6A CN111666842B (en) 2020-05-25 2020-05-25 Shadow detection method based on double-current-cavity convolution neural network

Publications (2)

Publication Number Publication Date
CN111666842A true CN111666842A (en) 2020-09-15
CN111666842B CN111666842B (en) 2022-08-26

Family

ID=72384497

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010449023.6A Active CN111666842B (en) 2020-05-25 2020-05-25 Shadow detection method based on double-current-cavity convolution neural network

Country Status (1)

Country Link
CN (1) CN111666842B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112257766A (en) * 2020-10-16 2021-01-22 中国科学院信息工程研究所 Shadow recognition detection method under natural scene based on frequency domain filtering processing
CN112949829A (en) * 2021-03-05 2021-06-11 深圳海翼智新科技有限公司 Feature graph pooling method, data processing method and computing device
CN113052775A (en) * 2021-03-31 2021-06-29 华南理工大学 Image shadow removing method and device
CN113065578A (en) * 2021-03-10 2021-07-02 合肥市正茂科技有限公司 Image visual semantic segmentation method based on double-path region attention coding and decoding
CN113178010A (en) * 2021-04-07 2021-07-27 湖北地信科技集团股份有限公司 High-resolution image shadow region restoration and reconstruction method based on deep learning
CN113870124A (en) * 2021-08-25 2021-12-31 西北工业大学 Dual-network mutual excitation learning shadow removing method based on weak supervision
CN113920124A (en) * 2021-06-22 2022-01-11 西安理工大学 Brain neuron iterative segmentation method based on segmentation and error guidance

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108986124A (en) * 2018-06-20 2018-12-11 天津大学 In conjunction with Analysis On Multi-scale Features convolutional neural networks retinal vascular images dividing method
US20190130575A1 (en) * 2017-10-30 2019-05-02 Beijing Curacloud Technology Co., Ltd. Systems and methods for image segmentation using a scalable and compact convolutional neural network
CN109711448A (en) * 2018-12-19 2019-05-03 华东理工大学 Based on the plant image fine grit classification method for differentiating key field and deep learning
CN109711413A (en) * 2018-12-30 2019-05-03 陕西师范大学 Image, semantic dividing method based on deep learning
CN110084249A (en) * 2019-04-24 2019-08-02 哈尔滨工业大学 The image significance detection method paid attention to based on pyramid feature
CN110232394A (en) * 2018-03-06 2019-09-13 华南理工大学 A kind of multi-scale image semantic segmentation method
CN110781776A (en) * 2019-10-10 2020-02-11 湖北工业大学 Road extraction method based on prediction and residual refinement network
CN110852267A (en) * 2019-11-11 2020-02-28 复旦大学 Crowd density estimation method and device based on optical flow fusion type deep neural network
CN111028235A (en) * 2019-11-11 2020-04-17 东北大学 Image segmentation method for enhancing edge and detail information by utilizing feature fusion
CN111079649A (en) * 2019-12-17 2020-04-28 西安电子科技大学 Remote sensing image ground feature classification method based on lightweight semantic segmentation network
CN111179244A (en) * 2019-12-25 2020-05-19 汕头大学 Automatic crack detection method based on cavity convolution
CN111192245A (en) * 2019-12-26 2020-05-22 河南工业大学 Brain tumor segmentation network and method based on U-Net network

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190130575A1 (en) * 2017-10-30 2019-05-02 Beijing Curacloud Technology Co., Ltd. Systems and methods for image segmentation using a scalable and compact convolutional neural network
CN110232394A (en) * 2018-03-06 2019-09-13 华南理工大学 A kind of multi-scale image semantic segmentation method
CN108986124A (en) * 2018-06-20 2018-12-11 天津大学 In conjunction with Analysis On Multi-scale Features convolutional neural networks retinal vascular images dividing method
CN109711448A (en) * 2018-12-19 2019-05-03 华东理工大学 Based on the plant image fine grit classification method for differentiating key field and deep learning
CN109711413A (en) * 2018-12-30 2019-05-03 陕西师范大学 Image, semantic dividing method based on deep learning
CN110084249A (en) * 2019-04-24 2019-08-02 哈尔滨工业大学 The image significance detection method paid attention to based on pyramid feature
CN110781776A (en) * 2019-10-10 2020-02-11 湖北工业大学 Road extraction method based on prediction and residual refinement network
CN110852267A (en) * 2019-11-11 2020-02-28 复旦大学 Crowd density estimation method and device based on optical flow fusion type deep neural network
CN111028235A (en) * 2019-11-11 2020-04-17 东北大学 Image segmentation method for enhancing edge and detail information by utilizing feature fusion
CN111079649A (en) * 2019-12-17 2020-04-28 西安电子科技大学 Remote sensing image ground feature classification method based on lightweight semantic segmentation network
CN111179244A (en) * 2019-12-25 2020-05-19 汕头大学 Automatic crack detection method based on cavity convolution
CN111192245A (en) * 2019-12-26 2020-05-22 河南工业大学 Brain tumor segmentation network and method based on U-Net network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIANG-CHIEH CHEN; GEORGE PAPANDREOU; IASONAS KOKKINOS; KEVIN MUR: "(1) 《DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs》", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 》 *
PINGPING ZHANG; WEI LIU; YINJIE LEI; HONGYU WANG; HUCHUAN LU: "《RAPNet: Residual Atrous Pyramid Network for Importance-Aware Street Scene Parsing》", 《IEEE TRANSACTIONS ON IMAGE PROCESSING 》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112257766A (en) * 2020-10-16 2021-01-22 中国科学院信息工程研究所 Shadow recognition detection method under natural scene based on frequency domain filtering processing
CN112257766B (en) * 2020-10-16 2023-09-29 中国科学院信息工程研究所 Shadow recognition detection method in natural scene based on frequency domain filtering processing
CN112949829A (en) * 2021-03-05 2021-06-11 深圳海翼智新科技有限公司 Feature graph pooling method, data processing method and computing device
CN113065578A (en) * 2021-03-10 2021-07-02 合肥市正茂科技有限公司 Image visual semantic segmentation method based on double-path region attention coding and decoding
CN113065578B (en) * 2021-03-10 2022-09-23 合肥市正茂科技有限公司 Image visual semantic segmentation method based on double-path region attention coding and decoding
CN113052775A (en) * 2021-03-31 2021-06-29 华南理工大学 Image shadow removing method and device
CN113052775B (en) * 2021-03-31 2023-05-23 华南理工大学 Image shadow removing method and device
CN113178010A (en) * 2021-04-07 2021-07-27 湖北地信科技集团股份有限公司 High-resolution image shadow region restoration and reconstruction method based on deep learning
CN113920124A (en) * 2021-06-22 2022-01-11 西安理工大学 Brain neuron iterative segmentation method based on segmentation and error guidance
CN113870124A (en) * 2021-08-25 2021-12-31 西北工业大学 Dual-network mutual excitation learning shadow removing method based on weak supervision

Also Published As

Publication number Publication date
CN111666842B (en) 2022-08-26

Similar Documents

Publication Publication Date Title
CN111666842B (en) Shadow detection method based on double-current-cavity convolution neural network
CN109934200B (en) RGB color remote sensing image cloud detection method and system based on improved M-Net
TWI744283B (en) Method and device for word segmentation
CN114841972B (en) Transmission line defect identification method based on saliency map and semantic embedded feature pyramid
CN113887459B (en) Open-pit mining area stope change area detection method based on improved Unet +
CN111797712B (en) Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network
CN110033040B (en) Flame identification method, system, medium and equipment
CN106846339A (en) Image detection method and device
CN110309808B (en) Self-adaptive smoke root node detection method in large-scale space
CN111738054B (en) Behavior anomaly detection method based on space-time self-encoder network and space-time CNN
CN109840483B (en) Landslide crack detection and identification method and device
CN111681273A (en) Image segmentation method and device, electronic equipment and readable storage medium
CN111680690B (en) Character recognition method and device
CN114155527A (en) Scene text recognition method and device
CN114742799B (en) Industrial scene unknown type defect segmentation method based on self-supervision heterogeneous network
CN110555464A (en) Vehicle color identification method based on deep learning model
CN113516126A (en) Adaptive threshold scene text detection method based on attention feature fusion
CN114022408A (en) Remote sensing image cloud detection method based on multi-scale convolution neural network
CN115861756A (en) Earth background small target identification method based on cascade combination network
CN113159158A (en) License plate correction and reconstruction method and system based on generation countermeasure network
CN117557774A (en) Unmanned aerial vehicle image small target detection method based on improved YOLOv8
Cho et al. Modified perceptual cycle generative adversarial network-based image enhancement for improving accuracy of low light image segmentation
CN111931689B (en) Method for extracting video satellite data identification features on line
CN117392392B (en) Rubber cutting line identification and generation method
CN113610857B (en) Apple grading method and system based on residual error network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant