CN114511644B - Adaptive digital camouflage method based on deep learning - Google Patents

Adaptive digital camouflage method based on deep learning Download PDF

Info

Publication number
CN114511644B
CN114511644B CN202210069098.0A CN202210069098A CN114511644B CN 114511644 B CN114511644 B CN 114511644B CN 202210069098 A CN202210069098 A CN 202210069098A CN 114511644 B CN114511644 B CN 114511644B
Authority
CN
China
Prior art keywords
convolution
camouflage
gating
deep learning
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210069098.0A
Other languages
Chinese (zh)
Other versions
CN114511644A (en
Inventor
张静
霍一博
杜晓辉
刘娟秀
刘霖
刘永
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN202210069098.0A priority Critical patent/CN114511644B/en
Publication of CN114511644A publication Critical patent/CN114511644A/en
Application granted granted Critical
Publication of CN114511644B publication Critical patent/CN114511644B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a self-adaptive digital camouflage method based on deep learning, and belongs to the field of image processing. According to the invention, the surrounding environment background is continuously scanned by devices such as a camera, and the environment background state corresponding to the position of the camouflage target is rapidly analyzed and calculated by utilizing the relation between the space information and the front and rear frames of the video, so that the real-time pixel level camouflage is carried out on the camouflage target, and the effect of visual stealth is achieved. The technology can realize good camouflage of stationary objects and moving objects, and has great significance for natural exploration, latent investigation, national security and scientific research.

Description

Adaptive digital camouflage method based on deep learning
Technical Field
The invention belongs to the field of image processing, in particular to a technology for calculating the optimal camouflage color of a camouflage target through the color information of the surrounding environment, which can ensure that the camouflage target has great concealment.
Background
The conventional digital camouflage generation method is complicated in implementation steps, generally needs to generate digital camouflage patches through background dominant color extraction and patch extraction, and finally color filling is carried out on the generated patches to complete the whole digital camouflage generation process. Although the digital camouflage generated by the traditional method is widely applied, due to the visual persistence effect of human eyes, the method cannot realize good camouflage in the motion process of the camouflage object, and the self-adaptive digital camouflage technology can make up for the defects, so that the camouflage object can also realize good camouflage effect in the motion process. The self-adaptive digital camouflage algorithm based on deep learning is provided, and the generation of the end-to-end self-adaptive digital camouflage is realized. The method provided by the invention utilizes the deep neural network to extract time and space information from the reference frame, and can realize good camouflage in the motion process of the camouflage object.
Disclosure of Invention
The invention aims to solve the problem that the digital camouflage cannot ensure good camouflage of the moving object to be camouflaged in the moving process of the object to be camouflaged, and provides a solution for the problem, environmental information around the object to be camouflaged is acquired through equipment such as a camera, the camouflage color most similar to the surrounding environment is calculated for the moving object to be camouflaged at any time, the purpose of high speed and high efficiency is achieved, and a solid foundation is laid for further research of the self-adaptive digital camouflage technology.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows: an adaptive digital camouflage method based on deep learning, the method comprising:
step 1: shooting an environmental information video around a current target to be camouflaged through a camera, and obtaining a video image closest to the current time;
step 2: making mask information on the video images of the six frames, so that a camouflage target is black in the mask video image, and environment information is white in the mask video image;
step 3: the color information obtained in the step 1 and the mask information obtained in the step 2 are used as inputs and sent to an adaptive camouflage network based on deep learning;
step 4: calculating the input obtained in the step 3 by utilizing the weight of the deep learning neural network trained in advance to obtain a result of the self-adaptive camouflage algorithm based on the deep learning;
step 5: extracting the result to obtain camouflage color information of the current frame, and performing color camouflage by utilizing the variable color metamaterial of the target to be camouflaged and the flexible OLED;
step 6: and (3) after the camouflage at the moment is finished, repeating the steps 1-5 until the camouflage task is finished;
the adaptive camouflage network for deep learning comprises an encoder and a decoder, and has 17 layers; the encoder has 12 layers, namely a 6-layer serial space-time gating convolution module with different convolution kernels, a 4-layer serial cavity convolution module and a 2-layer serial space-time gating convolution module; the decoder has 5 layers, and is sequentially provided with a transposition convolution module, a gating convolution module, a transposition convolution module, a gating convolution module and a gating convolution module;
the space-time gating convolution module is used for obtaining a result by adopting a mode of combining gating convolution and space-time convolution simultaneously for continuous frame images input by the piece; the gating convolution is to obtain the result of one branch by using convolution operation on the original input, then to obtain the result of the other branch by using the same convolution mode and sigmoid activation function to perform feature selection on the original input, and then to multiply the results of the two branches to obtain the output of the time-space gating convolution; the space-time convolution is to move the front quarter of the channel number in the continuous multi-frame image back one frame, the first frame left is filled with 0, then move one frame forward for the next quarter, the information of the last frame left is filled with zero, and the rest half part is kept unchanged; then, carrying out feature extraction on the processed continuous multi-frame images by using two-dimensional convolution, and then merging feature information;
the cavity convolution module uses a convolution kernel to sample input continuous frame images at equal intervals;
and when the transposition convolution module is used, four sides of an input image are subjected to zero padding, then the convolution kernel is rotated by 180 degrees, and convolution is performed.
Further, in the step 1, color information of the surrounding environment of the sixty frames closest to the current time is obtained, the video images of the sixty frames are sampled at equal intervals, the sampling interval is ten frames, finally, 6 frames of video images are obtained, and the 6 frames of images are adopted for subsequent processing.
The invention aims to solve the problem that the digital camouflage cannot ensure good camouflage of the moving object to be camouflaged in the moving process of the object to be camouflaged, and provides a solution for the problem, environmental information around the object to be camouflaged is acquired through equipment such as a camera, the camouflage color most similar to the surrounding environment is calculated for the moving object to be camouflaged at any time, the purpose of high speed and high efficiency is achieved, and a solid foundation is laid for further research of the self-adaptive digital camouflage technology.
In order to meet the requirement of real-time performance, the invention adopts time-space gating convolution to replace 3D convolution, and can effectively reduce operation parameters; in the latter half of the neural network, a 2D convolution module of transpose convolution and gate control convolution is adopted, and only the image of the current frame is decoded to generate, so that the parameter number of the neural network can be reduced to the greatest extent, and the operation time is reduced. According to the experiment, under the computer condition that the CPU is AMD ryzen-5 3600CPU@3.60GHz, the display card is NVIDIA GeTorce GTX 1660Ti GPU, and the memory is 16G RAM, the operation speed can reach more than 12 frames/second.
Drawings
Fig. 1 is a flow chart of an adaptive digital camouflage technique.
Fig. 2 is a deep learning network structure of the present invention.
FIG. 3 is a graph of camouflage results from one experiment of the present invention.
Detailed Description
The invention is described in further detail below with reference to the drawings and examples, but embodiments of the invention are not limited thereto.
The algorithm flow chart of the invention is shown in fig. 1, and the self-adaptive digital camouflage method based on deep learning specifically comprises the following steps:
step 1: shooting an environmental information video around a current target to be camouflaged through a camera, acquiring color information of the surrounding environment of sixty frames closest to the current time, sampling the video images of the sixty frames at equal intervals, wherein the sampling interval is ten frames, and finally obtaining six frames of video images;
step 2: making mask information on the video images of the six frames so that a camouflage target is black (the pixel is 0) in the mask video image, and the environment information is white (the pixel is 255) in the mask video image;
step 3: the color information obtained in the step 1 and the mask information obtained in the step 2 are used as inputs and sent to an adaptive camouflage algorithm based on deep learning;
step 4: calculating the input obtained in the step 3 by utilizing the weight of the deep learning neural network trained in advance to obtain a result of the self-adaptive camouflage algorithm based on the deep learning;
step 5: and extracting the result to obtain camouflage color information of the current frame, and performing color camouflage by utilizing the variable color metamaterial of the target to be camouflaged and the flexible OLED.
Step 6: and (5) after the camouflage is finished at the moment, repeating the steps 1-5 until the camouflage task is finished.
For the adaptive camouflage algorithm based on deep learning of step 3 in the specific embodiment, the algorithm needs to be described in detail:
a. introduction to the overall structure of the algorithm. The algorithm is an end-to-end convolutional neural network and mainly comprises cavity convolution, time-space gating convolution and transposed convolution. The algorithm structure adopts a coder decoder structure, and has 17 layers, wherein the coder has 12 layers, namely a 6-layer serial space-time gating convolution module with different convolution kernels, a 2-layer serial cavity convolution module and a 2-layer serial space-time gating convolution module, and the decoder has 5 layers, namely a transposition convolution module, a gating convolution module, a transposition convolution module, a gating convolution module and a gating convolution module.
b. Introduction to hole convolution
Hole convolution is a convolution manner in which an input matrix is sampled at equal intervals using a convolution kernel. The cavity convolution has two main effects, namely, the first is that the receptive field can be enlarged while the spatial resolution is not reduced. Second is the ability to capture multi-scale context information.
c. Introduction to space-time gated convolution
The space-time gating convolution consists of a space-time convolution product and a gating part. Space-time convolution is to move the front quarter of the channel information back one frame, the first frame is filled with 0; and then moving one quarter forward by one frame, zero padding the information of the last frame, keeping the last half part unchanged, then carrying out feature extraction by using two-dimensional convolution, and then merging the feature information. The gating convolution is to divide the input into two parts, wherein one part adopts a convolution mode such as a space-time convolution mode, the other part uses a sigmoid activation function to perform feature selection after using the convolution, and then the result of the two parts is multiplied to obtain the output of the space-time gating convolution.
Gating y,x =∑∑W g *I
TsmFeature y,x =∑∑W f *I
Figure BDA0003481350140000041
Where Γ is a sigmoid function, outputting only 0 and 1,0 representing an invalid position, and 1 representing a valid position.
Figure BDA0003481350140000042
Is an activation function and may be a ReLU or an ELU, etc. W (W) f And W is g Representing two different spatio-temporal convolution kernels.
d. Introduction to transpose convolution
Transpose convolution is a special way of convolution that can output information that is larger than the input size, often used to upsample the image. The operation of transpose convolution is to zero-fill the four sides of the input, then rotate the convolution kernel by 180 degrees, and then perform normal convolution to obtain the final output.

Claims (2)

1. An adaptive digital camouflage method based on deep learning, the method comprising:
step 1: shooting an environmental information video around a current target to be camouflaged through a camera, and obtaining a video image closest to the current time;
step 2: making mask information on the video image so that the camouflage target is black in the mask video image and the environment information is white in the mask video image;
step 3: taking the video image obtained in the step 1 and the mask information obtained in the step 2 as inputs, and sending the inputs to a self-adaptive camouflage network based on deep learning;
step 4: calculating the input obtained in the step 3 by utilizing the weight of the deep learning neural network trained in advance to obtain a result of the self-adaptive camouflage algorithm based on the deep learning;
step 5: extracting the result to obtain camouflage color information of the current frame, and performing color camouflage by utilizing the variable color metamaterial of the target to be camouflaged and the flexible OLED;
step 6: and (3) after the camouflage at the moment is finished, repeating the steps 1-5 until the camouflage task is finished;
the adaptive camouflage network for deep learning comprises an encoder and a decoder, and has 17 layers; the encoder has 12 layers, namely a 6-layer serial space-time gating convolution module with different convolution kernels, a 4-layer serial cavity convolution module and a 2-layer serial space-time gating convolution module; the decoder has 5 layers, and is sequentially provided with a transposition convolution module, a gating convolution module, a transposition convolution module, a gating convolution module and a gating convolution module;
the space-time gating convolution module is used for simultaneously adopting a gating convolution and space-time convolution combined mode to obtain a result for the input continuous frame images; the gating convolution is to obtain the result of one branch by using convolution operation on the original input, then to obtain the result of the other branch by using the same convolution mode and sigmoid activation function to perform feature selection on the original input, and then to multiply the results of the two branches to obtain the output of the time-space gating convolution; the space-time convolution is to move the front quarter of the channel number in the continuous multi-frame image back one frame, the first frame left is filled with 0, then move one frame forward for the next quarter, the information of the last frame left is filled with zero, and the rest half part is kept unchanged; then, carrying out feature extraction on the processed continuous multi-frame images by using two-dimensional convolution, and then merging feature information;
the cavity convolution module uses a convolution kernel to sample input continuous frame images at equal intervals;
the transpose convolution module is used for carrying out zero padding on four sides of an input image, rotating a convolution kernel by 180 degrees and then carrying out convolution.
2. The adaptive digital camouflage method based on deep learning as claimed in claim 1, wherein the color information of the surrounding environment of the sixty frames closest to the current time is obtained in the step 1, the video images of the sixty frames are sampled at equal intervals, the sampling interval is ten frames, finally, 6 frames of video images are obtained, and the 6 frames of images are used for subsequent processing.
CN202210069098.0A 2022-01-21 2022-01-21 Adaptive digital camouflage method based on deep learning Active CN114511644B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210069098.0A CN114511644B (en) 2022-01-21 2022-01-21 Adaptive digital camouflage method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210069098.0A CN114511644B (en) 2022-01-21 2022-01-21 Adaptive digital camouflage method based on deep learning

Publications (2)

Publication Number Publication Date
CN114511644A CN114511644A (en) 2022-05-17
CN114511644B true CN114511644B (en) 2023-06-09

Family

ID=81549781

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210069098.0A Active CN114511644B (en) 2022-01-21 2022-01-21 Adaptive digital camouflage method based on deep learning

Country Status (1)

Country Link
CN (1) CN114511644B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104376236A (en) * 2014-12-02 2015-02-25 上海出版印刷高等专科学校 Scheme self-adaptive digital watermark embedding and extracting method based on camouflage technology
CN110245598A (en) * 2019-06-06 2019-09-17 北京瑞莱智慧科技有限公司 It fights sample generating method, device, medium and calculates equipment
WO2020037965A1 (en) * 2018-08-21 2020-02-27 北京大学深圳研究生院 Method for multi-motion flow deep convolutional network model for video prediction
CN112733744A (en) * 2021-01-14 2021-04-30 北京航空航天大学 Camouflage object detection model based on edge cooperative supervision and multi-level constraint
CN113538505A (en) * 2021-06-17 2021-10-22 杭州电子科技大学 Motion estimation system and method of single picture based on deep learning
WO2021244079A1 (en) * 2020-06-02 2021-12-09 苏州科技大学 Method for detecting image target in smart home environment

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4143655B2 (en) * 2005-11-24 2008-09-03 キヤノン株式会社 Image processing apparatus, image processing method, program, storage medium, and system
CN104835104B (en) * 2014-02-07 2017-11-03 湖南第一师范学院 Digital picture active Camouflage project based on point shape
GB2532458B (en) * 2014-11-19 2018-04-18 Folium Optics Ltd Adaptive camouflage device, system and method
US10282628B1 (en) * 2016-04-08 2019-05-07 Joshua LaRue System for producing camouflage pattern based upon perception machine vision
US10776923B2 (en) * 2018-06-21 2020-09-15 International Business Machines Corporation Segmenting irregular shapes in images using deep region growing
WO2020215236A1 (en) * 2019-04-24 2020-10-29 哈尔滨工业大学(深圳) Image semantic segmentation method and system
CN110348319B (en) * 2019-06-18 2021-05-04 武汉大学 Face anti-counterfeiting method based on face depth information and edge image fusion
CN111292220B (en) * 2020-01-19 2021-09-28 西北工业大学 Target camouflage image generation method for target image recognition of synthetic aperture radar
CN111142081B (en) * 2020-04-07 2021-02-26 湖南赛博诺格电子科技有限公司 Radar self-adaptive camouflage method based on digital coding super surface
CN112288647B (en) * 2020-10-13 2022-06-03 武汉大学 Remote sensing image cloud and shadow restoration method based on gating convolution
CN112446357B (en) * 2020-12-15 2022-05-03 电子科技大学 SAR automatic target recognition method based on capsule network
CN112801895B (en) * 2021-01-15 2022-10-14 南京邮电大学 Two-stage attention mechanism-based GAN network image restoration algorithm
CN112801914A (en) * 2021-02-09 2021-05-14 北京工业大学 Two-stage image restoration method based on texture structure perception
CN113450561B (en) * 2021-05-06 2022-05-24 浙江工业大学 Traffic speed prediction method based on space-time graph convolution-generation countermeasure network
CN113507625B (en) * 2021-06-22 2022-08-19 西安理工大学 Self-adaptive video restoration method
CN113298736B (en) * 2021-06-24 2022-03-04 河北工业大学 Face image restoration method based on face pattern

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104376236A (en) * 2014-12-02 2015-02-25 上海出版印刷高等专科学校 Scheme self-adaptive digital watermark embedding and extracting method based on camouflage technology
WO2020037965A1 (en) * 2018-08-21 2020-02-27 北京大学深圳研究生院 Method for multi-motion flow deep convolutional network model for video prediction
CN110245598A (en) * 2019-06-06 2019-09-17 北京瑞莱智慧科技有限公司 It fights sample generating method, device, medium and calculates equipment
WO2021244079A1 (en) * 2020-06-02 2021-12-09 苏州科技大学 Method for detecting image target in smart home environment
CN112733744A (en) * 2021-01-14 2021-04-30 北京航空航天大学 Camouflage object detection model based on edge cooperative supervision and multi-level constraint
CN113538505A (en) * 2021-06-17 2021-10-22 杭州电子科技大学 Motion estimation system and method of single picture based on deep learning

Also Published As

Publication number Publication date
CN114511644A (en) 2022-05-17

Similar Documents

Publication Publication Date Title
CN106683067B (en) Deep learning super-resolution reconstruction method based on residual sub-images
CN109544448B (en) Group network super-resolution image reconstruction method of Laplacian pyramid structure
Zhao et al. Invertible image decolorization
CN113284051B (en) Face super-resolution method based on frequency decomposition multi-attention machine system
CN113096017A (en) Image super-resolution reconstruction method based on depth coordinate attention network model
CN110717868A (en) Video high dynamic range inverse tone mapping model construction and mapping method and device
CN113658057A (en) Swin transform low-light-level image enhancement method
CN112288630A (en) Super-resolution image reconstruction method and system based on improved wide-depth neural network
Wang et al. Underwater image super-resolution and enhancement via progressive frequency-interleaved network
CN116188274A (en) Image super-resolution reconstruction method
CN113379606B (en) Face super-resolution method based on pre-training generation model
CN114511644B (en) Adaptive digital camouflage method based on deep learning
CN112435165A (en) Two-stage video super-resolution reconstruction method based on generation countermeasure network
CN113313133A (en) Training method for generating countermeasure network and animation image generation method
CN111861877A (en) Method and apparatus for video hyper-resolution
Zhang et al. Research on image super-resolution reconstruction based on deep learning
CN115512039A (en) 3D face construction method
CN116266336A (en) Video super-resolution reconstruction method, device, computing equipment and storage medium
CN115294199A (en) Underwater image enhancement and depth estimation method, device and storage medium
CN115375540A (en) Terahertz image super-resolution method based on deep learning algorithm
CN114582002A (en) Facial expression recognition method combining attention module and second-order pooling mechanism
Xu et al. Joint super-resolution and inverse tone-mapping: A feature decomposition aggregation network and a new benchmark
CN117635478B (en) Low-light image enhancement method based on spatial channel attention
Terai et al. Comparative study of Video/Image denoising algorithms based on Convolutional neural network CNN
CN117834899A (en) Video snapshot compression imaging method and system based on reconstruction difficulty perception

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant