CN113538702A - Method for generating underwater scene panoramic image of marine culture area - Google Patents

Method for generating underwater scene panoramic image of marine culture area Download PDF

Info

Publication number
CN113538702A
CN113538702A CN202110738641.7A CN202110738641A CN113538702A CN 113538702 A CN113538702 A CN 113538702A CN 202110738641 A CN202110738641 A CN 202110738641A CN 113538702 A CN113538702 A CN 113538702A
Authority
CN
China
Prior art keywords
feature map
mask
feature
underwater
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110738641.7A
Other languages
Chinese (zh)
Other versions
CN113538702B (en
Inventor
付先平
姚冰
苏知青
袁国良
王辉兵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian Maritime University
Original Assignee
Dalian Maritime University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian Maritime University filed Critical Dalian Maritime University
Priority to CN202110738641.7A priority Critical patent/CN113538702B/en
Publication of CN113538702A publication Critical patent/CN113538702A/en
Application granted granted Critical
Publication of CN113538702B publication Critical patent/CN113538702B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/003Navigation within 3D models or images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/32Indexing scheme for image data processing or generation, in general involving image mosaicing
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A40/00Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
    • Y02A40/80Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in fisheries management
    • Y02A40/81Aquaculture, e.g. of fish

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computer Hardware Design (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Computer Graphics (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method for generating a panoramic image of an underwater scene in a marine culture area, which comprises the steps of planning a walking path of an underwater robot, acquiring underwater images by using the underwater robot, and extracting marine product target characteristics of the acquired images to obtain a local characteristic image and a global characteristic image; and performing target feature extraction on the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map to obtain a target mask feature map, performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image, performing exposure and judgment on whether the images are the same image, and performing image splicing on the images of the same object by using a splicing algorithm to obtain an underwater spliced image. According to the method, after the underwater image splicing is completed in the current target area through the robot, the observation effect is clearer, more accurate and more precise.

Description

Method for generating underwater scene panoramic image of marine culture area
Technical Field
The invention relates to the technical field of underwater image acquisition, in particular to a method for generating a panoramic image of an underwater scene in a marine culture area.
Background
At present, when an underwater robot carries out close-range underwater target identification, the condition of a target needs to be observed clearly in real time. Due to the problems of dynamic change of underwater target distribution, serious attenuation and scattering effect of underwater light, limited visible range of an underwater camera and the like, complete observation of an identified target is difficult to realize by depending on information acquired and recorded by a single image obtained under the conditions of limited distance and view angle, so that observers are difficult to comprehensively master the condition of a submarine environment.
Disclosure of Invention
The invention provides a method for generating a panoramic view of an underwater scene in a marine culture area.
The technical means adopted by the invention are as follows:
a method for generating a panoramic view of an underwater scene in a marine culture area comprises the following steps:
step 1, planning a walking route of an underwater robot, and controlling the motion direction of the underwater robot according to the walking route;
2, acquiring a plurality of images of a visible area of the underwater robot through a plurality of high-definition cameras arranged on the underwater robot;
step 3, extracting marine product target features of the collected images by using a local feature extractor and a global feature extractor to obtain a local feature map and a global feature map;
step 4, extracting target features from the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map through a softmax function to obtain a target mask feature map, and performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image;
step 5, sending the representative feature diagram into a DCP network for local guidance operation, increasing the weight of the features with more information of the feature diagram channels, and reducing the weight of the unimportant features with less information;
step 6, connecting the global feature map with the representative feature map to obtain a fusion feature map;
step 7, carrying out target re-identification detection on the multiple images in the visible area, and if the same object appears in two or more images, carrying out exposure processing on the images in which the same object is detected to obtain exposure processing images;
step 8, judging whether the exposure processing images have objects with the same characteristics, if not, returning to the step 2, and if so, executing the step 9;
step 9, carrying out image splicing on the images of the detected same object by using a splicing algorithm;
step 10, obstacle avoidance is carried out by using acoustic equipment;
step 11, whether an obstacle is encountered or not is judged, if yes, the step 1 is returned, and if not, the step 12 is executed;
step 12, storing the spliced image;
and step 13, judging whether the process is finished or not, and if not, returning to the step 1.
Further, the backbone network of the global feature extractor comprises two standard 3 × 3 convolutions for global feature extraction, a ReLu function for preventing overfitting during training the network, and a non-local attention mechanism network for increasing correlation between pixels;
the local feature extractor includes a yolov4 single stage detector network for focusing on regions in the underwater image with obvious objects and a mask network for detecting the resulting local region feature map.
Further, the step 4 comprises the following steps:
step 40, setting a candidate frame with high detection precision and high confidence coefficient, wherein the candidate frame comprises the correlation attribute of the underwater image;
step 41, after selecting a plurality of candidate regions with high top-D detection precision, using an index i e to {1, 2.. multidot.D } to indicate each selected representative feature, wherein a spatial region covered by each representative feature is represented as AiFor each candidate region AiBy assigning 1 to the pixels in the region and 0 to the remaining pixels, a binary feature matrix is obtained, the formula is as follows:
Figure BDA0003142388600000021
wherein ,Mi∈{0,1}C×KAnd all AiIn the C x K range to ensure that all partial mask regions are within the image region, and then mapping the partial masks to global features, the global feature F is obtainedgAnd partial mask { Mi}D i1Then, mapping the partial mask to the global features to obtain a set of feature map sets { F ] based on the partial maski}D i1
Step 42, only retaining the most representative features by using the mask network, obtaining the mask features, and then combining the global features FgFor each partial region i, a mask feature map F can be obtainediNamely:
Figure BDA0003142388600000031
wherein an indication is in global feature FgOperate on each channel of,FiIs a mapping of the partial mask features of part i, at FiWith only the partial mask feature of part i activated and with Fi∈RC×K×GGlobal feature FgAnd feature map F for each maskiAnd connecting in channel dimensions, supplementing lost information of the underwater target after convolution, and obtaining a plurality of characteristics with strong robustness.
Further, the step 5 comprises the following steps:
step 50, for each mask feature map FiThe above operation is performed by using global average pooling to obtain each mask feature map FiEach channel corresponds to a different feature of the image;
step 51, for each mask feature map FiPerforming Softmax operation on the maximum value of the channel pixel to obtain a weight vector w for representing the importance of each representative feature, and normalizing w by taking the sum of w as 1 to make the relative importance among different features more obvious;
step 52, adding global feature FgTo enhance the importance of the representative feature region, thereby obtaining the most representative feature:
Figure BDA0003142388600000032
wherein ,
Figure BDA0003142388600000033
this can be predicted by the following equation:
Figure BDA0003142388600000034
where μ (·) denotes a learning function, σμIs a parameter of μ (·), mgap (·) represents the global average pooling. Further, the planning of the walking route of the underwater robot adopts an obstacle avoidance algorithm or a PID algorithm;
the advancing direction of the propeller is controlled through an operator-critic algorithm, the motion trail of the robot is planned, and the algorithm formula is as follows:
(1) critic is defined as the state action function Q
Figure BDA0003142388600000035
Where Q is the state action function, π is the reward penalty strategy, γ is the decay factor [0,1 ]],utIs an action state action function taken at time t, and can use state xtTo xt+1Is learned, state xtThe method comprises the following steps of (1) including the current movement direction and the current position of the underwater robot;
(2) when the target strategy is fixed, Q can be learned offline, and the formula for updating Q is as follows:
Figure BDA0003142388600000036
update QwUp to Qw≈Qπ, wherein ,QwA specific position coordinate point of the robot movement;
(3) actor is defined as the state cost function:
τ(μθ)=∫ρμr(xt,μ)dx=Ε[r(xtθ(xt))]
in neural networks, τ (u) is optimized by minimizing the loss functionθ) Then the loss function is:
Figure BDA0003142388600000041
where L (w) is a simple mean square error function, N represents the time range of the sample, yiIs a target state action value obtained from a target deep neural network Q, wherein,
yi=r(xi,u'i)-γQ'w(xi,ui),u'i=μ(xi|θ')
the loss function is graded as:
Figure BDA0003142388600000042
if the actor is represented by a neural network with parameter θ, then τ (μ)θ)=Qw(xi,μ(xi| θ) | w), graduating τ as:
Figure BDA0003142388600000043
compared with the prior art, the method for generating the underwater scene panoramic image of the marine culture area can provide a clear and complete underwater target panoramic image, so that remote control personnel on the water surface can better observe the comprehensive condition of an underwater target, and the method has very important practical significance for the underwater robot to realize target grabbing and the automatic and autonomous operation of an underwater operation robot.
Drawings
FIG. 1 is a flow chart of a method for generating a panoramic view of an underwater scene in a marine culture area, which is disclosed by the invention;
FIG. 2 is a schematic diagram of a movement path planning of an underwater robot;
FIG. 3 is a schematic diagram of an underwater target re-identification algorithm;
FIG. 4 is a schematic diagram of an underwater panoramic image generation algorithm.
Detailed Description
As shown in fig. 1 to 4, the method for generating a panoramic view of an underwater scene in a marine culture area disclosed by the invention comprises the following steps:
a method for generating a panoramic view of an underwater scene in a marine culture area comprises the following steps:
step 1, planning a walking route of an underwater robot, and controlling the motion direction of the underwater robot according to the walking route;
2, acquiring a plurality of images of a visible area of the underwater robot through a plurality of high-definition cameras arranged on the underwater robot;
step 3, extracting marine product target features of the collected images by using a local feature extractor and a global feature extractor to obtain a local feature map and a global feature map;
step 4, extracting target features from the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map through a softmax function to obtain a target mask feature map, and performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image;
step 5, sending the representative feature diagram into a DCP network for local guidance operation, increasing the weight of the features with more information of the feature diagram channels, and reducing the weight of the unimportant features with less information;
step 6, connecting the global feature map with the representative feature map to obtain a fusion feature map;
step 7, carrying out target re-identification detection on the multiple images in the visible area, and if the same object appears in two or more images, carrying out exposure processing on the images in which the same object is detected to obtain exposure processing images;
step 8, judging whether the exposure processing images have objects with the same characteristics, if not, returning to the step 2, and if so, executing the step 9;
step 9, carrying out image splicing on the images of the detected same object by using a splicing algorithm;
step 10, obstacle avoidance is carried out by using acoustic equipment;
step 11, whether an obstacle is encountered or not is judged, if yes, the step 1 is returned, and if not, the step 12 is executed;
step 12, storing the spliced image;
and step 13, judging whether the process is finished or not, and if not, returning to the step 1.
Specifically, a robot driving route is planned, and the movement direction is adjusted. In the invention, the obstacle avoidance algorithm or the PID algorithm is used for controlling the robot and planning the running path of the robot. The underwater robot has 8 propellers, realizes 6 degrees of freedom and supplies enough power to the robot. And controlling the advancing direction of the propeller through an actor-critic algorithm, and planning the motion track of the robot. The algorithm formula is as follows:
(1) critic is defined as the state action function Q
Figure BDA0003142388600000051
Q is the state action function, π is the strategy taken, γ is the decay factor [0,1 ]],utThe action state action function taken at time t may use state xtTo xt+1Is learned.
(2) When the target strategy is fixed, Q can be learned offline, and the formula for updating Q is as follows:
Figure BDA0003142388600000061
update QwUp to Qw≈Qπ
(3) actor is defined as the state cost function:
τ(μθ)=∫ρμr(xt,μ)dx=Ε[r(xtθ(xt))]
in neural networks, τ (u) is optimized by minimizing the loss functionθ) Then the loss function is:
Figure BDA0003142388600000062
where L (w) is a simple mean square error function, N represents the time range of the sample, yiIs a target state action value obtained from a target deep neural network Q, wherein,
yi=r(xi,u'i)-γQ'w(xi,ui),u'i=μ(xi|θ')
the loss function is graded as:
Figure BDA0003142388600000063
if the actor is represented by a neural network with parameter θ, then τ (μ)θ)=Qw(xi,μ(xi| θ) | w), graduating τ as:
Figure BDA0003142388600000064
the underwater high-definition camera is used for acquiring the image of the visual area of the robot, and the camera can only capture a part of light reflected from an object due to the attenuation of the light under water and the scattering effect of suspended particles, so that the underwater image has the visualization problems of blurring, color cast and the like. The resolution of underwater images acquired by a common underwater camera is insufficient, and the accuracy of underwater target detection is influenced. Therefore, a high-quality underwater image of the visible area of the robot is acquired by using the underwater high-definition camera. Simultaneously, a plurality of high-definition cameras are installed on the robot, so that the underwater environment can be observed at multiple angles.
Further, the backbone network of the global feature extractor comprises two standard 3 × 3 convolutions for global feature extraction, a ReLu function for preventing overfitting during training the network, and a non-local attention mechanism network for increasing correlation between pixels;
the local feature extractor comprises a yolov4 single-stage detector network for paying attention to the region with the obvious target in the underwater image and a mask network for detecting the obtained local region feature map, wherein the yolov4 single-stage detector network only pays attention to the region with the obvious target in the underwater image (the target can be fish and plankton in the sea, and the features (size, color, shape and the like) are obvious.
Specifically, feature blocks are extracted for marine product objects in an image of a visible area of the underwater robot.
(1) First, a backbone network (Global Feature Module) is composed of two standard 3 × 3 convolutions, ReLu functions and a non-local attention mechanism network, and a network structure diagram 3 is shown in the figure. The two standard convolution blocks are used for extracting the global features of the underwater shot image, after the global features are extracted, the attention mechanism network is used for increasing the correlation among pixels, and the global features F with strong robustness are extractedg. The global features are then used for input, training and final optimization of the next local feature network.
Further, the step 4 comprises the following steps:
step 40, setting a candidate frame with high detection precision and high confidence coefficient, wherein the candidate frame comprises related attributes of the underwater image, specifically, the detector can circle the detected object by using one detection frame, the detection precision is high and the confidence coefficient are complementary, the confidence coefficient refers to a threshold value, whether the threshold value is set properly or not can indirectly influence the detection precision, meanwhile, the threshold value is a probability, and the threshold value range is greater than or equal to 0.5 and less than or equal to 1;
step 41, after selecting a plurality of candidate regions with high top-D detection precision, using an index i e to {1, 2.. multidot.D } to indicate each selected representative feature, wherein a spatial region covered by each representative feature is represented as AiFor each candidate region AiBy assigning 1 to the pixels in the region and 0 to the remaining pixels, a binary feature matrix is obtained, the formula is as follows:
Figure BDA0003142388600000071
wherein ,MiE {0, 1} CxK, and all AiIn the C x K range to ensure that all partial mask regions are within the image region, and then mapping the partial masks to global features, the global feature F is obtainedgAnd partial mask { Mi}D i=1Then, mapping the partial mask to the global features to obtain a set of feature map sets { F ] based on the partial maski}D i=1(ii) a The mask area refers to an area with an obvious target in the image (the target can be fish or marine plankton, generally refers to visible and touchable, and has obvious characteristics), and is detected by a detector;
step 42, only retaining the most representative features by using the mask network, obtaining the mask features, and then combining the global features FgFor each partial region i, a mask feature map F can be obtainediNamely:
Figure BDA0003142388600000072
wherein an indication is in global feature FgIs operated on each channel of FiIs a mapping of the partial mask features of part i, at FiWith only the partial mask feature of part i activated and with Fi∈RC×K×GGlobal feature FgAnd feature map F for each maskiAnd connecting in channel dimensions, supplementing lost information of the underwater target after convolution, and obtaining a plurality of characteristics with strong robustness.
Further, the step 5 comprises the following steps:
step 50, for each mask feature map FiThe above operation is performed by using global average pooling to obtain each mask feature map FiExtracting a plurality of characteristics for an image according to the maximum value of the channel pixel, wherein each channel corresponds to different characteristics of the image;
step 51, for each mask feature map FiPerforming Softmax operation on the maximum value of the channel pixel to obtain a weight vector w for representing the importance of each representative feature, and normalizing w by taking the sum of w as 1 to make the relative importance among different features more obvious;
step 52, adding global feature FgTo enhance the importance of the representative feature region, thereby obtaining the most representative feature:
Figure BDA0003142388600000081
wherein ,
Figure BDA0003142388600000082
this can be predicted by the following equation:
Figure BDA0003142388600000083
where μ (·) denotes a learning function, σμIs a parameter of μ (·), mgap (·) represents the global average pooling. The importance of each feature is highlighted by using weight ratio, and the obtained result is added with a global feature map FgThe aim is to further highlight the most representative features of underwater targets. Finally, the feature fusion module compensates the global feature F because the global feature and the local feature can compensate each othergAnd the most representative feature FpConnected in channel dimension to obtain a fusion feature Ff. Since the global feature and the part feature based on the representative feature provide complementary information, the global feature and the most representative feature are concat connected in the channel dimension, and the formula is as follows:
Ff=concat(Fg,Fp),Ff∈RH×W×2C
the specific process of exposing the image of the same object in step 6 is as follows: the method comprises the steps that a camera collects multiple images in real time, the images are sent to an underwater image re-recognition network, in the network training process, the network extracts features of underwater image target objects such as sea cucumbers, scallops and the like through a local module and a global module, then feature recombination is carried out on the local features and the global features to obtain robust features, then the features are put into a DCP network to further enhance the features, finally the output features and the global features of the network are connected in channel dimensions, whether the images shot by the camera are the same object or not is judged through the features, if yes, the next step is carried out, and otherwise, the step 2 is carried out in a circulating mode. The step realizes the function of detecting whether the multiple pictures contain the same object, realizes whether the multiple pictures have the same object mark, namely whether each picture contains the same object, and considers that the multiple pictures contain the same target if the contact ratio is high (the threshold value is more than or equal to 0.5), otherwise, the steps are opposite.
And 8, performing image splicing by adopting a Multi-Band algorithm.
When fusing the macroscopic features, a large smooth gradual change area is adopted, and when fusing the local details, a small smooth gradual change area is adopted to decompose the image into component weighted sums of different frequency bands, wherein the macroscopic features of the image are in the low frequency band of the image, and the local details are in the high frequency band of the image. And the image is expanded according to the frequency to form a pyramid, the high and low frequency components are respectively weighted smoothly and superposed according to different modes, and the frequency band components are added again to obtain the final fusion effect.
And (3) image preprocessing, namely after the robot completes the splicing of the underwater images in the current target area, the super-resolution reconstruction technology is used for carrying out the clarification processing on the panoramic image, so that the observation effect is clearer, and the obtained underwater data is more accurate and precise.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art should be considered to be within the technical scope of the present invention, and the technical solutions and the inventive concepts thereof according to the present invention should be equivalent or changed within the scope of the present invention.

Claims (5)

1. A method for generating a panoramic image of an underwater scene in a marine culture area is characterized by comprising the following steps: the method comprises the following steps:
step 1, planning a walking route of an underwater robot, and controlling the motion direction of the underwater robot according to the walking route;
2, acquiring a plurality of images of a visible area of the underwater robot through a plurality of high-definition cameras arranged on the underwater robot;
step 3, extracting marine product target features of the collected images by using a local feature extractor and a global feature extractor to obtain a local feature map and a global feature map;
step 4, extracting target features from the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map through a softmax function to obtain a target mask feature map, and performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image;
step 5, sending the representative feature diagram into a DCP network for local guidance operation, increasing the weight of the features with more information of the feature diagram channels, and reducing the weight of the unimportant features with less information;
step 6, connecting the global feature map with the representative feature map to obtain a fusion feature map;
step 7, carrying out target re-identification detection on the multiple images in the visible area, and if the same object appears in two or more images, carrying out exposure processing on the images in which the same object is detected to obtain exposure processing images;
step 8, judging whether the exposure processing images have objects with the same characteristics, if not, returning to the step 2, and if so, executing the step 9;
step 9, carrying out image splicing on the images of the detected same object by using a splicing algorithm;
step 10, obstacle avoidance is carried out by using acoustic equipment;
step 11, whether an obstacle is encountered or not is judged, if yes, the step 1 is returned, and if not, the step 12 is executed;
step 12, storing the spliced image;
and step 13, judging whether the process is finished or not, and if not, returning to the step 1.
2. The method for generating the underwater scene panorama of the mariculture area as claimed in claim 1, wherein: the backbone network of the global feature extractor comprises two standard 3 x 3 convolutions for global feature extraction, a ReLu function for preventing overfitting during network training and a non-local attention mechanism network for increasing correlation between pixels;
the local feature extractor includes a yolov4 single stage detector network for focusing on regions in the underwater image with obvious objects and a mask network for detecting the resulting local region feature map.
3. The method for generating the underwater scene panorama of the mariculture area as claimed in claim 1, wherein: the step 4 comprises the following steps:
step 40, setting a candidate frame with high detection precision and high confidence coefficient, wherein the candidate frame comprises the correlation attribute of the underwater image;
step 41, after selecting a plurality of candidate regions with high top-D detection precision, using an index i e to {1, 2.. multidot.D } to indicate each selected representative feature, wherein a spatial region covered by each representative feature is represented as AiFor each candidate region AiBy assigning 1 to the pixels in the region and 0 to the remaining pixels, a binary feature matrix is obtained, the formula is as follows:
Figure FDA0003142388590000021
wherein ,Mi∈{0,1}C×KAnd all AiIn the C x K range to ensure that all partial mask regions are within the image region, and then mapping the partial masks to global features, the global feature F is obtainedgAnd partial mask { Mi}D i=1Then, mapping the partial mask to the global features to obtain a set of feature map sets { F ] based on the partial maski}D i=1
Step 42, only retaining the most representative features by using the mask network, obtaining the mask features, and then combining the global features FgFor each partial region i, a mask feature map F can be obtainediNamely:
Fi=Mi⊙Fg
Figure FDA0003142388590000022
wherein an indication is in global feature FgIs operated on each channel of FiIs a mapping of the partial mask features of part i, at FiWith only the partial mask feature of part i activated and with Fi∈RC×K×GGlobal feature FgAnd feature map F for each maskiAnd connecting in channel dimensions, supplementing lost information of the underwater target after convolution, and obtaining a plurality of characteristics with strong robustness.
4. The method for generating the panoramic view of the underwater scene in the mariculture area as claimed in claim 3, wherein: the step 5 comprises the following steps:
step 50, for each mask feature map FiThe above operation is performed by using global average pooling to obtain each mask feature map FiEach channel corresponds to a different feature of the image;
step 51, for each mask feature map FiPerforming Softmax operation on the maximum value of the channel pixel to obtain a weight vector w for representing the importance of each representative feature, and normalizing w by taking the sum of w as 1 to make the relative importance among different features more obvious;
step 52, adding global feature FgTo enhance the importance of the representative feature region, thereby obtaining the most representative feature:
Figure FDA0003142388590000031
wherein ,
Figure FDA0003142388590000032
this can be predicted by the following equation:
Figure FDA0003142388590000033
where μ (·) denotes a learning function, σμIs a parameter of μ (·), mgap (·) represents the global average pooling.
5. The method for generating the underwater scene panorama of the mariculture area as claimed in claim 4, wherein:
the planning of the walking route of the underwater robot adopts an obstacle avoidance algorithm or a PID algorithm;
the advancing direction of the propeller is controlled through an operator-critic algorithm, the motion trail of the robot is planned, and the algorithm formula is as follows:
(1) critic is defined as the state action function Q
Figure FDA0003142388590000034
Where Q is the state action function, π is the reward penalty strategy, γ is the decay factor [0,1 ]],utIs an action state action function taken at time t, and can use state xtTo xt+1Is learned, state xtThe method comprises the following steps of (1) including the current movement direction and the current position of the underwater robot;
(2) when the target strategy is fixed, Q can be learned offline, and the formula for updating Q is as follows:
Figure FDA0003142388590000035
update QwUp to Qw≈Qπ, wherein ,QwA specific position coordinate point of the robot movement;
(3) actor is defined as the state cost function:
τ(μθ)=∫ρμr(xt,μ)dx=Ε[r(xtθ(xt))]
in neural networksOptimizing τ (u) by minimizing the loss functionθ) Then the loss function is:
Figure FDA0003142388590000036
where L (w) is a simple mean square error function, N represents the time range of the sample, yiIs a target state action value obtained from a target deep neural network Q, wherein,
yi=r(xi,u'i)-γQ'w(xi,ui),u'i=μ(xi|θ')
the loss function is graded as:
Figure FDA0003142388590000041
if the actor is represented by a neural network with parameter θ, then τ (μ)θ)=Qw(xi,μ(xi| θ |) w), graduating τ as:
Figure FDA0003142388590000042
CN202110738641.7A 2021-06-30 2021-06-30 Ocean cultivation area underwater scene panorama generation method Active CN113538702B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110738641.7A CN113538702B (en) 2021-06-30 2021-06-30 Ocean cultivation area underwater scene panorama generation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110738641.7A CN113538702B (en) 2021-06-30 2021-06-30 Ocean cultivation area underwater scene panorama generation method

Publications (2)

Publication Number Publication Date
CN113538702A true CN113538702A (en) 2021-10-22
CN113538702B CN113538702B (en) 2023-05-23

Family

ID=78097402

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110738641.7A Active CN113538702B (en) 2021-06-30 2021-06-30 Ocean cultivation area underwater scene panorama generation method

Country Status (1)

Country Link
CN (1) CN113538702B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114332566A (en) * 2021-12-28 2022-04-12 中国航天空气动力技术研究院 Target detection method, system and device for underwater image
CN115359378A (en) * 2022-10-22 2022-11-18 长岛国家海洋公园管理中心(庙岛群岛海豹省级自然保护区管理中心) Ocean fishing equipment for determining fishing path based on offshore marine garbage distribution
CN117094895A (en) * 2023-09-05 2023-11-21 杭州一隅千象科技有限公司 Image panorama stitching method and system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013116100A1 (en) * 2012-01-30 2013-08-08 Google Inc. Apparatus and method for acquiring underwater images
US20140192144A1 (en) * 2013-01-05 2014-07-10 Patrick A. St. Clair Spherical panoramic image camera rig
US20170214836A1 (en) * 2016-01-27 2017-07-27 Kyocera Corporation Electronic apparatus, method for controlling electronic apparatus, and non-transitory computer readable recording medium
CN109993091A (en) * 2019-03-25 2019-07-09 浙江大学 A kind of monitor video object detection method eliminated based on background

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013116100A1 (en) * 2012-01-30 2013-08-08 Google Inc. Apparatus and method for acquiring underwater images
US20140192144A1 (en) * 2013-01-05 2014-07-10 Patrick A. St. Clair Spherical panoramic image camera rig
US20170214836A1 (en) * 2016-01-27 2017-07-27 Kyocera Corporation Electronic apparatus, method for controlling electronic apparatus, and non-transitory computer readable recording medium
CN109993091A (en) * 2019-03-25 2019-07-09 浙江大学 A kind of monitor video object detection method eliminated based on background

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114332566A (en) * 2021-12-28 2022-04-12 中国航天空气动力技术研究院 Target detection method, system and device for underwater image
CN115359378A (en) * 2022-10-22 2022-11-18 长岛国家海洋公园管理中心(庙岛群岛海豹省级自然保护区管理中心) Ocean fishing equipment for determining fishing path based on offshore marine garbage distribution
CN117094895A (en) * 2023-09-05 2023-11-21 杭州一隅千象科技有限公司 Image panorama stitching method and system
CN117094895B (en) * 2023-09-05 2024-03-26 杭州一隅千象科技有限公司 Image panorama stitching method and system

Also Published As

Publication number Publication date
CN113538702B (en) 2023-05-23

Similar Documents

Publication Publication Date Title
CN113538702A (en) Method for generating underwater scene panoramic image of marine culture area
WO2021142902A1 (en) Danet-based unmanned aerial vehicle coastline floating garbage inspection system
Wang et al. Real-time underwater onboard vision sensing system for robotic gripping
KR102530691B1 (en) Device and method for monitoring a berthing
CN114248893B (en) Operation type underwater robot for sea cucumber fishing and control method thereof
CN110889844B (en) Coral distribution and health condition assessment method based on deep clustering analysis
CN109919026B (en) Surface unmanned ship local path planning method
Levy et al. Automated analysis of marine video with limited data
KR20210007767A (en) Autonomous navigation ship system for removing sea waste based on deep learning-vision recognition
CN109859202B (en) Deep learning detection method based on USV water surface optical target tracking
CN110941996A (en) Target and track augmented reality method and system based on generation of countermeasure network
CN110335245A (en) Cage netting damage monitoring method and system based on monocular space and time continuous image
CN116255908B (en) Underwater robot-oriented marine organism positioning measurement device and method
CN112927264B (en) Unmanned aerial vehicle tracking shooting system and RGBD tracking method thereof
Rasmussen et al. Deep census: AUV-based scallop population monitoring
CN111723823B (en) Underwater target detection method based on third party transfer learning
CN110866548A (en) Infrared intelligent matching identification and distance measurement positioning method and system for insulator of power transmission line
CN113591592B (en) Overwater target identification method and device, terminal equipment and storage medium
CN116659516B (en) Depth three-dimensional attention visual navigation method and device based on binocular parallax mechanism
JP7375901B2 (en) Display device, display method, program and image projection system
Chen et al. Robust autonomous landing of UAVs in non-cooperative environments based on comprehensive terrain understanding
CN116824319A (en) Fusion method, device and storage medium of infrared image and visible light image
Liu et al. Deep underwater monocular depth estimation with single-beam echosounder
CN110393165A (en) A kind of off-lying sea cultivation net cage bait-throwing method based on Autoamtic bait putting ship
Xin et al. ULL-SLAM: underwater low-light enhancement for the front-end of visual SLAM

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant