CN113538702A - Method for generating underwater scene panoramic image of marine culture area - Google Patents
Method for generating underwater scene panoramic image of marine culture area Download PDFInfo
- Publication number
- CN113538702A CN113538702A CN202110738641.7A CN202110738641A CN113538702A CN 113538702 A CN113538702 A CN 113538702A CN 202110738641 A CN202110738641 A CN 202110738641A CN 113538702 A CN113538702 A CN 113538702A
- Authority
- CN
- China
- Prior art keywords
- feature map
- mask
- feature
- underwater
- target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 238000001514 detection method Methods 0.000 claims abstract description 17
- 238000012545 processing Methods 0.000 claims abstract description 14
- 239000011159 matrix material Substances 0.000 claims abstract description 7
- 238000000605 extraction Methods 0.000 claims abstract description 4
- 230000006870 function Effects 0.000 claims description 32
- 230000009471 action Effects 0.000 claims description 15
- 230000033001 locomotion Effects 0.000 claims description 12
- 238000010586 diagram Methods 0.000 claims description 10
- 238000013507 mapping Methods 0.000 claims description 9
- 238000013528 artificial neural network Methods 0.000 claims description 8
- 230000004927 fusion Effects 0.000 claims description 6
- 238000011176 pooling Methods 0.000 claims description 6
- 230000007246 mechanism Effects 0.000 claims description 5
- 230000008569 process Effects 0.000 claims description 5
- 238000012549 training Methods 0.000 claims description 5
- 230000001502 supplementing effect Effects 0.000 claims description 3
- 238000009364 mariculture Methods 0.000 claims 4
- 230000001537 neural effect Effects 0.000 claims 1
- 230000000694 effects Effects 0.000 abstract description 5
- 230000008859 change Effects 0.000 description 3
- 241000251468 Actinopterygii Species 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 241000251511 Holothuroidea Species 0.000 description 1
- 241000237503 Pectinidae Species 0.000 description 1
- 238000005352 clarification Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000006798 recombination Effects 0.000 description 1
- 238000005215 recombination Methods 0.000 description 1
- 235000020637 scallop Nutrition 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/003—Navigation within 3D models or images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/32—Indexing scheme for image data processing or generation, in general involving image mosaicing
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A40/00—Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
- Y02A40/80—Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in fisheries management
- Y02A40/81—Aquaculture, e.g. of fish
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computer Hardware Design (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Computer Graphics (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a method for generating a panoramic image of an underwater scene in a marine culture area, which comprises the steps of planning a walking path of an underwater robot, acquiring underwater images by using the underwater robot, and extracting marine product target characteristics of the acquired images to obtain a local characteristic image and a global characteristic image; and performing target feature extraction on the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map to obtain a target mask feature map, performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image, performing exposure and judgment on whether the images are the same image, and performing image splicing on the images of the same object by using a splicing algorithm to obtain an underwater spliced image. According to the method, after the underwater image splicing is completed in the current target area through the robot, the observation effect is clearer, more accurate and more precise.
Description
Technical Field
The invention relates to the technical field of underwater image acquisition, in particular to a method for generating a panoramic image of an underwater scene in a marine culture area.
Background
At present, when an underwater robot carries out close-range underwater target identification, the condition of a target needs to be observed clearly in real time. Due to the problems of dynamic change of underwater target distribution, serious attenuation and scattering effect of underwater light, limited visible range of an underwater camera and the like, complete observation of an identified target is difficult to realize by depending on information acquired and recorded by a single image obtained under the conditions of limited distance and view angle, so that observers are difficult to comprehensively master the condition of a submarine environment.
Disclosure of Invention
The invention provides a method for generating a panoramic view of an underwater scene in a marine culture area.
The technical means adopted by the invention are as follows:
a method for generating a panoramic view of an underwater scene in a marine culture area comprises the following steps:
2, acquiring a plurality of images of a visible area of the underwater robot through a plurality of high-definition cameras arranged on the underwater robot;
step 3, extracting marine product target features of the collected images by using a local feature extractor and a global feature extractor to obtain a local feature map and a global feature map;
step 4, extracting target features from the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map through a softmax function to obtain a target mask feature map, and performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image;
step 5, sending the representative feature diagram into a DCP network for local guidance operation, increasing the weight of the features with more information of the feature diagram channels, and reducing the weight of the unimportant features with less information;
step 6, connecting the global feature map with the representative feature map to obtain a fusion feature map;
step 7, carrying out target re-identification detection on the multiple images in the visible area, and if the same object appears in two or more images, carrying out exposure processing on the images in which the same object is detected to obtain exposure processing images;
step 8, judging whether the exposure processing images have objects with the same characteristics, if not, returning to the step 2, and if so, executing the step 9;
step 9, carrying out image splicing on the images of the detected same object by using a splicing algorithm;
step 10, obstacle avoidance is carried out by using acoustic equipment;
step 11, whether an obstacle is encountered or not is judged, if yes, the step 1 is returned, and if not, the step 12 is executed;
step 12, storing the spliced image;
and step 13, judging whether the process is finished or not, and if not, returning to the step 1.
Further, the backbone network of the global feature extractor comprises two standard 3 × 3 convolutions for global feature extraction, a ReLu function for preventing overfitting during training the network, and a non-local attention mechanism network for increasing correlation between pixels;
the local feature extractor includes a yolov4 single stage detector network for focusing on regions in the underwater image with obvious objects and a mask network for detecting the resulting local region feature map.
Further, the step 4 comprises the following steps:
step 40, setting a candidate frame with high detection precision and high confidence coefficient, wherein the candidate frame comprises the correlation attribute of the underwater image;
step 41, after selecting a plurality of candidate regions with high top-D detection precision, using an index i e to {1, 2.. multidot.D } to indicate each selected representative feature, wherein a spatial region covered by each representative feature is represented as AiFor each candidate region AiBy assigning 1 to the pixels in the region and 0 to the remaining pixels, a binary feature matrix is obtained, the formula is as follows:
wherein ,Mi∈{0,1}C×KAnd all AiIn the C x K range to ensure that all partial mask regions are within the image region, and then mapping the partial masks to global features, the global feature F is obtainedgAnd partial mask { Mi}D i=1Then, mapping the partial mask to the global features to obtain a set of feature map sets { F ] based on the partial maski}D i=1;
Step 42, only retaining the most representative features by using the mask network, obtaining the mask features, and then combining the global features FgFor each partial region i, a mask feature map F can be obtainediNamely:
wherein an indication is in global feature FgOperate on each channel of,FiIs a mapping of the partial mask features of part i, at FiWith only the partial mask feature of part i activated and with Fi∈RC×K×GGlobal feature FgAnd feature map F for each maskiAnd connecting in channel dimensions, supplementing lost information of the underwater target after convolution, and obtaining a plurality of characteristics with strong robustness.
Further, the step 5 comprises the following steps:
step 50, for each mask feature map FiThe above operation is performed by using global average pooling to obtain each mask feature map FiEach channel corresponds to a different feature of the image;
step 51, for each mask feature map FiPerforming Softmax operation on the maximum value of the channel pixel to obtain a weight vector w for representing the importance of each representative feature, and normalizing w by taking the sum of w as 1 to make the relative importance among different features more obvious;
step 52, adding global feature FgTo enhance the importance of the representative feature region, thereby obtaining the most representative feature:
where μ (·) denotes a learning function, σμIs a parameter of μ (·), mgap (·) represents the global average pooling. Further, the planning of the walking route of the underwater robot adopts an obstacle avoidance algorithm or a PID algorithm;
the advancing direction of the propeller is controlled through an operator-critic algorithm, the motion trail of the robot is planned, and the algorithm formula is as follows:
(1) critic is defined as the state action function Q
Where Q is the state action function, π is the reward penalty strategy, γ is the decay factor [0,1 ]],utIs an action state action function taken at time t, and can use state xtTo xt+1Is learned, state xtThe method comprises the following steps of (1) including the current movement direction and the current position of the underwater robot;
(2) when the target strategy is fixed, Q can be learned offline, and the formula for updating Q is as follows:
update QwUp to Qw≈Qπ, wherein ,QwA specific position coordinate point of the robot movement;
(3) actor is defined as the state cost function:
τ(μθ)=∫ρμr(xt,μ)dx=Ε[r(xt,μθ(xt))]
in neural networks, τ (u) is optimized by minimizing the loss functionθ) Then the loss function is:
where L (w) is a simple mean square error function, N represents the time range of the sample, yiIs a target state action value obtained from a target deep neural network Q, wherein,
yi=r(xi,u'i)-γQ'w(xi,ui),u'i=μ(xi|θ')
the loss function is graded as:
if the actor is represented by a neural network with parameter θ, then τ (μ)θ)=Qw(xi,μ(xi| θ) | w), graduating τ as:
compared with the prior art, the method for generating the underwater scene panoramic image of the marine culture area can provide a clear and complete underwater target panoramic image, so that remote control personnel on the water surface can better observe the comprehensive condition of an underwater target, and the method has very important practical significance for the underwater robot to realize target grabbing and the automatic and autonomous operation of an underwater operation robot.
Drawings
FIG. 1 is a flow chart of a method for generating a panoramic view of an underwater scene in a marine culture area, which is disclosed by the invention;
FIG. 2 is a schematic diagram of a movement path planning of an underwater robot;
FIG. 3 is a schematic diagram of an underwater target re-identification algorithm;
FIG. 4 is a schematic diagram of an underwater panoramic image generation algorithm.
Detailed Description
As shown in fig. 1 to 4, the method for generating a panoramic view of an underwater scene in a marine culture area disclosed by the invention comprises the following steps:
a method for generating a panoramic view of an underwater scene in a marine culture area comprises the following steps:
2, acquiring a plurality of images of a visible area of the underwater robot through a plurality of high-definition cameras arranged on the underwater robot;
step 3, extracting marine product target features of the collected images by using a local feature extractor and a global feature extractor to obtain a local feature map and a global feature map;
step 4, extracting target features from the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map through a softmax function to obtain a target mask feature map, and performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image;
step 5, sending the representative feature diagram into a DCP network for local guidance operation, increasing the weight of the features with more information of the feature diagram channels, and reducing the weight of the unimportant features with less information;
step 6, connecting the global feature map with the representative feature map to obtain a fusion feature map;
step 7, carrying out target re-identification detection on the multiple images in the visible area, and if the same object appears in two or more images, carrying out exposure processing on the images in which the same object is detected to obtain exposure processing images;
step 8, judging whether the exposure processing images have objects with the same characteristics, if not, returning to the step 2, and if so, executing the step 9;
step 9, carrying out image splicing on the images of the detected same object by using a splicing algorithm;
step 10, obstacle avoidance is carried out by using acoustic equipment;
step 11, whether an obstacle is encountered or not is judged, if yes, the step 1 is returned, and if not, the step 12 is executed;
step 12, storing the spliced image;
and step 13, judging whether the process is finished or not, and if not, returning to the step 1.
Specifically, a robot driving route is planned, and the movement direction is adjusted. In the invention, the obstacle avoidance algorithm or the PID algorithm is used for controlling the robot and planning the running path of the robot. The underwater robot has 8 propellers, realizes 6 degrees of freedom and supplies enough power to the robot. And controlling the advancing direction of the propeller through an actor-critic algorithm, and planning the motion track of the robot. The algorithm formula is as follows:
(1) critic is defined as the state action function Q
Q is the state action function, π is the strategy taken, γ is the decay factor [0,1 ]],utThe action state action function taken at time t may use state xtTo xt+1Is learned.
(2) When the target strategy is fixed, Q can be learned offline, and the formula for updating Q is as follows:
update QwUp to Qw≈Qπ。
(3) actor is defined as the state cost function:
τ(μθ)=∫ρμr(xt,μ)dx=Ε[r(xt,μθ(xt))]
in neural networks, τ (u) is optimized by minimizing the loss functionθ) Then the loss function is:
where L (w) is a simple mean square error function, N represents the time range of the sample, yiIs a target state action value obtained from a target deep neural network Q, wherein,
yi=r(xi,u'i)-γQ'w(xi,ui),u'i=μ(xi|θ')
the loss function is graded as:
if the actor is represented by a neural network with parameter θ, then τ (μ)θ)=Qw(xi,μ(xi| θ) | w), graduating τ as:
the underwater high-definition camera is used for acquiring the image of the visual area of the robot, and the camera can only capture a part of light reflected from an object due to the attenuation of the light under water and the scattering effect of suspended particles, so that the underwater image has the visualization problems of blurring, color cast and the like. The resolution of underwater images acquired by a common underwater camera is insufficient, and the accuracy of underwater target detection is influenced. Therefore, a high-quality underwater image of the visible area of the robot is acquired by using the underwater high-definition camera. Simultaneously, a plurality of high-definition cameras are installed on the robot, so that the underwater environment can be observed at multiple angles.
Further, the backbone network of the global feature extractor comprises two standard 3 × 3 convolutions for global feature extraction, a ReLu function for preventing overfitting during training the network, and a non-local attention mechanism network for increasing correlation between pixels;
the local feature extractor comprises a yolov4 single-stage detector network for paying attention to the region with the obvious target in the underwater image and a mask network for detecting the obtained local region feature map, wherein the yolov4 single-stage detector network only pays attention to the region with the obvious target in the underwater image (the target can be fish and plankton in the sea, and the features (size, color, shape and the like) are obvious.
Specifically, feature blocks are extracted for marine product objects in an image of a visible area of the underwater robot.
(1) First, a backbone network (Global Feature Module) is composed of two standard 3 × 3 convolutions, ReLu functions and a non-local attention mechanism network, and a network structure diagram 3 is shown in the figure. The two standard convolution blocks are used for extracting the global features of the underwater shot image, after the global features are extracted, the attention mechanism network is used for increasing the correlation among pixels, and the global features F with strong robustness are extractedg. The global features are then used for input, training and final optimization of the next local feature network.
Further, the step 4 comprises the following steps:
step 40, setting a candidate frame with high detection precision and high confidence coefficient, wherein the candidate frame comprises related attributes of the underwater image, specifically, the detector can circle the detected object by using one detection frame, the detection precision is high and the confidence coefficient are complementary, the confidence coefficient refers to a threshold value, whether the threshold value is set properly or not can indirectly influence the detection precision, meanwhile, the threshold value is a probability, and the threshold value range is greater than or equal to 0.5 and less than or equal to 1;
step 41, after selecting a plurality of candidate regions with high top-D detection precision, using an index i e to {1, 2.. multidot.D } to indicate each selected representative feature, wherein a spatial region covered by each representative feature is represented as AiFor each candidate region AiBy assigning 1 to the pixels in the region and 0 to the remaining pixels, a binary feature matrix is obtained, the formula is as follows:
wherein ,MiE {0, 1} CxK, and all AiIn the C x K range to ensure that all partial mask regions are within the image region, and then mapping the partial masks to global features, the global feature F is obtainedgAnd partial mask { Mi}D i=1Then, mapping the partial mask to the global features to obtain a set of feature map sets { F ] based on the partial maski}D i=1(ii) a The mask area refers to an area with an obvious target in the image (the target can be fish or marine plankton, generally refers to visible and touchable, and has obvious characteristics), and is detected by a detector;
step 42, only retaining the most representative features by using the mask network, obtaining the mask features, and then combining the global features FgFor each partial region i, a mask feature map F can be obtainediNamely:
wherein an indication is in global feature FgIs operated on each channel of FiIs a mapping of the partial mask features of part i, at FiWith only the partial mask feature of part i activated and with Fi∈RC×K×GGlobal feature FgAnd feature map F for each maskiAnd connecting in channel dimensions, supplementing lost information of the underwater target after convolution, and obtaining a plurality of characteristics with strong robustness.
Further, the step 5 comprises the following steps:
step 50, for each mask feature map FiThe above operation is performed by using global average pooling to obtain each mask feature map FiExtracting a plurality of characteristics for an image according to the maximum value of the channel pixel, wherein each channel corresponds to different characteristics of the image;
step 51, for each mask feature map FiPerforming Softmax operation on the maximum value of the channel pixel to obtain a weight vector w for representing the importance of each representative feature, and normalizing w by taking the sum of w as 1 to make the relative importance among different features more obvious;
step 52, adding global feature FgTo enhance the importance of the representative feature region, thereby obtaining the most representative feature:
where μ (·) denotes a learning function, σμIs a parameter of μ (·), mgap (·) represents the global average pooling. The importance of each feature is highlighted by using weight ratio, and the obtained result is added with a global feature map FgThe aim is to further highlight the most representative features of underwater targets. Finally, the feature fusion module compensates the global feature F because the global feature and the local feature can compensate each othergAnd the most representative feature FpConnected in channel dimension to obtain a fusion feature Ff. Since the global feature and the part feature based on the representative feature provide complementary information, the global feature and the most representative feature are concat connected in the channel dimension, and the formula is as follows:
Ff=concat(Fg,Fp),Ff∈RH×W×2C。
the specific process of exposing the image of the same object in step 6 is as follows: the method comprises the steps that a camera collects multiple images in real time, the images are sent to an underwater image re-recognition network, in the network training process, the network extracts features of underwater image target objects such as sea cucumbers, scallops and the like through a local module and a global module, then feature recombination is carried out on the local features and the global features to obtain robust features, then the features are put into a DCP network to further enhance the features, finally the output features and the global features of the network are connected in channel dimensions, whether the images shot by the camera are the same object or not is judged through the features, if yes, the next step is carried out, and otherwise, the step 2 is carried out in a circulating mode. The step realizes the function of detecting whether the multiple pictures contain the same object, realizes whether the multiple pictures have the same object mark, namely whether each picture contains the same object, and considers that the multiple pictures contain the same target if the contact ratio is high (the threshold value is more than or equal to 0.5), otherwise, the steps are opposite.
And 8, performing image splicing by adopting a Multi-Band algorithm.
When fusing the macroscopic features, a large smooth gradual change area is adopted, and when fusing the local details, a small smooth gradual change area is adopted to decompose the image into component weighted sums of different frequency bands, wherein the macroscopic features of the image are in the low frequency band of the image, and the local details are in the high frequency band of the image. And the image is expanded according to the frequency to form a pyramid, the high and low frequency components are respectively weighted smoothly and superposed according to different modes, and the frequency band components are added again to obtain the final fusion effect.
And (3) image preprocessing, namely after the robot completes the splicing of the underwater images in the current target area, the super-resolution reconstruction technology is used for carrying out the clarification processing on the panoramic image, so that the observation effect is clearer, and the obtained underwater data is more accurate and precise.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art should be considered to be within the technical scope of the present invention, and the technical solutions and the inventive concepts thereof according to the present invention should be equivalent or changed within the scope of the present invention.
Claims (5)
1. A method for generating a panoramic image of an underwater scene in a marine culture area is characterized by comprising the following steps: the method comprises the following steps:
step 1, planning a walking route of an underwater robot, and controlling the motion direction of the underwater robot according to the walking route;
2, acquiring a plurality of images of a visible area of the underwater robot through a plurality of high-definition cameras arranged on the underwater robot;
step 3, extracting marine product target features of the collected images by using a local feature extractor and a global feature extractor to obtain a local feature map and a global feature map;
step 4, extracting target features from the local feature map through a target detection algorithm model to obtain a local mask feature map with strong robustness, performing mask processing on the target feature map through a softmax function to obtain a target mask feature map, and performing matrix mask operation on the target mask feature map and the global feature map to obtain a representative feature map in the image;
step 5, sending the representative feature diagram into a DCP network for local guidance operation, increasing the weight of the features with more information of the feature diagram channels, and reducing the weight of the unimportant features with less information;
step 6, connecting the global feature map with the representative feature map to obtain a fusion feature map;
step 7, carrying out target re-identification detection on the multiple images in the visible area, and if the same object appears in two or more images, carrying out exposure processing on the images in which the same object is detected to obtain exposure processing images;
step 8, judging whether the exposure processing images have objects with the same characteristics, if not, returning to the step 2, and if so, executing the step 9;
step 9, carrying out image splicing on the images of the detected same object by using a splicing algorithm;
step 10, obstacle avoidance is carried out by using acoustic equipment;
step 11, whether an obstacle is encountered or not is judged, if yes, the step 1 is returned, and if not, the step 12 is executed;
step 12, storing the spliced image;
and step 13, judging whether the process is finished or not, and if not, returning to the step 1.
2. The method for generating the underwater scene panorama of the mariculture area as claimed in claim 1, wherein: the backbone network of the global feature extractor comprises two standard 3 x 3 convolutions for global feature extraction, a ReLu function for preventing overfitting during network training and a non-local attention mechanism network for increasing correlation between pixels;
the local feature extractor includes a yolov4 single stage detector network for focusing on regions in the underwater image with obvious objects and a mask network for detecting the resulting local region feature map.
3. The method for generating the underwater scene panorama of the mariculture area as claimed in claim 1, wherein: the step 4 comprises the following steps:
step 40, setting a candidate frame with high detection precision and high confidence coefficient, wherein the candidate frame comprises the correlation attribute of the underwater image;
step 41, after selecting a plurality of candidate regions with high top-D detection precision, using an index i e to {1, 2.. multidot.D } to indicate each selected representative feature, wherein a spatial region covered by each representative feature is represented as AiFor each candidate region AiBy assigning 1 to the pixels in the region and 0 to the remaining pixels, a binary feature matrix is obtained, the formula is as follows:
wherein ,Mi∈{0,1}C×KAnd all AiIn the C x K range to ensure that all partial mask regions are within the image region, and then mapping the partial masks to global features, the global feature F is obtainedgAnd partial mask { Mi}D i=1Then, mapping the partial mask to the global features to obtain a set of feature map sets { F ] based on the partial maski}D i=1;
Step 42, only retaining the most representative features by using the mask network, obtaining the mask features, and then combining the global features FgFor each partial region i, a mask feature map F can be obtainediNamely:
wherein an indication is in global feature FgIs operated on each channel of FiIs a mapping of the partial mask features of part i, at FiWith only the partial mask feature of part i activated and with Fi∈RC×K×GGlobal feature FgAnd feature map F for each maskiAnd connecting in channel dimensions, supplementing lost information of the underwater target after convolution, and obtaining a plurality of characteristics with strong robustness.
4. The method for generating the panoramic view of the underwater scene in the mariculture area as claimed in claim 3, wherein: the step 5 comprises the following steps:
step 50, for each mask feature map FiThe above operation is performed by using global average pooling to obtain each mask feature map FiEach channel corresponds to a different feature of the image;
step 51, for each mask feature map FiPerforming Softmax operation on the maximum value of the channel pixel to obtain a weight vector w for representing the importance of each representative feature, and normalizing w by taking the sum of w as 1 to make the relative importance among different features more obvious;
step 52, adding global feature FgTo enhance the importance of the representative feature region, thereby obtaining the most representative feature:
where μ (·) denotes a learning function, σμIs a parameter of μ (·), mgap (·) represents the global average pooling.
5. The method for generating the underwater scene panorama of the mariculture area as claimed in claim 4, wherein:
the planning of the walking route of the underwater robot adopts an obstacle avoidance algorithm or a PID algorithm;
the advancing direction of the propeller is controlled through an operator-critic algorithm, the motion trail of the robot is planned, and the algorithm formula is as follows:
(1) critic is defined as the state action function Q
Where Q is the state action function, π is the reward penalty strategy, γ is the decay factor [0,1 ]],utIs an action state action function taken at time t, and can use state xtTo xt+1Is learned, state xtThe method comprises the following steps of (1) including the current movement direction and the current position of the underwater robot;
(2) when the target strategy is fixed, Q can be learned offline, and the formula for updating Q is as follows:
update QwUp to Qw≈Qπ, wherein ,QwA specific position coordinate point of the robot movement;
(3) actor is defined as the state cost function:
τ(μθ)=∫ρμr(xt,μ)dx=Ε[r(xt,μθ(xt))]
in neural networksOptimizing τ (u) by minimizing the loss functionθ) Then the loss function is:
where L (w) is a simple mean square error function, N represents the time range of the sample, yiIs a target state action value obtained from a target deep neural network Q, wherein,
yi=r(xi,u'i)-γQ'w(xi,ui),u'i=μ(xi|θ')
the loss function is graded as:
if the actor is represented by a neural network with parameter θ, then τ (μ)θ)=Qw(xi,μ(xi| θ |) w), graduating τ as:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110738641.7A CN113538702B (en) | 2021-06-30 | 2021-06-30 | Ocean cultivation area underwater scene panorama generation method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110738641.7A CN113538702B (en) | 2021-06-30 | 2021-06-30 | Ocean cultivation area underwater scene panorama generation method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113538702A true CN113538702A (en) | 2021-10-22 |
CN113538702B CN113538702B (en) | 2023-05-23 |
Family
ID=78097402
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110738641.7A Active CN113538702B (en) | 2021-06-30 | 2021-06-30 | Ocean cultivation area underwater scene panorama generation method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113538702B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114332566A (en) * | 2021-12-28 | 2022-04-12 | 中国航天空气动力技术研究院 | Target detection method, system and device for underwater image |
CN115359378A (en) * | 2022-10-22 | 2022-11-18 | 长岛国家海洋公园管理中心(庙岛群岛海豹省级自然保护区管理中心) | Ocean fishing equipment for determining fishing path based on offshore marine garbage distribution |
CN117094895A (en) * | 2023-09-05 | 2023-11-21 | 杭州一隅千象科技有限公司 | Image panorama stitching method and system |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013116100A1 (en) * | 2012-01-30 | 2013-08-08 | Google Inc. | Apparatus and method for acquiring underwater images |
US20140192144A1 (en) * | 2013-01-05 | 2014-07-10 | Patrick A. St. Clair | Spherical panoramic image camera rig |
US20170214836A1 (en) * | 2016-01-27 | 2017-07-27 | Kyocera Corporation | Electronic apparatus, method for controlling electronic apparatus, and non-transitory computer readable recording medium |
CN109993091A (en) * | 2019-03-25 | 2019-07-09 | 浙江大学 | A kind of monitor video object detection method eliminated based on background |
-
2021
- 2021-06-30 CN CN202110738641.7A patent/CN113538702B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013116100A1 (en) * | 2012-01-30 | 2013-08-08 | Google Inc. | Apparatus and method for acquiring underwater images |
US20140192144A1 (en) * | 2013-01-05 | 2014-07-10 | Patrick A. St. Clair | Spherical panoramic image camera rig |
US20170214836A1 (en) * | 2016-01-27 | 2017-07-27 | Kyocera Corporation | Electronic apparatus, method for controlling electronic apparatus, and non-transitory computer readable recording medium |
CN109993091A (en) * | 2019-03-25 | 2019-07-09 | 浙江大学 | A kind of monitor video object detection method eliminated based on background |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114332566A (en) * | 2021-12-28 | 2022-04-12 | 中国航天空气动力技术研究院 | Target detection method, system and device for underwater image |
CN115359378A (en) * | 2022-10-22 | 2022-11-18 | 长岛国家海洋公园管理中心(庙岛群岛海豹省级自然保护区管理中心) | Ocean fishing equipment for determining fishing path based on offshore marine garbage distribution |
CN117094895A (en) * | 2023-09-05 | 2023-11-21 | 杭州一隅千象科技有限公司 | Image panorama stitching method and system |
CN117094895B (en) * | 2023-09-05 | 2024-03-26 | 杭州一隅千象科技有限公司 | Image panorama stitching method and system |
Also Published As
Publication number | Publication date |
---|---|
CN113538702B (en) | 2023-05-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113538702A (en) | Method for generating underwater scene panoramic image of marine culture area | |
WO2021142902A1 (en) | Danet-based unmanned aerial vehicle coastline floating garbage inspection system | |
Wang et al. | Real-time underwater onboard vision sensing system for robotic gripping | |
KR102530691B1 (en) | Device and method for monitoring a berthing | |
CN114248893B (en) | Operation type underwater robot for sea cucumber fishing and control method thereof | |
CN110889844B (en) | Coral distribution and health condition assessment method based on deep clustering analysis | |
CN109919026B (en) | Surface unmanned ship local path planning method | |
Levy et al. | Automated analysis of marine video with limited data | |
KR20210007767A (en) | Autonomous navigation ship system for removing sea waste based on deep learning-vision recognition | |
CN109859202B (en) | Deep learning detection method based on USV water surface optical target tracking | |
CN110941996A (en) | Target and track augmented reality method and system based on generation of countermeasure network | |
CN110335245A (en) | Cage netting damage monitoring method and system based on monocular space and time continuous image | |
CN116255908B (en) | Underwater robot-oriented marine organism positioning measurement device and method | |
CN112927264B (en) | Unmanned aerial vehicle tracking shooting system and RGBD tracking method thereof | |
Rasmussen et al. | Deep census: AUV-based scallop population monitoring | |
CN111723823B (en) | Underwater target detection method based on third party transfer learning | |
CN110866548A (en) | Infrared intelligent matching identification and distance measurement positioning method and system for insulator of power transmission line | |
CN113591592B (en) | Overwater target identification method and device, terminal equipment and storage medium | |
CN116659516B (en) | Depth three-dimensional attention visual navigation method and device based on binocular parallax mechanism | |
JP7375901B2 (en) | Display device, display method, program and image projection system | |
Chen et al. | Robust autonomous landing of UAVs in non-cooperative environments based on comprehensive terrain understanding | |
CN116824319A (en) | Fusion method, device and storage medium of infrared image and visible light image | |
Liu et al. | Deep underwater monocular depth estimation with single-beam echosounder | |
CN110393165A (en) | A kind of off-lying sea cultivation net cage bait-throwing method based on Autoamtic bait putting ship | |
Xin et al. | ULL-SLAM: underwater low-light enhancement for the front-end of visual SLAM |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |