CN110826575A - Underwater target identification method based on machine learning - Google Patents
Underwater target identification method based on machine learning Download PDFInfo
- Publication number
- CN110826575A CN110826575A CN201910950105.6A CN201910950105A CN110826575A CN 110826575 A CN110826575 A CN 110826575A CN 201910950105 A CN201910950105 A CN 201910950105A CN 110826575 A CN110826575 A CN 110826575A
- Authority
- CN
- China
- Prior art keywords
- target
- underwater
- box
- scale
- machine learning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 238000010801 machine learning Methods 0.000 title claims abstract description 15
- 230000005764 inhibitory process Effects 0.000 claims abstract description 6
- 238000012545 processing Methods 0.000 claims abstract description 5
- 238000004364 calculation method Methods 0.000 claims description 10
- 230000006870 function Effects 0.000 claims description 10
- 239000011159 matrix material Substances 0.000 claims description 8
- 238000010586 diagram Methods 0.000 claims description 4
- 238000010606 normalization Methods 0.000 claims description 3
- 238000012216 screening Methods 0.000 claims description 3
- 238000007781 pre-processing Methods 0.000 claims description 2
- 238000001514 detection method Methods 0.000 abstract description 21
- 230000001629 suppression Effects 0.000 abstract description 7
- 230000000007 visual effect Effects 0.000 abstract description 2
- 238000013403 standard screening design Methods 0.000 description 14
- 241000251468 Actinopterygii Species 0.000 description 7
- 238000013527 convolutional neural network Methods 0.000 description 6
- 238000012549 training Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 230000007547 defect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 239000013598 vector Substances 0.000 description 2
- 241000238557 Decapoda Species 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000007717 exclusion Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 230000010355 oscillation Effects 0.000 description 1
- 230000036544 posture Effects 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
- G06V10/464—Salient features, e.g. scale invariant feature transforms [SIFT] using a plurality of salient features, e.g. bag-of-words [BoW] representations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
An underwater target identification method based on machine learning belongs to the technical field of underwater machine vision detection processing. The core of the underwater target recognition algorithm is an SSD target detection algorithm, a feedforward convolution network structure is adopted, convolution boxes with different sizes and scales are adopted between different layers to carry out convolution to obtain feature maps with different scales, regression is carried out according to the feature maps, and finally a result is obtained through a non-maximum suppression algorithm. The detection of the target object can be realized through the non-maximum inhibition algorithm, the recognition accuracy of the underwater target can be greatly improved, effective visual information is provided for the underwater robot to observe and operate the underwater target, and the intelligent recognition capability of the underwater target is improved.
Description
Technical Field
The invention belongs to the technical field of underwater machine vision detection, and particularly relates to an underwater target identification method based on machine learning.
Background
The ocean covers most of the earth's area, and has great resources and secret. In recent times, with the progress of the development technology of marine equipment, people begin to further recognize and develop marine resources, and various countries and regions in the world begin to develop marine equipment and exploit underwater resources without much energy. China has coastlines of nearly 2 kilometers, coastal economic sea areas have abundant ocean resources, good ocean development conditions and high requirements, and computer vision equipment serving as one of underwater sensing equipment is increasingly put into ocean detection. The complete vision system covers various disciplines and technologies such as optics, computer science control theory and the like, and is widely carried on underwater detection, operation and manned equipment, wherein the underwater tracking identification technology based on vision has very important research value.
The underwater robot is provided with the vision equipment, so that the underwater environment can be better sensed, and meanwhile, the underwater equipment can be saved more space by single-purpose installation, so that effective conditions are provided for precise and maneuvering operation of the underwater equipment. The related research of the invention aims to provide target and space information for underwater equipment operation, underwater robot obstacle avoidance, path planning and the like, so that the underwater monocular vision technology has practical research and practical application values.
The current underwater target recognition algorithm has the main defects that:
firstly, the precision of the current underwater vision detection is not enough, and the traditional recognition algorithm is not enough for the positioning precision of the target, so that the target detection is greatly influenced.
At present, algorithms based on machine learning are mainly divided into two types, one is a two-stage method such as R-CNN, the other is a one-stage, SSD belongs to the latter, and compared with Yolo, the SSD algorithm is much better than that in terms of accuracy and speed. In contrast to Yolo, SSD uses CNN to detect directly, rather than after the full connectivity layer as does Yolo. The method is characterized in that the detection is directly carried out by convolution only at one difference of SSD compared with Yolo, and two important changes are carried out, namely, the SSD extracts feature maps with different scales for detection, a large-scale feature map (a feature map closer to the front) can be used for detecting small objects, and a small-scale feature map (a feature map closer to the rear) is used for detecting large objects; the second is that the SSD uses Prior boxes (the former boxes, Default boxes, called Anchors in Faster R-CNN) of different dimensions and aspect ratios. The Yolo algorithm has the disadvantages of difficulty in detecting small targets and inaccurate positioning, but the important improvements enable the SSD to overcome the disadvantages to some extent.
Compared with the traditional CNN, the neural network, namely the network with an input layer, an output layer and a plurality of hidden layers, belongs to a sub-field of machine learning. The main principle is that after the training data enters the input layer, the training data is processed by a plurality of hidden layers, and a large amount of training is carried out to extract certain input characteristics, and the model is used for carrying out classification or prediction in practice. The traditional method has several problems in underwater target identification, such as low processing speed, low identification rate and the like.
SSDs, a single-shot detector for multiple classes, were introduced that are faster than the most advanced for single-shot detectors (yolo) and others before, with similar accuracy but faster speed compared to the fast-Rcnn and others identification networks. The core of the SSD is to predict the class scores and box offsets for a set of default bounding boxes using a small volume filter applied to the feature map. The design characteristics lead to simple end-to-end training and high precision, even in low-resolution input images, the conditions of unclear images and low resolution in an underwater environment can be met, and the balance between speed and accuracy is further improved.
Disclosure of Invention
The invention aims to provide an underwater target identification method based on machine learning, which can detect underwater target information more accurately.
The purpose of the invention is realized as follows:
an underwater target identification method based on machine learning comprises the following steps:
step 1: preprocessing an underwater target image, and performing retinex restoration on the image to obtain an input image;
step 2: carrying out convolution by adopting convolution boxes with different sizes and scales among different layers to obtain characteristic graphs with different scales;
and step 3: performing regression according to the feature maps obtained in the step two, performing normalization processing on each feature map, and obtaining defuelt boxes with different scales and offset through detectors and classifiers with different sizes;
and 4, step 4: obtaining a final result through an NMS inhibition algorithm;
in the step2, a region result obtained by RPN calculation in a region selection algorithm is adopted, the feature maps with different sizes are obtained by dividing according to different scales, then the RPN convolution core is used for moving the region on the feature map and obtaining the confidence value of the region, and a matrix with the confidence is obtained by continuously moving defualt box with different scales.
The width and height of the defuelt box in the step3 are as follows:
whereinAndwidth and height of defualt box, respectively, the scale being ar∈{1,2,1/2,3,1/3},Is the scale of different layers, wherein m is the number, the default scale is realized by m different scale characteristic diagram predictions, SminIs of the smallest scale, SmaxIs the scale of the largest feature map.
The central position of the selection frame in the step3 is as follows:
wherein x and y represent the horizontal frame, the central point of the longitudinal axis, | fkI represents the k-th feature graph scale, i.j E [0, fk]。
The method for NMS inhibition in step4 comprises the following steps:
step 4.1: arranging elements in the matrix according to the conf size;
step 4.2: IoU calculation is carried out on all the cross areas according to the sequence from large to small of the calculation result in the step 4.1, Th values are set, comparison is carried out on the Th values according to the sequence and IoU, and classification are carried out according to the sizes of the Th values;
step 4.3: the execution returns to the step 4.2 again at the second big frame position of the column;
step 4.4: repeating the step 4.3 until all the defuelt boxes in the row are executed;
step 4.5: performing traversal on the matrix, namely performing NMS of all categories;
step 4.6: and further residual screening is carried out, and all the final residual categories are selected according to the confidence coefficient.
The loss function of the whole model in the step4 is
Wherein x is used to determine if a designed feature capture box is presentThere is a corresponding target for the purpose of,whether the ith box is matched with the jth target frame of the pth object or not is represented, the matching is 1, and otherwise, the matching is 0; if it isIndicating that at least one box matches the jth target bounding box; n represents the number of matching sums;to measure the performance of the recognition;used to measure bounding box prediction performance; wherein And (2) representing the deviation between the real target frame of the jth target and the frame of the characteristic grabbing box, wherein m belongs to { cx, cy, w, h }, wherein (cx, cy) represents the coordinate of the center point of the frame, and (w, h) represents the width and the height of the frame.
The invention has the beneficial effects that:
(1) the method adopts a multi-scale and anchor point mode to solve the problem of low precision of the regional proposal, adopts multi-scale characteristic vectors, greatly improves the good effect of combining small targets and large targets, simultaneously has great help to improve the overall identification accuracy, and can obtain more accurate position information compared with the prior proposal methods;
(2) the algorithm carries out 20000 iterations to calculate an error curve according to a loss function, the initial error is defined as 500, namely the convergence range is (0,500), the final convergence is about 20, the error rate is less than about one percent, and the accuracy of the algorithm is greatly improved.
Drawings
FIG. 1 is a diagram of a multi-scale detection implementation;
FIG. 2 is a region selection method;
FIG. 3 is a NMS algorithm flow chart;
FIG. 4 is a graph of stack rate calculations;
FIG. 5 is a view of individual fish and multiple fish detection at multiple angles;
fig. 6 is a graph of error variation.
Detailed Description
A detailed embodiment and effect of the present invention will be illustrated by the following examples in conjunction with the summary of the invention.
Aiming at the defects in the prior art, the invention aims to provide an underwater target detection algorithm with high reliability and good real-time performance, and underwater target information can be detected more accurately. The algorithm can meet the requirements of underwater observation and operation, and provides accurate identification of underwater targets for the underwater robot. The underwater target recognition method can greatly improve the recognition accuracy of the underwater target, provide effective visual information for the underwater target observation and operation of the underwater robot, and improve the intelligent recognition capability of the underwater target.
Implementation 1: as shown in the attached figure 1, the invention realizes an underwater target recognition algorithm based on machine learning according to the requirements of underwater observation and operation. The core of the algorithm is an SSD target detection algorithm, SSD is a method for detecting targets with very high precision, it adopts a feedforward convolution network structure, firstly preprocesses an underwater target image, performs retinex restoration on the image to obtain an input image, then, convolution is carried out by adopting convolution boxes with different sizes and scales among different layers to obtain feature maps with different scales, regression is carried out according to the obtained feature maps, and finally a final result is obtained by an NMS (network management system) inhibition method, the SSD target detection method adopts a multi-scale and anchor point mode to solve the problem of low precision of region suggestion, the adopted multi-scale feature vector greatly improves the good effect on both small targets and large targets, meanwhile, the method greatly helps to improve the overall identification accuracy, and can obtain more accurate position information compared with the conventional suggestion method. And the subsequent convolution network uses convolution templates with different scales to perform feature fusion so as to obtain default boxes with different scales and offset, and finally, an extreme value suppression algorithm is added to realize a final detection classification result. The specific steps are that a standard VGG-16 network is adopted, feature detection is carried out through convolution templates with different sizes to obtain different feature maps, normalization processing is carried out on each feature map, then defuelt box with different scales and offset is obtained through detectors and classifiers with different sizes, and finally detection of a target object is achieved through a non-maximum suppression algorithm.
The SSD multi-scale implementation process is shown in FIG. 1, and it is described here that a standard convolutional neural network VGG-16 is adopted as a network for feature extraction at the front end, and the subsequent convolutional network performs feature fusion by using convolutional templates of different scales, so as to obtain default boxes of different scales and with offset, and finally, a final detection classification result is realized by adding an extremum suppression algorithm.
Implementation 2: as shown in fig. 2, in the convolution network algorithm of SSD, where RPN (region pro posal network) can be used to calculate the obtained region result, and divide the region result according to different scales to obtain multiple regions with feature map sizes 38 × 38 × 512, 19 × 19 × 1024, 10 × 10 × 512, 5 × 5 × 256, 3 × 3 × 256, 1 × 1 × 256, etc., the RPN process is illustrated according to fig. 2, where 5 × 5 × 256 is taken as an example, where k default boxes can be seen to be generated by different scales and ratios, where k ═ 6 can generate 5 × 5 × 256 convolution boxes and 4 default boxes with offset have corresponding confidence classes.
The default scale is accomplished here by m distinct scale feature map predictions, where the smallest scale is set to SminSet the maximum profile scale to S0.2maxIf it is 0.95, the dimensions of all layers can be obtained in a recursive manner according to the corresponding dimensions, specifically:
here according to different ratios to arWherein the scale is arE {1,2,1/2,3,1/3}, and is in accordance with arThe width and height of the corresponding default box can be found.
WhereinAndwidth and height of the defuelt box, respectively, and in the case of radio equal to 1, it is necessary to re-scale the box toThus, default boxes with 6 different dimensions can be obtained.
Meanwhile, the center position of the selection frame can be obtained by the following formula (3).
Wherein, | fkI represents the k-th characteristic diagram scale, x and y represent the center points of the horizontal and vertical axes of the frame respectively, i.j E [0, fk]。
The NMS (Non-extreme suppression) algorithm cannot accurately locate a true position in a convolutional neural network, and a Non-extreme suppression NMS (Non-maximum suppression) method needs to be adopted to alleviate the situation.
Implementation 3: as shown in figures 3 and 4 of the drawings,
step 1: the elements in the 8732 × 21 matrix are arranged according to the Conf size.
Step 2: IoU calculation is carried out on all the intersection areas according to the descending order of the calculation results in the first step, Th values are set, comparison is carried out on the Th values according to the order of IoU, and classification are carried out according to the sizes of the Th values.
Step 3: execution resumes at Step2 at the second big box position of the column.
Step 4: step3 is repeated until all default boxes in the row have been executed.
Step 5: the traversal of the 8732 x 21 matrix is performed, i.e. all classes of NMS are performed.
Step 6: and further residual screening is carried out, and all the final residual categories are selected according to the confidence coefficient.
And designing a feature capture box according to the coordinate points of each feature map to extract features, and using the features to predict the type and the bounding box of the target. Here, the convolution kernel of 3 × 3 is used to extract the features in each feature capture box, the convolution box used for each feature map is 3 × 6 (class +4), 6 is the number of capture boxes at each feature map coordinate point, and 4 is the deviation between the predicted target boundary and the target boundary box. If a feature map is m x n in size, with 6 boxes per coordinate, then an output of m x n 6 x (class +4) is ultimately produced:
and (4) calculating a loss function, wherein the loss function of the whole model is as follows:
where x is used to determine whether the designed feature capture box has a corresponding target,indicating whether the ith box is matched with the jth target frame of the pth object, wherein the matching is 1, and otherwise, the matching is 0. If it isIndicating that at least one box matches it for the jth target bounding box. Where N represents the number of matching sums. The first part of equation (5) is used to measure the performance of the recognition, and is mainly a class of softmax loss function, detail
WhereinIn equation (6), the second partial equation is used to measure the predicted performance of the bounding box, and the loss function used is as follows:
whereinAnd (2) representing the deviation between the real target frame of the jth target and the frame of the characteristic grabbing box, wherein m belongs to { cx, cy, w, h }, wherein (cx, cy) represents the coordinate of the center point of the frame, and (w, h) represents the width and the height of the frame. The position information of the last frame can be expressed by the following formula (7):
meanwhile, the integrated confidence loss function of the plurality of regions may be expressed as:
implementation 5: as shown in fig. 5, for the case of multi-angle recognition of underwater environment, 200 underwater photos of different forms of fishes and crabs are collected through internet and experimental pool experiments. Training the network, and performing experiments in the aspects of loss functions, recognition accuracy and positioning accuracy IOU. Here, the Conf of the target is set to 0.5 or more, and recursive detection is performed on exclusion in the image region center 1/2 region, thereby performing detection and recognition for complex reality such as dense fish distribution and target crossing. The target fishes under different postures and different positions can be accurately identified, the conditions of far and near multi-angle and complicated parallel distribution of multiple fishes are accurately identified, multi-target identification can be realized, and the accurate identification can be realized for some partially shielded fishes. 20000 iterations and calculation of an error curve based on the loss function are shown in fig. 6, where the initial error is defined as 500, i.e. the convergence range is (0,500), and the final convergence is around 20, and the error rate is less than about 1%. In the positioning accuracy map, it is found that the positioning accuracy converges in an oscillation mode and runs through the whole training process, which is caused by the fuzzy and virtual images of underwater imaging. Finally, the method can well overcome the problem of false recognition caused by low recognition accuracy due to the existing underwater imaging blur.
Claims (6)
1. An underwater target identification method based on machine learning is characterized by comprising the following steps:
step1, preprocessing an underwater target image, and performing retinex restoration on the image to obtain an input image;
step2, carrying out convolution by adopting convolution boxes with different sizes and scales among different layers to obtain feature maps with different scales;
step3, performing regression according to the feature maps obtained in the step two, performing normalization processing on each feature map, and obtaining defuelt boxes with different scales and offset through detectors and classifiers with different sizes;
and 4, obtaining a final result through an NMS (network management system) inhibition algorithm.
2. The machine learning-based underwater target recognition method according to claim 1, characterized in that: in the step2, a region result obtained by RPN calculation in a region selection algorithm is adopted, the feature maps with different sizes are obtained by dividing according to different scales, then the RPN convolution core is used for moving the region on the feature map and obtaining the confidence value of the region, and a matrix with the confidence is obtained by continuously moving defualt boxes with different scales.
3. The machine learning-based underwater target recognition method according to claim 1, characterized in that: the width and height of the defuelt box in the step3 are as follows:
whereinAndwidth and height of defualt box, respectively, the scale being ar∈{1,2,1/2,3,1/3},Is the scale of different layers, wherein m is the number, the default scale is realized by m different scale characteristic diagram predictions, SminIs of the smallest scale, SmaxIs the scale of the largest feature map.
4. The machine learning-based underwater target recognition method according to claim 1, characterized in that: the central position of the selection frame in the step3 is as follows:
wherein x and y represent the horizontal frame, the central point of the longitudinal axis, | fkI represents the k-th feature graph scale, i.j E [0, | fk|]。
5. The machine learning-based underwater target recognition method according to claim 1, characterized in that: the method for NMS inhibition in step4 comprises
Step 4.1: arranging elements in the matrix according to the conf size;
step 4.2: IoU calculation is carried out on all the cross areas according to the sequence from large to small of the calculation result in the step 4.1, Th values are set, comparison is carried out on the Th values according to the sequence and IoU, and classification are carried out according to the sizes of the Th values;
step 4.3: the execution returns to the step 4.2 again at the second big frame position of the column;
step 4.4: repeating the step 4.3 until all the defuelt boxes in the row are executed;
step 4.5: performing traversal on the matrix, namely performing NMS of all categories;
step 4.6: and further residual screening is carried out, and all the final residual categories are selected according to the confidence coefficient.
6. The machine learning-based underwater target recognition method according to claim 1, characterized in that: the loss function of the whole model in the step4 is
Where x is used to determine whether the designed feature capture box has a corresponding target,whether the ith box is matched with the jth target frame of the pth object or not is represented, the matching is 1, and otherwise, the matching is 0; if it isIndicating that at least one box matches the jth target bounding box; n represents the number of matching sums;to measure the performance of the recognition;used to measure bounding box prediction performance; wherein And (2) representing the deviation between the real target frame of the jth target and the frame of the characteristic grabbing box, wherein m belongs to { cx, cy, w, h }, wherein (cx, cy) represents the coordinate of the center point of the frame, and (w, h) represents the width and the height of the frame.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910950105.6A CN110826575A (en) | 2019-12-13 | 2019-12-13 | Underwater target identification method based on machine learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910950105.6A CN110826575A (en) | 2019-12-13 | 2019-12-13 | Underwater target identification method based on machine learning |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110826575A true CN110826575A (en) | 2020-02-21 |
Family
ID=69548686
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910950105.6A Pending CN110826575A (en) | 2019-12-13 | 2019-12-13 | Underwater target identification method based on machine learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110826575A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111666822A (en) * | 2020-05-13 | 2020-09-15 | 飒铂智能科技有限责任公司 | Low-altitude unmanned aerial vehicle target detection method and system based on deep learning |
CN112906458A (en) * | 2021-01-08 | 2021-06-04 | 浙江大学 | Swarm intelligent optimized underwater laser multi-target end-to-end automatic identification system |
CN113313116A (en) * | 2021-06-20 | 2021-08-27 | 西北工业大学 | Vision-based accurate detection and positioning method for underwater artificial target |
CN114882346A (en) * | 2021-01-22 | 2022-08-09 | 中国科学院沈阳自动化研究所 | Underwater robot target autonomous identification method based on vision |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107423760A (en) * | 2017-07-21 | 2017-12-01 | 西安电子科技大学 | Based on pre-segmentation and the deep learning object detection method returned |
CN110458160A (en) * | 2019-07-09 | 2019-11-15 | 北京理工大学 | A kind of unmanned boat waterborne target recognizer based on depth-compression neural network |
-
2019
- 2019-12-13 CN CN201910950105.6A patent/CN110826575A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107423760A (en) * | 2017-07-21 | 2017-12-01 | 西安电子科技大学 | Based on pre-segmentation and the deep learning object detection method returned |
CN110458160A (en) * | 2019-07-09 | 2019-11-15 | 北京理工大学 | A kind of unmanned boat waterborne target recognizer based on depth-compression neural network |
Non-Patent Citations (3)
Title |
---|
WEI LIU 等: ""SSD: Single Shot MultiBox Detector"", 《COMPUTER VISION-ECCV 2016》 * |
YTUSDC: ""SSD算法详解"", 《HTTPS://BLOG.CSDN.NET/YTUSDC/ARTICLE/DETAILS/86577939》 * |
徐岩等: "基于卷积神经网络的水下图像增强方法", 《吉林大学学报(工学版)》 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111666822A (en) * | 2020-05-13 | 2020-09-15 | 飒铂智能科技有限责任公司 | Low-altitude unmanned aerial vehicle target detection method and system based on deep learning |
CN112906458A (en) * | 2021-01-08 | 2021-06-04 | 浙江大学 | Swarm intelligent optimized underwater laser multi-target end-to-end automatic identification system |
CN112906458B (en) * | 2021-01-08 | 2022-07-05 | 浙江大学 | Group intelligent optimized underwater laser multi-target end-to-end automatic identification system |
CN114882346A (en) * | 2021-01-22 | 2022-08-09 | 中国科学院沈阳自动化研究所 | Underwater robot target autonomous identification method based on vision |
CN114882346B (en) * | 2021-01-22 | 2024-07-09 | 中国科学院沈阳自动化研究所 | Underwater robot target autonomous identification method based on vision |
CN113313116A (en) * | 2021-06-20 | 2021-08-27 | 西北工业大学 | Vision-based accurate detection and positioning method for underwater artificial target |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Yang et al. | Real-time face detection based on YOLO | |
CN110059558B (en) | Orchard obstacle real-time detection method based on improved SSD network | |
CN109800689B (en) | Target tracking method based on space-time feature fusion learning | |
CN110147743B (en) | Real-time online pedestrian analysis and counting system and method under complex scene | |
CN110232350B (en) | Real-time water surface multi-moving-object detection and tracking method based on online learning | |
CN110135243B (en) | Pedestrian detection method and system based on two-stage attention mechanism | |
CN106127204B (en) | A kind of multi-direction meter reading Region detection algorithms of full convolutional neural networks | |
CN107067415B (en) | A kind of object localization method based on images match | |
CN110826575A (en) | Underwater target identification method based on machine learning | |
CN108805906A (en) | A kind of moving obstacle detection and localization method based on depth map | |
CN107481315A (en) | A kind of monocular vision three-dimensional environment method for reconstructing based on Harris SIFT BRIEF algorithms | |
CN110991274B (en) | Pedestrian tumbling detection method based on Gaussian mixture model and neural network | |
CN111753682B (en) | Hoisting area dynamic monitoring method based on target detection algorithm | |
CN112149591B (en) | SSD-AEFF automatic bridge detection method and system for SAR image | |
CN109242019B (en) | Rapid detection and tracking method for optical small target on water surface | |
CN113743385A (en) | Unmanned ship water surface target detection method and device and unmanned ship | |
CN113888461A (en) | Method, system and equipment for detecting defects of hardware parts based on deep learning | |
CN112883850A (en) | Multi-view aerospace remote sensing image matching method based on convolutional neural network | |
CN111998862A (en) | Dense binocular SLAM method based on BNN | |
Han et al. | Research on remote sensing image target recognition based on deep convolution neural network | |
Zhang et al. | Nearshore vessel detection based on Scene-mask R-CNN in remote sensing image | |
CN117197676A (en) | Target detection and identification method based on feature fusion | |
CN113989672B (en) | SAR image ship detection method based on balance learning | |
Cheng et al. | C 2-YOLO: Rotating Object Detection Network for Remote Sensing Images with Complex Backgrounds | |
Zhang et al. | Lidar odometry and mapping based on two-stage feature extraction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200221 |