CN111178267A - Video behavior identification method for monitoring illegal fishing - Google Patents

Video behavior identification method for monitoring illegal fishing Download PDF

Info

Publication number
CN111178267A
CN111178267A CN201911395639.3A CN201911395639A CN111178267A CN 111178267 A CN111178267 A CN 111178267A CN 201911395639 A CN201911395639 A CN 201911395639A CN 111178267 A CN111178267 A CN 111178267A
Authority
CN
China
Prior art keywords
default
frame
picture
video
frames
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911395639.3A
Other languages
Chinese (zh)
Inventor
不公告发明人
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Shuzhilian Technology Co Ltd
Original Assignee
Chengdu Shuzhilian Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Shuzhilian Technology Co Ltd filed Critical Chengdu Shuzhilian Technology Co Ltd
Priority to CN201911395639.3A priority Critical patent/CN111178267A/en
Publication of CN111178267A publication Critical patent/CN111178267A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a video behavior identification method for monitoring illegal fishing, which comprises the steps of collecting video data of a monitored water area, cutting the video data to obtain a multi-frame picture group and generating a real label frame; inputting the multi-frame picture group into a convolutional neural network, extracting picture characteristics and generating a characteristic picture; extracting a feature map of a 6-layer SSD network, and then generating a default frame at each point of the feature map; matching each default frame with a real tag frame; gathering the matched default frames, inhibiting and screening through a non-maximum value, and outputting the screened default frames to obtain a target positioning picture with a final target area; and inputting the target positioning picture into the behavior recognition model, and judging whether the target positioning picture is an illegal fishing behavior. The method can effectively remove the video frames irrelevant to behavior recognition in the video under the complex environment to obtain the specific position of the target in the video, so as to ensure the accuracy of a video behavior recognition processing area and further improve the accuracy and the real-time performance of behavior recognition.

Description

Video behavior identification method for monitoring illegal fishing
Technical Field
The invention belongs to the technical field of video behavior recognition, and particularly relates to a video behavior recognition method for monitoring illegal fishing.
Background
Illegal fishing is a strict practice in China at present, but still many people can illegally fish by virtue of their own interests, and the ecological environment is seriously damaged. When illegal fishing is performed in rivers and seas, video monitoring can be adopted at present, but due to the limitation of environmental conditions, the detection target is usually small, and the influence of the background environment is very large; it is still very difficult to identify smaller fishing targets in complex and large background environments.
At present, video behavior recognition generally classifies pre-segmented short videos, and videos in a real environment are generally not pre-segmented and contain a large amount of irrelevant information. In practical application, however, the complex environmental information has a great influence on behavior recognition, and the accuracy of behavior recognition is greatly reduced.
Disclosure of Invention
In order to solve the problems, the invention provides a video behavior recognition method for monitoring illegal fishing, which can effectively remove video frames irrelevant to behavior recognition in a video under a complex environment to obtain the specific position of a target in the video, so as to ensure the accuracy of a video behavior recognition processing area and further improve the accuracy and the real-time performance of behavior recognition.
In order to achieve the purpose, the invention adopts the technical scheme that: a video behavior recognition method for monitoring illegal fishing detects the specific position of an illegal fishing target in a video and carries out behavior recognition on the target, and comprises the following steps:
collecting video data of a monitored water area, cutting the video data to obtain a multi-frame picture group, marking a target in the picture group, and generating a real label frame;
inputting the multi-frame picture group into a convolutional neural network, extracting picture characteristics and generating a characteristic picture;
generating a default frame, extracting a feature map of a 6-layer SSD network, and then generating the default frame at each point of the feature map, wherein the number of the default frames at each layer is different, but each point has the default frame;
matching default frames, and matching each default frame with a real label frame;
screening the default frame, gathering the matched default frame, inhibiting screening through a non-maximum value, and outputting the screened default frame to obtain a target positioning picture with a final target area;
and (4) behavior identification, namely inputting the target positioning picture into a behavior identification model and judging whether the target positioning picture is an illegal fishing behavior.
Further, in the process of matching the default boxes, each default box is matched with a real tag box, each default box selects the real tag box corresponding to the largest IOU value, one default box can only correspond to one real tag box, and one real tag box corresponds to a plurality of default boxes.
Further, in the process after the default box is matched;
if the ith default box is matched to the jth real tag box, calculating the softmax loss value of the default box belonging to the background category, the softmax loss function is as follows:
Figure BDA0002346223230000021
wherein, ViIs the output of the preceding stage output unit of the classifier, i represents the category index, and C is the total number of categories; siThe ratio of the index of the current element to the sum of the indexes of all elements, namely the confidence corresponding to the detected default frame;
and obtaining the confidence degrees corresponding to all the detected default frames to prepare for screening the default frames.
Further, screening all generated default frames by using a maximum suppression method to obtain a final target area;
the non-maximum value inhibition algorithm firstly removes the frame with the confidence coefficient lower than a preset threshold value, selects the frame with the maximum confidence coefficient from the rest frames, then compares the intersection ratio IOU of the other frames and the frame, inhibits the frame which is not the local maximum value if the intersection ratio IOU is larger than the preset threshold value, then selects the frame with the maximum confidence coefficient from the rest frames, repeats the steps until no default frame is inhibited, and finally obtains the frame which is the result frame;
and determining the position of the target in the picture according to the obtained target area to obtain a target positioning picture.
Further, all the generated default frames are screened by using a maximum suppression method, and a final calculation formula of the target area is obtained as follows:
Figure BDA0002346223230000022
wherein N istIs a suppression threshold of the maximum value, biIs the default box of detection, SiIs biConfidence corresponding to the box, M is the union of all box detections.
Further, the target positioning picture is input into the behavior recognition model, and whether the behavior recognition model is an illegal fishing behavior or not is judged, and the method comprises the following steps:
sequentially sequencing the pictures obtained by target positioning according to a time sequence, continuously extracting 16 pictures as a picture group, and overlapping 8 continuous pictures between two adjacent picture groups;
inputting the picture groups into a behavior recognition model, and extracting picture characteristics of each picture group through a convolution layer of a network to generate a characteristic graph;
connecting all the characteristics of the full-connection layer of the model, and sending an output value to a softmax classifier;
and outputting the category of the target behavior through the softmax layer of the model.
Further, the behavior recognition model comprises 8 convolutional layers, 5 pooling layers and 3 full-connection layers, and the model is trained by utilizing time information;
the first layer of convolutional layers uses 1 x 3 convolutional kernels, steps 0 x 1, and pooling layers with kernel size 1 x 2, steps 1 x 2; the second layer of convolution layers uses convolution kernels of 3 x 3, steps 1 x 1, and pooling layers with kernel size of 1 x 2, steps 1 x 2 are performed; the rest of the convolution layers use 3 × 3 convolution kernels with a step size of 1 × 1, and pooling layers with kernel sizes of 2 × 2 and a step size of 2 × 2 are performed; the seventh layer and the eighth layer are full connection layers, and the ninth layer is a softmax layer, so that a final classification result is obtained.
Further, in order to maintain early time information settings, 3D pooling layers are employed in the behavior recognition model, the kernel sizes of the first and second pooling layers being 1 x 2, step size 1 x 2; all the other 3D pooling layers were 2 x 2 with step size of 2 x 2; the first number of pooling layers is time depth, pooling over individual frames when time depth is 1, pooling between multiple frames when time depth is greater than 1, the former being advantageous to preserve the time characteristics of the initial stage.
The beneficial effects of the technical scheme are as follows:
firstly, positioning a target in a video to obtain a specific position of the target in the video; and then analyzing the position area by using a behavior recognition algorithm, and finally judging the behavior of the target. The efficiency of video behavior recognition is greatly improved, behavior recognition is not performed on invalid targets and invalid areas in the video, and meanwhile, behavior recognition is performed on specific area positions of the targets under the condition that the valid targets exist, so that the influence of complex backgrounds on behavior recognition is reduced, and the accuracy and the real-time performance of behavior recognition are greatly improved.
The method comprises the steps of cutting a video into a frame and a frame, inputting the frame and the frame into a network, obtaining target areas through model detection, then cutting the image areas, inputting the cut image into a rear-end behavior recognition model of the model, training the network, and finishing behavior recognition; the background of the behavior recognition area is purified through the front-end processing of the model, and the improvement of recognition accuracy is facilitated.
Drawings
Fig. 1 is a flow chart of a video behavior recognition method for monitoring illegal fishing according to the present invention.
Fig. 2 is a schematic structural diagram of a model used in the embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described with reference to the accompanying drawings.
In this embodiment, referring to fig. 1, the present invention provides a video behavior recognition method for monitoring illegal fishing, which detects specific positions of illegal fishing targets in a video and performs behavior recognition on the targets, and includes the steps of:
collecting video data of a monitored water area, cutting the video data to obtain a multi-frame picture group, marking a target in the picture group, and generating a real label frame;
inputting the multi-frame picture group into a convolutional neural network, extracting picture characteristics and generating a characteristic picture;
generating a default frame, extracting a feature map of a 6-layer SSD network, and then generating the default frame at each point of the feature map, wherein the number of the default frames at each layer is different, but each point has the default frame;
matching default frames, and matching each default frame with a real label frame;
screening the default frame, gathering the matched default frame, inhibiting screening through a non-maximum value, and outputting the screened default frame to obtain a target positioning picture with a final target area;
and (4) behavior identification, namely inputting the target positioning picture into a behavior identification model and judging whether the target positioning picture is an illegal fishing behavior.
As an optimization scheme of the above embodiment, in the process of matching default frames, each default frame is matched with a real tag frame, each default frame is selected to correspond to the real tag frame with the largest IOU value, one default frame can only correspond to one real tag frame, and one real tag frame corresponds to a plurality of default frames.
In the process after matching the default box;
if the ith default box is matched to the jth real tag box, calculating the softmax loss value of the default box belonging to the background category, the softmax loss function is as follows:
Figure BDA0002346223230000041
wherein, ViIs the output of the preceding stage output unit of the classifier, i represents the category index, and C is the total number of categories; siThe ratio of the index of the current element to the sum of the indexes of all elements, namely the confidence corresponding to the detected default frame;
and obtaining the confidence degrees corresponding to all the detected default frames to prepare for screening the default frames.
As an optimization scheme of the above embodiment, all generated default frames are screened by using a maximum suppression method to obtain a final target area;
the non-maximum value inhibition algorithm firstly removes the frame with the confidence coefficient lower than a preset threshold value, selects the frame with the maximum confidence coefficient from the rest frames, then compares the intersection ratio IOU of the other frames and the frame, inhibits the frame which is not the local maximum value if the intersection ratio IOU is larger than the preset threshold value, then selects the frame with the maximum confidence coefficient from the rest frames, repeats the steps until no default frame is inhibited, and finally obtains the frame which is the result frame;
and determining the position of the target in the picture according to the obtained target area to obtain a target positioning picture.
Screening all the generated default frames by using a maximum suppression method to obtain a final target area by the following calculation formula:
Figure BDA0002346223230000051
wherein N istIs a suppression threshold of the maximum value, biIs the default box of detection, SiIs biConfidence corresponding to the box, M is the union of all box detections.
As the optimization scheme of the embodiment, the method for inputting the target positioning picture into the behavior recognition model to judge whether the target positioning picture is an illegal fishing behavior comprises the following steps:
sequentially sequencing the pictures obtained by target positioning according to a time sequence, continuously extracting 16 pictures as a picture group, and overlapping 8 continuous pictures between two adjacent picture groups;
inputting the picture groups into a behavior recognition model, and extracting picture characteristics of each picture group through a convolution layer of a network to generate a characteristic graph;
connecting all the characteristics of the full-connection layer of the model, and sending an output value to a softmax classifier;
and outputting the category of the target behavior through the softmax layer of the model.
The behavior recognition model comprises 8 convolutional layers, 5 pooling layers and 3 full-connection layers, and is trained by utilizing time information;
the first layer of convolutional layers uses 1 x 3 convolutional kernels, steps 0 x 1, and pooling layers with kernel size 1 x 2, steps 1 x 2; the second layer of convolution layers uses convolution kernels of 3 x 3, steps 1 x 1, and pooling layers with kernel size of 1 x 2, steps 1 x 2 are performed; the rest of the convolution layers use 3 × 3 convolution kernels with a step size of 1 × 1, and pooling layers with kernel sizes of 2 × 2 and a step size of 2 × 2 are performed; the seventh layer and the eighth layer are full connection layers, and the ninth layer is a softmax layer, so that a final classification result is obtained.
In order to maintain early temporal information settings, 3D pooling layers were employed in the behavior recognition model, the kernel sizes of the first and second pooling layers being 1 x 2, step 1 x 2; all the other 3D pooling layers were 2 x 2 with step size of 2 x 2; the first number of pooling layers is time depth, pooling over individual frames when time depth is 1, pooling between multiple frames when time depth is greater than 1, the former being advantageous to preserve the time characteristics of the initial stage.
The following illustrates an exemplary video behavior recognition method for monitoring illegal fishing, which is used by the present invention, to detect the specific position of the illegal fishing target in the video and perform behavior recognition on the target. The network in fig. 2 is based on a convolutional network, which generates a fixed size bounding box set and object class scores in the box, then uses a non-maximization suppression step to generate final detection, and then performs behavior recognition on the detected target region, and the specific steps are as follows:
1. inputting 300 × 3 pictures into a network, extracting picture features based on a VGG-16 network, and generating a feature map; extracting feature maps of 6 convolutional layers in the network, and generating default frames on each point of the feature maps, wherein the number of the default frames on each layer is different, but each point has the default frame; where the convolutional layer 6 has 38 × 4 — 5776 default frames, the convolutional layer 7 has 19 × 6 — 2166 default frames, the convolutional layer 8 has 10 × 6 — 600 default frames, the convolutional layer 9 has 5 × 6 — 150 default frames, the convolutional layer 10 has 3 × 4 — 36 default frames, the convolutional layer 11 has 1 × 4 — 4 default frames, and there are 8732 default frames in total, and then we send these default frames to the NMS (maximum suppression) module to obtain the final detection result, i.e. the specific position of the target (person) in the video.
2. Adjusting the size of the detected picture obtained by target positioning, wherein the size is changed from 300 × 3 to 224 × 3, and the input picture is a three-channel picture with 224 × 3 pixels; the first layer uses 1 x 3 convolution kernels with step size 0 x 1 and pooling layers with kernel size 1 x 2 and step size 1 x 2; the second layer uses a convolution kernel of 3 x 3 with a step size of 1 x 1 and pooled layers with kernel size of 1 x 2 with a step size of 1 x 2; the remaining layers used 3 x 3 convolution kernels with step size 1 x 1 and pooled layers with kernel size 2 x 2; the sixth layer is a full connection layer and comprises 8192 hidden layers and 4096 hidden layers respectively, namely only 4096 characteristic values are left by the full connection layer of the seventh layer; and finally, the eighth layer is a full connection layer, and a final classification result is obtained.
The foregoing shows and describes the general principles and broad features of the present invention and advantages thereof. It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are described in the specification and illustrated only to illustrate the principle of the present invention, but that various changes and modifications may be made therein without departing from the spirit and scope of the present invention, which fall within the scope of the invention as claimed. The scope of the invention is defined by the appended claims and equivalents thereof.

Claims (8)

1. A video behavior identification method for monitoring illegal fishing is characterized in that the specific position of an illegal fishing target in a video is detected and behavior identification is carried out on the target, and the method comprises the following steps:
collecting video data of a monitored water area, cutting the video data to obtain a multi-frame picture group, marking a target in the picture group, and generating a real label frame;
inputting the multi-frame picture group into a convolutional neural network, extracting picture characteristics and generating a characteristic picture;
generating a default frame, extracting a feature map of a 6-layer SSD network, and then generating the default frame at each point of the feature map, wherein the number of the default frames at each layer is different, but each point has the default frame;
matching default frames, and matching each default frame with a real label frame;
screening the default frame, gathering the matched default frame, inhibiting screening through a non-maximum value, and outputting the screened default frame to obtain a target positioning picture with a final target area;
and (4) behavior identification, namely inputting the target positioning picture into a behavior identification model and judging whether the target positioning picture is an illegal fishing behavior.
2. The video behavior recognition method for monitoring illegal fishing according to claim 1, wherein in the process of matching default boxes, each default box is matched with a real tag box, each default box is selected to correspond to the real tag box with the largest IOU value, one default box can only correspond to one real tag box, and one real tag box corresponds to a plurality of default boxes.
3. The video behavior recognition method for monitoring illegal fishing according to claim 2, wherein in the process after matching the default frame;
if the ith default box is matched to the jth real tag box, calculating the softmax loss value of the default box belonging to the background category, the softmax loss function is as follows:
Figure FDA0002346223220000011
wherein, ViIs the output of the preceding stage output unit of the classifier, i represents the category index, and C is the total number of categories; siThe ratio of the index of the current element to the sum of the indexes of all elements, namely the confidence corresponding to the detected default frame;
and obtaining the confidence degrees corresponding to all the detected default frames to prepare for screening the default frames.
4. The video behavior recognition method for monitoring illegal fishing according to claim 3, wherein all the generated default frames are screened by using a maximum suppression method to obtain a final target area;
the non-maximum value inhibition algorithm firstly removes the frame with the confidence coefficient lower than a preset threshold value, selects the frame with the maximum confidence coefficient from the rest frames, then compares the intersection ratio IOU of the other frames and the frame, inhibits the frame which is not the local maximum value if the intersection ratio IOU is larger than the preset threshold value, then selects the frame with the maximum confidence coefficient from the rest frames, repeats the steps until no default frame is inhibited, and finally obtains the frame which is the result frame;
and determining the position of the target in the picture according to the obtained target area to obtain a target positioning picture.
5. The video behavior recognition method for monitoring illegal fishing according to claim 4, wherein the maximum suppression method is used to screen all the generated default frames to obtain the final target area according to the formula:
Figure FDA0002346223220000021
wherein N istIs a suppression threshold of the maximum value, biIs the default box of detection, SiIs biConfidence corresponding to the box, M is the union of all box detections.
6. The video behavior recognition method for monitoring illegal fishing according to claim 1, wherein the target positioning picture is inputted into the behavior recognition model to determine whether it is illegal fishing behavior, comprising the steps of:
sequentially sequencing the pictures obtained by target positioning according to a time sequence, continuously extracting 16 pictures as a picture group, and overlapping 8 continuous pictures between two adjacent picture groups;
inputting the picture groups into a behavior recognition model, and extracting picture characteristics of each picture group through a convolution layer of a network to generate a characteristic graph;
connecting all the characteristics of the full-connection layer of the model, and sending an output value to a softmax classifier;
and outputting the category of the target behavior through the softmax layer of the model.
7. The video behavior recognition method for monitoring illegal fishing according to claim 6, wherein the behavior recognition model comprises 8 convolutional layers, 5 pooling layers and 3 full-link layers, and the model is trained by using time information;
the first layer of convolutional layers uses 1 x 3 convolutional kernels, steps 0 x 1, and pooling layers with kernel size 1 x 2, steps 1 x 2; the second layer of convolution layers uses convolution kernels of 3 x 3, steps 1 x 1, and pooling layers with kernel size of 1 x 2, steps 1 x 2 are performed; the rest of the convolution layers use 3 × 3 convolution kernels with a step size of 1 × 1, and pooling layers with kernel sizes of 2 × 2 and a step size of 2 × 2 are performed; the seventh layer and the eighth layer are full connection layers, and the ninth layer is a softmax layer, so that a final classification result is obtained.
8. The video behavior recognition method for monitoring illegal fishing according to claim 7, characterized in that in order to maintain the early time information setting, 3D pooling layers are used in the behavior recognition model, the kernel size of the first and second pooling layers is 1 x 2, step 1 x 2; all the other 3D pooling layers were 2 x 2 with step size of 2 x 2; the first number of pooling layers is time depth, pooling over individual frames when time depth is 1, pooling between multiple frames when time depth is greater than 1, the former being advantageous to preserve the time characteristics of the initial stage.
CN201911395639.3A 2019-12-30 2019-12-30 Video behavior identification method for monitoring illegal fishing Pending CN111178267A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911395639.3A CN111178267A (en) 2019-12-30 2019-12-30 Video behavior identification method for monitoring illegal fishing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911395639.3A CN111178267A (en) 2019-12-30 2019-12-30 Video behavior identification method for monitoring illegal fishing

Publications (1)

Publication Number Publication Date
CN111178267A true CN111178267A (en) 2020-05-19

Family

ID=70652262

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911395639.3A Pending CN111178267A (en) 2019-12-30 2019-12-30 Video behavior identification method for monitoring illegal fishing

Country Status (1)

Country Link
CN (1) CN111178267A (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030117652A1 (en) * 1999-09-17 2003-06-26 Paul Lapstun Rotationally symmetric tags
CN107301376A (en) * 2017-05-26 2017-10-27 浙江大学 A kind of pedestrian detection method stimulated based on deep learning multilayer
CN107679469A (en) * 2017-09-22 2018-02-09 东南大学—无锡集成电路技术研究所 A kind of non-maxima suppression method based on deep learning
CN109147254A (en) * 2018-07-18 2019-01-04 武汉大学 A kind of video outdoor fire disaster smog real-time detection method based on convolutional neural networks
CN109409283A (en) * 2018-10-24 2019-03-01 深圳市锦润防务科技有限公司 A kind of method, system and the storage medium of surface vessel tracking and monitoring
CN109740463A (en) * 2018-12-21 2019-05-10 沈阳建筑大学 A kind of object detection method under vehicle environment
CN110232319A (en) * 2019-05-07 2019-09-13 杭州电子科技大学 A kind of ship Activity recognition method based on deep learning
CN110309702A (en) * 2019-04-18 2019-10-08 成都数之联科技有限公司 A kind of shops's counter video surveillance management system
CN110334574A (en) * 2019-04-26 2019-10-15 武汉理工大学 A method of automatically extracting traffic accident key frame in traffic video
CN110516605A (en) * 2019-08-28 2019-11-29 北京观微科技有限公司 Any direction Ship Target Detection method based on cascade neural network

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030117652A1 (en) * 1999-09-17 2003-06-26 Paul Lapstun Rotationally symmetric tags
CN107301376A (en) * 2017-05-26 2017-10-27 浙江大学 A kind of pedestrian detection method stimulated based on deep learning multilayer
CN107679469A (en) * 2017-09-22 2018-02-09 东南大学—无锡集成电路技术研究所 A kind of non-maxima suppression method based on deep learning
CN109147254A (en) * 2018-07-18 2019-01-04 武汉大学 A kind of video outdoor fire disaster smog real-time detection method based on convolutional neural networks
CN109409283A (en) * 2018-10-24 2019-03-01 深圳市锦润防务科技有限公司 A kind of method, system and the storage medium of surface vessel tracking and monitoring
CN109740463A (en) * 2018-12-21 2019-05-10 沈阳建筑大学 A kind of object detection method under vehicle environment
CN110309702A (en) * 2019-04-18 2019-10-08 成都数之联科技有限公司 A kind of shops's counter video surveillance management system
CN110334574A (en) * 2019-04-26 2019-10-15 武汉理工大学 A method of automatically extracting traffic accident key frame in traffic video
CN110232319A (en) * 2019-05-07 2019-09-13 杭州电子科技大学 A kind of ship Activity recognition method based on deep learning
CN110516605A (en) * 2019-08-28 2019-11-29 北京观微科技有限公司 Any direction Ship Target Detection method based on cascade neural network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
宁煜西等: "基于卷积神经网络的航班跟踪视频关键信息识别", 《空军预警学院学报》 *

Similar Documents

Publication Publication Date Title
CN109766830B (en) Ship target identification system and method based on artificial intelligence image processing
CN110119728B (en) Remote sensing image cloud detection method based on multi-scale fusion semantic segmentation network
CN107016357B (en) Video pedestrian detection method based on time domain convolutional neural network
CN111797712B (en) Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network
CN109145872B (en) CFAR and Fast-RCNN fusion-based SAR image ship target detection method
CN108491854B (en) Optical remote sensing image target detection method based on SF-RCNN
CN112598713A (en) Offshore submarine fish detection and tracking statistical method based on deep learning
CN112733614B (en) Pest image detection method with similar size enhanced identification
CN110334703B (en) Ship detection and identification method in day and night image
CN111160407A (en) Deep learning target detection method and system
CN109766823A (en) A kind of high-definition remote sensing ship detecting method based on deep layer convolutional neural networks
CN111046827A (en) Video smoke detection method based on convolutional neural network
CN112149591A (en) SSD-AEFF automatic bridge detection method and system for SAR image
CN114140665A (en) Dense small target detection method based on improved YOLOv5
CN108734200A (en) Human body target visible detection method and device based on BING features
CN111950357A (en) Marine water surface garbage rapid identification method based on multi-feature YOLOV3
Mu et al. Salient object detection in low contrast images via global convolution and boundary refinement
CN115700737A (en) Oil spill detection method based on video monitoring
CN117475353A (en) Video-based abnormal smoke identification method and system
Yang et al. SAR image target detection and recognition based on deep network
CN117557937A (en) Monitoring camera image anomaly detection method and system
CN110349119B (en) Pavement disease detection method and device based on edge detection neural network
CN108985216B (en) Pedestrian head detection method based on multivariate logistic regression feature fusion
CN114037737B (en) Neural network-based offshore submarine fish detection and tracking statistical method
CN107341456B (en) Weather sunny and cloudy classification method based on single outdoor color image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 610000 No. 270, floor 2, No. 8, Jinxiu street, Wuhou District, Chengdu, Sichuan

Applicant after: Chengdu shuzhilian Technology Co.,Ltd.

Address before: No.2, 4th floor, building 1, Jule road crossing, Section 1, West 1st ring road, Chengdu, Sichuan 610000

Applicant before: CHENGDU SHUZHILIAN TECHNOLOGY Co.,Ltd.