WO2022241803A1 - Attention mechanism-based system and method for detecting feature in target, and storage medium - Google Patents

Attention mechanism-based system and method for detecting feature in target, and storage medium Download PDF

Info

Publication number
WO2022241803A1
WO2022241803A1 PCT/CN2021/095956 CN2021095956W WO2022241803A1 WO 2022241803 A1 WO2022241803 A1 WO 2022241803A1 CN 2021095956 W CN2021095956 W CN 2021095956W WO 2022241803 A1 WO2022241803 A1 WO 2022241803A1
Authority
WO
WIPO (PCT)
Prior art keywords
detection
attention
module
target
image
Prior art date
Application number
PCT/CN2021/095956
Other languages
French (fr)
Chinese (zh)
Inventor
黄宇恒
魏东
岳许要
金晓峰
徐天适
Original Assignee
广州广电运通金融电子股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 广州广电运通金融电子股份有限公司 filed Critical 广州广电运通金融电子股份有限公司
Publication of WO2022241803A1 publication Critical patent/WO2022241803A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • G06F18/24137Distances to cluster centroïds
    • G06F18/2414Smoothing the distance, e.g. radial basis function networks [RBFN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • the invention relates to the field of intelligent security, in particular to an attention mechanism-based feature detection system, method and storage medium in an object.
  • Object detection is one of the research hotspots in machine vision
  • object feature detection refers to the technology of locating the components in the object to further analyze the structured information of the object in the video and image.
  • In-target feature detection is one of the important technologies for video/image structured analysis. For example, in the vehicle structured task, it is necessary to locate the features of the vehicle target such as the face, windows, lights, logos, and luggage racks, and for further analysis.
  • the feature detection in the target mainly relies on general scene detection methods such as SSD and YOLO, and its shortcomings are as follows:
  • In-target feature detection usually locates multiple feature locations from within a fixed type of target.
  • the existing general detection framework does not consider the guidance of target attributes for in-target feature detection.
  • the existing detection framework uses the same feature map to perform detection frame regression and classification on the target, which is not conducive to performance improvement.
  • the anchor box selection method of the existing detection framework will generate a large number of negative samples, resulting in the problem of unbalanced samples in the training phase and difficult classification.
  • the object of the present invention is to provide a system, method and storage medium for feature detection in an object based on attention mechanism, which can solve the above problems.
  • a feature detection system based on an attention mechanism includes a semantic extraction module, an attention map module, and a detection module, wherein: the semantic extraction module includes a multi-layer deep convolutional network, which is responsible for extracting high-level semantics from input images information, and share the extracted high-level semantic information to the attention map module and the detection module; the attention map module includes a classification sub-module and an attention sub-module; wherein, each attribute branch of the classification sub-module includes multiple Convolution layer, global pooling layer, global connection layer and softmax layer are responsible for classifying the global attributes of the target and supervising the training of the attention sub-module; wherein the attention sub-module includes multiple convolution layers and deconvolution layer, is responsible for constructing the attention map; the detection module includes an anchor frame filtering layer, a target detection layer and a parsing layer, and the anchor frame filtering layer performs data filtering on the received result of the attention map module and sends it to the target detection layer and The parsing layer performs detection and analysis,
  • the present invention also provides a method for detecting features within an object based on an attention mechanism, the method comprising the following steps:
  • Step S1 sample preparation: obtain the training image, mark the global attribute label of the image, the feature position of the image and the corresponding classification label;
  • Step S2 attention map training: the attention map module uses the training image and feature location information to generate attention map label information, and uses the attention map label information and image global attribute labels to supervise the attention module training;
  • Step S3 detection network training: fix the parameters of the attention module and the semantic extraction module, use the image feature position and image feature label to supervise the detection network training, and use the attention map to generate anchor frames, and introduce the attention module into the detection framework;
  • Step S4 global network optimization, according to the training, to obtain an optimized network framework
  • Step S5. For the new detection target, the image or video is acquired through the optical system, and the optimized network framework is imported to realize target positioning, analysis and detection.
  • the present invention also provides a computer-readable storage medium, on which computer instructions are stored, and the above-mentioned method is executed when the computer instructions are executed.
  • the beneficial effect of the present invention is that: the present invention adopts a multi-task learning method based on a deep convolutional network, and introduces attention learning and a single-scale detection mechanism to detect and locate the internal features of the target, and the target Classification and recognition of global attributes solves the problems of unbalanced sample distribution, multi-anchor boxes and multi-scales in the training phase of traditional solutions, which lead to high computing power requirements, and improves detection efficiency and accuracy.
  • Fig. 1 is a schematic diagram of the object feature detection system based on the attention mechanism of the present invention
  • Figure 2 is a schematic diagram of the process flow of the object feature detection method based on the attention mechanism
  • Figure 3 is a schematic diagram of candidate anchor box generation.
  • system means for distinguishing different components, elements, parts, parts or assemblies of different levels.
  • the words may be replaced by other expressions if other words can achieve the same purpose.
  • An attention mechanism-based feature detection system within an object see Figure 1, the system consists of a semantic extraction module, an attention map module, and a detection module.
  • the semantic extraction module includes a multi-layer deep convolutional network, which is responsible for extracting high-level semantic information from the input image, and sharing the extracted high-level semantic information to the attention map module and the detection module.
  • the attention map module includes a classification sub-module and an attention sub-module.
  • Each attribute branch of the classification submodule includes a plurality of convolutional layers, a global pooling layer, a global connection layer and a softmax layer, which are responsible for carrying out global attribute classification to the target and supervising the training of the attention submodule; wherein the attention
  • the force sub-module includes multiple convolutional layers and deconvolutional layers, responsible for constructing attention maps;
  • the detection module includes an anchor frame filtering layer, a target detection layer and an analysis layer, and the anchor frame filtering layer performs data filtering on the received results of the attention map module and sends them to the target detection layer and the analysis layer for detection and analysis , output the detection result.
  • the global attributes of a motor vehicle include the direction, model, and body color of the motor vehicle, and the vehicle features include windows, logos, lights, luggage racks, and sunroofs.
  • the training steps are divided into sample preparation, attention map multi-task training, detection network training, global network tuning, etc.
  • a kind of feature detection method in the target based on attention mechanism is implemented by the system of the first embodiment, referring to Fig. 2, method comprises the following steps.
  • Step S1 sample preparation: obtain the training image, mark the global attribute label of the image, the feature position of the image and the corresponding classification label.
  • Step S2 attention map training: the attention map module uses the training image and feature location information to generate attention map label information, and uses the attention map label information and image global attribute labels to supervise the attention module training; wherein, the generation method of the attention map label information include the following steps.
  • x and y represent the coordinates of the pixel in the attention map
  • x s , x e , y s , and y e represent the starting and ending positions of the target feature in the horizontal and vertical directions of the image, respectively.
  • Step S3 detection network training: fix the parameters of the attention module and semantic extraction module, use the image feature position and image feature label to supervise the detection network training, and use the attention map to generate anchor boxes, and introduce the attention module into the detection framework; among them, the anchor box
  • the generation method of is:
  • l i, j, k , t i, j, k , w i, j, k , h i, j, k are respectively the abscissa of the upper left corner of the anchor frame, the ordinate of the upper left corner, the width of the anchor frame and the anchor frame box height.
  • f is the value of each corresponding point in the anchor box area of the attention map.
  • T is the confidence filtering threshold.
  • Step S4 global network optimization, according to the training, to obtain an optimized network framework.
  • Step S5. For the new detection target, the image or video is acquired through the optical system, and the optimized network framework is imported to realize target positioning, analysis and detection.
  • the present invention also provides a computer-readable storage medium, on which computer instructions are stored, and the steps of the aforementioned method are executed when the computer instructions are run.
  • a computer-readable storage medium on which computer instructions are stored, and the steps of the aforementioned method are executed when the computer instructions are run.
  • Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory (ROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Flash memory or other memory technology, Compact Disc Read-Only Memory (CD-ROM), Digital Versatile Disc (DVD) or other optical storage, Magnetic tape cartridge, tape magnetic disk storage or other magnetic storage device or any other non-transmission medium that can be used to store information that can be accessed by a computing device.
  • computer-readable media excludes transitory computer-readable media, such as modulated data signals and carrier waves.
  • the computer program codes required for the operation of each part of this application can be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python etc., conventional procedural programming languages such as C language, VisualBasic, Fortran2003, Perl, COBOL2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages.
  • the program code may run entirely on the user's computer, or as a stand-alone software package, or run partly on the user's computer and partly on a remote computer, or entirely on the remote computer or processing device.
  • the remote computer can be connected to the user computer through any form of network, such as a local area network (LAN) or wide area network (WAN), or to an external computer (such as through the Internet), or in a cloud computing environment, or as a service Use software as a service (SaaS).
  • LAN local area network
  • WAN wide area network
  • SaaS service Use software as a service
  • the possible beneficial effects may be any one or a combination of the above, or any other possible beneficial effects.

Abstract

The present invention provides an attention mechanism-based system and method for detecting a feature in a target, and a storage medium, and relates to the field of intelligent security. The system comprises a semantic extraction module, an attention map module and a detection module, and a classification sub-module is responsible for carrying out global attribute classification on a target and for supervising the training of an attention sub-module; the attention sub-module is responsible for constructing an attention map; and the detection module comprises an anchor frame filter layer, a target detection layer and a parsing layer, the anchor frame filter layer performing data filtering on a received result of the attention map module, sending the result to the target detection layer and the parsing layer for detection and analysis, and outputting a detection result. In the present invention, a multi-task learning method based on a deep convolutional network is employed, and attention learning and a single-scale detection mechanism are introduced, a feature in a target is detected and positioned, and global attributes of the target are classified and recognized, which solves the problem that in training stages of conventional solutions, sample distribution is unbalanced and the requirement for computing power is high due to multiple anchor frames and multiple scales, thereby improving detection efficiency and precision.

Description

基于注意力机制的目标内特征检测系统、方法和存储介质In-target feature detection system, method and storage medium based on attention mechanism 技术领域technical field
本发明涉及智能安防领域,具体涉及一种基于注意力机制的目标内特征检测系统、方法和存储介质。The invention relates to the field of intelligent security, in particular to an attention mechanism-based feature detection system, method and storage medium in an object.
背景技术Background technique
目标检测是机器视觉的研究热点之一,而目标内特征检测是指,对目标内组件定位,以进一步分析视频、图像中的目标的结构化信息的技术。目标内特征检测是视频/图像结构化解析的重要技术之一,如车辆结构化任务中,需对车辆目标中的车脸、车窗、车灯、车标、行李架等特征进行定位,并做进一步分析。Object detection is one of the research hotspots in machine vision, and object feature detection refers to the technology of locating the components in the object to further analyze the structured information of the object in the video and image. In-target feature detection is one of the important technologies for video/image structured analysis. For example, in the vehicle structured task, it is necessary to locate the features of the vehicle target such as the face, windows, lights, logos, and luggage racks, and for further analysis.
现阶段目标内特征检测,主要借助于SSD、YOLO等通用场景检测方法,其不足如下:At this stage, the feature detection in the target mainly relies on general scene detection methods such as SSD and YOLO, and its shortcomings are as follows:
1.目标内特征检测通常从固定类型目标内定位多个特征位置,现有通用检测框架未考虑目标属性对目标内特征检测的指导性。1. In-target feature detection usually locates multiple feature locations from within a fixed type of target. The existing general detection framework does not consider the guidance of target attributes for in-target feature detection.
2.使用现有通用检测框架,需要构建多尺度的图像金字塔或特征金字塔,耗时较久,不利于边缘设备部署。2. Using the existing general detection framework, it is necessary to build a multi-scale image pyramid or feature pyramid, which takes a long time and is not conducive to the deployment of edge devices.
3.现有检测框架,借助深度卷积网络强大的拟合能力,使用同一特征图对目标进行检测框回归和分类,不利于性能的提升。3. The existing detection framework, with the help of the powerful fitting ability of the deep convolutional network, uses the same feature map to perform detection frame regression and classification on the target, which is not conducive to performance improvement.
4.现有检测框架的锚框选取方式,会生成大量的负样本,导致训练阶段样本不均衡问题,分类困难。4. The anchor box selection method of the existing detection framework will generate a large number of negative samples, resulting in the problem of unbalanced samples in the training phase and difficult classification.
发明内容Contents of the invention
为了克服现有技术的不足,本发明的目的在于提供一种基于注意力机制的 目标内特征检测系统、方法和存储介质,其能解决上述问题。In order to overcome the deficiencies in the prior art, the object of the present invention is to provide a system, method and storage medium for feature detection in an object based on attention mechanism, which can solve the above problems.
一种基于注意力机制的目标内特征检测系统,系统包括语义抽取模块、注意力图模块、和检测模块组成,其中:所述语义抽取模块包括多层深度卷积网络,负责从输入图像抽取高层语义信息,并将抽取的高层语义信息分享给所述注意力图模块和检测模块;所述注意力图模块包括分类子模块和注意力子模块;其中,所述分类子模块的每个属性分支包括多个卷积层、全局池化层、全局连接层和softmax层,负责对目标进行全局属性分类并监督注意力子模块的训练;其中,所述注意力子模块包括多个卷积层和反卷积层,负责构建注意力图;所述检测模块包括锚框过滤层、目标检测层和解析层,所述锚框过滤层将接收的所述注意力图模块的结果进行数据过滤并发送给目标检测层和解析层进行检测分析,输出检测结果。A feature detection system based on an attention mechanism, the system includes a semantic extraction module, an attention map module, and a detection module, wherein: the semantic extraction module includes a multi-layer deep convolutional network, which is responsible for extracting high-level semantics from input images information, and share the extracted high-level semantic information to the attention map module and the detection module; the attention map module includes a classification sub-module and an attention sub-module; wherein, each attribute branch of the classification sub-module includes multiple Convolution layer, global pooling layer, global connection layer and softmax layer are responsible for classifying the global attributes of the target and supervising the training of the attention sub-module; wherein the attention sub-module includes multiple convolution layers and deconvolution layer, is responsible for constructing the attention map; the detection module includes an anchor frame filtering layer, a target detection layer and a parsing layer, and the anchor frame filtering layer performs data filtering on the received result of the attention map module and sends it to the target detection layer and The parsing layer performs detection and analysis, and outputs the detection results.
本发明还提供了一种基于注意力机制的目标内特征检测方法,方法包括以下步骤:The present invention also provides a method for detecting features within an object based on an attention mechanism, the method comprising the following steps:
步骤S1、样本准备:获取训练图像,标注图像全局属性标签、图像特征位置及对应的分类标签;Step S1, sample preparation: obtain the training image, mark the global attribute label of the image, the feature position of the image and the corresponding classification label;
步骤S2、注意力图训练:注意力图模块使用训练图像以及特征位置信息,生成注意力图标签信息,并利用注意力图标签信息及图像全局属性标签监督注意力模块训练;Step S2, attention map training: the attention map module uses the training image and feature location information to generate attention map label information, and uses the attention map label information and image global attribute labels to supervise the attention module training;
步骤S3、检测网络训练:固定注意力模块及语义提取模块参数,使用图像特征位置、图像特征标签监督检测网络训练,并使用注意力图生成锚框,将注意力模块引入检测框架;Step S3, detection network training: fix the parameters of the attention module and the semantic extraction module, use the image feature position and image feature label to supervise the detection network training, and use the attention map to generate anchor frames, and introduce the attention module into the detection framework;
步骤S4、全局网络调优,根据训练,获得优化的网络框架;Step S4, global network optimization, according to the training, to obtain an optimized network framework;
步骤S5、针对新的检测目标,通过光学系统获取图像或视频,导入优化后的网络框架,实现目标定位、分析和检测。Step S5. For the new detection target, the image or video is acquired through the optical system, and the optimized network framework is imported to realize target positioning, analysis and detection.
本发明还提供了一种计算机可读存储介质,其上存储有计算机指令,所述计算机指令运行时执行前述的方法。The present invention also provides a computer-readable storage medium, on which computer instructions are stored, and the above-mentioned method is executed when the computer instructions are executed.
相比现有技术,本发明的有益效果在于:本发明采用基于深度卷积网络的多任务学习方法,并引入注意力学习,以及单尺度检测的机制,对目标内特征进行检测定位,对目标全局属性进行分类识别,解决了传统方案训练阶段样本分布不均衡、多锚框和多尺度导致算力需求高的问题,提高了检测效率和精度。Compared with the prior art, the beneficial effect of the present invention is that: the present invention adopts a multi-task learning method based on a deep convolutional network, and introduces attention learning and a single-scale detection mechanism to detect and locate the internal features of the target, and the target Classification and recognition of global attributes solves the problems of unbalanced sample distribution, multi-anchor boxes and multi-scales in the training phase of traditional solutions, which lead to high computing power requirements, and improves detection efficiency and accuracy.
附图说明Description of drawings
图1为本发明基于注意力机制的目标内特征检测系统示意图;Fig. 1 is a schematic diagram of the object feature detection system based on the attention mechanism of the present invention;
图2为基于注意力机制的目标内特征检测方法流程示意图;Figure 2 is a schematic diagram of the process flow of the object feature detection method based on the attention mechanism;
图3为候选锚框生成示意图。Figure 3 is a schematic diagram of candidate anchor box generation.
具体实施方式Detailed ways
为使本发明实施例的目的、技术方案和优点更加清楚,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本发明的一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本发明保护的范围。In order to make the purpose, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments It is a part of embodiments of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.
应当理解,本说明书中所使用的“系统”、“装置”、“单元”和/或“模组”是用于区分不同级别的不同组件、元件、部件、部分或装配的一种方法。然而,如果其他词语可实现相同的目的,则可通过其他表达来替换所述词语。It should be understood that "system", "device", "unit" and/or "module" used in this specification is a method for distinguishing different components, elements, parts, parts or assemblies of different levels. However, the words may be replaced by other expressions if other words can achieve the same purpose.
如本说明书和权利要求书中所示,除非上下文明确提示例外情形,“一”、“一个”、“一种”和/或“该”等词并非特指单数,也可包括复数。一般说来,术语“包括”与“包含”仅提示包括已明确标识的步骤和元素,而这些步骤和 元素不构成一个排它性的罗列,方法或者设备也可能包含其它的步骤或元素。As indicated in the specification and claims, the terms "a", "an", "an" and/or "the" are not specific to the singular and may include the plural unless the context clearly indicates an exception. Generally speaking, the terms "comprising" and "comprising" only suggest the inclusion of explicitly identified steps and elements, and these steps and elements do not constitute an exclusive list, and the method or device may also contain other steps or elements.
本说明书中使用了流程图用来说明根据本说明书的实施例的系统所执行的操作。应当理解的是,前面或后面操作不一定按照顺序来精确地执行。相反,可以按照倒序或同时处理各个步骤。同时,也可以将其他操作添加到这些过程中,或从这些过程移除某一步或数步操作。The flowchart is used in this specification to illustrate the operations performed by the system according to the embodiment of this specification. It should be understood that the preceding or following operations are not necessarily performed in the exact order. Instead, various steps may be processed in reverse order or simultaneously. At the same time, other operations can be added to these procedures, or a certain step or steps can be removed from these procedures.
第一实施例first embodiment
一种基于注意力机制的目标内特征检测系统,参见图1,系统包括语义抽取模块、注意力图模块、和检测模块组成。An attention mechanism-based feature detection system within an object, see Figure 1, the system consists of a semantic extraction module, an attention map module, and a detection module.
其中,语义抽取模块包括多层深度卷积网络,负责从输入图像抽取高层语义信息,并将抽取的高层语义信息分享给所述注意力图模块和检测模块。Among them, the semantic extraction module includes a multi-layer deep convolutional network, which is responsible for extracting high-level semantic information from the input image, and sharing the extracted high-level semantic information to the attention map module and the detection module.
其中,所述注意力图模块包括分类子模块和注意力子模块。Wherein, the attention map module includes a classification sub-module and an attention sub-module.
所述分类子模块的每个属性分支包括多个卷积层、全局池化层、全局连接层和softmax层,负责对目标进行全局属性分类并监督注意力子模块的训练;其中,所述注意力子模块包括多个卷积层和反卷积层,负责构建注意力图;Each attribute branch of the classification submodule includes a plurality of convolutional layers, a global pooling layer, a global connection layer and a softmax layer, which are responsible for carrying out global attribute classification to the target and supervising the training of the attention submodule; wherein the attention The force sub-module includes multiple convolutional layers and deconvolutional layers, responsible for constructing attention maps;
其中,所述检测模块包括锚框过滤层、目标检测层和解析层,所述锚框过滤层将接收的所述注意力图模块的结果进行数据过滤并发送给目标检测层和解析层进行检测分析,输出检测结果。Wherein, the detection module includes an anchor frame filtering layer, a target detection layer and an analysis layer, and the anchor frame filtering layer performs data filtering on the received results of the attention map module and sends them to the target detection layer and the analysis layer for detection and analysis , output the detection result.
以机动车为例,图1中,机动车的全局属性包括机动车方向、车型、车身颜色等,车辆特征包括车窗、车标、车灯、行李架、天窗等。训练步骤分为样本准备、注意力图多任务训练、检测网络训练、全局网络调优等。Taking a motor vehicle as an example, in Figure 1, the global attributes of a motor vehicle include the direction, model, and body color of the motor vehicle, and the vehicle features include windows, logos, lights, luggage racks, and sunroofs. The training steps are divided into sample preparation, attention map multi-task training, detection network training, global network tuning, etc.
第二实施例second embodiment
一种基于注意力机制的目标内特征检测方法,该方法由第一实施例的系统 实施,参见图2,方法包括以下步骤。A kind of feature detection method in the target based on attention mechanism, this method is implemented by the system of the first embodiment, referring to Fig. 2, method comprises the following steps.
步骤S1、样本准备:获取训练图像,标注图像全局属性标签、图像特征位置及对应的分类标签。Step S1, sample preparation: obtain the training image, mark the global attribute label of the image, the feature position of the image and the corresponding classification label.
步骤S2、注意力图训练:注意力图模块使用训练图像以及特征位置信息,生成注意力图标签信息,并利用注意力图标签信息及图像全局属性标签监督注意力模块训练;其中,注意力图标签信息的生成方法包括以下步骤。Step S2, attention map training: the attention map module uses the training image and feature location information to generate attention map label information, and uses the attention map label information and image global attribute labels to supervise the attention module training; wherein, the generation method of the attention map label information Include the following steps.
S21、计算标签每类特征的均值图像,表示为:
Figure PCTCN2021095956-appb-000001
S21. Calculate the mean image of each feature of the label, expressed as:
Figure PCTCN2021095956-appb-000001
S22、根据样本均值图像
Figure PCTCN2021095956-appb-000002
计算样本差异图重心(x c,y c);样本差异图重心(x c,y c)的计算公式为:
S22, according to the sample mean image
Figure PCTCN2021095956-appb-000002
Calculate the center of gravity (x c , y c ) of the sample difference map; the formula for calculating the center of gravity (x c , y c ) of the sample difference map is:
Figure PCTCN2021095956-appb-000003
Figure PCTCN2021095956-appb-000003
式中,
Figure PCTCN2021095956-appb-000004
为样本均值图像
Figure PCTCN2021095956-appb-000005
对应坐标(i,j)处的像素值,p i,j为特征图像(i,j)处的坐标值。
In the formula,
Figure PCTCN2021095956-appb-000004
is the sample mean image
Figure PCTCN2021095956-appb-000005
Corresponding to the pixel value at the coordinate (i, j), p i, j is the coordinate value at the feature image (i, j).
S23、根据差异图重心(x c,y c)生成注意力图G(x,y)。注意力图G(x,y)的计算公式为: S23. Generate an attention map G(x, y) according to the center of gravity (x c , y c ) of the difference map. The calculation formula of the attention map G(x, y) is:
Figure PCTCN2021095956-appb-000006
Figure PCTCN2021095956-appb-000006
式中,x、y表示像素在注意力图中的坐标,x s、x e、y s、y e分别表示目标特征在图像中横轴、纵轴方向的起始、终止位置。 In the formula, x and y represent the coordinates of the pixel in the attention map, and x s , x e , y s , and y e represent the starting and ending positions of the target feature in the horizontal and vertical directions of the image, respectively.
步骤S3、检测网络训练:固定注意力模块及语义提取模块参数,使用图像特征位置、图像特征标签监督检测网络训练,并使用注意力图生成锚框,将注意力模块引入检测框架;其中,锚框的生成方法为:Step S3, detection network training: fix the parameters of the attention module and semantic extraction module, use the image feature position and image feature label to supervise the detection network training, and use the attention map to generate anchor boxes, and introduce the attention module into the detection framework; among them, the anchor box The generation method of is:
S31、生成候选锚框,参见图3,以注意力图中的每个位置为锚点,并以锚点为中心生成不同尺度的矩形框,作为候选框;每个坐标为(i,j)处的锚点对应的多个候选框中的第k个候选框Bbox i,j,k为: S31. Generate candidate anchor boxes, see Figure 3, use each position in the attention map as an anchor point, and generate rectangular boxes of different scales centered on the anchor point as candidate boxes; each coordinate is (i, j) The k-th candidate box Bbox i, j, k among multiple candidate boxes corresponding to the anchor point of is:
Bbox i,j,k={l i,j,k,t i,j,k,w i,j,k,h i,j,k}…………式3; Bbox i, j, k = {l i, j, k , t i, j, k , w i, j, k , h i, j, k }…… Formula 3;
式中,l i,j,k、t i,j,k、w i,j,k、h i,j,k分别为该锚框左上角横坐标、左上角纵坐标、锚框宽度及锚框高度。 In the formula, l i, j, k , t i, j, k , w i, j, k , h i, j, k are respectively the abscissa of the upper left corner of the anchor frame, the ordinate of the upper left corner, the width of the anchor frame and the anchor frame box height.
S32、计算每个候选框的置信度C i,j,kS32. Calculate the confidence C i,j,k of each candidate frame:
Figure PCTCN2021095956-appb-000007
Figure PCTCN2021095956-appb-000007
式中,f为注意力图中锚框区域内各对应点的值。where f is the value of each corresponding point in the anchor box area of the attention map.
S33、根据候选框的置信度,对候选框进行过滤得到最终的候选框集合Bboxes:S33. According to the confidence of the candidate boxes, filter the candidate boxes to obtain the final candidate box set Bboxes:
Bboxes={C i,j,k≥T}………………………………式5; Bboxes={C i, j, k ≥ T}…………………… Formula 5;
式中,T为置信度过滤阈值。In the formula, T is the confidence filtering threshold.
步骤S4、全局网络调优,根据训练,获得优化的网络框架。Step S4, global network optimization, according to the training, to obtain an optimized network framework.
步骤S5、针对新的检测目标,通过光学系统获取图像或视频,导入优化后的网络框架,实现目标定位、分析和检测。Step S5. For the new detection target, the image or video is acquired through the optical system, and the optimized network framework is imported to realize target positioning, analysis and detection.
第三实施例third embodiment
本发明还提供了一种计算机可读存储介质,其上存储有计算机指令,所述计算机指令运行时执行前述方法的步骤。其中,所述方法请参见前述部分的详细介绍,此处不再赘述。The present invention also provides a computer-readable storage medium, on which computer instructions are stored, and the steps of the aforementioned method are executed when the computer instructions are run. Wherein, for the method, please refer to the detailed introduction in the foregoing part, and details will not be repeated here.
本领域普通技术人员可以理解上述实施例的各种方法中的全部或部分步骤是可以通过程序来指令相关的硬件来完成,该程序可以存储于计算机可读存储介质中,计算机可读介质包括永久性和非永久性、可移动和非可移动媒体可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括,但不限于相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器 (EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。按照本文中的界定,计算机可读介质不包括暂存电脑可读媒体(transitory media),如调制的数据信号和载波。Those of ordinary skill in the art can understand that all or part of the steps in the various methods of the above-mentioned embodiments can be completed by instructing related hardware through a program, and the program can be stored in a computer-readable storage medium, and the computer-readable medium includes permanent Both non-permanent and non-permanent, removable and non-removable media can be implemented by any method or technology for information storage. Information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory (ROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Flash memory or other memory technology, Compact Disc Read-Only Memory (CD-ROM), Digital Versatile Disc (DVD) or other optical storage, Magnetic tape cartridge, tape magnetic disk storage or other magnetic storage device or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, computer-readable media excludes transitory computer-readable media, such as modulated data signals and carrier waves.
本申请各部分操作所需的计算机程序编码可以用任意一种或多种程序语言编写,包括面向对象编程语言如Java、Scala、Smalltalk、Eiffel、JADE、Emerald、C++、C#、VB.NET、Python等,常规程序化编程语言如C语言、VisualBasic、Fortran2003、Perl、COBOL2002、PHP、ABAP,动态编程语言如Python、Ruby和Groovy,或其他编程语言等。该程序编码可以完全在用户计算机上运行、或作为独立的软件包在用户计算机上运行、或部分在用户计算机上运行部分在远程计算机运行、或完全在远程计算机或处理设备上运行。在后种情况下,远程计算机可以通过任何网络形式与用户计算机连接,比如局域网(LAN)或广域网(WAN),或连接至外部计算机(例如通过因特网),或在云计算环境中,或作为服务使用如软件即服务(SaaS)。The computer program codes required for the operation of each part of this application can be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python etc., conventional procedural programming languages such as C language, VisualBasic, Fortran2003, Perl, COBOL2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages. The program code may run entirely on the user's computer, or as a stand-alone software package, or run partly on the user's computer and partly on a remote computer, or entirely on the remote computer or processing device. In the latter case, the remote computer can be connected to the user computer through any form of network, such as a local area network (LAN) or wide area network (WAN), or to an external computer (such as through the Internet), or in a cloud computing environment, or as a service Use software as a service (SaaS).
需要说明的是,不同实施例可能产生的有益效果不同,在不同的实施例里,可能产生的有益效果可以是以上任意一种或几种的组合,也可以是其他任何可能获得的有益效果。It should be noted that different embodiments may have different beneficial effects. In different embodiments, the possible beneficial effects may be any one or a combination of the above, or any other possible beneficial effects.
最后应说明的是:以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。Finally, it should be noted that: the above embodiments are only used to illustrate the technical solutions of the present invention, rather than to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: it can still be Modifications are made to the technical solutions described in the foregoing embodiments, or equivalent replacements are made to some of the technical features; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the various embodiments of the present invention.

Claims (7)

  1. 一种基于注意力机制的目标内特征检测系统,其特征在于:系统包括语义抽取模块、注意力图模块、和检测模块组成,其中:A feature detection system based on an attention mechanism, characterized in that: the system includes a semantic extraction module, an attention map module, and a detection module, wherein:
    所述语义抽取模块包括多层深度卷积网络,负责从输入图像抽取高层语义信息,并将抽取的高层语义信息分享给所述注意力图模块和检测模块;The semantic extraction module includes a multi-layer deep convolutional network, which is responsible for extracting high-level semantic information from the input image, and sharing the extracted high-level semantic information to the attention map module and detection module;
    所述注意力图模块包括分类子模块和注意力子模块;其中,所述分类子模块的每个属性分支包括多个卷积层、全局池化层、全局连接层和softmax层,负责对目标进行全局属性分类并监督注意力子模块的训练;其中,所述注意力子模块包括多个卷积层和反卷积层,负责构建注意力图;The attention map module includes a classification sub-module and an attention sub-module; wherein, each attribute branch of the classification sub-module includes a plurality of convolutional layers, a global pooling layer, a global connection layer and a softmax layer, which are responsible for the target Global attribute classification and supervision of the training of the attention sub-module; wherein the attention sub-module includes a plurality of convolutional layers and deconvolution layers, responsible for constructing the attention map;
    所述检测模块包括锚框过滤层、目标检测层和解析层,所述锚框过滤层将接收的所述注意力图模块的结果进行数据过滤并发送给目标检测层和解析层进行检测分析,输出检测结果。The detection module includes an anchor frame filter layer, a target detection layer and an analysis layer, and the anchor frame filter layer performs data filtering on the received result of the attention map module and sends it to the target detection layer and the analysis layer for detection and analysis, and outputs Test results.
  2. 一种基于注意力机制的目标内特征检测方法,其特征在于,方法包括以下步骤:A method for feature detection in a target based on an attention mechanism, characterized in that the method comprises the following steps:
    步骤S1、样本准备:获取训练图像,标注图像全局属性标签、图像特征位置及对应的分类标签;Step S1, sample preparation: obtain the training image, mark the global attribute label of the image, the feature position of the image and the corresponding classification label;
    步骤S2、注意力图训练:注意力图模块使用训练图像以及特征位置信息,生成注意力图标签信息,并利用注意力图标签信息及图像全局属性标签监督注意力模块训练;Step S2, attention map training: the attention map module uses the training image and feature location information to generate attention map label information, and uses the attention map label information and image global attribute labels to supervise the attention module training;
    步骤S3、检测网络训练:固定注意力模块及语义提取模块参数,使用图像特征位置、图像特征标签监督检测网络训练,并使用注意力图生成锚框,将注意力模块引入检测框架;Step S3, detection network training: fix the parameters of the attention module and the semantic extraction module, use the image feature position and image feature label to supervise the detection network training, and use the attention map to generate anchor frames, and introduce the attention module into the detection framework;
    步骤S4、全局网络调优,根据训练,获得优化的网络框架;Step S4, global network optimization, according to the training, to obtain an optimized network framework;
    步骤S5、针对新的检测目标,通过光学系统获取图像或视频,导入优化后的网络框架,实现目标定位、分析和检测。Step S5. For the new detection target, the image or video is acquired through the optical system, and the optimized network framework is imported to realize target positioning, analysis and detection.
  3. 根据权利要求2所述的检测方法,其特征在于,步骤S2中的注意力图标签信息的生成方法包括以下步骤:The detection method according to claim 2, wherein the generation method of the attention map label information in step S2 comprises the following steps:
    S21、计算标签每类特征的均值图像,表示为:
    Figure PCTCN2021095956-appb-100001
    S21. Calculate the mean image of each feature of the label, expressed as:
    Figure PCTCN2021095956-appb-100001
    S22、根据样本均值图像
    Figure PCTCN2021095956-appb-100002
    计算样本差异图重心(x c,y c);
    S22, according to the sample mean image
    Figure PCTCN2021095956-appb-100002
    Calculate the center of gravity of the sample difference map (x c , y c );
    S23、根据差异图重心(x c,y c)生成注意力图G(x,y)。 S23. Generate an attention map G(x, y) according to the center of gravity (x c , y c ) of the difference map.
  4. 根据权利要求3所述的检测方法,其特征在于:样本差异图重心(x c,y c)的计算公式为: The detection method according to claim 3, characterized in that: the calculation formula for the center of gravity (x c , y c ) of the sample difference map is:
    Figure PCTCN2021095956-appb-100003
    Figure PCTCN2021095956-appb-100003
    式中,
    Figure PCTCN2021095956-appb-100004
    为样本均值图像
    Figure PCTCN2021095956-appb-100005
    对应坐标(i,j)处的像素值,p i,j为特征图像(i,j)处的坐标值。
    In the formula,
    Figure PCTCN2021095956-appb-100004
    is the sample mean image
    Figure PCTCN2021095956-appb-100005
    Corresponding to the pixel value at the coordinate (i, j), p i, j is the coordinate value at the feature image (i, j).
  5. 根据权利要求4所述的检测方法,其特征在于:注意力图G(x,y)的计算公式为:The detection method according to claim 4, characterized in that: the calculation formula of the attention map G (x, y) is:
    Figure PCTCN2021095956-appb-100006
    Figure PCTCN2021095956-appb-100006
    式中,x、y表示像素在注意力图中的坐标,x s、x e、y s、y e分别表示目标特征在图像中横轴、纵轴方向的起始、终止位置。 In the formula, x and y represent the coordinates of the pixel in the attention map, and x s , x e , y s , and y e represent the starting and ending positions of the target feature in the horizontal and vertical directions of the image, respectively.
  6. 根据权利要求5所述的检测方法,其特征在于,步骤S3中锚框的生成方法为:The detection method according to claim 5, wherein the generation method of the anchor frame in step S3 is:
    S31、生成候选锚框,以注意力图中的每个位置为锚点,并以锚点为中心生成不同尺度的矩形框,作为候选框;每个坐标为(i,j)处的锚点对应的多个候选框中的第k个候选框Bbox i,j,k为: S31. Generate a candidate anchor frame, use each position in the attention map as an anchor point, and generate rectangular frames of different scales centered on the anchor point as a candidate frame; each coordinate corresponds to the anchor point at (i, j) The k-th candidate box Bbox i, j, k in multiple candidate boxes of is:
    Bbox i,j,k={l i,j,k,t i,j,k,w i,j,k,h i,j,k}…………式3; Bbox i, j, k = {l i, j, k , t i, j, k , w i, j, k , h i, j, k }…… Formula 3;
    式中,l i,j,k、t i,j,k、w i,j,k、h i,j,k分别为该锚框左上角横坐标、左上角纵坐标、锚框宽度及锚框高度; In the formula, l i, j, k , t i, j, k , w i, j, k , h i, j, k are respectively the abscissa of the upper left corner of the anchor frame, the ordinate of the upper left corner, the width of the anchor frame and the anchor frame frame height;
    S32、计算每个候选框的置信度C i,j,kS32. Calculate the confidence C i,j,k of each candidate frame:
    Figure PCTCN2021095956-appb-100007
    Figure PCTCN2021095956-appb-100007
    式中,f为注意力图中锚框区域内各对应点的值;In the formula, f is the value of each corresponding point in the anchor frame area of the attention map;
    S33、根据候选框的置信度,对候选框进行过滤得到最终的候选框集合Bboxes:S33. According to the confidence of the candidate boxes, filter the candidate boxes to obtain the final candidate box set Bboxes:
    Bboxes={C i,j,k≥T}………………………………式5; Bboxes={C i, j, k ≥ T}…………………… Formula 5;
    式中,T为置信度过滤阈值。In the formula, T is the confidence filtering threshold.
  7. 一种计算机可读存储介质,其上存储有计算机指令,其特征在于:所述计算机指令运行时执行权利要求2-6任一项所述的方法。A computer-readable storage medium on which computer instructions are stored, wherein the method according to any one of claims 2-6 is executed when the computer instructions are run.
PCT/CN2021/095956 2021-05-20 2021-05-26 Attention mechanism-based system and method for detecting feature in target, and storage medium WO2022241803A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110554342.8A CN113255759B (en) 2021-05-20 2021-05-20 In-target feature detection system, method and storage medium based on attention mechanism
CN202110554342.8 2021-05-20

Publications (1)

Publication Number Publication Date
WO2022241803A1 true WO2022241803A1 (en) 2022-11-24

Family

ID=77183214

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/095956 WO2022241803A1 (en) 2021-05-20 2021-05-26 Attention mechanism-based system and method for detecting feature in target, and storage medium

Country Status (2)

Country Link
CN (1) CN113255759B (en)
WO (1) WO2022241803A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113792632A (en) * 2021-09-02 2021-12-14 广州广电运通金融电子股份有限公司 Finger vein identification method, system and storage medium based on multi-party cooperation
CN114973064A (en) * 2022-04-29 2022-08-30 华为技术有限公司 Method and device for generating pseudo label frame and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993101A (en) * 2019-03-28 2019-07-09 华南理工大学 The vehicle checking method returned based on branch intensive loop from attention network and circulation frame
US20200410235A1 (en) * 2018-09-21 2020-12-31 Ancestry.Com Operations Inc. Ventral-dorsal neural networks: object detection via selective attention
CN112183414A (en) * 2020-09-29 2021-01-05 南京信息工程大学 Weak supervision remote sensing target detection method based on mixed hole convolution
CN112200089A (en) * 2020-10-12 2021-01-08 西南交通大学 Dense vehicle detection method based on vehicle counting perception attention
US20210012146A1 (en) * 2019-07-12 2021-01-14 Wuyi University Method and apparatus for multi-scale sar image recognition based on attention mechanism

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9858496B2 (en) * 2016-01-20 2018-01-02 Microsoft Technology Licensing, Llc Object detection and classification in images
CN111985284A (en) * 2019-05-21 2020-11-24 天津科技大学 Single-stage target detection device without anchor box based on attention mechanism and semantic weak supervision
CN111178213B (en) * 2019-12-23 2022-11-18 大连理工大学 Aerial photography vehicle detection method based on deep learning
CN111242071B (en) * 2020-01-17 2023-04-07 陕西师范大学 Attention remote sensing image target detection method based on anchor frame
CN111275688B (en) * 2020-01-19 2023-12-12 合肥工业大学 Small target detection method based on context feature fusion screening of attention mechanism
CN111401201B (en) * 2020-03-10 2023-06-20 南京信息工程大学 Aerial image multi-scale target detection method based on spatial pyramid attention drive
CN111539469B (en) * 2020-04-20 2022-04-08 东南大学 Weak supervision fine-grained image identification method based on vision self-attention mechanism
CN111666836B (en) * 2020-05-22 2023-05-02 北京工业大学 High-resolution remote sensing image target detection method of M-F-Y type light convolutional neural network
CN112232231B (en) * 2020-10-20 2024-02-02 城云科技(中国)有限公司 Pedestrian attribute identification method, system, computer equipment and storage medium
CN112308019B (en) * 2020-11-19 2021-08-17 中国人民解放军国防科技大学 SAR ship target detection method based on network pruning and knowledge distillation
CN112200161B (en) * 2020-12-03 2021-03-02 北京电信易通信息技术股份有限公司 Face recognition detection method based on mixed attention mechanism
CN112580664A (en) * 2020-12-15 2021-03-30 哈尔滨理工大学 Small target detection method based on SSD (solid State disk) network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200410235A1 (en) * 2018-09-21 2020-12-31 Ancestry.Com Operations Inc. Ventral-dorsal neural networks: object detection via selective attention
CN109993101A (en) * 2019-03-28 2019-07-09 华南理工大学 The vehicle checking method returned based on branch intensive loop from attention network and circulation frame
US20210012146A1 (en) * 2019-07-12 2021-01-14 Wuyi University Method and apparatus for multi-scale sar image recognition based on attention mechanism
CN112183414A (en) * 2020-09-29 2021-01-05 南京信息工程大学 Weak supervision remote sensing target detection method based on mixed hole convolution
CN112200089A (en) * 2020-10-12 2021-01-08 西南交通大学 Dense vehicle detection method based on vehicle counting perception attention

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
WANG, LI ET AL.: "S-AT GCN: Spatial-Attention Graph Convolution Network based Feature Enhancement for 3D Object Detection", HTTPS://ARXIV.ORG/PDF/2103.08439, 15 March 2021 (2021-03-15), XP093007671 *

Also Published As

Publication number Publication date
CN113255759B (en) 2023-08-22
CN113255759A (en) 2021-08-13

Similar Documents

Publication Publication Date Title
Wiley et al. Computer vision and image processing: a paper review
Chen et al. Learning context flexible attention model for long-term visual place recognition
US10963632B2 (en) Method, apparatus, device for table extraction based on a richly formatted document and medium
CN111460968B (en) Unmanned aerial vehicle identification and tracking method and device based on video
CN111080693A (en) Robot autonomous classification grabbing method based on YOLOv3
Chen et al. Object-level motion detection from moving cameras
WO2021238019A1 (en) Real-time traffic flow detection system and method based on ghost convolutional feature fusion neural network
CN110929593B (en) Real-time significance pedestrian detection method based on detail discrimination
CN113076871B (en) Fish shoal automatic detection method based on target shielding compensation
WO2022241803A1 (en) Attention mechanism-based system and method for detecting feature in target, and storage medium
CN109447979B (en) Target detection method based on deep learning and image processing algorithm
CN111460927B (en) Method for extracting structured information of house property evidence image
US20230267735A1 (en) Method for structuring pedestrian information, device, apparatus and storage medium
Shen et al. Vehicle detection in aerial images based on lightweight deep convolutional network and generative adversarial network
CN110751232A (en) Chinese complex scene text detection and identification method
Li et al. ComNet: Combinational neural network for object detection in UAV-borne thermal images
CN111813997A (en) Intrusion analysis method, device, equipment and storage medium
CN112488229A (en) Domain self-adaptive unsupervised target detection method based on feature separation and alignment
CN113869264A (en) Commodity identification method, commodity identification system, storage medium and server
CN113591746A (en) Document table structure detection method and device
CN115471773B (en) Intelligent classroom-oriented student tracking method and system
CN116152696A (en) Intelligent security image identification method and system for industrial control system
CN113139540B (en) Backboard detection method and equipment
CN114898290A (en) Real-time detection method and system for marine ship
CN110427920B (en) Real-time pedestrian analysis method oriented to monitoring environment

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21940256

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE