CN113673273A - Design method of quantifiable front-end vehicle detection network structure - Google Patents
Design method of quantifiable front-end vehicle detection network structure Download PDFInfo
- Publication number
- CN113673273A CN113673273A CN202010400217.7A CN202010400217A CN113673273A CN 113673273 A CN113673273 A CN 113673273A CN 202010400217 A CN202010400217 A CN 202010400217A CN 113673273 A CN113673273 A CN 113673273A
- Authority
- CN
- China
- Prior art keywords
- layer
- excitation function
- input data
- result
- output
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
- Navigation (AREA)
Abstract
The invention provides a design method of a quantifiable front-end vehicle detection network structure, which comprises the following steps: s1, adopting a two-level network, wherein the first level input is a gray scale map of 47X47, and the second level network input is a gray scale map of 49X 49; defining an excitation function by self, quantizing the result of each layer to 4 bits, and quantizing the boundary data; s2, designing a secondary network structure: a first-level network: the step size of the first three layers is 2, the step size of the fourth layer is 1, and the fourth layer uses a custom excitation function and a 3 multiplied by 3 convolution kernel; the fifth layer uses the stimulus function is sigmol; the sixth layer is not processed using the stimulus function; performing convolution on each layer by using a non-alignment processing mode; a second-level network: the step size of the first layer is 1, the convolution size is 3 multiplied by 3, and a custom excitation function is used; the step length from the second layer to the fifth layer is 2, the convolution size is 3 multiplied by 3, and a custom excitation function is used; the sixth layer does not use an excitation function; the seventh layer uses the fully-connected excitation function is sigmol; the eighth layer uses full connectivity without using a stimulus function.
Description
Technical Field
The invention relates to the technical field of neural networks, in particular to a design method of a quantifiable front-end vehicle detection network structure.
Background
In the current society, the development of the neural network technology in the field of artificial intelligence is rapid. MTCNN technology is also one of the more popular technologies in recent years. MTCNN, Multi-task convolutional neural network, puts face region detection and face keypoint detection together, and can be generally divided into three-layer network structures of P-Net, R-Net and O-Net. The multi-task neural network model for the face detection task mainly adopts three cascaded networks and adopts the idea of adding a classifier into a candidate frame to carry out rapid and efficient face detection. The three cascaded networks are respectively P-Net for quickly generating candidate windows, R-Net for filtering and selecting high-precision candidate windows and O-Net for generating final bounding boxes and key points of the human face.
However, MTCNN cascade detection has the following drawbacks:
1. and the three-level cascade network is used, so that the running time for processing a fixed size picture is long. The missed detection is more.
2. The excitation function used is not suitable for the quantization process.
3. The three-level network is used in face detection and is not suitable for vehicle detection.
4. The pooling process which can not be quantitatively used on chips produced by Beijing Junzhen integrated circuit GmbH, for example, is used, so that only floating-point operation can be performed on some chips (such as Junzhen chips), and the operation time is long.
In addition, the following commonly used technical terms are also included in the prior art:
1. cascading: the mode that several detectors detect in series is called cascade.
2. And (3) convolution kernel: the convolution kernel is a parameter used for performing an operation on a matrix and an original image during image processing. The convolution kernel is typically a matrix of column numbers (e.g., a 3 x 3 matrix) with a weight value for each square on the region. The matrix shape is typically 1 × 1,3 × 3,5 × 5,7 × 7,1 × 3,3 × 1,2 × 2,1 × 5,5 × 1, … ….
3. Convolution: the centre of the convolution kernel is placed on the pixel to be calculated, the products of each element in the kernel and its covered image pixel value are calculated once and summed, and the resulting structure is the new pixel value at that location, a process called convolution.
4. Excitation function: a function that processes the convolved results.
5. Characteristic diagram: the result of the convolution calculation of the input data is called a feature map, and the result of the full connection of the data is also called a feature map. The feature size is typically expressed as length x width x depth, or 1 x depth.
6. Step length: the length by which the center position of the convolution kernel is shifted in coordinates.
7. And (3) carrying out non-alignment treatment on two ends: when the image or data is processed by the convolution kernel with the size of 3 × 3, if one convolution kernel is not enough, the data on two sides is insufficient, and the phenomenon of discarding the data on two sides or one side is adopted, which is called non-alignment processing on two ends.
Disclosure of Invention
In order to solve the problems of the prior art, the invention aims to:
1. and by using two-stage cascade, the size of the detection target is suitable for vehicle detection. The method has the advantages of short running time, less false detection and high accuracy when fixed-size pictures are processed, and meets the actual use requirements. The detection distance increases.
2. The pooling process is eliminated using an excitation function according to a quantifiable design.
Specifically, the invention provides a design method of a quantifiable front-end vehicle detection network structure, which comprises the following steps:
s1, designing a network to adopt a two-stage network, wherein an excitation function is a self-defined function, the first-stage input size is a gray scale image of 47X47, and the second-stage network input size is a gray scale image of 49X 49; defining an excitation function by self, quantizing the calculated results of each layer to 4 bits, and quantizing the boundary data;
s2, designing a secondary network structure:
s2.1, a first-level network:
the step size used by the first three layers is 2, the step size used by the fourth layer is 1, the four layers use custom excitation functions, and each layer uses a convolution kernel of 3 multiplied by 3;
the last two layers, i.e. the fifth layer and the sixth layer, are parallel layers, and the calculation results of the fourth layer are used; the fifth layer calculation result is used for judging whether the vehicle is a value of the vehicle, and the used excitation function is sigmol; the result of the sixth layer of calculation is used for fine tuning the value of the vehicle detection frame without any excitation function processing;
performing convolution on each layer by using a non-alignment processing mode;
s2.2, a second-level network:
the step length used by the first layer is 1, the convolution size is 3 multiplied by 3, and the user-defined excitation function is used for processing;
the step length from the second layer to the fifth layer, namely the second layer, the third layer, the fourth layer and the fifth layer is 2, the convolution size is 3 multiplied by 3, the user-defined excitation function is used for processing, and the input data of each layer from the second layer is the output result of the previous layer;
the sixth layer of input data is the output result of the fifth layer, and full connection is used without any excitation function;
the last two layers, namely the seventh layer and the eighth layer, are parallel layers, and the calculation result of the sixth layer is used; the seventh layer uses full connection, and the excitation function is sigmol and is used for judging whether the result is a target result; the eighth layer uses full connectivity, without using an excitation function, for fine tuning the values of the coordinates.
In S1, the custom excitation function is that, let x be the value of the input excitation function, y be the value of the output function, and the formula is:
thus, the present application has the advantages that: the method is simple, the recall rate and the accuracy rate of the face detection are improved with a little time cost, and the network can be quantized.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the principles of the invention.
FIG. 1 is a schematic flow diagram of the process of the present invention.
Fig. 2 is a schematic diagram of a first level network model of the network architecture in the method of the present invention.
Fig. 3 is a schematic diagram of a second level network model of the network architecture in the method of the present invention.
Detailed Description
In order that the technical contents and advantages of the present invention can be more clearly understood, the present invention will now be described in further detail with reference to the accompanying drawings.
As shown in fig. 1, the present invention relates to a design method of a quantifiable front-end vehicle detection network structure, which comprises the following steps:
s1, designing a network to adopt a two-stage network, wherein an excitation function is a self-defined function, the first-stage input size is a gray scale image of 47X47, and the second-stage network input size is a gray scale image of 49X 49; defining an excitation function by self, quantizing the calculated results of each layer to 4 bits, and quantizing the boundary data;
s2, designing a secondary network structure:
s2.1, a first-level network:
the step size used by the first three layers is 2, the step size used by the fourth layer is 1, the four layers use custom excitation functions, and each layer uses a convolution kernel of 3 multiplied by 3;
the last two layers, i.e. the fifth layer and the sixth layer, are parallel layers, and the calculation results of the fourth layer are used; the fifth layer calculation result is used for judging whether the vehicle is a value of the vehicle, and the used excitation function is sigmol; the result of the sixth layer of calculation is used for fine tuning the value of the vehicle detection frame without any excitation function processing;
performing convolution on each layer by using a non-alignment processing mode;
s2.2, a second-level network:
the step length used by the first layer is 1, the convolution size is 3 multiplied by 3, and the user-defined excitation function is used for processing;
the step length from the second layer to the fifth layer, namely the second layer, the third layer, the fourth layer and the fifth layer is 2, the convolution size is 3 multiplied by 3, the user-defined excitation function is used for processing, and the input data of each layer from the second layer is the output result of the previous layer;
the sixth layer of input data is the output result of the fifth layer, and full connection is used without any excitation function;
the last two layers, namely the seventh layer and the eighth layer, are parallel layers, and the calculation result of the sixth layer is used; the seventh layer uses full connection, and the excitation function is sigmol and is used for judging whether the result is a target result; the eighth layer uses full connections, not adapted to the excitation function, for fine tuning the values of the coordinates.
In S1, the custom excitation function is that, let x be the value of the input excitation function, y be the value of the output function, and the formula is:
in said S2.1
The input data of the first layer is a gray scale image 47 multiplied by 1, the depth is 1, the user-defined excitation function is used for processing, and the depth of an output feature image is 16;
the depth of input data is 16, the user-defined excitation function is used for processing, and the depth of an output feature map is 16;
the fourth layer input data depth is 16, processed using the custom excitation function, and the output feature map depth is 32.
The fifth layer of the S2.1 has an input data depth of 32 and an output data depth of 1.
In S2.1, the sixth layer input data depth is 32, and the output data depth is 4.
In said S2.2
The first layer input data is a gray scale map 49 × 49 × 1, and the output result is 47 × 47 × 16;
the second layer input data is 47 × 47 × 16, and the output feature map result is 23 × 23 × 32;
the third layer of input data is 23 × 23 × 32, and the output feature map result is 11 × 11 × 48;
the fourth layer input data is 11 × 11 × 48, and the output feature map result is 5 × 5 × 64;
the fifth layer input data is 5 × 5 × 64, and the output profile result is 2 × 2 × 64.
The sixth layer of input data in S2.2 is 2 multiplied by 64, and the output characteristic diagram result is 1 multiplied by 128; the seventh layer input data is 1 × 128, and the output characteristic diagram is 1 × 1; the eighth layer input data is 1 × 128 and the output profile is 1 × 4.
The eighth layer in S2.2 is used to fine tune four values of coordinates.
The technical solution of the present invention can be further explained as follows:
1. the network adopts a secondary network. The excitation function is a custom function, the first level input size is a gray scale map of 47X47, and the second level network input size is a gray scale map of 49X 49. Defining an excitation function by self, setting x as the value of an input excitation function and y as the value of an output function, and adopting the formula as follows:
because the calculated results of each layer need to be quantized to 4 bits, through theoretical analysis and various experimental tests, the excitation function is fixed between 0 and 4, and better accuracy can be obtained. Since the boundary data is quantized, when x <0.1, y is 0.
2. A two-level network architecture.
1) A first level network. The first three layers use a step size of 2 and the fourth layer uses a step size of 1, the four layers use custom excitation functions, and each layer uses a convolution kernel of 3 × 3.
The first layer of input data is a gray scale map 47 × 47 × 1 with a depth of 1, processed using a custom excitation function, and the output feature map depth is 16.
The second layer and the third layer have the input data depth of 16, are processed by using the custom excitation function, and have the output feature map depth of 16. The fourth layer input data depth is 16, processed using the custom excitation function, and the output feature map depth is 32.
The last two layers, i.e. the fifth and sixth layers, are parallel layers, using the calculation of the fourth layer. The fifth layer of calculation results is used for judging whether the vehicle is a value of the vehicle, and the used excitation function is sigmol, which is determined by the specificity of classification, otherwise, great precision loss is caused. The layer input data depth is 32 and the output data depth is 1.
The result of the sixth layer of calculations is to fine tune the values of the vehicle detection box without using any excitation function processing. The input data depth is 32 and the output data depth is 4.
Each layer is convoluted by using a non-alignment processing mode, and all calculation results can be effectively utilized by non-alignment and the design of the network, so that the waste of the calculation results is not caused.
The whole network structure needs to be considered from the whole calculated amount and practical application, not only the actual precision is met, but also the calculated amount is less, a first-level network is designed, the calculated amount of the network is small, and the precision is not reduced. The network flow diagram is shown in figure 2 of the network architecture of the drawings.
2) A second level network.
The first layer input data is the grayscale map 49 × 49 × 1, the step size is 1, the convolution size is 3 × 3, and the output result is 47 × 47 × 16 using custom excitation function processing.
The second layer input data is 47 × 47 × 16, the step size is 2, the convolution size is 3 × 3, and the output profile result is 23 × 23 × 32 using custom excitation function processing.
The third layer of input data is 23 × 23 × 32, the step size is 2, the convolution size is 3 × 3, and the output profile result is 11 × 11 × 48 using custom excitation function processing.
The fourth layer input data is 11 × 11 × 48, the step size is 2, the convolution size is 3 × 3, and the output profile result is 5 × 5 × 64 using custom excitation function processing.
The fifth layer input data is 5 × 5 × 64, the step size is 2, the convolution size is 3 × 3, the result of the output feature map is 2 × 2 × 64 by using custom excitation function processing.
The sixth layer input data is 2 × 2 × 64, with full concatenation, without any excitation function, and the output profile result is 1 × 128. The seventh and eighth layers are parallel layers.
In the seventh layer, the input data is 1 × 128, the excitation function is sigmol using full concatenation, and the output characteristic map is 1 × 1, which is a result of determining whether or not the target is present.
In the eighth layer, the input data is 1 × 128, full concatenation is used, no excitation function is used, and the input feature map is 1 × 4, which is four values for fine tuning coordinates. The network flow diagram is shown in figure 3 of the network architecture of the drawings.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the present invention, and various modifications and changes may be made to the embodiment of the present invention by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (8)
1. A method for designing a quantifiable front-end vehicle detection network structure, the method comprising the steps of:
s1, designing a network to adopt a two-stage network, wherein an excitation function is a self-defined function, the first-stage input size is a gray scale image of 47X47, and the second-stage network input size is a gray scale image of 49X 49; defining an excitation function by self, quantizing the calculated results of each layer to 4 bits, and quantizing the boundary data;
s2, designing a secondary network structure:
s2.1, a first-level network:
the step size used by the first three layers is 2, the step size used by the fourth layer is 1, the four layers use custom excitation functions, and each layer uses a convolution kernel of 3 multiplied by 3;
the last two layers, i.e. the fifth layer and the sixth layer, are parallel layers, and the calculation results of the fourth layer are used; the fifth layer calculation result is used for judging whether the vehicle is a value of the vehicle, and the used excitation function is sigmol; the result of the sixth layer of calculation is used for fine tuning the value of the vehicle detection frame without any excitation function processing;
performing convolution on each layer by using a non-alignment processing mode;
s2.2, a second-level network:
the step length used by the first layer is 1, the convolution size is 3 multiplied by 3, and the user-defined excitation function is used for processing;
the step length from the second layer to the fifth layer, namely the second layer, the third layer, the fourth layer and the fifth layer is 2, the convolution size is 3 multiplied by 3, the user-defined excitation function is used for processing, and the input data of each layer from the second layer is the output result of the previous layer;
the sixth layer of input data is the output result of the fifth layer, and full connection is used without any excitation function;
the last two layers, namely the seventh layer and the eighth layer, are parallel layers, and the calculation result of the sixth layer is used; the seventh layer uses full connection, and the excitation function is sigmol and is used for judging whether the result is a target result; the eighth layer uses full connectivity, without using an excitation function, for fine tuning the values of the coordinates.
3. the method of claim 1, wherein the step S2.1 is performed by a quantifiable design method for a front-end vehicle inspection network structure
The input data of the first layer is a gray scale image 47 multiplied by 1, the depth is 1, the user-defined excitation function is used for processing, and the depth of an output feature image is 16;
the depth of input data is 16, the user-defined excitation function is used for processing, and the depth of an output feature map is 16;
the fourth layer input data depth is 16, processed using the custom excitation function, and the output feature map depth is 32.
4. The method as claimed in claim 1, wherein the fifth layer of the input data depth of S2.1 is 32, and the output data depth is 1.
5. The method of claim 1, wherein the sixth layer of the S2.1 is 32 in input data depth and 4 in output data depth.
6. The method of claim 1, wherein the step S2.2 is performed by a quantifiable design method for a front-end vehicle inspection network structure
The first layer input data is a gray scale map 49 × 49 × 1, and the output result is 47 × 47 × 16;
the second layer input data is 47 × 47 × 16, and the output feature map result is 23 × 23 × 32;
the third layer of input data is 23 × 23 × 32, and the output feature map result is 11 × 11 × 48;
the fourth layer input data is 11 × 11 × 48, and the output feature map result is 5 × 5 × 64;
the fifth layer input data is 5 × 5 × 64, and the output profile result is 2 × 2 × 64.
7. The method of claim 1, wherein the sixth layer of input data in S2.2 is 2 x 64, and the output feature map result is 1 x 128; the seventh layer input data is 1 × 128, and the output characteristic diagram is 1 × 1; the eighth layer input data is 1 × 128 and the output profile is 1 × 4.
8. The method of claim 1, wherein the eighth layer of S2.2 is used for fine tuning four values of coordinates.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010400217.7A CN113673273B (en) | 2020-05-13 | 2020-05-13 | Design method of quantifiable front-end vehicle detection network structure |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010400217.7A CN113673273B (en) | 2020-05-13 | 2020-05-13 | Design method of quantifiable front-end vehicle detection network structure |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113673273A true CN113673273A (en) | 2021-11-19 |
CN113673273B CN113673273B (en) | 2023-05-12 |
Family
ID=78536808
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010400217.7A Active CN113673273B (en) | 2020-05-13 | 2020-05-13 | Design method of quantifiable front-end vehicle detection network structure |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113673273B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018003212A1 (en) * | 2016-06-30 | 2018-01-04 | クラリオン株式会社 | Object detection device and object detection method |
US20180157938A1 (en) * | 2016-12-07 | 2018-06-07 | Samsung Electronics Co., Ltd. | Target detection method and apparatus |
US20190377930A1 (en) * | 2018-06-11 | 2019-12-12 | Zkteco Usa, Llc | Method and System for Face Recognition Via Deep Learning |
CN110569971A (en) * | 2019-09-09 | 2019-12-13 | 吉林大学 | convolutional neural network single-target identification method based on LeakyRelu activation function |
CN110619319A (en) * | 2019-09-27 | 2019-12-27 | 北京紫睛科技有限公司 | Improved MTCNN model-based face detection method and system |
-
2020
- 2020-05-13 CN CN202010400217.7A patent/CN113673273B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018003212A1 (en) * | 2016-06-30 | 2018-01-04 | クラリオン株式会社 | Object detection device and object detection method |
US20180157938A1 (en) * | 2016-12-07 | 2018-06-07 | Samsung Electronics Co., Ltd. | Target detection method and apparatus |
US20190377930A1 (en) * | 2018-06-11 | 2019-12-12 | Zkteco Usa, Llc | Method and System for Face Recognition Via Deep Learning |
CN110569971A (en) * | 2019-09-09 | 2019-12-13 | 吉林大学 | convolutional neural network single-target identification method based on LeakyRelu activation function |
CN110619319A (en) * | 2019-09-27 | 2019-12-27 | 北京紫睛科技有限公司 | Improved MTCNN model-based face detection method and system |
Non-Patent Citations (2)
Title |
---|
包晓安;胡玲玲;张娜;吴彪;桂江生;: "基于级联网络的快速人脸检测算法", 浙江理工大学学报(自然科学版) * |
白博;谢刚;续欣莹;: "多尺度级联R-FCN的尾灯检测算法研究", 计算机工程与应用 * |
Also Published As
Publication number | Publication date |
---|---|
CN113673273B (en) | 2023-05-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111462126B (en) | Semantic image segmentation method and system based on edge enhancement | |
Pan et al. | Scalable vision transformers with hierarchical pooling | |
KR940007346B1 (en) | Edge detection apparatus for image processing system | |
EP3792838A2 (en) | Methods and systems for converting weights of a deep neural network from a first number format to a second number format | |
Zhu et al. | Adaptive layerwise quantization for deep neural network compression | |
CN105550701A (en) | Real-time image extraction and recognition method and device | |
CN113344179B (en) | IP core of binary convolution neural network algorithm based on FPGA | |
CN115147598A (en) | Target detection segmentation method and device, intelligent terminal and storage medium | |
CN110503152A (en) | Two-way neural network training method and image processing method for target detection | |
Chen et al. | Vip: Virtual pooling for accelerating cnn-based image classification and object detection | |
CN113673273A (en) | Design method of quantifiable front-end vehicle detection network structure | |
Reyna et al. | Implementation of the SVM neural network generalization function for image processing | |
Park et al. | GRLC: Grid-based run-length compression for energy-efficient CNN accelerator | |
CN113343861B (en) | Remote sensing image water body region extraction method based on neural network model | |
CN113673271A (en) | Method for calculating secondary loss based on double-layer labeling of pet detection | |
Li et al. | CNN weight parameter quantization method for FPGA | |
CN111190952B (en) | Method for extracting and persistence of multi-scale features of city portrait based on image pyramid | |
CN113743593A (en) | Neural network quantization method, system, storage medium and terminal | |
Boudabous et al. | HW/SW design and FPGA implementation of The GCM for an efficient text extraction from complex images | |
Karafyllidis et al. | Geometrical shape recognition using a cellular automaton architecture and its VLSI implementation | |
Chen et al. | Condensation-Net: memory-efficient network architecture with cross-channel pooling layers and virtual feature maps | |
Keshavarz | A FPGA implementation of neural/wavelet face detection system | |
CN113673668A (en) | Calculation method of secondary loss function in vehicle detection training | |
CN113673667A (en) | Design method of network structure in vehicle detection training | |
CN114005150B (en) | Design method for quantifiable front-end face detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |