CN108038455A - Bionic machine peacock image-recognizing method based on deep learning - Google Patents

Bionic machine peacock image-recognizing method based on deep learning Download PDF

Info

Publication number
CN108038455A
CN108038455A CN201711374581.5A CN201711374581A CN108038455A CN 108038455 A CN108038455 A CN 108038455A CN 201711374581 A CN201711374581 A CN 201711374581A CN 108038455 A CN108038455 A CN 108038455A
Authority
CN
China
Prior art keywords
face
convolutional neural
neural network
sample
deep learning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711374581.5A
Other languages
Chinese (zh)
Inventor
李成荣
胡耀聪
周世久
徐玉龙
李名扬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201711374581.5A priority Critical patent/CN108038455A/en
Publication of CN108038455A publication Critical patent/CN108038455A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/166Detection; Localisation; Normalisation using acquisition arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/008Artificial life, i.e. computing arrangements simulating life based on physical entities controlled by simulated intelligence so as to replicate intelligent life forms, e.g. based on robots replicating pets or humans in their appearance or behaviour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Robotics (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses the image-recognizing method of the bionic machine peacock based on deep learning, comprise the following steps:Image data set of the Face datection database disclosed in collection as training and verification;The deep learning framework based on convolutional neural networks is designed, Face datection function is realized in deep learning framework;Image scene taken by collection bionic machine peacock camera is finely adjusted trained convolutional Neural net net, realizes the Face datection function under indoor complex environment;Empirical parameter is obtained to determine that the dressing of spectators positions, and counts the corresponding proportion shared by a variety of colors;The present invention realizes the Face datection and colour recognition of amusement bio-robot precise and high efficiency under complex environment, and robustness is good;And small parameter perturbations are carried out to trained deep learning framework to image scene;Finally realize that the image scene captured for camera carries out real-time face detection and dressing identification;Can be applied to science and technology center, hotel, market, for tourist visit, amusement.

Description

Bionic machine peacock image recognition method based on deep learning
Technical Field
The invention relates to the technical field of computer identification, in particular to a bionic machine peacock image identification method based on deep learning.
Background
The bionic robot is a combined product of bionics and application requirements in the robot field. From the perspective of a robot, a bionic robot is a high-level stage of robot development, and biological characteristics provide many beneficial references for the design of the robot. The existing bionic robots are various in types, such as bionic robot fish and bionic robot dog. And is widely used in military and industrial fields, but is currently less applicable as an entertainment robot. The visual system is an important component of the bionic robot, and is equivalent to the 'eyes' of the bionic robot. The system generally captures surrounding environment information through a high-definition camera arranged on a robot, and analyzes and processes the captured image through the existing computer vision algorithm to realize the motion control of a robot actuating mechanism, such as underwater fishing of robot fish, gait planning of robot dogs and the like. At present, a vision system of a bionic robot depends on a traditional image processing algorithm, generally, image features of a current environment are firstly extracted, then the extracted features are described, and finally the current environment is identified, detected and other related tasks are carried out through feature description. Although a specific image feature has a good effect in the field of the bionic robot, the traditional image feature depends on manual experience, and in a complex environment, the feature extraction complexity is high and the robustness is low.
In view of the above, it is urgently needed to provide a bionic machine peacock visual identification method based on deep learning, which has good robustness and high real-time performance, is applied to places such as science and technology museums, superstores, hotels and the like, and teaches through lively activities.
Disclosure of Invention
In order to solve the technical problem, the technical scheme adopted by the invention is to provide an image recognition method of a bionic machine peacock based on deep learning, which comprises the following steps:
s1, collecting a public face detection database as a training and verification image data set;
s2, designing a deep learning framework based on a convolutional neural network, and realizing a face detection function in the deep learning framework;
s3, collecting the field image shot by the peacock camera of the bionic machine to finely adjust the trained convolutional neural network;
s4, testing according to the fine-tuned deep learning framework in the step S3, and achieving a face detection function in an indoor complex environment;
and S5, obtaining empirical parameters to determine the dressing position of the audience according to the position relation between the face frame position obtained in the step S4 and the position relation between the camera and the audience, and counting the corresponding proportion occupied by each color.
In the above method, the step S1 specifically includes the following steps:
s11, selecting a public wire _ face data set and a Celeba _ face data set as training samples for face detection by a face detection database; normalizing the original image in the image data set to a uniform size;
a large amount of face detection data are provided in the widget _ face data set and the Celeba _ face data set, and position information of a face frame is provided in a picture;
s12, randomly selecting a frame for the face image, and calculating the repetition degree IOU of the selected frame and the real frame;
s13, dividing the face detection data into three types, namely a face positive sample, a face negative sample and a face partial sample, wherein the proportion of the face positive sample, the face negative sample and the face partial sample is 1: 3: 1;
and S14, generating a file path from the training sample, and making labels corresponding to the file path, wherein the label content comprises positive and negative sample labels and offset values of a randomly generated frame and a real frame.
In the above method, in step S12, IOU > 0.65 is a positive face sample, and IOU < 0.4 is a negative face sample; the face part samples are between 0.4 IOU 0.65.
In the above method, the deep learning architecture is specifically as follows:
the deep learning framework is formed by cascading three convolutional neural networks, wherein the three convolutional neural networks are a first convolutional neural network PNet, a second convolutional neural network RNet and a third convolutional neural network ONet respectively;
the first convolution neural network PNet inputs image blocks with the size of 12 × 3 in the training stage, and the first convolution neural network PNet is a three-layer full convolution network and does not comprise a full connection layer;
the second convolutional neural network RNet inputs image blocks with the size of 24 × 3, and is a four-layer convolutional network comprising three convolutional layers and a full-connection layer;
image blocks of input size 48 x 3 of the third convolutional neural network ONet; the four-layer convolution network comprises three convolution layers and a full connection layer;
firstly, a training sample is sent to a first convolutional neural network PNet, and the output of the training sample is the type of the sample, namely a human face positive sample, a human face negative sample and a human face positioning predicted value;
the second convolutional neural network RNet is used to refine the output result of the PNet network;
the third convolutional neural network ONet is used for refining the output result of the RNet network and outputting the sample category, the predicted value of the face positioning and the predicted value of the face key point.
In the above method, the step S3 specifically includes the following steps:
s31, selecting a plurality of images including audiences and not including audiences collected on the scene of the bionic machine peacock to be divided into training samples and testing samples; selecting images in the training samples and taking random frames to obtain face positive samples, face negative samples and face part samples;
and S32, sending the determined training samples and labels to a deep learning framework, and finely adjusting the parameters of three convolutional neural networks in the deep learning framework.
In the above method, the step S4 specifically includes the following steps:
s41, sending the test sample into a first convolutional neural network PNet, and acquiring a candidate face frame and four point coordinates corresponding to each face;
s42, sending the candidate face frame obtained and output by the first convolutional neural network PNet into the second convolutional neural network RNet, carrying out thinning screening on the candidate face frame, carrying out non-maximum value suppression, and removing false detection;
and S43, screening the second convolutional neural network RNet and outputting a result to a third convolutional neural network ONet for further carrying out refinement screening on the candidate face frame, removing false detection and finally obtaining a face detection result, wherein the face detection result comprises the face frame positioning and the position relation between the camera and the audience.
In the above method, the step S5 specifically includes the following steps:
s51, according to multiple times of measurement and calibration of the camera, selecting the deviation of the dressing candidate frame with the positioning of 1.8 personal face frames in the image, wherein the size of the dressing candidate frame is 2 personal face frames;
s52, setting colors participating in the statistics of the proportion to be red, yellow, blue, green, purple, white, black and gray, and counting the corresponding proportion occupied by each color according to the selected dressing candidate frame.
The invention provides an image recognition method of a bionic robot peacock based on deep learning, aiming at realizing accurate and efficient face detection and color recognition of an entertainment bionic robot in a complex environment and having good robustness; performing parameter fine adjustment on the trained deep learning architecture by using the field image acquired by the bionic machine peacock; finally, real-time face detection and dressing identification are carried out on the field image captured by the camera; can be applied to science and technology museums, hotels, shopping malls, visitors and entertainment.
Drawings
FIG. 1 is a flow chart provided by the present invention;
FIG. 2 is a schematic structural diagram of a deep learning architecture according to the present invention; wherein,
(a) a first convolutional neural network PNet; (b) is a second convolutional neural network RNet; (c) is a third convolutional neural network ONet;
FIG. 3 is a block diagram of an image recognition process of a bionic machine peacock based on deep learning provided by the invention;
fig. 4 is an exemplary diagram of an image acquired by the camera and a statistical color proportion.
Detailed Description
The invention is described in detail below with reference to specific embodiments and the accompanying drawings.
As shown in fig. 1-3, the invention provides an image recognition method of a bionic machine peacock based on deep learning, which comprises the following steps:
s1, collecting the public face detection database as the training and verification image data set. Wherein, step S1 specifically includes the following steps:
s11, selecting a public wire _ face data set and a Celeba _ face data set as training samples for face detection by a face detection database. The widget _ face data set and the Celeba _ face data set provide a large amount of face detection data, position information of a face frame is provided in a figure, and more than 20 ten thousand face detection data can be obtained in total.
S12, randomly selecting a frame for the face image, and calculating the repetition degree IOU (intersection ratio) of the selected frame and the real frame. Defining the face positive sample with IOU more than 0.65 and the face negative sample with IOU less than 0.4; the face part samples are between 0.4 IOU 0.65.
S13, dividing the face detection data into three types, namely a face positive sample, a face negative sample and a face partial sample, wherein the proportion of the face positive sample, the face negative sample and the face partial sample is 1: 3: 1.
and S14, generating a file path from the training sample, and making labels corresponding to the file path, wherein the label content comprises positive and negative sample labels and offset values of a randomly generated frame and a real frame. The raw images in the image dataset are normalized to a uniform size, such as 800 x 600 x 3. And labels each image having a face. For example, 0001.png 204301264390 indicates that the image 0001.png contains a human face in a rectangular pixel region (204, 301, 264, 390), all labels are written in a txt file, and the image paths and the labels are in one-to-one correspondence.
S2, designing a deep learning framework based on the convolutional neural network, and realizing the face detection function in the deep learning framework. The deep learning architecture is specifically as follows:
as shown in fig. 2, the deep learning architecture is composed of three convolutional neural networks in cascade, the three convolutional neural networks are named as a first convolutional neural network PNet, a second convolutional neural network RNet and a third convolutional neural network ONet respectively; the first convolutional neural network PNet inputs image blocks of size 12 × 3 in the training phase, the second convolutional neural network RNet inputs image blocks of size 24 × 3, and the third convolutional neural network ONet inputs image blocks of size 48 × 3.
Firstly, a training sample is sent to a first convolution neural network PNet, the first convolution neural network PNet is a three-layer full convolution network and does not comprise a full connection layer, and the output of the first convolution neural network PNet is respectively the type (positive and negative samples) of the sample and the predicted value of face positioning.
The second convolutional neural network RNet is a four-layer convolutional network comprising three convolutional layers and a full-link layer. This network can be used to refine the output results of the PNet network.
The third convolutional neural network ONet is a four-layer convolutional network comprising three convolutional layers and a full-link layer. The network can be used to refine the output results of the RNet network and output predicted values of the classes of the samples (positive and negative samples), the face location and the face key points respectively.
S3, acquiring the field image shot by the peacock camera of the bionic machine, and finely adjusting the trained deep learning architecture. The captured live images include scenes with and without viewers. Wherein, step S3 specifically includes the following steps:
and S31, selecting 10 million images including audiences and not including audiences collected by the bionic machine peacock scene to be divided into training samples and testing samples. Selecting images in the training sample and taking a random frame to obtain a face positive sample, a face negative sample and a face partial sample, specifically including steps S12-S14.
And S32, sending the determined training samples and labels to a deep learning framework, and finely adjusting the parameters of three convolutional neural networks in the deep learning framework.
And S4, testing according to the fine-tuned deep learning framework in the step S3, and realizing the face detection function in the indoor complex environment. Wherein, step S4 specifically includes the following steps:
s41, the test sample is firstly sent to a first convolutional neural network PNet, and a candidate face frame and four point coordinates corresponding to each face are obtained.
And S42, sending the candidate face frame acquired and output by the first convolutional neural network PNet into the second convolutional neural network RNet, carrying out thinning screening on the candidate face frame, carrying out non-maximum value suppression, and removing false detection.
And S43, screening the second convolutional neural network RNet, outputting the result to a third convolutional neural network ONet, wherein the operation of the second convolutional neural network RNet is the same as that of the second convolutional neural network RNet, and the second convolutional neural network RNet is used for further carrying out thinning screening on the candidate face frame, removing false detection and finally obtaining a face detection result which comprises the face frame positioning and the position relation (the face key point position) of the camera to the audience.
The number of layers of the three convolution network cascades is gradually deepened, the size of a convolution kernel is gradually increased, and the detection of the human face is also gradually refined. The first convolutional neural network PNet can obtain a rough face candidate frame, and the second convolutional neural network RNet and the third convolutional neural network ONet gradually screen the candidate face frame obtained by the first convolutional neural network PNet and perform non-maximum suppression to obtain a final face detection result.
The deep learning architecture adopts a multi-task learning strategy to optimize the network and consists of three paths of supervised learning signals: A. and learning whether the input image block is a human face or not by using the cross loss entropy of the softmax classifier. B. The euclidean regression loss of the rectangular box learns the location of the rectangular box. C. Regression loss learns the locations of five face key points.
The construction of the deep learning architecture specifically comprises the following steps:
the whole network is trained by using CAFFE open source codes, an Intel core I7 processor is configured by using a computer, and a graphics card is configured as a NIVIDA GTX TITAN 1080TI GPU. The operating system is Linux 16.04.
The learning parameters of the network are: the iteration times are 100000 times, the sample number sent in each iteration is 100, and the initial learning rate: 0.01, impulse term: 0.9, counter-propagating mode: random gradient descent method.
And S5, obtaining empirical parameters to determine the dressing position of the audience according to the position relation between the face frame position obtained in the step S4 and the position relation between the camera and the audience, and counting the corresponding proportion occupied by each color. Wherein, step S5 specifically includes the following steps:
s51, according to multiple measurements and calibrations by the camera, selecting a deviation of the dressing candidate frame with a location of 1.8 individual face frames in the image, and a size of the dressing candidate frame is 2 individual face frames, as shown in fig. 4, which is an exemplary diagram of the image collected by the camera and the statistical color proportion provided by the present invention. The distance between the installation of the camera of the bionic machine peacock and the field audience is set to be 2.5 meters.
S52, setting colors participating in the statistics of the proportion to be red, yellow, blue, green, purple, white, black and gray, and counting the corresponding proportion occupied by each color according to the selected dressing candidate frame.
The bionic machine peacock provided by the invention makes decisions on whether to spread the screen or not and the size of the spread screen according to the captured visual information (the number of audience on site and the specific weight of dressing color) so as to realize entertainment.
The construction of the whole bionic machine peacock image recognition system is as follows:
the method is characterized in that a Logitech Pro C920 camera is mounted at the head of the bionic machinery peacock and used as an 'eye' of the bionic machinery peacock to capture a scene environment picture, wherein the distance between the 'eye' and an audience is controlled to be about 2.5 meters.
The MFC interface program is written using a Visual Stdio 2013 development environment. And putting the trained model file and the network script file in the CAFFE into the whole project to realize the functions of face detection and dressing identification of the bionic machine peacock.
The codes with the image recognition function are packaged into an SDK library form, so that other function modules of the bionic machine peacock can be called mutually conveniently, such as a motion execution module, a voice conversation module, a face recognition module, a color screening module and the like, the whole function of the bionic machine peacock is tested, and the bionic machine peacock is put into a science and technology museum, a hotel or a market to play the entertainment of the bionic machine peacock.
The invention provides an image recognition method of a bionic robot peacock based on deep learning, and aims to realize accurate and efficient face detection and color recognition of an entertainment bionic robot in a complex environment and to achieve good robustness. First, a face detection library is used to make pre-training samples of the convolutional neural network. Then, designing a structure of the convolutional neural network, wherein the structure comprises the number of layers of a network convolutional layer, a pooling layer and a full-connection layer, parameters of each layer, such as the number and the size of convolutional cores and the number of neurons output by the full-connection layer, and compiling a network structure and a trained script program on a deep learning architecture CAFFE. And (3) carrying out parameter fine adjustment on the trained deep learning architecture by using the live images (including audiences and non-audiences) acquired by the bionic machine peacock. And finally, writing a test program to realize real-time face detection and dressing identification on the field image captured by the camera.
The present invention is not limited to the above-mentioned preferred embodiments, and any structural changes made under the teaching of the present invention shall fall within the protection scope of the present invention, which has the same or similar technical solutions as the present invention.

Claims (7)

1. The image recognition method of the bionic machine peacock based on deep learning is characterized by comprising the following steps:
s1, collecting a public face detection database as a training and verification image data set;
s2, designing a deep learning framework based on a convolutional neural network, and realizing a face detection function in the deep learning framework;
s3, collecting the field image shot by the peacock camera of the bionic machine to finely adjust the trained convolutional neural network;
s4, testing according to the fine-tuned deep learning framework in the step S3, and achieving a face detection function in an indoor complex environment;
and S5, obtaining empirical parameters to determine the dressing position of the audience according to the position relation between the face frame position obtained in the step S4 and the position relation between the camera and the audience, and counting the corresponding proportion occupied by each color.
2. The image recognition method according to claim 1, wherein the step S1 specifically includes the steps of:
s11, selecting a public wire _ face data set and a Celeba _ face data set as training samples for face detection by a face detection database; normalizing the original image in the image data set to a uniform size;
a large amount of face detection data are provided in the widget _ face data set and the Celeba _ face data set, and position information of a face frame is provided in a picture;
s12, randomly selecting a frame for the face image, and calculating the repetition degree IOU of the selected frame and the real frame;
s13, dividing the face detection data into three types, namely a face positive sample, a face negative sample and a face partial sample, wherein the proportion of the face positive sample, the face negative sample and the face partial sample is 1: 3: 1;
and S14, generating a file path from the training sample, and making labels corresponding to the file path, wherein the label content comprises positive and negative sample labels and offset values of a randomly generated frame and a real frame.
3. The image recognition method according to claim 2, wherein in step S12, IOU > 0.65 is a positive face sample, and IOU < 0.4 is a negative face sample; the face part samples are between 0.4 IOU 0.65.
4. The image recognition method of claim 1, wherein the deep learning architecture is specifically as follows:
the deep learning framework is formed by cascading three convolutional neural networks, wherein the three convolutional neural networks are a first convolutional neural network PNet, a second convolutional neural network RNet and a third convolutional neural network ONet respectively;
the first convolution neural network PNet inputs image blocks with the size of 12 × 3 in the training stage, and the first convolution neural network PNet is a three-layer full convolution network and does not comprise a full connection layer;
the second convolutional neural network RNet inputs image blocks with the size of 24 × 3, and is a four-layer convolutional network comprising three convolutional layers and a full-connection layer;
image blocks of input size 48 x 3 of the third convolutional neural network ONet; the four-layer convolution network comprises three convolution layers and a full connection layer;
firstly, a training sample is sent to a first convolutional neural network PNet, and the output of the training sample is the type of the sample, namely a human face positive sample, a human face negative sample and a human face positioning predicted value;
the second convolutional neural network RNet is used to refine the output result of the PNet network;
the third convolutional neural network ONet is used for refining the output result of the RNet network and outputting the sample category, the predicted value of the face positioning and the predicted value of the face key point.
5. The image recognition method according to claim 4, wherein the step S3 specifically includes the steps of:
s31, selecting a plurality of images including audiences and not including audiences collected on the scene of the bionic machine peacock to be divided into training samples and testing samples; selecting images in the training samples and taking random frames to obtain face positive samples, face negative samples and face part samples;
and S32, sending the determined training samples and labels to a deep learning framework, and finely adjusting the parameters of three convolutional neural networks in the deep learning framework.
6. The image recognition method according to claim 5, wherein the step S4 specifically includes the steps of:
s41, sending the test sample into a first convolutional neural network PNet, and acquiring a candidate face frame and four point coordinates corresponding to each face;
s42, sending the candidate face frame obtained and output by the first convolutional neural network PNet into the second convolutional neural network RNet, carrying out thinning screening on the candidate face frame, carrying out non-maximum value suppression, and removing false detection;
and S43, screening the second convolutional neural network RNet and outputting a result to a third convolutional neural network ONet for further carrying out refinement screening on the candidate face frame, removing false detection and finally obtaining a face detection result, wherein the face detection result comprises the face frame positioning and the position relation between the camera and the audience.
7. The image recognition method according to claim 1, wherein the step S5 specifically includes the steps of:
s51, according to multiple times of measurement and calibration of the camera, selecting the deviation of the dressing candidate frame with the positioning of 1.8 personal face frames in the image, wherein the size of the dressing candidate frame is 2 personal face frames;
s52, setting colors participating in the statistics of the proportion to be red, yellow, blue, green, purple, white, black and gray, and counting the corresponding proportion occupied by each color according to the selected dressing candidate frame.
CN201711374581.5A 2017-12-19 2017-12-19 Bionic machine peacock image-recognizing method based on deep learning Pending CN108038455A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711374581.5A CN108038455A (en) 2017-12-19 2017-12-19 Bionic machine peacock image-recognizing method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711374581.5A CN108038455A (en) 2017-12-19 2017-12-19 Bionic machine peacock image-recognizing method based on deep learning

Publications (1)

Publication Number Publication Date
CN108038455A true CN108038455A (en) 2018-05-15

Family

ID=62099866

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711374581.5A Pending CN108038455A (en) 2017-12-19 2017-12-19 Bionic machine peacock image-recognizing method based on deep learning

Country Status (1)

Country Link
CN (1) CN108038455A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108985147A (en) * 2018-05-31 2018-12-11 成都通甲优博科技有限责任公司 Object detection method and device
CN109035221A (en) * 2018-07-10 2018-12-18 成都先进金属材料产业技术研究院有限公司 Macrostructure intelligence ranking method based on image recognition depth learning technology
CN109685018A (en) * 2018-12-26 2019-04-26 深圳市捷顺科技实业股份有限公司 A kind of testimony of a witness method of calibration, system and relevant device
CN110119738A (en) * 2019-05-21 2019-08-13 中电健康云科技有限公司 Indentation detection method, device, electronic equipment and readable storage medium storing program for executing
CN110222623A (en) * 2019-05-31 2019-09-10 深圳市恩钛控股有限公司 Micro- expression analysis method and system
CN110458025A (en) * 2019-07-11 2019-11-15 南京邮电大学 A kind of personal identification and localization method based on binocular camera
CN110458005A (en) * 2019-07-02 2019-11-15 重庆邮电大学 It is a kind of based on the progressive invariable rotary method for detecting human face with pseudo-crystalline lattice of multitask
CN111783653A (en) * 2020-07-01 2020-10-16 创新奇智(西安)科技有限公司 Trademark label detection method, detection model, device, equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070013791A1 (en) * 2005-07-05 2007-01-18 Koichi Kinoshita Tracking apparatus
CN105809198A (en) * 2016-03-10 2016-07-27 西安电子科技大学 SAR image target recognition method based on deep belief network
CN105930796A (en) * 2016-04-21 2016-09-07 中国人民解放军信息工程大学 Single-sample face image recognition method based on depth self-encoder
CN106022220A (en) * 2016-05-09 2016-10-12 西安北升信息科技有限公司 Method for performing multi-face tracking on participating athletes in sports video
CN106658169A (en) * 2016-12-18 2017-05-10 北京工业大学 Universal method for segmenting video news in multi-layered manner based on deep learning
CN107368670A (en) * 2017-06-07 2017-11-21 万香波 Stomach cancer pathology diagnostic support system and method based on big data deep learning

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070013791A1 (en) * 2005-07-05 2007-01-18 Koichi Kinoshita Tracking apparatus
CN105809198A (en) * 2016-03-10 2016-07-27 西安电子科技大学 SAR image target recognition method based on deep belief network
CN105930796A (en) * 2016-04-21 2016-09-07 中国人民解放军信息工程大学 Single-sample face image recognition method based on depth self-encoder
CN106022220A (en) * 2016-05-09 2016-10-12 西安北升信息科技有限公司 Method for performing multi-face tracking on participating athletes in sports video
CN106658169A (en) * 2016-12-18 2017-05-10 北京工业大学 Universal method for segmenting video news in multi-layered manner based on deep learning
CN107368670A (en) * 2017-06-07 2017-11-21 万香波 Stomach cancer pathology diagnostic support system and method based on big data deep learning

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108985147A (en) * 2018-05-31 2018-12-11 成都通甲优博科技有限责任公司 Object detection method and device
CN109035221A (en) * 2018-07-10 2018-12-18 成都先进金属材料产业技术研究院有限公司 Macrostructure intelligence ranking method based on image recognition depth learning technology
CN109685018A (en) * 2018-12-26 2019-04-26 深圳市捷顺科技实业股份有限公司 A kind of testimony of a witness method of calibration, system and relevant device
CN110119738A (en) * 2019-05-21 2019-08-13 中电健康云科技有限公司 Indentation detection method, device, electronic equipment and readable storage medium storing program for executing
CN110222623A (en) * 2019-05-31 2019-09-10 深圳市恩钛控股有限公司 Micro- expression analysis method and system
CN110458005A (en) * 2019-07-02 2019-11-15 重庆邮电大学 It is a kind of based on the progressive invariable rotary method for detecting human face with pseudo-crystalline lattice of multitask
CN110458005B (en) * 2019-07-02 2022-12-27 重庆邮电大学 Rotation-invariant face detection method based on multitask progressive registration network
CN110458025A (en) * 2019-07-11 2019-11-15 南京邮电大学 A kind of personal identification and localization method based on binocular camera
CN110458025B (en) * 2019-07-11 2022-10-14 南京邮电大学 Target identification and positioning method based on binocular camera
CN111783653A (en) * 2020-07-01 2020-10-16 创新奇智(西安)科技有限公司 Trademark label detection method, detection model, device, equipment and storage medium
CN111783653B (en) * 2020-07-01 2024-04-26 创新奇智(西安)科技有限公司 Trademark label detection method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN108038455A (en) Bionic machine peacock image-recognizing method based on deep learning
CN111310862B (en) Image enhancement-based deep neural network license plate positioning method in complex environment
Tran et al. Two-stream flow-guided convolutional attention networks for action recognition
CN110414432A (en) Training method, object identifying method and the corresponding device of Object identifying model
CN108875708A (en) Behavior analysis method, device, equipment, system and storage medium based on video
CN109919032B (en) Video abnormal behavior detection method based on motion prediction
CN106127702A (en) A kind of image mist elimination algorithm based on degree of depth study
CN108780519A (en) Structure learning in convolutional neural networks
JP2019512827A (en) System and method for training an object classifier by machine learning
CN107330364A (en) A kind of people counting method and system based on cGAN networks
CN111898406A (en) Face detection method based on focus loss and multitask cascade
Li et al. Visual stability prediction and its application to manipulation
CN106951870A (en) The notable event intelligent detecting prewarning method of monitor video that active vision notes
CN106650617A (en) Pedestrian abnormity identification method based on probabilistic latent semantic analysis
CN110008961A (en) Text real-time identification method, device, computer equipment and storage medium
CN105007525A (en) Interactive situation event correlation smart perception method based on application of smart television
CN108764447A (en) A kind of group robot Majiang game intelligence dynamicization system and mahjong identification learning algorithm
CN113988179A (en) Target segmentation method, system and equipment based on improved attention and loss function
CN109086664A (en) A kind of polymorphic gesture identification method of sound state fusion
CN107247952A (en) The vision significance detection method for the cyclic convolution neutral net supervised based on deep layer
Wasim et al. A novel deep learning based automated academic activities recognition in cyber-physical systems
Wang et al. Single shot multibox detector with deconvolutional region magnification procedure
CN117094895A (en) Image panorama stitching method and system
Vrskova et al. Education of video classification based by neural networks
CN111898492A (en) Intelligent campus study room monitoring and management system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20180515

RJ01 Rejection of invention patent application after publication