CN108491073A - A kind of good man-machine interactive system of interaction effect - Google Patents
A kind of good man-machine interactive system of interaction effect Download PDFInfo
- Publication number
- CN108491073A CN108491073A CN201810185554.1A CN201810185554A CN108491073A CN 108491073 A CN108491073 A CN 108491073A CN 201810185554 A CN201810185554 A CN 201810185554A CN 108491073 A CN108491073 A CN 108491073A
- Authority
- CN
- China
- Prior art keywords
- image
- children
- subsystem
- pseudo
- fusion
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 230000002452 interceptive effect Effects 0.000 title claims abstract description 33
- 230000000694 effects Effects 0.000 title claims abstract description 19
- 230000003993 interaction Effects 0.000 title claims abstract description 18
- 238000012549 training Methods 0.000 claims abstract description 40
- 238000000034 method Methods 0.000 claims abstract description 10
- 230000004044 response Effects 0.000 claims abstract description 9
- 230000008569 process Effects 0.000 claims abstract description 8
- 230000004927 fusion Effects 0.000 claims description 32
- 238000002156 mixing Methods 0.000 claims description 23
- 239000003086 colorant Substances 0.000 claims description 21
- 238000012545 processing Methods 0.000 claims description 19
- 238000005070 sampling Methods 0.000 claims description 6
- 238000000701 chemical imaging Methods 0.000 claims description 3
- 230000009466 transformation Effects 0.000 claims description 3
- 238000006243 chemical reaction Methods 0.000 claims description 2
- 238000000354 decomposition reaction Methods 0.000 claims description 2
- 230000010354 integration Effects 0.000 claims description 2
- 235000013399 edible fruits Nutrition 0.000 claims 1
- 230000006399 behavior Effects 0.000 abstract description 7
- 230000009286 beneficial effect Effects 0.000 abstract description 3
- 230000000007 visual effect Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 230000037081 physical activity Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000003542 behavioural effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/20—Education
- G06Q50/205—Education administration or guidance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/011—Emotion or mood input determined on the basis of sensed human body parameters such as pulse, heart rate or beat, temperature of skin, facial expressions, iris, voice pitch, brain activity patterns
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Tourism & Hospitality (AREA)
- Health & Medical Sciences (AREA)
- Strategic Management (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Economics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- General Business, Economics & Management (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The present invention provides a kind of good man-machine interactive systems of interaction effect, including authentication subsystem, training subsystem, data process subsystem and interactive subsystem, the authentication subsystem is verified for treating trained children's identity, the trained subsystem is used for being identified by the human body behavior of authentication children, obtain Human bodys' response result, the data process subsystem is using Human bodys' response result as children's physical training basic data, determine children's physical training situation, and children's physical training situation is sent to interactive subsystem, the interactive subsystem is used to play children's physical training instructional video according to children's physical training situation.Beneficial effects of the present invention are:A kind of good man-machine interactive system of interaction effect is provided, which realizes effective training to children, and children can obtain instructional video by interactive subsystem, help to create good Learning atmosphere, improve the training enthusiasm of children.
Description
Technical field
The present invention relates to child teaching technical fields, and in particular to a kind of good man-machine interactive system of interaction effect.
Background technology
With the improvement of living standards, people increasingly focus on children's health, carrying out physical training to children seems particularly
It is important.How to allow children warm to training, how to be assessed training condition is all the problem of pendulum is in face of people.
Human bodys' response is an emerging research direction in artificial intelligence field, be with a wide range of applications with it is non-
The economic value of Chang Keguan, the application field being related to include mainly:Video monitoring, medical diagnosis and monitoring, motion analysis, intelligence
Human-computer interaction, virtual reality etc..The corresponding groundwork flow of Human bodys' response is:Various kinds of sensors is selected to obtain human body row
For data information, and the behavioral trait of sensor characteristics and people is combined to establish rational behavior model, on this basis from original
Extracted in gathered data to behavior type have stronger descriptive power feature, and using suitable method to these features into
Row training, and then realize the pattern-recognition to human body behavior.The image preprocessing of high quality is the key that Activity recognition research, existing
The ineffective very big reason of somebody's body Activity recognition is not obtain the image of high quality.
Invention content
In view of the above-mentioned problems, the present invention is intended to provide a kind of good man-machine interactive system of interaction effect.
The purpose of the present invention is realized using following technical scheme:
Provide a kind of good man-machine interactive system of interaction effect, including authentication subsystem, training subsystem, number
According to processing subsystem and interactive subsystem, the authentication subsystem is verified for treating trained children's identity, institute
Trained subsystem is stated for being identified by the human body behavior of authentication children, obtaining Human bodys' response as a result, institute
Data process subsystem is stated using Human bodys' response result as children's physical training basic data, determines children's physical training feelings
Condition, and children's physical training situation is sent to interactive subsystem, the interactive subsystem is used for according to children's physical training feelings
Condition plays children's physical training instructional video.
Beneficial effects of the present invention are:A kind of good man-machine interactive system of interaction effect is provided, which realizes
Effective training to children, children can obtain instructional video by interactive subsystem, help to create good Learning atmosphere,
Improve the training enthusiasm of children.
Description of the drawings
Using attached drawing, the invention will be further described, but the embodiment in attached drawing does not constitute any limit to the present invention
System, for those of ordinary skill in the art, without creative efforts, can also obtain according to the following drawings
Other attached drawings.
Fig. 1 is the structural schematic diagram of the present invention;
Reference numeral:
Authentication subsystem 1, training subsystem 2, data process subsystem 3, interactive subsystem 4.
Specific implementation mode
The invention will be further described with the following Examples.
Referring to Fig. 1, a kind of good man-machine interactive system of interaction effect of the present embodiment, including authentication subsystem 1,
Training subsystem 2, data process subsystem 3 and interactive subsystem 4, the authentication subsystem 1 is for treating trained youngster
Virgin identity is verified, and the trained subsystem 2 is used to, to being identified by the human body behavior of authentication children, obtain people
Body Activity recognition as a result, the data process subsystem 3 using Human bodys' response result as children's physical training basic data,
It determines children's physical training situation, and children's physical training situation is sent to interactive subsystem 4, the interactive subsystem 4 is used
According to children's physical training situation broadcasting children's physical training instructional video.
A kind of good man-machine interactive system of interaction effect is present embodiments provided, which realizes to the effective of children
Training, children can obtain instructional video by interactive subsystem, help to create good Learning atmosphere, improve the instruction of children
Practice enthusiasm.
Preferably, the trained subsystem 2 includes first processing module, Second processing module, third processing module, the 4th
Processing module, the first processing module is acquired using visible light, infrared multi-spectral imaging system on human body image, described
Second processing module obtains color fusion image, the third processing for being merged to visible images and infrared image
Module is used to extract human body target profile according to color fusion image, and the fourth processing module is used for according to human body target profile
Human body behavior is identified.
This preferred embodiment trains subsystem 2 to obtain human body image and right using visible light, infrared multi-spectral imaging system
Image carries out fusion treatment, obtains the image of high quality, contributes to the detectivity and tracing property that promote follow-up human body, uses
Color fusion image meets the visual signature of the mankind.
Preferably, the Second processing module includes the first integrated unit, the second integrated unit, third integrated unit, institute
It states the first integrated unit to merge visible images and infrared image in non-down sampling contourlet transformation domain, obtains gray scale and melt
Image is closed, second integrated unit obtains pseudo-colours blending image, the third integrated unit root according to grayscale fusion image
Color fusion image is obtained according to pseudo-colours blending image;
First integrated unit merges visible images and infrared image in non-down sampling contourlet transformation domain,
Specially:Non-down sampling contourlet decomposition is carried out to visible images P and infrared image Q, obtains corresponding sub-band division coefficientWithRLPAnd RLQThe low frequency sub-band coefficient of visible images and infrared image is indicated respectively,WithThe sub-band coefficients in k-th of direction in j-th of scale high-frequency sub-band of visible light and infrared image are indicated respectively;
Low frequency sub-band is merged using following formula: In formula, RLR(x, y) indicates the corresponding low frequency sub-band coefficients of grayscale fusion image R,
Wherein, p indicates that the average gray value of infrared image, HX (x, y) indicate the gray value of pixel (x, y) in infrared image;
High-frequency sub-band is merged using following formula: In formula,Indicate the direction of fusion grayscale fusion image R
Sub-band coefficients, vP(x, y) indicates the side of visible images directional subband coefficient in n × n windows centered on pixel (x, y)
Difference, vQ(x, y) indicates the variance yields of infrared image directional subband coefficient in n × n windows centered on pixel (x, y);
Grayscale fusion image R is reconstructed according to the low frequency sub-band coefficient of grayscale fusion image and high-frequency sub-band coefficient;
The blending image that traditional image interfusion method obtains usually exist target-to-background contrast is relatively low, image more
The deficiencies of fuzzy.This preferred embodiment is improved by carrying out multiple dimensioned, multi-direction fusion to visible images and infrared image
Image co-registration is horizontal, by determining image co-registration mode, can preferably merge the information of the image of different-waveband, the ash of acquisition
It is more abundant to spend blending image details, textural characteristics.
Preferably, second integrated unit obtains pseudo-colours blending image according to grayscale fusion image, specially:Using
Following formula obtains pseudo-colours blending image in YUV color spaces:In formula, Y (x, y),
U (x, y), V (x, y) indicate respectively pseudo-colours blending image YUV color spaces component, R (x, y) indicate visible images and
The grayscale fusion image of infrared image, P (x, y) indicate that visible images, Q (x, y) indicate infrared image;
The third integrated unit obtains color fusion image according to pseudo-colours blending image, specially:By width nature
The color visible image shot under sunshine condition, which is used as, refers to image, and the reference picture is converted into YUV color spaces, root
According to gray average and variance of the reference picture in each channel of YUV color spaces, adjustment pseudo-colours blending image is YUV points corresponding
Magnitude, the pseudo-colours blending image after being adjusted specifically are carried out using following formula: In formula, S
Correspond to reference picture and pseudo-colours blending image, Y respectively with W1(x,y)、U1(x,y)、V1(x, y) indicates the puppet after adjustment respectively
For color fusion image in the component of YUV color spaces, μ and σ indicate the gray average of each Color Channel in YUV color spaces respectively
And variance;By the pseudo-colours blending image after adjustment from YUV color notation conversion space to RGB color, color integration figure is obtained
Picture.
Complementary information between this preferred embodiment organic combination different-waveband, enriches the detailed information of image, makes one
Body target is enhanced, to improve accuracy and robustness to target acquisition and tracking;Meanwhile blending image can be meter
Calculation machine visual analysis provides higher-quality source images;In addition, the color fusion image after color adjusts has Natural color
Color visual effect can improve the degree of fatigue that observer watches the perception of scene, reduction observer video, this is for certain
The Activity recognition application that observer participates in is needed to be of great significance.
Using the good man-machine interactive system of interaction effect of the present invention to children carry out physical training, choose 5 children into
Row experiment, respectively children 1, children 2, children 3, children 4, children 5 unite to children training enthusiasm and training cost
Meter, is compared, generation has the beneficial effect that shown in table compared with children's physical training system:
Children training enthusiasm improves | Training cost reduction | |
Children 1 | 29% | 27% |
Children 2 | 27% | 26% |
Children 3 | 26% | 26% |
Children 4 | 25% | 24% |
Children 5 | 24% | 22% |
Finally it should be noted that the above embodiments are merely illustrative of the technical solutions of the present invention, rather than the present invention is protected
The limitation of range is protected, although being explained in detail to the present invention with reference to preferred embodiment, those skilled in the art answer
Work as understanding, technical scheme of the present invention can be modified or replaced equivalently, without departing from the reality of technical solution of the present invention
Matter and range.
Claims (6)
1. a kind of good man-machine interactive system of interaction effect, which is characterized in that including authentication subsystem, training subsystem
System, data process subsystem and interactive subsystem, the authentication subsystem are tested for treating trained children's identity
Card, the trained subsystem are used to, to being identified by the human body behavior of authentication children, obtain Human bodys' response knot
Fruit, the data process subsystem determine children's physical efficiency using Human bodys' response result as children's physical training basic data
Training, and children's physical training situation is sent to interactive subsystem, the interactive subsystem is used for according to children's physical efficiency
Training plays children's physical training instructional video.
2. the good man-machine interactive system of interaction effect according to claim 1, which is characterized in that the trained subsystem
Including first processing module, Second processing module, third processing module, fourth processing module, the first processing module uses
Visible light, infrared multi-spectral imaging system on human body image are acquired, and the Second processing module is used for visible images
It is merged with infrared image, obtains color fusion image, the third processing module according to color fusion image for extracting
Human body target profile, the fourth processing module is for being identified human body behavior according to human body target profile.
3. the good man-machine interactive system of interaction effect according to claim 2, which is characterized in that the second processing mould
Block includes the first integrated unit, the second integrated unit, third integrated unit, and first integrated unit is in non-down sampling contourlet
Transform domain merges visible images and infrared image, obtains grayscale fusion image, second integrated unit is according to ash
It spends blending image and obtains pseudo-colours blending image, the third integrated unit obtains color integration figure according to pseudo-colours blending image
Picture.
4. the good man-machine interactive system of interaction effect according to claim 3, which is characterized in that first fusion is single
Member merges visible images and infrared image in non-down sampling contourlet transformation domain, specially:To visible images P and
Infrared image Q carries out non-down sampling contourlet decomposition, obtains corresponding sub-band division coefficientWith
RLPAnd RLQThe low frequency sub-band coefficient of visible images and infrared image is indicated respectively,WithRespectively indicate visible light and
The sub-band coefficients in k-th of direction in j-th of scale high-frequency sub-band of infrared image;
Low frequency sub-band is merged using following formula: In formula, RLR(x, y) indicates the corresponding low frequency sub-band coefficients of grayscale fusion image R,
Wherein, p indicates that the average gray value of infrared image, HX (x, y) indicate the gray value of pixel (x, y) in infrared image;
High-frequency sub-band is merged using following formula: In formula,Indicate the direction of fusion grayscale fusion image R
Sub-band coefficients, vP(x, y) indicates the side of visible images directional subband coefficient in n × n windows centered on pixel (x, y)
Difference, vQ(x, y) indicates the variance yields of infrared image directional subband coefficient in n × n windows centered on pixel (x, y);
Grayscale fusion image R is reconstructed according to the low frequency sub-band coefficient of grayscale fusion image and high-frequency sub-band coefficient.
5. the good man-machine interactive system of interaction effect according to claim 4, which is characterized in that second fusion is single
Member obtains pseudo-colours blending image according to grayscale fusion image, specially:Pseudo-colours is obtained using following formula in YUV color spaces to melt
Close image:In formula, Y (x, y), U (x, y), V (x, y) indicate that pseudo-colours is melted respectively
Image is closed in the component of YUV color spaces, the grayscale fusion image of R (x, y) expression visible images and infrared image, P (x, y)
Indicate that visible images, Q (x, y) indicate infrared image.
6. the good man-machine interactive system of interaction effect according to claim 5, which is characterized in that the third fusion
Unit obtains color fusion image according to pseudo-colours blending image, specially:The coloured silk that will be shot under the conditions of a width natural daylight
Color visible images, which are used as, refers to image, and the reference picture is converted into YUV color spaces, according to reference picture in YUV face
Gray average in each channel of the colour space and variance, the corresponding YUV component values of adjustment pseudo-colours blending image, after being adjusted
Pseudo-colours blending image is specifically carried out using following formula: In formula, S and W are corresponded to respectively
Reference picture and pseudo-colours blending image, Y1(x,y)、U1(x,y)、V1(x, y) indicates the pseudo-colours blending image after adjustment respectively
In the component of YUV color spaces, μ and σ indicate the gray average and variance of each Color Channel in YUV color spaces respectively;It will adjust
Pseudo-colours blending image after whole obtains color fusion image from YUV color notation conversion space to RGB color.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810185554.1A CN108491073A (en) | 2018-03-06 | 2018-03-06 | A kind of good man-machine interactive system of interaction effect |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810185554.1A CN108491073A (en) | 2018-03-06 | 2018-03-06 | A kind of good man-machine interactive system of interaction effect |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108491073A true CN108491073A (en) | 2018-09-04 |
Family
ID=63341629
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810185554.1A Withdrawn CN108491073A (en) | 2018-03-06 | 2018-03-06 | A kind of good man-machine interactive system of interaction effect |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108491073A (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106693348A (en) * | 2016-12-28 | 2017-05-24 | 巢湖学院 | Physical training monitoring system |
CN107253485A (en) * | 2017-05-16 | 2017-10-17 | 北京交通大学 | Foreign matter invades detection method and foreign matter intrusion detection means |
-
2018
- 2018-03-06 CN CN201810185554.1A patent/CN108491073A/en not_active Withdrawn
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106693348A (en) * | 2016-12-28 | 2017-05-24 | 巢湖学院 | Physical training monitoring system |
CN107253485A (en) * | 2017-05-16 | 2017-10-17 | 北京交通大学 | Foreign matter invades detection method and foreign matter intrusion detection means |
Non-Patent Citations (1)
Title |
---|
吴燕燕等: "结合NSST和颜色对比度增强的彩色夜视方法", 《光电工程》 * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107292813B (en) | A kind of multi-pose Face generation method based on generation confrontation network | |
CN107194904B (en) | NSCT area image fusion method based on supplement mechanism and PCNN | |
CN105787439B (en) | A kind of depth image human synovial localization method based on convolutional neural networks | |
CN104318569B (en) | Space salient region extraction method based on depth variation model | |
CN105069746B (en) | Video real-time face replacement method and its system based on local affine invariant and color transfer technology | |
CN109086754A (en) | A kind of human posture recognition method based on deep learning | |
CN105843386A (en) | Virtual fitting system in shopping mall | |
CN106570474A (en) | Micro expression recognition method based on 3D convolution neural network | |
CN104504673A (en) | Visible light and infrared images fusion method based on NSST and system thereof | |
CN108764066A (en) | A kind of express delivery sorting working specification detection method based on deep learning | |
CN106056564B (en) | Edge clear image interfusion method based on joint sparse model | |
CN109241830B (en) | Classroom lecture listening abnormity detection method based on illumination generation countermeasure network | |
CN1873693B (en) | Method based on Contourlet transformation, modified type pulse coupling neural network, and image amalgamation | |
CN108363973A (en) | A kind of unconfined 3D expressions moving method | |
CN103473801A (en) | Facial expression editing method based on single camera and motion capturing data | |
CN104899921A (en) | Single-view video human body posture recovery method based on multi-mode self-coding model | |
Li et al. | [Retracted] Analysis of Tai Chi Ideological and Political Course in University Based on Big Data and Graph Neural Networks | |
CN107392131A (en) | A kind of action identification method based on skeleton nodal distance | |
CN103971329A (en) | Cellular nerve network with genetic algorithm (GACNN)-based multisource image fusion method | |
CN103020933A (en) | Multi-source image fusion method based on bionic visual mechanism | |
CN110135277B (en) | Human behavior recognition method based on convolutional neural network | |
CN107748798A (en) | A kind of hand-drawing image search method based on multilayer visual expression and depth network | |
CN109492668A (en) | MRI based on multichannel convolutive neural network not same period multi-mode image characterizing method | |
CN107066979A (en) | A kind of human motion recognition method based on depth information and various dimensions convolutional neural networks | |
CN105930793A (en) | Human body detection method based on SAE characteristic visual learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20180904 |
|
WW01 | Invention patent application withdrawn after publication |