CN108491073A - A kind of good man-machine interactive system of interaction effect - Google Patents

A kind of good man-machine interactive system of interaction effect Download PDF

Info

Publication number
CN108491073A
CN108491073A CN201810185554.1A CN201810185554A CN108491073A CN 108491073 A CN108491073 A CN 108491073A CN 201810185554 A CN201810185554 A CN 201810185554A CN 108491073 A CN108491073 A CN 108491073A
Authority
CN
China
Prior art keywords
image
children
subsystem
pseudo
fusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201810185554.1A
Other languages
Chinese (zh)
Inventor
邱林新
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Kaida Photoelectric Technology Co Ltd
Original Assignee
Shenzhen Kaida Photoelectric Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Kaida Photoelectric Technology Co Ltd filed Critical Shenzhen Kaida Photoelectric Technology Co Ltd
Priority to CN201810185554.1A priority Critical patent/CN108491073A/en
Publication of CN108491073A publication Critical patent/CN108491073A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/20Education
    • G06Q50/205Education administration or guidance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/011Emotion or mood input determined on the basis of sensed human body parameters such as pulse, heart rate or beat, temperature of skin, facial expressions, iris, voice pitch, brain activity patterns

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Tourism & Hospitality (AREA)
  • Health & Medical Sciences (AREA)
  • Strategic Management (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Economics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • General Business, Economics & Management (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The present invention provides a kind of good man-machine interactive systems of interaction effect, including authentication subsystem, training subsystem, data process subsystem and interactive subsystem, the authentication subsystem is verified for treating trained children's identity, the trained subsystem is used for being identified by the human body behavior of authentication children, obtain Human bodys' response result, the data process subsystem is using Human bodys' response result as children's physical training basic data, determine children's physical training situation, and children's physical training situation is sent to interactive subsystem, the interactive subsystem is used to play children's physical training instructional video according to children's physical training situation.Beneficial effects of the present invention are:A kind of good man-machine interactive system of interaction effect is provided, which realizes effective training to children, and children can obtain instructional video by interactive subsystem, help to create good Learning atmosphere, improve the training enthusiasm of children.

Description

A kind of good man-machine interactive system of interaction effect
Technical field
The present invention relates to child teaching technical fields, and in particular to a kind of good man-machine interactive system of interaction effect.
Background technology
With the improvement of living standards, people increasingly focus on children's health, carrying out physical training to children seems particularly It is important.How to allow children warm to training, how to be assessed training condition is all the problem of pendulum is in face of people.
Human bodys' response is an emerging research direction in artificial intelligence field, be with a wide range of applications with it is non- The economic value of Chang Keguan, the application field being related to include mainly:Video monitoring, medical diagnosis and monitoring, motion analysis, intelligence Human-computer interaction, virtual reality etc..The corresponding groundwork flow of Human bodys' response is:Various kinds of sensors is selected to obtain human body row For data information, and the behavioral trait of sensor characteristics and people is combined to establish rational behavior model, on this basis from original Extracted in gathered data to behavior type have stronger descriptive power feature, and using suitable method to these features into Row training, and then realize the pattern-recognition to human body behavior.The image preprocessing of high quality is the key that Activity recognition research, existing The ineffective very big reason of somebody's body Activity recognition is not obtain the image of high quality.
Invention content
In view of the above-mentioned problems, the present invention is intended to provide a kind of good man-machine interactive system of interaction effect.
The purpose of the present invention is realized using following technical scheme:
Provide a kind of good man-machine interactive system of interaction effect, including authentication subsystem, training subsystem, number According to processing subsystem and interactive subsystem, the authentication subsystem is verified for treating trained children's identity, institute Trained subsystem is stated for being identified by the human body behavior of authentication children, obtaining Human bodys' response as a result, institute Data process subsystem is stated using Human bodys' response result as children's physical training basic data, determines children's physical training feelings Condition, and children's physical training situation is sent to interactive subsystem, the interactive subsystem is used for according to children's physical training feelings Condition plays children's physical training instructional video.
Beneficial effects of the present invention are:A kind of good man-machine interactive system of interaction effect is provided, which realizes Effective training to children, children can obtain instructional video by interactive subsystem, help to create good Learning atmosphere, Improve the training enthusiasm of children.
Description of the drawings
Using attached drawing, the invention will be further described, but the embodiment in attached drawing does not constitute any limit to the present invention System, for those of ordinary skill in the art, without creative efforts, can also obtain according to the following drawings Other attached drawings.
Fig. 1 is the structural schematic diagram of the present invention;
Reference numeral:
Authentication subsystem 1, training subsystem 2, data process subsystem 3, interactive subsystem 4.
Specific implementation mode
The invention will be further described with the following Examples.
Referring to Fig. 1, a kind of good man-machine interactive system of interaction effect of the present embodiment, including authentication subsystem 1, Training subsystem 2, data process subsystem 3 and interactive subsystem 4, the authentication subsystem 1 is for treating trained youngster Virgin identity is verified, and the trained subsystem 2 is used to, to being identified by the human body behavior of authentication children, obtain people Body Activity recognition as a result, the data process subsystem 3 using Human bodys' response result as children's physical training basic data, It determines children's physical training situation, and children's physical training situation is sent to interactive subsystem 4, the interactive subsystem 4 is used According to children's physical training situation broadcasting children's physical training instructional video.
A kind of good man-machine interactive system of interaction effect is present embodiments provided, which realizes to the effective of children Training, children can obtain instructional video by interactive subsystem, help to create good Learning atmosphere, improve the instruction of children Practice enthusiasm.
Preferably, the trained subsystem 2 includes first processing module, Second processing module, third processing module, the 4th Processing module, the first processing module is acquired using visible light, infrared multi-spectral imaging system on human body image, described Second processing module obtains color fusion image, the third processing for being merged to visible images and infrared image Module is used to extract human body target profile according to color fusion image, and the fourth processing module is used for according to human body target profile Human body behavior is identified.
This preferred embodiment trains subsystem 2 to obtain human body image and right using visible light, infrared multi-spectral imaging system Image carries out fusion treatment, obtains the image of high quality, contributes to the detectivity and tracing property that promote follow-up human body, uses Color fusion image meets the visual signature of the mankind.
Preferably, the Second processing module includes the first integrated unit, the second integrated unit, third integrated unit, institute It states the first integrated unit to merge visible images and infrared image in non-down sampling contourlet transformation domain, obtains gray scale and melt Image is closed, second integrated unit obtains pseudo-colours blending image, the third integrated unit root according to grayscale fusion image Color fusion image is obtained according to pseudo-colours blending image;
First integrated unit merges visible images and infrared image in non-down sampling contourlet transformation domain, Specially:Non-down sampling contourlet decomposition is carried out to visible images P and infrared image Q, obtains corresponding sub-band division coefficientWithRLPAnd RLQThe low frequency sub-band coefficient of visible images and infrared image is indicated respectively,WithThe sub-band coefficients in k-th of direction in j-th of scale high-frequency sub-band of visible light and infrared image are indicated respectively;
Low frequency sub-band is merged using following formula: In formula, RLR(x, y) indicates the corresponding low frequency sub-band coefficients of grayscale fusion image R, Wherein, p indicates that the average gray value of infrared image, HX (x, y) indicate the gray value of pixel (x, y) in infrared image;
High-frequency sub-band is merged using following formula: In formula,Indicate the direction of fusion grayscale fusion image R Sub-band coefficients, vP(x, y) indicates the side of visible images directional subband coefficient in n × n windows centered on pixel (x, y) Difference, vQ(x, y) indicates the variance yields of infrared image directional subband coefficient in n × n windows centered on pixel (x, y); Grayscale fusion image R is reconstructed according to the low frequency sub-band coefficient of grayscale fusion image and high-frequency sub-band coefficient;
The blending image that traditional image interfusion method obtains usually exist target-to-background contrast is relatively low, image more The deficiencies of fuzzy.This preferred embodiment is improved by carrying out multiple dimensioned, multi-direction fusion to visible images and infrared image Image co-registration is horizontal, by determining image co-registration mode, can preferably merge the information of the image of different-waveband, the ash of acquisition It is more abundant to spend blending image details, textural characteristics.
Preferably, second integrated unit obtains pseudo-colours blending image according to grayscale fusion image, specially:Using Following formula obtains pseudo-colours blending image in YUV color spaces:In formula, Y (x, y), U (x, y), V (x, y) indicate respectively pseudo-colours blending image YUV color spaces component, R (x, y) indicate visible images and The grayscale fusion image of infrared image, P (x, y) indicate that visible images, Q (x, y) indicate infrared image;
The third integrated unit obtains color fusion image according to pseudo-colours blending image, specially:By width nature The color visible image shot under sunshine condition, which is used as, refers to image, and the reference picture is converted into YUV color spaces, root According to gray average and variance of the reference picture in each channel of YUV color spaces, adjustment pseudo-colours blending image is YUV points corresponding Magnitude, the pseudo-colours blending image after being adjusted specifically are carried out using following formula: In formula, S Correspond to reference picture and pseudo-colours blending image, Y respectively with W1(x,y)、U1(x,y)、V1(x, y) indicates the puppet after adjustment respectively For color fusion image in the component of YUV color spaces, μ and σ indicate the gray average of each Color Channel in YUV color spaces respectively And variance;By the pseudo-colours blending image after adjustment from YUV color notation conversion space to RGB color, color integration figure is obtained Picture.
Complementary information between this preferred embodiment organic combination different-waveband, enriches the detailed information of image, makes one Body target is enhanced, to improve accuracy and robustness to target acquisition and tracking;Meanwhile blending image can be meter Calculation machine visual analysis provides higher-quality source images;In addition, the color fusion image after color adjusts has Natural color Color visual effect can improve the degree of fatigue that observer watches the perception of scene, reduction observer video, this is for certain The Activity recognition application that observer participates in is needed to be of great significance.
Using the good man-machine interactive system of interaction effect of the present invention to children carry out physical training, choose 5 children into Row experiment, respectively children 1, children 2, children 3, children 4, children 5 unite to children training enthusiasm and training cost Meter, is compared, generation has the beneficial effect that shown in table compared with children's physical training system:
Children training enthusiasm improves Training cost reduction
Children 1 29% 27%
Children 2 27% 26%
Children 3 26% 26%
Children 4 25% 24%
Children 5 24% 22%
Finally it should be noted that the above embodiments are merely illustrative of the technical solutions of the present invention, rather than the present invention is protected The limitation of range is protected, although being explained in detail to the present invention with reference to preferred embodiment, those skilled in the art answer Work as understanding, technical scheme of the present invention can be modified or replaced equivalently, without departing from the reality of technical solution of the present invention Matter and range.

Claims (6)

1. a kind of good man-machine interactive system of interaction effect, which is characterized in that including authentication subsystem, training subsystem System, data process subsystem and interactive subsystem, the authentication subsystem are tested for treating trained children's identity Card, the trained subsystem are used to, to being identified by the human body behavior of authentication children, obtain Human bodys' response knot Fruit, the data process subsystem determine children's physical efficiency using Human bodys' response result as children's physical training basic data Training, and children's physical training situation is sent to interactive subsystem, the interactive subsystem is used for according to children's physical efficiency Training plays children's physical training instructional video.
2. the good man-machine interactive system of interaction effect according to claim 1, which is characterized in that the trained subsystem Including first processing module, Second processing module, third processing module, fourth processing module, the first processing module uses Visible light, infrared multi-spectral imaging system on human body image are acquired, and the Second processing module is used for visible images It is merged with infrared image, obtains color fusion image, the third processing module according to color fusion image for extracting Human body target profile, the fourth processing module is for being identified human body behavior according to human body target profile.
3. the good man-machine interactive system of interaction effect according to claim 2, which is characterized in that the second processing mould Block includes the first integrated unit, the second integrated unit, third integrated unit, and first integrated unit is in non-down sampling contourlet Transform domain merges visible images and infrared image, obtains grayscale fusion image, second integrated unit is according to ash It spends blending image and obtains pseudo-colours blending image, the third integrated unit obtains color integration figure according to pseudo-colours blending image Picture.
4. the good man-machine interactive system of interaction effect according to claim 3, which is characterized in that first fusion is single Member merges visible images and infrared image in non-down sampling contourlet transformation domain, specially:To visible images P and Infrared image Q carries out non-down sampling contourlet decomposition, obtains corresponding sub-band division coefficientWith RLPAnd RLQThe low frequency sub-band coefficient of visible images and infrared image is indicated respectively,WithRespectively indicate visible light and The sub-band coefficients in k-th of direction in j-th of scale high-frequency sub-band of infrared image;
Low frequency sub-band is merged using following formula: In formula, RLR(x, y) indicates the corresponding low frequency sub-band coefficients of grayscale fusion image R, Wherein, p indicates that the average gray value of infrared image, HX (x, y) indicate the gray value of pixel (x, y) in infrared image;
High-frequency sub-band is merged using following formula: In formula,Indicate the direction of fusion grayscale fusion image R Sub-band coefficients, vP(x, y) indicates the side of visible images directional subband coefficient in n × n windows centered on pixel (x, y) Difference, vQ(x, y) indicates the variance yields of infrared image directional subband coefficient in n × n windows centered on pixel (x, y); Grayscale fusion image R is reconstructed according to the low frequency sub-band coefficient of grayscale fusion image and high-frequency sub-band coefficient.
5. the good man-machine interactive system of interaction effect according to claim 4, which is characterized in that second fusion is single Member obtains pseudo-colours blending image according to grayscale fusion image, specially:Pseudo-colours is obtained using following formula in YUV color spaces to melt Close image:In formula, Y (x, y), U (x, y), V (x, y) indicate that pseudo-colours is melted respectively Image is closed in the component of YUV color spaces, the grayscale fusion image of R (x, y) expression visible images and infrared image, P (x, y) Indicate that visible images, Q (x, y) indicate infrared image.
6. the good man-machine interactive system of interaction effect according to claim 5, which is characterized in that the third fusion Unit obtains color fusion image according to pseudo-colours blending image, specially:The coloured silk that will be shot under the conditions of a width natural daylight Color visible images, which are used as, refers to image, and the reference picture is converted into YUV color spaces, according to reference picture in YUV face Gray average in each channel of the colour space and variance, the corresponding YUV component values of adjustment pseudo-colours blending image, after being adjusted Pseudo-colours blending image is specifically carried out using following formula: In formula, S and W are corresponded to respectively Reference picture and pseudo-colours blending image, Y1(x,y)、U1(x,y)、V1(x, y) indicates the pseudo-colours blending image after adjustment respectively In the component of YUV color spaces, μ and σ indicate the gray average and variance of each Color Channel in YUV color spaces respectively;It will adjust Pseudo-colours blending image after whole obtains color fusion image from YUV color notation conversion space to RGB color.
CN201810185554.1A 2018-03-06 2018-03-06 A kind of good man-machine interactive system of interaction effect Withdrawn CN108491073A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810185554.1A CN108491073A (en) 2018-03-06 2018-03-06 A kind of good man-machine interactive system of interaction effect

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810185554.1A CN108491073A (en) 2018-03-06 2018-03-06 A kind of good man-machine interactive system of interaction effect

Publications (1)

Publication Number Publication Date
CN108491073A true CN108491073A (en) 2018-09-04

Family

ID=63341629

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810185554.1A Withdrawn CN108491073A (en) 2018-03-06 2018-03-06 A kind of good man-machine interactive system of interaction effect

Country Status (1)

Country Link
CN (1) CN108491073A (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106693348A (en) * 2016-12-28 2017-05-24 巢湖学院 Physical training monitoring system
CN107253485A (en) * 2017-05-16 2017-10-17 北京交通大学 Foreign matter invades detection method and foreign matter intrusion detection means

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106693348A (en) * 2016-12-28 2017-05-24 巢湖学院 Physical training monitoring system
CN107253485A (en) * 2017-05-16 2017-10-17 北京交通大学 Foreign matter invades detection method and foreign matter intrusion detection means

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
吴燕燕等: "结合NSST和颜色对比度增强的彩色夜视方法", 《光电工程》 *

Similar Documents

Publication Publication Date Title
CN107292813B (en) A kind of multi-pose Face generation method based on generation confrontation network
CN107194904B (en) NSCT area image fusion method based on supplement mechanism and PCNN
CN105787439B (en) A kind of depth image human synovial localization method based on convolutional neural networks
CN104318569B (en) Space salient region extraction method based on depth variation model
CN105069746B (en) Video real-time face replacement method and its system based on local affine invariant and color transfer technology
CN109086754A (en) A kind of human posture recognition method based on deep learning
CN105843386A (en) Virtual fitting system in shopping mall
CN106570474A (en) Micro expression recognition method based on 3D convolution neural network
CN104504673A (en) Visible light and infrared images fusion method based on NSST and system thereof
CN108764066A (en) A kind of express delivery sorting working specification detection method based on deep learning
CN106056564B (en) Edge clear image interfusion method based on joint sparse model
CN109241830B (en) Classroom lecture listening abnormity detection method based on illumination generation countermeasure network
CN1873693B (en) Method based on Contourlet transformation, modified type pulse coupling neural network, and image amalgamation
CN108363973A (en) A kind of unconfined 3D expressions moving method
CN103473801A (en) Facial expression editing method based on single camera and motion capturing data
CN104899921A (en) Single-view video human body posture recovery method based on multi-mode self-coding model
Li et al. [Retracted] Analysis of Tai Chi Ideological and Political Course in University Based on Big Data and Graph Neural Networks
CN107392131A (en) A kind of action identification method based on skeleton nodal distance
CN103971329A (en) Cellular nerve network with genetic algorithm (GACNN)-based multisource image fusion method
CN103020933A (en) Multi-source image fusion method based on bionic visual mechanism
CN110135277B (en) Human behavior recognition method based on convolutional neural network
CN107748798A (en) A kind of hand-drawing image search method based on multilayer visual expression and depth network
CN109492668A (en) MRI based on multichannel convolutive neural network not same period multi-mode image characterizing method
CN107066979A (en) A kind of human motion recognition method based on depth information and various dimensions convolutional neural networks
CN105930793A (en) Human body detection method based on SAE characteristic visual learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20180904

WW01 Invention patent application withdrawn after publication