CN110524559B - Intelligent man-machine interaction system and method based on personnel behavior data - Google Patents

Intelligent man-machine interaction system and method based on personnel behavior data Download PDF

Info

Publication number
CN110524559B
CN110524559B CN201910815748.XA CN201910815748A CN110524559B CN 110524559 B CN110524559 B CN 110524559B CN 201910815748 A CN201910815748 A CN 201910815748A CN 110524559 B CN110524559 B CN 110524559B
Authority
CN
China
Prior art keywords
data
behavior
personnel
face
skeleton
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910815748.XA
Other languages
Chinese (zh)
Other versions
CN110524559A (en
Inventor
甘小勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Weizhi Technology Co ltd
Original Assignee
Chengdu Weizhi Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Weizhi Technology Co ltd filed Critical Chengdu Weizhi Technology Co ltd
Priority to CN201910815748.XA priority Critical patent/CN110524559B/en
Publication of CN110524559A publication Critical patent/CN110524559A/en
Application granted granted Critical
Publication of CN110524559B publication Critical patent/CN110524559B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/0005Manipulators having means for high-level communication with users, e.g. speech generator, face recognition means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1602Programme controls characterised by the control system, structure, architecture

Abstract

The invention discloses an intelligent human-computer interaction system and method based on personnel behavior data, wherein the human-computer interaction system comprises a local subsystem and a cloud subsystem; the local subsystem comprises a local detection module, a local processing module, a voice synthesis and output module and a mobile unit; the cloud subsystem comprises: the system comprises a behavior recognition module, an identity recognition module and a personnel identity behavior database; the invention can identify the interactive scene according to the field or historical behavior data, make different interactive responses according to the identity of the person, and realize richer and more natural interactive responses according to the current behavior and the historical behavior data of the person.

Description

Intelligent man-machine interaction system and method based on personnel behavior data
Technical Field
The invention relates to a human-computer interaction robot, in particular to an intelligent human-computer interaction system and method based on personnel behavior data.
Background
With the rapid breakthrough of the algorithm and the calculation power of the artificial intelligence technology, the demand of the artificial intelligence product is more and more intense. Typical representatives at home and abroad are intelligent sound products taking an intelligent voice technology as a core, and the products mainly take passive voice response or mechanical response according to preset starting scenes or time as main characteristics and cannot perform active intelligent interaction according to behavior data of personnel such as identity, action, expression, social contact and the like.
The traditional voice interaction technology mainly considers the application of standardized scenes such as security, e-commerce and hotels, and the patent with the publication number of CN 108877797A discloses an active interactive intelligent voice system, which can make active or passive response according to preset scenes or time, but cannot make intelligent response according to on-site or historical behavior (action, expression and social contact) data, and cannot meet the requirements of personalized artificial intelligent interaction scenes in the fields of education, medical treatment and old age.
Disclosure of Invention
The invention aims to provide an intelligent human-computer interaction system based on personnel behavior data to solve the problems, which comprises a local subsystem and a cloud subsystem;
the local subsystem comprises:
the local detection module is used for acquiring face and skeleton images, scene images and voice print data;
the local processing module is used for analyzing the image face and the skeleton, processing the voice data and tracking a face target; the system captures a face image, obtains face characteristic parameters, matches personnel identity information through an identity recognition module, obtains personnel behavior data, recognizes the personnel behavior data through the behavior recognition module and adds the recognition data into the personnel identity behavior database; when the characteristic parameters accord with a set value, marking the characteristic points of the face image, and tracking the characteristic points;
And the voice synthesis and output module is used for outputting voice interaction data.
The mobile unit comprises a motor and a limiting device, the system sets the initial rotation angle of the motor, and after the human face and the skeleton are detected, the rotation angle of the motor is adjusted to enable the human face and the skeleton image to be positioned in the center of the picture;
the cloud subsystem comprises:
the behavior recognition module is used for detecting the human skeleton image, recognizing human body actions through the action model and performing action model training;
the identity recognition module is used for detecting the face image, extracting face characteristic points and recognizing the identity information of the personnel through characteristic matching;
the personnel identity behavior database stores personnel identity data and current behavior and historical behavior data of personnel;
and the interactive content output module processes the current behavior and historical behavior data of the personnel and outputs interactive data to the voice synthesis and output module.
The intelligent man-machine interaction method based on the personnel behavior data comprises the following steps:
s1, the system collects image data and voice print data;
s2, extracting the face feature and the skeleton feature by the system, recognizing the voice voiceprint data and recognizing the scene;
s3, judging whether the face and the skeleton are in the center position of the image, if so, turning to S4, otherwise, calculating the positions of the face and the skeleton by the system, controlling the motor to rotate to enable the face and the skeleton to be in the center position of the image, and tracking the face target;
S4, the system identifies the personnel identity and the behavior according to the personnel identity behavior database;
s5, the system generates voice interaction data according to the personnel identity, the behavior and the scene data;
s6, the system outputs the voice interaction data.
The invention has the beneficial effects that: the invention can identify the interactive scene according to the data of the on-site or historical behaviors (actions, expressions and social contact), make different interactive responses according to the identity of the person, and realize richer and more natural interactive responses according to the data of the current behaviors and the historical behaviors of the person.
Drawings
FIG. 1 is a system diagram of an intelligent human-computer interaction system based on human behavior data;
FIG. 2 is a flow chart of an intelligent human-computer interaction method based on human behavior data.
Detailed Description
The invention will be further described with reference to the accompanying drawings in which:
as shown in fig. 1, the intelligent human-computer interaction system based on personnel behavior data comprises a local subsystem and a cloud subsystem;
the local subsystem includes:
the local detection module is used for acquiring face and skeleton images, scene images and voice print data;
the local processing module is used for analyzing the image human face and the skeleton, processing the voice data and tracking the human face target; the system captures a face image, obtains face characteristic parameters, matches personnel identity information through an identity recognition module, obtains personnel behavior data, recognizes the personnel behavior data through the behavior recognition module and adds the recognition data into the personnel identity behavior database; when the characteristic parameters accord with a set value, marking the characteristic points of the face image, and tracking the characteristic points;
And the voice synthesis and output module is used for outputting voice interaction data.
The mobile unit comprises a motor and a limiting device, the system sets the initial rotation angle of the motor, and after the face and the skeleton are detected, the rotation angle of the motor is adjusted to enable the face and the skeleton to be positioned in the center of the picture;
the cloud subsystem comprises:
the behavior recognition module is used for detecting the human skeleton image, recognizing human body motions through the motion model and performing motion model training;
the identity recognition module is used for detecting the face image, extracting face characteristic points and recognizing the identity information of the personnel through characteristic matching;
the personnel identity behavior database stores personnel identity data and current behavior and historical behavior data of personnel;
and the interactive content output module processes the current behavior and historical behavior data of the personnel and outputs interactive data to the voice synthesis and output module.
The intelligent man-machine interaction method based on the personnel behavior data comprises the following steps:
s1, the system collects image data and voice print data;
s2, extracting the face feature and the skeleton feature by the system, recognizing the voice voiceprint data and recognizing the scene;
s3, judging whether the face and the skeleton are in the center position of the image, if so, turning to S4, otherwise, calculating the positions of the face and the skeleton by the system, controlling the motor to rotate to enable the face and the skeleton to be in the center position of the image, and tracking the face target;
S4, the system identifies the personnel identity and the behavior according to the personnel identity behavior database;
s5, the system generates voice interaction data according to the personnel identity, the behavior and the scene data;
s6, the system outputs the voice interaction data.
Further, the face target tracking process includes:
s31, detecting a local target, extracting target feature points, and completing feature modeling of the target according to the target feature points to generate a target address;
s32, inputting a first frame, initializing the detected target and creating a new tracker, and marking out the target address;
and S33, inputting the subsequent frame, predicting the track through a Kalman filter: calculating the intersection and comparison (IOU) of all target state predictions of the tracker and the detection box of the current frame according to the state and covariance predictions generated by the box of the previous frame, obtaining the maximum matching of the IOU through a Hungarian assignment algorithm, and then removing the matching pairs with the matching values lower than the threshold value of the IOU; updating a Kalman tracker by using a target box matched with the frame, calculating Kalman gain, state update and covariance update, and outputting a state update value as a tracking box of the frame; and if the target box is not matched, the tracker is reinitialized.
The local detection module adopts a high-definition intelligent camera and a high-sensitivity microphone matrix, and carries out image and voice data sampling with different frequencies on site according to specific scenes; the local algorithm detects the face, the skeleton and the voice semantics on the basis of the high-speed CPU and the GPU respectively, if the data are abnormal, the detected data are uploaded to the cloud subsystem for processing, if the data are abnormal, the face, the skeleton and the voice semantics data are detected again, and the data are processed through the cloud subsystem, so that the pressure of data transmission of a local data processing network can be reduced, and the data processing efficiency of the local subsystem is improved. Capturing 72 parameter characteristics of human face eyes, mouth, nose, ears and the like from the video stream, and when the confidence coefficient of the characteristic parameters exceeds 80%, determining that a detection target accords with human face judgment; and simultaneously, the ID of the face is marked, and real-time tracking and reproduction of the marked feature points are realized.
And detecting a human body in the image, returning to the position of a rectangular frame of the human body, accurately positioning key points including the nose, the neck and main joint parts of four limbs, and analyzing personnel behavior data based on the position characteristics of the joint points.
The local subsystem is an intelligent robot end, and the robot head is provided with a brushless motor to drive the robot head to rotate. When the robot is started, networked and self-checked, the setting of the initial angle (0 degree) of the motor is finished according to the brushless motor and the limiting device of the head. The motor rotation range takes the angle as a starting point and can rotate by plus or minus 90 degrees. After the initial angle is set, starting local detection to detect whether a human face and a skeleton exist, if the human face ID is detected to be more than or equal to 1, finely adjusting the angle, placing the central point of the portrait at the left center and the right center of the picture (50% of the picture), then starting zooming until the human skeleton is equal to 30% of the picture, and staying for 45 seconds; and zooming the picture with the face area less than or equal to 30% again, and staying for 45 seconds again. And after the two times of zooming are finished, rotating to the next observation angle, and starting the next human face and skeleton detection. If the human face and the skeleton are not detected, zooming is gradually carried out until the human face and the skeleton can be detected, and then fine adjustment and zooming stop are carried out.
The face target tracking is realized based on an OpenCV visual library, after the detection is finished, target face characteristic points are extracted, a target ID is generated according to the characteristic points, when a first frame is input, a new tracker is initialized and created according to the detected target, and the ID is marked out; the subsequent frame comes in, the trajectory is predicted by a Kalman filter: i.e. the state and covariance predictions produced by the box of the previous frame. And solving the intersection and parallel ratio (IOU) of all target state predictions of the tracker and the detection box of the current frame, obtaining the maximum matching of the IOU through a Hungarian assignment algorithm, and then removing the matching pairs with the matching values lower than the threshold value of the IOU. And updating the Kalman tracker by using the target box matched with the frame, calculating Kalman gain, state update and covariance update, outputting a state update value as the tracking box of the frame, and re-initializing the tracker if no target is matched.
The personnel identity behavior database of the cloud subsystem comprises face identity, current expression, current action description, current voice content, current time, current course and historical behavior characteristics.
And the identity recognition module of the cloud subsystem judges the identity of the person according to the detected data information such as the skeleton, the face, the voice voiceprint and the like (if the person is a stranger, new person identity information is added to the person behavior database), the behavior recognition module recognizes the behavior of the person, and the person is stored in the person identity behavior database after the recognition is finished.
And the interactive content output module outputs the interactive content after processing the current behavior data and the historical data. And the cloud subsystem transmits the processed data back to the local voice synthesis and output module to realize voice interaction. And after the personnel respond to the interactive contents, the local detection module collects the on-site personnel response data again and enters the next intelligent interaction turn. During man-machine interaction, the system can realize rich and natural interaction according to personnel preference and characteristics.
The communication between the cloud subsystem and the local subsystem is realized through a high-speed and low-delay WIFI or 4G or 5G communication module.
The invention can identify the interactive scene according to the data of the on-site or historical behaviors (actions, expressions and social contact), makes interactive response according to the identity of the person, and can realize richer and more natural interactive response according to the current behavior and the historical behavior data of the person.
The technical solution of the present invention is not limited to the limitations of the above specific embodiments, and all technical modifications made according to the technical solution of the present invention fall within the protection scope of the present invention.

Claims (4)

1. The intelligent human-computer interaction system based on the personnel behavior data is characterized by comprising a local subsystem and a cloud subsystem;
The local subsystem includes:
the local detection module is used for acquiring face and skeleton images, scene images and voice print data;
the local processing module is used for analyzing the image human face and the skeleton, processing the voice data and tracking the human face target; the system captures a face image, obtains face characteristic parameters, matches personnel identity information through a personnel identity recognition module, obtains personnel behavior data, recognizes the personnel behavior data through a behavior recognition module and adds the recognition data into a personnel identity behavior database; when the characteristic parameters accord with a set value, marking the characteristic points of the face image, and tracking the characteristic points;
the voice synthesis and output module is used for outputting voice interaction data;
the mobile unit comprises a motor and a limiting device, the system sets the initial rotation angle of the motor, and after the human face and the skeleton are detected, the rotation angle of the motor is adjusted to enable the human face and the skeleton image to be positioned in the center of the picture;
the cloud subsystem comprises:
the behavior recognition module is used for detecting the human skeleton image, recognizing human body actions through the action model and performing action model training;
the personnel identity recognition module is used for detecting the face image, extracting the face characteristic points and recognizing the personnel identity information through characteristic matching by the system;
The personnel identity behavior database stores personnel identity data and current behavior and historical behavior data of personnel;
and the interactive content output module processes the current behavior and historical behavior data of the personnel and outputs the interactive data to the voice synthesis and output module.
2. The intelligent man-machine interaction method based on the personnel behavior data is characterized by comprising the following steps:
s1, the system collects image data and voice print data;
s2, extracting the face feature and the skeleton feature by the system, recognizing the voice voiceprint data and recognizing the scene;
s3, judging whether the face and the skeleton are in the center position of the image, if so, turning to S4, otherwise, calculating the positions of the face and the skeleton by the system, controlling the motor to rotate to enable the face and the skeleton to be in the center position of the image, and tracking the face target;
s4, the system identifies the personnel identity and the action according to the personnel identity action database;
s5, the system generates voice interaction data according to the personnel identity, the behavior and the scene data;
s6, outputting voice interaction data by the system;
the face target tracking process comprises the following steps:
s31, detecting a local target, extracting target feature points, and completing feature modeling of the target according to the target feature points to generate a target address;
S32, inputting a first frame, initializing the detected target and creating a new tracker, and marking out the target address;
and S33, inputting the subsequent frame, predicting the track through a Kalman filter: predicting states and covariance generated by a box of a previous frame, solving the intersection and comparison between all target state predictions of the tracker and a detection box of the current frame, obtaining the maximum matching of the intersection and comparison through a Hungarian assignment algorithm, and removing the matching pairs with the matching values lower than the threshold of the intersection and comparison; updating a Kalman tracker by using a target box matched with the frame, calculating Kalman gain, state update and covariance update, and outputting a state update value as a tracking box of the frame; and if the target box is not matched, the tracker is reinitialized.
3. The intelligent human-computer interaction method based on personnel behavior data as claimed in claim 2, wherein the behavior recognition specifically comprises:
s41, constructing an LSTM neural network action database, wherein one action corresponds to one or more data sets;
s42, establishing an action model of the relation among joints, a framework time domain and a space domain;
s43, training the action model, adjusting the parameters of the action model, when the training degree of the parameters of the action model reaches the set value, the action model is successfully trained;
S44, collecting joint and skeleton characteristic data, and inputting the joint and skeleton characteristic data and the identity of the person as an action training model;
and S45, outputting the action recognition result and updating the action recognition result to the LSTM neural network action database.
4. The intelligent human-computer interaction method based on personnel behavior data as claimed in claim 2, wherein the S5 specifically comprises:
s51, the system builds a corpus database;
s52, the system searches the corpus database according to the personnel identity, the behavior and the scene data to obtain the voice interaction data for man-machine interaction;
and S53, outputting the voice data and updating the corpus.
CN201910815748.XA 2019-08-30 2019-08-30 Intelligent man-machine interaction system and method based on personnel behavior data Active CN110524559B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910815748.XA CN110524559B (en) 2019-08-30 2019-08-30 Intelligent man-machine interaction system and method based on personnel behavior data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910815748.XA CN110524559B (en) 2019-08-30 2019-08-30 Intelligent man-machine interaction system and method based on personnel behavior data

Publications (2)

Publication Number Publication Date
CN110524559A CN110524559A (en) 2019-12-03
CN110524559B true CN110524559B (en) 2022-06-10

Family

ID=68665760

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910815748.XA Active CN110524559B (en) 2019-08-30 2019-08-30 Intelligent man-machine interaction system and method based on personnel behavior data

Country Status (1)

Country Link
CN (1) CN110524559B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112001275A (en) * 2020-08-09 2020-11-27 成都未至科技有限公司 Robot for collecting student information
CN114326625B (en) * 2021-12-28 2023-08-25 毕马智能科技(上海)有限公司 Monitoring system and method for potential safety risk in power grid infrastructure construction

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105093986A (en) * 2015-07-23 2015-11-25 百度在线网络技术(北京)有限公司 Humanoid robot control method based on artificial intelligence, system and the humanoid robot
CN105182983A (en) * 2015-10-22 2015-12-23 深圳创想未来机器人有限公司 Face real-time tracking method and face real-time tracking system based on mobile robot
CN105447466A (en) * 2015-12-01 2016-03-30 深圳市图灵机器人有限公司 Kinect sensor based identity comprehensive identification method
CN105957521A (en) * 2016-02-29 2016-09-21 青岛克路德机器人有限公司 Voice and image composite interaction execution method and system for robot
KR20180054407A (en) * 2016-11-15 2018-05-24 주식회사 로보러스 Apparatus for recognizing user emotion and method thereof, and robot system using the same
CN108510577A (en) * 2018-01-31 2018-09-07 中国科学院软件研究所 A kind of sense of reality action migration and generation method and system based on existing action data
CN108520205A (en) * 2018-03-21 2018-09-11 安徽大学 A kind of human motion recognition method based on Citation-KNN
CN109330494A (en) * 2018-11-01 2019-02-15 珠海格力电器股份有限公司 Sweeping robot control method based on action recognition, system, sweeping robot
CN109623848A (en) * 2019-02-26 2019-04-16 江苏艾萨克机器人股份有限公司 A kind of hotel service robot

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2933067B1 (en) * 2014-04-17 2019-09-18 Softbank Robotics Europe Method of performing multi-modal dialogue between a humanoid robot and user, computer program product and humanoid robot for implementing said method

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105093986A (en) * 2015-07-23 2015-11-25 百度在线网络技术(北京)有限公司 Humanoid robot control method based on artificial intelligence, system and the humanoid robot
CN105182983A (en) * 2015-10-22 2015-12-23 深圳创想未来机器人有限公司 Face real-time tracking method and face real-time tracking system based on mobile robot
CN105447466A (en) * 2015-12-01 2016-03-30 深圳市图灵机器人有限公司 Kinect sensor based identity comprehensive identification method
CN105957521A (en) * 2016-02-29 2016-09-21 青岛克路德机器人有限公司 Voice and image composite interaction execution method and system for robot
KR20180054407A (en) * 2016-11-15 2018-05-24 주식회사 로보러스 Apparatus for recognizing user emotion and method thereof, and robot system using the same
CN108510577A (en) * 2018-01-31 2018-09-07 中国科学院软件研究所 A kind of sense of reality action migration and generation method and system based on existing action data
CN108520205A (en) * 2018-03-21 2018-09-11 安徽大学 A kind of human motion recognition method based on Citation-KNN
CN109330494A (en) * 2018-11-01 2019-02-15 珠海格力电器股份有限公司 Sweeping robot control method based on action recognition, system, sweeping robot
CN109623848A (en) * 2019-02-26 2019-04-16 江苏艾萨克机器人股份有限公司 A kind of hotel service robot

Also Published As

Publication number Publication date
CN110524559A (en) 2019-12-03

Similar Documents

Publication Publication Date Title
Zhang et al. Empowering things with intelligence: a survey of the progress, challenges, and opportunities in artificial intelligence of things
WO2019174439A1 (en) Image recognition method and apparatus, and terminal and storage medium
CN110135249B (en) Human behavior identification method based on time attention mechanism and LSTM (least Square TM)
WO2018188453A1 (en) Method for determining human face area, storage medium, and computer device
CN107784294B (en) Face detection and tracking method based on deep learning
CN105426723A (en) Voiceprint identification, face identification and synchronous in-vivo detection-based identity authentication method and system
CN105426827A (en) Living body verification method, device and system
CN107894836B (en) Human-computer interaction method for processing and displaying remote sensing image based on gesture and voice recognition
CN110263603B (en) Face recognition method and device based on central loss and residual error visual simulation network
CN110524559B (en) Intelligent man-machine interaction system and method based on personnel behavior data
CN111128157B (en) Wake-up-free voice recognition control method for intelligent household appliance, computer readable storage medium and air conditioner
US20200302162A1 (en) Method and apparatus for recognizing an organism action, server, and storage medium
CN105042789B (en) The control method and system of a kind of intelligent air condition
CN111382648A (en) Method, device and equipment for detecting dynamic facial expression and storage medium
CN109299690B (en) Method capable of improving video real-time face recognition precision
CN111708663A (en) Cloud computing safety monitoring system based on artificial intelligence
Yargıç et al. A lip reading application on MS Kinect camera
CN106926252A (en) A kind of hotel's intelligent robot method of servicing
CN110852306A (en) Safety monitoring system based on artificial intelligence
Hafeez et al. Multi-fusion sensors for action recognition based on discriminative motion cues and random forest
CN113627256B (en) False video inspection method and system based on blink synchronization and binocular movement detection
CN112700568B (en) Identity authentication method, equipment and computer readable storage medium
CN113159002A (en) Facial expression recognition method based on self-attention weight auxiliary module
CN112818796A (en) Intelligent posture discrimination method and storage device suitable for online invigilation scene
CN106997449A (en) Robot and face identification method with face identification functions

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant