CN105550679B - A kind of judgment method of robot cycle monitoring recording - Google Patents
A kind of judgment method of robot cycle monitoring recording Download PDFInfo
- Publication number
- CN105550679B CN105550679B CN201610109288.5A CN201610109288A CN105550679B CN 105550679 B CN105550679 B CN 105550679B CN 201610109288 A CN201610109288 A CN 201610109288A CN 105550679 B CN105550679 B CN 105550679B
- Authority
- CN
- China
- Prior art keywords
- robot
- waiting time
- judgment method
- monitored
- monitoring
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
Abstract
A kind of judgment method of robot cycle monitoring recording, using robot in the prior art as ontology, waiting time function is monitored in setting;The robot is set, can judge the age according to the sound of people, and the waiting time that setting circulation is monitored automatically;Design face recognition detection function, pass through recognition of face comparison database data, judge the user's age exchanged with the emotional robot, the waiting time that automatic setting circulation is monitored, the present invention adopts special optimization design, realize the judgment method of automatic setting circulation monitoring recording, existing emotional robot is enabled preferably to generate more abundant expression for the difference of user, in the case where effectively the age of identification user, expression and word speed, accurate, appropriate response is made.
Description
Technical field
The invention patent relates to a kind of robot human-machine intelligence alternating-current technique field, especially a kind of robot cycle is monitored
The judgment method of recording, this method use unique optimization design, are monitored between monitoring recording by the automatic cycle of setting
Every the time, to not have to go to wake up robot every time, also interaction can be constantly carried out with robot, greatly reinforced existing
The using effect of class and children education study class robot is accompanied in technology.
Background technique
In today of material progress and cultural and ideological progress high development, intelligent robot is increasingly paid attention to and is popularized, people
Life and amusement in also gradually using intelligent robot participate in some families education, cleaning and other work things
Business, in the robot of general company or children education, Intelligent dialogue function is only limitted to user and passes through some specific word
Language goes after waking up robot once be exchanged with people, and after the completion of robot and user's dialogue, people rethinks and machine
People speaks, and robot also cannot achieve monitor function, also has no idea to answer the problem of user proposes certainly, that is to say, that
The educational robot of reading partner in the prior art can only interact once after being waken up with user;
On the other hand, traditional robot and do not have and remove to judge and be arranged circulation monitoring recording according to word speed and sound
Interval time, it is even more impossible to pass through age of recognition of face people and interval time of circulation monitoring recording is arranged, therefore can not be with abundant
Lively aptitude manner realization people interacts with robot, therefore traditional robot is limited to the condition of itself, can not present
Too many active, changeableization, lively content, thus limited on interacting with user;
Summary of the invention
It is an object of the present invention to provide a kind of judgment methods of robot cycle monitoring recording, solve in the prior art
Emotional robot has a single function, cannot detect face, cannot understand the people with robot interactive by the sentiment analysis of voice
The emotion of class can not carry out the deficiency of abundant affective interaction with the mankind, and this method is by optimization design, using the sound of user
Sound, expression, the tone are to judge medium, and the emotion of user is identified with this, so that the interval time of circulation monitoring recording be arranged;
A kind of judgment method of robot cycle monitoring recording, including following Technology design:
Step 1: using emotional robot in the prior art as ontology, waiting time function is monitored in setting;
Illustrating as an example, the monitoring waiting time function is, according to the word speed speed of people, automatic setting circulation
Monitor the working method of waiting time;
Illustrate as an example, the monitoring waiting time function is completed using the processing of chess module unit is monitored;
Illustrate as an example, the monitoring modular unit flies to monitor module using news;
Step 2: the emotional robot is arranged, the age can be judged according to the sound of people, and setting recycles monitoring automatically
Waiting time;
For example, what is heard is that old man or child speak, robot will be arranged long the time that circulation is monitored;
Illustrate as an example, the waiting time of the old man and child's circulation monitoring is set as 5 seconds, and robot is allowed to have
Time enough waits old man or child to speak;
For another example, the word speed of young man is fast, and the time for recycling monitoring is just shorter;
Illustrate as a kind of applicating example, the time that the young man recycles monitoring is set as 1 second or 2 seconds, that is to say, that
It begins listening for waiting, does not hear that the sound of people just ceases listening for also within 1 second or 2 seconds;
Step 3: design face recognition detection function is judged and the feelings by recognition of face comparison database data
Feel user's age of robot exchange, the waiting time that automatic setting circulation is monitored, achievees the purpose that dual fail-safe;
Illustrate as an example, a large amount of human face data not of the same grade of typing, is sentenced in the comparison database with improving
Disconnected accuracy rate;
A kind of judgment method of robot cycle monitoring recording provided by the invention identifies judgement design, sound by word speed
Sound identification judgement design and recognition of face judgement design to perceive the variation of external users, can pass through the sound of user
The media such as sound, expression, word speed identification, circulation identification recording, preferably judges the emotion of user, then pass through the emotion machine
The modes such as expression, computer vision, the interactive voice of device people itself exchange naturally with the mankind, reach optimal exchange and interdynamic effect
Fruit.
Beneficial effects of the present invention:
The present invention adopts special optimization design, the judgment method of automatic setting circulation monitoring recording is realized, so that existing
There is emotional robot that can preferably generate more abundant expression for the difference of user, effectively identification user's
In the case where age, expression and word speed, accurate, appropriate response is made, expression is naturally, judgement is accurate.
Detailed description of the invention
Fig. 1 is a kind of theory structure schematic diagram of the judgment method of robot cycle monitoring recording of the present invention
Specific embodiment
With reference to the accompanying drawing to a preferred embodiment of the present invention will be described in detail.
Shown in referring to Fig.1, a kind of judgment method of robot cycle monitoring recording, including following Technology design:
Step 1: using emotional robot in the prior art as ontology, waiting time function is monitored in setting;
Illustrating as an example, the monitoring waiting time function is, according to the word speed speed of people, automatic setting circulation
Monitor the working method of waiting time;
Illustrate as an example, the monitoring waiting time function is completed using modular unit processing is monitored;
Illustrate as an example, the monitoring modular unit flies to monitor module using news;
Step 2: the emotional robot is arranged, the age can be judged according to the sound of people, and setting recycles monitoring automatically
Waiting time;
For example, what is heard is that old man or child speak, robot will be arranged long the time that circulation is monitored;
Illustrate as an example, the waiting time of the old man and child's circulation monitoring is set as 5 seconds, and robot is allowed to have
Time enough waits old man or child to speak;
For another example, the word speed of young man is fast, and the time for recycling monitoring is just shorter;
Illustrate as a kind of applicating example, the time that the young man recycles monitoring is set as 1 second or 2 seconds, that is to say, that
It begins listening for waiting, does not hear that the sound of people just ceases listening for also within 1 second or 2 seconds;
Step 3: design face recognition detection function is judged and the feelings by recognition of face comparison database data
Feel user's age of robot exchange, the waiting time that automatic setting circulation is monitored, achievees the purpose that dual fail-safe;
Illustrate as an example, a large amount of human face data not of the same grade of typing, is sentenced in the comparison database with improving
Disconnected accuracy rate;
A kind of judgment method of robot cycle monitoring recording provided by the invention identifies judgement design, sound by word speed
Sound identification judgement design and recognition of face judgement design to perceive the variation of external users, can pass through the sound of user
The media such as sound, expression, word speed identification, circulation identification recording, preferably judges the emotion of user, then pass through the emotion machine
The modes such as expression, computer vision, the interactive voice of device people itself exchange naturally with the mankind, reach optimal exchange and interdynamic effect
Fruit.
The present invention adopts special optimization design, the judgment method of automatic setting circulation monitoring recording is realized, so that existing
There is emotional robot that can preferably generate more abundant expression for the difference of user, effectively identification user's
In the case where age, expression and word speed, accurate, appropriate response is made, expression is naturally, judgement is accurate.
It is above-described to be merely a preferred embodiment of the present invention, it should be understood that the explanation of above embodiments is only used
In facilitating the understanding of the method and its core concept of the invention, it is not intended to limit the scope of protection of the present invention, it is all of the invention
Any modification for being made within thought and principle, equivalent replacement etc., should all be included in the protection scope of the present invention.
Claims (8)
1. a kind of judgment method of robot cycle monitoring recording, which is characterized in that including following Technology design:
Step 1: using the robot as ontology, waiting time function is monitored in setting;The monitoring waiting time function
For according to the word speed speed of people, automatic setting circulation monitors the working method of waiting time;
When Step 2: the robot is arranged, can judge the age according to the sound of people, and the waiting of circulation monitoring being set automatically
It is long;The waiting time setting that old man and child recycle monitoring is relatively long, and phase is arranged in the waiting time that the circulation of young man is monitored
To shorter;
Step 3: design face recognition detection function is judged and the robot by recognition of face comparison database data
User's age of exchange, the waiting time that automatic setting circulation is monitored.
2. a kind of judgment method of robot cycle monitoring recording according to claim 1, which is characterized in that the old man
The waiting time for recycling monitoring with child is set as 5 seconds.
3. a kind of judgment method of robot cycle monitoring recording according to claim 1, which is characterized in that the youth
People recycles the time monitored and is set as 1 second or 2 seconds, that is to say, that begins listening for waiting, does not hear within 1 second or 2 seconds the sound of people also
Sound just ceases listening for.
4. a kind of judgment method of robot cycle monitoring recording according to claim 1, which is characterized in that the machine
People uses emotional robot.
5. a kind of judgment method of robot cycle monitoring recording according to claim 1, which is characterized in that the comparison
The typing human face data of a large amount of all ages and classes in database, to improve the accuracy rate of judgement.
6. a kind of judgment method of robot cycle monitoring recording according to claim 1, which is characterized in that the monitoring
Waiting time function is completed using modular unit processing is monitored.
7. a kind of judgment method of robot cycle monitoring recording according to claim 6, which is characterized in that the monitoring
Modular unit flies to monitor module using news.
8. a kind of judgment method of robot cycle monitoring recording according to claim 4, which is characterized in that the emotion
Robot itself has the expression way that expression, computer vision, interactive voice etc. exchange naturally with the mankind.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610109288.5A CN105550679B (en) | 2016-02-29 | 2016-02-29 | A kind of judgment method of robot cycle monitoring recording |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610109288.5A CN105550679B (en) | 2016-02-29 | 2016-02-29 | A kind of judgment method of robot cycle monitoring recording |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105550679A CN105550679A (en) | 2016-05-04 |
CN105550679B true CN105550679B (en) | 2019-02-15 |
Family
ID=55829862
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610109288.5A Active CN105550679B (en) | 2016-02-29 | 2016-02-29 | A kind of judgment method of robot cycle monitoring recording |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105550679B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107040450B (en) | 2016-07-20 | 2018-06-01 | 平安科技(深圳)有限公司 | Automatic reply method and device |
CN111292146B (en) * | 2018-12-07 | 2023-06-13 | 泰康保险集团股份有限公司 | Insurance recommendation method and apparatus, computer storage medium and electronic device |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1759436A (en) * | 2003-05-21 | 2006-04-12 | 松下电器产业株式会社 | Voice output device and voice output method |
CN104021373A (en) * | 2014-05-27 | 2014-09-03 | 江苏大学 | Semi-supervised speech feature variable factor decomposition method |
CN104915000A (en) * | 2015-05-27 | 2015-09-16 | 天津科技大学 | Multisensory biological recognition interaction method for naked eye 3D advertisement |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030177102A1 (en) * | 2001-09-21 | 2003-09-18 | Timothy Robinson | System and method for biometric authorization for age verification |
-
2016
- 2016-02-29 CN CN201610109288.5A patent/CN105550679B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1759436A (en) * | 2003-05-21 | 2006-04-12 | 松下电器产业株式会社 | Voice output device and voice output method |
CN104021373A (en) * | 2014-05-27 | 2014-09-03 | 江苏大学 | Semi-supervised speech feature variable factor decomposition method |
CN104915000A (en) * | 2015-05-27 | 2015-09-16 | 天津科技大学 | Multisensory biological recognition interaction method for naked eye 3D advertisement |
Also Published As
Publication number | Publication date |
---|---|
CN105550679A (en) | 2016-05-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108000526B (en) | Dialogue interaction method and system for intelligent robot | |
CN102354349B (en) | Human-machine interaction multi-mode early intervention system for improving social interaction capacity of autistic children | |
Lany et al. | Interactions between statistical and semantic information in infant language development | |
CN106486122A (en) | A kind of intelligent sound interacts robot | |
Garg et al. | The last decade of HCI research on children and voice-based conversational agents | |
Hung et al. | Towards a method for evaluating naturalness in conversational dialog systems | |
Tahon et al. | Real-life emotion detection from speech in human-robot interaction: Experiments across diverse corpora with child and adult voices | |
CN105550679B (en) | A kind of judgment method of robot cycle monitoring recording | |
Caza et al. | Pragmatic bootstrapping: A neural network model of vocabulary acquisition | |
Catania et al. | Emoty: an emotionally sensitive conversational agent for people with neurodevelopmental disorders | |
Zhang et al. | Design of an intelligent agent to measure collaboration and verbal-communication skills of children with autism spectrum disorder in collaborative puzzle games | |
Catania et al. | What is the Best Action for Children to" Wake Up" and" Put to Sleep" a Conversational Agent? A Multi-Criteria Decision Analysis Approach | |
CN106844675A (en) | A kind of robot multi-modal output intent and robot for children | |
Gibson et al. | The processing and acquisition of reference | |
Lehman | Robo fashion world: a multimodal corpus of multi-child human-computer interaction | |
Croft | Language as a process | |
Mirnig et al. | Face-to-face with a robot: What do we actually talk about? | |
Navarretta | Transfer learning in multimodal corpora | |
Gunson et al. | It's Good to Chat? Evaluation and Design Guidelines for Combining Open-Domain Social Conversation with Task-Based Dialogue in Intelligent Buildings | |
Godwin-Jones | 4 Smart devices and informal language learning | |
Vanel et al. | A Survey of Socio-Emotional Strategies for Generation-Based Conversational Agents. | |
Weixiang et al. | The role of English language and literature in cultivating students' language skills | |
Khoo et al. | Spill the tea: When robot conversation agents support well-being for older adults | |
Zhang et al. | Development of a Portable Psychological Support Robot System | |
Goh | Comprehending speech genres for the listening classroom |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CP03 | Change of name, title or address |
Address after: 518000 Guangdong, Shenzhen, Baoan District, Baoan District Air Town Street Development Area, 2A, building 4 and 4, 4 Patentee after: SHENZHEN YYD ROBO Co.,Ltd. Address before: 518054, C building, 5 software industry base, Xuefu Road, Shenzhen, Guangdong, Nanshan District Patentee before: SHENZHEN QIANHAI YONGYIDA ROBOT Co.,Ltd. |
|
CP03 | Change of name, title or address |