CN110018738B - Emotion conversion system based on real scene emotion expression - Google Patents

Emotion conversion system based on real scene emotion expression Download PDF

Info

Publication number
CN110018738B
CN110018738B CN201910159043.7A CN201910159043A CN110018738B CN 110018738 B CN110018738 B CN 110018738B CN 201910159043 A CN201910159043 A CN 201910159043A CN 110018738 B CN110018738 B CN 110018738B
Authority
CN
China
Prior art keywords
emotion
scene
visual
current
real
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910159043.7A
Other languages
Chinese (zh)
Other versions
CN110018738A (en
Inventor
青春美
欧浩春
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201910159043.7A priority Critical patent/CN110018738B/en
Publication of CN110018738A publication Critical patent/CN110018738A/en
Application granted granted Critical
Publication of CN110018738B publication Critical patent/CN110018738B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/001Texturing; Colouring; Generation of texture or colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/011Emotion or mood input determined on the basis of sensed human body parameters such as pulse, heart rate or beat, temperature of skin, facial expressions, iris, voice pitch, brain activity patterns

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an emotion conversion system based on real scene emotion expression, which is used for a VR (virtual reality) system, reduces stimulation of a visual environment to a user, mainly carries out emotion calculation through a current visual field focal region, simultaneously detects a local image of the visual field focal region, weakens an object with potential strong visual stimulation to a wearer through a weakening visual region stimulation network and then combines the object with a real scene, carries out emotion conversion on the scene when the current scene emotion does not meet the selected scene emotion so that the emotion of the scene meets the requirement, and displays the converted scene on VR equipment in real time.

Description

Emotion conversion system based on real scene emotion expression
Technical Field
The invention relates to the technical field of information, in particular to an emotion conversion system based on real scene emotion expression.
Background
Human emotion affects human cognition and daily behavior, and emotion calculation enables a computer to quantitatively understand human emotional feeling. At present, objects for emotion calculation mainly include electroencephalograms, multimedia data (such as audio and images), text materials and the like. For images, emotion calculation depends mainly on the tone and texture features of the incoming image. Efficient emotion calculation enables a computer to extract features of the incoming object's emotional expression and to change the emotional expression of the incoming object according to the specified emotion. Currently, emotion calculation is mainly applied to multimedia data mining, such as social platforms. However, emotion calculation is not well applied to psychotherapy due to the lack of targeted methods and effective protocols.
Disclosure of Invention
In order to overcome the defects and shortcomings of the prior art, the invention provides an emotion conversion system based on real scene emotion expression. According to the invention, the real scene is converted into the selected scene emotion and displayed on the VR equipment, so that the stimulation of the real scene to the user is reduced, and accidents are avoided.
The invention adopts the following technical scheme:
an emotion conversion system based on real scene emotion expression is used for a VR system, the VR system is provided with a first camera for collecting real-time scene videos watched by a user and a second camera for collecting eye movements of the user, and the emotion conversion system comprises
The module for acquiring the visual field focal region judges the scene difference of the image data of the real-time scene video stream watched by the user to obtain the visual field focal region of the current frame image;
the scene emotion calculation module is used for carrying out scene calculation on the image data of the focus area of the visual field to obtain emotion distribution vectors for carrying out scene emotion correction, judging whether the current scene emotion accords with the user selected scene emotion or not, and entering the scene emotion conversion module if the current scene emotion does not accord with the user selected scene emotion;
the visual stimulation object detection module is used for detecting a visual stimulation object in the visual field focal region and obtaining the position of a potential visual stimulation object in the visual field focal region;
the weakening visual area stimulation module is used for carrying out edge detection on the potential visual stimulation object, fuzzifying the object after the outline of the potential visual object is obtained, and then rendering a new layer of texture for the fuzzified object;
and the real-time scene emotion conversion module is used for converting the potential visual stimulus object after the real scene image is combined with the rendered texture into the scene emotion expression selected by the user, displaying the scene emotion expression on VR equipment in real time, calculating the overall effect of the converted result and judging the conversion effect.
The step of obtaining the visual field focal region by the visual field focal region obtaining module is as follows:
firstly, carrying out scene difference judgment on the transmitted current frame image data and the previous frame image data, specifically comprising the following steps:
let the current frame image data be IcurThe last frame of image data is IpreW is the scene difference of the upper and lower frame data, H1Is the threshold value of the scene difference, g is the result after the convolution operation of the frame data,
convolution operation result of current frame image data:
Figure BDA0001983928800000021
where h is the convolution kernel
Convolution operation result of previous frame image data:
Figure BDA0001983928800000022
where h is the convolution kernel
W=∑(gcur-gpre);
When W is not less than H1Then, calculating to obtain the visual field focus area of the current frame, specifically:
let the camera shoot any eye of the user, the left eye angular coordinate of the eye is (x1, y1), the right eye angular coordinate is (x2, y2), the pupil center coordinate is (x, y), and the horizontal relative position S of the pupil in the eye is:
Figure BDA0001983928800000023
let the length h and width of the current frame image be w, the coordinates of the upper left corner of the current field of view focus region be ((S-0.25) × w,0.25 × h), the coordinates of the lower right corner be ((S +0.25) × w,0.75 × h), and the focus region be a matrix of two coordinates.
The visual stimulation object detection module is used for detecting visual stimulation objects in the visual field focal area and obtaining the position of potential visual stimulation objects in the visual field focal area, and specifically comprises the following steps:
and dividing the visual field focal region image into local images, judging whether the area of the local image corresponding to the current visual field focal region contains visual stimulation objects, and if so, determining the position of the local image in the visual field focal region and the position of the visual field focal region in the scene.
The scene emotion calculation module transmits the image of the focus area of the visual field to a scene calculation network, calculates the emotion distribution vector of the current scene, and then performs scene emotion proofreading on the emotion distribution vector of the scene, wherein the scene emotion proofreading specifically comprises the following steps:
let W2Is the difference value of scene emotion distribution, E is the emotion distribution vector, EmaxIs the maximum value in the emotion distribution vector, EiFor the ith element in the emotion distribution vector, H2 is the threshold of emotion distribution difference value
Figure BDA0001983928800000024
When W2 is not less than H2, EmaxThe corresponding emotion label is an emotion label of the current scene; otherwise, judging that the current scene watched by the current user is the fuzzy emotion label.
The new texture is particularly a texture that the emotion label is pleasant.
The emotion distribution vector is a vector consisting of 6 elements and corresponds to anger, disgust, fear, pleasure, sadness and surprise respectively, the maximum value of the emotion distribution vector does not absolutely represent the emotion label of the current real scene, and the maximum value of the scene emotion distribution vector must meet the condition that the difference value of the scene emotion distribution is greater than a threshold value H2The emotion label corresponding to the maximum value of the emotion distribution vector can be determined to be the emotion label of the current scene, and when the maximum value of the scene emotion distribution vector does not meet the requirement that the difference value is larger than the threshold value H2, the current scene is considered to be a fuzzy emotion scene.
The real-time scene emotion conversion module comprises a real-time scene emotion conversion network, in particular to a confrontation network for training a real scene picture data set with emotion labels.
The invention has the beneficial effects that:
(1) by the module for acquiring the focus area of the visual field, the next operation is not performed on the system when the difference between the upper frame scene and the lower frame scene is insufficient, the focus area of the visual field currently seen by the wearer is extracted on the occasion that the difference between the upper frame scene and the lower frame scene is large enough, and the next operation is performed on the system; emotion calculation is performed using only the focal region of the visual field and potential visual field stimulating objects in the visual field are detected. Therefore, the calculation cost of the system can be effectively reduced, and the fluency of the system on VR equipment is improved;
(2) when the current scene emotion is determined in the scene emotion calculating and detecting module, the scene emotion proofreading operation is used, the emotion label corresponding to the maximum value in the calculated scene emotion distribution vector is not directly used as the scene emotion label, but the emotion label is determined as the current emotion label only when the difference between the maximum value in the scene emotion distribution vector and the values of other elements is large enough, so that the emotion expression of the current scene can be effectively determined to be definitely consistent for most viewers, and the immersion of the scene and the recognition degree of the scene emotion are improved.
(3) According to the visual area blurring module, the visual stimulation of the object to the viewer is weakened in a mode of blurring the object with the potential visual stimulation in the scene and rendering a new texture on the object, and the emotional conversion is not only carried out on the color texture of the scene by the object, so that the visual stimulation of the environment to the wearer is reduced to a greater extent, and the pleasure degree of the wearer is improved;
(4) the real-time scene emotion conversion module comprises integral effect calculation after scene emotion conversion, whether results influencing wearer experience such as unnatural transition between obvious color blocks and colors can appear after scene conversion can be detected, parameters of the real-time scene emotion conversion network are adjusted through the integral effect calculation results, the conversion effect of the system can be continuously improved in the use process of a wearer, and the adaptability of the system to different environments is improved.
Drawings
FIG. 1 is a flow chart of the operation of the present invention;
fig. 2(a) and 2(b) are schematic views of installation positions of the first camera and the second camera according to the present invention, respectively.
Detailed Description
The present invention will be described in further detail with reference to examples and drawings, but the present invention is not limited to these examples.
Examples
An emotion conversion system based on real scene emotion expression is suitable for a VR system. The system is used for converting the emotion of the directional real scene of a user and mainly comprises a visual field focal region acquisition module, a scene emotion calculation module, a visual stimulus object detection module, a weakened visual region stimulus module and a real-time scene emotion conversion module.
As shown in fig. 1, 2(a) and 2(b), when a user uses a VR device equipped with the system, the user enters the system and selects one of six emotion tags of anger, disgust, fear, pleasure, sadness and surprise as scene emotion. After the scene emotion is selected, the system starts to work, the system enters a preparation state, a first camera C1 is arranged in front of the VR device to start collecting real scene images watched by a user, a second camera C2 in the VR device starts shooting images of the left eyeball, real-time positioning is carried out on the pupil of the left eyeball, and the relative position of the pupil of the left eyeball in the left eyeball is calculated. In the embodiment, the left eyeball and the right eyeball of the wearer are consistent by default, and the motion of the left eyeball is selected to represent the motion tracks of the eyes.
The module for acquiring the visual field focal region judges scene differences of image data of the real-time scene video stream watched by the user to obtain the visual field focal region of the current frame image, and specifically comprises the following steps:
firstly, carrying out scene difference judgment on the transmitted current frame image data and the previous frame image data, specifically comprising the following steps:
let the current frame image data be IcurThe last frame of image data is IpreW is the scene difference of the upper and lower frame data, H1Is the threshold value of the scene difference, g is the result after the convolution operation of the frame data,
convolution operation result of current frame image data:
Figure BDA0001983928800000041
where h is the convolution kernel (1)
Convolution operation result of previous frame image data:
Figure BDA0001983928800000042
where h is the convolution kernel (2)
W=∑(gcur-gpre);(3)
When W is not less than H1Then, calculating to obtain the visual field focus area of the current frame, specifically:
let the camera shoot any eye of the user, the left eye angular coordinate of the eye is (x1, y1), the right eye angular coordinate is (x2, y2), the pupil center coordinate is (x, y), and the horizontal relative position S of the pupil in the eye is:
Figure BDA0001983928800000043
let the length h and width of the current frame image be w, the coordinates of the upper left corner of the current field of view focus region be ((S-0.25) × w,0.25 × h), and the coordinates of the lower right corner be ((S +0.25) × w,0.75 × h).
The default system preparation is done and the first operation is to calculate the difference between the current frame and the blank frame.
The scene emotion calculating module and the visual stimulus object detecting module simultaneously transmit the image of the visual field focal region to the internal visual stimulus object detection and scene emotion calculation.
The scene emotion calculation module is used for carrying out scene calculation on the image data of the focus area of the visual field to obtain emotion distribution vectors for carrying out scene emotion correction, judging whether the current scene emotion accords with the user selected scene emotion or not, and entering the scene emotion conversion module if the current scene emotion does not accord with the user selected scene emotion;
the visual stimulation object detection module is used for detecting a visual stimulation object in the visual field focal region and obtaining the position of a potential visual stimulation object in the visual field focal region;
and transmitting the view focus area image to a scene calculation network by scene emotion calculation, calculating a current scene emotion distribution vector, and transmitting the obtained scene emotion distribution vector to a scene emotion proofreading. The scene emotion proofreading formula is as follows:
let W2 be the difference value of scene emotion distribution, E be emotion distribution vector, EmaxIs the maximum value in the emotion distribution vector, EiFor the ith element in the emotion distribution vector, H2 is the threshold of emotion distribution difference value
Figure BDA0001983928800000051
When W2 is not less than H2, EmaxThe corresponding emotion label is an emotion label of the current scene; otherwise, judging that the current scene watched by the current wearer is a fuzzy emotion tag;
the scene emotion calculation network is configured as a convolutional neural network trained on a large-scale real scene picture data set with emotion distribution vectors, the last layer is a full connection layer, the input is a real scene picture, and the output is an emotion distribution vector;
in this embodiment, the visual stimulus object detection network is configured as a convolutional neural network trained based on a large-scale visual stimulus object image training set, and the input is image data, the output is a classification result, and whether a potential visual stimulus object is included is judged;
the weakening visual area visual stimulation module is connected with the visual stimulation object detection, and the visual stimulation object detection transmits the local image position containing the potential stimulation object to the weakening visual area stimulation module. After the approximate position of the potential stimulus object is obtained (i.e., the position of the layout image including the potential stimulus object), the edge detection is performed on the local region to obtain the contour coordinates of the object. Pooling image data of the object according to the contour coordinates of the object to enable the image of the object to be blurred, and rendering a layer of texture with positive emotion labels on the object according to the contour coordinates of the object.
The texture with the positive emotion label in the embodiment refers to a texture with the emotion label being pleasant.
When a potential visual stimulus object exists and the object is weakened through the weakening visual area stimulus module, the input result is an image R, the image data of the potential visual stimulus object is replaced by weakened object image data R with positive emotion label texture at the corresponding position of the real scene image I of the current frame by combining the position of the potential stimulus object in the visual field focal area and the position of the visual field focal area in the real scene shot by the C1 camera, and I is obtainedcombine. Then adding IcombineAnd transmitting the real-time scene emotion conversion module. When the current visual field focus area has no potential visual stimulation object, the real-time field is transmittedScene emotion conversion module IcombineIs equal to I. The real-time scene emotion conversion module is connected with scene emotion proofreading and real-time video stream data in the weakening visual area stimulation module and the scene emotion calculation and detection module. The real-time scene emotion conversion module obtains a real scene I after weakened stimulation from the weakened visual area stimulation module and the real-time video streamcombineObtaining whether the emotion label of the current real scene meets the selected scene emotion from the scene emotion correction, and when the emotion label of the current real scene does not meet the selected scene emotion, displaying the image IcombineI of incoming real-time scene change network to outputcombine_newThe emotion expression of (a) meets the emotion requirement of the selected scene. After the conversion network completes one conversion, the result I is convertedcombine_newDisplaying the result on VR equipment in real time, calculating the integral effect of the converted result, judging whether the converted result has obvious unnatural transition between color blocks and colors and the like, feeding the calculated result back to a real-time scene conversion network, and improving the effect of the network by adjusting the parameters of the network;
the real-time scene emotion conversion network described in this embodiment is configured as a countermeasure network trained based on a large-scale real scene picture data set with emotion tags, and performs scene emotion conversion on an input real scene image according to a selected scene emotion.
In this embodiment, the emotion distribution vector is characterized by a vector consisting of 6 elements, where each element is a probability value corresponding to six Ekman's emotion classifications (Ekman's emotions) tags of anger, disgust, fear, pleasure, sadness, and surprise, respectively, and the Ekman's emotion classifications (Ekman's emotions) are associated with the probability values
Figure BDA0001983928800000061
Has a value of 1. And the maximum value of the emotion distribution vector does not absolutely represent the emotion label of the current real scene, and the emotion label corresponding to the maximum value of the emotion distribution vector can be determined to be the emotion label of the current scene only when the maximum value of the scene emotion distribution vector meets the requirement that the difference value of the scene emotion distribution is greater than the threshold value H2. When the maximum value of the scene emotion distribution vector is notWhen the requirement that the difference value is larger than the threshold value H2 is met, the current scene is considered as the fuzzy emotion scene.
The above embodiments are preferred embodiments of the present invention, but the present invention is not limited to the above embodiments, and any other changes, modifications, substitutions, combinations, and simplifications which do not depart from the spirit and principle of the present invention should be construed as equivalents thereof, and all such changes, modifications, substitutions, combinations, and simplifications are intended to be included in the scope of the present invention.

Claims (7)

1. The emotion conversion system is used for a VR system, the VR system is provided with a first camera for collecting real-time scene videos watched by a user and a second camera for collecting eye movements of the user, and the emotion conversion system comprises
The module for acquiring the visual field focal region judges the scene difference of the image data of the real-time scene video stream watched by the user to obtain the visual field focal region of the current frame image;
the scene emotion calculation module is used for carrying out scene calculation on the image data of the focus area of the visual field to obtain emotion distribution vectors for scene emotion correction, judging whether the current scene emotion accords with the scene emotion selected by the user or not, and entering the real-time scene emotion conversion module if the current scene emotion does not accord with the scene emotion selected by the user;
the visual stimulation object detection module is used for detecting a visual stimulation object in the visual field focal region and obtaining the position of a potential visual stimulation object in the visual field focal region;
the weakening visual area stimulation module is used for carrying out edge detection on the potential visual stimulation object, fuzzifying the object after the outline of the potential visual object is obtained, and then rendering a new layer of texture for the fuzzified object;
and the real-time scene emotion conversion module is used for converting the potential visual stimulus object after the real scene image is combined with the rendered texture into the scene emotion expression selected by the user, displaying the scene emotion expression on VR equipment in real time, calculating the overall effect of the converted result and judging the conversion effect.
2. The emotion conversion system of claim 1, wherein the step of obtaining the focal region of the field of view by the module for obtaining the focal region of the field of view is as follows:
firstly, carrying out scene difference judgment on the transmitted current frame image data and the previous frame image data, specifically comprising the following steps:
let the current frame image data be IcurThe last frame of image data is IpreW is the scene difference of the upper and lower frame data, H1Is the threshold value of the scene difference, g is the result after the convolution operation of the frame data,
convolution operation result of current frame image data:
Figure FDA0003084319900000011
where h is the convolution kernel
Convolution operation result of previous frame image data:
Figure FDA0003084319900000012
where h is the convolution kernel
W=∑(gcur-gpre);
When W is not less than H1Then, calculating to obtain the visual field focus area of the current frame, specifically:
let the camera shoot any eye of the user, the left eye angular coordinate of the eye is (x1, y1), the right eye angular coordinate is (x2, y2), the pupil center coordinate is (x, y), and the horizontal relative position S of the pupil in the eye is:
Figure FDA0003084319900000013
let the length h and width of the current frame image be w, the coordinates of the upper left corner of the current field of view focus region be ((S-0.25) × w,0.25 × h), the coordinates of the lower right corner of the current field of view focus region be ((S +0.25) × w,0.75 × h), and the field of view focus region be a matrix defined by these two coordinates.
3. The emotion conversion system according to claim 1, wherein the visual stimulus object detection module is configured to perform visual stimulus object detection on the visual field focal region and obtain the position of the potential visual stimulus object in the visual field focal region, and specifically:
and dividing the visual field focal region image into local images, judging whether the area of the local image corresponding to the current visual field focal region contains visual stimulation objects, and if so, determining the position of the local image in the visual field focal region and the position of the visual field focal region in the scene.
4. The emotion conversion system according to claim 1, wherein the scene emotion calculation module transmits the view focal region image to a scene calculation network, calculates a current scene emotion distribution vector, and then performs scene emotion proofreading on the scene emotion distribution vector, and the scene emotion proofreading specifically includes:
let W2 be the difference value of scene emotion distribution, E be emotion distribution vector, EmaxIs the maximum value in the emotion distribution vector, EiIs the ith element, H, in the emotion distribution vector2As a threshold for the value of the variance of the emotional distribution, i.e.
Figure FDA0003084319900000021
When W2 is more than or equal to H2When E is greatermaxThe corresponding emotion label is an emotion label of the current scene; otherwise, judging that the current scene watched by the current user is the fuzzy emotion label.
5. An emotion conversion system as claimed in claim 1, wherein the new texture is specifically a texture for which the emotion tag is pleasant.
6. The emotion conversion system of claim 4, wherein the emotion distribution vector is a vector consisting of 6 elements, and corresponds to anger, disgust, fear, pleasure, sadness and surprise, respectively, and the maximum value of the emotion distribution vector does not absolutely represent the emotion label of the current real scene, when the scene emotion distribution vector is presentThe maximum value of the vector satisfies that the difference value of the scene emotion distribution is larger than the threshold value H2Then, the emotion label corresponding to the maximum value of the emotion distribution vector can be determined as the emotion label of the current scene; when the maximum value of the scene emotion distribution vector does not meet the requirement that the difference value is greater than the threshold value H2Then the current scene is considered as a blurry emotion scene.
7. The emotion conversion system of claim 1, wherein the real-time scene emotion conversion module comprises a real-time scene emotion conversion network, in particular a confrontation network trained on real scene picture dataset with emotion labels.
CN201910159043.7A 2019-03-04 2019-03-04 Emotion conversion system based on real scene emotion expression Active CN110018738B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910159043.7A CN110018738B (en) 2019-03-04 2019-03-04 Emotion conversion system based on real scene emotion expression

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910159043.7A CN110018738B (en) 2019-03-04 2019-03-04 Emotion conversion system based on real scene emotion expression

Publications (2)

Publication Number Publication Date
CN110018738A CN110018738A (en) 2019-07-16
CN110018738B true CN110018738B (en) 2021-09-21

Family

ID=67189251

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910159043.7A Active CN110018738B (en) 2019-03-04 2019-03-04 Emotion conversion system based on real scene emotion expression

Country Status (1)

Country Link
CN (1) CN110018738B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111259674B (en) * 2020-01-13 2023-07-25 山东浪潮科学研究院有限公司 Text proofreading and emotion analysis method, equipment and medium based on GAN network
CN111339878B (en) * 2020-02-19 2023-06-20 华南理工大学 Correction type real-time emotion recognition method and system based on eye movement data

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102902356A (en) * 2012-09-18 2013-01-30 华南理工大学 Gesture control system and control method thereof
CN104257380A (en) * 2014-10-22 2015-01-07 南京邮电大学 Electroencephalograph collecting and processing system
CN104915658A (en) * 2015-06-30 2015-09-16 东南大学 Emotion component analyzing method and system based on emotion distribution learning
CN106648107A (en) * 2016-12-30 2017-05-10 包磊 VR scene control method and apparatus
CN106874914A (en) * 2017-01-12 2017-06-20 华南理工大学 A kind of industrial machinery arm visual spatial attention method based on depth convolutional neural networks
CN107578807A (en) * 2017-07-17 2018-01-12 华南理工大学 A kind of creation method of virtual reality emotion stimulating system
CN109416842A (en) * 2016-05-02 2019-03-01 华纳兄弟娱乐公司 Geometric match in virtual reality and augmented reality
CN109887095A (en) * 2019-01-22 2019-06-14 华南理工大学 A kind of emotional distress virtual reality scenario automatic creation system and method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9781392B2 (en) * 2015-09-16 2017-10-03 Intel Corporation Facilitating personal assistance for curation of multimedia and generation of stories at computing devices

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102902356A (en) * 2012-09-18 2013-01-30 华南理工大学 Gesture control system and control method thereof
CN104257380A (en) * 2014-10-22 2015-01-07 南京邮电大学 Electroencephalograph collecting and processing system
CN104915658A (en) * 2015-06-30 2015-09-16 东南大学 Emotion component analyzing method and system based on emotion distribution learning
CN109416842A (en) * 2016-05-02 2019-03-01 华纳兄弟娱乐公司 Geometric match in virtual reality and augmented reality
CN106648107A (en) * 2016-12-30 2017-05-10 包磊 VR scene control method and apparatus
CN106874914A (en) * 2017-01-12 2017-06-20 华南理工大学 A kind of industrial machinery arm visual spatial attention method based on depth convolutional neural networks
CN107578807A (en) * 2017-07-17 2018-01-12 华南理工大学 A kind of creation method of virtual reality emotion stimulating system
CN109887095A (en) * 2019-01-22 2019-06-14 华南理工大学 A kind of emotional distress virtual reality scenario automatic creation system and method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
情感虚拟人技术在人机交互中的应用;林本敬;《中国优秀硕士学位论文全文数据库 信息科技辑》;20100731;全文 *
虚拟现实增强技术综述;周忠等;《中国科学:信息科学》;20150220;全文 *

Also Published As

Publication number Publication date
CN110018738A (en) 2019-07-16

Similar Documents

Publication Publication Date Title
US10939034B2 (en) Imaging system and method for producing images via gaze-based control
CN109952594B (en) Image processing method, device, terminal and storage medium
CN104199544B (en) Advertisement orientation put-on method based on eye tracking
CN112884637B (en) Special effect generation method, device, equipment and storage medium
US20200023157A1 (en) Dynamic digital content delivery in a virtual environment
CN111770299B (en) Method and system for real-time face abstract service of intelligent video conference terminal
US6885761B2 (en) Method and device for generating a person's portrait, method and device for communications, and computer product
CN108449596B (en) 3D stereoscopic image quality evaluation method integrating aesthetics and comfort
KR20210142177A (en) Methods and devices for detecting children's conditions, electronic devices, memory
JP2002501234A (en) Human face tracking system
CN110827193A (en) Panoramic video saliency detection method based on multi-channel features
CN110018738B (en) Emotion conversion system based on real scene emotion expression
CN113192132B (en) Eye catch method and device, storage medium and terminal
WO2018225518A1 (en) Image processing device, image processing method, program, and telecommunication system
CN110837750A (en) Human face quality evaluation method and device
Petajan et al. Robust face feature analysis for automatic speachreading and character animation
CN110378234A (en) Convolutional neural networks thermal imagery face identification method and system based on TensorFlow building
CN114374882A (en) Barrage information processing method and device, terminal and computer-readable storage medium
CN106778576A (en) A kind of action identification method based on SEHM feature graphic sequences
US9323981B2 (en) Face component extraction apparatus, face component extraction method and recording medium in which program for face component extraction method is stored
CN108399358B (en) Expression display method and system for video chat
WO2020263277A1 (en) Landmark temporal smoothing
CN115657859A (en) Intelligent interaction system based on virtual reality
CN114882553A (en) Micro-expression recognition method and system based on deep learning
Chuang et al. Automatic facial feature extraction in model-based coding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant