WO2020125493A1 - Wearable augmented-reality remote video system and video call method - Google Patents

Wearable augmented-reality remote video system and video call method Download PDF

Info

Publication number
WO2020125493A1
WO2020125493A1 PCT/CN2019/124331 CN2019124331W WO2020125493A1 WO 2020125493 A1 WO2020125493 A1 WO 2020125493A1 CN 2019124331 W CN2019124331 W CN 2019124331W WO 2020125493 A1 WO2020125493 A1 WO 2020125493A1
Authority
WO
WIPO (PCT)
Prior art keywords
facial expression
user
current
augmented reality
signals
Prior art date
Application number
PCT/CN2019/124331
Other languages
French (fr)
Chinese (zh)
Inventor
王鹏
Original Assignee
中国科学院深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院深圳先进技术研究院 filed Critical 中国科学院深圳先进技术研究院
Publication of WO2020125493A1 publication Critical patent/WO2020125493A1/en

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/08Detecting, measuring or recording devices for evaluating the respiratory organs
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/318Heart-related electrical modalities, e.g. electrocardiography [ECG]
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/389Electromyography [EMG]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working

Definitions

  • the invention belongs to the technical field of wearable devices, and particularly relates to a wearable augmented reality remote video system and a video call method.
  • mobile terminals now have the function of remote video calling.
  • the user selects the object of the remote video call, after dialing the video call of the other party, through the front camera of the mobile terminal during the chat process
  • the facial expressions of the user can be obtained, and the facial expressions of the other party can be displayed through the display screen of the mobile terminal, so the face-to-face chat experience can be enjoyed.
  • the user needs to hold the mobile terminal and face the screen.
  • the user needs to hold the mobile terminal and make the mobile terminal 'S camera is aimed at the user's face.
  • embodiments of the present invention provide a wearable augmented reality remote video system and a video call method to solve the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face in order to perform effective video when making a remote video call The problem with the call.
  • the first aspect of the present invention provides a wearable augmented reality remote video system, which includes an augmented reality smart glasses system, a facial expression recognition system, and a facial expression construction system;
  • the augmented reality smart glasses system and the facial expression recognition system are set in the augmented reality smart glasses, and the facial expression construction system is set in the wearable smart bracelet;
  • the augmented reality smart glasses system and the facial expression recognition system are respectively A communication connection with the facial expression construction system, and a communication connection with the facial expression recognition system and the facial expression construction system;
  • the augmented reality smart glasses system is used to capture the current environment information of the user and construct a three-dimensional geographic image of the current environment information of the user according to the current environment, and project the three-dimensional geographic image to the augmented reality display of the opposite end of the call;
  • the facial expression recognition system is used to recognize the user's current facial expression and current emotion
  • the facial expression construction system is used to capture and construct a three-dimensional virtual facial expression of a user.
  • a second aspect of the present invention provides a video call method, including:
  • the augmented reality smart glasses system is used to capture the current environment information of the user and construct a three-dimensional geographic image of the current environment information of the user according to the current environment;
  • the present invention provides a wearable augmented reality remote video system and a video call method, by constructing a three-dimensional virtual facial expression image, and then identifying the user's current facial expression and current emotion based on the collected physiological signals, thereby driving the facial expression construction system to
  • the user's current facial expressions and current emotions update the three-dimensional virtual facial expressions, and by constructing a three-dimensional geographic image of the user's environment as the background of the video call, the updated three-dimensional virtual facial expression image and the three-dimensional geographic image are projected to the opposite end of the call
  • the display realizes a holographic and live video call with the opposite end of the call, which effectively solves the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face to make an effective video call when making a remote video call.
  • FIG. 1 is a schematic structural diagram of a wearable augmented reality remote video system according to Embodiment 1 of the present invention
  • FIG. 2 is a schematic structural diagram of an augmented reality smart glasses system 110 of a wearable augmented reality remote video system according to Embodiment 1 of the present invention
  • FIG. 3b is a schematic diagram of the distribution of the collection positions of the face corresponding to the facial expression recognition system 120 of the wearable augmented reality remote video system provided in Embodiment 1 of the present invention
  • FIG. 4 is a schematic structural diagram of a facial expression construction system 130 of a wearable augmented reality remote video system according to Embodiment 1 of the present invention
  • FIG. 5 is a schematic diagram of an implementation process of a video call method according to Embodiment 2 of the present invention.
  • FIG. 6 is a schematic flowchart of an implementation process corresponding to step S101 of Embodiment 2 provided by Embodiment 3 of the present invention.
  • FIG. 7 is a schematic flowchart of an implementation process corresponding to step S102 of Embodiment 2 provided by Embodiment 4 of the present invention.
  • FIG. 8 is a schematic structural diagram of a terminal device according to Embodiment 5 of the present invention.
  • this embodiment provides a wearable augmented reality remote video system 10, which includes an augmented reality smart glasses system 110, a facial expression recognition system 120, and a facial expression construction system 130;
  • the augmented reality smart glasses system 110 and the facial expression recognition system 120 are set in the augmented reality smart glasses, and the facial expression construction system 130 is set in the wearable smart bracelet; the augmented reality smart glasses system 110 is respectively connected with the facial expression recognition system 120 and the facial expression
  • the construction system 130 is in communication connection, and the facial expression recognition system 120 and the facial expression construction system 130 are in communication connection.
  • the augmented reality smart glasses system 110 is used to capture the user's current environmental information and construct a three-dimensional geographic image of the user's current environmental information according to the current environmental information, and project the three-dimensional geographic image onto the augmented reality display of the opposite end of the call.
  • the facial expression recognition system 120 is used to recognize the user's current facial expression and current emotion.
  • the facial expression construction system 130 is used to capture and construct a three-dimensional virtual facial expression of the user.
  • the augmented reality smart glasses system 110, the facial expression recognition system 120, and the facial expression construction system 130 may be wired communication connections or wireless communication connections, which are not limited herein.
  • the facial expression recognition system 120 and the facial expression construction system 130 may be a wired communication connection or a wireless communication connection, which is not limited herein.
  • a wireless connection can be established between the above-mentioned augmented reality smart glasses and the above-mentioned wearable smart bracelet to achieve communication and data transmission.
  • the wireless connection methods include but are not limited to: wireless Bluetooth connection method, wireless LAN connection method, The short-distance wireless communication method and the like enhance display how the smart glasses and the wearable smart bracelet realize the wireless connection, and how to perform communication and data transmission through the wireless connection link, as existing technical means, which will not be repeated here.
  • the above-mentioned augmented reality smart glasses are worn on the user's face, and the current environment information of the user is obtained through the dual current-view wide-angle camera of the augmented reality smart glasses, and through the facial expression recognition system provided on the side of the augmented reality smart glasses near the face
  • the user's physiological information, the wearable smart bracelet is worn on the user's wrist, and the user's face image information is collected through the 3D structured light collection camera provided on the wearable smart bracelet, and according to the collected face image information Construct three-dimensional virtual facial expressions.
  • the augmented reality smart glasses system 110 includes augmented reality smart glasses 111 and a host 112.
  • the augmented reality smart glasses 111 include an optical fiber scanning projector 111A, a binocular infrared gesture recognition camera 111B, an eye tracker 111C, a dual current viewing wide-angle camera 111D, an acceleration sensor 111E, a bone conduction headset, and a noise reduction microphone 111F .
  • the above-mentioned host 112 includes a processor 112A, a controller 112A and an optical fiber scanning projector 111A, a binocular infrared gesture recognition camera 111B, an eye tracker 111C, a dual current viewing wide-angle camera 111D, an acceleration sensor 111E, a bone conduction headset and a noise reduction microphone 111F is electrically connected.
  • the host 112 is used to collect the projection light source and project the three-dimensional geographic image on the augmented reality display at the opposite end of the call through the optical fiber.
  • the processor 112A controls the optical fiber scanning projector 111A, the binocular infrared gesture recognition camera 111B, the eye tracker 111C, the dual current viewing wide-angle camera 111D, the acceleration sensor 111E, the bone conduction headset and the noise reduction microphone 111F to work , And receive the data acquired by the optical fiber scanning projector 111A, the binocular infrared gesture recognition camera 111B, the eye tracker 111C, the dual-view wide-angle camera 111D, the acceleration sensor 111E, the bone conduction headset, and the noise reduction microphone 111F.
  • the optical fiber scanning projector 111A, the binocular infrared gesture recognition camera 111B, the eye tracker 111C, the dual current viewing wide-angle camera 111D, the acceleration sensor 111E, the bone conduction headset, and the noise reduction microphone 111F are respectively provided in the above enhancement Reality smart glasses 111 on.
  • the above-mentioned optical fiber scanning projector 111A is used for optical fiber scanning imaging
  • the binocular infrared gesture recognition camera 111B is used for photographing and recognizing the user's gestures
  • the eye tracker 111C is used to capture the intersection point where the user's eyes are focused
  • the dual current wide-angle camera 111D It is used to capture the current environment and construct a three-dimensional geographic image.
  • the acceleration sensor 111E is used to detect the user's head posture.
  • the bone conduction headset and noise reduction microphone 111F are used to play voice data of the opposite end of the call and receive voice data input by the user.
  • the facial expression recognition system 120 includes an EMG sensor 121, an ESR sensor 122, an ECG signal sensor 123, a respiratory signal sensor 124, and a data processor 125; the data processor 125 and the EMG sensor 121, respectively,
  • the skin electrical response sensor 122, the electrocardiographic signal sensor 123, and the respiratory signal sensor 124 are electrically connected.
  • the facial expression recognition system includes 4 pairs of 8 sensors, namely a first myoelectric sensor 121A, a second myoelectric sensor 121B, a first cutaneous electrical response sensor 122A, and a second cutaneous electrical sensor 122B,
  • the first ECG signal sensor 123A, the second ECG signal sensor 123B, the first respiration signal sensor 124A, and the second respiration signal sensor 124B, the above eight sensors are respectively in contact with the user's face muscle body surface, and are used to obtain the user's Physiological signals.
  • the physiological signals include skin electrical response signals, myoelectric signals, respiratory signals, and electrocardiographic signals.
  • the data processor 125 performs data processing and analysis on the physiological signals collected by each sensor to identify the user's current facial expression and current emotion.
  • the facial expression recognition system collects physiological signals corresponding to 8 collection positions through 8 sensors respectively.
  • the specific first EMG sensor 121A collects EMG signals at 01 collection positions
  • the second EMG sensor 121B collects 02
  • the EMG signal at the collection position the first galvanic skin response sensor 122A acquires the galvanic skin response signal at the 03 acquisition location
  • the second galvanic skin response sensor 122B acquires the galvanic skin response signal at the 04 acquisition location
  • the first ECG signal sensor 123A acquires the 05
  • the second ECG signal sensor 123B collects the ECG signal of 06 collection position
  • the first respiratory signal sensor 124A collects the respiratory signal of position 07
  • the second respiratory signal sensor 124B collects the 08 respiratory signal of collection position.
  • the above sensors can use existing sensors to collect physiological signals.
  • the muscle tension of each collection position is analyzed according to the physiological signals of each collection position, and then the muscle tension of the 8 collection positions is combined
  • the analysis determines the current facial expression of the user. It should be noted that the foregoing analysis of the muscle tension according to physiological signals and the analysis of the muscle tension according to the 8 collection positions to determine the current facial expression of the user can be implemented by an existing analysis algorithm, which will not be repeated here.
  • the facial expression construction system 130 includes a 3D structured light collection camera 131 and an EMG signal sensor 132.
  • the above 3D structured light collection camera 131 and EMG signal sensor 132 are integrated in a wearable smart bracelet.
  • the 3D structured light collection camera 131 is used to collect user's face image information
  • the EMG signal sensor 132 is used Used to obtain user's gesture information.
  • the above-mentioned augmented reality smart glasses system further includes an augmented reality display, and the augmented reality display is used to display virtual video data sent from a call end.
  • the wearable augmented reality remote video system constructs a three-dimensional virtual facial expression image through a facial expression construction system, and then recognizes the user's current facial expression and current emotion according to the physiological signals collected by the facial expression recognition system, thereby driving facial expressions
  • the construction system updates the three-dimensional virtual facial expression according to the user's current facial expression and current emotion, and constructs a three-dimensional geographic image of the user's environment through the augmented reality smart glasses system, which serves as the background of the video call.
  • the updated three-dimensional virtual facial expression image and The three-dimensional geographic image is projected on the display of the opposite end of the call to realize a holographic and live video call with the opposite end of the call, effectively solving the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face when making a remote video call. Problems with effective video calls.
  • this embodiment provides a video call method.
  • the above video call method is implemented based on the above-mentioned wearable enhanced display remote video system, which specifically includes:
  • Step S101 Capture and construct a three-dimensional virtual facial expression image of the user through the facial expression construction system.
  • the facial expression construction system includes a 3D structured light collection camera, which scans the user's face through the 3D structured light collection camera to obtain the user's expression information, and constructs a three-dimensional in a virtual environment based on the user's expression information Virtual facial expression image.
  • Step S102 Recognize the user's current facial expression and current emotion through the facial expression recognition system.
  • the user's current facial expression and current emotion are recognized through the facial expression recognition system.
  • the user's physiological information is collected through the facial expression recognition system, and the user's current facial expression and current emotion are analyzed according to the physiological information.
  • Step S103 Capture the current environment information of the user through the augmented reality smart glasses system and construct a three-dimensional geographic image of the current environment information of the user according to the current environment.
  • the front-view binocular wide-angle camera of the augmented reality smart glasses system of the augmented reality smart glasses system captures the current environment information of the user, and constructs a three-dimensional geographic image of the current environment of the user according to the binocular vision 3D reconstruction algorithm as a chat background .
  • the user's three-dimensional virtual facial expression image can be mapped to the three-dimensional geographic image, that is, the user's avatar can be mapped to the chat background.
  • the dual-view camera also captures the user's gestures and interacts with the call peer according to the user's gestures.
  • an eye tracker captures the intersection area focused by the user's eyes in real time, and switches the chat background according to the intersection area focused by the user's eyes.
  • Step S104 Update the three-dimensional virtual facial expression image based on the user's current facial expression and current emotion recognized by the facial expression recognition system, and merge the updated three-dimensional virtual facial expression image with the three-dimensional geographic image to generate virtual video data .
  • the user's physiological information is collected through the facial expression recognition system, and the user's current facial expression and current emotion are analyzed according to the physiological information to drive the facial expression construction system to update the user's three-dimensional virtual facial expression image.
  • the current facial expression of the user is determined by collecting the myoelectric signal and skin reaction signal of the face, and the collected human respiratory signal and the electrocardiographic signal are jointly identified by the emotion recognition algorithm in multiple dimensions to determine the user's current facial expression mood.
  • the updated 3D virtual facial expression image is fused with the 3D geographic image at the corresponding moment, that is, the 3D geographic image of the current environment is obtained in real time, and then the 3D virtual facial expression image corresponding to the current facial expression and the current emotion is projected Go to the three-dimensional geographic image of the current environment and generate virtual video data.
  • Step S105 Project the virtual video data to the augmented reality display at the opposite end of the call.
  • the virtual video information with the three-dimensional virtual facial expression image and the three-dimensional geographic image is sent to the call peer through the communication link with the call peer, and displayed through the enhanced display monitor of the call peer.
  • the communication of the wearable smart bracelet at the opposite end may also be performed through the wearable smart bracelet and the wearable smart bracelet at the opposite end of the call, which is not limited herein.
  • the synchronized audio data is also transmitted to the opposite end of the call, and the synchronized audio is played through the bone conduction headset of the augmented reality smart glasses at the opposite end of the call data.
  • the above-mentioned video call method before step S101 further includes:
  • Step S106 the user's gesture action is captured through the wearable smart bracelet worn on the user's wrist, and the call object is switched according to the gesture action.
  • the user's gestures and movements are recognized by the myoelectric sensor on the wrist, and the call object is switched according to the gestures of the user's left and right waving hands.
  • motion information is obtained through the myoelectric sensor of the wrist, such as the angle of rotation of the wrist, the swing amplitude of the wrist up and down, and the swing width of the wrist, etc., and the user's gesture movement is analyzed and recognized based on the motion information, and the user's gesture movement Switch to the corresponding call partner.
  • the gesture action is to swing the arm to the left, then select the previous call object of the current call object from the call list, and cut off the video call with the current call object, and establish a video call with the previous call object; if the gesture The action is to swing the arm to the right, then select the next call object of the current call object from the call list, cut off the video call with the current call object, and establish a video call with the next call object. It should be noted that the correspondence between the gesture action and the switching of the call object can be determined according to the settings of the user, and is not limited here.
  • the above-mentioned wearable augmented reality remote video system can also detect the user's head posture through the acceleration sensor of the augmented reality smart glasses, and recognize the user's gestures through the binocular infrared gesture recognition camera of the augmented reality smart glasses, thereby achieving Real-time interaction with the opposite end of the call.
  • the video call method provided in this embodiment can also construct a three-dimensional virtual facial expression image through the facial expression construction system, and then recognize the user's current facial expression and current emotion according to the physiological signals collected by the facial expression recognition system, thereby driving the facial expression construction system
  • Update the three-dimensional virtual facial expression according to the user's current facial expression and current emotion and construct the three-dimensional geographic image of the user's environment through the augmented reality smart glasses system, as the background of the video call, the updated three-dimensional virtual facial expression image and the three-dimensional geographic
  • the image is projected on the display of the opposite end of the call to realize a holographic and live video call with the opposite end of the call, effectively solving the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face in order to perform effective video when making a remote video call The problem with the call.
  • step S101 in Embodiment 1 specifically includes:
  • Step S201 Scan the user's facial expression through the 3D structured light collection camera of the facial expression construction system.
  • Step S202 Construct a three-dimensional virtual facial image of the user in the virtual space according to the scanned facial expression of the user.
  • the user's face is scanned by the 3D structured light collection camera to obtain the user's expression information, and a three-dimensional virtual facial expression image is constructed in the virtual environment according to the user's expression information.
  • step S102 in Embodiment 1 specifically includes:
  • Step S301 Collect physiological signals of the user, the physiological signals including skin electrical response signals, myoelectric signals, respiratory signals and electrocardiographic signals.
  • the physiological information of the user is collected through the eight sensors of the facial expression recognition system, and the physiological signals include skin electrical response signals, myoelectric signals, respiratory signals, and electrocardiographic signals.
  • the EMG signals are collected by the first EMG sensor and the second EMG sensor
  • the skin reaction signals are collected by the first EEG sensor and the second EEG sensor
  • the first EKG signal sensor and the second EKG signal are collected
  • the signal sensor collects the electrocardiographic signal
  • the first respiratory signal sensor and the second respiratory signal sensor collect the respiratory signal.
  • Step S302 Perform multi-dimensional joint recognition based on the user's physiological signal to determine the current facial expression and current emotion of the current user.
  • analyzing the user's current facial expression and current emotion based on physiological information drives the facial expression construction system to update the user's three-dimensional virtual facial expression image.
  • the current facial expression of the user is determined by collecting the myoelectric signal and the skin reaction signal of the face, and the collected human respiratory signal and the electrocardiographic signal are jointly identified by the emotion recognition algorithm in multiple dimensions to determine the user's current facial expression mood.
  • step S302 specifically includes the following steps:
  • Step S3021 Initialize the physiological signal, filter out interference signals and perform data sampling to obtain discrete physiological signals.
  • Step S3022 Extract feature values of various discrete physiological signals.
  • the feature value of each discrete physiological signal is extracted through a statistical feature algorithm, and the feature value of the physiological signal is extracted by specific classification, that is, the feature value of the myoelectric signal, the feature value of the skin reaction signal, and the feature of the respiratory signal are extracted Value and extract the characteristic value of ECG signal.
  • Step S3023 Based on the discrete binary particle swarm optimization algorithm, the effective characteristic signals of various discrete physiological signals are determined according to the characteristic values of various discrete physiological signals.
  • the effective characteristic signals of various discrete physiological signals are determined according to the characteristic values of various discrete physiological signals.
  • Effective feature signals refer to signal features useful for classifying emotions.
  • Step S3024 Perform emotion recognition based on the effective characteristic signals of the various types of discrete physiological signals, and obtain corresponding emotion categories.
  • the effective feature signals of various discrete physiological signals are screened out, and a feature set of valid feature signals is formed, and emotion recognition is performed according to the feature set, and the category to which the pattern with the highest similarity belongs belongs is output as the recognition result.
  • the similarity calculation is performed by various emotion feature values of the emotion database and the effective feature signals of various discrete physiological signals, and the emotion attribute with the highest similarity is used as the emotion category of the current user.
  • each feature is defined as a one-dimensional discrete binary variable of the granule, the variable length is equal to the number of all features, if the i-th bit is 1, then the i-th feature is selected, Otherwise, this feature is blocked. Therefore, each particle represents a different subset of features.
  • (a) Classification error rate Use the features determined in the feature subset to train the classifier, use the results of the classification to evaluate the performance of the particles, and use this to guide the target search of the BPSO algorithm;
  • Each feature subset contains a certain number of features. If the two feature subsets have the same accuracy, the subset with fewer features is selected.
  • Accuracy in the formula represents the accuracy rate obtained by each individual. Ones is the number of "1"s contained in the particle position, that is, the number of selected features.
  • the BPSO algorithm is to find the global minimum value and locate the weight of the accuracy rate to 100,000 to increase the importance of the accuracy rate. High accuracy means that the fitness value is small, and this feature subset is likely to win the competition.
  • k is a constant greater than O, and is a compromise parameter between accuracy and the number of features. The larger the value, the more attention is paid to the number of features. In this embodiment, K takes a value of 0.5.
  • the feature selection method based on the BPSO algorithm is as follows:
  • the feature vector can use a binary vector to represent particles.
  • the expression is as follows:
  • X represents the number of original features. If the i-th bit of x is l, the feature is selected, if it is O, the feature is not selected, and T (speed) represents the probability that the feature is selected.
  • Step 1 Set the group size to m, the maximum number of iterations, the iteration closing value t and the minimum error threshold;
  • Step 2 Initialize the position and velocity of the particle swarm, calculate the fitness of each particle as the initial individual extremum pbesti, and assign the minimum value of all pbesti to gbest as the initial global extremum, the initial value of Vi is set to O;
  • Step 3 According to the classification result of the nearest neighbor classifier, perform performance evaluation on the particle swarm;
  • Step 4 Update the speed and position of particles
  • Step 5 Use Hamming distance to judge the similarity of particles
  • Step 6 Determine whether the minimum error threshold or the maximum number of iterations is reached. If so, perform step 7; otherwise, go to step 3;
  • Step 7 Output the optimal target value and particle, and the feature corresponding to the particle position is the optimal feature combination found.
  • the terminal device 8 is a schematic diagram of a terminal device according to Embodiment 7 of the present invention.
  • the terminal device 8 of this embodiment includes: a processor 80, a memory 81, and a computer program 82 stored in the memory 81 and executable on the processor 80, for example, a program.
  • the processor 80 executes the computer program 82, the steps in the above embodiments of each picture processing method are implemented, for example, steps S101 to S105 shown in FIG. 5.
  • the computer program 82 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 81 and executed by the processor 80 to complete this invention.
  • the one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program 82 in the terminal device 8.
  • the computer program 82 may be divided into an expression construction module, a recognition module, an environment construction module, a fusion module, and a projection module.
  • the specific functions of each module are as follows:
  • Expression construction module used to capture and construct the user's three-dimensional virtual facial expression image through the facial expression construction system
  • Recognition module used to recognize the user's current facial expression and current emotion through the facial expression recognition system
  • An environment construction module configured to capture the user's current environment information through the augmented reality smart glasses system and construct a three-dimensional geographic image of the user's current environment information according to the current environment;
  • the fusion module is used to update the three-dimensional virtual facial expression image based on the current facial expression and current emotion of the user recognized by the facial expression recognition system, and fuse the updated three-dimensional virtual facial expression image with the three-dimensional geographic image to generate a virtual Video data
  • the projection module is used to project the virtual video data to the augmented reality display at the opposite end of the call.
  • the terminal device 8 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud management server.
  • the terminal device may include, but is not limited to, a processor 80 and a memory 81.
  • FIG. 8 is only an example of the terminal device 8 and does not constitute a limitation on the terminal device 8, and may include more or less components than the illustration, or a combination of certain components or different components.
  • the terminal device may further include an input and output device, a network access device, a bus, and the like.
  • the so-called processor 80 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application-specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the memory 81 may be an internal storage unit of the terminal device 8, such as a hard disk or a memory of the terminal device 8.
  • the memory 81 may also be an external storage device of the terminal device 8, for example, a plug-in hard disk equipped on the terminal device 8, a smart memory card (Smart, Media, Card, SMC), and a secure digital (SD) Cards, flash cards, etc.
  • the memory 81 may also include both an internal storage unit of the terminal device 8 and an external storage device.
  • the memory 81 is used to store the computer program and other programs and data required by the terminal device.
  • the memory 81 can also be used to temporarily store data that has been or will be output.
  • each functional unit and module is used as an example for illustration.
  • the above-mentioned functions can be allocated by different functional units
  • Module completion means that the internal structure of the system is divided into different functional units or modules to complete all or part of the functions described above.
  • the functional units and modules in the embodiment may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above integrated unit may use hardware It can also be implemented in the form of software functional units.
  • the specific names of each functional unit and module are only for the purpose of distinguishing each other, and are not used to limit the protection scope of the present application.
  • For the specific working processes of the above units and modules in the wireless terminal reference may be made to the corresponding processes in the foregoing method embodiments, which will not be repeated here.
  • the disclosed system/terminal device and method may be implemented in other ways.
  • the system/terminal device embodiments described above are only schematic.
  • the division of the module or unit is only a logical function division.
  • there may be another division manner such as multiple units Or components can be combined or integrated into another system, or some features can be ignored or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, systems or units, and may be in electrical, mechanical or other forms.
  • the unit described as a separate component may or may not be physically separated, and the component displayed as a unit may or may not be a physical unit, that is, it may be located in one place, or may be distributed to multiple network units on. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or software function unit.
  • the integrated module/unit is implemented in the form of a software functional unit and set as an independent product for sale or use, it may be stored in a computer-readable storage medium.
  • the present invention can realize all or part of the processes in the methods of the above embodiments, and can also be completed by a computer program instructing relevant hardware.
  • the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, the steps of the foregoing method embodiments may be implemented.
  • the computer program includes computer program code, and the computer program code may be in a source code form, an object code form, an executable file, or some intermediate form, etc.
  • the computer-readable medium may include any entity or system capable of carrying the computer program code, a recording medium, a USB flash drive, a mobile hard disk, a magnetic disk, an optical disk, a computer memory, and a read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signals, telecommunications signals and software distribution media, etc.
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • electrical carrier signals telecommunications signals and software distribution media, etc.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Veterinary Medicine (AREA)
  • Surgery (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Pathology (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Physics & Mathematics (AREA)
  • Physiology (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Child & Adolescent Psychology (AREA)
  • Developmental Disabilities (AREA)
  • Educational Technology (AREA)
  • Hospice & Palliative Care (AREA)
  • Psychiatry (AREA)
  • Psychology (AREA)
  • Social Psychology (AREA)
  • Cardiology (AREA)
  • Dentistry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Pulmonology (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The present application is applicable to the technical field of wearable devices, and provides a wearable augmented-reality remote video system and a video call method. The system comprises an augmented-reality smart glasses system, a facial expression recognition system and a facial expression construction system, wherein the augmented-reality smart glasses system is used for capturing the current environment information of a user, and constructing, according to the current environment information, a three-dimensional geographic image of the current environment information of the user; the facial expression recognition system is used for recognizing the current facial expression and current emotion of the user; and the facial expression construction system is used for capturing and constructing a three-dimensional virtual facial expression of the user. A holographic video call is realized by means of recognizing the current facial expression and current emotion of the user, and updating the three-dimensional virtual facial expression according to the current facial expression and current emotion of the user, and by means of constructing the three-dimensional geographic image of an environment where the user is located as the background of a video call, and projecting an image of the updated three-dimensional virtual facial expression and the three-dimensional geographic image into a display of an opposite call end.

Description

一种可穿戴增强现实远程视频系统及视频通话方法Wearable augmented reality remote video system and video call method 技术领域Technical field
本发明属于穿戴设备技术领域,尤其涉及一种可穿戴增强现实远程视频系统及视频通话方法。The invention belongs to the technical field of wearable devices, and particularly relates to a wearable augmented reality remote video system and a video call method.
背景技术Background technique
随着移动通信技术的不断发展,目前移动终端已经具备远程视频通话的功能,用户通过选择远程视频通话的对象,在拨通对方的视频通话后,在聊天的过程中通过移动终端的前置摄像头就能获取用户的面部表情,同时通过移动终端的显示屏就能显示对方的面部表情,因此可以享受面对面的聊天感受。为了使通话过程能够实时看到对方和对方的面部表情,就需要用户手持移动终端并面对屏幕,要使对应能够实时看到用户和用户的面部表情,就需要用户手持移动终端并使移动终端的摄像头对准用户的脸部。With the continuous development of mobile communication technology, mobile terminals now have the function of remote video calling. The user selects the object of the remote video call, after dialing the video call of the other party, through the front camera of the mobile terminal during the chat process The facial expressions of the user can be obtained, and the facial expressions of the other party can be displayed through the display screen of the mobile terminal, so the face-to-face chat experience can be enjoyed. In order to enable the user to see the other party and the other party's facial expressions in real time during the call, the user needs to hold the mobile terminal and face the screen. To enable the user to see the user and the user's facial expressions in real time, the user needs to hold the mobile terminal and make the mobile terminal 'S camera is aimed at the user's face.
综上所述,目前用户进行远程视频通话时需要手持移动终端并使移动终端对着用户脸部才能进行有效视频通话的问题。In summary, at present, when a user performs a remote video call, it is necessary to hold the mobile terminal and make the mobile terminal face the user's face to make an effective video call.
技术问题technical problem
有鉴于此,本发明实施例提供了一种可穿戴增强现实远程视频系统及视频通话方法,以解决目前用户进行远程视频通话时需要手持移动终端并使移动终端对着用户脸部才能进行有效视频通话的问题。In view of this, embodiments of the present invention provide a wearable augmented reality remote video system and a video call method to solve the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face in order to perform effective video when making a remote video call The problem with the call.
技术解决方案Technical solution
本发明的第一方面提供了一种可穿戴增强现实远程视频系统,其包括增强现实智能眼镜系统、面部表情识别系统以及面部表情构建系统;The first aspect of the present invention provides a wearable augmented reality remote video system, which includes an augmented reality smart glasses system, a facial expression recognition system, and a facial expression construction system;
所述增强现实智能眼镜系统和所述面部表情识别系统设置在增强现实智能眼镜中,面部表情构建系统设置在可穿戴智能手环中;所述增强现实智能眼镜系统分别与所述面部表情识别系统和所述面部表情构建系统通信连接,所述面部表情识别系统和所述面部表情构建系统通信连接;The augmented reality smart glasses system and the facial expression recognition system are set in the augmented reality smart glasses, and the facial expression construction system is set in the wearable smart bracelet; the augmented reality smart glasses system and the facial expression recognition system are respectively A communication connection with the facial expression construction system, and a communication connection with the facial expression recognition system and the facial expression construction system;
所述增强现实智能眼镜系统用于捕捉用户当前环境信息并根据所述当前环境构建用户当前环境信息的三维地理图像,并将所述三维地理图像投射至通话对端的增强现实显示器中;The augmented reality smart glasses system is used to capture the current environment information of the user and construct a three-dimensional geographic image of the current environment information of the user according to the current environment, and project the three-dimensional geographic image to the augmented reality display of the opposite end of the call;
所述面部表情识别系统用于识别用户的当前面部表情和当前情感;The facial expression recognition system is used to recognize the user's current facial expression and current emotion;
所述面部表情构建系统用于捕捉和构建用户的三维虚拟面部表情。The facial expression construction system is used to capture and construct a three-dimensional virtual facial expression of a user.
本发明的第二方面提供了一种视频通话方法,包括:A second aspect of the present invention provides a video call method, including:
通过面部表情构建系统捕捉和构建用户的三维虚拟面部表情图像;Capture and construct the user's three-dimensional virtual facial expression image through the facial expression construction system;
通过所述面部表情识别系统识别用户的当前面部表情和当前情感;Identifying the user's current facial expression and current emotion through the facial expression recognition system;
所述增强现实智能眼镜系统用于捕捉用户当前环境信息并根据所述当前环境构建用户当前环境信息的三维地理图像;The augmented reality smart glasses system is used to capture the current environment information of the user and construct a three-dimensional geographic image of the current environment information of the user according to the current environment;
基于面部表情识别系统识别的用户的当前面部表情和当前情感更新所述三维虚拟面部表情图像,并将更新后的三维虚拟面部表情图像与所述三维地理图像进行融合,生成虚拟视频数据;Updating the three-dimensional virtual facial expression image based on the user's current facial expression and current emotion recognized by the facial expression recognition system, and fusing the updated three-dimensional virtual facial expression image with the three-dimensional geographic image to generate virtual video data;
将所述虚拟视频数据投射至通话对端的增强现实显示器中。Project the virtual video data into the augmented reality display at the opposite end of the call.
有益效果Beneficial effect
本发明提供的一种可穿戴增强现实远程视频系统及视频通话方法,通过构建三维虚拟面部表情图像,再根据采集到的生理信号识别用户的当前面部表情和当前情感,从而驱动面部表情构建系统根据用户的当前面部表情和当前情绪更新三维虚拟面部表情,并且通过构建用户所处环境的三维地理图像,作为视频通话的背景,将更新后的三维虚拟面部表情图像与三维地理图像投射到通话对端的显示器中,实现与通话对端进行全息、实景的视频通话,有效地解决了目前用户进行远程视频通话时需要手持移动终端并使移动终端对着用户脸部才能进行有效视频通话的问题。The present invention provides a wearable augmented reality remote video system and a video call method, by constructing a three-dimensional virtual facial expression image, and then identifying the user's current facial expression and current emotion based on the collected physiological signals, thereby driving the facial expression construction system to The user's current facial expressions and current emotions update the three-dimensional virtual facial expressions, and by constructing a three-dimensional geographic image of the user's environment as the background of the video call, the updated three-dimensional virtual facial expression image and the three-dimensional geographic image are projected to the opposite end of the call The display realizes a holographic and live video call with the opposite end of the call, which effectively solves the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face to make an effective video call when making a remote video call.
附图说明BRIEF DESCRIPTION
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly explain the technical solutions in the embodiments of the present invention, the drawings required in the embodiments or the description of the prior art will be briefly introduced below. Obviously, the drawings in the following description are only for the invention. In some embodiments, for those of ordinary skill in the art, without paying creative labor, other drawings may be obtained based on these drawings.
图1是本发明实施例一提供的一种可穿戴增强现实远程视频系统的结构示意图;1 is a schematic structural diagram of a wearable augmented reality remote video system according to Embodiment 1 of the present invention;
图2是本发明实施例一提供的可穿戴增强现实远程视频系统的增强现实智能眼镜系统110的结构示意图;2 is a schematic structural diagram of an augmented reality smart glasses system 110 of a wearable augmented reality remote video system according to Embodiment 1 of the present invention;
图3a是本发明实施例一提供的可穿戴增强现实远程视频系统的面部表情识别系统120在增强现实智能眼镜上的分布示意图;FIG. 3a is a schematic diagram of a facial expression recognition system 120 of a wearable augmented reality remote video system provided in Embodiment 1 of the present invention on augmented reality smart glasses;
图3b是本发明实施例一提供的可穿戴增强现实远程视频系统的面部表情识别系统120对应人脸的采集位置的分布示意图;FIG. 3b is a schematic diagram of the distribution of the collection positions of the face corresponding to the facial expression recognition system 120 of the wearable augmented reality remote video system provided in Embodiment 1 of the present invention;
图4是本发明实施例一提供的可穿戴增强现实远程视频系统的面部表情构建系统130的结构示意图;FIG. 4 is a schematic structural diagram of a facial expression construction system 130 of a wearable augmented reality remote video system according to Embodiment 1 of the present invention;
图5是本发明实施例二提供的一种视频通话方法的实现流程示意图;5 is a schematic diagram of an implementation process of a video call method according to Embodiment 2 of the present invention;
图6是本发明实施例三提供的对应实施例二的步骤S101的实现流程示意图;6 is a schematic flowchart of an implementation process corresponding to step S101 of Embodiment 2 provided by Embodiment 3 of the present invention;
图7是本发明实施例四提供的对应实施例二的步骤S102的实现流程示意图;7 is a schematic flowchart of an implementation process corresponding to step S102 of Embodiment 2 provided by Embodiment 4 of the present invention;
图8是本发明实施例五提供的一种终端设备的结构示意图。8 is a schematic structural diagram of a terminal device according to Embodiment 5 of the present invention.
本发明的实施方式Embodiments of the invention
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本发明实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本发明。在其它情况中,省略对众所周知的系统、系统、电路以及方法的详细说明,以免不必要的细节妨碍本发明的描述。In the following description, for the purpose of illustration rather than limitation, specific details such as specific system structures and technologies are proposed to thoroughly understand the embodiments of the present invention. However, those skilled in the art should understand that the present invention can also be implemented in other embodiments without these specific details. In other cases, detailed descriptions of well-known systems, systems, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary details.
为了说明本发明所述的技术方案,下面通过具体实施例来进行说明。In order to explain the technical solution of the present invention, the following will be described with specific embodiments.
实施例一:Example one:
如图1所示,本实施例提供了一种可穿戴增强现实远程视频系统10,其包括增强现实智能眼镜系统110、面部表情识别系统120以及面部表情构建系统130;As shown in FIG. 1, this embodiment provides a wearable augmented reality remote video system 10, which includes an augmented reality smart glasses system 110, a facial expression recognition system 120, and a facial expression construction system 130;
增强现实智能眼镜系统110和面部表情识别系统120设置在增强现实智能眼镜中,面部表情构建系统130设置在可穿戴智能手环中;增强现实智能眼镜系统110分别与面部表情识别系统120和面部表情构建系统130通信连接,面部表情识别系统120和面部表情构建系统130通信连接。The augmented reality smart glasses system 110 and the facial expression recognition system 120 are set in the augmented reality smart glasses, and the facial expression construction system 130 is set in the wearable smart bracelet; the augmented reality smart glasses system 110 is respectively connected with the facial expression recognition system 120 and the facial expression The construction system 130 is in communication connection, and the facial expression recognition system 120 and the facial expression construction system 130 are in communication connection.
增强现实智能眼镜系统110用于捕捉用户当前环境信息并根据当前环境信息构建用户当前环境信息的三维地理图像,并将三维地理图像投射至通话对端的增强现实显示器中。The augmented reality smart glasses system 110 is used to capture the user's current environmental information and construct a three-dimensional geographic image of the user's current environmental information according to the current environmental information, and project the three-dimensional geographic image onto the augmented reality display of the opposite end of the call.
面部表情识别系统120用于识别用户的当前面部表情和当前情感。The facial expression recognition system 120 is used to recognize the user's current facial expression and current emotion.
面部表情构建系统130用于捕捉和构建用户的三维虚拟面部表情。The facial expression construction system 130 is used to capture and construct a three-dimensional virtual facial expression of the user.
需要说明的是,上述增强现实智能眼镜系统110与上述面部表情识别系统120以及面部表情构建系统130可以是有线通信连接,也可以是无线通信连接,在此不加以限制。上述面部表情识别系统120和面部表情构建系统130可以是有线通信连接,也可以是无线通信连接,在此不加以限制。It should be noted that the augmented reality smart glasses system 110, the facial expression recognition system 120, and the facial expression construction system 130 may be wired communication connections or wireless communication connections, which are not limited herein. The facial expression recognition system 120 and the facial expression construction system 130 may be a wired communication connection or a wireless communication connection, which is not limited herein.
还需要说明的是,上述增强现实智能眼镜与上述可穿戴智能手环之间可以通过建立无线连接来实现通信和数据传输,无线连接方式包括但不限于:无线蓝牙连接方式、无线局域网连接方式、近距离无线通信方式等,增强显示智能眼镜与可穿戴智能手环如何实现无线连接,如何通过无线连接的链路进行通信和数据传输,作为现有的技术手段,在此不加以赘述。It should also be noted that a wireless connection can be established between the above-mentioned augmented reality smart glasses and the above-mentioned wearable smart bracelet to achieve communication and data transmission. The wireless connection methods include but are not limited to: wireless Bluetooth connection method, wireless LAN connection method, The short-distance wireless communication method and the like enhance display how the smart glasses and the wearable smart bracelet realize the wireless connection, and how to perform communication and data transmission through the wireless connection link, as existing technical means, which will not be repeated here.
在具体应用中,上述增强现实智能眼镜佩戴在用户脸上,通过增强现实智能眼镜的双目前视广角摄像头获取用户当前环境信息,通过设置在增强现实智能眼镜靠近人脸侧的面 部表情识别系统获取用户的生理信息,可穿戴智能手环佩戴在所述用户的手腕处,通过设置在可穿戴智能手环的3D结构光采集摄像头采集用户的人脸图像信息,并根据采集到的人脸图像信息构建三维虚拟面部表情。In specific applications, the above-mentioned augmented reality smart glasses are worn on the user's face, and the current environment information of the user is obtained through the dual current-view wide-angle camera of the augmented reality smart glasses, and through the facial expression recognition system provided on the side of the augmented reality smart glasses near the face The user's physiological information, the wearable smart bracelet is worn on the user's wrist, and the user's face image information is collected through the 3D structured light collection camera provided on the wearable smart bracelet, and according to the collected face image information Construct three-dimensional virtual facial expressions.
在一个实施例中,增强现实智能眼镜系统110包括增强现实智能眼镜111和主机112。In one embodiment, the augmented reality smart glasses system 110 includes augmented reality smart glasses 111 and a host 112.
在一个实施例中,增强现实智能眼镜111包括光纤扫描投射器111A、双目红外手势识别摄像头111B、眼动仪111C、双目前视广角摄像头111D、加速度传感器111E、骨传导耳机及降噪麦克风111F。In one embodiment, the augmented reality smart glasses 111 include an optical fiber scanning projector 111A, a binocular infrared gesture recognition camera 111B, an eye tracker 111C, a dual current viewing wide-angle camera 111D, an acceleration sensor 111E, a bone conduction headset, and a noise reduction microphone 111F .
上述主机112包括处理器112A,控制器112A分别与光纤扫描投射器111A、双目红外手势识别摄像头111B、眼动仪111C、双目前视广角摄像头111D、加速度传感器111E、骨传导耳机及降噪麦克风111F电连接。The above-mentioned host 112 includes a processor 112A, a controller 112A and an optical fiber scanning projector 111A, a binocular infrared gesture recognition camera 111B, an eye tracker 111C, a dual current viewing wide-angle camera 111D, an acceleration sensor 111E, a bone conduction headset and a noise reduction microphone 111F is electrically connected.
上述主机112用于对投影光源进行收集,并通过光纤将三维地理图像投影在通话对端的增强现实显示器中。The host 112 is used to collect the projection light source and project the three-dimensional geographic image on the augmented reality display at the opposite end of the call through the optical fiber.
在具体应用中,上述处理器112A控制光纤扫描投射器111A、双目红外手势识别摄像头111B、眼动仪111C、双目前视广角摄像头111D、加速度传感器111E、骨传导耳机及降噪麦克风111F进行工作,并接收光纤扫描投射器111A、双目红外手势识别摄像头111B、眼动仪111C、双目前视广角摄像头111D、加速度传感器111E、骨传导耳机及降噪麦克风111F获取到的数据。In specific applications, the processor 112A controls the optical fiber scanning projector 111A, the binocular infrared gesture recognition camera 111B, the eye tracker 111C, the dual current viewing wide-angle camera 111D, the acceleration sensor 111E, the bone conduction headset and the noise reduction microphone 111F to work , And receive the data acquired by the optical fiber scanning projector 111A, the binocular infrared gesture recognition camera 111B, the eye tracker 111C, the dual-view wide-angle camera 111D, the acceleration sensor 111E, the bone conduction headset, and the noise reduction microphone 111F.
如图2所示,上述光纤扫描投射器111A、双目红外手势识别摄像头111B、眼动仪111C、双目前视广角摄像头111D、加速度传感器111E、骨传导耳机及降噪麦克风111F分别设置在上述增强现实智能眼镜111上。As shown in FIG. 2, the optical fiber scanning projector 111A, the binocular infrared gesture recognition camera 111B, the eye tracker 111C, the dual current viewing wide-angle camera 111D, the acceleration sensor 111E, the bone conduction headset, and the noise reduction microphone 111F are respectively provided in the above enhancement Reality smart glasses 111 on.
上述光纤扫描投射器111A用于光纤扫描成像,双目红外手势识别摄像头111B用于拍摄识别用户的手势动作,眼动仪111C用于捕捉用户的眼睛所聚焦的交点位置,双目前视广角摄像头111D用于拍摄当前环境和构建三维地理图像,加速度传感器111E用于检测用户的头部姿态,骨传导耳机及降噪麦克风111F用于播放通话对端的语音数据和接收用户输入的语音数据。The above-mentioned optical fiber scanning projector 111A is used for optical fiber scanning imaging, the binocular infrared gesture recognition camera 111B is used for photographing and recognizing the user's gestures, the eye tracker 111C is used to capture the intersection point where the user's eyes are focused, and the dual current wide-angle camera 111D It is used to capture the current environment and construct a three-dimensional geographic image. The acceleration sensor 111E is used to detect the user's head posture. The bone conduction headset and noise reduction microphone 111F are used to play voice data of the opposite end of the call and receive voice data input by the user.
在一个实施例中,面部表情识别系统120包括肌电传感器121、皮肤电反应传感器122、心电信号传感器123、呼吸信号传感器124及数据处理器125;数据处理器125分别与肌电传感器121、皮肤电反应传感器122、心电信号传感器123及呼吸信号传感器124电连接。In one embodiment, the facial expression recognition system 120 includes an EMG sensor 121, an ESR sensor 122, an ECG signal sensor 123, a respiratory signal sensor 124, and a data processor 125; the data processor 125 and the EMG sensor 121, respectively, The skin electrical response sensor 122, the electrocardiographic signal sensor 123, and the respiratory signal sensor 124 are electrically connected.
如图3a所示,上述面部表情识别系统包括4对共8个传感器,分别为第一肌电传感器121A、第二肌电传感器121B、第一皮肤电反应传感器122A、第二皮肤电传感器122B、第一心电信号传感器123A、第二心电信号传感器123B、第一呼吸信号传感器124A以及第二 呼吸信号传感器124B,上述8个传感器分别与用户的脸部肌体表面相接触,用于获取用户的生理信号,上述生理信号包括皮肤电反应信号、肌电信号、呼吸信号和心电信号。上述数据处理器125对各个传感器采集到的生理信号进行数据处理和分析,以识别用户的当前面部表情和当前情感。As shown in FIG. 3a, the facial expression recognition system includes 4 pairs of 8 sensors, namely a first myoelectric sensor 121A, a second myoelectric sensor 121B, a first cutaneous electrical response sensor 122A, and a second cutaneous electrical sensor 122B, The first ECG signal sensor 123A, the second ECG signal sensor 123B, the first respiration signal sensor 124A, and the second respiration signal sensor 124B, the above eight sensors are respectively in contact with the user's face muscle body surface, and are used to obtain the user's Physiological signals. The physiological signals include skin electrical response signals, myoelectric signals, respiratory signals, and electrocardiographic signals. The data processor 125 performs data processing and analysis on the physiological signals collected by each sensor to identify the user's current facial expression and current emotion.
如图3b所示,面部表情识别系统通过8个传感器分别对应采集8个采集位置的生理信号,具体的第一肌电传感器121A采集01采集位置的肌电信号,第二肌电传感器121B采集02采集位置的肌电信号,第一皮肤电反应传感器122A采集03采集位置的皮肤电反应信号,第二皮肤电反应传感器122B采集04采集位置的皮肤电反应信号,第一心电信号传感器123A采集05采集位置的心电信号,第二心电信号传感器123B采集06采集位置的心电信号,第一呼吸信号传感器124A采集07位置的呼吸信号,第二呼吸信号传感器124B采集08采集位置的呼吸信号。需要说明的是,上述各个传感器可以使用现有的传感器进行生理信号采集。As shown in FIG. 3b, the facial expression recognition system collects physiological signals corresponding to 8 collection positions through 8 sensors respectively. The specific first EMG sensor 121A collects EMG signals at 01 collection positions, and the second EMG sensor 121B collects 02 The EMG signal at the collection position, the first galvanic skin response sensor 122A acquires the galvanic skin response signal at the 03 acquisition location, the second galvanic skin response sensor 122B acquires the galvanic skin response signal at the 04 acquisition location, and the first ECG signal sensor 123A acquires the 05 The ECG signal of the collection position, the second ECG signal sensor 123B collects the ECG signal of 06 collection position, the first respiratory signal sensor 124A collects the respiratory signal of position 07, and the second respiratory signal sensor 124B collects the 08 respiratory signal of collection position. It should be noted that the above sensors can use existing sensors to collect physiological signals.
在具体应用中,上述8个传感器分别对应采集人脸的8个采集位置的生理信号后,根据各个采集位置的生理信号分析各个采集位置的肌肉紧张程度,再结合8个采集位置的肌肉紧张程度分析确定用户的当前面部表情。需要说明的是,上述根据生理信号分析肌肉紧张程度以及根据8个采集位置的肌肉紧张程度分析确定用户的当前面部表情可以通过现有的分析算法来实现,在此不加以赘述。In specific applications, after the above 8 sensors respectively correspond to the physiological signals of the 8 collection positions of the face, the muscle tension of each collection position is analyzed according to the physiological signals of each collection position, and then the muscle tension of the 8 collection positions is combined The analysis determines the current facial expression of the user. It should be noted that the foregoing analysis of the muscle tension according to physiological signals and the analysis of the muscle tension according to the 8 collection positions to determine the current facial expression of the user can be implemented by an existing analysis algorithm, which will not be repeated here.
在一个实施例中,面部表情构建系统130包括3D结构光采集摄像头131和肌电信号传感器132。In one embodiment, the facial expression construction system 130 includes a 3D structured light collection camera 131 and an EMG signal sensor 132.
如图4所示,上述3D结构光采集摄像头131和肌电信号传感器132集成在可穿戴智能手环中,3D结构光采集摄像头131用于采集用户的人脸图像信息,肌电信号传感器132用于获取用户的手势信息。As shown in FIG. 4, the above 3D structured light collection camera 131 and EMG signal sensor 132 are integrated in a wearable smart bracelet. The 3D structured light collection camera 131 is used to collect user's face image information, and the EMG signal sensor 132 is used Used to obtain user's gesture information.
在一个实施例中,上述增强现实智能眼镜系统还包括增强现实显示器,上述增强现实显示器用于显示通话对端发送过来的虚拟视频数据。In one embodiment, the above-mentioned augmented reality smart glasses system further includes an augmented reality display, and the augmented reality display is used to display virtual video data sent from a call end.
本实施例提供的可穿戴增强现实远程视频系统,通过面部表情构建系统构建三维虚拟面部表情图像,再根据面部表情识别系统采集到的生理信号识别用户的当前面部表情和当前情感,从而驱动面部表情构建系统根据用户的当前面部表情和当前情绪更新三维虚拟面部表情,并且通过增强现实智能眼镜系统构建用户所处环境的三维地理图像,作为视频通话的背景,将更新后的三维虚拟面部表情图像与三维地理图像投射到通话对端的显示器中,实现与通话对端进行全息、实景的视频通话,有效地解决了目前用户进行远程视频通话时需要手持移动终端并使移动终端对着用户脸部才能进行有效视频通话的问题。The wearable augmented reality remote video system provided in this embodiment constructs a three-dimensional virtual facial expression image through a facial expression construction system, and then recognizes the user's current facial expression and current emotion according to the physiological signals collected by the facial expression recognition system, thereby driving facial expressions The construction system updates the three-dimensional virtual facial expression according to the user's current facial expression and current emotion, and constructs a three-dimensional geographic image of the user's environment through the augmented reality smart glasses system, which serves as the background of the video call. The updated three-dimensional virtual facial expression image and The three-dimensional geographic image is projected on the display of the opposite end of the call to realize a holographic and live video call with the opposite end of the call, effectively solving the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face when making a remote video call. Problems with effective video calls.
实施例二:Example two:
如图5所示,本实施例提供了一种视频通话方法,上述视频通话方式基于上述可穿戴增强显示远程视频系统来实现,其具体包括:As shown in FIG. 5, this embodiment provides a video call method. The above video call method is implemented based on the above-mentioned wearable enhanced display remote video system, which specifically includes:
步骤S101:通过面部表情构建系统捕捉和构建用户的三维虚拟面部表情图像。Step S101: Capture and construct a three-dimensional virtual facial expression image of the user through the facial expression construction system.
在具体应用中,面部表情构建系统包括3D结构光采集摄像头,通过3D结构光采集摄像头对着用户的脸部进行扫描,获取用户的表情信息,并根据用户的表情信息在虚拟环境中构建一个三维虚拟面部表情图像。In specific applications, the facial expression construction system includes a 3D structured light collection camera, which scans the user's face through the 3D structured light collection camera to obtain the user's expression information, and constructs a three-dimensional in a virtual environment based on the user's expression information Virtual facial expression image.
步骤S102:通过面部表情识别系统识别用户的当前面部表情和当前情感。Step S102: Recognize the user's current facial expression and current emotion through the facial expression recognition system.
在具体应用中,当生成了三维虚拟面部表情图像后,通过面部表情识别系统对用户的当前面部表情和当前情感进行识别。通过面部表情识别系统采集用户的生理信息,根据生理信息分析用户当前面部表情和当前情感。In a specific application, after the three-dimensional virtual facial expression image is generated, the user's current facial expression and current emotion are recognized through the facial expression recognition system. The user's physiological information is collected through the facial expression recognition system, and the user's current facial expression and current emotion are analyzed according to the physiological information.
步骤S103:通过增强现实智能眼镜系统捕捉用户当前环境信息并根据所述当前环境构建用户当前环境信息的三维地理图像。Step S103: Capture the current environment information of the user through the augmented reality smart glasses system and construct a three-dimensional geographic image of the current environment information of the user according to the current environment.
在具体应用中,通过增强现实智能眼镜系统的增强现实智能眼镜的前视双目广角摄像头拍摄获取用户的当前环境信息,根据双目视觉三维重建算法构建用户当前环境的三维地理图像,作为聊天背景。与此同时,可以将用户的三维虚拟面部表情图像映射到上述三维地理图像中,即将用户的头像映射到聊天背景中。In specific applications, the front-view binocular wide-angle camera of the augmented reality smart glasses system of the augmented reality smart glasses system captures the current environment information of the user, and constructs a three-dimensional geographic image of the current environment of the user according to the binocular vision 3D reconstruction algorithm as a chat background . At the same time, the user's three-dimensional virtual facial expression image can be mapped to the three-dimensional geographic image, that is, the user's avatar can be mapped to the chat background.
在具体应用中,双目前视摄像头还用户捕捉用户的手势动作,根据用户的手势动作与通话对端进行交互。In a specific application, the dual-view camera also captures the user's gestures and interacts with the call peer according to the user's gestures.
在具体应用中,通过眼动仪实时捕捉用户的眼睛聚焦的交点区域,并根据用户眼睛聚焦的交点区域切换聊天背景。In a specific application, an eye tracker captures the intersection area focused by the user's eyes in real time, and switches the chat background according to the intersection area focused by the user's eyes.
步骤S104:基于面部表情识别系统识别的用户的当前面部表情和当前情感更新所述三维虚拟面部表情图像,并将更新后的三维虚拟面部表情图像与所述三维地理图像进行融合,生成虚拟视频数据。Step S104: Update the three-dimensional virtual facial expression image based on the user's current facial expression and current emotion recognized by the facial expression recognition system, and merge the updated three-dimensional virtual facial expression image with the three-dimensional geographic image to generate virtual video data .
在具体应用中,通过面部表情识别系统采集用户的生理信息,根据生理信息分析用户当前面部表情和当前情感驱使面部表情构建系统更新用户的三维虚拟面部表情图像。具体的,通过对面部的肌电信号和皮肤反应信号的采集,判断用户的当前面部表情,通过采集到的人体呼吸信号和心电信号通过情感识别算法进行多维度联合识别,从而判断用户的当前情绪。In specific applications, the user's physiological information is collected through the facial expression recognition system, and the user's current facial expression and current emotion are analyzed according to the physiological information to drive the facial expression construction system to update the user's three-dimensional virtual facial expression image. Specifically, the current facial expression of the user is determined by collecting the myoelectric signal and skin reaction signal of the face, and the collected human respiratory signal and the electrocardiographic signal are jointly identified by the emotion recognition algorithm in multiple dimensions to determine the user's current facial expression mood.
在具体应用中,将更新后的三维虚拟面部表情图像与对应时刻的三维地理图像进行融合,即实时获取当前环境的三维地理图像,再将当前面部表情和当前情感对应的三维虚拟 面部表情图像投射到当前环境的三维地理图像中,生成虚拟视频数据。In specific applications, the updated 3D virtual facial expression image is fused with the 3D geographic image at the corresponding moment, that is, the 3D geographic image of the current environment is obtained in real time, and then the 3D virtual facial expression image corresponding to the current facial expression and the current emotion is projected Go to the three-dimensional geographic image of the current environment and generate virtual video data.
步骤S105:将所述虚拟视频数据投射至通话对端的增强现实显示器中。Step S105: Project the virtual video data to the augmented reality display at the opposite end of the call.
在具体应用中,通过与通话对端的通信链路将具备三维虚拟面部表情图像与三维地理图像的虚拟视频信息发送给通话对端,并通过通话对端的增强显示显示器进行显示。In a specific application, the virtual video information with the three-dimensional virtual facial expression image and the three-dimensional geographic image is sent to the call peer through the communication link with the call peer, and displayed through the enhanced display monitor of the call peer.
在具体应用中,可以通过增强现实智能眼镜与通话对端的增强现实智能眼镜进行通信,也可以通过可穿戴智能手环与通话对端的增强现实智能眼镜进行通信,还可以通过增强现实智能眼镜与通话对端的可穿戴智能手环进行通信,还可以是通过可穿戴智能手环与通话对端的可穿戴智能手环进行通信,在此不加以限制。In specific applications, you can communicate with the augmented reality smart glasses at the opposite end of the call through augmented reality smart glasses, or you can communicate with the augmented reality smart glasses at the opposite end of the call through the wearable smart bracelet, and you can also communicate with the call through the augmented reality smart glasses The communication of the wearable smart bracelet at the opposite end may also be performed through the wearable smart bracelet and the wearable smart bracelet at the opposite end of the call, which is not limited herein.
在具体应用中,在将虚拟视频数据投射到通话对端的增强现实显示器的同时,将同步的音频数据也传输至通话对端,并通过通话对端的增强现实智能眼镜的骨传导耳机播放同步播放音频数据。In specific applications, while projecting virtual video data to the augmented reality display at the opposite end of the call, the synchronized audio data is also transmitted to the opposite end of the call, and the synchronized audio is played through the bone conduction headset of the augmented reality smart glasses at the opposite end of the call data.
在一个实施例中,上述视频通话方法在步骤S101之前还包括:In one embodiment, the above-mentioned video call method before step S101 further includes:
步骤S106:通过佩戴在用户腕部的可穿戴智能手环捕捉用户的手势动作,根据所述手势动作切换通话对象。Step S106: the user's gesture action is captured through the wearable smart bracelet worn on the user's wrist, and the call object is switched according to the gesture action.
在具体应用中,通过腕部的肌电传感器识别用户的手势动作,根据用户的左右摆手的手势动作切换通话对象。In a specific application, the user's gestures and movements are recognized by the myoelectric sensor on the wrist, and the call object is switched according to the gestures of the user's left and right waving hands.
在具体应用中,通过腕部的肌电传感器获取动作信息,如腕部旋转角度,腕部上下摆动幅度及腕部左右摆动幅度等,根据动作信息分析识别用户的手势动作,根据用户的手势动作切换到对应的通话对象。In specific applications, motion information is obtained through the myoelectric sensor of the wrist, such as the angle of rotation of the wrist, the swing amplitude of the wrist up and down, and the swing width of the wrist, etc., and the user's gesture movement is analyzed and recognized based on the motion information, and the user's gesture movement Switch to the corresponding call partner.
示例性的,若手势动作为向左摆动手臂,则从通话列表中选择当前通话对象的上一个通话对象,并切断与当前通话对象的视频通话,建立与上一个通话对象的视频通话;若手势动作为向右摆动手臂,则从通话列表中选择当前通话对象的下一个通话对象,并切断与当前通话对象的视频通话,建立与下一个通话对象的视频通话。需要说明的是,手势动作与切换通话对象的对应关系可以根据用户的设置进行确定,在此不加以限制。Exemplarily, if the gesture action is to swing the arm to the left, then select the previous call object of the current call object from the call list, and cut off the video call with the current call object, and establish a video call with the previous call object; if the gesture The action is to swing the arm to the right, then select the next call object of the current call object from the call list, cut off the video call with the current call object, and establish a video call with the next call object. It should be noted that the correspondence between the gesture action and the switching of the call object can be determined according to the settings of the user, and is not limited here.
在具体应用中,上述可穿戴增强现实远程视频系统还可以通过增强现实智能眼镜的加速度传感器检测用户的头部姿态,通过增强现实智能眼镜的双目红外手势识别摄像头识别用户的手势动作,进而实现与通话对端的实时交互。In specific applications, the above-mentioned wearable augmented reality remote video system can also detect the user's head posture through the acceleration sensor of the augmented reality smart glasses, and recognize the user's gestures through the binocular infrared gesture recognition camera of the augmented reality smart glasses, thereby achieving Real-time interaction with the opposite end of the call.
本实施例提供的视频通话方法,同样能够通过面部表情构建系统构建三维虚拟面部表情图像,再根据面部表情识别系统采集到的生理信号识别用户的当前面部表情和当前情感,从而驱动面部表情构建系统根据用户的当前面部表情和当前情绪更新三维虚拟面部表情,并且通过增强现实智能眼镜系统构建用户所处环境的三维地理图像,作为视频通话的背景, 将更新后的三维虚拟面部表情图像与三维地理图像投射到通话对端的显示器中,实现与通话对端进行全息、实景的视频通话,有效地解决了目前用户进行远程视频通话时需要手持移动终端并使移动终端对着用户脸部才能进行有效视频通话的问题。The video call method provided in this embodiment can also construct a three-dimensional virtual facial expression image through the facial expression construction system, and then recognize the user's current facial expression and current emotion according to the physiological signals collected by the facial expression recognition system, thereby driving the facial expression construction system Update the three-dimensional virtual facial expression according to the user's current facial expression and current emotion, and construct the three-dimensional geographic image of the user's environment through the augmented reality smart glasses system, as the background of the video call, the updated three-dimensional virtual facial expression image and the three-dimensional geographic The image is projected on the display of the opposite end of the call to realize a holographic and live video call with the opposite end of the call, effectively solving the problem that the current user needs to hold the mobile terminal and make the mobile terminal face the user's face in order to perform effective video when making a remote video call The problem with the call.
实施例三:Example three:
如图6所示,在本实施例中,实施例一中的步骤S101具体包括:As shown in FIG. 6, in this embodiment, step S101 in Embodiment 1 specifically includes:
步骤S201:通过面部表情构建系统的3D结构光采集摄像头对用户的面部表情进行扫描。Step S201: Scan the user's facial expression through the 3D structured light collection camera of the facial expression construction system.
步骤S202:并根据扫描到的用户的面部表情在虚拟空间中构建所述用户的三维虚拟面部图像。Step S202: Construct a three-dimensional virtual facial image of the user in the virtual space according to the scanned facial expression of the user.
在具体应用中,通过3D结构光采集摄像头对着用户的脸部进行扫描,获取用户的表情信息,并根据用户的表情信息在虚拟环境中构建一个三维虚拟面部表情图像。In specific applications, the user's face is scanned by the 3D structured light collection camera to obtain the user's expression information, and a three-dimensional virtual facial expression image is constructed in the virtual environment according to the user's expression information.
实施例四:Example 4:
如图7所示,在本实施例中,实施例一中的步骤S102具体包括:As shown in FIG. 7, in this embodiment, step S102 in Embodiment 1 specifically includes:
步骤S301:采集所述用户的生理信号,所述生理信号包括皮肤电反应信号、肌电信号、呼吸信号和心电信号。Step S301: Collect physiological signals of the user, the physiological signals including skin electrical response signals, myoelectric signals, respiratory signals and electrocardiographic signals.
在具体应用中,通过面部表情识别系统的8个传感器采集用户的生理信息,生理信号包括皮肤电反应信号、肌电信号、呼吸信号和心电信号。具体的,通过第一肌电传感器和第二肌电传感器采集肌电信号,通过第一皮肤电反应传感器和第二皮肤电传感器采集皮肤反应信号,通过第一心电信号传感器和第二心电信号传感器采集心电信号,通过第一呼吸信号传感器和第二呼吸信号传感器采集呼吸信号。In specific applications, the physiological information of the user is collected through the eight sensors of the facial expression recognition system, and the physiological signals include skin electrical response signals, myoelectric signals, respiratory signals, and electrocardiographic signals. Specifically, the EMG signals are collected by the first EMG sensor and the second EMG sensor, the skin reaction signals are collected by the first EEG sensor and the second EEG sensor, and the first EKG signal sensor and the second EKG signal are collected The signal sensor collects the electrocardiographic signal, and the first respiratory signal sensor and the second respiratory signal sensor collect the respiratory signal.
步骤S302:根据所述用户的生理信号进行多维度联合识别,判断当前用户的当前面部表情和当前情感。Step S302: Perform multi-dimensional joint recognition based on the user's physiological signal to determine the current facial expression and current emotion of the current user.
在具体应用中,根据生理信息分析用户当前面部表情和当前情感驱使面部表情构建系统更新用户的三维虚拟面部表情图像。In a specific application, analyzing the user's current facial expression and current emotion based on physiological information drives the facial expression construction system to update the user's three-dimensional virtual facial expression image.
具体的,通过对面部的肌电信号和皮肤反应信号的采集,判断用户的当前面部表情,通过采集到的人体呼吸信号和心电信号通过情感识别算法进行多维度联合识别,从而判断用户的当前情绪。Specifically, the current facial expression of the user is determined by collecting the myoelectric signal and the skin reaction signal of the face, and the collected human respiratory signal and the electrocardiographic signal are jointly identified by the emotion recognition algorithm in multiple dimensions to determine the user's current facial expression mood.
在一个实施例中,上述步骤S302具体包括以下步骤:In one embodiment, the above step S302 specifically includes the following steps:
步骤S3021:对所述生理信号进行初始化处理,滤除干扰信号并进行数据采样,获取离散生理信号。步骤S3022:提取各类离散生理信号的特征值。Step S3021: Initialize the physiological signal, filter out interference signals and perform data sampling to obtain discrete physiological signals. Step S3022: Extract feature values of various discrete physiological signals.
在具体应用中,通过统计特征算法提取各个离散生理信号的特征值,具体的分类提取 生理信号的特征值,即提取肌电信号的特征值、提取皮肤反应信号的特征值、提取呼吸信号的特征值以及提取心电信号的特征值。In specific applications, the feature value of each discrete physiological signal is extracted through a statistical feature algorithm, and the feature value of the physiological signal is extracted by specific classification, that is, the feature value of the myoelectric signal, the feature value of the skin reaction signal, and the feature of the respiratory signal are extracted Value and extract the characteristic value of ECG signal.
步骤S3023:基于离散二进制粒子群算法根据各类离散生理信号的特征值确定各类离散生理信号的有效特征信号。Step S3023: Based on the discrete binary particle swarm optimization algorithm, the effective characteristic signals of various discrete physiological signals are determined according to the characteristic values of various discrete physiological signals.
在具体应用中,为了减小计算量,基于离散二进制粒子群算法根据各类离散生理信号的特征值确定各类离散生理信号的有效特征信号。有效特征信号是指对分类情感有用的信号特征。In specific applications, in order to reduce the amount of calculation, based on the discrete binary particle swarm algorithm, the effective characteristic signals of various discrete physiological signals are determined according to the characteristic values of various discrete physiological signals. Effective feature signals refer to signal features useful for classifying emotions.
步骤S3024:根据所述各类离散生理信号的有效特征信号进行情感识别,获取对应的情感类别。Step S3024: Perform emotion recognition based on the effective characteristic signals of the various types of discrete physiological signals, and obtain corresponding emotion categories.
在具体应用中,筛选出各类的离散生理信号的有效特征信号,并形成有效特征信号的特征集,根据特征集进行情感识别,将相似度最高的模式所属的类别作为识别结果输出。通过情感数据库的各类情感特征值与各类离散生理信号的有效特征信号进行相似度计算,将相似度最高的情感属性作为当前用户的情感类别。In specific applications, the effective feature signals of various discrete physiological signals are screened out, and a feature set of valid feature signals is formed, and emotion recognition is performed according to the feature set, and the category to which the pattern with the highest similarity belongs belongs is output as the recognition result. The similarity calculation is performed by various emotion feature values of the emotion database and the effective feature signals of various discrete physiological signals, and the emotion attribute with the highest similarity is used as the emotion category of the current user.
在具体应用中,上述提取各类离散生理信号的特征值的具体实施方式如下:In specific applications, the specific implementation manners for extracting the characteristic values of various types of discrete physiological signals are as follows:
(1)编码:根据特征选择的特点,把每一个特征定义为粒的一维离散二进制变量,变量长度就等于所有特征的数量,如果第i位为1,那么第i个特征就被选中,否则这个特征就被屏蔽。因此,每一个粒子就代表了一个不同的特征子集。(1) Coding: According to the characteristics of feature selection, each feature is defined as a one-dimensional discrete binary variable of the granule, the variable length is equal to the number of all features, if the i-th bit is 1, then the i-th feature is selected, Otherwise, this feature is blocked. Therefore, each particle represents a different subset of features.
(2)初始化:为了获得不同数量的特征,首先随机产生每个粒子所含“1”的个数,然后再把这些“1”随机分布在这个粒子的所有维度中。(2) Initialization: In order to obtain different numbers of features, the number of "1"s contained in each particle is first randomly generated, and then these "1"s are randomly distributed in all dimensions of this particle.
(3)适应度的评价:基于BPSO算法搜索目标得到全局最小值,上述适应度的评价包含两部分内容:(3) Evaluation of fitness: Search for the target based on the BPSO algorithm to obtain the global minimum. The evaluation of fitness above includes two parts:
(a)分类错误率。使用特征子集中确定的特征来训练分类器,用分类的结果来评价粒子的性能,并以此来指导BPSO算法的目标搜索;(a) Classification error rate. Use the features determined in the feature subset to train the classifier, use the results of the classification to evaluate the performance of the particles, and use this to guide the target search of the BPSO algorithm;
(b)使用的特征数量。每个特征子集包含一定数量的特征,如果两个特征子集获得的准确度相同,包含特征比较少的子集被选中。(b) The number of features used. Each feature subset contains a certain number of features. If the two feature subsets have the same accuracy, the subset with fewer features is selected.
需要说明的是在准确率和特征数量这两个因素中,需要重点考虑的是准确率,因此适应度函数确定为如下形式:It should be noted that in the two factors of accuracy and number of features, the accuracy is the key consideration, so the fitness function is determined as follows:
fitness=10 5×(1-Accuracy)+k×Ones fitness=10 5 ×(1-Accuracy)+k×Ones
式中Accuracy表示每个个体获得的准确率,ones是粒子位置中包含“1”的数量,即被选中的特征的数量。BPSO算法是寻找全局最小值,准确率的权值定位100000,以提高准 确率的重要性。高准确率就意味着适应度值小,该特征子集就有可能在竞争中获胜。k是大于O的常数,是准确率和特征数量的折中参数,它的值越大,表示特征数量越获得重视,在本实施例中K取值为0.5。Accuracy in the formula represents the accuracy rate obtained by each individual. Ones is the number of "1"s contained in the particle position, that is, the number of selected features. The BPSO algorithm is to find the global minimum value and locate the weight of the accuracy rate to 100,000 to increase the importance of the accuracy rate. High accuracy means that the fitness value is small, and this feature subset is likely to win the competition. k is a constant greater than O, and is a compromise parameter between accuracy and the number of features. The larger the value, the more attention is paid to the number of features. In this embodiment, K takes a value of 0.5.
在具体应用中,基于BPSO算法的特征选择方法如下:In specific applications, the feature selection method based on the BPSO algorithm is as follows:
首先,在离散二进制PSO算法中,特征向量可以用一个二进制向量来表示粒子。表达式如下:First of all, in the discrete binary PSO algorithm, the feature vector can use a binary vector to represent particles. The expression is as follows:
X=(x 1,x 2,…,x D) T,x i∈{0,1} X = (x 1 , x 2 ,..., x D ) T , x i ∈{0, 1}
X表示原始特征的数量,如果x的第i位为l,则此特征被选中,如果为O,则此特征未被选中,T(速度)表示特征被选中的概率。X represents the number of original features. If the i-th bit of x is l, the feature is selected, if it is O, the feature is not selected, and T (speed) represents the probability that the feature is selected.
因此,上述提取特征值的步骤如下所示:Therefore, the above steps for extracting feature values are as follows:
Step 1:设置群体大小为m、最大迭代次数、迭代闭值t和最小误差阂值;Step 1: Set the group size to m, the maximum number of iterations, the iteration closing value t and the minimum error threshold;
Step 2:初始化粒子群的位置和速度,计算每个粒子的适配度将其作为初始个体极值pbesti,将所有pbesti中的最小值赋给gbest作为初始全局极值,Vi的初始值设为O;Step 2: Initialize the position and velocity of the particle swarm, calculate the fitness of each particle as the initial individual extremum pbesti, and assign the minimum value of all pbesti to gbest as the initial global extremum, the initial value of Vi is set to O;
Step 3:根据最近邻分类器的分类结果,对粒子群进行性能评价;Step 3: According to the classification result of the nearest neighbor classifier, perform performance evaluation on the particle swarm;
Step 4:更新粒子的速度和位置;Step 4: Update the speed and position of particles;
Step 5:用海明距离判断粒子的相似性;Step 5: Use Hamming distance to judge the similarity of particles;
Step 6:判断是否达到最小误差闽值或最大迭代次数,如果达到,执行step7;否则转向step3;Step 6: Determine whether the minimum error threshold or the maximum number of iterations is reached. If so, perform step 7; otherwise, go to step 3;
Step 7:输出最优目标值和粒子,该粒子位置所对应的特征即为找到的最优特征组合。Step 7: Output the optimal target value and particle, and the feature corresponding to the particle position is the optimal feature combination found.
实施例五:Example 5:
图8是本发明实施例七提供的终端设备的示意图。如图8所示,该实施例的终端设备8包括:处理器80、存储器81以及存储在所述存储器81中并可在所述处理器80上运行的计算机程序82,例如程序。所述处理器80执行所述计算机程序82时实现上述各个图片处理方法实施例中的步骤,例如图5所示的步骤S101至S105。8 is a schematic diagram of a terminal device according to Embodiment 7 of the present invention. As shown in FIG. 8, the terminal device 8 of this embodiment includes: a processor 80, a memory 81, and a computer program 82 stored in the memory 81 and executable on the processor 80, for example, a program. When the processor 80 executes the computer program 82, the steps in the above embodiments of each picture processing method are implemented, for example, steps S101 to S105 shown in FIG. 5.
示例性的,所述计算机程序82可以被分割成一个或多个模块/单元,所述一个或者多个模块/单元被存储在所述存储器81中,并由所述处理器80执行,以完成本发明。所述一个或多个模块/单元可以是能够完成特定功能的一系列计算机程序指令段,该指令段用于描述所述计算机程序82在所述终端设备8中的执行过程。例如,所述计算机程序82可以被分割成表情构建模块、识别模块、环境构建模块、融合模块以及投射模块,各模块具体功能如下:Exemplarily, the computer program 82 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 81 and executed by the processor 80 to complete this invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program 82 in the terminal device 8. For example, the computer program 82 may be divided into an expression construction module, a recognition module, an environment construction module, a fusion module, and a projection module. The specific functions of each module are as follows:
表情构建模块,用于通过面部表情构建系统捕捉和构建用户的三维虚拟面部表情图像;Expression construction module, used to capture and construct the user's three-dimensional virtual facial expression image through the facial expression construction system;
识别模块,用于通过面部表情识别系统识别用户的当前面部表情和当前情感;Recognition module, used to recognize the user's current facial expression and current emotion through the facial expression recognition system;
环境构建模块,用于通过增强现实智能眼镜系统捕捉用户当前环境信息并根据所述当前环境构建用户当前环境信息的三维地理图像;An environment construction module, configured to capture the user's current environment information through the augmented reality smart glasses system and construct a three-dimensional geographic image of the user's current environment information according to the current environment;
融合模块,用于基于面部表情识别系统识别的用户的当前面部表情和当前情感更新所述三维虚拟面部表情图像,并将更新后的三维虚拟面部表情图像与所述三维地理图像进行融合,生成虚拟视频数据;The fusion module is used to update the three-dimensional virtual facial expression image based on the current facial expression and current emotion of the user recognized by the facial expression recognition system, and fuse the updated three-dimensional virtual facial expression image with the three-dimensional geographic image to generate a virtual Video data
投射模块,用于将所述虚拟视频数据投射至通话对端的增强现实显示器中。The projection module is used to project the virtual video data to the augmented reality display at the opposite end of the call.
所述终端设备8可以是桌上型计算机、笔记本、掌上电脑及云端管理服务器等计算设备。所述终端设备可包括,但不仅限于,处理器80、存储器81。本领域技术人员可以理解,图8仅仅是终端设备8的示例,并不构成对终端设备8的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述终端设备还可以包括输入输出设备、网络接入设备、总线等。The terminal device 8 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud management server. The terminal device may include, but is not limited to, a processor 80 and a memory 81. Those skilled in the art may understand that FIG. 8 is only an example of the terminal device 8 and does not constitute a limitation on the terminal device 8, and may include more or less components than the illustration, or a combination of certain components or different components. For example, the terminal device may further include an input and output device, a network access device, a bus, and the like.
所称处理器80可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 80 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application-specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
所述存储器81可以是所述终端设备8的内部存储单元,例如终端设备8的硬盘或内存。所述存储器81也可以是所述终端设备8的外部存储设备,例如所述终端设备8上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器81还可以既包括所述终端设备8的内部存储单元也包括外部存储设备。所述存储器81用于存储所述计算机程序以及所述终端设备所需的其他程序和数据。所述存储器81还可以用于暂时地存储已经输出或者将要输出的数据。The memory 81 may be an internal storage unit of the terminal device 8, such as a hard disk or a memory of the terminal device 8. The memory 81 may also be an external storage device of the terminal device 8, for example, a plug-in hard disk equipped on the terminal device 8, a smart memory card (Smart, Media, Card, SMC), and a secure digital (SD) Cards, flash cards, etc. Further, the memory 81 may also include both an internal storage unit of the terminal device 8 and an external storage device. The memory 81 is used to store the computer program and other programs and data required by the terminal device. The memory 81 can also be used to temporarily store data that has been or will be output.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述系统的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能 单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述无线终端中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for convenience and conciseness of description, only the above-mentioned division of each functional unit and module is used as an example for illustration. In practical applications, the above-mentioned functions can be allocated by different functional units, Module completion means that the internal structure of the system is divided into different functional units or modules to complete all or part of the functions described above. The functional units and modules in the embodiment may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above integrated unit may use hardware It can also be implemented in the form of software functional units. In addition, the specific names of each functional unit and module are only for the purpose of distinguishing each other, and are not used to limit the protection scope of the present application. For the specific working processes of the above units and modules in the wireless terminal, reference may be made to the corresponding processes in the foregoing method embodiments, which will not be repeated here.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above embodiments, the description of each embodiment has its own emphasis. For a part that is not detailed or recorded in an embodiment, you can refer to the related descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。Those of ordinary skill in the art may realize that the units and algorithm steps of the examples described in conjunction with the embodiments disclosed herein can be implemented by electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are executed in hardware or software depends on the specific application of the technical solution and design constraints. Professional technicians can use different methods to implement the described functions for each specific application, but such implementation should not be considered beyond the scope of the present invention.
在本发明所提供的实施例中,应该理解到,所揭露的系统/终端设备和方法,可以通过其它的方式实现。例如,以上所描述的系统/终端设备实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,系统或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided by the present invention, it should be understood that the disclosed system/terminal device and method may be implemented in other ways. For example, the system/terminal device embodiments described above are only schematic. For example, the division of the module or unit is only a logical function division. In actual implementation, there may be another division manner, such as multiple units Or components can be combined or integrated into another system, or some features can be ignored or not implemented. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, systems or units, and may be in electrical, mechanical or other forms.
所述设置为分离部件说明的单元可以是或者也可以不是物理上分开的,设置为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The unit described as a separate component may or may not be physically separated, and the component displayed as a unit may or may not be a physical unit, that is, it may be located in one place, or may be distributed to multiple network units on. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or software function unit.
所述集成的模块/单元如果以软件功能单元的形式实现并设置为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或系统、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取 存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括是电载波信号和电信信号。If the integrated module/unit is implemented in the form of a software functional unit and set as an independent product for sale or use, it may be stored in a computer-readable storage medium. Based on this understanding, the present invention can realize all or part of the processes in the methods of the above embodiments, and can also be completed by a computer program instructing relevant hardware. The computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, the steps of the foregoing method embodiments may be implemented. Wherein, the computer program includes computer program code, and the computer program code may be in a source code form, an object code form, an executable file, or some intermediate form, etc. The computer-readable medium may include any entity or system capable of carrying the computer program code, a recording medium, a USB flash drive, a mobile hard disk, a magnetic disk, an optical disk, a computer memory, and a read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signals, telecommunications signals and software distribution media, etc. It should be noted that the content contained in the computer-readable medium can be appropriately increased or decreased according to the requirements of legislation and patent practice in jurisdictions. For example, in some jurisdictions, according to legislation and patent practice, computer-readable media Excluded are electrical carrier signals and telecommunications signals.
以上所述实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围,均应包含在本发明的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present invention, not to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, persons of ordinary skill in the art should understand that they can still implement the foregoing The technical solutions described in the examples are modified, or some of the technical features are equivalently replaced; and these modifications or replacements do not deviate from the essence and scope of the technical solutions of the embodiments of the present invention, and should be included in Within the protection scope of the present invention.

Claims (10)

  1. 一种可穿戴增强现实远程视频系统,其特征在于,所述可穿戴增强现实远程视频系统包括增强现实智能眼镜系统、面部表情识别系统以及面部表情构建系统;A wearable augmented reality remote video system, characterized in that the wearable augmented reality remote video system includes an augmented reality smart glasses system, a facial expression recognition system, and a facial expression construction system;
    所述增强现实智能眼镜系统和所述面部表情识别系统设置在增强现实智能眼镜中,面部表情构建系统设置在可穿戴智能手环中;所述增强现实智能眼镜系统分别与所述面部表情识别系统和所述面部表情构建系统通信连接,所述面部表情识别系统和所述面部表情构建系统通信连接;The augmented reality smart glasses system and the facial expression recognition system are set in the augmented reality smart glasses, and the facial expression construction system is set in the wearable smart bracelet; the augmented reality smart glasses system and the facial expression recognition system are respectively A communication connection with the facial expression construction system, and a communication connection with the facial expression recognition system and the facial expression construction system;
    所述增强现实智能眼镜系统用于捕捉用户当前环境信息并根据所述当前环境构建用户当前环境信息的三维地理图像,并将所述三维地理图像投射至通话对端的增强现实显示器中;The augmented reality smart glasses system is used to capture the current environment information of the user and construct a three-dimensional geographic image of the current environment information of the user according to the current environment, and project the three-dimensional geographic image to the augmented reality display of the opposite end of the call;
    所述面部表情识别系统用于识别用户的当前面部表情和当前情感;The facial expression recognition system is used to recognize the user's current facial expression and current emotion;
    所述面部表情构建系统用于捕捉和构建用户的三维虚拟面部表情。The facial expression construction system is used to capture and construct a three-dimensional virtual facial expression of a user.
  2. 根据权利要求1所述的可穿戴增强现实远程视频系统,其特征在于,所述增强现实智能眼镜系统包括增强现实智能眼镜和主机;The wearable augmented reality remote video system according to claim 1, wherein the augmented reality smart glasses system includes augmented reality smart glasses and a host;
    所述增强现实智能眼镜包括光纤扫描投射器、双目红外手势识别摄像头、眼动仪、双目前视广角摄像头、加速度传感器、骨传导耳机及降噪麦克风;The augmented reality smart glasses include an optical fiber scanning projector, a binocular infrared gesture recognition camera, an eye tracker, a dual current viewing wide-angle camera, an acceleration sensor, a bone conduction headset, and a noise reduction microphone;
    所述主机包括处理器,所述控制器分别与所述光纤扫描投射器、所述双目红外手势识别摄像头、所述眼动仪、所述双目前视广角摄像头、所述加速度传感器、所述骨传导耳机及所述降噪麦克风电连接;The host includes a processor, the controller and the optical fiber scanning projector, the binocular infrared gesture recognition camera, the eye tracker, the dual current viewing wide-angle camera, the acceleration sensor, the The bone conduction earphone and the noise reduction microphone are electrically connected;
    所述主机用于对投影光源进行收集,并通过光纤将三维地理图像投影在通话对端的增强现实显示器中。The host is used to collect the projection light source and project the three-dimensional geographic image on the augmented reality display at the opposite end of the call through the optical fiber.
  3. 根据权利要求1所述的可穿戴增强现实远程视频系统,其特征在于,所述面部表情识别系统包括肌电传感器、皮肤电反应传感器、心电信号传感器、呼吸信号传感器及数据处理器;The wearable augmented reality remote video system according to claim 1, wherein the facial expression recognition system includes an electromyography sensor, a skin electrical response sensor, an electrocardiographic signal sensor, a respiratory signal sensor, and a data processor;
    所述数据处理器分别与所述肌电传感器、所述皮肤电反应传感器、所述心电信号传感器及所述呼吸信号传感器电连接。The data processor is electrically connected to the myoelectric sensor, the skin electrical response sensor, the electrocardiographic signal sensor and the respiration signal sensor, respectively.
  4. 根据权利要求1所述的可穿戴增强现实远程视频系统,其特征在于,所述面部表情构建系统包括3D结构光采集摄像头和肌电信号传感器。The wearable augmented reality remote video system according to claim 1, wherein the facial expression construction system includes a 3D structured light collection camera and an EMG signal sensor.
  5. 根据权利要求4所述的可穿戴增强现实远程视频系统,其特征在于,所述3D结构光采集摄像头和所述肌电信号传感器集成在可穿戴智能手环中。The wearable augmented reality remote video system according to claim 4, wherein the 3D structured light collection camera and the EMG signal sensor are integrated in a wearable smart bracelet.
  6. 一种基于权利要求1所述的可穿戴增强现实远程视频系统的视频通话方法,其特征 在于,包括:A video call method based on the wearable augmented reality remote video system according to claim 1, characterized in that it includes:
    通过面部表情构建系统捕捉和构建用户的三维虚拟面部表情图像;Capture and construct the user's three-dimensional virtual facial expression image through the facial expression construction system;
    通过面部表情识别系统识别用户的当前面部表情和当前情感;Recognize the user's current facial expression and current emotion through the facial expression recognition system;
    通过增强现实智能眼镜系统捕捉用户当前环境信息并根据所述当前环境构建用户当前环境信息的三维地理图像;Capture the current environment information of the user through the augmented reality smart glasses system and construct a three-dimensional geographic image of the current environment information of the user according to the current environment;
    基于面部表情识别系统识别的用户的当前面部表情和当前情感更新所述三维虚拟面部表情图像,并将更新后的三维虚拟面部表情图像与所述三维地理图像进行融合,生成虚拟视频数据;Updating the three-dimensional virtual facial expression image based on the user's current facial expression and current emotion recognized by the facial expression recognition system, and fusing the updated three-dimensional virtual facial expression image with the three-dimensional geographic image to generate virtual video data;
    将所述虚拟视频数据投射至通话对端的增强现实显示器中。Project the virtual video data into the augmented reality display at the opposite end of the call.
  7. 根据权利要求6所述的视频通话方法,其特征在于,还包括:The video call method according to claim 6, further comprising:
    通过佩戴在用户腕部的可穿戴智能手环捕捉用户的手势动作,根据所述手势动作切换通话对象。The wearable smart bracelet worn on the user's wrist captures the user's gestures and switches the call object according to the gestures.
  8. 根据权利要求6所述的视频通话方法,其特征在于,所述通过面部表情构建系统捕捉和构建用户的三维虚拟面部表情图像包括:The video call method according to claim 6, wherein the capturing and constructing the three-dimensional virtual facial expression image of the user through the facial expression construction system includes:
    通过面部表情构建系统的3D结构光采集摄像头对用户的面部表情进行扫描;Scan the user's facial expressions through the 3D structured light collection camera of the facial expression construction system;
    并根据扫描到的用户的面部表情在虚拟空间中构建所述用户的三维虚拟面部图像。According to the scanned facial expressions of the user, a three-dimensional virtual facial image of the user is constructed in the virtual space.
  9. 根据权利要求6所述的视频通话方法,其特征在于,所述通过面部表情识别系统识别用户的当前面部表情和当前情感包括:The video call method according to claim 6, wherein the recognizing the user's current facial expression and current emotion through the facial expression recognition system includes:
    采集所述用户的生理信号,所述生理信号包括皮肤电反应信号、肌电信号、呼吸信号和心电信号;Collecting physiological signals of the user, the physiological signals including skin electrical response signals, myoelectric signals, respiratory signals and electrocardiographic signals;
    根据所述用户的生理信号进行多维度联合识别,判断当前用户的当前面部表情和当前情感。Multi-dimensional joint recognition is performed based on the user's physiological signals to determine the current facial expression and current emotion of the current user.
  10. 根据权利要求9所述的视频通话方法,其特征在于,根据所述用户的生理信号进行多维度联合识别,判断当前用户的当前面部表情和当前情感,包括:The video call method according to claim 9, wherein the multi-dimensional joint recognition is performed based on the user's physiological signal to determine the current facial expression and current emotion of the current user, including:
    对所述生理信号进行初始化处理,滤除干扰信号并进行数据采样,获取离散生理信号;Initialize the physiological signals, filter out interference signals and perform data sampling to obtain discrete physiological signals;
    提取各类离散生理信号的特征值;Extract the characteristic values of various discrete physiological signals;
    基于离散二进制粒子群算法根据各类离散生理信号的特征值确定各类离散生理信号的有效特征信号;Based on discrete binary particle swarm optimization algorithm to determine the effective characteristic signals of various discrete physiological signals based on the characteristic values of various discrete physiological signals
    根据所述各类离散生理信号的有效特征信号进行情感识别,获取对应的情感类别。Emotion recognition is performed according to the effective characteristic signals of the various types of discrete physiological signals to obtain corresponding emotion categories.
PCT/CN2019/124331 2018-12-17 2019-12-10 Wearable augmented-reality remote video system and video call method WO2020125493A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811544975.5A CN109803109B (en) 2018-12-17 2018-12-17 Wearable augmented reality remote video system and video call method
CN201811544975.5 2018-12-17

Publications (1)

Publication Number Publication Date
WO2020125493A1 true WO2020125493A1 (en) 2020-06-25

Family

ID=66556993

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/124331 WO2020125493A1 (en) 2018-12-17 2019-12-10 Wearable augmented-reality remote video system and video call method

Country Status (2)

Country Link
CN (1) CN109803109B (en)
WO (1) WO2020125493A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023222164A1 (en) * 2022-05-17 2023-11-23 Vimelio GmbH Method, module system and system for reducing noise levels caused by medical instruments
US11995776B2 (en) 2021-01-19 2024-05-28 Samsung Electronics Co., Ltd. Extended reality interaction in synchronous virtual spaces using heterogeneous devices

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109348075B (en) * 2018-09-17 2021-02-02 北京盘古影艺文化传播有限公司 Incoming call processing system
CN109803109B (en) * 2018-12-17 2020-07-31 中国科学院深圳先进技术研究院 Wearable augmented reality remote video system and video call method
CN110413106B (en) * 2019-06-18 2024-02-09 中国人民解放军军事科学院国防科技创新研究院 Augmented reality input method and system based on voice and gestures
CN113763531B (en) * 2020-06-05 2023-11-28 北京达佳互联信息技术有限公司 Three-dimensional face reconstruction method and device, electronic equipment and storage medium
CN114339117A (en) * 2021-12-31 2022-04-12 北京字跳网络技术有限公司 Video call method and device, electronic equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105933637A (en) * 2016-04-26 2016-09-07 上海与德通讯技术有限公司 Video communication method and system
CN107203263A (en) * 2017-04-11 2017-09-26 北京峰云视觉技术有限公司 A kind of virtual reality glasses system and image processing method
CN108337466A (en) * 2018-02-01 2018-07-27 联想(北京)有限公司 A kind of video call method and electronic equipment
WO2018141408A1 (en) * 2017-02-03 2018-08-09 Vestel Elektronik Sanayi Ve Ticaret A.S. IMPROVED METHOD AND SYSTEM FOR VIDEO CONFERENCES WITH HMDs
CN109803109A (en) * 2018-12-17 2019-05-24 中国科学院深圳先进技术研究院 A kind of wearable augmented reality remote video system and video call method

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103415003A (en) * 2013-08-26 2013-11-27 苏州跨界软件科技有限公司 Virtual figure communication system
CN103647922A (en) * 2013-12-20 2014-03-19 百度在线网络技术(北京)有限公司 Virtual video call method and terminals
US9277180B2 (en) * 2014-06-30 2016-03-01 International Business Machines Corporation Dynamic facial feature substitution for video conferencing
CN106037750A (en) * 2016-06-20 2016-10-26 国网上海市电力公司 Power line inspection system based on wearable intelligent equipment
WO2018041780A1 (en) * 2016-08-29 2018-03-08 Koninklijke Kpn N.V. Communicating in a virtual reality environment
CN108063909B (en) * 2016-11-08 2021-02-09 阿里巴巴集团控股有限公司 Video conference system, image tracking and collecting method and device
CN108646920A (en) * 2018-05-16 2018-10-12 Oppo广东移动通信有限公司 Identify exchange method, device, storage medium and terminal device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105933637A (en) * 2016-04-26 2016-09-07 上海与德通讯技术有限公司 Video communication method and system
WO2018141408A1 (en) * 2017-02-03 2018-08-09 Vestel Elektronik Sanayi Ve Ticaret A.S. IMPROVED METHOD AND SYSTEM FOR VIDEO CONFERENCES WITH HMDs
CN107203263A (en) * 2017-04-11 2017-09-26 北京峰云视觉技术有限公司 A kind of virtual reality glasses system and image processing method
CN108337466A (en) * 2018-02-01 2018-07-27 联想(北京)有限公司 A kind of video call method and electronic equipment
CN109803109A (en) * 2018-12-17 2019-05-24 中国科学院深圳先进技术研究院 A kind of wearable augmented reality remote video system and video call method

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11995776B2 (en) 2021-01-19 2024-05-28 Samsung Electronics Co., Ltd. Extended reality interaction in synchronous virtual spaces using heterogeneous devices
WO2023222164A1 (en) * 2022-05-17 2023-11-23 Vimelio GmbH Method, module system and system for reducing noise levels caused by medical instruments

Also Published As

Publication number Publication date
CN109803109B (en) 2020-07-31
CN109803109A (en) 2019-05-24

Similar Documents

Publication Publication Date Title
WO2020125493A1 (en) Wearable augmented-reality remote video system and video call method
US20220245961A1 (en) Training method for expression transfer model, expression transfer method and apparatus
CN109086726B (en) Local image identification method and system based on AR intelligent glasses
TWI751161B (en) Terminal equipment, smart phone, authentication method and system based on face recognition
JP2019531014A (en) Face model capture by wearable device
KR20190133204A (en) Accumulation and Reliability Allocation of Iris Codes
CN110956691A (en) Three-dimensional face reconstruction method, device, equipment and storage medium
CN112198959A (en) Virtual reality interaction method, device and system
Wang et al. Intelligent wearable virtual reality (VR) gaming controller for people with motor disabilities
TW202014992A (en) System and method for simulating expression of virtual facial model
CN105653020A (en) Time traveling method and apparatus and glasses or helmet using same
CN109144245B (en) Equipment control method and related product
CN106598211A (en) Gesture interaction system and recognition method for multi-camera based wearable helmet
Chalasani et al. Egocentric gesture recognition for head-mounted ar devices
CN111311733A (en) Three-dimensional model processing method and device, processor, electronic device and storage medium
JP2017224166A (en) Image generation device, image generation program, and image generation method
CN114615455A (en) Teleconference processing method, teleconference processing device, teleconference system, and storage medium
CN110007765A (en) A kind of man-machine interaction method, device and equipment
US10664689B2 (en) Determining user activity based on eye motion
CN104635930A (en) Information processing method and electronic device
Enikeev et al. Recognition of sign language using leap motion controller data
CN110321009B (en) AR expression processing method, device, equipment and storage medium
CN111597926A (en) Image processing method and device, electronic device and storage medium
CN114967937B (en) Virtual human motion generation method and system
CN113407031B (en) VR (virtual reality) interaction method, VR interaction system, mobile terminal and computer readable storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19898776

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19898776

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 10.11.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19898776

Country of ref document: EP

Kind code of ref document: A1