CN110013260B - Emotion theme regulation and control method, equipment and computer-readable storage medium - Google Patents

Emotion theme regulation and control method, equipment and computer-readable storage medium Download PDF

Info

Publication number
CN110013260B
CN110013260B CN201910360042.9A CN201910360042A CN110013260B CN 110013260 B CN110013260 B CN 110013260B CN 201910360042 A CN201910360042 A CN 201910360042A CN 110013260 B CN110013260 B CN 110013260B
Authority
CN
China
Prior art keywords
wearing
emotion
state
theme
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910360042.9A
Other languages
Chinese (zh)
Other versions
CN110013260A (en
Inventor
张圣杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Good Feeling Health Industry Group Co ltd
Original Assignee
Good Feeling Health Industry Group Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Good Feeling Health Industry Group Co ltd filed Critical Good Feeling Health Industry Group Co ltd
Priority to CN201910360042.9A priority Critical patent/CN110013260B/en
Publication of CN110013260A publication Critical patent/CN110013260A/en
Application granted granted Critical
Publication of CN110013260B publication Critical patent/CN110013260B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/165Evaluating the state of mind, e.g. depression, anxiety
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/742Details of notification to user or communication with user or patient ; user input means using visual displays

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Animal Behavior & Ethology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Veterinary Medicine (AREA)
  • Public Health (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Physics & Mathematics (AREA)
  • Surgery (AREA)
  • Biophysics (AREA)
  • Pathology (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Medical Informatics (AREA)
  • Developmental Disabilities (AREA)
  • Child & Adolescent Psychology (AREA)
  • Social Psychology (AREA)
  • Psychology (AREA)
  • Hospice & Palliative Care (AREA)
  • Educational Technology (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a method, equipment and a computer-readable storage medium for regulating and controlling emotion topics, wherein the method comprises the following steps: acquiring the wearing state and wearing environment of the wearable equipment; then, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment; then, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information; finally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature. The humanized emotion theme regulation and control scheme is realized, so that theme regulation and control of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.

Description

Emotion theme regulation and control method, equipment and computer-readable storage medium
Technical Field
The present application relates to the field of mobile communications, and in particular, to a method and apparatus for emotion theme adjustment and control, and a computer-readable storage medium.
Background
In the prior art, with the rapid development of intelligent terminal equipment, wearable equipment different from conventional smart phones, such as wearable equipment like a smart watch or a smart bracelet, appears. Because wearing formula equipment compares in traditional smart mobile phone, its particularities such as software, hardware environment, operation mode and operation environment, if the scheme of controlling of traditional smart mobile phone is transferred to wearing formula equipment, can bring inconvenience, user experience bad for user's operation.
Disclosure of Invention
In order to solve the technical defects in the prior art, the invention provides a method for regulating and controlling emotion topics, which comprises the following steps:
acquiring the wearing state and wearing environment of the wearable equipment;
acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment;
learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information;
and regulating and controlling the display theme category and the display theme characteristic of the wearable equipment according to the emotion characteristic in the wearing state and the wearing environment.
Optionally, the acquiring the wearing state and the wearing environment of the wearable device includes:
acquiring a wearing state of the wearable device, wherein the wearing state comprises a wearing position and a motion state;
according to the touch control signal and/or the buckling signal of the wearable equipment, adjusting the wearing position;
and adjusting the motion state according to the motion sensing signal of the wearable equipment.
Optionally, the acquiring the wearing state and the wearing environment of the wearable device further includes:
acquiring a wearing environment of the wearable device, wherein the wearing environment comprises an internal environment and an external environment;
determining the internal environment through the system running state and the application running state of the wearable device;
and determining the external environment through the light sensing assembly, the temperature sensing assembly and the humidity sensing assembly of the wearable device.
Optionally, the acquiring heart rate information of the wearing object according to the wearing state, and simultaneously acquiring state information of the wearing object according to the wearing environment, includes:
positioning a detection area corresponding to heart rate detection in the wearable device according to the wearing position;
And acquiring the heart rate information through a heart rate detection component in the detection area.
Optionally, the acquiring the heart rate information of the wearing object according to the wearing state, and simultaneously acquiring the state information of the wearing object according to the wearing environment, further includes:
determining a first emotional characteristic affecting the emotion of the wearing object according to the internal environment;
and determining a second emotion characteristic affecting emotion of the wearing object according to the external environment.
Optionally, the learning, in combination with the heart rate information and the state information, obtains the emotional characteristic of the wearing object, includes:
acquiring basic attribute information of the wearing object, wherein the basic attribute information comprises age information, gender information, work information and physiological information;
and learning to obtain the emotion characteristics of the wearing object in the current state by combining the heart rate information, the state information and the basic attribute information.
Optionally, in the wearing state and the wearing environment, the adjusting the display theme category and the display theme feature of the wearable device according to the emotion feature includes:
analyzing the emotion characteristics;
And determining the display interface attribute and the display content attribute corresponding to the emotion characteristics.
Optionally, in the wearing state and the wearing environment, the adjusting the display theme category and the display theme feature of the wearable device according to the emotion feature further includes:
determining the display theme category according to the display interface attribute;
and determining the display theme characteristics according to the display content attributes, wherein the display theme characteristics comprise theme contents.
The invention also provides emotion theme regulation and control equipment, which comprises:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
The present invention also proposes a computer-readable storage medium having stored thereon an emotion theme regulation program which, when executed by a processor, implements the steps of the emotion theme regulation method as set forth in any one of the above.
The wearable device has the beneficial effects that the wearing state and the wearing environment of the wearable device are obtained; then, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment; then, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information; finally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature. The humanized emotion theme regulation and control scheme is realized, so that theme regulation and control of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention.
In order to more clearly illustrate the embodiments of the invention or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, and it will be obvious to a person skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of an implementation manner of a wearable device according to an embodiment of the present invention;
fig. 2 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 3 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 4 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 5 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
FIG. 6 is a flowchart of a first embodiment of the emotion theme regulation method of the present invention;
FIG. 7 is a flowchart of a second embodiment of the emotion theme regulation method of the present invention;
FIG. 8 is a flowchart of a third embodiment of the emotion theme regulation method of the present invention;
FIG. 9 is a flowchart of a fourth embodiment of the emotion theme regulation method of the present invention;
FIG. 10 is a flowchart of a fifth embodiment of the emotion theme regulation method of the present invention;
FIG. 11 is a flowchart of a sixth embodiment of the emotion theme regulation method of the present invention;
FIG. 12 is a flowchart of a seventh embodiment of the emotion theme regulation method of the present invention;
fig. 13 is a flowchart of an eighth embodiment of the emotion theme regulation method of the present invention.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present invention. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. In this embodiment, the device may be a wristwatch, a bracelet, or a wearable device, and the device includes a screen and a connection portion. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to this embodiment.
Example 1
Fig. 6 is a flowchart of a first embodiment of the emotion theme regulation method of the present invention. A method of emotional topic regulation, the method comprising:
s1, acquiring a wearing state and a wearing environment of a wearable device;
s2, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment;
s3, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information;
s4, adjusting and controlling the display theme category and the display theme characteristic of the wearable device according to the emotion characteristic in the wearing state and the wearing environment.
In this embodiment, first, a wearing state and a wearing environment of a wearable device are acquired; then, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment; then, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information; finally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature.
Optionally, determining a device type of the wearable device, for example, the device type includes a wrist device type and a finger device type, and determining a corresponding wearing state and wearing environment according to different device types, respectively;
optionally, determining other terminal devices associated with the wearable device, and acquiring the wearing state and/or the wearing environment of the wearable device through the other terminal devices;
optionally, acquiring a heart rate identification component of the wearing object according to the wearing state, and then acquiring heart rate information through the heart rate identification component, or determining other terminal equipment associated with the wearing device, and acquiring heart rate information through the other terminal equipment;
optionally, acquiring state information of the wearing object according to the wearing environment, or determining other terminal devices associated with the wearing device, and acquiring heart rate information through the other terminal devices;
optionally, learning to obtain the emotional characteristics of the wearing object in a preset period by combining the heart rate information and the state information;
optionally, when the fluctuation amplitude of the heart rate information is greater than a preset value, acquiring the emotion characteristics of the wearing object by combining the heart rate information and the state information;
Optionally, when the state information meets a preset trigger condition, acquiring an emotion feature of the wearing object, wherein the preset condition comprises that the wearing position is switched and/or the touch pressure sensing value is greater than a preset threshold;
optionally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature;
optionally, in the wearing state and the wearing environment, the display theme category of the wearable device is regulated and controlled according to the emotion feature, then the heart rate information and the state information are detected again, and if a preset regulation and control condition is met, the subsequent regulation and control of the display theme feature is executed.
The method has the beneficial effects that the wearing state and the wearing environment of the wearable equipment are obtained; then, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment; then, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information; finally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature. The humanized emotion theme regulation and control scheme is realized, so that theme regulation and control of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example two
Fig. 7 is a flowchart of a second embodiment of the emotion theme adjustment and control method of the present invention, based on the foregoing embodiment, optionally, the acquiring the wearing state and the wearing environment of the wearable device includes:
s11, acquiring a wearing state of the wearable device, wherein the wearing state comprises a wearing position and a motion state;
s12, adjusting the wearing position according to the touch signal and/or the buckling signal of the wearable device;
s13, adjusting the motion state according to the motion sensing signal of the wearable device.
In this embodiment, first, a wearing state of the wearable device is obtained, where the wearing state includes a wearing position and a motion state; then, according to the touch signal and/or the buckling signal of the wearable device, adjusting the wearing position; finally, the motion state is adjusted according to the motion sensing signal of the wearable device.
Optionally, the wearing position includes a body position of the wearing object, and a line of sight orientation corresponding to the body position and the wearing object;
optionally, the movement state includes a movement state of a body position of the wearing object, and further includes a movement type of the wearing object itself;
Optionally, the current wearing position is verified, adjusted or determined according to the buckling signal of the edge of the wearable device or the touch signal in the screen area of the wearable device.
The method and the device have the beneficial effects that the wearing state of the wearable device is obtained, wherein the wearing state comprises the wearing position and the motion state; then, according to the touch signal and/or the buckling signal of the wearable device, adjusting the wearing position; finally, the motion state is adjusted according to the motion sensing signal of the wearable device. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example III
Fig. 8 is a flowchart of a third embodiment of the emotion theme adjustment and control method according to the present invention, based on the foregoing embodiment, optionally, the acquiring the wearing state and the wearing environment of the wearable device further includes:
s14, acquiring a wearing environment of the wearable device, wherein the wearing environment comprises an internal environment and an external environment;
S15, determining the internal environment through the system running state and the application running state of the wearable device;
s16, determining the external environment through the light sensing assembly, the temperature sensing assembly and the humidity sensing assembly of the wearable device.
In this embodiment, first, a wearing environment of the wearable device is obtained, where the wearing environment includes an internal environment and an external environment; then, determining the internal environment through the system running state and the application running state of the wearable device; and finally, determining the external environment through the light sensing assembly, the temperature sensing assembly and the humidity sensing assembly of the wearable equipment.
Optionally, the internal environment includes a current foreground application program and a recently started application program, and further includes a current system setting item, such as a volume level, a light intensity of the display screen, and a color setting type of the display screen;
optionally, the external environment includes a current light intensity, a current ambient temperature, and a current ambient humidity.
The method and the device have the beneficial effects that the wearing environment of the wearable device is obtained, wherein the wearing environment comprises an internal environment and an external environment; then, determining the internal environment through the system running state and the application running state of the wearable device; and finally, determining the external environment through the light sensing assembly, the temperature sensing assembly and the humidity sensing assembly of the wearable equipment. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example IV
Fig. 9 is a flowchart of a fourth embodiment of the emotion theme adjustment and control method according to the present invention, based on the foregoing embodiment, optionally, the acquiring heart rate information of the wearing object according to the wearing state, and simultaneously acquiring state information of the wearing object according to the wearing environment, includes:
s21, positioning a detection area corresponding to heart rate detection in the wearable device according to the wearing position;
s22, acquiring the heart rate information through a heart rate detection component in the detection area.
In this embodiment, first, according to the wearing position, a detection area corresponding to heart rate detection is positioned in the wearable device; the heart rate information is then acquired by a heart rate detection component within the detection zone.
Optionally, in this scheme, in order to obtain more accurate heart rate information, the wearing object sits still on the chair, and the health is in natural relaxation state, presents corresponding video material according to emotion induction experiment paradigm, carries out emotion induction. When a wearing object watches a video clip of the wearing device, the video clip is evaluated to evaluate the strength of the video clip for exciting the emotion. Pulse and electrocardiosignals are synchronously acquired by utilizing a corresponding sensor and an electrode of a physiological signal acquisition instrument, and the signals are sent to a database for storage. The wearing object respectively executes 3 emotion induction and physiological signal acquisition experiments at different times, and each emotion acquisition obtains 300 electrocardio signals and 300 pulse signals. Each emotional physiological signal acquired 3min data, and the data processing only intercepted pulse and electrocardiographic data within 30s from the 2 nd min as sample data.
The method has the beneficial effects that the detection area corresponding to heart rate detection is positioned in the wearable device according to the wearing position; the heart rate information is then acquired by a heart rate detection component within the detection zone. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example five
Fig. 10 is a flowchart of a fifth embodiment of the emotion theme adjustment and control method according to the present invention, based on the foregoing embodiment, optionally, the acquiring heart rate information of the wearing object according to the wearing state, and simultaneously acquiring state information of the wearing object according to the wearing environment, further includes:
s23, determining a first emotion feature affecting emotion of the wearing object according to the internal environment;
s24, determining a second emotion feature affecting the emotion of the wearing object according to the external environment.
In this embodiment, first, a first emotional characteristic affecting the emotion of the wearing object is determined according to the internal environment; then, a second emotional characteristic affecting the emotion of the wearing object is determined according to the external environment.
And extracting a time domain basic feature PPG wave peak value and a PPG wave peak interval by adopting a peak detection analysis method, and obtaining a PPG power spectrum by adopting a periodic chart power spectrum analysis method. And respectively calculating the time domain and frequency domain statistical feature quantities, namely the mean value, the maximum value, the minimum value, the median value, the variance and the standard deviation, and generating a multi-dimensional pulse feature vector.
The extraction of the mean value of the PPG wave crest is to calculate the mean value of all wave crests detected in a period of time (such as 30 s) by detecting the waveform data of each sample, and then 300 samples of each emotion are extracted to obtain 300 mean values, and the extraction of other statistical characteristic quantities is similar to a mean value extraction method. Detecting a main pulse P wave crest by a differential threshold method, detecting a maximum value and a minimum value of the preprocessed pulse signals, wherein the difference value is the maximum waveform height, and defining a threshold rule as follows: (1) the peak magnitude fluctuation range is not more than 0.3 times of the maximum waveform height; (2) the peak is larger than all amplitude values of about 200 sample points. When both threshold conditions are satisfied, the peak point may be determined to be the P-wave to be detected. And detecting the peak point of each 600 sample points of the pulse signal at intervals, and recording the interval between adjacent peaks while detecting the position of the peak point, thereby obtaining the cycle sequence of the pulse wave.
A deep learning neural network model is built, a multi-layer network is adopted for learning, more data are continuously used for training and testing until the passing rate of the tested data exceeds a certain interval, and the model is considered to be built to be completed if 85%.
The method has the advantages that the first emotion characteristics affecting the emotion of the wearing object are determined according to the internal environment; then, a second emotional characteristic affecting the emotion of the wearing object is determined according to the external environment. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example six
Fig. 11 is a flowchart of a sixth embodiment of the emotion theme adjustment and control method according to the present invention, based on the foregoing embodiment, optionally, the learning to obtain the emotion feature of the wearing object includes:
s31, acquiring basic attribute information of the wearing object, wherein the basic attribute information comprises age information, gender information, work information and physiological information;
S32, learning to obtain the emotion characteristics of the wearing object in the current state by combining the heart rate information, the state information and the basic attribute information.
In this embodiment, first, basic attribute information of the wearing object is obtained, where the basic attribute information includes age information, gender information, work information, and physiological information; and then, learning to obtain the emotion characteristics of the wearing object in the current state by combining the heart rate information, the state information and the basic attribute information.
Optionally, in this embodiment, a data tag is added, for example, a first emotion feature and a second emotion feature in this embodiment, and corresponding heart rate information, state information and basic attribute information are added, a detection waveform of each tested person is labeled with a time corresponding to each emotion, unified classification is performed, statistics is performed in a multidimensional feature vector, and a corresponding mode corresponding to the same emotion is found.
The method has the advantages that basic attribute information of the wearing object is obtained, wherein the basic attribute information comprises age information, gender information, work information and physiological information; and then, learning to obtain the emotion characteristics of the wearing object in the current state by combining the heart rate information, the state information and the basic attribute information. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example seven
Fig. 12 is a flowchart of a seventh embodiment of the emotion theme adjustment and control method according to the present invention, based on the foregoing embodiment, optionally, the adjusting, in the wearing state and in the wearing environment, the display theme category and the display theme feature of the wearable device according to the emotion feature includes:
s41, analyzing the emotion characteristics;
s42, determining display interface attributes and display content attributes corresponding to the emotion characteristics.
In this embodiment, first, the emotional characteristics are parsed; and then, determining the display interface attribute and the display content attribute corresponding to the emotion characteristics.
Optionally, in the daily use process of the wearable device, the heart rate waveform change of the user is continuously detected, the background neural network is matched and judged, different display subjects are switched according to the emotion judgment result, and targeted relief is detected, so that the visual angle experience of the user is improved, and the effects of relieving anxiety and calming mood of the user are achieved.
The embodiment has the beneficial effects that the emotion characteristics are analyzed; and then, determining the display interface attribute and the display content attribute corresponding to the emotion characteristics. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example eight
Fig. 13 is a flowchart of an eighth embodiment of the emotion theme adjustment and control method according to the present invention, based on the foregoing embodiment, optionally, in the wearing state and in the wearing environment, the adjusting and controlling, according to the emotion characteristics, a display theme category and a display theme characteristic of the wearable device, further includes:
s43, determining the display theme category according to the display interface attribute;
s44, determining the display theme characteristics according to the display content attribute, wherein the display theme characteristics comprise theme contents.
In this embodiment, first, the display theme category is determined according to the display interface attribute; and then determining the display theme characteristics according to the display content attributes, wherein the display theme characteristics comprise theme contents.
Optionally, in the wearing state and the wearing environment, firstly adjusting and controlling the display theme category of the wearable device according to the emotion characteristics, then detecting the heart rate information and the state information again, and if the heart rate information and the state information meet preset adjusting and controlling conditions, executing subsequent adjustment and control of the display theme characteristics;
optionally, after performing subsequent adjustment and control of the display theme feature, detecting the heart rate information and the state information again, and if the heart rate information and the state information meet preset adjustment and control conditions, repeating the step of adjusting and controlling the display theme category of the wearable device according to the emotion feature.
The method has the advantages that the display theme category is determined according to the display interface attribute; and then determining the display theme characteristics according to the display content attributes, wherein the display theme characteristics comprise theme contents. According to the method, the device and the system, a more humanized emotion theme regulation scheme is realized, so that theme regulation of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
Example nine
The invention also proposes an audio processing device comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any of the preceding claims when executed by the processor.
Specifically, in this embodiment, first, a wearing state and a wearing environment of a wearable device are acquired; then, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment; then, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information; finally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature.
Optionally, determining a device type of the wearable device, for example, the device type includes a wrist device type and a finger device type, and determining a corresponding wearing state and wearing environment according to different device types, respectively;
optionally, determining other terminal devices associated with the wearable device, and acquiring the wearing state and/or the wearing environment of the wearable device through the other terminal devices;
optionally, acquiring a heart rate identification component of the wearing object according to the wearing state, and then acquiring heart rate information through the heart rate identification component, or determining other terminal equipment associated with the wearing device, and acquiring heart rate information through the other terminal equipment;
optionally, acquiring state information of the wearing object according to the wearing environment, or determining other terminal devices associated with the wearing device, and acquiring heart rate information through the other terminal devices;
optionally, learning to obtain the emotional characteristics of the wearing object in a preset period by combining the heart rate information and the state information;
optionally, when the fluctuation amplitude of the heart rate information is greater than a preset value, acquiring the emotion characteristics of the wearing object by combining the heart rate information and the state information;
Optionally, when the state information meets a preset trigger condition, acquiring an emotion feature of the wearing object, wherein the preset condition comprises that the wearing position is switched and/or the touch pressure sensing value is greater than a preset threshold;
optionally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature;
optionally, in the wearing state and the wearing environment, the display theme category of the wearable device is regulated and controlled according to the emotion feature, then the heart rate information and the state information are detected again, and if a preset regulation and control condition is met, the subsequent regulation and control of the display theme feature is executed.
In this embodiment, first, acquiring a wearing state of the wearable device, where the wearing state includes a wearing position and a motion state; then, according to the touch signal and/or the buckling signal of the wearable device, adjusting the wearing position; finally, the motion state is adjusted according to the motion sensing signal of the wearable device.
Optionally, the wearing position includes a body position of the wearing object, and a line of sight orientation corresponding to the body position and the wearing object;
optionally, the movement state includes a movement state of a body position of the wearing object, and further includes a movement type of the wearing object itself;
optionally, the current wearing position is verified, adjusted or determined according to the buckling signal of the edge of the wearable device or the touch signal in the screen area of the wearable device.
In this embodiment, first, acquiring a wearing environment of the wearable device, where the wearing environment includes an internal environment and an external environment; then, determining the internal environment through the system running state and the application running state of the wearable device; and finally, determining the external environment through the light sensing assembly, the temperature sensing assembly and the humidity sensing assembly of the wearable equipment.
Optionally, the internal environment includes a current foreground application program and a recently started application program, and further includes a current system setting item, such as a volume level, a light intensity of the display screen, and a color setting type of the display screen;
Optionally, the external environment includes a current light intensity, a current ambient temperature, and a current ambient humidity.
In this embodiment, first, according to the wearing position, a detection area corresponding to heart rate detection is located in the wearable device; the heart rate information is then acquired by a heart rate detection component within the detection zone.
Optionally, in this scheme, in order to obtain more accurate heart rate information, the wearing object sits still on the chair, and the health is in natural relaxation state, presents corresponding video material according to emotion induction experiment paradigm, carries out emotion induction. When a wearing object watches a video clip of the wearing device, the video clip is evaluated to evaluate the strength of the video clip for exciting the emotion. Pulse and electrocardiosignals are synchronously acquired by utilizing a corresponding sensor and an electrode of a physiological signal acquisition instrument, and the signals are sent to a database for storage. The wearing object respectively executes 3 emotion induction and physiological signal acquisition experiments at different times, and each emotion acquisition obtains 300 electrocardio signals and 300 pulse signals. Each emotional physiological signal acquired 3min data, and the data processing only intercepted pulse and electrocardiographic data within 30s from the 2 nd min as sample data.
In this embodiment, first, determining, according to the internal environment, a first emotional characteristic affecting an emotion of the wearing object; then, a second emotional characteristic affecting the emotion of the wearing object is determined according to the external environment.
And extracting a time domain basic feature PPG wave peak value and a PPG wave peak interval by adopting a peak detection analysis method, and obtaining a PPG power spectrum by adopting a periodic chart power spectrum analysis method. And respectively calculating the time domain and frequency domain statistical feature quantities, namely the mean value, the maximum value, the minimum value, the median value, the variance and the standard deviation, and generating a multi-dimensional pulse feature vector.
The extraction of the mean value of the PPG wave crest is to calculate the mean value of all wave crests detected in a period of time (such as 30 s) by detecting the waveform data of each sample, and then 300 samples of each emotion are extracted to obtain 300 mean values, and the extraction of other statistical characteristic quantities is similar to a mean value extraction method. Detecting a main pulse P wave crest by a differential threshold method, detecting a maximum value and a minimum value of the preprocessed pulse signals, wherein the difference value is the maximum waveform height, and defining a threshold rule as follows: (1) the peak magnitude fluctuation range is not more than 0.3 times of the maximum waveform height; (2) the peak is larger than all amplitude values of about 200 sample points. When both threshold conditions are satisfied, the peak point may be determined to be the P-wave to be detected. And detecting the peak point of each 600 sample points of the pulse signal at intervals, and recording the interval between adjacent peaks while detecting the position of the peak point, thereby obtaining the cycle sequence of the pulse wave.
A deep learning neural network model is built, a multi-layer network is adopted for learning, more data are continuously used for training and testing until the passing rate of the tested data exceeds a certain interval, and the model is considered to be built to be completed if 85%.
The method and the device have the beneficial effects that through realizing a more humanized emotion theme regulation and control scheme, theme regulation and control of the wearable device is matched with emotion of the wearing object in the current state, so that the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
In this embodiment, first, basic attribute information of the wearing object is obtained, where the basic attribute information includes age information, gender information, work information, and physiological information; and then, learning to obtain the emotion characteristics of the wearing object in the current state by combining the heart rate information, the state information and the basic attribute information.
Optionally, in this embodiment, a data tag is added, for example, a first emotion feature and a second emotion feature in this embodiment, and corresponding heart rate information, state information and basic attribute information are added, a detection waveform of each tested person is labeled with a time corresponding to each emotion, unified classification is performed, statistics is performed in a multidimensional feature vector, and a corresponding mode corresponding to the same emotion is found.
In this embodiment, the method further includes, first, parsing the emotional characteristics; and then, determining the display interface attribute and the display content attribute corresponding to the emotion characteristics.
Optionally, in the daily use process of the wearable device, the heart rate waveform change of the user is continuously detected, the background neural network is matched and judged, different display subjects are switched according to the emotion judgment result, and targeted relief is detected, so that the visual angle experience of the user is improved, and the effects of relieving anxiety and calming mood of the user are achieved.
In this embodiment, first, determining the display theme category according to the display interface attribute; and then determining the display theme characteristics according to the display content attributes, wherein the display theme characteristics comprise theme contents.
Optionally, in the wearing state and the wearing environment, firstly adjusting and controlling the display theme category of the wearable device according to the emotion characteristics, then detecting the heart rate information and the state information again, and if the heart rate information and the state information meet preset adjusting and controlling conditions, executing subsequent adjustment and control of the display theme characteristics;
optionally, after performing subsequent adjustment and control of the display theme feature, detecting the heart rate information and the state information again, and if the heart rate information and the state information meet preset adjustment and control conditions, repeating the step of adjusting and controlling the display theme category of the wearable device according to the emotion feature.
Examples ten
Based on the above embodiments, the present invention also proposes a computer-readable storage medium, on which an emotion theme adjustment program is stored, which when executed by a processor, implements the steps of the emotion theme adjustment method as set forth in any one of the above.
According to the emotion theme regulation and control method, equipment and computer readable storage medium, the wearing state and the wearing environment of the wearable equipment are obtained; then, acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment; then, learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information; finally, in the wearing state and the wearing environment, the display theme category and the display theme feature of the wearable device are regulated and controlled according to the emotion feature. The humanized emotion theme regulation and control scheme is realized, so that theme regulation and control of the wearable device is matched with emotion of a wearing object in a current state, the wearable device has stronger affinity, user viscosity is improved, and user experience is enhanced.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, the element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article or apparatus that comprises the element.
The foregoing embodiment numbers of the present invention are merely for the purpose of description, and do not represent the advantages or disadvantages of the embodiments.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM, magnetic disk, optical disk) comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the method according to the embodiments of the present invention.
The embodiments of the present invention have been described above with reference to the accompanying drawings, but the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many forms may be made by those having ordinary skill in the art without departing from the spirit of the present invention and the scope of the claims, which are to be protected by the present invention.

Claims (7)

1. A method for regulating emotion theme, said method comprising:
acquiring the wearing state and wearing environment of the wearable equipment;
acquiring heart rate information of a wearing object according to the wearing state, and acquiring state information of the wearing object according to the wearing environment;
learning to obtain the emotion characteristics of the wearing object by combining the heart rate information and the state information;
regulating and controlling the display theme category and the display theme characteristic of the wearable equipment according to the emotion characteristic in the wearing state and the wearing environment;
wherein,,
when the fluctuation amplitude of the heart rate information is larger than a preset value or the state information meets a preset trigger condition, acquiring the emotion characteristics of the wearing object, wherein the trigger condition comprises switching the wearing position of the wearing equipment and/or switching the touch pressure sensing value of the wearing equipment to be larger than a preset threshold;
Firstly regulating and controlling the display theme category of the wearable equipment according to the emotion characteristics in the wearing state and the wearing environment, then detecting the heart rate information and the state information again, and executing subsequent display theme characteristic regulation if preset regulation conditions are met;
after the subsequent regulation and control of the display theme characteristics are executed, detecting the heart rate information and the state information again, and if the preset regulation and control conditions are met, repeating the step of regulating and controlling the display theme category of the wearable equipment according to the emotion characteristics;
the acquiring the wearing state and the wearing environment of the wearable device includes:
acquiring a wearing environment of the wearable device, wherein the wearing environment comprises an internal environment and an external environment;
determining the internal environment through the system running state and the application running state of the wearable device, wherein the internal environment further comprises a current foreground application program, a recently started application program and a current system setting item, and the system setting item comprises volume, light intensity of a display screen and color setting type of the display screen;
Determining the external environment through a light sensing assembly, a temperature sensing assembly and a humidity sensing assembly of the wearable device, wherein the external environment further comprises current illumination intensity, current environment temperature and current environment humidity;
the step of acquiring the heart rate information of the wearing object according to the wearing state, and simultaneously acquiring the state information of the wearing object according to the wearing environment comprises the following steps:
determining a first emotional characteristic affecting the emotion of the wearing object according to the internal environment;
determining a second emotional characteristic affecting the emotion of the wearing object according to the external environment;
the learning to obtain the emotional characteristics of the wearing object by combining the heart rate information and the state information comprises the following steps:
acquiring basic attribute information of the wearing object, wherein the basic attribute information comprises age information, gender information, work information and physiological information;
learning to obtain the emotion characteristics of the wearing object in the current state by combining the heart rate information, the state information and the basic attribute information;
wherein,,
and adding a data tag to the first emotion feature, the second emotion feature, the corresponding heart rate information, the corresponding state information and the corresponding basic attribute information, labeling the time corresponding to each emotion to the detection waveform of each wearing object, uniformly classifying, and counting in a multidimensional feature vector to find out the corresponding mode corresponding to the same emotion.
2. The emotion theme regulation method according to claim 1, wherein the acquiring the wearing state and the wearing environment of the wearable apparatus includes:
acquiring a wearing state of the wearable device, wherein the wearing state comprises a wearing position and a motion state;
according to the touch control signal and/or the buckling signal of the wearable equipment, adjusting the wearing position;
and adjusting the motion state according to the motion sensing signal of the wearable equipment.
3. The emotion theme regulation and control method according to claim 2, wherein the acquiring heart rate information of the wearing object according to the wearing state, and simultaneously acquiring state information of the wearing object according to the wearing environment, includes:
positioning a detection area corresponding to heart rate detection in the wearable device according to the wearing position;
and acquiring the heart rate information through a heart rate detection component in the detection area.
4. The emotion theme regulation method according to claim 3, wherein regulating the display theme category and the display theme feature of the wearable apparatus according to the emotion feature in the wearing state and the wearing environment includes:
Analyzing the emotion characteristics;
and determining the display interface attribute and the display content attribute corresponding to the emotion characteristics.
5. The emotion theme regulation method of claim 4, wherein the regulating the display theme category and the display theme feature of the wearable apparatus according to the emotion feature in the wearing state and the wearing environment further includes:
determining the display theme category according to the display interface attribute;
and determining the display theme characteristics according to the display content attributes, wherein the display theme characteristics comprise theme contents.
6. An emotion theme regulation apparatus, characterized by comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program implementing the steps of the method according to any one of claims 1 to 5 when executed by the processor.
7. A computer-readable storage medium, wherein an emotion theme regulation program is stored on the computer-readable storage medium, which when executed by a processor, implements the steps of the emotion theme regulation method as claimed in any one of claims 1 to 5.
CN201910360042.9A 2019-04-30 2019-04-30 Emotion theme regulation and control method, equipment and computer-readable storage medium Active CN110013260B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910360042.9A CN110013260B (en) 2019-04-30 2019-04-30 Emotion theme regulation and control method, equipment and computer-readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910360042.9A CN110013260B (en) 2019-04-30 2019-04-30 Emotion theme regulation and control method, equipment and computer-readable storage medium

Publications (2)

Publication Number Publication Date
CN110013260A CN110013260A (en) 2019-07-16
CN110013260B true CN110013260B (en) 2023-05-05

Family

ID=67192904

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910360042.9A Active CN110013260B (en) 2019-04-30 2019-04-30 Emotion theme regulation and control method, equipment and computer-readable storage medium

Country Status (1)

Country Link
CN (1) CN110013260B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111166354B (en) * 2020-01-23 2022-11-18 北京津发科技股份有限公司 Method for analyzing factors influencing emotion change and electronic equipment
CN111166293A (en) * 2020-01-23 2020-05-19 北京津发科技股份有限公司 Analysis device for factors influencing emotional changes
CN114287899A (en) * 2021-11-30 2022-04-08 歌尔光学科技有限公司 Wearable device control method and device, wearable device and medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9418390B2 (en) * 2012-09-24 2016-08-16 Intel Corporation Determining and communicating user's emotional state related to user's physiological and non-physiological data
CN105224172A (en) * 2015-09-25 2016-01-06 百度在线网络技术(北京)有限公司 A kind of theme output intent and system
US20170351768A1 (en) * 2016-06-03 2017-12-07 Intertrust Technologies Corporation Systems and methods for content targeting using emotional context information
CN107436763A (en) * 2017-07-04 2017-12-05 广东小天才科技有限公司 A kind of method and wearable device for switching wearable device interface subject
CN107997753A (en) * 2017-11-15 2018-05-08 惠州市桑莱士智能科技股份有限公司 A kind of heart rate wears detection sensor module
CN108594991A (en) * 2018-03-28 2018-09-28 努比亚技术有限公司 A kind of method, apparatus and computer storage media that help user to adjust mood

Also Published As

Publication number Publication date
CN110013260A (en) 2019-07-16

Similar Documents

Publication Publication Date Title
CN109938720B (en) Heart rate-based reminding method, wearable device and computer-readable storage medium
CN109982179B (en) Audio signal output method and device, wearable device and storage medium
CN110013260B (en) Emotion theme regulation and control method, equipment and computer-readable storage medium
CN109947249B (en) Interaction method of wearable device, wearable device and computer storage medium
CN107734170B (en) Notification message processing method, mobile terminal and wearable device
CN109799912B (en) Display control method, device and computer readable storage medium
CN110225282B (en) Video recording control method, device and computer readable storage medium
CN110177209B (en) Video parameter regulation and control method, device and computer readable storage medium
CN110177208B (en) Video recording association control method, equipment and computer readable storage medium
CN110069193B (en) Interface switching method of wearable device, wearable device and storage medium
CN110086808B (en) Account login control method and device and computer readable storage medium
CN109933400B (en) Display interface layout method, wearable device and computer readable storage medium
CN110570625A (en) safety reminding method of intelligent wearable device, intelligent wearable device and storage medium
CN110086929A (en) Breath screen display methods, mobile phone, wearable device and computer readable storage medium
CN110096150B (en) Search interaction control method, equipment and computer readable storage medium
CN110650289B (en) Shooting depth of field control method, equipment and computer readable storage medium
CN110083427B (en) Application program interaction control method, device and computer readable storage medium
CN110083513B (en) Lamp effect control method, equipment and computer readable storage medium
CN110069102B (en) Display area regulation and control method and device and computer readable storage medium
CN110187950B (en) Method for adjusting picture display position, wearable device and storage medium
CN110174981B (en) Icon interaction control method and device and computer readable storage medium
CN110049253B (en) Focusing control method and device and computer readable storage medium
CN110049248B (en) Shot object regulation and control method and device and computer readable storage medium
CN110109605B (en) Image sharing method and device and computer readable storage medium
CN109639894B (en) Information prompting method, terminal equipment and computer readable medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20230403

Address after: Room 1502, 13th floor, No. 52 North Fourth Ring West Road, Haidian District, Beijing, 100080

Applicant after: Good Feeling Health Industry Group Co.,Ltd.

Address before: 518000 6-8 / F, 10-11 / F, 6 / F, 6-10 / F, C zone, Han's innovation building, No. 9018, Beihuan Avenue, high tech Zone, Nanshan District, Shenzhen City, Guangdong Province

Applicant before: NUBIA TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant