CN110299100B - Display direction adjustment method, wearable device and computer readable storage medium - Google Patents

Display direction adjustment method, wearable device and computer readable storage medium Download PDF

Info

Publication number
CN110299100B
CN110299100B CN201910585501.3A CN201910585501A CN110299100B CN 110299100 B CN110299100 B CN 110299100B CN 201910585501 A CN201910585501 A CN 201910585501A CN 110299100 B CN110299100 B CN 110299100B
Authority
CN
China
Prior art keywords
face image
wearing state
preset
wearable device
calculation result
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910585501.3A
Other languages
Chinese (zh)
Other versions
CN110299100A (en
Inventor
李荣涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nubia Technology Co Ltd
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201910585501.3A priority Critical patent/CN110299100B/en
Publication of CN110299100A publication Critical patent/CN110299100A/en
Application granted granted Critical
Publication of CN110299100B publication Critical patent/CN110299100B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G3/00Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
    • G09G3/20Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters

Abstract

The application discloses a display direction adjustment method, a wearable device and a computer readable storage medium, wherein the method comprises the following steps: detecting whether the wearable device is in a wearing state; if the wearable equipment is in the wearing state, judging that the wearing state is a left hand wearing state or a right hand wearing state; and adjusting the display direction of the screen of the wearable device based on the judgment result. Through the application, when the wearable device is detected to be in the wearing state, whether the wearable device is worn on the left hand or the right hand is automatically judged, and then according to the judging result, the display direction of the screen of the wearable device is automatically adjusted, so that the convenience of using the wearable device by a user is improved.

Description

Display direction adjustment method, wearable device and computer readable storage medium
Technical Field
The application relates to the technical field of wearable equipment, in particular to a display direction adjusting method, wearable equipment and a computer readable storage medium.
Background
With rapid development of electronic technology, wearable devices such as intelligent electronic bracelets and watches are favored by more and more users. Because different users have different wearing preferences, some people prefer to wear the wearable device in the left hand and some people prefer to wear the wearable device in the right hand. Some wearable device manufacturers set a screen display mode of wearing corresponding wearable devices by the left hand and a screen display mode of wearing corresponding wearable devices by the right hand in order to meet wearing requirements of different users. After wearing the wearable device, the user needs to further manually select left-hand wearing or right-hand wearing, so as to set a corresponding screen display mode.
Disclosure of Invention
The main aim of the application is that after wearing wearable equipment, users need to further manually select left hand wearing or right hand wearing in the prior art, so that the technical problem of corresponding screen display modes is solved.
In order to achieve the above object, the present application provides a display direction adjustment method, including the steps of:
detecting whether the wearable device is in a wearing state;
if the wearable equipment is in the wearing state, judging that the wearing state is a left hand wearing state or a right hand wearing state;
and adjusting the display direction of the screen of the wearable device based on the judgment result.
Optionally, the step of detecting whether the wearable device is in a wearing state includes:
detecting whether the pressure sensor continuously collects pressure data which is larger than a preset pressure value, wherein the duration is longer than the preset duration;
if the pressure sensor continuously collects the pressure data which is larger than the preset pressure value and the duration is longer than the preset duration, the wearable device is determined to be in a wearing state.
Optionally, if the wearable device is in the wearing state, the step of determining that the wearing state is a left-hand wearing state or a right-hand wearing state includes:
If the wearable equipment is in a wearing state, controlling a camera device of the wearable equipment to be started and shoot;
and acquiring a picture shot by the image pickup device, and judging whether the wearing state is a left-hand wearing state or a right-hand wearing state according to the picture.
Optionally, the step of determining, according to the picture, whether the wearing state is a left-hand wearing state or a right-hand wearing state includes:
detecting whether a face image exists in the picture;
if the face image exists in the picture, calculating the similarity between the face image and a preset left face image and the similarity between the face image and a preset right face image respectively to obtain a similarity calculation result between the face image and the preset left face image and a similarity calculation result between the face image and the preset right face image;
if the similarity calculation result of the face image and the preset left face image is larger than the similarity calculation result of the face image and the preset right face image, determining that the wearing state is a left hand wearing state;
and if the similarity calculation result of the face image and the preset right face image is larger than the similarity calculation result of the face image and the preset left face image, determining that the wearing state is a right hand wearing state.
Optionally, the step of calculating the similarity between the face image and the preset left face image and the similarity between the face image and the preset right face image, respectively, to obtain a similarity calculation result between the face image and the preset left face image and a similarity calculation result between the face image and the preset right face image includes:
acquiring characteristic information of the face image, characteristic information of the preset left face image and characteristic information of the preset right face image;
calculating Euclidean distance and cosine distance between the feature information of the face image and the feature information of the preset left face image to obtain a first Euclidean distance and a first cosine distance;
calculating the average value of the first Euclidean distance and the first cosine distance to obtain a first average value calculation result, and taking the first average value calculation result as a similarity calculation result of the face image and a preset left face image;
calculating Euclidean distance and cosine distance between the feature information of the face image and the feature information of the preset right face image to obtain a second Euclidean distance and a second cosine distance;
and calculating the average value of the second Euclidean distance and the second cosine distance to obtain a second average value calculation result, and taking the second average value calculation result as a similarity calculation result of the face image and a preset right face image.
Optionally, after the step of detecting whether a face image exists in the picture, the method further includes:
if no face image exists in the picture, when a hand lifting action is detected, the friction force direction of the wearable equipment and the arm is obtained;
and judging whether the wearing state is a left hand wearing state or a right hand wearing state according to the friction force direction.
Optionally, the step of determining that the wearing state is a left-hand wearing state or a right-hand wearing state according to the friction force direction includes:
when the direction component of the friction force direction on the preset shaft points to the positive direction of the preset shaft, determining that the wearing state is a right-hand wearing state;
and when the direction component of the friction force direction on the preset shaft points to the negative direction of the preset shaft, determining that the wearing state is a left-hand wearing state.
Optionally, after the step of adjusting the display direction of the screen of the wearable device based on the determination result, the method further includes:
if the wearing state of the wearable device is not detected to be changed, the current display direction is kept unchanged.
In addition, to achieve the above object, the present application further provides a wearable device, including: a memory, a processor, and a computer program stored on the memory and executable on the processor; the computer program when executed by the processor implements the steps of the display orientation adjustment method as described above.
In addition, in order to achieve the above object, the present application further provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the steps of the display direction adjustment method as described above.
In the application, whether the wearable equipment is in a wearing state is detected; if the wearable equipment is in the wearing state, judging that the wearing state is a left hand wearing state or a right hand wearing state; and adjusting the display direction of the screen of the wearable device based on the judgment result. Through the application, when the wearable device is detected to be in the wearing state, whether the wearable device is worn on the left hand or the right hand is automatically judged, and then according to the judging result, the display direction of the screen of the wearable device is automatically adjusted, so that the convenience of using the wearable device by a user is improved.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention.
In order to more clearly illustrate the embodiments of the invention or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, and it will be obvious to a person skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a schematic hardware structure of an implementation manner of a wearable device according to an embodiment of the present invention;
fig. 2 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 3 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 4 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
fig. 5 is a hardware schematic of an implementation of a wearable device provided in an embodiment of the present application;
FIG. 6 is a flow chart illustrating an embodiment of a direction adjustment method according to the present application;
fig. 7 is a schematic diagram of coordinate axis directions relative to a wearable device.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
In the following description, suffixes such as "module", "component", or "unit" for representing elements are used only for facilitating the description of the present invention, and have no specific meaning per se. Thus, "module," "component," or "unit" may be used in combination.
The wearable device provided by the embodiment of the invention comprises a mobile terminal such as an intelligent bracelet, an intelligent watch and an intelligent mobile phone. With the continuous development of screen technology, mobile terminals such as smart phones and the like can also be used as wearable devices due to the appearance of screen forms such as flexible screens, folding screens and the like. The wearable device provided in the embodiment of the invention can comprise: RF (Radio Frequency) unit, wiFi module, audio output unit, A/V (audio/video) input unit, sensor, display unit, user input unit, interface unit, memory, processor, and power supply.
In the following description, a wearable device will be taken as an example, please refer to fig. 1, which is a schematic hardware structure of a wearable device implementing various embodiments of the present invention, where the wearable device 100 may include: an RF (Radio Frequency) unit 101, a WiFi module 102, an audio output unit 103, an a/V (audio/video) input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, a processor 110, and a power supply 111. Those skilled in the art will appreciate that the wearable device structure shown in fig. 1 does not constitute a limitation of the wearable device, and that the wearable device may include more or fewer components than shown, or certain components in combination, or a different arrangement of components.
The following describes the various components of the wearable device in detail with reference to fig. 1:
the radio frequency unit 101 may be used to send and receive information or send signals in a call process, specifically, the radio frequency unit 101 may send uplink information to the base station, or may send downlink information sent by the base station to the processor 110 of the wearable device to process the downlink information, where the downlink information sent by the base station to the radio frequency unit 101 may be generated according to the uplink information sent by the radio frequency unit 101, or may be actively pushed to the radio frequency unit 101 after detecting that the information of the wearable device is updated, for example, after detecting that the geographic position where the wearable device is located changes, the base station may send a notification of the change of the geographic position to the radio frequency unit 101 of the wearable device, after receiving the notification of the message, the radio frequency unit 101 may send the notification of the message to the processor 110 of the wearable device to process, and the processor 110 of the wearable device may control the notification of the message to be displayed on the display panel 1061 of the wearable device; typically, the radio frequency unit 101 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may also communicate with a network and other devices through wireless communication, which may specifically include: through wireless communication with a server in a network system, for example, the wearable device can download file resources from the server through wireless communication, for example, an application program can be downloaded from the server, after the wearable device finishes downloading a certain application program, if the file resources corresponding to the application program in the server are updated, the server can push a message notification of the resource update to the wearable device through wireless communication so as to remind a user to update the application program. The wireless communication may use any communication standard or protocol, including but not limited to GSM (Global System of Mobile communication, global System for Mobile communications), GPRS (General Packet Radio Service ), CDMA2000 (Code Division Multiple Access, CDMA 2000), WCDMA (Wideband Code Division Multiple Access ), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time Division synchronous code Division multiple Access), FDD-LTE (Frequency Division Duplexing-Long Term Evolution, frequency Division Duplex Long term evolution), and TDD-LTE (Time Division Duplexing-Long Term Evolution, time Division Duplex Long term evolution), etc.
In one embodiment, the wearable device 100 may access an existing communication network by inserting a SIM card.
In another embodiment, the wearable device 100 may access an existing communication network by setting an esim card (Embedded-SIM), and by adopting the esim card, the internal space of the wearable device may be saved and the thickness may be reduced.
It will be appreciated that although fig. 1 shows a radio frequency unit 101, it will be appreciated that the radio frequency unit 101 is not an essential component of a wearable device and may be omitted entirely as required within the scope of not changing the essence of the invention. The wearable device 100 may implement communication connection with other devices or communication networks through the wifi module 102 alone, which is not limited by the embodiment of the present invention.
WiFi belongs to a short-distance wireless transmission technology, and the wearable device can help a user to send and receive emails, browse webpages, access streaming media and the like through the WiFi module 102, so that wireless broadband Internet access is provided for the user. Although fig. 1 shows a WiFi module 102, it is understood that it does not belong to the necessary constitution of the wearable device, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the WiFi module 102 or stored in the memory 109 into an audio signal and output as sound when the wearable device 100 is in a call signal reception mode, a talk mode, a recording mode, a voice recognition mode, a broadcast reception mode, or the like. Also, the audio output unit 103 may also provide audio output (e.g., call signal reception sound, message reception sound, etc.) related to a specific function performed by the wearable device 100. The audio output unit 103 may include a speaker, a buzzer, and the like.
The a/V input unit 104 is used to receive an audio or video signal. The a/V input unit 104 may include a graphics processor (Graphics Processing Unit, GPU) 1041 and a microphone 1042, the graphics processor 1041 processing image data of still pictures or video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The processed image frames may be displayed on the display unit 106. The image frames processed by the graphics processor 1041 may be stored in the memory 109 (or other storage medium) or transmitted via the radio frequency unit 101 or the WiFi module 102. The microphone 1042 can receive sound (audio data) via the microphone 1042 in a phone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound into audio data. The processed audio (voice) data may be converted into a format output that can be transmitted to the mobile communication base station via the radio frequency unit 101 in the case of a telephone call mode. The microphone 1042 may implement various types of noise cancellation (or suppression) algorithms to cancel (or suppress) noise or interference generated in the course of receiving and transmitting the audio signal.
In one embodiment, the wearable device 100 includes one or more cameras, and by opening the cameras, capturing of images, photographing, video recording and other functions can be achieved, and the positions of the cameras can be set as required.
The wearable device 100 further comprises at least one sensor 105, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 1061 according to the brightness of ambient light, and the proximity sensor can turn off the display panel 1061 and/or the backlight when the wearable device 100 moves to the ear. As one type of motion sensor, the accelerometer sensor can detect the acceleration in all directions (typically three axes), and can detect the gravity and direction when stationary, and can be used for applications for recognizing the gesture of a mobile phone (such as horizontal-vertical screen switching, related games, magnetometer gesture calibration), vibration recognition related functions (such as pedometer, knocking), and the like.
In one embodiment, the wearable device 100 further comprises a proximity sensor, by employing the proximity sensor, the wearable device is able to achieve non-contact manipulation, providing more modes of operation.
In one embodiment, the wearable device 100 further comprises a heart rate sensor, which when worn, enables detection of heart rate by being in close proximity to the user.
In one embodiment, the wearable device 100 may further include a fingerprint sensor, by reading a fingerprint, security verification or the like can be achieved.
The display unit 106 is used to display information input by a user or information provided to the user. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display (Liquid Crystal Display, LCD), an Organic Light-Emitting Diode (OLED), or the like.
In one embodiment, the display panel 1061 employs a flexible display screen, and the wearable device employing the flexible display screen is capable of bending when worn, thereby fitting more. Optionally, the flexible display screen may be an OLED screen body and a graphene screen body, and in other embodiments, the flexible display screen may also be other display materials, which is not limited to this embodiment.
In one embodiment, the display panel 1061 of the wearable device may take a rectangular shape for ease of wrapping when worn. In other embodiments, other approaches may be taken as well.
The user input unit 107 may be used to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the wearable device. In particular, the user input unit 107 may include a touch panel 1071 and other input devices 1072. The touch panel 1071, also referred to as a touch screen, may collect touch operations thereon or thereabout by a user (e.g., operations of the user on the touch panel 1071 or thereabout by using any suitable object or accessory such as a finger, a stylus, etc.) and drive the corresponding connection device according to a predetermined program. The touch panel 1071 may include two parts of a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device, converts it into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute commands sent from the processor 110. Further, the touch panel 1071 may be implemented in various types such as resistive, capacitive, infrared, and surface acoustic wave. The user input unit 107 may include other input devices 1072 in addition to the touch panel 1071. In particular, other input devices 1072 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc., as specifically not limited herein.
In one embodiment, the sides of the wearable device 100 may be provided with one or more buttons. The button can realize a plurality of modes such as short pressing, long pressing, rotation and the like, thereby realizing a plurality of operation effects. The number of the buttons can be multiple, and different buttons can be combined for use, so that multiple operation functions are realized.
Further, the touch panel 1071 may overlay the display panel 1061, and when the touch panel 1071 detects a touch operation thereon or thereabout, the touch panel 1071 is transferred to the processor 110 to determine the type of touch event, and then the processor 110 provides a corresponding visual output on the display panel 1061 according to the type of touch event. Although in fig. 1, the touch panel 1071 and the display panel 1061 are two independent components for implementing the input and output functions of the wearable device, in some embodiments, the touch panel 1071 may be integrated with the display panel 1061 to implement the input and output functions of the wearable device, which is not limited herein. For example, when a message notification of a certain application is received through the rf unit 101, the processor 110 may control the message notification to be displayed in a certain preset area of the display panel 1061, where the preset area corresponds to a certain area of the touch panel 1071, and may control the message notification displayed in the corresponding area on the display panel 1061 by performing a touch operation on the certain area of the touch panel 1071.
The interface unit 108 serves as an interface through which at least one external device can be connected with the wearable apparatus 100. For example, the external devices may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification module, an audio input/output (I/O) port, a video I/O port, an earphone port, and the like. The interface unit 108 may be used to receive input (e.g., data information, power, etc.) from an external device and transmit the received input to one or more elements within the wearable apparatus 100 or may be used to transmit data between the wearable apparatus 100 and the external device.
In one embodiment, the interface unit 108 of the wearable device 100 adopts a contact structure, and is connected with other corresponding devices through the contact, so as to realize functions of charging, connection and the like. The contact can also be waterproof.
Memory 109 may be used to store software programs as well as various data. The memory 109 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), and a storage data area; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the handset, etc. In addition, memory 109 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
The processor 110 is a control center of the wearable device, connects various parts of the entire wearable device with various interfaces and lines, performs various functions of the wearable device and processes data by running or executing software programs and/or modules stored in the memory 109, and invoking data stored in the memory 109, thereby performing overall monitoring of the wearable device. Processor 110 may include one or more processing units; preferably, the processor 110 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The wearable device 100 may further include a power source 111 (such as a battery) for powering the various components, and preferably, the power source 111 may be logically connected to the processor 110 through a power management system, so as to perform functions of managing charging, discharging, and power consumption management through the power management system.
Although not shown in fig. 1, the wearable device 100 may further include a bluetooth module or the like, which is not described herein. The wearable device 100 can be connected with other terminal devices through bluetooth to realize communication and information interaction.
Fig. 2 to fig. 4 are schematic structural diagrams of a wearable device according to an embodiment of the present invention. The wearable device comprises a flexible screen. When the wearable device is unfolded, the flexible screen is in a strip shape; when the wearable device is in a wearing state, the flexible screen is bent to be annular. Fig. 2 and 3 show schematic structural diagrams of the wearable device screen when unfolded, and fig. 4 shows schematic structural diagrams of the wearable device screen when bent.
Based on the above embodiments, it can be seen that if the device is a wristwatch, a bracelet, or a wearable device, the screen of the device may not cover the watchband area of the device, or may cover the watchband area of the device. In this embodiment, the device may be a wristwatch, a bracelet, or a wearable device, and the device includes a screen and a connection portion. The screen may be a flexible screen and the connection may be a wristband. Alternatively, the screen of the device or the display area of the screen may be partially or fully overlaid on the wristband of the device. Fig. 5 is a schematic hardware diagram of an implementation manner of a wearable device according to an embodiment of the present application, where a screen of the device extends to two sides, and a part of the screen is covered on a watchband of the device. In other embodiments, the screen of the device may also be entirely covered on the watchband of the device, which is not limited to this embodiment.
Referring to fig. 6, fig. 6 is a flowchart illustrating an embodiment of a direction adjustment method according to the present application. In one embodiment, the display direction adjustment method includes the steps of:
step S10, detecting whether the wearable device is in a wearing state;
in an alternative embodiment, whether the wearable device is in a wearing state may be determined by detecting whether the wearable device is in contact with an arm. For example, the capacitance value is collected through a capacitance sensor arranged on a contact surface of the wearable device, which is in contact with the arm, and if the capacitance value which is smaller than the air reference capacitance value is continuously collected by the capacitance sensor and the duration time is longer than a preset time (for example, 30 seconds), the wearable device is determined to be in a wearing state.
In another alternative embodiment, step S10 includes:
detecting whether the pressure sensor continuously collects pressure data which is larger than a preset pressure value, wherein the duration is longer than the preset duration;
if the pressure sensor continuously collects the pressure data which is larger than the preset pressure value and the duration is longer than the preset duration, the wearable device is determined to be in a wearing state.
In this embodiment, a pressure sensor is provided at the bottom of the display portion and/or the bottom of the wristband portion of the wearable device. It will be readily appreciated that "bottom" refers to the side that is in contact with the arm. If the wearable device is worn on the arm, the pressure sensor can continuously detect the pressure data due to the supporting force of the arm on the wearable device. In this embodiment, if the pressure sensor continuously collects pressure data greater than a preset pressure value and the duration is greater than a preset duration, it is determined that the wearable device is in a wearing state. The magnitude of the preset pressure value and the magnitude of the preset duration are flexibly set according to actual conditions.
Step S20, if the wearable equipment is in a wearing state, judging that the wearing state is a left hand wearing state or a right hand wearing state;
in this embodiment, when the wearable device is detected to be in the wearing state, the wearing state is further determined to be a left-hand wearing state or a right-hand wearing state. For example, the wearing state of the wearable device is judged to be the left-hand wearing state or the right-hand wearing state by the picture shot by the camera of the wearable device, or the wearing state of the wearable device is judged to be the left-hand wearing state or the right-hand wearing state according to the spatial position change of the wearable device in the using process.
In an alternative embodiment, step S20 includes:
if the wearable equipment is in a wearing state, controlling a camera device of the wearable equipment to be started and shoot;
and acquiring a picture shot by the image pickup device, and judging whether the wearing state is a left-hand wearing state or a right-hand wearing state according to the picture.
In this embodiment, when it is detected that the wearable device is in a wearing state, the camera device of the wearable device is controlled to be turned on and shot. Wherein, camera device (such as camera) and the display screen of wearable equipment set up in same one side. And acquiring a picture shot by the image pickup device, and judging whether the wearing state is a left-hand wearing state or a right-hand wearing state according to the picture. It is easy to understand that when the wearable device is worn on the left hand and the right hand of the user, the photographed pictures are different, so that the wearing state can be judged to be the left hand wearing state or the right hand wearing state according to the difference of the pictures.
In an embodiment, the step of determining that the wearing state is a left-hand wearing state or a right-hand wearing state according to the picture includes:
detecting whether a face image exists in the picture;
in this embodiment, whether a face image exists in the picture can be detected through OpenCV, or whether a face image exists in the picture can be detected through a neural network model.
If the face image exists in the picture, calculating the similarity between the face image and a preset left face image and the similarity between the face image and a preset right face image respectively to obtain a similarity calculation result between the face image and the preset left face image and a similarity calculation result between the face image and the preset right face image;
in this embodiment, if a face image is detected to exist in the picture, the similarity between the face image and the preset left face image and the similarity between the face image and the preset right face image are calculated through a face similarity algorithm, so as to obtain a similarity calculation result between the face image and the preset left face image and a similarity calculation result between the face image and the preset right face image. The preset left face image is a left face image of a user to which the wearable device belongs, and the corresponding preset right face image is a right face image of the user to which the wearable device belongs.
If the similarity calculation result of the face image and the preset left face image is larger than the similarity calculation result of the face image and the preset right face image, determining that the wearing state is a left hand wearing state;
in this embodiment, if the similarity calculation result of the face image and the preset left face image is greater than the similarity calculation result of the face image and the preset right face image, it is indicated that the image is captured at the position on the left side of the face, so that the wearing state is determined to be the left-hand wearing state.
And if the similarity calculation result of the face image and the preset right face image is larger than the similarity calculation result of the face image and the preset left face image, determining that the wearing state is a right hand wearing state.
In this embodiment, if the similarity calculation result of the face image and the preset right face image is greater than the similarity calculation result of the face image and the preset left face image, it is indicated that the image is captured at the right side of the face, so that the wearing state is determined to be the right-hand wearing state.
Step S30, based on the judging result, adjusting the display direction of the screen of the wearable device.
In this embodiment, a screen display direction corresponding to the left-hand wearing state is preset and recorded as a first display direction; the screen display direction corresponding to the right hand wearing state is preset and is recorded as a second display direction. If the judgment result obtained in the step S20 is in a left-hand wearing state, adjusting the display direction of the screen of the wearable device to be a first display direction; and if the right-hand wearing state is determined according to the determination result obtained in the step S20, adjusting the display direction of the screen of the wearable device to be the second display direction.
In this embodiment, whether the wearable device is in a wearing state is detected; if the wearable equipment is in the wearing state, judging that the wearing state is a left hand wearing state or a right hand wearing state; and adjusting the display direction of the screen of the wearable device based on the judgment result. Through the embodiment, when the wearable device is detected to be in the wearing state, whether the wearable device is worn on the left hand or the right hand is automatically judged, and then according to the judging result, the display direction of the screen of the wearable device is automatically adjusted, so that the convenience of using the wearable device by a user is improved.
Further, in an embodiment of the display direction adjustment method of the present application, the step of calculating the similarity between the face image and the preset left face image and the similarity between the face image and the preset right face image, respectively, to obtain a similarity calculation result between the face image and the preset left face image and a similarity calculation result between the face image and the preset right face image include:
acquiring characteristic information of the face image, characteristic information of the preset left face image and characteristic information of the preset right face image;
in this embodiment, face feature extraction is performed on a preset left face image and a preset right face image in advance, so as to obtain feature information of the preset left face image and feature information of the preset right face image, and the feature information of the preset left face image and the feature information of the preset right face image are stored in a memory. Therefore, the characteristic information of the preset left face image and the characteristic information of the preset right face image can be directly obtained from the memory. The feature information of the face image is obtained by extracting the face features of the face image.
Calculating Euclidean distance and cosine distance between the feature information of the face image and the feature information of the preset left face image to obtain a first Euclidean distance and a first cosine distance; calculating the average value of the first Euclidean distance and the first cosine distance to obtain a first average value calculation result, and taking the first average value calculation result as a similarity calculation result of the face image and a preset left face image;
calculating Euclidean distance and cosine distance between the feature information of the face image and the feature information of the preset right face image to obtain a second Euclidean distance and a second cosine distance; and calculating the average value of the second Euclidean distance and the second cosine distance to obtain a second average value calculation result, and taking the second average value calculation result as a similarity calculation result of the face image and a preset right face image.
In this embodiment, in order to improve accuracy of the similarity calculation result, the euclidean distance and the cosine distance between the feature information of the face image and the feature information of the preset left face image are calculated, a first euclidean distance and a first cosine distance are obtained, an average value of the first euclidean distance and the first cosine distance is further calculated, a first average value calculation result is obtained, and the first average value calculation result is used as the similarity calculation result of the face image and the preset left face image. And similarly, calculating the Euclidean distance and the cosine distance of the characteristic information of the face image and the characteristic information of the preset right face image to obtain a second Euclidean distance and a second cosine distance, further calculating the average value of the second Euclidean distance and the second cosine distance to obtain a second average value calculation result, and taking the second average value calculation result as a similarity calculation result of the face image and the preset right face image.
In this embodiment, the euclidean distance and cosine distance between the face image and the preset left face/right face image are calculated, the average value of the euclidean distance and cosine distance obtained by calculation is taken as the similarity calculation result of the face image and the preset left face/right face image, and the accuracy of the similarity calculation result is improved.
Further, in an embodiment of the display direction adjustment method of the present application, after the step of detecting whether a face image exists in the picture, the method further includes:
if no face image exists in the picture, when a hand lifting action is detected, the friction force direction of the wearable equipment and the arm is obtained;
in this embodiment, whether the hand lifting action occurs may be determined by the accelerations in the three coordinate axis directions detected by the triaxial gyroscope provided in the wearable device. Or judging whether the hand lifting action occurs or not through the distance change detected by the infrared distance measuring sensor arranged on the wearable equipment.
And judging whether the wearing state is a left hand wearing state or a right hand wearing state according to the friction force direction.
In this embodiment, considering that in some cases, when the wearable device is in the wearing state, the image capturing device of the wearable device is controlled to be turned on and capture images, and if there is no face image in the image captured by the image capturing device, it is necessary to determine that the wearing state is a left-hand wearing state or a right-hand wearing state in other manners.
In this embodiment, when the wearable device is in the wearing state, if the user performs the hand lifting action, the wearable device and the arm tend to generate friction force, and when the worn arm is different, the direction of the friction force generated by the wearable device and the arm force is different, so that when the hand lifting action is detected, the friction force direction of the wearable device and the arm can be obtained through the friction force detector arranged on the contact surface of the wearable device and the arm, and the wearing state is judged to be the left hand wearing state or the right hand wearing state according to the friction force direction.
Further, in an embodiment of the display direction adjustment method of the present application, the step of determining that the wearing state is a left-hand wearing state or a right-hand wearing state according to the friction force direction includes:
when the direction component of the friction force direction on the preset shaft points to the positive direction of the preset shaft, determining that the wearing state is a right-hand wearing state;
and when the direction component of the friction force direction on the preset shaft points to the negative direction of the preset shaft, determining that the wearing state is a left-hand wearing state.
In an embodiment, referring to fig. 7, fig. 7 is a schematic diagram of a coordinate axis direction relative to a wearable device. In the figure, the X axis and the Y axis are in the same plane with the display screen of the wearable device, the positive direction of the X axis points to the right side of the display screen, and the positive direction of the Y axis points to the upper side of the display screen. In this embodiment, the X axis is set as a preset axis, and if the direction component of the friction force direction on the preset axis points to the positive direction of the preset axis, the movement direction of the wearable device on the X axis is left when the hand lifting action is performed, that is, the wearable device is worn on the right hand, that is, the wearing state is the right hand wearing state; if the direction component of the friction force direction on the preset axis points to the negative direction of the preset axis, the movement direction of the wearable device on the X axis is rightward when the hand lifting action is performed, namely the wearable device is worn on the left hand, namely the wearing state is the left hand wearing state.
Further, in an embodiment of the present application, after step S30, the method further includes:
if the wearing state of the wearable device is not detected to be changed, the current display direction is kept unchanged.
In this embodiment, the wearing state of the wearable device changes from the wearing state to the unworn state, if the wearing state changes, the pressure value collected by the pressure sensor will be less than the preset pressure value, so if the pressure sensor continuously collects the pressure value less than the preset pressure value, the wearing state of the wearable device is determined to change. If the above situation does not occur, that is, the wearing state of the wearable device is not detected to change, that is, the wearing hand is not changed, the current display direction needs to be kept unchanged.
In addition, the embodiment of the application also provides a wearable device, which comprises:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program when executed by the processor implements the steps of the various embodiments of the display orientation adjustment method described above.
The specific embodiments of the wearable device are substantially the same as the embodiments of the above-mentioned display direction adjustment method, and are not described herein.
In addition, the embodiment of the application further provides a computer readable storage medium, wherein the computer readable storage medium stores a display direction adjusting program, and the display direction adjusting program realizes the steps of each embodiment of the display direction adjusting method when being executed by a processor.
The specific embodiments of the computer readable storage medium are basically the same as the embodiments of the display direction adjustment method described above, and are not described herein.

Claims (7)

1. A display direction adjustment method, characterized in that the display direction adjustment method comprises the steps of:
detecting whether the wearable device is in a wearing state;
if the wearable equipment is in the wearing state, judging that the wearing state is a left hand wearing state or a right hand wearing state;
based on the judgment result, adjusting the display direction of the screen of the wearable device;
wherein,
if the wearable device is in the wearing state, the step of judging whether the wearing state is a left-hand wearing state or a right-hand wearing state comprises the following steps:
if the wearable equipment is in a wearing state, controlling a camera device of the wearable equipment to be started and shoot;
acquiring a picture shot by the camera device;
Detecting whether a face image exists in the picture;
if the face image exists in the picture, judging whether the wearing state is a left hand wearing state or a right hand wearing state according to the picture; if no face image exists in the picture, when a hand lifting action is detected, the friction force direction of the wearable equipment and the arm is obtained;
judging whether the wearing state is a left hand wearing state or a right hand wearing state according to the friction force direction;
wherein,
setting an X axis and a Y axis to be in the same plane with a display screen of the wearable device, wherein the positive direction of the X axis points to the right side of the display screen, and the positive direction of the Y axis points to the upper side of the display screen;
setting the X axis as a preset axis;
when the direction component of the friction force direction on the preset axis points to the positive direction of the preset axis, the movement direction of the wearable equipment on the X axis is leftwards, and the wearing state is determined to be a right-hand wearing state;
when the direction component of the friction force direction on the preset axis points to the negative direction of the preset axis, the movement direction of the wearable equipment on the X axis is rightward, and the wearing state is determined to be a left-hand wearing state.
2. The display direction adjustment method according to claim 1, wherein the step of detecting whether the wearable device is in a wearing state includes:
Detecting whether the pressure sensor continuously collects pressure data which is larger than a preset pressure value, wherein the duration is longer than the preset duration;
if the pressure sensor continuously collects the pressure data which is larger than the preset pressure value and the duration is longer than the preset duration, the wearable device is determined to be in a wearing state.
3. The display direction adjustment method according to claim 1, wherein the step of judging whether the wearing state is a left-hand wearing state or a right-hand wearing state according to the picture includes:
respectively calculating the similarity of the face image and a preset left face image and the similarity of the face image and a preset right face image to obtain a similarity calculation result of the face image and the preset left face image and a similarity calculation result of the face image and the preset right face image;
if the similarity calculation result of the face image and the preset left face image is larger than the similarity calculation result of the face image and the preset right face image, determining that the wearing state is a left hand wearing state;
and if the similarity calculation result of the face image and the preset right face image is larger than the similarity calculation result of the face image and the preset left face image, determining that the wearing state is a right hand wearing state.
4. A display direction adjustment method according to claim 3, wherein the step of calculating the similarity of the face image to a preset left face image and a preset right face image, respectively, to obtain a similarity calculation result of the face image to the preset left face image and a similarity calculation result of the face image to the preset right face image comprises:
acquiring characteristic information of the face image, characteristic information of the preset left face image and characteristic information of the preset right face image;
calculating Euclidean distance and cosine distance between the feature information of the face image and the feature information of the preset left face image to obtain a first Euclidean distance and a first cosine distance;
calculating the average value of the first Euclidean distance and the first cosine distance to obtain a first average value calculation result, and taking the first average value calculation result as a similarity calculation result of the face image and a preset left face image;
calculating Euclidean distance and cosine distance between the feature information of the face image and the feature information of the preset right face image to obtain a second Euclidean distance and a second cosine distance;
and calculating the average value of the second Euclidean distance and the second cosine distance to obtain a second average value calculation result, and taking the second average value calculation result as a similarity calculation result of the face image and a preset right face image.
5. The display direction adjustment method according to any one of claims 1 to 4, characterized by further comprising, after the step of adjusting the display direction of the screen of the wearable device based on the determination result:
if the wearing state of the wearable device is not detected to be changed, the current display direction is kept unchanged.
6. A wearable device, the wearable device comprising:
a memory, a processor, and a computer program stored on the memory and executable on the processor;
the computer program, when executed by the processor, implements the steps of the display orientation adjustment method according to any one of claims 1 to 5.
7. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, implements the steps of the display direction adjustment method according to any one of claims 1 to 5.
CN201910585501.3A 2019-07-01 2019-07-01 Display direction adjustment method, wearable device and computer readable storage medium Active CN110299100B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910585501.3A CN110299100B (en) 2019-07-01 2019-07-01 Display direction adjustment method, wearable device and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910585501.3A CN110299100B (en) 2019-07-01 2019-07-01 Display direction adjustment method, wearable device and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN110299100A CN110299100A (en) 2019-10-01
CN110299100B true CN110299100B (en) 2024-03-22

Family

ID=68029757

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910585501.3A Active CN110299100B (en) 2019-07-01 2019-07-01 Display direction adjustment method, wearable device and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN110299100B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114098755A (en) * 2020-08-25 2022-03-01 华为终端有限公司 ECG waveform display method, medium and electronic device thereof
CN112162595B (en) * 2020-09-23 2022-12-06 深圳市爱都科技有限公司 Vertical arm rotation identification method and wearable terminal
CN114089814A (en) * 2021-11-08 2022-02-25 广东乐心医疗电子股份有限公司 Display method and device and electronic equipment
CN114485739B (en) * 2021-12-27 2022-11-29 荣耀终端有限公司 Parameter calibration method and wearable device
CN115061354A (en) * 2022-06-08 2022-09-16 深圳市万如夏科技有限公司 Screen awakening and sleeping method of smart watch
CN116567141B (en) * 2023-07-10 2023-11-07 荣耀终端有限公司 Data processing method and electronic equipment

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013061176A (en) * 2011-09-12 2013-04-04 Seiko Epson Corp Wrist wearing type electronic equipment and control method therefor
CN103049736A (en) * 2011-10-17 2013-04-17 天津市亚安科技股份有限公司 Face identification method based on maximum stable extremum area
CN103744531A (en) * 2013-12-20 2014-04-23 深圳市奋达科技股份有限公司 Smart watch left or right function button conversion method and system
CN104484049A (en) * 2014-12-31 2015-04-01 联想(北京)有限公司 Information processing method and electronic equipment
CN104573642A (en) * 2014-12-26 2015-04-29 小米科技有限责任公司 Face recognition method and device
CN105631408A (en) * 2015-12-21 2016-06-01 小米科技有限责任公司 Video-based face album processing method and processing device
CN105955479A (en) * 2016-04-29 2016-09-21 联想(北京)有限公司 Information processing method and wearable device
CN106291121A (en) * 2016-07-29 2017-01-04 歌尔股份有限公司 The wearing state detection method of a kind of wearable device and wearable device
JP2017207848A (en) * 2016-05-17 2017-11-24 レノボ・シンガポール・プライベート・リミテッド Portable information terminal, method of determining worn arm, method of determining worn direction, and program
CN107944020A (en) * 2017-12-11 2018-04-20 深圳云天励飞技术有限公司 Facial image lookup method and device, computer installation and storage medium
CN108848427A (en) * 2018-04-26 2018-11-20 Oppo广东移动通信有限公司 Wearing state detection method and relevant device
CN109445267A (en) * 2018-12-29 2019-03-08 北京金山安全软件有限公司 Intelligent watch
CN109614027A (en) * 2018-12-04 2019-04-12 青岛真时科技有限公司 A kind of configuration method and smartwatch of smartwatch

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9606721B2 (en) * 2013-07-22 2017-03-28 Lg Electronics Inc. Mobile terminal and control method thereof
JP2015141293A (en) * 2014-01-28 2015-08-03 ソニー株式会社 Display control apparatus, display control method, program, and display device

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013061176A (en) * 2011-09-12 2013-04-04 Seiko Epson Corp Wrist wearing type electronic equipment and control method therefor
CN103049736A (en) * 2011-10-17 2013-04-17 天津市亚安科技股份有限公司 Face identification method based on maximum stable extremum area
CN103744531A (en) * 2013-12-20 2014-04-23 深圳市奋达科技股份有限公司 Smart watch left or right function button conversion method and system
CN104573642A (en) * 2014-12-26 2015-04-29 小米科技有限责任公司 Face recognition method and device
CN104484049A (en) * 2014-12-31 2015-04-01 联想(北京)有限公司 Information processing method and electronic equipment
CN105631408A (en) * 2015-12-21 2016-06-01 小米科技有限责任公司 Video-based face album processing method and processing device
CN105955479A (en) * 2016-04-29 2016-09-21 联想(北京)有限公司 Information processing method and wearable device
JP2017207848A (en) * 2016-05-17 2017-11-24 レノボ・シンガポール・プライベート・リミテッド Portable information terminal, method of determining worn arm, method of determining worn direction, and program
CN106291121A (en) * 2016-07-29 2017-01-04 歌尔股份有限公司 The wearing state detection method of a kind of wearable device and wearable device
CN107944020A (en) * 2017-12-11 2018-04-20 深圳云天励飞技术有限公司 Facial image lookup method and device, computer installation and storage medium
CN108848427A (en) * 2018-04-26 2018-11-20 Oppo广东移动通信有限公司 Wearing state detection method and relevant device
CN109614027A (en) * 2018-12-04 2019-04-12 青岛真时科技有限公司 A kind of configuration method and smartwatch of smartwatch
CN109445267A (en) * 2018-12-29 2019-03-08 北京金山安全软件有限公司 Intelligent watch

Also Published As

Publication number Publication date
CN110299100A (en) 2019-10-01

Similar Documents

Publication Publication Date Title
CN110299100B (en) Display direction adjustment method, wearable device and computer readable storage medium
CN109982179B (en) Audio signal output method and device, wearable device and storage medium
CN110213810B (en) Wearable device control method, wearable device and computer-readable storage medium
CN109953435B (en) Method for automatically adjusting tightness of watchband, wearable device and storage medium
CN110177209B (en) Video parameter regulation and control method, device and computer readable storage medium
CN110187769B (en) Preview image viewing method, equipment and computer readable storage medium
CN109947524B (en) Interface display method, wearable device and computer readable storage medium
CN110225282B (en) Video recording control method, device and computer readable storage medium
CN110139270B (en) Wearable device pairing method, wearable device and computer readable storage medium
CN110177208B (en) Video recording association control method, equipment and computer readable storage medium
CN110113529B (en) Shooting parameter regulation and control method and device and computer readable storage medium
CN110072071B (en) Video recording interaction control method, equipment and computer readable storage medium
CN110620875B (en) Screenshot control method, equipment and computer readable storage medium in video shooting process
CN110071866B (en) Instant messaging application control method, wearable device and storage medium
CN110657794A (en) Compass calibration method of wearable device, wearable device and storage medium
CN109901812B (en) Volume adjustment method for wearable device, wearable device and readable storage medium
CN110650289B (en) Shooting depth of field control method, equipment and computer readable storage medium
CN110191282B (en) Shooting parameter regulation and control method and device and computer readable storage medium
CN110062167B (en) Linkage shooting control method and device and computer readable storage medium
CN110187950B (en) Method for adjusting picture display position, wearable device and storage medium
CN109976613B (en) Text input implementation method, wearable device and computer-readable storage medium
CN110049248B (en) Shot object regulation and control method and device and computer readable storage medium
CN110049253B (en) Focusing control method and device and computer readable storage medium
CN110138965B (en) Wearable device, number formatting method and device thereof, and storage medium
CN110109605B (en) Image sharing method and device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant