WO2018099017A1 - 显示方法、装置及终端 - Google Patents

显示方法、装置及终端 Download PDF

Info

Publication number
WO2018099017A1
WO2018099017A1 PCT/CN2017/085744 CN2017085744W WO2018099017A1 WO 2018099017 A1 WO2018099017 A1 WO 2018099017A1 CN 2017085744 W CN2017085744 W CN 2017085744W WO 2018099017 A1 WO2018099017 A1 WO 2018099017A1
Authority
WO
WIPO (PCT)
Prior art keywords
screen display
display area
target
preset
result
Prior art date
Application number
PCT/CN2017/085744
Other languages
English (en)
French (fr)
Inventor
童碧峰
陈浩
陈晓晓
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to CN201780009703.2A priority Critical patent/CN108604143B/zh
Publication of WO2018099017A1 publication Critical patent/WO2018099017A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means

Definitions

  • the present invention relates to the field of mobile terminals, and in particular, to a display method, apparatus, and terminal.
  • each screen is displayed differently according to the needs of the user during use.
  • the content required by the user is displayed on an inappropriate screen, the privacy may be leaked, and the user experience may be adversely affected; on the other hand, if the user simultaneously opens multiple screens while using the multi-screen mobile terminal Displaying will also increase the power consumption of the terminal and affect the endurance effect.
  • the gravity sensor provided in the mobile terminal is used to discriminate the usage state of the user, and the vertical upward screen is used as the screen that the user is using to perform screen display.
  • this method of gravity sensing can only make a rough judgment on the user's use state, and cannot accurately determine the screen that the user is using for display, and cannot handle the use scenario of the multi-screen mobile terminal.
  • the embodiment of the invention provides a display method, device and terminal, which acquires external information through a sensor, accurately identifies a user's location and usage state, and performs subsequent corresponding operations, thereby facilitating the use of the user.
  • a first aspect of the embodiments of the present invention discloses a display method, which is applied to a mobile terminal having at least two screen display areas, and the method includes:
  • the mobile terminal receives an open command, where the open command is used to indicate that the target interface is opened;
  • the mobile terminal acquires data reported by the target sensor;
  • the target sensor includes at least one sensor selected according to a preset sensor selection mechanism; and the sensor selection mechanism is a selection mechanism determined by a sensor type supported by the mobile terminal;
  • the mobile terminal determines a first screen display area in the at least two screen display areas according to the data reported by the target sensor, and presents the target interface in the first screen display area.
  • the mobile terminal selects at least one sensor as a target sensor by using the supported sensor type, and then performs a relatively accurate determination on the orientation and usage state and orientation of the user according to the data reported by the target sensor, and can fully integrate the sensor type of the mobile terminal.
  • the characteristics are used to judge and realize the intelligent selection of the appropriate screen to display the target interface.
  • the presenting the target in the first screen display area After the interface also includes:
  • the intervention operation is for determining a second screen display area
  • the detection is yes, determining a second screen display area in the at least two screen display areas according to the intervention operation, withdrawing the target interface presented in the first screen display area, and displaying on the second screen The target interface is presented in the zone.
  • the mobile terminal After the mobile terminal selects a suitable screen according to the reported data to display the target interface, if the selected screen display area is not the screen display area that the user wants to select, the user may perform an intervention operation and manually select a corresponding screen display area for correction. , further improving the accuracy of the display.
  • the mobile terminal according to the data reported by the target sensor, in the at least two The first screen display area is determined in the screen display area, including:
  • the mobile terminal detects, according to the data reported by the target sensor, whether there is user operation feature information for the screen display area;
  • the detection is yes, determining a target screen display area in the at least two screen display areas according to the user operation feature information, and determining whether user sign information exists in a vertical direction of the target screen display area;
  • the mobile terminal according to the data reported by the target sensor, before detecting whether there is user operation feature information for the screen display area, further includes:
  • Determining, in the at least two screen display areas, the target screen display area according to the user operation feature information including:
  • the pitch angle coordinates of the at least two screen display areas can be calculated by a gyroscope/gravity sensor, can quickly determine a first target screen display area that the user may use, and collect correspondingly based on the first target screen display area. Sensor data can increase productivity and reduce power consumption.
  • the mobile terminal judges the orientation and the use state of the user according to the reported user operation feature information and the user's physical information, and selects an appropriate screen display area to display the target screen.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, Determining, in the at least two screen display areas, a target screen display area according to the user operation feature information, and determining whether there is user vital sign information in a vertical direction of the target screen display area;
  • the first screen display area is determined in the target screen display area, including:
  • the screen display area where the touch pressure does not exist is determined as the target screen display area in the at least two screen display areas;
  • the target screen display area in which the ranging result is not greater than the preset distance threshold is determined as the first screen display area.
  • the mobile terminal is in the at least two screen display areas according to the data reported by the target sensor. Before determining the first screen display area, it also includes:
  • a first display decision is determined based on a sensor type of the mobile terminal.
  • the display decision result of the first display decision is that the target screen display area that the ranging result is not greater than the preset distance threshold is determined as the first screen display area.
  • the first display decision is determined according to the sensor type of the mobile terminal, including the mobile terminal automatically selecting a sensor according to the sensor type and determining the first display decision, or the user may select the sensor type determined by the mobile terminal. One shows the decision.
  • the third possible implementation manner of the first aspect provides a display method decision under low-precision conditions, and the mobile terminal performs low-precision conditions according to the acquired touch pressure detection data and the ranging detection data.
  • the fuzzy decision determines the screen display area.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, Determining, in the at least two screen display areas, a target screen display area according to the user operation feature information, and determining whether there is user vital sign information in a vertical direction of the target screen display area;
  • the first screen display area is determined in the target screen display area, including:
  • the screen display area where the touch pressure does not exist is determined as the target screen display area in the at least two screen display areas;
  • the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the mobile terminal before determining the first screen display area in the at least two screen display areas, according to the data reported by the target sensor, further includes:
  • a second display decision is determined based on the sensor type of the mobile terminal.
  • the display decision result of the second display decision is that the target screen display area that the ranging result is not greater than the preset distance threshold is determined as the first screen display area.
  • the second display decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a second display decision, and may also select, by the user, the type of the sensor included in the mobile terminal.
  • the second shows the decision.
  • the third possible implementation manner of the first aspect provides a display method decision under a low-precision condition, and the mobile terminal performs low-accuracy conditions according to the acquired touch pressure detection, ranging detection, and imaging detection.
  • the precise decision determines the screen display area.
  • the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area After that, it also includes:
  • the count is the sum of the number of times the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas;
  • the target screen display area whose ranging result is not greater than the preset distance threshold is determined as the first screen display area.
  • the mobile terminal By comparing the fuzzy decision and the accurate decision result under low precision conditions, the mobile terminal obtains the same proportion of the fuzzy decision and the accurate decision result.
  • the display of the mobile terminal adopts the result of the accurate decision. And study the judgment results.
  • the result of the fuzzy decision has a certain degree of precision. In the subsequent judgment, the result of the fuzzy decision can be directly used, and no need to make an accurate judgment. Simplified the process of judgment.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, Determining, in the at least two screen display areas, a target screen display area according to the user operation feature information, and determining whether there is user vital sign information in a vertical direction of the target screen display area;
  • the first screen display area is determined in the target screen display area, including:
  • the screen display area having the temperature matching the preset temperature interval in the vertical direction is determined as the target screen display area in the at least two screen display areas;
  • the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the mobile terminal before determining the first screen display area in the at least two screen display areas, according to the data reported by the target sensor, further includes:
  • a third display decision is determined based on the sensor type of the mobile terminal.
  • the display decision result of the third display decision is that the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the third display decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a third display decision, and may also select, by the user, the type of the sensor included in the mobile terminal. Three display decisions.
  • the sixth possible implementation manner of the first aspect provides a display method decision under medium precision conditions, and the mobile terminal performs fuzzy decision under medium precision condition according to the acquired temperature detection data and the imaging detection data. Determine the screen display area.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, Determining, in the at least two screen display areas, a target screen display area according to the user operation feature information, and determining whether there is user vital sign information in a vertical direction of the target screen display area;
  • the first screen display area is determined in the target screen display area, including:
  • the screen display area having the temperature matching the preset temperature interval in the vertical direction is determined as the target screen display area in the at least two screen display areas;
  • the target screen display area in which the result of the face recognition coincides with the preset face is determined as the first screen display area.
  • the mobile terminal before determining the first screen display area in the at least two screen display areas, according to the data reported by the target sensor, further includes:
  • a fourth display decision is determined based on the sensor type of the mobile terminal.
  • the display decision result of the fourth display decision is that the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the fourth display decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a fourth display decision, and may also select, by the user, the type of the sensor included in the mobile terminal. Four shows the decision.
  • the seventh possible implementation of the first aspect provides a display method decision under medium accuracy conditions, wherein the mobile terminal performs the process according to the acquired temperature detection data, the imaging detection data, and the face recognition detection data. Precise decisions under precision conditions to determine the screen display area.
  • the target screen display area that matches the result of the face recognition with the preset face is determined as the first After the screen display area, it also includes:
  • the two target screen display areas are respectively an imaging result and a preset body image a matching target screen display area and a target screen display area in which the result of the face recognition is consistent with the preset face;
  • the target screen display area that matches the imaging result with the preset body image is determined as the first screen display area.
  • the result of the fuzzy decision and the accurate decision under the medium precision condition keep the same ratio reaching the preset second proportional threshold, the result of the fuzzy decision also has A certain degree of precision, in the subsequent judgment, the result of the fuzzy decision can be directly used, without having to make an accurate judgment, simplifying the process of the judgment.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, Determining, in the at least two screen display areas, a target screen display area according to the user operation feature information, and determining whether there is user vital sign information in a vertical direction of the target screen display area;
  • the first screen display area is determined in the target screen display area, including:
  • the screen display area of the sound having the specific frequency range in the direction is determined as the target screen display area in the at least two screen display areas;
  • the target display area in which the result of the face recognition coincides with the preset face is determined as the first screen display area.
  • the mobile terminal before determining the first screen display area in the at least two screen display areas, according to the data reported by the target sensor, further includes:
  • a fifth display decision is determined based on the sensor type of the mobile terminal.
  • the display decision result of the fifth display decision is that the target screen display area that matches the imaging result with the preset body image is determined as the first screen display area.
  • the fifth display decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a fifth display decision, and may also select, by the user, the type of the sensor included in the mobile terminal. Five shows the decision.
  • the ninth possible implementation manner of the first aspect provides a display method decision under high-precision conditions, and the mobile terminal performs high-precision conditions according to the acquired sound source identification data and face recognition data.
  • the fuzzy decision determines the screen display area.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, Determining, in the at least two screen display areas, a target screen display area according to the user operation feature information, and determining whether there is user vital sign information in a vertical direction of the target screen display area;
  • the first screen display area is determined in the target screen display area, including:
  • the screen display area of the sound having the specific frequency range in the direction is determined as the target screen display area in the at least two screen display areas;
  • the target screen display area in which the result of the voice recognition coincides with the preset sound is determined as the first screen display area.
  • the mobile terminal before determining the first screen display area in the at least two screen display areas, according to the data reported by the target sensor, further includes:
  • a sixth display decision is determined based on the sensor type of the mobile terminal.
  • the display decision result of the sixth display decision is that the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the sixth display decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a sixth display decision, and may also select, by the user, the type of the sensor included in the mobile terminal. Six shows the decision.
  • the tenth possible implementation manner of the first aspect provides a display method decision under high-precision conditions, and the mobile terminal performs high according to the acquired sound source identification data, face recognition data, and voice recognition data. Precise decisions under precision conditions to determine the screen display area.
  • the target screen display area that determines whether the result of the voice recognition is consistent with the preset sound is determined as the first After the screen display area, it also includes:
  • Counting the number of times the two target screen display areas are the same screen display area and the two target screen display areas are different screen displays The number of times of the area, wherein the two target screen display areas are respectively a target display area in which the result of the face recognition is consistent with the preset face and whether the result of the voice recognition is consistent with the preset sound;
  • the target display area that matches the result of the face recognition with the preset face is determined as the first screen display area.
  • the preset ratio is reached.
  • the result of the fuzzy decision has a certain degree of precision. In the subsequent judgment, the result of the fuzzy decision can be directly used, and it is not necessary to make an accurate judgment, which simplifies the decision process.
  • the detecting whether there is user operation feature information for the screen display area if the detection is yes, And determining, according to the user operation feature information, a target screen display area in the at least two screen display areas, and determining whether the user's physical sign information exists in a vertical direction of the target screen display area; Determining the first screen display area in the target screen display area, including:
  • the screen display area of the sound having the specific frequency range in the direction is determined as the target screen display area in the at least two screen display areas;
  • Iris recognition is performed in the vertical direction of the target screen display area in which the result of the face recognition coincides with the preset face, and it is determined whether the result of the iris recognition is consistent with the preset iris;
  • the target screen display area in which the result of the iris recognition coincides with the preset iris is determined as the first screen display area.
  • the mobile terminal before determining the first screen display area in the at least two screen display areas, according to the data reported by the target sensor, further includes:
  • a seventh display decision is determined based on the sensor type of the mobile terminal.
  • the display decision result of the seventh display decision is that the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the seventh display decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a seventh display decision, and may also select, by the user, the type of the sensor included in the mobile terminal. Seven shows the decision.
  • the twelfth possible implementation manner of the first aspect provides a display method decision under another high-precision condition, the mobile terminal, according to the acquired sound source identification data, face recognition data, and iris recognition data, Accurate decisions under high precision conditions are made to determine the screen display area.
  • the target screen display area that matches the result of the iris recognition with the preset iris is determined as the first After the screen display area, it also includes:
  • the two target screen display areas are the same screen display area and the two target screen display areas are different screen displays The number of times of the area, wherein the two target screen display areas are the target display area in which the result of the face recognition is consistent with the preset face and the target screen display area in which the result of the iris recognition is consistent with the preset iris;
  • the target display area that matches the result of the face recognition with the preset face is determined as the first screen display area.
  • the fuzzy decision and the precision under the high precision condition When the result of the judgment keeps the same proportion reaching the preset fourth ratio threshold, the result of the fuzzy judgment also has a certain degree of precision. In the subsequent judgment, the result of the fuzzy decision can be directly used, and no need to make an accurate judgment. Simplified the process of judgment.
  • a second aspect of the embodiment of the present invention discloses a display method, which is applied to a mobile terminal having at least two screen display areas, and the method includes:
  • the mobile terminal acquires data reported by the target sensor;
  • the target sensor includes at least one sensor selected according to a preset sensor selection mechanism; and the sensor selection mechanism is a sensor type supported by the mobile terminal Determined selection mechanism;
  • the mobile terminal determines a first screen closing area in the screen display area in an activated state according to the data reported by the target sensor, and closes the first screen closing area.
  • a second aspect of the embodiment of the present invention provides a method for automatically turning off a screen in which a mobile terminal has been activated, and the mobile terminal determines, by using data uploaded by the sensor, whether the user has stopped using a part of the screen display area of the mobile terminal, thereby Specifying a shutdown decision turns off the corresponding screen display area, saving power usage.
  • the method further includes:
  • the intervention operation is for determining a second screen closure area
  • the mobile terminal After the mobile terminal makes a screen close decision to close the corresponding screen display area, it is possible that the screen display area that is closed due to the inaccurate decision is not the screen display area that the user wants to close, the user can manually perform the manual intervention operation. Close, thus closing the correct screen display area.
  • the mobile terminal is in the active state according to the data reported by the target sensor
  • the first screen off area is determined in the screen display area, including:
  • the mobile terminal detects, according to the data reported by the at least one target sensor, whether there is user operation information in each of the screen display areas in the activated screen display area;
  • a first screen close area is determined based on the user operation information.
  • the mobile terminal judges the state of use of the user according to the data reported by the target sensor, selects the screen display area in the activated state that the user does not use, and closes it.
  • the mobile terminal detects the screen display area that is in an active state according to the data reported by the target sensor Whether there is user operation information in each of the screen display areas, and determining the first screen closing area according to the user operation information, including:
  • the screen display area in which the ranging result is greater than the preset distance threshold is determined as the first screen closing area.
  • the mobile terminal before determining the first screen closure area in the screen display area in the activated state, according to the data reported by the target sensor, further includes:
  • the result of the closing decision of the first closing decision is that the screen display area in the active state that the ranging result is greater than the preset distance threshold is determined as the first screen closing area.
  • a first shutdown decision including the mobile terminal automatically selecting a sensor according to a sensor type and determining a first shutdown decision, or manually selecting, by the user, the type of the sensor included in the mobile terminal. A decision to close.
  • the third possible implementation manner of the second aspect provides a fuzzy close decision, and the mobile terminal obtains a fuzzy close decision according to the action detection data, the temperature detection data, and the ranging detection data of the acquired activated screen. Determine the first screen off area.
  • the target screen close-up area in which the user's vital information is not present in the vertical direction is determined as the first screen-off area, including:
  • the screen display area in the active state in which the face recognition result coincides with the preset face is determined as the first screen close area.
  • the mobile terminal before determining the first screen closure area in the screen display area in the activated state, according to the data reported by the target sensor, further includes:
  • the result of the closing decision of the second closing decision is that the screen display area in the active state that the ranging result is greater than the preset distance threshold is determined as the first screen closing area.
  • the second closing decision is determined according to the sensor type of the mobile terminal, and the mobile terminal automatically selects a sensor according to the sensor type and determines a second shutdown decision, and the user may select the sensor type determined by the mobile terminal. Second, close the decision.
  • the fourth possible implementation manner of the second aspect provides an accurate close decision, and the mobile terminal detects motion data, temperature detection data, ranging detection data, and face recognition data according to the action of acquiring the activated screen. , get the exact close decision to determine the first screen off area.
  • the screen display area that is in an active state that matches the face recognition result with the preset face is determined as After the first screen is closed, it also includes:
  • Counting the number of times the two screen display areas in the active state are the same screen display area and the number of times the two screen display areas in the active state are different screen display areas, wherein the two screen display areas in the active state The screen display area in the active state in which the ranging result is greater than the preset distance threshold and the screen display area in the active state in which the face recognition result is inconsistent with the preset face;
  • the screen display area in the active state in which the ranging result is greater than the preset distance threshold is determined as the first screen closed area.
  • the mobile terminal By comparing the results of the fuzzy close decision and the precise close decision, the mobile terminal obtains the same ratio of the two decision results.
  • the result of the fuzzy close decision and the precise close decision are different, the closing of the mobile terminal adopts the result of the accurate decision, and Conduct the study of the judgment results.
  • the result of the fuzzy close decision also has a certain degree of precision. In the subsequent judgment, the result of the fuzzy close decision can be directly used, without having to perform precise Closing the judgment simplifies the process of the judgment.
  • a third aspect of the embodiments of the present invention discloses a display device, including:
  • Opening a module configured to receive, by the mobile terminal, an open command, where the open command is used to indicate that the target interface is opened;
  • a display data acquisition module configured to acquire, by the mobile terminal, data reported by the target sensor;
  • the target sensor includes at least one sensor selected according to a preset sensor selection mechanism; and the sensor selection mechanism is a sensor supported by the mobile terminal Type selection mechanism;
  • a display screen determining module configured to determine, by the mobile terminal, a first screen display area in the at least two screen display areas according to the data reported by the target sensor;
  • a display module configured to present the target interface in the first screen display area.
  • the method further includes:
  • Displaying an intervention operation detecting module configured to: after the interface of the first application is presented in the first screen display area, the detection is Whether there is an intervention operation; the intervention operation is for determining a second screen display area;
  • An intervention display module configured to determine a second screen display area in the at least two screen display areas according to the intervention operation, withdraw a target interface presented in the first screen display area, and display on the second screen The target interface is presented in the zone.
  • the display screen determining module includes:
  • a feature information detecting unit configured to: detect, by the mobile terminal, whether there is user operation feature information for the screen display area according to data reported by the target sensor;
  • a target screen determining unit configured to determine a target screen display area in the at least two screen display areas according to the user operation feature information if the feature information detecting unit detects YES;
  • a vital sign information detecting unit configured to determine whether there is user physical sign information in a vertical direction of the target screen display area
  • the first screen display area determining unit determines the target screen display area in which the user's vital sign information exists in the vertical direction as the first screen display area.
  • the feature information detecting unit includes:
  • a pressure detecting subunit configured to detect whether there is a touch pressure in each of the at least two screen display areas
  • the target screen determining unit determines the screen display area where the touch pressure does not exist as the target screen display area
  • the vital sign information detecting unit includes:
  • a ranging subunit configured to perform ranging in a vertical direction of the target screen display area, and determine whether the ranging result is greater than a preset distance threshold
  • the first screen display area determining unit determines the target screen display area whose ranging result is not greater than the preset distance threshold as the first screen display area.
  • the operation information detecting unit includes:
  • a pressure detecting subunit configured to detect whether there is a touch pressure in each of the at least two screen display areas
  • the target screen determining unit determines the screen display area where the touch pressure does not exist as the target screen display area
  • the vital sign information detecting unit includes:
  • a ranging subunit configured to perform ranging in a vertical direction of the target screen display area, and determine whether the ranging result is greater than a preset distance threshold
  • the vital sign information detecting unit further includes:
  • An imaging subunit configured to perform imaging in a vertical direction of a target screen display area where the ranging result is not greater than a preset distance threshold if the determination result of the ranging subunit is not greater than a preset distance threshold, and detect whether the imaging result is Matches the preset body image;
  • the first screen display area determining unit determines the target screen display area whose ranging result is not greater than the preset distance threshold as the first screen display area.
  • the method further includes:
  • a first statistic module configured to count the number of times that the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas, wherein the two target screen display areas are respectively a target screen display area in which the ranging result is not greater than a preset distance threshold and a target screen display area in which the imaging result matches the preset human body image;
  • a first calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • the first determining module is configured to determine, when the ratio calculated by the first calculating module is greater than the first preset ratio threshold, the target screen display area whose ranging result is not greater than the preset distance threshold as the first screen display area.
  • the operation information detecting unit includes:
  • a temperature subunit configured to detect whether there is a temperature matching the preset temperature interval in a vertical direction of each of the at least two screen display areas
  • the target screen determining unit determines a screen display area having a temperature matching the preset temperature interval in the vertical direction as the target screen display area;
  • the vital sign information detecting unit includes:
  • An imaging subunit configured to perform imaging in a vertical direction of the target screen display area, to detect whether the imaging result matches a preset human body image
  • the first screen display area determining unit determines the target screen display area in which the imaging result matches the preset body image as the first screen display area.
  • the operation information detecting unit includes:
  • a temperature subunit configured to detect whether there is a temperature matching the preset temperature interval in a vertical direction of each of the at least two screen display areas
  • the target screen determining unit determines a screen display area having a temperature matching the preset temperature interval in the vertical direction as the target screen display area;
  • the vital sign information detecting unit includes:
  • An imaging subunit configured to perform imaging in a vertical direction of the target screen display area, to detect whether the imaging result matches a preset human body image
  • the vital sign information detecting unit further includes:
  • a face recognition subunit configured to perform face recognition in a vertical direction of a target screen display area where the imaging result matches the preset body image, and if the detection result of the imaging subunit is YES, the result of detecting the face recognition Whether it is consistent with the preset face;
  • the first screen display area determining unit determines the target screen display area in which the result of the face recognition is consistent with the preset face as the first screen display area.
  • the method further includes:
  • a second statistic module configured to count the number of times that the two target screen display areas are the same screen display area and the two target screen displays
  • the area is the number of times of different screen display areas, wherein the two target screen display areas are respectively a target screen display area in which the imaging result matches the preset body image and the target screen display result of the face recognition is consistent with the preset face.
  • a second calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • a second determining module configured to determine, as the first screen display area, the target screen display area that matches the imaging result with the preset body image when the ratio is greater than the second preset ratio threshold.
  • the operation information detecting unit includes:
  • a sound subunit configured to detect whether a sound of a specific frequency range exists in a direction of each of the screen display areas of the at least two screen display areas;
  • the target screen determination unit determines a screen display area in which a sound of a specific frequency range exists as a target screen display area
  • the vital sign information detecting unit includes:
  • a face recognition subunit configured to perform face recognition in a vertical direction of the target screen display area after detecting the screen display area where the sound of the specific frequency range is determined as the target screen display area, and detecting the face recognition Whether the result is consistent with the preset face;
  • the first screen display area determining unit determines the target screen display area in which the result of the face recognition is consistent with the preset face as the first screen display area.
  • the operation information detecting unit includes:
  • a sound subunit configured to detect whether a sound of a specific frequency range exists in a direction of each of the screen display areas of the at least two screen display areas;
  • the target screen determination unit determines a screen display area in which a sound of a specific frequency range exists as a target screen display area
  • the vital sign information detecting unit includes:
  • a face recognition subunit configured to perform face recognition in a vertical direction of the target screen display area after detecting the screen display area where the sound of the specific frequency range is determined as the target screen display area, and detecting the face recognition Whether the result is consistent with the preset face;
  • the vital sign information detecting unit further includes:
  • a voice recognition subunit configured to perform voice recognition in a vertical direction of the target screen display area in which the result of the face recognition is consistent with the preset face, and determine whether the result of the voice recognition is consistent with the preset sound
  • the first screen display area determining unit determines the target screen display area in which the result of the sound recognition coincides with the preset sound as the first screen display area.
  • the method further includes:
  • the third statistic module is configured to count the number of times that the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas, wherein the two target screen display areas are respectively The result of face recognition is consistent with the preset face Whether the target display area and the result of the voice recognition are the target screen display area that is consistent with the preset sound;
  • a third calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • the third determining module is configured to determine, as the first screen display area, the target display area that matches the result of the face recognition with the preset face when the ratio is greater than the third preset ratio threshold.
  • the operation information detecting unit includes:
  • a sound subunit configured to detect whether a sound of a specific frequency range exists in a direction of each of the screen display areas of the at least two screen display areas;
  • the target screen determination unit determines a screen display area in which a sound of a specific frequency range exists as a target screen display area
  • the vital sign information detecting unit includes:
  • a face recognition subunit configured to perform face recognition in a vertical direction of the target screen display area after detecting the screen display area where the sound of the specific frequency range is determined as the target screen display area, and detecting the face recognition Whether the result is consistent with the preset face;
  • the vital sign information detecting unit further includes:
  • the iris recognition subunit is configured to perform iris recognition in a vertical direction of the target screen display area in which the result of the face recognition is consistent with the preset face, and determine whether the result of the iris recognition is consistent with the preset iris;
  • the first screen display area determining unit determines the target screen display area in which the result of the iris recognition is consistent with the preset iris as the first screen display area.
  • the method further includes:
  • the fourth statistic module is configured to count the number of times that the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas, wherein the two target screen display areas are respectively
  • the result of the face recognition is the same as the preset face and the target display area and the iris recognition result are the same as the preset iris display target display area;
  • a fourth calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • a fourth determining module configured to determine, as the first screen display area, the target display area that matches the result of the face recognition with the preset face when the ratio is greater than the fourth preset ratio threshold.
  • a fourth aspect of the embodiments of the present invention provides a display device for a multi-screen mobile terminal, including:
  • a screen detecting module configured to detect whether the mobile terminal has a screen display area in an activated state
  • the mobile terminal acquires data reported by the target sensor;
  • the target sensor includes at least one sensor selected according to a preset sensor selection mechanism;
  • the mechanism is a selection mechanism determined by the type of sensor supported by the mobile terminal;
  • a closing module configured, by the mobile terminal, according to the data reported by the target sensor, in the active screen display area
  • the first screen closing area is determined, and the first screen closing area is closed.
  • the method further includes:
  • Close the intervention detection module configured to detect whether there is a user intervention operation after the closing the first screen closure area; the intervention operation is used to determine a second screen closure area;
  • the closing module includes:
  • a closing operation detecting unit configured to detect, by the mobile terminal, whether there is user operation information in each of the screen display areas in the screen display area in which the target is in an active state according to the data reported by the target sensor;
  • a closing unit configured to determine a first screen closing area according to the user operation information.
  • the shutoff operation detecting unit includes:
  • An action subunit configured to detect whether there is an action change in a direction of each of the screen display areas in the activated screen display area
  • a second temperature subunit for detecting whether there is a temperature matching the preset temperature interval in a vertical direction of the screen display area in an active state in which the action changes;
  • a second ranging subunit configured to perform ranging in a vertical direction of an active screen display area in a vertical direction without a temperature matching the preset temperature interval, and detecting whether the ranging result is greater than a preset distance threshold ;
  • the closing unit determines the first screen closing area of the active screen display area that the ranging result is greater than the preset distance threshold.
  • the shutoff operation detecting unit includes:
  • An action subunit configured to detect whether there is an action change in a direction of each of the screen display areas in the activated screen display area
  • a second temperature subunit for detecting whether there is a temperature matching the preset temperature interval in a vertical direction of the screen display area in an active state in which the action changes;
  • a second ranging subunit configured to perform ranging in a vertical direction of an active screen display area in a vertical direction without a temperature matching the preset temperature interval, and detecting whether the ranging result is greater than a preset distance threshold ;
  • a second face recognition sub-unit configured to perform face recognition in a vertical direction of the screen display area in which the ranging result is greater than a preset distance threshold, and detect whether the face recognition result is consistent with the preset face
  • the closing unit determines the first screen closing area by the screen display area in the active state that is consistent with the preset face.
  • the method further includes:
  • a fifth statistic module configured to count the number of times that the two screen display areas in the active state are the same screen display area and the number of times that the two screen display areas in the activated state are different screen display areas, wherein the two The screen display areas in the activated state are the screen display areas in the active state in which the ranging result is greater than the preset distance threshold, and the screen display area in the active state in which the face recognition result is inconsistent with the preset face;
  • a fifth calculating module calculating a ratio of the number of times that the two screen display areas in the active state are the same screen display area to the total number of statistical times, wherein the total number of statistics is the two screens that are in an active state The sum of the number of times the display area is the same screen display area and the number of times the two screen display areas in the active state are different screen display areas;
  • the fifth determining module is configured to determine, when the ratio is greater than the fifth preset ratio threshold, that the screen display area in which the ranging result is greater than the preset distance threshold is determined as the first screen closing area.
  • a fifth aspect of the embodiment of the present invention discloses a terminal, including:
  • a memory for storing display instructions
  • a processor for invoking a display instruction in the memory the display instruction being applied to a mobile terminal having at least two screen display areas, the processor invoking the display instruction and executing the display method provided by the first aspect above .
  • a sixth aspect of the embodiment of the present invention discloses a terminal, including:
  • a memory for storing a shutdown instruction
  • a processor for invoking a shutdown instruction in the memory the shutdown instruction being applied to a mobile terminal having at least two screen display areas, the processor invoking the shutdown instruction and performing the display method provided by the second aspect above .
  • a seventh aspect of an embodiment of the present invention discloses a computer readable storage medium for storing one or more computer programs, the one or more computer programs comprising instructions, when the computer program is run on a computer, The instructions are for performing the display method provided by the first aspect or the second aspect described above.
  • An eighth aspect of the embodiments of the present invention discloses a computer program, the computer program comprising instructions for performing the display method provided by the first aspect or the second aspect when the computer program is executed on a computer .
  • a mobile terminal having at least two screen display areas receives an open command input by a user, the open command is used to indicate that the target interface is opened, and the mobile terminal receives data reported by the target sensor.
  • the mobile terminal determines a first screen display area in the at least two screen display areas according to the data reported by the target sensor, and presents the target interface in the first screen display area. It can accurately identify the user's location and usage status, and perform subsequent operations such as displaying the interface, lighting the screen, and opening multimedia, which is convenient for the user.
  • the multi-screen mobile terminal decides to select a corresponding sensor according to the display type of the sensor according to the type of the sensor, and the user may select a specific sensor, and the external data acquired by the sensor will be The data is reported to the mobile terminal to determine the location and usage status of the user, and the corresponding screen display area is selected for display, which can be fully combined with the characteristics of the sensor type of the mobile terminal to perform judgment, thereby intelligently selecting an appropriate screen-to-target interface. Display.
  • the display method can be divided into multiple precisions, and each precision display method decision includes fuzzy decision and accurate decision; when the ratio of the fuzzy decision and the accurate decision is consistent with the preset ratio threshold, then If the fuzzy decision is considered to be sufficiently accurate, then As a result of the fuzzy decision, there is no need to make an accurate decision.
  • the multi-screen mobile terminal determines the first screen display area in the at least two screen display areas according to the data reported by the sensor, and presents the target interface in the first screen display area, if The screen display area presented is not the screen display area that the user desires to present, and the user can also perform an intervention operation to correct and manually select the screen display area that is to be displayed.
  • 1 is a schematic diagram of mobile terminals of several different types of screen display areas
  • FIG. 2 is a schematic structural diagram of a terminal according to an embodiment of the present invention.
  • FIG. 3 is a positional relationship diagram of each sensor and a screen display area according to an embodiment of the present invention.
  • FIG. 4 is a schematic flowchart of a display method according to an embodiment of the present invention.
  • FIG. 5 is an interface diagram of a sensor type and display decision according to an embodiment of the present invention.
  • FIG. 6 is a schematic flowchart diagram of a display method according to another embodiment of the present invention.
  • FIG. 7a is a schematic flowchart of a decision of a low-precision display method according to an embodiment of the present invention.
  • FIG. 7b is a schematic flowchart of a decision of a low-precision display method according to another embodiment of the present invention.
  • FIG. 8a is a schematic flowchart of a method for determining a medium precision display method according to an embodiment of the present invention.
  • FIG. 8b is a schematic flowchart of a method for determining a medium precision display method according to another embodiment of the present invention.
  • FIG. 9a is a schematic flowchart of a method for determining a high-precision display method according to an embodiment of the present invention.
  • FIG. 9b is a schematic flowchart of a method for determining a high-precision display method according to another embodiment of the present invention.
  • FIG. 10a is a schematic flowchart of a high-precision display method decision according to another embodiment of the present invention.
  • FIG. 10b is a schematic flowchart of a method for determining a high-precision display method according to another embodiment of the present invention.
  • FIG. 11 is a schematic flowchart diagram of a display method according to another embodiment of the present invention.
  • FIG. 12 is a schematic diagram of a decision process of a display method according to an embodiment of the present invention.
  • FIG. 12b is a schematic diagram of a decision process of a display method according to another embodiment of the present invention.
  • FIG. 13 is a schematic structural diagram of a multi-screen mobile terminal according to an embodiment of the present disclosure.
  • FIG. 14 is a schematic structural diagram of a display device according to an embodiment of the present disclosure.
  • FIG. 15 is a schematic structural diagram of a display screen determining module according to an embodiment of the present disclosure.
  • FIG. 16 is a schematic structural diagram of a display device according to another embodiment of the present invention.
  • FIG. 17 is a schematic structural diagram of a shutdown module according to an embodiment of the present invention.
  • FIG. 1 it is a schematic diagram of mobile terminals of several different types of screen display areas
  • a is a double-sided screen device, that is, an electronic device having two screens, including but not limited to two screens that are two screens in front and back, and that maintain a certain angle;
  • b is a flexible screen device, that is, an electronic device whose screen can be folded into a certain angle;
  • c is a multi-screen display device, that is, an electronic device having a plurality of display screens, and each display screen is in a different plane;
  • d is a multi-screen display device, that is, an electronic device having a plurality of display screens, and each display screen is in the same plane.
  • the double-sided screen device and the flexible screen device shown in a and b in the prior art, the user's use state can be roughly judged by gravity sensing, and the user is not accurately determined that the user is using.
  • the screen is thus displayed, and in the case of a multi-screen display device in c and d, if it is only by means of gravity sensing, it is difficult to determine the position of the user and the state of use, so that the multi-screen display device cannot accurately perceive that the user is The screen used is thus displayed.
  • the acquired data is reported to the mobile terminal by using a sensor provided by the mobile terminal, and the mobile terminal determines the display screen that the user is using according to the reported data, and performs corresponding display.
  • the mobile terminal determines the orientation and the use state of the user according to the data uploaded by the equipped sensor, thereby selecting an appropriate screen display area to display the target interface.
  • the mobile terminal may be equipped with multiple types of sensors, including but not limited to sensors that sense the posture of the device, sensors for user actions and orientations, and sensors for user biometric identification.
  • the multi-screen mobile terminal selects a corresponding sensor according to a specific hardware sensor, and selects a corresponding sensor to obtain data according to the display method, and the user may select the specified sensor and select the acquired data.
  • the report is reported to the mobile terminal, and the mobile terminal makes a decision according to the reported data, thereby selecting a corresponding screen display area for display.
  • the embodiment of the present invention can formulate three display method decisions of low precision, medium precision and high precision according to possible sensor configurations of different multi-screen mobile terminals, and each of the precision display method decisions includes a fuzzy decision and Precise judgment meets various user usage scenarios.
  • the existing form of the double-sided screen device, the flexible screen device, and the multi-screen display device is a specific embodiment, and in the actual application, there may be more morphological products; There are other combinations of sensor configurations for display devices, and are not limited to the specific sensors listed in the examples.
  • the terminal 100 may include: a baseband chip 101, a memory 105 (one or more computer readable storage media), a radio frequency module 111, and a periphery. System 106. These components can communicate over one or more communication buses 112.
  • the peripheral system 106 is mainly used to implement the interaction function between the terminal 100 and the user/external environment, and mainly includes the input and output devices of the terminal 100.
  • the peripheral system 106 can include a screen display area 107, a camera controller 108, an audio controller 109, and a sensor 110.
  • the screen display area 107 may include multiple screen display areas, and the plurality of screen display areas may not be on the same plane or on the same plane; the sensor 110 may include multiple types of sensors, such as: gyroscope/ The gravity sensor 1101, the camera sensor 1102, the infrared array sensor 1103, the ultrasonic sensor 1104, the laser sensor 1105, the light/proximity sensor 1106, the pressure sensor 1107, the iris recognition sensor 1108, the acoustic sensor 1109, and the like.
  • sensors such as: gyroscope/ The gravity sensor 1101, the camera sensor 1102, the infrared array sensor 1103, the ultrasonic sensor 1104, the laser sensor 1105, the light/proximity sensor 1106, the pressure sensor 1107, the iris recognition sensor 1108, the acoustic sensor 1109, and the like.
  • the gyroscope/gravity sensor 1101 can be used to measure the screen orientation, and the gyroscope/gravity sensor 1101 can be located inside the terminal 100.
  • the gyroscope detects the direction of the screen by measuring the operating angular velocity when the terminal 100 is deflected and tilted, and the gravity sensor passes.
  • the gravity value determines which side of the screen is vertically up to detect the direction of the screen.
  • the camera sensor 1102 can be used for camera imaging, face recognition, etc., and can be located in the upper left corner of the screen display area 107, by comparing the image captured by the camera with a target image (such as a user face image, etc.) stored in the memory 105. In turn, face recognition is realized.
  • a target image such as a user face image, etc.
  • the infrared array sensor 1103 can be used for infrared imaging, temperature detection, motion recognition, etc., and can be located on both sides of the screen display area 107.
  • the infrared array sensor 1103 can detect the infrared specific band signal of the target object heat radiation, and convert the signal. An image or graphic that can be visually distinguished by the human eye and further calculate the temperature value.
  • the ultrasonic sensor 1104 can be used for ultrasonic imaging and ranging, etc., and can be located below the screen display area 107.
  • the ultrasonic sensor 1104 calculates the distance by the time when the ultrasonic wave propagates in the air and the propagation speed of the sound wave, and the ultrasonic beam 1104 transmits the pulse beam. After focusing, the ray is measured, and the direction of the ultrasonic beam and the depth of the focus are controlled to three-dimensionally scan the object, thereby obtaining a stereoscopic image of the object.
  • the laser sensor 1105, the light/proximity sensor 1106 can be used for ranging, and the laser sensor 1105 can be located at the lower right corner of the screen display area 107, and the target distance is measured by recording and processing the time elapsed from the time the light pulse is emitted until the return is received;
  • the proximity sensor 1106 can be located in the upper right corner of the screen display area 107.
  • the pressure sensor 1107 can be used to measure the touch pressure and the like, and can be located on the surface of the screen display area 107.
  • the strain gauges connected to the object to be tested are also deformed together, so that the strain gauge is blocked.
  • the value changes so that the voltage applied to the resistor changes to obtain a pressure value.
  • the iris recognition sensor 1108 can be used for iris recognition, and can be located above the screen display area 107.
  • the iris recognition sensor 1108 can collect and preserve the biometric features of the user's eyeball, and identify the collected iris to be detected during use.
  • the illumination can be performed by infrared LED light while the acquisition is achieved by the front camera of the device.
  • the sound sensor 1109 can be used for voice recognition and sound source recognition, etc., and can be located at the lower left corner of the screen display area 107.
  • a sound-sensitive capacitive electret microphone By embedding a sound-sensitive capacitive electret microphone in the sound sensor 1109, the sound wave causes the electret film in the microphone. Vibration, which causes a change in capacitance, and produces a small voltage that changes accordingly.
  • the specific positional relationship between each sensor and the screen display area 107 is as shown in FIG. 3.
  • the A-side screen and the B-side screen are two screen display areas, and the A and B are facing away from each other. Displayed outside.
  • the gyro/gravity sensor 1101 is located inside the terminal 100, and the pressure sensor 1107 is located outside the surface of the screen display area 107.
  • Other sensors may be located at the same position of the A-side screen and the B-side screen, respectively.
  • peripheral system 106 may also include other peripherals.
  • the baseband chip 101 can be integrated to include one or more processors 102, a clock module 103, and a power management module 104.
  • the clock module 103 integrated in the baseband chip 101 is primarily used to generate the clocks required for data transfer and timing control for the processor 102.
  • the power management module 104 integrated in the baseband chip 101 is primarily used to provide a stable, high precision voltage to the processor 102, the RF module 111, and the peripheral system 106.
  • the radio frequency module 111 is configured to receive and transmit radio frequency signals, and mainly integrates the receiver and the transmitter of the terminal 100.
  • the radio frequency module 111 communicates with the communication network and other communication devices via radio frequency signals.
  • the radio frequency module 111 may include, but is not limited to: an antenna system, an RF transceiver, one or more amplifiers, a tuner, one or more oscillators, a digital signal processor, a CODEC chip, a SIM card, and Storage media, etc.
  • the radio frequency module 111 can be implemented on a separate chip.
  • Memory 105 is coupled to processor 102 for storing various software programs or sets of instructions.
  • memory 105 may include high speed random access memory and may include non-volatile memory, such as one or more magnetic disk storage devices, flash memory devices, or other non-volatile solid state storage devices.
  • the memory 105 can store an operating system (hereinafter referred to as a system) such as an embedded operating system such as ANDROID, IOS, WINDOWS, or LINUX.
  • the memory 105 can also store a network communication program that can be used to communicate with one or more additional devices, one or more terminal devices, one or more network devices.
  • the memory 105 can also store a user interface program, which can realistically display the content of the application through a graphical operation interface, and pass the dish. Input controls such as singles, dialogs, and buttons receive user control operations on the application.
  • FIG. 4 is a schematic flowchart diagram of a display method according to an embodiment of the present disclosure, where the method includes but is not limited to the following steps:
  • S101 The mobile terminal receives an open command input by a user, where the open command is used to indicate that the target interface is opened.
  • the opening instruction is used to indicate that the target interface is opened, including an operation of activating a screen, entering a software application, playing multimedia, and the like.
  • S102 The mobile terminal receives data reported by the target sensor.
  • the mobile terminal may be equipped with hardware such as a sensor, and the types and types of sensors provided by different mobile terminals are different, and the type and function of the reported data are different.
  • the data collected by the gyroscope is the angular velocity when the terminal is deflected or tilted
  • the data collected by the pressure sensor is the pressure generated when the user touches the screen.
  • the target sensor in the embodiment of the present invention includes at least one sensor selected according to a preset sensor selection mechanism; the sensor selection mechanism is a selection mechanism determined by a sensor type supported by the mobile terminal.
  • the mobile terminal may select a target sensor according to a preset display decision according to a sensor type supported by the mobile terminal, or may select a target sensor manually according to a required display decision after the user knows the sensor type supported by the mobile terminal.
  • the setting may be performed in advance, and different target sensors are selected to correspond to different display decisions.
  • the sensor type A and the sensor type B correspond to the first display decision
  • the sensor type A and the sensor type C correspond to the second display decision, the sensor type A, the sensor type B, and the sensor type C.
  • the sensor type A and the sensor type B may be selected as the target sensor according to the first display decision, or the sensor type may be selected according to the second display decision.
  • a and sensor type B are used as target sensors
  • sensor type A, sensor type B, and sensor type C can also be selected as target sensors according to the third display decision.
  • Table 1 The specific correspondence between different sensor type combinations and display decisions is shown in Table 1.
  • the mobile terminal corresponds to different display decisions according to different sensor types, that is, different types of reporting data are obtained by calling different sensors, thereby correspondingly making different display judgments or display analysis.
  • sensor type A may be a device attitude sensor, such as a gyroscope/gravity sensor
  • sensor type B may be a user action/orientation sensor, such as a pressure sensor
  • sensor type C may be a user biometric sensor, such as an iris recognition sensor . Therefore, the device attitude sensor and the user action/orientation sensor can be corresponding to the first display decision; the device attitude sensor and the user biometric sensor are corresponding to the second display decision; the device attitude sensor, the user action/orientation sensor, and the user creature
  • the recognition sensor corresponds to a third display decision, and so on.
  • the mobile terminal may correspond to multiple display decisions according to different sensors, and the user can manually select the required display decision, as shown in FIG. 5:
  • the sensor types supported by the terminal include A, B, and C.
  • the color area represents on the left and on the right, and corresponds to three display decisions, namely the first display decision, the second display decision and the third display decision.
  • the user can click on the corresponding area of the display to further select the need.
  • Display decision (as shown on the right).
  • the system automatically selects, and the default first display decision is the final display decision.
  • the selection mechanism may further be that the user customizes the sensor type corresponding to different display decisions according to the sensor type supported by the terminal itself.
  • the system lists different sensor types supported by the mobile terminal under various display decisions.
  • the user can manually set different sensor types corresponding to the display decision, and slide the gray area to the left.
  • the representative turns on the corresponding sensor type, and sliding to the right means that the corresponding sensor type is turned off.
  • the user can manually set the sensor types corresponding to the first display decision to be A and C; the sensor types corresponding to the second display decision are B and C; the sensor types corresponding to the third display decision are A and B, and so on.
  • the mobile terminal can trigger or control the target sensor to acquire data, thereby receiving the data reported by the target sensor.
  • the mobile terminal determines a first screen display area in the at least two screen display areas according to the data reported by the target sensor, and presents the target interface in the first screen display area.
  • the mobile terminal selects a target sensor according to the supported sensor type, and according to the data reported by the sensor, determines the current user's usage state and location through its own decision mechanism, and determines a screen that matches the state of the user, and The open target interface is presented in a matching screen.
  • the target interface may not be presented on the screen desired by the user due to the inaccurate determination, so the user is required to actively intervene to correct the screen display area. Switching.
  • the mobile terminal After receiving the intervention operation of the user, the mobile terminal determines the second screen display area according to the intervention operation of the user, and displays the target interface on the screen designated by the user, and withdraws the presentation in the first screen display area. Target interface. If the mobile terminal does not receive the access operation of the user, the mobile terminal determines that the screen that should be presented is accurately determined according to the data reported by the sensor, and continues to remain in the first screen display area according to S106. Target interface.
  • the specific implementation manner of the intervention operation can be, for example, but not limited to, double-clicking the second screen display area, long pressing the second screen display area, or waking up the second screen display area, and the terminal provides whether to switch to the screen display area. Button, the user clicks to confirm the screen switch and so on.
  • FIG. 6 is a schematic flowchart diagram of a display method according to another embodiment of the present invention, illustrating in detail, in step S103 of FIG. 6, the mobile terminal reports data according to the target sensor on the at least two screens.
  • the case where the first screen display area is determined in the display area includes the following steps:
  • S201 The mobile terminal detects, according to the data reported by the target sensor, whether there is user operation feature information for the screen display area.
  • the mobile terminal has at least two screen display areas, and the sensor provided by the mobile terminal detects the user operation feature information on each screen display area, and reports the acquired information to the terminal.
  • the user operation feature information is operation information generated by the user when the multi-screen mobile terminal is used.
  • the user operation feature information may include a user's touch pressure on the screen, whether there is a human body temperature, a sound, or the like in the vertical direction of the screen.
  • the user's touch pressure on the screen can be measured by the pressure sensor, whether there is human body temperature in the vertical direction of the screen Measured by the infrared array sensor, whether there is sound in the vertical direction of the screen can be measured by the sound sensor.
  • the mobile terminal by processing and analyzing the user operation feature information reported by the sensor, the mobile terminal has an initial judgment on the orientation and usage state of the user, and may have a screen operated by the user as the target screen display area.
  • S203 Determine whether user sign information exists in a vertical direction of the target screen display area.
  • the detection is performed in the vertical direction of the aforementioned determined target screen display area, and it is further determined whether or not the user's vital information exists, thereby determining the screen that the user is currently using or will use.
  • the user vital sign information is that the mobile terminal detects the user's vital sign information when the user uses the multi-screen mobile terminal.
  • the user's physical information may include a distance between the user and the screen display area in the vertical direction of the screen display area detected by the mobile terminal, and whether the user's facial features match the preset face, Whether the user's iris matches the preset iris, etc.
  • the distance between the user in the vertical direction of the screen display area and the screen display area can be measured by a laser sensor, a light/proximity sensor or an ultrasonic sensor, and the facial features of the user can be measured by the camera sensor, the user's iris It can be measured by an iris recognition sensor.
  • S204 Determine a first screen display area in the target screen display area according to the user vital sign information.
  • a screen that is most likely to be used by the user is selected in the target screen display area, and is determined as a first screen display area, where the first screen is The target interface is presented in the display area.
  • the display method is divided into three levels of low, medium and high precision by different sensor configurations.
  • the display method will be described in more detail below with reference to the specific sensors invoked in the specific embodiments through Figures 7a to 10b.
  • the determination flow of the low-precision and medium-precision display method decision can be respectively described in FIG. 7a and FIG. 8a, and the descriptions of the high-precision display method can be described in FIG. 9a and FIG. 10a.
  • the process of judgment can be described in FIG. 7a and FIG. 8a.
  • FIG. 7a is a schematic flowchart of a method for making a low-precision display method according to the present disclosure, which includes the following steps:
  • S301 The mobile terminal receives data reported by the target sensor.
  • the mobile terminal specifies a corresponding decision according to the type and quantity of the sensor that is equipped by itself, and invokes a corresponding sensor to perform data collection and report.
  • S302 Detect whether there is touch pressure in each of the at least two screen display areas.
  • the user operation feature information in the embodiment of the present invention is described by taking an example of detecting whether there is user touch pressure on the screen display area of the mobile terminal.
  • the pressure sensor is called to detect whether there is touch pressure on each screen display area of the mobile terminal. If there is touch pressure, the determination may be the user's hand-held part, and the screen of the hand-held part is not the user's use screen, so the screen display area of the hand-held part is not included in the target screen display area.
  • step S311 is performed to end the subsequent detection in the screen display area.
  • S303 Determine a screen display area where there is no touch pressure in the at least two screen display areas as a target screen display area.
  • the screen display area may be the user's use screen, so it is determined as the target screen display area, and the number of the target screen display area may be at least one.
  • S304 Determine whether the ranging result in the vertical direction of the target screen display area is greater than a preset distance threshold.
  • the user's vital sign information in the embodiment of the present invention is an example for detecting the distance between the user and the user in the vertical direction of the screen display area of the mobile terminal.
  • the screen without the touch pressure may be the user's use screen, so further detection is performed on the screen where there is no touch pressure, and it is determined whether the distance measurement result in the vertical direction of the target screen display area is greater than the preset distance.
  • Threshold An ultrasonic sensor, a laser sensor, or a light/proximity sensor is called to measure whether a person is used within a preset distance threshold in the vertical direction of the target screen display area.
  • the preset distance threshold can be adjusted according to the user's habits, that is, the terminal can automatically collect the vertical distance from the screen when the user uses, and intelligently set the preset distance threshold according to the usual usage habits of the user to meet the different habits of different people. If the ranging result is not greater than the preset distance threshold, step S305 is performed; if the ranging result is greater than the preset distance threshold, step S312 is performed to end subsequent detection on the screen display area greater than the preset distance threshold.
  • S305 Determine a target screen display area whose ranging result is not greater than a preset distance threshold as the first screen display area.
  • the mobile terminal obtains the fuzzy decision result under the condition of low precision: the target screen display area not greater than the preset distance threshold is determined as the first screen display area, and the fuzzy decision is made The result is saved in the memory of the terminal.
  • S306 Detect whether the imaging result in the vertical direction of the target screen display area where the ranging result is not greater than the preset distance threshold matches the preset human body image.
  • the condition detection is performed again under the condition of the low-precision fuzzy decision.
  • the further detected information is to detect that the ranging result is not greater than the preset distance threshold. Whether the imaging result in the vertical direction of the target screen display area matches the preset human body image, that is, whether the obstacle in front of the screen of the mobile terminal is detected as a human body.
  • the infrared sensor or the ultrasonic sensor is called to perform imaging in a vertical direction of the target screen display area where the ranging result is not greater than the preset distance threshold.
  • the preset human body image may be any human body image acquired from the network, and the user's human body image obtained from the local network.
  • the user's human body image may be pre-acquired and stored offline in the terminal memory, if the imaging result and the preset are If the human body image is matched, step S307 is performed; if the imaging result does not match the preset human body image, step S313 is performed to end the subsequent detection of the screen display area that does not match the preset human body image.
  • S307 Determine a target screen display area that matches the imaging result with the preset human body image as the first screen display area.
  • the mobile terminal makes an accurate decision result under low precision conditions: the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area.
  • the two target screen display areas are respectively a target screen display area in which the ranging result is not greater than the preset distance threshold and a target screen display area in which the imaging result matches the preset human body image.
  • the result of the fuzzy decision is invoked from the memory, and the fuzzy decision result and the accurate decision result under the low precision condition are compared and learned, and the number of times the fuzzy decision result and the accurate decision result are consistent and inconsistent in the current stage are counted.
  • the current stage is from the first comparison and learning, to the comparison and learning.
  • the result of the fuzzy decision is consistent with the result of the accurate decision, that is, the target screen display area obtained after the step S305 is not greater than the preset distance threshold, and the target screen display area where the imaging result is matched with the preset body image in step S307.
  • the screen display area; the difference between the fuzzy decision result and the accurate decision result is the target screen display area in which the ranging result obtained after step S305 is not greater than the preset distance threshold and the target screen display in which the imaging result matches the preset human body image in step S307
  • the area is a different screen display area.
  • S309 Calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistics.
  • the total number of statistics is the sum of the number of times the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas.
  • the first preset ratio threshold may be automatically set by the mobile terminal, or may be set by the user.
  • the ratio is greater than the first preset ratio threshold, it indicates that the result of the fuzzy decision has a certain degree of accuracy.
  • the result of the fuzzy decision is directly displayed, so there is no need to make an accurate decision.
  • the target screen display area whose ranging result is not greater than the preset distance threshold is directly determined as the first screen display area, and the result of the ranging is not greater than the preset.
  • the step of whether the imaging result in the vertical direction of the target screen display area of the threshold value matches the preset human body image.
  • the memory invokes the results of the current phase comparison and learning to calculate the ratio of the result of the fuzzy decision and the result of the accurate decision to the ratio of the total number of comparisons and learning.
  • the first preset ratio threshold is 80%.
  • the number of times the statistical fuzzy decision result and the accurate judgment result are consistent is 8 times, and the number of inconsistencies is 2 times, that is, The total number of comparisons and learnings is 10, so in the next comparison and learning, if the fuzzy judgment result is consistent with the accurate judgment result, then the number of times the result of the fuzzy judgment is consistent with the result of the accurate judgment is compared.
  • the result of the fuzzy decision is directly Displaying, the condition detection is not performed again, that is, after the fuzzy decision result is obtained, the target screen display area whose ranging result is not greater than the preset distance threshold is directly determined as the first screen display area, and the execution is not performed.
  • the step of whether the imaging result in the vertical direction of the target screen display area where the ranging result is not greater than the preset distance threshold matches the preset human body image; if the next comparison and In the middle, the fuzzy decision result is inconsistent with the accurate judgment result, then the ratio of the result of the fuzzy decision and the result of the accurate judgment at this time accounts for 8/(10+1) 72.7% of the total number of comparisons and learning, not more than
  • the first preset ratio threshold is 80%, then the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning, and the results of the current phase comparison and learning are recalled from the memory to calculate the fuzzy decision.
  • the number of times the result of the comparison with the result of the accurate judgment is the ratio of the total number of comparisons and learning.
  • the result of the fuzzy decision is inaccurate, causing the user to manually set the displayed screen, and the second screen display area is determined according to the user's intervention operation, and the target interface is in the second Displayed in the screen display area while withdrawing the target interface presented in the first screen display area.
  • S301-S313 is re-executed, and the comparison and learning of the fuzzy decision and the accurate decision result are performed.
  • FIG. 8a is a schematic flowchart of a method for determining a medium precision display method according to an embodiment of the present invention, including the following steps:
  • S401 The mobile terminal receives data reported by the target sensor.
  • the mobile terminal specifies a corresponding decision according to the type and quantity of sensors equipped by itself, and invokes the phase.
  • the sensor should be collected and reported.
  • S402 Detect whether there is a temperature matching the preset temperature interval in a vertical direction of each of the at least two screen display areas.
  • the user operates the feature information in the embodiment of the present invention to detect whether there is a temperature matching the human body on the screen display area of the mobile terminal as an example.
  • the infrared array sensor is called to detect whether the vertical direction of each screen display area of the mobile terminal matches the preset temperature interval.
  • the preset temperature interval is a normal temperature interval of the human body. If the detection result is that there is a temperature matching the normal temperature interval of the human body, it is proved that the front end of the screen display area is used by the user.
  • step S411 is performed to end the subsequent detection in the screen display area.
  • S403 Determine, in the at least two screen display areas, a screen display area in which a temperature matching the preset temperature interval in the vertical direction exists as a target screen display area.
  • the number of target screen display areas may be at least one.
  • the user's vital sign information in the embodiment of the present invention is used to detect whether the imaging result in the vertical direction of the screen display area of the mobile terminal matches the preset human body image as an example.
  • the specific detection process is similar to step S306, in which an infrared sensor or an ultrasonic sensor is called to perform imaging in a vertical direction of the target screen display area where the imaging result does not match the preset human body image.
  • the preset human body image may be any human body image acquired from the network, and the human body image of the user may be acquired from the local network.
  • the human body image of the user may be pre-acquired and stored offline in the terminal memory. If the imaging result matches the preset human body image, step S405 is performed; if the imaging result does not match the preset human body image, step S412 is performed to end the subsequent detection of the screen display area that does not match the preset human body image.
  • S405 Determine a target screen display area that matches the imaging result with the preset body image as the first screen display area.
  • the mobile terminal obtains the fuzzy decision result under the medium precision condition: the target screen display area that matches the imaging result with the preset human body image is determined as the first screen display area, and the fuzzy decision is The result is saved in the memory of the terminal.
  • S406 Detect whether the face recognition result in the vertical direction of the target screen display area that matches the imaging result and the preset body image is consistent with the preset face.
  • the information of the further detection of the accurate determination under the medium precision condition is whether the face recognition result in the vertical direction of the screen display area of the mobile terminal is consistent with the preset face;
  • the preset face is a face image of a user stored in advance in the mobile terminal, and the determination is performed to identify whether the user who uses the mobile terminal is the user of the mobile terminal.
  • the camera sensor is called to perform face recognition in the imaging direction to identify whether it is a user.
  • the user's facial image may be pre-acquired and stored in the memory of the terminal, and the facial image collected by the camera sensor is compared with the facial image of the user. If the comparison result is consistent, the user who uses the mobile terminal is the user of the mobile terminal. I perform step S407; if the comparison result does not match, it means that the user who uses the mobile terminal is not the user of the mobile terminal, and then step S413 is executed to end the subsequent detection of the screen display area that is not consistent with the preset face. .
  • S407 Determine a target screen display area that matches the result of the face recognition with the preset face as the first screen display area.
  • the mobile terminal obtains an accurate judgment under the medium precision condition: determining the target screen display area as the first screen display area .
  • the two target screen display areas are respectively a target screen display area in which the imaging result matches the preset human body image and a target screen display area in which the result of the face recognition is consistent with the preset human face.
  • the result of the fuzzy decision is invoked from the memory, and the fuzzy decision result and the accurate decision result under the medium precision condition are compared and learned, and the number of times the fuzzy decision result and the accurate decision result are consistent and inconsistent in the current stage are counted.
  • the current stage is from the first comparison and learning, to the comparison and learning.
  • the result of the fuzzy decision is consistent with the result of the accurate decision, that is, the target screen display area in which the imaging result obtained in step S405 matches the preset human body image and the target screen display area in which the result of the face recognition in step S407 is consistent with the preset face is The same screen display area; the difference between the fuzzy decision result and the accurate decision result is the target screen display area in which the imaging result obtained in step S405 matches the preset human body image, and the result of the face recognition in step S407 is consistent with the preset face.
  • the screen display area is a different screen display area.
  • S409 Calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times.
  • the total number of statistics is the sum of the number of times the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas.
  • the second preset ratio threshold may be automatically set by the mobile terminal, or may be set by the user.
  • the ratio is greater than the second preset ratio threshold, it indicates that the result of the fuzzy decision has a certain degree of accuracy.
  • the result of the fuzzy decision is directly displayed, so there is no need to make an accurate decision.
  • the accuracy of the fuzzy decision is not enough, the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning.
  • the memory invokes the results of the current phase comparison and learning to calculate the ratio of the result of the fuzzy decision and the result of the accurate decision to the ratio of the total number of comparisons and learning.
  • the second preset ratio threshold is 80%.
  • the number of times the statistical fuzzy decision result and the accurate decision result are consistent with each other is 8 times, and the number of inconsistencies is 2 times, that is, The total number of comparisons and learnings is 10, so in the next comparison and learning, if the fuzzy judgment result is consistent with the accurate judgment result, then the number of times the result of the fuzzy judgment is consistent with the result of the accurate judgment is compared.
  • the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning.
  • the results of the current phase comparison and learning are recalled from the memory to calculate the result of the fuzzy decision and the accurate decision.
  • the results of the two are consistent with the proportion of the total number of comparisons and learning.
  • executing S419 determining whether there is an intervention operation, if not, keeping the fuzzy decision result unchanged, and displaying the result by the fuzzy decision; if yes, executing S420: determining according to the intervention operation Two screen display area.
  • the result of the fuzzy decision is inaccurate, causing the user to manually set the displayed screen, and the second screen display area is determined according to the user's intervention operation, and the target interface is in the second Displayed in the screen display area while withdrawing the target interface presented in the first screen display area.
  • S401-S413 is re-executed, and the comparison and learning of the fuzzy decision and the accurate judgment result are performed.
  • FIG. 9a is a schematic flowchart of a method for determining a high-precision display method according to an embodiment of the present invention, including the following steps:
  • S501 The mobile terminal receives data reported by the target sensor.
  • the mobile terminal specifies a corresponding decision according to the type and quantity of the sensor that is equipped by itself, and invokes a corresponding sensor to perform data collection and report.
  • S502 Detect whether there is a sound of a specific frequency range in a direction of each of the at least two screen display areas.
  • the user operates the feature information in the embodiment of the present invention to detect whether there is a sound matching the voice frequency on the screen display area of the mobile terminal as an example.
  • the sound sensor is called to detect whether there is a sound of a specific frequency range above each screen display area of the mobile terminal.
  • the sound of the specific frequency range is a frequency interval range that matches the vocal frequency.
  • the detection step is to confirm whether there is a user in front of the screen display area.
  • step S511 is performed to end the subsequent detection on the screen display area.
  • S503 Determine, in the at least two screen display areas, a screen display area in which a sound of a specific frequency range exists in a direction as a target screen display area.
  • the detection result is that the sound has a specific frequency range
  • the screen display area is determined as the target screen display area, and the number of the target screen display area may be at least One.
  • S504 Detect whether the result of the face recognition in the vertical direction of the target screen display area is consistent with the preset face.
  • the user's physical information in the embodiment of the present invention is used to detect whether the result of the face recognition in the vertical direction of the screen display area of the mobile terminal is consistent with the preset face.
  • the specific detection process is similar to step S406, and the camera sensor is called to perform face recognition in the imaging direction to identify whether it is a user.
  • the facial image of the user may be pre-acquired and stored in the memory of the terminal, and the facial image collected by the camera sensor is compared with the facial image of the user. If the comparison result is consistent, step S505 is performed; if the comparison result is inconsistent, Step S512 is executed to end the subsequent detection on the target screen display area.
  • S505 Determine a target screen display area that matches the result of the face recognition with the preset face as the first screen display area.
  • the mobile terminal performs a fuzzy decision under a high-precision condition: the target screen display area that matches the result of the face recognition with the preset face is determined as the first The screen displays the area and saves the result of the fuzzy decision in the memory of the terminal.
  • S506 Determine whether the result of the voice recognition in the vertical direction of the target screen display area consistent with the preset face is consistent with the preset sound.
  • the further detected information is that the voice recognition in the vertical direction of the screen display area of the mobile terminal is detected. Whether the result is consistent with the preset sound; the preset sound is a sound of a user stored in advance in the mobile terminal, and the detecting step is for detecting whether the user on the target screen display area is the user of the mobile terminal.
  • the sound sensor is called to perform voice recognition in the vertical direction of the screen display area in which the result of the face recognition coincides with the preset face, and whether the recognition is consistent with the user's voice. If the recognition result is consistent with the user voice, step S507 is performed; if the recognition result is inconsistent with the user voice, step S513 is performed to end the subsequent detection of the screen display area that is not consistent with the preset sound.
  • S507 Determine a target screen display area in which the result of the voice recognition is consistent with the preset sound as the first screen display area.
  • the target The screen display area serves as the first screen display area.
  • the two target screen display areas are respectively a target display area in which the result of the face recognition is consistent with the preset face and whether the result of the voice recognition is consistent with the preset sound.
  • the result of the fuzzy decision is invoked from the memory, and the fuzzy decision result and the accurate decision result under high-precision conditions are compared and learned, and the number of times the fuzzy decision result and the accurate decision result are consistent and inconsistent in the current stage are counted.
  • the current stage is from the first comparison and learning, to the comparison and learning.
  • the result of the fuzzy decision is consistent with the result of the accurate decision, that is, the target screen display area in which the result of the face recognition obtained in step S505 is consistent with the preset face and the target screen display area in which the result of the voice recognition in step S507 is consistent with the preset sound.
  • the same screen display area; the fuzzy decision result and the accurate decision result are inconsistent, that is, the result of the face recognition obtained in step S505 is consistent with the preset face and the result of the voice recognition in step S507 is consistent with the preset sound.
  • the target screen display area is a different screen display area.
  • S509 Calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistics.
  • the total number of statistics is the sum of the number of times the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas.
  • the third preset ratio threshold may be automatically set by the mobile terminal, or may be set by the user. When the ratio is greater than the third preset ratio threshold, it indicates that the result of the fuzzy decision has a certain degree of accuracy. In the subsequent high-precision display method decision, the result of the fuzzy decision is directly displayed, so there is no need to make an accurate decision.
  • the condition detection that is, after the fuzzy decision result is obtained
  • the target display area that directly matches the result of the face recognition with the preset face is determined as the first screen display area, and the result of the face recognition is not performed. The step of performing voice recognition in the vertical direction of the target screen display area in which the face is consistent.
  • the ratio is not greater than the third preset ratio threshold, it indicates that the accuracy of the fuzzy decision is not enough, the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning.
  • the memory invokes the results of the current phase comparison and learning to calculate the ratio of the result of the fuzzy decision and the result of the accurate decision to the ratio of the total number of comparisons and learning.
  • the third preset ratio threshold is 80%.
  • the target display area that directly matches the result of the face recognition with the preset face is determined as the first screen display area, and the result of the face recognition and the preset face are not executed.
  • the results of the comparison and learning are saved to the memory for the next comparison and learning when the result of the current phase comparison and learning is recalled from the memory to calculate the result of the fuzzy decision and the result of the accurate decision are both consistent with the total number of comparisons and learning.
  • executing S519 determining whether there is an intervention operation, if not, keeping the fuzzy decision result unchanged, and displaying the result by the fuzzy decision; if yes, executing S520: determining according to the intervention operation Two screen display area.
  • the result of the fuzzy decision is inaccurate, causing the user to manually set the displayed screen, and the second screen display area is determined according to the user's intervention operation, and the target interface is in the second Displayed in the screen display area while withdrawing the target interface presented in the first screen display area.
  • S501-S513 is re-executed, and the comparison and learning of the fuzzy decision and the accurate decision result are performed.
  • FIG. 10 is a schematic diagram of a process for determining a high-precision display method according to another embodiment of the present invention, wherein the detection processes and methods of steps S601 to S605 are the same as corresponding S501 to S505.
  • step S606 is performed to determine whether the result of the iris recognition in the vertical direction of the target screen display area in which the result of the face recognition is consistent with the preset face is Set the iris to be consistent.
  • the information for further detecting the accurate determination under the high-precision condition is whether the iris recognition result in the vertical direction of the screen display area of the mobile terminal is consistent with the preset iris; It is assumed that the iris is the iris of the user stored in advance in the mobile terminal, and the detecting step is to detect whether the user on the target screen display area is a user, corresponding to the foregoing step S506.
  • the iris recognition sensor is called to perform iris recognition in the vertical direction of the target screen display area in which the result of the face recognition coincides with the preset face, and whether the recognition is consistent with the iris of the user. If the recognition result is consistent, step S607 is performed; if the recognition result is inconsistent, it is proved that the user in the vertical direction of the screen display area is not the user, and step S613 is executed to end the subsequent detection on the screen display area.
  • the preset target screen display area that matches the result of the iris recognition with the preset iris is determined as the first screen display area.
  • the target The screen display area serves as the first screen display area.
  • the two target screen display areas are a target display area in which the result of the face recognition is consistent with the preset face and a target screen display area in which the result of the iris recognition is consistent with the preset iris.
  • the result of the fuzzy decision is called from the memory, and the fuzzy decision result under the high precision condition is compared with the accurate decision result. Compared with learning, the number of times the fuzzy decision result and the accurate judgment result are consistent and inconsistent in the current stage are counted. Among them, the current stage is from the first comparison and learning, to the comparison and learning.
  • the result of the fuzzy decision is consistent with the result of the accurate decision, that is, the target screen display area in which the result of the face recognition obtained in step S605 is consistent with the preset face and the target screen display area in which the result of the iris recognition in step S607 is consistent with the preset iris.
  • the display area of the same screen; the result of the fuzzy decision and the result of the accurate decision are that the result of the face recognition obtained in step S605 is consistent with the target face display area of the preset face and the result of the iris recognition in step S607 is consistent with the preset iris.
  • the target screen display area is a different screen display area.
  • S609 Calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistics.
  • the total number of statistics is the sum of the number of times the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas.
  • the fourth preset ratio threshold may be automatically set by the mobile terminal, or may be set by the user. When the ratio is greater than the fourth preset ratio threshold, it indicates that the result of the fuzzy decision has a certain degree of accuracy. In the subsequent high-precision display method decision, the result of the fuzzy decision is directly displayed, so there is no need to make an accurate decision.
  • the condition detection that is, after the fuzzy decision result is obtained
  • the target display area that directly matches the result of the face recognition with the preset face is determined as the first screen display area, and the result of the face recognition is not performed. The step of performing iris recognition in the vertical direction of the target screen display area where the face is consistent.
  • the accuracy of the fuzzy decision is not enough, the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning.
  • the memory invokes the results of the current phase comparison and learning to calculate the ratio of the result of the fuzzy decision and the result of the accurate decision to the ratio of the total number of comparisons and learning.
  • the fourth preset ratio threshold is 80%.
  • the number of times the statistical fuzzy decision result and the accurate judgment result are consistent is 8 times, and the number of inconsistencies is 2 times, that is, The total number of comparisons and learnings is 10, so in the next comparison and learning, if the fuzzy judgment result is consistent with the accurate judgment result, then the number of times the result of the fuzzy judgment is consistent with the result of the accurate judgment is compared.
  • the result of the fuzzy decision is directly Displaying, the condition detection is no longer performed for the precise decision, that is, after the fuzzy decision result is obtained, the target display area that directly matches the result of the face recognition with the preset face is determined as the first screen display area, and the execution is not performed.
  • the step of performing iris recognition in the vertical direction of the target screen display area in which the result of the face recognition coincides with the preset face; if the next comparison and learning, the fuzzy decision If the result of the accurate judgment is inconsistent, then the ratio of the result of the fuzzy decision and the result of the accurate judgment is proportional to the total number of comparisons and the total number of times of learning 8/(10+1) 72.7%, which is not greater than the fourth preset ratio.
  • the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning.
  • the results of the current phase comparison and learning are recalled from the memory to calculate the result of the fuzzy decision and the accurate decision.
  • the results of the two are consistent with the proportion of the total number of comparisons and learning.
  • executing S619 determining whether there is an intervention operation, if not, keeping the fuzzy decision result unchanged, and displaying the result by the fuzzy decision; if yes, executing S620: determining according to the intervention operation Two screen display area.
  • the result of the fuzzy decision is inaccurate, causing the user to manually set the displayed screen, and the second screen display area is determined according to the user's intervention operation, and the target interface is in the second Displayed in the screen display area At the same time, the target interface presented in the first screen display area is withdrawn. And in the next high-precision display method decision, S601-S613 is re-executed, and the comparison and learning of the fuzzy decision and the accurate decision result are performed.
  • the gyroscope/gravity sensor may be first used to calculate the pitch angle coordinates of each screen.
  • the gyroscope/gravity sensor may be first used to calculate the pitch angle coordinates of each screen.
  • determine which screens may be the screens that the user is using according to the posture position of the terminal, and then based on the pre-selected display method decision, call the corresponding sensors on the screens that may be used to collect data, and make a judgment. There is no need to call the corresponding sensor on the screen that is not possible to use to collect data, reducing power consumption.
  • the multi-screen mobile terminal calls different sensors to perform low-precision, medium-precision and high-precision display method decisions, which ensures the judgment. Accuracy meets the needs of mobile terminals with different sensor configurations, and provides users with more choices for manually selecting the specified sensors.
  • the face information can be collected by the camera sensor; the imaging is performed by the infrared array sensor; by the ultrasonic sensor, the laser sensor or the light/proximity The sensor measures the distance between the face and each screen display area to determine the screen display area that the user is using.
  • the face sensor information may be collected by the camera sensor, and the screen display area that the face may face may be analyzed according to the face information, thereby determining the screen display area that the user is using; the imaging result obtained by the infrared array sensor may be used to determine the user is using the image.
  • the screen display area; the distance between the face and each screen display area can be measured by an ultrasonic sensor, a laser sensor or a light/proximity sensor, and the distance between the face and each screen display area is compared, and the screen display area with the smallest distance is determined as The screen display area that the user is using.
  • the multi-screen display terminal includes four screen display areas A, B, C, and D (A side screen, B side screen, C side screen, and D side screen), and the four screen display areas are in the same flat. Assuming that the user is using the multi-screen display terminal, when it is detected by the camera sensor that the orientation of the face may face the screen display area A, the screen display area A is determined to be the screen that the user is using, and the screen display area A is activated for display.
  • the screen display areas A and C may be activated for display at the same time, or the ultrasonic sensor, the laser sensor or the light/proximity sensor may be further used to measure the face.
  • the distance between the screen display areas A and C is activated, and the screen display area with a small activation distance is displayed. It should be noted that whether the further detection is determined depends on the type of sensor supported by the terminal and selected by the user.
  • other sensors may also be used to determine the screen display area that the user is using.
  • the user can set which sensors correspond to which display method decisions.
  • the user can set the camera sensor and the infrared array sensor corresponding to the first display method decision; the camera sensor and the laser sensor correspond to the second display method decision; the camera sensor, the infrared array sensor, and the ultrasonic sensor correspond to the third display method decision, and the like.
  • FIG. 11 is a schematic flowchart of a display method of a multi-screen mobile terminal according to another embodiment of the present invention. It is described that when the screen display area of the multi-screen mobile terminal is in an active state, it is determined by its own sensor whether it is to be activated. The case where the screen is turned off is another implementation manner of the display method decision of the mobile terminal. As shown in FIG. 11, when the screen display area of the mobile terminal is in an active state, the method for performing screen shutdown by the data uploaded by the sensor includes the following steps:
  • S701 Detect whether the mobile terminal has a screen display area in an activated state.
  • the senor is equipped with hardware of the mobile terminal, and different types of sensors are different in type and quantity, and the type and function of the reported data are different.
  • the mobile terminal formulates different decisions according to its own sensor configuration, and calls different sensors to obtain different types of reported data to make display judgments. Among them, the type of sensor called and the corresponding The display decision can be selected by the mobile terminal according to the actual situation, or can be manually selected by the user.
  • the target sensor includes at least one sensor selected according to a preset sensor selection mechanism; the sensor selection mechanism is a selection mechanism determined by a sensor type supported by the mobile terminal.
  • S703 The mobile terminal determines a first screen closing area in the activated screen display area according to the data reported by the target sensor, and closes the first screen closing area.
  • the mobile terminal determines, according to the data reported by the sensor, whether the user is currently using the screen display area that is currently in the active state, and determines whether the current screen display area needs to be closed, and needs to be closed.
  • the screen display area is closed.
  • the accuracy may not be different from the screen area that the user wants to close, and then the user intervenes to Correct the behavior of the shutdown.
  • the specific implementation of the intervention operation may be, for example, but not limited to, double-clicking the second screen closure area, long pressing the second screen closure area, or displaying a button to close the screen in the second screen closure area, the user Click OK to close and so on.
  • FIG. 12a is a schematic diagram of a decision process of a display method according to an embodiment of the present invention.
  • the sensor is specifically invoked to describe the shutdown process in more detail; and the following steps are included:
  • S801 Detect whether the mobile terminal has an on-screen display area in an activated state.
  • the mobile terminal receives the data reported by the target sensor, specifically, whether there is an action change in the direction of detecting the screen display area of the mobile terminal, whether there is a temperature that matches the temperature of the human body, and the measurement Whether the result is greater than the preset distance threshold and whether the result of the face recognition is consistent with the preset face.
  • the specific detection steps are as follows.
  • the infrared array sensor is called to detect whether there is an action change in the direction of the screen display area in the activated state. If there is no action change, it is proved that the user on the screen display area is always used, and step S811 is performed. End the follow-up test and keep the display of the screen unchanged. If the detection result is that there is an action change, it is proved that the user on the screen display area may leave the screen display area, so the subsequent detection is continued, and step S803 is performed.
  • the infrared array sensor is invoked to perform temperature detection in a direction of the screen display area where the motion changes, and detect whether there is a temperature matching the human body temperature interval in the direction of the screen display area; if the detection result is the presence and the human body temperature interval The matching temperature proves that there is a user in use in the direction of the screen display area, and step S812 is performed to end the subsequent detection and keep the display unchanged. If the detection result is that there is no temperature matching the human body temperature interval, the user in the direction of the screen display area may leave, so the subsequent detection is continued, and step S804 is performed.
  • S804 Detect whether it is greater than a preset distance threshold.
  • the laser sensor, the ultrasonic sensor or the light/proximity sensor is called to perform ranging in the vertical direction of the screen display area where there is no temperature matching the human body temperature interval, and the detection result is greater than the preset distance threshold, if the detection result is If it is not greater than, it is proved that there is a user in the vertical direction of the screen display area, and step S813 is performed to end the subsequent detection and keep the screen display unchanged; If the detection result is greater than the preset distance threshold, it is proved that the user in front of the screen display area may have left, and the mobile terminal makes a fuzzy close decision according to step S805: the distance measurement result is greater than the preset distance threshold.
  • the screen display area is determined as the first screen close area, and the result of the fuzzy decision is saved in the memory of the terminal.
  • the face recognition is detected.
  • the camera sensor is invoked to be greater than the preset distance threshold.
  • the face recognition is performed in the direction of the screen display area to detect whether the result of the face recognition is consistent with the preset face; wherein the preset face is a face image of the user stored in advance in the mobile terminal.
  • step S814 is performed to end the subsequent detection, and the screen display remains unchanged
  • the detection result is Inconsistent with the preset face, it is proved that the user has left the screen display area, or the user in the direction of the screen display area is not the user himself, then according to step S807, an accurate close decision is made: the face recognition result and the preset
  • the screen display area in which the face is in an active state is determined as the first screen close area.
  • S808 Count the number of times that the two screen display areas in the active state are the same screen display area and the number of times the two screen display areas in the active state are different screen display areas.
  • the two screen display areas in the activated state are the screen display area in the active state in which the ranging result is greater than the preset distance threshold, and the screen display area in the active state in which the face recognition result is inconsistent with the preset face. .
  • the result of the fuzzy decision is invoked from the memory, and the fuzzy decision result and the accurate decision result are compared and learned, and the number of times the fuzzy decision result and the accurate decision result are consistent and inconsistent in the current stage are counted.
  • the current stage is from the first comparison and learning, to the comparison and learning.
  • the result of the fuzzy decision is consistent with the result of the accurate decision, that is, the screen display area in which the ranging result obtained in step S804 is greater than the preset distance threshold is in an active state, and the face recognition result is inactive with the preset face in step S806.
  • the screen display area is the same screen display area; the fuzzy decision result and the accurate decision result are inconsistent, that is, the screen display area in the active state in which the ranging result obtained in step S804 is greater than the preset distance threshold and the face recognition result in step S806 and
  • the screen display areas in which the preset faces are inconsistent are in different screen display areas.
  • the total number of statistics is the sum of the number of times the two screen display areas in the active state are the same screen display area and the number of times the two screen display areas in the active state are different screen display areas.
  • the fifth preset ratio threshold may be automatically set by the mobile terminal, or may be set by the user.
  • the ratio is greater than the fifth preset ratio threshold, it indicates that the result of the fuzzy decision has a certain degree of accuracy.
  • the result of the fuzzy decision is directly displayed, so there is no need to make an accurate decision again.
  • Condition detection that is, after the fuzzy decision result is obtained, the screen display area in the active state in which the ranging result is greater than the preset distance threshold is directly determined as the first screen closed area, and whether the detection is consistent with the preset face is not performed. A step of.
  • the accuracy of the fuzzy decision is not enough, the current comparison and learning are ended, and the results of the comparison and learning are saved to the memory for the next comparison and learning.
  • the memory invokes the results of the current phase comparison and learning to calculate the ratio of the result of the fuzzy decision and the result of the accurate decision to the ratio of the total number of comparisons and learning.
  • the fifth preset ratio threshold is 80%.
  • the number of times the statistical fuzzy decision result and the accurate judgment result are consistent is 8 times, and the number of inconsistencies is 2 times, that is, The total number of current comparisons and learning is 10, then In the next comparison and learning, if the fuzzy decision result is consistent with the accurate decision result, then the ratio of the result of the fuzzy decision and the result of the accurate decision at this time is the ratio of the total number of times of comparison and learning (8+1).
  • executing S820 After executing S815-S819 to obtain the result of the fuzzy decision, executing S820: determining whether there is an intervention operation, if not, keeping the fuzzy decision result unchanged, and displaying the result by the fuzzy decision; if yes, executing S821: determining according to the intervention operation Two screen display area.
  • the result of the fuzzy decision is inaccurate, causing the user to manually set the displayed screen, and the second screen display area is determined according to the user's intervention operation, and the target interface is in the second Displayed in the screen display area while withdrawing the target interface presented in the first screen display area.
  • S801-S814 is re-executed, and the comparison and learning of the fuzzy decision and the accurate decision result are performed.
  • the detection means adopted in the method flow is only a specific embodiment in the implementation process.
  • the mobile terminal may call other sensors to detect and judge according to its own hardware configuration.
  • the detection process here is only a method description, and is not limited herein.
  • the embodiment of the present invention further discloses a multi-screen mobile terminal.
  • the multi-screen mobile terminal may at least include: a sensor environment data decision module, an integrated decision module, and a user assist decision module; among them:
  • the sensor environment data decision module can be used for multi-sensor data classification processing and reporting.
  • the senor may include a gyroscope/gravity sensor, a laser sensor, a camera sensor, an infrared array sensor, an ultrasonic sensor, a light/proximity sensor, a pressure sensor, an iris recognition sensor, and an acoustic sensor wait.
  • a gyroscope/gravity sensor a laser sensor, a camera sensor, an infrared array sensor, an ultrasonic sensor, a light/proximity sensor, a pressure sensor, an iris recognition sensor, and an acoustic sensor wait.
  • the integrated decision module can be used to comprehensively analyze the first screen display area or the first screen closed area according to the data uploaded by the multi-sensor.
  • the user assisted decision module can be used for manual setting or intervention by the user.
  • the user-assisted decision-making module can provide the user with a multi-sensor intelligent display switch to the user, and the user determines which sensors are enabled to collect data through the switch; the user-assisted decision-making module can also provide the user with an on-screen number control switch for the user to select, the user can Select which screens to use specifically; the user-assisted decision-making module can also provide the user's intervention decision-making function to judge the user, that is, the comparison and learning of the results of the fuzzy judgment and the accurate judgment, directly display the result of the fuzzy judgment, After making an accurate decision, if the result of the fuzzy decision is incorrect, the user can intervene to switch the screen.
  • different sensors are combined to achieve different precision decisions, including: low-precision decision making, medium-precision decision making, and high-precision decision making.
  • Each decision can be divided into two stages, including : Fuzzy judgments and precise judgments.
  • the terminal can make an intelligent decision output according to different decisions, and can also comprehensively make decisions in combination with the user's intervention operation.
  • FIG. 14 is a schematic structural diagram of a display device according to an embodiment of the present invention.
  • the display device 900 can include at least an opening module 901, a display data receiving module 902, a display screen determining module 903, and a display module 904.
  • An interventional operation detection module 905, and an intervention display module 906 are displayed, wherein:
  • the opening module 901 is configured to receive, by the mobile terminal, an open command input by the user, where the open command is used to indicate that the target interface is opened.
  • the display data receiving module 902 is configured to receive, by the mobile terminal, data reported by the target sensor.
  • the display screen determining module 903 is configured to determine, by the mobile terminal, the first screen display area in the at least two screen display areas according to the data reported by the target sensor.
  • the display module 904 is configured to present the target interface in the first screen display area.
  • the intervention operation detection module 905 is configured to detect whether there is a user intervention operation after the interface of the first application is presented in the first screen display area; the access operation is used to indicate to determine the second screen display Area.
  • the intervention display module 906 is configured to determine a second screen display area in the at least two screen display areas according to the user intervention operation, withdraw the target interface presented in the first screen display area, and simultaneously in the The target interface is presented in the second screen display area.
  • FIG. 15 is a schematic structural diagram of a display screen determining module according to an embodiment of the present invention.
  • the display screen determining module 903 may at least include: a feature information detecting unit 9031, a target screen determining unit 9032, and a physical sign information detecting unit. 9033, and a first screen display area determining unit 9034, wherein:
  • the feature information detecting unit 9031 is configured to detect, by the mobile terminal, whether there is user operation feature information for the screen display area according to data reported by the target sensor.
  • the target screen determining unit 9032 is configured to determine a target screen display area in the at least two screen display areas according to the user operation feature information if the feature information detecting unit detects YES.
  • the vital sign information detecting unit 9033 is configured to determine whether the user's physical sign information exists in the vertical direction of the target screen display area.
  • the first screen display area determining unit 9034 determines the target screen display area in which the user's vital sign information exists in the vertical direction as the first screen display area.
  • the pressure detecting subunit in the feature information detecting unit 9031 is used to detect whether there is touch pressure in each of the at least two screen display areas when the low-precision decision is made.
  • the ranging subunit and the imaging subunit in the vital sign information detecting unit 9033 are called, wherein:
  • the ranging subunit is configured to perform ranging in a vertical direction of the target screen display area, and determine whether the ranging result is greater than a preset distance threshold.
  • the ranging subunit and the pressure detecting subunit here are used by the mobile terminal to make a fuzzy decision under low precision conditions.
  • An imaging subunit configured to perform imaging in a vertical direction of a target screen display area where the ranging result is not greater than a preset distance threshold if the determination result of the ranging subunit is not greater than a preset distance threshold, and detect whether the imaging result is Matches the preset body image.
  • the imaging subunit here combines the above-mentioned pressure detecting subunit and the ranging subunit for the mobile terminal to make an accurate decision under low precision conditions.
  • the display device 900 may further include: a first statistic module, a first calculation module, and a first determination module, where:
  • a first statistic module configured to count the number of times that the two target screen display areas are the same screen display area and the two target screen displays The area is the number of times of the different screen display areas, wherein the two target screen display areas are the target screen display area where the ranging result is not greater than the preset distance threshold and the target screen display area where the imaging result matches the preset body image.
  • a first calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of screen display areas and the number of times the two target screen display areas are different screen display areas.
  • the first determining module is configured to determine, when the ratio calculated by the first calculating module is greater than the first preset ratio threshold, the target screen display area whose ranging result is not greater than the preset distance threshold as the first screen display area.
  • the temperature subunit in the feature information detecting unit 9031 is used to detect whether there is a preset temperature in the vertical direction of each of the at least two screen display areas when the mobile terminal performs the medium precision determination The interval matches the temperature.
  • the imaging subunit and the face recognition subunit in the vital sign information detecting unit 9033 are called, wherein:
  • an imaging subunit configured to perform imaging in a vertical direction of the target screen display area to detect whether the imaging result matches the preset human body image.
  • the temperature subunit here and the imaging subunit are used by the mobile terminal to make a fuzzy decision under medium precision conditions.
  • a face recognition subunit configured to perform face recognition in a vertical direction of a target screen display area where the imaging result matches the preset body image, and if the detection result of the imaging subunit is YES, the result of detecting the face recognition Whether it is consistent with the preset face.
  • the face recognition subunit here combines the above-described temperature subunit and imaging subunit for the mobile terminal to make an accurate decision under medium precision conditions.
  • the display device 900 may further include: a second statistic module, a second calculation module, and a second determination module, where:
  • a second statistic module configured to count the number of times that the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas, wherein the two target screen display areas are respectively a target screen display area in which the imaging result matches the preset human body image and a target screen display area in which the result of the face recognition is consistent with the preset face;
  • a second calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • a second determining module configured to determine, as the first screen display area, the target screen display area that matches the imaging result with the preset body image when the ratio is greater than the second preset ratio threshold.
  • the sound subunit in the feature information detecting unit is called when the mobile terminal performs high-precision decision for detecting whether there is a sound of a specific frequency range in the direction of each screen display area of the at least two screen display areas.
  • the face recognition subunit in the vital sign detection unit and the voice recognition subunit are invoked, wherein:
  • a face recognition subunit configured to perform face recognition in a vertical direction of the target screen display area after detecting the screen display area where the sound of the specific frequency range is determined as the target screen display area, and detecting the face recognition Whether the result is consistent with the preset face.
  • the sound subunit and the face recognition subunit here are used for the mobile terminal to make a fuzzy decision under high precision conditions; the sound recognition subunit is used for a target screen whose result of the face recognition is consistent with the preset face The sound recognition is performed in the vertical direction of the display area to determine whether the result of the sound recognition is consistent with the preset sound.
  • the voice recognition subunit here combines the above-described sound subunit and face recognition subunit for the mobile terminal to make an accurate decision under high precision conditions.
  • the display device 900 may further include: a third statistic module, a third calculation module, and a third determination module, where:
  • the third statistic module is configured to count the number of times that the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas, wherein the two target screen display areas are respectively a result of the face recognition, a target display area consistent with the preset face and a target screen display area in which the result of the sound recognition is consistent with the preset sound;
  • a third calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • the third determining module is configured to determine, as the first screen display area, the target display area that matches the result of the face recognition with the preset face when the ratio is greater than the third preset ratio threshold.
  • the vital sign information detecting unit 9033 includes: a face recognition subunit, and an iris recognition subunit, wherein:
  • a face recognition subunit configured to perform face recognition in a vertical direction of the target screen display area after detecting the screen display area where the sound of the specific frequency range is determined as the target screen display area, and detecting the face recognition Whether the result is consistent with the preset face.
  • the iris recognition subunit is configured to perform iris recognition in a vertical direction of the target screen display area in which the result of the face recognition is consistent with the preset face, and determine whether the result of the iris recognition is consistent with the preset iris.
  • the iris recognition subunit herein combines the above-described sound subunit and face recognition subunit for the mobile terminal to make an accurate decision under high precision conditions.
  • the display device 900 may further include: a fourth statistic module, a fourth calculation module, and a fourth determination module, where:
  • the fourth statistic module is configured to count the number of times that the two target screen display areas are the same screen display area and the number of times the two target screen display areas are different screen display areas, wherein the two target screen display areas are respectively
  • the result of the face recognition is the same as the preset face and the target display area and the iris recognition result are the same as the preset iris display target display area;
  • a fourth calculating module configured to calculate a ratio of the number of times that the two target screen display areas are the same screen display area to the total number of statistical times, wherein the total number of statistics is that the two target screen display areas are the same The sum of the number of times the screen display area is different from the number of times the two target screen display areas are different screen display areas;
  • a fourth determining module configured to determine, as the first screen display area, the target display area that matches the result of the face recognition with the preset face when the ratio is greater than the fourth preset ratio threshold.
  • FIG. 16 is a schematic structural diagram of a display device according to another embodiment of the present invention.
  • the display device 900 can at least include: a screen detection module 915, a shutdown data receiving module 916, a shutdown module 917, a closed intervention detection module 918, and a shutdown intervention.
  • Module 919 wherein:
  • the screen detecting module 915 is configured to detect whether the mobile terminal has a screen display area in an activated state.
  • the data receiving module 916 is closed, and when the screen detecting module detects that it is YES, the mobile terminal receives data reported by the target sensor.
  • the closing module 917 is configured to determine, by the mobile terminal, the first screen closing area in the screen display area in the activated state according to the data reported by the target sensor, and close the first screen closing area.
  • the intervention detection module 918 is configured to detect whether there is a user intervention operation after the closing the first screen closure area; and the intervention operation is used to indicate determining the second screen closure area.
  • the intervention module 919 is configured to: when the closed intervention detection module detects YES, determine a second screen closure area in the active screen display area according to the user intervention operation, and reactivate the first screen The area is closed and the second screen off area is simultaneously closed.
  • FIG. 17 is a schematic structural diagram of a shutdown module according to an embodiment of the present invention.
  • the shutdown module 917 may at least include: a shutdown operation detecting unit 9171, and a shutdown unit 9172, wherein:
  • the operation detecting unit 9171 is configured to detect, by the mobile terminal, whether there is user operation information in each of the at least one screen display area in the activated state according to the data reported by the target sensor.
  • the closing unit 9172 is configured to determine the first screen closing area according to the user operation information.
  • the shutdown operation detecting unit 9171 may include at least: an action subunit, a second temperature subunit, a second ranging subunit, and a second face recognition subunit, wherein:
  • an action subunit configured to detect whether there is an action change in a direction of each of the screen display areas in the active screen display area.
  • the second temperature subunit is configured to detect whether there is a temperature matching the preset temperature interval in a vertical direction of the screen display area in the active state where the motion change exists.
  • a second ranging subunit configured to perform ranging in a vertical direction of an active screen display area in a vertical direction without a temperature matching the preset temperature interval, and detecting whether the ranging result is greater than a preset distance threshold .
  • the second face recognition sub-unit is configured to perform face recognition in a vertical direction of the screen display area in an active state in which the ranging result is greater than a preset distance threshold, and detect whether the face recognition result is consistent with the preset face.
  • the display apparatus 900 may further include: a fifth statistic module, a fifth calculation module, and a fifth determination module, where:
  • a fifth statistic module configured to count the number of times that the two screen display areas in the active state are the same screen display area and the number of times that the two screen display areas in the activated state are different screen display areas, wherein the two The screen display areas in the activated state are the screen display areas in the active state in which the ranging result is greater than the preset distance threshold, and the screen display area in the active state in which the face recognition result is inconsistent with the preset face;
  • a fifth calculating module calculating a ratio of the number of times that the two screen display areas in the active state are the same screen display area to the total number of statistical times, wherein the total number of statistics is the two screens that are in an active state The sum of the number of times the display area is the same screen display area and the number of times the two screen display areas in the active state are different screen display areas;
  • the fifth determining module is configured to determine, when the ratio is greater than the fifth preset ratio threshold, that the screen display area in which the ranging result is greater than the preset distance threshold is determined as the first screen closing area.
  • the multi-screen mobile terminal judges that it is in use by using the data reported by the sensor provided by itself.
  • the state and position of the user in the process so as to display the target interface on the correct screen; at the same time, through various judgments of judgment and fuzzy comparison and accurate judgment, the accuracy of the judgment is improved, and the judgment of the mobile terminal is more accurate. Improve the user experience.
  • the program can be stored in a computer readable storage medium, when the program is executed
  • the flow of the method embodiments as described above may be included.
  • the foregoing storage medium includes various media that can store program codes, such as a ROM or a random access memory RAM, a magnetic disk, or an optical disk.

Abstract

本发明公开了一种显示方法、装置及终端,其中,所述显示方法应用于具有至少两个屏幕显示区的移动终端,包括:移动终端接收用户输入的开启指令,所述开启指令用于指示打开目标界面;所述移动终端接收目标传感器上报的数据;所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,并在所述第一屏幕显示区中呈现所述目标界面。能够准确识别用户的位置和使用状态,进行后续相应操作,方便了用户的使用。

Description

显示方法、装置及终端
本申请要求于2016年11月30日提交中国专利局、申请号为201611095183.5、发明名称为“一种多屏移动终端的显示方法和设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
本申请要求于2017年03月30日提交中国专利局、申请号为201710204290.5、发明名称为“一种多屏移动终端的显示方法和设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本发明涉及移动终端领域,尤其涉及一种显示方法、装置及终端。
背景技术
随着科技的发展,双屏、众屏、以及柔性显示屏在智能移动终端上越来越普及。同时,移动终端上也集成了越来越多的,基于不同感知原理的传感器模块,用于感知各种信息,帮助移动终端获取外界信息从而做出各种决策。
对于多屏移动终端设备,由于具有多个可供显示的屏幕,在使用的过程中每个屏幕会根据用户的需求进行不同的显示。但是如果将用户需要的内容显示在不适当的屏幕上,会造成隐私被泄露、用户体验不好等负面影响;另一方面,如果用户在使用多屏移动终端的过程中将多个屏幕同时打开进行显示,还会使得终端的功耗增加,影响续航效果。在目前现有的技术中,采用的是通过移动终端中配备的重力传感器来判别用户的使用状态,以垂直向上的屏幕作为用户正在使用的屏幕,进行画面显示。但是,这种通过重力感应的方式只能对用户的使用状态进行粗略的判断,并不能准确判断出用户正在使用的屏幕从而进行显示,且无法处理多屏移动终端的使用场景。
发明内容
本发明实施例提供了一种显示方法、装置及终端,通过传感器获取外界信息,准确识别用户的位置和使用状态,进行后续相应操作,方便了用户的使用。
本发明实施例第一方面公开了一种显示方法,应用于具有至少两个屏幕显示区的移动终端,该方法包括:
所述移动终端接收开启指令,所述开启指令用于指示打开目标界面;
所述移动终端获取目标传感器上报的数据;所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制;
所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,并在所述第一屏幕显示区中呈现所述目标界面。
所述移动终端通过所支持的传感器类型来选取至少一个传感器作为目标传感器,然后根据目标传感器上报的数据,对用户的方位和使用状态和方位进行较为准确的判定,可以充分结合移动终端的传感器类型的特性来进行判断,实现了智能地选取合适的屏幕对目标界面进行显示。
结合第一方面,在第一方面的第一种可能的实现方式中,所述在所述第一屏幕显示区中呈现所述目 标界面之后,还包括:
检测是否有介入操作;所述介入操作用于确定第二屏幕显示区;
若检测为是,则根据所述介入操作在所述至少两个屏幕显示区中确定第二屏幕显示区,撤回所述第一屏幕显示区中呈现的目标界面,并在所述第二屏幕显示区中呈现所述目标界面。
在移动终端根据上报的数据选取合适的屏幕对目标界面进行显示之后,如果选取的屏幕显示区不是用户想要选择的屏幕显示区,则用户可以进行介入操作,手动选取相应的屏幕显示区进行修正,进一步提高了显示的准确度。
结合第一方面或第一方面的第一种可能的实现方式,在第一方面的第二种可能的实现方式中,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,包括:
所述移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息;
若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;
根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区。
作为一种可选的方式,所述移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息之前,还包括:
计算所述至少两个屏幕显示区的俯仰角坐标;
根据所述至少两个屏幕显示区的俯仰角坐标确定第一目标屏幕显示区;其中,所述第一目标屏幕显示区至少为一个;
所述根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,包括:
根据所述用户操作信息在所述第一目标屏幕显示区中确定目标屏幕显示区。
其中,所述至少两个屏幕显示区的俯仰角坐标可以由陀螺仪/重力传感器计算得到,能够快速确定用户可能使用的第一目标屏幕显示区,并基于所述第一目标屏幕显示区采集相应传感器的数据,可以提高工作效率,并降低功耗。
移动终端根据上报的用户操作特征信息和用户体征信息,从而对用户的方位和使用状态进行判断,选取合适的屏幕显示区对目标屏幕进行显示。
结合第一方面的第二种可能的实现方式,在第一方面的第三种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
若检测为是,则在所述至少两个屏幕显示区中将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值;
将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
作为一种可选的方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区 中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第一显示决策。
其中,所述第一显示决策的显示决策结果为所述将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第一显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第一显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第一显示决策。
所述第一方面的第三种可能的实现方式,提供了一种低精度条件下的显示方法决策,所述移动终端根据获取到触摸压力检测数据以及测距检测数据,进行低精度条件下的模糊判决以确定屏幕显示区。
结合第一方面的第二种可能的实现方式,在第一方面的第四种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
若检测为是,则在所述至少两个屏幕显示区中将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否不大于预设距离阈值;
在测距结果不大于预设距离阈值的目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
作为一种可选的方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第二显示决策。
其中,所述第二显示决策的显示决策结果为所述将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第二显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第二显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第二显示决策。
所述第一方面的第三种可能的实现方式,提供了一种低精度条件下的显示方法决策,所述移动终端根据获取到触摸压力检测、测距检测以及成像检测,进行低精度条件下的精确判决确定屏幕显示区。
结合第一方面的第四种可能的实现方式,在第一方面的第五种可能的实现方式中,所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为测距结果不大于预设距离阈值的目标屏幕显示区和成像结果与预设人体图像匹配的目标屏幕显示区;
计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统 计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
当所述比例大于第一预设比例阈值时,将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
移动终端通过对比低精度条件下的模糊判决和精确判决结果,得到模糊判决和精确判决结果保持相同的比例,当模糊判决和精确判决的结果不同时,所述移动终端的显示采用精确判决的结果,并进行判决结果的学习。当所述保持相同的比例达到预设的第一比例阈值时,说明模糊判决的结果也具有了一定的精确度,在之后的判断中,可以直接采用模糊判决的结果,不必再进行精确判断,简化了判决的过程。
结合第一方面的第二种可能的实现方式,在第一方面的第六种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
若检测为是,则在所述至少两个屏幕显示区中将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
作为一种可选的实现方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第三显示决策。
其中,所述第三显示决策的显示决策结果为所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第三显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第三显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第三显示决策。
所述第一方面的第六种可能的实现方式,提供了一种中精度条件下的显示方法决策,所述移动终端根据获取到温度检测数据以及成像检测数据,进行中精度条件下的模糊判决以确定屏幕显示区。
结合第一方面的第二种可能的实现方式,在第一方面的第七种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
若检测为是,则在所述至少两个屏幕显示区中将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
作为一种可选的实现方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第四显示决策。
其中,所述第四显示决策的显示决策结果为所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第四显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第四显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第四显示决策。
所述第一方面的第七种可能的实现方式,提供了一种中精度条件下的显示方法决策,所述移动终端根据获取到温度检测数据、成像检测数据以及人脸识别检测数据,进行中精度条件下的精确判决以确定屏幕显示区。
结合第一方面的第七种可能的实现方式,在第一方面的第八种可能的实现方式中,所述将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为成像结果与预设人体图像匹配的目标屏幕显示区和人脸识别的结果与预设人脸一致的目标屏幕显示区;
计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
当所述比例大于第二预设比例阈值时,将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
类似于第一方面的第五种可能的实现方式,当所述中精度条件下的模糊判决和精确判决的结果保持相同的比例达到预设的第二比例阈值时,说明模糊判决的结果也具有了一定的精确度,在之后的判断中,可以直接采用模糊判决的结果,不必再进行精确判断,简化了判决的过程。
结合第一方面的第二种可能的实现方式,在第一方面的第九种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
若检测为是,则在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
作为一种可选的实现方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第五显示决策。
其中,所述第五显示决策的显示决策结果为所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第五显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第五显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第五显示决策。
所述第一方面的第九种可能的实现方式,提供了一种高精度条件下的显示方法决策,所述移动终端根据获取到的音源识别数据以及人脸识别数据,进行高精度条件下的模糊判决以确定屏幕显示区。
结合第一方面的第二种可能的实现方式,在第一方面的第十种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
若检测为是,则在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行声音识别,判断声音识别的结果是否与预设声音一致;
将声音识别的结果是否与预设声音一致的目标屏幕显示区确定为第一屏幕显示区。
作为一种可选的实现方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第六显示决策。
其中,所述第六显示决策的显示决策结果为所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第六显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第六显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第六显示决策。
所述第一方面的第十种可能的实现方式,提供了一种高精度条件下的显示方法决策,所述移动终端根据获取到的音源识别数据、人脸识别数据以及声音识别数据,进行高精度条件下的精确判决以确定屏幕显示区。
结合第一方面的第十种可能的实现方式,在第一方面的第十一种可能的实现方式中,所述将声音识别的结果是否与预设声音一致的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示 区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和声音识别的结果是否与预设声音一致的目标屏幕显示区;
计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
当所述比例大于第三预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
类似于第一方面的第五种可能的实现方式以及第一方面的第七种可能的实现方式,当所述高精度条件下的模糊判决和精确判决的结果保持相同的比例达到预设的第三比例阈值时,说明模糊判决的结果也具有了一定的精确度,在之后的判断中,可以直接采用模糊判决的结果,不必再进行精确判断,简化了判决的过程。
结合第一方面的第二种可能的实现方式,在第一方面的第十二种可能的实现方式中,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
若检测为是,则在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别,判断虹膜识别的结果是否与预设虹膜一致;
将虹膜识别的结果与预设虹膜一致的目标屏幕显示区确定为第一屏幕显示区。
作为一种可选的实现方式,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区之前,还包括:
根据所述移动终端的传感器类型确定第七显示决策。
其中,所述第七显示决策的显示决策结果为所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
其中,根据所述移动终端的传感器类型确定第七显示决策,包括所述移动终端根据传感器类型自动选取传感器并确定第七显示决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第七显示决策。
所述第一方面的第十二种可能的实现方式,提供了另一种高精度条件下的显示方法决策,所述移动终端根据获取到的音源识别数据、人脸识别数据以及虹膜识别数据,进行高精度条件下的精确判决以确定屏幕显示区。
结合第一方面的第十二种可能的实现方式,在第一方面的第十三种可能的实现方式中,所述将虹膜识别的结果与预设虹膜一致的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示 区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和虹膜识别的结果与预设虹膜一致的目标屏幕显示区;
计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
当所述比例大于第四预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
类似于第一方面的第五种可能的实现方式以及第一方面的第七种可能的实现方式以及第一方面的第十种可能的实现方式,当所述高精度条件下的模糊判决和精确判决的结果保持相同的比例达到预设的第四比例阈值时,说明模糊判决的结果也具有了一定的精确度,在之后的判断中,可以直接采用模糊判决的结果,不必再进行精确判断,简化了判决的过程。
本发明实施例第二方面公开了一种显示方法,应用于具有至少两个屏幕显示区的移动终端,所述方法包括:
检测所述移动终端的是否有处于激活状态的屏幕显示区;
若检测为是,则所述移动终端获取目标传感器上报的数据;所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制;
所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,并关闭所述第一屏幕关闭区。
本发明实施例的第二方面为移动终端已被激活的屏幕提供了一种自动关闭的方法,所述移动终端通过传感器上传的数据,判断用户是否已经停止使用移动终端的部分屏幕显示区,从而指定关闭决策关闭相应的屏幕显示区,节约了电源的使用。
结合第二方面,在第二方面的第一种可能的实现方式中,在所述关闭所述第一屏幕关闭区之后,还包括:
检测是否有介入操作;所述介入操作用于确定第二屏幕关闭区;
若检测为是,则根据所述介入操作在所述处于激活状态的屏幕显示区中确定第二屏幕关闭区,重新激活所述第一屏幕关闭区,并关闭所述第二屏幕关闭区。
当所述移动终端做出屏幕关闭决策关闭相应的屏幕显示区之后,有可能由于判决不够精确关闭的屏幕显示区并非用户想要关闭的屏幕显示区,则用户可以通过手动介入操作的方式进行手动关闭,从而关闭正确的屏幕显示区。
结合第二方面或第二方面的第一种可能的实现方式,在第二方面的第二种可能的实现方式中,所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,包括:
所述移动终端根据所述至少一个目标传感器上报的数据,检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息;
根据所述用户操作信息确定第一屏幕关闭区。
移动终端根据目标传感器上报的数据,对用户的使用状态进行判断,选取用户没有进行使用的处在激活状态下的屏幕显示区并将其关闭。
结合第二方面的第二种可能的实现方式,在第二方面的第三种可能的实现方式中,所述移动终端根据所述目标传感器上报的数据,检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息,根据所述用户操作信息确定第一屏幕关闭区,包括:
检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
作为一种可选的方式,所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区之前,还包括:
根据所述移动终端的传感器类型确定第一关闭决策;
其中,所述第一关闭决策的关闭决策结果为将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
其中,根据所述移动终端的传感器类型确定第一关闭决策,包括所述移动终端根据传感器类型自动选取传感器并确定第一关闭决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第一关闭决策。
所述第二方面第三种可能的实现方式,提供了一种模糊关闭判决,所述移动终端根据获取已被激活屏幕的动作检测数据、温度检测数据以及测距检测数据,得到模糊关闭决策以确定第一屏幕关闭区。
结合第二方面的第二种可能的实现方式,在第二方面的第四种可能的实现方式中,所述检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息;若检测为是,则根据所述用户操作特征信息在所述处于激活状态的屏幕显示区中确定目标屏幕关闭区,并判断所述目标屏幕关闭区的垂直方向上是否存在用户的体征信息;将垂直方向上不存在用户体征信息的目标屏幕关闭区确定为第一屏幕关闭区,包括:
检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
在测距结果大于预设距离阈值的处于激活状态的屏幕显示区的垂直方向上进行人脸识别,检测人脸识别结果是否与预设人脸一致;
将人脸识别结果与预设人脸一致的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
作为一种可选的方式,所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区之前,还包括:
根据所述移动终端的传感器类型确定第二关闭决策;
其中,所述第二关闭决策的关闭决策结果为将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
其中,根据所述移动终端的传感器类型确定第二关闭决策,包括所述移动终端根据传感器类型自动选取传感器并确定第二关闭决策,也可以由用户自主选择所述移动终端包含的传感器种类确定第二关闭决策。
所述第二方面的第四种可能的实现方式,提供了一种精确关闭判决,所述移动终端根据获取已被激活屏幕的动作检测数据、温度检测数据、测距检测数据以及人脸识别数据,得到精确关闭决策以确定第一屏幕关闭区。
结合第二方面的第四种可能的实现方式,在第二方面的第五种可能的实现方式中,所述将人脸识别结果与预设人脸一致的处于激活状态的屏幕显示区确定为第一屏幕关闭区之后,还包括:
统计两个处于激活状态的屏幕显示区为同一屏幕显示区的次数以及所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数,其中,所述两个处于激活状态的屏幕显示区分别为测距结果大于预设距离阈值的处于激活状态的屏幕显示区和人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区;
计算所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数与所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数之和;
当所述比例大于第五预设比例阈值时,将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
移动终端通过对比模糊关闭判决和精确关闭判决的结果,得到两种判决结果保持相同的比例,当模糊关闭判决和精确关闭判决的结果不同时,所述移动终端的关闭采用精确判决的结果,并进行判决结果的学习。当所述保持相同的比例达到预设的第五比例阈值时,说明模糊关闭判决的结果也具有了一定的精确度,在之后的判断中,可以直接采用模糊关闭判决的结果,不必再进行精确关闭判断,简化了判决的过程。
本发明实施例第三方面公开了一种显示装置,包括:
开启模块,用于移动终端接收开启指令,所述开启指令用于指示打开目标界面;
显示数据获取模块,用于所述移动终端获取目标传感器上报的数据;所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制;
显示屏幕确定模块,用于所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区;
显示模块,用于在所述第一屏幕显示区中呈现所述目标界面。
结合第三方面,在第一种可能的实现方式中,还包括:
显示介入操作检测模块,用于所述在所述第一屏幕显示区中呈现所述第一应用的界面之后,检测是 否有介入操作;所述介入操作用于确定第二屏幕显示区;
介入显示模块,用于根据所述介入操作在所述至少两个屏幕显示区中确定第二屏幕显示区,撤回所述第一屏幕显示区中呈现的目标界面,并在所述第二屏幕显示区中呈现所述目标界面。
结合第三方面或第三方面的第一种可能的实现方式,在第二种可能的实现方式中,所述显示屏幕确定模块包括:
特征信息检测单元,用于所述移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息;
目标屏幕确定单元,用于若所述特征信息检测单元检测为是时,根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区;
体征信息检测单元,用于判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;
第一屏幕显示区确定单元,将垂直方向上存在用户体征信息的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第三种可能的实现方式中,所述特征信息检测单元包括:
压力检测子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
若检测为是,则所述目标屏幕确定单元将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
测距子单元,用于在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值;
若判断为是,则所述第一屏幕显示区确定单元将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第四种可能的实现方式中,所述操作信息检测单元包括:
压力检测子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
若检测为是,则所述目标屏幕确定单元将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
测距子单元,用于在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值;
所述体征信息检测单元还包括:
成像子单元,用于若所述测距子单元判断结果为不大于预设距离阈值时,在测距结果不大于预设距离阈值的目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
若判断为是,则所述第一屏幕显示区确定单元将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第四种可能的实现方式,在第三方面的第五种可能的实现方式中,还包括:
第一统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为测距结果不大于预设距离阈值的目标屏幕显示区和成像结果与预设人体图像匹配的目标屏幕显示区;
第一计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第一确定模块,用于当所述第一计算模块计算的比例大于第一预设比例阈值时,将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第六种可能的实现方式中,所述操作信息检测单元包括:
温度子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
若检测为是,则所述目标屏幕确定单元将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
成像子单元,用于在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
若成像结果为是,则所述第一屏幕显示区确定单元将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第七种可能的实现方式中,所述操作信息检测单元包括:
温度子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
若检测为是,则所述目标屏幕确定单元将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
成像子单元,用于在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
所述体征信息检测单元还包括:
人脸识别子单元,用于若所述成像子单元的检测结果为是时,在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
若检测结果为与预设人脸一致,则所述第一屏幕显示区确定单元将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第七种可能的实现方式,在第三方面的第八种可能的实现方式中,还包括:
第二统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示 区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为成像结果与预设人体图像匹配的目标屏幕显示区和人脸识别的结果与预设人脸一致的目标屏幕显示区;
第二计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第二确定模块,用于当所述比例大于第二预设比例阈值时,将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第九种可能的实现方式中,所述操作信息检测单元包括:
声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
若检测为是,则所述目标屏幕确定单元将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
若检测结果为与预设人脸一致,则所述第一屏幕显示区确定单元将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第十种可能的实现方式中,所述操作信息检测单元包括:
声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
若检测为是,则所述目标屏幕确定单元将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
所述体征信息检测单元还包括:
声音识别子单元,用于在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行声音识别,判断声音识别的结果是否与预设声音一致;
若检测结果为与预设声音一致,则所述第一屏幕显示区确定单元将声音识别的结果与预设声音一致的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第十种可能的实现方式,在第三方面的第十一种可能的实现方式中,还包括:
第三统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的 目标显示区和声音识别的结果是否与预设声音一致的目标屏幕显示区;
第三计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第三确定模块,用于当所述比例大于第三预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
结合第三方面的第二种可能的实现方式,在第三方面的第十二种可能的实现方式中,所述操作信息检测单元包括:
声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
若检测为是,则所述目标屏幕确定单元将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
所述体征信息检测单元包括:
人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
所述体征信息检测单元还包括:
虹膜识别子单元,用于在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别,判断虹膜识别的结果是否与预设虹膜一致;
若检测结果为与预设虹膜一致,则所述第一屏幕显示区确定单元将虹膜识别的结果与预设虹膜一致的目标屏幕显示区确定为第一屏幕显示区。
结合第三方面的第十二种可能的实现方式,在第三方面的第十三种可能的实现方式中,还包括:
第四统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和虹膜识别的结果与预设虹膜一致的目标屏幕显示区;
第四计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第四确定模块,用于当所述比例大于第四预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
本发明实施例第四方面公开了一种多屏移动终端的显示装置,包括:
屏幕检测模块,用于检测所述移动终端的是否有处于激活状态的屏幕显示区;
关闭数据获取模块,用于所述屏幕检测模块检测为是时,所述移动终端获取目标传感器上报的数据;所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制;
关闭模块,用于所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区 中确定第一屏幕关闭区,并关闭所述第一屏幕关闭区。
结合第四方面,在第一种可能的实现方式中,还包括:
关闭介入检测模块,用于在所述关闭所述第一屏幕关闭区之后,检测是否有用户介入操作;所述介入操作用于确定第二屏幕关闭区;
关闭介入模块,用于若所述关闭介入检测模块检测为是时,根据所述介入操作在所述处于激活状态的屏幕显示区中确定第二屏幕关闭区,重新激活所述第一屏幕关闭区,并关闭所述第二屏幕关闭区。
结合第四方面或第四方面第一种可能的实现方式,在第二种可能的实现方式中,所述关闭模块包括:
关闭操作检测单元,用于所述移动终端根据所述目标传感器上报的数据,检测所述目标处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息;
关闭单元,用于根据所述用户操作信息确定第一屏幕关闭区。
结合第四方面的第二种可能的实现方式,在第四方面的第三种可能的实现方式中,所述关闭操作检测单元包括:
动作子单元,用于检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
第二温度子单元,用于在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
第二测距子单元,用于在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
若检测结果为大于预设距离阈值,则所述关闭单元将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定第一屏幕关闭区。
结合第四方面的第三种可能的实现方式,在第四方面的第四种可能的实现方式中,所述关闭操作检测单元包括:
动作子单元,用于检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
第二温度子单元,用于在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
第二测距子单元,用于在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
第二人脸识别子单元,用于在测距结果大于预设距离阈值的处于激活状态的屏幕显示区的垂直方向上进行人脸识别,检测人脸识别结果是否与预设人脸一致;
若检测结果为与预设人脸一致,则所述关闭单元将与预设人脸一致的处于激活状态的屏幕显示区确定第一屏幕关闭区。
结合第四方面的第四种可能的实现方式,在第四方面的第五种可能的实现方式中,还包括:
第五统计模块,用于统计两个处于激活状态的屏幕显示区为同一屏幕显示区的次数以及所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数,其中,所述两个处于激活状态的屏幕显示区分别为测距结果大于预设距离阈值的处于激活状态的屏幕显示区和人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区;
第五计算模块,计算所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数与所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数之和;
第五确定模块,用于当所述比例大于第五预设比例阈值时,将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
本发明实施例第五方面公开了一种终端,包括:
存储器,用于存储显示指令;
处理器,用于调用所述存储器中的显示指令,所述显示指令应用于具有至少两个屏幕显示区的移动终端,所述处理器调用所述显示指令并执行上述第一方面提供的显示方法。
本发明实施例第六方面公开了一种终端,包括:
存储器,用于存储关闭指令;
处理器,用于调用所述存储器中的关闭指令,所述关闭指令应用于具有至少两个屏幕显示区的移动终端,所述处理器调用所述关闭指令并执行上述第二方面提供的显示方法。
本发明实施例第七方面公开了一种计算机可读的存储介质,用于存储一个或多个计算机程序,所述一个或多个计算机程序包括指令,当所述计算机程序在计算机上运行时,所述指令用于执行上述第一方面或第二方面提供的显示方法。
本发明实施例第八方面公开了一种计算机程序,所述计算机程序包括指令,当所述计算机程序在计算机上执行时,所述指令用于执行上述第一方面或第二方面提供的显示方法。
综上所述,通过实施本发明实施例,具有至少两个屏幕显示区的移动终端接收用户输入的开启指令,所述开启指令用于指示打开目标界面;所述移动终端接收目标传感器上报的数据;所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,并在所述第一屏幕显示区中呈现所述目标界面。能够准确识别用户的位置和使用状态,进行后续的显示界面、点亮屏幕以及打开多媒体等相关操作,方便了用户的使用。
在本发明实施例中,多屏移动终端根据自身配备的传感器类型,根据不同精度的显示方法决策选取相应的传感器,也可以由用户自主选择特定的传感器,所述传感器获取的外界数据之后将所述数据上报至移动终端,以确定用户的位置和使用状态,选取相应的屏幕显示区进行显示,可以充分结合移动终端的传感器类型的特性来进行判断,实现了智能地选取合适的屏幕对目标界面进行显示。其中,可以分多种精度的显示方法决策,每一种精度的显示方法决策都包含模糊判决和精确判决;当模糊判决和精确判决的结果保持一致的比例大于预设比例阈值的情况下,则认为模糊判决已经具有足够的精确程度,则可 将模糊判决的结果作为最终结果,无需再进行精确判决。
另外,当所述多屏移动终端根据传感器上报的数据在所述至少两个屏幕显示区中确定第一屏幕显示区,并在所述第一屏幕显示区中呈现所述目标界面之后,如果所述呈现的屏幕显示区不是用户所希望进行呈现的屏幕显示区,用户还可以进行介入操作进行纠正,手动选择想要进行显示的屏幕显示区。
附图说明
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍。
图1为几种不同种类屏幕显示区的移动终端的示意图;
图2为本发明实施例公开的一种终端的结构示意图;
图3为本发明实施例公开的各传感器与屏幕显示区的位置关系图;
图4为本发明实施例公开的一种显示方法的流程示意图;
图5为本发明实施例公开的一种传感器类型和显示决策的界面图;
图6为本发明另一实施例公开的一种显示方法的流程示意图;
图7a为本发明实施例公开的一种低精度显示方法决策的流程示意图;
图7b为本发明另一实施例公开的一种低精度显示方法决策的流程示意图;
图8a为本发明实施例公开的一种中精度显示方法决策的流程示意图;
图8b为本发明另一实施例公开的一种中精度显示方法决策的流程示意图;
图9a为本发明实施例公开的一种高精度显示方法决策的流程示意图;
图9b为本发明另一实施例公开的一种高精度显示方法决策的流程示意图;
图10a为本发明另一实施例公开的一种高精度显示方法决策的流程示意图;
图10b为本发明另一实施例公开的一种高精度显示方法决策的流程示意图;
图11为本发明另一实施例公开的一种显示方法的流程示意图;
图12a为本发明实施例公开的一种显示方法的决策流程示意图;
图12b为本发明另一实施例公开的一种显示方法的决策流程示意图;
图13为本发明实施例公开的一种多屏移动终端结构示意图;
图14为本发明实施例公开的一种显示装置的结构示意图;
图15为本发明实施例公开的一种显示屏幕确定模块的结构示意图;
图16为本发明另一实施例公开的一种显示装置的结构示意图;
图17为本发明实施例公开的一种关闭模块的结构示意图。
具体实施方式
下面将结合附图对本发明实施例中的技术方案进行清楚、详尽地描述。
如图1所示,是几种不同种类屏幕显示区的移动终端的示意图;其中,
a为双面屏设备,即具备两块屏幕的电子设备,包括但不限于为正反两块屏幕、保持一定夹角的两块屏幕;
b为柔性屏设备,即屏幕可被折叠成一定角度的电子设备;
c为多屏显示设备,即具备若干块显示屏幕的电子设备,且各个显示屏幕处于不同的平面;
d为多屏显示设备,即具备若干块显示屏幕的电子设备,且各个显示屏幕处于同一平面。
如图1,a和b所示的双面屏设备和柔性屏设备,在现有的技术中,可以通过重力感应的方式对用户的使用状态进行粗略的判断,并不能准确判断出用户正在使用的屏幕从而进行显示,且在c和d中多屏显示设备的情况下,如果只凭借重力感应的方式,难以判断用户的位置以及使用状态,使得所述多屏显示设备不能准确地感知用户正在使用的屏幕从而进行显示。
在本发明实施例中,通过移动终端配备的传感器,将获取的数据上报至所述移动终端,所述移动终端根据上报的数据,判断用户正在使用的显示屏,并进行相应的显示。
当用户在使用移动终端时,移动终端根据配备的传感器上传的数据,对用户的方位和使用状态进行判定,从而选取合适的屏幕显示区来显示目标界面。
其中,所述移动终端可以配备多种类型的传感器,包括但不限于感知设备姿态的传感器、用户动作及方位的传感器、用户生物信息识别的传感器。用户在使用移动终端的过程中,所述多屏移动终端根据具体的硬件传感器配备,选取相应的传感器根据显示方法决策获取数据,也可以由用户自主选择指定的传感器,并将所述获取的数据上报至移动终端,移动终端根据上报的数据做出判决,从而选取相应的屏幕显示区进行显示。其中,本发明实施例可以根据不同多屏移动终端可能的传感器配备,制定了低精度、中精度和高精度三种显示方法决策,并且每一种精度的显示方法决策中又包含了模糊判决和精确判决,满足了用户各种不同的使用场景。
需要说明的是,此处的双面屏设备、柔性屏设备和多屏显示设备的存在形态为一种具体的实施例,在实际应用中还可以有更多样的形态产品;其中,多屏显示设备的传感器配备也存在其他的组合方式,并不限制于实施例中列举的具体的传感器。
图2为本发明实施例公开的一种终端的结构示意图,如图2所示,终端100可以包括:基带芯片101、存储器105(一个或多个计算机可读存储介质)、射频模块111、外围系统106。这些部件可以在一个或多个通信总线112上通信。
外围系统106主要用于实现终端100和用户/外部环境之间的交互功能,主要包括终端100的输入输出装置。具体地,外围系统106可包括:屏幕显示区107、摄像头控制器108、音频控制器109以及传感器110。其中,屏幕显示区107可以包含多个屏幕显示区,且多个屏幕显示区可以不在同一个平面上,也可以在同一个平面上;传感器110可以包括多个种类的传感器,例如:陀螺仪/重力传感器1101、摄像头传感器1102、红外阵列传感器1103、超声波传感器1104、激光传感器1105、光/接近传感器1106、压力传感器1107、虹膜识别传感器1108、声音传感器1109等。
其中,陀螺仪/重力传感器1101可以用于测量屏幕方向,陀螺仪/重力传感器1101可以位于终端100的内部,陀螺仪通过测量终端100偏转、倾斜时的运转角速度来检测屏幕的方向,重力传感器通过重力值判断哪一面屏幕垂直向上来检测屏幕的方向。
摄像头传感器1102可以用于相机成像、人脸识别等,可以位于屏幕显示区107的左上角,通过将摄像头采集的图像与保存于存储器105中的目标图像(例如用户人脸图像等)相比较,进而实现人脸识别。
红外阵列传感器1103可以用于进行红外成像、温度探测以及动作识别等,可以位于屏幕显示区107的两侧,通过红外阵列传感器1103可以检测出目标物体热辐射的红外线特定波段信号,将该信号转换成可供人眼视觉分辩的图像或图形,并进一步计算出温度值。
超声波传感器1104可以用于超声波成像以及测距等,可以位于屏幕显示区107的下方,超声波传感器1104通过超声波在空气中传播的时间和声波的传播速度计算得到距离,由超声波传感器1104发射脉冲波束经聚焦后射线被测物体,控制超声波束的方向和聚焦的深度对物体进行三维扫描,从而得到物体的立体图像。
激光传感器1105、光/接近传感器1106可以用于测距,激光传感器1105可以位于屏幕显示区107的右下角,通过记录并处理从光脉冲发出到返回被接收所经历的时间,测定目标距离;光/接近传感器1106可以位于屏幕显示区107的右上角,通过感知周围光线的情况,可以通过发射特别短的光脉冲,并测量此光脉冲从发射到被物体反射回来的时间,通过测时间间隔来计算与物体之间的距离。
压力传感器1107可以用于测量触摸压力等,可以位于屏幕显示区107的表面,当待测物体受力发生应力变化时,与待测物体相连的电阻应变片也一起产生形变,使应变片的阻值发生改变,从而使加在电阻上的电压发生变化,得到压力值。
虹膜识别传感器1108可以用于虹膜识别,可以位于屏幕显示区107的上方,通过虹膜识别传感器1108可以采集用户眼球的生物特征并保存,在使用过程中将采集到的待检测虹膜进行识别,通过比对可以实现虹膜识别,在一种具体的实施例中可以通过红外LED光线进行照射,同时通过设备的前置摄像头实现采集。
声音传感器1109可以用于声音识别以及音源识别等,可以位于屏幕显示区107的左下角,通过在声音传感器1109内置一个对声音敏感的电容式驻极体话筒,声波使话筒内的驻极体薄膜振动,导致电容的变化,而产生与之对应变化的微小电压。
以双面屏为例,各传感器与屏幕显示区107的具体位置关系如图3所示,图3中A面屏和B面屏为正反两个屏幕显示区,且A和B背对向外进行显示。除陀螺仪/重力传感器1101位于终端100的内部,压力传感器1107位于屏幕显示区107的表面之外,其他传感器均可分别位于A面屏和B面屏的相同位置,图3中主要示出各传感器在A面屏上的位置。可以知道的是,各传感器与屏幕显示区107的具体位置关系在此只是一种具体的实施方式,不能完全限定各传感器与屏幕显示区的位置关系。
在另一些实施例中,外围系统106还可以包含其他外设。
基带芯片101可集成包括:一个或多个处理器102、时钟模块103以及电源管理模块104。集成于基带芯片101中的时钟模块103主要用于为处理器102产生数据传输和时序控制所需要的时钟。集成于基带芯片101中的电源管理模块104主要用于为处理器102、射频模块111以及外围系统106提供稳定的、高精确度的电压。
射频模块111用于接收和发送射频信号,主要集成了终端100的接收器和发射器。射频模块111通过射频信号与通信网络和其他通信设备通信。在具体的实现方式中,射频模块111可包括但不限于:天线系统、RF收发器、一个或多个放大器、调谐器、一个或多个振荡器、数字信号处理器、CODEC芯片、SIM卡和存储介质等。在一些实施例中,可在单独的芯片上实现射频模块111。
存储器105和处理器102耦合,用于存储各种软件程序或多组指令。具体地,存储器105可包括高速随机存取的存储器,并且可包括非易失性存储器,例如一个或多个磁盘存储设备、闪存设备或其他非易失性固态存储设备。存储器105可以存储操作系统(下述简称系统),例如ANDROID、IOS、WINDOWS,或者LINUX等嵌入式操作系统。存储器105还可以存储网络通信程序,该网络通信程序可用于与一个或多个附加设备,一个或多个终端设备,一个或多个网络设备进行通信。存储器105还可以存储用户接口程序,该用户接口程序可以通过图形化的操作界面将应用程序的内容形象逼真的显示出来,并通过菜 单、对话框以及按键等输入控件接收用户对应用程序的控制操作。
基于图1以及图2实施例描述的终端100,下面结合图4至图12来详细说明本发明实施例提供的显示方法的流程示意图。
请参见图4,图4为本发明实施例公开的一种显示方法的流程示意图,该方法包括但不限于以下步骤:
S101:移动终端接收用户输入的开启指令,所述开启指令用于指示打开目标界面。
具体地,所述开启指令用于指示打开目标界面,包括激活屏幕、进入软件应用、播放多媒体等操作。
S102:所述移动终端接收目标传感器上报的数据。
具体地,移动终端可以配备传感器等硬件,不同的移动终端所配备的传感器类型和数量都不一样,上报的数据类型和作用也不同。例如,陀螺仪采集到的数据是终端偏转或倾斜时的角速度,而压力传感器采集到的数据是用户触摸屏幕时产生的压力值。本发明实施例中的目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制。其中可以为移动终端根据自身支持的传感器类型通过预先设置的显示决策来自行选取目标传感器,也可以为用户获知移动终端所支持的传感器类型后,根据所需要的显示决策手动选取目标传感器。
需要说明的是,可以预先进行设置,选取不同的目标传感器对应不同的显示决策。
在一种可能的实现方式中,预先设置好了传感器类型A和传感器类型B对应第一显示决策,传感器类型A和传感器类型C对应第二显示决策,传感器类型A、传感器类型B和传感器类型C对应第三显示决策,等等。那么当移动终端支持传感器类型A、传感器类型B和传感器类型C的情况下,可以根据第一显示决策来选取传感器类型A、传感器类型B作为目标传感器,也可以根据第二显示决策来选取传感器类型A和传感器类型B作为目标传感器,还可以根据第三显示决策来选取传感器类型A、传感器类型B和传感器类型C作为目标传感器。不同传感器类型组合与显示决策的具体对应关系如表1所示。在具体的应用中,所述移动终端根据选取不同的传感器类型,对应不同的显示决策,也就是说通过调用不同的传感器得到不同类型的上报数据,从而对应做出不同的显示判断或显示分析。
表1不同传感器类型组合与显示决策对应关系表
选取的传感器类型 显示决策
A、B 第一显示决策
A、C 第二显示决策
A、B、C 第三显示决策
具体地,传感器类型A可以是设备姿态传感器,例如陀螺仪/重力传感器;传感器类型B可以是使用者动作/方位传感器,例如压力传感器;传感器类型C可以是使用者生物识别传感器,例如虹膜识别传感器。因此可以将设备姿态传感器和使用者动作/方位传感器对应第一显示决策;将设备姿态传感器和使用者生物识别传感器对应第二显示决策;将设备姿态传感器、使用者动作/方位传感器和使用者生物识别传感器对应第三显示决策,等等。
此外,移动终端根据选取不同的传感器,可能会对应多种显示决策,此时用户可以手动选择需要的显示决策,如图5所示:
假设终端支持的传感器类型包括A、B、C三种,当用户选取传感器类型A、B、C之后(左图中灰 色区域在左边代表开启,在右边代表关闭),对应了三种显示决策,即第一显示决策、第二显示决策和第三显示决策,此时用户可以点击显示屏的相应区域,进一步选择需要的显示决策(如右图)。当然,还可以是系统自动进行选择,默认第一显示决策为最终的显示决策。
在另外一种可能的实现方式中,选取机制还可以是用户根据终端本身支持的传感器类型自定义不同的显示决策对应的传感器类型。
当用户开启自定义显示决策模式之后,系统会分别在各种显示决策下面列出移动终端支持的不同传感器类型,用户可以手动设置该显示决策对应的不同的传感器类型,将灰色区域往左滑时代表开启对应的传感器类型,往右滑代表关闭对应的传感器类型。例如,用户可以手动设置第一显示决策对应的传感器类型是A和C;第二显示决策对应的传感器类型是B和C;第三显示决策对应的传感器类型是A和B,等等。
还需要说明的是,在移动终端选取了目标传感器后,移动终端可以触发或控制目标传感器获取数据,从而接收到该目标传感器上报的数据。
S103:所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,并在所述第一屏幕显示区中呈现所述目标界面。
具体地,所述移动终端根据支持的传感器类型选取目标传感器,并根据传感器上报的数据,通过自身的决策机制判断出当前使用者的使用状态和位置,确定与使用者的状态匹配的屏幕,并在匹配的屏幕中呈现所述打开的目标界面。
S104:检测是否有用户介入操作。
具体地,在移动终端通过判断呈现出了目标界面之后,有可能会由于判断不准确而导致没有在用户希望的屏幕上呈现所述目标界面,所以需要用户主动介入操作进行纠正,进行屏幕显示区的切换。
移动终端在接收到用户的介入操作之后,如S105所述,根据用户的介入操作确定第二屏幕显示区,将目标界面在用户指定的屏幕进行显示,同时撤回前述第一屏幕显示区中呈现的目标界面。若移动终端没有接收到用户的接入操作,说明所述移动终端根据传感器上报的数据对应该呈现的屏幕做出了准确的判断,根据S106所述,继续保持在第一屏幕显示区呈现所述目标界面。
可以知道的是,介入操作具体的实现方式例如可以但不限于是双击第二屏幕显示区、长按第二屏幕显示区,或者是唤醒第二屏幕显示区后终端提供是否切换至该屏幕显示区的按钮,用户点击确认进行屏幕切换等等。
图6为本发明另一实施例公开的一种显示方法的流程示意图,详细说明了图6的步骤S103中的“所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区”的情况,包括如下步骤:
S201:移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息。
具体地,所述移动终端具有至少两个屏幕显示区,移动终端所配备的传感器通过检测每一个屏幕显示区上的用户操作特征信息,并将获取的信息上报给终端。
其中,用户操作特征信息为用户在使用多屏移动终端时对所述移动终端产生的操作信息。在具体的实施例当中,用户操作特征信息可以包括用户对屏幕的触摸压力、在屏幕的垂直方向是否有人体温度、声音等。其中,用户对屏幕的触摸压力可以由压力传感器测得,在屏幕的垂直方向是否有人体温度可以 由红外阵列传感器测得,在屏幕的垂直方向是否有声音可以有声音传感器测得。
S202:若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区。
具体地,通过对传感器上报的用户操作特征信息进行处理和分析,所述移动终端对用户的方位和使用状态有了初级的评判,将可能存在用户操作的屏幕作为目标屏幕显示区。
S203:判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息。
具体地,在前述确定的目标屏幕显示区的垂直方向上进行检测,进一步判断是否存在用户体征信息,从而确定用户当前正在使用或将要使用的屏幕。
其中,用户体征信息为用户在使用多屏移动终端时所述移动终端检测到用户的体征信息。在具体的实施例当中,用户体征信息可以包括所述移动终端检测到的屏幕显示区垂直方向上的使用者与屏幕显示区之间的距离、使用者的面部特征是否与预设人脸匹配、使用者的虹膜是否与预设虹膜匹配等。其中,屏幕显示区垂直方向上的使用者与屏幕显示区之间的距离可以由激光传感器、光/接近传感器或者超声波传感器测得,使用者的面部特征可以由摄像头传感器测得,使用者的虹膜可以由虹膜识别传感器测得。
S204:根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区。
具体地,通过对传感器上报的用户体征信息进行分析和判定,在所述目标屏幕显示区中选定用户最有可能正在使用的屏幕,并确定为第一屏幕显示区,在所述第一屏幕显示区中呈现目标界面。
在多屏移动终端的具体使用当中,当移动终端的各个屏幕显示区处于不同的平面时,通过不同的传感器配备将所述显示方法分为了低、中、高三种精度的决策。下面通过图7a至图10b,结合具体实施例中调用的具体传感器,对所述显示方法做出更加详尽的描述。具体地,在本发明的实施例附图中,图7a、图8a描述的可以分别是低精度和中精度显示方法决策的判断流程,图9a和图10a描述的可以均为高精度显示方法决策的判断流程。
如图7a为本发明公开的一种低精度显示方法决策的流程示意图,包括以下步骤:
S301:移动终端接收目标传感器上报的数据。
具体地,如步骤S102所述,移动终端根据自身配备的传感器类型和数量指定相应的决策,调用相应的传感器进行数据采集并上报。
S302:检测至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力。
具体地,本发明实施例中的用户操作特征信息以检测所述移动终端的屏幕显示区上是否存在用户触摸压力为例进行说明。
调用压力传感器,检测所述移动终端的每一个屏幕显示区上是否存在有触摸压力。如果存在触摸压力,则判定可能为用户手持部位,手持部位的屏幕不会是用户的使用屏幕,故不将手持部位的屏幕显示区列入目标屏幕显示区。
具体地,若检测结果为有屏幕显示区存在触摸压力,则该屏幕显示区为用户手持部位,执行步骤S311,结束在该屏幕显示区的后续检测。
S303:在所述至少两个屏幕显示区中将不存在触摸压力的屏幕显示区确定为目标屏幕显示区。
具体地,若检测出有屏幕显示区不存在触摸压力,说明该屏幕显示区可能用户的使用屏幕,因此将其确定为目标屏幕显示区,目标屏幕显示区的个数至少可以为一个。
S304:判断所述目标屏幕显示区的垂直方向上的测距结果是否大于预设距离阈值。
具体地,本发明实施例中的用户体征信息为,检测所述移动终端的屏幕显示区的垂直方向上与用户之间的距离为例进行说明。
具体地,不存在触摸压力的屏幕可能为用户的使用屏幕,所以在不存在触摸压力的屏幕上进行进一步的检测,判断所述目标屏幕显示区的垂直方向上的测距结果是否大于预设距离阈值。调用超声波传感器、激光传感器或者光/接近传感器,测量目标屏幕显示区的垂直方向上的预设距离阈值内是否有人使用。预设距离阈值可以按照用户的习惯进行调整,即终端可以自动采集用户使用时离屏幕的垂直距离,根据用户平时的使用习惯智能设置预设距离阈值,以满足不同人群不同习惯的需求。若测距结果不大于预设距离阈值,则执行步骤S305;若测距结果为大于预设距离阈值,则执行步骤S312,结束在所述大于预设距离阈值的屏幕显示区上的后续检测。
S305:将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
具体地,根据所述测距结果,如果测距结果为不大于所述预设距离阈值,则可能在此屏幕的垂直方向上存在障碍物,很可能是用户正在使用。通过触摸压力和垂直测距两次检测,此时移动终端得出低精度条件下的模糊判决结果:将不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区,并将模糊判决的结果保存在终端的存储器中。
S306:检测所述测距结果不大于预设距离阈值的目标屏幕显示区垂直方向的成像结果是否与预设人体图像匹配。
具体地,为了得到更加精确的结果,在低精度模糊判决的条件下进行又一次的条件检测,在本发明实施例中,进一步检测的信息为检测所述测距结果不大于预设距离阈值的目标屏幕显示区垂直方向的成像结果是否与预设人体图像匹配,即检测所述移动终端的屏幕前的障碍物是否为人体。
调用红外传感器或者超声波传感器,在测距结果不大于预设距离阈值的目标屏幕显示区垂直方向进行成像。预设人体图像可以是从网络获取的任意的人体图像,还可以从本地网络获取的用户的人体图像,用户的人体图像可以是预先采集的,离线保存于终端存储器中,若成像结果与预设人体图像匹配,则执行步骤S307;若成像结果与预设人体图像不匹配,则执行步骤S313,结束在所述与预设人体图像不匹配的屏幕显示区的后续检测。
S307:将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
具体地,如果成像结果为与预设人体图像匹配,则移动终端做出低精度条件下的精确判决结果:将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
S308:统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数。
其中,两个目标屏幕显示区分别为测距结果不大于预设距离阈值的目标屏幕显示区和成像结果与预设人体图像匹配的目标屏幕显示区。
具体地,从存储器中调用模糊判决的结果,将所述低精度条件下的模糊判决结果和精确判决结果进行比较和学习,统计当前阶段内模糊判决结果和精确判决结果一致以及不一致的次数。其中,当前阶段为从第一次比较和学习开始,到此次比较和学习为止。模糊判决结果和精确判决结果保持一致即为执行步骤S305后得到的测距结果不大于预设距离阈值的目标屏幕显示区和步骤S307中成像结果与预设人体图像匹配的目标屏幕显示区为同一屏幕显示区;模糊判决结果和精确判决结果不一致即为执行步骤S305后得到的测距结果不大于预设距离阈值的目标屏幕显示区和步骤S307中成像结果与预设人体图像匹配的目标屏幕显示区为不同的屏幕显示区。
S309:计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例。
其中,总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和。
S310:当所述比例大于第一预设比例阈值时,将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
具体地,所述第一预设比例阈值可以由移动终端自动设定,也可以由用户自主设定。当比例大于第一预设比例阈值时,说明模糊判决的结果已经具有了一定的准确度,在后续的低精度显示方法决策中,直接以模糊判决的结果进行显示,故无需再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区,不执行所述在测距结果不大于预设距离阈值的目标屏幕显示区垂直方向的成像结果是否与预设人体图像匹配的步骤。当比例不大于第一预设比例阈值时,说明模糊判决的精确度还不够,结束当前的比较和学习,并将此次比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
例如,第一预设比例阈值为80%,当在当前的比较和学习的阶段内,统计的模糊判决结果和精确判决结果保持一致的次数为8次,两者不一致的次数为2次,即当前比较和学习的总次数为10次,那么在下一次的比较和学习中,若模糊判决结果和精确判决结果保持一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例为(8+1)/(10+1)=81.8%,大于第一预设比例阈值80%,则在后续的低精度显示方法决策中,直接以模糊判决的结果进行显示,不再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区,不执行所述在测距结果不大于预设距离阈值的目标屏幕显示区垂直方向的成像结果是否与预设人体图像匹配的步骤;若在下一次的比较和学习中,模糊判决结果和精确判决结果不一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例8/(10+1)=72.7%,不大于第一预设比例阈值80%,则结束当前的比较和学习,并将比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
在经过一段时间的比较和学习,得出模糊判决的结果精确度足够,并在后续检测中直接以模糊判决的结果进行显示,不再进行精确判决之后,在接下来的低精度显示方法决策中,至少可以包括以下几个步骤,如图7b所示:
执行S314-S318得出模糊判决的结果之后,执行S319:判断是否有介入操作,若否,则保持模糊判决结果不变,以模糊判决结果进行显示;若是,则执行S320:根据介入操作确定第二屏幕显示区。
具体地,若接收到用户的介入操作进行屏幕显示区的切换,说明模糊判决的结果不准确导致用户手动设置显示的屏幕,根据用户的介入操作确定第二屏幕显示区,将目标界面在第二屏幕显示区中进行显示,同时撤回在第一屏幕显示区中呈现的目标界面。且在下一次的低精度显示方法决策中,重新执行S301-S313,进行模糊判决和精确判决结果的比较和学习。
图8a为本发明实施例公开的一种中精度显示方法决策的流程示意图,包括以下步骤:
S401:移动终端接收目标传感器上报的数据。
具体地,如步骤S102所述,移动终端根据自身配备的传感器类型和数量指定相应的决策,调用相 应的传感器进行数据采集并上报。
S402:检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度。
具体地,本发明实施例中的用户操作特征信息以检测所述移动终端的屏幕显示区上是否存在与人体相匹配的温度为例进行说明。
调用红外阵列传感器,检测所述移动终端的每一个屏幕显示区垂直方向是否与预设温度区间相匹配的温度。具体地,所述预设温度区间为人体的正常温度区间,如果检测结果为存在与人体正常温度区间相匹配的温度,则证明屏幕显示区前端有用户正在使用。
具体地,若检测结果为不存在与人体正常温度区间相匹配的温度,则执行步骤S411,结束在该屏幕显示区的后续检测。
S403:在所述至少两个屏幕显示区中将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区。
具体地,将可能存在用户使用的屏幕显示区作为目标屏幕显示区,目标屏幕显示区的个数至少可以为一个。
S404:检测所述目标屏幕显示区垂直方向上的成像结果是否与预设人体图像匹配。
具体地,本发明实施例中的用户体征信息,以检测所述移动终端的屏幕显示区的垂直方向上的成像结果是否与预设人体图像匹配为例进行说明。
具体的检测过程与步骤S306相似,调用红外传感器或者超声波传感器,在成像结果与预设人体图像不匹配的目标屏幕显示区垂直方向进行成像。预设人体图像可以是从网络获取的任意的人体图像,还可以从本地网络获取的用户的人体图像,用户的人体图像可以是预先采集的,离线保存于终端存储器中。若成像结果与预设人体图像匹配,则执行步骤S405;若成像结果与预设人体图像不匹配,则执行步骤S412,结束在所述与预设人体图像不匹配的屏幕显示区的后续检测。
S405:将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
具体地,根据成像结果,如果成像结果为与预设人体图像匹配,则可能在此屏幕显示区的垂直方向存在人体,很有可能是用户正在使用。通过温度和成像两次检测,此时移动终端得出中精度条件下的模糊判决结果:将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区,并将模糊判决的结果保存在终端的存储器中。
S406:检测成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上的人脸识别结果是否与预设人脸一致。
具体地,在本发明实施例中,进行中精度条件下的精确判决的进一步检测的信息为,所述移动终端的屏幕显示区垂直方向上的人脸识别结果是否与预设人脸一致;所述预设人脸为预先存储在移动终端内的用户的面部图像,执行此次判断是为了识别使用该移动终端的用户是否为移动终端的用户本人。
调用摄像头传感器在成像方向进行人脸识别,识别是否为用户。用户的面部图像可以预先采集保存于终端的存储器中,将摄像头传感器采集到的面部图像与用户的面部图像进行比对,若比对结果相符,则说明使用该移动终端的用户为移动终端的用户本人,执行步骤S407;若比对结果不相符,则说明使用该移动终端的用户不是移动终端的用户本人,则执行步骤S413,结束所述不与预设人脸一致的屏幕显示区的后续检测。
S407:将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
具体地,如果人脸识别的结果为使用该移动终端的用户为用户本人,则所述移动终端得出中精度条件下的精确判决结过:将该目标屏幕显示区确定为第一屏幕显示区。
S408:统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数。
其中,所述两个目标屏幕显示区分别为成像结果与预设人体图像匹配的目标屏幕显示区和人脸识别的结果与预设人脸一致的目标屏幕显示区。
具体地,从存储器中调用模糊判决的结果,将中精度条件下的模糊判决结果和精确判决结果进行比较和学习,统计当前阶段内模糊判决结果和精确判决结果一致以及不一致的次数。其中,当前阶段为从第一次比较和学习开始,到此次比较和学习为止。模糊判决结果和精确判决结果保持一致即为步骤S405中得到的成像结果与预设人体图像匹配的目标屏幕显示区和步骤S407中人脸识别的结果与预设人脸一致的目标屏幕显示区为同一屏幕显示区;模糊判决结果和精确判决结果不一致即为步骤S405中得到的成像结果与预设人体图像匹配的目标屏幕显示区和步骤S407中人脸识别的结果与预设人脸一致的目标屏幕显示区为不同的屏幕显示区。
S409:计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例。
其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和。
S410:当所述比例大于第二预设比例阈值时,将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
具体地,所述第二预设比例阈值可以由移动终端自动设定,也可以由用户自主设定。当比例大于第二预设比例阈值时,说明模糊判决的结果已经具有了一定的准确度,在后续的中精度显示方法决策中,直接以模糊判决的结果进行显示,故无需再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区,不执行所述在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别的步骤。当比例不大于第二预设比例阈值时,说明模糊判决的精确度还不够,结束当前的比较和学习,并将此次比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
例如,第二预设比例阈值为80%,当在当前的比较和学习的阶段内,统计的模糊判决结果和精确判决结果保持一致的次数为8次,两者不一致的次数为2次,即当前比较和学习的总次数为10次,那么在下一次的比较和学习中,若模糊判决结果和精确判决结果保持一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例为(8+1)/(10+1)=81.8%,大于第二预设比例阈值80%,则在后续的中精度显示方法决策中,直接以模糊判决的结果进行显示,不再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区,不执行所述在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别的步骤;若在下一次的比较和学习中,模糊判决结果和精确判决结果不一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例8/(10+1)=72.7%,不大于第二预设比例阈值80%,则结束当前的比较和学习,并将比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
在经过一段时间的比较和学习,得出模糊判决的结果精确度足够,并在后续检测中直接以模糊判决的结果进行显示,不再进行精确判决之后,在接下来的中精度显示方法决策中,至少可以包括以下几个步骤,如图8b所示:
执行S414-S418得出模糊判决的结果之后,执行S419:判断是否有介入操作,若否,则保持模糊判决结果不变,以模糊判决结果进行显示;若是,则执行S420:根据介入操作确定第二屏幕显示区。
具体地,若接收到用户的介入操作进行屏幕显示区的切换,说明模糊判决的结果不准确导致用户手动设置显示的屏幕,根据用户的介入操作确定第二屏幕显示区,将目标界面在第二屏幕显示区中进行显示,同时撤回在第一屏幕显示区中呈现的目标界面。且在下一次的中精度显示方法决策中,重新执行S401-S413,进行模糊判决和精确判决结果的比较和学习。
图9a为本发明实施例公开的一种高精度显示方法决策的流程示意图,包括以下步骤:
S501:移动终端接收目标传感器上报的数据。
具体地,如步骤S102所述,移动终端根据自身配备的传感器类型和数量指定相应的决策,调用相应的传感器进行数据采集并上报。
S502:检测所述至少两个屏幕显示区中的每一个屏幕显示区的方向上是否存在特定频率范围的声音。
具体地,本发明实施例中的用户操作特征信息以检测所述移动终端的屏幕显示区上是否存在与人声频率相匹配的声音为例进行说明。
调用声音传感器,检测所述移动终端的每一个屏幕显示区的上方是否存在特定频率范围的声音。具体地,所述特定频率范围的声音为与人声频率相匹配的频率区间范围。该检测步骤是为了确认该屏幕显示区前是否有用户使用。
具体地,若检测结果为不存在特定频率范围的声音,则证明该屏幕显示区垂直方向上没有用户使用,执行步骤S511,结束在该屏幕显示区上的后续检测。
S503:在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区。
具体地,若检测结果为存在特定频率范围的声音,则证明该屏幕显示区垂直方向上有用户正在使用,将该屏幕显示区确定为目标屏幕显示区,目标屏幕显示区的个数至少可以为一个。
S504:检测所述目标屏幕显示区垂直方向上的人脸识别的结果是否与预设人脸一致。
具体地,本发明实施例中的用户体征信息,以检测所述移动终端的屏幕显示区的垂直方向上的人脸识别的结果是否与预设人脸一致为例进行说明。
具体的检测过程与步骤S406相似,调用摄像头传感器在成像方向进行人脸识别,识别是否为用户。用户的面部图像可以预先采集保存于终端的存储器中,将摄像头传感器采集到的面部图像与用户的面部图像进行比对,若比对结果为一致,则执行步骤S505;若比对结果不一致,则执行步骤S512,结束在该目标屏幕显示区上的后续检测。
S505:将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
具体地,若检测结果为与预设人脸一致,则所述移动终端做出高精度条件下的模糊判决:将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区,并将模糊判决的结果保存在终端的存储器中。
S506:判断所述人脸识别的结果与预设人脸一致的目标屏幕显示区垂直方向上的声音识别的结果是否与预设声音一致。
具体地,为了得到更加精确的结果,在高精度条件下进行又一次的条件检测,在本发明实施例中,进一步检测的信息为,检测所述移动终端的屏幕显示区垂直方向上的声音识别结果是否与预设声音一致;所述预设声音为预先存储在所述移动终端中的用户的声音,该检测步骤是为了检测目标屏幕显示区上的使用用户是否为移动终端的用户本人。
调用声音传感器,在人脸识别的结果与预设人脸一致的屏幕显示区的垂直方向上进行声音识别,识别是否与用户的声音一致。若识别结果为与用户声音一致,则执行步骤S507;若识别结果为与用户声音不一致,则执行步骤S513,结束在所述不与预设声音一致的屏幕显示区的后续检测。
S507:将声音识别的结果与预设声音一致的目标屏幕显示区确定为第一屏幕显示区。
具体地,若声音识别的结果与预设声音一致,则说明目标屏幕显示区前使用所述移动终端的用户为用户本人,则所述移动终端做出高精度条件下的精确判决:将该目标屏幕显示区作为第一屏幕显示区。
S508:统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数。
其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和声音识别的结果是否与预设声音一致的目标屏幕显示区。
具体地,从存储器中调用模糊判决的结果,将高精度条件下的模糊判决结果和精确判决结果进行比较和学习,统计当前阶段内模糊判决结果和精确判决结果一致以及不一致的次数。其中,当前阶段为从第一次比较和学习开始,到此次比较和学习为止。模糊判决结果和精确判决结果保持一致即为步骤S505中得到的人脸识别的结果与预设人脸一致的目标屏幕显示区和步骤S507中声音识别的结果与预设声音一致的目标屏幕显示区为同一屏幕显示区;模糊判决结果和精确判决结果不一致即为步骤S505中得到的人脸识别的结果与预设人脸一致的目标屏幕显示区和步骤S507中声音识别的结果与预设声音一致的目标屏幕显示区为不同的屏幕显示区。
S509:计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例。
其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和。
S510:当所述比例大于第三预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
具体地,所述第三预设比例阈值可以由移动终端自动设定,也可以由用户自主设定。当比例大于第三预设比例阈值时,说明模糊判决的结果已经具有了一定的准确度,在后续的高精度显示方法决策中,直接以模糊判决的结果进行显示,故无需再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区,不执行所述在人脸识别的结果与预设人脸一致的目标屏幕显示区垂直方向上进行声音识别的步骤。当比例不大于第三预设比例阈值时,说明模糊判决的精确度还不够,结束当前的比较和学习,并将此次比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
例如,第三预设比例阈值为80%,当在当前的比较和学习的阶段内,统计的模糊判决结果和精确判决结果保持一致的次数为8次,两者不一致的次数为2次,即当前比较和学习的总次数为10次,那么 在下一次的比较和学习中,若模糊判决结果和精确判决结果保持一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例为(8+1)/(10+1)=81.8%,大于第三预设比例阈值80%,则在后续的高精度显示方法决策中,直接以模糊判决的结果进行显示,不再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区,不执行所述在人脸识别的结果与预设人脸一致的目标屏幕显示区垂直方向上进行声音识别的步骤;若在下一次的比较和学习中,模糊判决结果和精确判决结果不一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例8/(10+1)=72.7%,不大于第三预设比例阈值80%,则结束当前的比较和学习,并将比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
在经过一段时间的比较和学习,得出模糊判决的结果精确度足够,并在后续检测中直接以模糊判决的结果进行显示,不再进行精确判决之后,在接下来的高精度显示方法决策中,至少可以包括以下几个步骤,如图9b所示:
执行S514-S518得出模糊判决的结果之后,执行S519:判断是否有介入操作,若否,则保持模糊判决结果不变,以模糊判决结果进行显示;若是,则执行S520:根据介入操作确定第二屏幕显示区。
具体地,若接收到用户的介入操作进行屏幕显示区的切换,说明模糊判决的结果不准确导致用户手动设置显示的屏幕,根据用户的介入操作确定第二屏幕显示区,将目标界面在第二屏幕显示区中进行显示,同时撤回在第一屏幕显示区中呈现的目标界面。且在下一次的高精度显示方法决策中,重新执行S501-S513,进行模糊判决和精确判决结果的比较和学习。
图10a为本发明另一实施例公开的一种高精度显示方法决策流程示意图,其中,所述步骤S601至S605的检测流程和方法均与对应的S501至S505相同。当所述移动终端做出高精度条件下的模糊判决之后,执行步骤S606:判断所述人脸识别的结果与预设人脸一致的目标屏幕显示区垂直方向上的虹膜识别的结果是否与预设虹膜一致。
具体地,在本发明实施例中,进行高精度条件下的精确判决的进一步检测的信息为,所述移动终端的屏幕显示区垂直方向上的虹膜识别结果是否与预设虹膜一致;所述预设虹膜为预先存储在所述移动终端中的用户的虹膜,与前述步骤S506相应,该检测步骤是为了检测目标屏幕显示区上的使用用户是否为用户。
调用虹膜识别传感器,在人脸识别的结果与预设人脸一致的目标屏幕显示区垂直方向上进行虹膜识别,识别是否与用户的虹膜一致。若识别结果为一致,则执行步骤S607;若识别结果为不一致,则证明该屏幕显示区的垂直方向上的使用用户不为用户,执行步骤S613,结束在该屏幕显示区上的后续检测。
S607:将虹膜识别的结果与预设虹膜一致的预设目标屏幕显示区确定为第一屏幕显示区。
具体地,若虹膜识别的结果与预设虹膜一致,则说明目标屏幕显示区前使用所述移动终端的用户为用户本人,则所述移动终端做出高精度条件下的精确判决:将该目标屏幕显示区作为第一屏幕显示区。
S608:统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数。
其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和虹膜识别的结果与预设虹膜一致的目标屏幕显示区。
具体地,从存储器中调用模糊判决的结果,将高精度条件下的模糊判决结果和精确判决结果进行比 较和学习,统计当前阶段内模糊判决结果和精确判决结果一致以及不一致的次数。其中,当前阶段为从第一次比较和学习开始,到此次比较和学习为止。模糊判决结果和精确判决结果保持一致即为步骤S605中得到的人脸识别的结果与预设人脸一致的目标屏幕显示区和步骤S607中虹膜识别的结果与预设虹膜一致的目标屏幕显示区为同一屏幕显示区;模糊判决结果和精确判决结果不一致即为步骤S605中得到的人脸识别的结果与预设人脸一致的目标屏幕显示区和步骤S607中虹膜识别的结果与预设虹膜一致的目标屏幕显示区为不同的屏幕显示区。
S609:计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例。
其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和。
S610:当所述比例大于第四预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
具体地,所述第四预设比例阈值可以由移动终端自动设定,也可以由用户自主设定。当比例大于第四预设比例阈值时,说明模糊判决的结果已经具有了一定的准确度,在后续的高精度显示方法决策中,直接以模糊判决的结果进行显示,故无需再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区,不执行所述在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别的步骤。当比例不大于第四预设比例阈值时,说明模糊判决的精确度还不够,结束当前的比较和学习,并将此次比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
例如,第四预设比例阈值为80%,当在当前的比较和学习的阶段内,统计的模糊判决结果和精确判决结果保持一致的次数为8次,两者不一致的次数为2次,即当前比较和学习的总次数为10次,那么在下一次的比较和学习中,若模糊判决结果和精确判决结果保持一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例为(8+1)/(10+1)=81.8%,大于第四预设比例阈值80%,则在后续的高精度显示方法决策中,直接以模糊判决的结果进行显示,不再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区,不执行所述在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别的步骤;若在下一次的比较和学习中,模糊判决结果和精确判决结果不一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例8/(10+1)=72.7%,不大于第四预设比例阈值80%,则结束当前的比较和学习,并将比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
在经过一段时间的比较和学习,得出模糊判决的结果精确度足够,并在后续检测中直接以模糊判决的结果进行显示,不再进行精确判决之后,在接下来的高精度显示方法决策中,至少可以包括以下几个步骤,如图10b所示:
执行S614-S618得出模糊判决的结果之后,执行S619:判断是否有介入操作,若否,则保持模糊判决结果不变,以模糊判决结果进行显示;若是,则执行S620:根据介入操作确定第二屏幕显示区。
具体地,若接收到用户的介入操作进行屏幕显示区的切换,说明模糊判决的结果不准确导致用户手动设置显示的屏幕,根据用户的介入操作确定第二屏幕显示区,将目标界面在第二屏幕显示区中进行显 示,同时撤回在第一屏幕显示区中呈现的目标界面。且在下一次的高精度显示方法决策中,重新执行S601-S613,进行模糊判决和精确判决结果的比较和学习。
在另外一种可能的实现方式中,在上述低精度显示方法决策、中精度显示方法决策以及高精度显示方法决策实施例的基础上,可以首先采用陀螺仪/重力传感器计算各屏幕的俯仰角坐标,从而得知终端自身的姿态位置,根据终端的姿态位置确定哪些屏幕可能为用户正在使用的屏幕,再基于预选的显示方法决策,调用这些可能正在使用的屏幕上相应的传感器采集数据,进行判断,此处无需调用其他不可能正在使用的屏幕上相应的传感器采集数据,降低功耗。
如图7a至图10b所述的实施例所述,给出了在具体的应用场景中,多屏移动终端调用不同的传感器进行低精度、中精度和高精度的显示方法决策,保证了判断的精度,满足了具有不同的传感器配备的移动终端的需求,也给用户手动选择指定的传感器提供了更多的选择。
在多屏移动终端的具体使用当中,当移动终端的各个屏幕显示区处于同一的平面时,可以通过摄像头传感器采集人脸信息;通过红外阵列传感器进行成像;通过超声波传感器、激光传感器或者光/接近传感器测量人脸与各屏幕显示区之间距离等来确定用户正在使用的屏幕显示区。
具体地,可以通过摄像头传感器采集人脸信息,根据人脸信息分析人脸可能朝向的屏幕显示区,进而确定用户正在使用的屏幕显示区;可以通过红外阵列传感器得到的成像结果确定用户正在使用的屏幕显示区;可以通过超声波传感器、激光传感器或者光/接近传感器测量人脸与各屏幕显示区之间距离,比较人脸与各屏幕显示区之间距离的大小,确定距离最小的屏幕显示区为用户正在使用的屏幕显示区。
如图1d所示,多屏显示终端包括A、B、C、D四个屏幕显示区(A面屏、B面屏、C面屏和D面屏),且这四个屏幕显示区处于同一平面。假设用户正在使用该多屏显示终端,当通过摄像头传感器检测出人脸的朝向可能朝向屏幕显示区A时,将屏幕显示区A确定为用户正在使用的屏幕,激活屏幕显示区A进行显示。
当通过摄像头传感器检测出人脸的朝向可能朝向屏幕显示区A和C时,可以同时将屏幕显示区A和C激活进行显示,也可以进一步采用超声波传感器、激光传感器或者光/接近传感器测量人脸与屏幕显示区A和C之间距离,激活距离较小的屏幕显示区进行显示。需要说明的是,是否进一步检测取决于终端支持的且用户选取的传感器类型来决定。
在另外一种可能的实现方式中,还可以用其他传感器来确定用户正在使用的屏幕显示区。
可以知道的是,用户可以自己设置哪些传感器对应哪种显示方法决策。例如,用户可以设置摄像头传感器和红外阵列传感器对应第一显示方法决策;摄像头传感器和激光传感器对应第二显示方法决策;摄像头传感器、红外阵列传感器以及超声波传感器对应第三显示方法决策,等等。
图11为本发明另一实施例公开的一种多屏移动终端的显示方法的流程示意图,描述了所述多屏移动终端的屏幕显示区处于激活状态时,通过自身的传感器判断是否要对激活的屏幕进行关闭的情况,是移动终端的显示方法决策的另一种实施方式。如图11所示,为移动终端的屏幕显示区处于激活状态时,通过传感器上传的数据进行屏幕关闭的方法流程,包括以下步骤:
S701:检测所述移动终端是否有处于激活状态的屏幕显示区。
S702:若检测为是,则所述移动终端接收目标传感器上报的数据。
具体地,所述传感器为移动终端的硬件配备,不同的移动终端所配备的传感器类型和数量都不一样,上报的数据类型和作用也不同。在具体的应用中,所述移动终端根据自身的传感器配备,制定不同的决策,调用不同的传感器从而得到不同类型的上报数据做出显示判断。其中,调用的传感器类型和相应的 显示决策可以由移动终端根据实际情况自行选择,也可以由用户手动选择。
需要说明的是,所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制。
S703:所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,并关闭所述第一屏幕关闭区。
具体地,所述移动终端根据传感器上报的数据,通过自身的决策机制判断出当前处于激活状态的屏幕显示区是否还有用户正在使用,判断是否需要将当前屏幕显示区关闭,并将需要关闭的屏幕显示区关闭。
S704:判断是否有用户介入操作。
具体地,当所述移动终端经过决策判断关闭了所述移动终端的第一屏幕关闭区之后,可能由于判断的准确性不够与用户想要关闭的屏幕区不同,这时由用户介入操作,以对关闭的行为进行纠正。如S705所述,根据所述用户介入操作在所述处于激活状态的屏幕显示区中确定第二屏幕关闭区,重新激活所述第一屏幕关闭区,并同时关闭所述第二屏幕关闭区;若在所述移动终端关闭第一屏幕关闭区之后没有用户介入操作,则如步骤S706所述:保持关闭所述第一屏幕关闭区。
可以知道的是,介入操作具体的实现方式例如可以但不限于是双击第二屏幕关闭区、长按第二屏幕关闭区,或者是在第二屏幕关闭区中显示是否关闭该屏幕的按钮,用户点击确认进行关闭等等。
图12a为本发明实施例公开的一种显示方法的决策流程示意图,结合了具体调用的传感器,对关闭流程做出了更加详尽的描述;包括以下步骤:
S801:检测移动终端是否有处于激活状态的屏幕显示区。
具体地,在本发明实施例中,所述移动终端接收目标传感器上报的数据具体为检测所述移动终端的屏幕显示区的方向上是否存在动作变化、是否存在与人体温度相匹配的温度、测距结果是否大于预设距离阈值以及人脸识别的结果是否与预设人脸一致。具体的检测步骤如下所示。
S802:检测是否有动作变化。
具体地,调用红外阵列传感器,在所述处于激活状态的屏幕显示区的方向上检测是否存在动作变化,如果不存在动作变化,则证明该屏幕显示区上的用户一直在使用,执行步骤S811,结束后续检测,保持屏幕的显示不变。如果检测结果为存在动作变化,则证明该屏幕显示区上的用户可能会离开该屏幕显示区,故继续进行后续检测,执行步骤S803。
S803:检测是否匹配温度。
具体地,调用红外阵列传感器,在存在动作变化的屏幕显示区的方向上进行温度检测,检测所述屏幕显示区方向上是否存在与人体温度区间匹配的温度;如果检测结果为存在与人体温度区间匹配的温度,则证明该屏幕显示区方向上有用户正在使用,执行步骤S812,结束后续检测,保持显示不变。如果检测结果为不存在与人体温度区间相匹配的温度,说明该屏幕显示区方向上的用户可能离开,故继续进行后续检测,执行步骤S804。
S804:检测是否大于预设距离阈值。
具体地,调用激光传感器、超声波传感器或者光/接近传感器在不存在与人体温度区间匹配的温度的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值,若检测结果为不大于,则证明所述屏幕显示区的垂直方向上有用户正在使用,执行步骤S813,结束后续检测,保持屏幕显示不变; 若检测结果为大于预设距离阈值,则证明所述屏幕显示区前的用户可能已经离开,所述移动终端根据步骤S805做出模糊关闭决策:将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区,并将模糊判决的结果保存在终端的存储器中。
S806:检测是否与预设人脸一致。
具体地,为了保证判决的结果更加精确,在所述移动终端做出模糊关闭决策之后,再进行人脸识别的检测,具体地,调用摄像头传感器在所述测距结果为大于预设距离阈值的屏幕显示区的方向上进行人脸识别,检测所述人脸识别的结果是否与预设人脸一致;其中,预设人脸为预先存储在所述移动终端中的用户的人脸图像。若检测结果为与预设人脸一致,则说明该屏幕显示区的方向上有用户正在使用,且该用户为用户本人,执行步骤S814,结束后续检测,保持屏幕显示不变,如果检测结果为与预设人脸不一致,则证明用户已经离开该屏幕显示区,或者该屏幕显示区方向上的使用者不是用户本人,则根据步骤S807,做出精确关闭决策:将人脸识别结果与预设人脸一致的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
S808:统计两个处于激活状态的屏幕显示区为同一屏幕显示区的次数以及所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数。
其中,所述两个处于激活状态的屏幕显示区分别为测距结果大于预设距离阈值的处于激活状态的屏幕显示区和人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区。
具体地,从存储器中调用模糊判决的结果,将模糊判决结果和精确判决结果进行比较和学习,统计当前阶段内模糊判决结果和精确判决结果一致以及不一致的次数。其中,当前阶段为从第一次比较和学习开始,到此次比较和学习为止。模糊判决结果和精确判决结果保持一致即为步骤S804中得到的测距结果大于预设距离阈值的处于激活状态的屏幕显示区和步骤S806中人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区为同一屏幕显示区;模糊判决结果和精确判决结果不一致即为步骤S804中得到的测距结果大于预设距离阈值的处于激活状态的屏幕显示区和步骤S806中人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区为不同的屏幕显示区。
S809:计算所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例。
其中,所述总的统计次数为所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数与所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数之和。
S810:当所述比例大于第五预设比例阈值时,将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
具体地,所述第五预设比例阈值可以由移动终端自动设定,也可以由用户自主设定。当比例大于第五预设比例阈值时,说明模糊判决的结果已经具有了一定的准确度,在后续的显示方法决策中,直接以模糊判决的结果进行显示,故无需再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区,不执行所述检测是否与预设人脸一致的步骤。当比例不大于第五预设比例阈值时,说明模糊判决的精确度还不够,结束当前的比较和学习,并将此次比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
例如,第五预设比例阈值为80%,当在当前的比较和学习的阶段内,统计的模糊判决结果和精确判决结果保持一致的次数为8次,两者不一致的次数为2次,即当前比较和学习的总次数为10次,那么 在下一次的比较和学习中,若模糊判决结果和精确判决结果保持一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例为(8+1)/(10+1)=81.8%,大于第五预设比例阈值80%,则在后续的显示方法决策中,直接以模糊判决的结果进行显示,不再进行精确决策的再一次条件检测,即在得出模糊判决结果之后,直接将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区,不执行所述检测是否与预设人脸一致的步骤;若在下一次的比较和学习中,模糊判决结果和精确判决结果不一致,那么此时模糊判决的结果和精确判决的结果保持一致的次数占比较和学习的总次数的比例8/(10+1)=72.7%,不大于第五预设比例阈值80%,则结束当前的比较和学习,并将比较和学习的结果保存至存储器,以备下一次比较和学习时从存储器调用当前阶段比较和学习的结果来计算模糊判决的结果和精确判决的结果两者保持一致的次数占比较和学习的总次数的比例。
在经过一段时间的比较和学习,得出模糊判决的结果精确度足够,并在后续检测中直接以模糊判决的结果进行显示,不再进行精确判决之后,在接下来的显示方法决策中,至少可以包括以下几个步骤,如图12b所示:
执行S815-S819得出模糊判决的结果之后,执行S820:判断是否有介入操作,若否,则保持模糊判决结果不变,以模糊判决结果进行显示;若是,则执行S821:根据介入操作确定第二屏幕显示区。
具体地,若接收到用户的介入操作进行屏幕显示区的切换,说明模糊判决的结果不准确导致用户手动设置显示的屏幕,根据用户的介入操作确定第二屏幕显示区,将目标界面在第二屏幕显示区中进行显示,同时撤回在第一屏幕显示区中呈现的目标界面。且在下一次的显示方法决策中,重新执行S801-S814,进行模糊判决和精确判决结果的比较和学习。
需要说明的是,该方法流程中采取的检测手段只是在实施过程中的一种具体实施例,在实际的实施过程中,所述移动终端根据自身的硬件配备,可以调用其他的传感器进行检测判断,此处的检测流程仅为方法描述,在此不做任何限定。
鉴于上述实施例的描述,本发明实施例还公开了一种多屏移动终端,如图13所示,多屏移动终端至少可以包括:传感器环境数据决策模块、综合决策模块以及用户辅助决策模块;其中:
传感器环境数据决策模块可以用于多传感器数据分类处理与上报。
具体地,传感器可以包括陀螺仪/重力传感器、激光传感器、摄像头传感器、红外阵列传感器、超声波传感器、光/接近传感器、压力传感器、虹膜识别传感器以及声音传感器等待。
综合决策模块可以用于根据多传感器上传的数据,综合分析确定第一屏幕显示区或者第一屏幕关闭区。
用户辅助决策模块可以用于用户手动设置或者介入操作。
具体地,用户辅助决策模块可以为用户提供多传感器智能显示开关给用户,用户通过开关决定启用哪些传感器收集数据;用户辅助决策模块还可以为用户提供屏显个数控制开关给用户选择,用户可以选择具体启用哪几个屏幕使用;用户辅助决策模块还可以提供用户介入决策功能给用户进行判决,即在经过模糊判决和精确判决的结果的比较和学习,直接以模糊判决的结果进行显示,不再进行精确判决之后,若模糊判决结果有误,用户可以介入操作来切换屏幕。
在一种可能的实现方式中,将不同的传感器进行组合,实现不同精度的决策,包括:低精度决策、中精度决策以及高精度决策,每种决策又可以分为两个阶段的判决,包括:模糊判决和精确判决。终端可以根据不同的决策智能判决输出,同时还可以结合用户的介入操作综合决策。
图14为本发明实施例公开的一种显示装置的结构示意图;如图14所示,显示装置900至少可以包括:开启模块901、显示数据接收模块902、显示屏幕确定模块903、显示模块904、显示介入操作检测模块905、以及介入显示模块906,其中:
开启模块901,用于移动终端接收用户输入的开启指令,所述开启指令用于指示打开目标界面。
显示数据接收模块902,用于所述移动终端接收目标传感器上报的数据。
显示屏幕确定模块903,用于所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区。
显示模块904,用于在所述第一屏幕显示区中呈现所述目标界面。
显示介入操作检测模块905,用于所述在所述第一屏幕显示区中呈现所述第一应用的界面之后,检测是否有用户介入操作;所述接入操作用于指示确定第二屏幕显示区。
介入显示模块906,用于根据所述用户介入操作在所述至少两个屏幕显示区中确定第二屏幕显示区,撤回所述第一屏幕显示区中呈现的目标界面,并同时在所述第二屏幕显示区中呈现所述目标界面。
图15为本发明实施例公开的一种显示屏幕确定模块的结构示意图;如图15所示,显示屏幕确定模块903至少可以包括:特征信息检测单元9031、目标屏幕确定单元9032、体征信息检测单元9033、以及第一屏幕显示区确定单元9034,其中:
特征信息检测单元9031,用于所述移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息。
目标屏幕确定单元9032,用于若所述特征信息检测单元检测为是时,根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区。
体征信息检测单元9033,用于判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息。
第一屏幕显示区确定单元9034,将垂直方向上存在用户体征信息的目标屏幕显示区确定为第一屏幕显示区。
其中,在进行低精度决策的时候调用特征信息检测单元9031中的压力检测子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力。
在进行过特征信息检测之后,调用体征信息检测单元9033中的测距子单元和成像子单元,其中:
测距子单元,用于在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值。
此处的测距子单元以及压力检测子单元用于所述移动终端做出低精度条件下的模糊判决。
成像子单元,用于若所述测距子单元判断结果为不大于预设距离阈值时,在测距结果不大于预设距离阈值的目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配。
其中,此处的成像子单元结合上述的压力检测子单元以及测距子单元,用于所述移动终端做出低精度条件下的精确判决。
在一种可能的实现方式中,显示装置900还可以包括:第一统计模块、第一计算模块以及第一确定模块,其中:
第一统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示 区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为测距结果不大于预设距离阈值的目标屏幕显示区和成像结果与预设人体图像匹配的目标屏幕显示区。
第一计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和。
第一确定模块,用于当所述第一计算模块计算的比例大于第一预设比例阈值时,将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
在所述移动终端进行中精度决策时调用特征信息检测单元9031中的温度子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度。
在进行过特征信息检测之后,调用体征信息检测单元9033中的成像子单元以及人脸识别子单元,其中:
成像子单元,用于在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配。
此处的温度子单元以及成像子单元用于所述移动终端做出中精度条件下的模糊判决。
人脸识别子单元,用于若所述成像子单元的检测结果为是时,在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致。
此处的人脸识别子单元结合上述的温度子单元以及成像子单元,用于所述移动终端做出中精度条件下的精确判决。
在一种可能的实现方式中,显示装置900还可以包括:第二统计模块、第二计算模块以及第二确定模块,其中:
第二统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为成像结果与预设人体图像匹配的目标屏幕显示区和人脸识别的结果与预设人脸一致的目标屏幕显示区;
第二计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第二确定模块,用于当所述比例大于第二预设比例阈值时,将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
在所述移动终端进行高精度决策时调用特征信息检测单元中的声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音。
在进行过特征信息检测之后,调用体征信息检测单元中的人脸识别子单元、以及声音识别子单元,其中:
人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致。
此处的声音子单元以及人脸识别子单元用于所述移动终端做出高精度条件下的模糊判决;声音识别子单元,用于在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行声音识别,判断声音识别的结果是否与预设声音一致。
此处的声音识别子单元结合上述的声音子单元以及人脸识别子单元,用于所述移动终端做出高精度条件下的精确判决。
在一种可能的实现方式中,显示装置900还可以包括:第三统计模块、第三计算模块以及第三确定模块,其中:
第三统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和声音识别的结果是否与预设声音一致的目标屏幕显示区;
第三计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第三确定模块,用于当所述比例大于第三预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
作为一种可选的方案,在高精度条件下的精确判决中,体征信息检测单元9033包括:人脸识别子单元、以及虹膜识别子单元,其中:
人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致。
虹膜识别子单元,用于在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别,判断虹膜识别的结果是否与预设虹膜一致。
此处的虹膜识别子单元结合上述的声音子单元以及人脸识别子单元,用于所述移动终端做出高精度条件下的精确判决。
在一种可能的实现方式中,显示装置900还可以包括:第四统计模块、第四计算模块以及第四确定模块,其中:
第四统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和虹膜识别的结果与预设虹膜一致的目标屏幕显示区;
第四计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
第四确定模块,用于当所述比例大于第四预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
图16为本发明另一实施例公开的一种显示装置的结构示意图;显示装置900至少可以包括:屏幕检测模块915、关闭数据接收模块916、关闭模块917、关闭介入检测模块918、以及关闭介入模块919,其中:
屏幕检测模块915,用于检测所述移动终端的是否有处于激活状态的屏幕显示区。
关闭数据接收模块916,用于所述屏幕检测模块检测为是时,所述移动终端接收目标传感器上报的数据。
关闭模块917,用于所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,并关闭所述第一屏幕关闭区。
关闭介入检测模块918,用于在所述关闭所述第一屏幕关闭区之后,检测是否有用户介入操作;所述介入操作用于指示确定第二屏幕关闭区。
关闭介入模块919,用于若所述关闭介入检测模块检测为是时,根据所述用户介入操作在所述处于激活状态的屏幕显示区中确定第二屏幕关闭区,重新激活所述第一屏幕关闭区,并同时关闭所述第二屏幕关闭区。
图17为本发明实施例公开的一种关闭模块的结构示意图;关闭模块917至少可以包括:关闭操作检测单元9171、以及关闭单元9172,其中:
关闭操作检测单元9171,用于所述移动终端根据所述目标传感器上报的数据,检测所述至少一个处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息。
关闭单元9172,用于根据所述用户操作信息确定第一屏幕关闭区。
在一种可能的实现方式中,关闭操作检测单元9171至少可以包括:动作子单元、第二温度子单元、第二测距子单元以及第二人脸识别子单元,其中:
动作子单元,用于检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化。
第二温度子单元,用于在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度。
第二测距子单元,用于在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值。
第二人脸识别子单元,用于在测距结果大于预设距离阈值的处于激活状态的屏幕显示区的垂直方向上进行人脸识别,检测人脸识别结果是否与预设人脸一致。
在另外一种可能的实现方式中,显示装置900还可以包括:第五统计模块、第五计算模块以及第五确定模块,其中:
第五统计模块,用于统计两个处于激活状态的屏幕显示区为同一屏幕显示区的次数以及所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数,其中,所述两个处于激活状态的屏幕显示区分别为测距结果大于预设距离阈值的处于激活状态的屏幕显示区和人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区;
第五计算模块,计算所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数与所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数之和;
第五确定模块,用于当所述比例大于第五预设比例阈值时,将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
综上所述,通过实施本发明实施例,多屏移动终端通过自身配备的传感器上报的数据,判断在使用 过程中用户的状态和位置,从而在正确的屏幕上显示目标界面;同时通过多种精度的判断决策以及模糊判断和精确判断的对比学习,来提高判断的精度,使得移动终端的判断更加准确,提升用户的体验感受。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,该流程可以由计算机程序来指令相关的硬件完成,该程序可存储于计算机可读取存储介质中,该程序在执行时,可包括如上述各方法实施例的流程。而前述的存储介质包括:ROM或随机存储记忆体RAM、磁碟或者光盘等各种可存储程序代码的介质。
尽管在此结合各实施例对本发明进行了描述,然而不能以此来限定本发明之权利范围,在实施所要求保护的本发明过程中,本领域技术人员通过查看所述附图、公开内容、以及所附权利要求书,可理解并实现上述实施例的全部或部分流程,并依本发明权利要求所作的等同变化,仍属于发明所涵盖的范围。在权利要求中,“包括”一词不排除其他组成部分或步骤,“一”或“一个”不排除多个的情况。单个控制器或其他单元可以实现权利要求中列举的若干项功能。相互不同的从属权利要求所记载了某些措辞,但这并不表示这些措辞不能组合起来产生良好的效果。

Claims (44)

  1. 一种显示方法,其特征在于,应用于具有至少两个屏幕显示区的移动终端,所述方法包括:
    所述移动终端接收开启指令,所述开启指令用于指示打开目标界面;
    所述移动终端获取目标传感器上报的数据;所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制;
    所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,并在所述第一屏幕显示区中呈现所述目标界面。
  2. 如权利要求1所述的方法,其特征在于,所述在所述第一屏幕显示区中呈现所述目标界面之后,还包括:
    检测是否有介入操作;所述介入操作用于确定第二屏幕显示区;
    若检测为是,则根据所述介入操作在所述至少两个屏幕显示区中确定第二屏幕显示区,撤回所述第一屏幕显示区中呈现的目标界面,并在所述第二屏幕显示区中呈现所述目标界面。
  3. 如权利要求1或2所述的方法,其特征在于,所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区,包括:
    所述移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息;
    若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;
    根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区。
  4. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
    若检测为是,则在所述至少两个屏幕显示区中将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值;
    将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
  5. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
    若检测为是,则在所述至少两个屏幕显示区中将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值;
    在测距结果不大于预设距离阈值的目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
    将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
  6. 如权利要求5所述的方法,其特征在于,所述将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
    统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为测距结果不大于预设距离阈值的目标屏幕显示区和成像结果与预设人体图像匹配的目标屏幕显示区;
    计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    当所述比例大于第一预设比例阈值时,将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
  7. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
    若检测为是,则将所述至少两个屏幕显示区中将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
    将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
  8. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
    若检测为是,则在所述至少两个屏幕显示区中将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
    在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
  9. 如权利要求8所述的方法,其特征在于,所述将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
    统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为成像结果与预设人体图像匹配的目标屏幕显示区和人脸识别的结果与预设人脸一致的目标屏幕显示区;
    计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    当所述比例大于第二预设比例阈值时,将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
  10. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
    若检测为是,则在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
  11. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
    若检测为是,则在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行声音识别,判断声音识别的结果是否与预设声音一致;
    将声音识别的结果是否与预设声音一致的目标屏幕显示区确定为第一屏幕显示区。
  12. 如权利要求11所述的方法,其特征在于,所述将声音识别的结果是否与预设声音一致的目标 屏幕显示区确定为第一屏幕显示区之后,还包括:
    统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和声音识别的结果是否与预设声音一致的目标屏幕显示区;
    计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    当所述比例大于第三预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
  13. 如权利要求3所述的方法,其特征在于,所述检测是否存在针对所述屏幕显示区的用户操作特征信息;若检测为是,则根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区,并判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;根据所述用户体征信息在所述目标屏幕显示区中确定第一屏幕显示区,包括:
    检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
    若检测为是,则在所述至少两个屏幕显示区中将方向上存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
    在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别,判断虹膜识别的结果是否与预设虹膜一致;
    将虹膜识别的结果与预设虹膜一致的目标屏幕显示区确定为第一屏幕显示区。
  14. 如权利要求13所述的方法,其特征在于,所述将虹膜识别的结果与预设虹膜一致的目标屏幕显示区确定为第一屏幕显示区之后,还包括:
    统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和虹膜识别的结果与预设虹膜一致的目标屏幕显示区;
    计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    当所述比例大于第四预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
  15. 一种显示方法,其特征在于,应用于具有至少两个屏幕显示区的移动终端,所述方法包括:
    检测所述移动终端的是否有处于激活状态的屏幕显示区;
    若检测为是,则所述移动终端获取目标传感器上报的数据;所述目标传感器包括根据预设的传感器选取机制选取的至少一个传感器;所述传感器选取机制为通过所述移动终端支持的传感器类型确定的选取机制;
    所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,并关闭所述第一屏幕关闭区。
  16. 如权利要求15所述的方法,其特征在于,在所述关闭所述第一屏幕关闭区之后,还包括:
    检测是否有介入操作;所述介入操作用于确定第二屏幕关闭区;
    若检测为是,则根据所述介入操作在所述处于激活状态的屏幕显示区中确定第二屏幕关闭区,重新激活所述第一屏幕关闭区,并关闭所述第二屏幕关闭区。
  17. 如权利要求15或16所述的方法,其特征在于,所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,包括:
    所述移动终端根据所述目标传感器上报的数据,检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息;
    根据所述用户操作信息确定第一屏幕关闭区。
  18. 如权利要求17所述的方法,其特征在于,所述移动终端根据所述目标传感器上报的数据,检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息,根据所述用户操作信息确定第一屏幕关闭区,包括:
    检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
    在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
    在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
    将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
  19. 如权利要求17所述的方法,其特征在于,所述检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息;若检测为是,则根据所述用户操作特征信息在所述处于激活状态的屏幕显示区中确定目标屏幕关闭区,并判断所述目标屏幕关闭区的垂直方向上是否存在用户的体征信息;将垂直方向上不存在用户体征信息的目标屏幕关闭区确定为第一屏幕关闭区,包括:
    检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
    在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
    在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
    在测距结果大于预设距离阈值的处于激活状态的屏幕显示区的垂直方向上进行人脸识别,检测人脸识别结果是否与预设人脸一致;
    将人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
  20. 如权利要求19所述的方法,其特征在于,所述将人脸识别结果与预设人脸一致的处于激活状 态的屏幕显示区确定为第一屏幕关闭区之后,还包括:
    统计两个处于激活状态的屏幕显示区为同一屏幕显示区的次数以及所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数,其中,所述两个处于激活状态的屏幕显示区分别为测距结果大于预设距离阈值的处于激活状态的屏幕显示区和人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区;
    计算所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数与所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数之和;
    当所述比例大于第五预设比例阈值时,将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
  21. 一种显示装置,其特征在于,包括:
    开启模块,用于移动终端接收开启指令,所述开启指令用于指示打开目标界面;
    显示数据获取模块,用于所述移动终端获取目标传感器上报的数据;
    显示屏幕确定模块,用于所述移动终端根据所述目标传感器上报的数据,在所述至少两个屏幕显示区中确定第一屏幕显示区;
    显示模块,用于在所述第一屏幕显示区中呈现所述目标界面。
  22. 如权利要求21所述的装置,其特征在于,所述装置还包括:
    显示介入操作检测模块,用于所述在所述第一屏幕显示区中呈现所述第一应用的界面之后,检测是否有介入操作;所述介入操作用于确定第二屏幕显示区;
    介入显示模块,用于根据所述介入操作在所述至少两个屏幕显示区中确定第二屏幕显示区,撤回所述第一屏幕显示区中呈现的目标界面,并在所述第二屏幕显示区中呈现所述目标界面。
  23. 如权利要求21或22所述的装置,其特征在于,所述显示屏幕确定模块包括:
    特征信息检测单元,用于所述移动终端根据所述目标传感器上报的数据,检测是否存在针对所述屏幕显示区的用户操作特征信息;
    目标屏幕确定单元,用于若所述特征信息检测单元检测为是时,根据所述用户操作特征信息在所述至少两个屏幕显示区中确定目标屏幕显示区;
    体征信息检测单元,用于判断所述目标屏幕显示区的垂直方向上是否存在用户体征信息;
    第一屏幕显示区确定单元,将垂直方向上存在用户体征信息的目标屏幕显示区确定为第一屏幕显示区。
  24. 如权利要求23所述的装置,其特征在于,所述特征信息检测单元包括:
    压力检测子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
    若检测为是,则所述目标屏幕确定单元将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
    所述体征信息检测单元包括:
    测距子单元,用于在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离 阈值;
    若判断为是,则所述第一屏幕显示区确定单元将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
  25. 如权利要求23所述的装置,其特征在于,所述特征信息检测单元包括:
    压力检测子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区是否存在触摸压力;
    若检测为是,则所述目标屏幕确定单元将不存在触摸压力的屏幕显示区确定为目标屏幕显示区;
    所述体征信息检测单元包括:
    测距子单元,用于在所述目标屏幕显示区的垂直方向上进行测距,判断测距结果是否大于预设距离阈值;
    所述体征信息检测单元还包括:
    成像子单元,用于若所述测距子单元判断结果为不大于预设距离阈值时,在测距结果不大于预设距离阈值的目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
    若判断为是,则所述第一屏幕显示区确定单元将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
  26. 如权利要求25所述的装置,其特征在于,所述装置还包括:
    第一统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为测距结果不大于预设距离阈值的目标屏幕显示区和成像结果与预设人体图像匹配的目标屏幕显示区;
    第一计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    第一确定模块,用于当所述第一计算模块计算的比例大于第一预设比例阈值时,将测距结果不大于预设距离阈值的目标屏幕显示区确定为第一屏幕显示区。
  27. 如权利要求23所述的装置,其特征在于,所述特征信息检测单元包括:
    温度子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
    若检测为是,则所述目标屏幕确定单元将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
    所述体征信息检测单元包括:
    成像子单元,用于在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
    若成像结果为是,则所述第一屏幕显示区确定单元将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
  28. 如权利要求23所述的装置,其特征在于,所述操作信息检测单元包括:
    温度子单元,用于检测所述至少两个屏幕显示区中的每一个屏幕显示区的垂直方向上是否存在与预设温度区间相匹配的温度;
    若检测为是,则所述目标屏幕确定单元将垂直方向上存在与预设温度区间相匹配的温度的屏幕显示区确定为目标屏幕显示区;
    所述体征信息检测单元包括:
    成像子单元,用于在所述目标屏幕显示区的垂直方向上进行成像,检测成像结果是否与预设人体图像匹配;
    所述体征信息检测单元还包括:
    人脸识别子单元,用于若所述成像子单元的检测结果为是时,在成像结果与预设人体图像匹配的目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    若检测结果为与预设人脸一致,则所述第一屏幕显示区确定单元将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
  29. 如权利要求28所述的装置,其特征在于,所述装置还包括:
    第二统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为成像结果与预设人体图像匹配的目标屏幕显示区和人脸识别的结果与预设人脸一致的目标屏幕显示区;
    第二计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    第二确定模块,用于当所述比例大于第二预设比例阈值时,将成像结果与预设人体图像匹配的目标屏幕显示区确定为第一屏幕显示区。
  30. 如权利要求23所述的装置,其特征在于,所述操作信息检测单元包括:
    声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
    若检测为是,则所述目标屏幕确定单元将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
    所述体征信息检测单元包括:
    人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    若检测结果为与预设人脸一致,则所述第一屏幕显示区确定单元将人脸识别的结果与预设人脸一致的目标屏幕显示区确定为第一屏幕显示区。
  31. 如权利要求23所述的装置,其特征在于,所述操作信息检测单元包括:
    声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
    若检测为是,则所述目标屏幕确定单元将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显 示区;
    所述体征信息检测单元包括:
    人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    所述体征信息检测单元还包括:
    声音识别子单元,用于在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行声音识别,判断声音识别的结果是否与预设声音一致;
    若检测结果为与预设声音一致,则所述第一屏幕显示区确定单元将声音识别的结果与预设声音一致的目标屏幕显示区确定为第一屏幕显示区。
  32. 如权利要求31所述的装置,其特征在于,所述装置还包括:
    第三统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和声音识别的结果是否与预设声音一致的目标屏幕显示区;
    第三计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    第三确定模块,用于当所述比例大于第三预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
  33. 如权利要求23所述的装置,其特征在于,所述操作信息检测单元包括:
    声音子单元,用于检测所述至少两个屏幕显示区的每一个屏幕显示区的方向上是否存在特定频率范围的声音;
    若检测为是,则所述目标屏幕确定单元将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区;
    所述体征信息检测单元包括:
    人脸识别子单元,用于在所述将存在特定频率范围的声音的屏幕显示区确定为目标屏幕显示区之后,在所述目标屏幕显示区的垂直方向上进行人脸识别,检测人脸识别的结果是否与预设人脸一致;
    所述体征信息检测单元还包括:
    虹膜识别子单元,用于在人脸识别的结果与预设人脸一致的目标屏幕显示区的垂直方向上进行虹膜识别,判断虹膜识别的结果是否与预设虹膜一致;
    若检测结果为与预设虹膜一致,则所述第一屏幕显示区确定单元将虹膜识别的结果与预设虹膜一致的目标屏幕显示区确定为第一屏幕显示区。
  34. 如权利要求33所述的装置,其特征在于,所述装置还包括:
    第四统计模块,用于统计两个目标屏幕显示区为同一屏幕显示区的次数以及所述两个目标屏幕显示区为不同屏幕显示区的次数,其中,所述两个目标屏幕显示区分别为人脸识别的结果与预设人脸一致的目标显示区和虹膜识别的结果与预设虹膜一致的目标屏幕显示区;
    第四计算模块,用于计算所述两个目标屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个目标屏幕显示区为同一屏幕显示区的次数与所述两个目标屏幕显示区为不同屏幕显示区的次数之和;
    第四确定模块,用于当所述比例大于第四预设比例阈值时,将人脸识别的结果与预设人脸一致的目标显示区确定为第一屏幕显示区。
  35. 一种显示装置,其特征在于,包括:
    屏幕检测模块,用于检测所述移动终端的是否有处于激活状态的屏幕显示区;
    关闭数据获取模块,用于所述屏幕检测模块检测为是时,所述移动终端获取目标传感器上报的数据;
    关闭模块,用于所述移动终端根据所述目标传感器上报的数据,在所述处于激活状态的屏幕显示区中确定第一屏幕关闭区,并关闭所述第一屏幕关闭区。
  36. 如权利要求35所述的装置,其特征在于,所述装置还包括:
    关闭介入检测模块,用于在所述关闭所述第一屏幕关闭区之后,检测是否有介入操作;所述介入操作用于确定第二屏幕关闭区;
    关闭介入模块,用于若所述关闭介入检测模块检测为是时,根据所述介入操作在所述处于激活状态的屏幕显示区中确定第二屏幕关闭区,重新激活所述第一屏幕关闭区,并关闭所述第二屏幕关闭区。
  37. 如权利要求35或36所述的装置,其特征在于,所述关闭模块包括:
    关闭操作检测单元,用于所述移动终端根据所述目标传感器上报的数据,检测所述至少一个处于激活状态的屏幕显示区中的每一个屏幕显示区是否存在用户操作信息;
    关闭单元,用于根据所述用户操作信息确定第一屏幕关闭区。
  38. 如权利要求37所述的装置,其特征在于,所述关闭操作检测单元包括:
    动作子单元,用于检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
    第二温度子单元,用于在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
    第二测距子单元,用于在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
    若检测结果为大于预设距离阈值,则所述关闭单元将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定第一屏幕关闭区。
  39. 如权利要求38所述的装置,其特征在于,所述关闭操作检测单元包括:
    动作子单元,用于检测所述处于激活状态的屏幕显示区中的每一个屏幕显示区的方向上是否存在动作变化;
    第二温度子单元,用于在存在动作变化的处于激活状态的屏幕显示区的垂直方向上检测是否存在与预设温度区间相匹配的温度;
    第二测距子单元,用于在垂直方向上不存在与预设温度区间相匹配的温度的处于激活状态的屏幕显示区的垂直方向上进行测距,检测测距结果是否大于预设距离阈值;
    第二人脸识别子单元,用于在测距结果大于预设距离阈值的处于激活状态的屏幕显示区的垂直方向上进行人脸识别,检测人脸识别结果是否与预设人脸一致;
    若检测结果为与预设人脸不一致,则所述关闭单元将与预设人脸不一致的处于激活状态的屏幕显示区确定第一屏幕关闭区。
  40. 如权利要求39所述的装置,其特征在于,所述装置还包括:
    第五统计模块,用于统计两个处于激活状态的屏幕显示区为同一屏幕显示区的次数以及所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数,其中,所述两个处于激活状态的屏幕显示区分别为测距结果大于预设距离阈值的处于激活状态的屏幕显示区和人脸识别结果与预设人脸不一致的处于激活状态的屏幕显示区;
    第五计算模块,计算所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数占总的统计次数的比例,其中,所述总的统计次数为所述两个处于激活状态的屏幕显示区为同一屏幕显示区的次数与所述两个处于激活状态的屏幕显示区为不同屏幕显示区的次数之和;
    第五确定模块,用于当所述比例大于第五预设比例阈值时,将测距结果大于预设距离阈值的处于激活状态的屏幕显示区确定为第一屏幕关闭区。
  41. 一种终端,其特征在于,所述终端包括:
    存储器,用于存储显示指令;
    处理器,用于调用所述存储器中的显示指令,所述显示指令应用于具有至少两个屏幕显示区的移动终端,所述处理器调用所述显示指令并执行如权利要求1-14任一项所述的显示方法。
  42. 一种终端,其特征在于,所述终端包括:
    存储器,用于存储关闭指令;
    处理器,用于调用所述存储器中的关闭指令,所述关闭指令应用于具有至少两个屏幕显示区的移动终端,所述处理器调用所述关闭指令并执行如权利要求15-20任一项所述的显示方法。
  43. 一种计算机可读的存储介质,用于存储一个或多个计算机程序,所述一个或多个计算机程序包括指令,当所述计算机程序在计算机上运行时,所述指令用于执行上述权利要求1-20任一项所述的显示方法。
  44. 一种计算机程序,所述计算机程序包括指令,当所述计算机程序在计算机上执行时,所述指令用于执行权利要求1-20任一项所述的显示方法。
PCT/CN2017/085744 2016-11-30 2017-05-24 显示方法、装置及终端 WO2018099017A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201780009703.2A CN108604143B (zh) 2016-11-30 2017-05-24 显示方法、装置及终端

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201611095183.5 2016-11-30
CN201611095183 2016-11-30
CN201710204290.5 2017-03-30
CN201710204290 2017-03-30

Publications (1)

Publication Number Publication Date
WO2018099017A1 true WO2018099017A1 (zh) 2018-06-07

Family

ID=62241150

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/085744 WO2018099017A1 (zh) 2016-11-30 2017-05-24 显示方法、装置及终端

Country Status (2)

Country Link
CN (1) CN108604143B (zh)
WO (1) WO2018099017A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109451336A (zh) * 2018-10-09 2019-03-08 Oppo(重庆)智能科技有限公司 视频播放方法及相关产品
CN113542453A (zh) * 2020-03-31 2021-10-22 北京小米移动软件有限公司 一种折叠屏终端设备、其状态检测方法和电子设备
CN113760212A (zh) * 2021-09-06 2021-12-07 西安超越申泰信息科技有限公司 一种基于win7实现屏幕自适应翻转的方法及装置

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101674347A (zh) * 2009-09-28 2010-03-17 明基电通有限公司 电子装置以及电子装置状态判断方法
CN105765500A (zh) * 2013-11-13 2016-07-13 戴尔产品有限公司 双显示系统中的动态悬浮敏感度及手势适应
CN105867860A (zh) * 2016-03-28 2016-08-17 联想(北京)有限公司 一种信息处理方法及电子设备

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103365393B (zh) * 2012-03-27 2018-04-27 联想(北京)有限公司 一种显示方法及电子设备
KR101386218B1 (ko) * 2012-05-24 2014-04-17 삼성디스플레이 주식회사 듀얼 디스플레이 장치 및 그 구동 방법
CN105955610A (zh) * 2016-04-27 2016-09-21 长沙丰灼通讯科技有限公司 一种双屏手持终端的目标屏识别方法
CN106095312B (zh) * 2016-06-08 2020-12-01 泾县谷声信息科技有限公司 一种屏幕解锁方法和装置
CN106155325B (zh) * 2016-07-22 2019-08-30 努比亚技术有限公司 一种双屏显示唤醒装置及方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101674347A (zh) * 2009-09-28 2010-03-17 明基电通有限公司 电子装置以及电子装置状态判断方法
CN105765500A (zh) * 2013-11-13 2016-07-13 戴尔产品有限公司 双显示系统中的动态悬浮敏感度及手势适应
CN105867860A (zh) * 2016-03-28 2016-08-17 联想(北京)有限公司 一种信息处理方法及电子设备

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109451336A (zh) * 2018-10-09 2019-03-08 Oppo(重庆)智能科技有限公司 视频播放方法及相关产品
CN109451336B (zh) * 2018-10-09 2021-08-17 Oppo(重庆)智能科技有限公司 视频播放方法及相关产品
CN113542453A (zh) * 2020-03-31 2021-10-22 北京小米移动软件有限公司 一种折叠屏终端设备、其状态检测方法和电子设备
CN113760212A (zh) * 2021-09-06 2021-12-07 西安超越申泰信息科技有限公司 一种基于win7实现屏幕自适应翻转的方法及装置
CN113760212B (zh) * 2021-09-06 2024-03-15 超越科技股份有限公司 一种基于win7实现屏幕自适应翻转的方法及装置

Also Published As

Publication number Publication date
CN108604143B (zh) 2020-12-01
CN108604143A (zh) 2018-09-28

Similar Documents

Publication Publication Date Title
US20180313695A1 (en) Smart device and method for controlling same
US11172838B2 (en) Sensing body information apparatus for volume and blood flow via light reflectance
US20160162039A1 (en) Method and system for touchless activation of a device
KR20170007285A (ko) 이미지들을 캡처하는 것을 돕기 위해 햅틱 피드백을 제공하기 위한 시스템 및 방법
WO2018099017A1 (zh) 显示方法、装置及终端
WO2019218843A1 (zh) 按键配置方法、装置、移动终端及存储介质
WO2019024644A1 (zh) 接近检测方法、装置、存储介质及电子设备
EP3407230A1 (en) Electronic apparatus and control method therefor
CN109634429A (zh) 一种电子设备及信息获取方法
US20190329364A1 (en) Object Recognition and Presentation for the Visually Impaired
EP3285049B1 (en) Method and device for microphone selection
CN105607738B (zh) 确定单手模式的方法及装置
CN108156729A (zh) 照明设备、照明设备的控制方法及装置
WO2022143384A1 (zh) 智能手环及其控制方法
US11520375B2 (en) Foldable electronic device and method of estimating bioinformation using the same
CN105526992A (zh) 水位测量方法及装置、容器
CN105266756A (zh) 瞳距测量方法、装置及终端
CN105323480A (zh) 一种基于超声波的拍照方法及装置
US20150160743A1 (en) Displacement detection device with no hovering function and computer system including the same
CN110290576A (zh) 电子设备控制方法及装置
WO2016047136A1 (ja) 測定装置及び測定方法
CN106873939B (zh) 电子设备及其使用方法
CN108509127A (zh) 启动录屏任务的方法、装置及计算机设备
US20150062321A1 (en) Device control by facial feature recognition
US11234101B2 (en) Determining an orientation and body location of a wearable device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17877330

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17877330

Country of ref document: EP

Kind code of ref document: A1