WO2023280020A1 - 一种系统模式切换方法、电子设备及计算机可读存储介质 - Google Patents

一种系统模式切换方法、电子设备及计算机可读存储介质 Download PDF

Info

Publication number
WO2023280020A1
WO2023280020A1 PCT/CN2022/101983 CN2022101983W WO2023280020A1 WO 2023280020 A1 WO2023280020 A1 WO 2023280020A1 CN 2022101983 W CN2022101983 W CN 2022101983W WO 2023280020 A1 WO2023280020 A1 WO 2023280020A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
system mode
image
electronic device
smart
Prior art date
Application number
PCT/CN2022/101983
Other languages
English (en)
French (fr)
Inventor
李乐
高晓强
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2023280020A1 publication Critical patent/WO2023280020A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/441Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
    • H04N21/4415Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning

Definitions

  • the present application relates to the field of electronic equipment, and in particular to a system mode switching method, electronic equipment, and a computer-readable storage medium.
  • the system modes of a smart TV generally include a normal mode and a children's mode.
  • the system mode of the smart TV can be switched from the normal mode to the children's mode, that is, the smart TV exits the normal mode (starting the children's mode), so that the children at home can control the scope, duration, sitting posture, etc. .
  • the system mode of the smart TV can be switched from the child mode to the normal mode, that is, the smart TV exits the child mode (starts the normal mode).
  • the embodiment of the present application provides a system mode switching method, which is used to improve the operation convenience and accuracy of switching the system mode of an electronic device.
  • the embodiment of the present application discloses a system mode switching method, which is applied to an electronic device (such as a smart TV).
  • the method includes: running the first system mode; acquiring the voice of the user of the electronic device, and analyzing the Instruction; determine that the instruction includes a preset wake-up word for switching the first system mode; acquire a user image collected for a user in front of the electronic device, and determine according to the user image that the user includes a user of a preset user type; according to the instruction, and According to the user image and the determined user of the preset user type, the electronic device is switched from the first system mode to the second system mode corresponding to the preset user type.
  • the above-mentioned first system mode is described by taking the child mode as an example, and the second system mode is described by taking the normal mode as an example.
  • the command to exit the children's mode is triggered by recognizing the user's voice, and the presence of an adult is determined by recognizing the user's image. Therefore, according to the instruction to exit the children's mode and the presence of an adult, the smart TV can be completed. Exit action for child mode. Therefore, the child mode exit operation path is short, and the usability is high. Moreover, user identification is accurate, and it is not easy to be bypassed by children.
  • the above-mentioned system mode switching method is suitable for exiting the guest mode of the mobile phone.
  • the user image includes a first user image and a second user image
  • Switching the electronic device from the first system mode to the second system mode corresponding to the preset user type according to the user image and the determined user of the preset user type according to the instruction includes:
  • the action instruction is based on a static action.
  • the action instruction is based on a dynamic action
  • the dynamic action includes a user's body action
  • the first user image includes a static image
  • the users determined according to the static image include users of a preset user type.
  • the user of the preset user type is the sender of the voice.
  • the electronic device is switched from the first system mode to the second system mode corresponding to the preset user type according to the instruction and according to the user image and the determined user of the preset user type ,include:
  • the electronic device is switched from the first system mode to the second system mode corresponding to the preset user type.
  • the user image includes a static image
  • the users determined according to the static image include users of a preset user type.
  • the user image includes a dynamic image
  • the dynamic image includes body movements of the user.
  • the body movement includes a hand waving gesture.
  • the preset user type is an adult
  • the first system mode is a child mode
  • the second system mode is a normal mode
  • the embodiment of the present application provides an electronic device, including: a pickup, used to collect the voice of the user; a camera, used to collect the user image of the user; a processor; a memory, the memory includes instructions, and the instructions are processed by the processor During execution, the electronic device is made to execute the system mode switching method provided in any implementation manner of the first aspect above.
  • the embodiments of the present application provide a computer-readable storage medium, on which instructions are stored, and when the instructions are executed on a computer, the computer executes the method provided by any implementation manner of the above-mentioned first aspect. System mode switching method.
  • Fig. 1 shows a first schematic diagram of a smart home scene according to some embodiments of the present application
  • Fig. 2 shows a schematic flow diagram of a smart TV exiting the children's mode according to some embodiments of the present application
  • FIG. 3 is a schematic structural diagram of an electronic device provided in an embodiment of the present application.
  • FIG. 4 shows a flowchart of a system mode switching method provided by an embodiment of the present application
  • Fig. 5 shows a second schematic diagram of a smart home scene according to some embodiments of the present application.
  • Fig. 6 shows a third schematic diagram of a smart home scene according to some embodiments of the present application.
  • Fig. 7 shows a fourth schematic diagram of a smart home scene according to some embodiments of the present application.
  • FIG. 8 shows a block diagram of an electronic device provided by an embodiment of the present application.
  • FIG. 9 shows a block diagram of a system on chip (SoC) provided by an embodiment of the present application.
  • SoC system on chip
  • the present application provides a system mode switching method, which is used in electronic equipment such as televisions, and is beneficial to improving the convenience and accuracy of switching the system mode of the electronic equipment.
  • Fig. 1 shows an example of a smart home scenario including multiple electronic devices.
  • specific electronic devices include: a smart TV 100 and a sound box 200 .
  • the smart TV 100 can interact with the cloud or other devices (for example, interact with the speaker 200 ) through the network.
  • the smart TV 100 can be equipped with an operating system, so that users can install and uninstall various application software by themselves while enjoying ordinary TV content, and continuously expand and upgrade functions.
  • the smart TV 100 may have interactive applications in various ways, such as human-computer interaction, multi-screen interaction, content sharing, and the like. It should be noted that the TV in each embodiment of the present application may be the smart TV 100 mentioned above, or a smart screen with a larger screen.
  • the speaker 200 may be a smart speaker.
  • the smart speaker 200 has a sound pickup function and can collect a user's voice.
  • the above-mentioned system modes of the smart TV 100 include a normal mode and a children's mode.
  • the smart TV 100 activates a corresponding system mode.
  • the smart TV 100 is switched to the children's mode. As shown in FIG. 1 , the smart TV 100 is in the children's mode, and the children at home sit on the sofa and watch a video of children playing football.
  • the smart TV 100 When an adult uses the smart TV 100, the smart TV 100 exits the children's mode, that is, the system mode of the smart TV 100 is switched from the children's mode to the normal mode.
  • the process is mainly divided into the following three parts:
  • the user reads the confirmation message on the confirmation pop-up box for exiting child mode, and performs user identity authentication.
  • user identity authentication there are currently three main forms of user identity authentication: 1user preset user identity password; 2simple mathematical calculation; 3verification of recorded voiceprint.
  • the smart TV 100 exits the child mode. Otherwise, the smart TV 100 does not perform the process of exiting the child mode.
  • the process of exiting the children’s mode of the smart TV 100 mainly has the following problems: (1) The operation path for exiting the children’s mode is long; authentication form) is easily bypassed by children; (3) the success rate of voiceprint recognition is low (the third identity authentication form); (4) relies on remote control operation; (5) it is relatively difficult for the elderly to operate.
  • the present application provides a system mode switching method, which generates an instruction to trigger exiting the children's mode by recognizing the user's voice, and determines that an adult is present by recognizing the user's image, so that according to the instruction to exit the children's mode and the presence of an adult, The exit operation of the child mode of the smart TV can be completed. Therefore, the child mode exit operation path is short, and the usability is high. Moreover, user identification is accurate, and it is not easy to be bypassed by children.
  • a smart TV 100 is provided as an example of a body of an electronic device.
  • the electronic device may be various electronic devices with a camera 101 and a pickup.
  • various smart home devices such as speaker 200 shown in FIG. 1 , smart alarm clock, etc.
  • mobile phones tablet computers, desktop computers, laptop computers, vehicle-mounted terminals, artificial intelligence (artificial intelligence, AI) intelligent voice terminals , wearable devices, augmented reality (augmented reality, AR) equipment, virtual reality (virtual reality, VR) equipment, ultra-mobile personal computer (ultra-mobile personal computer, UMPC), handheld computer, netbook, personal digital assistant (personal digital assistant) assistant, PDA) and other electronic devices.
  • augmented reality augmented reality, AR
  • virtual reality virtual reality
  • ultra-mobile personal computer ultra-mobile personal computer
  • UMPC ultra-mobile personal computer
  • PDA personal digital assistant
  • Exemplary embodiments of electronic devices include, but are not limited to, various electronic devices equipped with IOS, Android, Microsoft, HarmonyOS, or other operating systems.
  • Fig. 3 shows a schematic structural diagram of an electronic device 001 according to an embodiment of the present application.
  • the structure of the smart TV 100 in the above embodiments may be the same as that of the electronic device 001 .
  • the electronic device 001 may include:
  • Display 102 has a display interface.
  • the display interface is used to play videos or display screen projections from computers or mobile phones, etc.
  • the display screen 102 displays an icon of the children's mode, and plays the video of children playing football described in the above embodiments.
  • the display screen 102 may be a capacitive touch screen.
  • Processor 110 may include one or more processing units, for example, may include a central processing unit CPU (Central Processing Unit), an image processor GPU (Graphics Processing Unit), a digital signal processor DSP, a microprocessor MCU (Micro- Programmed Control Unit), AI (Artificial Intelligence, artificial intelligence) processor or programmable logic device FPGA (Field Programmable Gate Array) and other processing modules or processing circuits. Wherein, different processing units may be independent devices, or may be integrated in one or more processors. The processor can switch the system mode according to the recognized voice and image of the user.
  • CPU Central Processing Unit
  • GPU Graphics Processing Unit
  • DSP digital signal processor
  • microprocessor MCU Micro- Programmed Control Unit
  • AI Artificial Intelligence, artificial intelligence
  • FPGA Field Programmable Gate Array
  • a memory 180 or a cache may be provided in the processor 110 for storing instructions and data.
  • the memory 180 or the buffer may include a correspondence between user images and user types, or may include a correspondence between user voices, user types, and system modes. The above correspondence can be pre-stored in the memory 180.
  • the processor 110 After the processor 110 recognizes the user's voice and user image, it searches the memory 180 whether there is a corresponding system mode. If so, the processor 110 switches the system mode.
  • the power module 140 may include a power supply, power management components, and the like.
  • the power source can be a battery.
  • the power management component is used to manage the charging of the power supply and the power supply from the power supply to other modules.
  • the power management component includes a charge management module and a power management module.
  • the charging management module is used to receive charging input from the charger; the power management module is used to connect the power supply, the charging management module and the processor 110 .
  • the power management module receives the input of the power supply and/or the charging management module, and supplies power for the processor 110, the display screen 102, the camera 170, and the wireless communication module 120, etc.
  • the wireless communication module 120 may include an antenna, and transmit and receive electromagnetic waves via the antenna.
  • the wireless communication module 120 can provide wireless local area networks (wireless local area networks, WLAN) (such as wireless fidelity (Wireless Fidelity, Wi-Fi) network), bluetooth (bluetooth, BT), global navigation satellite system, etc. (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), near field communication technology (near field communication, NFC), infrared technology (infrared, IR) and other wireless communication solutions.
  • the electronic device 001 can communicate with the network and other devices (eg mobile phone, speaker 200 ) through wireless communication technology.
  • the audio module 150 is used to convert digital audio information into analog audio signal output, or convert analog audio input into digital audio signal.
  • the audio module 150 may also be used to encode and decode audio signals.
  • the audio module 150 may be set in the processor 110 , or some functional modules of the audio module 150 may be set in the processor 110 .
  • the audio module 150 may include a speaker, an earpiece, a pickup, and an earphone jack.
  • the camera 170 is used to capture still images or dynamic images.
  • the object generates an optical image through the lens and projects it to the photosensitive element.
  • the photosensitive element converts the light signal into an electrical signal, and then transmits the electrical signal to the ISP (Image Signal Processing, Image Signal Processing) to convert it into a digital image signal.
  • the electronic device 001 can realize the shooting function through an ISP, a camera 170, a video codec, a GPU (Graphic Processing Unit, graphics processor), a display screen 102, and an application processor.
  • the interface module 160 includes an external memory interface, a universal serial bus (universal serial bus, USB) interface, and the like.
  • the external memory interface can be used to connect an external memory card, such as a Micro SD card, to realize the expansion of the storage capacity of the electronic device 001.
  • the external memory card communicates with the processor 110 through the external memory interface to realize the data storage function.
  • the universal serial bus interface is used for the electronic device 001 to communicate with other electronic devices.
  • the electronic device 001 further includes a button 101 .
  • the key 101 may include a volume key, an on/off key, and the like.
  • the electronic device 001 further includes a touch detection device 190 .
  • the touch detection device 190 can detect the position of the user's touch point, and recognize the corresponding touch gesture according to the position of the user's touch point.
  • the structure shown in the embodiment of the present invention does not constitute a specific limitation on the electronic device 001 .
  • the electronic device 001 may include more or fewer components than shown in the figure, or combine certain components, or separate certain components, or arrange different components.
  • the illustrated components can be realized in hardware, software or a combination of software and hardware.
  • the application provides a system mode switching method, which can complete the exit operation of the children's mode of the smart TV by recognizing the user's voice and the user's image. Therefore, the child mode exit operation path is short, and the usability is high. Moreover, user identification is accurate, and it is not easy to be bypassed by children.
  • the method provided in this embodiment includes the following steps:
  • the current system mode of the smart TV 100 is the children's mode (ie, the first system mode).
  • the smart TV 100 runs the children's mode
  • the range, duration, and sitting posture of children watching the smart TV 100 are controlled.
  • the smart TV 100 is running in the children's mode
  • the children need to sit and watch the video, which is a children's video
  • the duration of watching the video is limited to 30 minutes.
  • the child sits on the sofa and can normally watch the video of the child playing football played on the smart TV 100, and the duration is 30 minutes.
  • the camera 101 of the smart TV 100 detects that the child is not sitting on the sofa, and the smart TV 100 stops playing the current video.
  • the smart TV 100 also stops playing the current video.
  • S200 The smart TV acquires the voice of the user, and parses out the instructions from the voice.
  • the adult When an adult needs to use the smart TV 100 to watch a video or after a child watches the video, and the adult switches the system mode of the smart TV 100 , the adult performs a voice wake-up to exit the children's mode. For example, as shown in FIG. 5, an adult says “Xiaoyi Xiaoyi, exit the child mode". "Xiaoyi Xiaoyi, exit child mode" is the wake-up voice (wake-up word + command word) sent by the user. Among them, the wake-up word is "Xiaoyi Xiaoyi", and the command word is "exit child mode".
  • the above-mentioned wake-up voice is acquired by other electronic devices communicatively connected with the smart TV 100 , such as the sound box 200 (indicated by path 1 in FIG. 5 ).
  • the speaker 200 can transmit the wake-up voice to the smart TV 100, and the smart TV 100 analyzes the wake-up voice.
  • the speaker 200 analyzes the wake-up voice by itself, and then sends the analyzed instruction to the smart TV 100 .
  • Both the sound box 200 and the smart TV 100 analyze the wake-up voice, and finally the command parsed from the wake-up voice is: the smart TV 100 exits the child mode.
  • the wake-up voice may also be acquired by other electronic devices communicatively connected with the smart TV 100 , such as a remote controller or a wearable device connected via Bluetooth.
  • the process of waking up the speech analysis is the same as the above-mentioned way of using the speaker 200 .
  • the wake-up voice may also be directly obtained by a built-in pickup in the smart TV 100 .
  • the above-mentioned wake-up voice may be pre-set in the electronic device by the user according to his own needs, or may be set before the electronic device leaves the factory.
  • the embodiment of the present application does not limit the setting method of the wake-up voice.
  • S300 The smart TV determines that the instruction includes a preset wake-up word for switching the child mode.
  • exit children's mode in the above instruction is a wake-up word for switching children's mode.
  • S400 The smart TV acquires a user image, and determines that the user includes an adult according to the user image.
  • the smart TV 100 confirms that the user corresponding to the acquired user image and the user who issued the wake-up voice are the same person.
  • the camera 101 of the smart TV 100 acquires the image of the user by detecting the sound source image, that is, the camera 101 collects the image of the user who made the wake-up voice (shown as the path 2 in FIG. 5 ). Wherein, what the camera 101 acquires is the user image of the user in front of the smart TV 100 . "Before the smart TV 100" refers to "where the camera 101 of the smart TV 100 can reach”.
  • the processor of the smart TV 100 judges the user image captured by the camera 101 and determines that the user who made the wake-up voice is an adult, not a child. Then, the smart TV 100 satisfies the wake-up operation condition that triggers exiting the child mode.
  • the processor of the smart TV 100 combines the received wake-up voice and the collected portrait data to perform calculations and determine whether the conditions for triggering the exit from the children's mode scene are met. That is, when the user who makes the wake-up voice is an adult, the condition for triggering the exit of the child mode scene is met.
  • the above "camera 101 collects images of the user who made the wake-up voice” is realized, for example, in the following manner: after the smart TV 100 acquires the wake-up voice from the user, the smart TV 100 performs sound source recognition on the user's wake-up voice to determine The user's location when the voice is spoken, specifically the azimuth angle. That is, whether the user who sends out the wake-up voice is located on the left side of the sound collector (such as the above-mentioned speaker 200), or is facing the position of the sound collector, etc., and then adjusts the position of the camera 101 of the smart TV 100 according to the user's location. Shooting angles to accurately capture images of the user who made the wake-up voice.
  • the above-mentioned "camera 101 collects images of the user who issued the wake-up voice” is realized, for example, in the following manner: the camera 101 of the smart TV 100 captures image information in real time, and the processor of the smart TV 100 processes the captured image The information is used for face recognition based on face recognition technology. When it is determined to be a portrait, the mouth opening recognition is performed. When it is determined that the recognition result is an opening of the mouth ("opening the mouth" indicates that a wake-up voice is being issued), the portrait of the mouth is obtained to complete the identification. The wake-up voice is used for image acquisition.
  • the user image collected by the camera 101 for the user who made the wake-up voice is a face
  • face detection technology may be used to distinguish whether the user who made the wake-up voice is an adult or a child.
  • the processor's image-based AI face recognition can automatically extract facial feature values, which include, for example, skin texture, skin color, brightness, and wrinkle texture. Confirm whether the portrait is an adult or a child by means of facial feature values.
  • the above-mentioned user image (human face) is a static image, and the processor determines that the user includes an adult (user of a preset user type) according to the static image.
  • the user corresponding to the acquired user image and the user who makes the wake-up voice are not necessarily the same person.
  • the camera 101 of the smart TV 100 collects the user image of the user in front of the smart TV, and the processor confirms whether the user image collected by the camera 101 contains an adult, and the adult may be The person who wakes up the voice may or may not be the one who made the wake-up voice, as long as there is an adult present.
  • the processor may confirm that the user corresponding to the user image is an adult or a child through the aforementioned AI face recognition.
  • the user is in a still standing posture
  • the camera 101 of the smart TV 100 acquires an image of the user in a still standing posture
  • the processor determines the height corresponding to the user image according to the height of the standing posture corresponding to the user image.
  • the user is an adult or a child.
  • the processor of the smart TV 100 determines that the user has issued a wake-up voice, and the users in front of the smart TV 100 include adults, that is, an adult is present to supervise the operation of exiting the children's mode, the condition for triggering the scene of exiting the children's mode is met.
  • the processor of the smart TV 100 uses voiceprint recognition technology to confirm whether the user who made the wake-up voice is an adult, and at the same time, after the user makes the wake-up voice, the processor of the smart TV 100
  • the camera 101 collects a user image of the user in front of the smart TV, and the processor determines whether the user image collected by the camera 101 includes an adult.
  • the process of the processor confirming whether the image of the user contains an adult is the same as that described in any of the above embodiments.
  • each voice has unique characteristics, through which the voices of different people can be effectively distinguished.
  • Such unique characteristics are, for example, the size of the vocal cavities, specifically the throat, nasal cavity, and oral cavity.
  • the shape, size, and position of these organs determine the tension of the vocal cords and the range of sound frequencies. That is, the magnitude of vocal cord tension and the range of sound frequencies are different between adults and children. Therefore, the collected user's vocal fold tension and sound frequency can be identified through the voiceprint recognition technology, so as to confirm whether the user who made the wake-up voice is an adult.
  • S500 The processor of the smart TV switches the smart TV 100 from the child mode to the normal mode according to the instruction, and according to the user image and the determined adult user.
  • the processor of the smart TV 100 successfully triggers the scene of exiting the children's mode according to the above-mentioned "exiting the children's mode" instruction and determining that the instruction is issued by an adult.
  • the smart TV 100 is switched from the child mode to the normal mode (ie, the second system mode). That is, the operation of exiting the child mode of the smart TV 100 is completed.
  • the processor of the smart TV 100 confirms that an adult has issued the above-mentioned "exit children's mode" command, and as described in the above embodiment, it is determined that there is an adult in front of the smart TV 100 (the adult is not necessarily the one who issued the instruction). If an adult who "exits the children's mode" is present, the smart TV 100 is switched from the children's mode to the normal mode.
  • the processor of the smart TV 100 confirms that a user (adult or child) has issued the above-mentioned "exit children's mode" instruction, and the above-mentioned embodiment describes that there is an adult in front of the smart TV 100 (the An adult (not necessarily the one who issued "exit kid mode") is present to switch the smart TV 100 from the kid mode to the normal mode.
  • the operation of exiting the child mode is performed through the adult image and the wake-up voice issued by the adult, it can be ensured that an adult is present to supervise the exit of the child mode. Therefore, the switching security and accuracy of the system mode of the present application are high. Moreover, it is convenient for the user to switch the system mode of the smart TV 100 . Moreover, the physical components involved in this application are all existing components of the smart TV 100 , and no new physical components are added.
  • the acquisition of the above-mentioned user image is obtained through the camera 101 that comes with the smart TV 100 , but this application is not limited thereto. Image acquisition.
  • a confirmation window will be generated on the display screen of the smart TV 100 (for example, in the middle of the screen). That is, as described in S400 above, after the adult sends out the wake-up voice, and it is determined that the user image captured by the camera 101 is an adult, it means that the condition for triggering the scene of exiting the child mode is met. Then, a confirmation window will be displayed on the smart TV 100 . As shown in FIG. 6 , the confirmation window displays "Please wave your hand towards the screen to confirm exiting the child mode" and "gesture pattern". Equivalently, the smart TV 100 prompts the user to perform an operation of confirming to exit the child mode through the confirmation window.
  • the user makes a gesture of waving according to the prompt in the confirmation window. That is, the user's body movement is a hand-waving gesture.
  • the smart TV 100 acquires the user image of the user in front of the smart TV 100 (shown in the path 3 in FIG. path shown).
  • the processor of the smart TV 100 determines according to the user image that the user's action instruction is: confirm to exit the child mode.
  • the processor of the smart TV 100 determines whether the user image containing the "waving hand gesture" corresponds to an adult. That is, the smart TV 100 collects portraits and gesture operations through the camera 101, and jointly detects whether it is an adult performing the exit gesture operation.
  • the above-mentioned waving gesture can be static (single-frame analysis is sufficient), or dynamic (multi-frame coherent analysis is required).
  • the processor determines that the user is an adult according to the above-mentioned action instruction of "confirming to exit the child mode" and from the user image. If the processor determines that the exit gesture operation is performed by an adult, the condition for confirming to exit the children's mode is met, and the smart TV 100 is switched from the children's mode to the normal mode. If the conditions for confirming to exit the children's mode are not satisfied, the smart TV 100 does not perform the operation of exiting the children's mode, and continues the previous play, etc. User identification is accurate and it is not easy to be bypassed by children.
  • the camera 101 of the smart TV 100 completes the collection of the user's image described in the above-mentioned embodiment.
  • the smart TV 100 combines voice and portrait detection to determine whether a wake-up condition for exiting the children's mode is met.
  • the user makes a specific gesture (such as the waving gesture of goodbye) according to the prompt of the smart TV 100, and the smart TV 100 jointly judges the portrait captured by the camera 101 and the associated body gesture to determine whether the exit condition is met. Satisfy the conditions to complete the exit, otherwise continue to play before waking up.
  • a specific gesture such as the waving gesture of goodbye
  • the smart TV 100 realizes switching from the child mode to the normal mode, and the system mode of the smart TV 100 is switched from the first system mode to the second system mode.
  • the smart TV 100 exits the children's mode, and the system mode of the smart TV 100 is switched from the children's mode to the normal mode.
  • the source range of the smart TV 100 for adults to watch is not limited.
  • an adult sits on a sofa and watches a video of a person crossing a cliff.
  • the above-mentioned body movement for the user to confirm exiting the child mode is a hand waving gesture.
  • the gestures corresponding to the above body movements are, for example, two-finger gestures, one-finger sliding gestures, page-turning gestures, two-finger pinching movements, left-to-right gesture, right to left gesture, up gesture, down gesture, press gesture, open clenched fist, tap gesture, clap gesture, reverse clap gesture, fisted hand, pinch gesture, reverse pinch gesture, Spread finger gesture, Reverse spread finger gesture, etc.
  • the above gestures are all dynamic actions, that is, the generated action instructions are based on the dynamic actions.
  • the present application is not limited thereto.
  • the above action instructions are based on static actions.
  • the user makes a static gesture: a thumbs up gesture, and remains still.
  • the camera 101 of the smart TV 100 collects the static action, and determines that the action command is: confirm to exit the child mode.
  • the camera 101 acquires the user image of the user in front of the smart TV 100, and confirms that the user image contains a "waving hand gesture" (shown by 4 path in FIG. 6 ).
  • the processor of the smart TV 100 determines according to the user image that the user's action instruction is: confirm to exit the child mode. And determine whether the user image containing the "waving hand gesture" corresponds to an adult. That is, the smart TV 100 collects portraits and gesture operations through the camera 101, and jointly detects whether it is an adult performing an exit gesture operation.
  • the display screen of the smart TV 100 does not generate a confirmation window, and the processor of the smart TV 100, according to the above-mentioned action instruction of "confirming and exiting the children's mode" and the user determined from the user image as an adult, satisfies the conditions for confirming and exiting the children's mode.
  • the television 100 switches from the child mode to the normal mode.
  • the awakening of exiting the children’s mode is completed through double joint recognition of voice and adult portrait; the confirmation of exiting the children’s mode is completed through the joint recognition of adult portrait and specific body gestures.
  • the existing physical devices without increasing the physical cost, it is more accurate and convenient for the personnel who meet the identity verification to quickly complete the exit of the child mode.
  • the first system mode is not limited to the children's mode.
  • the first system mode is a guest mode, such as a guest mode of a mobile phone.
  • the switching method of the above-mentioned system mode is also applicable to exiting the guest mode.
  • the trigger method for exiting the visitor mode double recognition of portrait + voice; the method for confirming the exit of the visitor mode: joint detection of adult portrait + associated body gestures.
  • the above-mentioned embodiment describes switching from the first system mode taking the child mode as an example to the second system mode taking the normal mode as an example. In some possible implementation manners, it may be switched from the first system mode exemplified by the normal mode to the second system mode exemplified by the child mode.
  • the method for system switching mode of the present application is applicable to a scenario involving switching from a low-privilege system mode to a high-privilege system mode.
  • Switching from the first system mode to the second system mode requires higher permissions.
  • the first system mode will not be easily switched to the second system mode.
  • Certain authority verification is required to switch from the first system mode to the second system mode.
  • the trigger method for switching from the first system mode to the second system mode dual recognition of portrait + voice; the method of confirming the switch from the first system mode to the second system mode: joint detection of adult portrait + associated body gestures. This mode switching process is convenient and not easily bypassed.
  • Electronic device 400 may include one or more processors 401 coupled to controller hub 403 .
  • the controller hub 403 communicates with a multi-branch bus such as a Front Side Bus (FSB, Front Side Bus), a point-to-point interface such as a QuickPath Interconnect (QPI, QuickPath Interconnect), or a similar connection 406.
  • a multi-branch bus such as a Front Side Bus (FSB, Front Side Bus)
  • QPI QuickPath Interconnect
  • QPI QuickPath Interconnect
  • Processor 401 communicates.
  • Processor 401 executes instructions that control data processing operations of a general type.
  • the controller hub 403 includes, but is not limited to, a graphics memory controller hub (GMCH, Graphics & Memory Controller Hub) (not shown) and an input/output hub (IOH, Input Output Hub) (which can be in on a separate chip) (not shown), where the GMCH includes the memory and graphics controller and is coupled to the IOH.
  • GMCH graphics memory controller hub
  • IOH Input Output Hub
  • Electronic device 400 may also include coprocessor 402 and memory 404 coupled to controller hub 403 .
  • one or both of the memory and the GMCH may be integrated within the processor (as described in this application), with the memory 404 and coprocessor 402 directly coupled to the processor 401 and the controller hub 403, which 403 and IOH are in a single chip.
  • the memory 404 may be, for example, a dynamic random access memory (DRAM, Dynamic Random Access Memory), a phase change memory (PCM, Phase Change Memory), or a combination of the two.
  • DRAM Dynamic Random Access Memory
  • PCM Phase Change Memory
  • One or more tangible, non-transitory computer-readable media for storing data and/or instructions may be included in memory 404 .
  • Computer-readable storage media have stored therein instructions, and in particular, temporary and permanent copies of the instructions.
  • the instructions may include: instructions that cause the electronic device 400 to implement the method shown in FIG. 4 when executed by at least one of the processors. When the instruction is run on the computer, the computer is made to execute the method disclosed in any one or combined embodiments above.
  • the coprocessor 402 is a special-purpose processor, such as, for example, a high-throughput MIC (Many Integrated Core, integrated many cores) processor, a network or communication processor, a compression engine, a graphics processor, a GPGPU (General- purpose computing on graphics processing units, general computing on graphics processing units), or embedded processors, etc.
  • a high-throughput MIC Many Integrated Core, integrated many cores
  • a network or communication processor such as, a network or communication processor, a compression engine, a graphics processor, a GPGPU (General- purpose computing on graphics processing units, general computing on graphics processing units), or embedded processors, etc.
  • GPGPU General- purpose computing on graphics processing units, general computing on graphics processing units
  • embedded processors etc.
  • the optional nature of coprocessor 402 is shown in Figure 8 with dashed lines.
  • the electronic device 400 may further include a network interface (NIC, Network Interface Controller) 406.
  • the network interface 406 may include a transceiver for providing a radio interface for the electronic device 400 to communicate with any other suitable devices (such as front-end modules, antennas, etc.).
  • network interface 406 may be integrated with other components of electronic device 400 .
  • the network interface 406 can implement the functions of the communication unit in the above-mentioned embodiments.
  • the electronic device 400 may further include an input/output (I/O, Input/Output) device 405 .
  • I/O 405 may include: a user interface designed to enable a user to interact with electronic device 400; a peripheral component interface designed to enable peripheral components to also interact with electronic device 400; and/or sensors designed to determine environmental conditions and/or location information.
  • Fig. 8 is only exemplary. That is, although it is shown in FIG. 8 that the electronic device 400 includes multiple components such as a processor 401, a controller hub 403, and a memory 404, in actual applications, the devices using the methods of the present application may only include the electronic device 400 Some of the devices, for example, may only include the processor 401 and the network interface 406 . The properties of optional components in Figure 8 are shown with dashed lines.
  • SoC 500 includes: interconnection unit 550, which is coupled to processor 510; system agent unit 580; bus controller unit 590; integrated memory controller unit 540; , which may include integrated graphics logic, image processor, audio processor and video processor; static random access memory (SRAM, Static Random-Access Memory) unit 530; direct memory access (DMA, Direct Memory Access) unit 560 .
  • interconnection unit 550 which is coupled to processor 510; system agent unit 580; bus controller unit 590; integrated memory controller unit 540; , which may include integrated graphics logic, image processor, audio processor and video processor; static random access memory (SRAM, Static Random-Access Memory) unit 530; direct memory access (DMA, Direct Memory Access) unit 560 .
  • SRAM Static Random-Access Memory
  • DMA Direct Memory Access
  • the coprocessor 520 includes a special-purpose processor, such as, for example, a network or communication processor, a compression engine, a GPGPU (General-purpose computing on graphics processing units, general-purpose computing on graphics processing units), a high-throughput MIC processor, or embedded processor, etc.
  • a special-purpose processor such as, for example, a network or communication processor, a compression engine, a GPGPU (General-purpose computing on graphics processing units, general-purpose computing on graphics processing units), a high-throughput MIC processor, or embedded processor, etc.
  • Static random access memory (SRAM) unit 530 may include one or more tangible, non-transitory computer-readable media for storing data and/or instructions.
  • Computer-readable storage media have stored therein instructions, and in particular, temporary and permanent copies of the instructions.
  • the instructions may include instructions that, when executed by at least one of the processors, cause the SoC to implement the method shown in FIG. 4 .
  • the computer is made to execute the methods disclosed in the above-mentioned embodiments.
  • the method implementations of the present application can be implemented in software, magnetic components, firmware, and the like.
  • Program code can be applied to input instructions to perform the functions described herein and generate output information.
  • the output information may be applied to one or more output devices in known manner.
  • a processing system includes any system having a processor such as, for example, a Digital Signal Processor (DSP), microcontroller, application specific integrated circuit (ASIC), or microprocessor.
  • DSP Digital Signal Processor
  • ASIC application specific integrated circuit
  • the program code can be implemented in a high-level procedural language or an object-oriented programming language to communicate with the processing system.
  • Program code can also be implemented in assembly or machine language, if desired.
  • the mechanisms described in this paper are not limited in scope to any particular programming language. In either case, the language may be a compiled or interpreted language.
  • One or more aspects of at least one embodiment can be implemented by representative instructions stored on a computer-readable storage medium, the instructions representing various logic in a processor, the instructions, when read by a machine, cause the machine to make The logic that implements the techniques of this article.
  • These representations known as "IP (Intellectual Property) cores," may be stored on a tangible computer-readable storage medium and provided to multiple customers or production facilities to load into the actual manufacturing of the logic or processor. in the manufacturing machine.
  • an instruction converter may be used to convert instructions from a source instruction set to a target instruction set.
  • an instruction converter may transform (eg, using static binary translation, dynamic binary translation including dynamic compilation), warp, emulate, or otherwise convert an instruction into one or more other instructions to be processed by the core.
  • the instruction converter can be implemented in software, hardware, firmware, or a combination thereof.
  • the instruction converter can be on-processor, off-processor, or partly on-processor and partly off-processor.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本申请涉及电子设备领域,公开了一种系统模式切换方法、电子设备及计算机可读存储介质。方法包括:运行第一系统模式;获取电子设备的用户的语音,并从语音中解析出指令;确定指令中包括预设的用于切换第一系统模式的唤醒词;获取针对电子设备前的用户采集的用户图像,并根据用户图像确定用户包括预设用户类型的用户;根据指令,并根据用户图像以及确定的预设用户类型的用户,将电子设备从第一系统模式切换为与预设用户类型对应的第二系统模式。这有利于提升切换该电子设备的系统模式的操作便捷度和准确性。

Description

一种系统模式切换方法、电子设备及计算机可读存储介质
本申请要求于2021年07月07日提交中国专利局、申请号为202110767081.8、申请名称为“一种系统模式切换方法、电子设备及计算机可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及电子设备领域,特别涉及一种系统模式切换方法、电子设备及计算机可读存储介质。
背景技术
随着智能家居设备的普及,智能电视已经进入到了千家万户。智能电视的系统模式一般包括普通模式和儿童模式。
当家中儿童使用智能电视时,可以将智能电视的系统模式由普通模式切换为儿童模式,即智能电视退出普通模式(启动儿童模式),这样可以控制家中儿童观看电视片源范围、时长、坐姿等。当家中成人使用智能电视时,可以将智能电视的系统模式由儿童模式切换为普通模式,即智能电视退出儿童模式(启动普通模式)。
上述智能电视退出儿童模式时存在诸多问题,例如儿童易仿照大人完成退出、老人不易操作、声纹识别错误率高等。
发明内容
本申请的实施例提供一种系统模式切换方法,用于提升切换电子设备的系统模式的操作便捷度和准确性。
为达到上述目的,本申请的实施例采用如下技术方案:
第一方面,本申请实施例公开了一种系统模式切换方法,应用于电子设备(例如智能电视),方法包括:运行第一系统模式;获取电子设备的用户的语音,并从语音中解析出指令;确定指令中包括预设的用于切换第一系统模式的唤醒词;获取针对电子设备前的用户采集的用户图像,并根据用户图像确定用户包括预设用户类型的用户;根据指令,并根据用户图像以及确定的预设用户类型的用户,将电子设备从第一系统模式切换为与预设用户类型对应的第二系统模式。
上述的第一系统模式以儿童模式为示例说明,第二系统模式以普通模式为示例说明。根据本申请的实施方式,通过识别用户的语音生成触发退出儿童模式的指令,并通过识别用户图像确定有成人在场,从而,根据退出儿童模式的指令以及确定有成人在场,即可完成智能电视的儿童模式的退出操作。从而,儿童模式退出操作路径短,易用性较高。并且,用户身份识别准确,不易被儿童绕过。
在一些可能的实施方式中,上述的系统模式切换方法适用于手机访客模式的退出。
在上述第一方面的一种可能实现中,用户图像包括第一用户图像和第二用户图像;
根据指令,并根据用户图像以及确定的预设用户类型的用户,将电子设备从第一系统模式切换为与预设用户类型对应的第二系统模式,包括:
根据指令和从第一用户图像中确定的预设用户类型的用户,生成确认窗口;
响应于确认窗口的生成,获取针对电子设备前的用户采集的用户图像作为第二用户图像,根据第二用户图像确定动作指令;
根据动作指令和从第二用户图像中确定的预设用户类型的用户,将电子设备从第一系统模式切换为与预设用户类型对应的第二系统模式。
在上述第一方面的一种可能实现中,动作指令是基于静态的动作。
在上述第一方面的一种可能实现中,动作指令是基于动态的动作,动态的动作包括用户的肢体动作。
在上述第一方面的一种可能实现中,第一用户图像包括静态的图像,根据静态的图像确定用户包括预设用户类型的用户。
在上述第一方面的一种可能实现中,预设用户类型的用户是语音的发出者。
在上述第一方面的一种可能实现中,根据指令,并根据用户图像以及确定的预设用户类型的用户,将电子设备从第一系统模式切换为与预设用户类型对应的第二系统模式,包括:
根据用户图像确定动作指令;
根据指令,并根据确定的预设用户类型的用户和动作指令,将电子设备从第一系统模式切换为与预设用户类型对应的第二系统模式。
在上述第一方面的一种可能实现中,用户图像包括静态的图像,根据静态的图像确定用户包括预设用户类型的用户。
在上述第一方面的一种可能实现中,用户图像包括动态的图像,动态的图像包括用户的肢体动作。
在上述第一方面的一种可能实现中,肢体动作包括挥手手势。
在上述第一方面的一种可能实现中,预设用户类型的用户为成人,第一系统模式为儿童模式,第二系统模式为普通模式。
第二方面,本申请的实施例提供了一种电子设备,包括:拾音器,用于采集用户的语音;摄像头,用于采集用户的用户图像;处理器;存储器,存储器包括指令,指令被处理器执行时,使得电子设备执行上述第一方面任一实现方式所提供的系统模式切换方法。
第三方面,本申请的实施例提供了一种计算机可读存储介质,计算机可读存储介质上存储有指令,该指令在计算机上执行时使得计算机执行上述第一方面任一实现方式所提供的系统模式切换方法。
附图说明
图1根据本申请的一些实施例,示出了智能家居的场景示意图一;
图2根据本申请的一些实施例,示出了智能电视退出儿童模式的流程示意图;
图3为本申请实施方式提供的电子设备的构造示意图;
图4示出了本申请一个实施例提供的系统模式切换方法的流程图;
图5根据本申请的一些实施例,示出了智能家居的场景示意图二;
图6根据本申请的一些实施例,示出了智能家居的场景示意图三;
图7根据本申请的一些实施例,示出了智能家居的场景示意图四;
图8示出了本申请一个实施例提供的电子设备的框图;
图9示出了本申请一个实施例提供的一种片上系统(SoC)的框图。
具体实施方式
以下将参考附图详细说明本申请的具体实施方式。
本申请提供了一种系统模式切换方法,用于电视等电子设备,有利于提升切换该电子设备的系统模式的操作便捷度和准确性。
图1示出了一种包含多个电子设备的智能家居场景示例。在图1所示的场景中,具体电子设备有:智能电视100和音箱200。
根据本申请的一些实施例,智能电视100能够通过网络实现与云端或者其他设备交互(例如与音箱200交互)。并且,智能电视100可以搭载操作系统,使用户在欣赏普通电视内容的同时,可自行安装和卸载各类应用软件,持续对功能进行扩充和升级。例如,智能电视100可以具备多种方式的交互式应用,如人机交互方式、多屏互动、内容共享等。需要说明的是,本申请各实施例中的电视可以是上述的智能电视100,也可以是具有较大屏幕的智慧屏。
根据本申请的一些实施例,音箱200可以是智能音箱。示例性的,智能音箱200具有拾音功能,能够采集用户的语音。
根据本申请的一些实施例,上述的智能电视100的系统模式包括普通模式和儿童模式。根据使用智能电视100的用户的类型(成人或儿童)智能电视100启动相应的系统模式。当儿童使用智能电视100时,智能电视100切换为儿童模式。如图1所示,智能电视100处于儿童模式,家中的儿童坐在沙发上观看儿童踢足球的视频。
当成人使用智能电视100时,智能电视100退出儿童模式,即智能电视100的系统模式由儿童模式切换为普通模式。参考图2,根据一种智能电视100退出儿童模式的方案,流程主要分为以下三部分:
(1)唤醒触发
用户(例如成人)通过智能电视100的遥控器退出智能电视100的当前视频播放后,选择智能电视100上的退出儿童模式图标,或者,用户语音唤醒触发退出儿童模式,智能电视100的显示屏上显示退出儿童模式的确认弹框。
(2)校验
用户阅读退出儿童模式的确认弹框上的确认信息,并进行用户身份认证。其中,用户身份认证,目前主要有三种形式:①用户预设的用户身份密码;②简单的数学计算题;③已录制声纹的验证。
(3)响应
若用户身份认证满足校验要求,智能电视100退出儿童模式。否则,智能电视100不做退出儿童模式处理。
由上可知,上述方案中,智能电视100退出儿童模式的流程主要存在以下问题: (1)退出儿童模式的操作路径较长;(2)用户身份的认证方式(第①种、第②种身份认证形式)易被儿童绕过;(3)声纹识别成功率较低(第③种身份认证形式);(4)依赖遥控器操作;(5)老人操作相对困难。
总而言之,智能电视100退出儿童模式的流程如存在以上问题,则会导致操作不方便、准确性不高的用户体验问题。
为此,本申请提供一种系统模式切换方法,通过识别用户的语音生成触发退出儿童模式的指令,并通过识别用户图像确定有成人在场,从而,根据退出儿童模式的指令以及确定有成人在场,即可完成智能电视的儿童模式的退出操作。从而,儿童模式退出操作路径短,易用性较高。并且,用户身份识别准确,不易被儿童绕过。
在如图1所示的场景中,智能电视100作为电子设备的本体的示例被提供。但是本申请不限于此,电子设备可以是带有摄像头101和拾音器的各种电子设备。例如,各种智能家居设备(例如图1中示出的音箱200、智能闹钟等)、手机、平板电脑、台式计算机、膝上型计算机、车载终端、人工智能(artificial intelligence,AI)智能语音终端、可穿戴设备、增强现实(augmented reality,AR)设备、虚拟现实(virtual reality,VR)设备、超级移动个人计算机(ultra-mobile personal computer,UMPC)、手持计算机、上网本、个人数字助理(personal digital assistant,PDA)等各种电子设备。电子设备的示例性实施例包括但不限于搭载IOS、Android、Microsoft、HarmonyOS或者其它操作系统的各种电子设备。
图3示出了根据本申请一实施例的电子设备001的结构示意图。上述实施例中的智能电视100的结构可以与该电子设备001相同。具体地,电子设备001可以包括:
显示屏102:具有显示界面。显示界面用来播放视频或者显示电脑或手机等的投屏画面等。例如,显示屏102显示儿童模式的图标,并播放上述实施例中所描述的儿童踢足球的视频。示例性的,显示屏102可以为电容式触摸屏。
处理器110:可以包括一个或多个处理单元,例如,可以包括中央处理器CPU(Central Processing Unit)、图像处理器GPU(Graphics Processing Unit)、数字信号处理器DSP、微处理器MCU(Micro-programmed Control Unit)、AI(Artificial Intelligence,人工智能)处理器或可编程逻辑器件FPGA(Field Programmable Gate Array)等的处理模块或处理电路。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。处理器能够根据识别的用户的语音和用户图像,完成系统模式的切换。
处理器110中可以设置存储器180或缓存器,用于存储指令和数据。可实施的,存储器180或缓存器中可以包括用户图像和用户类型的对应关系,或者可以包括用户的语音、用户类型与系统模式的对应关系。上述对应关系可以预先存储在存储器180中,处理器110在识别到用户的语音和用户图像后,从存储器180查找是否有对应的系统模式,如果有,则处理器110进行系统模式的切换。
电源模块140可以包括电源、电源管理部件等。电源可以为电池。电源管理部件用于管理电源的充电和电源向其他模块的供电。在一些实施例中,电源管理部件包括充电管理模块和电源管理模块。充电管理模块用于从充电器接收充电输入;电源管理模块用于连接电源,充电管理模块与处理器110。电源管理模块接收电源和/或充电管 理模块的输入,为处理器110,显示屏102,摄像头170,及无线通信模块120等供电。
无线通信模块120可以包括天线,并经由天线实现对电磁波的收发。无线通信模块120可以提供应用在电子设备001上的包括无线局域网(wireless localarea networks,WLAN)(如无线保真(wireless fidelity,Wi-Fi)网络),蓝牙(bluetooth,BT),全球导航卫星系统(global navigation satellite system,GNSS),调频(frequency modulation,FM),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。电子设备001可以通过无线通信技术与网络以及其他设备(例如手机、音箱200)进行通信。
音频模块150用于将数字音频信息转换成模拟音频信号输出,或者将模拟音频输入转换为数字音频信号。音频模块150还可以用于对音频信号编码和解码。在一些实施例中,音频模块150可以设置于处理器110中,或将音频模块150的部分功能模块设置于处理器110中。在一些实施例中,音频模块150可以包括扬声器、听筒、拾音器以及耳机接口。
摄像头170用于捕获静态图像或动态图像。物体通过镜头生成光学图像投射到感光元件。感光元件把光信号转换成电信号,之后将电信号传递给ISP(Image Signal Processing,图像信号处理)转换成数字图像信号。电子设备001可以通过ISP,摄像头170,视频编解码器,GPU(Graphic Processing Unit,图形处理器),显示屏102以及应用处理器等实现拍摄功能。
接口模块160包括外部存储器接口、通用串行总线(universal serial bus,USB)接口等。其中外部存储器接口可以用于连接外部存储卡,例如Micro SD卡,实现扩展电子设备001的存储能力。外部存储卡通过外部存储器接口与处理器110通信,实现数据存储功能。通用串行总线接口用于电子设备001和其他电子设备进行通信。
在一些实施例中,电子设备001还包括按键101。其中,按键101可以包括音量键、开/关机键等。
在一些实施例中,电子设备001还包括触控检测装置190。其中,触控检测装置190可以检测用户的触摸点位置,根据用户的触摸点位置识别对应的触摸手势。
可以理解的是,本发明实施例示意的结构并不构成对电子设备001的具体限定。在本申请另一些实施例中,电子设备001可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
以下将以智能电视100作为电子设备举例,结合附图详细阐述本申请实施例提供的一种系统模式切换方法。本申请提供的一种系统模式切换方法,通过识别用户的语音和用户图像即可完成智能电视的儿童模式的退出操作。从而,儿童模式退出操作路径短,易用性较高。并且,用户身份识别准确,不易被儿童绕过。
以下结合图4所示的流程图,详细介绍智能电视100的系统模式切换方法。
具体地,如图4所示,本实施例提供的方法包括以下步骤:
S100:智能电视运行儿童模式。
如图5所示,智能电视100当前的系统模式为儿童模式(即第一系统模式)。智能电视100运行儿童模式时,儿童观看智能电视100片源范围、时长、坐姿等受到控 制。例如,智能电视100运行儿童模式时,儿童需坐着看视频,看的是儿童类的视频,看视频的时长控制为30分钟。那么,儿童坐在沙发上,可以正常观看智能电视100播放的儿童踢足球的视频,时长为30分钟。当儿童躺着看视频时,智能电视100的摄像头101检测到儿童没有坐在沙发上,则智能电视100停止播放当前的视频。又例如,儿童观看当前视频的时长达到30分钟时,智能电视100也停止播放当前的视频。
S200:智能电视获取用户的语音,并从语音中解析出指令。
当成人需要使用智能电视100观看视频或者儿童看完视频后,成人对智能电视100进行系统模式的切换时,成人进行退出儿童模式的语音唤醒。例如,如图5所示,成人说出“小艺小艺,退出儿童模式”。“小艺小艺,退出儿童模式”就是用户发出的唤醒语音(唤醒词+命令语)。其中,唤醒词是“小艺小艺”,命令语是“退出儿童模式”。
示例性的,如图5所示,上述的唤醒语音被与智能电视100通信连接的其它电子设备例如音箱200所获取(图5中①路径所示)。音箱200可以将这段唤醒语音传输给智能电视100,由智能电视100进行唤醒语音的解析。或者,音箱200自行进行唤醒语音的解析,然后将解析出的指令发送给智能电视100。无论是音箱200还是智能电视100对这段唤醒语音进行解析,最终从这段唤醒语音中解析出的指令是:智能电视100退出儿童模式。
另外,示例性的,唤醒语音也可以是被与智能电视100通信连接的其他电子设备获取,例如通过蓝牙方式连接的遥控器、或者可穿戴设备等。唤醒语言的解析的过程和上述利用音箱200的方式一样。另外,唤醒语音也可以是直接由智能电视100中内置的拾音器获取。
上述的唤醒语音可以是用户根据自身需要预先设置在电子设备中的,也可以是电子设备在出厂之前设置好的,本申请的实施例中对唤醒语音的设置方式不作限定。
S300:智能电视确定指令中包括预设的用于切换儿童模式的唤醒词。
示例性的,上述指令中的“退出儿童模式”是切换儿童模式的唤醒词。
S400:智能电视获取用户图像,并根据用户图像确定用户包括成人。
示例性地,智能电视100确认获取的用户图像对应的用户和发出唤醒语音的用户是同一个人。
在用户发出唤醒语音后,智能电视100的摄像头101通过检测声源人像获取用户图像,即摄像头101对发出唤醒语音的用户进行图像采集(图5中②路径所示)。其中,摄像头101获取的是智能电视100前的用户的用户图像。“智能电视100前”是指“智能电视100的摄像头101的摄像范围可及之处”。智能电视100的处理器对摄像头101获取的用户图像进行判断,确定发出唤醒语音的用户是成人,而非儿童。则,智能电视100满足触发退出儿童模式的唤醒操作条件。
相当于,智能电视100的处理器结合接收到唤醒语音和采集的人像数据,一起进行计算和判断是否满足触发退出儿童模式场景的条件。也即,当发出唤醒语音的用户是由成人进行的,则满足触发退出儿童模式场景的条件。
上述“摄像头101对发出唤醒语音的用户进行图像采集”例如是通过如下方式实现:在智能电视100获取到用户发出的唤醒语音后,智能电视100对用户的唤醒语音 进行声源识别,以判断出用户在发出语音时的所处位置,具体为方位角度。即,发出唤醒语音的用户是位于声音采集器(例如上述的音箱200)的左侧、右侧还是正对声音采集器的位置等,进而根据用户的所处位置调整智能电视100的摄像头101的拍摄角度,以准确对于发出唤醒语音的用户进行图像采集。
在一些可能的实施方式中,上述“摄像头101对发出唤醒语音的用户进行图像采集”例如是通过如下方式实现:智能电视100的摄像头101实时捕捉图像信息,智能电视100的处理器对捕捉的图像信息根据人脸识别技术进行人像识别,在确定为人像时,再进行张嘴识别,在确定为识别结果为张嘴时(“张嘴”说明是在发出唤醒语音),获取该张嘴的人像,以完成对发出唤醒语音的用于进行图像采集。
示例性的,上述摄像头101针对发出唤醒语音的用户采集的用户图像是人脸,可以使用人脸检测技术区分发出唤醒语音的用户是成人还是小孩。具体是处理器基于图像的AI人脸识别。示例性的,AI人脸识别可以自动提取人脸特征值,人脸特征值例如包括皮肤纹理、皮肤颜色、光亮程度和皱纹纹理等方面。通过人脸特征值来确认人像是成人或小孩。示例性的,上述的用户图像(人脸)是静态的图像,处理器根据静态的图像确定用户包括成人(预设用户类型的用户)。
另外,示例性的,获取的用户图像对应的用户和发出唤醒语音的用户不必须是同一个人。
在一些可能的实施方式中,在用户发出唤醒语音后,智能电视100的摄像头101采集智能电视前的用户的用户图像,处理器确认摄像头101采集的用户图像中是否包含成人,这个成人可以是发出唤醒语音的人,也可以不是发出唤醒语音的人,只要确认是否有成人在场。示例性的,处理器可以通过上述的AI人脸识别确认用户图像对应的用户是成人或小孩。或者,在一些可能的实施方式中,用户处于静止的站立姿势,智能电视100的摄像头101获取静止的站立姿势的用户图像,处理器根据用户图像对应的站立姿势的高度,确定该用户图像对应的用户为成人或小孩。
当智能电视100的处理器确定用户发出了唤醒语音,并且智能电视100前的用户包括成人,即有成人在场监管退出儿童模式的操作,则满足触发退出儿童模式场景的条件。
此外,在一些可能的实施方式中,在用户发出唤醒语音后,智能电视100的处理器通过声纹识别技术确认发出唤醒语音的用户是否为成人,同时在用户发出唤醒语音后,智能电视100的摄像头101采集智能电视前的用户的用户图像,处理器确认摄像头101采集的用户图像中是否包含成人。处理器确认用户图像中是否包含成人的过程和上述任一实施例中所描述的确认是否包含成人的方式相同。
声纹识别的理论基础是每一个声音都具有独特的特征,通过该特征能将不同人的声音进行有效的区分。例如,这种独特的特征例如是声腔的尺寸,具体包括咽喉、鼻腔和口腔等,这些器官的形状、尺寸和位置决定了声带张力的大小和声音频率的范围。即成人和儿童的声带张力的大小和声音频率的范围是不同的。因此,可以通过声纹识别技术识别所采集的用户的声带张力的大小和声音频率,以确认发出唤醒语音的用户是否为成人。
S500:智能电视的处理器根据指令,并根据用户图像以及确定的成人用户,将智 能电视100从儿童模式切换为普通模式。
示例性的,智能电视100的处理器根据上述“退出儿童模式”的指令以及确定是成人发出的指令,成功触发退出儿童模式的场景。在此场景下,根据用户的图像,将智能电视100从儿童模式切换为普通模式(即第二系统模式)。即,完成了智能电视100退出儿童模式的操作。
或者,在一些可能的实施方式中,智能电视100的处理器确认有成人发出上述“退出儿童模式”的指令,以及上述实施例所描述的确定智能电视100前有成人(该成人不一定是发出“退出儿童模式”的成人)在场,则将智能电视100从儿童模式切换为普通模式。
或者,在一些可能的实施方式中,智能电视100的处理器确认有用户(成人或儿童)发出上述“退出儿童模式”的指令,以及上述实施例所描述的确定智能电视100前有成人(该成人不一定是发出“退出儿童模式”的成人)在场,将智能电视100从儿童模式切换为普通模式。
由于是通过成人图像和成人发出的唤醒语音来进行退出儿童模式的操作,能够确保有成人在场监管儿童模式的退出。因此,本申请的系统模式的切换安全性和准确性高。并且,方便用户进行智能电视100系统模式的切换。而且,本申请所涉及物理元器件均为智能电视100已有的元器件,没有新增物理元器件等。
上述用户图像的获取是通过智能电视100自带的摄像头101获取的,但本申请不限于此,在一些可能的实施方式中,可以通过与智能电视100通信连接的其它电子设备的摄像头101进行用户图像的获取。
如上所述,在利用静态的用户图像进行成人判断的情况下,成人的静态图像存在仿造的可能性,例如人性纸板、模型等,就会导致上述方案容易绕过。为了防止出现因仿造成人的静态图像,而成功触发退出儿童模式的情形出现,结合图6,在一些可能的实施方式中,需结合成人用户的确认动作,才能成功触发退出儿童模式。
示例性的,上述智能电视100满足触发退出儿童模式的唤醒操作条件后,参考图6,智能电视100的显示屏上会生成确认窗口(如:屏幕中间位置)。即如上述S400中所述描述的,在成人发出唤醒语音后,并确定摄像头101获取的用户图像是成人,说明满足触发退出儿童模式场景的条件。则,智能电视100上会显示确认窗口。如图6所示,确认窗口中显示“请面向屏幕挥挥手确认退出儿童模式”以及“手势图样”。相当于,智能电视100通过确认窗口提示用户进行确认退出儿童模式的操作。
此时,用户根据确认窗口的提示,做挥手手势。即,用户的肢体动作是挥手手势。智能电视100响应于确认窗口的生成,获取智能电视100前的用户的用户图像(图6中③路径所示),智能电视100的处理器确认用户图像中包含“挥手手势”(图6中④路径所示)。智能电视100的处理器根据该用户图像确定用户的动作指令是:确认退出儿童模式。并且智能电视100的处理器确定包含“挥手手势”的用户图像对应的是否是成人。即智能电视100通过摄像头101收集人像和手势操作,联合检测是否是成人进行退出手势操作。
上述的挥手手势可以是静态的(单帧分析即可),也可以是动态的(需要多帧连贯地分析)。
从而,处理器根据上述“确认退出儿童模式”的动作指令以及从用户图像中确定的用户为成人。处理器确定是成人进行退出手势操作,则满足确认退出儿童模式条件,将智能电视100从儿童模式切换为普通模式。若不满足确认退出儿童模式条件,智能电视100不做退出儿童模式操作,继续之前的播放等。用户身份识别准确,不易被儿童绕过。
相当于,本实施例中,用户通过特定唤醒词(例如“小艺小艺,退出儿童模式”)唤醒退出儿童模式的同时,智能电视100的摄像头101完成上述实施例所描述的用户图像的采集,智能电视100结合语音和人像检测,判断是否满足退出儿童模式的唤醒条件。
若满足唤醒条件,用户根据智能电视100提示,做特定手势(如:再见的挥手手势),智能电视100联合判断摄像头101采集的人像和关联的肢体手势,判断是否满足退出条件。满足条件完成退出,否则继续唤醒前的播放等。
即,通过成人人像以及智能电视100语音唤醒的联合识别,完成退出儿童模式的唤醒;通过成人人像以及相关联的退出手势,联合识别并完成儿童模式的退出。智能电视100实现了由儿童模式切换为普通模式,及智能电视100的系统模式由第一系统模式切换为第二系统模式。
如图7所示,智能电视100退出儿童模式,智能电视100的系统模式由儿童模式切换为普通模式。在普通模式下,成人观看智能电视100的片源范围不受限,如图7所示,成人坐在沙发上观看人跨越悬崖的视频。
需说明的是,上述的用户确认退出儿童模式的肢体动作是挥手手势。但本申请不限于此,在一些可能的实施方式中,上述的肢体动作对应的手势例如是双指并拢移动手势、单指滑动手势、翻页手势、两个手指的对捏运动、左到右手势、右到左手势、向上手势、向下手势、按压手势、打开握紧的拳头、轻敲手势、鼓掌手势、反向鼓掌手势、手握成拳、对捏手势、反向对捏手势、张开手指的手势、反向张开手指的手势等。
上述的手势都是动态的动作,即生成的动作指令是基于动态的动作。但本申请不限于此,在一些可能的实施方式中,上述的动作指令是基于静态的动作。例如,用户做静态的动作:竖起大拇指手势,并保持不动。智能电视100的摄像头101采集该静态的动作,并确定动作指令是:确认退出儿童模式。
如前所述,上述智能电视100满足触发退出儿童模式的唤醒操作条件后,参考图6,智能电视100的显示屏上会生成确认窗口,然后用户根据确认窗口的提示做出确认退出儿童模式的“手势图样”。但本申请不限于此,在一些可能的实施方式中,上述智能电视100满足触发退出儿童模式的唤醒操作条件后,智能电视100的显示屏上不会生成确认窗口。用户在发出唤醒语音后,用户直接做确认退出儿童模式的上述“手势图样”(例如挥手手势)。
摄像头101获取智能电视100前的用户的用户图像,确认用户图像中包含“挥手手势”(图6中④路径所示)。同样,智能电视100的处理器根据该用户图像确定用户的动作指令是:确认退出儿童模式。并且确定包含“挥手手势”的用户图像对应的是否是成人。即智能电视100通过摄像头101收集人像和手势操作,联合检测是否是 成人进行退出手势操作。
从而,智能电视100的显示屏不生成确认窗口,智能电视100的处理器根据上述“确认退出儿童模式”的动作指令以及从用户图像中确定的用户为成人,满足确认退出儿童模式条件,将智能电视100从儿童模式切换为普通模式。
综上,本申请的智能电视100在儿童模式下,退出儿童模式的唤醒通过语音和成人人像双重联合识别完成;确认退出儿童模式通过成人人像和特定肢体手势联合识别完成。并且使用已有物理器件,不增加物理成本的情况下,更加精确便捷为满足身份验证的人员快速完成儿童模式的退出。
需说明的是,上述第一系统模式不限于是儿童模式。例如,在一些可能的实施方式中,第一系统模式为访客模式,例如是手机的访客模式。上述系统模式的切换方法同样适用于退出访客模式。示例性地,访客模式退出的触发方式:人像+语音双重识别;确认访客模式退出的方式:成人人像+相关联的肢体手势联合检测方式。
需说明的是,上述实施例描述的是由以儿童模式为示例的第一系统模式切换为以普通模式为示例的第二系统模式。在一些可能的实施方式中,可以是由以普通模式为示例的第一系统模式切换为以儿童模式为示例的第二系统模式。
示例性地,本申请的系统切换模式的方法适用于涉及由低权限的系统模式切换为高权限的系统模式的场景。由第一系统模式切换为第二系统模式需要的权限较高,第一系统模式不会轻易切换为第二系统模式,需要得到一定的权限验证才能实现由第一系统模式切换为第二系统模式。那么,第一系统模式切换为第二系统模式的触发方式:人像+语音双重识别;确认第一系统模式切换为第二系统模式的方式:成人人像+相关联的肢体手势联合检测方式。这种模式切换过程便捷、不易被绕过。
现在参考图8,所示为根据本申请的一个实施例的电子设备400的框图。电子设备400可以包括耦合到控制器中枢403的一个或多个处理器401。对于至少一个实施例,控制器中枢403经由诸如前端总线(FSB,Front Side Bus)之类的多分支总线、诸如快速通道连(QPI,QuickPath Interconnect)之类的点对点接口、或者类似的连接406与处理器401进行通信。处理器401执行控制一般类型的数据处理操作的指令。在一实施例中,控制器中枢403包括,但不局限于,图形存储器控制器中枢(GMCH,Graphics&Memory Controller Hub)(未示出)和输入/输出中枢(IOH,Input Output Hub)(其可以在分开的芯片上)(未示出),其中GMCH包括存储器和图形控制器并与IOH耦合。
电子设备400还可包括耦合到控制器中枢403的协处理器402和存储器404。或者,存储器和GMCH中的一个或两者可以被集成在处理器内(如本申请中所描述的),存储器404和协处理器402直接耦合到处理器401以及控制器中枢403,控制器中枢403与IOH处于单个芯片中。
存储器404可以是例如动态随机存取存储器(DRAM,Dynamic Random Access Memory)、相变存储器(PCM,Phase Change Memory)或这两者的组合。存储器404中可以包括用于存储数据和/或指令的一个或多个有形的、非暂时性计算机可读介质。计算机可读存储介质中存储有指令,具体而言,存储有该指令的暂时和永久副本。该指令可以包括:由处理器中的至少一个执行时导致电子设备400实施如图4所示方法 的指令。当指令在计算机上运行时,使得计算机执行上述任一实施例或组合实施例公开的方法。
在一个实施例中,协处理器402是专用处理器,诸如例如高吞吐量MIC(Many Integrated Core,集成众核)处理器、网络或通信处理器、压缩引擎、图形处理器、GPGPU(General-purpose computing on graphics processing units,图形处理单元上的通用计算)、或嵌入式处理器等等。协处理器402的任选性质用虚线表示在图8中。
在一个实施例中,电子设备400可以进一步包括网络接口(NIC,Network Interface Controller)406。网络接口406可以包括收发器,用于为电子设备400提供无线电接口,进而与任何其他合适的设备(如前端模块,天线等)进行通信。在各种实施例中,网络接口406可以与电子设备400的其他组件集成。网络接口406可以实现上述实施例中的通信单元的功能。
电子设备400可以进一步包括输入/输出(I/O,Input/Output)设备405。I/O405可以包括:用户界面,该设计使得用户能够与电子设备400进行交互;外围组件接口的设计使得外围组件也能够与电子设备400交互;和/或传感器设计用于确定与电子设备400相关的环境条件和/或位置信息。
值得注意的是,图8仅是示例性的。即虽然图8中示出了电子设备400包括处理器401、控制器中枢403、存储器404等多个器件,但是,在实际的应用中,使用本申请各方法的设备,可以仅包括电子设备400各器件中的一部分器件,例如,可以仅包含处理器401和网络接口406。图8中可选器件的性质用虚线示出。
现在参考图9,所示为根据本申请的一实施例的SoC(System on Chip,片上系统)500的框图。在图9中,相似的部件具有同样的附图标记。另外,虚线框是更先进的SoC的可选特征。在图9中,SoC500包括:互连单元550,其被耦合至处理器510;系统代理单元580;总线控制器单元590;集成存储器控制器单元540;一组或一个或多个协处理器520,其可包括集成图形逻辑、图像处理器、音频处理器和视频处理器;静态随机存取存储器(SRAM,Static Random-Access Memory)单元530;直接存储器存取(DMA,Direct Memory Access)单元560。在一个实施例中,协处理器520包括专用处理器,诸如例如网络或通信处理器、压缩引擎、GPGPU(General-purpose computing on graphics processing units,图形处理单元上的通用计算)、高吞吐量MIC处理器、或嵌入式处理器等。
静态随机存取存储器(SRAM)单元530可以包括用于存储数据和/或指令的一个或多个有形的、非暂时性计算机可读介质。计算机可读存储介质中存储有指令,具体而言,存储有该指令的暂时和永久副本。该指令可以包括:由处理器中的至少一个执行时导致SoC实施如图4所示方法的指令。当指令在计算机上运行时,使得计算机执行上述实施例中公开的方法。
本申请的各方法实施方式均可以以软件、磁件、固件等方式实现。
可将程序代码应用于输入指令,以执行本文描述的各功能并生成输出信息。可以按已知方式将输出信息应用于一个或多个输出设备。为了本申请的目的,处理系统包括具有诸如例如数字信号处理器(DSP,Digital Signal Processor)、微控制器、专用集成电路(ASIC)或微处理器之类的处理器的任何系统。
程序代码可以用高级程序化语言或面向对象的编程语言来实现,以便与处理系统通信。在需要时,也可用汇编语言或机器语言来实现程序代码。事实上,本文中描述的机制不限于任何特定编程语言的范围。在任一情形下,该语言可以是编译语言或解释语言。
至少一个实施例的一个或多个方面可以由存储在计算机可读存储介质上的表示性指令来实现,指令表示处理器中的各种逻辑,指令在被机器读取时使得该机器制作用于执行本文的技术的逻辑。被称为“IP(Intellectual Property,知识产权)核”的这些表示可以被存储在有形的计算机可读存储介质上,并被提供给多个客户或生产设施以加载到实际制造该逻辑或处理器的制造机器中。
在一些情况下,指令转换器可用来将指令从源指令集转换至目标指令集。例如,指令转换器可以变换(例如使用静态二进制变换、包括动态编译的动态二进制变换)、变形、仿真或以其它方式将指令转换成将由核来处理的一个或多个其它指令。指令转换器可以用软件、硬件、固件、或其组合实现。指令转换器可以在处理器上、在处理器外、或者部分在处理器上且部分在处理器外。

Claims (13)

  1. 一种系统模式切换方法,应用于电子设备,其特征在于,所述方法包括:
    运行第一系统模式;
    获取所述电子设备的用户的语音,并从所述语音中解析出指令;
    确定所述指令中包括预设的用于切换所述第一系统模式的唤醒词;
    获取针对所述电子设备前的所述用户采集的用户图像,并根据所述用户图像确定所述用户包括预设用户类型的用户;
    根据所述指令,并根据所述用户图像以及确定的预设用户类型的用户,将所述电子设备从第一系统模式切换为与所述预设用户类型对应的第二系统模式。
  2. 如权利要求1所述的系统模式切换方法,其特征在于,所述用户图像包括第一用户图像和第二用户图像;
    所述根据所述指令,并根据所述用户图像以及确定的预设用户类型的用户,将所述电子设备从第一系统模式切换为与所述预设用户类型对应的第二系统模式,包括:
    根据所述指令和从所述第一用户图像中确定的预设用户类型的用户,生成确认窗口;
    响应于所述确认窗口的生成,获取针对所述电子设备前的所述用户采集的用户图像作为所述第二用户图像,根据所述第二用户图像确定动作指令;
    根据所述动作指令和从所述第二用户图像中确定的预设用户类型的用户,将所述电子设备从第一系统模式切换为与所述预设用户类型对应的第二系统模式。
  3. 如权利要求2所述的系统模式切换方法,其特征在于,所述动作指令是基于静态的动作。
  4. 如权利要求2所述的系统模式切换方法,其特征在于,所述动作指令是基于动态的动作,所述动态的动作包括所述用户的肢体动作。
  5. 如权利要求2所述的系统模式切换方法,其特征在于,所述第一用户图像包括静态的图像,根据所述静态的图像确定所述用户包括预设用户类型的用户。
  6. 如权利要求1所述的系统模式切换方法,其特征在于,所述预设用户类型的用户是所述语音的发出者。
  7. 如权利要求1所述的系统模式切换方法,其特征在于,所述根据所述指令,并根据所述用户图像以及确定的预设用户类型的用户,将所述电子设备从第一系统模式切换为与所述预设用户类型对应的第二系统模式,包括:
    根据所述用户图像确定动作指令;
    根据所述指令,并根据确定的预设用户类型的用户和所述动作指令,将所述电子设备从第一系统模式切换为与所述预设用户类型对应的第二系统模式。
  8. 如权利要求7所述的系统模式切换方法,其特征在于,所述用户图像包括静态的图像,根据所述静态的图像确定所述用户包括预设用户类型的用户。
  9. 如权利要求7所述的系统模式切换方法,其特征在于,所述用户图像包括动态的图像,所述动态的图像包括所述用户的肢体动作。
  10. 如权利要求4或9所述的系统模式切换方法,其特征在于,所述肢体动作包括挥手手势。
  11. 如权利要求1至10任一项所述的系统模式切换方法,其特征在于,所述预设用户类型的用户为成人,所述第一系统模式为儿童模式,所述第二系统模式为普通模式。
  12. 一种电子设备,其特征在于,包括:
    拾音器,用于采集用户的语音;
    摄像头,用于采集用户的用户图像;
    处理器;
    存储器,所述存储器包括指令,所述指令被所述处理器执行时,使得电子设备执行权利要求1至11任一项所述的系统模式切换方法。
  13. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有指令,该指令在计算机上执行时使得计算机执行权利要求1至11任一项所述的系统模式切换方法。
PCT/CN2022/101983 2021-07-07 2022-06-28 一种系统模式切换方法、电子设备及计算机可读存储介质 WO2023280020A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110767081.8 2021-07-07
CN202110767081.8A CN115604513A (zh) 2021-07-07 2021-07-07 一种系统模式切换方法、电子设备及计算机可读存储介质

Publications (1)

Publication Number Publication Date
WO2023280020A1 true WO2023280020A1 (zh) 2023-01-12

Family

ID=84800364

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/101983 WO2023280020A1 (zh) 2021-07-07 2022-06-28 一种系统模式切换方法、电子设备及计算机可读存储介质

Country Status (2)

Country Link
CN (1) CN115604513A (zh)
WO (1) WO2023280020A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117037790B (zh) * 2023-10-10 2024-01-09 朗朗教育科技股份有限公司 一种ai交互智慧屏控制系统及方法

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104750252A (zh) * 2015-03-09 2015-07-01 联想(北京)有限公司 一种信息处理方法及电子设备
CN106201663A (zh) * 2015-05-06 2016-12-07 中兴通讯股份有限公司 移动终端的儿童模式切换方法及装置
CN106921780A (zh) * 2017-03-09 2017-07-04 广东小天才科技有限公司 智能终端运行模式的切换方法、装置及智能终端
CN108600796A (zh) * 2018-03-09 2018-09-28 百度在线网络技术(北京)有限公司 智能电视的控制模式切换方法、设备及计算机可读介质
US20210090562A1 (en) * 2019-09-19 2021-03-25 Baidu Online Network Technology (Beijing) Co., Ltd. Speech recognition control method and apparatus, electronic device and readable storage medium
CN112770186A (zh) * 2020-12-17 2021-05-07 深圳Tcl新技术有限公司 电视机观看模式的确定方法、电视机和存储介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104750252A (zh) * 2015-03-09 2015-07-01 联想(北京)有限公司 一种信息处理方法及电子设备
CN106201663A (zh) * 2015-05-06 2016-12-07 中兴通讯股份有限公司 移动终端的儿童模式切换方法及装置
CN106921780A (zh) * 2017-03-09 2017-07-04 广东小天才科技有限公司 智能终端运行模式的切换方法、装置及智能终端
CN108600796A (zh) * 2018-03-09 2018-09-28 百度在线网络技术(北京)有限公司 智能电视的控制模式切换方法、设备及计算机可读介质
US20210090562A1 (en) * 2019-09-19 2021-03-25 Baidu Online Network Technology (Beijing) Co., Ltd. Speech recognition control method and apparatus, electronic device and readable storage medium
CN112770186A (zh) * 2020-12-17 2021-05-07 深圳Tcl新技术有限公司 电视机观看模式的确定方法、电视机和存储介质

Also Published As

Publication number Publication date
CN115604513A (zh) 2023-01-13

Similar Documents

Publication Publication Date Title
WO2021000814A1 (zh) 一种语音控制方法及相关装置
CN107103905B (zh) 用于语音识别的方法和产品以及信息处理设备
WO2021135685A1 (zh) 身份认证的方法以及装置
EP4006713A1 (en) Voice-controlled split-screen display method and electronic device
CN105575395A (zh) 语音唤醒方法及装置、终端及其处理方法
CN108600796A (zh) 智能电视的控制模式切换方法、设备及计算机可读介质
CN108766438A (zh) 人机交互方法、装置、存储介质及智能终端
CN109712621A (zh) 一种语音交互控制方法及终端
WO2021212388A1 (zh) 一种交互沟通实现方法、设备和存储介质
WO2023280020A1 (zh) 一种系统模式切换方法、电子设备及计算机可读存储介质
WO2020087895A1 (zh) 语音交互处理方法及装置
EP4199488A1 (en) Voice interaction method and electronic device
CN109558243A (zh) 虚拟数据的处理方法、装置、存储介质及终端
CN112634895A (zh) 语音交互免唤醒方法和装置
CN115206306A (zh) 语音交互方法、装置、设备及系统
KR20210040856A (ko) 스마트 백미러의 인터랙션 방법, 장치, 전자기기와 저장매체
WO2024179425A1 (zh) 语音交互方法及相关设备
CN114333774B (zh) 语音识别方法、装置、计算机设备及存储介质
US20220284738A1 (en) Target user locking method and electronic device
WO2023155607A1 (zh) 终端设备和语音唤醒方法
WO2022012602A1 (zh) 一种电子设备的屏幕交互方法及装置
CN115116437B (zh) 语音识别方法、装置、计算机设备、存储介质及产品
KR20230023520A (ko) 장거리 모션 제스처 인식장치
CN209453563U (zh) 一种智能机器人
CN114765026A (zh) 一种语音控制方法、装置及系统

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22836774

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22836774

Country of ref document: EP

Kind code of ref document: A1