CN113823280A - Intelligent device control method, electronic device and system - Google Patents

Intelligent device control method, electronic device and system Download PDF

Info

Publication number
CN113823280A
CN113823280A CN202010568806.6A CN202010568806A CN113823280A CN 113823280 A CN113823280 A CN 113823280A CN 202010568806 A CN202010568806 A CN 202010568806A CN 113823280 A CN113823280 A CN 113823280A
Authority
CN
China
Prior art keywords
position information
voice
user
equipment
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010568806.6A
Other languages
Chinese (zh)
Inventor
龙嘉裕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN202010568806.6A priority Critical patent/CN113823280A/en
Publication of CN113823280A publication Critical patent/CN113823280A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2803Home automation networks
    • H04L12/2816Controlling appliance services of a home automation network by calling their functionalities
    • H04L12/282Controlling appliance services of a home automation network by calling their functionalities based on user interaction within the home
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2803Home automation networks
    • H04L12/2816Controlling appliance services of a home automation network by calling their functionalities
    • H04L12/2821Avoiding conflicts related to the use of home appliances
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The application provides an intelligent device control method, an electronic device and a system; the intelligent home equipment control system relates to the field of intelligent home, and can control target intelligent home equipment more accurately based on equipment position information. The method is applied to a communication system comprising first equipment and second equipment, wherein the first equipment and the second equipment are intelligent equipment with the same equipment type. The first device and the second device have different position information, the first position information of the first device is the relative position relationship between the first device and the first object, and the second position information of the second device is the relative position relationship between the second device and the second object. The method comprises the following steps: a first voice of a user is received, wherein the first voice comprises equipment type, position information and a first operation. And according to the first voice, determining the position information contained in the first voice as the first position information or the second position information. And controlling the first equipment corresponding to the first position information to execute the first operation or controlling the second equipment corresponding to the second position information to execute the first operation.

Description

Intelligent device control method, electronic device and system
Technical Field
The application relates to the field of smart homes, in particular to a control method of smart equipment, electronic equipment and a system.
Background
The intelligent home is a system which connects various devices (such as audio and video devices, lighting systems, security systems and the like) in the home together through the internet of things technology and provides multiple functions of home appliance control, lighting control, anti-theft alarm and the like. The voice recognition technology is combined with the intelligent home, and the intelligent home voice control equipment is defined. The intelligent household voice control equipment is used as control equipment in an intelligent household, and more convenient intelligent household equipment management can be realized through a voice interaction process with a user. For example, the intelligent home voice control device recognizes a user voice command "please help me to turn on a desk lamp", and controls the desk lamp in the intelligent home device to be automatically turned on.
In a practical application scenario, when a user issues a voice command, a device is usually referred to as a device type. However, if there are multiple devices with the same device type around, the smart home voice control device cannot determine which device the user needs to control according to the voice command. For example, assume that there are two table lamps in a room, table lamp a and table lamp B, respectively. The user wants to turn on the desk lamp A, and sends a voice command 'please help me to turn on the desk lamp' to the intelligent household voice control device. At this time, because both the desk lamp a and the desk lamp B meet the requirement of the voice command, the current voice command is ambiguous and cannot be correctly executed.
In the prior art, there are generally two solutions to this problem. In the first solution, the smart home voice control device enumerates the brands and device names of all devices that satisfy the condition, and further queries the user as to which device the user specifically desires to operate. In the current solution, if there are more devices of the same type, the single feedback information amount of the smart home voice control device is large, which causes the pressure of the user to be large. According to the second solution, the intelligent home voice control equipment opens all the equipment meeting the requirements. In the current solution, the execution result is inconsistent with the user expectation, reducing the user satisfaction.
Disclosure of Invention
The intelligent equipment control method, the electronic equipment and the system can realize more accurate control of target intelligent household equipment based on equipment position information.
In order to achieve the purpose, the technical scheme is as follows:
in a first aspect, the present application provides an intelligent device control method, which is applied to a communication system including a first device and a second device. Wherein the first device and the second device both correspond to a first category. The method includes that a first device corresponds to first position information, the first position information is a relative position relation between the first device and a first object, a second device corresponds to second position information, and the second position information is a relative position relation between the second device and a second object, and the method includes the following steps: receiving a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation. And controlling first equipment corresponding to the first position information to execute a first operation according to the first voice. And receiving a second voice of the user, wherein the second voice comprises the first category, the second position information and the first operation. And controlling second equipment corresponding to the second position information to execute the first operation according to the second voice.
In the prior art, a user typically refers to a device with a first category when issuing a voice command. The first category includes, for example, device type and/or device name, etc. However, if there are a plurality of devices of the first type around, the control device cannot determine which device the user needs to control from the voice command. In the embodiment of the application, the control device may determine the device specifically performing the first operation based on the first location information or the second location information in the user voice command. That is, if the user voice command includes the first location information, the first device corresponding to the first location information is controlled to execute the first operation. And if the user voice command comprises the second position information, controlling second equipment corresponding to the second position information to execute the first operation.
The first object is an object (e.g., furniture, etc.) for positioning a first device. Typically the first object is at a distance from the first device which is less than the threshold a, i.e. the first object is adjacent to the first device. The second object is an object (e.g., furniture, etc.) for positioning a second device. Typically the second object is less than the threshold a from the second device, i.e. the second object is adjacent to the second device. Generally, when there are a plurality of devices of the same device type, a user is accustomed to describing the devices by device location, and device location information of each device is different. For example, a desk lamp on a desk, a desk lamp on a tea table, and a desk lamp of the same equipment type, but the corresponding position information is different.
Illustratively, the first device and the second device are both table lamps, i.e. the first category is table lamps. The first device is located above the desk, namely the first position information is above the desk. The second equipment is positioned above the tea table, namely the second position information is above the tea table. The mobile phone receives a first voice of the user, and if the first voice is 'turning on a desk lamp above a desk'. The mobile phone recognizes that the intention in the first voice, namely the first operation, is 'open', the first slot position, namely the first category, is 'desk lamp', and the second slot position, namely the first position information, is 'over desk'. The mobile phone controls the desk lamp above the desk to be turned on according to the first voice. The mobile phone receives a second voice of the user, and if the second voice is 'turning on a desk lamp above the tea table'. The mobile phone recognizes that the intention in the second voice, namely the first operation, is 'open', the first slot position, namely the first category, is 'table lamp', and the second slot position, namely the second position information, is 'above the tea table'. And the mobile phone controls the desk lamp above the tea table to be turned on according to the second voice.
Therefore, the description of the equipment position information is added into the voice command, and the mobile phone identifies the equipment position information, so that the only target intelligent home equipment corresponding to the current equipment position information can be determined. The method and the system ensure that the mobile phone can also correctly and quickly execute the user command in a home system with a plurality of devices of the same type, and improve the user experience.
In one possible implementation, the method further includes: and receiving third voice of the user, wherein the third voice comprises the first category and the first operation and does not contain the first position information or the second position information. And playing prompt voice to prompt the user to input the first position information or the second position information. And receiving fourth voice of the user, wherein the fourth voice comprises the first position information or the second position information. And controlling first equipment corresponding to the first position information to execute the first operation or controlling second equipment corresponding to the second position information to execute the first operation according to the fourth voice.
Optionally, the voice command does not include the device location of the target smart home device or the device location is ambiguous, and if the number of the smart home devices determined by the mobile phone according to the voice command is one, the currently determined smart home device is the target smart home device. If the number of the smart homes determined by the mobile phone according to the voice command is multiple, namely, multiple smart home devices with the same device type exist, the mobile phone guides the user to determine the smart home devices by using the device positions through voice prompt.
For example, assume that there are 2 table lamps in the home system, located on a desk and an end table, respectively, and the handset receives the user's voice command. For example, the voice command is "turn on a desk lamp", the intention in the voice command is recognized as "turn on", the first slot includes "desk lamp", and the second slot is missing. And the mobile phone determines that two table lamps exist in the home system, and broadcasts a voice prompt of which table lamp needs to be turned on, or broadcasts a voice prompt of whether the user wants to turn on a table lamp on a desk or a table lamp on a tea table desk, so as to guide the user to select and obtain the position information of the intelligent home equipment needing to be turned on. The mobile phone receives a voice command input again by the user, such as 'desk lamp on desk', determines that the target intelligent household equipment is the desk lamp on the desk, and controls the desk lamp on the desk to be turned on.
In one possible implementation, the method further includes: first position information of a first device and second position information of a second device input by a user are received. Or, instructing the user to take a first image containing the first device and the first object, and to take a second image containing the second device and the second object; and acquiring first position information and second position information according to the first image and the second image obtained by shooting respectively.
Optionally, the mobile phone directly receives the position information input by a plurality of modes, such as keyboard input, handwriting input, voice input, and the like. Or, the mobile phone acquires an image containing the target intelligent home equipment, and automatically acquires the position information of the target intelligent home equipment in the currently acquired image through an object recognition technology and a scene semantic recognition technology.
In one possible implementation manner, acquiring the first position information and the second position information according to the first image and the second image obtained by shooting respectively includes: determining the coordinates of the first device and the coordinates of the first object contained in the first image according to the shot first image; and determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image. Determining the relative position relation between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring first position information; and determining the relative position relationship between the second equipment and the second object according to the coordinates of the second equipment and the coordinates of the second object, and acquiring second position information.
Specifically, the mobile phone obtains coordinate information of the target intelligent home equipment in the acquired image and coordinate information of an object adjacent to the target intelligent home equipment and used for positioning through a scene semantic recognition technology. And determining the position information of the intelligent household equipment according to the coordinate relation between the coordinates of the intelligent household equipment and the coordinates of the object for positioning.
In one possible implementation, the first operation includes one or more of: opening, closing, parameter adjustment and mode setting.
In one possible implementation manner, the first object includes a first type object and a second type object, the first type object is an object whose moving frequency is less than or equal to a first threshold, and the second type object is an object whose moving frequency is greater than the first threshold; the first positioning information comprises first positioning information and second positioning information, the first positioning information is the positioning information determined according to the first type of object, and the second positioning information is the positioning information determined according to the second type of object; the first positioning information has a higher priority than the second positioning information. The second object comprises a first type object and a second type object, the second position information comprises third positioning information and fourth positioning information, the third positioning information is the position information determined according to the first type object, and the fourth positioning information is the position information determined according to the second type object; the priority of the third positioning information is higher than the priority of the fourth positioning information.
Optionally, the mobile phone may pre-configure a priority order for the device locations of the smart home devices, and if the device location of the target smart home device includes multiple device location descriptions, the mobile phone may display the sorted device locations according to a preset rule and the priority order. The preset rule may include, for example, determining a priority based on whether the positioning object in the device location description is easy to move. For example, whether the object is easy to move is determined according to the relationship between the moving frequency of the object and the first threshold. The first threshold is a moving frequency threshold, which may be an empirical value. If the desk does not move normally, the movement frequency is less than the first threshold value, and the object belongs to the first type. A laptop at a desk is often moved with a frequency greater than a first threshold, belonging to a second type of object.
For example, the priority of the positioning object which is not easy to move may be set higher, and the priority of the positioning object which is easy to move may be set lower, so as to avoid that the corresponding device location is invalid after the positioning object moves.
In one possible implementation, the method further includes: and receiving fifth voice of the user, wherein the fifth voice comprises the first category and the first operation, and the fifth voice also comprises second positioning information or fourth positioning information. And playing prompt voice to prompt the user to input the first positioning information or the third positioning information. And receiving sixth voice of the user, wherein the sixth voice comprises the first positioning information or the third positioning information. And controlling first equipment corresponding to the first positioning information to execute the first operation or controlling second equipment corresponding to the third positioning information to execute the first operation according to the sixth voice.
Optionally, if the positioning object corresponding to the device location included in the voice command is an easily-moved positioning object, the mobile phone determines, according to the current positioning object, the positioning object location information that is not easily moved in the corresponding device location list or the positioning object location information with the highest priority, and confirms the target smart home device with the user through voice interaction. Thereby avoiding execution errors caused by movement of the positioning object.
In one possible implementation, after acquiring the first position information and the second position information respectively according to the first image and the second image obtained by shooting, the method further includes: and receiving the correction information of the first position information and/or the second position information by the user, and obtaining the corrected first position information and/or the corrected second position information.
For example, when the mobile phone detects that the user edits the device position, the position information of the smart home device is corrected according to the user setting. In this way, a channel for correcting the position of the equipment according to the actual environment is provided for the user, so that errors caused by image recognition are avoided.
In one possible implementation, the method further includes: and receiving a first execution result of the first operation executed by the first equipment, and broadcasting the first execution result by voice. And receiving a second execution result of the first operation executed by the second equipment, and broadcasting the second execution result by voice.
For example, the mobile phone or the smart home voice control device receives an execution result of the target smart home device executing the first operation, matches a preconfigured reply voice according to the execution result, and broadcasts the reply voice in a voice mode, so that the user can know the execution result of the current voice command.
In a second aspect, the present application provides a communication system comprising an electronic device, a first device and a second device; the first equipment and the second equipment are intelligent equipment of a first category; the first device corresponds to first position information, the first position information is the relative position relationship between the first device and the first object, the second device corresponds to second position information, and the second position information is the relative position relationship between the second device and the second object. The electronic equipment is used for receiving a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation; determining first equipment corresponding to the first position information according to the first voice; and sending a first indication to the first device, wherein the first indication is used for indicating the first device to execute a first operation. And the first equipment is used for receiving the first indication and executing the first operation according to the first indication. The electronic equipment is also used for receiving a second voice of the user, wherein the second voice comprises a first category, second position information and a first operation; determining second equipment corresponding to the second position information according to the second voice; and sending a second instruction to the second device, wherein the second instruction is used for instructing the second device to execute the first operation. And the second equipment is used for receiving the second indication and executing the first operation according to the second indication.
In a possible implementation manner, the electronic device is further configured to receive a third voice of the user, where the third voice includes the first category and the first operation, and does not include the first location information or the second location information; playing prompt voice to prompt a user to input first position information or second position information; receiving fourth voice of the user, wherein the fourth voice comprises first position information or second position information; according to the fourth voice, determining first equipment corresponding to the first position information, and sending a first instruction to the first equipment; or according to the fourth voice, determining second equipment corresponding to the second position information, and sending a second instruction to the second equipment.
In a possible implementation manner, the electronic device is further configured to receive first location information of the first device and second location information of the second device, which are input by a user; or, instructing the user to take a first image containing the first device and the first object, and to take a second image containing the second device and the second object; and acquiring first position information and second position information according to the first image and the second image obtained by shooting respectively.
In a possible implementation manner, the electronic device is specifically configured to determine, according to a first image obtained by shooting, coordinates of the first device and coordinates of a first object included in the first image; determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image; determining the relative position relation between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring first position information; and determining the relative position relationship between the second equipment and the second object according to the coordinates of the second equipment and the coordinates of the second object, and acquiring second position information.
In one possible implementation, the first category includes: device type and/or device name.
In one possible implementation, the first operation includes one or more of: opening, closing, parameter adjustment and mode setting.
In addition, for technical effects of the communication system according to the second aspect, reference may be made to the technical effects of the intelligent device control method according to the first aspect, and details are not repeated here.
In a third aspect, the present application provides a communication system comprising an electronic device, a first device, a second device and a server; the first equipment and the second equipment are intelligent equipment of a first category; the first device corresponds to first position information, the first position information is the relative position relationship between the first device and the first object, the second device corresponds to second position information, and the second position information is the relative position relationship between the second device and the second object. The electronic equipment is used for receiving a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation; and transmits the first voice to the server. The server is used for receiving the first voice and determining first equipment corresponding to the first position information according to the first voice; and sending a first indication to the first device, wherein the first indication is used for indicating the first device to execute a first operation. And the first equipment is used for receiving the first instruction sent by the server and executing the first operation according to the first instruction. The electronic equipment is also used for receiving a second voice of the user, wherein the second voice comprises a first category, second position information and a first operation; and sends the second voice to the server. The server is further used for receiving the second voice and determining second equipment corresponding to the second position information according to the second voice; and sending a second instruction to the second device, wherein the second instruction is used for instructing the second device to execute the first operation. And the second equipment is used for receiving a second instruction sent by the server and executing the first operation according to the second instruction.
Wherein the electronic device sending the voice to the server may include: the electronic equipment directly forwards the received audio information to the server. Or the electronic equipment processes the received voice, such as recognizing the voice as text information, and then forwards the processed voice to the server.
In a possible implementation manner, the electronic device is further configured to receive a third voice of the user, where the third voice includes the first category and the first operation, and does not include the first location information or the second location information; and transmits the third voice to the server. And the server is also used for receiving third voice and sending a third instruction to the electronic equipment according to the third voice, wherein the third instruction is used for instructing the electronic equipment to acquire the first position information or the second position information. The electronic equipment is also used for receiving a third instruction, playing prompt voice according to the third instruction and prompting a user to input the first position information or the second position information; receiving fourth voice of the user, wherein the fourth voice comprises first position information or second position information; and transmits the fourth voice to the server. The server is further used for determining first equipment corresponding to the first position information according to the fourth voice and sending a first instruction to the first equipment; or according to the fourth voice, determining second equipment corresponding to the second position information, and sending a second instruction to the second equipment.
In a possible implementation manner, the electronic device is further configured to receive first location information of the first device and second location information of the second device, which are input by a user; or, instructing the user to take a first image containing the first device and the first object, and to take a second image containing the second device and the second object; and sends the first image and the second image to the server. And the server is also used for acquiring the first position information and the second position information according to the shot first image and second image respectively.
In a possible implementation manner, the server is specifically configured to determine, according to a first image obtained by shooting, coordinates of the first device and coordinates of the first object included in the first image; determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image; determining the relative position relation between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring first position information; and determining the relative position relationship between the second equipment and the second object according to the coordinates of the second equipment and the coordinates of the second object, and acquiring second position information.
In one possible implementation, the first category includes: device type and/or device name.
In one possible implementation, the first operation includes one or more of: opening, closing, parameter adjustment and mode setting.
In addition, for technical effects of the communication system according to the third aspect, reference may be made to the technical effects of the intelligent device control method according to the first aspect, and details are not repeated here.
In a fourth aspect, the present application provides an electronic device comprising: a receiving unit and a processing unit. The receiving unit is used for receiving a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation. The first device corresponds to a first category, the first device corresponds to first position information, and the first position information is a relative position relation between the first device and the first object. And the processing unit is used for controlling the first equipment corresponding to the first position information to execute the first operation according to the first voice. And the receiving unit is also used for receiving a second voice of the user, wherein the second voice comprises the first category, the second position information and the first operation. The second device corresponds to the first category, the second device corresponds to second position information, and the second position information is a relative position relationship between the second device and a second object. And the processing unit is further used for controlling second equipment corresponding to the second position information to execute the first operation according to the second voice.
In a possible implementation manner, the receiving unit is further configured to receive a third voice of the user, where the third voice includes the first category and the first operation, and does not include the first location information or the second location information. The electronic equipment further comprises a playing unit used for playing prompt voice and prompting the user to input the first position information or the second position information. The receiving unit is further configured to receive a fourth voice of the user, where the fourth voice includes the first location information or the second location information. And the processing unit is further used for controlling first equipment corresponding to the first position information to execute the first operation or controlling second equipment corresponding to the second position information to execute the first operation according to the fourth voice.
In a possible implementation manner, the receiving unit is further configured to receive first location information of the first device and second location information of the second device, which are input by a user. Or, the processing unit is further used for instructing the user to shoot a first image containing the first device and the first object and shoot a second image containing the second device and the second object; and acquiring first position information and second position information according to the first image and the second image obtained by shooting respectively.
In a possible implementation manner, the processing unit is specifically configured to determine, according to a first image obtained by shooting, coordinates of a first device and coordinates of a first object included in the first image; and determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image. Determining the relative position relation between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring first position information; and determining the relative position relationship between the second equipment and the second object according to the coordinates of the second equipment and the coordinates of the second object, and acquiring second position information.
In one possible implementation, the first category includes: device type and/or device name.
In one possible implementation manner, the first object includes a first type object and a second type object, the first type object is an object whose moving frequency is less than or equal to a first threshold, and the second type object is an object whose moving frequency is greater than the first threshold; the first positioning information comprises first positioning information and second positioning information, the first positioning information is the positioning information determined according to the first type of object, and the second positioning information is the positioning information determined according to the second type of object; the first positioning information has a higher priority than the second positioning information. The second object comprises a first type of object and a second type of object; the second position information comprises third positioning information and fourth positioning information, the third positioning information is position information determined according to the first type of object, and the fourth positioning information is position information determined according to the second type of object; the priority of the third positioning information is higher than the priority of the fourth positioning information.
In a possible implementation manner, the receiving unit is further configured to receive a fifth voice of the user, where the fifth voice includes the first category and the first operation, and the fifth voice further includes the second positioning information or the fourth positioning information. And the playing unit is also used for playing prompt voice to prompt a user to input the first positioning information or the third positioning information. The receiving unit is further configured to receive a sixth voice of the user, where the sixth voice includes the first positioning information or the third positioning information. And the processing unit is further used for controlling the first device corresponding to the first positioning information to execute the first operation or controlling the second device corresponding to the third positioning information to execute the first operation according to the sixth voice.
In a possible implementation manner, the processing unit is further configured to receive correction information of the first location information and/or the second location information from the user, and obtain the corrected first location information and/or the corrected second location information.
In one possible implementation, the first operation includes one or more of: opening, closing, parameter adjustment and mode setting.
In a possible implementation manner, the receiving unit is further configured to receive a first execution result of the first device executing the first operation, and the playing unit is further configured to voice-broadcast the first execution result. The receiving unit is further used for receiving a second execution result of the first operation executed by the second device, and the playing unit is further used for voice broadcasting the second execution result.
In addition, for technical effects of the electronic device according to the fourth aspect, reference may be made to the technical effects of the intelligent device control method according to the first aspect, and details are not repeated here.
In a fifth aspect, the present application provides an electronic device, comprising: a processor and a memory; a memory coupled to the processor, the memory for storing computer program code, the computer program code including computer instructions that, when read from the memory by the processor, cause the electronic device to perform operations comprising: receiving a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation; the first device corresponds to a first category, the first device corresponds to first position information, and the first position information is a relative position relation between the first device and the first object. And controlling first equipment corresponding to the first position information to execute a first operation according to the first voice. Receiving a second voice of the user, wherein the second voice comprises a first category, second position information and a first operation; the second device corresponds to the first category, the second device corresponds to second position information, and the second position information is a relative position relationship between the second device and a second object. And controlling second equipment corresponding to the second position information to execute the first operation according to the second voice.
In one possible implementation, when the processor reads the computer instructions from the memory, the electronic device is further caused to perform the following operations: and receiving third voice of the user, wherein the third voice comprises the first category and the first operation and does not contain the first position information or the second position information. And playing prompt voice to prompt the user to input the first position information or the second position information. And receiving fourth voice of the user, wherein the fourth voice comprises the first position information or the second position information. And controlling first equipment corresponding to the first position information to execute the first operation or controlling second equipment corresponding to the second position information to execute the first operation according to the fourth voice.
In one possible implementation, before receiving the first voice of the user, when the processor reads the computer instructions from the memory, the electronic device is further caused to perform the following operations: first position information of a first device and second position information of a second device input by a user are received. Or, instructing the user to take a first image containing the first device and the first object, and to take a second image containing the second device and the second object; and acquiring first position information and second position information according to the first image and the second image obtained by shooting respectively.
In one possible implementation manner, acquiring the first position information and the second position information according to the first image and the second image obtained by shooting respectively includes: determining the coordinates of the first device and the coordinates of the first object contained in the first image according to the shot first image; and determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image. Determining the relative position relation between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring first position information; and determining the relative position relationship between the second equipment and the second object according to the coordinates of the second equipment and the coordinates of the second object, and acquiring second position information.
In one possible implementation, the first category includes: device type and/or device name.
In one possible implementation, the first operation includes one or more of: opening, closing, parameter adjustment and mode setting.
In one possible implementation manner, the first object includes a first type object and a second type object, the first type object is an object whose moving frequency is less than or equal to a first threshold, and the second type object is an object whose moving frequency is greater than the first threshold; the first positioning information comprises first positioning information and second positioning information, the first positioning information is the positioning information determined according to the first type of object, and the second positioning information is the positioning information determined according to the second type of object; the first positioning information has a higher priority than the second positioning information. The second object comprises a first type of object and a second type of object; the second position information comprises third positioning information and fourth positioning information, the third positioning information is position information determined according to the first type of object, and the fourth positioning information is position information determined according to the second type of object; the priority of the third positioning information is higher than the priority of the fourth positioning information.
In one possible implementation, when the processor reads the computer instructions from the memory, the electronic device is further caused to perform the following operations: and receiving fifth voice of the user, wherein the fifth voice comprises the first category and the first operation, and the fifth voice also comprises second positioning information or fourth positioning information. And playing prompt voice to prompt the user to input the first positioning information or the third positioning information. And receiving sixth voice of the user, wherein the sixth voice comprises the first positioning information or the third positioning information. And controlling first equipment corresponding to the first positioning information to execute the first operation or controlling second equipment corresponding to the third positioning information to execute the first operation according to the sixth voice.
In one possible implementation, after acquiring the first position information and the second position information respectively according to the first image and the second image obtained by shooting, when the processor reads the computer instructions from the memory, the electronic device is further caused to perform the following operations: and receiving the correction information of the first position information and/or the second position information by the user, and obtaining the corrected first position information and/or the corrected second position information.
In one possible implementation, when the processor reads the computer instructions from the memory, the electronic device is further caused to perform the following operations: and receiving a first execution result of the first operation executed by the first equipment, and broadcasting the first execution result by voice. And receiving a second execution result of the first operation executed by the second equipment, and broadcasting the second execution result by voice.
In addition, for technical effects of the electronic device according to the fifth aspect, reference may be made to the technical effects of the intelligent device control method according to the first aspect, and details are not repeated here.
In a sixth aspect, the present application provides a terminal device having a function of implementing the intelligent device control method according to the first aspect and any one of the possible implementation manners. The function can be realized by hardware, and can also be realized by executing corresponding software by hardware. The hardware or software includes one or more modules corresponding to the functions described above.
In a seventh aspect, the present application provides a computer storage medium, which includes computer instructions, and when the computer instructions are executed on an electronic device, the electronic device is caused to execute the intelligent device control method according to any one of the first aspect and any one of the possible implementation manners of the first aspect.
In an eighth aspect, the present application provides a computer program product, which, when run on an electronic device, causes the electronic device to execute the intelligent device control method according to any one of the first aspect and any one of its possible implementations.
In a ninth aspect, there is provided circuitry comprising processing circuitry configured to perform the smart device control method as described in the first aspect and any one of its possible implementations.
In a tenth aspect, an embodiment of the present application provides a chip system, including at least one processor and at least one interface circuit, where the at least one interface circuit is configured to perform a transceiving function and send an instruction to the at least one processor, and when the at least one processor executes the instruction, the at least one processor performs the method for controlling an intelligent device as described in the first aspect and any one of the possible implementations of the first aspect.
Drawings
Fig. 1 is a schematic diagram of an object identification result provided in an embodiment of the present application;
FIG. 2 is a schematic diagram of a system architecture provided by an embodiment of the present application;
fig. 3 is a first schematic structural diagram of an electronic device according to an embodiment of the present disclosure;
fig. 4 is a block diagram illustrating a software structure of an electronic device according to an embodiment of the present disclosure;
fig. 5 is a schematic flowchart of a control method of an intelligent device according to an embodiment of the present application;
fig. 6 is a first schematic interface diagram of an electronic device according to an embodiment of the present disclosure;
fig. 7 is a second schematic interface diagram of an electronic device according to an embodiment of the present application;
FIG. 8 is a schematic diagram of an image segmentation process provided in an embodiment of the present application;
fig. 9 is a schematic view of an application scenario in which an electronic device determines a position of an object according to an embodiment of the present application;
FIG. 10 is a schematic diagram of object positions provided by an embodiment of the present application;
fig. 11 is a third schematic interface diagram of an electronic device according to an embodiment of the present application;
FIG. 12 is a schematic diagram of probabilities of different objects being included in a room according to an embodiment of the present application;
fig. 13 is a fourth schematic interface diagram of an electronic device according to an embodiment of the present application;
fig. 14 is a fifth schematic interface diagram of an electronic device according to an embodiment of the present application;
fig. 15 is a schematic structural diagram of an electronic device according to an embodiment of the present application;
fig. 16 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The following describes in detail a method, an electronic device, and a system for controlling an intelligent device according to an embodiment of the present application with reference to the accompanying drawings.
The terms "comprising" and "having," and any variations thereof, as referred to in the description of the present application, are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those steps or elements but may alternatively include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
It should be noted that in the embodiments of the present application, words such as "exemplary" or "for example" are used to indicate examples, illustrations or explanations. Any embodiment or design described herein as "exemplary" or "e.g.," is not necessarily to be construed as preferred or advantageous over other embodiments or designs. Rather, use of the word "exemplary" or "such as" is intended to present concepts related in a concrete fashion.
In the description of the present application, the meaning of "a plurality" means two or more unless otherwise specified. "and/or" herein is merely an association describing an associated object, and means that there may be three relationships, e.g., a and/or B, which may mean: a exists alone, A and B exist simultaneously, and B exists alone.
First, for the sake of understanding, the following description is made of terms and concepts related to the embodiments of the present application.
(1) Smart home
Smart homes may be defined as a target or a system. The intelligent home is realized by organically combining various application subsystems related to home life by taking a house as a platform and utilizing advanced technologies such as computers, network communication, automatic control and the like. Through comprehensive management, home safety, convenience, comfortableness and artistry are improved, and the environment-friendly and energy-saving living environment is realized.
Briefly, the smart home is a system which connects various devices (such as audio and video devices, lighting systems, security systems and the like) in a living environment together through the internet of things technology and provides multiple functions of household appliance control, lighting control, anti-theft alarm and the like.
(2) Intelligent household equipment
The intelligent home equipment is used as equipment in the intelligent home system, and compared with the traditional home equipment, information exchange and interconnection among the equipment can be realized. The intelligent household equipment can be connected with other electronic equipment, receives information sent by the electronic equipment and executes corresponding command actions.
For example, assuming that the air conditioner is used as a smart home device and is connected to a mobile phone, the air conditioner may receive a control command sent by a user through the mobile phone. For example, the air conditioner is automatically started when receiving an 'opening' command input by a user through a mobile phone. For another example, if the air conditioner receives a command of "adjust temperature to 26 ℃" input by the user through the mobile phone, the temperature is automatically adjusted to 26 ℃.
(3) Smart home application
The applications in the electronic equipment (such as a mobile phone, a tablet and the like) can be paired with the intelligent household equipment and manage the intelligent household equipment. Generally, intelligent home applications are preconfigured in electronic equipment, and visual management of the intelligent home equipment is realized.
(4) Voice control
One of the man-machine interaction modes is to receive a voice command expressed by a user through a natural language, convert audio information in the voice command into character information by utilizing a natural language identification and processing technology, and acquire the intention and the slot position of the user. And then, making a corresponding feedback interactive mode according to the user intention and the slot position.
In a smart home application scenario, the intent is to express user expectations, which may include, for example: on, off, parameter adjustment, mode setting, etc. The slot is used for representing key information corresponding to the intention, and may include: device name, device type, etc. The device name may include, for example, a factory name or a user-defined name of the device. The device type may also be described as a class of devices, one of which is used to represent a consumer's needs. The intelligent household equipment can be a lamp, an air conditioner, a television and the like.
In some embodiments, since the user can easily obtain the device type of the smart home device, the user generally does not name the smart home device or memorize the device model of the smart home device. Thus, the device name may be referred to by the device type. For example, if the device type is a table lamp, the device name is the table lamp. And if the equipment type is an air conditioner, the equipment name is the air conditioner. Accordingly, the device types can be distinguished directly according to the device names.
Optionally, the voice control technology is combined with the smart home, so that the smart home equipment is operated through voice control. For example, assume that an air conditioner is connected to a mobile phone as a smart home device. The mobile phone receives a voice command 'turn on the air conditioner' sent by the user. Performing voice recognition on the voice command; or sending the voice command to a server, and recognizing the voice command by the server. The intent is "open" and the slot is "air conditioner". And then, the mobile phone or the server sends the operation to be executed in the voice command to the target equipment, such as the air conditioner, so that the air conditioner executes the corresponding opening operation.
Further, a reply dialog corresponding to the operation result can be pre-configured in the electronic device for feeding back the receiving or execution result of the current operation of the user. E.g., "good," "open," "close," "currently this operation cannot be performed, please try again later," etc. And after the intelligent household equipment executes corresponding operation, feeding the result back to the electronic equipment. And the electronic equipment matches the corresponding reply dialect according to the execution result and converts the text information into voice. And the user can know the execution result of the current voice command in a voice broadcasting mode. In the above example, after the air conditioner is turned on, the reply after the execution is completed is returned to the mobile phone, and the mobile phone matches the corresponding reply word technique according to the execution result, such as "opened", and broadcasts the reply word technique by voice.
(5) RGB camera
RGB cameras are the most common cameras, and are arranged in various electronic devices such as mobile phones, computers, cameras, and the like. In the shooting process, after light enters the camera through the lens, infrared light is filtered, optical signals are converted into electric signals through the sensor, then the electric signals are converted into digital signals through the circuit, and RGB image information is generated.
Optionally, the electronic device may acquire the color image including the target smart home device through the RGB camera.
(6) Pixel
A pixel is a basic unit of image display, and is generally regarded as the smallest complete sampling unit of an image. The pixel value is generally expressed by a three-dimensional array, that is, the values of two-dimensional pixel points and RGB channels on a plane.
(7) Super pixel
The super-pixel is a small area formed by a series of pixel points which are adjacent in position and similar in characteristics such as color, brightness, texture and the like. Most of these small regions retain effective information for further image segmentation, and generally do not destroy the boundary information of objects in the image.
(8) Depth camera
Compared with a conventional camera, the depth camera can acquire depth information of a photographic subject, namely three-dimensional position and size information, in addition to acquiring a planar image. Common depth camera technologies include structured light technology, time of flight (TOF) technology, binocular stereo imaging technology, and the like. As shown in table 1 below, relevant features of three depth camera technologies are introduced. The electronic equipment combines the RGB camera shooting technology, the super-pixel recognition technology and the depth camera technology, so that the image can be more accurately segmented, and the object and the position relation in the image can be recognized.
TABLE 1
Figure BDA0002548742410000101
Figure BDA0002548742410000111
(9) Object recognition technique
The neural network carried in the electronic device is trained by using a Convolutional Neural Network (CNN) technology, so that the electronic device can analyze an object in a picture or a segment of video stream and mark the object. And if the object in the image is divided by a square frame, the object recognition is realized.
Among them, the CNN technology is to imitate the characteristics of human brain and construct a multi-layer neural network. The neural network includes an input layer, a hidden layer, and an output layer. Wherein the input layer of the neural network can process multidimensional data. Taking image processing as an example, the input layer may receive pixel values of an image. Hidden layers of a neural network include one or more convolutional layers (convolutional layers), one or more pooling layers (pooling layers), and one or more fully-connected layers (fully-connected layers). Typically, one or more convolutional layers are followed by a pooling layer. In some examples, the hidden layer of the neural network may also not contain a pooling layer. The structure and the working principle of the output layer of the neural network are the same as the output of the traditional feedforward neural network. For example, for a neural network for object recognition, the output layer may be designed to output center coordinates, size, classification, and the like of the object.
The primary image features are identified by utilizing the lower layers of the hidden layers in the neural network, the image features of a plurality of lower layers form the image features of the upper layer, and the final object classification is realized at the top layer through the combination of a plurality of layers. And finally, the output layer outputs an object recognition result. Specific CNN image recognition techniques can be referred to in the prior art, and embodiments of the present application are not specifically described.
For example, as shown in fig. 1, it is assumed that an electronic device acquires an image, and inputs the current image into a neural network by using a CNN technique to identify an object included in the image. The recognition result is output, as in fig. 1, an image containing frames in which recognized objects, such as cars, traffic lights and signposts, are selected is output.
(10) Computer vision
Computer vision refers to the process of using a camera and a computer to replace human eyes to perform machine vision such as identification, tracking and measurement on a target object, and further performing image processing to obtain an image or data more suitable for human eye observation or transmission to an instrument for detection.
(11) Scene semantic recognition technology
The scene semantic recognition technology is to recognize objects, object attributes (such as color and shape), an attachment relationship (such as a support relationship and a suspension relationship) between the objects and a relative position relationship (such as front, back, left, right, top, bottom and the like) between the objects in an image through a computer vision technology, so that a computer understands a scene where the objects in the image are located.
Optionally, the electronic device obtains an image including the target smart home device by using a photographing function, analyzes an object included in the current image by using an object recognition technology, and obtains a scene and position information of the target smart home device by combining a scene semantic recognition technology. For example, as shown in fig. 1, the electronic device recognizes that the objects in the current image are cars, traffic lights and indicators, and determines that the current scene is a street scene according to the types of the objects.
Fig. 2 is a schematic diagram of a system architecture according to an embodiment of the present application. As shown in fig. 2, the system architecture includes a first electronic device 100, one or more smart home devices 200 (e.g., smart home device 1, smart home device 2, and smart home device 3 shown in fig. 2), and a server 300.
The first electronic device 100 may be, for example, an electronic device with a photographing function, such as a mobile phone, a tablet computer, a Personal Computer (PC), a Personal Digital Assistant (PDA), a netbook, a wearable electronic device, and an artificial intelligence (artificial intelligence) terminal, and the specific form of the electronic device is not particularly limited in this application.
In some embodiments, the first electronic device 100 is connected to one or more smart home devices 200 to obtain device information of the smart home devices 200. The first electronic device 100 provides a human-computer interaction interface, displays the device information of the smart home device 200 for the user through the human-computer interaction interface, and receives the user's editing of the device information of the smart home device 200.
Optionally, the first application is installed in the first electronic device 100. The first application is an intelligent home application which can be connected with the intelligent home equipment and used for editing and managing the intelligent home equipment. As shown in fig. 1, a first electronic device 100 is connected to one or more smart home devices 200 through a first application.
In some embodiments, the location information of the target smart home device may be obtained by means of image analysis. The first electronic device 100 prompts a user to acquire an image containing the smart home device through a photographing function through a human-computer interaction interface. Then, through the scene semantic recognition technology, the first electronic device 100 obtains the scene and the location information where the smart home device is located. Based on this, the first electronic device 100 may obtain the respective corresponding location information of each of the plurality of smart home devices. Subsequently, when the smart home devices are selected and controlled through the first electronic device 100, the corresponding target smart home devices can be determined according to the position information, so that the target smart home devices can be selected and controlled more accurately.
The server 300 may be a device or a network device having a computing function, such as a cloud server or a network server. The server 300 may be a server, a server cluster composed of a plurality of servers, or a cloud computing service center. The server 300 may also be described as an intelligent home cloud platform, and is configured to manage the intelligent home devices included in the intelligent home system, receive and process service requests of the intelligent home devices 200, and the like.
For example, the first electronic device 100 receives a command input by a user to start a certain smart home device, and forwards the command to the server 300 for processing. The server 300 analyzes the command, determines the target smart home device, and issues the command to the target smart home device to start the target smart home device.
For another example, after the first electronic device 100 collects an image including the target smart home device, the image is uploaded to the server 300, and the server 300 performs scene semantic analysis on the image to obtain the location information of the target smart home device. The server 300 issues the analysis result to the first electronic device 100, so that the first electronic device 100 acquires the location information of the target smart home device.
In some embodiments, if the system architecture may not include the server 300, the first electronic device 100 processes the received command and analyzes the received command to obtain the location information of the target smart home device.
Optionally, as shown in fig. 2, a second electronic device 400 may be further included in the system architecture. The second electronic device 400 is connected to one or more smart home devices 200, and is configured to control the smart home devices 200. The second electronic device 400 may also be described as a smart home voice control device. The first electronic device 100 synchronizes the device information of the smart home device 200 in the second electronic device 400 in advance, and subsequently, the second electronic device 400 directly receives a user command to select and control the smart home device 200.
In some embodiments, the first electronic device 100 may also be the second electronic device 400, and the first electronic device 100 receives a user command to select and control the smart home device 200.
The second electronic device 400 may include a dedicated device for controlling the smart home device, or a device including a function for controlling the smart home device. Such as smart sounds, smart watches, etc. The specific form of the intelligent household voice control equipment is not particularly limited.
Fig. 3 is a schematic structural diagram of a first electronic device 100 according to an embodiment of the present application.
The first electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a Universal Serial Bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a radio frequency module 150, a communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, a button 190, a motor 191, an indicator 192, a camera 193, a display screen 194, a Subscriber Identification Module (SIM) card interface 195, and the like. The sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, and the like.
It is to be understood that the illustrated structure of the embodiment of the present application does not specifically limit the first electronic device 100. In other embodiments of the present application, the first electronic device 100 may include more or fewer components than shown, or combine certain components, or split certain components, or a different arrangement of components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
Processor 110 may include one or more processing units, such as: the processor 110 may include an Application Processor (AP), a modem processor, a Graphics Processing Unit (GPU), an Image Signal Processor (ISP), a controller, a memory, a video codec, a Digital Signal Processor (DSP), a baseband processor, and/or a neural-Network Processing Unit (NPU), etc. The different processing units may be separate devices or may be integrated into one or more processors.
Wherein the controller may be a neural center and a command center of the first electronic device 100. The controller can generate an operation control signal according to the instruction operation code and the timing signal to complete the control of instruction fetching and instruction execution.
A memory may also be provided in processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that have just been used or recycled by the processor 110. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Avoiding repeated accesses reduces the latency of the processor 110, thereby increasing the efficiency of the system.
In some embodiments, processor 110 may include one or more interfaces. The interface may include an integrated circuit (I2C) interface, an integrated circuit built-in audio (I2S) interface, a Pulse Code Modulation (PCM) interface, a universal asynchronous receiver/transmitter (UART) interface, a Mobile Industry Processor Interface (MIPI), a general-purpose input/output (GPIO) interface, a Subscriber Identity Module (SIM) interface, and/or a Universal Serial Bus (USB) interface, etc.
The I2C interface is a bi-directional synchronous serial bus that includes a serial data line (SDA) and a Serial Clock Line (SCL). In some embodiments, processor 110 may include multiple sets of I2C buses. The processor 110 may be coupled to the touch sensor 180K, the charger, the flash, the camera 193, etc. through different I2C bus interfaces, respectively. For example: the processor 110 may be coupled to the touch sensor 180K through an I2C interface, such that the processor 110 and the touch sensor 180K communicate through an I2C bus interface to implement the touch function of the first electronic device 100.
The I2S interface may be used for audio communication. In some embodiments, processor 110 may include multiple sets of I2S buses. The processor 110 may be coupled to the audio module 170 via an I2S bus to enable communication between the processor 110 and the audio module 170. In some embodiments, the audio module 170 may communicate audio signals to the communication module 160 via the I2S interface, enabling answering of calls via a bluetooth headset.
The PCM interface may also be used for audio communication, sampling, quantizing and encoding analog signals. In some embodiments, the audio module 170 and the communication module 160 may be coupled by a PCM bus interface. In some embodiments, the audio module 170 may also transmit audio signals to the communication module 160 through the PCM interface, so as to implement a function of answering a call through a bluetooth headset. Both the I2S interface and the PCM interface may be used for audio communication.
The UART interface is a universal serial data bus used for asynchronous communications. The bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication. In some embodiments, a UART interface is generally used to connect the processor 110 with the communication module 160. For example: the processor 110 communicates with a bluetooth module in the communication module 160 through a UART interface to implement a bluetooth function. In some embodiments, the audio module 170 may transmit the audio signal to the communication module 160 through the UART interface, so as to realize the function of playing music through the bluetooth headset.
MIPI interfaces may be used to connect processor 110 with peripheral devices such as display screen 194, camera 193, and the like. The MIPI interface includes a Camera Serial Interface (CSI), a Display Serial Interface (DSI), and the like. In some embodiments, the processor 110 and the camera 193 communicate through a CSI interface to implement the shooting function of the first electronic device 100. The processor 110 and the display screen 194 communicate through the DSI interface to implement the display function of the first electronic device 100.
The GPIO interface may be configured by software. The GPIO interface may be configured as a control signal and may also be configured as a data signal. In some embodiments, a GPIO interface may be used to connect the processor 110 with the camera 193, the display 194, the communication module 160, the audio module 170, the sensor module 180, and the like. The GPIO interface may also be configured as an I2C interface, an I2S interface, a UART interface, a MIPI interface, and the like.
The USB interface 130 is an interface conforming to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, or the like. The USB interface 130 may be used to connect a charger to charge the first electronic device 100, and may also be used to transmit data between the first electronic device 100 and a peripheral device. And the earphone can also be used for connecting an earphone and playing audio through the earphone. The interface may also be used to connect other electronic devices, such as AR devices and the like.
It should be understood that the interfacing relationship between the modules illustrated in the embodiment of the present application is only an illustration, and does not constitute a limitation on the structure of the first electronic device 100. In other embodiments of the present application, the first electronic device 100 may also adopt different interface connection manners or a combination of multiple interface connection manners in the above embodiments.
The charging management module 140 is configured to receive charging input from a charger. The charger may be a wireless charger or a wired charger. In some wired charging embodiments, the charging management module 140 may receive charging input from a wired charger via the USB interface 130. In some wireless charging embodiments, the charging management module 140 may receive a wireless charging input through a wireless charging coil of the first electronic device 100. The charging management module 140 may also supply power to the electronic device through the power management module 141 while charging the battery 142.
The power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110. The power management module 141 receives input from the battery 142 and/or the charge management module 140 and provides power to the processor 110, the internal memory 121, the external memory, the display 194, the camera 193, the communication module 160, and the like. The power management module 141 may also be used to monitor parameters such as battery capacity, battery cycle count, battery state of health (leakage, impedance), etc. In some other embodiments, the power management module 141 may also be disposed in the processor 110. In other embodiments, the power management module 141 and the charging management module 140 may be disposed in the same device.
The wireless communication function of the first electronic device 100 can be implemented by the antenna 1, the antenna 2, the radio frequency module 150, the communication module 160, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the first electronic device 100 may be used to cover a single or multiple communication bands. Different antennas can also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The radio frequency module 150 may provide a solution including 2G/3G/4G/5G wireless communication applied on the first electronic device 100. The rf module 150 may include at least one filter, switch, power amplifier, Low Noise Amplifier (LNA), and the like. The rf module 150 may receive the electromagnetic wave from the antenna 1, and filter, amplify, etc. the received electromagnetic wave, and transmit the filtered electromagnetic wave to the modem processor for demodulation. The rf module 150 may also amplify the signal modulated by the modem processor, and convert the signal into electromagnetic wave through the antenna 1 to radiate the electromagnetic wave. In some embodiments, at least some of the functional modules of the rf module 150 may be disposed in the processor 110. In some embodiments, at least some functional modules of the rf module 150 may be disposed in the same device as at least some modules of the processor 110.
The modem processor may include a modulator and a demodulator. The modulator is used for modulating a low-frequency baseband signal to be transmitted into a medium-high frequency signal. The demodulator is used for demodulating the received electromagnetic wave signal into a low-frequency baseband signal. The demodulator then passes the demodulated low frequency baseband signal to a baseband processor for processing. The low frequency baseband signal is processed by the baseband processor and then transferred to the application processor. The application processor outputs a sound signal through an audio device (not limited to the speaker 170A, the receiver 170B, etc.) or displays an image or video through the display screen 194. In some embodiments, the modem processor may be a stand-alone device. In other embodiments, the modem processor may be separate from the processor 110 and may be disposed in the same device as the rf module 150 or other functional modules.
The communication module 160 may provide a solution for wireless communication applied to the first electronic device 100, including Wireless Local Area Networks (WLANs) (e.g., wireless fidelity (Wi-Fi) networks), bluetooth (bluetooth, BT), Global Navigation Satellite System (GNSS), Frequency Modulation (FM), Near Field Communication (NFC), Infrared (IR), and the like. The communication module 160 may be one or more devices integrating at least one communication processing module. The communication module 160 receives electromagnetic waves via the antenna 2, performs frequency modulation and filtering processing on electromagnetic wave signals, and transmits the processed signals to the processor 110. The communication module 160 may also receive a signal to be transmitted from the processor 110, frequency-modulate it, amplify it, and convert it into electromagnetic waves via the antenna 2 to radiate it.
In some embodiments, the antenna 1 of the first electronic device 100 is coupled to the radio frequency module 150 and the antenna 2 is coupled to the communication module 160 so that the first electronic device 100 can communicate with a network and other devices through wireless communication techniques. The wireless communication technology may include global system for mobile communications (GSM), General Packet Radio Service (GPRS), code division multiple access (code division multiple access, CDMA), Wideband Code Division Multiple Access (WCDMA), time-division code division multiple access (time-division code division multiple access, TD-SCDMA), Long Term Evolution (LTE), LTE, BT, GNSS, WLAN, NFC, FM, and/or IR technologies, etc. The GNSS may include a Global Positioning System (GPS), a global navigation satellite system (GLONASS), a beidou navigation satellite system (BDS), a quasi-zenith satellite system (QZSS), and/or a Satellite Based Augmentation System (SBAS).
The first electronic device 100 implements the display function through the GPU, the display screen 194, and the application processor. The GPU is a microprocessor for image processing, and is connected to the display screen 194 and an application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. The processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
The display screen 194 is used to display images, video, and the like. The display screen 194 includes a display panel. The display panel may adopt a Liquid Crystal Display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (active-matrix organic light-emitting diode, AMOLED), a flexible light-emitting diode (FLED), a miniature, a Micro-oeld, a quantum dot light-emitting diode (QLED), and the like. In some embodiments, the first electronic device 100 may include 1 or N display screens 194, N being a positive integer greater than 1.
The first electronic device 100 may implement a photographing function through the ISP, the camera 193, the video codec, the GPU, the display screen 194, the application processor, and the like.
The ISP is used to process the data fed back by the camera 193. For example, when a photo is taken, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing and converting into an image visible to naked eyes. The ISP can also carry out algorithm optimization on the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in camera 193.
The camera 193 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image to the photosensitive element. The photosensitive element may be a Charge Coupled Device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor. The light sensing element converts the optical signal into an electrical signal, which is then passed to the ISP where it is converted into a digital image signal. And the ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into image signal in standard RGB, YUV and other formats. In some embodiments, the first electronic device 100 may include 1 or N cameras 193, N being a positive integer greater than 1.
In some embodiments, the first electronic device 100 captures an image containing the target smart home device using the camera 193. And subsequently, obtaining the position information of the target intelligent household equipment through an image recognition technology. Furthermore, the first electronic device 100 may determine the target smart home device from a plurality of smart home devices of the same device type according to the location information.
The digital signal processor is used for processing digital signals, and can process digital image signals and other digital signals. For example, when the first electronic device 100 selects a frequency bin, the digital signal processor is used to perform fourier transform or the like on the frequency bin energy.
Video codecs are used to compress or decompress digital video. The first electronic device 100 may support one or more video codecs. In this way, the first electronic device 100 can play or record video in a plurality of encoding formats, such as: moving Picture Experts Group (MPEG) 1, MPEG2, MPEG3, MPEG4, and the like.
The NPU is a neural-network (NN) computing processor that processes input information quickly by using a biological neural network structure, for example, by using a transfer mode between neurons of a human brain, and can also learn by itself continuously. The NPU may implement applications such as intelligent recognition of the first electronic device 100, for example: image recognition, face recognition, speech recognition, text understanding, and the like.
The external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, to extend the storage capability of the first electronic device 100. The external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music, video, etc. are saved in an external memory card.
The internal memory 121 may be used to store computer-executable program code, which includes instructions. The processor 110 executes various functional applications and data processing of the first electronic device 100 by executing instructions stored in the internal memory 121. The internal memory 121 may include a program storage area and a data storage area. The storage program area may store an operating system, an application program (such as a sound playing function, an image playing function, etc.) required by at least one function, and the like. The storage data area may store data (such as audio data, a phone book, etc.) created during the use of the first electronic device 100, and the like. In addition, the internal memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (UFS), and the like.
The first electronic device 100 can implement an audio function through the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the earphone interface 170D, and the application processor. Such as music playing, recording, etc.
The audio module 170 is used to convert digital audio information into an analog audio signal output and also to convert an analog audio input into a digital audio signal. The audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be disposed in the processor 110, or some functional modules of the audio module 170 may be disposed in the processor 110.
The speaker 170A, also called a "horn", is used to convert the audio electrical signal into an acoustic signal. The first electronic device 100 can listen to music through the speaker 170A or listen to a hands-free call.
The receiver 170B, also called "earpiece", is used to convert the electrical audio signal into an acoustic signal. When the first electronic device 100 receives a call or voice information, it can receive voice by placing the receiver 170B close to the ear of the person.
The microphone 170C, also referred to as a "microphone," is used to convert sound signals into electrical signals. When making a call or transmitting voice information, the user can input a voice signal to the microphone 170C by speaking the user's mouth near the microphone 170C. The first electronic device 100 may be provided with at least one microphone 170C. In other embodiments, the first electronic device 100 may be provided with two microphones 170C to achieve a noise reduction function in addition to collecting sound signals. In other embodiments, the first electronic device 100 may further include three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, implement directional recording functions, and so on.
In some embodiments, the first electronic device 100 receives a voice command of the user by using the microphone 170C, and controls the target smart home device according to the voice command.
The headphone interface 170D is used to connect a wired headphone. The headset interface 170D may be the USB interface 130, or may be a 3.5mm open mobile electronic device platform (OMTP) standard interface, a cellular telecommunications industry association (cellular telecommunications industry association of the USA, CTIA) standard interface.
The keys 190 include a power-on key, a volume key, and the like. The keys 190 may be mechanical keys. Or may be touch keys. The first electronic device 100 may receive a key input, and generate a key signal input related to user setting and function control of the first electronic device 100.
The motor 191 may generate a vibration cue. The motor 191 may be used for incoming call vibration cues, as well as for touch vibration feedback. For example, touch operations applied to different applications (e.g., photographing, audio playing, etc.) may correspond to different vibration feedback effects. The motor 191 may also respond to different vibration feedback effects for touch operations applied to different areas of the display screen 194. Different application scenes (such as time reminding, receiving information, alarm clock, game and the like) can also correspond to different vibration feedback effects. The touch vibration feedback effect may also support customization.
Indicator 192 may be an indicator light that may be used to indicate a state of charge, a change in charge, or a message, missed call, notification, etc.
The SIM card interface 195 is used to connect a SIM card. The SIM card can be brought into and out of contact with the first electronic device 100 by being inserted into the SIM card interface 195 or being pulled out of the SIM card interface 195. The first electronic device 100 may support 1 or N SIM card interfaces, where N is a positive integer greater than 1. The SIM card interface 195 may support a Nano SIM card, a Micro SIM card, a SIM card, etc. The same SIM card interface 195 can be inserted with multiple cards at the same time. The types of the plurality of cards may be the same or different. The SIM card interface 195 may also be compatible with different types of SIM cards. The SIM card interface 195 may also be compatible with external memory cards. The first electronic device 100 interacts with the network through the SIM card to implement functions such as a call and data communication. In some embodiments, the first electronic device 100 employs esims, namely: an embedded SIM card. The eSIM card may be embedded in the first electronic device 100 and cannot be separated from the first electronic device 100.
The software system of the first electronic device 100 may adopt a layered architecture, an event-driven architecture, a micro-core architecture, a micro-service architecture, or a cloud architecture. In the embodiment of the present application, a software structure of the first electronic device 100 is exemplarily described by taking an Android system with a layered architecture as an example.
Fig. 4 is a block diagram of a software structure of the first electronic device 100 according to an embodiment of the present application.
The layered architecture divides the software into several layers, each layer having a clear role and division of labor. The layers communicate with each other through a software interface. In some embodiments, the Android system is divided into four layers, an application layer, an application framework layer, an Android runtime (Android runtime) and system library, and a kernel layer from top to bottom.
The application layer may include a series of applications.
As shown in fig. 4, the application programs may include APP (application) such as call, memo, browser, contact, camera, gallery, calendar, map, bluetooth, music, video, short message, etc. The applications may be pre-installed in the first electronic device 100 when it leaves the factory, or may be downloaded and installed from an application market when the user uses the first electronic device 100, which is not limited in this embodiment of the application.
In some embodiments, the application program includes a first application, and after the first electronic device 100 is connected to the smart home device, the first application controls the smart home device to start, close, or adjust functions. In addition, the first electronic device 100 can edit the specific attribute of the smart home device through the first application. The specific attribute may include, for example, a device name of the smart home device, a room where the device is located, location information of the device, and the like.
The application framework layer provides an Application Programming Interface (API) and a programming framework for the application program of the application layer. The application framework layer includes a number of predefined functions.
As shown in FIG. 4, the application framework layers may include a window manager, content provider, view system, resource manager, notification manager, and the like.
Wherein, the window manager is used for managing the window program. The window manager can obtain the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like. The content provider is used to store and retrieve data and make it accessible to applications. The data may include video, images, audio, calls made and received, browsing history and bookmarks, phone books, etc. The view system includes visual controls such as controls to display text, controls to display pictures, and the like. The view system may be used to build applications. The display interface may be composed of one or more views. The resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and the like. The notification manager enables the application to display notification information in the status bar, can be used to convey notification-type messages, can disappear automatically after a short dwell, and does not require user interaction. Such as a notification manager used to inform download completion, message alerts, etc. The notification manager may also be a notification that appears in the form of a chart or scroll bar text at the top status bar of the system, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, prompting text information in the status bar, sounding a prompt tone, vibrating the electronic device, flashing an indicator light, etc.
The system library may include a plurality of functional modules. For example: surface managers (surface managers), Media Libraries (Media Libraries), three-dimensional graphics processing Libraries (e.g., OpenGL ES), 2D graphics engines (e.g., SGL), and the like.
The surface manager is used to manage the display subsystem and provide fusion of 2D and 3D layers for multiple applications.
The media library supports a variety of commonly used audio, video format playback and recording, and still image files, among others. The media library may support a variety of audio-video encoding formats, such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, and the like.
The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like.
The 2D graphics engine is a drawing engine for 2D drawing.
The Android Runtime comprises a core library and a virtual machine. The Android runtime is responsible for scheduling and managing an Android system.
The core library comprises two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. And executing java files of the application program layer and the application program framework layer into a binary file by the virtual machine. The virtual machine is used for performing the functions of object life cycle management, stack management, thread management, safety and exception management, garbage collection and the like.
The kernel layer is a layer between hardware and software. The kernel layer at least comprises a display driver, a camera driver, an audio driver, a sensor driver and the like, and the embodiment of the application does not limit the display driver, the camera driver, the audio driver, the sensor driver and the like.
The following embodiments provide technical solutions that can be applied to the first electronic device with the structure shown in fig. 3 and/or the first electronic device with the structure shown in fig. 4.
The following description specifically describes, by way of example, a first electronic device 100 with reference to the drawings, a smart device control method provided in an embodiment of the present application.
In the prior art, after receiving a control command sent by a user, a mobile phone determines a target smart home device according to a device type included in the control command. And if the control command is 'light on', the mobile phone recognizes that the target intelligent household equipment is a light. However, a plurality of smart home devices of the same device type may exist in the home environment, and after receiving the control command of the user, the mobile phone cannot determine which smart home device to control according to the device type in the control command.
For example, it is assumed that a home environment includes 5 lamps, that is, the 5 pieces of smart home equipment are all lamps with the same equipment type. If the mobile phone receives a control command of turning on the light, the specific light which the user desires to turn on cannot be determined directly according to the control command. For the situation that multiple devices are of the same type, currently, a processing scheme configured for a mobile phone generally distinguishes smart home devices of the same device type by using brands and device names of the smart home devices, the brands and the device names are broadcasted by the mobile phone, a target smart home device is determined according to the re-selection of a user, and therefore the user needs to remember the brands and the device names of the smart home devices and perform selection control, certain difficulty exists, and user pressure is high. And if the number of the intelligent household devices with the same device type is large, the time for re-selection is long, and the efficiency is low. Or the mobile phone executes the same operation on all the intelligent household devices of the same device type according to the control command, and the execution result is not in accordance with the user expectation, so that the user satisfaction is reduced. Therefore, in the prior art, target intelligent household equipment which is controlled by a user and is expected to be controlled conveniently and accurately cannot be controlled.
Generally, when a user desires to control one of a plurality of smart home devices of the same device type, the user usually tends to use the location of the smart home device to describe a target smart home device. For example, the user tends to distinguish the smart home devices by using the position relationship between the smart home devices and other objects, such as "a lamp on a desk", "a lamp on the left of a computer", "a lamp on a wall", and the like. However, in the prior art, the electronic device cannot recognize the position relationship between the smart home device and other objects included in the voice command.
Based on the above, the application provides an intelligent device control method, which combines a voice control technology to divide a slot position into a first slot position and a second slot position in the user intention and the slot position obtained by converting a voice command into text information. The first slot corresponds to information corresponding to the existing slot, such as a device name, a device type, and the like. The second slot position is used for representing the position relation between the intelligent household equipment and other objects.
For example, assuming that a plurality of smart home devices of the same device type exist in the home environment, the mobile phone receives a voice command of the user, obtains the position relationship between the target smart home device and other objects included in the second slot according to the voice command, and distinguishes the plurality of smart home devices of the same device type according to the position relationship to determine the target smart home device. If two table lamps exist in the home environment, one table lamp is located on the cabinet on the left side of the bed head, and the other table lamp is located on the cabinet on the right side of the bed head, the second slot position information in the voice commands corresponding to the two table lamps is different.
The following describes in detail a control method of an intelligent device according to an embodiment of the present application with reference to the accompanying drawings.
Fig. 5 is a schematic flowchart illustrating a control method for an intelligent device according to an embodiment of the present application, and referring to fig. 5, the method may include S501-S504:
s501, the mobile phone is connected with the intelligent household equipment.
One or more intelligent home devices exist in the intelligent home system. The intelligent household equipment comprises a plurality of pieces of intelligent household equipment with the same equipment type. The mobile phone is connected with the intelligent household devices through the intelligent household application, and after the connection is successful, the intelligent household devices are set through the intelligent household application. The connection mode may be a bluetooth connection mode, a Wi-Fi wireless connection mode, or the like, and the embodiment of the present application is not particularly limited.
Illustratively, as shown in (a) of fig. 6, a main interface 601 of the mobile phone displays icons of applications included in a part of the mobile phone, for example, icons 61 of smart life applications on the main interface 601. The cellular phone detects an operation of the user clicking the icon 61, and displays an interface 602 as shown in (b) in fig. 6. The interface 602 is a main page of the smart life application, and the mobile phone meets the requirement of adding or deleting the smart home device on the interface 602. The cell phone detects an operation of the user clicking the add control 62, and displays an interface 603 as shown in (c) in fig. 6. A menu bar 63 is displayed in the interface 603, and when the mobile phone detects that the user clicks an option of adding equipment in the menu bar 63, the smart home equipment is automatically searched. Moreover, the mobile phone displays an interface 604 shown in (d) in fig. 6, where the interface 604 is used to display a scanning progress of the smart home application establishing the connection between the mobile phone and the smart home device. When the scanning progress reaches 100%, the connection between the mobile phone and the intelligent household equipment is successfully established. At this time, the mobile phone displays an interface 605 shown in (e) in fig. 6, which is used for prompting the user that the current smart home device is successfully connected. Then, the mobile phone automatically jumps to display an interface 606 shown in (f) of fig. 6, and the smart home devices that have been added to the current smart home system are displayed in the interface 606. As shown in the interface 606, the smart home device included in the current smart home system is a desk lamp.
S502, the mobile phone obtains information of the intelligent household equipment.
The information of the smart home device includes a device name of the smart home device, device location information, a device room where the smart home device is located, a device type, and the like. The device position information is used for representing the relative position relation between the intelligent household device and other objects. After the connection between the mobile phone and the smart home device is established, the mobile phone can acquire basic information of the smart home device, such as a device type, a device model and the like. In addition, the mobile phone can edit the equipment information of the intelligent household equipment.
For example, after the mobile phone is connected to the smart home device, the connected smart home device information may be displayed and edited by clicking the option 64 in the interface 606 shown in (f) in fig. 6. If the currently connected smart home device is a desk lamp, after clicking the option 64 in the interface 606, displaying an interface 701 as shown in (a) of fig. 7, and displaying device information of the smart home device, such as a device name, a device room, and a device location of the smart home device, in the interface 701. The mobile phone can receive the editing of the device information of the smart home device by the user at the interface 701.
Optionally, the device location information of the smart home device is determined through a scene semantic recognition technology. The mobile phone can prompt a user to collect images containing the intelligent household equipment in a man-machine interaction mode, analyze image information and further obtain equipment position information. Or, the mobile phone forwards the acquired image to a server, such as an intelligent home cloud platform. And analyzing the image by the server to obtain the position information of the equipment. And the mobile phone receives the equipment position information of the intelligent household equipment sent by the server. For example, a first device and a second device with the same device type exist in a home system, and a mobile phone instructs a user to take a first image containing the first device and a first object and a second image containing the second device and a second object. And acquiring first position information and second position information according to the first image and the second image obtained by shooting respectively. Wherein the first object is an object (e.g., furniture, etc.) for positioning the first device. Typically the first object is at a distance from the first device which is less than the threshold a, i.e. the first object is adjacent to the first device. The second object is an object (e.g., furniture, etc.) for positioning a second device. Typically the second object is less than the threshold a from the second device, i.e. the second object is adjacent to the second device. The threshold a is a preset distance threshold. And then subsequently, a plurality of intelligent household devices with the same device type can be distinguished by using the device positions, and the target intelligent household device is determined in the intelligent household devices with the same device type.
For example, as shown in an interface 701 in fig. 7 (a), after detecting that a user clicks an icon 71 corresponding to a device position on the interface 701, a mobile phone displays an interface 702 shown in fig. 7 (b), where the interface 702 is a shooting interface and is used to prompt the user to shoot a scene where the smart home device currently being edited is located. The shot view range of the view frame must include the smart home device being edited and other objects for positioning, where the other objects for positioning are adjacent objects, for example, objects whose distance from the smart home device being edited is smaller than a preset threshold. As shown in the interface 702, the smart home device being edited is the desk lamp 72, and the current shooting interface includes the desk lamp 72 and objects for positioning near the desk lamp 72, such as a desk 73, a computer 74, and a photo frame 75. After the mobile phone detects that the user clicks the photographing icon 76, it is determined that the user has finished photographing the scene where the smart home device is located. And combining the images shot by the RGB camera and the depth camera, acquiring the outline of the object according to the superpixel, and acquiring the three-dimensional space coordinates of all pixels of the object in the outline to realize accurate image segmentation. And judging the position relation between the target intelligent household equipment and other objects by utilizing a scene semantic recognition technology according to the image segmentation result and the coordinates of the objects in the image, and obtaining the equipment position of the target intelligent household equipment.
The mobile phone determines the equipment position of the intelligent household equipment through the following steps from one step to three.
Firstly, the mobile phone carries out image segmentation on the acquired image and determines an object contained in the image.
Illustratively, as shown in fig. 8 (a), the RGB image 801 is an RGB image 801 obtained by a mobile phone through an RGB camera, and the RGB image 801 is a general color image including a target object. As shown in fig. 8 (b), a depth image 802 obtained by the mobile phone through the depth camera is shown. Different depth information of the object is represented in the depth image 802 by different gray scales. If the depths of the sofa 81 and the bookshelf 82 are different, the gray scale displayed in the depth image 802 is different. The depth of each pixel point in the image can be obtained through the depth image shot by the depth camera, and then the object contour segmentation of the superpixel is obtained by subsequently combining the RGB image. As shown in fig. 8 (c), a super pixel image 803 is obtained based on the RGB image 801 and the depth image 802 for the mobile phone. In the superpixel image 803, pixel points with adjacent positions and similar characteristics such as color, brightness, texture and the like form different small regions, and further an object contour is obtained. For example, the two differently colored bolster 83 and bolster 84 on the sofa are two different small areas, respectively. For another example, in the superpixel image 803 due to the shooting angle, the sofa is divided into the area 85 and the area 86. After that, the mobile phone obtains a segmentation map 804 as shown in (d) in fig. 8 based on the RGB image 801, the depth image 802, and the super pixel image 803. Accurate segmentation information for objects in the image is contained in the segmentation map 804. Such as sofa 87, floor 88, wall 89, photo 810, bookshelf 811, bolster 812, bolster 813 in section 804.
It is understood that the segmentation graph 804 should contain accurate segmentation information of the object, and the dashed box in (d) in fig. 8 is only schematic to indicate the position of the object and is not used to indicate the contour of the object.
And step two, the mobile phone determines the position relation of the object in the image according to the image segmentation result.
Specifically, the positional relationship of the object may include, for example, a support relationship. The supporting relationship may include, for example, an up-down supporting relationship and a fore-aft supporting relationship. As shown in the divided view 804 in fig. 8 (d), in the divided view 804, the floor 88 has the sofa 87, and the floor 88 and the sofa 87 are in the up-down supporting relationship. The sofa 87 is provided with a pillow 812, and the sofa 87 and the pillow 812 are in a vertical supporting relationship. A picture 810 is suspended from the wall 89, and the wall 89 and the picture 810 are in a front-to-back supporting relationship.
Wherein, the support relationship should follow the following rule (r) -rule (r).
The coordinate of the lowest point of the support is less than or equal to the coordinate of the highest point of the supported object.
The supporting object and the supported object are adjacent, namely, no other object exists between the supporting object and the supported object. If the distance between the lowest point of the supported object and the highest point of the support is equal to 0, the support and the supported object are in an up-and-down supporting relationship. If the supporting object and the supported object have no distance in the depth direction, the supporting object and the supported object are in a front-back supporting relationship.
And thirdly, the floor is the bottom object, namely the floor can only be used as a support and has no upper and lower supporting relation lower than the floor.
And fourthly, prior knowledge of other support relations. If the window is on the wall, the window is in a front-back supporting relationship.
Illustratively, as shown in fig. 9, three-dimensional coordinate x, y, z-axis directions of the image are defined. Where the x-axis and z-axis represent two axes parallel to the ground and the y-axis represents an axis directed perpendicular to the ground towards the ceiling. Then, in fig. 9, an x coordinate is used to indicate the length of the object in the left-right direction, a y coordinate is used to indicate the height of the object, and a z coordinate is used to indicate the depth of the object in the front-rear direction. Assuming that an object i and an object j exist in the image, the position relationship between the object i and the object j can be divided into upper or lower, front or rear, left or right, and the support relationship between the object i and the object j does not violate the above rule (r)-rule (r). Wherein the three-dimensional space coordinates of the object i are x respectivelyi min,xi max,yi min,yi max,zi min,zi maxThe three-dimensional space coordinates of the object j are respectively xj min,xj max,yj min,yj max,zj min,zj max. The object i and the object j are solid figures, and projections exist in space. I is the projection of the coincident part of the object I and the object j on the x axisx i,jThe projection of object I and object j on the y-axis is shown as Iy i,jThe projection of object I and object j on the z-axis is shown as Iz i,jAnd (4) showing.
In one possible implementation, a possible positional relationship graph of object i and object j is shown, as shown in FIG. 10. Where circles are used to represent object i and rectangles are used to represent object j. It is understood that the positional relationship of the two objects has spatial relativity, and therefore, the positional relationship of the objects is exemplarily described below by taking one of the objects i as an example, and the positional relationship of the other object j is not specifically described. The possible relative positional relationship of the object i and the object j is that the object i is above the object j (i.e., the object j is below the object i), the object i is behind the object j (i.e., the object j is in front of the object i), and the object i is to the right of the object j (i.e., the object j is to the left of the object i).
Illustratively, as shown in fig. 10 (a), the lowest point coordinate of the object i on the y-axis is greater than or equal to the highest point coordinate of the object j on the y-axis, i.e., yi min≥yj max. Alternatively, as shown in fig. 10 (b), the center point coordinate of the object i on the y-axis is greater than or equal to the highest point coordinate of the object j on the y-axis, that is, (y)i min+yi max)/2≥yj max. Further, as shown in fig. 10 (c), the projection of the object I and the projection of the object j overlap each other on the plane composed of the x-axis and the z-axis, i.e., Ix i,jAnd Iz i,jAre all not empty. Then, from the positional relationship shown in fig. 10 (a) and 10 (c), or from the positional relationship shown in fig. 10 (b) and 10 (c), it is known that the object i is above the object j.
As shown in FIG. 10 (d), the coordinate of the foremost point of the object i on the z-axis is greater than or equal to the coordinate of the rearmost point of the object j on the z-axis, i.e., zi min≥zj max. Alternatively, as shown in (e) of fig. 10, the center point coordinate of the object i on the z-axis is greater than or equal to the last point coordinate of the object j on the z-axis, that is, (z)i min+zi max)/2≥zj max. Further, as shown in (f) of fig. 10, on the plane composed of the x-axis and the y-axis, the projection of the object I and the projection of the object j overlap, i.e., Ix i,jAnd Iy i,jAre all not empty. Then, from the positional relationship shown in (d) in fig. 10 and (f) in fig. 10, or from the positional relationship shown in (e) in fig. 10 and (f) in fig. 10, it is known that the object i is behind the object j.
As shown in FIG. 10 (h), the leftmost point coordinate of the object i on the x-axis is greater than or equal to the rightmost point coordinate of the object j on the x-axis, i.e., xi min≥xj max. Alternatively, as shown in fig. 10 (k), the center point coordinate of the object i on the x-axis is greater than or equal to the rightmost point coordinate of the object j on the x-axis, i.e., (x)i min+xi max)/2≥xj max. Further, as shown in (l) of FIG. 10, the projection of object I and the projection of object j coincide with each other on the plane formed by the y-axis and the z-axis, i.e., Iy i,jAnd Iz i,jAre all not empty. Then, from the positional relationship shown in (h) in fig. 10 and (i) in fig. 10, or from the positional relationship shown in (k) in fig. 10 and (l) in fig. 10, it is known that the object i is to the right of the object j.
Fig. 10 is only for illustrating the relative positional relationship between the object i and the object j, and is not limited to the size and the specific shape of the object i and the object j.
That is, in the handAfter the image segmentation result is determined, the coordinate of the object contour in the three-dimensional space can be determined, and the edge coordinate of the object can be determined. I.e. the three-dimensional spatial coordinates x of each object can be determinedmin,xmax,ymin,ymax,zmin,zmax. And then, according to the judgment rule, determining the position relation of the object in the image.
And step three, the mobile phone determines the position of the target object according to the position relation of the object in the image.
Specifically, the mobile phone determines the position information of the target object based on the position relationship. The mobile phone determines the coordinates of the first device and the coordinates of the first object contained in the first image according to the shot first image; and determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image. Then, the mobile phone determines the relative position relationship between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and obtains first position information; and determining the relative position relationship between the second equipment and the second object according to the coordinates of the second equipment and the coordinates of the second object, and acquiring second position information.
For example, as shown in fig. 9, assuming that the target object is a sofa 91, after determining the position relationship according to the three-dimensional space coordinates of each object in fig. 9, the position information of the sofa 91 is acquired, including that the sofa 91 is located above the floor 92, the sofa 91 is located on the right side of the wall 93, and the sofa 91 is located in front of the bookshelf 94.
For another example, in the interface 702 shown in (b) in fig. 7, in the image acquired by the mobile phone, the device location information of the smart home device is obtained according to the method in the above step one to step three. The smart home device desk lamp 72 is located above a desk 73, the desk lamp 72 is located on the right side of a computer 74, and the desk lamp 72 is located on the right side of a photo frame 75. After the device location of the smart home device is determined by the mobile phone, an interface 1101 shown in fig. 11 (a) is displayed, and the device location of the smart home device being edited is displayed in the interface 1101. When the mobile phone detects that the user clicks the saving icon 111, it indicates that the user has confirmed that the position and position information of the current smart home device is correct, and the saving is confirmed.
Therefore, by using the scene semantic recognition technology, the mobile phone can quickly obtain the equipment position of the intelligent household equipment in a shooting mode, the time for inputting the equipment position by a user is reduced, the user operation is convenient, and the efficiency is improved.
In some embodiments, through the image shot by the mobile phone, the device room of the smart home device can be determined, and thus the target smart home device can be determined more accurately.
For example, assume that the target object is a sofa, i.e., the smart home device being edited is a sofa. As shown in the segmentation map 804 in fig. 8 (d), objects included in the current image, such as the sofa 87, the floor 88, the wall 89, the bolster 812, the bolster 813, and the bookshelf 810, have been identified.
And then, the mobile phone judges the scene information of the target object according to the identified object information. In the intelligent home system, the scene information of the target object is room information. Therefore, the scene information of the target object is determined, that is, the type of the room in which the target object is located is determined. If the scene information of the target object is living room or bedroom, etc.
In a possible implementation manner, a priori knowledge base is preconfigured in the mobile phone, and the mobile phone determines scene information of the target object according to the priori knowledge in the priori knowledge base. Wherein, the prior knowledge base comprises prior knowledge of the correlation between the object type and the room type. For example, the possibility that various objects are contained in different rooms is different, and the images of different rooms have different image characteristics (such as light, texture, and the like), the prior knowledge base can be constructed according to the indoor environment.
Illustratively, as shown in FIG. 12, the contents of one possible a priori knowledge base are shown. Including the probability that different room classes contain different objects. Such as the presence of a cupboard in the kitchen, the presence of a bed in a bedroom, a small probability of a bed in the kitchen, etc. And storing the related prior knowledge into a prior knowledge base of the mobile phone, and after the mobile phone identifies the object contained in the current image, directly matching the room type of the current target object according to the prior knowledge. For example, the interface 804 shown in fig. 8 (d) is based on the sofa 87, the floor 88, the wall 89, the pillow 812, the pillow 813 and the bookshelf 810 included in the current image. Based on the prior knowledge shown in fig. 12, the mobile phone determines that the room type in which the current target object sofa 87 is located is the living room. For another example, as shown in (b) of fig. 7, the interface 702 includes an object table lamp 72, a desk 73, a computer 74, and a photo frame 75 in the scene currently being photographed. Based on the prior knowledge shown in fig. 12, the mobile phone determines that the room type where the smart home device desk lamp 72 currently being edited is located is a study room. Then, as shown in an interface 1101 in fig. 11 (a), after detecting that the user clicks the save icon 111, the cellular phone displays an interface 1102 shown in fig. 11 (b), and automatically displays the recognition result of the device room, such as a study, obtained from the photographed image at the device room 112.
Or after the mobile phone acquires the position information of the smart home device, the device room of the smart home device is not directly acquired, but the device room needs to be set independently. As shown in an interface 1101 in (a) of fig. 11, after detecting that the user clicks the save icon 1101, the mobile phone displays an interface 701 as shown in (a) of fig. 7, and the device room is to be confirmed. After the mobile phone detects that the user clicks the icon 77 corresponding to the equipment room on the interface 701, the mobile phone can prompt the user to shoot the image containing the intelligent household equipment again, and the equipment room is confirmed through the method. For example, in some scenes, an equipment room cannot be analyzed based on an image of an obtained equipment position, and a user needs to shoot an image containing more objects adjacent to the smart home equipment again, so as to judge the equipment room.
It should be noted that, in the embodiments of the present application, the confirmation order of the device room and the device location is not specifically limited. The mobile phone collects images once, and the equipment room and the equipment position are determined according to the collected images. Alternatively, the cell phone may capture a first image to confirm the device room and a second image to determine the device location. Still alternatively, the cell phone may capture a first image to determine the device location and a second image to determine the device room.
In some embodiments, the mobile phone receives the correction information of the first location information and/or the second location information from the user, and obtains the corrected first location information and/or the corrected second location information. For example, when the mobile phone detects that the user edits the device position, the position information of the smart home device is corrected according to the user setting. In this way, a channel for correcting the position of the equipment according to the actual environment is provided for the user, so that errors caused by image recognition are avoided.
Illustratively, as shown in an interface 1301 in (a) of fig. 13, when an operation of clicking the edit icon 131 by the user is detected on the interface 1301, an interface 1302 as shown in (b) of fig. 13 is displayed. The device location information corresponding to the edit icon 131 can be edited on the interface 1302. For example, as shown in the captured image displayed in the interface 702 shown in (b) of fig. 7, the smart home device desk lamp 72 is located on the right side of the computer 74, but on the interface 1301 shown in (a) of fig. 13, the device position is displayed as the desk lamp is located on the left side of the computer, and an abnormality occurs, and the user can click the edit icon 131 to correct the abnormality information. As shown in an interface 1302 of (b) of fig. 13, the cellular phone receives a user's correction of the device position at the interface 1302. When the mobile phone detects that the user clicks the confirmation icon 132, it indicates that the user has completed correcting the current device location, and an interface 1303 shown in (c) of fig. 13 is displayed. When the mobile phone detects that the user clicks the save icon 133 in the interface 1303, it indicates that the user has confirmed that the current device locations are all correct, and the save operation can be executed.
Similarly, if there is a possibility that the room category determined from the image has an error, the mobile phone may receive a user's modification of the device room in an interface 1102 shown in fig. 11 (b), and display the modified device room.
Optionally, the mobile phone directly receives the device location input by the user. For example, as shown in an interface 1401 in (a) of fig. 14, when the mobile phone detects an operation of clicking the icon 141 corresponding to the device position on the interface 1401 by the user, an interface 1402 shown in (b) of fig. 14 is displayed, and the mobile phone receives the device position input by the user through the keyboard 142 at the interface 1402. Further, when the mobile phone detects that the user clicks the add device location control 143, a location description of the smart home device being edited may be added to the device location. And determining the position of the equipment according to the position relation between more positioning objects and the intelligent household equipment which is being edited. After the editing is completed, an interface 1101 shown in (a) of fig. 11 is displayed, and the device position information input by the user is displayed in the interface 1101. And if the control command of the subsequent user comprises any position in the equipment positions, the intelligent household equipment can be controlled.
Likewise, the handset may directly receive user input from the device room. For example, as shown in an interface 1401 (a) in fig. 14, when the mobile phone detects an operation of clicking the icon 144 corresponding to the device room on the interface 1401, an interface 1403 shown in (c) in fig. 14 is displayed, and the mobile phone receives the device room input by the user through the keyboard 145 at the interface 1403.
It is understood that the manner in which the user edits the device location and the device room may also include handwriting input, voice input, and the like.
In some embodiments, the handset will define the device type as the device name. Assuming that the smart home device currently being edited is a desk lamp, as shown in (a) in fig. 14, an interface 1401 displays the device name as a desk lamp. In addition, the mobile phone can also receive the modification of the device name by the user, and the user defines the device name by the user. As shown in an interface 1401 in fig. 14 (a), when the mobile phone detects an operation of clicking the icon 146 corresponding to the device name on the interface 1401, the device name can be edited. Similarly, the editing mode may include a plurality of modes such as keyboard input, handwriting input, voice input, and the like.
Optionally, the mobile phone may automatically determine the device room corresponding to the smart home device according to the type of the object in the environment where the smart home device is located. For example, if the mobile phone determines that the bed exists in the environment where the smart home device is located, the device room is automatically corresponding to a bedroom. For another example, if the mobile phone determines that a bathtub exists in the environment where the smart home device is located, the device room is automatically corresponding to a bathroom and the like. As shown in an interface 701 in fig. 7 (a), the device room is displayed as a default room, which is a device room corresponding to the smart home device in the priori knowledge. Alternatively, the default room may be a default room set by the user. If the device room displayed in the interface 701 does not match the device room where the smart home device is located, the user may edit and correct the device room. Or the mobile phone is displayed to be empty at the position of the equipment room and is input by the user.
Optionally, the mobile phone may pre-configure a priority order for the device locations of the smart home devices, and if the device location of the target smart home device includes multiple device location descriptions, the mobile phone may display the sorted device locations according to a preset rule and the priority order. The preset rule may include, for example, determining a priority based on whether the positioning object in the device location description is easy to move. For example, whether the object is easy to move is determined according to the relationship between the moving frequency of the object and the first threshold. The first threshold is a moving frequency threshold, which may be an empirical value. If the desk does not move normally, the movement frequency is less than the first threshold value, and the object belongs to the first type. A laptop at a desk is often moved with a frequency greater than a first threshold, belonging to a second type of object. For example, the priority of the positioning object which is not easy to move may be set higher, and the priority of the positioning object which is easy to move may be set lower, so as to avoid that the corresponding device location is invalid after the positioning object moves.
That is, the first object for positioning the first device includes a first type object that is an object whose moving frequency is less than or equal to the first threshold value and a second type object that is an object whose moving frequency is greater than the first threshold value. The first location information includes first location information and second location information, the first location information being location information determined according to a first type of object, the second location information being location information determined according to a second type of object. The first positioning information has a higher priority than the second positioning information. The second object for positioning the second device as described above also comprises the first type of object and the second type of object. The second location information includes third location information and fourth location information, the third location information being location information determined according to the first type of object, the fourth location information being location information determined according to the second type of object. The priority of the third positioning information is higher than the priority of the fourth positioning information.
Illustratively, as shown in an interface 1101 in fig. 11 (a), 3 pieces of device location description information are displayed in the interface 1101. The positioning objects are a desk, a computer and a photo frame respectively. The user generally does not move the desk, and the device position description priority corresponding to the desk can be defined to be the highest and displayed at the top of the device position. Secondly, both the computer and the photo frame have the risk of movement, and the photo frame is most easily moved. Therefore, the device location description priority corresponding to the photo frame is the lowest and is displayed at the lowest position of the device location.
Subsequently, when the mobile phone receives the user control command, the target intelligent household equipment can be judged according to the equipment position priority. Specific implementations are described in detail below.
Optionally, the positions of the multiple devices all correspond to the same target smart home device, so that the implementation manner of controlling the target smart home device by the user can be increased, and meanwhile, the possibility of misoperation is increased. Therefore, only the device position information corresponding to one or a limited number of positioning objects which are not easy to move can be reserved, and the control abnormity caused by the movement of the positioning objects is reduced.
Therefore, the information of one or more intelligent household devices obtained by the method contains the device position information, and then the target intelligent household device can be accurately determined according to the device position information in the subsequent process of controlling the target intelligent household device, so that the abnormal control condition of the intelligent household devices with the same device type is avoided.
S503, receiving a user voice command, and determining the target intelligent household equipment according to the user voice command.
Specifically, the mobile phone receives a voice command of the user, analyzes the voice command, and acquires the user intention and the slot position. Or, the mobile phone is connected with the target intelligent household equipment, and after the equipment information of the target intelligent household equipment is obtained, the information of the target intelligent household equipment is synchronized to the intelligent household voice control equipment. Subsequently, the intelligent household voice control equipment receives the voice command of the user, analyzes the voice command and obtains the intention and the slot position of the user. Or after receiving the user voice command, the mobile phone or the smart home voice control device uploads the voice command to the smart home cloud platform, the smart home cloud platform analyzes the voice command, obtains the user intention and the slot position, and sends an analysis result to the mobile phone or the smart home voice control device. Wherein the slot positions include a first slot position and a second slot position. The first slot includes a device name, a device type, and the like. The second slot includes a device location.
The following description will take the example of the mobile phone processing the voice command.
Optionally, the voice command includes a device position of the target smart home device, and the mobile phone converts audio information in the voice command into text information, so as to obtain the device position of the target smart home device included in the second slot. Therefore, even if a plurality of intelligent household devices of the same device type exist, the corresponding only one target intelligent household device can be matched according to the device position, and accurate control is achieved.
Suppose that the smart home system includes a first device and a second device, and the first device and the second device both correspond to a first category. Wherein the first category includes, for example, a device type and/or a device name. That is to say, first equipment and second equipment are the smart home devices of the same type. The first device corresponds to first position information, the first position information is the relative position relationship between the first device and the first object, the second device corresponds to second position information, and the second position information is the relative position relationship between the second device and the second object. The mobile phone receives a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation, and the mobile phone controls first equipment corresponding to the first position information to execute the first operation according to the first voice. The mobile phone receives a second voice of the user, the second voice comprises a first category, second position information and a first operation, and the mobile phone controls second equipment corresponding to the second position information to execute the first operation according to the second voice.
Illustratively, the first device and the second device are both table lamps, i.e. the first category is table lamps. The first device is located above the desk, namely the first position information is above the desk. The second equipment is positioned above the tea table, namely the second position information is above the tea table. The mobile phone receives a first voice of the user, and if the first voice is 'turning on a desk lamp above a desk'. The mobile phone recognizes that the intention in the first voice, namely the first operation, is 'open', the first slot position, namely the first category, is 'desk lamp', and the second slot position, namely the first position information, is 'over desk'. The mobile phone controls the desk lamp above the desk to be turned on according to the first voice. The mobile phone receives a second voice of the user, and if the second voice is 'turning on a desk lamp above the tea table'. The mobile phone recognizes that the intention in the second voice, namely the first operation, is 'open', the first slot position, namely the first category, is 'table lamp', and the second slot position, namely the second position information, is 'above the tea table'. And the mobile phone controls the desk lamp above the tea table to be turned on according to the second voice. Therefore, even if the first device and the second device are the same in device type and are both table lamps, the mobile phone can determine which table lamp the user desires to control according to different position information of the two table lamps, and the user requirements are met.
Optionally, the voice command does not include the device location of the target smart home device or the device location is ambiguous, and if the number of the smart home devices determined by the mobile phone according to the voice command is one, the currently determined smart home device is the target smart home device. If the number of the smart homes determined by the mobile phone according to the voice command is multiple, namely, multiple smart home devices with the same device type exist, the mobile phone guides the user to determine the smart home devices by using the device positions through voice prompt. For example, if the slot position obtained by recognizing the voice command does not include the second slot position after the mobile phone receives the voice command of the user, and the mobile phone determines that the number of the smart home devices corresponding to the current voice command is multiple, the mobile phone automatically plays the prompt voice to guide the user to determine the device position of the target smart home device. For example, the mobile phone receives a third voice of the user, where the third voice includes the first category and the first operation, and does not include the first location information or the second location information. Then, the mobile phone plays the prompt voice to prompt the user to input the first position information or the second position information. And then, the mobile phone receives fourth voice of the user, wherein the fourth voice comprises the first position information or the second position information. And the mobile phone controls the first device corresponding to the first position information to execute the first operation or controls the second device corresponding to the second position information to execute the first operation according to the fourth voice.
For example, assume that there are 2 table lamps in the home system, located on a desk and an end table, respectively, and the handset receives the user's voice command. For example, the voice command is "turn on a desk lamp", the intention in the voice command is recognized as "turn on", the first slot includes "desk lamp", and the second slot is missing. And the mobile phone determines that two table lamps exist in the home system, and broadcasts a voice prompt of which table lamp needs to be turned on, or broadcasts a voice prompt of whether the user wants to turn on a table lamp on a desk or a table lamp on a tea table desk, so as to guide the user to select and obtain the position information of the intelligent home equipment needing to be turned on. The mobile phone receives a voice command input again by the user, such as 'desk lamp on desk', determines that the target intelligent household equipment is the desk lamp on the desk, and controls the desk lamp on the desk to be turned on.
Optionally, the mobile phone may further extract slot position information including the device room in the voice command, and the mobile phone may distinguish the target smart home device according to the device room. For example, suppose there are two table lamps located in different rooms in a home system. The mobile phone receives a voice command of opening a desk lamp of a study room. The mobile phone extracts the equipment room as a study according to the voice command, determines the target intelligent home equipment, and controls the desk lamp in the study to be turned on.
Similarly, if the user voice command includes the device room and lacks the location information, the user can be guided to further select the device room through the broadcast voice prompt. For example, the voice command is "open desk lamp in study", the intention in the recognition voice command is "open", the first slot includes "desk lamp", the equipment room is study, and the position information is missing. If the mobile phone confirms that two table lamps exist in the study, the mobile phone can broadcast 'you want to open the table lamp on the desk or the table lamp on the tea table' through voice, guide the user to select, receive the voice command selected by the user, and determine the target intelligent home equipment.
Optionally, if the missing equipment room or the equipment position in the voice command of the user does not affect the execution of the voice command, the mobile phone does not need to play the prompt voice, and directly controls the target smart home equipment according to the voice command, so that the efficiency is improved. For example, the user voice command is "turn off the desk lamp", and the mobile phone determines that a plurality of desk lamps exist in the current home system. However, if only one of the table lamps is in the on state, the mobile phone determines that the target smart home device is the table lamp currently in the on state according to the first operation, and then directly sends the instruction command to the table lamp to control the table lamp to be turned off.
Optionally, the mobile phone receives a fifth voice of the user, where the fifth voice includes the first category and the first operation, and the fifth voice further includes the second positioning information or the fourth positioning information. And the mobile phone plays prompt voice to prompt the user to input the first positioning information or the third positioning information. And the mobile phone receives sixth voice of the user, wherein the sixth voice comprises the first positioning information or the third positioning information. And the mobile phone controls the first equipment corresponding to the first positioning information to execute the first operation or controls the second equipment corresponding to the third positioning information to execute the first operation according to the sixth voice. The second positioning information and the fourth positioning information are position information determined according to a positioning object which is easy to move, and the first positioning information and the third positioning information are position information determined according to a positioning object which is not easy to move. That is to say, if the positioning object corresponding to the device location included in the voice command is an easily movable positioning object, the mobile phone determines the positioning object location information that is not easily movable in the corresponding device location list or the positioning object location information with the highest priority according to the current positioning object, and confirms the target smart home device with the user through voice interaction. Thereby avoiding execution errors caused by movement of the positioning object.
Illustratively, as shown in the captured scene displayed in the interface 702 in (b) of fig. 7, the mobile phone receives a user voice command, such as "turn on the desk lamp on the right side of the photo frame". The mobile phone judges that the second slot position 'right side of the photo frame' contained in the voice command, wherein the photo frame is a positioning object easy to move and has a moving risk. If the photo frame is moved at this time, the object on the right side of the photo frame observed by the user is inconsistent with the object on the right side of the photo frame stored in the mobile phone, the execution result is wrong, and the user requirements cannot be met. Therefore, as shown in an interface 1101 in fig. 11 (a), a device location list of the target smart home device is displayed in the interface 1101, and the mobile phone determines, according to the device location list, the location object location information that is not easily moved or the location object location information with the highest priority among the location object location information. As shown in the interface 1101, if the mobile phone determines that the object information is located above the desk, the voice broadcast "please ask for the desk lamp above the desk", so as to guide the user to determine the correct target smart home device.
Optionally, if the smart home system does not include multiple smart home devices of the same device type, that is, one device type only includes one smart home device. If the mobile phone can directly determine the unique smart home device according to the user intention and the first slot position even if the second slot position is missing in the voice command of the user received by the mobile phone, the mobile phone does not need to guide the user to perform secondary selection, but directly controls the corresponding target smart home device.
Optionally, the user voice command is not limited to the canonical description, and the user may issue the voice command based on his own habits. For example, in a sectional view 804 shown in fig. 8 (d), the sofa 87 is positioned in front of the bookshelf 811. The voice command received by the mobile phone may be "sofa beside bookshelf". Therefore, the mobile phone can carry out generalization processing on the voice command, and the position relations with similar meanings refer to the same target object according to the position relation of the object, so that the difficulty of controlling the target object by the user is further reduced. For example, the mobile phone determines that the sofa beside the bookshelf is the sofa 87 in front of the bookshelf 811.
Illustratively, the mobile phone performs generalization processing on the voice command, and positions of devices with similar meanings in the voice command correspond to the same target smart home device. For example, as shown in (b) of fig. 7, in the scene displayed in the interface 702, the mobile phone receives the user voice commands of "turn on the desk lamp on the right side of the computer", "turn on the desk lamp beside the computer", and "turn on the desk lamp on the right side of the photo frame". The mobile phone can determine that the corresponding target smart home devices are all the table lamps 72 according to the voice command.
And S504, controlling the target intelligent household equipment to execute a first operation according to the voice command.
Specifically, the first operation is an operation determined according to the user's intention, such as an operation of turning on, turning off, parameter adjustment, mode setting, or the like. And the mobile phone or the intelligent household voice control equipment sends the user intention in the voice command to the determined target intelligent household equipment, and controls the target intelligent household equipment to execute the first operation. Or after the mobile phone or the smart home voice control device receives the analysis result of the smart home cloud platform on the voice command, the user intention in the voice command is sent to the determined target smart home device, and the target smart home device is controlled to execute the first operation. Or the mobile phone or the smart home voice control device forwards the voice command to the smart home cloud platform, and the smart home cloud platform analyzes the voice command and then directly controls the target smart home device to execute the first operation.
Optionally, the mobile phone or the smart home voice control device receives a first execution result of the first device executing the first operation, and broadcasts the first execution result in a voice mode. And the mobile phone or the intelligent household voice control equipment receives a second execution result of the second equipment executing the first operation, and broadcasts the second execution result in a voice mode. That is to say, the mobile phone or the smart home voice control device receives an execution result of the target smart home device executing the first operation, matches a preconfigured reply voice according to the execution result, and broadcasts the reply voice in a voice mode, so that the user can know the execution result of the current voice command.
Illustratively, as shown in the scene displayed in the interface 702 in fig. 7 (b), the handset receives a user semantic command of "turn on the desk lamp above the study desk". Then, the first operation is a turning-on operation, and after the mobile phone sends a turning-on command to the desk lamp 72, the desk lamp 72 executes the first operation and turns on automatically. The mobile phone receives the execution result of the desk lamp 72, and after the corresponding answer phone is matched, automatic voice broadcasting is turned on.
Therefore, the intelligent device control method provided by the embodiment of the application can be used for additionally identifying the device position information of the intelligent household device in the identification of the voice command. Based on the device position information, the target intelligent household device can be selected and controlled more accurately. The situation that a user needs to confirm for multiple times or cannot accurately control certain intelligent household equipment in a scene of the intelligent household equipment with the same equipment type is avoided. The intelligent device control method can improve working efficiency and reduce user pressure.
As shown in fig. 15, an embodiment of the present application discloses an electronic device. The electronic device 1500 may be configured to implement the methods described in the above method embodiments. Illustratively, the electronic device 1500 may specifically include: a receiving unit 1501 and a processing unit 1502. Wherein, the receiving unit 1501 is configured to support the electronic device 1500 to execute step S503 in fig. 5; the processing unit 1502 is configured to support the electronic device 1500 to execute step S501, step S502, step S503, and step S504 in fig. 5.
Optionally, the electronic device 1500 may further include a playing unit 1503, which is used to support the electronic device 1500 to play the voice prompt.
All relevant contents of each step related to the above method embodiment may be referred to the functional description of the corresponding functional module, and are not described herein again.
As shown in fig. 16, an embodiment of the present application discloses an electronic device, including: one or more processors 1601; a memory 1602; one or more computer programs 1603; and a microphone 1604. The various devices described above may be connected by one or more communication buses 1605. Wherein the one or more computer programs 1603 are stored in the memory 1602 and configured to be executed by the one or more processors 1601, the one or more computer programs 1603 include instructions that may be used to perform the steps of the present embodiment.
Optionally, as also shown in fig. 16, the electronic device may further include a camera 1606. The camera 1606 may be used to capture images including the target smart home device.
Optionally, as also shown in fig. 16, the electronic device may further include a communication module 1607. The communication module 1607 may be used for interacting with the smart home devices. The communication module 1607 may also be used to interact with a server, such as a cloud platform of smart home devices.
For example, the processor 1601 may specifically be the processor 110 shown in fig. 3, the memory 1602 may specifically be the internal memory 121 shown in fig. 3, the microphone 1604 may specifically be used in the audio module shown in fig. 3, and may be the microphone 170C or the receiver 170B shown in fig. 3, the camera 1606 may specifically be the camera 193 shown in fig. 3, and the communication module 1607 may specifically be the mobile phone number radio frequency module 150 and/or the communication module 160 shown in fig. 3, which is not limited in this embodiment of the present invention.
Embodiments of the present application further provide a chip system, where the chip system includes at least one processor and at least one interface circuit. The processor and the interface circuit may be interconnected by wires. For example, the interface circuit may be used to receive signals from other devices. As another example, the interface circuit may be used to send signals to other devices. Illustratively, the interface circuit may read instructions stored in the memory and send the instructions to the processor. The instructions, when executed by the processor, may cause the electronic device to perform the steps performed by the handset in the embodiments described above. Of course, the chip system may further include other discrete devices, which is not specifically limited in this embodiment of the present application.
The embodiment of the present application further provides a computer storage medium, where a computer instruction is stored in the computer storage medium, and when the computer instruction runs on an electronic device, the electronic device is enabled to execute the relevant method steps to implement the intelligent device control method in the foregoing embodiment.
The embodiment of the present application further provides a computer program product, which when running on a computer, causes the computer to execute the above related steps, so as to implement the intelligent device control method in the above embodiment.
In addition, embodiments of the present application also provide an apparatus, which may be specifically a component or a module, and may include a processor and a memory connected to each other; the memory is used for storing computer execution instructions, and when the apparatus runs, the processor can execute the computer execution instructions stored in the memory, so that the apparatus executes the intelligent device control method in the above-mentioned method embodiments.
In addition, the electronic device, the computer storage medium, the computer program product, or the chip provided in the embodiments of the present application are all configured to execute the corresponding method provided above, so that the beneficial effects achieved by the electronic device, the computer storage medium, the computer program product, or the chip may refer to the beneficial effects in the corresponding method provided above, and are not described herein again.
Through the above description of the embodiments, it is clear to those skilled in the art that, for convenience and simplicity of description, the foregoing division of the functional modules is merely used as an example, and in practical applications, the above function distribution may be performed by different functional modules according to needs, that is, the internal structure of the device may be divided into different functional modules to perform all or part of the above described functions. For the specific working processes of the system, the apparatus and the unit described above, reference may be made to the corresponding processes in the foregoing method embodiments, and details are not described here again.
In the embodiments provided in the present application, it should be understood that the disclosed method can be implemented in other ways. For example, the above-described embodiments of the electronic device are merely illustrative, and for example, the division of the modules or units is only one logical division, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or may be integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of modules or units through some interfaces, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application may be substantially implemented or contributed to by the prior art, or all or part of the technical solution may be embodied in a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) or a processor to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: flash memory, removable hard drive, read only memory, random access memory, magnetic or optical disk, and the like.
The above description is only an embodiment of the present application, but the scope of the present application is not limited thereto, and any changes or substitutions within the technical scope of the present disclosure should be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (26)

1. An intelligent device control method is applied to a communication system comprising a first device and a second device, wherein the first device and the second device both correspond to a first class; the first device corresponds to first position information, the first position information is a relative position relationship between the first device and a first object, the second device corresponds to second position information, and the second position information is a relative position relationship between the second device and a second object, the method includes:
receiving a first voice of a user, wherein the first voice comprises the first category, the first position information and a first operation;
controlling the first equipment corresponding to the first position information to execute the first operation according to the first voice;
receiving a second voice of a user, wherein the second voice comprises the first category, the second position information and the first operation;
and controlling the second equipment corresponding to the second position information to execute the first operation according to the second voice.
2. The method of claim 1, further comprising:
receiving a third voice of the user, wherein the third voice comprises the first category and the first operation and does not contain the first position information or the second position information;
playing prompt voice to prompt a user to input the first position information or the second position information;
receiving fourth voice of a user, wherein the fourth voice comprises the first position information or the second position information;
and controlling the first equipment corresponding to the first position information to execute the first operation or controlling the second equipment corresponding to the second position information to execute the first operation according to the fourth voice.
3. The method of claim 1 or 2, wherein prior to said receiving the first speech of the user, the method further comprises:
receiving the first position information of the first device and the second position information of the second device input by a user;
or, instructing a user to take a first image containing the first device and the first object and to take a second image containing the second device and the second object; and acquiring the first position information and the second position information according to the first image and the second image obtained by shooting respectively.
4. The method according to claim 3, wherein the acquiring the first position information and the second position information from the first image and the second image obtained by shooting respectively comprises:
determining the coordinates of the first device and the coordinates of the first object contained in the first image according to the first image obtained by shooting; determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image;
determining the relative position relationship between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring the first position information; and determining the relative position relationship between the second device and the second object according to the coordinates of the second device and the coordinates of the second object, and acquiring the second position information.
5. The method according to any of claims 1-4, wherein the first category comprises: device type and/or device name.
6. The method according to any of claims 1-5, wherein the first operation comprises one or more of: opening, closing, parameter adjustment and mode setting.
7. A communication system, comprising an electronic device, a first device and a second device; the first equipment and the second equipment are both intelligent household equipment of a first category; the first device corresponds to first position information, the first position information is the relative position relationship between the first device and a first object, the second device corresponds to second position information, and the second position information is the relative position relationship between the second device and a second object;
the electronic equipment is used for receiving a first voice of a user, wherein the first voice comprises the first category, the first position information and a first operation; determining the first equipment corresponding to the first position information according to the first voice; sending a first indication to the first device, wherein the first indication is used for indicating the first device to execute the first operation;
the first device is used for receiving the first indication and executing the first operation according to the first indication;
the electronic device is further configured to receive a second voice of the user, where the second voice includes the first category, the second location information, and the first operation; determining the second equipment corresponding to the second position information according to the second voice; sending a second indication to the second device, wherein the second indication is used for indicating the second device to execute the first operation;
and the second device is used for receiving the second indication and executing the first operation according to the second indication.
8. The communication system of claim 7,
the electronic device is further configured to receive a third voice of the user, where the third voice includes the first category and the first operation, and does not include the first location information or the second location information; playing prompt voice to prompt a user to input the first position information or the second position information; receiving fourth voice of a user, wherein the fourth voice comprises the first position information or the second position information; according to the fourth voice, the first equipment corresponding to the first position information is determined, and the first indication is sent to the first equipment; or, according to the fourth voice, determining the second device corresponding to the second location information, and sending the second indication to the second device.
9. The communication system according to claim 7 or 8,
the electronic device is further used for receiving the first position information of the first device and the second position information of the second device input by a user;
or, instructing a user to take a first image containing the first device and the first object and to take a second image containing the second device and the second object; and acquiring the first position information and the second position information according to the first image and the second image obtained by shooting respectively.
10. The communication system of claim 9,
the electronic device is specifically configured to determine, according to the first image obtained by shooting, coordinates of the first device and coordinates of the first object included in the first image; determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image; determining the relative position relationship between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring the first position information; and determining the relative position relationship between the second device and the second object according to the coordinates of the second device and the coordinates of the second object, and acquiring the second position information.
11. The communication system according to any of claims 7-10, wherein the first category comprises: device type and/or device name.
12. A communication system according to any of claims 7-11, wherein the first operation comprises one or more of: opening, closing, parameter adjustment and mode setting.
13. A communication system, comprising an electronic device, a first device, a second device and a server; the first equipment and the second equipment are intelligent equipment of a first category; the first device corresponds to first position information, the first position information is the relative position relationship between the first device and a first object, the second device corresponds to second position information, and the second position information is the relative position relationship between the second device and a second object;
the electronic equipment is used for receiving a first voice of a user, wherein the first voice comprises the first category, the first position information and a first operation; and sending the first voice to the server;
the server is used for receiving the first voice and determining the first equipment corresponding to the first position information according to the first voice; sending a first indication to the first device, wherein the first indication is used for indicating the first device to execute the first operation;
the first device is configured to receive the first instruction sent by the server, and execute the first operation according to the first instruction;
the electronic device is further configured to receive a second voice of the user, where the second voice includes the first category, the second location information, and the first operation; and sending the second voice to the server;
the server is further configured to receive the second voice, and determine, according to the second voice, the second device corresponding to the second location information; sending a second indication to the second device, wherein the second indication is used for indicating the second device to execute the first operation;
and the second device is used for receiving the second instruction sent by the server and executing the first operation according to the second instruction.
14. The communication system of claim 13,
the electronic device is further configured to receive a third voice of the user, where the third voice includes the first category and the first operation, and does not include the first location information or the second location information; and sending the third voice to the server;
the server is further configured to receive the third voice, and send a third instruction to the electronic device according to the third voice, where the third instruction is used to instruct the electronic device to acquire the first location information or the second location information;
the electronic equipment is further used for receiving the third instruction, playing prompt voice according to the third instruction and prompting a user to input the first position information or the second position information; receiving fourth voice of a user, wherein the fourth voice comprises the first position information or the second position information; and sending the fourth voice to the server;
the server is further configured to determine, according to the fourth voice, the first device corresponding to the first location information, and send the first indication to the first device; or, according to the fourth voice, determining the second device corresponding to the second location information, and sending the second indication to the second device.
15. The communication system according to claim 13 or 14,
the electronic device is further used for receiving the first position information of the first device and the second position information of the second device input by a user; or, instructing a user to take a first image containing the first device and the first object and to take a second image containing the second device and the second object; and sending the first image and the second image to the server;
the server is further configured to obtain the first position information and the second position information according to the first image and the second image obtained through shooting respectively.
16. The communication system of claim 15,
the server is specifically configured to determine, according to the first image obtained by shooting, coordinates of the first device and coordinates of the first object included in the first image; determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image; determining the relative position relationship between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring the first position information; and determining the relative position relationship between the second device and the second object according to the coordinates of the second device and the coordinates of the second object, and acquiring the second position information.
17. The communication system according to any of claims 13-16, wherein the first category comprises: device type and/or device name.
18. A communication system according to any of claims 13-17, wherein the first operation comprises one or more of: opening, closing, parameter adjustment and mode setting.
19. An electronic device, comprising: a processor, a microphone, and a memory, the memory coupled with the processor, the memory for storing computer program code, the computer program code comprising computer instructions that, when read from the memory by the processor, cause the electronic device to perform operations comprising:
receiving a first voice of a user, wherein the first voice comprises a first category, first position information and a first operation; the first equipment corresponds to the first category, the first equipment corresponds to the first position information, and the first position information is the relative position relation between the first equipment and a first object;
controlling the first equipment corresponding to the first position information to execute the first operation according to the first voice;
receiving a second voice of the user, wherein the second voice comprises the first category, second position information and the first operation; the second device corresponds to the first category, the second device corresponds to second position information, and the second position information is a relative position relationship between the second device and a second object;
and controlling the second equipment corresponding to the second position information to execute the first operation according to the second voice.
20. The electronic device of claim 19, wherein the computer instructions, when read from the memory by the processor, further cause the electronic device to:
receiving a third voice of the user, wherein the third voice comprises the first category and the first operation and does not contain the first position information or the second position information;
playing prompt voice to prompt a user to input the first position information or the second position information;
receiving fourth voice of a user, wherein the fourth voice comprises the first position information or the second position information;
and controlling the first equipment corresponding to the first position information to execute the first operation or controlling the second equipment corresponding to the second position information to execute the first operation according to the fourth voice.
21. The electronic device of claim 19 or 20, wherein, when the processor reads the computer instructions from the memory prior to the receiving the first voice of the user, the electronic device is further caused to:
receiving the first position information of the first device and the second position information of the second device input by a user;
or, instructing a user to take a first image containing the first device and the first object and to take a second image containing the second device and the second object; and acquiring the first position information and the second position information according to the first image and the second image obtained by shooting respectively.
22. The electronic device according to claim 21, wherein the acquiring the first position information and the second position information from the first image and the second image obtained by shooting, respectively, comprises:
determining the coordinates of the first device and the coordinates of the first object contained in the first image according to the first image obtained by shooting; determining the coordinates of the second device and the coordinates of the second object contained in the second image according to the shot second image;
determining the relative position relationship between the first equipment and the first object according to the coordinates of the first equipment and the coordinates of the first object, and acquiring the first position information; and determining the relative position relationship between the second device and the second object according to the coordinates of the second device and the coordinates of the second object, and acquiring the second position information.
23. The electronic device of any one of claims 19-22, wherein the first category comprises: device type and/or device name.
24. The electronic device of any of claims 19-23, wherein the first operation comprises one or more of: opening, closing, parameter adjustment and mode setting.
25. A computer-readable storage medium having instructions stored therein, which when run on an electronic device, cause the electronic device to perform the smart device control method of any one of claims 1-6.
26. A computer program product comprising instructions for causing an electronic device to perform the smart device control method of any one of claims 1-6 when the computer program product is run on the electronic device.
CN202010568806.6A 2020-06-19 2020-06-19 Intelligent device control method, electronic device and system Pending CN113823280A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010568806.6A CN113823280A (en) 2020-06-19 2020-06-19 Intelligent device control method, electronic device and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010568806.6A CN113823280A (en) 2020-06-19 2020-06-19 Intelligent device control method, electronic device and system

Publications (1)

Publication Number Publication Date
CN113823280A true CN113823280A (en) 2021-12-21

Family

ID=78924605

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010568806.6A Pending CN113823280A (en) 2020-06-19 2020-06-19 Intelligent device control method, electronic device and system

Country Status (1)

Country Link
CN (1) CN113823280A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115457960A (en) * 2022-11-09 2022-12-09 广州小鹏汽车科技有限公司 Voice interaction method, server and computer readable storage medium
CN115512704A (en) * 2022-11-09 2022-12-23 广州小鹏汽车科技有限公司 Voice interaction method, server and computer readable storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105629747A (en) * 2015-09-18 2016-06-01 宇龙计算机通信科技(深圳)有限公司 Voice control method and device of smart home system
US20160378080A1 (en) * 2015-06-25 2016-12-29 Intel Corporation Technologies for conversational interfaces for system control
CN107528753A (en) * 2017-08-16 2017-12-29 捷开通讯(深圳)有限公司 Intelligent home voice control method, smart machine and the device with store function
CN108702313A (en) * 2018-04-28 2018-10-23 深圳魔耳智能声学科技有限公司 Intelligent home voice control method, device, equipment and system
CN110535732A (en) * 2019-07-29 2019-12-03 深圳绿米联创科技有限公司 A kind of apparatus control method, device, electronic equipment and storage medium
CN110738994A (en) * 2019-09-25 2020-01-31 北京爱接力科技发展有限公司 Control method, device, robot and system for smart homes
CN111028839A (en) * 2019-12-20 2020-04-17 苏州思必驰信息科技有限公司 Smart home control method and device and electronic equipment
US20200175981A1 (en) * 2018-11-30 2020-06-04 Beijing Xiaomi Intelligent Technology Co., Ltd. Method, apparatus, system, device for controlling intelligent device and storage medium
US20200294490A1 (en) * 2017-11-09 2020-09-17 Gree Electric Appliances (Wuhan) Co., Ltd Method, apparatus, system, and electronic device for voice controlling smart home electrical appliance

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160378080A1 (en) * 2015-06-25 2016-12-29 Intel Corporation Technologies for conversational interfaces for system control
CN105629747A (en) * 2015-09-18 2016-06-01 宇龙计算机通信科技(深圳)有限公司 Voice control method and device of smart home system
CN107528753A (en) * 2017-08-16 2017-12-29 捷开通讯(深圳)有限公司 Intelligent home voice control method, smart machine and the device with store function
US20200294490A1 (en) * 2017-11-09 2020-09-17 Gree Electric Appliances (Wuhan) Co., Ltd Method, apparatus, system, and electronic device for voice controlling smart home electrical appliance
CN108702313A (en) * 2018-04-28 2018-10-23 深圳魔耳智能声学科技有限公司 Intelligent home voice control method, device, equipment and system
US20200175981A1 (en) * 2018-11-30 2020-06-04 Beijing Xiaomi Intelligent Technology Co., Ltd. Method, apparatus, system, device for controlling intelligent device and storage medium
CN110535732A (en) * 2019-07-29 2019-12-03 深圳绿米联创科技有限公司 A kind of apparatus control method, device, electronic equipment and storage medium
CN110738994A (en) * 2019-09-25 2020-01-31 北京爱接力科技发展有限公司 Control method, device, robot and system for smart homes
CN111028839A (en) * 2019-12-20 2020-04-17 苏州思必驰信息科技有限公司 Smart home control method and device and electronic equipment

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115457960A (en) * 2022-11-09 2022-12-09 广州小鹏汽车科技有限公司 Voice interaction method, server and computer readable storage medium
CN115512704A (en) * 2022-11-09 2022-12-23 广州小鹏汽车科技有限公司 Voice interaction method, server and computer readable storage medium
CN115512704B (en) * 2022-11-09 2023-08-29 广州小鹏汽车科技有限公司 Voice interaction method, server and computer readable storage medium

Similar Documents

Publication Publication Date Title
CN110231905B (en) Screen capturing method and electronic equipment
WO2020078299A1 (en) Method for processing video file, and electronic device
CN111046680B (en) Translation method and electronic equipment
CN113645351B (en) Application interface interaction method, electronic device and computer-readable storage medium
CN113272745B (en) Smart home equipment sharing system and method and electronic equipment
CN111752443A (en) Method, related device and system for controlling page by display equipment
CN111669459B (en) Keyboard display method, electronic device and computer readable storage medium
CN112262563B (en) Image processing method and electronic device
CN112580400B (en) Image optimization method and electronic equipment
CN113542580B (en) Method and device for removing light spots of glasses and electronic equipment
CN113935898A (en) Image processing method, system, electronic device and computer readable storage medium
CN112237031B (en) Method for accessing intelligent household equipment to network and related equipment
CN113810542B (en) Control method applied to electronic equipment, electronic equipment and computer storage medium
CN112784174A (en) Method, device and system for determining pose
WO2023241209A9 (en) Desktop wallpaper configuration method and apparatus, electronic device and readable storage medium
CN114115770A (en) Display control method and related device
CN113823280A (en) Intelligent device control method, electronic device and system
WO2022007707A1 (en) Home device control method, terminal device, and computer-readable storage medium
CN113986369B (en) Internet of things equipment control method and system, electronic equipment and storage medium
CN114356195A (en) File transmission method and related equipment
CN115914461A (en) Position relation identification method and electronic equipment
CN113536834A (en) Pouch detection method and device
CN114995715B (en) Control method of floating ball and related device
CN113572798B (en) Device control method, system, device, and storage medium
CN115641867A (en) Voice processing method and terminal equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination