WO2020098421A1 - 智能设备控制方法、装置和存储介质 - Google Patents

智能设备控制方法、装置和存储介质 Download PDF

Info

Publication number
WO2020098421A1
WO2020098421A1 PCT/CN2019/110191 CN2019110191W WO2020098421A1 WO 2020098421 A1 WO2020098421 A1 WO 2020098421A1 CN 2019110191 W CN2019110191 W CN 2019110191W WO 2020098421 A1 WO2020098421 A1 WO 2020098421A1
Authority
WO
WIPO (PCT)
Prior art keywords
smart device
broadcast
distance
broadcast content
target
Prior art date
Application number
PCT/CN2019/110191
Other languages
English (en)
French (fr)
Inventor
张胜美
李金鑫
宋奔
Original Assignee
北京猎户星空科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京猎户星空科技有限公司 filed Critical 北京猎户星空科技有限公司
Priority to EP19884579.4A priority Critical patent/EP3866054B1/en
Priority to JP2021526383A priority patent/JP7204912B2/ja
Priority to US17/293,770 priority patent/US11247342B2/en
Publication of WO2020098421A1 publication Critical patent/WO2020098421A1/zh

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/0005Manipulators having means for high-level communication with users, e.g. speech generator, face recognition means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1694Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
    • B25J9/1697Vision controlled systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/008Artificial life, i.e. computing arrangements simulating life based on physical entities controlled by simulated intelligence so as to replicate intelligent life forms, e.g. based on robots replicating pets or humans in their appearance or behaviour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation

Definitions

  • This application relates to the technical field of smart devices, and in particular to smart device control methods, devices, and storage media.
  • childcare robots can obtain fairy tales, children's songs, poems, etc. from the server and play them to children.
  • the reception robot can help users handle some services (such as bank card opening business) by themselves.
  • the embodiments of the present application provide a smart device control method, device, and storage medium, which are used to solve the problems of low exposure rate and utilization rate of smart devices in the prior art.
  • an embodiment of the present application provides a method for controlling a smart device.
  • the method includes:
  • the smart device If a person feature is detected in the collected image and the distance between the target with the person feature and the smart device is within the first specified distance, the smart device is triggered to perform voice broadcast;
  • the smart device After the voice broadcast, if it is detected that the distance between the target and the smart device is within the second specified distance, the smart device is awakened, and the second specified distance is less than the first specified distance.
  • the character characteristics include at least one of the following characteristics: facial characteristics and human characteristics.
  • the collecting images of the surrounding environment of the smart device includes:
  • triggering the smart device to perform voice broadcast includes:
  • the smart device is triggered to perform voice broadcast.
  • the method further includes:
  • the voice broadcast process if it is detected that the distance between at least one other target with a character characteristic and the smart device is within the first specified distance, the voice broadcast is not repeated after the voice broadcast is completed.
  • the method further includes:
  • the voice broadcast is completed and the interval is set for a period of time, if it is detected that the distance between at least one other target with a character characteristic and the smart device is within the first specified distance, the voice broadcast is performed again.
  • the triggering the smart device to perform voice broadcast includes:
  • the triggering the smart device to perform voice broadcast includes:
  • the smart device is triggered to perform voice loop broadcast on the broadcast content in the preset broadcast content set.
  • the method further includes:
  • the method further includes:
  • a preset expression is displayed on the screen of the smart device.
  • the present application also provides a smart device control device, the device including:
  • An image collection module used to collect images of the surrounding environment of the smart device when the smart device is in a sleep state
  • the pre-wake-up module is configured to trigger the smart device to perform voice broadcast if a person feature is detected in the collected image and the distance between the target with the person feature and the smart device is within the first specified distance;
  • a wake-up module is used to wake up the smart device after detecting that the distance between the target and the smart device is within a second specified distance after the voice broadcast, the second specified distance is less than the first specified distance.
  • the character characteristics include at least one of the following characteristics: facial characteristics and human characteristics.
  • the image acquisition module is specifically configured to acquire color images and depth images of objects in front of the smart device in real time;
  • the pre-wake-up module is specifically used for:
  • the smart device is triggered to perform voice broadcast.
  • the pre-wake-up module is further configured to: during the voice broadcast process, if the distance between at least one other target with a character characteristic and the smart device is detected in the first specified Within the distance, after the voice broadcast is completed, the voice broadcast is not repeated.
  • the pre-wake-up module is further configured to: after the voice broadcast is completed and the interval is set for a set period of time, if the distance between at least one other target with character characteristics and the smart device is Within the first specified distance, voice broadcast is performed again.
  • the pre-wake-up module is specifically used to:
  • the pre-wake-up module is specifically used to:
  • the smart device is triggered to perform voice loop broadcast on the broadcast content in the preset broadcast content set.
  • the device further includes:
  • a broadcast content maintenance module configured to receive a broadcast content addition instruction and add the broadcast content included in the broadcast content addition instruction to the broadcast content set;
  • the device further includes:
  • the broadcast report emotion control module is used for performing voice broadcast while displaying a preset expression on the screen of the smart device.
  • the present application also provides a computing device, including at least one processor; and a memory communicatively connected to the at least one processor; wherein the memory stores instructions executable by the at least one processor , The instruction is executed by the at least one processor, so that the at least one processor can execute any smart device control method provided in the embodiments of the present application.
  • the present application also provides a computer-readable storage medium, wherein the storage medium stores computer-executable instructions, and the computer-executable instructions are used to cause the computer to execute any intelligence in the embodiments of the present application Device control method.
  • the present application also provides a computer program product, the computer program product includes a computer program stored on a computer-readable storage medium, the computer program includes program instructions, and the program instructions are executed by a processor Steps to implement any of the above smart device control methods.
  • the smart device control method, device and storage medium provided by this application. According to the image detection, it can determine whether the users in a long distance around have interaction intentions, and through the form of active voice broadcast, guide the users with potential interaction intentions to the smart device to operate, thereby improving the exposure rate and utilization rate of the smart device .
  • FIG. 1 is a schematic diagram of an application scenario of smart device control in an embodiment of this application.
  • FIG. 2 is one of the schematic flowcharts of the smart device control method in the embodiment of the present application.
  • FIG. 3 is a schematic diagram of a smart device control method in an embodiment of this application.
  • FIG. 4 is a second schematic flowchart of a method for controlling a smart device in an embodiment of the present application
  • FIG. 5 is a schematic diagram of a smart device control device in an embodiment of this application.
  • FIG. 6 is a schematic structural diagram of a computing device according to an embodiment of the present application.
  • the embodiments of the present application provide a smart device control method, device, and storage medium.
  • a smart device control method for controlling the exposure rate and usage rate of smart devices, so that people can interact with the smart device more naturally and use the services provided by the smart device more conveniently.
  • the smart device may include two modes, such as a pre-wake mode and a wake-up mode.
  • Pre-wake mode It is a functional design when the user is far away from the smart device.
  • the smart device collects images of the surrounding environment in real time, and when a face or body feature is detected, it is determined that there is a user around, and the voice broadcast function is turned on, such as voice greeting, broadcast of services that it can provide, etc. Users understand the functions of smart devices, attract users to approach smart devices, and use specific functions of smart devices.
  • Wake-up mode It is a functional design when the user is closer to the smart device.
  • the smart device can interact with the user, follow the user's instructions to complete the corresponding service, or actively guide the user to complete the corresponding service.
  • Depth image In 3D (Dimensions) computer graphics, Depth Map (depth image) is an image that contains distance information from the surface of the scene object of the viewpoint.
  • Characters used to describe human characteristics, such as facial features and human features.
  • the embodiments of the present application realize that the smart device uses a more natural interaction scheme to guide the user to use the smart device to complete the corresponding service.
  • the smart device may actively greet and announce its own function. For example, a bank reception robot can broadcast "Hi, I am a bank assistant, please ask what business you need to handle” in a natural interactive way. In this way, the smart device can greet the user through natural interaction, attract the user to interact with the smart device, and guide the user to complete the operation on the smart device.
  • FIG. 1 it is a schematic diagram of an application scenario of a smart device control method provided by an embodiment of the present application.
  • the application scenario includes a user 10 and a smart device 11.
  • the smart device 11 can collect a front image in real time, and then analyze whether there are people in front according to the image If someone is close to you, you will enter the pre-wake state and perform voice broadcast to guide users to use smart devices.
  • the smart device in the embodiment of the present application may be a mobile robot or a robot placed at a fixed position, and the functions provided by the smart device may be various functions in various industries, as described above with the growth of children Smart devices, bank reception robots, etc.
  • the method includes the following steps:
  • Step 201 When the smart device is in a sleep state, collect images of the surrounding environment of the smart device.
  • Step 202 If a person feature is detected in the collected image and the distance between the target with the person feature and the smart device is within the first specified distance, the smart device is triggered to perform voice broadcast.
  • the person feature detection can perform face detection and / or human body detection, as long as it can determine that the image contains a person portrait, it is applicable to the embodiments of the present application.
  • the broadcast voice can be set by the service provider with built-in functions of the smart device as needed.
  • the bank can set up its reception smart device to broadcast "Hi, I am the bank secret, and can handle ** business for individual users.”
  • Step 203 After the voice broadcast, if it is detected that the distance between the target and the smart device is within the second specified distance, the smart device is awakened, and the second specified distance is less than the first specified distance.
  • the first specified distance is 3 meters
  • the second specified distance is 1.5 meters, that is, the user is within 1.5-3 meters of the voice broadcast, after determining that the user is close and within 1.5 meters, you can enter the wake-up mode, and the user Interact.
  • the distance between the target and the smart device is within the first specified distance, it is determined that someone within a short distance may have an intention to interact, and then greet the target through voice broadcast to facilitate guiding the target to the smart device. Then, when it is determined that it is very close to the target (that is, within the second specified distance), it is determined that the target has an intention to interact, and then enters the wake-up mode for interaction. In this way, the user is diverted to the smart device for operation in a simple and natural way, thereby improving the utilization rate of the smart device.
  • the scheme for determining the distance between the smart device and the target may be specifically implemented as the following:
  • Option 1 Use a distance sensor to detect the distance between the target in the surrounding environment and the smart device.
  • the scheme of whether to perform voice broadcasting can be understood as when a person's feature is detected and the distance between the target detecting the surrounding environment and the smart device is within the first specified distance, the smart device can be triggered to perform voice broadcasting. Otherwise, even if a person characteristic is detected, but the distance sensor does not detect an object, the smart device will not be triggered to perform voice broadcast. That is, if the target distance is too far, it is determined that there is no user with intent to interact, and no voice broadcast is required for diversion.
  • the distance sensor detects an object (the object may be a table, an obstruction, etc.), it is considered that there is no user with intent to interact, and voice announcement will not be triggered.
  • RGB-D camera can be used to collect images and then judge.
  • the color image and the depth image of the object in front of the smart device are collected in real time; then, the person feature detection is performed based on the color image in real time; after the person feature is detected, each character feature detected is executed: determining the person feature The corresponding data of the target in the depth image; according to the corresponding data of the target with the character in the depth image, determine the distance between the target with the character and the smart device; if the distance is within the first specified distance, Then, the smart device is triggered to perform voice broadcast.
  • an infrared sensor and an infrared camera can be used to obtain a depth image (Depth image), that is, infrared light is emitted through the infrared sensor, and the infrared light will be reflected to the infrared camera when it encounters an obstruction, and the infrared camera images according to the captured infrared light to obtain The depth data of each pixel in the infrared image.
  • RGB-D camera color images ie RGB (red, green, blue, red, green, blue) images
  • Depth images are registered, so there is a one-to-one correspondence between the pixels of the two images. Therefore, according to each pixel of the detected person characteristic, the distance between the target with the person characteristic and the smart device can be determined.
  • the distance from the closest pixel can be used as the distance from the smart device, or the furthest distance can be used.
  • the average distance of each pixel of the person feature can also be used. This application does not limit this.
  • the target when the target is close enough, it can directly determine that the target has an interactive intention, and enter the wake-up mode.
  • the first specified distance is (1.5m-3m), and the second specified distance is 0-1.5m.
  • the specific first specified distance and the second specified distance may be determined according to experiments or actual needs, which is not limited in this application.
  • the intelligent device manager can input the broadcast file (the file can be a text file or an audio file) and save it through the voice broadcast page. Then when voice broadcast is needed, the smart device performs voice broadcast according to the pre-stored file. If it is a text file, text-to-speech (Text ToSpeech, TTS) synthesis technology can be used for broadcasting.
  • the broadcast file can be a text file or an audio file
  • TTS text-to-speech
  • multiple broadcast contents may be set to constitute a preset broadcast content set. Trigger the smart device to randomly select the broadcast content from the preset broadcast content set for voice broadcast during voice broadcast, or trigger the smart device to sequentially select the broadcast content from the preset broadcast content set for voice broadcast, or trigger the smart The device cyclically broadcasts the broadcast content in the preset broadcast content set.
  • the number of words in each voice has an upper limit, and the number of voices may also have an upper limit (such as 10).
  • the user can also update the broadcast content configured by himself, which may specifically include: receiving a broadcast content addition instruction, adding the broadcast content included in the broadcast content addition instruction to the broadcast content collection; or, receiving the broadcast content Content modification instruction, modify the corresponding broadcast content in the broadcast content set according to the broadcast content modification instruction; or, receive the broadcast content replacement instruction and replace the corresponding broadcast content set in accordance with the broadcast content replacement instruction Broadcast content.
  • the user can maintain the broadcast content through the operations of adding, deleting, and modifying.
  • the voice will not be repeatedly broadcasted continuously. In this way, it is possible to avoid continuously disturbing the user and causing the user to suffer from hearing fatigue. At the same time, it can also save resources consumed by voice broadcasting.
  • the smart device may be triggered to use a volume greater than a preset volume for voice broadcast, where the preset volume is the volume when the smart device wakes up and interacts with the target.
  • the voice content broadcast by the smart device can attract the user's attention, so as to be able to more effectively guide users with potential interaction intentions to the smart device for operation.
  • a preset expression can also be displayed on the screen of the smart device. For example, a smiley face is leaked, so that the user can not only be interested in the smart device visually, but also be further interested in the smart device visually, so as to guide the user to the smart device.
  • the character feature detection and voice broadcast can be two independent programs, that is, image collection and character feature detection can be performed in real time, and the target with the character feature is determined when the character feature is detected The distance to the smart device. If the distance is within the first specified distance, send a message to the voice broadcast program. The voice broadcast program determines whether to perform voice broadcast according to the received message.
  • the duration in the above 2) to perform voice broadcast, and explain the cooperation of these two programs. Specifically, it can be implemented assuming that the specified duration is 5 seconds.
  • the voice broadcast program starts to broadcast "Hi, I am receiving smart device XXX, what are the needs?"
  • the voice broadcast program starts to broadcast "Hi, I am receiving smart device XXX, what are the needs?"
  • the message is ignored.
  • the timer starts. If within 5 seconds, a message of the character feature detection program is received, the message is discarded. If the message is received after 5 seconds, a voice broadcast is performed.
  • Step 401 When the smart device is in the sleep state, collect images of the surrounding environment of the smart device.
  • Step 402 Perform feature detection based on the color image in real time.
  • Step 403 If a human face is detected, then for the detected human face, find the corresponding data of the human face in the depth image; and, according to the corresponding data, determine the distance between the target corresponding to the human face and the smart device.
  • Step 404 Determine whether there is at least one target within 1.5 meters from the smart device. If yes, perform step 405; otherwise, perform step 406.
  • Step 405 Wake up the smart device.
  • Step 406 Determine whether there is at least one target within 1.5-3 meters from the smart device. If yes, perform step 407; otherwise, perform step 401.
  • Step 407 Determine whether voice announcement is currently in progress. If yes, return to step 401. If no, perform step 408.
  • Step 408 Whether the duration from the last voice broadcast reaches the set duration, if not, return to step 401, and if yes, perform step 409.
  • Step 409 Randomly select a piece of broadcast content, and broadcast the selected broadcast content with a volume greater than a specified decibel.
  • Step 410 After the voice broadcast is completed, determine the distance between the target and the smart device, and then execute step 404.
  • the embodiments of the present application also provide a smart device control device.
  • the device includes:
  • the image collection module 501 is used to collect images of the surrounding environment of the smart device when the smart device is in a sleep state;
  • the pre-wake-up module 502 is configured to trigger the smart device to perform voice broadcast if a person feature is detected in the collected image and the distance between the target with the person feature and the smart device is within the first specified distance ;
  • the wake-up module 503 is used to wake up the smart device if it is detected that the distance between the target and the smart device is within the second specified distance after voice announcement, the second specified distance is less than the first Specify the distance.
  • the character characteristics include at least one of the following characteristics: facial characteristics and human characteristics.
  • the image acquisition module is specifically configured to acquire color images and depth images of objects in front of the smart device in real time;
  • the pre-wake-up module is specifically used for:
  • the smart device is triggered to perform voice broadcast.
  • the pre-wakeup module is also used to:
  • the voice broadcast process if it is detected that the distance between at least one other target with a character characteristic and the smart device is within the first specified distance, the voice broadcast is not repeated after the voice broadcast is completed.
  • the pre-wakeup module is also used to:
  • the voice broadcast is completed and the interval is set for a period of time, if it is detected that the distance between at least one other target with a character characteristic and the smart device is within the first specified distance, the voice broadcast is performed again.
  • the pre-wake-up module is specifically used to:
  • the pre-wake-up module is specifically used to:
  • the smart device is triggered to perform voice loop broadcast on the broadcast content in the preset broadcast content set.
  • the device further includes:
  • a broadcast content maintenance module configured to receive a broadcast content addition instruction and add the broadcast content included in the broadcast content addition instruction to the broadcast content set;
  • the device further includes:
  • the broadcast report emotion control module is used for performing voice broadcast while displaying a preset expression on the screen of the smart device.
  • the computing device may include at least one processor and at least one memory.
  • the memory stores program codes, and when the program codes are executed by the processor, the processor is caused to execute the steps in the smart device control method according to various exemplary embodiments of the present application described above in this specification.
  • the processor may perform steps 201-203 as shown in FIG.
  • the computing device 130 according to this embodiment of the present application is described below with reference to FIG. 6.
  • the computing device 130 shown in FIG. 6 is only an example, and should not bring any limitation to the functions and use scope of the embodiments of the present application.
  • the computing device 130 is expressed in the form of a general-purpose computing device.
  • the components of the computing device 130 may include, but are not limited to: the at least one processor 131, the at least one memory 132, and a bus 133 connecting different system components (including the memory 132 and the processor 131).
  • the bus 133 represents one or more of several types of bus structures, including a memory bus or a memory controller, a peripheral bus, a processor, or a local bus using any of a variety of bus structures.
  • the memory 132 may include a readable medium in the form of volatile memory, such as random access memory (RAM) 1321 and / or cache memory 1322, and may further include read only memory (ROM) 1323.
  • RAM random access memory
  • ROM read only memory
  • the memory 132 may further include a program / utility tool 1325 having a set of (at least one) program modules 1324.
  • program modules 1324 include, but are not limited to, an operating system, one or more application programs, other program modules, and program data. These Each of the examples or some combination may include an implementation of the network environment.
  • the computing device 130 may also communicate with one or more external devices 134 (eg, keyboard, pointing device, etc.), and may also communicate with one or more devices that enable a user to interact with the computing device 130, and / or with the computing device 130 Any device (eg, router, modem, etc.) capable of communicating with one or more other computing devices. Such communication may be performed through an input / output (I / O) interface 135.
  • the computing device 130 can also communicate with one or more networks (such as a local area network (LAN), a wide area network (WAN), and / or a public network, such as the Internet) through the network adapter 136. As shown, the network adapter 136 communicates with other modules for the computing device 130 via the bus 133.
  • LAN local area network
  • WAN wide area network
  • public network such as the Internet
  • computing device 130 may be used in conjunction with the computing device 130, including but not limited to: microcode, device drivers, redundant processors, external disk drive arrays, RAID systems, tape drives And data backup storage system.
  • various aspects of the smart device control method provided by the present application may also be implemented in the form of a program product, which includes program code.
  • the program product runs on a computer device
  • the program code is used to enable
  • the computer device performs the steps in the smart device control method according to various exemplary embodiments of the present application described above in this specification.
  • the computer device may perform steps 201-203 as shown in FIG. 2.
  • the program product may use any combination of one or more readable media.
  • the readable medium may be a readable signal medium or a readable storage medium.
  • the readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination of the above. More specific examples of readable storage media (non-exhaustive list) include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the foregoing.
  • the program product for smart device control of the embodiment of the present application may use a portable compact disk read-only memory (CD-ROM) and include program code, and may be run on a computing device.
  • CD-ROM portable compact disk read-only memory
  • the program product of the present application is not limited to this.
  • the readable storage medium may be any tangible medium containing or storing a program, which may be used by or in combination with an instruction execution system, apparatus, or device.
  • the readable signal medium may include a data signal that is propagated in baseband or as part of a carrier wave, in which readable program code is carried. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above.
  • the readable signal medium may also be any readable medium other than a readable storage medium, and the readable medium may send, propagate, or transmit a program for use by or in combination with an instruction execution system, apparatus, or device.
  • the program code contained on the readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • the program code for performing the operation of the present application can be written in any combination of one or more programming languages.
  • the programming language includes object-oriented programming languages such as Java, C ++, etc., as well as conventional procedural programming Language-such as "C" language or similar programming language.
  • the program code may be completely executed on the user computing device, partially executed on the user equipment, executed as an independent software package, partially executed on the user computing device, partially executed on the remote computing device, or entirely executed on the remote computing device or server On the implementation.
  • the remote computing device may be connected to the user computing device through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computing device (eg, using Internet services Provider to connect via the Internet).
  • LAN local area network
  • WAN wide area network
  • Internet services Provider to connect via the Internet
  • the embodiments of the present application may be provided as methods, systems, or computer program products. Therefore, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware. Moreover, the present application may take the form of a computer program product implemented on one or more computer usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer usable program code.
  • computer usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
  • These computer program instructions can be provided to the processor of a general-purpose computer, special-purpose computer, embedded processing machine, or other programmable data processing device to produce a machine that enables the generation of instructions executed by the processor of the computer or other programmable data processing device
  • These computer program instructions may also be stored in a computer-readable memory that can guide a computer or other programmable data processing device to work in a specific manner, so that the instructions stored in the computer-readable memory produce an article of manufacture including an instruction device, the instructions The device implements the functions specified in one block or multiple blocks of the flowchart one flow or multiple flows and / or block diagrams.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing device, so that a series of operating steps are performed on the computer or other programmable device to produce computer-implemented processing, which is executed on the computer or other programmable device
  • the instructions provide steps for implementing the functions specified in one block or multiple blocks of the flowchart one flow or multiple flows and / or block diagrams.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Robotics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Mechanical Engineering (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • User Interface Of Digital Computer (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Manipulator (AREA)
  • Image Analysis (AREA)

Abstract

智能设备控制方法、装置和存储介质,涉及智能设备技术领域。该方法中在智能设备处于休眠状态时,采集所述智能设备周围环境的图像(201);若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报(202);语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离(203)。由此,本方法能够根据图像检测,确定周围的用户是否有交互意图,并通过主动进行语音播报的形式,将有潜在交互意图的用户引导到智能设备进行操作,从而提高智能设备的曝光率和使用率。

Description

智能设备控制方法、装置和存储介质
相关申请的交叉引用
本申请要求在2018年11月13日提交中国专利局、申请号为201811346631.3、申请名称为“智能设备控制方法、装置和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及智能设备技术领域,尤其涉及智能设备控制方法、装置和存储介质。
背景技术
随着人工智能技术的不断发展,越来越多的智能设备(如机器人)进入人们的生活和工作领域。
如今各类的机器人能够提供不同的功能,例如育儿机器人能够从服务器获得童话故事、儿歌、诗歌等播放给小朋友。接待机器人能够帮助用户自助办理一些业务(如银行开卡业务)。
然而,智能设备提供的功能虽然较为全面了,但是人们面对陌生的智能设备仍无所适从。
发明内容
本申请实施例提供一种智能设备控制方法、装置和存储介质,用于解决现有技术中智能设备的曝光率和使用率较低等的问题。
第一方面,本申请实施例提供一种智能设备控制方法,该方法包括:
在智能设备处于休眠状态时,采集所述智能设备周围环境的图像;
若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报;
语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离。
在一种可能的实施方式中,所述人物特征包括以下特征的至少一种:人脸特征和人体特征。
在一种可能的实施方式中,所述采集所述智能设备周围环境的图像,包括:
实时采集所述智能设备前方物体的彩色图像和深度图像;
若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在所述第一指定距离内,则触发所述智能设备进行语音播报,包括:
实时基于所述彩色图像进行人物特征检测;
针对检测到的每个人物特征执行:
确定具有该人物特征的目标在深度图像中的对应数据;
根据具有该人物特征的目标在深度图像中的对应数据,确定具有该人物特征的目标与所述智能设备的距离;
若该距离在第一指定距离内,则触发所述智能设备进行语音播报。
在一种可能的实施方式中,所述方法还包括:
在语音播报过程中,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,在语音播报完毕后,不重复进行语音播报。
在一种可能的实施方式中,所述方法还包括:
在语音播报完毕且间隔设定时长后,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,再次进行语音播报。
在一种可能的实施方式中,所述触发所述智能设备进行语音播报,包括:
触发所述智能设备采用大于预设音量的音量进行语音播报,所述预设音量为所述智能设备唤醒后与所述目标交互时的音量。
在一种可能的实施方式中,所述触发所述智能设备进行语音播报,包括:
触发所述智能设备从预设播报内容集合中随机选择播报内容进行语音播报;或者
触发所述智能设备从预设播报内容集合中顺序选择播报内容进行语音播报;或者
触发所述智能设备对预设播报内容集合中的播报内容进行语音循环播报。
在一种可能的实施方式中,所述方法还包括:
接收播报内容增加指令,将所述播报内容增加指令中包含的播报内容添加至所述播报内容集合中;或者
接收播报内容修改指令,根据所述播报内容修改指令对所述播报内容集合中相应的播报内容进行修改;或者
接收播报内容替换指令,根据所述播报内容替换指令替换所述播报内容集合中相应的播报内容。
在一种可能的实施方式中,所述方法还包括:
进行语音播报的同时,在所述智能设备的屏幕上显示预设表情。
第二方面,本申请还提供一种智能设备控制装置,所述装置包括:
图像采集模块,用于在智能设备处于休眠状态时,采集所述智能设备周围环境的图像;
预唤醒模块,用于若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报;
唤醒模块,用于语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离。
在一种可能的实施方式中,所述人物特征包括以下特征的至少一种:人脸特征和人体特征。
在一种可能的实施方式中,所述图像采集模块,具体用于实时采集所述 智能设备前方物体的彩色图像和深度图像;
所述预唤醒模块具体用于:
实时基于所述彩色图像进行人物特征检测;
针对检测到的每个人物特征执行:
确定具有该人物特征的目标在深度图像中的对应数据;
根据具有该人物特征的目标在深度图像中的对应数据,确定具有该人物特征的目标与所述智能设备的距离;
若该距离在所述第一指定距离内,则触发所述智能设备进行语音播报。
在一种可能的实施方式中,所述预唤醒模块还用于:在语音播报过程中,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,在语音播报完毕后,不重复进行语音播报。
在一种可能的实施方式中,所述预唤醒模块还用于:在语音播报完毕且间隔设定时长后,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,再次进行语音播报。
在一种可能的实施方式中,所述预唤醒模块具体用于:
触发所述智能设备采用大于预设音量的音量进行语音播报,所述预设音量为所述智能设备唤醒后与所述目标交互时的音量。
在一种可能的实施方式中,所述预唤醒模块具体用于:
触发所述智能设备从预设播报内容集合中随机选择播报内容进行语音播报;或者
触发所述智能设备从预设播报内容集合中顺序选择播报内容进行语音播报;或者
触发所述智能设备对预设播报内容集合中的播报内容进行语音循环播报。
在一种可能的实施方式中,所述装置还包括:
播报内容维护模块,用于接收播报内容增加指令,将所述播报内容增加指令中包含的播报内容添加至所述播报内容集合中;或者
接收播报内容修改指令,根据所述播报内容修改指令对所述播报内容集 合中相应的播报内容进行修改;或者
接收播报内容替换指令,根据所述播报内容替换指令替换所述播报内容集合中相应的播报内容。
在一种可能的实施方式中,所述装置还包括:
播报表情控制模块,用于进行语音播报的同时,在所述智能设备的屏幕上显示预设表情。
第三方面,本申请还提供一种计算装置,包括至少一个处理器;以及与所述至少一个处理器通信连接的存储器;其中,所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行本申请实施例提供的任一智能设备控制方法。
第四方面,本申请还提供了一种计算机可读存储介质,其中,所述存储介质存储有计算机可执行指令,所述计算机可执行指令用于使计算机执行本申请实施例中的任一智能设备控制方法。
第五方面,本申请还提供了一种计算机程序产品,所述计算机程序产品包括存储在计算机可读存储介质上的计算机程序,所述计算机程序包括程序指令,所述程序指令被处理器执行时实现上述任一智能设备控制方法的步骤。
本申请提供的智能设备控制方法、装置和存储介质。能够根据图像检测,确定周围较远距离的用户是否有交互意图,并通过主动进行语音播报的形式,将有潜在交互意图的用户引导到智能设备进行操作,从而提高智能设备的曝光率和使用率。
本申请的其它特征和优点将在随后的说明书中阐述,并且,部分地从说明书中变得显而易见,或者通过实施本申请而了解。本申请的目的和其他优点可通过在所写的说明书、权利要求书、以及附图中所特别指出的结构来实现和获得。
附图说明
此处所说明的附图用来提供对本申请的进一步理解,构成本申请的一部 分,本申请的示意性实施例及其说明用于解释本申请,并不构成对本申请的不当限定。
图1为本申请实施例中的智能设备控制的应用场景示意图;
图2为本申请实施例中的智能设备控制方法的流程示意图之一;
图3为本申请实施例中的智能设备控制方法的示意图;
图4为本申请实施例中的智能设备控制方法的流程示意图之二;
图5为本申请实施例中的智能设备控制装置示意图;
图6为根据本申请实施方式的计算装置的结构示意图。
具体实施方式
为了提高智能设备的曝光率和使用率,让人们能够更加自然的和智能设备交互,方便的使用智能设备提供的业务,本申请实施例中提供一种智能设备控制方法、装置和存储介质。为了更好的理解本申请实施例提供的技术方案,这里对该方案采用的一些名词和基本原理做一下简单说明:
名词说明:
在睡眠状态之外,智能设备可以包括两个模式,如包括预唤醒模式和唤醒模式。
预唤醒模式:是用户距离智能设备较远时的功能设计。在预唤醒模式下,智能设备实时采集周围环境图像,并在检测到人脸或人体特征时,确定周围有用户存在,开启语音播报功能,例如语音打招呼,播报自身能够提供的业务等,以便于用户了解智能设备的功能,吸引用户走近智能设备,并使用智能设备的具体功能。
唤醒模式:是用户距离智能设备较近时的功能设计,在唤醒模式下,智能设备可以和用户进行交互,依照用户指令以完成相应的服务,也可以主动引导用户以完成相应的服务等。
深度图像:在3D(Dimensions,维度)计算机图形中,Depth Map(深度图像)是包含与视点的场景对象的表面的距离信息的图像。
人物特征:用于描述人的特征,如人脸特征,人体特征。
原理说明:
为了能够提高智能设备的使用效率,本申请实施例中实现了智能设备采用较为自然的交互方案,来引导用户使用智能设备完成相应业务。具体的,当用户靠近智能设备时,就像自然人靠近自然人一样,双方通过打招呼来发起谈话。本申请实施例中,当确定智能设备前方有人时,智能设备可以主动打招呼播报自身的功能。例如,银行接待机器人能够采用自然的交互方式播报“Hi,我是银行助理,请问您需要办理什么业务”。这样,智能设备能够通过自然交互的方式向用户打招呼,吸引用户和智能设备交互,并引导用户在智能设备上完成操作。
进一步的,本申请实施例中为了能够尽可能的少打扰用户,当判断用户离智能设备较近时,才主动打招呼播报语音,而距离较远时并不打招呼,这样可以采用更为自然的方式引导用户使用智能设备。
为了便于进一步理解本申请实施例提供的技术方案,下面结合附图对此进行说明。
如图1所示,为本申请实施例提供的智能设备控制方法的应用场景示意图,该应用场景中包括用户10、智能设备11,智能设备11可以实时采集前方图像,然后根据图像分析前方是否有人,若有人,且距离较近时,则进入预唤醒状态,进行语音播报,以此引导用户使用智能设备。
需要说明的是,本申请实施例中的智能设备可以是能够移动的机器人也可以是固定位置放置的机器人,智能设备提供的功能可以是各行各业各种功能,如前所述的伴随小朋友成长的智能设备,银行接待机器人等。
下面对本申请实施例提供的智能设备控制方法做进一步介绍,如图2所示,该方法包括以下步骤:
步骤201:在智能设备处于休眠状态时,采集所述智能设备周围环境的图像。
步骤202:若在采集的图像中检测到人物特征且具有所述人物特征的目标 与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报。
其中,具体实施时,人物特征检测,可以进行人脸检测和/或人体检测,只要能够确定图像中包含人像均适用于本申请实施例。
播报的语音可以为该智能设备内置功能的服务提供商自行按需设置的。例如银行可以设置其接待智能设备播报“Hi,我是银行小秘,可以为个人用户办理**业务”。
步骤203:语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离。
例如,第一指定距离为3米,第二指定距离为1.5米,即距离用户在1.5-3米内则进行语音播报,之后,确定用户靠近且距离在1.5米内,则可以进入唤醒模式,与用户进行交互。
这样,本申请实施例中若目标与智能设备的距离在第一指定距离内,则确定近距离内有人可能有交互意图,然后通过语音播报实现和目标的打招呼,以便于引导目标靠近智能设备。然后,在确定距离目标很近(即第二指定距离内)时,确定目标有交互意图,开始进入唤醒模式进行交互。这样,通过简单的自然的方式将用户导流到智能设备上进行操作,从而提高智能设备的使用率。
进一步的,确定智能设备与目标的距离的方案可具体实施为以下几种:
方案1、采用距离传感器探测周围环境的目标与智能设备的距离。
这样,是否进行语音播报的方案可理解为当检测到人物特征且探测周围环境的目标与智能设备的距离在第一指定距离内时,则可触发所述智能设备进行语音播报。否则,即使检测到人物特征、但距离传感器没有探测到物体时,也不触发所述智能设备进行语音播报。也即,目标距离太远则确定不存在具有交互意图的用户,无需语音播报进行导流。
相应的,若没有检测到人物特征,无论距离传感器是否探测到物体(该 物体可能是桌子,遮挡物等),都认为不存在具有交互意图的用户,同样不会触发语音播报。
方案2:可以采用RGB-D摄像机来采集图像,然后进行判断。
即实时采集所述智能设备前方物体的彩色图像和深度图像;然后,实时基于所述彩色图像进行人物特征检测;在检测到人物特征后,针对检测到的每个人物特征执行:确定该人物特征的目标在深度图像中的对应数据;根据具有该人物特征的目标在深度图像中的对应数据,确定具有该人物特征的目标与所述智能设备的距离;若该距离在第一指定距离内,则触发所述智能设备进行语音播报。
具体实施时,可以采用红外传感器和红外相机获取深度图像(Depth图像),即通过红外传感器发射红外光,红外光遇到遮挡物后会反射给红外相机,红外相机根据捕捉的红外光成像,得到红外图像中各像素的深度数据。RGB-D摄像机彩色图像(即RGB(red、green、blue,红绿蓝)图像)和Depth图像是配准的,因而两图像的像素点之间具有一一对应关系。故此,根据检测到的人物特征的各像素点,可以确定具有该人物特征的目标和智能设备的距离。由于人物特征像素点有多个,具体实施时,可以采用距离最近的像素点的距离作为距离智能设备的距离,也可以采用最远距离,当然也可以采用人物特征各像素点的距离平均值,本申请对此不作限定。
本申请实施例中,还可以当目标足够近时,直接判断该目标具有交互意图,而进入唤醒模式。如图3所示,第一指定距离为(1.5m-3m),第二指定距离为0-1.5m。连续获取图片后,分析具有人物特征的目标和智能设备的距离,当距离在1.5-3m内时,则执行预唤醒,当距离在1.5m以内时则直接执行唤醒。当然,具体实施时,具体的第一指定距离和第二指定距离可以根据实验或实际需要确定,本申请对此并不作限定。
关于语音播报,具体实施时,智能设备管理人员可以通过语音播报页面自行输入播报的文件(该文件的形式可以是文本文件或音频文件)并保存。然后需要语音播报时,由智能设备根据预存的文件进行语音播报。若是文本 文件,则可以采用文本到语音(Text To Speech,TTS)合成技术进行播报。
当然具体实施时可以设置多条播报内容构成预设播报内容集合。语音播报时触发智能设备从预设播报内容集合中随机选择播报内容进行语音播报,或者,触发所述智能设备从预设播报内容集合中顺序选择播报内容进行语音播报,亦或者,触发所述智能设备对预设播报内容集合中的播报内容进行语音循环播报。
当然,随机选择时,可以随机选择一条,也可以随机选择多条进行播报。具体实施时,每条语音的字数具有字数上限,语音的条数也可以具有上限(如10条)。
相应的,用户也可以对自己配置的播报内容进行更新,具体可包括:接收播报内容增加指令,将所述播报内容增加指令中包含的播报内容添加至所述播报内容集合中;或者,接收播报内容修改指令,根据所述播报内容修改指令对所述播报内容集合中相应的播报内容进行修改;亦或者,接收播报内容替换指令,根据所述播报内容替换指令替换所述播报内容集合中相应的播报内容。
由此,用户通过上述增删改的操作可以实现对播报内容的维护。
关于语音播报除了上述内容,下面将补充几中场景的对语音内容的播报进行补充说明,包括:
1)、在语音播报过程中,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,在语音播报完毕后,不重复进行语音播报。
也即,连续监测到具有人物特征的目标距离智能设备在第一指定距离内时,不会连续重复播报语音。这样,能够避免连续打扰用户,让用户产生听觉疲劳。同时,也能节约由于语音播报耗费的资源。
2)、本申请实施例中,在语音播报完毕且间隔设定时长后,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,再次进行语音播报。
例如,本次语音播报完毕后,间隔5秒钟,5秒钟之后若满足语音播报条件时,则再次进行语音播报。这样,采用周期性检测并确定是否进行语音播报,避免了持续播报语音导致电能和处理资源的损耗,实现在资源的使用和及时导流用户之间实现平衡。
3)、语音播报时,可以触发所述智能设备采用大于预设音量的音量进行语音播报,所述预设音量为所述智能设备唤醒后与目标交互时的音量。这样,智能设备播报的语音内容可以引起用户注意,以便于能够更加有效的将具有潜在交互意图的用户导流到智能设备上操作。
4)、为了便于能够让用户以更加自然的方式理解和收听播报的内容,本申请中进行语音播报的同时,还可以在所述智能设备的屏幕上显示预设表情。如,漏出一个笑脸,以便于用户除了在听觉上能够对智能设备产生兴趣,在视觉上也能进一步的对智能设备产生兴趣,以便于将用户导流到智能设备上。
本申请实施例中,具体实施时,人物特征检测和语音播报可以为独立的两个程序,即图像的采集和人物特征检测可以实时进行,并在检测到人物特征时确定具有该人物特征的目标与智能设备之间的距离,若该距离在第一指定距离范围内,则发送消息给语音播报程序,语音播报程序根据收到的消息确定是否进行语音播报。
以前述2)中的间隔设定时长进行语音播报,对这两个程序的配合使用进行说明。具体可实施为假设指定时长为5秒,一开始,语音播报程序在收到人物特征检测程序发送的消息后,开始播报“Hi,我是接待智能设备XXX,请问有什么需求?”,在该语音还没有播报完,又收到人物特征检测程序的消息时,则忽略该消息。播报完开始计时,若在5秒之内,收到人物特征检测程序的消息,则丢弃该消息。若在5秒之后,收到消息,则进行语音播报。
具体实施时,还可以根据采集的图像确定是否有人进入智能设备的视野或离开智能设备的视野。若在语音播报过程中,视野中所有人离开智能设备视野时,则可以终止播报,达到节约资源的目的。
为了便于系统的理解本申请实施例提供的智能设备的控制方法,结合图4 对此进行说明,包括以下步骤:
步骤401:在智能设备处于休眠状态时,采集所述智能设备周围环境的图像。
步骤402:实时基于所述彩色图像进行人物特征检测。
步骤403:若检测到人脸,则针对检测到的人脸,在深度图像中查找该人脸的对应数据;并,根据该对应数据,确定该人脸对应的目标与智能设备的距离。
步骤404:确定是否存在至少一个目标距离智能设备的距离在1.5米内,若是,执行步骤405,否则,执行步骤406。
步骤405:唤醒所述智能设备。
步骤406:确定是否存在至少一个目标距离智能设备的距离在1.5-3米内,若是,执行步骤407,否则,执行步骤401。
步骤407:确定当前是否正在进行语音播报,若是,则返回执行步骤401,若否,则执行步骤408。
步骤408:距离上次语音播报的时长是否达到设定时长,若否,则返回执行步骤401,是则,执行步骤409。
步骤409:随机选择一条播报内容,并采用大于指定分贝的音量播报选择的播报内容。
步骤410:语音播报完毕后,确定目标与智能设备的距离,之后执行步骤404。
基于相同的构思,本申请实施例还提供一种智能设备控制装置。如图5所示,该装置包括:
图像采集模块501,用于在智能设备处于休眠状态时,采集所述智能设备周围环境的图像;
预唤醒模块502,用于若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能 设备进行语音播报;
唤醒模块503,用于语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离。
在一种可能的实施方式中,所述人物特征包括以下特征的至少一种:人脸特征和人体特征。
在一种可能的实施方式中,所述图像采集模块,具体用于实时采集所述智能设备前方物体的彩色图像和深度图像;
所述预唤醒模块具体用于:
实时基于所述彩色图像进行人物特征检测;
针对检测到的每个人物特征执行:
确定具有该人物特征的目标在深度图像中的对应数据;
根据具有该人物特征的目标在深度图像中的对应数据,确定具有该人物特征的目标与所述智能设备的距离;
若该距离在所述第一指定距离内,则触发所述智能设备进行语音播报。
在一种可能的实施方式中,所述预唤醒模块还用于:
在语音播报过程中,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,在语音播报完毕后,不重复进行语音播报。
在一种可能的实施方式中,所述预唤醒模块还用于:
在语音播报完毕且间隔设定时长后,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,再次进行语音播报。
在一种可能的实施方式中,所述预唤醒模块具体用于:
触发所述智能设备采用大于预设音量的音量进行语音播报,所述预设音量为所述智能设备唤醒后与目标交互时的音量。
在一种可能的实施方式中,所述预唤醒模块具体用于:
触发所述智能设备从预设播报内容集合中随机选择播报内容进行语音播报;或者
触发所述智能设备从预设播报内容集合中顺序选择播报内容进行语音播报;或者
触发所述智能设备对预设播报内容集合中的播报内容进行语音循环播报。
在一种可能的实施方式中,所述装置还包括:
播报内容维护模块,用于接收播报内容增加指令,将所述播报内容增加指令中包含的播报内容添加至所述播报内容集合中;或者
接收播报内容修改指令,根据所述播报内容修改指令对所述播报内容集合中相应的播报内容进行修改;或者
接收播报内容替换指令,根据所述播报内容替换指令替换所述播报内容集合中相应的播报内容。
在一种可能的实施方式中,所述装置还包括:
播报表情控制模块,用于进行语音播报的同时,在所述智能设备的屏幕上显示预设表情。
在介绍了本申请示例性实施方式的智能设备控制方法、装置之后,接下来,介绍根据本申请的另一示例性实施方式的计算装置。
所属技术领域的技术人员能够理解,本申请的各个方面可以实现为系统、方法或程序产品。因此,本申请的各个方面可以具体实现为以下形式,即:完全的硬件实施方式、完全的软件实施方式(包括固件、微代码等),或硬件和软件方面结合的实施方式,这里可以统称为“电路”、“模块”或“系统”。
在一些可能的实施方式中,根据本申请的计算装置可以至少包括至少一个处理器、以及至少一个存储器。其中,存储器存储有程序代码,当程序代码被处理器执行时,使得处理器执行本说明书上述描述的根据本申请各种示例性实施方式的智能设备控制方法中的步骤。例如,处理器可以执行如图2中所示的步骤201-203。
下面参照图6来描述根据本申请的这种实施方式的计算装置130。图6显 示的计算装置130仅仅是一个示例,不应对本申请实施例的功能和使用范围带来任何限制。
如图6所示,计算装置130以通用计算装置的形式表现。计算装置130的组件可以包括但不限于:上述至少一个处理器131、上述至少一个存储器132、连接不同系统组件(包括存储器132和处理器131)的总线133。
总线133表示几类总线结构中的一种或多种,包括存储器总线或者存储器控制器、外围总线、处理器或者使用多种总线结构中的任意总线结构的局域总线。
存储器132可以包括易失性存储器形式的可读介质,例如随机存取存储器(RAM)1321和/或高速缓存存储器1322,还可以进一步包括只读存储器(ROM)1323。
存储器132还可以包括具有一组(至少一个)程序模块1324的程序/实用工具1325,这样的程序模块1324包括但不限于:操作系统、一个或者多个应用程序、其它程序模块以及程序数据,这些示例中的每一个或某种组合中可能包括网络环境的实现。
计算装置130也可以与一个或多个外部设备134(例如键盘、指向设备等)通信,还可与一个或者多个使得用户能与计算装置130交互的设备通信,和/或与使得该计算装置130能与一个或多个其它计算装置进行通信的任何设备(例如路由器、调制解调器等等)通信。这种通信可以通过输入/输出(I/O)接口135进行。并且,计算装置130还可以通过网络适配器136与一个或者多个网络(例如局域网(LAN),广域网(WAN)和/或公共网络,例如因特网)通信。如图所示,网络适配器136通过总线133与用于计算装置130的其它模块通信。应当理解,尽管图中未示出,可以结合计算装置130使用其它硬件和/或软件模块,包括但不限于:微代码、设备驱动器、冗余处理器、外部磁盘驱动阵列、RAID系统、磁带驱动器以及数据备份存储系统等。
在一些可能的实施方式中,本申请提供的智能设备控制方法的各个方面还可以实现为一种程序产品的形式,其包括程序代码,当程序产品在计算机 设备上运行时,程序代码用于使计算机设备执行本说明书上述描述的根据本申请各种示例性实施方式的智能设备控制方法中的步骤,例如,计算机设备可以执行如图2中所示的步骤201-203。
程序产品可以采用一个或多个可读介质的任意组合。可读介质可以是可读信号介质或者可读存储介质。可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。
本申请的实施方式的用于智能设备控制的程序产品可以采用便携式紧凑盘只读存储器(CD-ROM)并包括程序代码,并可以在计算装置上运行。然而,本申请的程序产品不限于此,在本文件中,可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。
可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了可读程序代码。这种传播的数据信号可以采用多种形式,包括——但不限于——电磁信号、光信号或上述的任意合适的组合。可读信号介质还可以是可读存储介质以外的任何可读介质,该可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。
可读介质上包含的程序代码可以用任何适当的介质传输,包括——但不限于——无线、有线、光缆、RF等等,或者上述的任意合适的组合。
可以以一种或多种程序设计语言的任意组合来编写用于执行本申请操作的程序代码,程序设计语言包括面向对象的程序设计语言—诸如Java、C++等,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算装置上执行、部分地在用户设备上执行、作 为一个独立的软件包执行、部分在用户计算装置上部分在远程计算装置上执行、或者完全在远程计算装置或服务器上执行。在涉及远程计算装置的情形中,远程计算装置可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算装置,或者,可以连接到外部计算装置(例如利用因特网服务提供商来通过因特网连接)。
应当注意,尽管在上文详细描述中提及了装置的若干单元或子单元,但是这种划分仅仅是示例性的并非强制性的。实际上,根据本申请的实施方式,上文描述的两个或更多单元的特征和功能可以在一个单元中具体化。反之,上文描述的一个单元的特征和功能可以进一步划分为由多个单元来具体化。
此外,尽管在附图中以特定顺序描述了本申请方法的操作,但是,这并非要求或者暗示必须按照该特定顺序来执行这些操作,或是必须执行全部所示的操作才能实现期望的结果。附加地或备选地,可以省略某些步骤,将多个步骤合并为一个步骤执行,和/或将一个步骤分解为多个步骤执行。
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设 备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。
尽管已描述了本申请的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例做出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本申请范围的所有变更和修改。
显然,本领域的技术人员可以对本申请进行各种改动和变型而不脱离本申请的精神和范围。这样,倘若本申请的这些修改和变型属于本申请权利要求及其等同技术的范围之内,则本申请也意图包含这些改动和变型在内。

Claims (12)

  1. 一种智能设备控制方法,其特征在于,所述方法包括:
    在智能设备处于休眠状态时,采集所述智能设备周围环境的图像;
    若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报;
    语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离。
  2. 根据权利要求1所述的方法,其特征在于,所述人物特征包括以下特征的至少一种:人脸特征和人体特征。
  3. 根据权利要求1所述的方法,其特征在于,所述采集所述智能设备周围环境的图像,包括:
    实时采集所述智能设备前方物体的彩色图像和深度图像;
    若在采集的图像中检测到人物特征且具有所述人物特征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报,包括:
    实时基于所述彩色图像进行人物特征检测;
    针对检测到的每个人物特征执行:
    确定具有该人物特征的目标在深度图像中的对应数据;
    根据具有该人物特征的目标在深度图像中的对应数据,确定具有该人物特征的目标与所述智能设备的距离;
    若该距离在所述第一指定距离内,则触发所述智能设备进行语音播报。
  4. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    在语音播报过程中,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,在语音播报完毕后,不重复进行语音播报。
  5. 根据权利要求1所述的方法,其特征在于,所述方法还包括:
    在语音播报完毕且间隔设定时长后,若检测到至少一个具有人物特征的其他目标与所述智能设备之间的距离在所述第一指定距离内,再次进行语音播报。
  6. 根据权利要求1所述的方法,其特征在于,所述触发所述智能设备进行语音播报,包括:
    触发所述智能设备采用大于预设音量的音量进行语音播报,所述预设音量为所述智能设备唤醒后与所述目标交互时的音量。
  7. 根据权利要求1-6任一所述的方法,其特征在于,所述触发所述智能设备进行语音播报,包括:
    触发所述智能设备从预设播报内容集合中随机选择播报内容进行语音播报;或者
    触发所述智能设备从预设播报内容集合中顺序选择播报内容进行语音播报;或者
    触发所述智能设备对预设播报内容集合中的播报内容进行语音循环播报。
  8. 根据权利要求7所述的方法,其特征在于,所述方法还包括:
    接收播报内容增加指令,将所述播报内容增加指令中包含的播报内容添加至所述播报内容集合中;或者
    接收播报内容修改指令,根据所述播报内容修改指令对所述播报内容集合中相应的播报内容进行修改;或者
    接收播报内容替换指令,根据所述播报内容替换指令替换所述播报内容集合中相应的播报内容。
  9. 根据权利要求1-6任一所述的方法,其特征在于,所述方法还包括:
    进行语音播报的同时,在所述智能设备的屏幕上显示预设表情。
  10. 一种智能设备控制装置,其特征在于,所述装置包括:
    图像采集模块,用于在智能设备处于休眠状态时,采集所述智能设备周围环境的图像;
    预唤醒模块,用于若在采集的图像中检测到人物特征且具有所述人物特 征的目标与所述智能设备之间的距离在第一指定距离内,则触发所述智能设备进行语音播报;
    唤醒模块,用于语音播报之后,若检测到所述目标与所述智能设备之间的距离在第二指定距离内,则唤醒所述智能设备,所述第二指定距离小于所述第一指定距离。
  11. 一种计算机可读存储介质,其特征在于,所述存储介质存储有计算机可执行指令,所述计算机可执行指令用于使计算机执行如权利要求1-9中任一权利要求所述的方法。
  12. 一种计算装置,其特征在于,包括:至少一个处理器;以及与所述至少一个处理器通信连接的存储器;其中,所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行如权利要求1-9中任一权利要求所述的方法。
PCT/CN2019/110191 2018-11-13 2019-10-09 智能设备控制方法、装置和存储介质 WO2020098421A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP19884579.4A EP3866054B1 (en) 2018-11-13 2019-10-09 Smart device control method and apparatus, and storage medium
JP2021526383A JP7204912B2 (ja) 2018-11-13 2019-10-09 インテリジェント機器を制御するための方法、デバイスおよび記憶媒体
US17/293,770 US11247342B2 (en) 2018-11-13 2019-10-09 Smart device control method and apparatus, and storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811346631.3A CN109543578B (zh) 2018-11-13 2018-11-13 智能设备控制方法、装置和存储介质
CN201811346631.3 2018-11-13

Publications (1)

Publication Number Publication Date
WO2020098421A1 true WO2020098421A1 (zh) 2020-05-22

Family

ID=65847106

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/110191 WO2020098421A1 (zh) 2018-11-13 2019-10-09 智能设备控制方法、装置和存储介质

Country Status (6)

Country Link
US (1) US11247342B2 (zh)
EP (1) EP3866054B1 (zh)
JP (1) JP7204912B2 (zh)
CN (1) CN109543578B (zh)
TW (1) TWI741406B (zh)
WO (1) WO2020098421A1 (zh)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109543578B (zh) * 2018-11-13 2020-12-22 北京猎户星空科技有限公司 智能设备控制方法、装置和存储介质
CN109940638B (zh) * 2019-04-26 2021-04-16 北京猎户星空科技有限公司 机器人、机器人控制方法、装置、存储介质和控制器
CN110253595B (zh) * 2019-06-21 2022-01-21 北京猎户星空科技有限公司 一种智能设备控制方法和装置
CN112207812A (zh) * 2019-07-12 2021-01-12 阿里巴巴集团控股有限公司 设备控制方法、设备、系统及存储介质
CN110427887B (zh) * 2019-08-02 2023-03-10 腾讯科技(深圳)有限公司 一种基于智能的会员身份识别方法及装置
CN113556649B (zh) * 2020-04-23 2023-08-04 百度在线网络技术(北京)有限公司 智能音箱的播报控制方法和装置
CN112130918A (zh) * 2020-09-25 2020-12-25 深圳市欧瑞博科技股份有限公司 智能设备唤醒方法、装置、系统及智能设备
CN114047901B (zh) * 2021-11-25 2024-03-15 阿里巴巴(中国)有限公司 人机交互方法及智能设备
CN114310909A (zh) * 2022-01-26 2022-04-12 美的集团(上海)有限公司 一种机器人及其控制方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140313303A1 (en) * 2013-04-18 2014-10-23 Digimarc Corporation Longitudinal dermoscopic study employing smartphone-based image registration
CN104305648A (zh) * 2014-10-28 2015-01-28 广东小天才科技有限公司 一种节能手环及应用于该节能手环的控制方法
CN108733420A (zh) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 智能设备的唤醒方法、装置、智能设备和存储介质
CN108733417A (zh) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 智能设备的工作模式选择方法和装置
CN109543578A (zh) * 2018-11-13 2019-03-29 北京猎户星空科技有限公司 智能设备控制方法、装置和存储介质

Family Cites Families (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04335399A (ja) * 1991-05-11 1992-11-24 Colleen Denshi Kk 音量自動調節機能を備えた音声出力装置
JP2003122382A (ja) * 2001-10-11 2003-04-25 Seiko Epson Corp 携帯型の複合型再生装置
JP5732632B2 (ja) * 2011-02-03 2015-06-10 株式会社国際電気通信基礎技術研究所 ロボットシステムおよびそれに用いる空間陣形認識装置
JP2013003221A (ja) * 2011-06-14 2013-01-07 Tomy Co Ltd 楽音出力装置
JP5803470B2 (ja) * 2011-09-15 2015-11-04 富士ゼロックス株式会社 電力供給制御装置、画像処理装置、電力供給制御プログラム
KR102158208B1 (ko) * 2013-07-26 2020-10-23 엘지전자 주식회사 전자기기 및 그것의 제어 방법
CN105845135A (zh) 2015-01-12 2016-08-10 芋头科技(杭州)有限公司 一种机器人系统的声音识别系统及方法
JP2017069876A (ja) * 2015-10-01 2017-04-06 富士ゼロックス株式会社 処理装置
JP6610349B2 (ja) * 2016-03-10 2019-11-27 大日本印刷株式会社 案内ロボット制御システム、プログラム及び案内ロボット
JP6707641B2 (ja) * 2015-12-11 2020-06-10 ソムニック インク. 定置状態の検出によってユーザおよび/または外部装置とのインタフェースを形成する装置、システムおよび方法
JP6599803B2 (ja) * 2016-03-08 2019-10-30 シャープ株式会社 発話装置
US20180053504A1 (en) * 2016-08-19 2018-02-22 Otis Elevator Company Intention recognition for triggering voice recognition system
JP2018067785A (ja) * 2016-10-19 2018-04-26 前川 博文 コミュニケーションロボットシステム
JP2018152704A (ja) * 2017-03-13 2018-09-27 富士ゼロックス株式会社 情報処理装置
JP6673276B2 (ja) * 2017-03-28 2020-03-25 カシオ計算機株式会社 音声検出装置、音声検出方法、及びプログラム
CN106956274A (zh) * 2017-03-28 2017-07-18 旗瀚科技有限公司 一种机器人唤醒方法
TWI621071B (zh) * 2017-08-31 2018-04-11 國立勤益科技大學 使用深度學習於車牌與人臉辨識之門禁管理系統
CN108153169A (zh) * 2017-12-07 2018-06-12 北京康力优蓝机器人科技有限公司 导览模式切换方法、系统和导览机器人
CN107972048A (zh) * 2017-12-28 2018-05-01 沈阳建筑大学 一种智能多功能导游机器人及其控制方法
US10709335B2 (en) * 2017-12-31 2020-07-14 Google Llc Infant monitoring system with observation-based system control and feedback loops
CN108519819A (zh) * 2018-03-30 2018-09-11 北京金山安全软件有限公司 智能设备的处理方法、装置、智能设备及介质

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140313303A1 (en) * 2013-04-18 2014-10-23 Digimarc Corporation Longitudinal dermoscopic study employing smartphone-based image registration
CN104305648A (zh) * 2014-10-28 2015-01-28 广东小天才科技有限公司 一种节能手环及应用于该节能手环的控制方法
CN108733420A (zh) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 智能设备的唤醒方法、装置、智能设备和存储介质
CN108733417A (zh) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 智能设备的工作模式选择方法和装置
CN109543578A (zh) * 2018-11-13 2019-03-29 北京猎户星空科技有限公司 智能设备控制方法、装置和存储介质

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3866054A4 *

Also Published As

Publication number Publication date
US20210354306A1 (en) 2021-11-18
JP2022507447A (ja) 2022-01-18
JP7204912B2 (ja) 2023-01-16
EP3866054A1 (en) 2021-08-18
US11247342B2 (en) 2022-02-15
TWI741406B (zh) 2021-10-01
TW202018587A (zh) 2020-05-16
CN109543578A (zh) 2019-03-29
EP3866054B1 (en) 2023-12-06
EP3866054C0 (en) 2023-12-06
CN109543578B (zh) 2020-12-22
EP3866054A4 (en) 2021-12-01

Similar Documents

Publication Publication Date Title
WO2020098421A1 (zh) 智能设备控制方法、装置和存储介质
CN107995101B (zh) 一种用于将语音消息转为文本信息的方法与设备
CN110730952B (zh) 处理网络上的音频通信的方法和系统
AU2014200039B2 (en) User interface method and system for use in mobile terminal
WO2022148083A1 (zh) 仿真3d数字人交互方法、装置、电子设备及存储介质
CN109940638B (zh) 机器人、机器人控制方法、装置、存储介质和控制器
CN108470533A (zh) 基于虚拟人的增强型智能交互广告系统及装置
TW201923737A (zh) 交互方法和設備
CN107704169B (zh) 虚拟人的状态管理方法和系统
CN204631465U (zh) 一种远距离语音控制的人性化智能家居控制系统
CN110253595B (zh) 一种智能设备控制方法和装置
WO2020062670A1 (zh) 电器设备的控制方法、装置、电器设备和介质
WO2018006496A1 (zh) 智能随身翻译器
CN105446953A (zh) 一种智能机器人与虚拟3d的交互系统及方法
CN106572131B (zh) 物联网中媒体数据分享的方法及系统
CN107301867A (zh) 一种语音重启控制系统
CN207134134U (zh) 一种语音内容实时展示的系统
CN104506735A (zh) 一种移动终端及其控制方法
CN105516649A (zh) 用于宠物的远程监控及人与宠物智能交互系统
CN104506736A (zh) 一种移动终端及其控制方法
CN104504853A (zh) 一种移动终端及其控制装置
CN111312243B (zh) 设备交互方法和装置
CN207337339U (zh) 透明显示互动系统
CN209767565U (zh) 一种基于新型智能管家系统
CN210038806U (zh) 基于人脸识别的学生课堂综合评估系统

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19884579

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2021526383

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2019884579

Country of ref document: EP

Effective date: 20210512