CN111312248A - Interaction method, device, system and storage medium - Google Patents

Interaction method, device, system and storage medium Download PDF

Info

Publication number
CN111312248A
CN111312248A CN202010103936.2A CN202010103936A CN111312248A CN 111312248 A CN111312248 A CN 111312248A CN 202010103936 A CN202010103936 A CN 202010103936A CN 111312248 A CN111312248 A CN 111312248A
Authority
CN
China
Prior art keywords
video stream
screen
information
sound box
equipment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010103936.2A
Other languages
Chinese (zh)
Inventor
王斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Wentai Information Technology Co Ltd
Original Assignee
Shanghai Wentai Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Wentai Information Technology Co Ltd filed Critical Shanghai Wentai Information Technology Co Ltd
Priority to CN202010103936.2A priority Critical patent/CN111312248A/en
Publication of CN111312248A publication Critical patent/CN111312248A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the invention discloses an interaction method, an interaction device, an interaction system and a storage medium. The method comprises the following steps: the method comprises the steps of collecting voice information of an interactive object, sending object information to a cloud server, receiving a voice analysis result of the cloud server on the voice information, and awakening a screen device when the voice analysis result meets a preset condition so as to display a video stream corresponding to the voice information on the screen device. Compared with the prior art, the embodiment of the invention fully utilizes the characteristics of the equipment with the screen, displays the video stream on the equipment with the screen, reduces the visual interaction cost of the intelligent sound box and enlarges the interaction range.

Description

Interaction method, device, system and storage medium
Technical Field
The embodiment of the invention relates to the technical field of human-computer interaction, in particular to an interaction method, an interaction device, an interaction system and a storage medium.
Background
The intelligent sound box is used as an upgrade product of the sound box, and the types of the intelligent sound box are more and more. However, the interaction between the user and the smart sound box is limited to simple voice interaction, the interaction between the mobile phone APP and the smart sound box, the sound box with a screen and the like. These several kinds of modes all have certain limitation, can't realize effectual mutual, for example, the unable quick transmission of realization information of simple voice interaction mode, cell-phone APP need rely on user's hand with intelligent audio amplifier's interaction mode, the interaction mode of taking the screen audio amplifier is the transmission of realization information that can be quick, but has increased the hardware cost, moreover because the limitation of audio amplifier volume for the screen undersize of taking the screen audio amplifier, interaction range has certain limitation.
Disclosure of Invention
The embodiment of the invention provides an interaction method, an interaction device, an interaction system and a storage medium, which are used for realizing effective interaction with an intelligent sound box.
In a first aspect, an embodiment of the present invention provides an interaction method, which is applied to a smart speaker, where the method includes:
collecting object information of an interactive object;
sending the object information to a cloud server;
receiving an analysis result of the cloud server on the object information;
and when the analysis result meets a preset condition, awakening the equipment with the screen so as to display the video stream corresponding to the object information on the equipment with the screen.
In a second aspect, an embodiment of the present invention further provides an interaction method, which is applied to a device with a screen, where the method includes:
receiving a wake-up instruction sent by the intelligent sound box;
when the equipment with the screen is awakened, feeding back a response result of the awakening instruction to the intelligent sound box;
and receiving and displaying the video stream sent by the intelligent sound box.
In a third aspect, an embodiment of the present invention further provides an interaction device, disposed in an intelligent sound box, including:
the information acquisition module is used for acquiring object information of the interactive object;
the information sending module is used for sending the object information to a cloud server;
the first analysis result receiving module is used for receiving the analysis result of the cloud server on the object information;
and the awakening module is used for awakening the equipment with the screen when the analysis result meets the preset condition so as to display the video stream corresponding to the object information on the equipment with the screen.
Optionally, the first parsing result receiving module is specifically configured to:
determining a video stream corresponding to the analysis result;
when the analysis result contains visual interaction information, generating and sending a wake-up instruction to the screen equipment;
and when a response result of the awakening instruction fed back by the screen equipment is received, sending the video stream to the screen equipment so as to display the video stream through the screen equipment.
Optionally, the apparatus further comprises:
a video stream update request receiving module, configured to receive a video stream update request sent by the device with screen after a video stream corresponding to the object information is displayed on the device with screen;
and the updating module is used for updating the video stream according to the video stream updating request and sending the video stream to the screen equipment.
Optionally, the apparatus further comprises:
the video stream updating information acquisition module is used for acquiring and sending the video stream updating information of the interactive object to the cloud server after the video stream corresponding to the object information is displayed on the screen-equipped device;
the second analysis result receiving module is used for receiving the analysis result of the cloud server on the video stream updating information;
and the target video stream determining module is used for determining a target video stream according to the analysis result of the video stream updating information and sending the target video stream to the screen equipment.
Optionally, the apparatus further comprises:
and the sound box awakening information receiving module is used for receiving the sound box awakening information of the interactive object before acquiring and sending the object information of the interactive object to the cloud server.
In a fourth aspect, an embodiment of the present invention further provides an interaction apparatus, which is disposed in a device with a screen, and includes:
the awakening instruction receiving module is used for receiving an awakening instruction sent by the intelligent sound box;
the response result feedback module is used for feeding back a response result of the awakening instruction to the intelligent sound box when the equipment with the screen is awakened;
and the display module is used for receiving and displaying the video stream sent by the intelligent sound box.
Optionally, the apparatus further comprises:
the video stream updating information capturing module is used for capturing video stream updating information of an interactive object after receiving and displaying the video stream sent by the intelligent sound box;
and the analysis module is used for analyzing the video stream updating information, generating a video stream updating request based on an analysis result and sending the video stream updating request to the intelligent sound box.
In a fifth aspect, an embodiment of the present invention further provides an interactive system, including: the intelligent sound box is respectively in communication connection with the cloud server and the screen equipment;
the intelligent sound box is used for collecting object information of an interactive object; sending the object information to a cloud server; receiving an analysis result of the cloud server on the object information; when the received analysis result sent by the cloud server meets a preset condition, awakening the equipment with the screen;
the cloud server is used for analyzing the object information sent by the intelligent sound box to obtain an analysis result;
the screen equipment is used for receiving a wake-up instruction sent by the intelligent sound box; when the equipment with the screen is awakened, feeding back a response result of the awakening instruction to the intelligent sound box; and receiving and displaying the video stream sent by the intelligent sound box.
In a sixth aspect, an embodiment of the present invention further provides a storage medium, on which a computer program is stored, where the computer program, when executed by a corresponding apparatus, implements the interaction method according to the first aspect or the second aspect.
The embodiment of the invention provides an interaction method, an interaction device, an interaction system and a storage medium, wherein voice information of an interaction object is collected, object information is sent to a cloud server, a voice analysis result of the cloud server on the voice information is received, and when the voice analysis result meets a preset condition, a screen device is awakened so as to display a video stream corresponding to the voice information on the screen device. Compared with the prior art, the embodiment of the invention fully utilizes the characteristics of the equipment with the screen, displays the video stream on the equipment with the screen, reduces the visual interaction cost of the intelligent sound box and enlarges the interaction range.
Drawings
FIG. 1 is a flow chart of an interaction method provided by an embodiment of the invention;
FIG. 2 is a flow chart of an interaction method according to another embodiment of the present invention;
FIG. 3 is a flow chart of an interaction method according to another embodiment of the invention;
FIG. 4 is a block diagram of an interactive apparatus according to an embodiment of the present invention;
FIG. 5 is a block diagram of an interactive apparatus according to another embodiment of the present invention;
fig. 6 is a block diagram of an interactive system according to an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be further noted that, for the convenience of description, only some of the structures related to the present invention are shown in the drawings, not all of the structures. In addition, the embodiments and features of the embodiments in the present invention may be combined with each other without conflict.
In an embodiment, fig. 1 is a flowchart of an interaction method according to an embodiment of the present invention, where the embodiment is applicable to a case of human-computer interaction, and the method may be executed by an interaction device, where the device may be implemented in a software and/or hardware manner and may be configured in an intelligent speaker, and referring to fig. 1, the method may include the following steps:
and S110, collecting object information of the interactive object.
Optionally, the interactive object includes a user of the smart speaker. The object information is information used by the user for interacting with the smart sound box, and may be, for example, voice information, gesture information, remote control information, or the like, where the gesture information may be a gesture action of the user, and different gesture actions represent different meanings, for example, a horizontal index finger represents opening music, a vertical index finger represents opening news, and interaction with the smart sound box through the gesture information is achieved, and is not limited to voice interaction. The association relationship between the gesture action and the meaning can be set according to the requirement, and the embodiment is not limited. The remote control information may be information that a user controls the smart speaker through a remote controller or the like. The smart sound box in this embodiment has a screen projection function, that is, a formed video stream may be projected onto other devices with screens, for example, a TV device, a mobile phone, or a projector, and the smart sound box may have a screen or may not have a screen, and when the smart sound box has a screen, the object information may also be touch information.
Optionally, before interacting with the smart speaker, the user needs to wake up the smart speaker first, for example, the user may send voice wake-up information to the smart speaker, where the voice wake-up information may include turning on the smart speaker, turning on music, or turning on shopping, and when the smart speaker recognizes that the voice wake-up information includes a keyword "turn on" or "turn on", the user may wake up the smart speaker. For another example, the smart speaker may be awakened through a gesture or a physical button, where the physical button may be an open button on the smart speaker. And after the intelligent sound box is awakened, the object information of the interactive object can be collected, and interaction with the user is carried out.
And S120, sending the object information to a cloud server.
The cloud server is used for analyzing the object information acquired by the intelligent sound box and determining the specific content of the object information. Illustratively, taking the object information as the voice information as an example, the smart sound box sends the collected voice information of the user to the cloud server, and the voice information is identified by the cloud server.
And S130, receiving an analysis result of the cloud server on the object information.
And S140, when the analysis result meets the preset condition, awakening the screen equipment so as to display the video stream corresponding to the object information on the screen equipment.
The screen-equipped equipment can be used for installing screen projection application programs and is provided with a screen, can interact with the intelligent sound box and display video streams formed by the intelligent sound box, so that the defect that the intelligent sound box does not have the screen can be overcome, the interaction range of a user can be expanded when the screen of the intelligent sound box is small, and the user does not need to be limited in the setting range of the intelligent sound box. The screen projecting application program in the screen projecting equipment can be self-contained when leaving a factory, and can also be self-contained by a user in the later period according to the requirement, and the screen projecting equipment can interact with the intelligent sound box after the screen projecting application program is started, so that the transmission and the display of video streams are completed. Optionally, when the screen-equipped device is a TV device or a projector, the large screen characteristic of the TV device or the projector is fully utilized, and the visual interaction cost of the smart sound box is reduced. The preset condition is a condition for starting a screen projection function of the intelligent sound box, for example, when a keyword in an analysis result relates to visual interaction, the screen projection function of the intelligent sound box is started, a wakeup instruction is sent to the equipment with the screen, the equipment with the screen can receive and display a video stream sent by the intelligent sound box after being awakened, and a user can obtain required information through the equipment with the screen, wherein the keyword relating to the visual interaction can be 'video', 'shopping', 'news', 'music', or the like.
The embodiment of the invention provides an interaction method, which is applied to an intelligent sound box, and is characterized in that voice information of an interaction object is collected, object information is sent to a cloud server, a voice analysis result of the cloud server on the voice information is received, and when the voice analysis result meets a preset condition, a screen device is awakened, so that a video stream corresponding to the voice information is displayed on the screen device. Compared with the prior art, the embodiment of the invention fully utilizes the characteristics of the equipment with the screen, displays the video stream on the equipment with the screen, reduces the visual interaction cost of the intelligent sound box and enlarges the interaction range.
Optionally, on the basis of the above embodiment, when the user has a new demand, the smart sound box may further collect video stream update information of the user, and update the currently displayed video stream according to the video stream update information. Correspondingly, after the video stream corresponding to the object information is displayed on the device with the screen, the method further comprises the following steps:
acquiring and sending video stream updating information of an interaction object to a cloud server;
receiving an analysis result of the video stream updating information by the cloud server;
and determining the target video stream according to the analysis result of the video stream updating information, and sending the target video stream to the equipment with the screen.
The video stream updating information can be voice information or gesture actions, the intelligent sound box sends the collected voice information or gesture actions to the cloud server, the voice information or gesture actions are analyzed by the cloud server, then the video stream of the intelligent sound box is changed and synchronously sent to the equipment with the screen according to the analysis result sent by the cloud server, and the updating requirements of users are met. Certainly, the analysis result of the voice information or the gesture action can also be determined by the smart sound box, and in the embodiment, the cloud server is used for analysis as an example, so that the accuracy of the analysis result can be improved.
In another embodiment, fig. 2 is a flowchart of an interaction method according to another embodiment of the present invention, where the present embodiment is optimized based on the foregoing embodiment, and with reference to fig. 2, the method includes the following steps:
s210, receiving sound box awakening information of the interactive object.
The sound box wake-up information includes but is not limited to voice information, gesture information or touch information, and the smart sound box is woken up through the sound box wake-up information. Optionally, when the sound box wake-up information is voice information, the wake-up authority of the user can be verified according to the collected voice information, when the user has the wake-up authority, the intelligent sound box wakes up, otherwise, the wake-up fails. Illustratively, the smart speaker extracts voiceprint features of the voice message, and when the voiceprint features are matched with the voiceprint features stored in the voiceprint feature table, the user is considered to have the authority to wake up the smart speaker, and the smart speaker can be woken up according to the voice message.
And S220, collecting object information of the interactive object.
And S230, sending the object information to a cloud server.
And S240, receiving an analysis result of the cloud server on the object information.
And S250, determining the video stream corresponding to the analysis result.
Illustratively, if the parsing result includes the keyword "music", the information related to the music is formed into a video stream and sent to the device with screen.
And S260, judging whether the analysis result contains visual interaction information or not, if so, executing S270, and otherwise, executing S2110.
And S270, generating and sending a wake-up instruction to the screen equipment.
When the analysis result contains visual interaction information, the connection between the intelligent sound box and the equipment with the screen is required to be established, and the information required by the user is displayed through the equipment with the screen. At this moment, the intelligent sound box generates a wake-up instruction and sends the wake-up instruction to the equipment with the screen, wherein the wake-up instruction comprises an equipment identifier of the equipment with the screen to be woken up and a video stream transmission request, the equipment identifier is used for uniquely identifying the equipment with the screen, and when the equipment identifier is matched with the equipment identifier stored in the equipment with the screen, the equipment with the screen is woken up. The video streaming request is used for requesting the screen device to allow the smart sound box to transmit the video stream. Alternatively, the video stream may be allowed to be transmitted by default when the on-screen device is awakened.
And S280, when a response result of the awakening instruction fed back by the screen equipment is received, sending the video stream to the screen equipment so as to display the video stream through the screen equipment.
When the on-screen device is awakened, a response result of successful awakening can be fed back to the smart speaker, and when the smart speaker receives the response result, the video stream is sent to the on-screen device.
And S290, receiving a video stream updating request sent by the equipment with the screen.
The intelligent sound box can receive video stream updating information of a user and can also receive a video stream updating request sent by the equipment with the screen. Optionally, when the device with the screen is awakened, the capture function is turned on, that is, the remote control operation of the user can be captured, the video stream update request is generated based on the remote control operation, and the video stream update request is sent to the smart sound box. Of course, when the screen device has a voice capture function or an image capture function, the video stream update request may also be generated based on captured voice information or gesture information.
And S2100, updating the video stream according to the video stream updating request and sending the video stream to the screen equipment.
The intelligent sound box updates the video stream of the intelligent sound box according to the video stream updating request, and synchronously sends the video stream to the equipment with the screen, so that the updating requirement of a user is met. The interaction mode provided by the embodiment only needs the user to pay attention to the equipment with the screen, free interaction can be achieved through various modes such as voice, remote control, gestures or touch control, and the interaction range can be expanded to the whole room.
And S2110, keeping the screen projection function in a closed state.
When the analysis result does not contain visual interaction information, the screen projection function of the intelligent sound box keeps in a closed state, and the power consumption of the intelligent sound box is saved.
On the basis of the embodiment, when the analysis result contains visual interaction information, a screen projection function of the intelligent sound box is started, meanwhile, a wake-up instruction is sent to the equipment with the screen to wake up the equipment with the screen, and after the equipment with the screen is woken up, the intelligent sound box sends a video stream to the equipment with the screen to be displayed by the equipment with the screen, so that the interaction requirements of users are met. Compared with the prior art, the intelligent sound box provided by the embodiment of the invention has the advantages that the visual interaction of a user is not influenced even if the screen is not arranged, and particularly, when the screen equipment is large-screen equipment such as TV equipment or a projector, the large-screen characteristic of the screen equipment can be fully utilized, so that the user does not need to be limited in the visual interaction range.
In another embodiment, fig. 3 is a flowchart of an interaction method provided by another embodiment of the present invention, the method is applied to a device with a screen, and referring to fig. 3, the method includes the following steps:
and S310, receiving a wake-up instruction sent by the intelligent sound box.
Optionally, the on-screen device of this embodiment uses a TV device as an example, and before the TV device interacts with the smart speaker, the on-screen device is awakened according to an awakening instruction sent by the smart speaker.
And S320, when the equipment with the screen is awakened, feeding back a response result of the awakening instruction to the intelligent sound box.
S330, receiving and displaying the video stream sent by the intelligent sound box.
The embodiment of the invention provides an interaction method, which is characterized in that a TV device receives a wake-up instruction sent by an intelligent sound box and receives and displays a video stream sent by the intelligent sound box. Compared with the prior art, the screen projection of the embodiment not only transmits the address of the video resource into the player of the TV equipment, but also transmits the video stream formed by the intelligent sound box codes to the TV equipment, and acquires the information of the TV equipment in real time to achieve the synchronous effect, thereby solving the problem that the intelligent sound box without the screen lacks visual interaction, and solving the problem that the visual interaction of the intelligent sound box with the screen is limited by the operation before the intelligent sound box with the screen.
On the basis of the foregoing embodiment, optionally, after receiving and displaying the video stream sent by the smart sound box, the method further includes:
capturing video stream update information of an interactive object;
and analyzing the video stream updating information, generating a video stream updating request based on an analysis result, and sending the video stream updating request to the intelligent sound box.
Specifically, after the TV equipment is awakened, the capture function is started, so that the video stream update information of the user can be captured in real time, and the update requirement of the user is met.
In an embodiment, fig. 4 is a structural diagram of an interaction apparatus provided in an intelligent sound box according to an embodiment of the present invention, where the apparatus is disposed in the intelligent sound box, and may execute the interaction method applied to the intelligent sound box provided in the foregoing embodiment, with reference to fig. 4, the apparatus includes:
an information collecting module 410, configured to collect object information of an interactive object;
the information sending module 420 is configured to send the object information to the cloud server;
a first analysis result receiving module 430, configured to receive an analysis result of the object information by the cloud server;
and the waking module 440 is configured to wake up the on-screen device when the parsing result meets a preset condition, so as to display the video stream corresponding to the object information on the on-screen device.
The embodiment of the invention provides an interaction device, which is characterized in that voice information of an interaction object is collected, object information is sent to a cloud server, a voice analysis result of the cloud server on the voice information is received, and when the voice analysis result meets a preset condition, a screen device is awakened so as to display a video stream corresponding to the voice information on the screen device. Compared with the prior art, the embodiment of the invention fully utilizes the characteristics of the equipment with the screen, displays the video stream on the equipment with the screen, reduces the visual interaction cost of the intelligent sound box and enlarges the interaction range.
On the basis of the foregoing embodiment, the wake-up module 440 is specifically configured to:
determining a video stream corresponding to the analysis result;
when the analysis result contains visual interaction information, generating and sending a wake-up instruction to the screen equipment;
and when a response result of the awakening instruction fed back by the screen equipment is received, sending the video stream to the screen equipment so as to display the video stream through the screen equipment.
On the basis of the above embodiment, the apparatus further includes:
the video stream updating request receiving module is used for receiving a video stream updating request sent by the screen-mounted equipment after the video stream corresponding to the object information is displayed on the screen-mounted equipment;
and the updating module is used for updating the video stream according to the video stream updating request and sending the video stream to the screen equipment.
On the basis of the above embodiment, the apparatus further includes:
the video stream updating information acquisition module is used for acquiring and sending video stream updating information of the interaction object to the cloud server after the video stream corresponding to the object information is displayed on the screen-equipped device;
the second analysis result receiving module is used for receiving the analysis result of the video stream updating information from the cloud server;
and the target video stream determining module is used for determining the target video stream according to the analysis result of the video stream updating information and sending the target video stream to the equipment with the screen.
On the basis of the above embodiment, the apparatus further includes:
and the sound box awakening information receiving module is used for receiving the sound box awakening information of the interactive object before acquiring and sending the object information of the interactive object to the cloud server.
The interaction device provided by the embodiment of the invention can execute the interaction method applied to the intelligent sound box provided by the embodiment of the invention, and has corresponding functional modules and beneficial effects of the execution method.
In another embodiment, fig. 5 is a structural diagram of an interaction apparatus according to another embodiment of the present invention, where the apparatus is disposed in a device with a screen, and is capable of executing the interaction method applied to the device with the screen, and referring to fig. 5, the apparatus includes:
a wake-up instruction receiving module 510, configured to receive a wake-up instruction sent by the smart speaker;
a response result feedback module 520, configured to feed back a response result of the wake-up instruction to the smart speaker when the device with the screen is woken up;
and the display module 530 is configured to receive and display the video stream sent by the smart speaker.
The embodiment of the invention provides an interactive device, which receives a wake-up instruction sent by an intelligent sound box through TV equipment, and receives and displays a video stream sent by the intelligent sound box. Compared with the prior art, the screen projection of the embodiment not only transmits the address of the video resource into the player of the TV equipment, but also transmits the video stream formed by the intelligent sound box codes to the TV equipment, and acquires the information of the TV equipment in real time to achieve the synchronous effect, thereby solving the problem that the intelligent sound box without the screen lacks visual interaction, and solving the problem that the visual interaction of the intelligent sound box with the screen is limited by the operation before the intelligent sound box with the screen.
On the basis of the above embodiment, the apparatus further includes:
the video stream updating information capturing module is used for capturing video stream updating information of the interactive object after receiving and displaying the video stream sent by the intelligent sound box;
and the analysis module is used for analyzing the video stream updating information, generating a video stream updating request based on the analysis result and sending the video stream updating request to the intelligent sound box.
The interaction device provided by the fifth embodiment of the invention can execute the interaction method applied to the equipment with the screen provided by the fifth embodiment of the invention, and has the corresponding functional modules and beneficial effects of the execution method.
In an embodiment, fig. 6 is a block diagram of an interactive system according to an embodiment of the present invention, and referring to fig. 6, the interactive system includes: the system comprises a cloud server 610, an intelligent sound box 620 and a device with a screen 630, wherein the intelligent sound box 620 is in communication connection with the cloud server 610 and the device with the screen 630 respectively;
the smart sound box 620 is used for collecting object information of an interactive object; sending the object information to the cloud server 610; receiving the analysis result of the object information by the cloud server 610; and when the received analysis result sent by the cloud server 610 meets a preset condition, waking up the on-screen device 630;
the cloud server 610 is configured to analyze the object information sent by the smart sound box 620 to obtain an analysis result; the on-screen device 630 is configured to receive a wake-up instruction sent by the smart sound box 620; when the on-screen device 630 wakes up, a response result of the wake-up instruction is fed back to the smart sound box 620; and receiving and displaying the video stream sent by smart sound box 620.
The embodiment does not limit the connection mode between the smart sound box 620 and the cloud server 610, and between the smart sound box 620 and the device with screen 630, for example, the smart sound box 620 and the cloud server 610 may be connected through the ethernet, and the smart sound box 620 and the device with screen 630 may be connected through the lan.
Optionally, smart speaker 620 is specifically configured to:
determining a video stream corresponding to the analysis result;
when the analysis result contains the visual interaction information, a wake-up instruction is generated and sent to the device with screen 630;
when a response result of the wake-up command fed back by the on-screen device 630 is received, the video stream is sent to the on-screen device 630 so as to be displayed by the on-screen device 630.
Optionally, after displaying the video stream corresponding to the object information on the device with screen 630, the smart speaker 620 is further configured to receive a video stream update request sent by the device with screen 630;
the video stream is updated according to the video stream update request and sent to the on-screen device 630.
Optionally, after the video stream corresponding to the object information is displayed on the on-screen device 630, the smart sound box 620 is further configured to acquire and send video stream update information of the interaction object to the cloud server 610;
receiving an analysis result of the video stream update information by the cloud server 610;
and determining a target video stream according to the analysis result of the video stream update information, and sending the target video stream to the screen equipment 630.
Optionally, the smart sound box 620 is further configured to receive sound box wake-up information of the interactive object before collecting and sending object information of the interactive object to the cloud server.
Optionally, the screen device 630 is specifically configured to:
receiving a wake-up instruction sent by the smart sound box 620, and feeding back a response result of the wake-up instruction to the smart sound box 620 when the smart sound box is woken up;
and receiving and displaying the video stream sent by smart sound box 620.
Optionally, the on-screen device 630 is further configured to capture video stream update information of the interactive object after receiving and displaying the video stream sent by the smart sound box 620;
and analyzing the video stream updating information, generating a video stream updating request based on the analysis result, and sending the video stream updating request to the smart sound box 620.
Optionally, the cloud server 610 is further configured to analyze video stream update information of the interactive object sent by the smart sound box 620, and send an analysis result to the smart sound box 620.
The interactive system of the present embodiment and the interactive method provided by the foregoing embodiment belong to the same inventive concept, and specific details may refer to the foregoing embodiment, which is not described herein again.
In an embodiment, the present invention further provides a storage medium having a computer program stored thereon, where the computer program is executed by a corresponding apparatus to implement the interaction method applied to a smart sound box or applied to a device with a screen according to the above embodiments of the present invention, where the interaction method applied to the smart sound box includes:
collecting object information of an interactive object;
sending the object information to a cloud server;
receiving an analysis result of the cloud server on the object information;
and when the analysis result meets a preset condition, awakening the equipment with the screen so as to display the video stream corresponding to the object information on the equipment with the screen.
Optionally, when the analysis result meets a preset condition, waking up a device with a screen to display a video stream corresponding to the object information on the device with the screen, including:
determining a video stream corresponding to the analysis result;
when the analysis result contains visual interaction information, generating and sending a wake-up instruction to the screen equipment;
and when a response result of the awakening instruction fed back by the screen equipment is received, sending the video stream to the screen equipment so as to display the video stream through the screen equipment.
Optionally, after the video stream corresponding to the object information is displayed on the device with screen, the method further includes:
receiving a video stream updating request sent by the equipment with the screen;
and updating the video stream according to the video stream updating request, and sending the video stream to the screen equipment.
Optionally, after the video stream corresponding to the object information is displayed on the device with screen, the method further includes:
collecting and sending video stream updating information of the interactive object to the cloud server;
receiving an analysis result of the video stream updating information by the cloud server;
and determining a target video stream according to the analysis result of the video stream updating information, and sending the target video stream to the screen equipment.
Optionally, before collecting and sending object information of the interaction object to the cloud server, the method further includes:
and receiving the sound box awakening information of the interactive object.
The interaction method applied to the screen equipment comprises the following steps:
receiving a wake-up instruction sent by the intelligent sound box;
when the equipment with the screen is awakened, feeding back a response result of the awakening instruction to the intelligent sound box;
and receiving and displaying the video stream sent by the intelligent sound box.
Optionally, after receiving and displaying the video stream sent by the smart speaker, the method further includes:
capturing video stream update information of an interactive object;
and analyzing the video stream updating information, generating a video stream updating request based on an analysis result, and sending the video stream updating request to the intelligent sound box.
Of course, the storage medium provided by the embodiment of the present invention includes computer-executable instructions, where the computer-executable instructions are not limited to the operations in the interaction method described above, and may also perform related operations in the interaction method provided by any embodiment of the present invention, and have corresponding functions and advantages.
From the above description of the embodiments, it is obvious for those skilled in the art that the present invention can be implemented by software and necessary general hardware, and certainly, can also be implemented by hardware, but the former is a better embodiment in many cases. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which may be stored in a computer-readable storage medium, such as a floppy disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a FLASH Memory (FLASH), a hard disk or an optical disk of a computer, and includes several instructions to enable a computer device (which may be a robot, a personal computer, a server, or a network device) to execute the interaction method according to the foregoing embodiments of the present invention.
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present invention and the technical principles employed. It will be understood by those skilled in the art that the present invention is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the invention. Therefore, although the present invention has been described in greater detail by the above embodiments, the present invention is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present invention, and the scope of the present invention is determined by the scope of the appended claims.

Claims (10)

1. An interaction method applied to a smart sound box, the method comprising:
collecting object information of an interactive object;
sending the object information to a cloud server;
receiving an analysis result of the cloud server on the object information;
and when the analysis result meets a preset condition, awakening the equipment with the screen so as to display the video stream corresponding to the object information on the equipment with the screen.
2. The method according to claim 1, wherein when the parsing result satisfies a preset condition, waking up a device with a screen to display a video stream corresponding to the object information on the device with the screen, includes:
determining a video stream corresponding to the analysis result;
when the analysis result contains visual interaction information, generating and sending a wake-up instruction to the screen equipment;
and when a response result of the awakening instruction fed back by the screen equipment is received, sending the video stream to the screen equipment so as to display the video stream through the screen equipment.
3. The method according to claim 1, further comprising, after presenting the video stream corresponding to the object information on the screen device:
receiving a video stream updating request sent by the equipment with the screen;
and updating the video stream according to the video stream updating request, and sending the video stream to the screen equipment.
4. The method according to claim 1, further comprising, after presenting the video stream corresponding to the object information on the screen device:
collecting and sending video stream updating information of the interactive object to the cloud server;
receiving an analysis result of the video stream updating information by the cloud server;
and determining a target video stream according to the analysis result of the video stream updating information, and sending the target video stream to the screen equipment.
5. The method according to any one of claims 1 to 4, before collecting and sending object information of the interaction object to the cloud server, further comprising:
and receiving the sound box awakening information of the interactive object.
6. An interaction method is applied to a screen device, and comprises the following steps:
receiving a wake-up instruction sent by the intelligent sound box;
when the equipment with the screen is awakened, feeding back a response result of the awakening instruction to the intelligent sound box;
and receiving and displaying the video stream sent by the intelligent sound box.
7. The method of claim 6, further comprising, after receiving and presenting the video stream sent by the smartspeaker:
capturing video stream update information of an interactive object;
and analyzing the video stream updating information, generating a video stream updating request based on an analysis result, and sending the video stream updating request to the intelligent sound box.
8. The utility model provides an interactive device, its characterized in that sets up in intelligent audio amplifier, includes:
the information acquisition module is used for acquiring object information of the interactive object;
the information sending module is used for sending the object information to a cloud server;
the first analysis result receiving module is used for receiving the analysis result of the cloud server on the object information;
and the awakening module is used for awakening the equipment with the screen when the analysis result meets the preset condition so as to display the video stream corresponding to the object information on the equipment with the screen.
9. An interactive system, comprising: the intelligent sound box is respectively in communication connection with the cloud server and the screen equipment;
the intelligent sound box is used for collecting object information of an interactive object; sending the object information to a cloud server; receiving an analysis result of the cloud server on the object information; when the received analysis result sent by the cloud server meets a preset condition, awakening the equipment with the screen;
the cloud server is used for analyzing the object information sent by the intelligent sound box to obtain an analysis result;
the screen equipment is used for receiving a wake-up instruction sent by the intelligent sound box; when the equipment with the screen is awakened, feeding back a response result of the awakening instruction to the intelligent sound box; and receiving and displaying the video stream sent by the intelligent sound box.
10. A storage medium having stored thereon a computer program, characterized in that the program, when executed by a corresponding apparatus, implements the interaction method according to any one of claims 1-7.
CN202010103936.2A 2020-02-20 2020-02-20 Interaction method, device, system and storage medium Pending CN111312248A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010103936.2A CN111312248A (en) 2020-02-20 2020-02-20 Interaction method, device, system and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010103936.2A CN111312248A (en) 2020-02-20 2020-02-20 Interaction method, device, system and storage medium

Publications (1)

Publication Number Publication Date
CN111312248A true CN111312248A (en) 2020-06-19

Family

ID=71147673

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010103936.2A Pending CN111312248A (en) 2020-02-20 2020-02-20 Interaction method, device, system and storage medium

Country Status (1)

Country Link
CN (1) CN111312248A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113823283A (en) * 2021-09-22 2021-12-21 百度在线网络技术(北京)有限公司 Information processing method, apparatus, storage medium, and program product

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105825568A (en) * 2016-03-16 2016-08-03 广东威创视讯科技股份有限公司 Portable intelligent interactive equipment
CN107103263A (en) * 2016-02-23 2017-08-29 芋头科技(杭州)有限公司 A kind of control information exchange method and intelligent robot
US20170303008A1 (en) * 2016-04-19 2017-10-19 Google Inc. Methods, systems and media for interacting with content using a second screen device
CN107608649A (en) * 2017-11-02 2018-01-19 泉州创景视迅数字科技有限公司 A kind of AR augmented realities intelligent image identification displaying content system and application method
CN107835444A (en) * 2017-11-16 2018-03-23 百度在线网络技术(北京)有限公司 Information interacting method, device, voice frequency terminal and computer-readable recording medium
CN108039172A (en) * 2017-12-01 2018-05-15 Tcl通力电子(惠州)有限公司 Smart bluetooth speaker voice interactive method, smart bluetooth speaker and storage medium
CN108320745A (en) * 2018-02-08 2018-07-24 北京小米移动软件有限公司 Control the method and device of display
CN108415561A (en) * 2018-02-11 2018-08-17 北京光年无限科技有限公司 Gesture interaction method based on visual human and system
CN108762512A (en) * 2018-08-17 2018-11-06 浙江核聚智能技术有限公司 Human-computer interaction device, method and system
CN108810595A (en) * 2018-06-29 2018-11-13 四川斐讯信息技术有限公司 A kind of video broadcasting method and system based on speaker
CN108899020A (en) * 2018-05-29 2018-11-27 北京小米移动软件有限公司 Information search method, device and computer readable storage medium
CN109788360A (en) * 2018-12-12 2019-05-21 百度在线网络技术(北京)有限公司 Voice-based TV control method and device

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107103263A (en) * 2016-02-23 2017-08-29 芋头科技(杭州)有限公司 A kind of control information exchange method and intelligent robot
CN105825568A (en) * 2016-03-16 2016-08-03 广东威创视讯科技股份有限公司 Portable intelligent interactive equipment
US20170303008A1 (en) * 2016-04-19 2017-10-19 Google Inc. Methods, systems and media for interacting with content using a second screen device
CN107608649A (en) * 2017-11-02 2018-01-19 泉州创景视迅数字科技有限公司 A kind of AR augmented realities intelligent image identification displaying content system and application method
CN107835444A (en) * 2017-11-16 2018-03-23 百度在线网络技术(北京)有限公司 Information interacting method, device, voice frequency terminal and computer-readable recording medium
CN108039172A (en) * 2017-12-01 2018-05-15 Tcl通力电子(惠州)有限公司 Smart bluetooth speaker voice interactive method, smart bluetooth speaker and storage medium
CN108320745A (en) * 2018-02-08 2018-07-24 北京小米移动软件有限公司 Control the method and device of display
CN108415561A (en) * 2018-02-11 2018-08-17 北京光年无限科技有限公司 Gesture interaction method based on visual human and system
CN108899020A (en) * 2018-05-29 2018-11-27 北京小米移动软件有限公司 Information search method, device and computer readable storage medium
CN108810595A (en) * 2018-06-29 2018-11-13 四川斐讯信息技术有限公司 A kind of video broadcasting method and system based on speaker
CN108762512A (en) * 2018-08-17 2018-11-06 浙江核聚智能技术有限公司 Human-computer interaction device, method and system
CN109788360A (en) * 2018-12-12 2019-05-21 百度在线网络技术(北京)有限公司 Voice-based TV control method and device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113823283A (en) * 2021-09-22 2021-12-21 百度在线网络技术(北京)有限公司 Information processing method, apparatus, storage medium, and program product
CN113823283B (en) * 2021-09-22 2024-03-08 百度在线网络技术(北京)有限公司 Information processing method, apparatus, storage medium, and program product

Similar Documents

Publication Publication Date Title
CN108520746B (en) Method and device for controlling intelligent equipment through voice and storage medium
CN107370649B (en) Household appliance control method, system, control terminal and storage medium
WO2020244266A1 (en) Remote control method for smart television, mobile terminal, and smart television
CN105634881B (en) Application scene recommendation method and device
WO2017024711A1 (en) Intelligent household appliance control method, device, system and apparatus
US20170060599A1 (en) Method and apparatus for awakening electronic device
CN111050199B (en) Display device and scheduling method of Bluetooth communication resources of display device
CN111277884B (en) Video playing method and device
CN103634683A (en) Screen capturing method and device for intelligent televisions
CN106569758B (en) Wireless screen projection method and device
CN111050198A (en) Display device and switching method of Bluetooth audio device
CN111314372A (en) Display equipment awakening method, control terminal, server and display equipment
CN106856433B (en) Equipment awakening method, device and system
WO2022105409A1 (en) Fault diagnosis method, terminal device, and display device
CN112153440B (en) Display equipment and display system
CN109032554B (en) Audio processing method and electronic equipment
CN103269445B (en) Intelligent television system control method
WO2024103926A1 (en) Voice control methods and apparatuses, storage medium, and electronic device
CN114067798A (en) Server, intelligent equipment and intelligent voice control method
US11553019B2 (en) Method, apparatus, electronic device and storage medium for acquiring programs in live streaming room
CN111312248A (en) Interaction method, device, system and storage medium
CN109408164B (en) Method, device and equipment for controlling screen display content and readable storage medium
CN111263223A (en) Media volume adjusting method and display device
CN113325980B (en) Control method, control device, electronic equipment and readable storage medium
CN112835506B (en) Display device and control method thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200619