CN114120998A - Interactive content presentation method and device - Google Patents

Interactive content presentation method and device Download PDF

Info

Publication number
CN114120998A
CN114120998A CN202111273084.2A CN202111273084A CN114120998A CN 114120998 A CN114120998 A CN 114120998A CN 202111273084 A CN202111273084 A CN 202111273084A CN 114120998 A CN114120998 A CN 114120998A
Authority
CN
China
Prior art keywords
operation instruction
audio resource
content
resource content
control information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111273084.2A
Other languages
Chinese (zh)
Inventor
陈皓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qingdao Haier Technology Co Ltd
Haier Smart Home Co Ltd
Original Assignee
Qingdao Haier Technology Co Ltd
Haier Smart Home Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qingdao Haier Technology Co Ltd, Haier Smart Home Co Ltd filed Critical Qingdao Haier Technology Co Ltd
Priority to CN202111273084.2A priority Critical patent/CN114120998A/en
Publication of CN114120998A publication Critical patent/CN114120998A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The invention provides an interactive content presentation method and an interactive content presentation device, wherein the interactive content presentation method is applied to voice interactive equipment, and comprises the following steps: acquiring a user operation instruction, wherein the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server; responding to a user operation instruction, and performing play control on the audio resource content; and sending play control information corresponding to the audio resource content to the pre-bound terminal equipment, wherein the terminal equipment is provided with a screen for displaying the play control information in real time. The invention provides an interactive content presentation method and device, which are applied to the interactive process of voice interactive equipment and a user, and when the voice interactive equipment plays audio resource content, the corresponding play control information can be sent to terminal equipment for the terminal equipment to display the play control information in real time, so that the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.

Description

Interactive content presentation method and device
Technical Field
The invention relates to the technical field of Internet of things, in particular to an interactive content presentation method and device.
Background
With the development of the internet of things technology, various household appliances such as air conditioners, refrigerators, water dispensers and the like can be networked, the household appliances can have a voice interaction function, besides the household appliances, the household appliances can also be multimedia playing equipment such as a sound box, and the equipment with the voice interaction function can be called voice interaction equipment.
When existing voice interaction equipment, such as an intelligent sound box, plays a voice resource, a user cannot see the playing progress and content introduction of the currently played voice resource content, and is difficult to perform operations such as resource switching and start-stop control, so that the played content is not intuitively grasped, and the control operation is inconvenient.
Disclosure of Invention
The invention provides an interactive content presentation method and device, which are used for solving the defects that in the prior art, when a sound resource is played, a user cannot see the playing progress and content introduction of the content of the currently played sound resource, the operations such as resource switching, start-stop control and the like are difficult to perform, the played content is not intuitively mastered, the control operation is inconvenient, the user can intuitively master the played content conveniently, and the simplicity and convenience of the control operation are improved.
The invention provides an interactive content presentation method, which is applied to voice interactive equipment and comprises the following steps: acquiring a user operation instruction, wherein the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server; responding to the user operation instruction, and performing play control on the audio resource content; and sending play control information corresponding to the audio resource content to a pre-bound terminal device, wherein the terminal device is provided with a screen for displaying the play control information in real time.
According to the interactive content presentation method provided by the present invention, the sending of the play control information corresponding to the audio resource content to the pre-bound terminal device includes: calling a message middleware through the server side, and uploading playing control information corresponding to the audio resource content to the message middleware in real time; and sending the play control information to the terminal equipment through the message middleware.
According to the interactive content presentation method provided by the present invention, the sending the play control information to the terminal device through the message middleware comprises: and sending the play control information to the terminal equipment based on the websocket long link through the message middleware.
According to the interactive content presentation method provided by the invention, the playing control of the audio resource content in response to the user operation instruction comprises the following steps: responding to the user operation instruction, and sending a resource calling request to a server side, wherein the resource calling request is used for indicating the server side to search the audio resource content; and receiving the audio resource content sent by the server side, and performing play control on the audio resource content.
According to the interactive content presentation method provided by the present invention, after sending the play control information corresponding to the audio resource content to the pre-bound terminal device, the method further includes: receiving a target operation instruction sent by the terminal equipment; and responding to the target operation instruction, and adjusting the content of the audio resource.
According to the interactive content presentation method provided by the invention, the adjusting the audio resource content in response to the target operation instruction comprises the following steps: under the condition that the target operation instruction is start-stop control or progress adjustment, starting or stopping playing the audio resource content or adjusting the playing progress of the audio resource content based on the target operation instruction; or, under the condition that the target operation instruction is to switch the audio resource content, switching to the updated resource content through a resource list of the server side based on the target operation instruction, and playing the updated resource content.
According to the interactive content presentation method provided by the invention, the responding to the user operation instruction and controlling the playing of the audio resource content comprises the following steps: responding to a request instruction sent by the terminal equipment, and determining the audio resource content from the server; and performing play control on the audio resource content.
The invention also provides an interactive content presentation device, which is applied to the voice interaction equipment, and the interactive content presentation device comprises: the system comprises an acquisition module, a processing module and a processing module, wherein the acquisition module is used for acquiring a user operation instruction, and the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server; the playing module is used for responding to the user operation instruction and performing playing control on the audio resource content; and the sending module is used for sending the playing control information corresponding to the audio resource content to a pre-bound terminal device, and the terminal device is provided with a screen for displaying the playing control information in real time.
The invention also provides an electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, wherein the processor implements the steps of the interactive content presentation method as described in any one of the above when executing the program.
The invention also provides a non-transitory computer readable storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of the interactive content presentation method as described in any one of the above.
The invention provides an interactive content presentation method and device, which are applied to the interactive process of voice interactive equipment and a user, and when the voice interactive equipment plays audio resource content, the corresponding play control information can be sent to terminal equipment for the terminal equipment to display the play control information in real time, so that the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.
Drawings
In order to more clearly illustrate the technical solutions of the present invention or the prior art, the drawings needed for the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and those skilled in the art can also obtain other drawings according to the drawings without creative efforts.
FIG. 1 is a flow chart diagram of an interactive content presentation method provided by the present invention;
FIG. 2 is a schematic structural diagram of an interactive content presentation apparatus provided by the present invention;
fig. 3 is a schematic structural diagram of an electronic device provided in the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention clearer, the technical solutions of the present invention will be clearly and completely described below with reference to the accompanying drawings, and it is obvious that the described embodiments are some, but not all embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The interactive content presenting method and apparatus of the present invention are described below with reference to fig. 1 to 3.
The invention provides an interactive content presentation method which is applied to voice interactive equipment.
The voice interaction equipment can be a sound box, household appliances such as an intelligent refrigerator or an intelligent air conditioner, the voice interaction equipment can be connected to the Internet, the voice interaction equipment can be provided with a sound pick-up, a loudspeaker or a processor, the sound pick-up and the loudspeaker can be electrically connected with the processor, the sound pick-up can collect voice signals of a user, the loudspeaker can play sound resource contents, for example, music can be played, the voice interaction equipment can be bound with terminal equipment, the terminal equipment can be a mobile phone, a mobile terminal such as a tablet computer or a notebook computer, and also can be fixed equipment such as a desktop computer or an intelligent television, the terminal equipment is provided with a display screen, and related contents can be presented to the user through the display screen.
As shown in fig. 1, the interactive contents presentation method includes the following steps 110 to 130.
Step 110, obtaining a user operation instruction, where the user operation instruction includes at least one of a voice control instruction of a user, a target operation instruction sent by a terminal device, and an on-demand instruction from a server;
it can be understood that the voice interaction device can acquire a voice control instruction of a user through the sound pickup, the voice control instruction can be corresponding to playing a certain type of song, for example, the voice control instruction can be 'playing a popular song in cantonese', the voice interaction device can be placed in a target area, the user can talk in the target area according to a certain rule, the voice interaction device can be triggered, and the voice interaction device can acquire the voice control instruction.
In a specific example, the voice interaction device may be a smart speaker, and the user may speak into the smart speaker to give a voice control instruction, the speaker may have a plurality of microphones, for example, the speaker may have a full-directional microphone and a single-directional microphone, and the speaker may collect the voice control instruction of the user through the microphones.
Of course, the target operation instruction may also be sent to the voice interaction device through the terminal device, or the on-demand instruction may be sent to the voice interaction device through the server, and the voice interaction device may perform a corresponding response according to the target operation instruction or the on-demand instruction.
And 120, responding to the user operation instruction, and performing play control on the audio resource content.
It can be understood that, after receiving the voice control instruction, the voice interaction device may perform voice recognition on the voice control instruction, and recognize a corresponding keyword, and may search for corresponding audio resource content in a local storage by using the keyword, or may search for corresponding audio resource content in a server end in communication connection with the voice interaction device, where the audio resource content may be a multimedia file with a sound signal, such as a song, a phase sound, a drama, a movie, or a tv series.
The voice interaction device may play the audio resource content through the speaker, and certainly, the voice interaction device may also play a video in the audio resource content, for example, the video may be displayed through a display screen, and when the audio resource content is played, the voice interaction device itself may not display the playing progress of the audio resource content.
For example, when the voice interaction device is an intelligent sound box, the intelligent sound box may extract a keyword according to the voice control instruction, search the corresponding audio resource content from the server by using the keyword, and play the audio resource content through a speaker.
Step 130, sending playing control information corresponding to the content of the audio resource to the pre-bound terminal device, where the terminal device is configured with a screen for displaying the playing control information in real time.
It can be understood that the terminal device may be bound to the voice interaction device in advance, for example, the terminal device may scan a two-dimensional code of the voice interaction device, so as to identify a device number of the voice interaction device, and thus, the device number is utilized to establish binding with the voice interaction device, so that the voice interaction device and the terminal device may establish stable communication connection through the internet, and the voice interaction device and the terminal device may be in wireless communication connection, for example, the voice interaction device and the terminal device may realize stable communication connection through WIFI.
When the voice interaction device plays the audio resource content, the playing control information corresponding to the audio resource content is generated, the playing control information may include detailed image-text introduction of the audio resource content and a corresponding playing progress, and may also be other content related to the audio resource content, and the playing control information may include playing detailed content.
Here, the sending of the playing detail content to the terminal device may be directly sending the playing detail content to the terminal device by the voice interaction device, or sending the playing detail content to the terminal device by the voice interaction device through the server as an intermediary, which is not specifically limited herein as long as the terminal device can be ensured to receive the playing detail content.
For example, when the voice interaction device is a sound box, the sound box can recognize the playing progress of the audio resource content, and also can obtain the detailed image-text introduction of the audio resource content from the server, the sound box can feed back the playing progress to the server in real time and cache the playing progress in the database integrated by the server, the server can send the playing progress and the detailed image-text introduction to the terminal device, for example, the terminal device here can be a mobile phone, the mobile phone can display the playing progress on the display screen of the mobile phone when receiving the playing progress of the audio resource content, the playing progress can be displayed in a progress bar manner, the playing progress can correspond to the indication mark on the progress bar, the indication mark can change in real time, thereby dynamically and visually showing the playing progress of the audio resource content, and the detailed image-text introduction can be presented in a software page manner, when the vocal resource content is a song, the detailed graphic introduction can comprise lyrics, along with the adjustment of the playing progress of the song, the playing progress position of the corresponding song in the lyrics can be highlighted, and the lyrics can also dynamically roll to adapt to the adjustment of the playing progress of the song.
The invention provides an interactive content presentation method, which is applied to the interactive process of voice interactive equipment and a user, and when the voice interactive equipment plays audio resource content, the corresponding play control information can be sent to the terminal equipment for the terminal equipment to display the play control information in real time, so that the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.
In some embodiments, the step 120 of performing playback control on the audio resource content in response to the user operation instruction includes: responding to a user operation instruction, sending a resource calling request to a server side, so that the server side can search the audio resource content based on the resource calling request; and receiving the audio resource content sent by the server side, and playing and controlling the audio resource content.
It can be understood that the voice interaction device may be in communication connection with a server side, the server side may be integrated with a database, and mass vocal resource contents may be stored in the database, where the voice interaction device may send a resource calling request to the server side when receiving a voice control instruction, and the server side may search for corresponding vocal resource contents in the database and return the vocal resource contents to the voice interaction device when receiving the resource calling request, where the voice interaction device may play the vocal resource contents.
The server side can be integrated with a Redis database, wherein the Redis database is a high-performance key-value database. The method supports the persistence of data, supports various data structures and supports the backup of data. The database pressure can be reduced, the overall performance is improved, and better user experience is brought.
For example, when the voice interaction device is a smart speaker, the voice control instruction may be to play a target song, when the smart speaker is triggered, the resource invocation request may be sent to the server, and when receiving the resource invocation request, the server may search for the target song in the database and send the target song to the smart speaker, and the smart speaker may automatically play the target song.
In some embodiments, the step 130 of sending the play control information corresponding to the audio resource content to the pre-bound terminal device includes: calling a message middleware through a server side, and uploading playing control information corresponding to the content of the sound resource to the message middleware in real time; and sending the playing control information to the terminal equipment through the message middleware.
It can be understood that the message middleware is a message forwarding device, and may be communicatively connected to the server side, or may be integrated at the server side, the message middleware may perform functions of data caching and data forwarding, and data to be forwarded may form a queue in the message middleware and be sent in sequence. The voice interaction equipment can be in real-time communication with the server side when playing the sound resource content, the message middleware can be called through the server side, the playing control information corresponding to the sound resource content can be uploaded to the message middleware in real time, the message forwarding speed of the message middleware can be increased, the real-time performance of sending the playing control information can be realized, and the playing progress of the voice interaction equipment and the terminal equipment can be kept synchronous.
The message middleware may be a rockmq, which is a piece of low-latency, highly reliable, scalable, and easy-to-use message middleware.
In some embodiments, sending the play control information to the terminal device through the message middleware includes: and sending the play control information to the terminal equipment based on the websocket long link through the message middleware.
It can be appreciated that the websocket is a protocol for full duplex communication over a single TCP connection, making data exchange between the client and the server simpler, allowing the server to actively push data to the client. In the WebSocket API, the browser and the server only need to complete one matching process, and persistent connection can be directly established between the browser and the server, and bidirectional data transmission is carried out.
The message middleware can establish a communication connection relation of the websocket with the terminal equipment, and the message middleware can send the play control information to the terminal equipment based on the websocket long link, so that stable transmission of the play control information is realized.
In some embodiments, after sending the play control information corresponding to the audio resource content to the pre-bound terminal device, the method further includes: receiving a target operation instruction sent by terminal equipment; and responding to the target operation instruction, and adjusting the content of the audio resource.
It can be understood that, after receiving the play control information, the terminal device may display the play control information on the terminal device, so as to achieve visual presentation of the play control information, and after seeing the play control information, the user may perform an operation on a display interface of the mobile terminal, for example, may give a target operation instruction, where the target operation instruction is a control command given corresponding to the play control information, and the terminal device may send the target operation instruction to the voice interaction device, and the voice interaction device may adjust the content of the sound resource according to the target operation instruction.
That is to say, the voice interaction device can not only respond to the voice control instruction given by the user through voice, but also respond to the target operation instruction given by the user through the terminal device, and the voice interaction device can perform voice interaction with the user and also perform communication interaction with the terminal device.
After the user sees the playing progress bar corresponding to the audio resource content on the display interface of the terminal device, the user can automatically adjust the playing progress of the audio resource content, can also switch to play other audio resource contents, and can also give other types of control commands through the terminal device, which is not specifically limited herein.
In some embodiments, adjusting the audio resource content in response to the target operation instruction includes: and under the condition that the target operation instruction is start-stop control or progress adjustment, starting or pausing playing of the audio resource content or adjusting the playing progress of the audio resource content based on the target operation instruction.
It can be understood that, when the user sees the playing progress of the audio resource content on the display screen of the terminal device, the playing state of the audio resource content may be adjusted, for example, the audio resource content may be controlled to start playing or pause playing, where, when the target operation instruction is start-stop control or progress control of the audio resource content, the audio resource content may be started or paused according to the target operation instruction.
In some embodiments, adjusting the audio resource content in response to the target operation instruction includes: and under the condition that the target operation instruction is to switch the audio resource content, switching the resource content to be updated through the resource list of the server side based on the target operation instruction, and playing the updated resource content.
It can be understood that the target operation instruction may also be to switch the audio resource content, that is, to switch the audio resource content to the updated resource content, which is similar to a "song switching" function in a music player, and when the target operation instruction corresponding to the audio resource content is received, the target operation instruction may be responded, and a resource switching request may be sent to the server, so that the server searches for the updated resource content in the resource list, and the server returns the updated resource content to the voice interaction device, and the voice interaction device may play the updated resource content.
In an actual scenario, when a user interacts with a voice interaction device, for example, listening to music or broadcasting through the voice interaction device, a dynamic play control bar may be displayed on the voice interaction device, where the dynamic play control bar may include basic song information, a play progress, and the like. After the play control bar is presented, the user can switch the content of the audio resource on the play control bar, and can give instruction control such as the previous or next, and play, pause or all play.
In some embodiments, the playing control of the audio resource content in response to the user operation instruction includes: responding to a request order sent by the terminal equipment, and determining the audio resource content from the server; and performing play control on the content of the audio resource. It can be understood that a user can directly give a request instruction in the terminal device, that is, select the audio resource content to be played in the terminal device, the terminal device can send the request instruction to the voice interaction device, and after receiving the request instruction, the voice interaction device can respond to the request instruction, and can determine the corresponding audio resource content, so as to play the audio resource content.
It should be noted that the terminal may search the audio resource content corresponding to the on-demand instruction from the local storage, or may search the audio resource content corresponding to the on-demand instruction from the server side, which is not specifically limited herein.
That is to say, the user can request the audio resource content to the voice interaction device through voice, and can also request the audio resource content through the terminal device, so that the selection of the user can be expanded, the interaction requirements of the user can be met in a diversified manner, and the interaction experience of the user is further improved.
It should be noted that, in the process of performing speech recognition on the speech control instruction, the speech control instructions of multiple users may be collected in advance, and then the multiple speech control instructions are split according to their actual meanings, that is, the speech control instructions are classified into one class with the same meaning.
It can be understood that, because the sentences used by each user to express the same meaning are different, some users are used to perform voice input according to a normal language sequence, some users are used to invert sentences, and for example, some users only speak keywords during voice input, and some users are used to input a plurality of vocabularies which are irrelevant to actual meanings, such as "en," and "o", but the actual meanings expressed by the users are consistent in reality. Then these actually equivalent terms are taken as a classification.
After the voice control commands are classified, the voice control commands of each type are converted into segmented text information according to the time sequence of the input voice control commands. If one end voice control instruction of one category is 'i want to listen to music', the voice control instruction is converted into segmented text information which is five texts, and the segmented text information is arranged according to the sequence of 'i want to listen to music'.
When the voice control instruction is converted, the converted segmented text is a single byte, for example, i want to listen to music is converted into a text message composed of five texts i want to listen to music, and the text message is arranged according to the sequence when the voice control instruction is input by the user.
Machine learning training can be carried out on the same type of segmented text information converted according to the time sequence to obtain a text recognition model capable of extracting keywords from the segmented text information and forming a keyword sequence according to the time sequence;
specifically, after the voice control instruction is converted into a segmented text, a plurality of segments of the same type are taken as input, and a keyword corresponding to the actual meaning of the voice control instruction is taken as a result for training, so that a text recognition model capable of outputting relevant detection according to the input segmented text information is formed. For example, when "i", "want", "listen", "sound" and "music" are input, the actual meaning of the voice control instruction is to open the music player, so the actual corresponding keywords are "listen", "sound" and "music", and the two keywords "i" and "want" are interfering words, which can be eliminated during output. And after extracting the keywords from the text recognition model, sequencing the keywords according to the time sequence to form a keyword sequence.
It can be understood that machine learning is a one-domain multi-domain cross discipline, and relates to multiple disciplines such as probability theory, statistics, approximation wheel, convex analysis, algorithm complexity theory and the like. The special research on how a computer simulates or realizes the learning behavior of human beings so as to acquire new knowledge or skills and reorganize the existing knowledge structure to continuously improve the performance of the computer. According to the similarity of the functions and forms of the algorithms, the algorithms can be classified, such as a regression algorithm, an example-based algorithm, a regularization algorithm, decision tree learning, a Bayesian method, a kernel-based algorithm, a clustering algorithm, association learning, an artificial neural network, a latitude reduction algorithm, an integration algorithm and the like.
It can be understood that after the text recognition model extracts the keywords, the keywords are directly sorted according to the time sequence when the voice control instruction is input.
The keyword sequence may be identified by a semantic processing module to execute commands of the voice control instruction.
Specifically, after the keyword sequence is obtained through the text recognition model, the keyword sequence is placed into the semantic processing module, and the text information of the keyword sequence is processed by the semantic processing module, so that the keyword sequence is converted into a command to be executed.
It can be understood that since the voice control command is selected by the keywords of the text recognition model, useless interference words are removed when the voice control command is processed by the semantic module, so that the voice control command can be converted into an accurate command to be executed no matter what habit the user inputs the voice command, and the experience of the user is improved.
It is understood that the semantic processing module employs any of the LSTM, CNN or transform models.
The interactive content presentation method can be constructed based on Spring boots, and the Spring boots are designed to simplify the initial construction and development process of new Spring applications. The framework uses a specific way to configure, thereby eliminating the need for developers to define a templated configuration. The method can also be constructed based on MyBatis-Plus (MP for short), the MP is an enhancement tool of MyBatis and is a developed MyBatis enhancement tool, only enhancement is carried out on the basis of MyBatis without change, and the method is used for simplifying development and improving efficiency and has the characteristics of no invasion, small loss, strong CRUD operation, support of XML hot loading, built-in code generator and the like.
It is worth mentioning that, currently, when a user interacts with a voice interaction device, for example, a song is played, the user cannot view a currently played resource on a mobile phone app, and when the user interacts with the voice interaction device, the user only supports operations such as voice switching and pausing at present, which is similar to that of a mobile phone music client, and the intuitive operation on the mobile phone app is not supported. Currently, users are not supported to check the playing progress of resources in real time.
In order to improve the user experience of the scene and make the interaction more humanized, the embodiment can realize that the user can check the played audio resource content in real time on the app of the terminal device when playing the audio resource. The user can autonomously switch resources and perform operations such as play control on the app of the terminal device. And based on the instantaneity of the message, middleware such as redis, websocket and the like are introduced for addition, so that the state consistency of the voice interaction equipment and the terminal equipment is ensured. The access efficiency is higher, the use is more convenient, and the user experience is better.
The following describes the interactive content presentation apparatus provided by the present invention, and the interactive content presentation apparatus described below and the interactive content presentation method described above may be referred to correspondingly.
As shown in fig. 2, the present invention provides an interactive content presenting apparatus, applied to a voice interaction device, the interactive content presenting apparatus comprising: an acquisition module 210, a playing module 220 and a sending module 230.
The obtaining module 210 is configured to obtain a user operation instruction, where the user operation instruction includes at least one of a voice control instruction of a user, a target operation instruction sent by a terminal device, and an on-demand instruction from a server.
And the playing module 220 is configured to perform playing control on the audio resource content in response to the user operation instruction.
A sending module 230, configured to send the play control information corresponding to the audio resource content to a pre-bound terminal device, where the terminal device is configured with a screen for displaying the play control information in real time.
The invention provides an interactive content presentation device, which is applied to the interactive process of voice interactive equipment and a user, and can send corresponding play control information to terminal equipment when the voice interactive equipment plays audio resource content, so that the terminal equipment can display the play control information in real time, the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.
In some embodiments, the playing module is further configured to invoke a message middleware through the server, and upload, to the message middleware, playing control information corresponding to the content of the audio resource in real time; and sending the play control information to the terminal equipment through the message middleware.
In some embodiments, the play module is further configured to send, through the message middleware, the play control information to the terminal device based on the websocket long link.
In some embodiments, the playing module is further configured to send a resource invocation request to the server end in response to the user operation instruction, where the resource invocation request is used to instruct the server end to search for the audio resource content; and receiving the audio resource content sent by the server side, and performing play control on the audio resource content. In some embodiments, the interactive content presentation device further comprises an adjustment module.
The adjustment module is used for: receiving a target operation instruction sent by the terminal equipment; and responding to the target operation instruction, and adjusting the content of the audio resource.
In some embodiments, the adjustment module is further configured to: under the condition that the target operation instruction is start-stop control or progress adjustment, starting or stopping playing the audio resource content or adjusting the playing progress of the audio resource content based on the target operation instruction; or, under the condition that the target operation instruction is to switch the audio resource content, switching to the updated resource content through a resource list of the server side based on the target operation instruction, and playing the updated resource content.
In some embodiments, the interactive content presentation method further comprises: and determining a module.
A determining module, configured to determine, from the server, the audio resource content in response to an on-demand instruction sent by the terminal device; and performing play control on the audio resource content.
Fig. 3 illustrates a physical structure diagram of an electronic device, which may include, as shown in fig. 3: a processor (processor)310, a communication Interface (communication Interface)320, a memory (memory)330 and a communication bus 340, wherein the processor 310, the communication Interface 320 and the memory 330 communicate with each other via the communication bus 340. The processor 310 may invoke logic instructions in the memory 330 to perform an interactive content presentation method for application to a voice interaction device, the interactive content presentation method comprising: acquiring a user operation instruction, wherein the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server; responding to the user operation instruction, and performing play control on the audio resource content; and sending play control information corresponding to the audio resource content to a pre-bound terminal device, wherein the terminal device is provided with a screen for displaying the play control information in real time.
The invention provides an interactive content presentation method, which is applied to the interactive process of voice interactive equipment and a user, and when the voice interactive equipment plays audio resource content, the corresponding play control information can be sent to the terminal equipment for the terminal equipment to display the play control information in real time, so that the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.
In addition, the logic instructions in the memory 330 may be implemented in the form of software functional units and stored in a computer readable storage medium when the software functional units are sold or used as independent products. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
In another aspect, the present invention further provides a computer program product, where the computer program product includes a computer program, the computer program can be stored on a non-transitory computer-readable storage medium, and when the computer program is executed by a processor, a computer can execute the interactive content presentation method provided by the above methods, and the method is applied to a voice interaction device, and the interactive content presentation method includes: acquiring a user operation instruction, wherein the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server; responding to the user operation instruction, and performing play control on the audio resource content; and sending play control information corresponding to the audio resource content to a pre-bound terminal device, wherein the terminal device is provided with a screen for displaying the play control information in real time.
The invention provides an interactive content presentation method, which is applied to the interactive process of voice interactive equipment and a user, and when the voice interactive equipment plays audio resource content, the corresponding play control information can be sent to the terminal equipment for the terminal equipment to display the play control information in real time, so that the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.
In still another aspect, the present invention also provides a non-transitory computer-readable storage medium, on which a computer program is stored, the computer program being implemented by a processor to perform the interactive content presentation method provided by the above methods, and the interactive content presentation method is applied to a voice interaction device, and the interactive content presentation method includes: acquiring a user operation instruction, wherein the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server; responding to the user operation instruction, and performing play control on the audio resource content; and sending play control information corresponding to the audio resource content to a pre-bound terminal device, wherein the terminal device is provided with a screen for displaying the play control information in real time. The invention provides an interactive content presentation method, which is applied to the interactive process of voice interactive equipment and a user, and when the voice interactive equipment plays audio resource content, the corresponding play control information can be sent to the terminal equipment for the terminal equipment to display the play control information in real time, so that the user can conveniently and intuitively master the play content, and the simplicity and convenience of control operation are improved.
The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment. One of ordinary skill in the art can understand and implement it without inventive effort.
Through the above description of the embodiments, those skilled in the art will clearly understand that each embodiment can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware. With this understanding in mind, the above-described technical solutions may be embodied in the form of a software product, which can be stored in a computer-readable storage medium such as ROM/RAM, magnetic disk, optical disk, etc., and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the methods described in the embodiments or some parts of the embodiments.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (10)

1. An interactive content presentation method, applied to a voice interaction device, includes:
acquiring a user operation instruction, wherein the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server;
responding to the user operation instruction, and performing play control on the audio resource content;
and sending play control information corresponding to the audio resource content to a pre-bound terminal device, wherein the terminal device is provided with a screen for displaying the play control information in real time.
2. The interactive content presentation method according to claim 1, wherein the sending of the play control information corresponding to the audio resource content to the pre-bound terminal device comprises:
calling a message middleware through the server side, and uploading playing control information corresponding to the audio resource content to the message middleware in real time;
and sending the play control information to the terminal equipment through the message middleware.
3. The interactive content presentation method according to claim 2, wherein the sending the play control information to the terminal device through the message middleware comprises:
and sending the play control information to the terminal equipment based on the websocket long link through the message middleware.
4. The interactive content presentation method according to claim 1, wherein the controlling playback of the audio resource content in response to the user operation instruction comprises:
responding to the user operation instruction, and sending a resource calling request to a server side, wherein the resource calling request is used for indicating the server side to search the audio resource content;
and receiving the audio resource content sent by the server side, and performing play control on the audio resource content.
5. The interactive content presentation method according to any one of claims 1 to 4, further comprising, after the sending the playback control information corresponding to the audio resource content to the pre-bound terminal device:
receiving a target operation instruction sent by the terminal equipment;
and responding to the target operation instruction, and adjusting the content of the audio resource.
6. The interactive content presentation method according to claim 5, wherein the adjusting the audio resource content in response to the target operation instruction comprises:
under the condition that the target operation instruction is start-stop control or progress adjustment, starting or stopping playing the audio resource content or adjusting the playing progress of the audio resource content based on the target operation instruction;
or, under the condition that the target operation instruction is to switch the audio resource content, switching to the updated resource content through a resource list of the server side based on the target operation instruction, and playing the updated resource content.
7. The interactive content presentation method according to any one of claims 1 to 4, wherein the controlling playback of the audio resource content in response to the user operation instruction comprises:
responding to a request instruction sent by the terminal equipment, and determining the audio resource content from the server;
and performing play control on the audio resource content.
8. An interactive content presentation apparatus applied to a voice interaction device, the interactive content presentation apparatus comprising:
the system comprises an acquisition module, a processing module and a processing module, wherein the acquisition module is used for acquiring a user operation instruction, and the user operation instruction comprises at least one of a voice control instruction of a user, a target operation instruction sent by terminal equipment and an on-demand instruction from a server;
the playing module is used for responding to the user operation instruction and performing playing control on the audio resource content;
and the sending module is used for sending the playing control information corresponding to the audio resource content to a pre-bound terminal device, and the terminal device is provided with a screen for displaying the playing control information in real time.
9. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the steps of the interactive content presentation method according to any one of claims 1 to 7 are implemented when the processor executes the program.
10. A non-transitory computer readable storage medium, having stored thereon a computer program, wherein the computer program, when executed by a processor, implements the steps of the interactive content presentation method according to any one of claims 1 to 7.
CN202111273084.2A 2021-10-29 2021-10-29 Interactive content presentation method and device Pending CN114120998A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111273084.2A CN114120998A (en) 2021-10-29 2021-10-29 Interactive content presentation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111273084.2A CN114120998A (en) 2021-10-29 2021-10-29 Interactive content presentation method and device

Publications (1)

Publication Number Publication Date
CN114120998A true CN114120998A (en) 2022-03-01

Family

ID=80379529

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111273084.2A Pending CN114120998A (en) 2021-10-29 2021-10-29 Interactive content presentation method and device

Country Status (1)

Country Link
CN (1) CN114120998A (en)

Similar Documents

Publication Publication Date Title
WO2022121601A1 (en) Live streaming interaction method and apparatus, and device and medium
US11853370B2 (en) Scene aware searching
WO2019228302A1 (en) Live broadcast room display method, apparatus and device, and storage medium
CN106098063B (en) Voice control method, terminal device and server
CN109147784B (en) Voice interaction method, device and storage medium
CN110519636B (en) Voice information playing method and device, computer equipment and storage medium
CN111683263B (en) Live broadcast guiding method, device, equipment and computer readable storage medium
CN112040263A (en) Video processing method, video playing method, video processing device, video playing device, storage medium and equipment
CN106796496A (en) Display device and its operating method
CN111372109B (en) Intelligent television and information interaction method
CN107463700B (en) Method, device and equipment for acquiring information
CN107040452B (en) Information processing method and device and computer readable storage medium
CN109036374B (en) Data processing method and device
CN109474843A (en) The method of speech control terminal, client, server
CN114025186A (en) Virtual voice interaction method and device in live broadcast room and computer equipment
CN112463106A (en) Voice interaction method, device and equipment based on intelligent screen and storage medium
WO2019047850A1 (en) Identifier displaying method and device, request responding method and device
CN109710799B (en) Voice interaction method, medium, device and computing equipment
US11568871B1 (en) Interactive media system using audio inputs
CN113438492B (en) Method, system, computer device and storage medium for generating title in live broadcast
CN110337041A (en) Video broadcasting method, device, computer equipment and storage medium
CN110309470A (en) A kind of virtual news main broadcaster system and its implementation based on air imaging
WO2022046517A1 (en) Systems and methods for reducing latency in cloud services
CN114120998A (en) Interactive content presentation method and device
CN114449301B (en) Item sending method, item sending device, electronic equipment and computer-readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination